id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
241895779
Looping in a state when there's no explicit jump No more of this nonsense 10: [...] => 10 Agreed https://github.com/acbabis/factorio-state-machine/pull/11
gharchive/issue
2017-07-11T01:12:14
2025-04-01T04:55:50.129946
{ "authors": [ "acbabis", "jheinemeyer" ], "repo": "acbabis/factorio-state-machine", "url": "https://github.com/acbabis/factorio-state-machine/issues/10", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1732793085
1721 automate the push from GitHub to gitlab Description This pull request adds a GitHub Actions workflows to mirror the dev, staging & production branch and tags to GitLab. Changes Made Created a new GitHub Actions workflow file: .github/workflows/mirror-dev.yml. Created a new GitHub Actions workflow file: .github/workflows/mirror-staging.yml. Created a new GitHub Actions workflow file: .github/workflows/mirror-production.yml. Added a workflow to mirror the dev, staging & production branch and tags to GitLab. Restricted the workflow to run only on the accessibility-exchange/platform repository. Testing Tested the workflow on a fork to ensure it successfully mirrors the branches to GitLab without limit for repository. After adding limit for repository made sure it doesn't work on fork. Verified that tags are mirrored as well. Checklist [x] The workflow file is located in the correct directory: .github/workflows/mirror-staging.yml. [x] The workflow is correctly restricted to the accessibility-exchange/platform repository. [ ] Git global user name and email are correctly set. [ ] The GitLab remote URL is correctly configured. [ ] SSH key setup is working and allows pushing to GitLab. [ ] The workflow has been tested and verified for successful execution @jobara This is pending setup of Secrets and Variables.
gharchive/pull-request
2023-05-30T18:49:58
2025-04-01T04:55:50.149627
{ "authors": [ "marvinroman" ], "repo": "accessibility-exchange/platform", "url": "https://github.com/accessibility-exchange/platform/pull/1725", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
205141149
Investigate regularization for deep models insensitivity to parameter changes: penalize with the gradient magnitude of the layer's output to its paramters L2-regularization on weights, but not on biases
gharchive/issue
2017-02-03T12:06:35
2025-04-01T04:55:50.150786
{ "authors": [ "accosmin" ], "repo": "accosmin/nano", "url": "https://github.com/accosmin/nano/issues/146", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1598498870
login test for return from login() method transkribus_utils.py, line 68 : ACDHTranskribusUtils.list_docs throws exception since requests.get(url, cookies=self.login_cookie) only returns <Response [401]> because server-auth failed. Hence ACDHTranskribusUtils.login() should check if the auth post request via equests.post() was succesfull. Authentication may fail because transkribus server side auth struggles with certain characters such as "…".
gharchive/issue
2023-02-24T11:56:19
2025-04-01T04:55:50.158030
{ "authors": [ "cfhaak" ], "repo": "acdh-oeaw/acdh-transkribus-utils", "url": "https://github.com/acdh-oeaw/acdh-transkribus-utils/issues/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1741880358
Gregor/refactor/entity table factory Describe your changes refactor: tables.py and entities-table-factory. Overview: Tables.py refactor of tables.py file to own python module within apis_entities refactor of get_entities_tables factory method into factory-class. factory-class persists already created table-classes, instead of re-creating them on each request. inner-generic-table-class removes constraints of former implementation its the broadest implementation of the table the calling view is now responsible for customizing the instance of the table class that will be send to the template, instead of hardcoding the customization in the class itself. added utility class get_entities_table, that uses the new FactoryClass, but catches all existing calls from other modules, to not break current implementations. GenericListViewNew in apis_entities.views.py: table specific code moved into own class / mixin that is now inherited by the GenericListViewNew refactored variable names moved seperate logic into own methods added docstrings: with context information that is not visible from the code itself. tried to name methods in a way that they explain what they do added inline comments on class vars and instance vars that are bound to implicit logic that happens in the parent classes - these should NOT be renamed, or things will break. Related issues and PRs #182 - implements factory refactor as an example solution for apis_entities Resolves issue: #169 ---> sorry, my formatter doesn't work currently and the files are not formatted correctly. <-- Checklist (Replace the space in square brackets with a lowercase x for all that apply) [x] My changes don't generate new warnings or errors [ ] My changes follow the project's code formatting rules and style guidelines [x] I have commented my code with Docstrings and code comments, particularly complex, unusual or hard-to-read code [ ] I have updated the project documentation to reflect the changes I introduce [ ] I have added new unit tests or updated existing ones to demonstrate my changes works Styling the columns (adjusting the width) was not possible atm, this must be done in a separate PR. headers are now removed for the two edit and detail columns. edit and detail columns are not sortable and don't participate in sorting order obviously, the two columns are now also implemented, which resolves #169 Also, there are merge conflics. Please resolve them You are totally right, I didn't notice that there where unnecessary commits, that weren't meant to be published, in the branches history. Will rework that and remember it in future commits. Closing again, for new PR with squashed commits and formatted files.
gharchive/pull-request
2023-06-05T14:12:47
2025-04-01T04:55:50.167524
{ "authors": [ "b1rger", "gregorpirgie" ], "repo": "acdh-oeaw/apis-core-rdf", "url": "https://github.com/acdh-oeaw/apis-core-rdf/pull/185", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1496183867
🛑 AceBlock Swagger UI is down In 299748f, AceBlock Swagger UI ($ACEBLOCK_SWAGGER_UI) was down: HTTP code: 0 Response time: 0 ms Resolved: AceBlock Swagger UI is back up in f17f113.
gharchive/issue
2022-12-14T09:38:53
2025-04-01T04:55:50.171991
{ "authors": [ "aceblockID" ], "repo": "aceblockID/monitoring", "url": "https://github.com/aceblockID/monitoring/issues/4241", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1203115444
🛑 AceBlock Swagger UI is down In b1be53e, AceBlock Swagger UI ($ACEBLOCK_SWAGGER_UI) was down: HTTP code: 0 Response time: 0 ms Resolved: AceBlock Swagger UI is back up in b95d99a.
gharchive/issue
2022-04-13T10:35:54
2025-04-01T04:55:50.174325
{ "authors": [ "aceblockID" ], "repo": "aceblockID/monitoring", "url": "https://github.com/aceblockID/monitoring/issues/445", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1556259749
🛑 AceBlock Swagger UI is down In b2528b2, AceBlock Swagger UI ($ACEBLOCK_SWAGGER_UI) was down: HTTP code: 0 Response time: 0 ms Resolved: AceBlock Swagger UI is back up in e7c8575.
gharchive/issue
2023-01-25T08:43:47
2025-04-01T04:55:50.176398
{ "authors": [ "aceblockID" ], "repo": "aceblockID/monitoring", "url": "https://github.com/aceblockID/monitoring/issues/5150", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1576686925
twitter share link'i çalışmıyor ana sayfadaki paylaş kısmındaki twitter share çalışmıyor. @mthocur bakabilir misin buraya bi ss atabilir misiniz @ondokuzz ben yardımcı olayım müsaitim @karasakalmt @hsyngul bu repository de henuz paylasma secenekleri yok musaitseniz paylasma seceneklerini ekleyebilirsiniz size assign edebilirim @karasakalmt assign edebilirsiniz hocam. Yalnız bi introya ihtiyacım olacak ilgili discord kanalına alırsanız çok sevinirim @hsyngul karasakalmt#1437 ulasabilirsiniz twitter kapatıldığı için çalışmıyomuş. vpn ile girince çalıştı. issue'yu kapatabiliriz. @ondokuzz Huseyin Gul#0295 discord dan ekleyebilirseniz diğer issue lar için yardımcı olabiliriz ben de ekli değilim. ekleyen olursa beni de eklesin. murat#1927 bu bu repoya ait degil bizim eklememiz gerekiyor @hsyngul ilgileniyor https://github.com/acikkaynak/deprem-io-frontend/pull/11 @karasakalmt issue yu kapatabiliriz
gharchive/issue
2023-02-08T19:28:25
2025-04-01T04:55:50.227641
{ "authors": [ "hsyngul", "karasakalmt", "ondokuzz" ], "repo": "acikkaynak/deprem-io-frontend", "url": "https://github.com/acikkaynak/deprem-io-frontend/issues/4", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1575832801
feat: seo meta tagler için yeni yapı oluşturulması İsteğiniz şeyi tanımlayın içerisinde bulunan meta taglerin dinamik hale getirilmesi sosyal mecralarda görünürlüğü arttırmak için paylaşım spesifik resimler üretilebilmesi üretilen resimlerin aynı şekilde image proplarının karşılığında gösterilmesinin sağlanması Değerlendirdiğiniz alternatifleri tanımlayın resim üretimi için gönderilen parametrelere göre sadece image outputu verebilecek bir edge function yaratılması Ek Bağlam Hedef görüntü olarak tekil bir kaydın sosyal mecralarda paylaşımı sonucu aşağıdaki gibi görünüm sağlanabilmesi thank you @dopplerDistortion tekrar açtım issue'yu tekil kayit icin bir sey tutuyor muyuz? herhangi bir sey yapildi mi orayla alakli? @eraygundogmus dc @dopplerDistortion bu iş tamamlandı mı? closed görüyorum ama in progress'te, teyit etmek istiyorum.
gharchive/issue
2023-02-08T10:06:23
2025-04-01T04:55:50.231474
{ "authors": [ "berkayvuran", "dopplerDistortion", "eraygundogmus", "fdemir" ], "repo": "acikkaynak/deprem-yardim-frontend", "url": "https://github.com/acikkaynak/deprem-yardim-frontend/issues/435", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
726298684
Eigen values Find the eigen values of a given matrix. Hint: If no function is present for determinant computation create one. I want to do this
gharchive/issue
2020-10-21T09:20:33
2025-04-01T04:55:50.251709
{ "authors": [ "anushka-gupta-31", "shreya1010svg" ], "repo": "acmpesuecc/laalibrary", "url": "https://github.com/acmpesuecc/laalibrary/issues/12", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1398287876
Expose Pod information to Containers through Environment Variables Description There may be cases where we'd like containers to have access to pod information through environment variables. Some example pod information that can be pulled through the downward api is the fieldRef and resourceFieldRef. A short list of available fields field description metadata.name the pod's name metadata.namespace the pod's namespace metadata.uid the pod's unique ID metadata.annotations['<KEY>'] the value of the pod's annotation named (for example, metadata.annotations['myannotation']) metadata.labels['<KEY>'] the text value of the pod's label named (for example, metadata.labels['mylabel']) spec.serviceAccountName the name of the pod's service account spec.nodeName the name of the node where the Pod is executing status.hostIP the primary IP address of the node to which the Pod is assigned status.podIP the pod's primary IP address (usually, its IPv4 address) Resources Kubernetes Downward API: https://kubernetes.io/docs/concepts/workloads/pods/downward-api/ Example Kubernetes downward API usage: https://kubernetes.io/docs/tasks/inject-data-application/environment-variable-expose-pod-information/ Previous discussion in Acorn Slack: https://acorn-users.slack.com/archives/C03R9ME0SKC/p1663138625938179 Closing in favor or #726
gharchive/issue
2022-10-05T19:44:30
2025-04-01T04:55:50.266885
{ "authors": [ "djcarpe" ], "repo": "acorn-io/acorn", "url": "https://github.com/acorn-io/acorn/issues/746", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1661136327
add regions to project output Checklist [ ] The title of this PR would make a good line in Acorn's Release Note's Changelog [ ] The title of this PR ends with a link to the main issue being address in parentheses, like: This is a title (#1216). Here's an example [ ] All relevant issues are referenced in the PR description. NOTE: don't use GitHub keywords that auto-close issues [ ] Commits follow contributing guidance [ ] Automated tests added to cover the changes. If tests couldn't be added, an explanation is provided in the Verification and Testing section [ ] Changes to user-facing functionality, API, CLI, and upgrade impacts are clearly called out in PR description [ ] PR has at least two approvals before merging (or a reasonable exception, like it's just a docs change) We just have a single column "Regions" with the default region having an * on it? currently it looks like this NAME DEFAULT DEFAULT REGION SUPPORTED REGIONS acorn * local [local] hub.dev.acornlabs.com/jsilverio22/acorn sandbox [sandbox] hub.dev.acornlabs.com/jsilverio22/test sandbox [sandbox] hub.dev.acornlabs.com/jsilverio22/test-region sandbox [sandbox] hub.stg.acornlabs.com/jsilverio22/acorn sandbox [sandbox] is the single column w/ * the desired? @ibuildthecloud I'd prefer just * and not (*). Also for the comma separated list don't put spaces after the ,.
gharchive/pull-request
2023-04-10T18:00:51
2025-04-01T04:55:50.271602
{ "authors": [ "ibuildthecloud", "jsilverio22" ], "repo": "acorn-io/acorn", "url": "https://github.com/acorn-io/acorn/pull/1465", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1825171795
Panic when deploying apps with images in autoupgrade format without registry. acorn version v0.8.0-alpha7-172-gff1e20f9+ff1e20f9 acorn run -n mytestnew sangeetha/myfirstacorn:v#.#.# panic: interface conversion: error is *errors.StatusError, not *imageallowrules.ErrImageNotAllowed goroutine 1 [running]: github.com/acorn-io/runtime/pkg/cli.(*Run).Run(0xc00081d880, 0xc0003f6300, {0xc0008f0330, 0x1, 0x3}) github.com/acorn-io/runtime/pkg/cli/run.go:289 +0xd06 github.com/acorn-io/runtime/pkg/cli/builder.bind.func1(0xc0003f6300?, {0xc0008f0330, 0x1, 0x3}) github.com/acorn-io/runtime/pkg/cli/builder/builder.go:431 +0x214 github.com/spf13/cobra.(*Command).execute(0xc0003f6300, {0xc0008f02d0, 0x3, 0x3}) github.com/spf13/cobra@v1.7.0/command.go:940 +0x862 github.com/spf13/cobra.(*Command).ExecuteC(0xc0001af200) github.com/spf13/cobra@v1.7.0/command.go:1068 +0x3bd github.com/spf13/cobra.(*Command).Execute(...) github.com/spf13/cobra@v1.7.0/command.go:992 github.com/spf13/cobra.(*Command).ExecuteContext(...) github.com/spf13/cobra@v1.7.0/command.go:985 github.com/acorn-io/runtime/pkg/cli.RunAndHandleError({0x48611f8?, 0xc00013fa90?}, 0xc0001af200?) github.com/acorn-io/runtime/pkg/cli/errors.go:17 +0x5d main.main() github.com/acorn-io/runtime/main.go:15 +0x2f Testing with 2bb41677, I get acorn run -n mytestnew 'sangeetha/myfirstacorn:v#.#.#' ✗ ERROR: unable to find an image for sangeetha/myfirstacorn:v#.#.# matching pattern v#.#.# - if you are trying to use a remote image, specify the full registry I assume that this was fixed along the way. Tested with acorn version v0.8.0-rc4+b6a95e8a User is presented with following error message as expected with/without have image allow rules enabled. acorn run -n mytestnew 'sangeetha/myfirstacorn:v#.#.#' ✗ ERROR: unable to find an image for sangeetha/myfirstacorn:v#.#.# matching pattern v#.#.# - if you are trying to use a remote image, specify the full registry
gharchive/issue
2023-07-27T20:32:53
2025-04-01T04:55:50.274484
{ "authors": [ "iwilltry42", "sangee2004" ], "repo": "acorn-io/runtime", "url": "https://github.com/acorn-io/runtime/issues/1989", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
346638814
how work tell me how work ! xD
gharchive/issue
2018-08-01T15:00:06
2025-04-01T04:55:50.320100
{ "authors": [ "CarlosCenturion", "sagedAlAbadla" ], "repo": "acrosa/bot", "url": "https://github.com/acrosa/bot/issues/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1178192479
fix(ingress): fix redirect by annotation on k8s >= 1.19 The API version change to the ingresses are broken when used with ports that are not numbers. E.g. ALB ingress controller annotations. Refs: https://github.com/acryldata/datahub-helm/pull/98/files/20521ec9661db7afdb3f9b96dbe53181c62947ad#r831117346 This is my relevant configuration, that worked before on a v1.21 cluster: ingress = { enabled = true hosts = [{ host = local.datahub_fqdn paths = ["/*"] redirectPaths = [{ path = "/*" name = "ssl-redirect" port = "use-annotation" // <- this is not a number! }] }] annotations = { "kubernetes.io/ingress.class" = "alb" "alb.ingress.kubernetes.io/target-type" = "ip" "alb.ingress.kubernetes.io/group.name" = "internal" "alb.ingress.kubernetes.io/listen-ports" = jsonencode([{ HTTP = 80 }, { HTTPS = 443 }]) "alb.ingress.kubernetes.io/actions.ssl-redirect" = jsonencode({ Type = "redirect", RedirectConfig = { Protocol = "HTTPS", Port = "443", StatusCode = "HTTP_301" } }) } } @dexter-mh-lee done. @dexter-mh-lee looks like 0.2.64 was already released a couple of hours ago. how do we want to fix it? open a new PR just with bumping the version? Ah. could you do that please? Merged without waiting for the actions.
gharchive/pull-request
2022-03-23T14:20:26
2025-04-01T04:55:50.323119
{ "authors": [ "dexter-mh-lee", "felixb" ], "repo": "acryldata/datahub-helm", "url": "https://github.com/acryldata/datahub-helm/pull/100", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
522490053
Routines API? Hello, I'm trying to build an app that reminds users to take their medication at a specific time or Routine. Is there a way to create a routine from my custom app without the user leaving the app and creating the routine outside of it? Hi @Ksaboor, Are you referring to a mobile app (iOS/Android), a web app, or a conversational Action for the Google Assistant? Actions on Google currently supports adding your conversational Action to a Google Assistant Routine using Routine suggestions. (See here for more detail.) Hello @taycaldwell, My apologies I meant a custom google home assistant app.
gharchive/issue
2019-11-13T21:43:56
2025-04-01T04:55:50.374024
{ "authors": [ "Ksaboor", "taycaldwell" ], "repo": "actions-on-google/actions-on-google-nodejs", "url": "https://github.com/actions-on-google/actions-on-google-nodejs/issues/369", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
392513798
Actions on Console: Need an information..How to get the Account linking logs.. Hi. Is there any way to check the logs (normal case without error), while i do the account linking? I have checked in LogsViewer i could not get the response from oAuth. For my account linking i have used the amazon oAuth server . For my analysis, i need the logs received from oAuth. Is there any way? If you're using a service like Amazon's for OAuth, you would need to check with them for logging. Hi Fleker, The basic reason for asking the same is, i need to get the agentuserid. Currently in a scenario, i have completed the OAuth - Account Linking using amazon. But during the SYNC response i could not find agentuserid, as smarthome receives only intent and other parameter. Let me know if you have inputs... The agentUserId is a value that you provide, it's not given to you. Yes, i have learned it. Using the OAuth - Access token, i need to get an AgentUserid (unique id) from OAuth Provider. I have found that from Amazon API gateway to my handler, the Authorisation headers are removed. I am checking with Amazon. Thanks a lot for your response.
gharchive/issue
2018-12-19T09:36:11
2025-04-01T04:55:50.376416
{ "authors": [ "Fleker", "sakthisg" ], "repo": "actions-on-google/smart-home-nodejs", "url": "https://github.com/actions-on-google/smart-home-nodejs/issues/240", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2665048130
MSVC ARM compiler fails CMake configuration Description Using the MSVC ARM compiler on the latest windows-2022 image fails CMake configuration Image: windows-2022 Version: 20241113.3.0 Included Software: https://github.com/actions/runner-images/blob/win22/20241113.3/images/windows/Windows2022-Readme.md Image Release: https://github.com/actions/runner-images/releases/tag/win22%2F20241113.3 Platforms affected [ ] Azure DevOps [X] GitHub Actions - Standard Runners [ ] GitHub Actions - Larger Runners Runner images affected [ ] Ubuntu 20.04 [ ] Ubuntu 22.04 [ ] Ubuntu 24.04 [ ] macOS 12 [ ] macOS 13 [ ] macOS 13 Arm64 [ ] macOS 14 [ ] macOS 14 Arm64 [ ] macOS 15 [ ] macOS 15 Arm64 [ ] Windows Server 2019 [X] Windows Server 2022 Image version and build link Image: windows-2022 Version: 20241113.3.0 Included Software: https://github.com/actions/runner-images/blob/win22/20241113.3/images/windows/Windows2022-Readme.md Image Release: https://github.com/actions/runner-images/releases/tag/win22%2F20241113.3 https://github.com/libsdl-org/SDL/actions/runs/11873470201/job/33088578294 Is it regression? https://github.com/libsdl-org/SDL/actions/runs/11873470201/job/33088868957 Expected behavior CMake configuration suceeds. kernel32.lib is a library that is expected to be available in any Windows SDK. -- The C compiler identification is MSVC 19.41.34123.0 -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Check for working C compiler: C:/Program Files/Microsoft Visual Studio/2022/Enterprise/VC/Tools/MSVC/14.41.34120/bin/Hostx64/arm/cl.exe - skipped -- Detecting C compile features -- Detecting C compile features - done Actual behavior -- The C compiler identification is MSVC 19.42.34433.0 -- Detecting C compiler ABI info -- Detecting C compiler ABI info - failed -- Check for working C compiler: C:/Program Files/Microsoft Visual Studio/2022/Enterprise/VC/Tools/MSVC/14.42.34433/bin/Hostx64/arm/cl.exe -- Check for working C compiler: C:/Program Files/Microsoft Visual Studio/2022/Enterprise/VC/Tools/MSVC/14.42.34433/bin/Hostx64/arm/cl.exe - broken CMake Error at C:/Program Files/CMake/share/cmake-3.31/Modules/CMakeTestCCompiler.cmake:67 (message): The C compiler "C:/Program Files/Microsoft Visual Studio/2022/Enterprise/VC/Tools/MSVC/14.42.34433/bin/Hostx64/arm/cl.exe" is not able to compile a simple test program. It fails with the following output: Change Dir: 'D:/a/SDL/SDL/build/CMakeFiles/CMakeScratch/TryCompile-rxem71' Run Build Command(s): D:/a/_temp/ninja-1.12.1-X64/ninja.exe -v cmTC_aa530 [1/2] C:\PROGRA~1\MICROS~2\2022\ENTERP~1\VC\Tools\MSVC\1442~1.344\bin\Hostx64\arm\cl.exe /nologo /DWIN32 /D_WINDOWS /Zi /Ob0 /Od -MTd /showIncludes /FoCMakeFiles\cmTC_aa530.dir\testCCompiler.c.obj /FdCMakeFiles\cmTC_aa530.dir\ /FS -c D:\a\SDL\SDL\build\CMakeFiles\CMakeScratch\TryCompile-rxem71\testCCompiler.c [2/2] C:\Windows\system32\cmd.exe /C "cd . && "C:\Program Files\CMake\bin\cmake.exe" -E vs_link_exe --msvc-ver=1942 --intdir=CMakeFiles\cmTC_aa530.dir --rc=C:\PROGRA~2\WI3CF2~1\10\bin\100261~1.0\x64\rc.exe --mt=C:\PROGRA~2\WI3CF2~1\10\bin\100261~1.0\x64\mt.exe --manifests -- C:\PROGRA~1\MICROS~2\2022\ENTERP~1\VC\Tools\MSVC\1442~1.344\bin\Hostx64\arm\link.exe /nologo CMakeFiles\cmTC_aa530.dir\testCCompiler.c.obj /out:cmTC_aa530.exe /implib:cmTC_aa530.lib /pdb:cmTC_aa530.pdb /version:0.0 -DEBUG /debug /INCREMENTAL /subsystem:console kernel32.lib user32.lib && cd ." FAILED: cmTC_aa530.exe C:\Windows\system32\cmd.exe /C "cd . && "C:\Program Files\CMake\bin\cmake.exe" -E vs_link_exe --msvc-ver=1942 --intdir=CMakeFiles\cmTC_aa530.dir --rc=C:\PROGRA~2\WI3CF2~1\10\bin\100261~1.0\x64\rc.exe --mt=C:\PROGRA~2\WI3CF2~1\10\bin\100261~1.0\x64\mt.exe --manifests -- C:\PROGRA~1\MICROS~2\2022\ENTERP~1\VC\Tools\MSVC\1442~1.344\bin\Hostx64\arm\link.exe /nologo CMakeFiles\cmTC_aa530.dir\testCCompiler.c.obj /out:cmTC_aa530.exe /implib:cmTC_aa530.lib /pdb:cmTC_aa530.pdb /version:0.0 -DEBUG /debug /INCREMENTAL /subsystem:console kernel32.lib user32.lib && cd ." LINK Pass 1: command "C:\PROGRA~1\MICROS~2\2022\ENTERP~1\VC\Tools\MSVC\1442~1.344\bin\Hostx64\arm\link.exe /nologo CMakeFiles\cmTC_aa530.dir\testCCompiler.c.obj /out:cmTC_aa530.exe /implib:cmTC_aa530.lib /pdb:cmTC_aa530.pdb /version:0.0 -DEBUG /debug /INCREMENTAL /subsystem:console kernel32.lib user32.lib /MANIFEST /MANIFESTFILE:CMakeFiles\cmTC_aa530.dir/intermediate.manifest CMakeFiles\cmTC_aa530.dir/manifest.res" failed (exit code 1104) with the following output: LINK : fatal error LNK1104: cannot open file 'kernel32.lib' ninja: build stopped: subcommand failed. CMake will not be able to correctly generate this project. Call Stack (most recent call first): CMakeLists.txt:4 (project) -- Configuring incomplete, errors occurred! Repro steps build: os: windows-latest steps: - use: ilammy/msvc-dev-cmd@v1 with: arch: x64_arm - run: | # download ninja - run: | cmake -S some-cmake-project -B build -GNinja See https://github.com/zufuliu/notepad4/issues/839, Windows SDK 10.0.26100.0 no longer install libs for 32-bit arm. Following is my workaround (copy libs from Windows SDK 10.0.22621.0). see https://github.com/zufuliu/notepad4/commit/e53e43305e7e0805670cb6d1abc5ef155fb36357: SET "Win10Lib=C:\Program Files (x86)\Windows Kits\10\Lib" robocopy "%Win10Lib%\10.0.22621.0\ucrt\arm" "%Win10Lib%\10.0.26100.0\ucrt\arm" /E 1>NUL robocopy "%Win10Lib%\10.0.22621.0\um\arm" "%Win10Lib%\10.0.26100.0\um\arm" /E 1>NUL Thanks, I worked around this issue my ci by pinning the Windows SDK version to 10.0.22621.0 (sdk input of ilammy/msvc-dev-cmd) Is there a public announcement about ARM32 on Windows being deprecated? Hi @madebr - Thank you for bringing this issue to our attention. We will look into this issue and will update you after investigating. Meh... this broke ReactOS ARM port build: https://github.com/actions/runner-images/issues/10978#issuecomment-2480636730 (at first I thought this is a problem with 20241113.3.0 image) I haven't tested yet but explicitly setting the SDK version in your project files (.vcxproj) might be a workaround. Setting it to 10.0 will make Visual Studio pick the latest installed version. <PropertyGroup Label="Globals"> <WindowsTargetPlatformVersion>10.0.22621.0</WindowsTargetPlatformVersion> ... https://learn.microsoft.com/en-us/gaming/gdk/_content/gc/tools-pc/visualstudio/gr-vs-2019-support-notes Hi @madebr , It is belongs to WDK update which is not supported 32 bit as mentioned by @zufuliu . Thanks you for understanding! Hence, We're closing this issue . Please feel free to reach out if you have any concern.
gharchive/issue
2024-11-16T22:48:55
2025-04-01T04:55:50.388843
{ "authors": [ "Prabhatkumar59", "binarymaster", "kishorekumar-anchala", "madebr", "maron2000", "zufuliu" ], "repo": "actions/runner-images", "url": "https://github.com/actions/runner-images/issues/10981", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1523980316
docs: update chart to reflect macos-latest The chart makes it seem like macos-latest is 11 when it's actually 12. Hello! We update docs manually as soon as it is fully migrated.
gharchive/pull-request
2023-01-07T16:58:04
2025-04-01T04:55:50.390599
{ "authors": [ "ckipp01", "mikhailkoliada" ], "repo": "actions/runner-images", "url": "https://github.com/actions/runner-images/pull/6888", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
954566015
Builds are failing at Ubuntu-latest based Microsoft hosted pool agents Description I have a privately hosted pool (OryxLinux) and I'm trying to move to Microsoft hosted pool where image is ubuntu-latest.As part of the test I create docker container with port mapping and then access those access those container ports from test (simple get http call) to see if they are working. I'm getting following error message and tests are failing left and right. Did not get the port mapping in expected format. at Microsoft.Oryx.Tests.Common.EndToEndTestHelper.RunAndAssertAppAsync(String imageName, ITestOutputHelper output, IEnumerable1 volumes, List1 environmentVariables, Int32 port, String link, String runCmd, String[] runArgs, Func2 assertAction, DockerCli dockerCli) in /home/vsts/work/1/s/tests/Oryx.Tests.Common/EndToEndTestHelper.cs:line 228 at Microsoft.Oryx.Tests.Common.EndToEndTestHelper.BuildRunAndAssertAppAsync(ITestOutputHelper output, IEnumerable1 volumes, String buildImage, String buildCmd, String[] buildArgs, String runtimeImageName, List1 environmentVariables, Int32 port, String link, String runCmd, String[] runArgs, Func2 assertAction) in /home/vsts/work/1/s/tests/Oryx.Tests.Common/EndToEndTestHelper.cs:line 182 at Microsoft.Oryx.Integration.Tests.DotNetCoreRuntimeVersion21Tests.CanBuildAndRunApp_WhenRecursiveLookUpIsDisabled_ButProjectSettingIsSupplied() in /home/vsts/work/1/s/tests/Oryx.Integration.Tests/DotNetCore/DotNetCoreRuntimeVersion21Tests.cs:line 683 --- End of stack trace from previous location where exception was thrown --- Standard Output Messages: Debugging Information: StdOut: Found build manifest file at '/oryxtests/a0ec176690a54017897a4006e2bfd16d/oryx-manifest.toml'. Deserializing it... Build Operation ID: |IM7kOncAsTk=.1ad74fbc_ Agent extension Before if loop >> DotNet Runtime 2.1.26 DotNet Runtime 2.1.26Writing output script to '/run.sh' Found startup DLL name from manifest file Running the command: dotnet "MyWebApp.dll" Hosting environment: Production Content root path: /oryxtests/a0ec176690a54017897a4006e2bfd16d Now listening on: http://[::]:3000 Application started. Press Ctrl+C to shut down. All these tests are running fine locally as well as in the privately hosted oryxlinux pool. I'm guessing some port restriction is there. Can you please take a look into it? Here is a sample build https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=5029380&view=logs&j=72dd53dc-cc7e-5359-419d-190601712c34&t=28831f2f-ff0f-5427-ca18-e78664bf3d01 Virtual environments affected [ ] Ubuntu 16.04 [ ] Ubuntu 18.04 [X] Ubuntu 20.04 [ ] macOS 10.15 [ ] macOS 11 [ ] Windows Server 2016 [ ] Windows Server 2019 Image version and build link image: ubuntu-latest failed build (hosted in azure pipelines) link: https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=5029380&view=results successful build (hosted in private pool - OryxLinux) link: https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=5029191&view=results relevant PR: https://github.com/microsoft/Oryx/pull/1064/files Is it regression? n/a Expected behavior all the tests should run successfully, container ports should be accessible from tests Actual behavior all the tests failing with following error message [xUnit.net 00:15:10.10] Microsoft.Oryx.Integration.Tests.DotNetCoreRuntimeVersion21Tests.CanBuildAndRunApp_WhenRecursiveLookUpIsDisabled_ButProjectSettingIsSupplied [FAIL] X Microsoft.Oryx.Integration.Tests.DotNetCoreRuntimeVersion21Tests.CanBuildAndRunApp_WhenRecursiveLookUpIsDisabled_ButProjectSettingIsSupplied [29s 753ms] Error Message: Did not get the port mapping in expected format. StdOut: 0.0.0.0:49206 :::49206 Repro steps Not sure if this works .. but the easiest repro would be ... rerun the failed job of the following build ... https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=5029380&view=results Hello @arroyc, We will check the issue @arroyc can you please point out the script or yml file where is the step As part of the test I create docker container with port mapping @Darleev any update? @dsame thanks for your reply .. i believe this is what you are asking .. this is where we populate docker arguments for port mapping https://github.com/microsoft/Oryx/blob/a389296f4a1a52b0581660c43a2ba641c04b77d8/tests/Oryx.Tests.Common/DockerCli.cs#L355 https://github.com/microsoft/Oryx/blob/a389296f4a1a52b0581660c43a2ba641c04b77d8/tests/Oryx.Tests.Common/DockerCli.cs#L291 This is how we run docker once we populate all the necessary arguments https://github.com/microsoft/Oryx/blob/a389296f4a1a52b0581660c43a2ba641c04b77d8/tests/Oryx.Tests.Common/DockerCli.cs#L121 https://github.com/microsoft/Oryx/blob/a389296f4a1a52b0581660c43a2ba641c04b77d8/src/BuildScriptGenerator.Common/ProcessHelper.cs#L97 let me know if thats not what you are looking for, I can share other details as you need @arroyc can you please point out the script or yml file where is the step As part of the test I create docker container with port mapping ? The root of the problem the docker listen on both IPv4 and IPv6 interfaces, so the detected ports string is 0.0.0.0:49263 :::49263 which cases the assert to fail: var portMapping = stdOut?.Split(":"); Assert.NotNull(portMapping); Assert.True( (portMapping.Length == 2), In order to provide the fix i need to know the command which runs the docker. so we are into that famous issue where docker is binding to both ipv4 and ipv6 and we want only ipv4, hmm doing a quick search on internet shows I need to set net.ipv6.bindv6only to 0. Is that the fix? I was able to find an workaround by installing a stable docker ce here is the build id https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=5035925&view=logs&s=ff05ad62-bb9a-53b6-ce9f-72f329a63e7c&j=65e01c55-aa95-5d41-4bd2-87f2a532974e
gharchive/issue
2021-07-28T07:32:35
2025-04-01T04:55:50.413219
{ "authors": [ "Darleev", "arroyc", "dsame" ], "repo": "actions/virtual-environments", "url": "https://github.com/actions/virtual-environments/issues/3807", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
859002762
WEBGL_compressed_texture_astc error is output to the console I have tried Rhodonite v0.3.19dev in a Windows environment. However, it seems that the error of WEBGL_compressed_texture_astc extension is output to the console. https://cx20.github.io/gltf-test/examples/rhodonite/index.html?category=tutorialModels&model=AlphaBlendModeTest&scale=0.5&type=glTF rhodonite.min.js:1 The library does not support this environment because the WEBGL_compressed_texture_astc is not available __getExtension @ rhodonite.min.js:1 p @ rhodonite.min.js:1 addWebGLContext @ rhodonite.min.js:1 setProcessApproachAndCanvas @ rhodonite.min.js:1 load @ index.js:53 async function (async) load @ index.js:49 load (async) (anonymous) @ index.js:341 This extension is only available in certain environments. It is better not to make an error because this extension is not available in Windows environments. ┆Issue is synchronized with this Wrike Item Currently, Those are now warning, not error actually. But certainly, It looks better to change them to 'info' output than 'warn'. @cx20 Since 01443f9d48726029abb5d1a69c8d13e71365a2c7, those warning-level messages are now info-level messages. This change would have made the console much cleaner and improved the psychological safety of users.
gharchive/issue
2021-04-15T15:33:54
2025-04-01T04:55:50.472810
{ "authors": [ "cx20", "emadurandal" ], "repo": "actnwit/RhodoniteTS", "url": "https://github.com/actnwit/RhodoniteTS/issues/764", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1607408300
Change recursive variable to integer if self.recursive == "not_recursive": to 0 Done in aaf1daf3b7e8f164e0caf3b4e13285d515649a83
gharchive/issue
2023-03-02T20:00:33
2025-04-01T04:55:50.489018
{ "authors": [ "acturtle", "zchmielewska" ], "repo": "acturtle/cashflower", "url": "https://github.com/acturtle/cashflower/issues/114", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
897494271
🛑 Processors is down In 32f7483, Processors ($STATUS_URL) was down: HTTP code: 200 Response time: 119 ms Resolved: Processors is back up in 9a12449.
gharchive/issue
2021-05-20T21:51:05
2025-04-01T04:55:50.491879
{ "authors": [ "max-acumen" ], "repo": "acumenlabs/status-page", "url": "https://github.com/acumenlabs/status-page/issues/1126", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
897712116
⚠️ Fetchers has degraded performance In 369a37a, Fetchers ($STATUS_URL) experienced degraded performance: HTTP code: 200 Response time: 388 ms Resolved: Fetchers performance has improved in 5b7b4fc.
gharchive/issue
2021-05-21T06:08:03
2025-04-01T04:55:50.494325
{ "authors": [ "max-acumen" ], "repo": "acumenlabs/status-page", "url": "https://github.com/acumenlabs/status-page/issues/1128", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1094700935
⚠️ Processors has degraded performance In 8348101, Processors ($STATUS_URL) experienced degraded performance: HTTP code: 200 Response time: 44 ms Resolved: Processors performance has improved in 59e08b6.
gharchive/issue
2022-01-05T20:03:46
2025-04-01T04:55:50.496459
{ "authors": [ "max-acumen" ], "repo": "acumenlabs/status-page", "url": "https://github.com/acumenlabs/status-page/issues/2211", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1327106427
⚠️ Processors has degraded performance In 26dfb80, Processors ($STATUS_URL) experienced degraded performance: HTTP code: 200 Response time: 113 ms Resolved: Processors performance has improved in d3f2ba1.
gharchive/issue
2022-08-03T11:54:19
2025-04-01T04:55:50.498549
{ "authors": [ "danielshir" ], "repo": "acumenlabs/status-page", "url": "https://github.com/acumenlabs/status-page/issues/3556", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2299141828
Add dictionary to run with fuzzers Relates to: https://github.com/ada-url/ada/pull/658 Is there anything that needs to be done to pass the dict to the fuzzer? Is there anything that needs to be done to pass the dict to the fuzzer? From what I was reading here, looks we just need this .options file. I saw some other examples, just updating the build.sh to copy these files too and we should be fine. But let's see... https://github.com/google/oss-fuzz/blob/ad88a2e5295d91251d15f8a612758cd9e5ad92db/projects/expat/build.sh @anonrig would you like to set a max value here? From examples, expat defined one but other projects were not. We can do 1024 @anonrig have you noticed any difference since last run? I haven't @anonrig have you noticed any difference since last run? I haven't I haven't as well
gharchive/pull-request
2024-05-16T01:45:53
2025-04-01T04:55:50.502961
{ "authors": [ "CarlosEduR", "anonrig" ], "repo": "ada-url/ada", "url": "https://github.com/ada-url/ada/pull/660", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
884520738
Monospaced blocks are rendered incorrectly in Outlook on the web (OWA) Monospaced blocks seem to behave differently in the side pane in Office 365 on the web (OWA) than they do if you double click the e-mail so it pops out. It looks fine when you compose the e-mail, but the issue shows up when you read it. Given the following example text: hello ``` pre text here ``` bye The following HTML is generated in the Outlook "side-pane": <div> <p style="margin:0 0 1.2em 0;">hello</p> <pre style="font-size:1em;font-family:Consolas,Inconsolata,Courier,monospace;margin:1.2em 0;line-height:1.2em;"><code style="font-size:0.85em;font-family:Consolas,Inconsolata,Courier,monospace;background-color:#F8F8F8;display:inline;white-space:pre;border-radius:3px;margin:0 0.15em;padding:0.5em 0.7em;border:1px solid #CCCCCC;overflow:auto;">pre text here </code></pre> <p style="margin:0 0 1.2em 0;">bye</p> <div title="MDH:redacted" style="font-size:0em;width:0;height:0;max-width:0;margin:0;padding:0;overflow:hidden;max-height:0;"> ​</div> </div> While this is generated in the popup window: <div class="x_markdown-here-wrapper" style=""> <p style="margin:0px 0px 1.2em!important">hello</p><pre style="font-size:0.85em; font-family:Consolas,Inconsolata,Courier,monospace; font-size:1em; line-height:1.2em; margin:1.2em 0px"><code style="font-size:0.85em; font-family:Consolas,Inconsolata,Courier,monospace; margin:0px 0.15em; padding:0px 0.3em; white-space:pre-wrap; border:1px solid rgb(234,234,234); background-color:rgb(248,248,248); border-radius:3px; display:inline; white-space:pre; overflow:auto; border-radius:3px; border:1px solid rgb(204,204,204); padding:0.5em 0.7em; display:block!important">pre text here </code></pre> <p style="margin:0px 0px 1.2em!important">bye</p> <div title="MDH:redacted" style="height:0; width:0; max-height:0; max-width:0; overflow:hidden; font-size:0em; padding:0; margin:0">​</div> </div> This gets extra messy with multiple lines: Side-pane: Popup: I found that I could fix this by editing the theme customization in the add-on preferences and taking out the !important in this line: display: block !important; /* added to counteract the Yahoo-specific `code` rule; without this, code blocks in Blogger are broken */ I did this with Markdown Here Revival, which I'm now using instead of this original one since it's no longer maintained. I recommend you switch. This is now fixed in Markdown Here Revival.
gharchive/issue
2021-05-10T15:52:20
2025-04-01T04:55:50.602833
{ "authors": [ "lgrn", "neilmayhew" ], "repo": "adam-p/markdown-here", "url": "https://github.com/adam-p/markdown-here/issues/659", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1839535961
Add support for 10Mhz clock First of all thanks for your work on this library! I was looking to use your library to support the CAN SPI Click board which uses the MCP2515, however I was initially unable to use it as the board uses a 10Mhz clock which is not supported by your library. This pull request adds support for 10Mhz clocks. Hello 👋 First of all, thank you for your contribution! I'll review the changes and come back to you early next week. Did you add the support for 25kbps CAN for other oscillator frequencies as well or just the 10MHz? Just for the 10Mhz. I didn’t want to risk modifying anything not directly related to my change.On 7 Aug 2023, at 18:48, Piotr Adamczyk @.***> wrote: Did you add the support for 25kbps CAN for other oscillator frequencies as well or just the 10MHz? —Reply to this email directly, view it on GitHub, or unsubscribe.You are receiving this because you authored the thread.Message ID: @.***>
gharchive/pull-request
2023-08-07T14:09:20
2025-04-01T04:55:50.628062
{ "authors": [ "Stephen-Bath", "adamczykpiotr" ], "repo": "adamczykpiotr/pico-mcp2515", "url": "https://github.com/adamczykpiotr/pico-mcp2515/pull/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1402929749
Renaming doesn't work properly. Updating the anidb library seems working. But when I try to rename my files, they always end up in " [] - -.mkv". I tried all possible combinations from the available tags but none of them seems to work. Could you provide the exact command you have used? Either it is wrong, or anidb does not return any data. anidbcli -r -e mkv,mp4 api -u "myuser" -p "mypw" -r "%a_romaji% - %a_kanji% - %a_english% - %ep_no% [%g_name%] (%crc32%)" --state 2 -a "." For the test I added all kind of % to see which one is working. Exact same command seems to be working just fine, did you try it with any other release? It's possible that anidb doesn't provide the data for that file (althought it should end up with error). I tried an anime from the last and one from the current season, both animes should be there as it was added to my anidb library. I tried to debug the python code, but its a bit difficult as you get banned after trying it 5-6 times. I am not an expert in python, if you can help me to help debug it for you - i maybe can help and find the issue. Sorry, I totally forgot about this issue. For debugging/running invoking anidbcli multiple times in a single batch you can invoke it with parameter -t, anidbcli will then reuse the previous session token (if available) and you will not get banned.
gharchive/issue
2022-10-10T10:17:04
2025-04-01T04:55:50.631597
{ "authors": [ "Xentros", "adameste" ], "repo": "adameste/anidbcli", "url": "https://github.com/adameste/anidbcli/issues/27", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
98599032
Possibility to respond to https requests without connecting upstream servers This enables a caching proxy for offline use, which is no problem with HTTP at the moment. Additionally HTTPS needs the hostname to create a certificate and must suppress the handshake to upstream without a connection. @jekh Please, let us discuss this. UnencryptedUDTChainedProxyTest fails in Travis-CI with Java 8, but works with Java 7 and 6. It works in my environment with Java 8, too: $ java -version java version "1.7.0_79" OpenJDK Runtime Environment (IcedTea 2.5.6) (7u79-2.5.6-1~deb8u1) OpenJDK 64-Bit Server VM (build 24.79-b02, mixed mode) $ /usr/lib/jvm/java-8-openjdk-amd64/bin/java -version openjdk version "1.8.0_45-internal" OpenJDK Runtime Environment (build 1.8.0_45-internal-b14) OpenJDK 64-Bit Server VM (build 25.45-b02, mixed mode) $ uname -sorv Linux 3.16.0-4-amd64 #1 SMP Debian 3.16.7-ckt11-1+deb8u2 (2015-07-17) GNU/Linux I have some reservations about this change. It's very specific to one particular use case, a caching proxy for offline use. I do see the utility of wanting to avoid making connections to the remote server. For non-SSL connections this is already possible using the filters, but for SSL connections it isn't, because the ConnectionFlow always attempts to establish the remote connection. In the long term, I'd like to see a more flexible way of controlling connections, so that users of LittleProxy can customize the connection flow as needed. For now, I'm not sure that checking remoteAddress.isUnresolved() is the best way to indicate that we should suppress the connection to the remote server. It's overloading the use of the remoteAddress.isUnresolved() status for an unrelated purpose. Can you think of a better way to indicate in the connection flow that we should not connect to the upstream server? Perhaps it could be integrated into one of the return values of the filters? For what it's worth, the Betamax testing library also has this requirement, as it works on the basis of replaying recorded responses in place of contacting the target server. I may try the options mentioned in #184 to see if they have an effect. I think this is going to become more and more important as websites default to SSL. @pledbrook You need special branches of LittleProxy and LittleProxy-mitm to enable MITM while offline. At the moment consider to use https://github.com/ganskef/LittleProxy-parent to build both. I'm closing this PR since it's replaced with #282.
gharchive/pull-request
2015-08-02T10:07:23
2025-04-01T04:55:50.643402
{ "authors": [ "ganskef", "jekh", "pledbrook" ], "repo": "adamfisk/LittleProxy", "url": "https://github.com/adamfisk/LittleProxy/pull/230", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
257670389
No exception with corrupt file Hi, actualy we decompress our GZIP TAR file in C# with the command line tool "Tar.exe". We start an process and parse the result of the command line. I want to change this by unsing Sharpcomress. We have sereval unit test for the decompression. When I use Sharpcomeress to decompress one of the unit tests failed. This test try to extract a corrupt TAR file. The "Tar.exe" thorws an exception. Sharpcomress decompress the file without an exception. Is this a known bug? This is the code I use to decompress the attached file: public bool Extract(string filename, string extractPath) { try { using (Stream stream = File.OpenRead(filename)) { var reader = ReaderFactory.Open(stream); reader.WriteAllToDirectory(extractPath, new ExtractionOptions { ExtractFullPath = true, Overwrite = true }); } return true; } catch (Exception e) { Console.WriteLine(e); return false; } } Kind regards Mike 20150427_140120_f4_3d_80_00_01_6e.zip SharpCompress does not do any validation of CRCs that GZip or whatever might have. The extraction will look successful. It's not a bug but the lack of a CRC validation feature. The TAR.exe say this: tar -tf 20150427_140120_f4_3d_80_00_01_6e.scd1 Warning: non-standard extension used on a compressed archive. Exception: System.Exception: header checksum is invalid. bei Ionic.Tar._internal_ListOrExtract(String archive, Boolean wantExtract) bei Ionic.TarApp.ListContents() bei Ionic.TarApp.Main(String[] args) Is it possible to make this to a feature request? Tar.zip It is possible. CRC is already a known missing feature. I'm accepting pull requests if you'd like to do it. How can I do that?
gharchive/issue
2017-09-14T10:28:43
2025-04-01T04:55:50.650835
{ "authors": [ "InBerg", "adamhathcock" ], "repo": "adamhathcock/sharpcompress", "url": "https://github.com/adamhathcock/sharpcompress/issues/297", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
620233780
Randomly fadeOut glitch on Component hide I open a new Issue to divide topics. Description When the component hide using internal function and fadeOut animation occurs a random glitch showing the component for just half a second after fadeOut is completed. Possibile Enhance I don't know if problem is here, but as the fadeOut is a Promise, personally I'd prefer waiting for both end before setting visible to false. So, instead writing: this.modalContainerRef.current.fadeOut(ANIMATION_TIME); await this.pickerContainerRef.current.fadeOut(ANIMATION_TIME); this.setState({ visible: false }, () => { this.cache.purge(); }); You probably could be more consistent and avoid any error writing await Promise.all([ /** Hide Modal Container */ new Promise((resolve) => { if (this.modalContainerRef.current?.fadeOut) { return this.modalContainerRef.current.fadeOut(ANIMATION_TIME).then(resolve); } return resolve(); }), /** Hide Picker Container */ new Promise((resolve) => { if (this.pickerContainerRef.current?.fadeOut) { return this.pickerContainerRef.current.fadeOut(ANIMATION_TIME).then(resolve); } return resolve(); }), ]); this.setState({ visible: false }, () => { this.cache.purge(); }); Hi @marcocavanna. I've been playing around with a few different options today - do you want to see if #16 solves the issues on your end? If it doesn't, I might need a little more info from you - because I can't seem to replicate it anymore. Note that I originally did something similar to you with the Promise.all() technique, but I realised that I didn't actually need to fade both elements out anyway - only the parent container! Also, note that those manual fadeOut methods are only used by Android as of the latest version (v1.1.0). iOS relies on the default React Native <Modal /> fade out functionality because I found smoother results using that instead. This should be fixed in v1.1.1. Please re-open if not.
gharchive/issue
2020-05-18T13:59:30
2025-04-01T04:55:50.654782
{ "authors": [ "adammcarth", "marcocavanna" ], "repo": "adammcarth/react-native-segmented-picker", "url": "https://github.com/adammcarth/react-native-segmented-picker/issues/15", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2311610208
🛑 relay.wagnersnetz.de is down In 6040b6a, relay.wagnersnetz.de (https://relay.wagnersnetz.de/actor) was down: HTTP code: 502 Response time: 726 ms Resolved: relay.wagnersnetz.de is back up in 43c16d6 after 9 minutes.
gharchive/issue
2024-05-22T22:39:56
2025-04-01T04:55:50.664510
{ "authors": [ "adamus1red" ], "repo": "adamus1red/ActivityPub-Relays", "url": "https://github.com/adamus1red/ActivityPub-Relays/issues/11894", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1573065931
🛑 relay.pissdichal.de is down In bcf2574, relay.pissdichal.de (https://relay.pissdichal.de/actor) was down: HTTP code: 0 Response time: 0 ms Resolved: relay.pissdichal.de is back up in 7fb20b3.
gharchive/issue
2023-02-06T18:23:18
2025-04-01T04:55:50.667726
{ "authors": [ "adamus1red" ], "repo": "adamus1red/ActivityPub-Relays", "url": "https://github.com/adamus1red/ActivityPub-Relays/issues/3269", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
153091586
Email regex too strict The regex requires a TLD, which isn't required in the email address spec[0]. There's been a lot written[1] about email regex that's too strict and I tend to agree that it's better to have false positives than false negatives, especially with a default validator. The regex actually fails a lot of valid email addresses (although most will likely rarely appear). Here's a list[2] of valid email address the regex fails, "much.more unusual"@example.com "very.unusual.@.unusual.com"@example.com "very.(),:;<>[]".VERY."very@\ "very".unusual"@strange.example.com admin@mailserver1 #!$%&'*+-/=?^_`{}|~@example.org "()<>[]:,;@\"!#$%&'*+-/=?^_`{}| ~.a"@example.org " "@example.org (space between the quotes) example@localhost (sent from localhost) user@com user@localserver user@[IPv6:2001:db8::1] [0] https://tools.ietf.org/html/rfc2822#section-3.4 [1] Email spec in ABNF and also includes the regex implementation https://html.spec.whatwg.org/multipage/forms.html#valid-e-mail-address [2] Taken from https://en.wikipedia.org/wiki/Email_address#Valid_email_addresses You also don't seem to be able to validate emails that use special characters e.g. scøtt@êxample.com Will get to this with https://github.com/adamwaite/Validator/issues/30. If anyone could provide an updated regex I'll be happy to implement it. I don't know enough about it to do it myself. Here's the regex implementation from the spec cited above, /^[a-zA-Z0-9.!#$%&'*+\/=?^_`{|}~-]+@[a-zA-Z0-9](?:[a-zA-Z0-9-]{0,61}[a-zA-Z0-9])?(?:\.[a-zA-Z0-9](?:[a-zA-Z0-9-]{0,61}[a-zA-Z0-9])?)*$/ @adamkuipers your rule is very complete but does not work for special chars like the one @slhodson969 highlighted this is a very blatant email validator ^.+@.+\\..+$ Remember that is better to check email on server side and not block the user on UI side. If we put a wrong validator, users with valid emails will be unable to register in your app and not being very happy about it Yeah, I agree that the simpler one is better. https://github.com/adamwaite/Validator/blob/master/Validator/Validator/EmailValidationPattern.swift @adamwaite A quick information: What is your plan to release a new version (for carthage ;) ) with these PR merged? Thanks Right now 🎉 Version 2.1.0 ready to roll.
gharchive/issue
2016-05-04T19:10:44
2025-04-01T04:55:50.677343
{ "authors": [ "adamkuipers", "adamwaite", "racer1988", "slhodson969" ], "repo": "adamwaite/Validator", "url": "https://github.com/adamwaite/Validator/issues/36", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
24652790
Support KB Server Allow AIM to connect to a KB server to retrieve / set new KB values. This will probably be a somewhat larger research item -- is it better for mobile devices to pull down larger chunks of the KB server / do a bulk update, or work with smaller bits of data? Hi Erik, Thanks for supplying this library note. It is the javascript library version of the python based avahi linux zeroconf library module I use to create standalone KB services for kbdiscovery testing for Bruce. It allows me to test discovery on linux whereas I use the bonjour dns-sd version for OSX & windows. Interestingly also shows the npm package downloader and installer ...equivalent of using pip on all 3 OS's . Good to see this thanks. On Tue, May 22, 2018 at 4:16 AM, Erik Brommers notifications@github.com wrote: Javascript library for local (LAN) KB server discovery: https://www.npmjs.com/package/bonjour — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/adapt-it/adapt-it-mobile/issues/8#issuecomment-390737980, or mute the thread https://github.com/notifications/unsubscribe-auth/ANW-JFbpxM-oPqsEO1Qa33Sd9MoYtgZzks5t0wSZgaJpZM4BVtco .
gharchive/issue
2013-12-20T22:49:57
2025-04-01T04:55:50.686796
{ "authors": [ "eb1", "gateslp" ], "repo": "adapt-it/adapt-it-mobile", "url": "https://github.com/adapt-it/adapt-it-mobile/issues/8", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2561584272
fix: set options.source before raising error on empty doc tree Made sure the options variable is properly backfilled before any exception can be raised This fixes #705 Thanks!
gharchive/pull-request
2024-10-02T13:00:36
2025-04-01T04:55:50.701652
{ "authors": [ "adbar", "dmoklaf" ], "repo": "adbar/trafilatura", "url": "https://github.com/adbar/trafilatura/pull/707", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
310677348
HASwitchPlate Connected automation triggers multiple times The HASwitchPlate_Connected automation triggers multiple (close to 15 times) when cycling power on the panel or when it reconnects to WiFi/MQTT file: HASwitchPlatePages.yaml I can't seem to reproduce this behavior. I've deployed a new instance of hass (currently 0.67.1) and deployed the stock config from the repo here with the following command: svn export https://github.com/aderusha/HASwitchPlate/trunk/Home_Assistant ~/.homeassistant --force I started up Hass and configured the default state of the page selection buttons with the following commands: curl -X POST -H "x-ha-access: YOUR_PASSWORD" -H "Content-Type: application/json" -d '{ "entity_id": "input_text.haswitchplate_pagebutton1_label", "value": "scenes" }' http://localhost:8123/api/services/input_text/set_value curl -X POST -H "x-ha-access: YOUR_PASSWORD" -H "Content-Type: application/json" -d '{ "entity_id": "input_text.haswitchplate_pagebutton2_label", "value": "status" }' http://localhost:8123/api/services/input_text/set_value curl -X POST -H "x-ha-access: YOUR_PASSWORD" -H "Content-Type: application/json" -d '{ "entity_id": "input_text.haswitchplate_pagebutton3_label", "value": "media" }' http://localhost:8123/api/services/input_text/set_value curl -X POST -H "x-ha-access: YOUR_PASSWORD" -H "Content-Type: application/json" -d '{ "entity_id": "input_number.haswitchplate_pagebutton1_page", "value": 1}' http://localhost:8123/api/services/input_number/set_value curl -X POST -H "x-ha-access: YOUR_PASSWORD" -H "Content-Type: application/json" -d '{ "entity_id": "input_number.haswitchplate_pagebutton2_page", "value": 2}' http://localhost:8123/api/services/input_number/set_value curl -X POST -H "x-ha-access: YOUR_PASSWORD" -H "Content-Type: application/json" -d '{ "entity_id": "input_number.haswitchplate_pagebutton3_page", "value": 8}' http://localhost:8123/api/services/input_number/set_value curl -X POST -H "x-ha-access: YOUR_PASSWORD" -H "Content-Type: application/json" -d '{ "entity_id": "input_number.haswitchplate_active_page", "value": 1}' http://localhost:8123/api/services/input_number/set_value Once Hass has initialized I powered on the HASwitchPlate and observed all automations running once. I power cycled the HASwitchPlate device several more times with Hass running and observed the same behavior each time. Can you test this with a clean configuration by chance? Is there any way another automation you've deployed might be interacting somehow? I've narrowed the issue down to two compounding problems. One is an issue with Hass for which I have submitted a bug report. With MQTT discovery enabled, hass is sending multiple replies in response to a single trigger. This behavior is compounded by the page switching automations, which in certain cases can result in the page command being sent twice. Because Hass is duplicating the automation response, the system can get stuck in a loop sending page commands back and forth, each of which triggers an additional cascade of MQTT messages until the Arduino code can't keep up and either resets or takes too long to respond to keep the loop going. The result can be 10 or more responses on some of the page flipping automations. this is still occurring. in fact I have the switch plate disconnected from power yet whenever I restart HA, the state goes back to "Connected" even though the switch plate is physically powered off, then the automation triggers now close to 25 times
gharchive/issue
2018-04-03T03:00:54
2025-04-01T04:55:50.730654
{ "authors": [ "aderusha", "squirtbrnr" ], "repo": "aderusha/HASwitchPlate", "url": "https://github.com/aderusha/HASwitchPlate/issues/7", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1337325297
Update to Python 3.9 Hi, Can you update to Python 3.9? Im getting this error: [ERROR] WebDriverException: Message: Service chromedriver unexpectedly exited. Status code was: 127 Traceback (most recent call last):   File "/var/task/lambda_function.py", line 80, in lambda_handler     driver = webdriver.Chrome(chrome_options=chrome_options)   File "/opt/python/selenium/webdriver/chrome/webdriver.py", line 68, in __init__     self.service.start()   File "/opt/python/selenium/webdriver/common/service.py", line 98, in start     self.assert_process_still_running()   File "/opt/python/selenium/webdriver/common/service.py", line 109, in assert_process_still_running     raise WebDriverException( Thanks. Seeing as how AWS has now removed 3.6 support it seems as if the only compatible version is 3.7 based on other issues that I've read. Hopefully this will be looked at so that we at least have 3.7 & 3.8 support. In any case some updates would be awesome.
gharchive/issue
2022-08-12T15:06:24
2025-04-01T04:55:50.735570
{ "authors": [ "paulocoutinhox", "selected-pixel-jameson" ], "repo": "adieuadieu/serverless-chrome", "url": "https://github.com/adieuadieu/serverless-chrome/issues/342", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2580419765
Add flags to specify the exact functionality required from a particular sub command When executed vc-analyze calc-stats "path/to/repo" Instead add flags to the cmd to specify the type of operation required vc-analyze calc-stats --author-stats "path/to/repo" vc-analyze calc-stats --commit-size "path/to/repo" /assign
gharchive/issue
2024-10-11T04:43:19
2025-04-01T04:55:50.737321
{ "authors": [ "Nikhils-179", "adigulalkari" ], "repo": "adigulalkari/VC-Analyzer", "url": "https://github.com/adigulalkari/VC-Analyzer/issues/25", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
136973590
"Parsing failed: Unexpected ï" if file has byte order mark If checkstyle is run on a file with a byte order mark (for example this file), it causes the following error: ./source/Test.hx:1:1: error: Parsing failed: Unexpected ï Stacktrace: Called from hxparse/Lexer.hx line 100 Called from Checker.hx line 104 Called from Checker.hx line 133 Checkstyle version: 191bf0c245 This should be fixed in hxparse, because if I remove the BOM from content before sending it to Lexer, all positions are off by three. And that would likely affect a lot of checks. So the solution for now is to remove the BOM from your files. Makes sense. And yeah, that's what I did. :) Ping @Simn. I don't think hxparse is the right place to fix this, haxeparser is more appropriate.
gharchive/issue
2016-02-27T22:00:40
2025-04-01T04:55:50.775781
{ "authors": [ "AlexHaxe", "Gama11", "Simn" ], "repo": "adireddy/haxe-checkstyle", "url": "https://github.com/adireddy/haxe-checkstyle/issues/44", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1014094649
What if we add jump to top button , so user easily jump to top whenever they need. If u can add it go ahead Sir May I ! Sir May I ! Assigned u ! @adithyapaib i would linke to work on this issue please assign this to me. Sir, I have made my pull request #40 and added a ( jump to top) button and is working fine,pls have a look @adithyapaib PR is done please cross check if it is working fine or not
gharchive/issue
2021-10-02T16:09:00
2025-04-01T04:55:50.783200
{ "authors": [ "AnuragRoshan", "Sameer1408", "adithyapaib", "kapilbhale" ], "repo": "adithyapaib/Neomorphic-PortfolioTemplate", "url": "https://github.com/adithyapaib/Neomorphic-PortfolioTemplate/issues/24", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2597525184
i have added the transition in the quicklinks Issues Identification Closes: #805 Description Summary i have added transition too quicklinks Details i have added transition Types of Changes Please check the boxes that apply [x ] Bugfix (non-breaking change that fixes an issue) [ ] New feature (non-breaking change that adds functionality) [ ] Breaking change (fix or feature that would cause existing functionality to not work as expected) [ ] Documentation update (Documentation content changed) [ ] Other (please describe): Checklist Please check the boxes that apply [x ] My code follows the style guidelines of this project [x ] I have performed a self-review of my own code [x ] I have commented my code, particularly in hard-to-understand areas [ ] I have made corresponding changes to the documentation [x ] My changes generate no new warnings [x ] My changes do not break the current system and pass all existing test cases [x ] I have added tests that prove my fix is effective or that my feature works [ x] New and existing unit tests pass locally with my changes Screenshots Additional Information i cant see any difference can you share recording from the deployment as i did not get any reply i am closing this will reopen when i get a reply
gharchive/pull-request
2024-10-18T13:29:18
2025-04-01T04:55:50.789063
{ "authors": [ "Ash182004", "aditya-bhaumik" ], "repo": "aditya-bhaumik/Pathsphere", "url": "https://github.com/aditya-bhaumik/Pathsphere/pull/807", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
923829538
🛑 Massar Login is down In 974d57b, Massar Login (https://massarservice.men.gov.ma/moutamadris/Account) was down: HTTP code: 0 Response time: 0 ms Resolved: Massar Login is back up in 47f1b98.
gharchive/issue
2021-06-17T11:46:32
2025-04-01T04:55:50.833377
{ "authors": [ "adnane-X-tebbaa" ], "repo": "adnane-X-tebbaa/DownTime-Score", "url": "https://github.com/adnane-X-tebbaa/DownTime-Score/issues/260", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1954283282
feat: article meta blocks EXLM-180 Please always provide the Jira Issue your PR is for, as well as test URLs where your change can be observed (before and after): Jira ID: https://jira.corp.adobe.com/browse/EXLM-180 Test URLs: Before: https://main--exlm--adobe-experience-league.hlx.page/docs/integrations-learn/experience-cloud/solution-categories/content-management After: https://exlm-180-article-meta--exlm--adobe-experience-league.hlx.page/docs/integrations-learn/experience-cloud/solution-categories/content-management @nitin-rachabathuni overall, I think decorating based on textContent is a bad idea. In the converter, can we split this into 2 blocks instead? one for topic-list and one for created-for ? @nitin-rachabathuni overall, I think decorating based on textContent is a bad idea. In the converter, can we split this into 2 blocks instead? one for topic-list and one for created-for ? Hi @ahmed-musallam I had to do it based on textContent as id/classes/meta-tags are all getting stripped off. Went ahead with textContent as there is no other way. cc @thomsebastin @devisree29 @nitin-rachabathuni overall, I think decorating based on textContent is a bad idea. In the converter, can we split this into 2 blocks instead? one for topic-list and one for created-for ? Hi @ahmed-musallam , I have updated the code, ty. @devisree29 could you please address this at the convertor side.
gharchive/pull-request
2023-10-20T13:21:48
2025-04-01T04:55:50.838383
{ "authors": [ "Sivaramvt", "ahmed-musallam", "nitin-rachabathuni" ], "repo": "adobe-experience-league/exlm", "url": "https://github.com/adobe-experience-league/exlm/pull/45", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
315403689
Adding testng variant, supporting through IExecutionListener Description @agudian Adding TestNG variant Related Issue Tasks [x] I have signed the CLA. [x] I have written tests and verified that they fail without my change. Thanks @agudian, can you please review again? Thanks @agudian, did that too, sorry have been more close to javascript, still catching up with best practices. Let me know if I can be of more help
gharchive/pull-request
2018-04-18T09:36:36
2025-04-01T04:55:50.845958
{ "authors": [ "santthosh" ], "repo": "adobe/S3Mock", "url": "https://github.com/adobe/S3Mock/pull/48", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2321893264
Fix test plugin-action-bar test failing in CI but passing locally https://github.com/adobe/aem-sidekick/blob/main/test/app/components/plugin/plugin-action-bar.test.js#L636 This test hasn't failed on me in a while, closing. @dylandepass if it's still failing for you, feel free to reopen.
gharchive/issue
2024-05-28T20:45:20
2025-04-01T04:55:50.847498
{ "authors": [ "dylandepass", "rofe" ], "repo": "adobe/aem-sidekick", "url": "https://github.com/adobe/aem-sidekick/issues/156", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
18582695
Focus lost after closing Project Settings File > Project Settings Click Cancel Try to type anything Result: Nothing happens. Keyboard focus is not in editor area. We just need a simple EditorManager.focusEditor() call to fix this, so marking as a starter bug. Or we could use this as an opportunity to finally fix post-dialog focus more generally :-) n00b here, searching for my first bug to tackle. Unless I'm mistaken, the status of this issue should no longer be open since @Jerhamre seems to have added a fix that's been merged in #12824? Thanks @webjocky 👍
gharchive/issue
2013-08-27T01:04:07
2025-04-01T04:55:50.849578
{ "authors": [ "peterflynn", "petetnt", "webjocky" ], "repo": "adobe/brackets", "url": "https://github.com/adobe/brackets/issues/4954", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
21929203
Mac 10.9: Brackets Helper process erroneously shown as "Not Responding" in Activity Monitor Description: Mac 10.9: Brackets Helper process Not Responding 15-20 seconds after launch in Activity Monitor. Shows up in red text in Activity Monitor. The same thing happens with Google Chrome on Mac 10.9, so this is a tracking bug. According to Peter Flynn, it's the GPU helper. We checked the Chrome process ID that was not responding > opened Chrome's task manager (hamburger icon > Tools > Task Manager) > mapped not responding process ID from Activity Monitor to process ID in Chrome Task Manager > GPU Process. Here is the Chromium bug for this issue: Issue 304860: Mavericks: Helper processes shown in Activity Monitor as "Not Responding" https://code.google.com/p/chromium/issues/detail?id=304860 Repro on Mac 10.9 using Brackets Sprint 33 10188 GM, Sprint 32 9791 GM, Sprint 31 9569 GM. UTR on Mac 10.8 using BK Sprint 33 GM 10188. Repro steps: Install Brackets on Mac 10.9. Launch Activity Monitor. Launch Brackets. Watch Activity Monitor for ~15-20 seconds. Actual results: Activity Monitor says Brackets Helper process is not responding. Expected results: Brackets Helper is responding. Workaround: None. Everything seems to be working fine, so it appears this is a false positive. Comment from Chromium bug: "This is an issue with how Activity Monitor determines responsiveness – these processes are running perfectly fine. (We changed chrome to work around that Activity Monitor issue, and the fix will reach you eventually, depending on which channel you use, how far we merge this, etc.)" From what I can see, this isn't related to chrome. I don't use Chrome, only Safari. Brackets Helper on my Mac isn't responding as well, on fresh restart, without the use of live preview. This bug is about brackets helper 'erroneously' appearing unresponsive in Activity monitor. The application should not be affected by this. Do you mean the same? Also which version of brackets are you using? I notice from the comments above that the issue may have been fixed in the chromium build we integrated some time back. Are you seeing this in Brackets 1.2? @prksingh Good catch! It does appear fixed now on Mac, so let's close it. @ElegantScripting If you're seeing slow performance when using Live Preview, that's a different issue. Could you file a new bug with more details? (E.g. does it happen with the simple Getting Started project that comes with Brackets too, or only with your own projects?) No live preview Extensions disabled Restarted brackets Using latest version Hmm, so apparently looks like the Chrome bug regressed on Yosemite: https://code.google.com/p/chromium/issues/detail?id=397642. They did land a fix on the Chrome 39 2171 branch shortly before we started out upgrade to CEF 2171, but I'm guessing we just narrowly missed the boat on picking up that fix. (If I'm reading the version numbers right, we need 2171.50 and we currently have 2171.36). Updating the bug title and reopening.
gharchive/issue
2013-10-31T21:44:19
2025-04-01T04:55:50.858974
{ "authors": [ "ElegantScripting", "dalcala", "peterflynn", "prksingh" ], "repo": "adobe/brackets", "url": "https://github.com/adobe/brackets/issues/5794", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
447096264
CIF-826 - Add integration tests Description This PR adds integration tests in it/http. To run the integration tests, a test content package at it/content needs to be installed. Also the tests rely on a mock server in it/mock-server. These tests are automatically executed on the master branch for any new change. Coverage is reported to codecov.io with the #integration flag. Unit tests are marked with the #unittests flag. External contributors are supposed to run the integration tests locally as described in the README.md. Pull requests will be validated using the integration tests running on a Jenkins instance on Adobe internal infrastructure. The tests need to be executed against AEM 6.4, 6.5 and master, the Jenkins job has to be triggered manually for each PR. Types of changes [ ] Bug fix (non-breaking change which fixes an issue) [X] New feature (non-breaking change which adds functionality) [ ] Breaking change (fix or feature that would cause existing functionality to change) Checklist: [X] I have signed the Adobe Open Source CLA. [X] My change requires a change to the documentation. [X] I have updated the documentation accordingly. [X] I have read the CONTRIBUTING document. [X] I have added tests to cover my changes and the overall coverage did not decrease. [X] All unit tests pass on CircleCi. [X] I ran all tests locally and they pass. Codecov Report Merging #22 into master will increase coverage by 7.1%. The diff coverage is n/a. @@ Coverage Diff @@ ## master #22 +/- ## =========================================== + Coverage 71.34% 78.45% +7.1% - Complexity 205 237 +32 =========================================== Files 16 16 Lines 904 905 +1 Branches 143 143 =========================================== + Hits 645 710 +65 + Misses 160 112 -48 + Partials 99 83 -16 Flag Coverage Δ Complexity Δ #integration 78.23% <ø> (?) 237 <ø> (?) #unittests 71.34% <ø> (?) 205 <ø> (?) Impacted Files Coverage Δ Complexity Δ ...mmerce/graphql/core/GraphqlProductViewHandler.java 58.67% <0%> (+0.34%) 11% <0%> (ø) :arrow_down: ...a/impl/CatalogDataResourceProviderManagerImpl.java 77.71% <0%> (+4%) 33% <0%> (+5%) :arrow_up: .../cq/commerce/graphql/resource/ProductResource.java 90.62% <0%> (+6.25%) 9% <0%> (+1%) :arrow_up: ...adobe/cq/commerce/graphql/core/MagentoProduct.java 64.36% <0%> (+6.89%) 37% <0%> (+4%) :arrow_up: ...e/cq/commerce/graphql/resource/ResourceMapper.java 79.88% <0%> (+8.87%) 34% <0%> (+4%) :arrow_up: ...mmerce/graphql/magento/GraphqlDataServiceImpl.java 88.88% <0%> (+9.09%) 39% <0%> (+3%) :arrow_up: ...atalog/admin/impl/CatalogIdentifierDatasource.java 100% <0%> (+11.11%) 12% <0%> (+2%) :arrow_up: ...aphql/resource/GraphqlResourceProviderFactory.java 100% <0%> (+11.11%) 10% <0%> (+2%) :arrow_up: ...ommerce/graphql/resource/RootCategoryResource.java 100% <0%> (+18.18%) 4% <0%> (+1%) :arrow_up: ...cq/commerce/graphql/resource/CategoryResource.java 85% <0%> (+20%) 5% <0%> (+2%) :arrow_up: ... and 2 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 5f27bd8...0661363. Read the comment docs.
gharchive/pull-request
2019-05-22T12:03:48
2025-04-01T04:55:50.882244
{ "authors": [ "codecov-io", "herzog31" ], "repo": "adobe/commerce-cif-connector", "url": "https://github.com/adobe/commerce-cif-connector/pull/22", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1826697256
Performance: fix CLS on main element in Doc Detail page Issue mentioned in this PR: https://github.com/adobe/helix-website/issues/353#issuecomment-1649418313 Details: main element still seems to move around quite a bit: . use.live URL as test URL to have a cached and more constant reference Test Urls to be used: Before: https://main--helix-website--adobe.hlx.page/docs/sidekick After: https://redesign-doc-page-perf-fix--helix-website--adobe.hlx.live/docs/sidekick @snowiewdev i think this was fixed https://github.com/adobe/helix-website/pull/388 @davidnuescheler Thank you so much! The fix on #388 is looking good to me: https://pagespeed.web.dev/analysis/https-www-hlx-live-docs-sidekick/lbesvr3spj?form_factor=desktop Closing this issue
gharchive/issue
2023-07-28T16:10:26
2025-04-01T04:55:50.890054
{ "authors": [ "davidnuescheler", "snowiewdev" ], "repo": "adobe/helix-website", "url": "https://github.com/adobe/helix-website/issues/356", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1883494099
fix: Add rum tracking to form block Adding rum tracking to form block. https://main--helix-website--adobe.hlx.page/developer/forms vs. https://main--helix-website--pareesh.hlx.page/developer/forms Types of changes [x] Bug fix (non-breaking change which fixes an issue) [ ] New feature (non-breaking change which adds functionality) [ ] Breaking change (fix or feature that would cause existing functionality to change) Checklist: [x] I have signed the Adobe Open Source CLA. [x] My code follows the code style of this project. [ ] My change requires a change to the documentation. [ ] I have updated the documentation accordingly. [ ] I have read the CONTRIBUTING document. [ ] I have added tests to cover my changes. [ ] All new and existing tests passed. Seems outdated. Closing for now, feel free to reopen and refresh if needed.
gharchive/pull-request
2023-09-06T08:16:24
2025-04-01T04:55:50.894787
{ "authors": [ "Pareesh", "kptdobe" ], "repo": "adobe/helix-website", "url": "https://github.com/adobe/helix-website/pull/371", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2291316697
chore: deprecate bookmarklet The Sidekick Bookmarklet is being deprecated due to lack of use (0.1%). Bookmarklet users should: no longer be able to install new bookmarklets receive a deprecation message when they click the bookmarklet Test URLs: Before: https://main--helix-website--adobe.aem.page/tools/sidekick/?giturl=https://github.com/adobecom/blog After: https://deprecate-bookmarklet--helix-website--adobe.aem.page/tools/sidekick/?giturl=https://github.com/adobecom/blog
gharchive/pull-request
2024-05-12T12:41:18
2025-04-01T04:55:50.897596
{ "authors": [ "rofe" ], "repo": "adobe/helix-website", "url": "https://github.com/adobe/helix-website/pull/508", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2054592026
TableHeaderProps.columns should be readonly T[] instead of T[] Provide a general summary of the issue here Basically to make types work better I as const the columns, which causes the <Row types to be better (I get a union of literals), but TableHeaderProps errors since it expects a mutable array 🤔 Expected Behavior? Allow readonly arrays being passed 😯 Current Behavior Readonly arrays not allowed 💁 Possible Solution No response 🔦 Context No response 🖥️ Steps to Reproduce The type 'readonly [{ readonly id: "teamName"; readonly name: "Team"; }, { readonly id: "creator"; readonly name: "Creator"; }, { readonly id: "createdAt"; readonly name: "Created"; }, { readonly id: "status"; readonly name: "Status"; }]' is 'readonly' and cannot be assigned to the mutable type 'object[]'.ts(4104) types.d.ts(1740, 5): The expected type comes from property 'columns' which is declared here on type 'IntrinsicAttributes & TableHeaderProps<object> & RefAttributes<HTMLTableSectionElement>' (property) TableHeaderProps<object>.columns?: object[] | undefined const columns = [{id: 1, col1: "col1"}] as const; <TableHeader columns={columns}> () => <div/> </TableHeader> Version 1.0.0 What browsers are you seeing the problem on? Safari If other, please specify. No response What operating system are you using? Mac OS 🧢 Your Company/Team No response 🕷 Tracking Issue No response Thanks for the issue, this seems fine. I don't think we mutate the columns.
gharchive/issue
2023-12-23T00:17:39
2025-04-01T04:55:50.902539
{ "authors": [ "sbdchd", "snowystinger" ], "repo": "adobe/react-spectrum", "url": "https://github.com/adobe/react-spectrum/issues/5625", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1235763437
CSF 3 Well & Switch Closes The goal is to outline a few simple patterns for some easy components. I've also taken some liberties in removing stories I think are covered well enough by Chromatic. And I've tried to combine some of our render functions to instead utilize CSF 3's property spreading as they recommend in their docs. See docs for v2 vs v3 comparisons: https://storybook.js.org/blog/component-story-format-3-0/ Open questions/opinions needed: Do we like this style over the Template.bind({}) syntax? Do we like spreading all the props or would we prefer more stories with custom render functions instead. For instance, should our Default Switch story render children, or should it take children as args? In both cases, you could override the children, however, it's a little different. export const Default: ComponentStoryObj<typeof Switch> = { render: (args) => ( <Switch {...args} /> ), args: {children: <>Switch Label</>} }; // two options here, omit the Default.args export const NoLabel: ComponentStoryObj<typeof Switch> = { ...Default, args: {'aria-label': 'This switch has no visible label'}, name: 'no label', parameters: {description: {data: 'Try me with a screen reader.'}} }; // or include them export const NoLabel: ComponentStoryObj<typeof Switch> = { ...Default, args: {...Default.args, children: null, 'aria-label': 'This switch has no visible label'}, name: 'no label', parameters: {description: {data: 'Try me with a screen reader.'}} }; vs export const Default: ComponentStoryObj<typeof Switch> = { render: (args) => ( <Switch {...args}>Switch Label</Switch> ), args: {} }; // two options here now to override the children for our story with no labels export const NoLabel: ComponentStoryObj<typeof Switch> = { ...Default, args: {children: null, 'aria-label': 'This switch has no visible label'}, name: 'no label', parameters: {description: {data: 'Try me with a screen reader.'}} }; // or export const NoLabel: ComponentStoryObj<typeof Switch> = { render: (args) => <Switch {...args} />, args: {'aria-label': 'This switch has no visible label'}, name: 'no label', parameters: {description: {data: 'Try me with a screen reader.'}} }; ✅ Pull Request Checklist: [ ] Included link to corresponding React Spectrum GitHub Issue. [ ] Added/updated unit tests and storybook for this change (for new code or code which already has tests). [ ] Filled out test instructions. [ ] Updated documentation (if it already exists for this component). [ ] Looked at the Accessibility Practices for this feature - Aria Practices 📝 Test Instructions: 🧢 Your Project: I think I like the first example better, but it feels like a pretty small difference overall. Think I would have to try using this format for a more complex component that has a more involved setup/varied setups (aka the collection ones) to form a strong opinion. Definitely like this over the bind syntax though, that syntax was always a bit confusing to me.
gharchive/pull-request
2022-05-13T22:26:30
2025-04-01T04:55:50.908465
{ "authors": [ "LFDanLu", "snowystinger" ], "repo": "adobe/react-spectrum", "url": "https://github.com/adobe/react-spectrum/pull/3129", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2423491660
Get past 30 days' data Please ensure your pull request adheres to the following guidelines: [ ] make sure to link the related issues in this description [ ] when merging / squashing, make sure the fixed issue references are visible in the commits, for easy compilation of release notes Related Issues Thanks for contributing! no activity, closing
gharchive/pull-request
2024-07-22T18:19:29
2025-04-01T04:55:50.910273
{ "authors": [ "rachel-kittens", "solaris007" ], "repo": "adobe/spacecat-audit-worker", "url": "https://github.com/adobe/spacecat-audit-worker/pull/330", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
756415658
Add In-line Alert component Description Let's implement this re-imagined Alert component https://spectrum-contributions.corp.adobe.com/page/in-line-alert-beta/ Why do you need this feature or component? @badimon asked for it Additional context Alert was previously deprecated and was nuked, this is a new version of it: https://github.com/adobe/spectrum-css/pull/757/files#diff-47c0d377df764997b3c618a2fd561fdd7d8737a96e35dc906ee0b921618ed819L1 This has been implemented and migrated to use @adobe/spectrum-tokens: https://github.com/adobe/spectrum-css/tree/main/components/inlinealert
gharchive/issue
2020-12-03T17:52:02
2025-04-01T04:55:50.913541
{ "authors": [ "lazd", "pfulton" ], "repo": "adobe/spectrum-css", "url": "https://github.com/adobe/spectrum-css/issues/1081", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
574986849
FieldLabel: Change left/right to start/end Description Currently, FieldLabel has left/right variants, it should probably be start/end Why do you need this feature or component? This makes the API make more sense. Right now, left/right work according to LTR/RTL, but it doesn't make sense. Additional context #532 synced to jira: https://jira.corp.adobe.com/browse/CSS-128 Closing this, it's tracked in Jira
gharchive/issue
2020-03-03T22:13:59
2025-04-01T04:55:50.915624
{ "authors": [ "lazd", "misterbrownlee" ], "repo": "adobe/spectrum-css", "url": "https://github.com/adobe/spectrum-css/issues/605", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1662531544
chore: separate out js functionality for each component Description Continue migrating the JS into enhancement.js files for the components they relate to Files Changed: // Load Inputgroup functionality from components/Inputgroup/metadata/enhancement.js // Load Stepper functionality from components/stepper/metadata/enhancement.js // Load Picker functionality from components/picker/metadata/enhancement.js // Load Treeview functionality from components/treeview/metadata/enhancement.js // Load Accordion functionality from components/accordion/metadata/enhancement.js // Load Cyclebutton functionality from components/cyclebutton/metadata/enhancement.js // Load Sliders functionality from components/slider/metadata/enhancement.js // Load Dial functionality from components/dial/metadata/enhancement.js // Load Dialog functionality from components/dialog/metadata/enhancement.js I'm sure you realize this and just haven't gotten to take care of it yet, but when I run this locally and try to view the docs site, I'm getting a setPickerValue is not a function error in the console, and it's preventing anything else on any of the component pages from loading properly. Again, you probably know about this already, but I thought I should call it out so that we can take care of it! Great catch Patrick! I suspect this comes from needing to load the enhancement.js assets first before tools/documentation/assets/js/site.js which references it. We'll also need a more graceful degradation if the function can't be found - an if check should do the trick to see if window.setPickerValue is undefined.
gharchive/pull-request
2023-04-11T13:49:27
2025-04-01T04:55:50.919035
{ "authors": [ "Rajdeepc", "castastrophe" ], "repo": "adobe/spectrum-css", "url": "https://github.com/adobe/spectrum-css/pull/1767", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1034988346
Deprecate profile-counters-v1 Please link to the issue #… https://jira.corp.adobe.com/browse/CJM-17091 @prabhum2 @Seth0115 when are we planning to merge this change? @prabhum2 @Seth0115 when are we planning to merge this change? Whenever XDM team deems possible?
gharchive/pull-request
2021-10-25T10:59:40
2025-04-01T04:55:50.921928
{ "authors": [ "Seth0115", "jainaks" ], "repo": "adobe/xdm", "url": "https://github.com/adobe/xdm/pull/1394", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
2193318721
[MWPW-143523] Accessibility Fixes Describe your specific features or fixes Bundles several accessibility fixes together. Modifies template-x to display a border when the hidden link is in focus Modifies the marquee to make the pause button keyboard accessible. Modifies the layout block to be an that links to the configs destination instead of using an onclick tracker. Modifies the Monthly / Yearly pricing radio button to be selectable via keyboard Modifies the hero animation on pages with the columns block such that it can be paused / unpaused by the enter key. Resolves: MWPW-143523 MWPW-143518 MWPW-143516 MWPW-143517 MWPW-143524 MWPW-143520 MWPW-143525 MWPW-143522 Test URLs: Before: https://main--express--adobecom.hlx.page/express/ -https://template-x-accessibility--express--adobecom.hlx.page/express/ -https://template-x-accessibility--express--adobecom.hlx.page/express/feature/image/resize -https://template-x-accessibility--express--adobecom.hlx.page/express/create/logo Codecov Report Attention: Patch coverage is 33.33333% with 2 lines in your changes are missing coverage. Please review. Project coverage is 36.03%. Comparing base (75a7616) to head (f50cce1). Files Patch % Lines express/blocks/layouts/layouts.js 0.00% 2 Missing :warning: Additional details and impacted files @@ Coverage Diff @@ ## stage #762 +/- ## ========================================== - Coverage 36.26% 36.03% -0.24% ========================================== Files 132 131 -1 Lines 25728 25611 -117 ========================================== - Hits 9331 9228 -103 + Misses 16397 16383 -14 :umbrella: View full report in Codecov by Sentry. :loudspeaker: Have feedback on the report? Share it here. nice! This also helps me see one issue with the updated billing-toggle. Let me push a fix directly to this branch tomorrow I can't tell which card in the layout block I'm tabbing through on https://template-x-accessibility--express--adobecom.hlx.page/express/feature/image/resize. Is it intended? I still can't perform the copy to clipboard action on the template-x templates. Do we plan to fix that in this PR?
gharchive/pull-request
2024-03-18T21:05:56
2025-04-01T04:55:50.935595
{ "authors": [ "JingleH", "codecov-commenter", "echen-adobe", "qiyundai" ], "repo": "adobecom/express", "url": "https://github.com/adobecom/express/pull/762", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2625766126
MWPW-157100: Merch card headers as p tags for Accessibility #3080 This PR changes the tag names of all headings to <p> except for the first heading <h3> as per accessibility requirements, as changed in this milo PR: https://github.com/adobecom/milo/pull/3080 Closing this PR since merch card block NALA tests has been moved to Milo repo.
gharchive/pull-request
2024-10-31T01:49:49
2025-04-01T04:55:50.937385
{ "authors": [ "Axelcureno" ], "repo": "adobecom/nala", "url": "https://github.com/adobecom/nala/pull/461", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
921305268
Kerberos Credential Retrieval from Cache not Working in Cross-Realm Setup Summary We have a problem with JDK and Kerberos in a cross-realm setup, of which we suspect the root cause to be in the JDK. We assume cross-realm support was implemented as described here: https://bugs.openjdk.java.net/browse/JDK-8005819 The ticket's test description states the following setup: client@REALM2 has established a kerberos context with krbservice1@REALM1 with delegcred=false, krbservice1@REALM1 delegates to krbservice2@REALM1 using S4U2PROXY However, in our setup krbservice2, the one we want to delegate to, is in a different realm than krbservice1. This means we have no official statement if our setup is supposed to work at all. The delegation to krbservice2 is configured as resource-based delegation. Which means in the AD for krbservice1, krbservice2 is not in the property list of "allowedToDelegateTo". Instead, we configured the trust relationship for delegation with the PowerShell command Set-ADUser. krbservice2 cannot be added to the allowedToDelegateTo property of krbservice1 because they are in different realms. Steps to reproduce On the first round the user "client" establishes a connection to our krbservice1. krbservice1 sends a delegated search request to krbservice2. The Java security module successfully retrieves a ticket on the user's behalf, which is attached to the search request to krbservice2 and search results are returned. However, on the second search query, in the class sun/security/krb5/internal/CredentialsUtil.java it restores the credentials from the cache, but it is not able to retrieve a valid ticket for krbservice2. Instead, we get the following error messages: KRBError: sTime is Tue Jun 01 14:03:37 CEST 2021 1622549017000 suSec is 204909 error code is 28 error Message is null sname is HTTP/krbservice2@REALM2 msgType is 30 and KRBError: sTime is Tue Jun 01 14:03:37 CEST 2021 1622549017000 suSec is 201727 error code is 13 error Message is KDC cannot accommodate requested option sname is HTTP/krbservice2@REALM1 eData provided. msgType is 30 When we look into the first error message with Wireshark, we found the error to be: KRB5KRB_AP_PATH_NOT_ACCEPTED When we decode the "Unknown eData field of KRB-ERROR" sent along with the first error message, we get: C:\Users\me\Downloads>.\Err_6.4.5.exe c0000272 for hex 0xc0000272 / decimal -1073741198 STATUS_NO_MATCH ntstatus.h There was no match for the specified key in the index. as an HRESULT: Severity: FAILURE (1), FACILITY_NULL (0x0), Code 0x272 for hex 0x272 / decimal 626 ERROR_NO_MORE_MATCHES winerror.h There are no more matches for the current index enumeration. 2 matches found for "c0000272" From these error messages we understand the following: For some unknown reason in the second search request with the credentials retrieved from ReferralsCache it is not possible to find krbservice2 in REALM2 anymore. Analysis We think that this is a bug in the JDK, because when we clear the ReferralsCache between search requests, it returns search results successfully. We did this manually with the remote debugging options of an IDE. When remote-debugging into the code of CredentialsUtil we see that on the first request, Credentials of type Proxy are not stored, but the ones with None are. A comment in the code explicitly states that Credentials for Proxy should not be cached. However, in the second search request, the Credentials for Proxy are retrieved from the Cache. We think this contradicts the warning in the comment, as this is basically a caching of Proxy credentials. Therefore we suspect this to be the location of the bug and Credentials for type Proxy should not be retrieved from the cache. Please fix. Triaging info This is the data of the middle service, krbservice1 in the above described setup. Java version: We tried both: openjdk version "1.8.0_292" OpenJDK Runtime Environment (AdoptOpenJDK)(build 1.8.0_292-b10) OpenJDK 64-Bit Server VM (AdoptOpenJDK)(build 25.292-b10, mixed mode) and openjdk version "11.0.11" 2021-04-20 OpenJDK Runtime Environment AdoptOpenJDK-11.0.11+9 (build 11.0.11+9) OpenJDK 64-Bit Server VM AdoptOpenJDK-11.0.11+9 (build 11.0.11+9, mixed mode) We encounter the same problem in both versions. What is your operating system and platform? Windows Server 2016 Datacenter Did it work before? Our setup works successfully with krbservice1 and 2 in the same realm with resource-based delegation configured. Here is our suggestion for a fix. We successfully tested it on our internal environment. #318_add_check_before_requesting_referrals_cache_to_prevent_usage_of_an_inappropriate_tick.zip Duplicated by https://bugs.java.com/bugdatabase/view_bug.do?bug_id=JDK-8270137 Thanks for reporting this bug. Just for the record, the OpenJDK discussion is taking place in the context of https://github.com/openjdk/jdk/pull/5036 merged upstream.
gharchive/issue
2021-06-15T11:40:49
2025-04-01T04:55:50.965822
{ "authors": [ "Raytion", "karianna", "martinuy" ], "repo": "adoptium/adoptium-support", "url": "https://github.com/adoptium/adoptium-support/issues/318", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1113552219
exclude jdk_tools testcases due to upstream issue Fixes #3061 @sophia-guo @llxia @ShelleyLambert @llxia here are the requested changes
gharchive/pull-request
2022-01-25T08:06:13
2025-04-01T04:55:50.967547
{ "authors": [ "LizyBbethy" ], "repo": "adoptium/aqa-tests", "url": "https://github.com/adoptium/aqa-tests/pull/3283", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1549838546
Reenable JDK19 tests after the fix for eclipse-openj9/openj9#15251 The following four tests are reenabled since eclipse-openj9/openj9#15251 has been fixed. Previously they were exclude in adoptium/aqa-tests#3750 SC_Softmx_JitAot SC_Softmx_JitAot_Linux SharedClasses.SCM23.MultiThread SharedClasses.SCM23.MultiThreadMultiCL Signed-off-by: Annabelle Huo Annabelle.Huo@ibm.com @llxia @Mesbah-Alam May I ask you to review? Thank you! @JasonFengJ9 @pshipton @0xdaryl fyi
gharchive/pull-request
2023-01-19T19:43:49
2025-04-01T04:55:50.969874
{ "authors": [ "a7ehuo" ], "repo": "adoptium/aqa-tests", "url": "https://github.com/adoptium/aqa-tests/pull/4261", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1227080599
Add gcc11.2 role for jdk19+ Fixes #2538 Signed-off-by: Adam Brousseau adam.brousseau88@gmail.com Checklist [ ] commit message has one of the standard prefixes [ ] faq.md updated if appropriate [ ] other documentation is changed or added (if applicable) [ ] playbook changes run through VPC or QPC (if you have access) [ ] for inventory.yml changes, bastillion/nagios/jenkins updated accordingly @sxa can you review or delegate? I basically copy/pasted the gcc10 role. Will need Adopt to do the binary work on the the other side. Is there an outlook on timeframe for completing this work? Ie. adding the gcc11 binary to Adopt's Jenkins' userContent. VPC run: https://ci.adoptopenjdk.net/job/VagrantPlaybookCheck/1458 - I'm not sure why the CentOS6 run failed to extract GCC10 in that run though - it /shouldn't/ be anything to do with this PR. We have GCC11 now available on all of the requiredLInux platforms now (I will update the ARM32 one next week, but hat won't require a playbook change), so this should be good to go in if you're happy to take it out of draf @AdamBrousseau Think this is in order now @sxa
gharchive/pull-request
2022-05-05T19:17:57
2025-04-01T04:55:50.974795
{ "authors": [ "AdamBrousseau", "sxa" ], "repo": "adoptium/infrastructure", "url": "https://github.com/adoptium/infrastructure/pull/2542", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
915944134
Problem uninstalling AOJDK 11.0.7_10 OpenJ0.20.0 (x64) LTS in Windows 10 2H2 x64 This is a copy/move from temurin-build Issue #2572 at github https://github.com/adoptium/temurin-build/issues/2572 (as the topic also concerns the installer, and it could be useful to all to open the issue here under /installer for everyones benefit, as kindly suggested by @karianna ) What are you trying to do? -to uninstall OLD LTS (to upgrade to a current JDK 11 LTS (11.0.10_9_openj9-0.24.0.msi)) - currently the installer and the associated uninstaller is stuck, impossible to uninstall it Expected behaviour: Old 11 LTS 11.0.7_10 to be uninstalled smoothly (and replaced with the new one) Observed behaviour: The old AOJDK can not be uninstalled from windows system by its uninstaller, even if it is exactly the same name dot subvairiant etc. When trying to circumvent the problem by installing new LTS supposedly to uninstall or rewrite registry entries pointing to the correct LTS - it fails the same by invoking the same old failing uninstaller script. It asks for .msi name AOJDK EOJ9 11.0.7_10 (x64) , and when matched - just refuses to proceed and continues to insist that it is not the right one, although the name and variant is the exact subvariant dots and an exact match of a filename (downloaded .msi of the 11.0.7.10 build from github while trying to correct the situation) Any other comments: By the way of trying to resolve this, under releases history I noticed some sub-variant versions of LTS 11.0.10_9 / openj9.0.24.0 installer on github - it would be good to get everything coherent in th final outcome of eradicating .msi uninstall problem for good. Useful observation: A solid manual removal instructions is missing - it was not possible to find it only by search engine query. Thank you in advance for considering this critical uninstaller bug Please give us some log files from the msi msiexec /x "C:\MyPackage\Example.msi" /L*V "C:\log\example.log" Then try to reboot and uninstall. @douph1 Thank you for the swift reply; as I am in daily workplace, will try to provide log answer ASAP in the work break. Thank you for the log cmd line as the next step, have a great day! Closed as no new input since long time
gharchive/issue
2021-06-09T08:37:08
2025-04-01T04:55:50.981668
{ "authors": [ "douph1", "succeedmr123" ], "repo": "adoptium/installer", "url": "https://github.com/adoptium/installer/issues/318", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2454073120
Another problem with isless As a complement to https://github.com/adrhill/SparseConnectivityTracer.jl/issues/152, with non-evenly distributed t: using SparseConnectivityTracer: TracerLocalSparsityDetector, jacobian_sparsity using DataInterpolations: LinearInterpolation detector = TracerLocalSparsityDetector() A = LinearInterpolation([0.0, 2.0, 10.0], [0.0, 2.0, 3.0]) function f!(output, input) output .= A.(input) end input = rand(5) output = zero(input) jacobian_sparsity(f!, output, input, detector) this gives ERROR: MethodError: isless(::SparseConnectivityTracer.Dual{Float64, SparseConnectivityTracer.GradientTracer{SparseConnectivityTracer.IndexSetGradientPattern{Int64, BitSet}}}, ::Float64) is ambiguous. Candidates: isless(dx::D, y::Real) where D<:SparseConnectivityTracer.Dual @ SparseConnectivityTracer path\to\.julia\packages\SparseConnectivityTracer\tpAG2\src\overloads\dual.jl:25 isless(x::Real, y::AbstractFloat) @ Base operators.jl:178 Possible fix, define isless(::D, ::AbstractFloat) where D<:SparseConnectivityTracer.Dual Stacktrace: [1] lt(o::Base.Order.ForwardOrdering, a::SparseConnectivityTracer.Dual{…}, b::Float64) @ Base.Order .\ordering.jl:117 [2] bracketstrictlymontonic @ path\to\.julia\packages\FindFirstFunctions\2w5qZ\src\FindFirstFunctions.jl:179 [inlined] [3] searchsortedlastcorrelated @ path\to\.julia\packages\FindFirstFunctions\2w5qZ\src\FindFirstFunctions.jl:266 [inlined] [4] get_idx(A::LinearInterpolation{…}, t::SparseConnectivityTracer.Dual{…}, iguess::Int64; lb::Int64, ub_shift::Int64, idx_shift::Int64, side::Symbol) @ DataInterpolations path\to\.julia\packages\DataInterpolations\i1xgb\src\interpolation_utils.jl:133 [5] get_idx @ path\to\.julia\packages\DataInterpolations\i1xgb\src\interpolation_utils.jl:120 [inlined] [6] _interpolate(A::LinearInterpolation{…}, t::SparseConnectivityTracer.Dual{…}, iguess::Int64) @ DataInterpolations path\to\.julia\packages\DataInterpolations\i1xgb\src\interpolation_methods.jl:19 [7] _interpolate @ path\to\.julia\packages\DataInterpolations\i1xgb\src\interpolation_methods.jl:5 [inlined] [8] AbstractInterpolation @ path\to\.julia\packages\DataInterpolations\i1xgb\src\DataInterpolations.jl:24 [inlined] [9] _broadcast_getindex_evalf @ .\broadcast.jl:709 [inlined] [10] _broadcast_getindex @ .\broadcast.jl:682 [inlined] [11] getindex @ .\broadcast.jl:636 [inlined] [12] macro expansion @ .\broadcast.jl:1004 [inlined] [13] macro expansion @ .\simdloop.jl:77 [inlined] [14] copyto! @ .\broadcast.jl:1003 [inlined] [15] copyto! @ .\broadcast.jl:956 [inlined] [16] materialize! @ .\broadcast.jl:914 [inlined] [17] materialize!(dest::Vector{…}, bc::Base.Broadcast.Broadcasted{…}) @ Base.Broadcast .\broadcast.jl:911 [18] f!(output::Vector{SparseConnectivityTracer.Dual{…}}, input::Vector{SparseConnectivityTracer.Dual{…}}) @ Main path\to\Ribasim_development\runners\runner.jl:35 [19] trace_function(::Type{SparseConnectivityTracer.Dual{…}}, f!::typeof(f!), y::Vector{Float64}, x::Vector{Float64}) @ SparseConnectivityTracer path\to\.julia\packages\SparseConnectivityTracer\tpAG2\src\interface.jl:49 [20] _local_jacobian_sparsity(f!::Function, y::Vector{…}, x::Vector{…}, ::Type{…}) @ SparseConnectivityTracer path\to\.julia\packages\SparseConnectivityTracer\tpAG2\src\interface.jl:94 [21] jacobian_sparsity(f!::Function, y::Vector{…}, x::Vector{…}, ::TracerLocalSparsityDetector{…}) @ SparseConnectivityTracer path\to\.julia\packages\SparseConnectivityTracer\tpAG2\src\adtypes.jl:122 [22] top-level scope @ path\to\Ribasim_development\runners\runner.jl:41 Some type information was truncated. Use `show(err)` to see complete types. Thanks a lot! Looks like the overload of isless(x::Dual, y::Real) added in #149 isn't specific enough for input types and we also need an additional overload on isless(x::Dual, y::AbstractFloat). Maybe we should run our Dual tests on more diverse input types to catch this type of problem.
gharchive/issue
2024-08-07T18:25:33
2025-04-01T04:55:50.988912
{ "authors": [ "SouthEndMusic", "adrhill" ], "repo": "adrhill/SparseConnectivityTracer.jl", "url": "https://github.com/adrhill/SparseConnectivityTracer.jl/issues/153", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1328610483
Open Weather Map Can't find open weather map API on rapid Api website and alternate or some sort of help will be appreciated. same problem here You can make a free account on the OpenWeather website. Assuming you have created the .env file to hide your api keys you can then update the getWeatherData function in the js file of the api folder: export const getWeatherData = async (lat, lng) => { try { const apiKey = process.env.REACT_APP_OPEN_WEATHER_API_KEY; const { data } = await axios.get( `https://api.openweathermap.org/data/2.5/weather?lat=${lat}&lon=${lng}&appid=${apiKey}` ); return data; } catch (error) { console.log(error); } }; Additionally, the img src will need to be updated in the Map.jsx file: {weatherData?.list?.map((data, i) => ( <div key={i} lat={data.coord.lat} lng={data.coord.lon}> <img src={`https://openweathermap.org/img/wn/${data.weather[0].icon}.png`} height={75} alt="" /> </div> ))} When console.log'd, the api calls return the expected data with these updates, however, I haven't been able to get the weather icons to successfully render on the map.
gharchive/issue
2022-08-04T13:16:46
2025-04-01T04:55:51.017327
{ "authors": [ "GeomaticaNet", "gitpk-0", "pankaj11koundal" ], "repo": "adrianhajdin/project_travel_advisor", "url": "https://github.com/adrianhajdin/project_travel_advisor/issues/40", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1515214437
Cannot reach myIP:3000 permission denied Is there an existing issue for this? [X] I have searched the existing issues What happened? I had tesla mate for a year use to work on debian base with docker, but lately cannot get into the UI the grafana log give me =eror msg="failed to determine the status of alerting engine. Enable either legacy or unified alerting explicitly and try again" err="failed to verify if the 'alert' table exists: unable to open database file: permission denied" and container Telamate log give me (File.Error) could not write to file "/opt/app/lib/tzdata-1.1.1/priv/latest_remote_poll.txt": permission denied no issue with ip :4000 also cannot update image: postgres:13 to image: postgres:14 not compatible? Expected Behavior Before could access the UI but not anymore Steps To Reproduce Start Teslamate as normal using step 3 The Grafana dashboards are available at http://your-ip-address:3000 Relevant log output [error] GenServer :tzdata_release_updater terminating ** (File.Error) could not write to file "/opt/app/lib/tzdata-1.1.1/priv/latest_remote_poll.txt": permission denied logger=migrator t=2022-12-31T21:42:31.39+0000 lvl=eror msg="failed to determine the status of alerting engine. Enable either legacy or unified alerting explicitly and try again" err="failed to verify if the 'alert' table exists: unable to open database file: permission denied" logger=settings t=2022-12-31T21:43:31.72+0000 lvl=info msg="Starting Grafana" version=8.5.6 commit=623255cf2d branch=HEAD compiled=2022-06-14T09:25:38Z logger=settings t=2022-12-31T21:43:31.72+0000 lvl=info msg="The state of unified alerting is still not defined. The decision will be made during as we run the database migrations" Client TESLAMATE_302E3936363 closed its connection. 1672521241: mosquitto version 2.0.15 terminating [42] FATAL: terminating connection due to administrator command 2022-12-31 21:14:01.663 UTC [41] FATAL: terminating connection due to administrator command 2022-12-31 21:14:01.663 UTC [46] FATAL: terminating connection due to administrator command 2022-12-31 21:14:01.663 UTC [43] FATAL: terminating connection due to administrator command 2022-12-31 21:14:01.663 UTC [38] FATAL: terminating connection due to administrator command 2022-12-31 21:14:01.664 UTC [44] FATAL: terminating connection due to administrator command 2022-12-31 21:14:01.664 UTC [45] FATAL: terminating connection due to administrator command 2022-12-31 21:14:01.665 UTC [1] LOG: background worker "logical replication launcher" (PID 33) exited with exit code 1 2022-12-31 21:14:01.665 UTC [37] FATAL: terminating connection due to administrator command 2022-12-31 21:14:01.667 UTC [40] FATAL: terminating connection due to administrator command 2022-12-31 21:14:01.667 UTC [39] FATAL: terminating connection due to administrator command Screenshots No response Additional data _teslamate-database-1_logs(1).txt _teslamate-teslamate-1_logs(1).txt _teslamate-grafana-1_logs(2).txt _teslamate-mosquitto-1_logs.txt Type of installation Docker Version 1.27.1 Do you have space available on the system hosting TeslaMate? Definitely not the issue, the partition where docker folder is got 430Gib free #3040 - similar errors. No fix, but possible collaboration. It looks like the both the teslamate and grafana containers have file permission problems. Were they upgraded from older versions? The output of docker exec tm_grafana_1 ls -la /var/lib/grafana docker exec tm_teslamate_1 ls -la /opt/app/lib/tzdata-1.1.1/priv might be useful (replace tm_grafana_1 and tm_teslamate_1 by the names of your containers). Just in case I removed _1 at the end of all my containers's names and relaunched the stack in portainer # docker exec teslamate-teslamate ls -la /opt/app/lib/tzdata-1.1.1/priv total 20 drwxr-xr-x 4 root users 4096 Jul 22 10:35 . drwxr-xr-x 4 root users 4096 Jul 22 10:35 .. -rw-r--r-- 1 root users 9 Jul 22 10:35 latest_remote_poll.txt drwxr-xr-x 2 root users 4096 Jul 22 10:35 release_ets drwxr-xr-x 2 root users 4096 Jul 22 10:35 tmp_downloads ~# docker exec teslamate-grafana ls -la /var/lib/grafana Error response from daemon: Container b1a475948dab1774bf8444a25ec1681ce75837921c3c9e06838b0ff3a2d714a3 is restarting, wait until the container is running In the teslamate container, you somehow have all these files owned by root. They are owned by nonroot for me with a container created using the instructions at https://docs.teslamate.org/docs/installation/docker on debian 11. Is that different from what you are doing? The command for the grafana container failed because the container was restarting. You could try it again but wait a little longer after starting it to avoid this error. I tried stopping the container and starting it after the other, this is not working because of permission issue I followed the instructions to the letter, you can see my install in the attached file is also running on a Debian 11 base : OMV. So for some reason the user became root, how can I change it back to nonroot without recreate everything and loosing all my data? teslamate.txt look like for docker exec teslamate-teslamate-1 ls -la /opt/app/lib/tzdata-1.1.1/priv the files owner is somehow fixed total 28 drwxr-xr-x 1 nonroot nonroot 4096 Jan 13 13:47 . drwxr-xr-x 1 nonroot nonroot 4096 Jan 13 13:47 .. -rw-r--r-- 1 nonroot nonroot 9 Jan 13 18:17 latest_remote_poll.txt drwxr-xr-x 1 nonroot nonroot 4096 Jan 13 18:17 release_ets drwxr-xr-x 1 nonroot nonroot 4096 Jan 13 18:17 tmp_downloads but beside this basicaly the same result, no fixed. _teslamate-teslamate-1_logs(2).txt _teslamate-mosquitto-1_logs(2).txt _teslamate-grafana-1_logs(4).txt _teslamate-database-1_logs(4).txt
gharchive/issue
2022-12-31T22:06:18
2025-04-01T04:55:51.077257
{ "authors": [ "cwanja", "dyxyl", "galosre" ], "repo": "adriankumpf/teslamate", "url": "https://github.com/adriankumpf/teslamate/issues/3035", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
748390701
Script de création de CSV en sortie de LiveCAM Créer un script permettant de créer un CSV avec comme information la date, le nombre de personnes, le chemin de la photo et le nom du fichier Ce CSV permettra de comparer le nombre de photo avec une (ou +) personne détectée et le nombre de photo total Excellent, on va bien s'amuser à analyser ce fichier 🚀 @adriens LiveCAM Stats.xlsx :bomb:
gharchive/issue
2020-11-23T00:35:52
2025-04-01T04:55:51.084854
{ "authors": [ "adriens", "gbertherat" ], "repo": "adriens/person-counter-api", "url": "https://github.com/adriens/person-counter-api/issues/43", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1710488973
🛑 adv-compass-pmdevel is down In 94ca810, adv-compass-pmdevel (https://api.github.com/orgs/advancedcsg/actions/runners/310) was down: HTTP code: 200 Response time: 161 ms Resolved: adv-compass-pmdevel is back up in 4e34375.
gharchive/issue
2023-05-15T17:07:37
2025-04-01T04:55:51.109161
{ "authors": [ "Abhyudaya1414" ], "repo": "advancedcsg-open/toolchain-status-dashboard", "url": "https://github.com/advancedcsg-open/toolchain-status-dashboard/issues/36", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1291889742
[BUG] Angle D3D11 glEnableVertexAttribArray() program state rendering error. This bug seems like it's up to google angle to fix it but just to throw it out there it seems sometimes when a shader is switched, it's vertex attributes do not switch from the old one to the new one, and an OpenGL error 0x0502 error is thrown, this bug was discovered by making a mesh render in 2D queue setForce2DQueue() with a sprite that exists in the same scene but with a lower local z value than the 2D queued mesh. Unexpected result: mesh will not be shown and an opengl error is thrown. CCRenderer.cpp:320 When flush2D is called (which renders all sprites in a scene in a single batch to minimize draw calls), it calls drawElements() in the same sort order GLOBALZ_XXX as the mesh, they both set up their vertex attributes in: CommandBufferGL.cpp:283 Now that triangle batch is drawn with it's according vertex attributes, and here comes the turn to drawMeshCommand() which does the same thing and sets up it's vertex attributes according to the program state just like before, BUT an opengl es angle error is thrown OpenGL error 0x0502 when the draw call is finally issued, which confused me as to how could a previous opengl state affect future states? I don't know if that bug affects only D3D11 Angle or OpenGL ES devices and if both are affected then mobile devices are affected by this bug too. Do you have a reproduce demo about this issue Somehow I can't reproduce the bug again with d3d11... The helloworld sprite seems to render fine behind the mesh of quads: I think it's a vertex attribute problem, the bug is really confusing. I'll close this issue for now.
gharchive/issue
2022-07-01T22:47:11
2025-04-01T04:55:51.149979
{ "authors": [ "DelinWorks", "halx99" ], "repo": "adxeproject/adxe", "url": "https://github.com/adxeproject/adxe/issues/697", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1259325029
Command line arguments for Operator In the docs https://github.com/adyanth/cloudflare-operator/blob/main/docs/configuration.md#operator it says you can add command line arguments like --overwrite-unmanaged-dns but it isn't clear to me where or how this should be done. Thanks Hey @acrogenesis, It is to be added to the deployment's spec.template.spec.containers[1].args after --leader-elect. Currently, it is a manual step, but I am planning to worry about making it easier to configure when I start looking into Helm charts for it since there are very few meaningful configurable parameters for now. Obligatory patch: (Note that it is not idempotent) kubectl -n cloudflare-operator-system patch deploy cloudflare-operator-controller-manager --type=json -p='[{"op": "add", "path": "/spec/template/spec/containers/1/args/-", "value": "--overwrite-unmanaged-dns=true"}]' Reference: https://github.com/adyanth/cloudflare-operator/blob/702e042e33c1cf802c80a0e3aaf73b20e515409e/main.go#L60 That makes sense. Thanks!
gharchive/issue
2022-06-03T03:18:42
2025-04-01T04:55:51.153238
{ "authors": [ "acrogenesis", "adyanth" ], "repo": "adyanth/cloudflare-operator", "url": "https://github.com/adyanth/cloudflare-operator/issues/61", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2125800393
[ENH] Loading unequal length, no missing values classification problems This PR: Fixes #1059 because I have reconfigured tsc.com to return 404 when file not found. Adapts load_classification to match load_regression, so that the equal length and no missing value variants of problems can be loaded directly from the zip. Allows suffixes _discr : discrete version of a regression problem _eq : equal length version of a variable length _nmv : no missing values version of set with missing values. so dataset Foo_disc_eq_nmv_TEST.ts is a variant of the regression problem with a discretised target, made equal length (method may vary on dataset) and missing values imputed. Removes the requirement for data to be on one of the data set lists (which are not really lists, Changes the name of tsc_data_lists to tsc_datasets. They are sets not lists (for query reasons) Adds a note book "data_unequal.ipynb" to demonstrate usage. Allows for loading the discrete versions of regression problems @TonyBagnall does this allow to load discrete version of regression problems using the load_classification or load_regression function ? and second question, the equal length fix follows the padding zero or the 2nd bake off ? I think we already discussed but completely forgot which is why i am asking here to keep track :) @TonyBagnall does this allow to load discrete version of regression problems using the load_classification or load_regression function ? and second question, the equal length fix follows the padding zero or the 2nd bake off ? I think we already discussed but completely forgot which is why i am asking here to keep track :) loading discrete versions with load_classification is the plan, but I think in future PR. Series padded with mean+small amount of noise
gharchive/pull-request
2024-02-08T18:13:41
2025-04-01T04:55:51.192106
{ "authors": [ "TonyBagnall", "hadifawaz1999" ], "repo": "aeon-toolkit/aeon", "url": "https://github.com/aeon-toolkit/aeon/pull/1157", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
652198623
build: use graphql-tools to load files Fixes failing integration tests in jobs/5744 Thanks for the fix, did anything to to make the test fail? I noticed the build is failing in https://github.com/aerogear/graphback/pull/1616. So it is blocking dependency updates. Though I see this is failing for a different reason now, will do some more checks. I noticed the build is failing in #1616. So it is blocking dependency updates. Though I see this is failing for a different reason now, will do some more checks. Thanks for the quick fix!!
gharchive/pull-request
2020-07-07T10:16:29
2025-04-01T04:55:51.223361
{ "authors": [ "craicoverflow", "machi1990" ], "repo": "aerogear/graphback", "url": "https://github.com/aerogear/graphback/pull/1664", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
908316950
enable VNC not working in Robot framework with selenoid we were using Protractor for test automation and we have setup Selenoid for remote execution. Everything is working perfectly fine. Now we are switching to Robot framework (Java version). I have setup the desired capabilities and remote url in "Open browser keyword" Code in robot framework for launching tests on remote with desired capabilities ${desired caps} getCapabilities // capabilities are returned in this in string format with enablevnc as true Open Browser ${SSO_URL} ${Browser} None http://0.0.0.0:4444/wd/hub ${desired caps} Tests are executing on Selenoid but not able to see the execution, VNC flag is missing. On the other hand i had done the setup in selenium 4 chrome standalone docker image and executed the robot tests on this image , i am able to see the execution even without the enable VNC flag. (tests are executed on noVNC). Could some one help me in viewing the robot test (Java Port) execution on Selenoid. @kiruparkuma see https://github.com/aerokube/selenoid/issues/554#issuecomment-432604950. Trying to run Selenoid on my Robot Framework tests but I am unable to view Video or record it. What am I doing wrong? @madb1ker see my comment above. @madb1ker see my comment above. desired_capabilities has been deprecated and removed. I can run my tests in Robot Framework with Selenoid, I just cannot find how to set enableVNC and enableVideo to True. @madb1ker Selenoid supports W3C version of the protocol which is current standard, so Selenoid capabilities can go under selenoid:options key. Not sure how to add this in robot-framework. https://www.w3.org/TR/webdriver2/#extensions-0 @madb1ker Selenoid supports W3C version of the protocol which is current standard, so Selenoid capabilities can go under selenoid:options key. Not sure how to add this in robot-framework. https://www.w3.org/TR/webdriver2/#extensions-0 Thank you! For the record this is the correct way to use Selenoid with Robot Framework passing the arguments in the browser options: ${selenoid_args} Create Dictionary enableVNC=${True} enableVideo=${True} Open Browser ${WMS_URL} ${BROWSER} remote_url=${SELENOID_SERVER} options=set_capability("selenoid:options",${selenoid_args}) Ok, closing then.
gharchive/issue
2021-06-01T13:51:42
2025-04-01T04:55:51.253846
{ "authors": [ "kiruparkuma", "madb1ker", "vania-pooh" ], "repo": "aerokube/selenoid", "url": "https://github.com/aerokube/selenoid/issues/1108", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1647301978
video rename failed cause no such file or dictionary I started Selenoid using CM, and overwrote the default video storage path, but when a task ended, an error was reported that the video file was not found. The following is the output log of Selenoid: ··· 023/03/30 09:50:24 [-] [INIT] [Using Docker API version: 1.41] 2023/03/30 09:50:24 [-] [INIT] [Timezone: UTC] 2023/03/30 09:50:24 [-] [INIT] [Listening on :4444] 2023/03/30 09:51:50 [-] [NEW_REQUEST] [unknown] [10.19.43.154] 2023/03/30 09:51:50 [-] [NEW_REQUEST_ACCEPTED] [unknown] [10.19.43.154] 2023/03/30 09:51:50 [0] [LOCATING_SERVICE] [chrome] [101.0] 2023/03/30 09:51:50 [0] [USING_DOCKER] [chrome] [101.0] 2023/03/30 09:51:50 [0] [CREATING_CONTAINER] [selenoid/vnc_chrome:101.0] 2023/03/30 09:51:50 [0] [STARTING_CONTAINER] [selenoid/vnc_chrome:101.0] [57a8fa68185b5b76751cb9adb93cd65f664a7b87031604e9d97ba4e5b89b70dc] 2023/03/30 09:51:50 [0] [CONTAINER_STARTED] [selenoid/vnc_chrome:101.0] [57a8fa68185b5b76751cb9adb93cd65f664a7b87031604e9d97ba4e5b89b70dc] [0.36s] 2023/03/30 09:51:50 [0] [CREATING_VIDEO_CONTAINER] [selenoid/video-recorder:latest-release] 2023/03/30 09:51:50 [0] [STARTING_VIDEO_CONTAINER] [selenoid/video-recorder:latest-release] [52d2a818239b4b3e3f17d9cdbc4017f7ba6c5a50d92339b4cdf00a95174b7376] 2023/03/30 09:51:51 [0] [VIDEO_CONTAINER_STARTED] [selenoid/video-recorder:latest-release] [52d2a818239b4b3e3f17d9cdbc4017f7ba6c5a50d92339b4cdf00a95174b7376] [0.57s] 2023/03/30 09:51:51 [0] [SERVICE_STARTED] [selenoid/vnc_chrome:101.0] [57a8fa68185b5b76751cb9adb93cd65f664a7b87031604e9d97ba4e5b89b70dc] [0.00s] 2023/03/30 09:51:51 [0] [PROXY_TO] [57a8fa68185b5b76751cb9adb93cd65f664a7b87031604e9d97ba4e5b89b70dc] [http://172.17.0.4:4444/] 2023/03/30 09:51:51 [0] [SESSION_ATTEMPTED] [http://172.17.0.4:4444/] [1] 2023/03/30 09:51:52 [0] [SESSION_CREATED] [a8535d93afac3bb6f2ceb066cd48fe46] [1] [1.75s] 2023/03/30 09:56:59 [4164] [SESSION_DELETED] [a8535d93afac3bb6f2ceb066cd48fe46] 2023/03/30 09:56:59 [0] [STOPPING_VIDEO_CONTAINER] [52d2a818239b4b3e3f17d9cdbc4017f7ba6c5a50d92339b4cdf00a95174b7376] 2023/03/30 09:57:00 [0] [REMOVING_CONTAINER] [52d2a818239b4b3e3f17d9cdbc4017f7ba6c5a50d92339b4cdf00a95174b7376] 2023/03/30 09:57:00 [0] [FAILED_TO_REMOVE_CONTAINER] [52d2a818239b4b3e3f17d9cdbc4017f7ba6c5a50d92339b4cdf00a95174b7376] [Error response from daemon: removal of container 52d2a818239b4b3e3f17d9cdbc4017f7ba6c5a50d92339b4cdf00a95174b7376 is already in progress] 2023/03/30 09:57:00 [0] [REMOVING_CONTAINER] [57a8fa68185b5b76751cb9adb93cd65f664a7b87031604e9d97ba4e5b89b70dc] 2023/03/30 09:57:00 [0] [CONTAINER_REMOVED] [57a8fa68185b5b76751cb9adb93cd65f664a7b87031604e9d97ba4e5b89b70dc] 2023/03/30 09:57:00 [0] [VIDEO_ERROR] [Failed to rename /data/apps/selenoid/video/selenoid4d7d695b45a07445ff6aa168a7e82297.mp4 to /data/apps/selenoid/video/chrome982Z.mp4: rename /data/apps/selenoid/video/selenoid4d7d695b45a07445ff6aa168a7e82297.mp4 /data/apps/selenoid/video/chrome982Z.mp4: no such file or directory] ··· The strange thing is that the video storage address I specified has this video. What is the reason? [@jsy19-43-152 /data/apps/selenoid/video]# ls -ll total 51300 -rw-r--r-- 1 root root 6061329 Mar 30 14:33 selenoid2f04e740e2924418e96f37c946892986.mp4 -rw-r--r-- 1 root root 5052357 Mar 30 16:30 selenoid3eafdd6088457654abfdae3f23754ade.mp4 -rw-r--r-- 1 root root 5030674 Mar 30 17:57 selenoid4d7d695b45a07445ff6aa168a7e82297.mp4 -rw-r--r-- 1 root root 15894839 Dec 9 17:22 selenoid529b1de5f2ab1b8791abbbca9801676e.mp4 -rw-r--r-- 1 root root 6139882 Mar 30 14:34 selenoid8459de847e5fb72108bead8abea1ec31.mp4 -rw-r--r-- 1 root root 14338547 Dec 13 12:56 selenoidf4c92d231db8b4a399d3a1e6260246cf.mp4 here is my cm command : ./cm selenoid start --config-dir /data/apps/selenoid/ --browsers-json /data/apps/selenoid/browsers.json --vnc --port 4444 --args "-limit 2 -timeout 3m -log-output-dir /data/apps/selenoid/logs/ -video-output-dir /data/apps/selenoid/video/ -session-attempt-timeout 2m -service-startup-timeout 2m -container-network bridge" @suzhenyu006 just using --config-dir should be sufficient. No need to override -log-output-dir and other flags as these directories are inside container and --config-dir is outside.
gharchive/issue
2023-03-30T10:02:48
2025-04-01T04:55:51.265805
{ "authors": [ "suzhenyu006", "vania-pooh" ], "repo": "aerokube/selenoid", "url": "https://github.com/aerokube/selenoid/issues/1327", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
158893486
How to hide/disable the Custom button in Color Palette [yes ] I have verified there are no duplicate active or recent bugs, questions, or requests I would like to remove the "custom" button in the Color Palette dialog. How can this be done? Reason: I have an array of preset colors. I want the user to pick only the preset colors not a custom color using seekbar Regards Hi @CeJienAJPC, this is not a bug you can deny the custom Color selection by set the Option "allowUserColorInput" to false. builder.allowUserColorInput(false); Best regards, Vall0n @Vall0n Thank you for the solution. I know it ain't a bug. I was just looking if there was a way. Thanks again
gharchive/issue
2016-06-07T10:49:52
2025-04-01T04:55:51.337522
{ "authors": [ "CeJienAJPC", "Vall0n" ], "repo": "afollestad/material-dialogs", "url": "https://github.com/afollestad/material-dialogs/issues/1094", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
477556612
Progress Dialog in v3 What module does this apply to? core Description what you'd like to happen: Google clearly stated in its design guidelines that progress dialogs are deprecated and that they should not be used. Despite this, almost every google app (Gmail, Drive, Sheets, Slides, Docs) make use of dialog progress bars to display downloads/uploads. Since this library is opinionated it may be worth the effort packing a progress layout and creating an extension function that transforms the dialog into a progress dialog. I don't want to include a progress dialog, cause I still prefer and want to encourage the use of inline progress indicators.
gharchive/issue
2019-08-06T19:32:33
2025-04-01T04:55:51.339544
{ "authors": [ "AndreaCioccarelli", "afollestad" ], "repo": "afollestad/material-dialogs", "url": "https://github.com/afollestad/material-dialogs/issues/1862", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
97100095
How to set a global style for the Dialogs I have different style per activity. When I use the Dialog, they show up with different colours in each activity . Read the README on the front page.
gharchive/issue
2015-07-24T17:38:44
2025-04-01T04:55:51.341169
{ "authors": [ "MrThiago", "afollestad" ], "repo": "afollestad/material-dialogs", "url": "https://github.com/afollestad/material-dialogs/issues/624", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
53951667
Additional fix for vibrating windows Previously we only checked if the listview could scroll down, which was incorrect when the listview had already been scrolled to the bottom Release 0.6.0 is now available Thanks @teslacoil and @hzsweers!
gharchive/pull-request
2015-01-10T08:46:05
2025-04-01T04:55:51.342358
{ "authors": [ "afollestad", "teslacoil" ], "repo": "afollestad/material-dialogs", "url": "https://github.com/afollestad/material-dialogs/pull/179", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
130319423
Provide an option to disable animations Maybe as a real Jenkins plugin that could be updated using the normal update center? The problem for me is that our users can't reach fonts.googleapis.com and afonsof.github.io from where they access the Jenkins server. Also it would be awesome if there were no animated graphics. Our clients have to be accessed using Remote Desktop and Firefox for example uses a ton of CPU to render the gif animations (It uses way less CPU for the CSS based animations) also since you use distinct shapes for yello-anime.gif and yellow.png the animations don't add anything. If I'm not mistaken this theme is only using SVG files, not gifs: https://github.com/afonsof/jenkins-material-theme/tree/master/images @jhasse you are right, the lack of gifs is a big reason for me to use this theme :) With the gifs in the original jenkins theme the animations are necessary because they look identical if I were to disable gif animations in my browser. With this theme however the animations are not needed because a Yellow and a Yellow animated (in progress or something) can be easily distinguished. I hope this explains my reference to the gif files :) Ah I understand :) I guess it would be really good to make this theme into a dedicated Jenkins plugin then, since a configuration option could then be added to disable things like animations. I like them though! Not having any CPU issues in Firefox with them. Well animations are one of those things... IMO the animated items are of no interest to me - while they are animated I don't care about them, but they grab my attention. When the build is finally done and may be interesting because it failed or has test results - they are not animated anymore. About that CPU use: Open 5 or more tabs, access your browser over Remote Desktop note CPU use multiply by the number of developers. Then cry for all the CPU cycles that your VM server has to spend to do really important stuff like animating icons... maybe you don't see it because your browser has access to one of those fancy GPUs (any actually because VMs don't have real GPUs). ;) Guys, thanks for the suggestions! I guess both of them are interesting to implement. What about splitting it in 2 different issues? Disable animation option Created a dedicated plugin for this theme Yes, good idea. I've published the Python plugin to Jenkins in the past, so I can help with the latter one :) Done. Version 1.1.7 Awesome thanks!
gharchive/issue
2016-02-01T10:18:52
2025-04-01T04:55:51.349239
{ "authors": [ "afonsof", "cobexer", "jhasse" ], "repo": "afonsof/jenkins-material-theme", "url": "https://github.com/afonsof/jenkins-material-theme/issues/22", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
122858694
Ensure ?mode=vr forces stereo https://aframe.io/aframe/examples/showcase-shopping/?mode=vr doesn't seem to force stereo. Perhaps we want ?mode=stereo, idk. Per https://webvr.slack.com/archives/declarative-vr/p1450387507000072 Yeah, I think its useful to force stereo. You wouldn't be able to so otherwise unless you are on mobile or a VR enabled browser. Also would be useful to have a function call to do the same. And a postMessage call. Hi all, I'm looking for a solution to be able to force the vr view on smartphone when my webpage is launched. Anyone can help me with that ? Thank you in advance I'm working on this now. I'll update the bug when it's ready.
gharchive/issue
2015-12-18T01:34:12
2025-04-01T04:55:51.355872
{ "authors": [ "caseyyee", "cvan", "evilgouky" ], "repo": "aframevr/aframe-core", "url": "https://github.com/aframevr/aframe-core/issues/704", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1473388798
⚠️ AGEG website (prod) has degraded performance In b367389, AGEG website (prod) (https://www.ageg.ca/) experienced degraded performance: HTTP code: 200 Response time: 1261 ms Resolved: AGEG website (prod) performance has improved in 06438d0.
gharchive/issue
2022-12-02T20:06:46
2025-04-01T04:55:51.417404
{ "authors": [ "mdaoustUdeS" ], "repo": "ageg-status/ageg-status.github.io", "url": "https://github.com/ageg-status/ageg-status.github.io/issues/822", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
387704778
Start Error Why i recive thies error ? root@vps-zap59814x-4:~/SnapchatBot-master/examples# python reflectorbot.py -u noxxxxx -p 19wr#%xxxxxxxxx [1] 21608 -bash: FeFOZyU9T@T2YRyUR%SxPz%Sg7vOCrdg5: command not found root@vps-zap59814-4:~/SnapchatBot-master/examples# [2018-12-05 11:02:36,165] Starting new HTTPS connection (1): feelinsonice-hrd.appspot.com:443 [2018-12-05 11:02:36,536] https://feelinsonice-hrd.appspot.com:443 "POST /loq/login HTTP/1.1" 400 None Traceback (most recent call last): File "reflectorbot.py", line 22, in bot = ReflectorBot(args.username, args.password) File "build/bdist.linux-x86_64/egg/snapchat_bots/bot.py", line 26, in init File "/usr/local/lib/python2.7/dist-packages/requests-2.20.1-py2.7.egg/requests/models.py", line 897, in json return complexjson.loads(self.text, **kwargs) File "/usr/lib/python2.7/json/init.py", line 339, in loads return _default_decoder.decode(s) File "/usr/lib/python2.7/json/decoder.py", line 364, in decode obj, end = self.raw_decode(s, idx=_w(s, 0).end()) File "/usr/lib/python2.7/json/decoder.py", line 382, in raw_decode raise ValueError("No JSON object could be decoded") ValueError: No JSON object could be decoded I have the same error and without any idea how to fix it. Did anybody got a fix? First line in the readme "This repo is deprecated due to changes in Snapchat's unofficial API." So there is no way to send snaps again ?
gharchive/issue
2018-12-05T11:04:50
2025-04-01T04:55:51.426362
{ "authors": [ "Nisutec", "Smadger1", "rhpo", "rydeom" ], "repo": "agermanidis/SnapchatBot", "url": "https://github.com/agermanidis/SnapchatBot/issues/67", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
349797693
chapter 11 Question 8 restore_model parameter function ? in chapter 11 Question 8 the code written to restore model parameter to value of best parameters in early stop is : def _restore_model_params(self, model_params): gvar_names = list(model_params.keys()) assign_ops = {gvar_name: self._graph.get_operation_by_name(gvar_name + "/Assign") for gvar_name in gvar_names} init_values = {gvar_name: assign_op.inputs[1] for gvar_name, assign_op in assign_ops.items()} feed_dict = {init_values[gvar_name]: model_params[gvar_name] for gvar_name in gvar_names} self._session.run(assign_ops, feed_dict=feed_dict) I'm not able to understand this code properly .Can you please explain assign_ops and init_values code ? Hi @nitml, Good question. When you create a variable, TensorFlow automatically creates several operations and tensors, including an assignment operation to initialize the variable to whatever initialization value you chose for that variable. The first line of the function gets the list of names of variables whose values we want to restore. The next line gets a reference to all the assignment operations for these variables. Indeed, if a variable is named "V", then the assignment operation is called "V/Assign". The third line gets a reference to the tensors that represent the initialization values for these variables. Indeed, an assignment operation has two inputs: (1) a reference to the variable, (2) the value to assign. Here, I am getting each assignment operation's second input (op.inputs[1]). The I prepare a feed_dict to tell TensorFlow to use the given model_params values instead of the initialization values for each variable. Then I run all the assignment operations, feeding the values to use instead of the default initialization values. Here's a simpler example, to help you understand what's going on: >>> import tensorflow as tf >>> V = tf.Variable(1, name="V") >>> graph = tf.get_default_graph() >>> sess = tf.InteractiveSession() >>> V.initializer.run() >>> V.eval() 1 >>> assign_init_value_op = graph.get_operation_by_name("V/Assign") >>> init_value = assign_init_value_op.inputs[1] >>> sess.run(assign_init_value_op, feed_dict={init_value: 2}) >>> V.eval() 2 Hope this helps, Aurélien Great Explanation, i have another doubt in this Question 8 to use RandomizedSearchCV we created Sklearn learn compatible class DNNClassifier and then used following code : `from sklearn.model_selection import RandomizedSearchCV def leaky_relu(alpha=0.01): def parametrized_leaky_relu(z, name=None): return tf.maximum(alpha * z, z, name=name) return parametrized_leaky_relu param_distribs = { "n_neurons": [10, 30, 50, 70, 90, 100, 120, 140, 160], "batch_size": [10, 50, 100, 500], "learning_rate": [0.01, 0.02, 0.05, 0.1], "activation": [tf.nn.relu, tf.nn.elu, leaky_relu(alpha=0.01), leaky_relu(alpha=0.1)], # you could also try exploring different numbers of hidden layers, different optimizers, etc. #"n_hidden_layers": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10], #"optimizer_class": [tf.train.AdamOptimizer, partial(tf.train.MomentumOptimizer, momentum=0.95)], } rnd_search = RandomizedSearchCV(DNNClassifier(random_state=42), param_distribs, n_iter=50, random_state=42, verbose=2) rnd_search.fit(X_train1, y_train1, X_valid=X_valid1, y_valid=y_valid1, n_epochs=1000)` My question is why have we defined leaked_elu function such that in has another function parametrized_leaky_relu and not simply : def relu_elu(z,alpha=0.01,name=None): return tf.maximum(alpha * z, z, name=name) Hi @nitml , With your leaky_relu() function, how would you define the param_distribs dict? For example, the following would NOT work, because the leaky_relu() function would complain that the argument z is missing: # THIS WILL NOT WORK param_distribs = { ... "activation": [tf.nn.relu, tf.nn.elu, leaky_relu(alpha=0.01), leaky_relu(alpha=0.1)], ... } Of course you could try to pass some inputs z, but which ones? We don't have that data yet, as it will only be computed during training, and it will be different for each training batch. However, you could do this: # THIS SHOULD WORK param_distribs = { ... "activation": [tf.nn.relu, tf.nn.elu, lambda z: leaky_relu(z, alpha=0.01), lambda z: leaky_relu(alpha=0.1)], ... } As you can see, the activation function needs to be, well, a function, not the result of calling the leaky_relu() function. My function creates a parametrized_leaky_relu() function that has the desired alpha hyperparameter already set to the desired value. So TensorFlow will just give it the input data, and it will compute the right thing. Your function directly returns a result, so it must be wrapped into a function, for example using a lambda. I hope this is clear, Aurélien Thanks for your Time and explanation,but things are not properly clear, suppose i want to use leaky_rule having parametrized_leaky_rule() function within it as activation function then my code will be : Z=tf.layers.dense(X,20,name="hidden_layer2") hidden2=leaky_rule(Z)` then will not it be a error as leaky_rule function does not have parameter "z "as it is parameter of parametrized_leaky_rule() function and not leaky_rule. In question 8 we used "dnn" function in "fit" function to make layers : `def _dnn(self,inputs): for layer in range(self.n_hidden_layers): if self.dropout_rate: inputs=tf.layers.dropout(inputs,self.dropout_rate,training=self.training) inputs=tf.layers.dense(inputs,self.n_neurons,kernel_initializer=self.initializer, name="hidden%d" % (layer + 1)) if self.batch_norm_momentum: inputs=tf.layers.batch_normalization(inputs,momentum=self.batch_norm_momentum,training=self.training) inputs = self.activation(inputs, name="hidden%d_out" % (layer + 1)) return inputs` so when we use "RandomizedSearchCV(DNNClassifier" with activation function as "leaky_rule" then second last line "inputs = self.activation(inputs, name="hidden%d_out" % (layer + 1))" should rise error as "inputs" is parameter of "parametrized_leaky_rule()" function and not "leaky_rule". i hope my question is clear ? Thanks for all your Time and support !! Hi @nitml , Sorry I wasn't clear. As you can see on the second to last line in the _dnn() method, the self.activation attribute needs to be a function that takes some inputs as its first argument, and also takes a name argument. For example, I could use this function if I want alpha=0.01: def leaky_relu_001(z, name=None): return tf.maximum(0.01 * z, z, name=name) Then I could just create a DNNClassifier like this: dnn_clf = DNNClassifier(activation=leaky_relu_001, ...) If I want to use alpha=0.1 instead, then I can create another function: def leaky_relu_01(z, name=None): return tf.maximum(0.1 * z, z, name=name) I could then create a param_distribs dict like this: param_distribs = { ... "activation": [tf.nn.relu, tf.nn.elu, leaky_relu_001, leaky_relu_01], ... } And then create a RandomizedSearchCV using this param_distrib, and everything should work fine. Now instead of manually copy/pasting the code to write a new function every time I want a different alpha value, I could write a create_leaky_relu_function() function, that creates a leaky_relu() function for me, with the desired alpha value: def create_leaky_relu_function(alpha): def leaky_relu(z, name=None): return tf.maximum(alpha * z, z, name=name) return leaky_relu Now if I want a new function for alpha=0.2, it is much easier to create: leaky_relu_02 = create_leaky_relu_function(0.2) Now this leaky_relu_02() function is a function that takes the inputs z and an optional name. I can use it when creating a DNNClassifier(activation=leaky_relu_02), and I can use it in param_distribs={"activation": [leaky_relu_02, ...], ...}. The main difficulty is to understand that the create_leaky_relu_function() returns a new function with its own alpha value that will never change once the function is created. Perhaps this simple example may help: >>> def create_multiplier(factor): ... def mul(x): ... return factor * x ... return mul ... >>> mul2 = create_multiplier(2) >>> mul3 = create_multiplier(3) >>> mul2(10) 20 >>> mul3(10) 30 First I create a function that can be used to create functions. Then I use it to create two functions mul2() and mul3(). The first one has factor=2, so it multiplies its argument by 2, while the second has factor=3 so it multiplies its argument by 3. Now let's look at your function: def your_leaky_relu(z, alpha=0.01, name=None): return tf.maximum(alpha * z, z, name=name) It can indeed be used as you showed: Z = tf.layers.dense(X, 20, name="hidden_layer2") hidden2 = your_leaky_relu(Z) You could even set alpha to any value you want: Z = tf.layers.dense(X, 20, name="hidden_layer2") hidden2 = your_leaky_relu(Z, alpha=0.2) However, suppose you want to pass it to the DNNClassifier constructor, and you want alpha=0.2, how can you do that? If you write the following code, then alpha is not set, so it will default to 0.01: dnn_clf = DNNClassifier(activation=your_leaky_relu, ...) But if you write the following code, then you will get an error, because activation is supposed to be a function, not a tensor (which is what your function returns): dnn_clf = DNNClassifier(activation=your_leaky_relu(alpha=0.2), ...) # ERROR The error will actually not happen right away, but only when self.activation is used (when you fit the DNNClassifier). So, as you can see, the problem is that there is no way to set the alpha argument in your function when you pass it to the DNNClassifier. This is why, if we want to try 3 different alpha values, we need 3 different functions, and this is why it is useful to have a function that creates them. Hope this helps, Aurélien Great Explanation, Thanks for your time and support
gharchive/issue
2018-08-12T09:17:23
2025-04-01T04:55:51.448160
{ "authors": [ "ageron", "nitml" ], "repo": "ageron/handson-ml", "url": "https://github.com/ageron/handson-ml/issues/281", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
124548093
SELECT 1 bug First bug to fix from here SELECT 1 FROM t1 WHERE 1 IN (SELECT 1) Error message: context is not defined Very strange error. It occurs from command line: alasql 'SELECT 1 FROM @[1,2] WHERE 1 IN (SELECT 1)' but does not inside the Mocha test.... Very strange error. It occurs from command line: alasql 'SELECT 1 FROM @[1,2] WHERE 1 IN (SELECT 1)' but does not inside the Mocha test.... Fixed. Hope now we can pass at least one SQLLOGICTEST file in2.test. @mathiasrw May I ask you to run the test again? Fixed. Hope now we can pass at least one SQLLOGICTEST file in2.test. @mathiasrw May I ask you to run the test again? Sure ! Sure ! ZAP - you have done it again! Solution to this issue tested and confirmed... The rest of the tests are still running. Ill update the repo as soon as they are done. Regarding in2.test we get the following at the moment: 2/125 ./test/evidence/in2.test Mimic sqlite SELECT 1 FROM t1 WHERE 1 IN (SELECT 1,2) No exception thrown Fail found for statement setting up data so skipping rest of tests ☓ Ran 53 tests as sqlite 3 skipped 1 failed 92% was OK Time: 465ms ZAP - you have done it again! Solution to this issue tested and confirmed... The rest of the tests are still running. Ill update the repo as soon as they are done. Regarding in2.test we get the following at the moment: 2/125 ./test/evidence/in2.test Mimic sqlite SELECT 1 FROM t1 WHERE 1 IN (SELECT 1,2) No exception thrown Fail found for statement setting up data so skipping rest of tests ☓ Ran 53 tests as sqlite 3 skipped 1 failed 92% was OK Time: 465ms So basically its expecting to get an error on the following 4 statements statement error SELECT 1 FROM t1 WHERE 1 IN (SELECT 1,2) statement error SELECT 1 FROM t1 WHERE 1 IN (SELECT x,y FROM t1) statement error SELECT 1 FROM t1 WHERE 1 IN (SELECT * FROM t1) statement error SELECT 1 FROM t1 WHERE 1 IN (SELECT min(x),max(x) FROM t1) So basically its expecting to get an error on the following 4 statements statement error SELECT 1 FROM t1 WHERE 1 IN (SELECT 1,2) statement error SELECT 1 FROM t1 WHERE 1 IN (SELECT x,y FROM t1) statement error SELECT 1 FROM t1 WHERE 1 IN (SELECT * FROM t1) statement error SELECT 1 FROM t1 WHERE 1 IN (SELECT min(x),max(x) FROM t1) I will check this tonight after LegoLand (try to get some inspiration there:) )... I will check this tonight after LegoLand (try to get some inspiration there:) )... No stress :) No stress :) Just remember that Lego is from Denmark ))) Just remember that Lego is from Denmark ))) It iiissss - and we are very proud of it. I have a feeling its not the one in Denmark you are at :) It iiissss - and we are very proud of it. I have a feeling its not the one in Denmark you are at :) https://github.com/agershun/alasql/blob/develop/TESTLOG.md is updated. https://github.com/agershun/alasql/blob/develop/TESTLOG.md is updated.
gharchive/issue
2016-01-01T15:29:42
2025-04-01T04:55:51.462008
{ "authors": [ "agershun", "mathiasrw" ], "repo": "agershun/alasql", "url": "https://github.com/agershun/alasql/issues/501", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
943810596
ModuleNotFoundError with welly version 0.4.10 I receive the error message below when trying to import welly version 0.4.10, I do not receive the same error with version 0.4.9: ModuleNotFoundError Traceback (most recent call last) /tf/eds_repos/eds-ppfg/ppfg_utils.py in <module> 7 8 import os ----> 9 import welly 10 import numpy as np 11 import matplotlib.pyplot as plt /usr/local/lib/python3.6/dist-packages/welly/__init__.py in <module> 4 ================== 5 """ ----> 6 from .project import Project 7 from .well import Well 8 from .header import Header /usr/local/lib/python3.6/dist-packages/welly/project.py in <module> 16 from tqdm import tqdm 17 ---> 18 from .well import Well, WellError 19 from . import utils 20 from .utils import deprecated /usr/local/lib/python3.6/dist-packages/welly/well.py in <module> 23 from .curve import Curve 24 from .header import Header ---> 25 from .location import Location 26 from .synthetic import Synthetic 27 from .canstrat import well_to_card_1 /usr/local/lib/python3.6/dist-packages/welly/location.py in <module> 19 from .fields import dev_fields 20 from .crs import CRS ---> 21 from .tools import compute_position_log 22 23 /usr/local/lib/python3.6/dist-packages/welly/tools.py in <module> 6 """ 7 import numpy as np ----> 8 import wellpathpy as wp 9 import warnings 10 ModuleNotFoundError: No module named 'wellpathpy'``` @kwinkunks @EvanBianco PR #162 is a fix for this
gharchive/issue
2021-07-13T20:41:25
2025-04-01T04:55:51.486309
{ "authors": [ "astroChance", "kinverarity1" ], "repo": "agile-geoscience/welly", "url": "https://github.com/agile-geoscience/welly/issues/165", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
945563392
DRY project name email handler uses message_details.project_name to get the project name from distutils. we also set the project name as part of the Supervisor object, but there is no connection between these two. either link or remove need. To me it sounds simpler to use the project name from distutils. Could the object sent to supervisor be optional and an override?
gharchive/issue
2021-07-15T16:30:42
2025-04-01T04:55:51.510839
{ "authors": [ "jacobdadams", "steveoh" ], "repo": "agrc/supervisor", "url": "https://github.com/agrc/supervisor/issues/8", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2102889573
Calculation "Time to solution + Progress" Hello, I would like to ask for support to calculate the time for solution + progress that is shown on the GLPI calls page, I still haven't been able to concatenate the data via the database and arrive at the same calculation used by GLPI, what bases and parameters could I use? Note: in my calendar all days and times are useful.
gharchive/issue
2024-01-26T20:55:06
2025-04-01T04:55:51.520407
{ "authors": [ "izaacmenezes" ], "repo": "agrorural/GLPI", "url": "https://github.com/agrorural/GLPI/issues/2", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2132737502
Request: add GNU patch Having patch would be useful on all platforms. https://savannah.gnu.org/projects/patch/ cli.zip https://github.com/ahgamut/superconfigure/releases/tag/z0.0.33 has patch.
gharchive/issue
2024-02-13T16:36:48
2025-04-01T04:55:51.558926
{ "authors": [ "ahgamut", "mattyclarkson" ], "repo": "ahgamut/superconfigure", "url": "https://github.com/ahgamut/superconfigure/issues/25", "license": "Unlicense", "license_type": "permissive", "license_source": "github-api" }
565289031
Typed channels support introduced a new method FromTypedChannel() From() calls FromChannel() if channel is of type chan interface{} and FromTypedChannel for all other channels benchmark for FromTypedChannel() and FromChannel() performance comparison calls FromChannel() if channel is of type chan interface{} but can't all chans be casted to chan interface{} ? calls FromChannel() if channel is of type chan interface{} but can't all chans be casted to chan interface{} ? No, I don't think so. You can think of chan as a type constructor that is parameterized by some inner type and it is invariant. So, chan interface{} and chan int are totally different types. Disclaimer. I am not 100% sure, since I am not a Go developer in real life :) So, there might be a way. @ahmetb I've remembered that we use a naming convention of -T() for the methods that identify user types with reflection, e.g. WhereT(predicateFn interface{}) vs Where(predicate func(interface{}) bool). So, probably the right name for FromTypedChannel() is FromChannelT(). What do you think? Yeah we currently don’t have FromXxxT, but this would reinforce the pattern. FromChannelT is good. Please make sure we explain the difference in its godoc. Yeah we currently don’t have FromXxxT, but this would reinforce the pattern. FromChannelT is good. Please make sure we explain the difference in its godoc. @ahmetb done.
gharchive/pull-request
2020-02-14T12:10:28
2025-04-01T04:55:51.594580
{ "authors": [ "ahmetb", "kalaninja" ], "repo": "ahmetb/go-linq", "url": "https://github.com/ahmetb/go-linq/pull/90", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1343602593
[Idea] Allow configuration of pattern groups through a config file What if there was a file that holds common queries, in a rc file somewhere on the home dir: e.g. context-groups: prod-new: include: "/prod/" exclude: "^/legacy/" stg: include: "/stg/" $ kubectl allctx -g prod -- get nodes # g used as a flag for context group It is a great idea! I will try to implement this feature and raise a PR. No need. This is basically just a note so that I don't forget someday later. It doesn't mean I'm interested in adding this.
gharchive/issue
2022-08-18T20:20:47
2025-04-01T04:55:51.596337
{ "authors": [ "ahmetb", "furkanpehli1" ], "repo": "ahmetb/kubectl-allctx", "url": "https://github.com/ahmetb/kubectl-allctx/issues/6", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
803096234
e621 doesn't work? ahoviewer won't let me add it. That booru doesn't seem to support an XML API, but provides a JSON one. This could be supported with a plugin similar to what I created for Sankaku. Gonna close this, and recommend you open a request issue over at ahoviewer-plugins repo.
gharchive/issue
2021-02-08T00:08:43
2025-04-01T04:55:51.598455
{ "authors": [ "ahodesuka", "blissfulVoid" ], "repo": "ahodesuka/ahoviewer", "url": "https://github.com/ahodesuka/ahoviewer/issues/115", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
379523556
tres compiler error in TDNSResolver::getResponse I'm on Ubuntu 16.04 with "gcc (Ubuntu 5.4.0-6ubuntu1~16.04.10) 5.4.0 20160609". Since 947908cb565cc4fb72517329a23b76ebd15eeeed make tres fails with this: $ nice make tres g++ -std=gnu++14 -Wall -O2 -MMD -MP -ggdb -Iext/simplesocket -Iext/simplesocket/ext/fmt-5.2.1/include -Iext/ -pthread -c -o tres.o tres.cc tres.cc: In member function ‘DNSMessageReader TDNSResolver::getResponse(const ComboAddress&, const DNSName&, const DNSType&, int)’: tres.cc:129:32: error: converting to ‘const key_type {aka const std::tuple<ComboAddress, DNSName, DNSType>}’ from initializer list would use explicit constructor ‘constexpr std::tuple< <template-parameter-1-1> >::tuple(const _Elements& ...) [with _Elements = {ComboAddress, DNSName, DNSType}]’ if(skips.count({server,dn,dt})) { ^ tres.cc:185:36: error: converting to ‘std::set<std::tuple<ComboAddress, DNSName, DNSType> >::value_type {aka std::tuple<ComboAddress, DNSName, DNSType>}’ from initializer list would use explicit constructor ‘constexpr std::tuple< <template-parameter-1-1> >::tuple(const _Elements& ...) [with _Elements = {ComboAddress, DNSName, DNSType}]’ skips.insert({server,dn,dt}); ^ <builtin>: recipe for target 'tres.o' failed make: *** [tres.o] Error 1 I will address this - some c++17-isms have crept in. For now, replace {server, dn, dt} by std::tie(server, dn, dt) since travis can compile tdns again, this should be resolved. Can you check? Hmm. It compiles, but the output is mostly on one line. And it segfaults at the end. $ ./tres mattnordhoff.net A Retrieved . NSSET from hints, have 26 addresses Result of query for mattnordhoff.net.|A (0 intermediate, 2 actual) mattnordhoff.net. 300 A 104.18.56.175 mattnordhoff.net. 300 A 104.18.57.175 Used 6 queries sh: 1: dot: not found {"answer":[{"content":"104.18.56.175","name":"mattnordhoff.net.","ttl":300,"type":"A"},{"content":"104.18.57.175","name":"mattnordhoff.net.","ttl":300,"type":"A"}],"dot":"","intermediate":[],"msec":143,"name":"mattnordhoff.net.","numformerrs":0,"numqueries":6,"numtimeouts":0,"rcode":0,"trace":"mattnordhoff.net.|A Starting query at authority = ., have 26 addresses to try\nmattnordhoff.net.|A Sending to server k.root-servers.net. on 2001:7fd::1\nmattnordhoff.net.|A Received a 838 byte response with RCode Noerror, qname mattnordhoff.net., qtype A, aa: 0\nmattnordhoff.net.|A Authority net. IN NS 172800 c.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 m.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 d.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 f.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 k.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 e.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 a.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 i.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 h.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 g.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 l.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 b.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 j.gtld-servers.net.\nmattnordhoff.net.|A Additional a.gtld-servers.net. IN AAAA 172800 2001:503:a83e::2:30\nmattnordhoff.net.|A Additional b.gtld-servers.net. IN AAAA 172800 2001:503:231d::2:30\nmattnordhoff.net.|A Additional c.gtld-servers.net. IN AAAA 172800 2001:503:83eb::30\nmattnordhoff.net.|A Additional d.gtld-servers.net. IN AAAA 172800 2001:500:856e::30\nmattnordhoff.net.|A Additional e.gtld-servers.net. IN AAAA 172800 2001:502:1ca1::30\nmattnordhoff.net.|A Additional f.gtld-servers.net. IN AAAA 172800 2001:503:d414::30\nmattnordhoff.net.|A Additional g.gtld-servers.net. IN AAAA 172800 2001:503:eea3::30\nmattnordhoff.net.|A Additional h.gtld-servers.net. IN AAAA 172800 2001:502:8cc::30\nmattnordhoff.net.|A Additional i.gtld-servers.net. IN AAAA 172800 2001:503:39c1::30\nmattnordhoff.net.|A Additional j.gtld-servers.net. IN AAAA 172800 2001:502:7094::30\nmattnordhoff.net.|A Additional k.gtld-servers.net. IN AAAA 172800 2001:503:d2d::30\nmattnordhoff.net.|A Additional l.gtld-servers.net. IN AAAA 172800 2001:500:d937::30\nmattnordhoff.net.|A Additional m.gtld-servers.net. IN AAAA 172800 2001:501:b1f9::30\nmattnordhoff.net.|A Additional a.gtld-servers.net. IN A 172800 192.5.6.30\nmattnordhoff.net.|A Additional b.gtld-servers.net. IN A 172800 192.33.14.30\nmattnordhoff.net.|A Additional c.gtld-servers.net. IN A 172800 192.26.92.30\nmattnordhoff.net.|A Additional d.gtld-servers.net. IN A 172800 192.31.80.30\nmattnordhoff.net.|A Additional e.gtld-servers.net. IN A 172800 192.12.94.30\nmattnordhoff.net.|A Additional f.gtld-servers.net. IN A 172800 192.35.51.30\nmattnordhoff.net.|A Additional g.gtld-servers.net. IN A 172800 192.42.93.30\nmattnordhoff.net.|A Additional h.gtld-servers.net. IN A 172800 192.54.112.30\nmattnordhoff.net.|A Additional i.gtld-servers.net. IN A 172800 192.43.172.30\nmattnordhoff.net.|A Additional j.gtld-servers.net. IN A 172800 192.48.79.30\nmattnordhoff.net.|A Additional k.gtld-servers.net. IN A 172800 192.52.178.30\nmattnordhoff.net.|A Additional l.gtld-servers.net. IN A 172800 192.41.162.30\nmattnordhoff.net.|A Additional m.gtld-servers.net. IN A 172800 192.55.83.30\nmattnordhoff.net.|A Additional . IN OPT 0 \\# 0\nmattnordhoff.net.|A We got delegated to 13 net. nameserver names \nmattnordhoff.net.|A Have 26 IP addresses to iterate to: a.gtld-servers.net.=2001:503:a83e::2:30 a.gtld-servers.net.=192.5.6.30 b.gtld-servers.net.=2001:503:231d::2:30 b.gtld-servers.net.=192.33.14.30 c.gtld-servers.net.=2001:503:83eb::30 c.gtld-servers.net.=192.26.92.30 d.gtld-servers.net.=2001:500:856e::30 d.gtld-servers.net.=192.31.80.30 e.gtld-servers.net.=2001:502:1ca1::30 e.gtld-servers.net.=192.12.94.30 f.gtld-servers.net.=2001:503:d414::30 f.gtld-servers.net.=192.35.51.30 g.gtld-servers.net.=2001:503:eea3::30 g.gtld-servers.net.=192.42.93.30 h.gtld-servers.net.=2001:502:8cc::30 h.gtld-servers.net.=192.54.112.30 i.gtld-servers.net.=2001:503:39c1::30 i.gtld-servers.net.=192.43.172.30 j.gtld-servers.net.=2001:502:7094::30 j.gtld-servers.net.=192.48.79.30 k.gtld-servers.net.=2001:503:d2d::30 k.gtld-servers.net.=192.52.178.30 l.gtld-servers.net.=2001:500:d937::30 l.gtld-servers.net.=192.41.162.30 m.gtld-servers.net.=2001:501:b1f9::30 m.gtld-servers.net.=192.55.83.30 \n mattnordhoff.net.|A Starting query at authority = net., have 26 addresses to try\n mattnordhoff.net.|A Sending to server b.gtld-servers.net. on 192.33.14.30\n mattnordhoff.net.|A Received a 188 byte response with RCode Noerror, qname mattnordhoff.net., qtype A, aa: 0\n mattnordhoff.net.|A Authority mattnordhoff.net. IN NS 172800 cody.ns.cloudflare.com.\n mattnordhoff.net.|A Authority mattnordhoff.net. IN NS 172800 dana.ns.cloudflare.com.\n mattnordhoff.net.|A Additional cody.ns.cloudflare.com. IN A 172800 173.245.59.107\n mattnordhoff.net.|A Not accepting IP address of cody.ns.cloudflare.com.: out of authority of this server\n mattnordhoff.net.|A Additional cody.ns.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::adf5:3b6b\n mattnordhoff.net.|A Not accepting IP address of cody.ns.cloudflare.com.: out of authority of this server\n mattnordhoff.net.|A Additional dana.ns.cloudflare.com. IN A 172800 173.245.58.105\n mattnordhoff.net.|A Not accepting IP address of dana.ns.cloudflare.com.: out of authority of this server\n mattnordhoff.net.|A Additional dana.ns.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::adf5:3a69\n mattnordhoff.net.|A Not accepting IP address of dana.ns.cloudflare.com.: out of authority of this server\n mattnordhoff.net.|A Additional . IN OPT 0 \\# 0\n mattnordhoff.net.|A We got delegated to 2 mattnordhoff.net. nameserver names \n mattnordhoff.net.|A Don't have a resolved nameserver to ask anymore, trying to resolve 2 names\n mattnordhoff.net.|A Attempting to resolve NS dana.ns.cloudflare.com.|A\n dana.ns.cloudflare.com.|A Starting query at authority = ., have 26 addresses to try\n dana.ns.cloudflare.com.|A Sending to server f.root-servers.net. on 2001:500:2f::f\n dana.ns.cloudflare.com.|A Received a 847 byte response with RCode Noerror, qname dana.ns.cloudflare.com., qtype A, aa: 0\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 l.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 b.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 c.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 d.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 e.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 f.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 g.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 a.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 h.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 i.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 j.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 k.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 m.gtld-servers.net.\n dana.ns.cloudflare.com.|A Additional l.gtld-servers.net. IN A 172800 192.41.162.30\n dana.ns.cloudflare.com.|A Additional l.gtld-servers.net. IN AAAA 172800 2001:500:d937::30\n dana.ns.cloudflare.com.|A Additional b.gtld-servers.net. IN A 172800 192.33.14.30\n dana.ns.cloudflare.com.|A Additional b.gtld-servers.net. IN AAAA 172800 2001:503:231d::2:30\n dana.ns.cloudflare.com.|A Additional c.gtld-servers.net. IN A 172800 192.26.92.30\n dana.ns.cloudflare.com.|A Additional c.gtld-servers.net. IN AAAA 172800 2001:503:83eb::30\n dana.ns.cloudflare.com.|A Additional d.gtld-servers.net. IN A 172800 192.31.80.30\n dana.ns.cloudflare.com.|A Additional d.gtld-servers.net. IN AAAA 172800 2001:500:856e::30\n dana.ns.cloudflare.com.|A Additional e.gtld-servers.net. IN A 172800 192.12.94.30\n dana.ns.cloudflare.com.|A Additional e.gtld-servers.net. IN AAAA 172800 2001:502:1ca1::30\n dana.ns.cloudflare.com.|A Additional f.gtld-servers.net. IN A 172800 192.35.51.30\n dana.ns.cloudflare.com.|A Additional f.gtld-servers.net. IN AAAA 172800 2001:503:d414::30\n dana.ns.cloudflare.com.|A Additional g.gtld-servers.net. IN A 172800 192.42.93.30\n dana.ns.cloudflare.com.|A Additional g.gtld-servers.net. IN AAAA 172800 2001:503:eea3::30\n dana.ns.cloudflare.com.|A Additional a.gtld-servers.net. IN A 172800 192.5.6.30\n dana.ns.cloudflare.com.|A Additional a.gtld-servers.net. IN AAAA 172800 2001:503:a83e::2:30\n dana.ns.cloudflare.com.|A Additional h.gtld-servers.net. IN A 172800 192.54.112.30\n dana.ns.cloudflare.com.|A Additional h.gtld-servers.net. IN AAAA 172800 2001:502:8cc::30\n dana.ns.cloudflare.com.|A Additional i.gtld-servers.net. IN A 172800 192.43.172.30\n dana.ns.cloudflare.com.|A Additional i.gtld-servers.net. IN AAAA 172800 2001:503:39c1::30\n dana.ns.cloudflare.com.|A Additional j.gtld-servers.net. IN A 172800 192.48.79.30\n dana.ns.cloudflare.com.|A Additional j.gtld-servers.net. IN AAAA 172800 2001:502:7094::30\n dana.ns.cloudflare.com.|A Additional k.gtld-servers.net. IN A 172800 192.52.178.30\n dana.ns.cloudflare.com.|A Additional k.gtld-servers.net. IN AAAA 172800 2001:503:d2d::30\n dana.ns.cloudflare.com.|A Additional m.gtld-servers.net. IN A 172800 192.55.83.30\n dana.ns.cloudflare.com.|A Additional m.gtld-servers.net. IN AAAA 172800 2001:501:b1f9::30\n dana.ns.cloudflare.com.|A Additional . IN OPT 0 \\# 0\n dana.ns.cloudflare.com.|A We got delegated to 13 com. nameserver names \n dana.ns.cloudflare.com.|A Have 26 IP addresses to iterate to: a.gtld-servers.net.=192.5.6.30 a.gtld-servers.net.=2001:503:a83e::2:30 b.gtld-servers.net.=192.33.14.30 b.gtld-servers.net.=2001:503:231d::2:30 c.gtld-servers.net.=192.26.92.30 c.gtld-servers.net.=2001:503:83eb::30 d.gtld-servers.net.=192.31.80.30 d.gtld-servers.net.=2001:500:856e::30 e.gtld-servers.net.=192.12.94.30 e.gtld-servers.net.=2001:502:1ca1::30 f.gtld-servers.net.=192.35.51.30 f.gtld-servers.net.=2001:503:d414::30 g.gtld-servers.net.=192.42.93.30 g.gtld-servers.net.=2001:503:eea3::30 h.gtld-servers.net.=192.54.112.30 h.gtld-servers.net.=2001:502:8cc::30 i.gtld-servers.net.=192.43.172.30 i.gtld-servers.net.=2001:503:39c1::30 j.gtld-servers.net.=192.48.79.30 j.gtld-servers.net.=2001:502:7094::30 k.gtld-servers.net.=192.52.178.30 k.gtld-servers.net.=2001:503:d2d::30 l.gtld-servers.net.=192.41.162.30 l.gtld-servers.net.=2001:500:d937::30 m.gtld-servers.net.=192.55.83.30 m.gtld-servers.net.=2001:501:b1f9::30 \n dana.ns.cloudflare.com.|A Starting query at authority = com., have 26 addresses to try\n dana.ns.cloudflare.com.|A Sending to server b.gtld-servers.net. on 2001:503:231d::2:30\n dana.ns.cloudflare.com.|A Received a 581 byte response with RCode Noerror, qname dana.ns.cloudflare.com., qtype A, aa: 0\n dana.ns.cloudflare.com.|A Authority cloudflare.com. IN NS 172800 ns3.cloudflare.com.\n dana.ns.cloudflare.com.|A Authority cloudflare.com. IN NS 172800 ns5.cloudflare.com.\n dana.ns.cloudflare.com.|A Authority cloudflare.com. IN NS 172800 ns4.cloudflare.com.\n dana.ns.cloudflare.com.|A Authority cloudflare.com. IN NS 172800 ns6.cloudflare.com.\n dana.ns.cloudflare.com.|A Authority cloudflare.com. IN NS 172800 ns7.cloudflare.com.\n dana.ns.cloudflare.com.|A Additional ns3.cloudflare.com. IN A 172800 162.159.0.33\n dana.ns.cloudflare.com.|A Additional ns3.cloudflare.com. IN A 172800 162.159.7.226\n dana.ns.cloudflare.com.|A Additional ns3.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:21\n dana.ns.cloudflare.com.|A Additional ns3.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:7e2\n dana.ns.cloudflare.com.|A Additional ns5.cloudflare.com. IN A 172800 162.159.2.9\n dana.ns.cloudflare.com.|A Additional ns5.cloudflare.com. IN A 172800 162.159.9.55\n dana.ns.cloudflare.com.|A Additional ns5.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:209\n dana.ns.cloudflare.com.|A Additional ns5.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:937\n dana.ns.cloudflare.com.|A Additional ns4.cloudflare.com. IN A 172800 162.159.1.33\n dana.ns.cloudflare.com.|A Additional ns4.cloudflare.com. IN A 172800 162.159.8.55\n dana.ns.cloudflare.com.|A Additional ns4.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:121\n dana.ns.cloudflare.com.|A Additional ns4.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:837\n dana.ns.cloudflare.com.|A Additional ns6.cloudflare.com. IN A 172800 162.159.3.11\n dana.ns.cloudflare.com.|A Additional ns6.cloudflare.com. IN A 172800 162.159.5.6\n dana.ns.cloudflare.com.|A Additional ns6.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:30b\n dana.ns.cloudflare.com.|A Additional ns6.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:506\n dana.ns.cloudflare.com.|A Additional ns7.cloudflare.com. IN A 172800 162.159.4.8\n dana.ns.cloudflare.com.|A Additional ns7.cloudflare.com. IN A 172800 162.159.6.6\n dana.ns.cloudflare.com.|A Additional ns7.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:408\n dana.ns.cloudflare.com.|A Additional ns7.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:606\n dana.ns.cloudflare.com.|A Additional . IN OPT 0 \\# 0\n dana.ns.cloudflare.com.|A We got delegated to 5 cloudflare.com. nameserver names \n dana.ns.cloudflare.com.|A Have 20 IP addresses to iterate to: ns3.cloudflare.com.=162.159.0.33 ns3.cloudflare.com.=162.159.7.226 ns3.cloudflare.com.=2400:cb00:2049:1::a29f:21 ns3.cloudflare.com.=2400:cb00:2049:1::a29f:7e2 ns4.cloudflare.com.=162.159.1.33 ns4.cloudflare.com.=162.159.8.55 ns4.cloudflare.com.=2400:cb00:2049:1::a29f:121 ns4.cloudflare.com.=2400:cb00:2049:1::a29f:837 ns5.cloudflare.com.=162.159.2.9 ns5.cloudflare.com.=162.159.9.55 ns5.cloudflare.com.=2400:cb00:2049:1::a29f:209 ns5.cloudflare.com.=2400:cb00:2049:1::a29f:937 ns6.cloudflare.com.=162.159.3.11 ns6.cloudflare.com.=162.159.5.6 ns6.cloudflare.com.=2400:cb00:2049:1::a29f:30b ns6.cloudflare.com.=2400:cb00:2049:1::a29f:506 ns7.cloudflare.com.=162.159.4.8 ns7.cloudflare.com.=162.159.6.6 ns7.cloudflare.com.=2400:cb00:2049:1::a29f:408 ns7.cloudflare.com.=2400:cb00:2049:1::a29f:606 \n dana.ns.cloudflare.com.|A Starting query at authority = cloudflare.com., have 20 addresses to try\n dana.ns.cloudflare.com.|A Sending to server ns6.cloudflare.com. on 2400:cb00:2049:1::a29f:506\n dana.ns.cloudflare.com.|A Received a 67 byte response with RCode Noerror, qname dana.ns.cloudflare.com., qtype A, aa: 1\n dana.ns.cloudflare.com.|A Answer says it is authoritative!\n dana.ns.cloudflare.com.|A Answer dana.ns.cloudflare.com. IN A 86400 173.245.58.105\n dana.ns.cloudflare.com.|A We got an answer to our question!\n dana.ns.cloudflare.com.|A Additional . IN OPT 0 \\# 0\n dana.ns.cloudflare.com.|A Done, returning 1 results, 0 intermediate\n mattnordhoff.net.|A Got 1 nameserver A addresses, adding to list\n mattnordhoff.net.|A We now have 1 resolved A addresses to try\n mattnordhoff.net.|A Starting query at authority = mattnordhoff.net., have 1 addresses to try\n mattnordhoff.net.|A Sending to server dana.ns.cloudflare.com. on 173.245.58.105\n mattnordhoff.net.|A Received a 77 byte response with RCode Noerror, qname mattnordhoff.net., qtype A, aa: 1\n mattnordhoff.net.|A Answer says it is authoritative!\n mattnordhoff.net.|A Answer mattnordhoff.net. IN A 300 104.18.56.175\n mattnordhoff.net.|A We got an answer to our question!\n mattnordhoff.net.|A Answer mattnordhoff.net. IN A 300 104.18.57.175\n mattnordhoff.net.|A We got an answer to our question!\n mattnordhoff.net.|A Additional . IN OPT 0 \\# 0\n mattnordhoff.net.|A Done, returning 2 results, 0 intermediate\n","type":"A"} Segmentation fault (this issue was fixed in e3891625e4c98c00cad97660a391104c73a767f8 )
gharchive/issue
2018-11-11T13:38:21
2025-04-01T04:55:51.608430
{ "authors": [ "ahupowerdns", "mnordhoff" ], "repo": "ahupowerdns/hello-dns", "url": "https://github.com/ahupowerdns/hello-dns/issues/38", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1330611115
Fix bug in shell meridional transect Short description of why the PR is needed and how it satisfies those requirements, in sentence form. (Delete unused sections) Added public API: symbol (e.g. vcm.my_function) or script and optional description of changes or why they are needed Can group multiple related symbols on a single bullet Refactored public API: Bulleted list of removed or refactored symbols, such as changes to name, type, behavior, argument, etc. Be cautious about doing these and discuss with team more broadly. Significant internal changes: Bulleted list of changes to non-public API Requirement changes: Bulleted list, if relevant, of any changes to setup.py, requirement.txt, environment.yml, etc [ ] Ran make lock_deps/lock_pip following these instructions [ ] Add PR review with license info for any additions to constraints.txt (example) [ ] Tests added Resolves # [JIRA-TAG] Coverage reports (updated automatically): Need to add test
gharchive/pull-request
2022-08-06T02:11:41
2025-04-01T04:55:51.629516
{ "authors": [ "nbren12" ], "repo": "ai2cm/fv3net", "url": "https://github.com/ai2cm/fv3net/pull/1990", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
174928442
Request - beetroot to biofuel (1.10) Beet roots should make perfectly good biofuel, right? Currently fairly useless (unless you like soup) but plentiful from villages. Ditto for beet seeds. Can't remember if I recommended it before but it would be nice to be able to whitelist other mods' items for crushing into biofuel, too. Added, thanks!
gharchive/issue
2016-09-04T02:55:18
2025-04-01T04:55:51.657783
{ "authors": [ "aidancbrady", "rdunham3" ], "repo": "aidancbrady/Mekanism", "url": "https://github.com/aidancbrady/Mekanism/issues/3736", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }