id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
241895779
|
Looping in a state when there's no explicit jump
No more of this nonsense
10:
[...]
=> 10
Agreed
https://github.com/acbabis/factorio-state-machine/pull/11
|
gharchive/issue
| 2017-07-11T01:12:14 |
2025-04-01T04:55:50.129946
|
{
"authors": [
"acbabis",
"jheinemeyer"
],
"repo": "acbabis/factorio-state-machine",
"url": "https://github.com/acbabis/factorio-state-machine/issues/10",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1732793085
|
1721 automate the push from GitHub to gitlab
Description
This pull request adds a GitHub Actions workflows to mirror the dev, staging & production branch and tags to GitLab.
Changes Made
Created a new GitHub Actions workflow file: .github/workflows/mirror-dev.yml.
Created a new GitHub Actions workflow file: .github/workflows/mirror-staging.yml.
Created a new GitHub Actions workflow file: .github/workflows/mirror-production.yml.
Added a workflow to mirror the dev, staging & production branch and tags to GitLab.
Restricted the workflow to run only on the accessibility-exchange/platform repository.
Testing
Tested the workflow on a fork to ensure it successfully mirrors the branches to GitLab without limit for repository.
After adding limit for repository made sure it doesn't work on fork.
Verified that tags are mirrored as well.
Checklist
[x] The workflow file is located in the correct directory: .github/workflows/mirror-staging.yml.
[x] The workflow is correctly restricted to the accessibility-exchange/platform repository.
[ ] Git global user name and email are correctly set.
[ ] The GitLab remote URL is correctly configured.
[ ] SSH key setup is working and allows pushing to GitLab.
[ ] The workflow has been tested and verified for successful execution
@jobara This is pending setup of Secrets and Variables.
|
gharchive/pull-request
| 2023-05-30T18:49:58 |
2025-04-01T04:55:50.149627
|
{
"authors": [
"marvinroman"
],
"repo": "accessibility-exchange/platform",
"url": "https://github.com/accessibility-exchange/platform/pull/1725",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
205141149
|
Investigate regularization for deep models
insensitivity to parameter changes: penalize with the gradient magnitude of the layer's output to its paramters
L2-regularization on weights, but not on biases
|
gharchive/issue
| 2017-02-03T12:06:35 |
2025-04-01T04:55:50.150786
|
{
"authors": [
"accosmin"
],
"repo": "accosmin/nano",
"url": "https://github.com/accosmin/nano/issues/146",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1598498870
|
login test for return from login() method
transkribus_utils.py, line 68 : ACDHTranskribusUtils.list_docs throws exception since requests.get(url, cookies=self.login_cookie) only returns <Response [401]> because server-auth failed. Hence ACDHTranskribusUtils.login() should check if the auth post request via equests.post() was succesfull.
Authentication may fail because transkribus server side auth struggles with certain characters such as "…".
|
gharchive/issue
| 2023-02-24T11:56:19 |
2025-04-01T04:55:50.158030
|
{
"authors": [
"cfhaak"
],
"repo": "acdh-oeaw/acdh-transkribus-utils",
"url": "https://github.com/acdh-oeaw/acdh-transkribus-utils/issues/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1741880358
|
Gregor/refactor/entity table factory
Describe your changes
refactor: tables.py and entities-table-factory.
Overview:
Tables.py
refactor of tables.py file to own python module within apis_entities
refactor of get_entities_tables factory method into factory-class.
factory-class persists already created table-classes, instead of
re-creating them on each request.
inner-generic-table-class removes constraints of former implementation
its the broadest implementation of the table
the calling view is now responsible for customizing the instance of
the table class that will be send to the template, instead of
hardcoding the customization in the class itself.
added utility class get_entities_table,
that uses the new FactoryClass, but catches all
existing calls from other modules, to not break current implementations.
GenericListViewNew in apis_entities.views.py:
table specific code moved into own class / mixin that is now
inherited by the GenericListViewNew
refactored variable names
moved seperate logic into own methods
added docstrings: with context information that is not visible
from the code itself.
tried to name methods in a way that they explain what they do
added inline comments on class vars and instance vars that are
bound to implicit logic that happens in the parent classes - these
should NOT be renamed, or things will break.
Related issues and PRs
#182 - implements factory refactor as an example solution for apis_entities
Resolves issue:
#169
---> sorry, my formatter doesn't work currently and the files are not formatted correctly. <--
Checklist (Replace the space in square brackets with a lowercase x for all that apply)
[x] My changes don't generate new warnings or errors
[ ] My changes follow the project's code formatting rules and style guidelines
[x] I have commented my code with Docstrings and code comments, particularly complex, unusual or hard-to-read code
[ ] I have updated the project documentation to reflect the changes I introduce
[ ] I have added new unit tests or updated existing ones to demonstrate my changes works
Styling the columns (adjusting the width) was not possible atm, this must be done in a separate PR.
headers are now removed for the two edit and detail columns.
edit and detail columns are not sortable and don't participate in sorting order
obviously, the two columns are now also implemented, which resolves #169
Also, there are merge conflics. Please resolve them
You are totally right, I didn't notice that there where unnecessary commits, that weren't meant to be published, in the branches history. Will rework that and remember it in future commits.
Closing again, for new PR with squashed commits and formatted files.
|
gharchive/pull-request
| 2023-06-05T14:12:47 |
2025-04-01T04:55:50.167524
|
{
"authors": [
"b1rger",
"gregorpirgie"
],
"repo": "acdh-oeaw/apis-core-rdf",
"url": "https://github.com/acdh-oeaw/apis-core-rdf/pull/185",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1496183867
|
🛑 AceBlock Swagger UI is down
In 299748f, AceBlock Swagger UI ($ACEBLOCK_SWAGGER_UI) was down:
HTTP code: 0
Response time: 0 ms
Resolved: AceBlock Swagger UI is back up in f17f113.
|
gharchive/issue
| 2022-12-14T09:38:53 |
2025-04-01T04:55:50.171991
|
{
"authors": [
"aceblockID"
],
"repo": "aceblockID/monitoring",
"url": "https://github.com/aceblockID/monitoring/issues/4241",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1203115444
|
🛑 AceBlock Swagger UI is down
In b1be53e, AceBlock Swagger UI ($ACEBLOCK_SWAGGER_UI) was down:
HTTP code: 0
Response time: 0 ms
Resolved: AceBlock Swagger UI is back up in b95d99a.
|
gharchive/issue
| 2022-04-13T10:35:54 |
2025-04-01T04:55:50.174325
|
{
"authors": [
"aceblockID"
],
"repo": "aceblockID/monitoring",
"url": "https://github.com/aceblockID/monitoring/issues/445",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1556259749
|
🛑 AceBlock Swagger UI is down
In b2528b2, AceBlock Swagger UI ($ACEBLOCK_SWAGGER_UI) was down:
HTTP code: 0
Response time: 0 ms
Resolved: AceBlock Swagger UI is back up in e7c8575.
|
gharchive/issue
| 2023-01-25T08:43:47 |
2025-04-01T04:55:50.176398
|
{
"authors": [
"aceblockID"
],
"repo": "aceblockID/monitoring",
"url": "https://github.com/aceblockID/monitoring/issues/5150",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1576686925
|
twitter share link'i çalışmıyor
ana sayfadaki paylaş kısmındaki twitter share çalışmıyor.
@mthocur bakabilir misin
buraya bi ss atabilir misiniz @ondokuzz ben yardımcı olayım müsaitim @karasakalmt
@hsyngul bu repository de henuz paylasma secenekleri yok musaitseniz paylasma seceneklerini ekleyebilirsiniz size assign edebilirim
@karasakalmt assign edebilirsiniz hocam. Yalnız bi introya ihtiyacım olacak ilgili discord kanalına alırsanız çok sevinirim
@hsyngul karasakalmt#1437 ulasabilirsiniz
twitter kapatıldığı için çalışmıyomuş. vpn ile girince çalıştı.
issue'yu kapatabiliriz.
@ondokuzz Huseyin Gul#0295 discord dan ekleyebilirseniz diğer issue lar için yardımcı olabiliriz
ben de ekli değilim. ekleyen olursa beni de eklesin. murat#1927
bu bu repoya ait degil bizim eklememiz gerekiyor @hsyngul ilgileniyor
https://github.com/acikkaynak/deprem-io-frontend/pull/11 @karasakalmt issue yu kapatabiliriz
|
gharchive/issue
| 2023-02-08T19:28:25 |
2025-04-01T04:55:50.227641
|
{
"authors": [
"hsyngul",
"karasakalmt",
"ondokuzz"
],
"repo": "acikkaynak/deprem-io-frontend",
"url": "https://github.com/acikkaynak/deprem-io-frontend/issues/4",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1575832801
|
feat: seo meta tagler için yeni yapı oluşturulması
İsteğiniz şeyi tanımlayın
içerisinde bulunan meta taglerin dinamik hale getirilmesi
sosyal mecralarda görünürlüğü arttırmak için paylaşım spesifik resimler üretilebilmesi
üretilen resimlerin aynı şekilde image proplarının karşılığında gösterilmesinin sağlanması
Değerlendirdiğiniz alternatifleri tanımlayın
resim üretimi için gönderilen parametrelere göre sadece image outputu verebilecek bir edge function yaratılması
Ek Bağlam
Hedef görüntü olarak tekil bir kaydın sosyal mecralarda paylaşımı sonucu aşağıdaki gibi görünüm sağlanabilmesi
thank you
@dopplerDistortion tekrar açtım issue'yu
tekil kayit icin bir sey tutuyor muyuz? herhangi bir sey yapildi mi orayla alakli?
@eraygundogmus dc
@dopplerDistortion bu iş tamamlandı mı? closed görüyorum ama in progress'te, teyit etmek istiyorum.
|
gharchive/issue
| 2023-02-08T10:06:23 |
2025-04-01T04:55:50.231474
|
{
"authors": [
"berkayvuran",
"dopplerDistortion",
"eraygundogmus",
"fdemir"
],
"repo": "acikkaynak/deprem-yardim-frontend",
"url": "https://github.com/acikkaynak/deprem-yardim-frontend/issues/435",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
726298684
|
Eigen values
Find the eigen values of a given matrix.
Hint: If no function is present for determinant computation create one.
I want to do this
|
gharchive/issue
| 2020-10-21T09:20:33 |
2025-04-01T04:55:50.251709
|
{
"authors": [
"anushka-gupta-31",
"shreya1010svg"
],
"repo": "acmpesuecc/laalibrary",
"url": "https://github.com/acmpesuecc/laalibrary/issues/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1398287876
|
Expose Pod information to Containers through Environment Variables
Description
There may be cases where we'd like containers to have access to pod information through environment variables.
Some example pod information that can be pulled through the downward api is the fieldRef and resourceFieldRef.
A short list of available fields
field
description
metadata.name
the pod's name
metadata.namespace
the pod's namespace
metadata.uid
the pod's unique ID
metadata.annotations['<KEY>']
the value of the pod's annotation named (for example, metadata.annotations['myannotation'])
metadata.labels['<KEY>']
the text value of the pod's label named (for example, metadata.labels['mylabel'])
spec.serviceAccountName
the name of the pod's service account
spec.nodeName
the name of the node where the Pod is executing
status.hostIP
the primary IP address of the node to which the Pod is assigned
status.podIP
the pod's primary IP address (usually, its IPv4 address)
Resources
Kubernetes Downward API: https://kubernetes.io/docs/concepts/workloads/pods/downward-api/
Example Kubernetes downward API usage: https://kubernetes.io/docs/tasks/inject-data-application/environment-variable-expose-pod-information/
Previous discussion in Acorn Slack: https://acorn-users.slack.com/archives/C03R9ME0SKC/p1663138625938179
Closing in favor or #726
|
gharchive/issue
| 2022-10-05T19:44:30 |
2025-04-01T04:55:50.266885
|
{
"authors": [
"djcarpe"
],
"repo": "acorn-io/acorn",
"url": "https://github.com/acorn-io/acorn/issues/746",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1661136327
|
add regions to project output
Checklist
[ ] The title of this PR would make a good line in Acorn's Release Note's Changelog
[ ] The title of this PR ends with a link to the main issue being address in parentheses, like: This is a title (#1216). Here's an example
[ ] All relevant issues are referenced in the PR description. NOTE: don't use GitHub keywords that auto-close issues
[ ] Commits follow contributing guidance
[ ] Automated tests added to cover the changes. If tests couldn't be added, an explanation is provided in the Verification and Testing section
[ ] Changes to user-facing functionality, API, CLI, and upgrade impacts are clearly called out in PR description
[ ] PR has at least two approvals before merging (or a reasonable exception, like it's just a docs change)
We just have a single column "Regions" with the default region having an * on it?
currently it looks like this
NAME DEFAULT DEFAULT REGION SUPPORTED REGIONS
acorn * local [local]
hub.dev.acornlabs.com/jsilverio22/acorn sandbox [sandbox]
hub.dev.acornlabs.com/jsilverio22/test sandbox [sandbox]
hub.dev.acornlabs.com/jsilverio22/test-region sandbox [sandbox]
hub.stg.acornlabs.com/jsilverio22/acorn sandbox [sandbox]
is the single column w/ * the desired? @ibuildthecloud
I'd prefer just * and not (*). Also for the comma separated list don't put spaces after the ,.
|
gharchive/pull-request
| 2023-04-10T18:00:51 |
2025-04-01T04:55:50.271602
|
{
"authors": [
"ibuildthecloud",
"jsilverio22"
],
"repo": "acorn-io/acorn",
"url": "https://github.com/acorn-io/acorn/pull/1465",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1825171795
|
Panic when deploying apps with images in autoupgrade format without registry.
acorn version v0.8.0-alpha7-172-gff1e20f9+ff1e20f9
acorn run -n mytestnew sangeetha/myfirstacorn:v#.#.#
panic: interface conversion: error is *errors.StatusError, not *imageallowrules.ErrImageNotAllowed
goroutine 1 [running]:
github.com/acorn-io/runtime/pkg/cli.(*Run).Run(0xc00081d880, 0xc0003f6300, {0xc0008f0330, 0x1, 0x3})
github.com/acorn-io/runtime/pkg/cli/run.go:289 +0xd06
github.com/acorn-io/runtime/pkg/cli/builder.bind.func1(0xc0003f6300?, {0xc0008f0330, 0x1, 0x3})
github.com/acorn-io/runtime/pkg/cli/builder/builder.go:431 +0x214
github.com/spf13/cobra.(*Command).execute(0xc0003f6300, {0xc0008f02d0, 0x3, 0x3})
github.com/spf13/cobra@v1.7.0/command.go:940 +0x862
github.com/spf13/cobra.(*Command).ExecuteC(0xc0001af200)
github.com/spf13/cobra@v1.7.0/command.go:1068 +0x3bd
github.com/spf13/cobra.(*Command).Execute(...)
github.com/spf13/cobra@v1.7.0/command.go:992
github.com/spf13/cobra.(*Command).ExecuteContext(...)
github.com/spf13/cobra@v1.7.0/command.go:985
github.com/acorn-io/runtime/pkg/cli.RunAndHandleError({0x48611f8?, 0xc00013fa90?}, 0xc0001af200?)
github.com/acorn-io/runtime/pkg/cli/errors.go:17 +0x5d
main.main()
github.com/acorn-io/runtime/main.go:15 +0x2f
Testing with 2bb41677, I get
acorn run -n mytestnew 'sangeetha/myfirstacorn:v#.#.#'
✗ ERROR: unable to find an image for sangeetha/myfirstacorn:v#.#.# matching pattern v#.#.# - if you are trying to use a remote image, specify the full registry
I assume that this was fixed along the way.
Tested with acorn version v0.8.0-rc4+b6a95e8a
User is presented with following error message as expected with/without have image allow rules enabled.
acorn run -n mytestnew 'sangeetha/myfirstacorn:v#.#.#'
✗ ERROR: unable to find an image for sangeetha/myfirstacorn:v#.#.# matching pattern v#.#.# - if you are trying to use a remote image, specify the full registry
|
gharchive/issue
| 2023-07-27T20:32:53 |
2025-04-01T04:55:50.274484
|
{
"authors": [
"iwilltry42",
"sangee2004"
],
"repo": "acorn-io/runtime",
"url": "https://github.com/acorn-io/runtime/issues/1989",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
346638814
|
how work
tell me how work !
xD
|
gharchive/issue
| 2018-08-01T15:00:06 |
2025-04-01T04:55:50.320100
|
{
"authors": [
"CarlosCenturion",
"sagedAlAbadla"
],
"repo": "acrosa/bot",
"url": "https://github.com/acrosa/bot/issues/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1178192479
|
fix(ingress): fix redirect by annotation on k8s >= 1.19
The API version change to the ingresses are broken when used with ports that are not numbers. E.g. ALB ingress controller annotations.
Refs: https://github.com/acryldata/datahub-helm/pull/98/files/20521ec9661db7afdb3f9b96dbe53181c62947ad#r831117346
This is my relevant configuration, that worked before on a v1.21 cluster:
ingress = {
enabled = true
hosts = [{
host = local.datahub_fqdn
paths = ["/*"]
redirectPaths = [{
path = "/*"
name = "ssl-redirect"
port = "use-annotation" // <- this is not a number!
}]
}]
annotations = {
"kubernetes.io/ingress.class" = "alb"
"alb.ingress.kubernetes.io/target-type" = "ip"
"alb.ingress.kubernetes.io/group.name" = "internal"
"alb.ingress.kubernetes.io/listen-ports" = jsonencode([{ HTTP = 80 }, { HTTPS = 443 }])
"alb.ingress.kubernetes.io/actions.ssl-redirect" = jsonencode({
Type = "redirect",
RedirectConfig = {
Protocol = "HTTPS",
Port = "443",
StatusCode = "HTTP_301"
}
})
}
}
@dexter-mh-lee done.
@dexter-mh-lee looks like 0.2.64 was already released a couple of hours ago. how do we want to fix it? open a new PR just with bumping the version?
Ah. could you do that please? Merged without waiting for the actions.
|
gharchive/pull-request
| 2022-03-23T14:20:26 |
2025-04-01T04:55:50.323119
|
{
"authors": [
"dexter-mh-lee",
"felixb"
],
"repo": "acryldata/datahub-helm",
"url": "https://github.com/acryldata/datahub-helm/pull/100",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
522490053
|
Routines API?
Hello,
I'm trying to build an app that reminds users to take their medication at a specific time or Routine. Is there a way to create a routine from my custom app without the user leaving the app and creating the routine outside of it?
Hi @Ksaboor,
Are you referring to a mobile app (iOS/Android), a web app, or a conversational Action for the Google Assistant?
Actions on Google currently supports adding your conversational Action to a Google Assistant Routine using Routine suggestions. (See here for more detail.)
Hello @taycaldwell, My apologies I meant a custom google home assistant app.
|
gharchive/issue
| 2019-11-13T21:43:56 |
2025-04-01T04:55:50.374024
|
{
"authors": [
"Ksaboor",
"taycaldwell"
],
"repo": "actions-on-google/actions-on-google-nodejs",
"url": "https://github.com/actions-on-google/actions-on-google-nodejs/issues/369",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
392513798
|
Actions on Console: Need an information..How to get the Account linking logs..
Hi.
Is there any way to check the logs (normal case without error), while i do the account linking?
I have checked in LogsViewer i could not get the response from oAuth. For my account linking i have used the amazon oAuth server . For my analysis, i need the logs received from oAuth. Is there any way?
If you're using a service like Amazon's for OAuth, you would need to check with them for logging.
Hi Fleker, The basic reason for asking the same is, i need to get the agentuserid. Currently in a scenario, i have completed the OAuth - Account Linking using amazon. But during the SYNC response i could not find agentuserid, as smarthome receives only intent and other parameter. Let me know if you have inputs...
The agentUserId is a value that you provide, it's not given to you.
Yes, i have learned it. Using the OAuth - Access token, i need to get an AgentUserid (unique id) from OAuth Provider. I have found that from Amazon API gateway to my handler, the Authorisation headers are removed. I am checking with Amazon. Thanks a lot for your response.
|
gharchive/issue
| 2018-12-19T09:36:11 |
2025-04-01T04:55:50.376416
|
{
"authors": [
"Fleker",
"sakthisg"
],
"repo": "actions-on-google/smart-home-nodejs",
"url": "https://github.com/actions-on-google/smart-home-nodejs/issues/240",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2665048130
|
MSVC ARM compiler fails CMake configuration
Description
Using the MSVC ARM compiler on the latest windows-2022 image fails CMake configuration
Image: windows-2022
Version: 20241113.3.0
Included Software: https://github.com/actions/runner-images/blob/win22/20241113.3/images/windows/Windows2022-Readme.md
Image Release: https://github.com/actions/runner-images/releases/tag/win22%2F20241113.3
Platforms affected
[ ] Azure DevOps
[X] GitHub Actions - Standard Runners
[ ] GitHub Actions - Larger Runners
Runner images affected
[ ] Ubuntu 20.04
[ ] Ubuntu 22.04
[ ] Ubuntu 24.04
[ ] macOS 12
[ ] macOS 13
[ ] macOS 13 Arm64
[ ] macOS 14
[ ] macOS 14 Arm64
[ ] macOS 15
[ ] macOS 15 Arm64
[ ] Windows Server 2019
[X] Windows Server 2022
Image version and build link
Image: windows-2022
Version: 20241113.3.0
Included Software: https://github.com/actions/runner-images/blob/win22/20241113.3/images/windows/Windows2022-Readme.md
Image Release: https://github.com/actions/runner-images/releases/tag/win22%2F20241113.3
https://github.com/libsdl-org/SDL/actions/runs/11873470201/job/33088578294
Is it regression?
https://github.com/libsdl-org/SDL/actions/runs/11873470201/job/33088868957
Expected behavior
CMake configuration suceeds. kernel32.lib is a library that is expected to be available in any Windows SDK.
-- The C compiler identification is MSVC 19.41.34123.0
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Check for working C compiler: C:/Program Files/Microsoft Visual Studio/2022/Enterprise/VC/Tools/MSVC/14.41.34120/bin/Hostx64/arm/cl.exe - skipped
-- Detecting C compile features
-- Detecting C compile features - done
Actual behavior
-- The C compiler identification is MSVC 19.42.34433.0
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - failed
-- Check for working C compiler: C:/Program Files/Microsoft Visual Studio/2022/Enterprise/VC/Tools/MSVC/14.42.34433/bin/Hostx64/arm/cl.exe
-- Check for working C compiler: C:/Program Files/Microsoft Visual Studio/2022/Enterprise/VC/Tools/MSVC/14.42.34433/bin/Hostx64/arm/cl.exe - broken
CMake Error at C:/Program Files/CMake/share/cmake-3.31/Modules/CMakeTestCCompiler.cmake:67 (message):
The C compiler
"C:/Program Files/Microsoft Visual Studio/2022/Enterprise/VC/Tools/MSVC/14.42.34433/bin/Hostx64/arm/cl.exe"
is not able to compile a simple test program.
It fails with the following output:
Change Dir: 'D:/a/SDL/SDL/build/CMakeFiles/CMakeScratch/TryCompile-rxem71'
Run Build Command(s): D:/a/_temp/ninja-1.12.1-X64/ninja.exe -v cmTC_aa530
[1/2] C:\PROGRA~1\MICROS~2\2022\ENTERP~1\VC\Tools\MSVC\1442~1.344\bin\Hostx64\arm\cl.exe /nologo /DWIN32 /D_WINDOWS /Zi /Ob0 /Od -MTd /showIncludes /FoCMakeFiles\cmTC_aa530.dir\testCCompiler.c.obj /FdCMakeFiles\cmTC_aa530.dir\ /FS -c D:\a\SDL\SDL\build\CMakeFiles\CMakeScratch\TryCompile-rxem71\testCCompiler.c
[2/2] C:\Windows\system32\cmd.exe /C "cd . && "C:\Program Files\CMake\bin\cmake.exe" -E vs_link_exe --msvc-ver=1942 --intdir=CMakeFiles\cmTC_aa530.dir --rc=C:\PROGRA~2\WI3CF2~1\10\bin\100261~1.0\x64\rc.exe --mt=C:\PROGRA~2\WI3CF2~1\10\bin\100261~1.0\x64\mt.exe --manifests -- C:\PROGRA~1\MICROS~2\2022\ENTERP~1\VC\Tools\MSVC\1442~1.344\bin\Hostx64\arm\link.exe /nologo CMakeFiles\cmTC_aa530.dir\testCCompiler.c.obj /out:cmTC_aa530.exe /implib:cmTC_aa530.lib /pdb:cmTC_aa530.pdb /version:0.0 -DEBUG /debug /INCREMENTAL /subsystem:console kernel32.lib user32.lib && cd ."
FAILED: cmTC_aa530.exe
C:\Windows\system32\cmd.exe /C "cd . && "C:\Program Files\CMake\bin\cmake.exe" -E vs_link_exe --msvc-ver=1942 --intdir=CMakeFiles\cmTC_aa530.dir --rc=C:\PROGRA~2\WI3CF2~1\10\bin\100261~1.0\x64\rc.exe --mt=C:\PROGRA~2\WI3CF2~1\10\bin\100261~1.0\x64\mt.exe --manifests -- C:\PROGRA~1\MICROS~2\2022\ENTERP~1\VC\Tools\MSVC\1442~1.344\bin\Hostx64\arm\link.exe /nologo CMakeFiles\cmTC_aa530.dir\testCCompiler.c.obj /out:cmTC_aa530.exe /implib:cmTC_aa530.lib /pdb:cmTC_aa530.pdb /version:0.0 -DEBUG /debug /INCREMENTAL /subsystem:console kernel32.lib user32.lib && cd ."
LINK Pass 1: command "C:\PROGRA~1\MICROS~2\2022\ENTERP~1\VC\Tools\MSVC\1442~1.344\bin\Hostx64\arm\link.exe /nologo CMakeFiles\cmTC_aa530.dir\testCCompiler.c.obj /out:cmTC_aa530.exe /implib:cmTC_aa530.lib /pdb:cmTC_aa530.pdb /version:0.0 -DEBUG /debug /INCREMENTAL /subsystem:console kernel32.lib user32.lib /MANIFEST /MANIFESTFILE:CMakeFiles\cmTC_aa530.dir/intermediate.manifest CMakeFiles\cmTC_aa530.dir/manifest.res" failed (exit code 1104) with the following output:
LINK : fatal error LNK1104: cannot open file 'kernel32.lib'
ninja: build stopped: subcommand failed.
CMake will not be able to correctly generate this project.
Call Stack (most recent call first):
CMakeLists.txt:4 (project)
-- Configuring incomplete, errors occurred!
Repro steps
build:
os: windows-latest
steps:
- use: ilammy/msvc-dev-cmd@v1
with:
arch: x64_arm
- run: |
# download ninja
- run: |
cmake -S some-cmake-project -B build -GNinja
See https://github.com/zufuliu/notepad4/issues/839, Windows SDK 10.0.26100.0 no longer install libs for 32-bit arm.
Following is my workaround (copy libs from Windows SDK 10.0.22621.0). see https://github.com/zufuliu/notepad4/commit/e53e43305e7e0805670cb6d1abc5ef155fb36357:
SET "Win10Lib=C:\Program Files (x86)\Windows Kits\10\Lib"
robocopy "%Win10Lib%\10.0.22621.0\ucrt\arm" "%Win10Lib%\10.0.26100.0\ucrt\arm" /E 1>NUL
robocopy "%Win10Lib%\10.0.22621.0\um\arm" "%Win10Lib%\10.0.26100.0\um\arm" /E 1>NUL
Thanks, I worked around this issue my ci by pinning the Windows SDK version to 10.0.22621.0 (sdk input of ilammy/msvc-dev-cmd)
Is there a public announcement about ARM32 on Windows being deprecated?
Hi @madebr - Thank you for bringing this issue to our attention. We will look into this issue and will update you after investigating.
Meh... this broke ReactOS ARM port build: https://github.com/actions/runner-images/issues/10978#issuecomment-2480636730
(at first I thought this is a problem with 20241113.3.0 image)
I haven't tested yet but explicitly setting the SDK version in your project files (.vcxproj) might be a workaround.
Setting it to 10.0 will make Visual Studio pick the latest installed version.
<PropertyGroup Label="Globals">
<WindowsTargetPlatformVersion>10.0.22621.0</WindowsTargetPlatformVersion>
...
https://learn.microsoft.com/en-us/gaming/gdk/_content/gc/tools-pc/visualstudio/gr-vs-2019-support-notes
Hi @madebr ,
It is belongs to WDK update which is not supported 32 bit as mentioned by @zufuliu .
Thanks you for understanding!
Hence, We're closing this issue .
Please feel free to reach out if you have any concern.
|
gharchive/issue
| 2024-11-16T22:48:55 |
2025-04-01T04:55:50.388843
|
{
"authors": [
"Prabhatkumar59",
"binarymaster",
"kishorekumar-anchala",
"madebr",
"maron2000",
"zufuliu"
],
"repo": "actions/runner-images",
"url": "https://github.com/actions/runner-images/issues/10981",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1523980316
|
docs: update chart to reflect macos-latest
The chart makes it seem like macos-latest is 11 when it's actually 12.
Hello! We update docs manually as soon as it is fully migrated.
|
gharchive/pull-request
| 2023-01-07T16:58:04 |
2025-04-01T04:55:50.390599
|
{
"authors": [
"ckipp01",
"mikhailkoliada"
],
"repo": "actions/runner-images",
"url": "https://github.com/actions/runner-images/pull/6888",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
954566015
|
Builds are failing at Ubuntu-latest based Microsoft hosted pool agents
Description
I have a privately hosted pool (OryxLinux) and I'm trying to move to Microsoft hosted pool where image is ubuntu-latest.As part of the test I create docker container with port mapping and then access those access those container ports from test (simple get http call) to see if they are working. I'm getting following error message and tests are failing left and right.
Did not get the port mapping in expected format.
at Microsoft.Oryx.Tests.Common.EndToEndTestHelper.RunAndAssertAppAsync(String imageName, ITestOutputHelper output, IEnumerable1 volumes, List1 environmentVariables, Int32 port, String link, String runCmd, String[] runArgs, Func2 assertAction, DockerCli dockerCli) in /home/vsts/work/1/s/tests/Oryx.Tests.Common/EndToEndTestHelper.cs:line 228 at Microsoft.Oryx.Tests.Common.EndToEndTestHelper.BuildRunAndAssertAppAsync(ITestOutputHelper output, IEnumerable1 volumes, String buildImage, String buildCmd, String[] buildArgs, String runtimeImageName, List1 environmentVariables, Int32 port, String link, String runCmd, String[] runArgs, Func2 assertAction) in /home/vsts/work/1/s/tests/Oryx.Tests.Common/EndToEndTestHelper.cs:line 182
at Microsoft.Oryx.Integration.Tests.DotNetCoreRuntimeVersion21Tests.CanBuildAndRunApp_WhenRecursiveLookUpIsDisabled_ButProjectSettingIsSupplied() in /home/vsts/work/1/s/tests/Oryx.Integration.Tests/DotNetCore/DotNetCoreRuntimeVersion21Tests.cs:line 683
--- End of stack trace from previous location where exception was thrown ---
Standard Output Messages:
Debugging Information:
StdOut: Found build manifest file at '/oryxtests/a0ec176690a54017897a4006e2bfd16d/oryx-manifest.toml'. Deserializing it...
Build Operation ID: |IM7kOncAsTk=.1ad74fbc_
Agent extension
Before if loop >> DotNet Runtime 2.1.26
DotNet Runtime 2.1.26Writing output script to '/run.sh'
Found startup DLL name from manifest file
Running the command: dotnet "MyWebApp.dll"
Hosting environment: Production
Content root path: /oryxtests/a0ec176690a54017897a4006e2bfd16d
Now listening on: http://[::]:3000
Application started. Press Ctrl+C to shut down.
All these tests are running fine locally as well as in the privately hosted oryxlinux pool. I'm guessing some port restriction is there. Can you please take a look into it?
Here is a sample build
https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=5029380&view=logs&j=72dd53dc-cc7e-5359-419d-190601712c34&t=28831f2f-ff0f-5427-ca18-e78664bf3d01
Virtual environments affected
[ ] Ubuntu 16.04
[ ] Ubuntu 18.04
[X] Ubuntu 20.04
[ ] macOS 10.15
[ ] macOS 11
[ ] Windows Server 2016
[ ] Windows Server 2019
Image version and build link
image: ubuntu-latest
failed build (hosted in azure pipelines) link: https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=5029380&view=results
successful build (hosted in private pool - OryxLinux) link: https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=5029191&view=results
relevant PR: https://github.com/microsoft/Oryx/pull/1064/files
Is it regression?
n/a
Expected behavior
all the tests should run successfully, container ports should be accessible from tests
Actual behavior
all the tests failing with following error message
[xUnit.net 00:15:10.10] Microsoft.Oryx.Integration.Tests.DotNetCoreRuntimeVersion21Tests.CanBuildAndRunApp_WhenRecursiveLookUpIsDisabled_ButProjectSettingIsSupplied [FAIL]
X Microsoft.Oryx.Integration.Tests.DotNetCoreRuntimeVersion21Tests.CanBuildAndRunApp_WhenRecursiveLookUpIsDisabled_ButProjectSettingIsSupplied [29s 753ms]
Error Message:
Did not get the port mapping in expected format. StdOut: 0.0.0.0:49206
:::49206
Repro steps
Not sure if this works .. but the easiest repro would be ...
rerun the failed job of the following build ...
https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=5029380&view=results
Hello @arroyc,
We will check the issue
@arroyc can you please point out the script or yml file where is the step As part of the test I create docker container with port mapping
@Darleev any update?
@dsame thanks for your reply .. i believe this is what you are asking .. this is where we populate docker arguments for port mapping https://github.com/microsoft/Oryx/blob/a389296f4a1a52b0581660c43a2ba641c04b77d8/tests/Oryx.Tests.Common/DockerCli.cs#L355
https://github.com/microsoft/Oryx/blob/a389296f4a1a52b0581660c43a2ba641c04b77d8/tests/Oryx.Tests.Common/DockerCli.cs#L291
This is how we run docker once we populate all the necessary arguments
https://github.com/microsoft/Oryx/blob/a389296f4a1a52b0581660c43a2ba641c04b77d8/tests/Oryx.Tests.Common/DockerCli.cs#L121
https://github.com/microsoft/Oryx/blob/a389296f4a1a52b0581660c43a2ba641c04b77d8/src/BuildScriptGenerator.Common/ProcessHelper.cs#L97
let me know if thats not what you are looking for, I can share other details as you need
@arroyc can you please point out the script or yml file where is the step As part of the test I create docker container with port mapping ?
The root of the problem the docker listen on both IPv4 and IPv6 interfaces, so the detected ports string is
0.0.0.0:49263
:::49263
which cases the assert to fail:
var portMapping = stdOut?.Split(":");
Assert.NotNull(portMapping);
Assert.True(
(portMapping.Length == 2),
In order to provide the fix i need to know the command which runs the docker.
so we are into that famous issue where docker is binding to both ipv4 and ipv6 and we want only ipv4, hmm doing a quick search on internet shows I need to set net.ipv6.bindv6only to 0. Is that the fix?
I was able to find an workaround by installing a stable docker ce
here is the build id https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=5035925&view=logs&s=ff05ad62-bb9a-53b6-ce9f-72f329a63e7c&j=65e01c55-aa95-5d41-4bd2-87f2a532974e
|
gharchive/issue
| 2021-07-28T07:32:35 |
2025-04-01T04:55:50.413219
|
{
"authors": [
"Darleev",
"arroyc",
"dsame"
],
"repo": "actions/virtual-environments",
"url": "https://github.com/actions/virtual-environments/issues/3807",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
859002762
|
WEBGL_compressed_texture_astc error is output to the console
I have tried Rhodonite v0.3.19dev in a Windows environment.
However, it seems that the error of WEBGL_compressed_texture_astc extension is output to the console.
https://cx20.github.io/gltf-test/examples/rhodonite/index.html?category=tutorialModels&model=AlphaBlendModeTest&scale=0.5&type=glTF
rhodonite.min.js:1
The library does not support this environment because the WEBGL_compressed_texture_astc is not available
__getExtension @ rhodonite.min.js:1
p @ rhodonite.min.js:1
addWebGLContext @ rhodonite.min.js:1
setProcessApproachAndCanvas @ rhodonite.min.js:1
load @ index.js:53
async function (async)
load @ index.js:49
load (async)
(anonymous) @ index.js:341
This extension is only available in certain environments. It is better not to make an error because this extension is not available in Windows environments.
┆Issue is synchronized with this Wrike Item
Currently, Those are now warning, not error actually.
But certainly, It looks better to change them to 'info' output than 'warn'.
@cx20 Since 01443f9d48726029abb5d1a69c8d13e71365a2c7, those warning-level messages are now info-level messages.
This change would have made the console much cleaner and improved the psychological safety of users.
|
gharchive/issue
| 2021-04-15T15:33:54 |
2025-04-01T04:55:50.472810
|
{
"authors": [
"cx20",
"emadurandal"
],
"repo": "actnwit/RhodoniteTS",
"url": "https://github.com/actnwit/RhodoniteTS/issues/764",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1607408300
|
Change recursive variable to integer
if self.recursive == "not_recursive":
to 0
Done in aaf1daf3b7e8f164e0caf3b4e13285d515649a83
|
gharchive/issue
| 2023-03-02T20:00:33 |
2025-04-01T04:55:50.489018
|
{
"authors": [
"acturtle",
"zchmielewska"
],
"repo": "acturtle/cashflower",
"url": "https://github.com/acturtle/cashflower/issues/114",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
897494271
|
🛑 Processors is down
In 32f7483, Processors ($STATUS_URL) was down:
HTTP code: 200
Response time: 119 ms
Resolved: Processors is back up in 9a12449.
|
gharchive/issue
| 2021-05-20T21:51:05 |
2025-04-01T04:55:50.491879
|
{
"authors": [
"max-acumen"
],
"repo": "acumenlabs/status-page",
"url": "https://github.com/acumenlabs/status-page/issues/1126",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
897712116
|
⚠️ Fetchers has degraded performance
In 369a37a, Fetchers ($STATUS_URL) experienced degraded performance:
HTTP code: 200
Response time: 388 ms
Resolved: Fetchers performance has improved in 5b7b4fc.
|
gharchive/issue
| 2021-05-21T06:08:03 |
2025-04-01T04:55:50.494325
|
{
"authors": [
"max-acumen"
],
"repo": "acumenlabs/status-page",
"url": "https://github.com/acumenlabs/status-page/issues/1128",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1094700935
|
⚠️ Processors has degraded performance
In 8348101, Processors ($STATUS_URL) experienced degraded performance:
HTTP code: 200
Response time: 44 ms
Resolved: Processors performance has improved in 59e08b6.
|
gharchive/issue
| 2022-01-05T20:03:46 |
2025-04-01T04:55:50.496459
|
{
"authors": [
"max-acumen"
],
"repo": "acumenlabs/status-page",
"url": "https://github.com/acumenlabs/status-page/issues/2211",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1327106427
|
⚠️ Processors has degraded performance
In 26dfb80, Processors ($STATUS_URL) experienced degraded performance:
HTTP code: 200
Response time: 113 ms
Resolved: Processors performance has improved in d3f2ba1.
|
gharchive/issue
| 2022-08-03T11:54:19 |
2025-04-01T04:55:50.498549
|
{
"authors": [
"danielshir"
],
"repo": "acumenlabs/status-page",
"url": "https://github.com/acumenlabs/status-page/issues/3556",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2299141828
|
Add dictionary to run with fuzzers
Relates to: https://github.com/ada-url/ada/pull/658
Is there anything that needs to be done to pass the dict to the fuzzer?
Is there anything that needs to be done to pass the dict to the fuzzer?
From what I was reading here, looks we just need this .options file.
I saw some other examples, just updating the build.sh to copy these files too and we should be fine. But let's see...
https://github.com/google/oss-fuzz/blob/ad88a2e5295d91251d15f8a612758cd9e5ad92db/projects/expat/build.sh
@anonrig would you like to set a max value here?
From examples, expat defined one but other projects were not.
We can do 1024
@anonrig have you noticed any difference since last run? I haven't
@anonrig have you noticed any difference since last run? I haven't
I haven't as well
|
gharchive/pull-request
| 2024-05-16T01:45:53 |
2025-04-01T04:55:50.502961
|
{
"authors": [
"CarlosEduR",
"anonrig"
],
"repo": "ada-url/ada",
"url": "https://github.com/ada-url/ada/pull/660",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
884520738
|
Monospaced blocks are rendered incorrectly in Outlook on the web (OWA)
Monospaced blocks seem to behave differently in the side pane in Office 365 on the web (OWA) than they do if you double click the e-mail so it pops out. It looks fine when you compose the e-mail, but the issue shows up when you read it.
Given the following example text:
hello
```
pre text here
```
bye
The following HTML is generated in the Outlook "side-pane":
<div>
<p style="margin:0 0 1.2em 0;">hello</p>
<pre style="font-size:1em;font-family:Consolas,Inconsolata,Courier,monospace;margin:1.2em 0;line-height:1.2em;"><code style="font-size:0.85em;font-family:Consolas,Inconsolata,Courier,monospace;background-color:#F8F8F8;display:inline;white-space:pre;border-radius:3px;margin:0 0.15em;padding:0.5em 0.7em;border:1px solid #CCCCCC;overflow:auto;">pre text here
</code></pre>
<p style="margin:0 0 1.2em 0;">bye</p>
<div title="MDH:redacted" style="font-size:0em;width:0;height:0;max-width:0;margin:0;padding:0;overflow:hidden;max-height:0;">
</div>
</div>
While this is generated in the popup window:
<div class="x_markdown-here-wrapper" style="">
<p style="margin:0px 0px 1.2em!important">hello</p><pre style="font-size:0.85em; font-family:Consolas,Inconsolata,Courier,monospace; font-size:1em; line-height:1.2em; margin:1.2em 0px"><code style="font-size:0.85em; font-family:Consolas,Inconsolata,Courier,monospace; margin:0px 0.15em; padding:0px 0.3em; white-space:pre-wrap; border:1px solid rgb(234,234,234); background-color:rgb(248,248,248); border-radius:3px; display:inline; white-space:pre; overflow:auto; border-radius:3px; border:1px solid rgb(204,204,204); padding:0.5em 0.7em; display:block!important">pre text here
</code></pre>
<p style="margin:0px 0px 1.2em!important">bye</p>
<div title="MDH:redacted" style="height:0; width:0; max-height:0; max-width:0; overflow:hidden; font-size:0em; padding:0; margin:0"></div>
</div>
This gets extra messy with multiple lines:
Side-pane:
Popup:
I found that I could fix this by editing the theme customization in the add-on preferences and taking out the !important in this line:
display: block !important; /* added to counteract the Yahoo-specific `code` rule; without this, code blocks in Blogger are broken */
I did this with Markdown Here Revival, which I'm now using instead of this original one since it's no longer maintained. I recommend you switch.
This is now fixed in Markdown Here Revival.
|
gharchive/issue
| 2021-05-10T15:52:20 |
2025-04-01T04:55:50.602833
|
{
"authors": [
"lgrn",
"neilmayhew"
],
"repo": "adam-p/markdown-here",
"url": "https://github.com/adam-p/markdown-here/issues/659",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1839535961
|
Add support for 10Mhz clock
First of all thanks for your work on this library!
I was looking to use your library to support the CAN SPI Click board which uses the MCP2515, however I was initially unable to use it as the board uses a 10Mhz clock which is not supported by your library.
This pull request adds support for 10Mhz clocks.
Hello 👋
First of all, thank you for your contribution! I'll review the changes and come back to you early next week.
Did you add the support for 25kbps CAN for other oscillator frequencies as well or just the 10MHz?
Just for the 10Mhz. I didn’t want to risk modifying anything not directly related to my change.On 7 Aug 2023, at 18:48, Piotr Adamczyk @.***> wrote:
Did you add the support for 25kbps CAN for other oscillator frequencies as well or just the 10MHz?
—Reply to this email directly, view it on GitHub, or unsubscribe.You are receiving this because you authored the thread.Message ID: @.***>
|
gharchive/pull-request
| 2023-08-07T14:09:20 |
2025-04-01T04:55:50.628062
|
{
"authors": [
"Stephen-Bath",
"adamczykpiotr"
],
"repo": "adamczykpiotr/pico-mcp2515",
"url": "https://github.com/adamczykpiotr/pico-mcp2515/pull/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1402929749
|
Renaming doesn't work properly.
Updating the anidb library seems working. But when I try to rename my files, they always end up in " [] - -.mkv". I tried all possible combinations from the available tags but none of them seems to work.
Could you provide the exact command you have used? Either it is wrong, or anidb does not return any data.
anidbcli -r -e mkv,mp4 api -u "myuser" -p "mypw" -r "%a_romaji% - %a_kanji% - %a_english% - %ep_no% [%g_name%] (%crc32%)" --state 2 -a "."
For the test I added all kind of % to see which one is working.
Exact same command seems to be working just fine, did you try it with any other release? It's possible that anidb doesn't provide the data for that file (althought it should end up with error).
I tried an anime from the last and one from the current season, both animes should be there as it was added to my anidb library.
I tried to debug the python code, but its a bit difficult as you get banned after trying it 5-6 times.
I am not an expert in python, if you can help me to help debug it for you - i maybe can help and find the issue.
Sorry, I totally forgot about this issue. For debugging/running invoking anidbcli multiple times in a single batch you
can invoke it with parameter -t, anidbcli will then reuse the previous session token (if available) and you will not get
banned.
|
gharchive/issue
| 2022-10-10T10:17:04 |
2025-04-01T04:55:50.631597
|
{
"authors": [
"Xentros",
"adameste"
],
"repo": "adameste/anidbcli",
"url": "https://github.com/adameste/anidbcli/issues/27",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
98599032
|
Possibility to respond to https requests without connecting upstream servers
This enables a caching proxy for offline use, which is no problem with HTTP at the moment.
Additionally HTTPS needs the hostname to create a certificate and must suppress the handshake to upstream without a connection.
@jekh Please, let us discuss this.
UnencryptedUDTChainedProxyTest fails in Travis-CI with Java 8, but works with Java 7 and 6. It works in my environment with Java 8, too:
$ java -version
java version "1.7.0_79"
OpenJDK Runtime Environment (IcedTea 2.5.6) (7u79-2.5.6-1~deb8u1)
OpenJDK 64-Bit Server VM (build 24.79-b02, mixed mode)
$ /usr/lib/jvm/java-8-openjdk-amd64/bin/java -version
openjdk version "1.8.0_45-internal"
OpenJDK Runtime Environment (build 1.8.0_45-internal-b14)
OpenJDK 64-Bit Server VM (build 25.45-b02, mixed mode)
$ uname -sorv
Linux 3.16.0-4-amd64 #1 SMP Debian 3.16.7-ckt11-1+deb8u2 (2015-07-17) GNU/Linux
I have some reservations about this change. It's very specific to one particular use case, a caching proxy for offline use.
I do see the utility of wanting to avoid making connections to the remote server. For non-SSL connections this is already possible using the filters, but for SSL connections it isn't, because the ConnectionFlow always attempts to establish the remote connection. In the long term, I'd like to see a more flexible way of controlling connections, so that users of LittleProxy can customize the connection flow as needed.
For now, I'm not sure that checking remoteAddress.isUnresolved() is the best way to indicate that we should suppress the connection to the remote server. It's overloading the use of the remoteAddress.isUnresolved() status for an unrelated purpose. Can you think of a better way to indicate in the connection flow that we should not connect to the upstream server? Perhaps it could be integrated into one of the return values of the filters?
For what it's worth, the Betamax testing library also has this requirement, as it works on the basis of replaying recorded responses in place of contacting the target server. I may try the options mentioned in #184 to see if they have an effect. I think this is going to become more and more important as websites default to SSL.
@pledbrook You need special branches of LittleProxy and LittleProxy-mitm to enable MITM while offline. At the moment consider to use https://github.com/ganskef/LittleProxy-parent to build both.
I'm closing this PR since it's replaced with #282.
|
gharchive/pull-request
| 2015-08-02T10:07:23 |
2025-04-01T04:55:50.643402
|
{
"authors": [
"ganskef",
"jekh",
"pledbrook"
],
"repo": "adamfisk/LittleProxy",
"url": "https://github.com/adamfisk/LittleProxy/pull/230",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
257670389
|
No exception with corrupt file
Hi,
actualy we decompress our GZIP TAR file in C# with the command line tool "Tar.exe".
We start an process and parse the result of the command line.
I want to change this by unsing Sharpcomress.
We have sereval unit test for the decompression.
When I use Sharpcomeress to decompress one of the unit tests failed.
This test try to extract a corrupt TAR file.
The "Tar.exe" thorws an exception.
Sharpcomress decompress the file without an exception. Is this a known bug?
This is the code I use to decompress the attached file:
public bool Extract(string filename, string extractPath)
{
try
{
using (Stream stream = File.OpenRead(filename))
{
var reader = ReaderFactory.Open(stream);
reader.WriteAllToDirectory(extractPath, new ExtractionOptions { ExtractFullPath = true, Overwrite = true });
}
return true;
}
catch (Exception e)
{
Console.WriteLine(e);
return false;
}
}
Kind regards Mike
20150427_140120_f4_3d_80_00_01_6e.zip
SharpCompress does not do any validation of CRCs that GZip or whatever might have. The extraction will look successful. It's not a bug but the lack of a CRC validation feature.
The TAR.exe say this:
tar -tf 20150427_140120_f4_3d_80_00_01_6e.scd1
Warning: non-standard extension used on a compressed archive.
Exception: System.Exception: header checksum is invalid.
bei Ionic.Tar._internal_ListOrExtract(String archive, Boolean wantExtract)
bei Ionic.TarApp.ListContents()
bei Ionic.TarApp.Main(String[] args)
Is it possible to make this to a feature request?
Tar.zip
It is possible. CRC is already a known missing feature. I'm accepting pull requests if you'd like to do it.
How can I do that?
|
gharchive/issue
| 2017-09-14T10:28:43 |
2025-04-01T04:55:50.650835
|
{
"authors": [
"InBerg",
"adamhathcock"
],
"repo": "adamhathcock/sharpcompress",
"url": "https://github.com/adamhathcock/sharpcompress/issues/297",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
620233780
|
Randomly fadeOut glitch on Component hide
I open a new Issue to divide topics.
Description
When the component hide using internal function and fadeOut animation occurs a random glitch showing the component for just half a second after fadeOut is completed.
Possibile Enhance
I don't know if problem is here, but as the fadeOut is a Promise, personally I'd prefer waiting for both end before setting visible to false.
So, instead writing:
this.modalContainerRef.current.fadeOut(ANIMATION_TIME);
await this.pickerContainerRef.current.fadeOut(ANIMATION_TIME);
this.setState({ visible: false }, () => {
this.cache.purge();
});
You probably could be more consistent and avoid any error writing
await Promise.all([
/** Hide Modal Container */
new Promise((resolve) => {
if (this.modalContainerRef.current?.fadeOut) {
return this.modalContainerRef.current.fadeOut(ANIMATION_TIME).then(resolve);
}
return resolve();
}),
/** Hide Picker Container */
new Promise((resolve) => {
if (this.pickerContainerRef.current?.fadeOut) {
return this.pickerContainerRef.current.fadeOut(ANIMATION_TIME).then(resolve);
}
return resolve();
}),
]);
this.setState({ visible: false }, () => {
this.cache.purge();
});
Hi @marcocavanna. I've been playing around with a few different options today - do you want to see if #16 solves the issues on your end? If it doesn't, I might need a little more info from you - because I can't seem to replicate it anymore.
Note that I originally did something similar to you with the Promise.all() technique, but I realised that I didn't actually need to fade both elements out anyway - only the parent container! Also, note that those manual fadeOut methods are only used by Android as of the latest version (v1.1.0). iOS relies on the default React Native <Modal /> fade out functionality because I found smoother results using that instead.
This should be fixed in v1.1.1. Please re-open if not.
|
gharchive/issue
| 2020-05-18T13:59:30 |
2025-04-01T04:55:50.654782
|
{
"authors": [
"adammcarth",
"marcocavanna"
],
"repo": "adammcarth/react-native-segmented-picker",
"url": "https://github.com/adammcarth/react-native-segmented-picker/issues/15",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2311610208
|
🛑 relay.wagnersnetz.de is down
In 6040b6a, relay.wagnersnetz.de (https://relay.wagnersnetz.de/actor) was down:
HTTP code: 502
Response time: 726 ms
Resolved: relay.wagnersnetz.de is back up in 43c16d6 after 9 minutes.
|
gharchive/issue
| 2024-05-22T22:39:56 |
2025-04-01T04:55:50.664510
|
{
"authors": [
"adamus1red"
],
"repo": "adamus1red/ActivityPub-Relays",
"url": "https://github.com/adamus1red/ActivityPub-Relays/issues/11894",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1573065931
|
🛑 relay.pissdichal.de is down
In bcf2574, relay.pissdichal.de (https://relay.pissdichal.de/actor) was down:
HTTP code: 0
Response time: 0 ms
Resolved: relay.pissdichal.de is back up in 7fb20b3.
|
gharchive/issue
| 2023-02-06T18:23:18 |
2025-04-01T04:55:50.667726
|
{
"authors": [
"adamus1red"
],
"repo": "adamus1red/ActivityPub-Relays",
"url": "https://github.com/adamus1red/ActivityPub-Relays/issues/3269",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
153091586
|
Email regex too strict
The regex requires a TLD, which isn't required in the email address spec[0]. There's been a lot written[1] about email regex that's too strict and I tend to agree that it's better to have false positives than false negatives, especially with a default validator. The regex actually fails a lot of valid email addresses (although most will likely rarely appear).
Here's a list[2] of valid email address the regex fails,
"much.more unusual"@example.com
"very.unusual.@.unusual.com"@example.com
"very.(),:;<>[]".VERY."very@\ "very".unusual"@strange.example.com
admin@mailserver1
#!$%&'*+-/=?^_`{}|~@example.org
"()<>[]:,;@\"!#$%&'*+-/=?^_`{}| ~.a"@example.org
" "@example.org (space between the quotes)
example@localhost (sent from localhost)
user@com
user@localserver
user@[IPv6:2001:db8::1]
[0] https://tools.ietf.org/html/rfc2822#section-3.4
[1] Email spec in ABNF and also includes the regex implementation https://html.spec.whatwg.org/multipage/forms.html#valid-e-mail-address
[2] Taken from https://en.wikipedia.org/wiki/Email_address#Valid_email_addresses
You also don't seem to be able to validate emails that use special characters e.g. scøtt@êxample.com
Will get to this with https://github.com/adamwaite/Validator/issues/30.
If anyone could provide an updated regex I'll be happy to implement it. I don't know enough about it to do it myself.
Here's the regex implementation from the spec cited above,
/^[a-zA-Z0-9.!#$%&'*+\/=?^_`{|}~-]+@[a-zA-Z0-9](?:[a-zA-Z0-9-]{0,61}[a-zA-Z0-9])?(?:\.[a-zA-Z0-9](?:[a-zA-Z0-9-]{0,61}[a-zA-Z0-9])?)*$/
@adamkuipers your rule is very complete but does not work for special chars like the one @slhodson969 highlighted
this is a very blatant email validator
^.+@.+\\..+$
Remember that is better to check email on server side and not block the user on UI side.
If we put a wrong validator, users with valid emails will be unable to register in your app and not being very happy about it
Yeah, I agree that the simpler one is better.
https://github.com/adamwaite/Validator/blob/master/Validator/Validator/EmailValidationPattern.swift
@adamwaite A quick information: What is your plan to release a new version (for carthage ;) )
with these PR merged?
Thanks
Right now 🎉
Version 2.1.0 ready to roll.
|
gharchive/issue
| 2016-05-04T19:10:44 |
2025-04-01T04:55:50.677343
|
{
"authors": [
"adamkuipers",
"adamwaite",
"racer1988",
"slhodson969"
],
"repo": "adamwaite/Validator",
"url": "https://github.com/adamwaite/Validator/issues/36",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
24652790
|
Support KB Server
Allow AIM to connect to a KB server to retrieve / set new KB values.
This will probably be a somewhat larger research item -- is it better for mobile devices to pull down larger chunks of the KB server / do a bulk update, or work with smaller bits of data?
Hi Erik,
Thanks for supplying this library note.
It is the javascript library version of the python based avahi linux
zeroconf library module I use
to create standalone KB services for kbdiscovery testing for Bruce.
It allows me to test discovery on linux whereas I use the bonjour dns-sd
version for OSX & windows.
Interestingly also shows the npm package downloader and installer
...equivalent of using pip on all
3 OS's .
Good to see this thanks.
On Tue, May 22, 2018 at 4:16 AM, Erik Brommers notifications@github.com
wrote:
Javascript library for local (LAN) KB server discovery:
https://www.npmjs.com/package/bonjour
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
https://github.com/adapt-it/adapt-it-mobile/issues/8#issuecomment-390737980,
or mute the thread
https://github.com/notifications/unsubscribe-auth/ANW-JFbpxM-oPqsEO1Qa33Sd9MoYtgZzks5t0wSZgaJpZM4BVtco
.
|
gharchive/issue
| 2013-12-20T22:49:57 |
2025-04-01T04:55:50.686796
|
{
"authors": [
"eb1",
"gateslp"
],
"repo": "adapt-it/adapt-it-mobile",
"url": "https://github.com/adapt-it/adapt-it-mobile/issues/8",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2561584272
|
fix: set options.source before raising error on empty doc tree
Made sure the options variable is properly backfilled before any exception can be raised
This fixes #705
Thanks!
|
gharchive/pull-request
| 2024-10-02T13:00:36 |
2025-04-01T04:55:50.701652
|
{
"authors": [
"adbar",
"dmoklaf"
],
"repo": "adbar/trafilatura",
"url": "https://github.com/adbar/trafilatura/pull/707",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
310677348
|
HASwitchPlate Connected automation triggers multiple times
The HASwitchPlate_Connected automation triggers multiple (close to 15 times) when cycling power on the panel or when it reconnects to WiFi/MQTT
file: HASwitchPlatePages.yaml
I can't seem to reproduce this behavior. I've deployed a new instance of hass (currently 0.67.1) and deployed the stock config from the repo here with the following command:
svn export https://github.com/aderusha/HASwitchPlate/trunk/Home_Assistant ~/.homeassistant --force
I started up Hass and configured the default state of the page selection buttons with the following commands:
curl -X POST -H "x-ha-access: YOUR_PASSWORD" -H "Content-Type: application/json" -d '{ "entity_id": "input_text.haswitchplate_pagebutton1_label", "value": "scenes" }' http://localhost:8123/api/services/input_text/set_value
curl -X POST -H "x-ha-access: YOUR_PASSWORD" -H "Content-Type: application/json" -d '{ "entity_id": "input_text.haswitchplate_pagebutton2_label", "value": "status" }' http://localhost:8123/api/services/input_text/set_value
curl -X POST -H "x-ha-access: YOUR_PASSWORD" -H "Content-Type: application/json" -d '{ "entity_id": "input_text.haswitchplate_pagebutton3_label", "value": "media" }' http://localhost:8123/api/services/input_text/set_value
curl -X POST -H "x-ha-access: YOUR_PASSWORD" -H "Content-Type: application/json" -d '{ "entity_id": "input_number.haswitchplate_pagebutton1_page", "value": 1}' http://localhost:8123/api/services/input_number/set_value
curl -X POST -H "x-ha-access: YOUR_PASSWORD" -H "Content-Type: application/json" -d '{ "entity_id": "input_number.haswitchplate_pagebutton2_page", "value": 2}' http://localhost:8123/api/services/input_number/set_value
curl -X POST -H "x-ha-access: YOUR_PASSWORD" -H "Content-Type: application/json" -d '{ "entity_id": "input_number.haswitchplate_pagebutton3_page", "value": 8}' http://localhost:8123/api/services/input_number/set_value
curl -X POST -H "x-ha-access: YOUR_PASSWORD" -H "Content-Type: application/json" -d '{ "entity_id": "input_number.haswitchplate_active_page", "value": 1}' http://localhost:8123/api/services/input_number/set_value
Once Hass has initialized I powered on the HASwitchPlate and observed all automations running once. I power cycled the HASwitchPlate device several more times with Hass running and observed the same behavior each time.
Can you test this with a clean configuration by chance? Is there any way another automation you've deployed might be interacting somehow?
I've narrowed the issue down to two compounding problems. One is an issue with Hass for which I have submitted a bug report. With MQTT discovery enabled, hass is sending multiple replies in response to a single trigger. This behavior is compounded by the page switching automations, which in certain cases can result in the page command being sent twice. Because Hass is duplicating the automation response, the system can get stuck in a loop sending page commands back and forth, each of which triggers an additional cascade of MQTT messages until the Arduino code can't keep up and either resets or takes too long to respond to keep the loop going. The result can be 10 or more responses on some of the page flipping automations.
this is still occurring. in fact I have the switch plate disconnected from power yet whenever I restart HA, the state goes back to "Connected" even though the switch plate is physically powered off, then the automation triggers now close to 25 times
|
gharchive/issue
| 2018-04-03T03:00:54 |
2025-04-01T04:55:50.730654
|
{
"authors": [
"aderusha",
"squirtbrnr"
],
"repo": "aderusha/HASwitchPlate",
"url": "https://github.com/aderusha/HASwitchPlate/issues/7",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1337325297
|
Update to Python 3.9
Hi,
Can you update to Python 3.9?
Im getting this error:
[ERROR] WebDriverException: Message: Service chromedriver unexpectedly exited. Status code was: 127
Traceback (most recent call last):
File "/var/task/lambda_function.py", line 80, in lambda_handler
driver = webdriver.Chrome(chrome_options=chrome_options)
File "/opt/python/selenium/webdriver/chrome/webdriver.py", line 68, in __init__
self.service.start()
File "/opt/python/selenium/webdriver/common/service.py", line 98, in start
self.assert_process_still_running()
File "/opt/python/selenium/webdriver/common/service.py", line 109, in assert_process_still_running
raise WebDriverException(
Thanks.
Seeing as how AWS has now removed 3.6 support it seems as if the only compatible version is 3.7 based on other issues that I've read. Hopefully this will be looked at so that we at least have 3.7 & 3.8 support. In any case some updates would be awesome.
|
gharchive/issue
| 2022-08-12T15:06:24 |
2025-04-01T04:55:50.735570
|
{
"authors": [
"paulocoutinhox",
"selected-pixel-jameson"
],
"repo": "adieuadieu/serverless-chrome",
"url": "https://github.com/adieuadieu/serverless-chrome/issues/342",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2580419765
|
Add flags to specify the exact functionality required from a particular sub command
When executed vc-analyze calc-stats "path/to/repo"
Instead add flags to the cmd to specify the type of operation required
vc-analyze calc-stats --author-stats "path/to/repo"
vc-analyze calc-stats --commit-size "path/to/repo"
/assign
|
gharchive/issue
| 2024-10-11T04:43:19 |
2025-04-01T04:55:50.737321
|
{
"authors": [
"Nikhils-179",
"adigulalkari"
],
"repo": "adigulalkari/VC-Analyzer",
"url": "https://github.com/adigulalkari/VC-Analyzer/issues/25",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
136973590
|
"Parsing failed: Unexpected ï" if file has byte order mark
If checkstyle is run on a file with a byte order mark (for example this file), it causes the following error:
./source/Test.hx:1:1: error: Parsing failed: Unexpected ï
Stacktrace:
Called from hxparse/Lexer.hx line 100
Called from Checker.hx line 104
Called from Checker.hx line 133
Checkstyle version: 191bf0c245
This should be fixed in hxparse, because if I remove the BOM from content before sending it to Lexer, all positions are off by three. And that would likely affect a lot of checks.
So the solution for now is to remove the BOM from your files.
Makes sense. And yeah, that's what I did. :)
Ping @Simn.
I don't think hxparse is the right place to fix this, haxeparser is more appropriate.
|
gharchive/issue
| 2016-02-27T22:00:40 |
2025-04-01T04:55:50.775781
|
{
"authors": [
"AlexHaxe",
"Gama11",
"Simn"
],
"repo": "adireddy/haxe-checkstyle",
"url": "https://github.com/adireddy/haxe-checkstyle/issues/44",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1014094649
|
What if we add jump to top button , so user easily jump to top whenever they need.
If u can add it go ahead
Sir May I !
Sir May I !
Assigned u !
@adithyapaib i would linke to work on this issue please assign this to me.
Sir, I have made my pull request #40
and added a ( jump to top) button and is working fine,pls have a look
@adithyapaib PR is done please cross check if it is working fine or not
|
gharchive/issue
| 2021-10-02T16:09:00 |
2025-04-01T04:55:50.783200
|
{
"authors": [
"AnuragRoshan",
"Sameer1408",
"adithyapaib",
"kapilbhale"
],
"repo": "adithyapaib/Neomorphic-PortfolioTemplate",
"url": "https://github.com/adithyapaib/Neomorphic-PortfolioTemplate/issues/24",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2597525184
|
i have added the transition in the quicklinks
Issues Identification
Closes: #805
Description
Summary
i have added transition too quicklinks
Details
i have added transition
Types of Changes
Please check the boxes that apply
[x ] Bugfix (non-breaking change that fixes an issue)
[ ] New feature (non-breaking change that adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
[ ] Documentation update (Documentation content changed)
[ ] Other (please describe):
Checklist
Please check the boxes that apply
[x ] My code follows the style guidelines of this project
[x ] I have performed a self-review of my own code
[x ] I have commented my code, particularly in hard-to-understand areas
[ ] I have made corresponding changes to the documentation
[x ] My changes generate no new warnings
[x ] My changes do not break the current system and pass all existing test cases
[x ] I have added tests that prove my fix is effective or that my feature works
[ x] New and existing unit tests pass locally with my changes
Screenshots
Additional Information
i cant see any difference can you share recording from the deployment
as i did not get any reply i am closing this
will reopen when i get a reply
|
gharchive/pull-request
| 2024-10-18T13:29:18 |
2025-04-01T04:55:50.789063
|
{
"authors": [
"Ash182004",
"aditya-bhaumik"
],
"repo": "aditya-bhaumik/Pathsphere",
"url": "https://github.com/aditya-bhaumik/Pathsphere/pull/807",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
923829538
|
🛑 Massar Login is down
In 974d57b, Massar Login (https://massarservice.men.gov.ma/moutamadris/Account) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Massar Login is back up in 47f1b98.
|
gharchive/issue
| 2021-06-17T11:46:32 |
2025-04-01T04:55:50.833377
|
{
"authors": [
"adnane-X-tebbaa"
],
"repo": "adnane-X-tebbaa/DownTime-Score",
"url": "https://github.com/adnane-X-tebbaa/DownTime-Score/issues/260",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1954283282
|
feat: article meta blocks EXLM-180
Please always provide the Jira Issue your PR is for, as well as test URLs where your change can be observed (before and after):
Jira ID: https://jira.corp.adobe.com/browse/EXLM-180
Test URLs:
Before: https://main--exlm--adobe-experience-league.hlx.page/docs/integrations-learn/experience-cloud/solution-categories/content-management
After: https://exlm-180-article-meta--exlm--adobe-experience-league.hlx.page/docs/integrations-learn/experience-cloud/solution-categories/content-management
@nitin-rachabathuni overall, I think decorating based on textContent is a bad idea.
In the converter, can we split this into 2 blocks instead? one for topic-list and one for created-for ?
@nitin-rachabathuni overall, I think decorating based on textContent is a bad idea.
In the converter, can we split this into 2 blocks instead? one for topic-list and one for created-for ?
Hi @ahmed-musallam
I had to do it based on textContent as id/classes/meta-tags are all getting stripped off. Went ahead with textContent as there is no other way. cc @thomsebastin @devisree29
@nitin-rachabathuni overall, I think decorating based on textContent is a bad idea.
In the converter, can we split this into 2 blocks instead? one for topic-list and one for created-for ?
Hi @ahmed-musallam , I have updated the code, ty.
@devisree29 could you please address this at the convertor side.
|
gharchive/pull-request
| 2023-10-20T13:21:48 |
2025-04-01T04:55:50.838383
|
{
"authors": [
"Sivaramvt",
"ahmed-musallam",
"nitin-rachabathuni"
],
"repo": "adobe-experience-league/exlm",
"url": "https://github.com/adobe-experience-league/exlm/pull/45",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
315403689
|
Adding testng variant, supporting through IExecutionListener
Description
@agudian Adding TestNG variant
Related Issue
Tasks
[x] I have signed the CLA.
[x] I have written tests and verified that they fail without my change.
Thanks @agudian, can you please review again?
Thanks @agudian, did that too, sorry have been more close to javascript, still catching up with best practices. Let me know if I can be of more help
|
gharchive/pull-request
| 2018-04-18T09:36:36 |
2025-04-01T04:55:50.845958
|
{
"authors": [
"santthosh"
],
"repo": "adobe/S3Mock",
"url": "https://github.com/adobe/S3Mock/pull/48",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2321893264
|
Fix test plugin-action-bar test failing in CI but passing locally
https://github.com/adobe/aem-sidekick/blob/main/test/app/components/plugin/plugin-action-bar.test.js#L636
This test hasn't failed on me in a while, closing. @dylandepass if it's still failing for you, feel free to reopen.
|
gharchive/issue
| 2024-05-28T20:45:20 |
2025-04-01T04:55:50.847498
|
{
"authors": [
"dylandepass",
"rofe"
],
"repo": "adobe/aem-sidekick",
"url": "https://github.com/adobe/aem-sidekick/issues/156",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
18582695
|
Focus lost after closing Project Settings
File > Project Settings
Click Cancel
Try to type anything
Result: Nothing happens. Keyboard focus is not in editor area.
We just need a simple EditorManager.focusEditor() call to fix this, so marking as a starter bug. Or we could use this as an opportunity to finally fix post-dialog focus more generally :-)
n00b here, searching for my first bug to tackle. Unless I'm mistaken, the status of this issue should no longer be open since @Jerhamre seems to have added a fix that's been merged in #12824?
Thanks @webjocky 👍
|
gharchive/issue
| 2013-08-27T01:04:07 |
2025-04-01T04:55:50.849578
|
{
"authors": [
"peterflynn",
"petetnt",
"webjocky"
],
"repo": "adobe/brackets",
"url": "https://github.com/adobe/brackets/issues/4954",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
21929203
|
Mac 10.9: Brackets Helper process erroneously shown as "Not Responding" in Activity Monitor
Description:
Mac 10.9: Brackets Helper process Not Responding 15-20 seconds after launch in Activity Monitor. Shows up in red text in Activity Monitor. The same thing happens with Google Chrome on Mac 10.9, so this is a tracking bug. According to Peter Flynn, it's the GPU helper. We checked the Chrome process ID that was not responding > opened Chrome's task manager (hamburger icon > Tools > Task Manager) > mapped not responding process ID from Activity Monitor to process ID in Chrome Task Manager > GPU Process. Here is the Chromium bug for this issue:
Issue 304860: Mavericks: Helper processes shown in Activity Monitor as "Not Responding"
https://code.google.com/p/chromium/issues/detail?id=304860
Repro on Mac 10.9 using Brackets Sprint 33 10188 GM, Sprint 32 9791 GM, Sprint 31 9569 GM. UTR on Mac 10.8 using BK Sprint 33 GM 10188.
Repro steps:
Install Brackets on Mac 10.9.
Launch Activity Monitor.
Launch Brackets.
Watch Activity Monitor for ~15-20 seconds.
Actual results:
Activity Monitor says Brackets Helper process is not responding.
Expected results:
Brackets Helper is responding.
Workaround:
None. Everything seems to be working fine, so it appears this is a false positive. Comment from Chromium bug: "This is an issue with how Activity Monitor determines responsiveness – these processes are running perfectly fine. (We changed chrome to work around that Activity Monitor issue, and the fix will reach you eventually, depending on which channel you use, how far we merge this, etc.)"
From what I can see, this isn't related to chrome. I don't use Chrome, only Safari.
Brackets Helper on my Mac isn't responding as well, on fresh restart, without the use of live preview.
This bug is about brackets helper 'erroneously' appearing unresponsive in Activity monitor.
The application should not be affected by this. Do you mean the same?
Also which version of brackets are you using? I notice from the comments above that the issue may have been fixed in the chromium build we integrated some time back. Are you seeing this in Brackets 1.2?
@prksingh Good catch! It does appear fixed now on Mac, so let's close it.
@ElegantScripting If you're seeing slow performance when using Live Preview, that's a different issue. Could you file a new bug with more details? (E.g. does it happen with the simple Getting Started project that comes with Brackets too, or only with your own projects?)
No live preview
Extensions disabled
Restarted brackets
Using latest version
Hmm, so apparently looks like the Chrome bug regressed on Yosemite: https://code.google.com/p/chromium/issues/detail?id=397642.
They did land a fix on the Chrome 39 2171 branch shortly before we started out upgrade to CEF 2171, but I'm guessing we just narrowly missed the boat on picking up that fix. (If I'm reading the version numbers right, we need 2171.50 and we currently have 2171.36).
Updating the bug title and reopening.
|
gharchive/issue
| 2013-10-31T21:44:19 |
2025-04-01T04:55:50.858974
|
{
"authors": [
"ElegantScripting",
"dalcala",
"peterflynn",
"prksingh"
],
"repo": "adobe/brackets",
"url": "https://github.com/adobe/brackets/issues/5794",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
447096264
|
CIF-826 - Add integration tests
Description
This PR adds integration tests in it/http. To run the integration tests, a test content package at it/content needs to be installed. Also the tests rely on a mock server in it/mock-server.
These tests are automatically executed on the master branch for any new change. Coverage is reported to codecov.io with the #integration flag. Unit tests are marked with the #unittests flag.
External contributors are supposed to run the integration tests locally as described in the README.md. Pull requests will be validated using the integration tests running on a Jenkins instance on Adobe internal infrastructure. The tests need to be executed against AEM 6.4, 6.5 and master, the Jenkins job has to be triggered manually for each PR.
Types of changes
[ ] Bug fix (non-breaking change which fixes an issue)
[X] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to change)
Checklist:
[X] I have signed the Adobe Open Source CLA.
[X] My change requires a change to the documentation.
[X] I have updated the documentation accordingly.
[X] I have read the CONTRIBUTING document.
[X] I have added tests to cover my changes and the overall coverage did not decrease.
[X] All unit tests pass on CircleCi.
[X] I ran all tests locally and they pass.
Codecov Report
Merging #22 into master will increase coverage by 7.1%.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #22 +/- ##
===========================================
+ Coverage 71.34% 78.45% +7.1%
- Complexity 205 237 +32
===========================================
Files 16 16
Lines 904 905 +1
Branches 143 143
===========================================
+ Hits 645 710 +65
+ Misses 160 112 -48
+ Partials 99 83 -16
Flag
Coverage Δ
Complexity Δ
#integration
78.23% <ø> (?)
237 <ø> (?)
#unittests
71.34% <ø> (?)
205 <ø> (?)
Impacted Files
Coverage Δ
Complexity Δ
...mmerce/graphql/core/GraphqlProductViewHandler.java
58.67% <0%> (+0.34%)
11% <0%> (ø)
:arrow_down:
...a/impl/CatalogDataResourceProviderManagerImpl.java
77.71% <0%> (+4%)
33% <0%> (+5%)
:arrow_up:
.../cq/commerce/graphql/resource/ProductResource.java
90.62% <0%> (+6.25%)
9% <0%> (+1%)
:arrow_up:
...adobe/cq/commerce/graphql/core/MagentoProduct.java
64.36% <0%> (+6.89%)
37% <0%> (+4%)
:arrow_up:
...e/cq/commerce/graphql/resource/ResourceMapper.java
79.88% <0%> (+8.87%)
34% <0%> (+4%)
:arrow_up:
...mmerce/graphql/magento/GraphqlDataServiceImpl.java
88.88% <0%> (+9.09%)
39% <0%> (+3%)
:arrow_up:
...atalog/admin/impl/CatalogIdentifierDatasource.java
100% <0%> (+11.11%)
12% <0%> (+2%)
:arrow_up:
...aphql/resource/GraphqlResourceProviderFactory.java
100% <0%> (+11.11%)
10% <0%> (+2%)
:arrow_up:
...ommerce/graphql/resource/RootCategoryResource.java
100% <0%> (+18.18%)
4% <0%> (+1%)
:arrow_up:
...cq/commerce/graphql/resource/CategoryResource.java
85% <0%> (+20%)
5% <0%> (+2%)
:arrow_up:
... and 2 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 5f27bd8...0661363. Read the comment docs.
|
gharchive/pull-request
| 2019-05-22T12:03:48 |
2025-04-01T04:55:50.882244
|
{
"authors": [
"codecov-io",
"herzog31"
],
"repo": "adobe/commerce-cif-connector",
"url": "https://github.com/adobe/commerce-cif-connector/pull/22",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1826697256
|
Performance: fix CLS on main element in Doc Detail page
Issue mentioned in this PR:
https://github.com/adobe/helix-website/issues/353#issuecomment-1649418313
Details:
main element still seems to move around quite a bit:
.
use.live URL as test URL to have a cached and more constant reference
Test Urls to be used:
Before: https://main--helix-website--adobe.hlx.page/docs/sidekick
After: https://redesign-doc-page-perf-fix--helix-website--adobe.hlx.live/docs/sidekick
@snowiewdev i think this was fixed https://github.com/adobe/helix-website/pull/388
@davidnuescheler Thank you so much! The fix on #388 is looking good to me:
https://pagespeed.web.dev/analysis/https-www-hlx-live-docs-sidekick/lbesvr3spj?form_factor=desktop
Closing this issue
|
gharchive/issue
| 2023-07-28T16:10:26 |
2025-04-01T04:55:50.890054
|
{
"authors": [
"davidnuescheler",
"snowiewdev"
],
"repo": "adobe/helix-website",
"url": "https://github.com/adobe/helix-website/issues/356",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1883494099
|
fix: Add rum tracking to form block
Adding rum tracking to form block.
https://main--helix-website--adobe.hlx.page/developer/forms
vs.
https://main--helix-website--pareesh.hlx.page/developer/forms
Types of changes
[x] Bug fix (non-breaking change which fixes an issue)
[ ] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to change)
Checklist:
[x] I have signed the Adobe Open Source CLA.
[x] My code follows the code style of this project.
[ ] My change requires a change to the documentation.
[ ] I have updated the documentation accordingly.
[ ] I have read the CONTRIBUTING document.
[ ] I have added tests to cover my changes.
[ ] All new and existing tests passed.
Seems outdated. Closing for now, feel free to reopen and refresh if needed.
|
gharchive/pull-request
| 2023-09-06T08:16:24 |
2025-04-01T04:55:50.894787
|
{
"authors": [
"Pareesh",
"kptdobe"
],
"repo": "adobe/helix-website",
"url": "https://github.com/adobe/helix-website/pull/371",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2291316697
|
chore: deprecate bookmarklet
The Sidekick Bookmarklet is being deprecated due to lack of use (0.1%). Bookmarklet users should:
no longer be able to install new bookmarklets
receive a deprecation message when they click the bookmarklet
Test URLs:
Before: https://main--helix-website--adobe.aem.page/tools/sidekick/?giturl=https://github.com/adobecom/blog
After: https://deprecate-bookmarklet--helix-website--adobe.aem.page/tools/sidekick/?giturl=https://github.com/adobecom/blog
|
gharchive/pull-request
| 2024-05-12T12:41:18 |
2025-04-01T04:55:50.897596
|
{
"authors": [
"rofe"
],
"repo": "adobe/helix-website",
"url": "https://github.com/adobe/helix-website/pull/508",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2054592026
|
TableHeaderProps.columns should be readonly T[] instead of T[]
Provide a general summary of the issue here
Basically to make types work better I as const the columns, which causes the <Row types to be better (I get a union of literals), but TableHeaderProps errors since it expects a mutable array
🤔 Expected Behavior?
Allow readonly arrays being passed
😯 Current Behavior
Readonly arrays not allowed
💁 Possible Solution
No response
🔦 Context
No response
🖥️ Steps to Reproduce
The type 'readonly [{ readonly id: "teamName"; readonly name: "Team"; }, { readonly id: "creator"; readonly name: "Creator"; }, { readonly id: "createdAt"; readonly name: "Created"; }, { readonly id: "status"; readonly name: "Status"; }]' is 'readonly' and cannot be assigned to the mutable type 'object[]'.ts(4104)
types.d.ts(1740, 5): The expected type comes from property 'columns' which is declared here on type 'IntrinsicAttributes & TableHeaderProps<object> & RefAttributes<HTMLTableSectionElement>'
(property) TableHeaderProps<object>.columns?: object[] | undefined
const columns = [{id: 1, col1: "col1"}] as const;
<TableHeader columns={columns}>
() => <div/>
</TableHeader>
Version
1.0.0
What browsers are you seeing the problem on?
Safari
If other, please specify.
No response
What operating system are you using?
Mac OS
🧢 Your Company/Team
No response
🕷 Tracking Issue
No response
Thanks for the issue, this seems fine. I don't think we mutate the columns.
|
gharchive/issue
| 2023-12-23T00:17:39 |
2025-04-01T04:55:50.902539
|
{
"authors": [
"sbdchd",
"snowystinger"
],
"repo": "adobe/react-spectrum",
"url": "https://github.com/adobe/react-spectrum/issues/5625",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1235763437
|
CSF 3 Well & Switch
Closes
The goal is to outline a few simple patterns for some easy components.
I've also taken some liberties in removing stories I think are covered well enough by Chromatic.
And I've tried to combine some of our render functions to instead utilize CSF 3's property spreading as they recommend in their docs.
See docs for v2 vs v3 comparisons: https://storybook.js.org/blog/component-story-format-3-0/
Open questions/opinions needed:
Do we like this style over the Template.bind({}) syntax? Do we like spreading all the props or would we prefer more stories with custom render functions instead. For instance, should our Default Switch story render children, or should it take children as args?
In both cases, you could override the children, however, it's a little different.
export const Default: ComponentStoryObj<typeof Switch> = {
render: (args) => (
<Switch {...args} />
),
args: {children: <>Switch Label</>}
};
// two options here, omit the Default.args
export const NoLabel: ComponentStoryObj<typeof Switch> = {
...Default,
args: {'aria-label': 'This switch has no visible label'},
name: 'no label',
parameters: {description: {data: 'Try me with a screen reader.'}}
};
// or include them
export const NoLabel: ComponentStoryObj<typeof Switch> = {
...Default,
args: {...Default.args, children: null, 'aria-label': 'This switch has no visible label'},
name: 'no label',
parameters: {description: {data: 'Try me with a screen reader.'}}
};
vs
export const Default: ComponentStoryObj<typeof Switch> = {
render: (args) => (
<Switch {...args}>Switch Label</Switch>
),
args: {}
};
// two options here now to override the children for our story with no labels
export const NoLabel: ComponentStoryObj<typeof Switch> = {
...Default,
args: {children: null, 'aria-label': 'This switch has no visible label'},
name: 'no label',
parameters: {description: {data: 'Try me with a screen reader.'}}
};
// or
export const NoLabel: ComponentStoryObj<typeof Switch> = {
render: (args) => <Switch {...args} />,
args: {'aria-label': 'This switch has no visible label'},
name: 'no label',
parameters: {description: {data: 'Try me with a screen reader.'}}
};
✅ Pull Request Checklist:
[ ] Included link to corresponding React Spectrum GitHub Issue.
[ ] Added/updated unit tests and storybook for this change (for new code or code which already has tests).
[ ] Filled out test instructions.
[ ] Updated documentation (if it already exists for this component).
[ ] Looked at the Accessibility Practices for this feature - Aria Practices
📝 Test Instructions:
🧢 Your Project:
I think I like the first example better, but it feels like a pretty small difference overall. Think I would have to try using this format for a more complex component that has a more involved setup/varied setups (aka the collection ones) to form a strong opinion. Definitely like this over the bind syntax though, that syntax was always a bit confusing to me.
|
gharchive/pull-request
| 2022-05-13T22:26:30 |
2025-04-01T04:55:50.908465
|
{
"authors": [
"LFDanLu",
"snowystinger"
],
"repo": "adobe/react-spectrum",
"url": "https://github.com/adobe/react-spectrum/pull/3129",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2423491660
|
Get past 30 days' data
Please ensure your pull request adheres to the following guidelines:
[ ] make sure to link the related issues in this description
[ ] when merging / squashing, make sure the fixed issue references are visible in the commits, for easy compilation of release notes
Related Issues
Thanks for contributing!
no activity, closing
|
gharchive/pull-request
| 2024-07-22T18:19:29 |
2025-04-01T04:55:50.910273
|
{
"authors": [
"rachel-kittens",
"solaris007"
],
"repo": "adobe/spacecat-audit-worker",
"url": "https://github.com/adobe/spacecat-audit-worker/pull/330",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
756415658
|
Add In-line Alert component
Description
Let's implement this re-imagined Alert component https://spectrum-contributions.corp.adobe.com/page/in-line-alert-beta/
Why do you need this feature or component?
@badimon asked for it
Additional context
Alert was previously deprecated and was nuked, this is a new version of it: https://github.com/adobe/spectrum-css/pull/757/files#diff-47c0d377df764997b3c618a2fd561fdd7d8737a96e35dc906ee0b921618ed819L1
This has been implemented and migrated to use @adobe/spectrum-tokens: https://github.com/adobe/spectrum-css/tree/main/components/inlinealert
|
gharchive/issue
| 2020-12-03T17:52:02 |
2025-04-01T04:55:50.913541
|
{
"authors": [
"lazd",
"pfulton"
],
"repo": "adobe/spectrum-css",
"url": "https://github.com/adobe/spectrum-css/issues/1081",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
574986849
|
FieldLabel: Change left/right to start/end
Description
Currently, FieldLabel has left/right variants, it should probably be start/end
Why do you need this feature or component?
This makes the API make more sense. Right now, left/right work according to LTR/RTL, but it doesn't make sense.
Additional context
#532
synced to jira: https://jira.corp.adobe.com/browse/CSS-128
Closing this, it's tracked in Jira
|
gharchive/issue
| 2020-03-03T22:13:59 |
2025-04-01T04:55:50.915624
|
{
"authors": [
"lazd",
"misterbrownlee"
],
"repo": "adobe/spectrum-css",
"url": "https://github.com/adobe/spectrum-css/issues/605",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1662531544
|
chore: separate out js functionality for each component
Description
Continue migrating the JS into enhancement.js files for the components they relate to
Files Changed:
// Load Inputgroup functionality from components/Inputgroup/metadata/enhancement.js
// Load Stepper functionality from components/stepper/metadata/enhancement.js
// Load Picker functionality from components/picker/metadata/enhancement.js
// Load Treeview functionality from components/treeview/metadata/enhancement.js
// Load Accordion functionality from components/accordion/metadata/enhancement.js
// Load Cyclebutton functionality from components/cyclebutton/metadata/enhancement.js
// Load Sliders functionality from components/slider/metadata/enhancement.js
// Load Dial functionality from components/dial/metadata/enhancement.js
// Load Dialog functionality from components/dialog/metadata/enhancement.js
I'm sure you realize this and just haven't gotten to take care of it yet, but when I run this locally and try to view the docs site, I'm getting a setPickerValue is not a function error in the console, and it's preventing anything else on any of the component pages from loading properly.
Again, you probably know about this already, but I thought I should call it out so that we can take care of it!
Great catch Patrick! I suspect this comes from needing to load the enhancement.js assets first before tools/documentation/assets/js/site.js which references it. We'll also need a more graceful degradation if the function can't be found - an if check should do the trick to see if window.setPickerValue is undefined.
|
gharchive/pull-request
| 2023-04-11T13:49:27 |
2025-04-01T04:55:50.919035
|
{
"authors": [
"Rajdeepc",
"castastrophe"
],
"repo": "adobe/spectrum-css",
"url": "https://github.com/adobe/spectrum-css/pull/1767",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1034988346
|
Deprecate profile-counters-v1
Please link to the issue #…
https://jira.corp.adobe.com/browse/CJM-17091
@prabhum2 @Seth0115 when are we planning to merge this change?
@prabhum2 @Seth0115 when are we planning to merge this change?
Whenever XDM team deems possible?
|
gharchive/pull-request
| 2021-10-25T10:59:40 |
2025-04-01T04:55:50.921928
|
{
"authors": [
"Seth0115",
"jainaks"
],
"repo": "adobe/xdm",
"url": "https://github.com/adobe/xdm/pull/1394",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2193318721
|
[MWPW-143523] Accessibility Fixes
Describe your specific features or fixes
Bundles several accessibility fixes together.
Modifies template-x to display a border when the hidden link is in focus
Modifies the marquee to make the pause button keyboard accessible.
Modifies the layout block to be an that links to the configs destination instead of using an onclick tracker.
Modifies the Monthly / Yearly pricing radio button to be selectable via keyboard
Modifies the hero animation on pages with the columns block such that it can be paused / unpaused by the enter key.
Resolves: MWPW-143523
MWPW-143518
MWPW-143516
MWPW-143517
MWPW-143524
MWPW-143520
MWPW-143525
MWPW-143522
Test URLs:
Before: https://main--express--adobecom.hlx.page/express/
-https://template-x-accessibility--express--adobecom.hlx.page/express/
-https://template-x-accessibility--express--adobecom.hlx.page/express/feature/image/resize
-https://template-x-accessibility--express--adobecom.hlx.page/express/create/logo
Codecov Report
Attention: Patch coverage is 33.33333% with 2 lines in your changes are missing coverage. Please review.
Project coverage is 36.03%. Comparing base (75a7616) to head (f50cce1).
Files
Patch %
Lines
express/blocks/layouts/layouts.js
0.00%
2 Missing :warning:
Additional details and impacted files
@@ Coverage Diff @@
## stage #762 +/- ##
==========================================
- Coverage 36.26% 36.03% -0.24%
==========================================
Files 132 131 -1
Lines 25728 25611 -117
==========================================
- Hits 9331 9228 -103
+ Misses 16397 16383 -14
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Have feedback on the report? Share it here.
nice! This also helps me see one issue with the updated billing-toggle. Let me push a fix directly to this branch tomorrow
I can't tell which card in the layout block I'm tabbing through on https://template-x-accessibility--express--adobecom.hlx.page/express/feature/image/resize. Is it intended?
I still can't perform the copy to clipboard action on the template-x templates. Do we plan to fix that in this PR?
|
gharchive/pull-request
| 2024-03-18T21:05:56 |
2025-04-01T04:55:50.935595
|
{
"authors": [
"JingleH",
"codecov-commenter",
"echen-adobe",
"qiyundai"
],
"repo": "adobecom/express",
"url": "https://github.com/adobecom/express/pull/762",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2625766126
|
MWPW-157100: Merch card headers as p tags for Accessibility #3080
This PR changes the tag names of all headings to <p> except for the first heading <h3> as per accessibility requirements, as changed in this milo PR: https://github.com/adobecom/milo/pull/3080
Closing this PR since merch card block NALA tests has been moved to Milo repo.
|
gharchive/pull-request
| 2024-10-31T01:49:49 |
2025-04-01T04:55:50.937385
|
{
"authors": [
"Axelcureno"
],
"repo": "adobecom/nala",
"url": "https://github.com/adobecom/nala/pull/461",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
921305268
|
Kerberos Credential Retrieval from Cache not Working in Cross-Realm Setup
Summary
We have a problem with JDK and Kerberos in a cross-realm setup, of which we suspect the root cause to be in the JDK.
We assume cross-realm support was implemented as described here: https://bugs.openjdk.java.net/browse/JDK-8005819
The ticket's test description states the following setup:
client@REALM2 has established a kerberos context with krbservice1@REALM1 with delegcred=false, krbservice1@REALM1 delegates to krbservice2@REALM1 using S4U2PROXY
However, in our setup krbservice2, the one we want to delegate to, is in a different realm than krbservice1. This means we have no official statement if our setup is supposed to work at all.
The delegation to krbservice2 is configured as resource-based delegation. Which means in the AD for krbservice1, krbservice2 is not in the property list of "allowedToDelegateTo". Instead, we configured the trust relationship for delegation with the PowerShell command Set-ADUser. krbservice2 cannot be added to the allowedToDelegateTo property of krbservice1 because they are in different realms.
Steps to reproduce
On the first round the user "client" establishes a connection to our krbservice1. krbservice1 sends a delegated search request to krbservice2. The Java security module successfully retrieves a ticket on the user's behalf, which is attached to the search request to krbservice2 and search results are returned.
However, on the second search query, in the class sun/security/krb5/internal/CredentialsUtil.java it restores the credentials from the cache, but it is not able to retrieve a valid ticket for krbservice2. Instead, we get the following error messages:
KRBError:
sTime is Tue Jun 01 14:03:37 CEST 2021 1622549017000
suSec is 204909
error code is 28
error Message is null
sname is HTTP/krbservice2@REALM2
msgType is 30
and
KRBError:
sTime is Tue Jun 01 14:03:37 CEST 2021 1622549017000
suSec is 201727
error code is 13
error Message is KDC cannot accommodate requested option
sname is HTTP/krbservice2@REALM1
eData provided.
msgType is 30
When we look into the first error message with Wireshark, we found the error to be:
KRB5KRB_AP_PATH_NOT_ACCEPTED
When we decode the "Unknown eData field of KRB-ERROR" sent along with the first error message, we get:
C:\Users\me\Downloads>.\Err_6.4.5.exe c0000272
for hex 0xc0000272 / decimal -1073741198
STATUS_NO_MATCH ntstatus.h
There was no match for the specified key in the index.
as an HRESULT: Severity: FAILURE (1), FACILITY_NULL (0x0), Code 0x272
for hex 0x272 / decimal 626
ERROR_NO_MORE_MATCHES winerror.h
There are no more matches for the current index
enumeration.
2 matches found for "c0000272"
From these error messages we understand the following: For some unknown reason in the second search request with the credentials retrieved from ReferralsCache it is not possible to find krbservice2 in REALM2 anymore.
Analysis
We think that this is a bug in the JDK, because when we clear the ReferralsCache between search requests, it returns search results successfully. We did this manually with the remote debugging options of an IDE.
When remote-debugging into the code of CredentialsUtil we see that on the first request, Credentials of type Proxy are not stored, but the ones with None are. A comment in the code explicitly states that Credentials for Proxy should not be cached. However, in the second search request, the Credentials for Proxy are retrieved from the Cache. We think this contradicts the warning in the comment, as this is basically a caching of Proxy credentials. Therefore we suspect this to be the location of the bug and Credentials for type Proxy should not be retrieved from the cache.
Please fix.
Triaging info
This is the data of the middle service, krbservice1 in the above described setup.
Java version: We tried both:
openjdk version "1.8.0_292"
OpenJDK Runtime Environment (AdoptOpenJDK)(build 1.8.0_292-b10)
OpenJDK 64-Bit Server VM (AdoptOpenJDK)(build 25.292-b10, mixed mode)
and
openjdk version "11.0.11" 2021-04-20
OpenJDK Runtime Environment AdoptOpenJDK-11.0.11+9 (build 11.0.11+9)
OpenJDK 64-Bit Server VM AdoptOpenJDK-11.0.11+9 (build 11.0.11+9, mixed mode)
We encounter the same problem in both versions.
What is your operating system and platform?
Windows Server 2016 Datacenter
Did it work before?
Our setup works successfully with krbservice1 and 2 in the same realm with resource-based delegation configured.
Here is our suggestion for a fix. We successfully tested it on our internal environment.
#318_add_check_before_requesting_referrals_cache_to_prevent_usage_of_an_inappropriate_tick.zip
Duplicated by https://bugs.java.com/bugdatabase/view_bug.do?bug_id=JDK-8270137
Thanks for reporting this bug.
Just for the record, the OpenJDK discussion is taking place in the context of https://github.com/openjdk/jdk/pull/5036
merged upstream.
|
gharchive/issue
| 2021-06-15T11:40:49 |
2025-04-01T04:55:50.965822
|
{
"authors": [
"Raytion",
"karianna",
"martinuy"
],
"repo": "adoptium/adoptium-support",
"url": "https://github.com/adoptium/adoptium-support/issues/318",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1113552219
|
exclude jdk_tools testcases due to upstream issue
Fixes #3061
@sophia-guo @llxia @ShelleyLambert
@llxia here are the requested changes
|
gharchive/pull-request
| 2022-01-25T08:06:13 |
2025-04-01T04:55:50.967547
|
{
"authors": [
"LizyBbethy"
],
"repo": "adoptium/aqa-tests",
"url": "https://github.com/adoptium/aqa-tests/pull/3283",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1549838546
|
Reenable JDK19 tests after the fix for eclipse-openj9/openj9#15251
The following four tests are reenabled since
eclipse-openj9/openj9#15251 has been fixed.
Previously they were exclude in adoptium/aqa-tests#3750
SC_Softmx_JitAot
SC_Softmx_JitAot_Linux
SharedClasses.SCM23.MultiThread
SharedClasses.SCM23.MultiThreadMultiCL
Signed-off-by: Annabelle Huo Annabelle.Huo@ibm.com
@llxia @Mesbah-Alam May I ask you to review? Thank you!
@JasonFengJ9 @pshipton @0xdaryl fyi
|
gharchive/pull-request
| 2023-01-19T19:43:49 |
2025-04-01T04:55:50.969874
|
{
"authors": [
"a7ehuo"
],
"repo": "adoptium/aqa-tests",
"url": "https://github.com/adoptium/aqa-tests/pull/4261",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1227080599
|
Add gcc11.2 role for jdk19+
Fixes #2538
Signed-off-by: Adam Brousseau adam.brousseau88@gmail.com
Checklist
[ ] commit message has one of the standard prefixes
[ ] faq.md updated if appropriate
[ ] other documentation is changed or added (if applicable)
[ ] playbook changes run through VPC or QPC (if you have access)
[ ] for inventory.yml changes, bastillion/nagios/jenkins updated accordingly
@sxa can you review or delegate? I basically copy/pasted the gcc10 role. Will need Adopt to do the binary work on the the other side.
Is there an outlook on timeframe for completing this work? Ie. adding the gcc11 binary to Adopt's Jenkins' userContent.
VPC run: https://ci.adoptopenjdk.net/job/VagrantPlaybookCheck/1458 - I'm not sure why the CentOS6 run failed to extract GCC10 in that run though - it /shouldn't/ be anything to do with this PR.
We have GCC11 now available on all of the requiredLInux platforms now (I will update the ARM32 one next week, but hat won't require a playbook change), so this should be good to go in if you're happy to take it out of draf @AdamBrousseau
Think this is in order now @sxa
|
gharchive/pull-request
| 2022-05-05T19:17:57 |
2025-04-01T04:55:50.974795
|
{
"authors": [
"AdamBrousseau",
"sxa"
],
"repo": "adoptium/infrastructure",
"url": "https://github.com/adoptium/infrastructure/pull/2542",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
915944134
|
Problem uninstalling AOJDK 11.0.7_10 OpenJ0.20.0 (x64) LTS in Windows 10 2H2 x64
This is a copy/move from temurin-build Issue #2572 at github https://github.com/adoptium/temurin-build/issues/2572
(as the topic also concerns the installer, and it could be useful to all to open the issue here under /installer for everyones benefit, as kindly suggested by @karianna )
What are you trying to do?
-to uninstall OLD LTS (to upgrade to a current JDK 11 LTS (11.0.10_9_openj9-0.24.0.msi)) - currently the installer and the associated uninstaller is stuck, impossible to uninstall it
Expected behaviour:
Old 11 LTS 11.0.7_10 to be uninstalled smoothly (and replaced with the new one)
Observed behaviour:
The old AOJDK can not be uninstalled from windows system by its uninstaller, even if it is exactly the same name dot subvairiant etc.
When trying to circumvent the problem by installing new LTS supposedly to uninstall or rewrite registry entries pointing to the correct LTS - it fails the same by invoking the same old failing uninstaller script.
It asks for .msi name AOJDK EOJ9 11.0.7_10 (x64) , and when matched - just refuses to proceed and continues to insist that it is not the right one, although the name and variant is the exact subvariant dots and an exact match of a filename (downloaded .msi of the 11.0.7.10 build from github while trying to correct the situation)
Any other comments:
By the way of trying to resolve this, under releases history I noticed some sub-variant versions of LTS 11.0.10_9 / openj9.0.24.0 installer on github - it would be good to get everything coherent in th final outcome of eradicating .msi uninstall problem for good.
Useful observation: A solid manual removal instructions is missing - it was not possible to find it only by search engine query.
Thank you in advance for considering this critical uninstaller bug
Please give us some log files from the msi
msiexec /x "C:\MyPackage\Example.msi" /L*V "C:\log\example.log"
Then try to reboot and uninstall.
@douph1
Thank you for the swift reply; as I am in daily workplace, will try to provide log answer ASAP in the work break.
Thank you for the log cmd line as the next step, have a great day!
Closed as no new input since long time
|
gharchive/issue
| 2021-06-09T08:37:08 |
2025-04-01T04:55:50.981668
|
{
"authors": [
"douph1",
"succeedmr123"
],
"repo": "adoptium/installer",
"url": "https://github.com/adoptium/installer/issues/318",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2454073120
|
Another problem with isless
As a complement to https://github.com/adrhill/SparseConnectivityTracer.jl/issues/152, with non-evenly distributed t:
using SparseConnectivityTracer: TracerLocalSparsityDetector, jacobian_sparsity
using DataInterpolations: LinearInterpolation
detector = TracerLocalSparsityDetector()
A = LinearInterpolation([0.0, 2.0, 10.0], [0.0, 2.0, 3.0])
function f!(output, input)
output .= A.(input)
end
input = rand(5)
output = zero(input)
jacobian_sparsity(f!, output, input, detector)
this gives
ERROR: MethodError: isless(::SparseConnectivityTracer.Dual{Float64, SparseConnectivityTracer.GradientTracer{SparseConnectivityTracer.IndexSetGradientPattern{Int64, BitSet}}}, ::Float64) is ambiguous.
Candidates:
isless(dx::D, y::Real) where D<:SparseConnectivityTracer.Dual
@ SparseConnectivityTracer path\to\.julia\packages\SparseConnectivityTracer\tpAG2\src\overloads\dual.jl:25
isless(x::Real, y::AbstractFloat)
@ Base operators.jl:178
Possible fix, define
isless(::D, ::AbstractFloat) where D<:SparseConnectivityTracer.Dual
Stacktrace:
[1] lt(o::Base.Order.ForwardOrdering, a::SparseConnectivityTracer.Dual{…}, b::Float64)
@ Base.Order .\ordering.jl:117
[2] bracketstrictlymontonic
@ path\to\.julia\packages\FindFirstFunctions\2w5qZ\src\FindFirstFunctions.jl:179 [inlined]
[3] searchsortedlastcorrelated
@ path\to\.julia\packages\FindFirstFunctions\2w5qZ\src\FindFirstFunctions.jl:266 [inlined]
[4] get_idx(A::LinearInterpolation{…}, t::SparseConnectivityTracer.Dual{…}, iguess::Int64; lb::Int64, ub_shift::Int64, idx_shift::Int64, side::Symbol)
@ DataInterpolations path\to\.julia\packages\DataInterpolations\i1xgb\src\interpolation_utils.jl:133
[5] get_idx
@ path\to\.julia\packages\DataInterpolations\i1xgb\src\interpolation_utils.jl:120 [inlined]
[6] _interpolate(A::LinearInterpolation{…}, t::SparseConnectivityTracer.Dual{…}, iguess::Int64)
@ DataInterpolations path\to\.julia\packages\DataInterpolations\i1xgb\src\interpolation_methods.jl:19
[7] _interpolate
@ path\to\.julia\packages\DataInterpolations\i1xgb\src\interpolation_methods.jl:5 [inlined]
[8] AbstractInterpolation
@ path\to\.julia\packages\DataInterpolations\i1xgb\src\DataInterpolations.jl:24 [inlined]
[9] _broadcast_getindex_evalf
@ .\broadcast.jl:709 [inlined]
[10] _broadcast_getindex
@ .\broadcast.jl:682 [inlined]
[11] getindex
@ .\broadcast.jl:636 [inlined]
[12] macro expansion
@ .\broadcast.jl:1004 [inlined]
[13] macro expansion
@ .\simdloop.jl:77 [inlined]
[14] copyto!
@ .\broadcast.jl:1003 [inlined]
[15] copyto!
@ .\broadcast.jl:956 [inlined]
[16] materialize!
@ .\broadcast.jl:914 [inlined]
[17] materialize!(dest::Vector{…}, bc::Base.Broadcast.Broadcasted{…})
@ Base.Broadcast .\broadcast.jl:911
[18] f!(output::Vector{SparseConnectivityTracer.Dual{…}}, input::Vector{SparseConnectivityTracer.Dual{…}})
@ Main path\to\Ribasim_development\runners\runner.jl:35
[19] trace_function(::Type{SparseConnectivityTracer.Dual{…}}, f!::typeof(f!), y::Vector{Float64}, x::Vector{Float64})
@ SparseConnectivityTracer path\to\.julia\packages\SparseConnectivityTracer\tpAG2\src\interface.jl:49
[20] _local_jacobian_sparsity(f!::Function, y::Vector{…}, x::Vector{…}, ::Type{…})
@ SparseConnectivityTracer path\to\.julia\packages\SparseConnectivityTracer\tpAG2\src\interface.jl:94
[21] jacobian_sparsity(f!::Function, y::Vector{…}, x::Vector{…}, ::TracerLocalSparsityDetector{…})
@ SparseConnectivityTracer path\to\.julia\packages\SparseConnectivityTracer\tpAG2\src\adtypes.jl:122
[22] top-level scope
@ path\to\Ribasim_development\runners\runner.jl:41
Some type information was truncated. Use `show(err)` to see complete types.
Thanks a lot! Looks like the overload of isless(x::Dual, y::Real) added in #149 isn't specific enough for input types and we also need an additional overload on isless(x::Dual, y::AbstractFloat).
Maybe we should run our Dual tests on more diverse input types to catch this type of problem.
|
gharchive/issue
| 2024-08-07T18:25:33 |
2025-04-01T04:55:50.988912
|
{
"authors": [
"SouthEndMusic",
"adrhill"
],
"repo": "adrhill/SparseConnectivityTracer.jl",
"url": "https://github.com/adrhill/SparseConnectivityTracer.jl/issues/153",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1328610483
|
Open Weather Map
Can't find open weather map API on rapid Api website and alternate or some sort of help will be appreciated.
same problem here
You can make a free account on the OpenWeather website.
Assuming you have created the .env file to hide your api keys you can then update the getWeatherData function in the js file of the api folder:
export const getWeatherData = async (lat, lng) => {
try {
const apiKey = process.env.REACT_APP_OPEN_WEATHER_API_KEY;
const { data } = await axios.get(
`https://api.openweathermap.org/data/2.5/weather?lat=${lat}&lon=${lng}&appid=${apiKey}`
);
return data;
} catch (error) {
console.log(error);
}
};
Additionally, the img src will need to be updated in the Map.jsx file:
{weatherData?.list?.map((data, i) => (
<div key={i} lat={data.coord.lat} lng={data.coord.lon}>
<img
src={`https://openweathermap.org/img/wn/${data.weather[0].icon}.png`}
height={75}
alt=""
/>
</div>
))}
When console.log'd, the api calls return the expected data with these updates, however, I haven't been able to get the weather icons to successfully render on the map.
|
gharchive/issue
| 2022-08-04T13:16:46 |
2025-04-01T04:55:51.017327
|
{
"authors": [
"GeomaticaNet",
"gitpk-0",
"pankaj11koundal"
],
"repo": "adrianhajdin/project_travel_advisor",
"url": "https://github.com/adrianhajdin/project_travel_advisor/issues/40",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1515214437
|
Cannot reach myIP:3000 permission denied
Is there an existing issue for this?
[X] I have searched the existing issues
What happened?
I had tesla mate for a year use to work on debian base with docker, but lately cannot get into the UI the grafana log give me
=eror msg="failed to determine the status of alerting engine. Enable either legacy or unified alerting explicitly and try again" err="failed to verify if the 'alert' table exists: unable to open database file: permission denied"
and container Telamate log give me
(File.Error) could not write to file "/opt/app/lib/tzdata-1.1.1/priv/latest_remote_poll.txt": permission denied
no issue with ip :4000
also cannot update image: postgres:13 to image: postgres:14 not compatible?
Expected Behavior
Before could access the UI but not anymore
Steps To Reproduce
Start Teslamate as normal using step 3 The Grafana dashboards are available at http://your-ip-address:3000
Relevant log output
[error] GenServer :tzdata_release_updater terminating
** (File.Error) could not write to file "/opt/app/lib/tzdata-1.1.1/priv/latest_remote_poll.txt": permission denied
logger=migrator t=2022-12-31T21:42:31.39+0000 lvl=eror msg="failed to determine the status of alerting engine. Enable either legacy or unified alerting explicitly and try again" err="failed to verify if the 'alert' table exists: unable to open database file: permission denied"
logger=settings t=2022-12-31T21:43:31.72+0000 lvl=info msg="Starting Grafana" version=8.5.6 commit=623255cf2d branch=HEAD compiled=2022-06-14T09:25:38Z
logger=settings t=2022-12-31T21:43:31.72+0000 lvl=info msg="The state of unified alerting is still not defined. The decision will be made during as we run the database migrations"
Client TESLAMATE_302E3936363 closed its connection.
1672521241: mosquitto version 2.0.15 terminating
[42] FATAL: terminating connection due to administrator command
2022-12-31 21:14:01.663 UTC [41] FATAL: terminating connection due to administrator command
2022-12-31 21:14:01.663 UTC [46] FATAL: terminating connection due to administrator command
2022-12-31 21:14:01.663 UTC [43] FATAL: terminating connection due to administrator command
2022-12-31 21:14:01.663 UTC [38] FATAL: terminating connection due to administrator command
2022-12-31 21:14:01.664 UTC [44] FATAL: terminating connection due to administrator command
2022-12-31 21:14:01.664 UTC [45] FATAL: terminating connection due to administrator command
2022-12-31 21:14:01.665 UTC [1] LOG: background worker "logical replication launcher" (PID 33) exited with exit code 1
2022-12-31 21:14:01.665 UTC [37] FATAL: terminating connection due to administrator command
2022-12-31 21:14:01.667 UTC [40] FATAL: terminating connection due to administrator command
2022-12-31 21:14:01.667 UTC [39] FATAL: terminating connection due to administrator command
Screenshots
No response
Additional data
_teslamate-database-1_logs(1).txt
_teslamate-teslamate-1_logs(1).txt
_teslamate-grafana-1_logs(2).txt
_teslamate-mosquitto-1_logs.txt
Type of installation
Docker
Version
1.27.1
Do you have space available on the system hosting TeslaMate?
Definitely not the issue, the partition where docker folder is got 430Gib free
#3040 - similar errors. No fix, but possible collaboration.
It looks like the both the teslamate and grafana containers have file permission problems. Were they upgraded from older versions?
The output of
docker exec tm_grafana_1 ls -la /var/lib/grafana
docker exec tm_teslamate_1 ls -la /opt/app/lib/tzdata-1.1.1/priv
might be useful (replace tm_grafana_1 and tm_teslamate_1 by the names of your containers).
Just in case I removed _1 at the end of all my containers's names and relaunched the stack in portainer
# docker exec teslamate-teslamate ls -la /opt/app/lib/tzdata-1.1.1/priv total 20 drwxr-xr-x 4 root users 4096 Jul 22 10:35 . drwxr-xr-x 4 root users 4096 Jul 22 10:35 .. -rw-r--r-- 1 root users 9 Jul 22 10:35 latest_remote_poll.txt drwxr-xr-x 2 root users 4096 Jul 22 10:35 release_ets drwxr-xr-x 2 root users 4096 Jul 22 10:35 tmp_downloads
~# docker exec teslamate-grafana ls -la /var/lib/grafana Error response from daemon: Container b1a475948dab1774bf8444a25ec1681ce75837921c3c9e06838b0ff3a2d714a3 is restarting, wait until the container is running
In the teslamate container, you somehow have all these files owned by root. They are owned by nonroot for me with a container created using the instructions at https://docs.teslamate.org/docs/installation/docker on debian 11. Is that different from what you are doing?
The command for the grafana container failed because the container was restarting. You could try it again but wait a little longer after starting it to avoid this error.
I tried stopping the container and starting it after the other, this is not working because of permission issue
I followed the instructions to the letter, you can see my install in the attached file is also running on a Debian 11 base : OMV.
So for some reason the user became root, how can I change it back to nonroot without recreate everything and loosing all my data?
teslamate.txt
look like for docker exec teslamate-teslamate-1 ls -la /opt/app/lib/tzdata-1.1.1/priv
the files owner is somehow fixed
total 28 drwxr-xr-x 1 nonroot nonroot 4096 Jan 13 13:47 . drwxr-xr-x 1 nonroot nonroot 4096 Jan 13 13:47 .. -rw-r--r-- 1 nonroot nonroot 9 Jan 13 18:17 latest_remote_poll.txt drwxr-xr-x 1 nonroot nonroot 4096 Jan 13 18:17 release_ets drwxr-xr-x 1 nonroot nonroot 4096 Jan 13 18:17 tmp_downloads
but beside this basicaly the same result, no fixed.
_teslamate-teslamate-1_logs(2).txt
_teslamate-mosquitto-1_logs(2).txt
_teslamate-grafana-1_logs(4).txt
_teslamate-database-1_logs(4).txt
|
gharchive/issue
| 2022-12-31T22:06:18 |
2025-04-01T04:55:51.077257
|
{
"authors": [
"cwanja",
"dyxyl",
"galosre"
],
"repo": "adriankumpf/teslamate",
"url": "https://github.com/adriankumpf/teslamate/issues/3035",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
748390701
|
Script de création de CSV en sortie de LiveCAM
Créer un script permettant de créer un CSV avec comme information la date, le nombre de personnes, le chemin de la photo et le nom du fichier
Ce CSV permettra de comparer le nombre de photo avec une (ou +) personne détectée et le nombre de photo total
Excellent, on va bien s'amuser à analyser ce fichier 🚀
@adriens LiveCAM Stats.xlsx
:bomb:
|
gharchive/issue
| 2020-11-23T00:35:52 |
2025-04-01T04:55:51.084854
|
{
"authors": [
"adriens",
"gbertherat"
],
"repo": "adriens/person-counter-api",
"url": "https://github.com/adriens/person-counter-api/issues/43",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1710488973
|
🛑 adv-compass-pmdevel is down
In 94ca810, adv-compass-pmdevel (https://api.github.com/orgs/advancedcsg/actions/runners/310) was down:
HTTP code: 200
Response time: 161 ms
Resolved: adv-compass-pmdevel is back up in 4e34375.
|
gharchive/issue
| 2023-05-15T17:07:37 |
2025-04-01T04:55:51.109161
|
{
"authors": [
"Abhyudaya1414"
],
"repo": "advancedcsg-open/toolchain-status-dashboard",
"url": "https://github.com/advancedcsg-open/toolchain-status-dashboard/issues/36",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1291889742
|
[BUG] Angle D3D11 glEnableVertexAttribArray() program state rendering error.
This bug seems like it's up to google angle to fix it but just to throw it out there it seems sometimes when a shader is switched, it's vertex attributes do not switch from the old one to the new one, and an OpenGL error 0x0502 error is thrown, this bug was discovered by making a mesh render in 2D queue setForce2DQueue() with a sprite that exists in the same scene but with a lower local z value than the 2D queued mesh.
Unexpected result: mesh will not be shown and an opengl error is thrown.
CCRenderer.cpp:320
When flush2D is called (which renders all sprites in a scene in a single batch to minimize draw calls), it calls drawElements() in the same sort order GLOBALZ_XXX as the mesh, they both set up their vertex attributes in:
CommandBufferGL.cpp:283
Now that triangle batch is drawn with it's according vertex attributes, and here comes the turn to drawMeshCommand() which does the same thing and sets up it's vertex attributes according to the program state just like before, BUT an opengl es angle error is thrown OpenGL error 0x0502 when the draw call is finally issued, which confused me as to how could a previous opengl state affect future states?
I don't know if that bug affects only D3D11 Angle or OpenGL ES devices and if both are affected then mobile devices are affected by this bug too.
Do you have a reproduce demo about this issue
Somehow I can't reproduce the bug again with d3d11...
The helloworld sprite seems to render fine behind the mesh of quads:
I think it's a vertex attribute problem, the bug is really confusing.
I'll close this issue for now.
|
gharchive/issue
| 2022-07-01T22:47:11 |
2025-04-01T04:55:51.149979
|
{
"authors": [
"DelinWorks",
"halx99"
],
"repo": "adxeproject/adxe",
"url": "https://github.com/adxeproject/adxe/issues/697",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1259325029
|
Command line arguments for Operator
In the docs https://github.com/adyanth/cloudflare-operator/blob/main/docs/configuration.md#operator it says you can add command line arguments like --overwrite-unmanaged-dns but it isn't clear to me where or how this should be done.
Thanks
Hey @acrogenesis,
It is to be added to the deployment's spec.template.spec.containers[1].args after --leader-elect. Currently, it is a manual step, but I am planning to worry about making it easier to configure when I start looking into Helm charts for it since there are very few meaningful configurable parameters for now.
Obligatory patch: (Note that it is not idempotent)
kubectl -n cloudflare-operator-system patch deploy cloudflare-operator-controller-manager --type=json -p='[{"op": "add", "path": "/spec/template/spec/containers/1/args/-", "value": "--overwrite-unmanaged-dns=true"}]'
Reference:
https://github.com/adyanth/cloudflare-operator/blob/702e042e33c1cf802c80a0e3aaf73b20e515409e/main.go#L60
That makes sense. Thanks!
|
gharchive/issue
| 2022-06-03T03:18:42 |
2025-04-01T04:55:51.153238
|
{
"authors": [
"acrogenesis",
"adyanth"
],
"repo": "adyanth/cloudflare-operator",
"url": "https://github.com/adyanth/cloudflare-operator/issues/61",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2125800393
|
[ENH] Loading unequal length, no missing values classification problems
This PR:
Fixes #1059 because I have reconfigured tsc.com to return 404 when file not found.
Adapts load_classification to match load_regression, so that the equal length and no missing value variants of problems can be loaded directly from the zip. Allows suffixes
_discr : discrete version of a regression problem
_eq : equal length version of a variable length
_nmv : no missing values version of set with missing values.
so dataset Foo_disc_eq_nmv_TEST.ts is a variant of the regression problem with a discretised target, made equal length (method may vary on dataset) and missing values imputed.
Removes the requirement for data to be on one of the data set lists (which are not really lists,
Changes the name of tsc_data_lists to tsc_datasets. They are sets not lists (for query reasons)
Adds a note book "data_unequal.ipynb" to demonstrate usage.
Allows for loading the discrete versions of regression problems
@TonyBagnall does this allow to load discrete version of regression problems using the load_classification or load_regression function ?
and second question, the equal length fix follows the padding zero or the 2nd bake off ? I think we already discussed but completely forgot which is why i am asking here to keep track :)
@TonyBagnall does this allow to load discrete version of regression problems using the load_classification or load_regression function ?
and second question, the equal length fix follows the padding zero or the 2nd bake off ? I think we already discussed but completely forgot which is why i am asking here to keep track :)
loading discrete versions with load_classification is the plan, but I think in future PR. Series padded with mean+small amount of noise
|
gharchive/pull-request
| 2024-02-08T18:13:41 |
2025-04-01T04:55:51.192106
|
{
"authors": [
"TonyBagnall",
"hadifawaz1999"
],
"repo": "aeon-toolkit/aeon",
"url": "https://github.com/aeon-toolkit/aeon/pull/1157",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
652198623
|
build: use graphql-tools to load files
Fixes failing integration tests in jobs/5744
Thanks for the fix, did anything to to make the test fail?
I noticed the build is failing in https://github.com/aerogear/graphback/pull/1616. So it is blocking dependency updates.
Though I see this is failing for a different reason now, will do some more checks.
I noticed the build is failing in #1616. So it is blocking dependency updates.
Though I see this is failing for a different reason now, will do some more checks.
Thanks for the quick fix!!
|
gharchive/pull-request
| 2020-07-07T10:16:29 |
2025-04-01T04:55:51.223361
|
{
"authors": [
"craicoverflow",
"machi1990"
],
"repo": "aerogear/graphback",
"url": "https://github.com/aerogear/graphback/pull/1664",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
908316950
|
enable VNC not working in Robot framework with selenoid
we were using Protractor for test automation and we have setup Selenoid for remote execution. Everything is working perfectly fine.
Now we are switching to Robot framework (Java version). I have setup the desired capabilities and remote url in "Open browser keyword"
Code in robot framework for launching tests on remote with desired capabilities
${desired caps} getCapabilities // capabilities are returned in this in string format with enablevnc as true
Open Browser ${SSO_URL} ${Browser} None http://0.0.0.0:4444/wd/hub ${desired caps}
Tests are executing on Selenoid but not able to see the execution, VNC flag is missing.
On the other hand i had done the setup in selenium 4 chrome standalone docker image and executed the robot tests on this image , i am able to see the execution even without the enable VNC flag. (tests are executed on noVNC).
Could some one help me in viewing the robot test (Java Port) execution on Selenoid.
@kiruparkuma see https://github.com/aerokube/selenoid/issues/554#issuecomment-432604950.
Trying to run Selenoid on my Robot Framework tests but I am unable to view Video or record it.
What am I doing wrong?
@madb1ker see my comment above.
@madb1ker see my comment above.
desired_capabilities has been deprecated and removed.
I can run my tests in Robot Framework with Selenoid, I just cannot find how to set enableVNC and enableVideo to True.
@madb1ker Selenoid supports W3C version of the protocol which is current standard, so Selenoid capabilities can go under selenoid:options key. Not sure how to add this in robot-framework. https://www.w3.org/TR/webdriver2/#extensions-0
@madb1ker Selenoid supports W3C version of the protocol which is current standard, so Selenoid capabilities can go under selenoid:options key. Not sure how to add this in robot-framework. https://www.w3.org/TR/webdriver2/#extensions-0
Thank you!
For the record this is the correct way to use Selenoid with Robot Framework passing the arguments in the browser options:
${selenoid_args} Create Dictionary enableVNC=${True} enableVideo=${True} Open Browser ${WMS_URL} ${BROWSER} remote_url=${SELENOID_SERVER} options=set_capability("selenoid:options",${selenoid_args})
Ok, closing then.
|
gharchive/issue
| 2021-06-01T13:51:42 |
2025-04-01T04:55:51.253846
|
{
"authors": [
"kiruparkuma",
"madb1ker",
"vania-pooh"
],
"repo": "aerokube/selenoid",
"url": "https://github.com/aerokube/selenoid/issues/1108",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1647301978
|
video rename failed cause no such file or dictionary
I started Selenoid using CM, and overwrote the default video storage path, but when a task ended, an error was reported that the video file was not found. The following is the output log of Selenoid:
···
023/03/30 09:50:24 [-] [INIT] [Using Docker API version: 1.41]
2023/03/30 09:50:24 [-] [INIT] [Timezone: UTC]
2023/03/30 09:50:24 [-] [INIT] [Listening on :4444]
2023/03/30 09:51:50 [-] [NEW_REQUEST] [unknown] [10.19.43.154]
2023/03/30 09:51:50 [-] [NEW_REQUEST_ACCEPTED] [unknown] [10.19.43.154]
2023/03/30 09:51:50 [0] [LOCATING_SERVICE] [chrome] [101.0]
2023/03/30 09:51:50 [0] [USING_DOCKER] [chrome] [101.0]
2023/03/30 09:51:50 [0] [CREATING_CONTAINER] [selenoid/vnc_chrome:101.0]
2023/03/30 09:51:50 [0] [STARTING_CONTAINER] [selenoid/vnc_chrome:101.0] [57a8fa68185b5b76751cb9adb93cd65f664a7b87031604e9d97ba4e5b89b70dc]
2023/03/30 09:51:50 [0] [CONTAINER_STARTED] [selenoid/vnc_chrome:101.0] [57a8fa68185b5b76751cb9adb93cd65f664a7b87031604e9d97ba4e5b89b70dc] [0.36s]
2023/03/30 09:51:50 [0] [CREATING_VIDEO_CONTAINER] [selenoid/video-recorder:latest-release]
2023/03/30 09:51:50 [0] [STARTING_VIDEO_CONTAINER] [selenoid/video-recorder:latest-release] [52d2a818239b4b3e3f17d9cdbc4017f7ba6c5a50d92339b4cdf00a95174b7376]
2023/03/30 09:51:51 [0] [VIDEO_CONTAINER_STARTED] [selenoid/video-recorder:latest-release] [52d2a818239b4b3e3f17d9cdbc4017f7ba6c5a50d92339b4cdf00a95174b7376] [0.57s]
2023/03/30 09:51:51 [0] [SERVICE_STARTED] [selenoid/vnc_chrome:101.0] [57a8fa68185b5b76751cb9adb93cd65f664a7b87031604e9d97ba4e5b89b70dc] [0.00s]
2023/03/30 09:51:51 [0] [PROXY_TO] [57a8fa68185b5b76751cb9adb93cd65f664a7b87031604e9d97ba4e5b89b70dc] [http://172.17.0.4:4444/]
2023/03/30 09:51:51 [0] [SESSION_ATTEMPTED] [http://172.17.0.4:4444/] [1]
2023/03/30 09:51:52 [0] [SESSION_CREATED] [a8535d93afac3bb6f2ceb066cd48fe46] [1] [1.75s]
2023/03/30 09:56:59 [4164] [SESSION_DELETED] [a8535d93afac3bb6f2ceb066cd48fe46]
2023/03/30 09:56:59 [0] [STOPPING_VIDEO_CONTAINER] [52d2a818239b4b3e3f17d9cdbc4017f7ba6c5a50d92339b4cdf00a95174b7376]
2023/03/30 09:57:00 [0] [REMOVING_CONTAINER] [52d2a818239b4b3e3f17d9cdbc4017f7ba6c5a50d92339b4cdf00a95174b7376]
2023/03/30 09:57:00 [0] [FAILED_TO_REMOVE_CONTAINER] [52d2a818239b4b3e3f17d9cdbc4017f7ba6c5a50d92339b4cdf00a95174b7376] [Error response from daemon: removal of container 52d2a818239b4b3e3f17d9cdbc4017f7ba6c5a50d92339b4cdf00a95174b7376 is already in progress]
2023/03/30 09:57:00 [0] [REMOVING_CONTAINER] [57a8fa68185b5b76751cb9adb93cd65f664a7b87031604e9d97ba4e5b89b70dc]
2023/03/30 09:57:00 [0] [CONTAINER_REMOVED] [57a8fa68185b5b76751cb9adb93cd65f664a7b87031604e9d97ba4e5b89b70dc]
2023/03/30 09:57:00 [0] [VIDEO_ERROR] [Failed to rename /data/apps/selenoid/video/selenoid4d7d695b45a07445ff6aa168a7e82297.mp4 to /data/apps/selenoid/video/chrome982Z.mp4: rename /data/apps/selenoid/video/selenoid4d7d695b45a07445ff6aa168a7e82297.mp4 /data/apps/selenoid/video/chrome982Z.mp4: no such file or directory]
···
The strange thing is that the video storage address I specified has this video. What is the reason?
[@jsy19-43-152 /data/apps/selenoid/video]# ls -ll
total 51300
-rw-r--r-- 1 root root 6061329 Mar 30 14:33 selenoid2f04e740e2924418e96f37c946892986.mp4
-rw-r--r-- 1 root root 5052357 Mar 30 16:30 selenoid3eafdd6088457654abfdae3f23754ade.mp4
-rw-r--r-- 1 root root 5030674 Mar 30 17:57 selenoid4d7d695b45a07445ff6aa168a7e82297.mp4
-rw-r--r-- 1 root root 15894839 Dec 9 17:22 selenoid529b1de5f2ab1b8791abbbca9801676e.mp4
-rw-r--r-- 1 root root 6139882 Mar 30 14:34 selenoid8459de847e5fb72108bead8abea1ec31.mp4
-rw-r--r-- 1 root root 14338547 Dec 13 12:56 selenoidf4c92d231db8b4a399d3a1e6260246cf.mp4
here is my cm command :
./cm selenoid start --config-dir /data/apps/selenoid/ --browsers-json /data/apps/selenoid/browsers.json --vnc --port 4444 --args "-limit 2 -timeout 3m -log-output-dir /data/apps/selenoid/logs/ -video-output-dir /data/apps/selenoid/video/ -session-attempt-timeout 2m -service-startup-timeout 2m -container-network bridge"
@suzhenyu006 just using --config-dir should be sufficient. No need to override -log-output-dir and other flags as these directories are inside container and --config-dir is outside.
|
gharchive/issue
| 2023-03-30T10:02:48 |
2025-04-01T04:55:51.265805
|
{
"authors": [
"suzhenyu006",
"vania-pooh"
],
"repo": "aerokube/selenoid",
"url": "https://github.com/aerokube/selenoid/issues/1327",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
158893486
|
How to hide/disable the Custom button in Color Palette
[yes ] I have verified there are no duplicate active or recent bugs, questions, or requests
I would like to remove the "custom" button in the Color Palette dialog. How can this be done?
Reason: I have an array of preset colors. I want the user to pick only the preset colors not a custom color using seekbar
Regards
Hi @CeJienAJPC,
this is not a bug you can deny the custom Color selection by set the Option "allowUserColorInput" to false.
builder.allowUserColorInput(false);
Best regards,
Vall0n
@Vall0n
Thank you for the solution. I know it ain't a bug. I was just looking if there was a way. Thanks again
|
gharchive/issue
| 2016-06-07T10:49:52 |
2025-04-01T04:55:51.337522
|
{
"authors": [
"CeJienAJPC",
"Vall0n"
],
"repo": "afollestad/material-dialogs",
"url": "https://github.com/afollestad/material-dialogs/issues/1094",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
477556612
|
Progress Dialog in v3
What module does this apply to?
core
Description what you'd like to happen:
Google clearly stated in its design guidelines that progress dialogs are deprecated and that they should not be used.
Despite this, almost every google app (Gmail, Drive, Sheets, Slides, Docs) make use of dialog progress bars to display downloads/uploads.
Since this library is opinionated it may be worth the effort packing a progress layout and creating an extension function that transforms the dialog into a progress dialog.
I don't want to include a progress dialog, cause I still prefer and want to encourage the use of inline progress indicators.
|
gharchive/issue
| 2019-08-06T19:32:33 |
2025-04-01T04:55:51.339544
|
{
"authors": [
"AndreaCioccarelli",
"afollestad"
],
"repo": "afollestad/material-dialogs",
"url": "https://github.com/afollestad/material-dialogs/issues/1862",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
97100095
|
How to set a global style for the Dialogs
I have different style per activity. When I use the Dialog, they show up with different colours in each activity .
Read the README on the front page.
|
gharchive/issue
| 2015-07-24T17:38:44 |
2025-04-01T04:55:51.341169
|
{
"authors": [
"MrThiago",
"afollestad"
],
"repo": "afollestad/material-dialogs",
"url": "https://github.com/afollestad/material-dialogs/issues/624",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
53951667
|
Additional fix for vibrating windows
Previously we only checked if the listview could scroll down, which was
incorrect when the listview had already been scrolled to the bottom
Release 0.6.0 is now available
Thanks @teslacoil and @hzsweers!
|
gharchive/pull-request
| 2015-01-10T08:46:05 |
2025-04-01T04:55:51.342358
|
{
"authors": [
"afollestad",
"teslacoil"
],
"repo": "afollestad/material-dialogs",
"url": "https://github.com/afollestad/material-dialogs/pull/179",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
130319423
|
Provide an option to disable animations
Maybe as a real Jenkins plugin that could be updated using the normal update center?
The problem for me is that our users can't reach fonts.googleapis.com and afonsof.github.io from where they access the Jenkins server.
Also it would be awesome if there were no animated graphics. Our clients have to be accessed using Remote Desktop and Firefox for example uses a ton of CPU to render the gif animations (It uses way less CPU for the CSS based animations) also since you use distinct shapes for yello-anime.gif and yellow.png the animations don't add anything.
If I'm not mistaken this theme is only using SVG files, not gifs: https://github.com/afonsof/jenkins-material-theme/tree/master/images
@jhasse you are right, the lack of gifs is a big reason for me to use this theme :)
With the gifs in the original jenkins theme the animations are necessary because they look identical if I were to disable gif animations in my browser.
With this theme however the animations are not needed because a Yellow and a Yellow animated (in progress or something) can be easily distinguished.
I hope this explains my reference to the gif files :)
Ah I understand :)
I guess it would be really good to make this theme into a dedicated Jenkins plugin then, since a configuration option could then be added to disable things like animations. I like them though! Not having any CPU issues in Firefox with them.
Well animations are one of those things...
IMO the animated items are of no interest to me - while they are animated I don't care about them, but they grab my attention. When the build is finally done and may be interesting because it failed or has test results - they are not animated anymore.
About that CPU use: Open 5 or more tabs, access your browser over Remote Desktop note CPU use multiply by the number of developers. Then cry for all the CPU cycles that your VM server has to spend to do really important stuff like animating icons... maybe you don't see it because your browser has access to one of those fancy GPUs (any actually because VMs don't have real GPUs). ;)
Guys, thanks for the suggestions! I guess both of them are interesting to implement. What about splitting it in 2 different issues?
Disable animation option
Created a dedicated plugin for this theme
Yes, good idea. I've published the Python plugin to Jenkins in the past, so I can help with the latter one :)
Done. Version 1.1.7
Awesome thanks!
|
gharchive/issue
| 2016-02-01T10:18:52 |
2025-04-01T04:55:51.349239
|
{
"authors": [
"afonsof",
"cobexer",
"jhasse"
],
"repo": "afonsof/jenkins-material-theme",
"url": "https://github.com/afonsof/jenkins-material-theme/issues/22",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
122858694
|
Ensure ?mode=vr forces stereo
https://aframe.io/aframe/examples/showcase-shopping/?mode=vr doesn't seem to force stereo. Perhaps we want ?mode=stereo, idk.
Per https://webvr.slack.com/archives/declarative-vr/p1450387507000072
Yeah, I think its useful to force stereo. You wouldn't be able to so otherwise unless you are on mobile or a VR enabled browser.
Also would be useful to have a function call to do the same.
And a postMessage call.
Hi all,
I'm looking for a solution to be able to force the vr view on smartphone when my webpage is launched.
Anyone can help me with that ?
Thank you in advance
I'm working on this now. I'll update the bug when it's ready.
|
gharchive/issue
| 2015-12-18T01:34:12 |
2025-04-01T04:55:51.355872
|
{
"authors": [
"caseyyee",
"cvan",
"evilgouky"
],
"repo": "aframevr/aframe-core",
"url": "https://github.com/aframevr/aframe-core/issues/704",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1473388798
|
⚠️ AGEG website (prod) has degraded performance
In b367389, AGEG website (prod) (https://www.ageg.ca/) experienced degraded performance:
HTTP code: 200
Response time: 1261 ms
Resolved: AGEG website (prod) performance has improved in 06438d0.
|
gharchive/issue
| 2022-12-02T20:06:46 |
2025-04-01T04:55:51.417404
|
{
"authors": [
"mdaoustUdeS"
],
"repo": "ageg-status/ageg-status.github.io",
"url": "https://github.com/ageg-status/ageg-status.github.io/issues/822",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
387704778
|
Start Error
Why i recive thies error ?
root@vps-zap59814x-4:~/SnapchatBot-master/examples# python reflectorbot.py -u noxxxxx -p 19wr#%xxxxxxxxx
[1] 21608
-bash: FeFOZyU9T@T2YRyUR%SxPz%Sg7vOCrdg5: command not found
root@vps-zap59814-4:~/SnapchatBot-master/examples# [2018-12-05 11:02:36,165] Starting new HTTPS connection (1): feelinsonice-hrd.appspot.com:443
[2018-12-05 11:02:36,536] https://feelinsonice-hrd.appspot.com:443 "POST /loq/login HTTP/1.1" 400 None
Traceback (most recent call last):
File "reflectorbot.py", line 22, in
bot = ReflectorBot(args.username, args.password)
File "build/bdist.linux-x86_64/egg/snapchat_bots/bot.py", line 26, in init
File "/usr/local/lib/python2.7/dist-packages/requests-2.20.1-py2.7.egg/requests/models.py", line 897, in json
return complexjson.loads(self.text, **kwargs)
File "/usr/lib/python2.7/json/init.py", line 339, in loads
return _default_decoder.decode(s)
File "/usr/lib/python2.7/json/decoder.py", line 364, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
File "/usr/lib/python2.7/json/decoder.py", line 382, in raw_decode
raise ValueError("No JSON object could be decoded")
ValueError: No JSON object could be decoded
I have the same error and without any idea how to fix it. Did anybody got a fix?
First line in the readme
"This repo is deprecated due to changes in Snapchat's unofficial API."
So there is no way to send snaps again ?
|
gharchive/issue
| 2018-12-05T11:04:50 |
2025-04-01T04:55:51.426362
|
{
"authors": [
"Nisutec",
"Smadger1",
"rhpo",
"rydeom"
],
"repo": "agermanidis/SnapchatBot",
"url": "https://github.com/agermanidis/SnapchatBot/issues/67",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
349797693
|
chapter 11 Question 8 restore_model parameter function ?
in chapter 11 Question 8 the code written to restore model parameter to value of best parameters in early stop is :
def _restore_model_params(self, model_params):
gvar_names = list(model_params.keys())
assign_ops = {gvar_name: self._graph.get_operation_by_name(gvar_name + "/Assign")
for gvar_name in gvar_names}
init_values = {gvar_name: assign_op.inputs[1] for gvar_name, assign_op in assign_ops.items()}
feed_dict = {init_values[gvar_name]: model_params[gvar_name] for gvar_name in gvar_names}
self._session.run(assign_ops, feed_dict=feed_dict)
I'm not able to understand this code properly .Can you please explain assign_ops and init_values code ?
Hi @nitml,
Good question. When you create a variable, TensorFlow automatically creates several operations and tensors, including an assignment operation to initialize the variable to whatever initialization value you chose for that variable.
The first line of the function gets the list of names of variables whose values we want to restore.
The next line gets a reference to all the assignment operations for these variables. Indeed, if a variable is named "V", then the assignment operation is called "V/Assign".
The third line gets a reference to the tensors that represent the initialization values for these variables. Indeed, an assignment operation has two inputs: (1) a reference to the variable, (2) the value to assign. Here, I am getting each assignment operation's second input (op.inputs[1]).
The I prepare a feed_dict to tell TensorFlow to use the given model_params values instead of the initialization values for each variable.
Then I run all the assignment operations, feeding the values to use instead of the default initialization values.
Here's a simpler example, to help you understand what's going on:
>>> import tensorflow as tf
>>> V = tf.Variable(1, name="V")
>>> graph = tf.get_default_graph()
>>> sess = tf.InteractiveSession()
>>> V.initializer.run()
>>> V.eval()
1
>>> assign_init_value_op = graph.get_operation_by_name("V/Assign")
>>> init_value = assign_init_value_op.inputs[1]
>>> sess.run(assign_init_value_op, feed_dict={init_value: 2})
>>> V.eval()
2
Hope this helps,
Aurélien
Great Explanation, i have another doubt in this Question 8 to use RandomizedSearchCV we created Sklearn learn compatible class DNNClassifier and then used following code :
`from sklearn.model_selection import RandomizedSearchCV
def leaky_relu(alpha=0.01):
def parametrized_leaky_relu(z, name=None):
return tf.maximum(alpha * z, z, name=name)
return parametrized_leaky_relu
param_distribs = {
"n_neurons": [10, 30, 50, 70, 90, 100, 120, 140, 160],
"batch_size": [10, 50, 100, 500],
"learning_rate": [0.01, 0.02, 0.05, 0.1],
"activation": [tf.nn.relu, tf.nn.elu, leaky_relu(alpha=0.01), leaky_relu(alpha=0.1)],
# you could also try exploring different numbers of hidden layers, different optimizers, etc.
#"n_hidden_layers": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10],
#"optimizer_class": [tf.train.AdamOptimizer, partial(tf.train.MomentumOptimizer, momentum=0.95)],
}
rnd_search = RandomizedSearchCV(DNNClassifier(random_state=42), param_distribs, n_iter=50,
random_state=42, verbose=2)
rnd_search.fit(X_train1, y_train1, X_valid=X_valid1, y_valid=y_valid1, n_epochs=1000)`
My question is why have we defined leaked_elu function such that in has another function parametrized_leaky_relu and not simply :
def relu_elu(z,alpha=0.01,name=None): return tf.maximum(alpha * z, z, name=name)
Hi @nitml ,
With your leaky_relu() function, how would you define the param_distribs dict? For example, the following would NOT work, because the leaky_relu() function would complain that the argument z is missing:
# THIS WILL NOT WORK
param_distribs = {
...
"activation": [tf.nn.relu, tf.nn.elu, leaky_relu(alpha=0.01), leaky_relu(alpha=0.1)],
...
}
Of course you could try to pass some inputs z, but which ones? We don't have that data yet, as it will only be computed during training, and it will be different for each training batch.
However, you could do this:
# THIS SHOULD WORK
param_distribs = {
...
"activation": [tf.nn.relu, tf.nn.elu, lambda z: leaky_relu(z, alpha=0.01), lambda z: leaky_relu(alpha=0.1)],
...
}
As you can see, the activation function needs to be, well, a function, not the result of calling the leaky_relu() function. My function creates a parametrized_leaky_relu() function that has the desired alpha hyperparameter already set to the desired value. So TensorFlow will just give it the input data, and it will compute the right thing. Your function directly returns a result, so it must be wrapped into a function, for example using a lambda.
I hope this is clear,
Aurélien
Thanks for your Time and explanation,but things are not properly clear,
suppose i want to use leaky_rule having parametrized_leaky_rule() function within it as activation function then my code will be :
Z=tf.layers.dense(X,20,name="hidden_layer2")
hidden2=leaky_rule(Z)`
then will not it be a error as leaky_rule function does not have parameter "z "as it is parameter of parametrized_leaky_rule() function and not leaky_rule.
In question 8 we used "dnn" function in "fit" function to make layers :
`def _dnn(self,inputs):
for layer in range(self.n_hidden_layers):
if self.dropout_rate:
inputs=tf.layers.dropout(inputs,self.dropout_rate,training=self.training)
inputs=tf.layers.dense(inputs,self.n_neurons,kernel_initializer=self.initializer,
name="hidden%d" % (layer + 1))
if self.batch_norm_momentum:
inputs=tf.layers.batch_normalization(inputs,momentum=self.batch_norm_momentum,training=self.training)
inputs = self.activation(inputs, name="hidden%d_out" % (layer + 1))
return inputs`
so when we use "RandomizedSearchCV(DNNClassifier" with activation function as "leaky_rule" then second last line "inputs = self.activation(inputs, name="hidden%d_out" % (layer + 1))" should rise error as "inputs" is parameter of "parametrized_leaky_rule()" function and not "leaky_rule".
i hope my question is clear ?
Thanks for all your Time and support !!
Hi @nitml ,
Sorry I wasn't clear. As you can see on the second to last line in the _dnn() method, the self.activation attribute needs to be a function that takes some inputs as its first argument, and also takes a name argument.
For example, I could use this function if I want alpha=0.01:
def leaky_relu_001(z, name=None):
return tf.maximum(0.01 * z, z, name=name)
Then I could just create a DNNClassifier like this:
dnn_clf = DNNClassifier(activation=leaky_relu_001, ...)
If I want to use alpha=0.1 instead, then I can create another function:
def leaky_relu_01(z, name=None):
return tf.maximum(0.1 * z, z, name=name)
I could then create a param_distribs dict like this:
param_distribs = {
...
"activation": [tf.nn.relu, tf.nn.elu, leaky_relu_001, leaky_relu_01],
...
}
And then create a RandomizedSearchCV using this param_distrib, and everything should work fine. Now instead of manually copy/pasting the code to write a new function every time I want a different alpha value, I could write a create_leaky_relu_function() function, that creates a leaky_relu() function for me, with the desired alpha value:
def create_leaky_relu_function(alpha):
def leaky_relu(z, name=None):
return tf.maximum(alpha * z, z, name=name)
return leaky_relu
Now if I want a new function for alpha=0.2, it is much easier to create:
leaky_relu_02 = create_leaky_relu_function(0.2)
Now this leaky_relu_02() function is a function that takes the inputs z and an optional name. I can use it when creating a DNNClassifier(activation=leaky_relu_02), and I can use it in param_distribs={"activation": [leaky_relu_02, ...], ...}.
The main difficulty is to understand that the create_leaky_relu_function() returns a new function with its own alpha value that will never change once the function is created. Perhaps this simple example may help:
>>> def create_multiplier(factor):
... def mul(x):
... return factor * x
... return mul
...
>>> mul2 = create_multiplier(2)
>>> mul3 = create_multiplier(3)
>>> mul2(10)
20
>>> mul3(10)
30
First I create a function that can be used to create functions. Then I use it to create two functions mul2() and mul3(). The first one has factor=2, so it multiplies its argument by 2, while the second has factor=3 so it multiplies its argument by 3.
Now let's look at your function:
def your_leaky_relu(z, alpha=0.01, name=None):
return tf.maximum(alpha * z, z, name=name)
It can indeed be used as you showed:
Z = tf.layers.dense(X, 20, name="hidden_layer2")
hidden2 = your_leaky_relu(Z)
You could even set alpha to any value you want:
Z = tf.layers.dense(X, 20, name="hidden_layer2")
hidden2 = your_leaky_relu(Z, alpha=0.2)
However, suppose you want to pass it to the DNNClassifier constructor, and you want alpha=0.2, how can you do that? If you write the following code, then alpha is not set, so it will default to 0.01:
dnn_clf = DNNClassifier(activation=your_leaky_relu, ...)
But if you write the following code, then you will get an error, because activation is supposed to be a function, not a tensor (which is what your function returns):
dnn_clf = DNNClassifier(activation=your_leaky_relu(alpha=0.2), ...) # ERROR
The error will actually not happen right away, but only when self.activation is used (when you fit the DNNClassifier).
So, as you can see, the problem is that there is no way to set the alpha argument in your function when you pass it to the DNNClassifier. This is why, if we want to try 3 different alpha values, we need 3 different functions, and this is why it is useful to have a function that creates them.
Hope this helps,
Aurélien
Great Explanation,
Thanks for your time and support
|
gharchive/issue
| 2018-08-12T09:17:23 |
2025-04-01T04:55:51.448160
|
{
"authors": [
"ageron",
"nitml"
],
"repo": "ageron/handson-ml",
"url": "https://github.com/ageron/handson-ml/issues/281",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
124548093
|
SELECT 1 bug
First bug to fix from here
SELECT 1 FROM t1 WHERE 1 IN (SELECT 1)
Error message:
context is not defined
Very strange error. It occurs from command line:
alasql 'SELECT 1 FROM @[1,2] WHERE 1 IN (SELECT 1)'
but does not inside the Mocha test....
Very strange error. It occurs from command line:
alasql 'SELECT 1 FROM @[1,2] WHERE 1 IN (SELECT 1)'
but does not inside the Mocha test....
Fixed. Hope now we can pass at least one SQLLOGICTEST file in2.test.
@mathiasrw May I ask you to run the test again?
Fixed. Hope now we can pass at least one SQLLOGICTEST file in2.test.
@mathiasrw May I ask you to run the test again?
Sure !
Sure !
ZAP - you have done it again!
Solution to this issue tested and confirmed...
The rest of the tests are still running. Ill update the repo as soon as they are done.
Regarding in2.test we get the following at the moment:
2/125 ./test/evidence/in2.test
Mimic sqlite
SELECT 1 FROM t1 WHERE 1 IN (SELECT 1,2)
No exception thrown
Fail found for statement setting up data so skipping rest of tests
☓ Ran 53 tests as sqlite
3 skipped
1 failed
92% was OK
Time: 465ms
ZAP - you have done it again!
Solution to this issue tested and confirmed...
The rest of the tests are still running. Ill update the repo as soon as they are done.
Regarding in2.test we get the following at the moment:
2/125 ./test/evidence/in2.test
Mimic sqlite
SELECT 1 FROM t1 WHERE 1 IN (SELECT 1,2)
No exception thrown
Fail found for statement setting up data so skipping rest of tests
☓ Ran 53 tests as sqlite
3 skipped
1 failed
92% was OK
Time: 465ms
So basically its expecting to get an error on the following 4 statements
statement error
SELECT 1 FROM t1 WHERE 1 IN (SELECT 1,2)
statement error
SELECT 1 FROM t1 WHERE 1 IN (SELECT x,y FROM t1)
statement error
SELECT 1 FROM t1 WHERE 1 IN (SELECT * FROM t1)
statement error
SELECT 1 FROM t1 WHERE 1 IN (SELECT min(x),max(x) FROM t1)
So basically its expecting to get an error on the following 4 statements
statement error
SELECT 1 FROM t1 WHERE 1 IN (SELECT 1,2)
statement error
SELECT 1 FROM t1 WHERE 1 IN (SELECT x,y FROM t1)
statement error
SELECT 1 FROM t1 WHERE 1 IN (SELECT * FROM t1)
statement error
SELECT 1 FROM t1 WHERE 1 IN (SELECT min(x),max(x) FROM t1)
I will check this tonight after LegoLand (try to get some inspiration there:) )...
I will check this tonight after LegoLand (try to get some inspiration there:) )...
No stress :)
No stress :)
Just remember that Lego is from Denmark )))
Just remember that Lego is from Denmark )))
It iiissss - and we are very proud of it. I have a feeling its not the one in Denmark you are at :)
It iiissss - and we are very proud of it. I have a feeling its not the one in Denmark you are at :)
https://github.com/agershun/alasql/blob/develop/TESTLOG.md is updated.
https://github.com/agershun/alasql/blob/develop/TESTLOG.md is updated.
|
gharchive/issue
| 2016-01-01T15:29:42 |
2025-04-01T04:55:51.462008
|
{
"authors": [
"agershun",
"mathiasrw"
],
"repo": "agershun/alasql",
"url": "https://github.com/agershun/alasql/issues/501",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
943810596
|
ModuleNotFoundError with welly version 0.4.10
I receive the error message below when trying to import welly version 0.4.10, I do not receive the same error with version 0.4.9:
ModuleNotFoundError Traceback (most recent call last)
/tf/eds_repos/eds-ppfg/ppfg_utils.py in <module>
7
8 import os
----> 9 import welly
10 import numpy as np
11 import matplotlib.pyplot as plt
/usr/local/lib/python3.6/dist-packages/welly/__init__.py in <module>
4 ==================
5 """
----> 6 from .project import Project
7 from .well import Well
8 from .header import Header
/usr/local/lib/python3.6/dist-packages/welly/project.py in <module>
16 from tqdm import tqdm
17
---> 18 from .well import Well, WellError
19 from . import utils
20 from .utils import deprecated
/usr/local/lib/python3.6/dist-packages/welly/well.py in <module>
23 from .curve import Curve
24 from .header import Header
---> 25 from .location import Location
26 from .synthetic import Synthetic
27 from .canstrat import well_to_card_1
/usr/local/lib/python3.6/dist-packages/welly/location.py in <module>
19 from .fields import dev_fields
20 from .crs import CRS
---> 21 from .tools import compute_position_log
22
23
/usr/local/lib/python3.6/dist-packages/welly/tools.py in <module>
6 """
7 import numpy as np
----> 8 import wellpathpy as wp
9 import warnings
10
ModuleNotFoundError: No module named 'wellpathpy'```
@kwinkunks @EvanBianco PR #162 is a fix for this
|
gharchive/issue
| 2021-07-13T20:41:25 |
2025-04-01T04:55:51.486309
|
{
"authors": [
"astroChance",
"kinverarity1"
],
"repo": "agile-geoscience/welly",
"url": "https://github.com/agile-geoscience/welly/issues/165",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
945563392
|
DRY project name
email handler uses message_details.project_name to get the project name from distutils. we also set the project name as part of the Supervisor object, but there is no connection between these two. either link or remove need.
To me it sounds simpler to use the project name from distutils. Could the object sent to supervisor be optional and an override?
|
gharchive/issue
| 2021-07-15T16:30:42 |
2025-04-01T04:55:51.510839
|
{
"authors": [
"jacobdadams",
"steveoh"
],
"repo": "agrc/supervisor",
"url": "https://github.com/agrc/supervisor/issues/8",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2102889573
|
Calculation "Time to solution + Progress"
Hello, I would like to ask for support to calculate the time for solution + progress that is shown on the GLPI calls page, I still haven't been able to concatenate the data via the database and arrive at the same calculation used by GLPI, what bases and parameters could I use?
Note: in my calendar all days and times are useful.
|
gharchive/issue
| 2024-01-26T20:55:06 |
2025-04-01T04:55:51.520407
|
{
"authors": [
"izaacmenezes"
],
"repo": "agrorural/GLPI",
"url": "https://github.com/agrorural/GLPI/issues/2",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2132737502
|
Request: add GNU patch
Having patch would be useful on all platforms. https://savannah.gnu.org/projects/patch/
cli.zip https://github.com/ahgamut/superconfigure/releases/tag/z0.0.33 has patch.
|
gharchive/issue
| 2024-02-13T16:36:48 |
2025-04-01T04:55:51.558926
|
{
"authors": [
"ahgamut",
"mattyclarkson"
],
"repo": "ahgamut/superconfigure",
"url": "https://github.com/ahgamut/superconfigure/issues/25",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
}
|
565289031
|
Typed channels support
introduced a new method FromTypedChannel()
From() calls FromChannel() if channel is of type chan interface{} and FromTypedChannel for all other channels
benchmark for FromTypedChannel() and FromChannel() performance comparison
calls FromChannel() if channel is of type chan interface{}
but can't all chans be casted to chan interface{} ?
calls FromChannel() if channel is of type chan interface{}
but can't all chans be casted to chan interface{} ?
No, I don't think so. You can think of chan as a type constructor that is parameterized by some inner type and it is invariant. So, chan interface{} and chan int are totally different types.
Disclaimer. I am not 100% sure, since I am not a Go developer in real life :) So, there might be a way.
@ahmetb I've remembered that we use a naming convention of -T() for the methods that identify user types with reflection, e.g. WhereT(predicateFn interface{}) vs Where(predicate func(interface{}) bool). So, probably the right name for FromTypedChannel() is FromChannelT(). What do you think?
Yeah we currently don’t have FromXxxT, but this would reinforce the pattern. FromChannelT is good. Please make sure we explain the difference in its godoc.
Yeah we currently don’t have FromXxxT, but this would reinforce the pattern. FromChannelT is good. Please make sure we explain the difference in its godoc.
@ahmetb done.
|
gharchive/pull-request
| 2020-02-14T12:10:28 |
2025-04-01T04:55:51.594580
|
{
"authors": [
"ahmetb",
"kalaninja"
],
"repo": "ahmetb/go-linq",
"url": "https://github.com/ahmetb/go-linq/pull/90",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1343602593
|
[Idea] Allow configuration of pattern groups through a config file
What if there was a file that holds common queries, in a rc file somewhere on the home dir:
e.g.
context-groups:
prod-new:
include: "/prod/"
exclude: "^/legacy/"
stg:
include: "/stg/"
$ kubectl allctx -g prod -- get nodes # g used as a flag for context group
It is a great idea! I will try to implement this feature and raise a PR.
No need.
This is basically just a note so that I don't forget someday later. It doesn't mean I'm interested in adding this.
|
gharchive/issue
| 2022-08-18T20:20:47 |
2025-04-01T04:55:51.596337
|
{
"authors": [
"ahmetb",
"furkanpehli1"
],
"repo": "ahmetb/kubectl-allctx",
"url": "https://github.com/ahmetb/kubectl-allctx/issues/6",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
803096234
|
e621 doesn't work?
ahoviewer won't let me add it.
That booru doesn't seem to support an XML API, but provides a JSON one.
This could be supported with a plugin similar to what I created for Sankaku.
Gonna close this, and recommend you open a request issue over at ahoviewer-plugins repo.
|
gharchive/issue
| 2021-02-08T00:08:43 |
2025-04-01T04:55:51.598455
|
{
"authors": [
"ahodesuka",
"blissfulVoid"
],
"repo": "ahodesuka/ahoviewer",
"url": "https://github.com/ahodesuka/ahoviewer/issues/115",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
379523556
|
tres compiler error in TDNSResolver::getResponse
I'm on Ubuntu 16.04 with "gcc (Ubuntu 5.4.0-6ubuntu1~16.04.10) 5.4.0 20160609".
Since 947908cb565cc4fb72517329a23b76ebd15eeeed make tres fails with this:
$ nice make tres
g++ -std=gnu++14 -Wall -O2 -MMD -MP -ggdb -Iext/simplesocket -Iext/simplesocket/ext/fmt-5.2.1/include -Iext/ -pthread -c -o tres.o tres.cc
tres.cc: In member function ‘DNSMessageReader TDNSResolver::getResponse(const ComboAddress&, const DNSName&, const DNSType&, int)’:
tres.cc:129:32: error: converting to ‘const key_type {aka const std::tuple<ComboAddress, DNSName, DNSType>}’ from initializer list would use explicit constructor ‘constexpr std::tuple< <template-parameter-1-1> >::tuple(const _Elements& ...) [with _Elements = {ComboAddress, DNSName, DNSType}]’
if(skips.count({server,dn,dt})) {
^
tres.cc:185:36: error: converting to ‘std::set<std::tuple<ComboAddress, DNSName, DNSType> >::value_type {aka std::tuple<ComboAddress, DNSName, DNSType>}’ from initializer list would use explicit constructor ‘constexpr std::tuple< <template-parameter-1-1> >::tuple(const _Elements& ...) [with _Elements = {ComboAddress, DNSName, DNSType}]’
skips.insert({server,dn,dt});
^
<builtin>: recipe for target 'tres.o' failed
make: *** [tres.o] Error 1
I will address this - some c++17-isms have crept in. For now, replace {server, dn, dt} by std::tie(server, dn, dt)
since travis can compile tdns again, this should be resolved. Can you check?
Hmm. It compiles, but the output is mostly on one line. And it segfaults at the end.
$ ./tres mattnordhoff.net A
Retrieved . NSSET from hints, have 26 addresses
Result of query for mattnordhoff.net.|A (0 intermediate, 2 actual)
mattnordhoff.net. 300 A 104.18.56.175
mattnordhoff.net. 300 A 104.18.57.175
Used 6 queries
sh: 1: dot: not found
{"answer":[{"content":"104.18.56.175","name":"mattnordhoff.net.","ttl":300,"type":"A"},{"content":"104.18.57.175","name":"mattnordhoff.net.","ttl":300,"type":"A"}],"dot":"","intermediate":[],"msec":143,"name":"mattnordhoff.net.","numformerrs":0,"numqueries":6,"numtimeouts":0,"rcode":0,"trace":"mattnordhoff.net.|A Starting query at authority = ., have 26 addresses to try\nmattnordhoff.net.|A Sending to server k.root-servers.net. on 2001:7fd::1\nmattnordhoff.net.|A Received a 838 byte response with RCode Noerror, qname mattnordhoff.net., qtype A, aa: 0\nmattnordhoff.net.|A Authority net. IN NS 172800 c.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 m.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 d.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 f.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 k.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 e.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 a.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 i.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 h.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 g.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 l.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 b.gtld-servers.net.\nmattnordhoff.net.|A Authority net. IN NS 172800 j.gtld-servers.net.\nmattnordhoff.net.|A Additional a.gtld-servers.net. IN AAAA 172800 2001:503:a83e::2:30\nmattnordhoff.net.|A Additional b.gtld-servers.net. IN AAAA 172800 2001:503:231d::2:30\nmattnordhoff.net.|A Additional c.gtld-servers.net. IN AAAA 172800 2001:503:83eb::30\nmattnordhoff.net.|A Additional d.gtld-servers.net. IN AAAA 172800 2001:500:856e::30\nmattnordhoff.net.|A Additional e.gtld-servers.net. IN AAAA 172800 2001:502:1ca1::30\nmattnordhoff.net.|A Additional f.gtld-servers.net. IN AAAA 172800 2001:503:d414::30\nmattnordhoff.net.|A Additional g.gtld-servers.net. IN AAAA 172800 2001:503:eea3::30\nmattnordhoff.net.|A Additional h.gtld-servers.net. IN AAAA 172800 2001:502:8cc::30\nmattnordhoff.net.|A Additional i.gtld-servers.net. IN AAAA 172800 2001:503:39c1::30\nmattnordhoff.net.|A Additional j.gtld-servers.net. IN AAAA 172800 2001:502:7094::30\nmattnordhoff.net.|A Additional k.gtld-servers.net. IN AAAA 172800 2001:503:d2d::30\nmattnordhoff.net.|A Additional l.gtld-servers.net. IN AAAA 172800 2001:500:d937::30\nmattnordhoff.net.|A Additional m.gtld-servers.net. IN AAAA 172800 2001:501:b1f9::30\nmattnordhoff.net.|A Additional a.gtld-servers.net. IN A 172800 192.5.6.30\nmattnordhoff.net.|A Additional b.gtld-servers.net. IN A 172800 192.33.14.30\nmattnordhoff.net.|A Additional c.gtld-servers.net. IN A 172800 192.26.92.30\nmattnordhoff.net.|A Additional d.gtld-servers.net. IN A 172800 192.31.80.30\nmattnordhoff.net.|A Additional e.gtld-servers.net. IN A 172800 192.12.94.30\nmattnordhoff.net.|A Additional f.gtld-servers.net. IN A 172800 192.35.51.30\nmattnordhoff.net.|A Additional g.gtld-servers.net. IN A 172800 192.42.93.30\nmattnordhoff.net.|A Additional h.gtld-servers.net. IN A 172800 192.54.112.30\nmattnordhoff.net.|A Additional i.gtld-servers.net. IN A 172800 192.43.172.30\nmattnordhoff.net.|A Additional j.gtld-servers.net. IN A 172800 192.48.79.30\nmattnordhoff.net.|A Additional k.gtld-servers.net. IN A 172800 192.52.178.30\nmattnordhoff.net.|A Additional l.gtld-servers.net. IN A 172800 192.41.162.30\nmattnordhoff.net.|A Additional m.gtld-servers.net. IN A 172800 192.55.83.30\nmattnordhoff.net.|A Additional . IN OPT 0 \\# 0\nmattnordhoff.net.|A We got delegated to 13 net. nameserver names \nmattnordhoff.net.|A Have 26 IP addresses to iterate to: a.gtld-servers.net.=2001:503:a83e::2:30 a.gtld-servers.net.=192.5.6.30 b.gtld-servers.net.=2001:503:231d::2:30 b.gtld-servers.net.=192.33.14.30 c.gtld-servers.net.=2001:503:83eb::30 c.gtld-servers.net.=192.26.92.30 d.gtld-servers.net.=2001:500:856e::30 d.gtld-servers.net.=192.31.80.30 e.gtld-servers.net.=2001:502:1ca1::30 e.gtld-servers.net.=192.12.94.30 f.gtld-servers.net.=2001:503:d414::30 f.gtld-servers.net.=192.35.51.30 g.gtld-servers.net.=2001:503:eea3::30 g.gtld-servers.net.=192.42.93.30 h.gtld-servers.net.=2001:502:8cc::30 h.gtld-servers.net.=192.54.112.30 i.gtld-servers.net.=2001:503:39c1::30 i.gtld-servers.net.=192.43.172.30 j.gtld-servers.net.=2001:502:7094::30 j.gtld-servers.net.=192.48.79.30 k.gtld-servers.net.=2001:503:d2d::30 k.gtld-servers.net.=192.52.178.30 l.gtld-servers.net.=2001:500:d937::30 l.gtld-servers.net.=192.41.162.30 m.gtld-servers.net.=2001:501:b1f9::30 m.gtld-servers.net.=192.55.83.30 \n mattnordhoff.net.|A Starting query at authority = net., have 26 addresses to try\n mattnordhoff.net.|A Sending to server b.gtld-servers.net. on 192.33.14.30\n mattnordhoff.net.|A Received a 188 byte response with RCode Noerror, qname mattnordhoff.net., qtype A, aa: 0\n mattnordhoff.net.|A Authority mattnordhoff.net. IN NS 172800 cody.ns.cloudflare.com.\n mattnordhoff.net.|A Authority mattnordhoff.net. IN NS 172800 dana.ns.cloudflare.com.\n mattnordhoff.net.|A Additional cody.ns.cloudflare.com. IN A 172800 173.245.59.107\n mattnordhoff.net.|A Not accepting IP address of cody.ns.cloudflare.com.: out of authority of this server\n mattnordhoff.net.|A Additional cody.ns.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::adf5:3b6b\n mattnordhoff.net.|A Not accepting IP address of cody.ns.cloudflare.com.: out of authority of this server\n mattnordhoff.net.|A Additional dana.ns.cloudflare.com. IN A 172800 173.245.58.105\n mattnordhoff.net.|A Not accepting IP address of dana.ns.cloudflare.com.: out of authority of this server\n mattnordhoff.net.|A Additional dana.ns.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::adf5:3a69\n mattnordhoff.net.|A Not accepting IP address of dana.ns.cloudflare.com.: out of authority of this server\n mattnordhoff.net.|A Additional . IN OPT 0 \\# 0\n mattnordhoff.net.|A We got delegated to 2 mattnordhoff.net. nameserver names \n mattnordhoff.net.|A Don't have a resolved nameserver to ask anymore, trying to resolve 2 names\n mattnordhoff.net.|A Attempting to resolve NS dana.ns.cloudflare.com.|A\n dana.ns.cloudflare.com.|A Starting query at authority = ., have 26 addresses to try\n dana.ns.cloudflare.com.|A Sending to server f.root-servers.net. on 2001:500:2f::f\n dana.ns.cloudflare.com.|A Received a 847 byte response with RCode Noerror, qname dana.ns.cloudflare.com., qtype A, aa: 0\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 l.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 b.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 c.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 d.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 e.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 f.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 g.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 a.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 h.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 i.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 j.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 k.gtld-servers.net.\n dana.ns.cloudflare.com.|A Authority com. IN NS 172800 m.gtld-servers.net.\n dana.ns.cloudflare.com.|A Additional l.gtld-servers.net. IN A 172800 192.41.162.30\n dana.ns.cloudflare.com.|A Additional l.gtld-servers.net. IN AAAA 172800 2001:500:d937::30\n dana.ns.cloudflare.com.|A Additional b.gtld-servers.net. IN A 172800 192.33.14.30\n dana.ns.cloudflare.com.|A Additional b.gtld-servers.net. IN AAAA 172800 2001:503:231d::2:30\n dana.ns.cloudflare.com.|A Additional c.gtld-servers.net. IN A 172800 192.26.92.30\n dana.ns.cloudflare.com.|A Additional c.gtld-servers.net. IN AAAA 172800 2001:503:83eb::30\n dana.ns.cloudflare.com.|A Additional d.gtld-servers.net. IN A 172800 192.31.80.30\n dana.ns.cloudflare.com.|A Additional d.gtld-servers.net. IN AAAA 172800 2001:500:856e::30\n dana.ns.cloudflare.com.|A Additional e.gtld-servers.net. IN A 172800 192.12.94.30\n dana.ns.cloudflare.com.|A Additional e.gtld-servers.net. IN AAAA 172800 2001:502:1ca1::30\n dana.ns.cloudflare.com.|A Additional f.gtld-servers.net. IN A 172800 192.35.51.30\n dana.ns.cloudflare.com.|A Additional f.gtld-servers.net. IN AAAA 172800 2001:503:d414::30\n dana.ns.cloudflare.com.|A Additional g.gtld-servers.net. IN A 172800 192.42.93.30\n dana.ns.cloudflare.com.|A Additional g.gtld-servers.net. IN AAAA 172800 2001:503:eea3::30\n dana.ns.cloudflare.com.|A Additional a.gtld-servers.net. IN A 172800 192.5.6.30\n dana.ns.cloudflare.com.|A Additional a.gtld-servers.net. IN AAAA 172800 2001:503:a83e::2:30\n dana.ns.cloudflare.com.|A Additional h.gtld-servers.net. IN A 172800 192.54.112.30\n dana.ns.cloudflare.com.|A Additional h.gtld-servers.net. IN AAAA 172800 2001:502:8cc::30\n dana.ns.cloudflare.com.|A Additional i.gtld-servers.net. IN A 172800 192.43.172.30\n dana.ns.cloudflare.com.|A Additional i.gtld-servers.net. IN AAAA 172800 2001:503:39c1::30\n dana.ns.cloudflare.com.|A Additional j.gtld-servers.net. IN A 172800 192.48.79.30\n dana.ns.cloudflare.com.|A Additional j.gtld-servers.net. IN AAAA 172800 2001:502:7094::30\n dana.ns.cloudflare.com.|A Additional k.gtld-servers.net. IN A 172800 192.52.178.30\n dana.ns.cloudflare.com.|A Additional k.gtld-servers.net. IN AAAA 172800 2001:503:d2d::30\n dana.ns.cloudflare.com.|A Additional m.gtld-servers.net. IN A 172800 192.55.83.30\n dana.ns.cloudflare.com.|A Additional m.gtld-servers.net. IN AAAA 172800 2001:501:b1f9::30\n dana.ns.cloudflare.com.|A Additional . IN OPT 0 \\# 0\n dana.ns.cloudflare.com.|A We got delegated to 13 com. nameserver names \n dana.ns.cloudflare.com.|A Have 26 IP addresses to iterate to: a.gtld-servers.net.=192.5.6.30 a.gtld-servers.net.=2001:503:a83e::2:30 b.gtld-servers.net.=192.33.14.30 b.gtld-servers.net.=2001:503:231d::2:30 c.gtld-servers.net.=192.26.92.30 c.gtld-servers.net.=2001:503:83eb::30 d.gtld-servers.net.=192.31.80.30 d.gtld-servers.net.=2001:500:856e::30 e.gtld-servers.net.=192.12.94.30 e.gtld-servers.net.=2001:502:1ca1::30 f.gtld-servers.net.=192.35.51.30 f.gtld-servers.net.=2001:503:d414::30 g.gtld-servers.net.=192.42.93.30 g.gtld-servers.net.=2001:503:eea3::30 h.gtld-servers.net.=192.54.112.30 h.gtld-servers.net.=2001:502:8cc::30 i.gtld-servers.net.=192.43.172.30 i.gtld-servers.net.=2001:503:39c1::30 j.gtld-servers.net.=192.48.79.30 j.gtld-servers.net.=2001:502:7094::30 k.gtld-servers.net.=192.52.178.30 k.gtld-servers.net.=2001:503:d2d::30 l.gtld-servers.net.=192.41.162.30 l.gtld-servers.net.=2001:500:d937::30 m.gtld-servers.net.=192.55.83.30 m.gtld-servers.net.=2001:501:b1f9::30 \n dana.ns.cloudflare.com.|A Starting query at authority = com., have 26 addresses to try\n dana.ns.cloudflare.com.|A Sending to server b.gtld-servers.net. on 2001:503:231d::2:30\n dana.ns.cloudflare.com.|A Received a 581 byte response with RCode Noerror, qname dana.ns.cloudflare.com., qtype A, aa: 0\n dana.ns.cloudflare.com.|A Authority cloudflare.com. IN NS 172800 ns3.cloudflare.com.\n dana.ns.cloudflare.com.|A Authority cloudflare.com. IN NS 172800 ns5.cloudflare.com.\n dana.ns.cloudflare.com.|A Authority cloudflare.com. IN NS 172800 ns4.cloudflare.com.\n dana.ns.cloudflare.com.|A Authority cloudflare.com. IN NS 172800 ns6.cloudflare.com.\n dana.ns.cloudflare.com.|A Authority cloudflare.com. IN NS 172800 ns7.cloudflare.com.\n dana.ns.cloudflare.com.|A Additional ns3.cloudflare.com. IN A 172800 162.159.0.33\n dana.ns.cloudflare.com.|A Additional ns3.cloudflare.com. IN A 172800 162.159.7.226\n dana.ns.cloudflare.com.|A Additional ns3.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:21\n dana.ns.cloudflare.com.|A Additional ns3.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:7e2\n dana.ns.cloudflare.com.|A Additional ns5.cloudflare.com. IN A 172800 162.159.2.9\n dana.ns.cloudflare.com.|A Additional ns5.cloudflare.com. IN A 172800 162.159.9.55\n dana.ns.cloudflare.com.|A Additional ns5.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:209\n dana.ns.cloudflare.com.|A Additional ns5.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:937\n dana.ns.cloudflare.com.|A Additional ns4.cloudflare.com. IN A 172800 162.159.1.33\n dana.ns.cloudflare.com.|A Additional ns4.cloudflare.com. IN A 172800 162.159.8.55\n dana.ns.cloudflare.com.|A Additional ns4.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:121\n dana.ns.cloudflare.com.|A Additional ns4.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:837\n dana.ns.cloudflare.com.|A Additional ns6.cloudflare.com. IN A 172800 162.159.3.11\n dana.ns.cloudflare.com.|A Additional ns6.cloudflare.com. IN A 172800 162.159.5.6\n dana.ns.cloudflare.com.|A Additional ns6.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:30b\n dana.ns.cloudflare.com.|A Additional ns6.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:506\n dana.ns.cloudflare.com.|A Additional ns7.cloudflare.com. IN A 172800 162.159.4.8\n dana.ns.cloudflare.com.|A Additional ns7.cloudflare.com. IN A 172800 162.159.6.6\n dana.ns.cloudflare.com.|A Additional ns7.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:408\n dana.ns.cloudflare.com.|A Additional ns7.cloudflare.com. IN AAAA 172800 2400:cb00:2049:1::a29f:606\n dana.ns.cloudflare.com.|A Additional . IN OPT 0 \\# 0\n dana.ns.cloudflare.com.|A We got delegated to 5 cloudflare.com. nameserver names \n dana.ns.cloudflare.com.|A Have 20 IP addresses to iterate to: ns3.cloudflare.com.=162.159.0.33 ns3.cloudflare.com.=162.159.7.226 ns3.cloudflare.com.=2400:cb00:2049:1::a29f:21 ns3.cloudflare.com.=2400:cb00:2049:1::a29f:7e2 ns4.cloudflare.com.=162.159.1.33 ns4.cloudflare.com.=162.159.8.55 ns4.cloudflare.com.=2400:cb00:2049:1::a29f:121 ns4.cloudflare.com.=2400:cb00:2049:1::a29f:837 ns5.cloudflare.com.=162.159.2.9 ns5.cloudflare.com.=162.159.9.55 ns5.cloudflare.com.=2400:cb00:2049:1::a29f:209 ns5.cloudflare.com.=2400:cb00:2049:1::a29f:937 ns6.cloudflare.com.=162.159.3.11 ns6.cloudflare.com.=162.159.5.6 ns6.cloudflare.com.=2400:cb00:2049:1::a29f:30b ns6.cloudflare.com.=2400:cb00:2049:1::a29f:506 ns7.cloudflare.com.=162.159.4.8 ns7.cloudflare.com.=162.159.6.6 ns7.cloudflare.com.=2400:cb00:2049:1::a29f:408 ns7.cloudflare.com.=2400:cb00:2049:1::a29f:606 \n dana.ns.cloudflare.com.|A Starting query at authority = cloudflare.com., have 20 addresses to try\n dana.ns.cloudflare.com.|A Sending to server ns6.cloudflare.com. on 2400:cb00:2049:1::a29f:506\n dana.ns.cloudflare.com.|A Received a 67 byte response with RCode Noerror, qname dana.ns.cloudflare.com., qtype A, aa: 1\n dana.ns.cloudflare.com.|A Answer says it is authoritative!\n dana.ns.cloudflare.com.|A Answer dana.ns.cloudflare.com. IN A 86400 173.245.58.105\n dana.ns.cloudflare.com.|A We got an answer to our question!\n dana.ns.cloudflare.com.|A Additional . IN OPT 0 \\# 0\n dana.ns.cloudflare.com.|A Done, returning 1 results, 0 intermediate\n mattnordhoff.net.|A Got 1 nameserver A addresses, adding to list\n mattnordhoff.net.|A We now have 1 resolved A addresses to try\n mattnordhoff.net.|A Starting query at authority = mattnordhoff.net., have 1 addresses to try\n mattnordhoff.net.|A Sending to server dana.ns.cloudflare.com. on 173.245.58.105\n mattnordhoff.net.|A Received a 77 byte response with RCode Noerror, qname mattnordhoff.net., qtype A, aa: 1\n mattnordhoff.net.|A Answer says it is authoritative!\n mattnordhoff.net.|A Answer mattnordhoff.net. IN A 300 104.18.56.175\n mattnordhoff.net.|A We got an answer to our question!\n mattnordhoff.net.|A Answer mattnordhoff.net. IN A 300 104.18.57.175\n mattnordhoff.net.|A We got an answer to our question!\n mattnordhoff.net.|A Additional . IN OPT 0 \\# 0\n mattnordhoff.net.|A Done, returning 2 results, 0 intermediate\n","type":"A"}
Segmentation fault
(this issue was fixed in e3891625e4c98c00cad97660a391104c73a767f8 )
|
gharchive/issue
| 2018-11-11T13:38:21 |
2025-04-01T04:55:51.608430
|
{
"authors": [
"ahupowerdns",
"mnordhoff"
],
"repo": "ahupowerdns/hello-dns",
"url": "https://github.com/ahupowerdns/hello-dns/issues/38",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1330611115
|
Fix bug in shell meridional transect
Short description of why the PR is needed and how it satisfies those requirements, in sentence form.
(Delete unused sections)
Added public API:
symbol (e.g. vcm.my_function) or script and optional description of changes or why they are needed
Can group multiple related symbols on a single bullet
Refactored public API:
Bulleted list of removed or refactored symbols, such as changes to name, type, behavior, argument, etc. Be cautious about doing these and discuss with team more broadly.
Significant internal changes:
Bulleted list of changes to non-public API
Requirement changes:
Bulleted list, if relevant, of any changes to setup.py, requirement.txt, environment.yml, etc
[ ] Ran make lock_deps/lock_pip following these instructions
[ ] Add PR review with license info for any additions to constraints.txt
(example)
[ ] Tests added
Resolves # [JIRA-TAG]
Coverage reports (updated automatically):
Need to add test
|
gharchive/pull-request
| 2022-08-06T02:11:41 |
2025-04-01T04:55:51.629516
|
{
"authors": [
"nbren12"
],
"repo": "ai2cm/fv3net",
"url": "https://github.com/ai2cm/fv3net/pull/1990",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
174928442
|
Request - beetroot to biofuel (1.10)
Beet roots should make perfectly good biofuel, right? Currently fairly useless (unless you like soup) but plentiful from villages.
Ditto for beet seeds.
Can't remember if I recommended it before but it would be nice to be able to whitelist other mods' items for crushing into biofuel, too.
Added, thanks!
|
gharchive/issue
| 2016-09-04T02:55:18 |
2025-04-01T04:55:51.657783
|
{
"authors": [
"aidancbrady",
"rdunham3"
],
"repo": "aidancbrady/Mekanism",
"url": "https://github.com/aidancbrady/Mekanism/issues/3736",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.