id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
1741080150
|
Update sbt-github-actions to 0.15.0
About this PR
📦 Updates com.codecommit:sbt-github-actions from 0.14.2 to 0.15.0
Usage
✅ Please merge!
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
🔍 Files still referring to the old version number
The following files still refer to the old version number (0.14.2).
You might want to review and update them manually.
project/Dependencies.scala
⚙ Adjust future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "com.codecommit", artifactId = "sbt-github-actions" } ]
Or, add this to slow down future updates of this dependency:
dependencyOverrides = [{
pullRequests = { frequency = "30 days" },
dependency = { groupId = "com.codecommit", artifactId = "sbt-github-actions" }
}]
labels: sbt-plugin-update, early-semver-major, semver-spec-minor, old-version-remains, commit-count:n:2
Superseded by #195.
|
gharchive/pull-request
| 2023-06-05T06:35:11 |
2025-04-01T06:40:40.536213
|
{
"authors": [
"scala-steward"
],
"repo": "trace4cats/trace4cats-dynamic-extras",
"url": "https://github.com/trace4cats/trace4cats-dynamic-extras/pull/160",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1994656026
|
Update bundler
What?
Update bundler to remove warning "`DidYouMean::SPELL_CHECKERS.merge!(error_name => spell_checker)' has been deprecated.", which is present many times when running spec tests.
Why?
To stop using the deprecated method.
|
gharchive/pull-request
| 2023-11-15T12:09:51 |
2025-04-01T06:40:40.541104
|
{
"authors": [
"alexdesi"
],
"repo": "trade-tariff/trade-tariff-frontend",
"url": "https://github.com/trade-tariff/trade-tariff-frontend/pull/1719",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
416223254
|
Add Filterer responseType option
What does this PR do?
This pull request adds a new FilterResponse class and adds a responseType option to the Filterer::filter function to allow an instance of this new class to be returned.
The default return is still an array, which means it is not backwards-breaking. Additionally, the FilterResponse object can be converted to the traditional array result with its toArray() function.
Example use:
$response = Filterer::filter($spec, $value, ['responseType' => Filterer::RESPONSE_TYPE_FILTER]);
if ($response->success === false) {
throw new FilterException($response->errorMessage);
}
return $response->filteredValue;
Checklist
[X] Pull request contains a clear definition of changes
[X] Tests (either unit, integration, or acceptance) written and passing
[X] Relevant documentation produced and/or updated
Coverage decreased (-0.4%) to 97.26% when pulling 859e1caecc2cbd91906dc80584b516dfe77b30ae on jncarver:filter-response into cc04632ec167679e7b0309b50242dab53ab9879a on traderinteractive:master.
@jncarver What advantage is gained by adding the response object?
@chadicus I find it easier to use an object with explicit properties than an array of results that are in a specific order. It's also more compact, which is helpful with line length restrictions.
The biggest difference, however, is that the original array of errors are placed into the response object, so you can iterate through them if you need to. In the traditional array response, they are combined into a newline-delimited string which can be a pain if you're trying to do anything with the errors, like provide an error count.
|
gharchive/pull-request
| 2019-03-01T18:20:44 |
2025-04-01T06:40:40.546866
|
{
"authors": [
"chadicus",
"coveralls",
"jncarver"
],
"repo": "traderinteractive/filter-php",
"url": "https://github.com/traderinteractive/filter-php/pull/84",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
592568661
|
支持在tweb平台上运行
根据: https://github.com/tradingAI/proto/tree/master/model 中的输入输出, 将结果转换成相应的proto文件存储到s3中
具体实现 trunner(>=1.0.1)中的upload方法
状态信息
https://github.com/tradingAI/tbase/pull/15
|
gharchive/issue
| 2020-04-02T11:45:36 |
2025-04-01T06:40:40.549098
|
{
"authors": [
"iminders"
],
"repo": "tradingAI/tbase",
"url": "https://github.com/tradingAI/tbase/issues/11",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1072366978
|
feat: badge cleanup
Description
Cleaning up badge component, NB: this applies no design system changes.
This will be done, hopefully in a future step.
added missing compoundVariants programatically
added all color variants in story programatically
default color variant (gray) is duplicated to make sure it's clear to the user
cleaned up theme file
Package changes
updated rollup-plugin-typescript2 from ^0.29.0 to ^0.30.0 due to a build issue
CI: https://github.com/traefik/faency/runs/4433331267?check_suite_focus=true
https://github.com/alexjoverm/typescript-library-starter/issues/310#issuecomment-811451090
:tada: This PR is included in version 1.0.0 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
|
gharchive/pull-request
| 2021-12-06T16:47:35 |
2025-04-01T06:40:40.556720
|
{
"authors": [
"seedy",
"traefiker"
],
"repo": "traefik/faency",
"url": "https://github.com/traefik/faency/pull/221",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2687671131
|
Having issues with Traefik v3.2.1's RFC2136 DNS challenge for automatic Let's Encrypt certificate generation.
Traefik RFC2136 DNS Challenge TSIG Authentication Failure
Description
Having issues with Traefik v3.2.1's RFC2136 DNS challenge for automatic Let's Encrypt certificate generation. DNS updates are failing with TSIG verification errors when attempting to add/remove TXT records for domain validation.
Environment
Traefik: v3.2.1
Container Runtime: Podman 5.3.0
DNS Server: BIND 9.18.28 (Extended Support Version)
Domains:
traefik.home.doemijdienamespacemaar.nl
gitlab-tst.home.doemijdienamespacemaar.nl
Environment Variables:
RFC2136_NAMESERVER=ns3.home.doemijdienamespacemaar.nl53
RFC2136_TSIG_ALGORITHM=hmac-sha256.
RFC2136_TSIG_KEY_NAME=tsig-key
RFC2136_TSIG_SECRET=[REDACTED]
LEGO_EXPERIMENTAL_CNAME_SUPPORT=true
LEGO_LOG_LEVEL=DEBUG
BIND TSIG Key:
key name: tsig-key
algorithm: hmac-sha256
secret: [REDACTED]
System Architecture
graph LR
T[Traefik v3.2.1]
B[ns3.home.doemijdienamespacemaar.nl]
LE[Let's Encrypt]
T -->|RFC2136 Provider| B
B -->|TSIG hmac-sha256| B
T -->|ACME| LE
LE -->|Verify *.home.doemijdienamespacemaar.nl| B
subgraph Flow
1[Request Cert] --> 2[DNS-01 Challenge]
2 -->|RFC2136| 3[Create TXT on ns3.home.doemijdienamespacemaar.nl]
3 --> 4[Verify ZONE home.doemijdienamespacemaar.nl]
4 --> 5[Issue Cert]
5 -->|RFC2136| 6[Remove TXT]
end
Problem
Manual DNS updates using nsupdate with the same TSIG key work successfully
Traefik/Lego DNS challenge fails with TSIG verification (BADKEY) errors
Unable to automatically obtain/renew Let's Encrypt certificates
Diagnostic Steps Performed
Verified BIND version and configuration
Tested manual DNS updates using nsupdate with identical TSIG key
Enabled debug logging in both Traefik and BIND
Verified network connectivity and DNS resolution
Confirmed TSIG key format and encoding
Findings
Manual DNS updates succeed, confirming basic TSIG key functionality
Traefik's RFC2136 provider fails to authenticate with the same TSIG key
Potential issue with TSIG algorithm format (trailing dot in configuration)
Network connectivity and DNS resolution working correctly
Potential Issues Identified
TSIG algorithm format includes trailing dot (hmac-sha256.) which may cause issues
Possible mismatch in key name or case sensitivity
Potential issue in how Traefik/Lego formats the TSIG authentication
Questions
Is the trailing dot in RFC2136_TSIG_ALGORITHM=hmac-sha256. causing authentication failures?
Are there any known issues with Traefik v3.2.1's RFC2136 provider implementation?
Could there be a case sensitivity issue with the TSIG key name?
Additional Information
Full debug logs available upon request
Manual nsupdate commands work successfully
Network connectivity verified between Traefik container and BIND server
DNS resolution working correctly for all domains
Possible Workarounds Attempted
Adjusted TSIG key format and encoding
Verified DNS server configuration
Tested with different delay settings
Confirmed network connectivity and DNS resolution
Next Steps
Would appreciate guidance on:
Correct format for TSIG algorithm string
Any known issues with RFC2136 provider in Traefik v3.2.1
Additional debugging steps or configuration changes to try
Technical Configuration Details
Traefik ACME Configuration (traefik.yaml)
certificatesResolvers:
myresolver:
acme:
email: "[REDACTED]"
storage: "/var/traefik/certs/acme.json"
dnsChallenge:
provider: rfc2136
delayBeforeCheck: 60
logs
1. Traefik attempts DNS-01 challenge
[INFO] [traefik.home.doemijdienamespacemaar.nl] acme: Obtaining bundled SAN certificate
[INFO] [traefik.home.doemijdienamespacemaar.nl] AuthURL: https://acme-v02.api.letsencrypt.org/acme/authz/2073916237/434830476007
[INFO] [traefik.home.doemijdienamespacemaar.nl] acme: Could not find solver for: tls-alpn-01
[INFO] [traefik.home.doemijdienamespacemaar.nl] acme: Could not find solver for: http-01
[INFO] [traefik.home.doemijdienamespacemaar.nl] acme: use dns-01 solver
[INFO] [traefik.home.doemijdienamespacemaar.nl] acme: Preparing to solve DNS-01
2. DNS update fails with REFUSED
[WARN] [traefik.home.doemijdienamespacemaar.nl] acme: cleaning up failed: rfc2136: failed to insert: DNS update failed: server replied: REFUSED
3. BIND server logs showing TSIG verification failures
24-Nov-2024 11:03:22.043 error: client @0x7f8525d1c0 192.168.33.110#59813: request has invalid signature: TSIG m3razwi/f61pvbyokiglfqnuvkzfezji25bl4qzomnm=: tsig verify failure (BADKEY)
4. Final error from Traefik
[ERR] Unable to obtain ACME certificate for domains error="unable to generate a certificate for the domains [traefik.home.doemijdienamespacemaar.nl]: error: one or more domains had a problem:
[traefik.home.doemijdienamespacemaar.nl] [traefik.home.doemijdienamespacemaar.nl] acme: error presenting token: rfc2136: failed to insert: DNS update failed: server replied: REFUSED"
5. Successful manual updates from same client IP nsupdate
24-Nov-2024 12:49:56.662 info: client @0x7f842062c0 192.168.33.110#34499/key tsig-key: updating zone 'home.doemijdienamespacemaar.nl/IN': adding an RR at '_acme-challenge.test.home.doemijdienamespacemaar.nl' TXT "test-manual-update"
6 Add TXT record
$ cat /tmp/test-add.txt
server ns3.home.doemijdienamespacemaar.nl
zone home.doemijdienamespacemaar.nl
update add _acme-challenge.test.home.doemijdienamespacemaar.nl. 60 IN TXT "nsupdate-verify-test"
send
$ nsupdate -k /tmp/tsig.key -v /tmp/test-add.txt
Outgoing update query:
;; ->>HEADER<<- opcode: UPDATE, status: NOERROR, id: 0
;; flags:; ZONE: 0, PREREQ: 0, UPDATE: 0, ADDITIONAL: 0
;; ZONE SECTION:
;home.doemijdienamespacemaar.nl. IN SOA
;; UPDATE SECTION:
_acme-challenge.test.home.doemijdienamespacemaar.nl. 60 IN TXT "nsupdate-verify-test"
7. Verify record was added
$ dig @ns3.home.doemijdienamespacemaar.nl _acme-challenge.test.home.doemijdienamespacemaar.nl TXT
; <<>> DiG 9.18.28-0ubuntu0.24.04.1-Ubuntu <<>> @ns3.home.doemijdienamespacemaar.nl _acme-challenge.test.home.doemijdienamespacemaar.nl TXT
;; global options: +cmd
;; Got answer:
;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 33094
;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 1
;; ANSWER SECTION:
_acme-challenge.test.home.doemijdienamespacemaar.nl. 60 IN TXT "nsupdate-verify-test"
8. Delete TXT record
$ cat /tmp/test-del.txt
server ns3.home.doemijdienamespacemaar.nl
zone home.doemijdienamespacemaar.nl
update delete _acme-challenge.test.home.doemijdienamespacemaar.nl. TXT
send
$ nsupdate -k /tmp/tsig.key -v /tmp/test-del.txt
Outgoing update query:
;; ->>HEADER<<- opcode: UPDATE, status: NOERROR, id: 0
;; flags:; ZONE: 0, PREREQ: 0, UPDATE: 0, ADDITIONAL: 0
;; ZONE SECTION:
;home.doemijdienamespacemaar.nl. IN SOA
;; UPDATE SECTION:
_acme-challenge.test.home.doemijdienamespacemaar.nl. 0 ANY TXT
4. Verify record was deleted
$ dig @ns3.home.doemijdienamespacemaar.nl _acme-challenge.test.home.doemijdienamespacemaar.nl TXT
; <<>> DiG 9.18.28-0ubuntu0.24.04.1-Ubuntu <<>> @ns3.home.doemijdienamespacemaar.nl _acme-challenge.test.home.doemijdienamespacemaar.nl TXT
;; global options: +cmd
;; Got answer:
;; ->>HEADER<<- opcode: QUERY, status: NXDOMAIN, id: 62118
;; flags: qr aa rd ra; QUERY: 1, ANSWER: 0, AUTHORITY: 1, ADDITIONAL: 1
;; AUTHORITY SECTION:
home.doemijdienamespacemaar.nl. 7200 IN SOA ns3.home.doemijdienamespacemaar.nl. info.doemijdiemailmaar.nl. 50 43200 900 1814400 7200
Hello,
There is no specific implementation of this DNS challenge inside Traefik, I recommend opening an issue on the lego repository, I will answer you in this new issue.
|
gharchive/issue
| 2024-11-24T12:40:24 |
2025-04-01T06:40:40.591044
|
{
"authors": [
"RobertWi",
"ldez"
],
"repo": "traefik/traefik",
"url": "https://github.com/traefik/traefik/issues/11303",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1672147809
|
wrong api response for healthcheck interval and timeout values
Welcome!
[X] Yes, I've searched similar issues on GitHub and didn't find any.
[X] Yes, I've searched similar issues on the Traefik community forum and didn't find any.
What did you do?
I configured healthcheck interval or timeout to "wrong" values, e.g. -10s for interval
What did you see instead?
In the api response (and webui) the values for interval and timeout are the same as in the config (in this case -10s), but in reality, they are some sort of fallback values (in this case 30s for interval)
What version of Traefik are you using?
Version: 2.9.10
Codename: banon
Go version: go1.20.3
Built: 2023-04-06T16:15:08Z
OS/Arch: linux/amd64
What is your environment & configuration?
services:
traefik:
image: traefik
command:
- --api.insecure=true
- --providers.docker=true
- --log.level=DEBUG
ports:
- "8080:8080"
volumes:
- /run/user/1000/docker.sock:/var/run/docker.sock
whoami:
image: traefik/whoami
labels:
- "traefik.http.services.myservice.loadbalancer.healthcheck.path=/health"
- "traefik.http.services.myservice.loadbalancer.healthcheck.interval=-10s"
If applicable, please paste the log output in DEBUG level
time="2023-04-18T00:10:28Z" level=info msg="Configuration loaded from flags."
time="2023-04-18T00:10:28Z" level=info msg="Traefik version 2.9.10 built on 2023-04-06T16:15:08Z"
time="2023-04-18T00:10:28Z" level=debug msg="Static configuration loaded {\"global\":{\"checkNewVersion\":true},\"serversTransport\":{\"maxIdleConnsPerHost\":200},\"entryPoints\":{\"http\":{\"address\":\":80\",\"transport\":{\"lifeCycle\":{\"graceTimeOut\":\"10s\"},\"respondingTimeouts\":{\"idleTimeout\":\"3m0s\"}},\"forwardedHeaders\":{},\"http\":{},\"http2\":{\"maxConcurrentStreams\":250},\"udp\":{\"timeout\":\"3s\"}},\"traefik\":{\"address\":\":8080\",\"transport\":{\"lifeCycle\":{\"graceTimeOut\":\"10s\"},\"respondingTimeouts\":{\"idleTimeout\":\"3m0s\"}},\"forwardedHeaders\":{},\"http\":{},\"http2\":{\"maxConcurrentStreams\":250},\"udp\":{\"timeout\":\"3s\"}}},\"providers\":{\"providersThrottleDuration\":\"2s\",\"docker\":{\"watch\":true,\"endpoint\":\"unix:///var/run/docker.sock\",\"defaultRule\":\"Host(`{{ normalize .Name }}`)\",\"exposedByDefault\":true,\"swarmModeRefreshSeconds\":\"15s\"}},\"api\":{\"insecure\":true,\"dashboard\":true},\"log\":{\"level\":\"DEBUG\",\"format\":\"common\"}}"
time="2023-04-18T00:10:28Z" level=info msg="\nStats collection is disabled.\nHelp us improve Traefik by turning this feature on :)\nMore details on: https://doc.traefik.io/traefik/contributing/data-collection/\n"
time="2023-04-18T00:10:28Z" level=info msg="Starting provider aggregator aggregator.ProviderAggregator"
time="2023-04-18T00:10:28Z" level=debug msg="Starting TCP Server" entryPointName=traefik
time="2023-04-18T00:10:28Z" level=debug msg="Starting TCP Server" entryPointName=http
time="2023-04-18T00:10:28Z" level=info msg="Starting provider *traefik.Provider"
time="2023-04-18T00:10:28Z" level=debug msg="*traefik.Provider provider configuration: {}"
time="2023-04-18T00:10:28Z" level=info msg="Starting provider *docker.Provider"
time="2023-04-18T00:10:28Z" level=debug msg="*docker.Provider provider configuration: {\"watch\":true,\"endpoint\":\"unix:///var/run/docker.sock\",\"defaultRule\":\"Host(`{{ normalize .Name }}`)\",\"exposedByDefault\":true,\"swarmModeRefreshSeconds\":\"15s\"}"
time="2023-04-18T00:10:28Z" level=info msg="Starting provider *acme.ChallengeTLSALPN"
time="2023-04-18T00:10:28Z" level=debug msg="*acme.ChallengeTLSALPN provider configuration: {}"
time="2023-04-18T00:10:28Z" level=debug msg="Configuration received: {\"http\":{\"routers\":{\"api\":{\"entryPoints\":[\"traefik\"],\"service\":\"api@internal\",\"rule\":\"PathPrefix(`/api`)\",\"priority\":2147483646},\"dashboard\":{\"entryPoints\":[\"traefik\"],\"middlewares\":[\"dashboard_redirect@internal\",\"dashboard_stripprefix@internal\"],\"service\":\"dashboard@internal\",\"rule\":\"PathPrefix(`/`)\",\"priority\":2147483645}},\"services\":{\"api\":{},\"dashboard\":{},\"noop\":{}},\"middlewares\":{\"dashboard_redirect\":{\"redirectRegex\":{\"regex\":\"^(http:\\\\/\\\\/(\\\\[[\\\\w:.]+\\\\]|[\\\\w\\\\._-]+)(:\\\\d+)?)\\\\/$\",\"replacement\":\"${1}/dashboard/\",\"permanent\":true}},\"dashboard_stripprefix\":{\"stripPrefix\":{\"prefixes\":[\"/dashboard/\",\"/dashboard\"]}}},\"serversTransports\":{\"default\":{\"maxIdleConnsPerHost\":200}}},\"tcp\":{},\"udp\":{},\"tls\":{}}" providerName=internal
time="2023-04-18T00:10:28Z" level=debug msg="Provider connection established with docker 23.0.3 (API 1.42)" providerName=docker
time="2023-04-18T00:10:28Z" level=debug msg="Configuration received: {\"http\":{\"routers\":{\"traefik-traefik\":{\"service\":\"traefik-traefik\",\"rule\":\"Host(`traefik-traefik`)\"},\"whoami-traefik\":{\"service\":\"myservice\",\"rule\":\"Host(`whoami-traefik`)\"}},\"services\":{\"myservice\":{\"loadBalancer\":{\"servers\":[{\"url\":\"http://172.20.0.2:80\"}],\"healthCheck\":{\"path\":\"/health\",\"interval\":\"-10s\",\"followRedirects\":true},\"passHostHeader\":true}},\"traefik-traefik\":{\"loadBalancer\":{\"servers\":[{\"url\":\"http://172.20.0.3:80\"}],\"passHostHeader\":true}}}},\"tcp\":{},\"udp\":{}}" providerName=docker
time="2023-04-18T00:10:28Z" level=debug msg="No default certificate, fallback to the internal generated certificate" tlsStoreName=default
time="2023-04-18T00:10:28Z" level=debug msg="Added outgoing tracing middleware dashboard@internal" middlewareType=TracingForwarder entryPointName=traefik routerName=dashboard@internal middlewareName=tracing
time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" entryPointName=traefik middlewareName=dashboard_stripprefix@internal middlewareType=StripPrefix routerName=dashboard@internal
time="2023-04-18T00:10:28Z" level=debug msg="Adding tracing to middleware" entryPointName=traefik middlewareName=dashboard_stripprefix@internal routerName=dashboard@internal
time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" entryPointName=traefik routerName=dashboard@internal middlewareName=dashboard_redirect@internal middlewareType=RedirectRegex
time="2023-04-18T00:10:28Z" level=debug msg="Setting up redirection from ^(http:\\/\\/(\\[[\\w:.]+\\]|[\\w\\._-]+)(:\\d+)?)\\/$ to ${1}/dashboard/" entryPointName=traefik routerName=dashboard@internal middlewareName=dashboard_redirect@internal middlewareType=RedirectRegex
time="2023-04-18T00:10:28Z" level=debug msg="Adding tracing to middleware" entryPointName=traefik middlewareName=dashboard_redirect@internal routerName=dashboard@internal
time="2023-04-18T00:10:28Z" level=debug msg="Added outgoing tracing middleware api@internal" entryPointName=traefik routerName=api@internal middlewareName=tracing middlewareType=TracingForwarder
time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" middlewareName=traefik-internal-recovery entryPointName=traefik middlewareType=Recovery
time="2023-04-18T00:10:28Z" level=debug msg="No entryPoint defined for this router, using the default one(s) instead: [http]" routerName=whoami-traefik
time="2023-04-18T00:10:28Z" level=debug msg="No entryPoint defined for this router, using the default one(s) instead: [http]" routerName=traefik-traefik
time="2023-04-18T00:10:28Z" level=debug msg="No default certificate, fallback to the internal generated certificate" tlsStoreName=default
time="2023-04-18T00:10:28Z" level=debug msg="Added outgoing tracing middleware api@internal" entryPointName=traefik routerName=api@internal middlewareName=tracing middlewareType=TracingForwarder
time="2023-04-18T00:10:28Z" level=debug msg="Added outgoing tracing middleware dashboard@internal" middlewareName=tracing middlewareType=TracingForwarder entryPointName=traefik routerName=dashboard@internal
time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" middlewareName=dashboard_stripprefix@internal middlewareType=StripPrefix entryPointName=traefik routerName=dashboard@internal
time="2023-04-18T00:10:28Z" level=debug msg="Adding tracing to middleware" entryPointName=traefik routerName=dashboard@internal middlewareName=dashboard_stripprefix@internal
time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" entryPointName=traefik routerName=dashboard@internal middlewareType=RedirectRegex middlewareName=dashboard_redirect@internal
time="2023-04-18T00:10:28Z" level=debug msg="Setting up redirection from ^(http:\\/\\/(\\[[\\w:.]+\\]|[\\w\\._-]+)(:\\d+)?)\\/$ to ${1}/dashboard/" entryPointName=traefik routerName=dashboard@internal middlewareType=RedirectRegex middlewareName=dashboard_redirect@internal
time="2023-04-18T00:10:28Z" level=debug msg="Adding tracing to middleware" entryPointName=traefik routerName=dashboard@internal middlewareName=dashboard_redirect@internal
time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" middlewareType=Recovery entryPointName=traefik middlewareName=traefik-internal-recovery
time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" entryPointName=http routerName=whoami-traefik@docker serviceName=myservice middlewareName=pipelining middlewareType=Pipelining
time="2023-04-18T00:10:28Z" level=debug msg="Creating load-balancer" entryPointName=http routerName=whoami-traefik@docker serviceName=myservice
time="2023-04-18T00:10:28Z" level=debug msg="Creating server 0 http://172.20.0.2:80" serverName=0 entryPointName=http routerName=whoami-traefik@docker serviceName=myservice
time="2023-04-18T00:10:28Z" level=debug msg="child http://172.20.0.2:80 now UP"
time="2023-04-18T00:10:28Z" level=debug msg="Propagating new UP status"
time="2023-04-18T00:10:28Z" level=debug msg="Added outgoing tracing middleware myservice" entryPointName=http middlewareName=tracing middlewareType=TracingForwarder routerName=whoami-traefik@docker
time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" entryPointName=http routerName=traefik-traefik@docker middlewareType=Pipelining middlewareName=pipelining serviceName=traefik-traefik
time="2023-04-18T00:10:28Z" level=debug msg="Creating load-balancer" routerName=traefik-traefik@docker serviceName=traefik-traefik entryPointName=http
time="2023-04-18T00:10:28Z" level=debug msg="Creating server 0 http://172.20.0.3:80" routerName=traefik-traefik@docker serverName=0 serviceName=traefik-traefik entryPointName=http
time="2023-04-18T00:10:28Z" level=debug msg="child http://172.20.0.3:80 now UP"
time="2023-04-18T00:10:28Z" level=debug msg="Propagating new UP status"
time="2023-04-18T00:10:28Z" level=debug msg="Added outgoing tracing middleware traefik-traefik" middlewareType=TracingForwarder entryPointName=http routerName=traefik-traefik@docker middlewareName=tracing
time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" middlewareName=traefik-internal-recovery middlewareType=Recovery entryPointName=http
time="2023-04-18T00:10:28Z" level=error msg="Health check interval smaller than zero for service 'myservice@docker'" serviceName=myservice@docker
time="2023-04-18T00:10:28Z" level=debug msg="Setting up healthcheck for service myservice@docker with [Hostname: Headers: map[] Path: /health Method: Port: 0 Interval: 30s Timeout: 5s FollowRedirects: true]" serviceName=myservice@docker
time="2023-04-18T00:10:28Z" level=debug msg="Initial health check for backend: \"myservice@docker\""
last three lines are relevant here
Hello @1-Samuel ,
We've looked into it, and we agree that the lack of consistency between the effective configuration of traefik and the "raw data" returned by the API is annoying and disturbing (although not very concerning in itself in our opinion).
So here is our plan:
for v3, we actually would like to take another route: in such a case of wonky input from the user (e.g. -10s) we're of the opinion that we should log a clear error about it, and simply fail. And not try to cover up the mistake with a "saner" default. We've actually already taken such steps for similar situations in v3, and we think we should keep that trend going.
for v2, we cannot do the same as above, because it would be a breaking change. We may want to go in the direction you suggested (i.e. fix the raw data so that it fits with the effective configuration), but we're not convinced yet that the exact fix you proposed is the way we want to go. And tbh, we're not even sure that changing anything at all is worth the effort for v2.
So we would like to think about it some more, and also gauge the interest of the community on that matter.
WDYT?
Thanks.
Hello,
thank you for looking into the issue. Sounds like a good plan to me. :+1:
I only stumbled upon the issue while looking into #6550 and I couldn't think of a good way to write a test for retry intervals, so I thought I give the bug fix a try, just to practice a bit. :) I don't think many people with a real use case will run into this issue anyway.
Thank you :)
Hello @mpl ,
yeah I will have a look at implementing the v3 changes (fail if implausible values are configured). But I think making a new PR would make more sense, since the changes would be different all together. I'll close it :)
|
gharchive/issue
| 2023-04-18T00:15:33 |
2025-04-01T06:40:40.604031
|
{
"authors": [
"1-Samuel",
"mpl"
],
"repo": "traefik/traefik",
"url": "https://github.com/traefik/traefik/issues/9848",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
405905377
|
What is the status of Binary Ninja support? DynInst support?
These are some questions and a documentation bug report (insofar as the README is a bit confusing or doesn't answer these questions).
The top of the README states:
Control flow recovery is performed using the mcsema-disass tool, which relies on IDA Pro, Binary Ninja, or DynInst to disassemble a binary file and produce a control flow graph.
and later...
Currently IDA and BinaryNinja are supported for control-flow recovery...
but not dyninst...? (I see there are several abandoned or pending PRs).
Then later...
Why do I need IDA Pro to use McSema ... we have a branch with experimental Binary Ninja support ...
It looks like the referenced branch is deleted (presumably merged?). I also see a ninja branch referenced in a PR, and another one hanging out. What is the status of Binary Ninja support?
I'm trying out McSema for fun. I might shell out $150 for binary ninja if I was sure it would work (but maybe McSema requires the $600 version for its API?).
Thanks!
Also getting slightly off topic, I tried out the freeware version of IDA on the off chance it worked and no one has noticed. It doesn't and I don't get any helpful errors (is idal64 the headless version that McSema needs? if so I guess they don't ship that with the freeware version), but I wonder if the free version could be made to work in some other way? Or maybe y'all could try to convince them to release the functionality McSema needs in the free version (still for non-commercial use)? Seems mutually beneficial, but what do I know
@jberryman I have not tested IDA Pro Free Version, but if it includes IDAPython it should work. Use of idal64 is not required; you can use idaq64, but you have to do it from an X session since it has to start a GUI. I do this when debugging the CFG recovery scripts.
Thanks. Hm, doesn't seem to. The IDA free tree looks like:
/home/me/.local/idafree-7.0/
├── appico64.png
├── assistant
├── cfg
│ ├── ida.cfg
│ └── idagui.cfg
├── dbgsrv
├── ida64
├── ida64.int
├── ida.hlp
├── idc
│ └── idc.idc
├── ids
│ ├── idsnames
│ └── win7.zip
├── libclpx.so
├── libdwarfx.so
├── libida64.so
├── libQt5CLucene.so.5
├── libQt5Core.so.5
├── libQt5DBus.so.5
├── libQt5Gui.so.5
├── libQt5Help.so.5
├── libQt5Network.so.5
├── libQt5PrintSupport.so.5
├── libQt5Sql.so.5
├── libQt5Widgets.so.5
├── libQt5XcbQpa.so.5
├── libX11-xcb.so.1
├── license.txt
├── loaders
│ ├── elf64.so
│ ├── macho64.so
│ └── pe64.so
├── plugins
│ ├── bdescr64.so
│ ├── dbg64.so
│ ├── dwarf64.so
│ ├── objc64.so
│ ├── platforms
│ │ ├── libqlinuxfb.so
│ │ ├── libqminimal.so
│ │ ├── libqoffscreen.so
│ │ └── libqxcb.so
│ ├── platformthemes
│ │ └── libqgtk2.so
│ ├── plugins.cfg
│ └── tds64.so
├── procs
│ └── pc64.so
├── qidahelpcollection.qhc
├── qidahelp.qch
├── qt.conf
├── qwingraph
├── sig
│ ├── list
│ └── pc
│ ├── bcb5rt.sig
│ ├── iclapp64.sig
│ ├── ms64wdk.sig
│ ├── msmfc64d.sig
│ ├── msmfc64.sig
│ ├── msmfc64u.sig
│ ├── mssdk64.sig
│ ├── pe64.sig
│ ├── vc32rtf.sig
│ ├── vc64_14.sig
│ ├── vc64atl.sig
│ ├── vc64extra.sig
│ ├── vc64mfc.sig
│ ├── vc64rtf.sig
│ ├── vc64seh.sig
│ ├── vc64ucrt.sig
│ ├── vcseh.sig
│ └── win64unx.sig
├── til
│ ├── gnuunx64.til
│ ├── macosx64.til
│ └── pc
│ ├── bcb5win.til
│ ├── gnulnx_x64.til
│ ├── mssdk.til
│ ├── ntddk64.til
│ ├── vc10_64.til
│ ├── vc6win.til
│ └── vc8amd64.til
├── uninstall
├── uninstall.dat
└── Uninstall IDA Free.desktop
Sadly it looks like the free version is missing IDAPython so the script will not work.
Closing for now.
|
gharchive/issue
| 2019-02-01T22:54:57 |
2025-04-01T06:40:40.799491
|
{
"authors": [
"artemdinaburg",
"jberryman",
"pgoodman"
],
"repo": "trailofbits/mcsema",
"url": "https://github.com/trailofbits/mcsema/issues/515",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1120698177
|
requirement, pypi: Add a --require-hashes flag
Closes #172
This needs cleanup + testing but I wanted to engage early on this one since the intended behaviour isn't straightforward.
@di You mentioned that the --require-hashes flag should skip dependency resolution. Did you mean that we shouldn't resolve dependencies for each requirement and instead, only audit what is explicitly listed in the requirements file?
I was also wondering how hashes should work when the --require-hashes flag isn't supplied. My thinking is that when the --require-hashes flag is supplied, every requirement in the file needs to have a hash. But if it isn't supplied, we should still check any hashes that we find against PyPI.
You mentioned that the --require-hashes flag should skip dependency resolution. Did you mean that we shouldn't resolve transitive dependencies for each requirement and instead, only audit what is explicitly listed in the requirements file?
I was also wondering how hashes should work when the --require-hashes flag isn't supplied. My thinking is that when the --require-hashes flag is supplied, every requirement in the file needs to have a hash AND we check them against PyPI. But if the flag isn't supplied, we should still check any hashes that we find against PyPI and fail if it doesn't match.
This is an indirect answer to both: we can skip dependency resolution entirely, since pip stipulates the following:
If any dependency has a content hash, then all dependencies have content hashes
(Transitively) All dependencies must be explicitly listed in the requirements file if hashing is enabled
In other words: our requirements parser in pip-api should preserve those properties: we should raise an error if we're given a requirements file with dependencies that are missing hashes, and we should not do dependency resolution because only the present hashed dependencies should be considered.
(Looking at things more, I'm not actually sure that we need the --require-hashes flag -- pip stipulates that passing it is identical to any requirement having hashes attached to it. But I guess we should preserve it as a no-op, since that's what pip seems to do?
In other words: our requirements parser in pip-api should preserve those properties: we should raise an error if we're given a requirements file with dependencies that are missing hashes, and we should not do dependency resolution because only the present hashed dependencies should be considered.
Ah! Right, I misunderstood how that works. Thanks for explaining.
(Looking at things more, I'm not actually sure that we need the --require-hashes flag -- pip stipulates that passing it is identical to any requirement having hashes attached to it. But I guess we should preserve it as a no-op, since that's what pip seems to do?
I think it's still important because your file may not have ANY hashes in it. If you pass in --require-hashes, it will complain whereas it otherwise would succeed.
I think it's still important because your file may not have ANY hashes in it. If you pass in --require-hashes, it will complain whereas it otherwise would succeed.
Ah! You're absolutely right. Yes, we need it for that case.
@di You mentioned that the --require-hashes flag should skip dependency resolution. Did you mean that we shouldn't resolve transitive dependencies for each requirement and instead, only audit what is explicitly listed in the requirements file?
Thinking about this more, I think we still need to do dependency resolution, and ensure that we have hashes for all subdependencies, because this is what pip does as well:
$ cat req.txt
sampleproject==2.0.0 \
--hash=sha256:2b0c55537193b792098977fdb62f0acbaeb2c3cfc56d0e24ccab775201462e04 \
--hash=sha256:d99de34ffae5515db43916ec47380d3c603e9dead526f96581b48c070cc816d3
$ python -m pip install -r req.txt
Collecting sampleproject==2.0.0 (from -r req.txt (line 1))
Using cached https://files.pythonhosted.org/packages/b8/f7/dd9223b39f683690c30f759c876df0944815e47b588cb517e4b9e652bcf7/sampleproject-2.0.0-py3-none-any.whl
Collecting peppercorn (from sampleproject==2.0.0->-r req.txt (line 1))
ERROR: In --require-hashes mode, all requirements must have their versions pinned with ==. These do not:
peppercorn from https://files.pythonhosted.org/packages/14/84/d8d9c3f17bda2b6f49406982546d6f6bc0fa188a43d4e3ba9169a457ee04/peppercorn-0.6-py3-none-any.whl#sha256=46125cad688a9cf3b08e463bcb797891ee73ece93602a8ea6f14e40d1042d454 (from sampleproject==2.0.0->-r req.txt (line 1))
I think it's still important because your file may not have ANY hashes in it. If you pass in --require-hashes, it will complain whereas it otherwise would succeed.
Agreed!
Thinking about this more, I think we still need to do dependency resolution, and ensure that we have hashes for all subdependencies, because this is what pip does as well:
Can we leave that validation to pip and not attempt to do that validation ourselves? I imagine that's going to introduce a significant chunk of code reinventing what pip does which will have to be removed during integration, since pip already contains this logic.
Can we leave that validation to pip and not attempt to do it ourselves? I imagine that's going to introduce a significant chunk of code reinventing what pip does which will have to be removed during integration, since pip already contains this logic.
I'm inclined to agree with this, although it does deviate our handling of --require-hashes from that of pip (our --require-hashes is more like --no-deps --require-hashes). My reasoning is that pip-audit isn't in the business of telling the user whether pip install ... would succeed, and that the overwhelming majority of hashed requirements files should be valid already (since, if they're like Warehouse's, they're checked in after successful local and CI usage).
Thoughts @di?
I suppose the pip install with an incompletely hashed requirements file, or a hashed requirements file that's missing dependencies, would fail to install anyways, so the user would be protected regardless.
|
gharchive/pull-request
| 2022-02-01T13:48:06 |
2025-04-01T06:40:40.812301
|
{
"authors": [
"di",
"tetsuo-cpp",
"woodruffw"
],
"repo": "trailofbits/pip-audit",
"url": "https://github.com/trailofbits/pip-audit/pull/229",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1514673788
|
Downgrades react-intl to fix importStar error
This fixes an issue with websites that have an old version of requirejs. react intl v5 > relies on import *, which the older version of requirejs does not support
see https://github.com/transcend-io/consent-manager-ui/pull/101
|
gharchive/pull-request
| 2022-12-30T17:36:54 |
2025-04-01T06:40:40.829825
|
{
"authors": [
"michaelfarrell76"
],
"repo": "transcend-io/consent-manager-ui",
"url": "https://github.com/transcend-io/consent-manager-ui/pull/100",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
370895267
|
WIP: Adds UI for confirming pre-approval request deletion [delivers #161129636]
Description
Delete UI, some spacing issues to work out. WIP.
Breaks out the PreApprovalRequest row into its own component that owns the state of whether the delete UI is shown
Code Review Verification Steps
[ ] End to end tests pass (make e2e_test).
[ ] Code follows the guidelines for Logging
[ ] The requirements listed in
Querying the Database Safely
have been satisfied.
Any new migrations/schema changes:
[ ] Follow our guidelines for zero-downtime deploys (see Zero-Downtime Deploys)
[ ] Have been communicated to #dp3-engineering
[ ] There are no aXe warnings for UI.
[ ] This works in IE.
Any new client dependencies (Google Analytics, hosted libraries, CDNs, etc) have been:
[ ] Communicated to @willowbl00
[ ] Added to the list of network dependencies
[ ] Request review from a member of a different team.
[ ] Have the Pivotal acceptance criteria been met for this change?
References
Pivotal story for this change
Screenshots
This is a great start, but I'm not sure how you plan on signaling that one of the deletion confirms is open to prevent this:
|
gharchive/pull-request
| 2018-10-17T04:33:43 |
2025-04-01T06:40:40.837019
|
{
"authors": [
"stangah",
"tinyels"
],
"repo": "transcom/mymove",
"url": "https://github.com/transcom/mymove/pull/1164",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2534607937
|
B 20441 int
Agility ticket
Summary
Updating migrations to add WHERE NOT EXIST to protect against any issues from running the migration more than once
Verification Steps for the Author
These are to be checked by the author.
[ ] Tested in the Experimental environment (for changes to containers, app startup, or connection to data stores)
[ ] Have the Agility acceptance criteria been met for this change?
Verification Steps for Reviewers
These are to be checked by a reviewer.
[ ] Has the branch been pulled in and checked out?
[ ] Have the BL acceptance criteria been met for this change?
[ ] Was the CircleCI build successful?
[ ] Has the code been reviewed from a standards and best practices point of view?
Setup to Run the Code
Instructions for starting storybook
Instructions for starting the MilMove application
Instructions for running tests
How to test
Access the
Login as a
Frontend
[ ] There are no aXe warnings for UI.
[ ] This works in Supported Browsers and their phone views (Chrome, Firefox, Edge).
[ ] There are no new console errors in the browser devtools.
[ ] There are no new console errors in the test output.
[ ] If this PR adds a new component to Storybook, it ensures the component is fully responsive, OR if it is intentionally not, a wrapping div using the officeApp class or custom min-width styling is used to hide any states the would not be visible to the user.
[ ] This change meets the standards for Section 508 compliance.
Backend
[ ] Code follows the guidelines for Logging.
[ ] The requirements listed in Querying the Database Safely have been satisfied.
Database
Any new migrations/schema changes:
[ ] Follows our guidelines for Zero-Downtime Deploys.
[ ] Have been communicated to #g-database.
[ ] Secure migrations have been tested following the instructions in our docs.
Screenshots
Warnings
:warning:
Files located in legacy directories (src/shared or src/scenes) have
been edited. Are you sure you don’t want to also relocate them to the new file structure?
View the frontend file org ADR for more information
Generated by :no_entry_sign: dangerJS against 912398a13be2e07babfe0d644614ac7640536f9b
|
gharchive/pull-request
| 2024-09-18T20:01:40 |
2025-04-01T06:40:40.849186
|
{
"authors": [
"CoryKleinjanCACI",
"robot-mymove"
],
"repo": "transcom/mymove",
"url": "https://github.com/transcom/mymove/pull/13739",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
933241127
|
[MB-8204] Refactor TXO MTO fetch to not query for MoveTaskOrder
Description
Explain a little about the changes at a high level.
Reviewer Notes
Is there anything you would like reviewers to give additional scrutiny?
Setup
Add any steps or code to run in this section to help others prepare to run your code:
echo "Code goes here"
Code Review Verification Steps
[ ] If the change is risky, it has been tested in experimental before merging.
[ ] Code follows the guidelines for Logging
[ ] The requirements listed in Querying the Database Safely have been satisfied.
Any new migrations/schema changes:
[ ] Follow our guidelines for zero-downtime deploys (see Zero-Downtime Deploys)
[ ] Have been communicated to #g-database
[ ] Secure migrations have been tested following the instructions in our docs
[ ] There are no aXe warnings for UI.
[ ] This works in Supported Browsers and their phone views (Chrome, Firefox, IE, Edge).
[ ] Tested in the Experimental environment (for changes to containers, app startup, or connection to data stores)
[ ] User facing changes have been reviewed by design.
[ ] Request review from a member of a different team.
[ ] Have the Jira acceptance criteria been met for this change?
References
Jira story for this change
this article explains more about the approach used.
Screenshots
If this PR makes visible UI changes, an image of the finished UI can help reviewers and casual
observers understand the context of the changes. A before image is optional and
can be included at the submitter's discretion.
Consider using an animated image to show an entire workflow instead of using multiple images. You may want to use GIPHY CAPTURE for this! 📸
Please frame screenshots to show enough useful context but also highlight the affected regions.
Warnings
:warning:
Files located in legacy directories (src/shared or src/scenes) have
been edited. Are you sure you don’t want to also relocate them to the new file structure?
View the frontend file org ADR for more information
Messages
:book:
:link: MB-8204
Generated by :no_entry_sign: dangerJS against 94d60bf5a36c4b8cb0ee8d8665c616b86dd29ecd
This looks good so far. I love removing code. Here are a few more things we need to remove:
https://github.com/transcom/mymove/blob/master/pkg/services/move_task_order.go#L37
https://github.com/transcom/mymove/blob/master/pkg/services/move_task_order.go#L58
https://github.com/transcom/mymove/blob/master/pkg/services/move_task_order/move_task_order_fetcher.go#L23-L40
And then we'll need to regenerate the mocks:
make mocks_generate
|
gharchive/pull-request
| 2021-06-30T02:09:58 |
2025-04-01T06:40:40.861864
|
{
"authors": [
"hannaliebl",
"monfresh",
"robot-mymove"
],
"repo": "transcom/mymove",
"url": "https://github.com/transcom/mymove/pull/6901",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
953298375
|
[MB-8827] Bump storybook-related dependencies to 6.3.6
Description
This PR consolidates the following dependabot PRs related to upgrading Storybook to version 6.3.6:
#7068
#7066
#7063
#7061
#7060
Setup
Just need to ensure the tests and build pass with no issues.
Messages
:book:
:link: MB-8827
Generated by :no_entry_sign: dangerJS against 822e9a2a4009fb08faf5fa03ce3c70fc897a2d6b
Can a front-end person confirm the happo diffs are okay? Especially the document viewer one (the diff shows it as loading instead of present)
cc: @YanZ777 @jenniferlynparsons
|
gharchive/pull-request
| 2021-07-26T21:20:15 |
2025-04-01T06:40:40.865841
|
{
"authors": [
"pearl-truss",
"robot-mymove"
],
"repo": "transcom/mymove",
"url": "https://github.com/transcom/mymove/pull/7071",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1100598229
|
[MB-11012] Fixes bug where NTS shipment turns into HHG when editing as a SC
Jira ticket for this change
Summary
We ran across a bug in the services counselor UI where editing an NTS shipment turns it into an HHG after submission. This PR addresses that issue (it looks like we had just hardcoded the HHG as an input to the component).
I also noticed that we were using some old naming for ntsSac and ntsTac in some places. That was causing those values (if present on the order record) to not show up on the edit form, for instance. Those are fixed as well in this PR.
Note to reviewers: Are there any particular tests I need to add/edit for this?
Setup to Run Your Code
💻 You will need to use three separate terminals to test this locally.
Terminal 1
Start the Storybook locally.
make storybook
Terminal 2
Start the UI locally.
make client_run
Terminal 3
Start the Go server locally.
make server_run
Additional steps
As a services counselor, try editing an NTS shipment (or NTS-R) and change something like the counselor remarks. Hit save and verify that the shipment remains the same type it was previously.
Also, you can add NTS TAC/SAC values to the associated order and verify that they show up in the edit interface.
Verification Steps for Author
These are to be checked by the author.
[ ] Tested in the Experimental environment (for changes to containers, app startup, or connection to data stores)
[ ] Request review from a member of a different team.
[ ] Have the Jira acceptance criteria been met for this change?
Verification Steps for Reviewers
These are to be checked by a reviewer.
Frontend
[ ] User facing changes have been reviewed by design.
[ ] There are no aXe warnings for UI.
[ ] This works in Supported Browsers and their phone views (Chrome, Firefox, IE, Edge).
[ ] There are no new console errors in the browser devtools
[ ] There are no new console errors in the test output
Warnings
:warning:
This PR does not include changes to unit tests, even though it affects app code.
Messages
:book:
:link: MB-11012
Generated by :no_entry_sign: dangerJS against aea9c431eafe9cc2bc9c8d4023bc31b6c8eca337
|
gharchive/pull-request
| 2022-01-12T17:20:42 |
2025-04-01T06:40:40.874671
|
{
"authors": [
"reggieriser",
"robot-mymove"
],
"repo": "transcom/mymove",
"url": "https://github.com/transcom/mymove/pull/8021",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1454075565
|
[MB-14464] Updated milmove-cypress hash to latest one, now with Cypress 8.5.0
Jira ticket for this change
Summary
First, reference this circleci-docker PR. This MilMove PR is just trying to get MilMove on the latest image with a downgraded Cypress (8.5.0). We're currently on the same version, but we're having to use a circleci-docker hash from back in April because cypress was updated after that to 9.x and 10.x in circleci-docker independently of MilMove. As a result, updating other dependencies becomes more difficult. So this just tries to get everything on the latest main branches back in sync.
Setup to Run Your Code
💻 You will need to use three separate terminals to test this locally.
Terminal 1
Start the Storybook locally.
make storybook
Terminal 2
Start the UI locally.
make client_run
Terminal 3
Start the Go server locally.
make server_run
Additional steps
Just make sure all E2E tests pass.
Verification Steps for Author
These are to be checked by the author.
[ ] Tested in the Experimental environment (for changes to containers, app startup, or connection to data stores)
[ ] Request review from a member of a different team.
[ ] Have the Jira acceptance criteria been met for this change?
Messages
:book:
:link: MB-14464
Generated by :no_entry_sign: dangerJS against ffdd48b6377fec23f2641a781a88ee0bcf573cd9
|
gharchive/pull-request
| 2022-11-17T22:15:13 |
2025-04-01T06:40:40.881581
|
{
"authors": [
"reggieriser",
"robot-mymove"
],
"repo": "transcom/mymove",
"url": "https://github.com/transcom/mymove/pull/9612",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2432740832
|
Modernization
The original project is great!
It relies on a lot of outdated dependencies, though. I've gone through the codebase and bumped dependencies, gotten rid of old code, and implemented a unified way to set the icons on the right bar and the bottom bar.
The key dependency react-notion-x is not altered here. I'm also working on that, but have hit some sort of a roadblock with some dependency.
Thanks @gauravmm for this! While we are waiting for @transitive-bullshit to decide how to continue with this starter I created my own fork with updated react-notion-x dependency.
See more here:
https://github.com/onnimonni/nextjs-notion-starter-kit/pull/2
You can see the forked react-notion-x here: https://github.com/onnimonni/react-notion-x/tree/publish-forked-versions
It works perfectly but few tests are still failing which are written with ts and try to import the internal packages eg `import "notion-utils"' are not working and fail with:
Uncaught exception in src/notion-api-universal.test.ts
Error [ERR_MODULE_NOT_FOUND]: Cannot find package '@onnimonni/notion-utils' imported from /Users/onnimonni/Projects/react-notion-x/packages/notion-client/src/notion-api.ts
I'm happy to add you as contributor to both of these forks if you want to help out 👍.
Fixed in https://github.com/transitive-bullshit/nextjs-notion-starter-kit/pull/641
|
gharchive/pull-request
| 2024-07-26T18:06:56 |
2025-04-01T06:40:40.895541
|
{
"authors": [
"gauravmm",
"onnimonni",
"transitive-bullshit"
],
"repo": "transitive-bullshit/nextjs-notion-starter-kit",
"url": "https://github.com/transitive-bullshit/nextjs-notion-starter-kit/pull/629",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2258024456
|
Services support
Hello, I would like to inquire if the Transient Robotics platform supports sending and receiving ROS services. Currently, I am developing a web interface that requires some services to control certain actions on the robot, and I would like to evaluate your solution for a series of additional features that you offer. I have read that managing services may be a limitation of the MQTT communication protocol. I'm not sure if this is the right place to ask this question; if it isn't, I can move it to where it's more appropriate. Thank you very much.
We don't currently have an existing capability to do this, but it has been requested by others, too, so we'll probably create one soon. Since you are asking this question here, on the github repo of the open-source Transitive framework, I'll also point out that you could write that capability yourself, if that's what you want. To get started with that, you'd follow these steps and then look at the starter code that is generated from it: https://transitiverobotics.com/docs/develop/creating_capabilities
To implement the service calls, I would use mqttSync.mqtt directly, i.e., not use mqttSync (which is for synchronizing stateful data), but the mqtt client itself. You can follow this example from the robot-agent itself, used to trigger agent actions, in this case restarting the agent, from the web:
https://github.com/transitiverobotics/transitive/blob/bf7aadaef4c6508b5195ee9f33079ac4e18370f6/cloud/app/web_components/robot-agent-device.jsx#L294-L298
MQTT does not have any built-in support for RPCs, so you'll need to think about the return value yourself. In the agent we do this by writing it into the synchronized data, where the web client will receive it, e.g., for receiving the package log upon request. This pattern is not ideal though and still needs some improvement.
But again, if this is more involved than you'd like, then you can also wait until we create a capability for it that you could then just install on your robots.
Thanks for your kind reply. I took a look at the solution proposed by you by searching for additional material online like async-mqtt or mqtt-json-rpc but, given my limited knowledge of the mqtt protocol, i would most likely not be able to obtain a good result. Currently I will therefore remain on the websocket protocol but keeping an eye on the developments and implementations of your platform. Thanks so much again
Makes sense. Just to clarify though: MQTTSync is not async-mqtt. The former is a protocol we've developed an it's documented here and here.
|
gharchive/issue
| 2024-04-23T06:07:29 |
2025-04-01T06:40:40.901646
|
{
"authors": [
"alexgiano",
"chfritz"
],
"repo": "transitiverobotics/transitive",
"url": "https://github.com/transitiverobotics/transitive/issues/38",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2342555849
|
Support LAMP based monthly csvs
There's a new dataset for monthly dumps based on LAMP with slightly different formatting
Use that new file and handle its formatting
Already using this for data already in the production dashboard, going to merge to avoid someone parsing data with the old version
|
gharchive/pull-request
| 2024-06-09T23:41:52 |
2025-04-01T06:40:40.902977
|
{
"authors": [
"devinmatte"
],
"repo": "transitmatters/mbta-performance",
"url": "https://github.com/transitmatters/mbta-performance/pull/27",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
606144836
|
Feature: Setting constant onset shift for one cachefile
Is your feature request related to a problem? Please describe.
Currently, it is possibly to shift the onset of MEPs in the GUI. However, for one cachefile it requires one to manually shift the onset for every single MEP.
Describe the solution you'd like
It would be great to have the possibility to set an onset shift and have it constant for the whole data of one Cachefile.
I believe using the python API is the best approach for such batch-manipulation. I wrote a couple of examples and added them to the documentation. The one for your use-case is at https://offline-inspect.readthedocs.io/en/latest/cachefile.html#batch-manipulate and repeated here for ease of use:
from offspect.api import CacheFile, decode, encode
cf = CacheFile("merged.hdf5")
for ix, (data, attrs) in enumerate(cf):
key = "onset_shift"
old = decode(attrs[key])
print(f"Trace {ix} {key}:", old, end=" ")
new = old + 5
attrs["onset_shift"] = encode(new)
cf.set_trace_attrs(ix, attrs)
test = decode(cf.get_trace_attrs(ix)["onset_shift"])
print("to", test)
In this example, we shift all onsets by 5 samples to the right.
Closing this, as we agreed a API approach is most suitable. Find updated documentations with examples for batch manipulation https://offline-inspect.readthedocs.io/en/latest/cachefile.html
|
gharchive/issue
| 2020-04-24T08:40:37 |
2025-04-01T06:40:40.931501
|
{
"authors": [
"BettinaTrunk",
"agricolab"
],
"repo": "translationalneurosurgery/tool-offspect",
"url": "https://github.com/translationalneurosurgery/tool-offspect/issues/26",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2365950350
|
🛑 TechinAsia_Laravel_API is down
In 833fcd5, TechinAsia_Laravel_API (https://www.techinasia.com/api/2.0/companies) was down:
HTTP code: 403
Response time: 172 ms
Resolved: TechinAsia_Laravel_API is back up in 93232f9 after 7 minutes.
|
gharchive/issue
| 2024-06-21T07:53:12 |
2025-04-01T06:40:40.950196
|
{
"authors": [
"traqy"
],
"repo": "traqy/upptime",
"url": "https://github.com/traqy/upptime/issues/11537",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2600780068
|
🛑 TechinAsia_WordPress_API is down
In 3514459, TechinAsia_WordPress_API (https://www.techinasia.com/wp-json/techinasia/2.0/posts) was down:
HTTP code: 403
Response time: 226 ms
Resolved: TechinAsia_WordPress_API is back up in 2b7888e after 7 minutes.
|
gharchive/issue
| 2024-10-20T18:46:59 |
2025-04-01T06:40:40.953011
|
{
"authors": [
"traqy"
],
"repo": "traqy/upptime",
"url": "https://github.com/traqy/upptime/issues/13001",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2059023544
|
🛑 TechinAsia is down
In eebba53, TechinAsia (https://www.techinasia.com) was down:
HTTP code: 403
Response time: 959 ms
Resolved: TechinAsia is back up in 078d2ea after 9 minutes.
|
gharchive/issue
| 2023-12-29T03:13:59 |
2025-04-01T06:40:40.955359
|
{
"authors": [
"traqy"
],
"repo": "traqy/upptime",
"url": "https://github.com/traqy/upptime/issues/1999",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2066054495
|
🛑 TechinAsia_Laravel_API is down
In 1aa71df, TechinAsia_Laravel_API (https://www.techinasia.com/api/2.0/companies) was down:
HTTP code: 403
Response time: 228 ms
Resolved: TechinAsia_Laravel_API is back up in bf63915 after 7 minutes.
|
gharchive/issue
| 2024-01-04T17:22:32 |
2025-04-01T06:40:40.958631
|
{
"authors": [
"traqy"
],
"repo": "traqy/upptime",
"url": "https://github.com/traqy/upptime/issues/2402",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2190266490
|
🛑 TechinAsia_Laravel_API is down
In c9f95b1, TechinAsia_Laravel_API (https://www.techinasia.com/api/2.0/companies) was down:
HTTP code: 403
Response time: 225 ms
Resolved: TechinAsia_Laravel_API is back up in 08f5cd2 after 7 minutes.
|
gharchive/issue
| 2024-03-16T21:40:36 |
2025-04-01T06:40:40.961258
|
{
"authors": [
"traqy"
],
"repo": "traqy/upptime",
"url": "https://github.com/traqy/upptime/issues/6624",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2196605939
|
🛑 TechinAsia_Laravel_API is down
In c5ea9d6, TechinAsia_Laravel_API (https://www.techinasia.com/api/2.0/companies) was down:
HTTP code: 403
Response time: 230 ms
Resolved: TechinAsia_Laravel_API is back up in 11b702b after 14 minutes.
|
gharchive/issue
| 2024-03-20T03:40:21 |
2025-04-01T06:40:40.963873
|
{
"authors": [
"traqy"
],
"repo": "traqy/upptime",
"url": "https://github.com/traqy/upptime/issues/6807",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2203867476
|
🛑 TechinAsia_Laravel_API is down
In 6cfb34a, TechinAsia_Laravel_API (https://www.techinasia.com/api/2.0/companies) was down:
HTTP code: 403
Response time: 227 ms
Resolved: TechinAsia_Laravel_API is back up in 07f7c1a after 7 minutes.
|
gharchive/issue
| 2024-03-23T12:53:45 |
2025-04-01T06:40:40.966416
|
{
"authors": [
"traqy"
],
"repo": "traqy/upptime",
"url": "https://github.com/traqy/upptime/issues/7012",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2033588404
|
🛑 TechinAsia_WordPress_API is down
In dc58ede, TechinAsia_WordPress_API (https://www.techinasia.com/wp-json/techinasia/2.0/posts) was down:
HTTP code: 403
Response time: 231 ms
Resolved: TechinAsia_WordPress_API is back up in bc82ed7 after 37 minutes.
|
gharchive/issue
| 2023-12-09T03:46:08 |
2025-04-01T06:40:40.969175
|
{
"authors": [
"traqy"
],
"repo": "traqy/upptime",
"url": "https://github.com/traqy/upptime/issues/774",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2267390167
|
🛑 TechinAsia is down
In ff66d5d, TechinAsia (https://www.techinasia.com) was down:
HTTP code: 403
Response time: 947 ms
Resolved: TechinAsia is back up in a00258e after 7 minutes.
|
gharchive/issue
| 2024-04-28T06:52:06 |
2025-04-01T06:40:40.971823
|
{
"authors": [
"traqy"
],
"repo": "traqy/upptime",
"url": "https://github.com/traqy/upptime/issues/8938",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
268483417
|
Code Coverage Plugin
@travisjeffery I would like to enable https://codecov.io/ on the repo. I don't think I can perform all the required steps to set it up but I can assist and add the required test setup. What do you think?
Good idea, I've set it up. Let's get the coverage up before we add the badge to the README though :D
@travisjeffery I messaged on gitter before looking at these issues - i will try and help with coverage too :)
|
gharchive/issue
| 2017-10-25T17:29:57 |
2025-04-01T06:40:41.011711
|
{
"authors": [
"jacktuck",
"ryanstinson",
"travisjeffery"
],
"repo": "travisjeffery/jocko",
"url": "https://github.com/travisjeffery/jocko/issues/69",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2571194352
|
x and y-units in mosaic and terrain example
In this example https://pdemtools.readthedocs.io/en/latest/examples/mosaic_and_terrain.html there are some very nice figures.
However, the x- and y- units are unknown. Would it be possible to show these as latitude/longitudes or something the reader can understand?
Thanks Adam. The merging process I use to combined the mosaics from different tiles loses the CF-compliant coordinate attributes that the PDT data normally comes with. New addition of code below enforces the CF-compliant attributes and makes the mosaic data match the rest of the data. This is also reflected in the notebooks.
dem['x'].attrs['axis'] = 'X'
dem['x'].attrs['long_name'] = 'x coordinate of projection'
dem['x'].attrs['standard_name'] = 'projection_x_coordinate'
dem['x'].attrs['units'] = 'metre'
dem['y'].attrs['axis'] = 'Y'
dem['y'].attrs['long_name'] = 'y coordinate of projection'
dem['y'].attrs['standard_name'] = 'projection_y_coordinate'
dem['y'].attrs['units'] = 'metre'
|
gharchive/issue
| 2024-10-07T18:46:09 |
2025-04-01T06:40:41.018532
|
{
"authors": [
"AdamRJensen",
"trchudley"
],
"repo": "trchudley/pdemtools",
"url": "https://github.com/trchudley/pdemtools/issues/31",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
195782945
|
Redundant parentheses cause memory corruption
Following code cause memory corruption:
main :- (foo, (bar(baz))), halt.
Yep,
main :- (bar(baz)), halt.
is enough.
Not sure if related, following code does not parse correctly:
main :- write(a), nl, (write(b), nl), halt.
Listing:
Trealla v0.1alpha, 64-bits (int=128-bits), Dec 16 2016
main :-
write(a),
nl(write(b),nl),
halt.
According to valgrind's result, those lines trealla.c#L2058-L2063 are suspicious.
After I commented them out, memory corruption disappeared.
Yes, very suspect. Can't imagine what it is supposed to be doing.
Perhaps it was supposed to be NLIST_POP_FRONT, not NLIST_FRONT.
Anyway, it still doesn't parse correctly.
Now your test case will cause an infinite loop:
main :- (bar(baz)), halt.
Now for this:
main :- write(a), nl, (write(b), nl), halt.
Seems ok now?
Cool! Fix confirmed in devel branch. I'll close this issue.
|
gharchive/issue
| 2016-12-15T11:33:15 |
2025-04-01T06:40:41.023217
|
{
"authors": [
"dram",
"infradig"
],
"repo": "trealla-lang/trealla",
"url": "https://github.com/trealla-lang/trealla/issues/54",
"license": "bsd-2-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1197117811
|
Unification segv
?- A=[A|A],B=[A|B],A=[A|B], false.
Segmentation fault (core dumped)
(This one blocks a lot)
I can only repeat that this blocks a lot, I just reencountered exactly this error.
|
gharchive/issue
| 2022-04-08T10:05:53 |
2025-04-01T06:40:41.024431
|
{
"authors": [
"UWN"
],
"repo": "trealla-prolog/trealla",
"url": "https://github.com/trealla-prolog/trealla/issues/520",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
753524358
|
Add "lakectl fs upload --recursive" flag
Behaves a lot like s3 cp --recursive: uploads files using their path relative from the local directory.
Example
Suppose you have these files:
/tmp/a/foo
/tmp/a/b/bar
/tmp/a/b/c/d/e/xyzzy
Then lakectl upload --recursive --source /tmp/a/ lakefs://repo@master/top/ will create these objects:
lakefs://repo@master/top/foo
lakefs://repo@master/top/b/bar
lakefs://repo@master/top/b/c/d/e/xyzzy
Alternative solution for https://lakefs.slack.com/archives/C016726JLJW/p1606735311131500.
Alternative to #977, probably don't want both...
Thanks! Pulling this one, throwing #977 away.
|
gharchive/pull-request
| 2020-11-30T14:28:14 |
2025-04-01T06:40:41.070132
|
{
"authors": [
"arielshaqed"
],
"repo": "treeverse/lakeFS",
"url": "https://github.com/treeverse/lakeFS/pull/979",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
230711591
|
CancellationException thrown + unhandled after Single subscription disposed (RxJava 2)
I'm not quite sure how to handle this issue.
I have a subscription to a Single that is bound to the lifecycle of a fragment, which I may at some point have to cancel/dispose. Once this has happened, if the fragment is closed the exception is thrown to the system and the app crashes. See below for the stack trace
java.util.concurrent.CancellationException
at io.reactivex.internal.operators.single.SingleTakeUntil$TakeUntilOtherSubscriber.onNext(SingleTakeUntil.java:150)
at io.reactivex.internal.subscriptions.DeferredScalarSubscription.complete(DeferredScalarSubscription.java:133)
at io.reactivex.internal.operators.single.SingleToFlowable$SingleToFlowableObserver.onSuccess(SingleToFlowable.java:63)
at io.reactivex.internal.operators.observable.ObservableElementAtSingle$ElementAtObserver.onNext(ObservableElementAtSingle.java:91)
at io.reactivex.internal.operators.observable.ObservableFilter$FilterObserver.onNext(ObservableFilter.java:52)
at io.reactivex.internal.operators.observable.ObservableOnErrorReturn$OnErrorReturnObserver.onNext(ObservableOnErrorReturn.java:66)
at io.reactivex.internal.operators.observable.ObservableCombineLatest$LatestCoordinator.drain(ObservableCombineLatest.java:246)
at io.reactivex.internal.operators.observable.ObservableCombineLatest$LatestCoordinator.combine(ObservableCombineLatest.java:199)
at io.reactivex.internal.operators.observable.ObservableCombineLatest$CombinerObserver.onNext(ObservableCombineLatest.java:316)
at io.reactivex.internal.operators.observable.ObservableSkip$SkipObserver.onNext(ObservableSkip.java:53)
at io.reactivex.internal.operators.observable.ObservableRefCount$ConnectionObserver.onNext(ObservableRefCount.java:169)
at io.reactivex.internal.operators.observable.ObservablePublish$PublishObserver.onNext(ObservablePublish.java:233)
at io.reactivex.internal.util.NotificationLite.accept(NotificationLite.java:246)
at io.reactivex.subjects.BehaviorSubject$BehaviorDisposable.test(BehaviorSubject.java:490)
at io.reactivex.subjects.BehaviorSubject$BehaviorDisposable.emitNext(BehaviorSubject.java:485)
at io.reactivex.subjects.BehaviorSubject.onNext(BehaviorSubject.java:184)
at com.trello.rxlifecycle2.components.support.RxFragment.onPause(RxFragment.java:88)
at android.support.v4.app.Fragment.performPause(Fragment.java:2371)
at android.support.v4.app.FragmentManagerImpl.moveToState(FragmentManager.java:1357)
at android.support.v4.app.FragmentManagerImpl.moveFragmentToExpectedState(FragmentManager.java:1528)
at android.support.v4.app.FragmentManagerImpl.moveToState(FragmentManager.java:1595)
at android.support.v4.app.FragmentManagerImpl.dispatchPause(FragmentManager.java:2920)
at android.support.v4.app.FragmentController.dispatchPause(FragmentController.java:234)
at android.support.v4.app.FragmentActivity.onPause(FragmentActivity.java:448)
at com.trello.rxlifecycle2.components.support.RxAppCompatActivity.onPause(RxAppCompatActivity.java:81)
at android.app.Activity.performPause(Activity.java:6819)
at android.app.Instrumentation.callActivityOnPause(Instrumentation.java:1322)
at android.app.ActivityThread.performPauseActivityIfNeeded(ActivityThread.java:3740)
at android.app.ActivityThread.performPauseActivity(ActivityThread.java:3717)
at android.app.ActivityThread.performPauseActivity(ActivityThread.java:3691)
at android.app.ActivityThread.handlePauseActivity(ActivityThread.java:3665)
at android.app.ActivityThread.-wrap16(ActivityThread.java)
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1489)
at android.os.Handler.dispatchMessage(Handler.java:102)
at android.os.Looper.loop(Looper.java:154)
at android.app.ActivityThread.main(ActivityThread.java:6119)
at java.lang.reflect.Method.invoke(Native Method)
at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:886)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:776)
The issue can be seen in SingleTakeUntil.TakeUntilMainObserver.otherError which will throw an exception to the default UncaughtExceptionHandler if the Disposable(subscription) is disposed
How do I deal with this?
Single emits CancellationException when it disposes: https://github.com/trello/RxLifecycle#unsubscription
Due to the way RxLifecycle is setup, there isn't a great way around it. I suggest checking out AutoDispose if you want to automatically dispose of Singles without getting an exception.
Seems like it should come with a clearer warning in the documentation.
In any case, an easy workaround seems to be to convert your Single to an Observable
|
gharchive/issue
| 2017-05-23T13:43:53 |
2025-04-01T06:40:41.074452
|
{
"authors": [
"dlew",
"kassim"
],
"repo": "trello/RxLifecycle",
"url": "https://github.com/trello/RxLifecycle/issues/217",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1419965042
|
Search bar
I think it would be useful to have a search bar component to filter rows of a table/list or other data. I've seen your examples of using a MultiSelectbox box for filtering, but I think the search functionality would be helpful too.
Hey @alexheiner, thanks for the comment! Did you have the chance to check out the SelectBox component already, which essentially provides a search-box to filter among the available options? Or are you looking for something different? Feel free to elaborate 🙂
I actually didn't notice the ability to search in the SelectBox, thanks for pointing that out. Sorry I don't know how I missed that. I was originally thinking of a more general purpose table/data search, such as searching for a specific address, phone number, etc in a table. But I thInk the SelectBox could probably be used to do something similar.
@alexheiner Let us know if a simple search bar (including further features, such as auto complete) still would be more appealing to your use cases! :)
Sounds good! I do think it could be a useful feature, but it could be something that is added down the line in a future release. Thanks for getting back to me about it! I think this is a really cool project, I'm excited to keep using it!
SearchBox does no appear to be documented in the list of components here through. https://www.tremor.so/docs/components
|
gharchive/issue
| 2022-10-23T22:49:26 |
2025-04-01T06:40:41.077971
|
{
"authors": [
"alexheiner",
"christopherkindl",
"mitrotasios",
"tonyxiao"
],
"repo": "tremorlabs/tremor",
"url": "https://github.com/tremorlabs/tremor/issues/145",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
58928164
|
Can I use my own form?
like:
$.prompt({
html: '<form action="/user/settings" method="post"......',
})
Impromptu wraps the all of the states into a single form, so doing so would create a form inside a form, which will cause issues.
|
gharchive/issue
| 2015-02-25T16:31:06 |
2025-04-01T06:40:41.099019
|
{
"authors": [
"trentrichardson",
"zackperdue"
],
"repo": "trentrichardson/jQuery-Impromptu",
"url": "https://github.com/trentrichardson/jQuery-Impromptu/issues/57",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
168201123
|
Query previous/next historical entry
One of the main use cases I've found with django-simple-history has been to get the previous historical entry to calculate deltas in values.
Issue #201 talks about the Django utility functions, which allows for getting the last value by a field. I think it would be great to add a previous property to the historical model that grabbed the get_previous_by_history_date that reflected the same tracked instance automatically. I feel like a next and previous method fits perfectly within the historical records context.
As a dependent though, comparing versions between points in time would be excellent, too, but I'll submit a second issue based on response to this.
What is the best way, given a historical entry, to get the "previous" entry?
In my use case, thing and thing.history.most_recent() have the same field values in all cases, except they don't compare identically because thing._state != thing.history.most_recent()._state, and id(thing) != id(thing.history.most_recent()).
:white_check_mark: https://github.com/treyhunner/django-simple-history/pull/365
|
gharchive/issue
| 2016-07-28T21:27:55 |
2025-04-01T06:40:41.109098
|
{
"authors": [
"SpainTrain",
"ransford",
"willjschmitt"
],
"repo": "treyhunner/django-simple-history",
"url": "https://github.com/treyhunner/django-simple-history/issues/230",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1922749330
|
fix: #526 CLI warn if using @trigger.dev/react package with Next.js pages directory
Closes #526
✅ Checklist
[x] I have followed every step in the contributing guide
[x] The PR title follows the convention.
[x] I ran and tested the code works
Screenshots
This is a bit stale now
|
gharchive/pull-request
| 2023-10-02T21:26:43 |
2025-04-01T06:40:41.138540
|
{
"authors": [
"matt-aitken",
"neo773"
],
"repo": "triggerdotdev/trigger.dev",
"url": "https://github.com/triggerdotdev/trigger.dev/pull/530",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
421820622
|
feat: Upgrade to material-components-web v1.1.0
WIP
closes #1842 #1845
Almost ready!
|
gharchive/pull-request
| 2019-03-16T15:41:05 |
2025-04-01T06:40:46.098722
|
{
"authors": [
"trimox"
],
"repo": "trimox/angular-mdc-web",
"url": "https://github.com/trimox/angular-mdc-web/pull/1845",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1333651515
|
sc-7976 GDS User UI : Remove blue border from css custom-styled button
Scope of changes
sc-7976 GDS User UI : Remove blue border from css custom-styled button
Type of change
[ ] bug fix
[ ] new feature
[ ] documentation
[ ] other (describe)
Acceptance criteria
Describe how reviewers can test this change to be sure that it works correctly. Add a checklist if possible
Author checklist
[x] I have manually tested the change and/or added automation in the form of unit tests or integration tests
[ ] I have updated the dependencies list
[ ] I have recompiled and included new protocol buffers to reflect changes I made
[ ] I have added new test fixtures as needed to support added tests
[x] Check this box if a reviewer can merge this pull request after approval (leave it unchecked if you want to do it yourself)
[x] I have moved the associated Shortcut story to "Ready for Review"
Reviewer(s) checklist
[ ] Any new user-facing content that has been added for this PR has been QA'ed to ensure correct grammar, spelling, and understandability.
Codecov Report
Merging #699 (916900f) into main (c9e02fb) will decrease coverage by 5.80%.
The diff coverage is n/a.
@@ Coverage Diff @@
## main #699 +/- ##
==========================================
- Coverage 43.89% 38.08% -5.81%
==========================================
Files 580 155 -425
Lines 17363 2434 -14929
Branches 1273 618 -655
==========================================
- Hits 7621 927 -6694
+ Misses 8557 1330 -7227
+ Partials 1185 177 -1008
Impacted Files
Coverage Δ
...eb/gds-user-ui/src/components/FileUpload/index.tsx
...i/src/components/NeedsAttention/AttentionAlert.tsx
web/gds-user-ui/src/theme/components/Button.ts
...github.com/trisacrypto/directory/pkg/gds/backup.go
...ficateManagement/CertificateManagement.stories.tsx
...isacrypto/directory/pkg/utils/logger/middleware.go
web/gds-user-ui/src/locales/fr/messages.js
web/gds-user-ui/src/components/NotFound/index.tsx
...om/trisacrypto/directory/pkg/gds/client/profile.go
...-user-ui/src/components/Navbar/Landing/NavItem.tsx
... and 415 more
Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.
|
gharchive/pull-request
| 2022-08-09T18:44:25 |
2025-04-01T06:40:46.193574
|
{
"authors": [
"codecov-commenter",
"elysee15"
],
"repo": "trisacrypto/directory",
"url": "https://github.com/trisacrypto/directory/pull/699",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1929918655
|
Output showing in s (Seconds) instead of micro or milli seconds And Output not in readable format
Added an Image to show the output
In the first you can see it's showing "s" which is not correct
And in the second log it's not in a readable format
Request:-
Add a UTC timestamp as a configurable option maybe because we can stream these logs or access from Docker logs we would be able to confirm with a timestamp
Hi 👋 , it this is accepted I would like to work on this, could you assign it to me 🙂!
Okay, I will take a look at it. The timing got messed up in the last update from a bug I didn't catch.
I could include that. Maybe a few other logging options too like JSON for structured loggers?
@tristanisham
Logger does look good and this will help a lot in future.
You can provide more options, also you can maybe as well use local timestamp so that people developing in their own timezone will be able to relate by the time easier. You can enable this as an option.
What do you have in mind for Structured loggers?
Also are you working on adding testing?
@Yashrocky96 could you make a pull request for the errant seconds label?
As for logging, I'll add some options to enable logging in UTC and support for the TZ variable.
@tristanisham I have raised a PR, Please review and let me know
#5 is almost done. One more change and it's good to merge @Yashrocky96
@tristanisham I have updated the PR
Also, Please check this out if you can.
Story:- When there's an error during execution of onError for example during development you are giving a wrong JSON and you get an error that the input is Failed to parse JSON.
In this scenario, onError hook is executing twice and logging the information twice.
For now, I am fairly new with Elysia and do not understand why that is happening.
@tristanisham Please close this issue and would I not join as a Contributor here?
@tristanisham
Hi,
I am on vacation and will review these changes when I return in
November.
Tristan Isham
On October 23, 2023, Tristan Isham @.***> wrote:
@tristanisham https://github.com/tristanisham
—
Reply to this email directly, view it on GitHub
<https://github.com/tristanisham/logysia/issues/4#issuecomment-
1774702710>, or unsubscribe
<https://github.com/notifications/unsubscribe-
auth/AFQNWUQIXLSLPUOWAMZVWC3YAYUTJAVCNFSM6AAAAAA5VUOQ4GVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTONZUG4YDENZRGA>.
You are receiving this because you were mentioned.Message ID:
@.***>
Hi Tristanisham,
Hope that you are back from your vacations, Would you please add me as a contributor and do a patch release please?
Just send them here? Is there a pull request?
Could you please send them here?
What do I send and how do I send, This is my first Open-Source Contribution @tristanisham . Please guide me. And thanks.
You have already merged my changes from my forked repository
Hi @tristanisham Please do a patch release and add me as a contributor to the package, if my application does use this package.
@tristanisham Please do a patch release
|
gharchive/issue
| 2023-10-06T10:59:16 |
2025-04-01T06:40:46.204365
|
{
"authors": [
"Yashrocky96",
"kareemmahlees",
"tristanisham"
],
"repo": "tristanisham/logysia",
"url": "https://github.com/tristanisham/logysia/issues/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1572484324
|
Do not call yarn directly in the package.json pre commands
Your Environment
Prettier version: 2.8.3
node version [12.x.x, 14.x.x]: 18.13
package manager: pnpm@7.26.3
IDE: VSCode
Describe the bug
When installed via pnpm install, pnpm expresses that the command fails due to yarn missing.
Install a fresh nvm version (like 18.13) and run npm install -g pnpm. Then create a new project add @trivago/prettier-plugin-sort-imports as a dep. Run install and see the error.
Expected behavior
Should work without the need to have a global yarn package manager.
Any ideas?
hello?
Please use require while loading the plugin if you use it pnpm. Something like this https://github.com/verdaccio/verdaccio/blob/c9bd22dcaa2f0d919338fb387ae261992f0e1116/prettier.config.js#L10
Feel free to reopen if you still face the issue.
|
gharchive/issue
| 2023-02-06T12:29:40 |
2025-04-01T06:40:46.226621
|
{
"authors": [
"ayusharma",
"matsko"
],
"repo": "trivago/prettier-plugin-sort-imports",
"url": "https://github.com/trivago/prettier-plugin-sort-imports/issues/206",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
916374962
|
Rename basis part of DG to operator or something similar?
This came up in https://github.com/trixi-framework/Trixi.jl/pull/617#discussion_r644467670. The name basis might not fit very well for FD-SBP methods...
@andrewwinters5000 suggested "spectral approximation", but that doesn't really fit for FD methods, I think. The "approximation" part is probably okay, though - "approximation space"? But that's also quite long compared to "basis"...
I like the approximation-related name. One could also use a name related to the reference element/domain since all quantities are defined there?
One could also use a name related to the reference element/domain since all quantities are defined there?
Like what?
I was thinking ReferenceElementData/ReferenceDomainData after deal.ii's FiniteElementData and GridAP's ReferenceFE, but those names are pretty long.
Hm. So it would be something like LobattoLegendreReferenceElementData? That is quite long...
Yeah, that's definitely too long. Maybe ReferenceElementData would be better as an abstract type, something like
LobattoLegendreElement <: AbstractReferenceElementData{RealT}.
Triage: Not necessary
|
gharchive/issue
| 2021-06-09T16:04:24 |
2025-04-01T06:40:46.231198
|
{
"authors": [
"jlchan",
"ranocha",
"sloede"
],
"repo": "trixi-framework/Trixi.jl",
"url": "https://github.com/trixi-framework/Trixi.jl/issues/633",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1182680642
|
Brew Tap is broken
~ brew install trntv/sshed/sshed 1 [15:45:25]
==> Tapping trntv/sshed
Cloning into '/usr/local/Homebrew/Library/Taps/trntv/homebrew-sshed'...
remote: Enumerating objects: 6, done.
remote: Counting objects: 100% (6/6), done.
remote: Compressing objects: 100% (4/4), done.
remote: Total 6 (delta 2), reused 0 (delta 0), pack-reused 0
Receiving objects: 100% (6/6), done.
Resolving deltas: 100% (2/2), done.
Error: Invalid formula: /usr/local/Homebrew/Library/Taps/trntv/homebrew-sshed/sshed.rb
sshed: Calling bottle :unneeded is disabled! There is no replacement.
Please report this issue to the trntv/sshed tap (not Homebrew/brew or Homebrew/core):
/usr/local/Homebrew/Library/Taps/trntv/homebrew-sshed/sshed.rb:9
Error: Cannot tap trntv/sshed: invalid syntax in tap!
Same for me with M1 Mac Pro
Same here
|
gharchive/issue
| 2022-03-27T22:13:21 |
2025-04-01T06:40:46.234176
|
{
"authors": [
"basarozcan",
"gabrielbryk",
"rinchen"
],
"repo": "trntv/sshed",
"url": "https://github.com/trntv/sshed/issues/16",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1103968108
|
🛑 PiHole is down
In e4592e5, PiHole (https://pihole.tronflix.app/admin/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: PiHole is back up in d0b14bc.
|
gharchive/issue
| 2022-01-14T18:01:42 |
2025-04-01T06:40:46.306625
|
{
"authors": [
"tronyx"
],
"repo": "tronyx/upptime",
"url": "https://github.com/tronyx/upptime/issues/152",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1437488943
|
🛑 Radarr Anime is down
In 2bf7d09, Radarr Anime (https://tronflix.app/radarr-anime/activity/queue/) was down:
HTTP code: 520
Response time: 103 ms
Resolved: Radarr Anime is back up in 150bc0e.
|
gharchive/issue
| 2022-11-06T18:02:59 |
2025-04-01T06:40:46.309116
|
{
"authors": [
"tronyx"
],
"repo": "tronyx/upptime",
"url": "https://github.com/tronyx/upptime/issues/1988",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1454414529
|
🛑 Sonarr is down
In fa8805e, Sonarr (https://tronflix.app/sonarr/activity/queue/) was down:
HTTP code: 521
Response time: 38 ms
Resolved: Sonarr is back up in cdfc73d.
|
gharchive/issue
| 2022-11-18T04:19:52 |
2025-04-01T06:40:46.311339
|
{
"authors": [
"tronyx"
],
"repo": "tronyx/upptime",
"url": "https://github.com/tronyx/upptime/issues/3618",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2670403934
|
🛑 Overseerr is down
In 0210f49, Overseerr (https://overseerr.tronflix.app/login) was down:
HTTP code: 523
Response time: 19906 ms
Resolved: Overseerr is back up in 1cafcc4 after 1 hour, 29 minutes.
|
gharchive/issue
| 2024-11-19T00:48:09 |
2025-04-01T06:40:46.313759
|
{
"authors": [
"tronyx"
],
"repo": "tronyx/upptime",
"url": "https://github.com/tronyx/upptime/issues/5850",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2738636405
|
🛑 Grafana is down
In be12d86, Grafana (https://grafana.tronflix.app) was down:
HTTP code: 521
Response time: 99 ms
Resolved: Grafana is back up in 8b12c10 after 3 hours, 20 minutes.
|
gharchive/issue
| 2024-12-13T15:22:12 |
2025-04-01T06:40:46.315998
|
{
"authors": [
"tronyx"
],
"repo": "tronyx/upptime",
"url": "https://github.com/tronyx/upptime/issues/5979",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
65128708
|
multiple instances of gridster
Hi Troolee,
I am trying to create 2 separate instances of gridster on one page how ever i have found some difficulties as they are using the same naming conventions. I have tried to initialize gridster with options (item_class ,placeholder_class) how ever I couldnt see any differences. Could you advise pls if gridster support multiple instances on one project and if yes ;) how correctly initialize each instance. I am trying :
for second instance I am trying :
var options = {
cell_height: 10,
vertical_margin: 10,
item_class: "my_iteam_class",
placeholder_class: "my_placeholder_class"
};
$('.grid-stack.test').gridstack(options);
Thanks for advance...
I'm also having problems with this.
@kozilla did you managed to put this to work?
@troolee Any suggestion?
Please check 2 grids demo http://troolee.github.io/gridstack.js/demo/two.html
|
gharchive/issue
| 2015-03-30T03:50:55 |
2025-04-01T06:40:46.318547
|
{
"authors": [
"jpsfs",
"kozilla",
"troolee"
],
"repo": "troolee/gridstack.js",
"url": "https://github.com/troolee/gridstack.js/issues/111",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
230477453
|
nerfing pin/multi-pin
Per issue #67 Attempt to balance marksman pin/multi-pin abilities.
Nerfed the bonus accuracy. Redesigned other things.
I tried to reword some text to remove vagueness.
What do you think of the following changes:
"If the target is fired at by you" -> "If you hit the selected target"
The effect should only trigger when a target is hit.
"The Pin/Multi-Pin ability can be activated by a pinned target up to 3 times per Marksman player's turn cycle" -> "This ability allows a marksman to react to a pinned target once per turn."
This version of the feat only allows one target to be pinned, so it doesn't makes sense that the only pinned target would trigger the marksman's reaction shot more than once per turn.
Also, "Only 1 target may be pinned by the marksman at a time." should be added to the first iteration of this feat.
"You may select and pin up to PER/2 (round down) targets at a time (you have to fire at a target to pin it).+30% to hit a Pinned target (instead of +10%). Only usable with long-rifles. " -> "You may monitor and react to up to PER/2 (round down) pinned targets at a time. +20% to hit a pinned target. Only usable with Long Rifles."
Cleared up some phrasing. A target is only pinned when hit, that is defined in the first section of the feat so we don't have to mention that again. The original +20% replaces any other bonus, this is the intended behavior of all class feats. if class X has a level 5 feat that says "+15% to do Y", and the level 10 variant says "+30% to do Y" that means the total bonus to performing Y is 30%. This phrasing is fairly ambiguous and I apologize for that. Maybe we should change all second-tier class feats so that bonuses are written as "The bonus for doing Y increases to +30%". That is fairly explicit.
Are we also adding in that this ability takes actions off of the marksman's turn, or leaving it as-is? I'm fine with either decision, I thought'd I'd bring it up since we'd discussed it in the past.
What do you think of the following changes:
"If the target is fired at by you" -> "If you hit the selected target"
The effect should only trigger when a target is hit.
Why should the effect only trigger when a target is hit?
My reasoning was that I didn't want to straight nerf marksman, so it gives them an advantage this way while other ways get nerfed. Also realism, 'pinning' sounds analogous to 'suppressing' a target, and that is more about psychological warfare than physical damage. Also, it would even make sense if you just aimed at him, for him to be pinned, since him coming out of cover would be stupid. I suppose you may be considering the first shot to be a maiming shot? Also, most long rifles are bolt action, making it much harder to pin targets if it's an on-hit effect (especially in regards to multi-pin).
"The Pin/Multi-Pin ability can be activated by a pinned target up to 3 times per Marksman player's turn cycle" -> "This ability allows a marksman to react to a pinned target once per turn."
This version of the feat only allows one target to be pinned, so it doesn't makes sense that the only pinned target would trigger the marksman's reaction shot more than once per turn.
Why doesn't it make sense? In my mind, the target could be darting from cover to cover on their turn. If I was trying to kill someone, I would be shooting every chance I get.
Also, "Only 1 target may be pinned by the marksman at a time." should be added to the first iteration of this feat.
"If the target dies you may select a new target." already says that, but we can make it more explicit if you prefer.
"You may select and pin up to PER/2 (round down) targets at a time (you have to fire at a target to pin it).+30% to hit a Pinned target (instead of +10%). Only usable with long-rifles. " -> "You may monitor and react to up to PER/2 (round down) pinned targets at a time. +20% to hit a pinned target. Only usable with Long Rifles."
Cleared up some phrasing. A target is only pinned when hit, that is defined in the first section of the feat so we don't have to mention that again. The original +20% replaces any other bonus, this is the intended behavior of all class feats. if class X has a level 5 feat that says "+15% to do Y", and the level 10 variant says "+30% to do Y" that means the total bonus to performing Y is 30%. This phrasing is fairly ambiguous and I apologize for that. Maybe we should change all second-tier class feats so that bonuses are written as "The bonus for doing Y increases to +30%". That is fairly explicit.
Ok.
Are we also adding in that this ability takes actions off of the marksman's turn, or leaving it as-is? I'm fine with either decision, I thought'd I'd bring it up since we'd discussed it in the past.
I put the following in, "The Pin/Multi-Pin ability can be activated by a pinned target up to 3 times per Marksman player's turn cycle" for that reason. I figured that it would be ok to grant the marksman the ability to react without extra expense.
I think this is good discussion!
I felt pinned by Alex's nerf monstrosity at nerf wars in past even when not hit by his shots.
If the trigger is the target leaving cover, how would it be possible for the marksman's ability to trigger more than once per turn per target? That also doesn't make sense to me.
We should make that much more explicit. Players sometimes create loopholes when they don't quite understand what's written. We should avoid attempting to communicate with nuance.
Whichever way we communicate it, we need to make it very clear.
I still don't understand how a target could proc 3 sniper reaction shots in a single turn.
@bleehu,
2. If a target re-enters cover and leaves it again.
Example:, I am pinned and hiding behind light cover in a hallway, I run ahead to then next cover in the hallway, proc'ing marksman Pin. Then I leave from that cover, to a third cover. Proc'ing marksman pin a 2nd time.
5. I tried to make it about multi-pin as well as 'single'-pin in one swoop. The way I meant it to be written is that the Marksman player cannot shoot more than 3 reaction shots per his full turn cycle, no matter how many people he is pinning.
That makes a little more sense. =)
@trowl223, fyi I won't do (don't want to do) any more work on this until you respond to the recent post(s) =)
(I have plenty of other things to do, so take your time)
So sorry! working on this now.
Makes sense.
I don't think so. The marksman is attuned to only the precise locations of the target he has pinned. if one leaves cover and is not hit by the reaction shot, the marksman has lost track of them temporarily and must readjust. The time frames seem to be in favor of the marksman too much. three shots in the span of their enemy's turn while they've already had a full turn is like granting them extra actions.
The situation will not always be that someone is moving in a straight line down a hallway towards you as in your given example. In that instance, it may be up to the DM to allow that if the Marksman has a semi-auto rifle since they don't have to readjust, but in a different example this may not be the case. What if the Marksman was providing cover to their squad from a bell-tower positioned 200 meters from combat, and they have to watch a 300m spread of land in front of them? A pinned target moving side-to-side form the Marksman's POV would require considerable time for the marksman to readjust if they miss the first shot as the target leaves cover.
My thoughts exactly. An ability of this caliber needs to be pretty explicit.
I still don't think that they should receive an extended amount of extra firing time. At the second tier, I think that with multiple targets the Marksman can have multiple reaction shots (as per my original edit) but at the first tier we're talking about a potential Level 5 ability, its too much out of turn action IMO.
@trowl223, sounds good.
Please review my latest commit
|
gharchive/pull-request
| 2017-05-22T18:18:16 |
2025-04-01T06:40:46.336477
|
{
"authors": [
"1sourcecontrol",
"bleehu",
"trowl223"
],
"repo": "trowl223/Compound_X",
"url": "https://github.com/trowl223/Compound_X/pull/107",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1822581803
|
NAS-123236 / 23.10 / Allow waiting for jobs to complete
Context
A new parameter has been added to MiddlewareCommand to allow waiting for a job to complete so we get the job's endpoints output and not the job number.
Jira URL: https://ixsystems.atlassian.net/browse/NAS-123236
|
gharchive/pull-request
| 2023-07-26T14:44:25 |
2025-04-01T06:40:46.390837
|
{
"authors": [
"Qubad786",
"bugclerk"
],
"repo": "truenas/ixdiagnose",
"url": "https://github.com/truenas/ixdiagnose/pull/29",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2274370803
|
broomwood, folks
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[b,f]
/trunk merge
|
gharchive/pull-request
| 2024-05-02T00:03:51 |
2025-04-01T06:40:46.474204
|
{
"authors": [
"mmatheson"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/12951",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2278645807
|
favillous, gigantostracous
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[f,g]
/trunk merge
|
gharchive/pull-request
| 2024-05-04T01:24:16 |
2025-04-01T06:40:46.476100
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/14364",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2279494188
|
fostered, crosses
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[c,f]
/trunk merge
|
gharchive/pull-request
| 2024-05-05T12:03:28 |
2025-04-01T06:40:46.477689
|
{
"authors": [
"epes"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/15406",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2386947651
|
drilling, enters
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[d,e]
/trunk merge
|
gharchive/pull-request
| 2024-07-02T19:05:48 |
2025-04-01T06:40:46.479230
|
{
"authors": [
"mmatheson"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/16514",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2389784460
|
aerometry, dominion
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[a,d]
/trunk merge
|
gharchive/pull-request
| 2024-07-04T02:17:30 |
2025-04-01T06:40:46.480647
|
{
"authors": [
"epes"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/17863",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2393776690
|
goats, exactingness
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[e,g]
/trunk merge
|
gharchive/pull-request
| 2024-07-06T23:08:15 |
2025-04-01T06:40:46.482222
|
{
"authors": [
"mmatheson"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/21343",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2398739126
|
galvanoplastical, descanting
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[d,g]
/trunk merge
|
gharchive/pull-request
| 2024-07-09T17:09:01 |
2025-04-01T06:40:46.483846
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/22916",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2401468170
|
culturization
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[c]
/trunk merge
|
gharchive/pull-request
| 2024-07-10T19:09:31 |
2025-04-01T06:40:46.485610
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/23420",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2412123985
|
antimedication, detrect
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[a,d]
/trunk merge
|
gharchive/pull-request
| 2024-07-16T22:09:27 |
2025-04-01T06:40:46.487116
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/25755",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2420639699
|
fiscalization, accounters
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[a,f]
/trunk merge
|
gharchive/pull-request
| 2024-07-20T07:07:52 |
2025-04-01T06:40:46.488655
|
{
"authors": [
"mmatheson"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/27175",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2421055504
|
dovecot, fuddledness
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[d,f]
/trunk merge
|
gharchive/pull-request
| 2024-07-20T21:04:17 |
2025-04-01T06:40:46.490229
|
{
"authors": [
"mmatheson"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/27497",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2421525688
|
atypicality, bractlets
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[a,b]
/trunk merge
|
gharchive/pull-request
| 2024-07-21T17:08:02 |
2025-04-01T06:40:46.491931
|
{
"authors": [
"mmatheson"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/28145",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2426547392
|
fireplough
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[f]
/trunk merge
|
gharchive/pull-request
| 2024-07-24T04:04:17 |
2025-04-01T06:40:46.493536
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/30714",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2428935389
|
emesidae, chromatophoric
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[c,e]
/trunk merge
|
gharchive/pull-request
| 2024-07-25T04:05:33 |
2025-04-01T06:40:46.495352
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/31916",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2434153771
|
criminals, excorticating
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[c,e]
/trunk merge
|
gharchive/pull-request
| 2024-07-28T23:07:52 |
2025-04-01T06:40:46.496967
|
{
"authors": [
"epes"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/36399",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2437184287
|
gendarmes, conceivability
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[c,g]
/trunk merge
|
gharchive/pull-request
| 2024-07-30T08:05:57 |
2025-04-01T06:40:46.498570
|
{
"authors": [
"mmatheson"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/37980",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2261023853
|
genteelest, epencephalic
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[e,g]
/trunk merge
|
gharchive/pull-request
| 2024-04-24T11:03:44 |
2025-04-01T06:40:46.500119
|
{
"authors": [
"epes"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/7572",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2262087802
|
equimultiple, frequence
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[e,f]
/trunk merge
|
gharchive/pull-request
| 2024-04-24T20:02:08 |
2025-04-01T06:40:46.501695
|
{
"authors": [
"epes"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/7819",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2237065894
|
aasvogel, disarray (logical-conflict)
This pull request was generated by the 'mq' tool
/trunk merge
|
gharchive/pull-request
| 2024-04-11T07:17:00 |
2025-04-01T06:40:46.502422
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/870",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2420912047
|
gurdy, blinking
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[b,g]
/trunk merge
|
gharchive/pull-request
| 2024-07-20T14:05:52 |
2025-04-01T06:40:46.504012
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/108428",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2233163453
|
bilimbi
This pull request was generated by the 'mq' tool
/trunk merge
|
gharchive/pull-request
| 2024-04-09T10:53:29 |
2025-04-01T06:40:46.504994
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/15773",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2243184927
|
foism, debussed
This pull request was generated by the 'mq' tool
/trunk merge
|
gharchive/pull-request
| 2024-04-15T09:34:28 |
2025-04-01T06:40:46.505677
|
{
"authors": [
"joshmarinacci"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/26303",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2245242536
|
emydian, diddering
This pull request was generated by the 'mq' tool
/trunk merge
|
gharchive/pull-request
| 2024-04-16T06:43:28 |
2025-04-01T06:40:46.506354
|
{
"authors": [
"joshmarinacci"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/28286",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2248128487
|
feater, desipience
This pull request was generated by the 'mq' tool
/trunk merge
|
gharchive/pull-request
| 2024-04-17T12:04:58 |
2025-04-01T06:40:46.507049
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/30508",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2249253383
|
cycloheptanone, goldeneye
This pull request was generated by the 'mq' tool
/trunk merge
|
gharchive/pull-request
| 2024-04-17T20:54:59 |
2025-04-01T06:40:46.507732
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/31398",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2253945564
|
desipiency, esparcet
This pull request was generated by the 'mq' tool
/trunk merge
|
gharchive/pull-request
| 2024-04-19T20:35:52 |
2025-04-01T06:40:46.508420
|
{
"authors": [
"mmatheson"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/36644",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2254591220
|
effervescence, garbure
This pull request was generated by the 'mq' tool
/trunk merge
|
gharchive/pull-request
| 2024-04-20T15:05:30 |
2025-04-01T06:40:46.509110
|
{
"authors": [
"mmatheson"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/38609",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2254895623
|
cosignificative, bowlers
This pull request was generated by the 'mq' tool
/trunk merge
|
gharchive/pull-request
| 2024-04-21T04:29:17 |
2025-04-01T06:40:46.509825
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/39979",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2255092666
|
farsakh
This pull request was generated by the 'mq' tool
/trunk merge
|
gharchive/pull-request
| 2024-04-21T13:55:59 |
2025-04-01T06:40:46.510595
|
{
"authors": [
"mmatheson"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/41007",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2255171996
|
grandmother, antifermentative
This pull request was generated by the 'mq' tool
/trunk merge
|
gharchive/pull-request
| 2024-04-21T16:55:01 |
2025-04-01T06:40:46.511297
|
{
"authors": [
"mmatheson"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/41339",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2256195102
|
fadridden, absorbition
This pull request was generated by the 'mq' tool
/trunk merge
|
gharchive/pull-request
| 2024-04-22T10:38:51 |
2025-04-01T06:40:46.512005
|
{
"authors": [
"mmatheson"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/43189",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2257404565
|
conventionality, excoecaria
This pull request was generated by the 'mq' tool
/trunk merge
|
gharchive/pull-request
| 2024-04-22T20:27:04 |
2025-04-01T06:40:46.512928
|
{
"authors": [
"joshmarinacci"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/44246",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2284951764
|
frangula, bureaucratical
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 2100s
close stale after: 24 hours
[pullrequest]
requests per hour: 100
deps=[b,f]
/trunk merge
|
gharchive/pull-request
| 2024-05-08T07:59:22 |
2025-04-01T06:40:46.514543
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/47745",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2287747019
|
appaloosa, flavobacterium
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 2100s
close stale after: 24 hours
[pullrequest]
requests per hour: 100
deps=[a,f]
/trunk merge
|
gharchive/pull-request
| 2024-05-09T13:53:08 |
2025-04-01T06:40:46.516157
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/48534",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2290289390
|
charcoalist
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 2100s
close stale after: 24 hours
[pullrequest]
requests per hour: 100
deps=[c]
/trunk merge
|
gharchive/pull-request
| 2024-05-10T19:26:42 |
2025-04-01T06:40:46.517760
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/49262",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2292419265
|
chaton, agruif
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 2100s
close stale after: 24 hours
[pullrequest]
requests per hour: 100
deps=[a,c]
/trunk merge
|
gharchive/pull-request
| 2024-05-13T10:25:53 |
2025-04-01T06:40:46.519344
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/51089",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2322143373
|
footnoted, anerethisia
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 2100s
close stale after: 24 hours
[pullrequest]
requests per hour: 100
deps=[a,f]
/trunk merge
|
gharchive/pull-request
| 2024-05-29T00:59:02 |
2025-04-01T06:40:46.520980
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/60563",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2325866079
|
emprizes, glitch
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 2100s
close stale after: 24 hours
[pullrequest]
requests per hour: 100
deps=[e,g]
/trunk merge
|
gharchive/pull-request
| 2024-05-30T14:48:47 |
2025-04-01T06:40:46.522588
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/61964",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2329148117
|
gallivorous, consentaneously
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 2100s
close stale after: 24 hours
[pullrequest]
requests per hour: 100
deps=[c,g]
/trunk merge
|
gharchive/pull-request
| 2024-06-01T13:31:36 |
2025-04-01T06:40:46.524424
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/63536",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2336647040
|
dissyllabify, foredoes
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[d,f]
/trunk merge
|
gharchive/pull-request
| 2024-06-05T19:12:17 |
2025-04-01T06:40:46.526038
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/66950",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2340395494
|
botong, apodous
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[a,b]
/trunk merge
|
gharchive/pull-request
| 2024-06-07T12:58:19 |
2025-04-01T06:40:46.527606
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/68596",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2342272863
|
betattered
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[b]
/trunk merge
|
gharchive/pull-request
| 2024-06-09T13:58:01 |
2025-04-01T06:40:46.529221
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/70536",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2342699236
|
excommunicable
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[e]
/trunk merge
|
gharchive/pull-request
| 2024-06-10T02:25:43 |
2025-04-01T06:40:46.530834
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/71026",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2350291037
|
azotise, exploit
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[a,e]
/trunk merge
|
gharchive/pull-request
| 2024-06-13T06:43:06 |
2025-04-01T06:40:46.532612
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/74010",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2351757096
|
belauds, dicyclopentadienyliron
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[b,d]
/trunk merge
|
gharchive/pull-request
| 2024-06-13T18:02:23 |
2025-04-01T06:40:46.534453
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/74460",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2352179224
|
franklinite, clarigation
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[c,f]
/trunk merge
|
gharchive/pull-request
| 2024-06-13T22:48:58 |
2025-04-01T06:40:46.536026
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/74649",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2354354842
|
arachnoidal, embolectomies
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[a,e]
/trunk merge
|
gharchive/pull-request
| 2024-06-15T01:03:32 |
2025-04-01T06:40:46.537693
|
{
"authors": [
"EliSchleifer"
],
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/75681",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.