id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
651032512
Token and expected token differ on a backslash ( %5C ) I am getting the following error: solver.go:79] cert-manager/acmesolver "msg"="comparing token" "base_path"="/.well-known/acme-challenge" "host"="example.com" "path"="/.well-known/acme-challenge/opBqXqa5UCniT9uUKkqSZyENLDR4aaX15oqJzt5S8MU%5C" "token"="opBqXqa5UCniT9uUKkqSZyENLDR4aaX15oqJzt5S8MU%5C" "expected_token"="opBqXqa5UCniT9uUKkqSZyENLDR4aaX15oqJzt5S8MU" The token and the expected token should match. But they differ on a backslash at the end (represented by %5C) If I curl without the backslash the token is there: curl "http://example.com/.well-known/acme-challenge/opBqXqa5UCniT9uUKkqSZyENLDR4aaX15oqJzt5S8MU" opBqXqa5UCniT9uUKkqSZyENLDR4aaX15oqJzt5S8MU.vlkT-oHqQGI2e1QtLfbS7OZfyt4o-9N72K7IkUc_LNo Environment details:: Kubernetes version 1.18.4+k3s1 Image: quay.io/jetstack/cert-manager-controller:v0.15.1 Install method: helm I am very curious where this backslash comes from... Can you get us the output of kubectl describe challenge and kubectl get challenge -o yaml. I assume you're using let's encrypt? Also what ingress controller are you using? I should preface by saying that I removed cert-manager a bit after reporting this bug, because I'm rebuilding my cluster. This is what I get: % kubectl get challenge -o yaml apiVersion: v1 items: [] kind: List metadata: resourceVersion: "" selfLink: "" Error from server: conversion webhook for acme.cert-manager.io/v1alpha2, Kind=Challenge failed: Post https://cert-manager-webhook.cert-manager.svc:443/convert?timeout=30s: service "cert-manager-webhook" not found However, I'd like to point out that I was getting those "service not found" errors before I got rid of cert-manager as well. I'm using nginx and I also have metallb. Also, I don't know if this is useful (or related to this, since I'm still learning all this stuff), but I had to edit the cert-manager deployment to override the DNS from ClusterFirst to None and define an external DNS resolver because core-dns / kube-dns wasn't able to resolve my domain. However, I'd like to point out that I was getting those "service not found" errors before I got rid of cert-manager as well. This indicates an issue with the installation, if you see those cert-manager might also not be able to function correctly /triage support I didn't do anything special for the installation. This is the command I ran: helm install cert-manager jetstack/cert-manager --namespace kube-system I'll rebuild my cluster and deploy cert-manager on its own cert-manager ns. I might be able to get that done today and I can report if I see any backslash issue. Can you try helm install cert-manager jetstack/cert-manager --namespace cert-manager It is not advised to install cert-manager in kube-system. Okay, I reinstalled cert-manager following the official installation guide. I installed from Helm like this: helm install cert-manager jetstack/cert-manager --namespace cert-manager --version v0.15.1 --set installCRDs=true % kubectl describe challenge -n www Name: www-example-com-tls-1737868453-886957241-3604311114 Namespace: www Labels: <none> Annotations: <none> API Version: acme.cert-manager.io/v1alpha3 Kind: Challenge Metadata: Creation Timestamp: 2020-07-08T01:04:06Z Finalizers: finalizer.acme.cert-manager.io Generation: 1 Owner References: API Version: acme.cert-manager.io/v1alpha2 Block Owner Deletion: true Controller: true Kind: Order Name: www-example-com-tls-1737868453-886957241 UID: 57e1c991-1569-43d0-8414-4729e0f18556 Resource Version: 35484 Self Link: /apis/acme.cert-manager.io/v1alpha3/namespaces/www/challenges/www-example-com-tls-1737868453-886957241-3604311114 UID: 07a0246d-a636-4bb4-8b15-5ece17c3e70b Spec: Authz URL: https://acme-v02.api.letsencrypt.org/acme/authz-v3/5735132020 Dns Name: www.example.com Issuer Ref: Group: cert-manager.io Kind: ClusterIssuer Name: letsencrypt-prod Key: 6flstbp94Nan2alAvXQ5GpkNHBXdJWxLDjguzu8RsIw.2t2Ll636CQnzqLtsb2Aicd3cgIwCIpcHb9-mzrfFn08 Solver: http01: Ingress: Class: nginx Token: 6flstbp94Nan2alAvXQ5GpkNHBXdJWxLDjguzu8RsIw Type: http-01 URL: https://acme-v02.api.letsencrypt.org/acme/chall-v3/5735132020/c9Lg9g Wildcard: false Status: Presented: true Processing: true Reason: Waiting for http-01 challenge propagation: failed to perform self check GET request 'http://www.example.com/.well-known/acme-challenge/6flstbp94Nan2alAvXQ5GpkNHBXdJWxLDjguzu8RsIw': Get "http://www.example.com/.well-known/acme-challenge/6flstbp94Nan2alAvXQ5GpkNHBXdJWxLDjguzu8RsIw": dial tcp [redacted external-facing IP]:80: connect: connection timed out State: pending Events: <none> % kubectl get challenge -n www -o yaml apiVersion: v1 items: - apiVersion: acme.cert-manager.io/v1alpha3 kind: Challenge metadata: creationTimestamp: "2020-07-08T01:04:06Z" finalizers: - finalizer.acme.cert-manager.io generation: 1 name: www-example-com-tls-1737868453-886957241-3604311114 namespace: www ownerReferences: - apiVersion: acme.cert-manager.io/v1alpha2 blockOwnerDeletion: true controller: true kind: Order name: www-example-com-tls-1737868453-886957241 uid: 57e1c991-1569-43d0-8414-4729e0f18556 resourceVersion: "35484" selfLink: /apis/acme.cert-manager.io/v1alpha3/namespaces/www/challenges/www-example-com-tls-1737868453-886957241-3604311114 uid: 07a0246d-a636-4bb4-8b15-5ece17c3e70b spec: authzURL: https://acme-v02.api.letsencrypt.org/acme/authz-v3/5735132020 dnsName: www.example.com issuerRef: group: cert-manager.io kind: ClusterIssuer name: letsencrypt-prod key: 6flstbp94Nan2alAvXQ5GpkNHBXdJWxLDjguzu8RsIw.2t2Ll636CQnzqLtsb2Aicd3cgIwCIpcHb9-mzrfFn08 solver: http01: ingress: class: nginx token: 6flstbp94Nan2alAvXQ5GpkNHBXdJWxLDjguzu8RsIw type: http-01 url: https://acme-v02.api.letsencrypt.org/acme/chall-v3/5735132020/c9Lg9g wildcard: false status: presented: true processing: true reason: 'Waiting for http-01 challenge propagation: failed to perform self check GET request ''http://www.example.com/.well-known/acme-challenge/6flstbp94Nan2alAvXQ5GpkNHBXdJWxLDjguzu8RsIw'': Get "http://www.example.com/.well-known/acme-challenge/6flstbp94Nan2alAvXQ5GpkNHBXdJWxLDjguzu8RsIw": dial tcp [redacted external-facing IP]:80: connect: connection timed out' state: pending kind: List metadata: resourceVersion: "" selfLink: "" From cert-manager pod's logs: E0708 01:19:36.554148 1 sync.go:185] cert-manager/controller/challenges "msg"="propagation check failed" "error"="failed to perform self check GET request 'http://www.example.com/.well-known/acme-challenge/6flstbp94Nan2alAvXQ5GpkNHBXdJWxLDjguzu8RsIw': Get \"http://www.example.com/.well-known/acme-challenge/6flstbp94Nan2alAvXQ5GpkNHBXdJWxLDjguzu8RsIw\": dial tcp [redacted external IP]:80: connect: connection timed out" "dnsName"="www.example.com" "resource_kind"="Challenge" "resource_name"="www-example-com-tls-1737868453-886957241-3604311114" "resource_namespace"="www" "type"="http-01" % curl http://www.example.com/.well-known/acme-challenge/6flstbp94Nan2alAvXQ5GpkNHBXdJWxLDjguzu8RsIw 6flstbp94Nan2alAvXQ5GpkNHBXdJWxLDjguzu8RsIw.2t2Ll636CQnzqLtsb2Aicd3cgIwCIpcHb9-mzrfFn08% % curl http://www.example.com/.well-known/acme-challenge/6flstbp94Nan2alAvXQ5GpkNHBXdJWxLDjguzu8RsIw%5C 404 page not found :80: connect: connection timed out indicates that the cluster itself is not able to reach the external IP. Can you run the curl command from inside a pod from inside the cluster? You are correct, I checked and I couldn't curl from inside the pod. The route was failing, so I ended up adding a DNS entry to my pihole hardcoding this to the internal address and everything automagically worked and I have a certificate. I still don't know where that backslash came from, but it doesn't seem to affect certificate generation. I have a feeling it could be some logging issue... Good to hear it works now! Will close this issue. Feel free to /reopen if you run into it again! /close
gharchive/issue
2020-07-05T09:12:41
2025-04-01T04:34:40.677696
{ "authors": [ "leorockway", "meyskens" ], "repo": "jetstack/cert-manager", "url": "https://github.com/jetstack/cert-manager/issues/3070", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1026336979
This software doesn't work at all Hello. Since yesterday almost 24 hours, I am trying to make cert-manager work and I didn't succeed at all. All pods are up: # kubectl -n cert-manager get deployment,pod,svc NAME READY UP-TO-DATE AVAILABLE AGE deployment.apps/cert-manager 1/1 1 1 11h deployment.apps/cert-manager-cainjector 1/1 1 1 11h deployment.apps/cert-manager-webhook 1/1 1 1 11h NAME READY STATUS RESTARTS AGE pod/cert-manager-65d4f87f59-mqzjr 1/1 Running 0 20m pod/cert-manager-cainjector-6fd58df9bd-9n7fn 1/1 Running 0 20m pod/cert-manager-webhook-5498c57bb5-vth7z 1/1 Running 0 20m NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE service/cert-manager ClusterIP 172.24.201.104 <none> 9402/TCP 11h service/cert-manager-webhook ClusterIP 172.24.14.11 <none> 443/TCP 11h I made all steps according by the doc: https://cert-manager.io/docs/configuration/acme/dns01/route53/ The certificate is Ready: # kubectl -n production get certificate/httpbin2 NAME READY SECRET AGE httpbin2 True httpbin2 5m19s However, this certificate is issued by cert-manager.local, not by Let's encrypt. This means that the cert-manager did something completely wrong. In the log file I see the next message: I1014 12:17:18.967271 1 conditions.go:261] Setting lastTransitionTime for CertificateRequest "httpbin2-ql7w8" condition "Approved" to 2021-10-14 12:17:18.96726296 +0000 UTC m=+1027.621572984 I1014 12:17:19.106569 1 conditions.go:261] Setting lastTransitionTime for CertificateRequest "httpbin2-ql7w8" condition "Ready" to 2021-10-14 12:17:19.106559386 +0000 UTC m=+1027.760869380 I1014 12:17:19.176015 1 conditions.go:261] Setting lastTransitionTime for CertificateRequest "httpbin2-ql7w8" condition "Ready" to 2021-10-14 12:17:19.176006326 +0000 UTC m=+1027.830316320 I1014 12:17:19.219653 1 controller.go:161] cert-manager/controller/certificaterequests-issuer-acme "msg"="re-queuing item due to optimistic locking on resource" "key"="production/httpbin2-ql7w8" "error"="Operation cannot be fulfilled on certificaterequests.cert-manager.io \"httpbin2-ql7w8\": the object has been modified; please apply your changes to the latest version and try again" I1014 12:17:19.286650 1 conditions.go:190] Found status change for Certificate "httpbin2" condition "Ready": "False" -> "True"; setting lastTransitionTime to 2021-10-14 12:17:19.286637385 +0000 UTC m=+1027.940947380 E1014 12:17:21.839988 1 sync.go:185] cert-manager/controller/challenges "msg"="propagation check failed" "error"="dial tcp 172.24.0.10:53: i/o timeout" "dnsName"="mydomain.domain.com" "resource_kind"="Challenge" "resource_name"="httpbin-csh7g-583384746-1109480142" "resource_namespace"="production" "resource_version"="v1" "type"="DNS-01" I1014 12:17:21.881038 1 controller.go:161] cert-manager/controller/challenges "msg"="re-queuing item due to optimistic locking on resource" "key"="production/httpbin-csh7g-583384746-1109480142" "error"="Operation cannot be fulfilled on challenges.acme.cert-manager.io \"httpbin-csh7g-583384746-1109480142\": the object has been modified; please apply your changes to the latest version and try again" E1014 12:18:08.709081 1 util.go:80] cert-manager/controller/orders/handleOwnedResource "msg"="error getting referenced owning resource" "error"="order.acme.cert-manager.io \"httpbin-csh7g-583384746\" not found" "related_resource_kind"="Order" "related_resource_name"="httpbin-csh7g-583384746" "related_resource_namespace"="production" "resource_kind"="Challenge" "resource_name"="httpbin-csh7g-583384746-1109480142" "resource_namespace"="production" "resource_version"="v1" E1014 12:18:08.767540 1 util.go:80] cert-manager/controller/orders/handleOwnedResource "msg"="error getting referenced owning resource" "error"="order.acme.cert-manager.io \"httpbin-csh7g-583384746\" not found" "related_resource_kind"="Order" "related_resource_name"="httpbin-csh7g-583384746" "related_resource_namespace"="production" "resource_kind"="Challenge" "resource_name"="httpbin-csh7g-583384746-1109480142" "resource_namespace"="production" "resource_version"="v1" E1014 12:18:08.768861 1 controller.go:211] cert-manager/controller/challenges "msg"="challenge in work queue no longer exists" "error"="challenge.acme.cert-manager.io \"httpbin-csh7g-583384746-1109480142\" not found" Would you please so kind tell me where is the real documentation and does this software work?! Software versions: kubernetes: v1.22.2 cert-manager: v1.5.4 As @wallrj mentions, this sounds like a DNS issue in your cluster (based on dial tcp 172.24.0.10:53: i/o timeout). Would you please so kind tell me where is the real documentation and does this software work?! Documentation can be found at https://cert-manager.io/ Please also take note of our code of conduct: https://github.com/jetstack/cert-manager/blob/master/CODE_OF_CONDUCT.md#our-standards specifically Showing empathy towards other community members
gharchive/issue
2021-10-14T12:27:07
2025-04-01T04:34:40.684974
{ "authors": [ "amkiryushkin", "munnerz" ], "repo": "jetstack/cert-manager", "url": "https://github.com/jetstack/cert-manager/issues/4521", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
498187171
make the webhook port configurable in GKE private clusters, by default masters are allowed to talk to the cluster nodes only on 443 and 10250. so configuring secure_port: 443, will work out of the box without needing to add fw rules What this PR does / why we need it: Which issue this PR fixes (optional, in fixes #<issue number>(, fixes #<issue_number>, ...) format, will close that issue when PR gets merged): fixes # Special notes for your reviewer: Release note: https://github.com/jetstack/cert-manager/issues/2108 /ok-to-test @nrobert13 are you able to follow up and address the above? We intend to cut the v0.11 release on Wednesday, so this PR is likely to miss the deadline for v0.11 otherwise 😀 Is this closed now because you could not agree on the portname?
gharchive/pull-request
2019-09-25T10:12:26
2025-04-01T04:34:40.689370
{ "authors": [ "monotek", "munnerz", "nrobert13" ], "repo": "jetstack/cert-manager", "url": "https://github.com/jetstack/cert-manager/pull/2107", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
505862160
Remove deprecated arguments from helm chart deployment What this PR does / why we need it: The controller arguments --default-acme-issuer-challenge-type and --default-acme-issuer-dns01-provider-name are not available in version 0.11. This causes the deployment pod to fail in a CrashLoopBackOff when a value was defined. This PR removes the ability to define these deprecated arguments. Which issue this PR fixes (optional, in fixes #<issue number>(, fixes #<issue_number>, ...) format, will close that issue when PR gets merged): fixes # Release note: /assign @kragniz Thanks for this! 🙈 /lgtm /approve /cherrypick release-0.11 /ok-to-test
gharchive/pull-request
2019-10-11T13:44:40
2025-04-01T04:34:40.692572
{ "authors": [ "Lucaber", "munnerz" ], "repo": "jetstack/cert-manager", "url": "https://github.com/jetstack/cert-manager/pull/2201", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
266171603
WIP: Cassandra Controller Fixes: https://gitlab.jetstack.net/marshal/navigator/issues/26 I'm splitting this into smaller branches.
gharchive/pull-request
2017-10-17T15:28:38
2025-04-01T04:34:40.693697
{ "authors": [ "wallrj" ], "repo": "jetstack/navigator", "url": "https://github.com/jetstack/navigator/pull/46", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
268103959
Test commit - do not merge What this PR does / why we need it: Which issue this PR fixes (optional, in fixes #<issue number>(, fixes #<issue_number>, ...) format, will close that issue when PR gets merged): fixes # Special notes for your reviewer: Release note: NONE /assign /test /hold /retest /retest /retest /retest /test
gharchive/pull-request
2017-10-24T16:39:55
2025-04-01T04:34:40.697265
{ "authors": [ "munnerz" ], "repo": "jetstack/navigator", "url": "https://github.com/jetstack/navigator/pull/54", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
71291274
android:windowTranslucentStatus is not working with the Slidingmenu I have added the following lines in styles.xml @android:color/transparent true But the toolbar in the layout is still staying below the status bar and the color of the status bar seen gray. I want to show my layout and the sliding menu full screened including the status bar with a transparent color. What change should i make to enable this feature. +1 mSlidingMenu.attachToActivity(activity, SlidingMenu.SLIDING_CONTENT, true); Set slideStyle to SlidingMenu.SLIDING_CONTENT and pass true as third parameter (boolean actionbarOverlay) works for me. @pguilbot how to fix ? thanks, I have salved it. This is a solution https://github.com/dalong982242260/FangTantan
gharchive/issue
2015-04-27T13:55:01
2025-04-01T04:34:40.763454
{ "authors": [ "Livinlawrence", "NoahZu", "chih-hao-li", "jjhesk" ], "repo": "jfeinstein10/SlidingMenu", "url": "https://github.com/jfeinstein10/SlidingMenu/issues/716", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2354308516
Elixir Resources Added Pull Request Description Added: Communities Documentation Youtube Playlists Youtube Channels Books Elixir resources like maps, structs, Blank tables for other categories to match the format closes #1171 Closes #1171 Type of change [x] Adding a new Resource [ ] Adding a new Category [x] Documentation enhancement or fixes [ ] Website enhancement or fixes Checklist: [x] I read carefully CONTRIBUTING.md [x] I have performed a self-review of my own code [x] I have commented my code, particularly in hard-to-understand areas [ ] I have made corresponding changes to the documentation Screenshots (if appropriate): Please add screenshots to help explain your changes. Additional It has been mentioned before any resource adding is limited to 150 points, so I will see any activities other technical works above limits will be reduced. @sau-mili @sanjay-kv you can check, I have stopped doing the documentation edits in this repo. I have only contributed to the website after the 150 points thing was informed by both GSSOC and The PA. Thanks
gharchive/pull-request
2024-06-15T00:35:11
2025-04-01T04:34:40.779496
{ "authors": [ "sanjay-kv", "sau-mili" ], "repo": "jfmartinz/ResourceHub", "url": "https://github.com/jfmartinz/ResourceHub/pull/1173", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
152023501
apparmor: fix template syntax error ERROR: Syntax Error: Invalid Regex @{PROC}/{,**^[0-9],sys/kernel/shm*} Did you take this from the docker one On Saturday, April 30, 2016, djtm notifications@github.com wrote: ERROR: Syntax Error: Invalid Regex @{PROC}/{,**^[0-9],sys/kernel/shm*} You can view, comment on, or merge this pull request online at: https://github.com/jfrazelle/bane/pull/9 Commit Summary apparmor: fix template syntax error File Changes M apparmor/template.go https://github.com/jfrazelle/bane/pull/9/files#diff-0 (3) Patch Links: https://github.com/jfrazelle/bane/pull/9.patch https://github.com/jfrazelle/bane/pull/9.diff — You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub https://github.com/jfrazelle/bane/pull/9 -- Jessie Frazelle 4096R / D4C4 DD60 0D66 F65A 8EFC 511E 18F3 685C 0022 BFF3 pgp.mit.edu http://pgp.mit.edu/pks/lookup?op=get&search=0x18F3685C0022BFF3 Yes Missed two lines.
gharchive/pull-request
2016-04-30T10:42:45
2025-04-01T04:34:40.785903
{ "authors": [ "djtm", "jfrazelle" ], "repo": "jfrazelle/bane", "url": "https://github.com/jfrazelle/bane/pull/9", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
158943877
Updated atom to v1.8.0 See https://github.com/atom/atom/releases/tag/v1.8.0 thanks! just waiting on CI thanks!
gharchive/pull-request
2016-06-07T14:53:45
2025-04-01T04:34:40.787290
{ "authors": [ "jfrazelle", "michael-k" ], "repo": "jfrazelle/dockerfiles", "url": "https://github.com/jfrazelle/dockerfiles/pull/145", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
101615359
SSL Parameter should be configurable outside of vhosts It may be common practice, but there is no point in configuring the ssl parameters in every vhosts. There should be a global configuration that sets secure defaults and only if a vhosts really needs a custom config, should the vhost parameters be used. This would make all vhosts more readable, mass updates more stable and analysis of the configuration by admins much easier. This seems even more important since the default value is a very verbose way of naming the currently mostly used ciphers. What does @jfryman think about this? If you can confirm that this would be helpful, i could work on a PR to implement this. But if this is not desired I will stick with a fork. Another alternative would to be to handle an undef or empty value of ciphers and handling that case by not adding a configuration line. I think its good to have a set of defaults but i would also like to be able to override this on the vhost @b4ldr That should not be an issue. For our production deployments we currently include a modules.d/*.conf, (where our ssl.conf is located ) within the http{ } section. Below that modules.d include we simply do an include of vhosts.d ( or sites-enabled/available, depending on the os ). Thus any setting can be overwritten by the vhosts. Thus using a wildcard include of conf.d or modules.d or simply placing a global ciphersuite within the http section before including virtual hosts one should have the best of both worlds. @omcnet I also think it makes a lot of sense to have SSL defaults on the http level. But as @b4ldr said any PR must still allow for SSL directives to be overridden at the server level too. Please have directives added directly to the nginx.conf file, don't put them into a separate file to include, so the generated configuration is easier to understand. I think this is sensible as well, but as best I know, a PR hasn't come up for this, and it's 1 year later. Other than keeping things shorter and more readable, I'd argue that it's more or less a cosmetic thing, since the current solution should give the same results. But I'm going to leave this open, since I would love to see this get implemented if it's not too difficult. We'd probably need to move some of the default parameters out of the 'vhost' resource and into params or config?
gharchive/issue
2015-08-18T09:42:25
2025-04-01T04:34:40.823535
{ "authors": [ "3flex", "b4ldr", "omcnet", "wyardley" ], "repo": "jfryman/puppet-nginx", "url": "https://github.com/jfryman/puppet-nginx/issues/670", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
219398316
Not getting any notifications for network requests Not able to get any notifications using the logging interceptor for Chuck. I have Okhttp3.3.0 and Android 6.0. Not able to get any notifications to view network traffic. figured out what was wrong. Had to include compile 'com.squareup.retrofit2:retrofit:2.2.0' compile "com.squareup.retrofit2:converter-gson:2.2.0" to my gradle as I have been using Jackson.
gharchive/issue
2017-04-04T22:03:57
2025-04-01T04:34:40.832262
{ "authors": [ "phundal209" ], "repo": "jgilfelt/chuck", "url": "https://github.com/jgilfelt/chuck/issues/41", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
774695839
Persistent zoom levels I have a large diagram over 50 pages. When I am working on it I like to see each page zoomed at 75%. Whenever I close and reopen the diagram I see that all 50+ pages have defaulted back to 100%, so before I begin work I need to resize each page to 75%, which is annoying and a waste of time. Is there a way to set the page default zoom to 75% for the entire diagram and have is save that way? Asked the same in https://github.com/jgraph/drawio-desktop/issues/723 We cannot find a good solution to this. Saving the zoom level per page in the diagram falls apart when users are collaboratively editing.
gharchive/issue
2020-12-25T09:46:08
2025-04-01T04:34:40.855844
{ "authors": [ "davidjgraph", "mararad" ], "repo": "jgraph/drawio", "url": "https://github.com/jgraph/drawio/issues/1661", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1477585019
New DrawIO app on Mac doesn't support multiple tabs. Preflight Checklist [x] I agree to follow the Code of Conduct that this project adheres to. [x] I have searched the issue tracker for a feature request that matches the one I want to file, without success. Describe the bug After I upgraded to the latest DrawIO app on Mac (downloaded from https://www.diagrams.net/), lots of functions have lost. i.e. can't create multiple tab, just few menu options are available. To Reproduce Steps to reproduce the behavior: Create a new pape You can't find a way to create a new tab like the web version. Expected behavior below is a tab function I want that is only available from the web version, but not the app. Screenshots If applicable, add screenshots to help explain your problem. draw.io version (In the Help->About menu of the draw.io editor): draw.io version [v20.6.0] Desktop (please complete the following information): OS: [Mac Ventura 13.0.1] Browser [ Chrome] Version [107] Exactly where did you download 20.6.0, what was the URL?
gharchive/issue
2022-12-05T22:27:45
2025-04-01T04:34:40.862058
{ "authors": [ "davidjgraph", "starzhanganz" ], "repo": "jgraph/drawio", "url": "https://github.com/jgraph/drawio/issues/3227", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1140181076
Update workflow for rendering for use with _Quizzes If the Quizzes are stored in a separate repository, the workflow as it is doesn't exactly work. Need to reconfigure so that you can tell it to only to render and not look for quizzes. See this example: https://github.com/jhudsl/Documentation_and_Usability/actions/runs/1853495226
gharchive/issue
2022-02-16T15:26:46
2025-04-01T04:34:41.046555
{ "authors": [ "cansavvy" ], "repo": "jhudsl/OTTR_Template", "url": "https://github.com/jhudsl/OTTR_Template/issues/478", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
415577765
Not preserving content outside head and body elements I'm using Jsoup parser. With prettyPrint option disabled, I'm loosing whitespaces between DOCTYPE definition and html tag. Also, I'm loosing whitespaces between html and head start tag. I think the main problem is that Jsoup is not keeping whitespaces between html childs, and between document and html element. I'm using HTML parser. Using XML parser the output is correct, so maybe HTML parser should do the same. If not, what are the reasons? Here is a test link: https://try.jsoup.org/~lPGQPj9gNjjfU1yN-pYRvQVfzMY I'm using HTML parser. Using XML parser the output is correct, so maybe HTML parser should do the same. Thanks, this was implemented in 9675a92b41593244d6d3ecd99425aeb19e2a6219. Hi, Thank you for implementing this! Now, I realised that whitespace between closing body tag and closing html tag has lost. Check the example: http://try.jsoup.org/~fETsCv8YuhVF_1ofxxcX9lUvngk
gharchive/issue
2019-02-28T11:30:31
2025-04-01T04:34:41.086320
{ "authors": [ "Vegeeto", "jhy" ], "repo": "jhy/jsoup", "url": "https://github.com/jhy/jsoup/issues/1185", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1441876746
Batch effect issue for multiple 10X visium samples Hello I am having batch effect issue for running multiple sample clustering in spaGCN as indicated in previous post https://github.com/jianhuupenn/SpaGCN/issues/37 Would you please help me if you can explain steps to perform z-score normalization for integrated dataset. I would appreciate all the suggestion. Thanks in advance For integrated samples , I used following steps : QC Filteration Integrated four samples with scanpy as suggested in https://nbisweden.github.io/workshop-scRNAseq/labs/compiled/scanpy/scanpy_07_spatial.html using scanorama method. Is it ok if I run clf=spg.multiSpaGCN on the integrated adata object ?
gharchive/issue
2022-11-09T10:57:37
2025-04-01T04:34:41.093036
{ "authors": [ "archanabhardwaj" ], "repo": "jianhuupenn/SpaGCN", "url": "https://github.com/jianhuupenn/SpaGCN/issues/54", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
166028190
请考虑对options的处理切入点。 如果在context.create({}),这里的选项,没有办法进行后继的一些处理,比如基于某些原因,项目中途使用此组件或者基于标准写法定义选项,后续是没有办法进行转换的。 dataContext.create({ getInitData: { method: "GET", url: "invoice/getInitData", cache: "default" }) 比如 cache: "default" 是按fetch标准写法,要转换成你的storage,是无法实现的。 我加了一个事件,可以处理了,还有根据method == "POST"就可以全局作一些处理,这都是很常见的作法, context.on("processOptions", function (config) { if (config.method == "POST") { config.header = Object.assign({}, config.header, { "Content-Type": "text/json" }); } if (config.cache === "default") { config.storage = { type: 'localStorage', key: config.url, tag: 'v1.0' }; } }); 还有一个问题,目前导致我不能继续使用这个组件了, 在sendAjax 中POST方法下处理主体数据,你在内部序列化时进行了 params 化,但是并不是所有数据都是这种处理方法,外部无法进行处理。比如,我项目的提交数据格式必须是json。像zepto,jquery都有这方法的机制,用processData标记是否需要内部进行处理。 第一个问题: 请求在发送之前会调用willFetch回调,接收两个参数vars和config,vars是每次请求的动态数据,config是内部process后的options的结果。可以通过在这个回调里修改config。 第二个问题: 你指出的应该是参数命名问题,你建议使用cache,是这样理解吗? 第三个问题: 好问题!希望传JSON字符串,你希望具体什么写?我看看能不能提供支持。 碰到同样的问题,什么时候可以提供直接传递json的功能?不然只能回到zepto了
gharchive/issue
2016-07-18T06:37:35
2025-04-01T04:34:41.098487
{ "authors": [ "Jias", "sparkleDai", "xuguogang" ], "repo": "jias/natty-fetch", "url": "https://github.com/jias/natty-fetch/issues/20", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1375613078
https://18comic.vip/album/255011/?shunt=1 特定漫画 255011卡在3977处理页面不动了 https://18comic.vip/album/255011/?shunt=1 特定漫画 255011卡在3977处理页面不动了 下载第一话就直接出错了,不知道是不是问题出在这里
gharchive/issue
2022-09-16T08:28:45
2025-04-01T04:34:41.100355
{ "authors": [ "waitfortea" ], "repo": "jiayaoO3O/18-comic-finder", "url": "https://github.com/jiayaoO3O/18-comic-finder/issues/69", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1791006258
PIT support for Opensearch 2.7/2.8 Hi gilles, This issue is to exchange thoughts on how to support the above. I would like to check a few things with you before I create a PR. In the release-notes for 2.0.4 you write: _pit API seems no longer supported in the latest Opensearch 2.7 release. I've restricted searchAfter to not allow OS1/OS2 anymore. If somebody needs this, please create a PR to add support for the new APIs in Opensearch. Or you can use scrolling searches This surprised me because it appears in the docs for both versions 2.7 and 2.8 I also checked the docs of the Elastic counterpart, and I see no differences between the two. What am I missing? Because from the looks of it, it seems to me it would 'just' work? Full disclosure: In the end, what I will need is PIT-based slicing support :) A few things I've learned: the restful-url paths for opensearch are different, creating a pit is at /my-index-pattern/_search/point_in_time as opposed to elastic's /my-index-pattern/_pit. object OpensearchPit { suspend fun createPointInTime(index: String, keepAlive: Duration, searchClient: SearchClient): String { val result = searchClient.restClient.post { path(index, "_search", "point_in_time") parameter("keep_alive", "${keepAlive.inWholeSeconds}s") } val id = searchClient.json.parseToJsonElement(result.getOrThrow().text).jsonObject["pit_id"]?.jsonPrimitive?.content require(id != null, lazyMessage = { "opensearch API did not contain pit id. Payload was ${result.getOrNull()}" }) return id } suspend fun deletePointInTime(id: String, searchClient: SearchClient): JsonObject { val body = mapOf("pit_id" to listOf(id)) return searchClient.restClient.delete { path("_search", "point_in_time") rawBody(searchClient.json.encodeToString(body)) }.parse(JsonObject.serializer()) } } see https://github.com/jillesvangurp/kt-search/issues/77
gharchive/issue
2023-07-06T07:46:50
2025-04-01T04:34:41.108940
{ "authors": [ "hanswesterbeek" ], "repo": "jillesvangurp/kt-search", "url": "https://github.com/jillesvangurp/kt-search/issues/77", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
575126872
Exception in TouchBarsManager onUpdateEditorHeader in 2020.1 EAP Version and Environment Details Operation system: MacOS IDE name and version: WebStorm 2020.1 EAP, but works in WebStorm 2019.x stable. Plugin version: 2.3.0 (current latest) Expected Behaviour Plugin does not throw. Actual Behaviour java.lang.Throwable: can't find project data to update header of editor: com.intellij.openapi.editor.impl.EditorImpl@110d4461, project: Project (name=flash-app, containerState=ACTIVE, componentStore=/Users/vasyl/code/FlashFX/flash-app) at com.intellij.openapi.diagnostic.Logger.error(Logger.java:146) at com.intellij.ui.mac.touchbar.TouchBarsManager.onUpdateEditorHeader(TouchBarsManager.java:372) at com.intellij.openapi.editor.impl.EditorImpl.setHeaderComponent(EditorImpl.java:1954) at com.intellij.lang.jsgraphql.v1.ide.project.JSGraphQLLanguageUIProjectService.createToolWindowResultEditor(JSGraphQLLanguageUIProjectService.java:591) at com.intellij.lang.jsgraphql.v1.ide.project.JSGraphQLLanguageUIProjectService.lambda$null$9(JSGraphQLLanguageUIProjectService.java:610) at com.intellij.openapi.application.TransactionGuardImpl$2.run(TransactionGuardImpl.java:205) at com.intellij.openapi.application.impl.ApplicationImpl.runIntendedWriteActionOnCurrentThread(ApplicationImpl.java:831) at com.intellij.openapi.application.impl.ApplicationImpl.lambda$invokeLater$4(ApplicationImpl.java:310) at com.intellij.openapi.application.impl.FlushQueue.doRun(FlushQueue.java:80) at com.intellij.openapi.application.impl.FlushQueue.runNextEvent(FlushQueue.java:128) at com.intellij.openapi.application.impl.FlushQueue.flushNow(FlushQueue.java:46) at com.intellij.openapi.application.impl.FlushQueue$FlushNow.run(FlushQueue.java:184) at java.desktop/java.awt.event.InvocationEvent.dispatch(InvocationEvent.java:313) at java.desktop/java.awt.EventQueue.dispatchEventImpl(EventQueue.java:776) at java.desktop/java.awt.EventQueue$4.run(EventQueue.java:727) at java.desktop/java.awt.EventQueue$4.run(EventQueue.java:721) at java.base/java.security.AccessController.doPrivileged(Native Method) at java.base/java.security.ProtectionDomain$JavaSecurityAccessImpl.doIntersectionPrivilege(ProtectionDomain.java:85) at java.desktop/java.awt.EventQueue.dispatchEvent(EventQueue.java:746) at com.intellij.ide.IdeEventQueue.defaultDispatchEvent(IdeEventQueue.java:974) at com.intellij.ide.IdeEventQueue._dispatchEvent(IdeEventQueue.java:847) at com.intellij.ide.IdeEventQueue.lambda$null$8(IdeEventQueue.java:449) at com.intellij.openapi.progress.impl.CoreProgressManager.computePrioritized(CoreProgressManager.java:728) at com.intellij.ide.IdeEventQueue.lambda$dispatchEvent$9(IdeEventQueue.java:448) at com.intellij.openapi.application.impl.ApplicationImpl.runIntendedWriteActionOnCurrentThread(ApplicationImpl.java:831) at com.intellij.ide.IdeEventQueue.dispatchEvent(IdeEventQueue.java:496) at com.intellij.cloudConfig.CloudConfigManager.waitDone(CloudConfigManager.java:1890) at com.intellij.cloudConfig.CloudConfigManager.getRepositoryPlugins(CloudConfigManager.java:1874) at com.intellij.cloudConfig.CloudConfigManager.updatePlugins(CloudConfigManager.java:1949) at com.intellij.cloudConfig.CloudConfigManager.safeUpdatePlugins(CloudConfigManager.java:1899) at com.intellij.cloudConfig.CloudConfigManager.lambda$doConnection$22(CloudConfigManager.java:1043) at com.intellij.openapi.application.TransactionGuardImpl$2.run(TransactionGuardImpl.java:205) at com.intellij.openapi.application.impl.ApplicationImpl.runIntendedWriteActionOnCurrentThread(ApplicationImpl.java:831) at com.intellij.openapi.application.impl.ApplicationImpl.lambda$invokeLater$4(ApplicationImpl.java:310) at com.intellij.openapi.application.impl.FlushQueue.doRun(FlushQueue.java:80) at com.intellij.openapi.application.impl.FlushQueue.runNextEvent(FlushQueue.java:128) at com.intellij.openapi.application.impl.FlushQueue.flushNow(FlushQueue.java:46) at com.intellij.openapi.application.impl.FlushQueue$FlushNow.run(FlushQueue.java:184) at java.desktop/java.awt.event.InvocationEvent.dispatch(InvocationEvent.java:313) at java.desktop/java.awt.EventQueue.dispatchEventImpl(EventQueue.java:776) at java.desktop/java.awt.EventQueue$4.run(EventQueue.java:727) at java.desktop/java.awt.EventQueue$4.run(EventQueue.java:721) at java.base/java.security.AccessController.doPrivileged(Native Method) at java.base/java.security.ProtectionDomain$JavaSecurityAccessImpl.doIntersectionPrivilege(ProtectionDomain.java:85) at java.desktop/java.awt.EventQueue.dispatchEvent(EventQueue.java:746) at com.intellij.ide.IdeEventQueue.defaultDispatchEvent(IdeEventQueue.java:974) at com.intellij.ide.IdeEventQueue._dispatchEvent(IdeEventQueue.java:847) at com.intellij.ide.IdeEventQueue.lambda$null$8(IdeEventQueue.java:449) at com.intellij.openapi.progress.impl.CoreProgressManager.computePrioritized(CoreProgressManager.java:739) at com.intellij.ide.IdeEventQueue.lambda$dispatchEvent$9(IdeEventQueue.java:448) at com.intellij.openapi.application.impl.ApplicationImpl.runIntendedWriteActionOnCurrentThread(ApplicationImpl.java:831) at com.intellij.ide.IdeEventQueue.dispatchEvent(IdeEventQueue.java:496) at java.desktop/java.awt.EventDispatchThread.pumpOneEventForFilters(EventDispatchThread.java:203) at java.desktop/java.awt.EventDispatchThread.pumpEventsForFilter(EventDispatchThread.java:124) at java.desktop/java.awt.EventDispatchThread.pumpEventsForHierarchy(EventDispatchThread.java:113) at java.desktop/java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:109) at java.desktop/java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:101) at java.desktop/java.awt.EventDispatchThread.run(EventDispatchThread.java:90) Steps to Reproduce / Link to Repo with Reproduction and Instructions Open a project with a .graphql file. Hi Vasyl, Thanks for reporting this. It appears to be caused by https://youtrack.jetbrains.com/issue/IDEA-231011 Is not reproducible since plugin version 3.0.0 and IDEA 2021.2.
gharchive/issue
2020-03-04T04:20:09
2025-04-01T04:34:41.141297
{ "authors": [ "jimkyndemeyer", "koresar", "vepanimas" ], "repo": "jimkyndemeyer/js-graphql-intellij-plugin", "url": "https://github.com/jimkyndemeyer/js-graphql-intellij-plugin/issues/319", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
972365779
feat: read from psql WIP Implemented in PSQL
gharchive/pull-request
2021-08-17T07:01:09
2025-04-01T04:34:41.153283
{ "authors": [ "cristianmtr" ], "repo": "jina-ai/jina-commons", "url": "https://github.com/jina-ai/jina-commons/pull/21", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
202317743
Foerign key by annotations vs Foerign key by "AddForeignKey" Hello everyone! What's the difference between use annotations and use the method "AddForeignKey"? By annotation example: type User struct { gorm.Model Profile Profile `gorm:"ForeignKey:ProfileID"` ProfileID int } type Profile struct { gorm.Model Name string } By method example: db.Model(&User{}).AddForeignKey("profile_id", "profiles(id)", "RESTRICT", "RESTRICT") Tags affect query construction, processing of associated models and auto-migrations, while method does only what it states.
gharchive/issue
2017-01-21T14:14:25
2025-04-01T04:34:41.168635
{ "authors": [ "rennanfelipe", "zardak" ], "repo": "jinzhu/gorm", "url": "https://github.com/jinzhu/gorm/issues/1322", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
64059522
FoundationDB dialect integration This is now ready for review. For "test_all.sh", at first I wasn't sure about adding foundation, but ultimately I figured it was best to include it. One interesting lesson I learned while implementing this dialect is that the FoundationDB DATETIME type does not support timezones. The unit-tests are passing for me, he is the output: jay@jays-mba:~/go/src/github.com/jinzhu/gorm$ ./test_all.sh testing postgres... === RUN TestRegisterCallback --- PASS: TestRegisterCallback (0.00s) === RUN TestRegisterCallbackWithOrder --- PASS: TestRegisterCallbackWithOrder (0.00s) === RUN TestRegisterCallbackWithComplexOrder --- PASS: TestRegisterCallbackWithComplexOrder (0.00s) === RUN TestReplaceCallback [info] replacing callback `create` from /Users/jay/go/src/github.com/jinzhu/gorm/callback_test.go:94 --- PASS: TestReplaceCallback (0.00s) === RUN TestRemoveCallback [info] removing callback `create` from /Users/jay/go/src/github.com/jinzhu/gorm/callback_test.go:107 --- PASS: TestRemoveCallback (0.00s) === RUN TestCloneSearch --- PASS: TestCloneSearch (0.00s) === RUN TestHasOneAndHasManyAssociation --- PASS: TestHasOneAndHasManyAssociation (0.11s) === RUN TestRelated --- PASS: TestRelated (0.02s) === RUN TestManyToMany --- PASS: TestManyToMany (0.07s) === RUN TestForeignKey --- PASS: TestForeignKey (0.00s) === RUN TestRunCallbacks --- PASS: TestRunCallbacks (0.01s) === RUN TestCallbacksWithErrors --- PASS: TestCallbacksWithErrors (0.02s) === RUN TestCreate --- PASS: TestCreate (0.00s) === RUN TestCreateWithNoGORMPrimayKey --- PASS: TestCreateWithNoGORMPrimayKey (0.00s) === RUN TestCreateWithNoStdPrimaryKeyAndDefaultValues --- PASS: TestCreateWithNoStdPrimaryKeyAndDefaultValues (0.00s) === RUN TestAnonymousScanner --- PASS: TestAnonymousScanner (0.00s) === RUN TestAnonymousField --- PASS: TestAnonymousField (0.00s) === RUN TestSelectWithCreate --- PASS: TestSelectWithCreate (0.01s) === RUN TestOmitWithCreate --- PASS: TestOmitWithCreate (0.01s) === RUN TestCustomizeColumn --- PASS: TestCustomizeColumn (0.01s) === RUN TestCustomColumnAndIgnoredFieldClash --- PASS: TestCustomColumnAndIgnoredFieldClash (0.01s) === RUN TestDelete --- PASS: TestDelete (0.01s) === RUN TestInlineDelete --- PASS: TestInlineDelete (0.01s) === RUN TestSoftDelete --- PASS: TestSoftDelete (0.02s) === RUN TestSaveAndQueryEmbeddedStruct --- PASS: TestSaveAndQueryEmbeddedStruct (0.01s) === RUN TestJoinTable --- PASS: TestJoinTable (0.06s) === RUN TestExceptionsWithInvalidSql --- PASS: TestExceptionsWithInvalidSql (0.00s) === RUN TestSetTable --- PASS: TestSetTable (0.06s) === RUN TestHasTable --- PASS: TestHasTable (0.01s) === RUN TestTableName --- PASS: TestTableName (0.00s) === RUN TestSqlNullValue --- PASS: TestSqlNullValue (0.04s) === RUN TestTransaction --- PASS: TestTransaction (0.01s) === RUN TestRow --- PASS: TestRow (0.00s) === RUN TestRows --- PASS: TestRows (0.00s) === RUN TestScan --- PASS: TestScan (0.01s) === RUN TestRaw --- PASS: TestRaw (0.01s) === RUN TestGroup --- PASS: TestGroup (0.00s) === RUN TestJoins --- PASS: TestJoins (0.00s) === RUN TestHaving --- PASS: TestHaving (0.00s) === RUN TestTimeWithZone --- PASS: TestTimeWithZone (0.01s) === RUN TestHstore --- PASS: TestHstore (0.01s) === RUN TestSetAndGet --- PASS: TestSetAndGet (0.00s) === RUN TestCompatibilityMode `testdb` is not officially supported, running under compatibility mode. --- PASS: TestCompatibilityMode (0.00s) === RUN TestOpenExistingDB --- PASS: TestOpenExistingDB (0.01s) === RUN TestIndexes --- PASS: TestIndexes (0.09s) === RUN TestAutoMigration --- PASS: TestAutoMigration (0.05s) === RUN TestPointerFields --- PASS: TestPointerFields (0.01s) === RUN TestPolymorphic --- PASS: TestPolymorphic (0.04s) === RUN TestPreload --- PASS: TestPreload (0.05s) === RUN TestFirstAndLast --- PASS: TestFirstAndLast (0.01s) === RUN TestFirstAndLastWithNoStdPrimaryKey --- PASS: TestFirstAndLastWithNoStdPrimaryKey (0.00s) === RUN TestUIntPrimaryKey --- PASS: TestUIntPrimaryKey (0.00s) === RUN TestFindAsSliceOfPointers --- PASS: TestFindAsSliceOfPointers (0.01s) === RUN TestSearchWithPlainSQL --- PASS: TestSearchWithPlainSQL (0.01s) === RUN TestSearchWithStruct --- PASS: TestSearchWithStruct (0.02s) === RUN TestSearchWithMap --- PASS: TestSearchWithMap (0.01s) === RUN TestSearchWithEmptyChain --- PASS: TestSearchWithEmptyChain (0.01s) === RUN TestSelect --- PASS: TestSelect (0.00s) === RUN TestOrderAndPluck --- PASS: TestOrderAndPluck (0.01s) === RUN TestLimit --- PASS: TestLimit (0.01s) === RUN TestOffset --- PASS: TestOffset (0.05s) === RUN TestOr --- PASS: TestOr (0.01s) === RUN TestCount --- PASS: TestCount (0.01s) === RUN TestNot --- PASS: TestNot (0.06s) === RUN TestFillSmallerStruct --- PASS: TestFillSmallerStruct (0.01s) === RUN TestFindOrInitialize --- PASS: TestFindOrInitialize (0.01s) === RUN TestFindOrCreate --- PASS: TestFindOrCreate (0.02s) === RUN TestSelectWithEscapedFieldName --- PASS: TestSelectWithEscapedFieldName (0.01s) === RUN TestSelectWithVariables --- PASS: TestSelectWithVariables (0.00s) === RUN TestSelectWithArrayInput --- PASS: TestSelectWithArrayInput (0.01s) === RUN TestScopes --- PASS: TestScopes (0.01s) === RUN TestScannableSlices --- PASS: TestScannableSlices (0.02s) === RUN TestUpdate --- PASS: TestUpdate (0.02s) === RUN TestUpdateWithNoStdPrimaryKeyAndDefaultValues --- PASS: TestUpdateWithNoStdPrimaryKeyAndDefaultValues (0.01s) === RUN TestUpdates --- PASS: TestUpdates (0.02s) === RUN TestUpdateColumn --- PASS: TestUpdateColumn (0.01s) === RUN TestSelectWithUpdate --- PASS: TestSelectWithUpdate (0.02s) === RUN TestSelectWithUpdateWithMap --- PASS: TestSelectWithUpdateWithMap (0.02s) === RUN TestOmitWithUpdate --- PASS: TestOmitWithUpdate (0.02s) === RUN TestOmitWithUpdateWithMap --- PASS: TestOmitWithUpdateWithMap (0.02s) === RUN TestSelectWithUpdateColumn --- PASS: TestSelectWithUpdateColumn (0.01s) === RUN TestOmitWithUpdateColumn --- PASS: TestOmitWithUpdateColumn (0.01s) PASS ok github.com/jinzhu/gorm 1.550s testing foundation... === RUN TestRegisterCallback --- PASS: TestRegisterCallback (0.00s) === RUN TestRegisterCallbackWithOrder --- PASS: TestRegisterCallbackWithOrder (0.00s) === RUN TestRegisterCallbackWithComplexOrder --- PASS: TestRegisterCallbackWithComplexOrder (0.00s) === RUN TestReplaceCallback [info] replacing callback `create` from /Users/jay/go/src/github.com/jinzhu/gorm/callback_test.go:94 --- PASS: TestReplaceCallback (0.00s) === RUN TestRemoveCallback [info] removing callback `create` from /Users/jay/go/src/github.com/jinzhu/gorm/callback_test.go:107 --- PASS: TestRemoveCallback (0.00s) === RUN TestCloneSearch --- PASS: TestCloneSearch (0.00s) === RUN TestHasOneAndHasManyAssociation --- PASS: TestHasOneAndHasManyAssociation (0.41s) === RUN TestRelated --- PASS: TestRelated (0.08s) === RUN TestManyToMany --- PASS: TestManyToMany (0.35s) === RUN TestForeignKey --- PASS: TestForeignKey (0.00s) === RUN TestRunCallbacks --- PASS: TestRunCallbacks (0.06s) === RUN TestCallbacksWithErrors --- PASS: TestCallbacksWithErrors (0.10s) === RUN TestCreate --- PASS: TestCreate (0.04s) === RUN TestCreateWithNoGORMPrimayKey --- PASS: TestCreateWithNoGORMPrimayKey (0.01s) === RUN TestCreateWithNoStdPrimaryKeyAndDefaultValues --- PASS: TestCreateWithNoStdPrimaryKeyAndDefaultValues (0.04s) === RUN TestAnonymousScanner --- PASS: TestAnonymousScanner (0.02s) === RUN TestAnonymousField --- PASS: TestAnonymousField (0.03s) === RUN TestSelectWithCreate --- PASS: TestSelectWithCreate (0.06s) === RUN TestOmitWithCreate --- PASS: TestOmitWithCreate (0.05s) === RUN TestCustomizeColumn --- PASS: TestCustomizeColumn (0.17s) === RUN TestCustomColumnAndIgnoredFieldClash --- PASS: TestCustomColumnAndIgnoredFieldClash (0.11s) === RUN TestDelete --- PASS: TestDelete (0.11s) === RUN TestInlineDelete --- PASS: TestInlineDelete (0.08s) === RUN TestSoftDelete --- PASS: TestSoftDelete (0.33s) === RUN TestSaveAndQueryEmbeddedStruct --- PASS: TestSaveAndQueryEmbeddedStruct (0.04s) === RUN TestJoinTable --- PASS: TestJoinTable (0.59s) === RUN TestExceptionsWithInvalidSql --- PASS: TestExceptionsWithInvalidSql (0.02s) === RUN TestSetTable --- PASS: TestSetTable (0.38s) === RUN TestHasTable --- PASS: TestHasTable (0.12s) === RUN TestTableName --- PASS: TestTableName (0.00s) === RUN TestSqlNullValue --- PASS: TestSqlNullValue (0.15s) === RUN TestTransaction --- PASS: TestTransaction (0.03s) === RUN TestRow --- PASS: TestRow (0.05s) === RUN TestRows --- PASS: TestRows (0.05s) === RUN TestScan --- PASS: TestScan (0.06s) === RUN TestRaw --- PASS: TestRaw (0.08s) === RUN TestGroup --- PASS: TestGroup (0.00s) === RUN TestJoins --- PASS: TestJoins (0.02s) === RUN TestHaving --- PASS: TestHaving (0.01s) === RUN TestTimeWithZone --- PASS: TestTimeWithZone (0.05s) === RUN TestHstore --- SKIP: TestHstore (0.00s) main_test.go:506: === RUN TestSetAndGet --- PASS: TestSetAndGet (0.00s) === RUN TestCompatibilityMode `testdb` is not officially supported, running under compatibility mode. --- PASS: TestCompatibilityMode (0.00s) === RUN TestOpenExistingDB --- PASS: TestOpenExistingDB (0.02s) === RUN TestIndexes --- PASS: TestIndexes (0.78s) === RUN TestAutoMigration --- PASS: TestAutoMigration (0.86s) === RUN TestPointerFields --- PASS: TestPointerFields (0.19s) === RUN TestPolymorphic --- PASS: TestPolymorphic (0.11s) === RUN TestPreload --- PASS: TestPreload (0.24s) === RUN TestFirstAndLast --- PASS: TestFirstAndLast (0.06s) === RUN TestFirstAndLastWithNoStdPrimaryKey --- PASS: TestFirstAndLastWithNoStdPrimaryKey (0.04s) === RUN TestUIntPrimaryKey --- PASS: TestUIntPrimaryKey (0.01s) === RUN TestFindAsSliceOfPointers --- PASS: TestFindAsSliceOfPointers (0.03s) === RUN TestSearchWithPlainSQL --- PASS: TestSearchWithPlainSQL (0.10s) === RUN TestSearchWithStruct --- PASS: TestSearchWithStruct (0.09s) === RUN TestSearchWithMap --- PASS: TestSearchWithMap (0.05s) === RUN TestSearchWithEmptyChain --- PASS: TestSearchWithEmptyChain (0.05s) === RUN TestSelect --- PASS: TestSelect (0.02s) === RUN TestOrderAndPluck --- PASS: TestOrderAndPluck (0.08s) === RUN TestLimit --- PASS: TestLimit (0.08s) === RUN TestOffset --- PASS: TestOffset (0.32s) === RUN TestOr --- PASS: TestOr (0.04s) === RUN TestCount --- PASS: TestCount (0.06s) === RUN TestNot --- PASS: TestNot (0.19s) === RUN TestFillSmallerStruct --- PASS: TestFillSmallerStruct (0.02s) === RUN TestFindOrInitialize --- PASS: TestFindOrInitialize (0.07s) === RUN TestFindOrCreate --- PASS: TestFindOrCreate (0.18s) === RUN TestSelectWithEscapedFieldName --- PASS: TestSelectWithEscapedFieldName (0.05s) === RUN TestSelectWithVariables --- PASS: TestSelectWithVariables (0.02s) === RUN TestSelectWithArrayInput --- PASS: TestSelectWithArrayInput (0.03s) === RUN TestScopes --- PASS: TestScopes (0.13s) === RUN TestScannableSlices --- PASS: TestScannableSlices (0.04s) === RUN TestUpdate --- PASS: TestUpdate (0.16s) === RUN TestUpdateWithNoStdPrimaryKeyAndDefaultValues --- PASS: TestUpdateWithNoStdPrimaryKeyAndDefaultValues (0.10s) === RUN TestUpdates --- PASS: TestUpdates (0.09s) === RUN TestUpdateColumn --- PASS: TestUpdateColumn (0.09s) === RUN TestSelectWithUpdate --- PASS: TestSelectWithUpdate (0.10s) === RUN TestSelectWithUpdateWithMap --- PASS: TestSelectWithUpdateWithMap (0.08s) === RUN TestOmitWithUpdate --- PASS: TestOmitWithUpdate (0.07s) === RUN TestOmitWithUpdateWithMap --- PASS: TestOmitWithUpdateWithMap (0.07s) === RUN TestSelectWithUpdateColumn --- PASS: TestSelectWithUpdateColumn (0.04s) === RUN TestOmitWithUpdateColumn --- PASS: TestOmitWithUpdateColumn (0.05s) PASS ok github.com/jinzhu/gorm 10.273s testing mysql... === RUN TestRegisterCallback --- PASS: TestRegisterCallback (0.00s) === RUN TestRegisterCallbackWithOrder --- PASS: TestRegisterCallbackWithOrder (0.00s) === RUN TestRegisterCallbackWithComplexOrder --- PASS: TestRegisterCallbackWithComplexOrder (0.00s) === RUN TestReplaceCallback [info] replacing callback `create` from /Users/jay/go/src/github.com/jinzhu/gorm/callback_test.go:94 --- PASS: TestReplaceCallback (0.00s) === RUN TestRemoveCallback [info] removing callback `create` from /Users/jay/go/src/github.com/jinzhu/gorm/callback_test.go:107 --- PASS: TestRemoveCallback (0.00s) === RUN TestCloneSearch --- PASS: TestCloneSearch (0.00s) === RUN TestHasOneAndHasManyAssociation --- PASS: TestHasOneAndHasManyAssociation (0.05s) === RUN TestRelated --- PASS: TestRelated (0.01s) === RUN TestManyToMany --- PASS: TestManyToMany (0.07s) === RUN TestForeignKey --- PASS: TestForeignKey (0.00s) === RUN TestRunCallbacks --- PASS: TestRunCallbacks (0.01s) === RUN TestCallbacksWithErrors --- PASS: TestCallbacksWithErrors (0.01s) === RUN TestCreate --- PASS: TestCreate (0.00s) === RUN TestCreateWithNoGORMPrimayKey --- PASS: TestCreateWithNoGORMPrimayKey (0.00s) === RUN TestCreateWithNoStdPrimaryKeyAndDefaultValues --- PASS: TestCreateWithNoStdPrimaryKeyAndDefaultValues (0.00s) === RUN TestAnonymousScanner --- PASS: TestAnonymousScanner (0.00s) === RUN TestAnonymousField --- PASS: TestAnonymousField (0.00s) === RUN TestSelectWithCreate --- PASS: TestSelectWithCreate (0.01s) === RUN TestOmitWithCreate --- PASS: TestOmitWithCreate (0.01s) === RUN TestCustomizeColumn --- PASS: TestCustomizeColumn (0.02s) === RUN TestCustomColumnAndIgnoredFieldClash --- PASS: TestCustomColumnAndIgnoredFieldClash (0.03s) === RUN TestDelete --- PASS: TestDelete (0.01s) === RUN TestInlineDelete --- PASS: TestInlineDelete (0.01s) === RUN TestSoftDelete --- PASS: TestSoftDelete (0.05s) === RUN TestSaveAndQueryEmbeddedStruct --- PASS: TestSaveAndQueryEmbeddedStruct (0.00s) === RUN TestJoinTable --- PASS: TestJoinTable (0.31s) === RUN TestExceptionsWithInvalidSql --- PASS: TestExceptionsWithInvalidSql (0.00s) === RUN TestSetTable --- PASS: TestSetTable (0.06s) === RUN TestHasTable --- PASS: TestHasTable (0.02s) === RUN TestTableName --- PASS: TestTableName (0.00s) === RUN TestSqlNullValue --- PASS: TestSqlNullValue (0.02s) === RUN TestTransaction --- PASS: TestTransaction (0.01s) === RUN TestRow --- PASS: TestRow (0.00s) === RUN TestRows --- PASS: TestRows (0.00s) === RUN TestScan --- PASS: TestScan (0.01s) === RUN TestRaw --- PASS: TestRaw (0.00s) === RUN TestGroup --- PASS: TestGroup (0.00s) === RUN TestJoins --- PASS: TestJoins (0.00s) === RUN TestHaving --- PASS: TestHaving (0.00s) === RUN TestTimeWithZone --- PASS: TestTimeWithZone (0.01s) === RUN TestHstore --- SKIP: TestHstore (0.00s) main_test.go:506: === RUN TestSetAndGet --- PASS: TestSetAndGet (0.00s) === RUN TestCompatibilityMode `testdb` is not officially supported, running under compatibility mode. --- PASS: TestCompatibilityMode (0.00s) === RUN TestOpenExistingDB --- PASS: TestOpenExistingDB (0.00s) === RUN TestIndexes --- PASS: TestIndexes (0.12s) === RUN TestAutoMigration --- PASS: TestAutoMigration (0.14s) === RUN TestPointerFields --- PASS: TestPointerFields (0.02s) === RUN TestPolymorphic --- PASS: TestPolymorphic (0.06s) === RUN TestPreload --- PASS: TestPreload (0.04s) === RUN TestFirstAndLast --- PASS: TestFirstAndLast (0.01s) === RUN TestFirstAndLastWithNoStdPrimaryKey --- PASS: TestFirstAndLastWithNoStdPrimaryKey (0.00s) === RUN TestUIntPrimaryKey --- PASS: TestUIntPrimaryKey (0.00s) === RUN TestFindAsSliceOfPointers --- PASS: TestFindAsSliceOfPointers (0.01s) === RUN TestSearchWithPlainSQL --- PASS: TestSearchWithPlainSQL (0.01s) === RUN TestSearchWithStruct --- PASS: TestSearchWithStruct (0.01s) === RUN TestSearchWithMap --- PASS: TestSearchWithMap (0.00s) === RUN TestSearchWithEmptyChain --- PASS: TestSearchWithEmptyChain (0.01s) === RUN TestSelect --- PASS: TestSelect (0.00s) === RUN TestOrderAndPluck --- PASS: TestOrderAndPluck (0.01s) === RUN TestLimit --- PASS: TestLimit (0.01s) === RUN TestOffset --- PASS: TestOffset (0.04s) === RUN TestOr --- PASS: TestOr (0.00s) === RUN TestCount --- PASS: TestCount (0.00s) === RUN TestNot --- PASS: TestNot (0.03s) === RUN TestFillSmallerStruct --- PASS: TestFillSmallerStruct (0.00s) === RUN TestFindOrInitialize --- PASS: TestFindOrInitialize (0.01s) === RUN TestFindOrCreate --- PASS: TestFindOrCreate (0.01s) === RUN TestSelectWithEscapedFieldName --- PASS: TestSelectWithEscapedFieldName (0.00s) === RUN TestSelectWithVariables --- PASS: TestSelectWithVariables (0.00s) === RUN TestSelectWithArrayInput --- PASS: TestSelectWithArrayInput (0.00s) === RUN TestScopes --- PASS: TestScopes (0.00s) === RUN TestScannableSlices --- PASS: TestScannableSlices (0.02s) === RUN TestUpdate --- PASS: TestUpdate (0.01s) === RUN TestUpdateWithNoStdPrimaryKeyAndDefaultValues --- PASS: TestUpdateWithNoStdPrimaryKeyAndDefaultValues (0.01s) === RUN TestUpdates --- PASS: TestUpdates (0.01s) === RUN TestUpdateColumn --- PASS: TestUpdateColumn (0.01s) === RUN TestSelectWithUpdate --- PASS: TestSelectWithUpdate (0.01s) === RUN TestSelectWithUpdateWithMap --- PASS: TestSelectWithUpdateWithMap (0.01s) === RUN TestOmitWithUpdate --- PASS: TestOmitWithUpdate (0.01s) === RUN TestOmitWithUpdateWithMap --- PASS: TestOmitWithUpdateWithMap (0.01s) === RUN TestSelectWithUpdateColumn --- PASS: TestSelectWithUpdateColumn (0.01s) === RUN TestOmitWithUpdateColumn --- PASS: TestOmitWithUpdateColumn (0.01s) PASS ok github.com/jinzhu/gorm 1.797s testing sqlite3... === RUN TestRegisterCallback --- PASS: TestRegisterCallback (0.00s) === RUN TestRegisterCallbackWithOrder --- PASS: TestRegisterCallbackWithOrder (0.00s) === RUN TestRegisterCallbackWithComplexOrder --- PASS: TestRegisterCallbackWithComplexOrder (0.00s) === RUN TestReplaceCallback [info] replacing callback `create` from /Users/jay/go/src/github.com/jinzhu/gorm/callback_test.go:94 --- PASS: TestReplaceCallback (0.00s) === RUN TestRemoveCallback [info] removing callback `create` from /Users/jay/go/src/github.com/jinzhu/gorm/callback_test.go:107 --- PASS: TestRemoveCallback (0.00s) === RUN TestCloneSearch --- PASS: TestCloneSearch (0.00s) === RUN TestHasOneAndHasManyAssociation --- PASS: TestHasOneAndHasManyAssociation (0.01s) === RUN TestRelated --- PASS: TestRelated (0.00s) === RUN TestManyToMany --- PASS: TestManyToMany (0.07s) === RUN TestForeignKey --- PASS: TestForeignKey (0.00s) === RUN TestRunCallbacks --- PASS: TestRunCallbacks (0.01s) === RUN TestCallbacksWithErrors --- PASS: TestCallbacksWithErrors (0.01s) === RUN TestCreate --- PASS: TestCreate (0.00s) === RUN TestCreateWithNoGORMPrimayKey --- PASS: TestCreateWithNoGORMPrimayKey (0.00s) === RUN TestCreateWithNoStdPrimaryKeyAndDefaultValues --- PASS: TestCreateWithNoStdPrimaryKeyAndDefaultValues (0.00s) === RUN TestAnonymousScanner --- PASS: TestAnonymousScanner (0.00s) === RUN TestAnonymousField --- PASS: TestAnonymousField (0.00s) === RUN TestSelectWithCreate --- PASS: TestSelectWithCreate (0.00s) === RUN TestOmitWithCreate --- PASS: TestOmitWithCreate (0.00s) === RUN TestCustomizeColumn --- PASS: TestCustomizeColumn (0.00s) === RUN TestCustomColumnAndIgnoredFieldClash --- PASS: TestCustomColumnAndIgnoredFieldClash (0.00s) === RUN TestDelete --- PASS: TestDelete (0.00s) === RUN TestInlineDelete --- PASS: TestInlineDelete (0.00s) === RUN TestSoftDelete --- PASS: TestSoftDelete (0.00s) === RUN TestSaveAndQueryEmbeddedStruct --- PASS: TestSaveAndQueryEmbeddedStruct (0.00s) === RUN TestJoinTable --- PASS: TestJoinTable (0.01s) === RUN TestExceptionsWithInvalidSql --- PASS: TestExceptionsWithInvalidSql (0.00s) === RUN TestSetTable --- PASS: TestSetTable (0.02s) === RUN TestHasTable --- PASS: TestHasTable (0.00s) === RUN TestTableName --- PASS: TestTableName (0.00s) === RUN TestSqlNullValue --- PASS: TestSqlNullValue (0.00s) === RUN TestTransaction --- PASS: TestTransaction (0.00s) === RUN TestRow --- PASS: TestRow (0.00s) === RUN TestRows --- PASS: TestRows (0.00s) === RUN TestScan --- PASS: TestScan (0.00s) === RUN TestRaw --- PASS: TestRaw (0.01s) === RUN TestGroup --- PASS: TestGroup (0.00s) === RUN TestJoins --- PASS: TestJoins (0.00s) === RUN TestHaving --- PASS: TestHaving (0.00s) === RUN TestTimeWithZone --- PASS: TestTimeWithZone (0.00s) === RUN TestHstore --- SKIP: TestHstore (0.00s) main_test.go:506: === RUN TestSetAndGet --- PASS: TestSetAndGet (0.00s) === RUN TestCompatibilityMode `testdb` is not officially supported, running under compatibility mode. --- PASS: TestCompatibilityMode (0.00s) === RUN TestOpenExistingDB `sqlite` is not officially supported, running under compatibility mode. --- PASS: TestOpenExistingDB (0.00s) === RUN TestIndexes --- PASS: TestIndexes (0.01s) === RUN TestAutoMigration --- PASS: TestAutoMigration (0.01s) === RUN TestPointerFields --- PASS: TestPointerFields (0.01s) === RUN TestPolymorphic --- PASS: TestPolymorphic (0.01s) === RUN TestPreload --- PASS: TestPreload (0.02s) === RUN TestFirstAndLast --- PASS: TestFirstAndLast (0.00s) === RUN TestFirstAndLastWithNoStdPrimaryKey --- PASS: TestFirstAndLastWithNoStdPrimaryKey (0.00s) === RUN TestUIntPrimaryKey --- PASS: TestUIntPrimaryKey (0.00s) === RUN TestFindAsSliceOfPointers --- PASS: TestFindAsSliceOfPointers (0.01s) === RUN TestSearchWithPlainSQL --- PASS: TestSearchWithPlainSQL (0.01s) === RUN TestSearchWithStruct --- PASS: TestSearchWithStruct (0.01s) === RUN TestSearchWithMap --- PASS: TestSearchWithMap (0.00s) === RUN TestSearchWithEmptyChain --- PASS: TestSearchWithEmptyChain (0.00s) === RUN TestSelect --- PASS: TestSelect (0.00s) === RUN TestOrderAndPluck --- PASS: TestOrderAndPluck (0.01s) === RUN TestLimit --- PASS: TestLimit (0.01s) === RUN TestOffset --- PASS: TestOffset (0.04s) === RUN TestOr --- PASS: TestOr (0.00s) === RUN TestCount --- PASS: TestCount (0.00s) === RUN TestNot --- PASS: TestNot (0.02s) === RUN TestFillSmallerStruct --- PASS: TestFillSmallerStruct (0.02s) === RUN TestFindOrInitialize --- PASS: TestFindOrInitialize (0.00s) === RUN TestFindOrCreate --- PASS: TestFindOrCreate (0.01s) === RUN TestSelectWithEscapedFieldName --- PASS: TestSelectWithEscapedFieldName (0.00s) === RUN TestSelectWithVariables --- PASS: TestSelectWithVariables (0.00s) === RUN TestSelectWithArrayInput --- PASS: TestSelectWithArrayInput (5.01s) === RUN TestScopes --- PASS: TestScopes (0.00s) === RUN TestScannableSlices --- PASS: TestScannableSlices (0.00s) === RUN TestUpdate --- PASS: TestUpdate (0.00s) === RUN TestUpdateWithNoStdPrimaryKeyAndDefaultValues --- PASS: TestUpdateWithNoStdPrimaryKeyAndDefaultValues (0.00s) === RUN TestUpdates --- PASS: TestUpdates (0.00s) === RUN TestUpdateColumn --- PASS: TestUpdateColumn (0.00s) === RUN TestSelectWithUpdate --- PASS: TestSelectWithUpdate (0.00s) === RUN TestSelectWithUpdateWithMap --- PASS: TestSelectWithUpdateWithMap (0.01s) === RUN TestOmitWithUpdate --- PASS: TestOmitWithUpdate (0.00s) === RUN TestOmitWithUpdateWithMap --- PASS: TestOmitWithUpdateWithMap (0.00s) === RUN TestSelectWithUpdateColumn --- PASS: TestSelectWithUpdateColumn (0.00s) === RUN TestOmitWithUpdateColumn --- PASS: TestOmitWithUpdateColumn (0.00s) PASS ok github.com/jinzhu/gorm 5.598s LGTM, thank you for your work.
gharchive/pull-request
2015-03-24T17:52:52
2025-04-01T04:34:41.178359
{ "authors": [ "jaytaylor", "jinzhu" ], "repo": "jinzhu/gorm", "url": "https://github.com/jinzhu/gorm/pull/436", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1840972753
multimonitor glitch i laptop is locked and suspended the i connect it to my monitor the monitor screen is not locked and i can see content and mouse work just keyboard input is freezed. This probably means that the swaylock-effects process crashed and the session lock protocol is probably not used or there’s some bug in the compositor because the mouse shouldn’t work either. Which Wayland compositor (e.g. Sway) do you use and which version? I use Hyprland v0.27.2 Well, I’ve looked into Hyprland issues and that compositor is so buggy that there’s a good chance it’s a bug in Hyprland, not in the swaylock effects. I’m still on Sway 1.7, which doesn’t implement the session lock protocol, but I’m about to upgrade to Sway 1.8, so I’ll see how it behaves with session locking and multi-monitors after that. that compositor is so buggy sway has almost twice as many open bug reports :) but it might be and most likely is an issue on my side, but I can't test multimon stuff atm (on holiday)
gharchive/issue
2023-08-08T09:50:58
2025-04-01T04:34:41.185159
{ "authors": [ "jirutka", "niksingh710", "vaxerski" ], "repo": "jirutka/swaylock-effects", "url": "https://github.com/jirutka/swaylock-effects/issues/43", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
58387935
Making consulate installation optional and minor fixes I made consulate installation optional as it was described in README.md, added archive download check to prevent file download if it's already existing. Also added default consul_node_name value as ansible_nodename. Confirmed that get_url does not do anything if file exists. will only download every time if you specify a directory: If dest is a directory, the file will always be downloaded (regardless of the force option), but replaced only if the contents changed Since this was not merge-able and we got another pull request I took that one, thanks for the time you put into this.
gharchive/pull-request
2015-02-20T17:58:32
2025-04-01T04:34:41.258079
{ "authors": [ "MrMMorris", "manul7", "savagegus" ], "repo": "jivesoftware/ansible-consul", "url": "https://github.com/jivesoftware/ansible-consul/pull/22", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2228696580
Support for CUDA Similar to #23 , would you be open to adding support for building for CUDA? I have not yet figured out how to do this (don't have much experience with cc/cxx/build scripts) but this would be great for deploying on infra such as HF Inference Endpoints. It would be nice if we supported CUDA. However, I'm a Mac user, and testing with NVIDIA's GPUs is not easy for me... It seems like bindgen_cuda has some logic for finding proper CUDA paths to link and it can probably be added as an optional dependency similar to intel-mkl-src in #23 . It needs an attached gpu to build but seems like that is also pretty common. I can take a look at this in a week or so
gharchive/issue
2024-04-05T19:39:55
2025-04-01T04:34:41.307629
{ "authors": [ "filiptibell", "jkawamoto" ], "repo": "jkawamoto/ctranslate2-rs", "url": "https://github.com/jkawamoto/ctranslate2-rs/issues/24", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2519651768
[CRP] [GPS] Novatel up and running Description: take over and integrate novatel driver into CRP (as subrepo), publish ego odometry topic. AC: subrepo added add launch file to publish the right topic test with docker (measurement stored in common drive) system spec doc updated Effort plan: 0.8SP Prio 1 A test docker with the and packages was built and ran successfully on the jetson. The docker was started with the following arguments: docker run -it --rm --network host novatel:latest Added novatel driver as a submodule and created a launcher package to have a launcher while leaving the driver repository untouched. Created a wrapper package for nova submodule to separate the launcher (novatel_gps_launcher). The CRP docker has been tested with the novatel driver in leaf (on jetson) with success. The docker should be started like this: docker run -it --rm --network host --ipc host --pid host crp_arm64:latest Also added installation of some useful tools to the docker like net-tools and mcap for recording bags. Documentation has been updated with the purpose, usage and launch parameters of the novatel launcher.
gharchive/issue
2024-09-11T12:45:45
2025-04-01T04:34:41.337964
{ "authors": [ "AnonymDavid" ], "repo": "jkk-research/CooperativeResearchPlatform", "url": "https://github.com/jkk-research/CooperativeResearchPlatform/issues/8", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
53289195
cleanconfig deprecated ? Hey, I'm implementing the last release on a project, but I can't find the way to config the clean function. Is it deprecated? I was using the example below: cleanconfig: { plugins: [ { removeViewBox: false }, { removeUselessStrokeAndFill: false }, { cleanupIDs: false }, { mergePaths: false } ] }, Hi @miguelmoraleda, I guess you answered your own question in the meantime, right? Happy spriting anyway! :) Cheers, Joschi @miguelmoraleda btw, please don't wipe out the issues fields — this makes it pretty impossible to reconstruct the issue history. Thanks! :+1: Hi @jkphl, You are right. Sorry about the wipe :). I'm running in a new problem... SVGO is adding a letter to each of my IDs. Do you have any idea how to avoid this behaviour? Thanks in advance. Best regards, Miguel It's not SVGO adding the letter, but svg-sprite itself. To avoid possible ID conflicts while combining multiple SVG files into a sprite, the IDs need to get "namespaced", which is done by prepending them with a (unique) letter. Otherwise ID clashes could happen ... (This isn't something new, by the way, has been the same with previous svg-sprite releases) I could imagine making the namespacing optional, however. In that case, you would have to avoid ID clashes manually. Interested in this? I could try to do this tomorrow (it's 2:30 a.m. over here, definitely time for going to bed ;)). Cheers, Joschi Yeah!! I'm very interested on it. I would be grateful if you can add namespacing as optional. In my current flow I have a "src" svg folder and a "dest" svg folder for my sprite... In the src folder I'm adding IDs to specific paths, so I can animate it later. The namespacing scary me a little bit because my IDs can change the name every time I regenerate the sprite. Have a nice rest and thank you for a great work :) Best regards, Miguel Closing this in favour of #27. Hey @miguelmoraleda, as of release v1.0.6 grunt-svg-sprite now supports disabling ID namespacing. Please see the namespaceIDs option for further details. Cheers, Joschi Hey @jkphl, Thank you!!! :)
gharchive/issue
2015-01-03T01:05:52
2025-04-01T04:34:41.349523
{ "authors": [ "jkphl", "miguelmoraleda" ], "repo": "jkphl/grunt-svg-sprite", "url": "https://github.com/jkphl/grunt-svg-sprite/issues/26", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2229177605
FLAG CONFIG_OAUTH2_VALIDATE_COMMON__NAME Current Behavior Validate common name flag is not working properly as in the logs of oauth the first 2 logs line showing blank common name ill be attaching the logs below and it failed to validate the common name on the other hand if i don't use this flag then first 2 logs doesn't show the common name but on successful authentication it show the common name in the logs LOGS WHEN VALIDATE_COMMON_NAME Flag is turned on with idtoken.email level=WARN msg="user validation: common_name mismatch: openvpn client is empty" ip=192.168.35.1:58566 cid=2 kid=1 session_id=zSZQiybIkkWLA58z common_name="" idtoken.subject=112549187817525499082 idtoken.email=mubasit.khan@arpatech.com idtoken.preferred_username="" user.subject=112549187817525499082 user.preferred_username="" error_id=91dc984f40fb6d4fef886db14928ece2d92101388e0a2fd7d68f60512876cf8d level=INFO msg="client disconnected" ip=: cid=2 common_name="" reason=DISCONNECT session_id="" session_state="" LOGS WHEN VALIDATE_COMMON_NAME Flag is turned off level=INFO msg="client established" ip=192.168.35.1:54232 vpn_ip=10.8.0.3 cid=4 common_name=mubasit.khan reason=ESTABLISHED session_id=UHYdv/QiDjo8fk3j session_state=Initial Expected Behavior Like my Common name should validate with idtoken.email and if validation is completed then should be able to connect Steps To Reproduce No response Environment openvpn-auth-oauth2 Version:1.19.1 OpenVPN Server Version: 2.6.9 Server OS: Linux Ubuntu OpenVPN Client (Windows, OS): OIDC Provider: Gsuite openvpn-auth-oauth2 logs Apr 06 06:11:08 ARP-792 openvpn-auth-oauth2[1019981]: time=2024-04-06T06:11:08.255-04:00 level=INFO msg="accept OpenVPN client cid 3, kid 1" ip=192.168.35.1:55137 cid=3 kid=1 session_id=dK/cFymJt3IzyJd7 common_name="" idtoken.subject=112549187817525499082 idtoken.email=mubasit.khan@arpatech.com idtoken.preferred_username="" user.subject=112549187817525499082 user.preferred_username="" Apr 06 06:11:13 ARP-792 openvpn-auth-oauth2[1019981]: time=2024-04-06T06:11:13.614-04:00 level=INFO msg="client established" ip=192.168.35.1:55137 vpn_ip=10.8.0.6 cid=3 common_name="" reason=ESTABLISHED session_id=dK/cFymJt3IzyJd7 session_state=Initial Apr 06 06:12:07 ARP-792 openvpn-auth-oauth2[1019981]: time=2024-04-06T06:12:07.632-04:00 level=INFO msg="client disconnected" ip=192.168.35.1:55137 cid=3 common_name="" reason=DISCONNECT session_id=dK/cFymJt3IzyJd7 session_state=Initial Apr 06 06:12:14 ARP-792 openvpn-auth-oauth2[1019981]: time=2024-04-06T06:12:14.166-04:00 level=INFO msg="new client connection" ip=192.168.35.1:54232 cid=4 kid=1 common_name="" reason=CONNECT session_id=UHYdv/QiDjo8fk3j session_state=Initial Apr 06 06:12:14 ARP-792 openvpn-auth-oauth2[1019981]: time=2024-04-06T06:12:14.166-04:00 level=INFO msg="start pending auth" ip=192.168.35.1:54232 cid=4 kid=1 common_name="" reason=CONNECT session_id=UHYdv/QiDjo8fk3j session_state=Initial Apr 06 06:12:14 ARP-792 openvpn-auth-oauth2[1019981]: time=2024-04-06T06:12:14.840-04:00 level=INFO msg="initialize authorization via oauth2" ip=192.168.35.1:54232 cid=4 kid=1 common_name="" Apr 06 06:12:16 ARP-792 openvpn-auth-oauth2[1019981]: time=2024-04-06T06:12:16.959-04:00 level=INFO msg="successful authorization via oauth2" ip=192.168.35.1:54232 cid=4 kid=1 session_id=UHYdv/QiDjo8fk3j common_name="" idtoken.subject=112549187817525499082 idtoken.email=mubasit.khan@arpatech.com idtoken.preferred_username="" user.subject=112549187817525499082 user.preferred_username="" Apr 06 06:12:16 ARP-792 openvpn-auth-oauth2[1019981]: time=2024-04-06T06:12:16.959-04:00 level=INFO msg="accept OpenVPN client cid 4, kid 1" ip=192.168.35.1:54232 cid=4 kid=1 session_id=UHYdv/QiDjo8fk3j common_name="" idtoken.subject=112549187817525499082 idtoken.email=mubasit.khan@arpatech.com idtoken.preferred_username="" user.subject=112549187817525499082 user.preferred_username="" Apr 06 06:12:23 ARP-792 openvpn-auth-oauth2[1019981]: time=2024-04-06T06:12:23.198-04:00 level=INFO msg="client established" ip=192.168.35.1:54232 vpn_ip=10.8.0.3 cid=4 common_name=mubasit.khan reason=ESTABLISHED session_id=UHYdv/QiDjo8fk3j session_state=Initial Apr 06 06:12:41 ARP-792 openvpn-auth-oauth2[1019981]: time=2024-04-06T06:12:41.876-04:00 level=INFO msg="client disconnected" ip=192.168.35.1:54232 cid=4 common_name=mubasit.khan reason=DISCONNECT session_id=UHYdv/QiDjo8fk3j session_state=Initial openvpn server logs No response Anything else? No response CONFIG_OAUTH2_VALIDATE_COMMON__NAME is not a boolean which requires true or false. Instead it accepts a string value which be any claim from your idtoken. If you wish to compare the command with the email field, use CONFIG_OAUTH2_VALIDATE_COMMON__NAME=email but on successful authentication it show the common name in the logs It sounds like you are using username-as-common-name, read the FAQ here what todo here: https://github.com/jkroepke/openvpn-auth-oauth2/wiki/FAQ#username-as-common-name
gharchive/issue
2024-04-06T10:17:19
2025-04-01T04:34:41.362069
{ "authors": [ "Mubasitkhan99", "jkroepke" ], "repo": "jkroepke/openvpn-auth-oauth2", "url": "https://github.com/jkroepke/openvpn-auth-oauth2/issues/247", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2010976707
[Feature Request] Different kinds of clock The digital clock is nice, but it might not be the least distractive option for everyone. I have implemented different kinds of clocks you can choose in the options. Currently, besides the standard digital clock, i have implemented the option to show no clock at all, to have an simple analog watch or (my favorite) to display the time in binary format. Would love to prepare a pull request, but with so many in the pipeline i am not sure where to base it on and prefer to wait until the way forward is clearer for me. For me they do work and i would like to share. What are your opinions, what can and should be improved? This is (of course) related to the issue #81, could include #65 and might replace pull request #146.
gharchive/issue
2023-11-26T09:58:42
2025-04-01T04:34:41.365014
{ "authors": [ "khwolf" ], "repo": "jkuester/unlauncher", "url": "https://github.com/jkuester/unlauncher/issues/173", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
196237334
emitting events Hi! What a nice component! I wonder if it's possible to emit the events for the the component. There is a outputEvents event, but it's not called. Cool events would be: -onInit() -onOpen() -onClose() Thank you! hi @egandro. thanks! just commited and published to npm version with input & output events. for more info please see https://github.com/jkuri/ng2-datepicker/commit/9b192e730d5db1341e085dea6cfaf90bfe9c01c3. Hi! Thank you that is of great help to me!
gharchive/issue
2016-12-17T19:02:45
2025-04-01T04:34:41.368871
{ "authors": [ "egandro", "jkuri" ], "repo": "jkuri/ng2-datepicker", "url": "https://github.com/jkuri/ng2-datepicker/issues/121", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
197266244
Not working with Angular 2 Universal When load the app, got below error: Exception: Call to Node module failed with error: TypeError: ctorParameters.map is not a function Can you provide more info about this please? I am using ASP.NET Core, developing in Visual Studio 2015 My package.json { "name": "Angular2Spa", "version": "0.0.0", "dependencies": { "@angular/common": "2.0.0", "@angular/compiler": "2.0.0", "@angular/core": "2.0.0", "@angular/forms": "2.0.0", "@angular/http": "2.0.0", "@angular/platform-browser": "2.0.0", "@angular/platform-browser-dynamic": "2.0.0", "@angular/platform-server": "2.0.0", "@angular/router": "3.0.0", "@types/node": "^6.0.38", "angular2-platform-node": "^2.1.0-rc.1", "angular2-universal": "^2.1.0-rc.1", "angular2-universal-polyfills": "^2.1.0-rc.1", "aspnet-prerendering": "^1.0.6", "aspnet-webpack": "^1.0.11", "bootstrap": "^3.3.7", "css": "^2.2.1", "css-loader": "^0.25.0", "es6-shim": "^0.35.1", "expose-loader": "^0.7.1", "extract-text-webpack-plugin": "^1.0.1", "file-loader": "^0.9.0", "isomorphic-fetch": "^2.2.1", "jquery": "^2.2.1", "ng2-datepicker": "1.4.7", "ng2-modal": "0.0.24", "preboot": "^4.5.2", "raw-loader": "^0.5.1", "reflect-metadata": "^0.1.9", "rxjs": "5.0.0-beta.12", "style-loader": "^0.13.0", "to-string-loader": "^1.1.5", "ts-loader": "^0.8.2", "typescript": "^2.0.0", "url-loader": "^0.5.7", "webpack": "^1.12.14", "webpack-externals-plugin": "^1.0.0", "webpack-hot-middleware": "^2.10.0", "webpack-merge": "^0.14.1", "zone.js": "^0.6.21" } } My app.moudule.ts: import { NgModule } from '@angular/core'; import { RouterModule } from '@angular/router'; import { FormsModule } from '@angular/forms'; import { UniversalModule } from 'angular2-universal'; import { ModalModule } from "ng2-modal"; import { DatePickerModule } from 'ng2-datepicker'; import { AppComponent } from './components/app/app.component' import { NavMenuComponent } from './components/navmenu/navmenu.component'; import { HomeComponent } from './components/home/home.component'; import { FetchDataComponent } from './components/fetchdata/fetchdata.component'; import { CounterComponent } from './components/counter/counter.component'; import { DataEntryComponent } from './components/dataentry/dataentry.component'; @NgModule({ bootstrap: [ AppComponent ], declarations: [ AppComponent, NavMenuComponent, CounterComponent, FetchDataComponent, DataEntryComponent, HomeComponent ], imports: [ UniversalModule, // Must be first import. This automatically imports BrowserModule, HttpModule, and JsonpModule too. FormsModule, ModalModule, DatePickerModule, RouterModule.forRoot([ { path: '', redirectTo: 'home', pathMatch: 'full' }, { path: 'home', component: HomeComponent }, { path: 'counter', component: CounterComponent }, { path: 'fetch-data', component: FetchDataComponent }, { path: 'data-entry', component: DataEntryComponent }, { path: '**', redirectTo: 'home' } ]) ] }) export class AppModule { } You can see I am using the UniversalModule to do the server-side angular rendering. The error I see is like this: An unhandled exception occurred while processing the request. Exception: Call to Node module failed with error: TypeError: ctorParameters.map is not a function at ReflectionCapabilities.parameters (C:\Projects\Learning\CoreAngular2\Angular2Application1\Angular2Application1\node_modules@angular\core\bundles\core.umd.js:2118:51) at Reflector.parameters (C:\Projects\Learning\CoreAngular2\Angular2Application1\Angular2Application1\node_modules@angular\core\bundles\core.umd.js:2319:52) at CompileMetadataResolver.getDependenciesMetadata (C:\Projects\Learning\CoreAngular2\Angular2Application1\Angular2Application1\node_modules@angular\compiler\bundles\compiler.umd.js:14336:81) at CompileMetadataResolver.getTypeMetadata (C:\Projects\Learning\CoreAngular2\Angular2Application1\Angular2Application1\node_modules@angular\compiler\bundles\compiler.umd.js:14301:28) at CompileMetadataResolver.getDirectiveMetadata (C:\Projects\Learning\CoreAngular2\Angular2Application1\Angular2Application1\node_modules@angular\compiler\bundles\compiler.umd.js:14074:30) at C:\Projects\Learning\CoreAngular2\Angular2Application1\Angular2Application1\node_modules@angular\compiler\bundles\compiler.umd.js:14143:51 at Array.forEach (native) at Object.Call (C:\Projects\Learning\CoreAngular2\Angular2Application1\Angular2Application1\node_modules\es6-shim\es6-shim.js:307:14) at Array.forEach (C:\Projects\Learning\CoreAngular2\Angular2Application1\Angular2Application1\node_modules\es6-shim\es6-shim.js:1313:17) at CompileMetadataResolver.getNgModuleMetadata (C:\Projects\Learning\CoreAngular2\Angular2Application1\Angular2Application1\node_modules@angular\compiler\bundles\compiler.umd.js:14136:46) I have the same problem
gharchive/issue
2016-12-22T21:05:12
2025-04-01T04:34:41.387125
{ "authors": [ "jkuri", "jmaocanada", "nenad-matic-devtech" ], "repo": "jkuri/ng2-datepicker", "url": "https://github.com/jkuri/ng2-datepicker/issues/130", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
161501987
Demo not working. Demo is not working.Tried using with Angular RC version and its not working. Working now.
gharchive/issue
2016-06-21T18:01:43
2025-04-01T04:34:41.388263
{ "authors": [ "jkuri", "uzumakinaruto123" ], "repo": "jkuri/ng2-slimscroll", "url": "https://github.com/jkuri/ng2-slimscroll/issues/9", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1998358015
Why does proto not return the proto read Why does proto not return the proto read , eg. 2, or 3 ? Nevermind, I was a noob :/ (still am doh)
gharchive/issue
2023-11-17T06:21:12
2025-04-01T04:34:41.424132
{ "authors": [ "boltex" ], "repo": "jlaine/node-jpickle", "url": "https://github.com/jlaine/node-jpickle/issues/17", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
577390884
this project is awesome, thanks! see title :) Wanted to say the same. Just amazing, thanks for this great idea and implementation!
gharchive/issue
2020-03-07T19:48:32
2025-04-01T04:34:41.424886
{ "authors": [ "tugdualenligne", "wbolster" ], "repo": "jlesage/docker-handbrake", "url": "https://github.com/jlesage/docker-handbrake/issues/107", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
725276816
Get list of phonemes from librispeech lexicon Issue: The librispeech lexicon contains a list of common english words and their composing phonemes. Need to know what phonemes exist Task: Create function that returns a list of phonemes given the librispeech_lexicon.txt file Desired input: Filepath to librispeech_lexicon.txt file Desired output: Python list of all phonemes sorted alphabetically. (The natsort library might be handy if the standard python sort doesn't work) The lexicon format follows [word] \t [phoneme_1] [phoneme_2] ... [phoneme_n] For example: ACCELERATE AE0 K S EH1 L ER0 EY2 T List of syllables and composing Phonemes http://web.archive.org/web/20160822211027/http:/semarch.linguistics.fas.nyu.edu/barker/Syllables/index.txt """ Description: Create a function that will get all the phonemes that exist in the file 'librispeech-lexicon.txt'. The code will take in the file 'librispeech-lexicon.txt' and extract each line from the file. The file is structured as [word] \t [phoneme1] ... [phonemeN]. After reading a line from the file, the first will be take out of the line leaving the phonemes in the line. The line will be split into each individual words/phonemes and be placed into a numpy array. Only phonemes which are in the numpy array will be added that way there will be no repetition of phonemes into the array. The expected output is a sorted array consisting on each phoneme that is in the file with no duplicates. Current inputs: librispeech-lexicon.txt Expected output: List of all the phonemes in the input file with no duplicates Error(s)/warning(s) to work on: FutureWarning: elementwise comparison failed; returning scalar instead, but in the future will perform elementwise comparison if word not in test_arr: """ from pathlib import Path import numpy as np def phoneme_list(): #Path of input file. Change at your will. text_path = Path('C:\\Deepcut\\SD1_Intervals_assignment\\librispeech-lexicon.txt') # creates an empty numpy array phoneme_arr = np.array([]) # Opens file to read and will be referenced as file in the loop with open(text_path, 'r') as file: # takes one line from the file. Will be referenced as line for line in file: # discards the file word file each letter discard, space, phoneme_line = line.partition(' ') # this will split the line (the one without the first word) # and place the individual words into the numpy array # the split line will be referenced as word for word in phoneme_line.split(): # if word is not in the numpy array then it will added in there if word not in phoneme_arr: phoneme_arr = np.append(phoneme_arr, word) # closes file file.close() # prints a sorted array with no duplicates print(np.sort(phoneme_arr, kind='mergesort')) if __name__ == '__main__': phoneme_list() This is a good solution. I'm may move the path to be an argument and have it return a list as the output, but this works. Thanks!
gharchive/issue
2020-10-20T07:11:23
2025-04-01T04:34:41.447150
{ "authors": [ "deliat258", "jluckenbaugh2" ], "repo": "jluckenbaugh2/Deepcut", "url": "https://github.com/jluckenbaugh2/Deepcut/issues/2", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1972308192
garbage output on small models spread to many GPUs when loading a small model on multiple GPUs, it produces garbage. the machine has 4 x 3070 (8GB) and an older i5-7400, UBU 22.04, Cuda 11.8 How to reproduce starting the server by hand ollama serve ollama run zephyr >>> why is the sky blue? acia##############################################################################^C nvidia-smi shows it loads 4 x 1,2 GB Workaround Till we get some flags to control the loading to a discrete GPU start the server with CUDA_VISIBLE_DEVICES CUDA_VISIBLE_DEVICES="3" ollama serve # loads the model on GPU No. 4 There seems to be a "memory-low-limit-point", under which this error occures. hitting same issue serve logs ❯ ollama serve 2023/11/11 13:14:54 images.go:824: total blobs: 10 2023/11/11 13:14:54 images.go:831: total unused blobs removed: 0 2023/11/11 13:14:54 routes.go:696: Listening on 127.0.0.1:11434 (version 0.1.9) [GIN] 2023/11/11 - 13:14:59 | 200 | 50.322µs | 127.0.0.1 | HEAD "/" [GIN] 2023/11/11 - 13:14:59 | 200 | 788.233µs | 127.0.0.1 | POST "/api/show" 2023/11/11 13:15:00 llama.go:290: 47146 MB VRAM available, loading up to 309 GPU layers 2023/11/11 13:15:00 llama.go:415: starting llama runner 2023/11/11 13:15:00 llama.go:473: waiting for llama runner to start responding ggml_init_cublas: found 2 CUDA devices: Device 0: NVIDIA GeForce RTX 3090, compute capability 8.6 Device 1: NVIDIA GeForce RTX 3090, compute capability 8.6 {"timestamp":1699704901,"level":"INFO","function":"main","line":1323,"message":"build info","build":219,"commit":"9e70cc0"} {"timestamp":1699704901,"level":"INFO","function":"main","line":1325,"message":"system info","n_threads":64,"n_threads_batch":-1,"total_threads":128,"system_info":"AVX = 1 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | FMA = 0 | NEON = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | "} llama_model_loader: loaded meta data with 19 key-value pairs and 291 tensors from /home/m/.ollama/models/blobs/sha256:22f7f8ef5f4c791c1b03d7eb414399294764d7cc82c7e94aa81a1feb80a983a2 (version GGUF V2 (latest)) llama_model_loader: - tensor 0: token_embd.weight q4_0 [ 4096, 32000, 1, 1 ] llama_model_loader: - tensor 1: blk.0.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 2: blk.0.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 3: blk.0.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 4: blk.0.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 5: blk.0.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 6: blk.0.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 7: blk.0.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 8: blk.0.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 9: blk.0.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 10: blk.1.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 11: blk.1.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 12: blk.1.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 13: blk.1.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 14: blk.1.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 15: blk.1.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 16: blk.1.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 17: blk.1.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 18: blk.1.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 19: blk.10.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 20: blk.10.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 21: blk.10.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 22: blk.10.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 23: blk.10.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 24: blk.10.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 25: blk.10.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 26: blk.10.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 27: blk.10.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 28: blk.11.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 29: blk.11.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 30: blk.11.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 31: blk.11.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 32: blk.11.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 33: blk.11.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 34: blk.11.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 35: blk.11.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 36: blk.11.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 37: blk.12.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 38: blk.12.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 39: blk.12.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 40: blk.12.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 41: blk.12.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 42: blk.12.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 43: blk.12.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 44: blk.12.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 45: blk.12.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 46: blk.13.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 47: blk.13.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 48: blk.13.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 49: blk.13.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 50: blk.13.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 51: blk.13.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 52: blk.13.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 53: blk.13.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 54: blk.13.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 55: blk.14.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 56: blk.14.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 57: blk.14.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 58: blk.14.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 59: blk.14.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 60: blk.14.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 61: blk.14.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 62: blk.14.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 63: blk.14.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 64: blk.15.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 65: blk.15.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 66: blk.15.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 67: blk.15.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 68: blk.15.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 69: blk.15.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 70: blk.15.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 71: blk.15.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 72: blk.15.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 73: blk.16.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 74: blk.16.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 75: blk.16.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 76: blk.16.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 77: blk.16.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 78: blk.16.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 79: blk.16.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 80: blk.16.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 81: blk.16.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 82: blk.17.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 83: blk.17.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 84: blk.17.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 85: blk.17.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 86: blk.17.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 87: blk.17.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 88: blk.17.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 89: blk.17.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 90: blk.17.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 91: blk.18.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 92: blk.18.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 93: blk.18.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 94: blk.18.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 95: blk.18.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 96: blk.18.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 97: blk.18.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 98: blk.18.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 99: blk.18.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 100: blk.19.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 101: blk.19.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 102: blk.19.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 103: blk.19.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 104: blk.19.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 105: blk.19.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 106: blk.19.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 107: blk.19.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 108: blk.19.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 109: blk.2.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 110: blk.2.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 111: blk.2.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 112: blk.2.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 113: blk.2.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 114: blk.2.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 115: blk.2.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 116: blk.2.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 117: blk.2.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 118: blk.20.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 119: blk.20.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 120: blk.20.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 121: blk.20.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 122: blk.20.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 123: blk.20.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 124: blk.20.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 125: blk.20.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 126: blk.20.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 127: blk.21.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 128: blk.21.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 129: blk.21.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 130: blk.21.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 131: blk.21.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 132: blk.21.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 133: blk.21.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 134: blk.21.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 135: blk.21.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 136: blk.22.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 137: blk.22.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 138: blk.22.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 139: blk.22.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 140: blk.22.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 141: blk.22.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 142: blk.22.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 143: blk.22.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 144: blk.22.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 145: blk.23.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 146: blk.23.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 147: blk.23.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 148: blk.23.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 149: blk.23.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 150: blk.23.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 151: blk.23.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 152: blk.23.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 153: blk.23.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 154: blk.3.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 155: blk.3.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 156: blk.3.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 157: blk.3.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 158: blk.3.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 159: blk.3.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 160: blk.3.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 161: blk.3.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 162: blk.3.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 163: blk.4.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 164: blk.4.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 165: blk.4.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 166: blk.4.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 167: blk.4.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 168: blk.4.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 169: blk.4.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 170: blk.4.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 171: blk.4.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 172: blk.5.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 173: blk.5.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 174: blk.5.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 175: blk.5.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 176: blk.5.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 177: blk.5.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 178: blk.5.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 179: blk.5.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 180: blk.5.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 181: blk.6.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 182: blk.6.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 183: blk.6.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 184: blk.6.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 185: blk.6.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 186: blk.6.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 187: blk.6.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 188: blk.6.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 189: blk.6.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 190: blk.7.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 191: blk.7.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 192: blk.7.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 193: blk.7.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 194: blk.7.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 195: blk.7.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 196: blk.7.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 197: blk.7.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 198: blk.7.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 199: blk.8.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 200: blk.8.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 201: blk.8.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 202: blk.8.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 203: blk.8.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 204: blk.8.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 205: blk.8.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 206: blk.8.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 207: blk.8.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 208: blk.9.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 209: blk.9.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 210: blk.9.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 211: blk.9.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 212: blk.9.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 213: blk.9.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 214: blk.9.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 215: blk.9.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 216: blk.9.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 217: output.weight q6_K [ 4096, 32000, 1, 1 ] llama_model_loader: - tensor 218: blk.24.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 219: blk.24.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 220: blk.24.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 221: blk.24.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 222: blk.24.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 223: blk.24.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 224: blk.24.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 225: blk.24.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 226: blk.24.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 227: blk.25.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 228: blk.25.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 229: blk.25.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 230: blk.25.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 231: blk.25.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 232: blk.25.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 233: blk.25.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 234: blk.25.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 235: blk.25.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 236: blk.26.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 237: blk.26.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 238: blk.26.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 239: blk.26.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 240: blk.26.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 241: blk.26.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 242: blk.26.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 243: blk.26.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 244: blk.26.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 245: blk.27.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 246: blk.27.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 247: blk.27.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 248: blk.27.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 249: blk.27.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 250: blk.27.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 251: blk.27.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 252: blk.27.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 253: blk.27.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 254: blk.28.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 255: blk.28.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 256: blk.28.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 257: blk.28.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 258: blk.28.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 259: blk.28.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 260: blk.28.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 261: blk.28.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 262: blk.28.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 263: blk.29.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 264: blk.29.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 265: blk.29.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 266: blk.29.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 267: blk.29.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 268: blk.29.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 269: blk.29.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 270: blk.29.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 271: blk.29.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 272: blk.30.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 273: blk.30.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 274: blk.30.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 275: blk.30.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 276: blk.30.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 277: blk.30.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 278: blk.30.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 279: blk.30.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 280: blk.30.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 281: blk.31.attn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 282: blk.31.ffn_down.weight q4_0 [ 11008, 4096, 1, 1 ] llama_model_loader: - tensor 283: blk.31.ffn_gate.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 284: blk.31.ffn_up.weight q4_0 [ 4096, 11008, 1, 1 ] llama_model_loader: - tensor 285: blk.31.ffn_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - tensor 286: blk.31.attn_k.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 287: blk.31.attn_output.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 288: blk.31.attn_q.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 289: blk.31.attn_v.weight q4_0 [ 4096, 4096, 1, 1 ] llama_model_loader: - tensor 290: output_norm.weight f32 [ 4096, 1, 1, 1 ] llama_model_loader: - kv 0: general.architecture str llama_model_loader: - kv 1: general.name str llama_model_loader: - kv 2: llama.context_length u32 llama_model_loader: - kv 3: llama.embedding_length u32 llama_model_loader: - kv 4: llama.block_count u32 llama_model_loader: - kv 5: llama.feed_forward_length u32 llama_model_loader: - kv 6: llama.rope.dimension_count u32 llama_model_loader: - kv 7: llama.attention.head_count u32 llama_model_loader: - kv 8: llama.attention.head_count_kv u32 llama_model_loader: - kv 9: llama.attention.layer_norm_rms_epsilon f32 llama_model_loader: - kv 10: general.file_type u32 llama_model_loader: - kv 11: tokenizer.ggml.model str llama_model_loader: - kv 12: tokenizer.ggml.tokens arr llama_model_loader: - kv 13: tokenizer.ggml.scores arr llama_model_loader: - kv 14: tokenizer.ggml.token_type arr llama_model_loader: - kv 15: tokenizer.ggml.bos_token_id u32 llama_model_loader: - kv 16: tokenizer.ggml.eos_token_id u32 llama_model_loader: - kv 17: tokenizer.ggml.unknown_token_id u32 llama_model_loader: - kv 18: general.quantization_version u32 llama_model_loader: - type f32: 65 tensors llama_model_loader: - type q4_0: 225 tensors llama_model_loader: - type q6_K: 1 tensors llm_load_vocab: special tokens definition check successful ( 259/32000 ). llm_load_print_meta: format = GGUF V2 (latest) llm_load_print_meta: arch = llama llm_load_print_meta: vocab type = SPM llm_load_print_meta: n_vocab = 32000 llm_load_print_meta: n_merges = 0 llm_load_print_meta: n_ctx_train = 4096 llm_load_print_meta: n_embd = 4096 llm_load_print_meta: n_head = 32 llm_load_print_meta: n_head_kv = 32 llm_load_print_meta: n_layer = 32 llm_load_print_meta: n_rot = 128 llm_load_print_meta: n_gqa = 1 llm_load_print_meta: f_norm_eps = 0.0e+00 llm_load_print_meta: f_norm_rms_eps = 1.0e-05 llm_load_print_meta: f_clamp_kqv = 0.0e+00 llm_load_print_meta: f_max_alibi_bias = 0.0e+00 llm_load_print_meta: n_ff = 11008 llm_load_print_meta: freq_base_train = 10000.0 llm_load_print_meta: freq_scale_train = 1 llm_load_print_meta: model type = 7B llm_load_print_meta: model ftype = mostly Q4_0 llm_load_print_meta: model params = 6.74 B llm_load_print_meta: model size = 3.56 GiB (4.54 BPW) llm_load_print_meta: general.name = LLaMA v2 llm_load_print_meta: BOS token = 1 '<s>' llm_load_print_meta: EOS token = 2 '</s>' llm_load_print_meta: UNK token = 0 '<unk>' llm_load_print_meta: LF token = 13 '<0x0A>' llm_load_tensors: ggml ctx size = 0.10 MB llm_load_tensors: using CUDA for GPU acceleration ggml_cuda_set_main_device: using device 0 (NVIDIA GeForce RTX 3090) as main device llm_load_tensors: mem required = 70.41 MB llm_load_tensors: offloading 32 repeating layers to GPU llm_load_tensors: offloading non-repeating layers to GPU llm_load_tensors: offloaded 35/35 layers to GPU llm_load_tensors: VRAM used: 3577.55 MB .................................................................................................. llama_new_context_with_model: n_ctx = 4096 llama_new_context_with_model: freq_base = 10000.0 llama_new_context_with_model: freq_scale = 1 llama_kv_cache_init: offloading v cache to GPU llama_kv_cache_init: offloading k cache to GPU llama_kv_cache_init: VRAM kv self = 2048.00 MB llama_new_context_with_model: kv self size = 2048.00 MB llama_new_context_with_model: compute buffer total size = 294.13 MB llama_new_context_with_model: VRAM scratch buffer: 288.00 MB llama_new_context_with_model: total VRAM used: 5913.56 MB (model: 3577.55 MB, context: 2336.00 MB) llama server listening at http://127.0.0.1:52490 {"timestamp":1699704904,"level":"INFO","function":"main","line":1746,"message":"HTTP server listening","hostname":"127.0.0.1","port":52490} {"timestamp":1699704904,"level":"INFO","function":"log_server_request","line":1233,"message":"request","remote_addr":"127.0.0.1","remote_port":59550,"status":200,"method":"HEAD","path":"/","params":{}} 2023/11/11 13:15:04 llama.go:487: llama runner started in 4.200890 seconds [GIN] 2023/11/11 - 13:15:04 | 200 | 4.623967608s | 127.0.0.1 | POST "/api/generate" run logs ❯ ollama run llama2 >>> hi mel#################################################^C >>> thanks for the workaround @chymian! This is a bug that's consistently reproducible. Possibly related: https://github.com/ggerganov/llama.cpp/issues/3772#issuecomment-1778754477 it's still happening with 0.1.10 According to ggerganov/llama.cpp#2470, we can define GGML_CUDA_PEER_MAX_BATCH_SIZE=0 (at compile time) to disable buggy peer access as a workaround. @jmorganca Would this be an option you may want to consider for the time being? @chymian could you retry on the latest version to see if this is still a problem? I'm on the latest and getting this now on an M2 Ultra. @briankwest the Apple M2's have a single logical GPU, so it's unlikely your issue is related to the multi-GPU splitting logic for NVIDIA/Radeon GPUs. Please file a new issue with more details about your scenario and include server logs so we can see what might be going wrong.
gharchive/issue
2023-11-01T12:43:10
2025-04-01T04:34:41.488405
{ "authors": [ "briankwest", "chymian", "dhiltgen", "mrasputin", "wookayin" ], "repo": "jmorganca/ollama", "url": "https://github.com/jmorganca/ollama/issues/961", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1323360891
Provide developer instructions Developer instructions would help me and others contribute. I would need to know How to install for development How to "build" and test The style guide (or automated checks used) Example I can see this project uses Poetry. I'm not familiar with poetry. I normally use pyproject.toml+setup.cfg+setup.py and can do pip install -e .. What would I do here to install for development? Solved by updating pip.
gharchive/issue
2022-07-31T04:45:42
2025-04-01T04:34:41.494410
{ "authors": [ "MarcSkovMadsen", "jmosbacher" ], "repo": "jmosbacher/pydantic-panel", "url": "https://github.com/jmosbacher/pydantic-panel/issues/8", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
655910865
Fix Numba deprecation warning Change @numba.jitclass(csr_spec) to @numba.experimental.jitclass(csr_spec) to get rid of the Numba deprecation warning. Coverage remained the same at 99.453% when pulling a4ea6920e69821284eda5ea996ef405a2ebdf64a on theodumont:numba-deprecator into ddfd73b7337223c62a035487a7b5fa6748d7f307 on jni:master. Thank you @theodumont!
gharchive/pull-request
2020-07-13T14:48:01
2025-04-01T04:34:41.523360
{ "authors": [ "coveralls", "jni", "theodumont" ], "repo": "jni/skan", "url": "https://github.com/jni/skan/pull/87", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
709048689
Removing dead link Hey @jnv just removing a dead project. Cheers! LGTM, thanks Peter! No worries Jan!
gharchive/pull-request
2020-09-25T15:57:06
2025-04-01T04:34:41.528826
{ "authors": [ "jnv", "spekulatius" ], "repo": "jnv/lists", "url": "https://github.com/jnv/lists/pull/161", "license": "cc0-1.0", "license_type": "permissive", "license_source": "bigquery" }
165582298
Docker install fails with error at Composer/SimpleSAML Step I've found this Dockerfile very useful as an exercise in learning Docker and in the process get my head around simpleSAML. I started with the Docker image available on the Docker hub which saved a lot of time in debugging my iDP. I did hit a problem when trying to create my own Dockerfile, and then later discovered that the problem exists here also, so it wasn't just my simpleSAML newbie status. When running the install the Composer step that performs the simpleSAML install fails now, presumably due to changes in simpleSAML dependencies since this script was created. Here is the output showing the problem Step 28 : RUN php composer.phar install ---> Running in b2c6110c2f1d PHP Warning: Module 'mcrypt' already loaded in Unknown on line 0 PHP Warning: Module 'mcrypt' already loaded in Unknown on line 0 Running composer as root/super user is highly discouraged as packages, plugins and scripts cannot always be trusted Loading composer repositories with package information Installing dependencies (including require-dev) from lock file Your requirements could not be resolved to an installable set of packages. Problem 1 - The requested PHP extension ext-pdo_sqlite * is missing from your system. Install or enable PHP's pdo_sqlite extension. The command '/bin/sh -c php composer.phar install' returned a non-zero code: 2 This may be all that is needed to fix this Add to Dockerfile before the Composer run RUN apt-get install -y php5-sqlite @paulfreeman much appreciated. I added this in and it's all good.
gharchive/issue
2016-07-14T14:57:24
2025-04-01T04:34:41.531213
{ "authors": [ "jnyryan", "paulfreeman" ], "repo": "jnyryan/docker-simplesamlphp", "url": "https://github.com/jnyryan/docker-simplesamlphp/issues/2", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1465075887
🛑 MegaConstruct is down In 2358740, MegaConstruct (https://www.megaconstruct.ro) was down: HTTP code: 0 Response time: 0 ms Resolved: MegaConstruct is back up in ea7c8f4.
gharchive/issue
2022-11-26T07:25:14
2025-04-01T04:34:41.536119
{ "authors": [ "joahn3" ], "repo": "joahn3/earthlink-uptime", "url": "https://github.com/joahn3/earthlink-uptime/issues/1651", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2304432544
🛑 CEI Sales Wiki is down In b522872, CEI Sales Wiki (https://wiki.ceisales.com/login) was down: HTTP code: 0 Response time: 0 ms Resolved: CEI Sales Wiki is back up in 5eac147 after 17 minutes.
gharchive/issue
2024-05-19T05:06:03
2025-04-01T04:34:41.538568
{ "authors": [ "joahn3" ], "repo": "joahn3/earthlink-uptime", "url": "https://github.com/joahn3/earthlink-uptime/issues/3347", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2754890255
🛑 Paicu & Sons Consulting is down In 4bf46bb, Paicu & Sons Consulting (https://paicucons.ro) was down: HTTP code: 0 Response time: 0 ms Resolved: Paicu & Sons Consulting is back up in a902456 after 17 minutes.
gharchive/issue
2024-12-22T23:11:50
2025-04-01T04:34:41.540936
{ "authors": [ "joahn3" ], "repo": "joahn3/earthlink-uptime", "url": "https://github.com/joahn3/earthlink-uptime/issues/4011", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2268448029
🛑 Security & Telecom Systems is down In 075b9c3, Security & Telecom Systems (https://security-telecom.ro) was down: HTTP code: 0 Response time: 0 ms Resolved: Security & Telecom Systems is back up in 689fcae after 5 minutes.
gharchive/issue
2024-04-29T08:40:11
2025-04-01T04:34:41.543275
{ "authors": [ "joahn3" ], "repo": "joahn3/upptime2", "url": "https://github.com/joahn3/upptime2/issues/937", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
300323430
pigpiod stops working after stack smashing Hi, Thank you very much for your great tool pigpio. I really like it! However, I ran into an issue and hopefully somebody can help me. In my application pigpiod regularly stops working after some time. The pigpiod process then is still alive, but one of its threads died after gcc's checking found a corrupted stack. This then requires killing and restarting pigpiod and my app: pi@raspi3:~ $ cat /dev/pigerr *** stack smashing detected ***: /usr/bin/pigpiod terminated 2018-02-24 15:20:18 sigHandler: Unhandled signal 6, terminating After removing the signal handler for SIGABRT 6 from the code and adding debug options in the Makefile (-g -O0, removing the strip commands) I was able to create a core dump. It shows an overwrite beyond the limit of array report[MAX_REPORT] in function alertEmit(): root@raspi3:~/PIGPIO/crash# gdb pigpiod core GNU gdb (Raspbian 7.12-6) 7.12.0.20161007-git ... Reading symbols from pigpiod...done. [New LWP 430] [New LWP 541] [New LWP 542] [New LWP 431] [New LWP 434] [New LWP 428] [Thread debugging using libthread_db enabled] Using host libthread_db library "/lib/arm-linux-gnueabihf/libthread_db.so.1". Core was generated by `pigpiod'. Program terminated with signal SIGABRT, Aborted. #0 __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:51 51 ../sysdeps/unix/sysv/linux/raise.c: Datei oder Verzeichnis nicht gefunden. [Current thread is 1 (Thread 0x767b7470 (LWP 430))] (gdb) bt #0 __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:51 #1 0x76cdc824 in __GI_abort () at abort.c:89 #2 0x76d15f78 in __libc_message (do_abort=do_abort@entry=2, fmt=<optimized out>) at ../sysdeps/posix/libc_fatal.c:175 #3 0x76d92264 in __GI___fortify_fail (msg=0x76dd1d90 "stack smashing detected") at fortify_fail.c:30 #4 0x76d9221c in __stack_chk_fail () at stack_chk_fail.c:28 #5 0x76e47a20 in alertEmit (sample=0x767af164, numSamples=120, changedBits=3211264, eTick=1108643017) at pigpio.c:6128 #6 0x243991fe in ?? () Backtrace stopped: previous frame identical to this frame (corrupt stack?) (gdb) fr 5 #5 0x76e47a20 in alertEmit (sample=0x767af164, numSamples=120, changedBits=3211264, eTick=1108643017) at pigpio.c:6128 ... (gdb) p report[0] $18 = {seqno = 21513, flags = 0, tick = 1108637207, level = 606704111} (gdb) p report[1] $19 = {seqno = 21514, flags = 0, tick = 1108637267, level = 607752687} (gdb) p report[119] $20 = {seqno = 21632, flags = 0, tick = 1108643017, level = 607752703} (gdb) p report[120] $21 = {seqno = 21633, flags = 44, tick = 1108643017, level = 607752703} (gdb) p report[121] $22 = {seqno = 21634, flags = 48, tick = 1108643017, level = 607752703} (gdb) p report[122] $23 = {seqno = 21635, flags = 51, tick = 1108643017, level = 607752703} (gdb) p report[123] $25 = {seqno = 21636, flags = 52, tick = 1108643017, level = 607752703} (gdb) p report[124] $26 = {seqno = 21637, flags = 58, tick = 1108643017, level = 607752703} (gdb) p report[125] $27 = {seqno = 0, flags = 0, tick = 2126326868, level = 0} (gdb) p report[126] $28 = {seqno = 4096, flags = 1081, tick = 33675, level = 1108646412} The array size is MAX_REPORT=120, it has been written until report[124] in this case. My setup: Raspberry PI 3 with up-to-date Raspbian (Jessie) and the latest PIGPIO V66. root@raspi3:~/PIGPIO# cat /etc/debian_version 9.3 root@raspi3:~/PIGPIO# uname -a Linux raspi3 4.9.59-v7+ #1047 SMP Sun Oct 29 12:19:23 GMT 2017 armv7l GNU/Linux root@raspi3:~/PIGPIO# pigpiod -v 66 I'm reading signals of 6 TSIC 306 temperature sensors connected to 6 GPIOs using callbacks for both signal edges in a Python 3 application. Each of the sensors sends two bytes 10 times a second with the ZacWire protocol. This is about 10*40 signal edges per sensor per second. The application also sets watchdog timers to detect the ends of transmissions. The problem seems to depend on timing. My tests with stripped down applications also caused the crash, but by far not as frequently. Please let me know how I can help to fix this issue. Best regards Holger The most useful thing would be a script which triggers the problem. Perhaps using PWM to generate edges similar to those being generated by the sensors. Ok, I'll have a look at how some stress test could be simulated. Meanwhile, I tested increasing the buffer size MAX_REPORT from 120 to 1200 and pigpiod worked without any problems for over 20 hours now. With the 120 entry buffer it usually crashed after 10 to 30 minutes. Of course that is not a clean solution. Don't worry about the stress test unless it is easy. Thanks for the very detailed report. I looked in more detail at your core dump and I think it shows the problem. If a watchdog or event is due for the notification I stupidly append the report to the array without checking that there is room. It appears to be a straightforward bug. I might simply dimension the array to be MAX_REPORT + 64 to allow for the maximum number of watchdogs and events (1 each per GPIO). I chose MAX_REPORT to be 120 for a reason i can no longer remember, may have been ethernet frame size. If you get a chance could you change gpioReport_t report[MAX_REPORT]; to gpioReport_t report[MAX_REPORT + 64]; in alertEmit() I will add a fix for this bug in V67. This is easy... the test with report[MAX_REPORT + 64] is already running, now. I'll give some feedback tomorrow. The test with buffer report[120 + 64] has been running for 15 hours without problems. Looks good! Thank you for your help! Good, thanks. In that case your core dump has probably identified the actual bug.
gharchive/issue
2018-02-26T17:16:46
2025-04-01T04:34:41.554659
{ "authors": [ "grillbaer", "joan2937" ], "repo": "joan2937/pigpio", "url": "https://github.com/joan2937/pigpio/issues/201", "license": "unlicense", "license_type": "permissive", "license_source": "bigquery" }
761852303
Update org.eclipse.jgit to 5.10.0.202012080955-r Updates org.eclipse.jgit:org.eclipse.jgit from 5.8.1.202007141445-r to 5.10.0.202012080955-r. I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! Ignore future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "org.eclipse.jgit", artifactId = "org.eclipse.jgit" } ] labels: library-update Superseded by #22.
gharchive/pull-request
2020-12-11T03:32:58
2025-04-01T04:34:41.558425
{ "authors": [ "scala-steward" ], "repo": "joan38/mill-git", "url": "https://github.com/joan38/mill-git/pull/21", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2157184612
Fix Creating-a-Crew-and-kick-it-off.md so it can run Update deps to include crewai[tools] Remove invalid max_inter arg from Task constructor call This is great, thanks!
gharchive/pull-request
2024-02-27T17:07:46
2025-04-01T04:34:41.562212
{ "authors": [ "barapa", "joaomdmoura" ], "repo": "joaomdmoura/crewAI", "url": "https://github.com/joaomdmoura/crewAI/pull/280", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
94531097
search of lowercase asN fails Not uppercasing the search value results in false results irrtree -s as2914 AS-OPENTRANSIT results in IRRTree (1.1.0) report for 'AS-OPENTRANSIT' (IPv4), using rr.ntt.net at 2015-07-12 04 :01 NOT_FOUND: as2914 not present in AS-OPENTRANSIT or any of its members but irrtree -s AS2914 AS-OPENTRANSIT returned IRRTree (1.1.0) report for 'AS-OPENTRANSIT' (IPv4), using rr.ntt.net at 2015-07-12 03 :11 AS-OPENTRANSIT (31302 ASNs) +-- AS-INTEROUTE (29504 ASNs) +-- AS-WIFIWEB-TRANSITO (6 ASNs) +-- AS2914 (2489 pfxs) Fixed in 4db9b198fb6e45f0a62b392fa0f5c77b1b93e618. thanks
gharchive/issue
2015-07-12T04:03:49
2025-04-01T04:34:41.566866
{ "authors": [ "job", "troy2914" ], "repo": "job/irrtree", "url": "https://github.com/job/irrtree/issues/3", "license": "bsd-2-clause", "license_type": "permissive", "license_source": "bigquery" }
370448700
wrong index? https://github.com/jocicmarko/ultrasound-nerve-segmentation/blob/df9eb77f0ac3b84ae125b6b177ef2e4429818d66/submission.py#L43 Should it be imgs_test[i,:,:,0]? I think so too. Otherwise, the shape of the img becomes (96, 1) giving the empty submission as the result.
gharchive/issue
2018-10-16T05:47:46
2025-04-01T04:34:41.590043
{ "authors": [ "GeorgeBatch", "zihaozhihao" ], "repo": "jocicmarko/ultrasound-nerve-segmentation", "url": "https://github.com/jocicmarko/ultrasound-nerve-segmentation/issues/82", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2234734960
Please update the image to use Libreoffice ^V24.2.0.2 Hi, We have a problem converting tiff images to PDF with jodconverter due to a bug (lacking feature) in LibreOffice, that has been reported to be fixed since release 24.2.0.2 (https://bugs.documentfoundation.org/show_bug.cgi?id=155444) Please update libreoffice on this container. Released https://github.com/jodconverter/docker-image-jodconverter-runtime/releases/tag/0.1.8 Hi, Eugen, I checked the latest jodconverter image on the repo and it still uses an outdated version of libreoffice: to use version 24 instead of 7.4 you might need to do steps as described in https://libre-software.net/linux/how-to-install-libreoffice-on-ubuntu-linux-mint/ sudo add-apt-repository ppa:libreoffice/ppa sudo apt update sudo apt upgrade I was playing a bit around, to get the latest libreoffice installed I needed to create a sources.list file in /etc/apt/sources.list.d with the line deb http://deb.debian.org/debian bookworm-backports main followed by apt update and apt install -t bookworm-backports libreoffice this installed libreoffice (4:24.2.0-1~bpo12+1) I will not go forward and switch to backports or any other frankendebian variant to upgrade libreoffice - this runtime should be the stable baseline, not an experimental spearhead. This said, surely i understand your need, so i would suggest what ever project you actually build uppon the runtime your won project / jodconverter solution something based on https://github.com/EugenMayer/officeconverter In case 1, just install the libreoffice version of your liking during the app-packing step. In case 2, create your own image based on the published version and upgrade the libre-office variant. Both are viable and do not need you to fork and maintain all the things, thus i think those are viable solutions in this case.
gharchive/issue
2024-04-10T04:31:26
2025-04-01T04:34:41.598176
{ "authors": [ "EugenMayer", "SchlauFuchs" ], "repo": "jodconverter/docker-image-jodconverter-runtime", "url": "https://github.com/jodconverter/docker-image-jodconverter-runtime/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
364687771
License? Hello! I'm sure this has come up before, but we're creating an Apache licensed project http://github.com/puppetlabs/bolt and would not like for it to be held to the GPL license. Is the license flexible at all? I believe that LGPL, GPLv3, or Apache would allow us to maintain Bolt's Apache license (or others, we're open to discussion) @joekhoobyar @jballanc My only contribution to this project is extremely minor. If you need approval, I would gladly re-license that single line change under a more permissive license. I wonder, though, this gem has not been touched in some years. Perhaps a clean-room re-implementation might be called for?
gharchive/issue
2018-09-27T23:23:32
2025-04-01T04:34:41.627330
{ "authors": [ "jballanc", "lucywyman" ], "repo": "joekhoobyar/net-ssh-kerberos", "url": "https://github.com/joekhoobyar/net-ssh-kerberos/issues/7", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
193283565
Enhance UI .. First of all I want to say I loved this app.The code was well documented and fun to work with..I was having some free time so I decided to enhance the UI.This is what I have done(looks better on hardware):- You can change anything you don't like or can contact me.The only trouble I am having is with the NNID stuff because I don't have a NNID currently . This would require the three heavenly libs(sftd,sfii and sf2d) for building from now onward.The file size has increased by about 1 mb but I feel thats OK for an improved UI. Hi, as much as I love supporting open source and submitting pull requests, I'll have to deny this one. The reason being I'm trying to make this project retain the same look and feel as PSPident, with a tiny bit of changes. Besides I already have an alternate GUI version in mind. It is very identical to my other project, VITAident, which has both the original and the GUI version. https://gbatemp.net/threads/release-3dsident.435755/page-5#post-6704153 Ok I understand and your gui version looks a lot better than mine 👍🏼👍🏼
gharchive/pull-request
2016-12-03T12:15:02
2025-04-01T04:34:41.630077
{ "authors": [ "Pirater12", "joel16" ], "repo": "joel16/3DSident", "url": "https://github.com/joel16/3DSident/pull/3", "license": "Zlib", "license_type": "permissive", "license_source": "github-api" }
1257688880
Enable Logging? How can i enable logging the call queries? I tried this, but there is no logging of the queries? $l = new MyCustomLogger(); $this->googleAds = new GoogleAds($this->clientId,$this->clientSecret, $this->developerToken, SupportedVersions::VERSION_10, 'rest', $l, 'debug'); I'm not sure this is a bug with this package. If you've passed a correct instance of the monolog logger interface, you should be seeing request logs... I'll see if I can replicate this issue myself
gharchive/issue
2022-06-02T05:02:30
2025-04-01T04:34:41.631558
{ "authors": [ "dominikraff", "joelbutcher" ], "repo": "joelbutcher/laravel-googleads", "url": "https://github.com/joelbutcher/laravel-googleads/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1352951170
useTitleAsClassname ignores character Schema definition: { "$schema": "http://json-schema.org/draft-04/schema#", "title": "myTestClassV1_0", } This will generate a class with the name myTestClassV10 instead of myTestClassV1_0 This is because of the default 'property word delimiters'. Underscore is assumed to separate words, and is then used to create camel case names. If you set property word delimiters to only space and hyphen (something like that) then you should see the underscore is retained. That did it, many thanks :)
gharchive/issue
2022-08-27T05:40:30
2025-04-01T04:34:41.638014
{ "authors": [ "Stromner", "joelittlejohn" ], "repo": "joelittlejohn/jsonschema2pojo", "url": "https://github.com/joelittlejohn/jsonschema2pojo/issues/1429", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
183140258
Object name of "_" causes exception I have some json I'm trying to process that uses the single character "_" as the name (not the value) of one of the fields in the json. If I had the option of changing the name of the field I would. But, it's a third party library (apache solr). At any rate, the normalizeName() function throws an exception. Exception in thread "main" java.lang.StringIndexOutOfBoundsException: String index out of range: 0 at java.lang.String.charAt(Unknown Source) at org.jsonschema2pojo.util.NameHelper.normalizeName(NameHelper.java:45) at org.jsonschema2pojo.util.NameHelper.getPropertyName(NameHelper.java:76) I guess we can fix this quite easily. Do you have any thoughts about what you'd like to see generated when we encounter a property called _? First, to be clear, I won't tell you I'm very animated about getting this fixed. I'm ok if you ignore it or wait and fix it when there are other more important fixes. The workaround is straight forward. Change the "_" in the example json to something like "underscore", generate the POJOs, and then go fix the annotations for this field. Not that hard. I've already done that. Second, if it had been almost anyone else other than someone big like apache solr generating this, I wouldn't have said anything. People who use names like this or (my bigger pet peave) use object names to provide some of the information about the object (like use the name of a server as the name of an object instead of having a string named "server" in the object with the name of the server as a value) are IMHO misusing the standard. It makes parsing much harder as you either have to write a custom deserializer or use a map and lose the safety of having a real pojo with types for the values and a list of getters/setters. OK, I feel better now. If you do want to do something, I'd probably have the name of the field as "underscore" or "hyphen" if the name is "-". But, again, I'm ok if you don't fix it. Finally, I wanted to say this. I have to generate parsers for some company's json on a regular basis. Being able to take their example json (nobody gives me schemas), run it through your code and generate pojos has saved me a lot of time. So, thank you for doing this program. Thanks @Smitty010, it's nice to get this kind of insight into how you're thinking about this change. I think if the string contains no valid characters we should use the unicode name for the chars within it. This could produce a long name, but it's an edge-case so as long as we have some way to support this I think that's okay. We can prefix field name with something to indicate that it's a special translation (e.g. with __ or $). In this case, your underscore field would have been translated to a field called: __low_line This kind of JSON: { "#!" : "foo", ";;" : "bar" } would produce fields named: __number_sign_exclamation_mark __semicolon_semicolon I encountered the same exception, and finally found that the error is caused by field name which is using Chinese Characters....After changing to PinYin or English characters, exception will be eliminated. Hope there could be some upgrade for such kind of situation:)) Faced the same exception when using Russian characters as keys. If I use English characters the problem goes away.
gharchive/issue
2016-10-14T20:11:40
2025-04-01T04:34:41.644870
{ "authors": [ "JohnLuCN", "Smitty010", "firsovroman", "joelittlejohn" ], "repo": "joelittlejohn/jsonschema2pojo", "url": "https://github.com/joelittlejohn/jsonschema2pojo/issues/632", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1035156748
ClientAuthError Hi, after obtaining token library is making to get requests and then throws error: client_info_empty_error: The client info was empty. Please review the trace to determine the root cause. node_modules/@azure/msal-common/dist/error/AuthError.js:27:1 ` 25 | var _this = this; 26 | var errorString = errorMessage ? errorCode + ": " + errorMessage : errorCode; 27 | _this = _super.call(this, errorString) || this; | ^ 28 | Object.setPrototypeOf(_this, AuthError.prototype); 29 | _this.errorCode = errorCode || Constants.EMPTY_STRING; 30 | _this.errorMessage = errorMessage || "";` stack trace: at ClientAuthError.AuthError [as constructor] (webpack:///./node_modules/@azure/msal-common/dist/error/AuthError.js:27:1) at new ClientAuthError (webpack:///./node_modules/@azure/msal-common/dist/error/ClientAuthError.js:198:1) at Function../node_modules/@azure/msal-common/dist/error/ClientAuthError.js.ClientAuthError.createClientInfoEmptyError (webpack:///./node_modules/@azure/msal-common/dist/error/ClientAuthError.js:215:1) at ResponseHandler../node_modules/@azure/msal-common/dist/response/ResponseHandler.js.ResponseHandler.generateAccountEntity (webpack:///./node_modules/@azure/msal-common/dist/response/ResponseHandler.js:209:34) at ResponseHandler../node_modules/@azure/msal-common/dist/response/ResponseHandler.js.ResponseHandler.generateCacheRecord (webpack:///./node_modules/@azure/msal-common/dist/response/ResponseHandler.js:160:1) at ResponseHandler.eval (webpack:///./node_modules/@azure/msal-common/dist/response/ResponseHandler.js:102:1) at step (webpack:///./node_modules/@azure/msal-common/dist/_virtual/_tslib.js:75:1) at Object.eval [as next] (webpack:///./node_modules/@azure/msal-common/dist/_virtual/_tslib.js:56:45) at eval (webpack:///./node_modules/@azure/msal-common/dist/_virtual/_tslib.js:49:1) at new Promise (<anonymous>) From previous event: at getRet (http://localhost:60245/__cypress/runner/cypress_runner.js:153093:21) From previous event: at Context.thenFn (http://localhost:60245/__cypress/runner/cypress_runner.js:153111:64) at Context.then (http://localhost:60245/__cypress/runner/cypress_runner.js:153563:22) at Context.<anonymous> (http://localhost:60245/__cypress/runner/cypress_runner.js:169833:24) at <unknown> (http://localhost:60245/__cypress/runner/cypress_runner.js:168527:18) From previous event: at runCommand (http://localhost:60245/__cypress/runner/cypress_runner.js:168506:11) at next (http://localhost:60245/__cypress/runner/cypress_runner.js:168649:17) at <unknown> (http://localhost:60245/__cypress/runner/cypress_runner.js:168678:19) From previous event: at next (http://localhost:60245/__cypress/runner/cypress_runner.js:168649:37) From previous event: at <unknown> (http://localhost:60245/__cypress/runner/cypress_runner.js:182790:80) From previous event: at Object.run (http://localhost:60245/__cypress/runner/cypress_runner.js:182785:24) at Object.run (http://localhost:60245/__cypress/runner/cypress_runner.js:168703:18) at $Cy.cy.<computed> [as msalLogin] (http://localhost:60245/__cypress/runner/cypress_runner.js:169873:20) at Context.runnable.fn (http://localhost:60245/__cypress/runner/cypress_runner.js:170097:24) at callFn (http://localhost:60245/__cypress/runner/cypress_runner.js:122959:22) at Test.../driver/node_modules/mocha/lib/runnable.js.Runnable.run (http://localhost:60245/__cypress/runner/cypress_runner.js:122946:8) at <unknown> (http://localhost:60245/__cypress/runner/cypress_runner.js:176627:31) From previous event: at Object.onRunnableRun (http://localhost:60245/__cypress/runner/cypress_runner.js:176612:20) at $Cypress.action (http://localhost:60245/__cypress/runner/cypress_runner.js:166178:29) at Test.Runnable.run (http://localhost:60245/__cypress/runner/cypress_runner.js:174346:14) at Runner.../driver/node_modules/mocha/lib/runner.js.Runner.runTest (http://localhost:60245/__cypress/runner/cypress_runner.js:123618:11) at <unknown> (http://localhost:60245/__cypress/runner/cypress_runner.js:123744:13) at next (http://localhost:60245/__cypress/runner/cypress_runner.js:123527:15) at <unknown> (http://localhost:60245/__cypress/runner/cypress_runner.js:123537:8) at next (http://localhost:60245/__cypress/runner/cypress_runner.js:123439:15) at <unknown> (http://localhost:60245/__cypress/runner/cypress_runner.js:123505:6) at timeslice (http://localhost:60245/__cypress/runner/cypress_runner.js:117431:28) have you got such issue? As this issue is stale and I have released an updated version I close this issue.
gharchive/issue
2021-10-25T13:42:53
2025-04-01T04:34:41.649668
{ "authors": [ "joelmeiller", "pazernh" ], "repo": "joelmeiller/cypress-msal-login", "url": "https://github.com/joelmeiller/cypress-msal-login/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1088447432
Value index not properly removing value https://atomicdata.dev/agents/publicAgent appears in https://atomicdata.dev/properties What could have caused this? The property existed before, but has been removed. The value cache is not properly updated. Something else? I tried running with --rebuild-index, and it fixed the issue. I think the value cache does not update properly in some cases [ ] Add value index test for changing a property-value (unless one already exists) [ ] Add value index test for removing a resource I was able to reproduce this by chaning the parent of the default /setup resource, which seems to be a simple thing to add to tests.
gharchive/issue
2021-12-24T16:32:27
2025-04-01T04:34:41.665878
{ "authors": [ "joepio" ], "repo": "joepio/atomic-data-rust", "url": "https://github.com/joepio/atomic-data-rust/issues/258", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1419934948
build docs in github ci Build and publish docs in github CI. This is a follow up ticket from #4 . Tooling docs for multiple versions, to be evaluated: sphinx RTD theme with GitHub versions sphinx multiversion sphinx versions sphinx contrib multiversion I used sphinx RTD theme with GitHub versions and set up github ci accordingly. As a result, documentation is built when pushing commits. The documentation is published using github pages.
gharchive/issue
2022-10-23T21:05:12
2025-04-01T04:34:41.710025
{ "authors": [ "johannesjh" ], "repo": "johannesjh/req2flatpak", "url": "https://github.com/johannesjh/req2flatpak/issues/9", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
97297631
resource and libvrary to handle acls A lib to manage acls on a consul server using the HTTP api and the resources that use it. The plan is to later add more resources/providers that use that same consul_client.rb lib instead of writing files. @InformatiQ Please take a look at the new libraries which were merged into #126 and refactor to follow this form.
gharchive/pull-request
2015-07-26T08:20:26
2025-04-01T04:34:41.719454
{ "authors": [ "InformatiQ", "johnbellone" ], "repo": "johnbellone/consul-cookbook", "url": "https://github.com/johnbellone/consul-cookbook/pull/203", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1691543226
Boolean in Infix Expressions Syntax: <expression> in <list> Returns true if the result of the expression is in the list, and false if the result of the expression is not in the list. Commit: https://github.com/johneastman/boomerang/commit/e7707b0888b91c25aa76923859acc871df04775c
gharchive/issue
2023-05-01T23:16:08
2025-04-01T04:34:41.724699
{ "authors": [ "johneastman" ], "repo": "johneastman/boomerang", "url": "https://github.com/johneastman/boomerang/issues/70", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
424415264
negative fetch size is mysql-specific I used MongoSyphon against SQL Server at a customer, and it turns out that setting the fetch size to a negative value, as in line 92 in RDBMSConnection.java, is a feature specific to the MySQL driver. It didn't work with the MSSQL 6.2.1 driver and I had to comment out that line, after which things seemed to be okay. But I don't know the full implications of that. Discussion on stackoverflow: https://stackoverflow.com/questions/25019410/confusion-with-setfetchsize-method-of-statement-object Link to the place in the code: https://github.com/johnlpage/MongoSyphon/blob/a66484c875b244c8e03ec0816fe7f22fe5e3295c/src/main/java/com/johnlpage/mongosyphon/RDBMSConnection.java#L92 Hi, someone had the same issue with oracle. The implication is you need to find a way to avoid it pulling the whole result set for a large set. Will set that as mysql only though. On Fri, 22 Mar 2019, 17:59 André Spiegel, notifications@github.com wrote: I used MongoSyphon against SQL Server at a customer, and it turns out that setting the fetch size to a negative value, as in line 92 in RDBMSConnection.java, is a feature specific to the MySQL driver. It didn't work with the MSSQL 6.2.1 driver and I had to comment out that line, after which things seemed to be okay. But I don't know the full implications of that. Discussion on stackoverflow: https://stackoverflow.com/questions/25019410/confusion-with-setfetchsize-method-of-statement-object Link to the place in the code: https://github.com/johnlpage/MongoSyphon/blob/a66484c875b244c8e03ec0816fe7f22fe5e3295c/src/main/java/com/johnlpage/mongosyphon/RDBMSConnection.java#L92 — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/johnlpage/MongoSyphon/issues/11, or mute the thread https://github.com/notifications/unsubscribe-auth/AEPhVrfNVkfB29e_2DY-xUlN1MYTTRK0ks5vZVJSgaJpZM4cEfEK . Fixed - that only kicks in for MySQL now.
gharchive/issue
2019-03-22T21:59:45
2025-04-01T04:34:41.740373
{ "authors": [ "drmirror", "johnlpage" ], "repo": "johnlpage/MongoSyphon", "url": "https://github.com/johnlpage/MongoSyphon/issues/11", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
251961767
Failed to generate .clang-format First of all, excellent project! I'm trying to run on our project: esbmc but I get the following: $ python3 unformat/ --root /home/mramalho/esbmc-github/src/clang-c-frontend/ "/home/mramalho/esbmc-github/src/clang-c-frontend/**/*.h" "/home/mramalho/esbmc-github/src/clang-c-frontend/**/*.cpp" Using the configuration file from the provided root, '/home/mramalho/esbmc-github/src/clang-c-frontend/' 0: .Traceback (most recent call last): File "/usr/lib64/python3.5/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/usr/lib64/python3.5/runpy.py", line 85, in _run_code exec(code, run_globals) File "unformat/__main__.py", line 85, in <module> main(args, pool) File "unformat/__main__.py", line 52, in main (generation_fittest, population) = generate(population, source_filenames, args, pool) File "unformat/__main__.py", line 40, in generate return recombine(scored_population, args) File "unformat/recombine.py", line 91, in recombine range(args.population - 1)] File "unformat/recombine.py", line 90, in <listcomp> recombined_configs = [mutate(ranked[int(random() * random() * len(ranked))][1], args.mutation) for _ in File "unformat/recombine.py", line 79, in mutate return {key: visit_line(key, value, mutation_rate) for key, value in config.items()} AttributeError: 'NoneType' object has no attribute 'items' I'm actually setting the root one subdir of the project because that's the place with the best code style so far. If I run on the actual root dir (esbmc/src), it goes out of memory but then keeps running, eventually failing with the same error. Thanks @mikhailramalho, what versions of Python and clang-format are you using? Also, how much RAM and what CPU do you have? To avoid the OOM issue, you might try reducing the number of parallel processes with -j N. Hi, $ clang-format -version clang-format version 3.9.1 (tags/RELEASE_391/final) $ python3 --version Python 3.5.3 Processor: Intel® Xeon(R) CPU W3520 @ 2.67GHz × 8 Memory: 23.5 GiB Fedora 25 Hmm, you have plenty of RAM. Nevertheless, -j 4 should reduce the chance of OOM. I'll look into how easy it is to catch this error and fail more gracefully. I've observed this problem and reducing jobs to one doesn't help. It seems to be a problem with files from my cmake build directory (which got pulled in when I specified the root of my project). There's likely a lot of extreme code in the average build folder. Could this be the case with you? I think I've reproduced the same issue with 60 files approx: .Traceback (most recent call last): File "/home/manuel-sanchez/Python-3.5.4/lib/python3.5/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/manuel-sanchez/Python-3.5.4/lib/python3.5/runpy.py", line 85, in _run_code exec(code, run_globals) File "/home/manuel-sanchez/Documentos/unformat/__main__.py", line 85, in <module> main(args, pool) File "/home/manuel-sanchez/Documentos/unformat/__main__.py", line 52, in main (generation_fittest, population) = generate(population, source_filenames, args, pool) File "/home/manuel-sanchez/Documentos/unformat/__main__.py", line 40, in generate return recombine(scored_population, args) File "/home/manuel-sanchez/Documentos/unformat/recombine.py", line 91, in recombine range(args.population - 1)] File "/home/manuel-sanchez/Documentos/unformat/recombine.py", line 90, in <listcomp> recombined_configs = [mutate(ranked[int(random() * random() * len(ranked))][1], args.mutation) for _ in File "/home/manuel-sanchez/Documentos/unformat/recombine.py", line 79, in mutate return {key: visit_line(key, value, mutation_rate) for key, value in config.items()} AttributeError: 'NoneType' object has no attribute 'items' after looking at the backtrace and the code I would say my problem is that make_initial_configs() is returning an empty set as population, since I called unformat this way: python3.5 $HOME/Documentos/unformat --root . "<my-sourcedir>/**/*.h" "<my-sourcedir>/**/*.cpp" --command clang-format-3.9 with the root directory containing an empty .clang-format file. Maybe make_initial_configs() should be able to return a valid structured (but empty) set of args, or I should not be that lazy and feed the tool with some initial config. Yes, I'm that guy, leaving all the work to write a .clang-format matching his code guidelines to a tool :) That's a good catch, thanks. I'm elbows-deep into a branch that adds uncrustify support or I'd address that case right now. That branch already has the fix.
gharchive/issue
2017-08-22T13:43:17
2025-04-01T04:34:41.747302
{ "authors": [ "Manu343726", "johnmcfarlane", "mikhailramalho" ], "repo": "johnmcfarlane/unformat", "url": "https://github.com/johnmcfarlane/unformat/issues/4", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
86154901
WebStorm - Error: Invalid or corrupt jarfile I have just read [Style Y252] and downloaded jar file with the WebStorm templates. However, when I try importing it into my Webstorm it is not presented in files which are available to be imported. When I try opening the archive by doubleclick I get this issue: Error: Invalid or corrupt jarfile d:\Users\username\Desktop\webstorm-angular-file-template.settings.jar Tried to clone the repo also but got the same issue. WebStorm version: 10.0.1 same as https://github.com/johnpapa/angular-styleguide/issues/395 ? Is this issue fixable? I also have WebStorm version: 10.0.1 and was able to get it working. Not sure if this is the reason you are having difficulties, but you do not need to open the archive prior to importing; JAR files look like they need to be unzipped prior to installation but WebStorm should take care of this for you. happy to take a PR for webstorm. any takers? I'll take it :) Should also be covered by #537 @MAustinMMDP this should be closed now that you ave your PR for https://github.com/johnpapa/angular-styleguide/pull/544
gharchive/issue
2015-06-08T12:23:05
2025-04-01T04:34:41.753707
{ "authors": [ "Akarienta", "MAustinMMDP", "johnpapa", "joshtoo", "victorwoo" ], "repo": "johnpapa/angular-styleguide", "url": "https://github.com/johnpapa/angular-styleguide/issues/421", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
453850195
Latest broccoli-static-compiler uses the broccoli .read / .rebuild API, which is deprecated. [API] Warning: The .read and .rebuild APIs will stop working in the next Broccoli version [API] Warning: Use broccoli-plugin instead: https://github.com/broccolijs/broccoli-plugin [API] Warning: Plugin uses .read/.rebuild API: StaticCompiler StaticCompiler: The .read/.rebuild API is no longer supported as of Broccoli 1.0. Plugins must now derive from broccoli-plugin. https://github.com/broccolijs/broccoli/blob/master/docs/broccoli-1-0-plugin-api.md This package is deprecated in favor of using broccoli-funnel instead. The best path to fixing is to find out which packages are depending on this library and update them to use broccoli-funnel.
gharchive/issue
2019-06-09T00:59:36
2025-04-01T04:34:41.866769
{ "authors": [ "mileslane", "rwjblue" ], "repo": "joliss/broccoli-static-compiler", "url": "https://github.com/joliss/broccoli-static-compiler/issues/18", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1625353120
Myprofile reserved rockets In this milestone, I rendered a list of all reserved rockets using array filter on the "My profile" page. Thank you for your very encouraging review.
gharchive/pull-request
2023-03-15T11:38:30
2025-04-01T04:34:41.875582
{ "authors": [ "jonas-45" ], "repo": "jonas-45/space-travellers-hub", "url": "https://github.com/jonas-45/space-travellers-hub/pull/36", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
581940648
Is it possible to have this bot send a message at a specific time everyday? I couldn't find any way to do that. Yes there is. Use the custom command interval trigger. It works like a charm! Thanks a bunch!
gharchive/issue
2020-03-16T02:24:28
2025-04-01T04:34:41.881291
{ "authors": [ "GenryMg", "cmarabate" ], "repo": "jonas747/yagpdb", "url": "https://github.com/jonas747/yagpdb/issues/659", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1022393274
🛑 Nextcloud is down In 50c9fba, Nextcloud (https://nextcloud.jonasled.de) was down: HTTP code: 0 Response time: 0 ms Resolved: Nextcloud is back up in dc720ac.
gharchive/issue
2021-10-11T08:07:22
2025-04-01T04:34:41.892224
{ "authors": [ "jonasled" ], "repo": "jonasled/status", "url": "https://github.com/jonasled/status/issues/146", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
931022922
Push multiple routes dynamically It would be nice to be able to push multiple routes from a single path. A common use case might be that the app receives a dynamic link of /machine/1234/settings. Given that path, I would want to push 3 routes onto the main navigation stack. On bottom, a list of machines. On top of that, the show page for machine 123. And on top of that, the settings for that machine. That way, the user can hit back and "zoom out", as it were, to see the show page, and then back again to see the all machines. Maybe that is possible now, but I don't see it in the documentation. Thanks for the fantastic work! this is exactly what we need too. other packages like auto_route offer this capability. there is no need to add /machine/:id/:id2 simple create GetPage( name: '/machine', page: () => const Machine(), ), and Get.toName("/machine/$id/id2"); on /machine/1234/settings. parse current routes
gharchive/issue
2021-06-27T21:33:12
2025-04-01T04:34:41.896478
{ "authors": [ "SogoGolf", "steven-spiel", "xTanveerAhmad" ], "repo": "jonataslaw/getx", "url": "https://github.com/jonataslaw/getx/issues/1581", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
682499642
Docs updated for logWriterCallback option Added docs for the logging option @Nipodemos maybe you can update the -BR version too? @ajatsoft maybe you can update the -ES version too? @szymon-filipiak maybe you can update the -PL version too? i noticed that one of the best ways to truly understand a topic is to translate it. only after i translated that i've noticed that this logWriter does not respect Get.isLogEnable. Why? Because then you can control it from the logging package to decide to collect the log. Also then there is also one location in the library to check if logging is needed (see log.dart). oh i understand, thanks for the reply. @jonataslaw since the other guys will need to make a new PR to translate, i think this can be approved now i am wondering if we should put this docs in the other languages files the english version so that someone that will read the translated docs can at least acknowledge that there is a logWriterCallback. What do you think @stefandevo , i'm wondering this for ages Yes, I will try to do this night El 20 ago 2020, a las 10:10, Stefan de Vogelaere notifications@github.com escribió:  @ajatsoft maybe you can update the -ES version too? — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub, or unsubscribe. oh i understand, thanks for the reply. @jonataslaw since the other guys will need to make a new PR to translate, i think this can be approved now i am wondering if we should put this docs in the other languages files the english version so that someone that will read the translated docs can at least acknowledge that there is a logWriterCallback. What do you think @stefandevo , i'm wondering this for ages I agree with the approval of this PR, as well as the inclusion, although English in the other docs I'll merge, then we insert the English version in the other docs while the Spanish and Polish versions are not ready. I update the Spanish version with it in last pr
gharchive/pull-request
2020-08-20T07:44:29
2025-04-01T04:34:41.902673
{ "authors": [ "Nipodemos", "ajatsoft", "jonataslaw", "stefandevo" ], "repo": "jonataslaw/getx", "url": "https://github.com/jonataslaw/getx/pull/484", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
192122119
Use u64 instead of i64 for the type for recorded values This will make it easier to port my tests over, since those all use unsigned values, and value has to be non-negative anyway. This diff is mostly pretty mechanical; 95% of it is just making obvious changes to keep the compiler happy. If you'd prefer I could submit this (and future) PRs to a different (non-master) branch if you'd like to accumulate this stuff elsewhere, though I don't plan on ever making a PR that would leave master in a broken state. I'm happy to compare them to master. I wasn't planning on cutting another minor release any time soon, so preparing master for the eventual major release that this will be seems fine. My biggest worry about this change is the danger of u64 underflow when we subtract. I've pointed out one already above, but there may be others. Could you make a pass and just check that there aren't other places where we're subtracting from a u64 where we don't know that the result isn't negative? Sure, I'll check over it and try to convince myself that we won't underflow. My implementation used u64 and didn't have any issues, but you also have more functionality so there may well be things that don't play well. Yeah, I have this nagging sensation that there was some case where I couldn't figure out whether a particular number was allowed to be negative or not, and I just decided to do everything with i64 (obviously a very lazy solution :p). I completely agree that the right thing to do (as you've started here) is to move to u64 and figure out where, if anywhere, we need to watch for underflow. I haven't forgotten about this; just been busy. OK, I think that's everything -- let me know if there's anything else you'd like addressed in this PR. Looks good to me! Thanks for all the work you've put in :)
gharchive/pull-request
2016-11-28T21:29:44
2025-04-01T04:34:41.956809
{ "authors": [ "jonhoo", "marshallpierce" ], "repo": "jonhoo/hdrsample", "url": "https://github.com/jonhoo/hdrsample/pull/1", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1174538169
import { Fountain } from 'fountain' doesn't work I ran npm install fountain and then import { Fountain } from 'fountain' [plugin vite:dep-scan] Failed to resolve entry for package "fountain". The package may have incorrect main/module/exports specified in its package.json: Failed to resolve entry for package "fountain". The package may have incorrect main/module/exports specified in its package.json. node_modules/vite/dist/node/chunks/dep-9c153816.js:29537:10: 29537 │ throw new Error(`Failed to resolve entry for package "${id}". ` + ╵ ^ at packageEntryFailure (/Users/pj/software/petersir/sveltekit/node_modules/vite/dist/node/chunks/dep-9c153816.js:29537:11) at resolvePackageEntry (/Users/pj/software/petersir/sveltekit/node_modules/vite/dist/node/chunks/dep-9c153816.js:29533:9) at tryNodeResolve (/Users/pj/software/petersir/sveltekit/node_modules/vite/dist/node/chunks/dep-9c153816.js:29349:20) at Context.resolveId (/Users/pj/software/petersir/sveltekit/node_modules/vite/dist/node/chunks/dep-9c153816.js:29166:28) at Object.resolveId (/Users/pj/software/petersir/sveltekit/node_modules/vite/dist/node/chunks/dep-9c153816.js:38271:55) at processTicksAndRejections (node:internal/process/task_queues:96:5) at async resolve (/Users/pj/software/petersir/sveltekit/node_modules/vite/dist/node/chunks/dep-9c153816.js:38474:26) at async /Users/pj/software/petersir/sveltekit/node_modules/vite/dist/node/chunks/dep-9c153816.js:38612:34 at async callback (/Users/pj/software/petersir/sveltekit/node_modules/esbuild/lib/main.js:917:28) at async handleRequest (/Users/pj/software/petersir/sveltekit/node_modules/esbuild/lib/main.js:697:30) This error came from the "onResolve" callback registered here: I'm trying to import it to a browser-based SvelteKit JS project. I'm looking into this, but just as an update for you. I ran my tests against Node v16.14.0 and they all came out passing. In fact I did a bit of upgrading to the testing suite just because there were some deprecated methods in there and everything still appears to be working fine. Reading into this, this seems to be a fairly common issue in Vite and I did find references that indicated using older versions of Vite like 2.7x could help resolve this issue in the meantime. Ultimately, it does seem like its something missing in my package like perhaps the modules field as best I can tell. Overall, I am still fairly new to all of this so I'm going to keep digging into this and let you know once I get something solid to work with. I don't think the new version is registered with npm. Apologies, I just meant that I updated my testing suite earlier, which would have not gone to NPM; however, I have just published v0.1.3 and will hopefully be up shortly. The issue was npm i fountain.ts (not npm i fountain) and also importing fountain.ts (not fountain) import from 'fountain' is a totally different package. Yup! I just realized my README was wrong too. 🤦‍♀️ This is now updated.
gharchive/issue
2022-03-20T12:56:15
2025-04-01T04:34:41.994331
{ "authors": [ "jonnygreenwald", "pjebs" ], "repo": "jonnygreenwald/Fountain.ts", "url": "https://github.com/jonnygreenwald/Fountain.ts/issues/8", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
219319816
Refresh accessToken or tokenID Hi, Your plugin works well, thank you very much, however I want to refresh accessToken and tokenID when it expired, how can I do that? Thanks Hi! I really recommend you to read at least the JavaScript parts of the code. It's pretty straightforward and gets you an overview of what the library is capable to do. You can refresh the token by calling signInSilentlyPromise. Is there any way to actually return the refreshToken?
gharchive/issue
2017-04-04T17:08:58
2025-04-01T04:34:42.400610
{ "authors": [ "ianataffinity", "nhhoang", "stami" ], "repo": "joonhocho/react-native-google-sign-in", "url": "https://github.com/joonhocho/react-native-google-sign-in/issues/24", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2676714501
TT-80-njrc-Update-Backend-Profile-Backend It works, for now... Merge main and i will review it
gharchive/pull-request
2024-11-20T17:54:40
2025-04-01T04:34:42.419777
{ "authors": [ "AbdulTur", "noahjrc" ], "repo": "jordyob03/TripTailor", "url": "https://github.com/jordyob03/TripTailor/pull/44", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
614280864
Doc still references tactician This file seems outdated since Messenger replaced Tactician https://github.com/jorge07/symfony-5-es-cqrs-boilerplate/blob/7272cd84931d43cfcf4173c459ac826186be43b2/doc/GetStarted/Buses.md Totally right. I'll find time this weekend to add the messenger documentation. Thanks for report it!
gharchive/issue
2020-05-07T19:10:21
2025-04-01T04:34:42.420999
{ "authors": [ "jon-ht", "jorge07" ], "repo": "jorge07/symfony-5-es-cqrs-boilerplate", "url": "https://github.com/jorge07/symfony-5-es-cqrs-boilerplate/issues/147", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2520375339
Does not send a buffer as a message to a host Although the connection and disconnection orders from the client to the host do not cause problems, when I try to send a buffer in the payload to the host, it simply does nothing, I cannot send the data to the host. Using the TCP node included in node-red works without errors, but I am interested in being able to close and open connections. Ok. Just had to pass the buffer to sting msg.payload = <buffer>.toString() Sorry.
gharchive/issue
2024-09-11T17:45:26
2025-04-01T04:34:42.426043
{ "authors": [ "GusanoGris" ], "repo": "jorgen-k/node-red-contrib-tcp-client2", "url": "https://github.com/jorgen-k/node-red-contrib-tcp-client2/issues/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
207890192
style component pass style prop to #21 Will be updated in latest version - thanks
gharchive/issue
2017-02-15T18:37:32
2025-04-01T04:34:42.429324
{ "authors": [ "EricPKerr", "davidpaulsson", "sibelius" ], "repo": "jorilallo/react-native-emoji", "url": "https://github.com/jorilallo/react-native-emoji/issues/11", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1000451805
Using e:becomes Is there a simple example someplace of how e:becomes is intended to work? I had thought that it would be used like log:implies but that it would retract the subject graph in addition to adding the object graph. But I can’t seem to get that to work. Here's the specific rule I am working on: { ?dtd time:unitType ?unit . ?unit time:years ?duration . ?duration e:numeral ?years . ?years log:equalTo 100 } => { ?dtd time:unitType time:unitCentury } . I'd like to change this to use e:becomes so that the triples in the subject are retracted; i.e. the conclusion replaces the antecedent. (Also, please let me know if there is someplace else where you would prefer I ask questions like this.) No problem to ask it here :-) There is an example of the use of e:becomes in https://github.com/josd/eye/blob/master/reasoning/gps/gps-plugin.n3 and you can test it in the directory eye/reasoning/gps via $ eye --quiet --nope gps-plugin.n3 gps-example1.n3 --query gps-query1.n3 2>/dev/null PREFIX math: <http://www.w3.org/2000/10/swap/math#> PREFIX list: <http://www.w3.org/2000/10/swap/list#> PREFIX log: <http://www.w3.org/2000/10/swap/log#> PREFIX e: <http://eulersharp.sourceforge.net/2003/03swap/log-rules#> PREFIX gps: <http://josd.github.io/eye/reasoning/gps/gps-schema#> PREFIX : <http://josd.github.io/eye/reasoning#> :i1 gps:path ((:drive_gent_brugge :drive_brugge_oostende) 2400.0 0.01 0.9408 0.99). :i1 gps:path ((:drive_gent_kortrijk :drive_kortrijk_brugge :drive_brugge_oostende) 4100.0 0.018000000000000002 0.903168 0.9801). It is a built-in and it retracts the triples in the subject graph and asserts the triples in the object graph and it is intented to implement linear logic implication. See also https://scholar.google.be/citations?view_op=view_citation&hl=en&user=-vOppWIAAAAJ&sortby=pubdate&citation_for_view=-vOppWIAAAAJ:6_hjMsCP8ZoC Thank you. I can see that the GPS example uses e:becomes, but I have trouble understanding it because it also uses other builtins I'm not familiar with like e:call and e:finalize. I'm not sure if I also need these, or if they are only necessary for the GPS case. I wonder if you might post a minimal example that could help me see how I can use e:becomes for my case? A simple example is $ cat eb.n3 PREFIX e: <http://eulersharp.sourceforge.net/2003/03swap/log-rules#> PREFIX : <http://example.org/test#> :Socrates a :PoisonDrinker. {{?S a :PoisonDrinker} e:becomes {?S a :DeadPerson}} => {?S a ?B}. which should give $ eye --quiet --nope eb.n3 --pass 2>/dev/null PREFIX e: <http://eulersharp.sourceforge.net/2003/03swap/log-rules#> PREFIX : <http://example.org/test#> :Socrates a :DeadPerson. So :Socrates a :PoisonDrinker is retracted and becoming :Socrates a :DeadPerson Built-ins can not be retracted nor asserted hence can't be part of the subject or object of e:becomes. Thank you again! That is very helpful. Playing around with this example, I noticed something odd: if I use the --pass-only-new flag, the triples asserted in the object of e:becomes are not output: $ eye --quiet --nope eb.n3 --pass-only-new 2>/dev/null PREFIX e: <http://eulersharp.sourceforge.net/2003/03swap/log-rules#> PREFIX : <http://example.org/test#> Is this because :Socrates a :DeadPerson is “asserted” rather than “derived”? (It still seems like a “new” triple, though.) I’m also a bit confused by this behavior: The following produces :Socrates a :DeadPerson as expected: PREFIX : <http://example.org/test#> PREFIX e: <http://eulersharp.sourceforge.net/2003/03swap/log-rules#> PREFIX log: <http://www.w3.org/2000/10/swap/log#> PREFIX xsd: <http://www.w3.org/2001/XMLSchema#> :Socrates a :PoisonDrinker . :Socrates :drinksPoisonAmount "10"^^xsd:decimal . {{?S a :PoisonDrinker ; :drinksPoisonAmount "10"^^xsd:decimal } e:becomes {?S a :DeadPerson}} => {?S a ?B} . But if the datatyped value is produced via the log:dtlit builtin: PREFIX : <http://example.org/test#> PREFIX e: <http://eulersharp.sourceforge.net/2003/03swap/log-rules#> PREFIX log: <http://www.w3.org/2000/10/swap/log#> PREFIX xsd: <http://www.w3.org/2001/XMLSchema#> :Socrates a :PoisonDrinker . { ("10" xsd:decimal) log:dtlit ?amount } => { :Socrates :drinksPoisonAmount ?amount } . {{?S a :PoisonDrinker ; :drinksPoisonAmount "10"^^xsd:decimal } e:becomes {?S a :DeadPerson}} => {?S a ?B} . … then the result is: :Socrates a :PoisonDrinker. :Socrates :drinksPoisonAmount "10"^^xsd:decimal. Is this also a case of “built-ins cannot be retracted,” because the datatyped literal in the subject was created via a builtin? Playing around with this example, I noticed something odd: if I use the --pass-only-new flag, the triples asserted in the object of e:becomes are not output: $ eye --quiet --nope eb.n3 --pass-only-new 2>/dev/null PREFIX e: <http://eulersharp.sourceforge.net/2003/03swap/log-rules#> PREFIX : <http://example.org/test#> Is this because :Socrates a :DeadPerson is “asserted” rather than “derived”? (It still seems like a “new” triple, though.) Thanks for the observation and this is now corrected in https://github.com/josd/eye/releases/tag/v21.0920.2018 ... But if the datatyped value is produced via the log:dtlit builtin: PREFIX : <http://example.org/test#> PREFIX e: <http://eulersharp.sourceforge.net/2003/03swap/log-rules#> PREFIX log: <http://www.w3.org/2000/10/swap/log#> PREFIX xsd: <http://www.w3.org/2001/XMLSchema#> :Socrates a :PoisonDrinker . { ("10" xsd:decimal) log:dtlit ?amount } => { :Socrates :drinksPoisonAmount ?amount } . {{?S a :PoisonDrinker ; :drinksPoisonAmount "10"^^xsd:decimal } e:becomes {?S a :DeadPerson}} => {?S a ?B} . … then the result is: :Socrates a :PoisonDrinker. :Socrates :drinksPoisonAmount "10"^^xsd:decimal. Is this also a case of “built-ins cannot be retracted,” because the datatyped literal in the subject was created via a builtin? It is actually retracted but a bit later again asserted by the rule { ("10" xsd:decimal) log:dtlit ?amount } => { :Socrates :drinksPoisonAmount ?amount } . which is the intrinsic behaviour of the eye forward chaining (rules are repeatedly running as long as new conclusions follow). It is actually retracted but a bit later again asserted I understand why :Socrates :drinksPoisonAmount "10"^^xsd:decimal is in the conclusion; what I don’t understand is why :Socrates a :PoisonDrinker is in the conclusion (rather than :Socrates a :DeadPerson). ... then with --pass I get only :Socrates a :DeadPerson (as expected), but with --pass-only-new I get :Socrates a :PoisonDrinker. :Socrates a :DeadPerson. Is this intended? :Socrates a :PoisonDrinker was derived but then retracted, so should it be “new”? As we said before, it is indeed retracted but asserted again and the --pass-only-new is rather primitive in that it outputs each time something is asserted - it is actually more something like a debugging thing ;-) (For context, I am trying to determine the best way to produce a file containing only derived triples but no retracted ones. Maybe I should just use --pass and then post-process with SPARQL to remove the non-derived (i.e. input) triples?) Maybe that is indeed better. --pass-only-new is rather primitive in that it outputs each time something is asserted OK, that’s what I figured—so I will find another solution. Thanks for your quick replies and fixes, I really appreciate it! With the latest eye https://github.com/josd/eye/releases/tag/v21.0921.1204 --pass-only-new should now produce only new triples: $ cat eb3.n3 PREFIX e: <http://eulersharp.sourceforge.net/2003/03swap/log-rules#> PREFIX : <http://example.org/test#> {{?S :drinks :poison ; a :PoisonDrinker } e:becomes {?S a :DeadPerson}} => {?S a ?B}. {?S :drinks :poison} => {?S a :PoisonDrinker}. :Socrates :drinks :poison. $ eye --nope eb3.n3 --pass-only-new eye --nope eb3.n3 --pass-only-new EYE v21.0921.1204 josd SWI-Prolog version 8.3.29-32-ga624386fa starting 32 [msec cputime] 29 [msec walltime] #Processed by EYE v21.0921.1204 josd #eye --nope eb3.n3 --pass-only-new GET file:///tmp/eb3.n3 SC=3 networking 1 [msec cputime] 1 [msec walltime] PREFIX e: <http://eulersharp.sourceforge.net/2003/03swap/log-rules#> PREFIX : <http://example.org/test#> :Socrates a :DeadPerson. reasoning 1 [msec cputime] 0 [msec walltime] #2021-09-21T12:12:04.711Z in=3 out=1 ent=1 step=2 brake=2 inf=18175 sec=0.034 inf/sec=534559 #ENDS 2021-09-21T12:12:04.711Z in=3 out=1 ent=1 step=2 brake=2 inf=18175 sec=0.034 inf/sec=534559 Thanks, that makes sense. I was looking at your paper “Predicting future state for adaptive clinical pathway management” and it seems very useful to be able to clearly separate 1) conditions for applying a transition rule, 2) statements to be retracted, and 3) statements to be asserted. I think part of the conceptual difficulty I was having with using e:becomes was confusion between 1) and 2). Another thing that still confuses me: Rules using e:becomes take the form {?X e:becomes ?Y} => ?Z. ?X is the graph of statements to be retracted and ?Y is the graph of statements to be asserted. But then what is ?Z? Thanks for the nice 1)2)3) summary. e:becomes is a built-in and buit-ins can only be used in the premis of a rule, so we need a rule and the conclusion ?Z is nothing specific and was just for demonstration purposes. So you can also write a rule like {{?S a :PoisonDrinker ; :drinksPoisonAmount "10"^^xsd:decimal} e:becomes {?S a :DeadPerson}} => {}. or {{?S :drinks :poison ; a :PoisonDrinker } e:becomes {?S a :DeadPerson}} => {{?S :drinks :poison ; a :PoisonDrinker } :became {?S a :DeadPerson}}. OK, I see. I tried changing ?Z to different things and it didn’t seem to have any effect, so it’s good to understand why. Thanks again! So your 1) conditions for applying a transition rule, 2) statements to be retracted, and 3) statements to be asserted all occur in the premis of rules like {Condition-triples. {From-triples} e:becomes {To-triples}} => {Conclusion-triples}. OK, I see. I tried changing ?Z to different things and it didn’t seem to have any effect, so it’s good to understand why. Thanks again! Well, technically speaking, the ?Z is concluded and it will be asserted when it is not yet known. So in the original example ?S a ?B will be concluded as :Socrates a _:sk_0 which means Socrates is something but that is already known as :Socrates a :DeadPerson With a rule like {{?S :drinks :poison ; a :PoisonDrinker } e:becomes {?S a :DeadPerson}} => {{?S :drinks :poison ; a :PoisonDrinker } :became {?S a :DeadPerson}}. you will see an additional triple concluded :Socrates a :DeadPerson. {:Socrates :drinks :poison. :Socrates a :PoisonDrinker} :became {:Socrates a :DeadPerson}. the ?Z is concluded and it will be asserted when it is not yet known Ah, I see. I guess in all the examples I tried it was already known, and so it just seemed to have no effect.
gharchive/issue
2021-09-19T23:41:29
2025-04-01T04:34:42.454749
{ "authors": [ "josd", "rybesh" ], "repo": "josd/eye", "url": "https://github.com/josd/eye/issues/22", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
49910489
Why are you uploading a new file before removing the ones who should be deleted ? Hi there, First, thanks for your work, you save us hours and hours ! Before I go through the pull request process, I was wondering what design decision made you unlink files to be removed after having handled the current file upload. Are there any reasons why you didn't call __unlinkFiles on top of the behavior's afterSave function ? Without removing files beforehand, it is impossible to replace a file with a same name. I've seen that you've been adding the ability to create a callback to handle the file upload manually, but to me, replacing an existing file with a new one with the same name is valid enough a use case to include it in the Behavior. Resulting code would look like the following <?php /** * After save method. Called after all saves * * Handles moving file uploads * * @param Model $model Model instance * @param bool $created True if this save created a new record * @param array $options Options passed from Model::save(). * @return bool * @throws UploadException */ public function afterSave(Model $model, $created, $options = array()) { $temp = array($model->alias => array()); $this->_unlinkFiles($model); foreach ($this->settings[$model->alias] as $field => $options) { if ($this->_shouldSkip($model, $field)) { continue; } $tempPath = $this->_getPath($model, $field); $path = $this->settings[$model->alias][$field]['path']; $thumbnailPath = $this->settings[$model->alias][$field]['thumbnailPath']; if (!empty($tempPath)) { $path .= $tempPath . DS; $thumbnailPath .= $tempPath . DS; } $tmp = $this->runtime[$model->alias][$field]['tmp_name']; $filePath = $path . $model->data[$model->alias][$field]; if (!$this->handleUploadedFile($model, $field, $tmp, $filePath)) { CakeLog::error(sprintf('Model %s, Field %s: Unable to move the uploaded file to %s', $model->alias, $field, $filePath)); $model->invalidate($field, sprintf('Unable to move the uploaded file to %s', $filePath)); $db = $model->getDataSource(); $db->rollback(); throw new UploadException('Unable to upload file'); } $this->_createThumbnails($model, $field, $path, $thumbnailPath); if ($model->hasField($options['fields']['dir'])) { if (!($created && $options['pathMethod'] == '_getPathFlat') && $options['saveDir']) { $db = $model->getDataSource(); $temp[$model->alias][$options['fields']['dir']] = $db->value($tempPath, 'string'); } } } $this->_updateRecord($model, $temp); return; } What do you think of it ? Hi, I have the same problem. When I want to replace a file with the same name, the upload fails. Can you fix it? Or tell me how should I do? Thanks a lot, and good work. Relates to #191 Closing as dupe.
gharchive/issue
2014-11-24T16:27:23
2025-04-01T04:34:42.523487
{ "authors": [ "CaptainQuirk", "davidyell", "josegonzalez", "vnt" ], "repo": "josegonzalez/cakephp-upload", "url": "https://github.com/josegonzalez/cakephp-upload/issues/265", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
250126252
Add Support for Secret Injection from Drone Secrets, and a couple other things This PR allows developers to inject arbitrary secrets from their Drone Secrets into their Container. It also adds support for Docker labels, setting a Log Driver and options, allowing override of the defaults. All these new features are documented and usage shown in DOCS.md. Closing this PR since all these chances are in the latest version.
gharchive/pull-request
2017-08-14T19:16:02
2025-04-01T04:34:42.597980
{ "authors": [ "josmo", "one000mph" ], "repo": "josmo/drone-ecs", "url": "https://github.com/josmo/drone-ecs/pull/16", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
34251983
Path to bash executable Hi, Using xlogin as Debian service requires replacement /usr/bin/bash to /bin/bash in three places or symbolic link the first to the second path. This is fixed in ed76919ae72bcd87efb8206625a197da2c428063. You can set the location of Bash when installing using make BASH=/bin/bash install.
gharchive/issue
2014-05-25T01:49:10
2025-04-01T04:34:42.603479
{ "authors": [ "joukewitteveen", "paprykarz" ], "repo": "joukewitteveen/xlogin", "url": "https://github.com/joukewitteveen/xlogin/issues/2", "license": "bsd-2-clause", "license_type": "permissive", "license_source": "bigquery" }
1760099862
CUDA error: device-side assert triggered I have the same error, when I use the command on coco dataset, as "GPUS_PER_NODE=4 ./tools/run_dist_launch.sh 4 ./configs/deta_swin_ft.sh --coco_path /mnt/home/dataset/coco --finetune /mnt/home/DETA/adet_swin_pt_o365.pth" or " ./configs/deta.sh --eval --coco_path ./data/coco --resume ./adet_checkpoint0011.pth". My environment is Pytorch=1.8.1 Cuda=11.1, and I train well on Deformable-DETR whithout this error. The detail of error is as follow: Test: [ 0/2500] eta: 1:19:14 class_error: 0.00 loss: 14.3390 (14.3390) loss_ce: 0.6692 (0.6692) loss_bbox: 0.2385 (0.2385) loss_giou: 0.8719 (0.8719) loss_ce_0: 0.7682 (0.7682) loss_bbox_0: 0.2413 (0.2413) loss_giou_0: 0.8721 (0.8721) loss_ce_1: 0.7386 (0.7386) loss_bbox_1: 0.2372 (0.2372) loss_giou_1: 0.8720 (0.8720) loss_ce_2: 0.7082 (0.7082) loss_bbox_2: 0.2383 (0.2383) loss_giou_2: 0.8715 (0.8715) loss_ce_3: 0.6925 (0.6925) loss_bbox_3: 0.2384 (0.2384) loss_giou_3: 0.8715 (0.8715) loss_ce_4: 0.6827 (0.6827) loss_bbox_4: 0.2385 (0.2385) loss_giou_4: 0.8716 (0.8716) loss_ce_enc: 1.1769 (1.1769) loss_bbox_enc: 0.4718 (0.4718) loss_giou_enc: 1.7679 (1.7679) loss_ce_unscaled: 0.6692 (0.6692) class_error_unscaled: 0.0000 (0.0000) loss_bbox_unscaled: 0.0477 (0.0477) loss_giou_unscaled: 0.4359 (0.4359) cardinality_error_unscaled: 889.5000 (889.5000) loss_ce_0_unscaled: 0.7682 (0.7682) loss_bbox_0_unscaled: 0.0483 (0.0483) loss_giou_0_unscaled: 0.4361 (0.4361) cardinality_error_0_unscaled: 886.5000 (886.5000) loss_ce_1_unscaled: 0.7386 (0.7386) loss_bbox_1_unscaled: 0.0474 (0.0474) loss_giou_1_unscaled: 0.4360 (0.4360) cardinality_error_1_unscaled: 889.5000 (889.5000) loss_ce_2_unscaled: 0.7082 (0.7082) loss_bbox_2_unscaled: 0.0477 (0.0477) loss_giou_2_unscaled: 0.4358 (0.4358) cardinality_error_2_unscaled: 889.5000 (889.5000) loss_ce_3_unscaled: 0.6925 (0.6925) loss_bbox_3_unscaled: 0.0477 (0.0477) loss_giou_3_unscaled: 0.4358 (0.4358) cardinality_error_3_unscaled: 889.5000 (889.5000) loss_ce_4_unscaled: 0.6827 (0.6827) loss_bbox_4_unscaled: 0.0477 (0.0477) loss_giou_4_unscaled: 0.4358 (0.4358) cardinality_error_4_unscaled: 889.5000 (889.5000) loss_ce_enc_unscaled: 1.1769 (1.1769) loss_bbox_enc_unscaled: 0.0944 (0.0944) loss_giou_enc_unscaled: 0.8839 (0.8839) cardinality_error_enc_unscaled: 22179.5000 (22179.5000) time: 1.9019 data: 0.6984 max mem: 1327 /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [96,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [97,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [98,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [99,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [100,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [101,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [102,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [103,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [104,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [105,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [106,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [107,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [108,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [109,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [110,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [111,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [112,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [113,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [114,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [115,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [116,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [117,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [118,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [119,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [120,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [121,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [122,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [123,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [124,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [125,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [126,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [127,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [32,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [33,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [34,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [35,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [36,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [37,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [38,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [39,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [40,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [41,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [42,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [43,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [44,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [45,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [46,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [47,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [48,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [49,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [50,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [51,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [52,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [53,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [54,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [55,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [56,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [57,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [58,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [59,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [60,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [61,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [62,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [63,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [0,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [1,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [2,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [3,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [4,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [5,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [6,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [7,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [8,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [9,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [10,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [11,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [12,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [13,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [14,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [15,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [16,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [17,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [18,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [19,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [20,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [21,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [22,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [23,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [24,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [25,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [26,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [27,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [28,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [29,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [30,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [31,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [64,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [65,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [66,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [67,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [68,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [69,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [70,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [71,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [72,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [73,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [74,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [75,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [76,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [77,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [78,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [79,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [80,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [81,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [82,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [83,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [84,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [85,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [86,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [87,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [88,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [89,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [90,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [91,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [92,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [93,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [94,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. /opt/conda/conda-bld/pytorch_1616554793803/work/aten/src/ATen/native/cuda/IndexKernel.cu:142: operator(): block: [0,0,0], thread: [95,0,0] Assertion index >= -sizes[i] && index < sizes[i] && "index out of bounds" failed. Traceback (most recent call last): File "main.py", line 346, in main(args) File "main.py", line 284, in main test_stats, coco_evaluator = evaluate(model, criterion, postprocessors, File "/opt/conda/lib/python3.8/site-packages/torch/autograd/grad_mode.py", line 27, in decorate_context return func(*args, **kwargs) File "/mnt/home/DETA/engine.py", line 110, in evaluate loss_dict = criterion(outputs, targets) File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 889, in _call_impl result = self.forward(*input, **kwargs) File "/mnt/home/DETA/models/deformable_detr.py", line 398, in forward indices = self.stg1_assigner(enc_outputs, bin_targets) File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 889, in _call_impl result = self.forward(*input, **kwargs) File "/mnt/home/DETA/models/assigner.py", line 328, in forward pos_pr_inds = all_pr_inds[matched_labels == 1] RuntimeError: CUDA error: device-side assert triggered Hi, This is indeed strange if you can run Deformable-DETR just fine. I'd first try running Deformable-DETR from this repo (configs/improved_baseline.sh). If that works, it might be an NMS version issue.
gharchive/issue
2023-06-16T07:34:31
2025-04-01T04:34:42.773702
{ "authors": [ "jjdbear", "jozhang97" ], "repo": "jozhang97/DETA", "url": "https://github.com/jozhang97/DETA/issues/22", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
936449774
Release your work on probe ? Hi, Can you release the helm chart with all the work you do aroun all the probe ? Im' currently facing some problems with the readinessProbe, which is a little bit to low... Thank you and for the work you. Hi, The customization of the probe has been released ! V2.1.5 available on the repo
gharchive/issue
2021-07-04T12:43:30
2025-04-01T04:34:42.777110
{ "authors": [ "jp-gouin", "thierrystiegler" ], "repo": "jp-gouin/helm-openldap", "url": "https://github.com/jp-gouin/helm-openldap/issues/34", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
55823913
Don't import StringIO or smart_text from rest_framework.compat. Fixes #3. I had to make a small change to tox.ini to get it to run. Should that be updated to 3.0.4? @brandoncazander thanks! I'll bump version up and publish this in a couple of minutes. I'll update the supported versions and tox againce once 3.1.0 is out Awesome, thanks @jpadilla! I'm consistently impressed by how fast you move. @brandoncazander heh, just released on PyPI and tagged v1.0.1. Thanks again!
gharchive/pull-request
2015-01-28T22:13:58
2025-04-01T04:34:42.779536
{ "authors": [ "brandoncazander", "jpadilla" ], "repo": "jpadilla/django-rest-framework-xml", "url": "https://github.com/jpadilla/django-rest-framework-xml/pull/4", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
606971616
authorize assertion removed from Ember Simple Auth 2.1.1. According to the ember-simple-auth repo the authorize assertion should not be used any more: https://github.com/simplabs/ember-simple-auth/pull/2010 The authorize assertian has been removed from Ember Simple Auth 2.1.1. This breaks the following code: https://github.com/jpadilla/ember-simple-auth-token/blob/facc3bea47196abc553d0248e3aca8c43254440d/addon/mixins/token-authorizer.js#L39-L48 I'll work on a PR to resolve this if is hasn't been fixed already and I'm just looking in the wrong place. @puremath Good point. I have been doing this manually without token-authorizer, but that isn't ideal. However, I'm not sure that we want to change this because we want to maintain backwards compatibility. Therefore, I have created a new token-adapter mixin to service this purpose. @fenichelar Good point, the new adapter is definitely the better way forward. I was also worried about breaking backwards compatiblity but wasn't sure how to fix this. My own fix looks pretty much like yours. There was one point I wasn't sure about, though. Maybe you can explain to me why this works: If the token is refreshed, the isAuthorized-Property of the session service doesn't change. That's why I don't understand how the Authorization header (i.e. the headers computed property) is updated in this case since the computed property doesn't have to be reevaluated. Thanks for your help. @puremath You are correct. Below is what I use in my applications. headers: computed('session.data.authenticated', function() { if (this.session.isAuthenticated) { return { Authorization: `Bearer ${this.session.data.authenticated.token}`, }; } else { return {}; } }), Computed should depend on session.data.authenticated not session.isAuthenticated. I don't use the variables for header name, prefix, and token property name. I also don't use this.get( because I am using a version of ember that does not require it. But here I think we should use this.get(. I have made the correction to this PR. Thank you for catching this. I ended up creating a factory function to create the computed property to make it depend on the tokenPropertyName: https://github.com/puremath/ember-simple-auth-token/blob/1e217e216cbf92629f1426b62db423bfa444e030/addon/mixins/token-authorizer.js#L52 that way I don't have to depend on the whole session.data.authenticated Object, but it makes the code more complex. I'm not sure whether this is worth it.
gharchive/issue
2020-04-26T09:58:38
2025-04-01T04:34:42.785856
{ "authors": [ "fenichelar", "puremath" ], "repo": "jpadilla/ember-simple-auth-token", "url": "https://github.com/jpadilla/ember-simple-auth-token/issues/273", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1889310525
🛑 Bikaner Electrical is down In 62d5bf6, Bikaner Electrical (bikanerelectrical.com) was down: HTTP code: 0 Response time: 0 ms Resolved: Bikaner Electrical is back up in 6b4fc5c after 6 minutes.
gharchive/issue
2023-09-10T20:17:38
2025-04-01T04:34:42.845265
{ "authors": [ "jpleasants" ], "repo": "jpleasants/plesk2", "url": "https://github.com/jpleasants/plesk2/issues/102", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2433930648
🛑 Tara Gems Galore is down In c19d1d2, Tara Gems Galore (https://taragemsgalorellc.com) was down: HTTP code: 0 Response time: 0 ms Resolved: Tara Gems Galore is back up in d235563 after 7 minutes.
gharchive/issue
2024-07-28T13:23:16
2025-04-01T04:34:42.847929
{ "authors": [ "jpleasants" ], "repo": "jpleasants/plesk9", "url": "https://github.com/jpleasants/plesk9/issues/1455", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1974718437
App Header Pattern publish to site and examples ### Tasks - [ ] Publish pattern doc to site from #2356 - [ ] Add examples to storybook - [ ] Migration QA to StackOverflow 17 Nov - Started dev work, design page + examples 22 Nov - PR dev review ready for example matching UITK 24 Nov - PR ready, page design needs discussion with @ivan-calderon
gharchive/issue
2023-11-02T17:18:51
2025-04-01T04:34:42.859065
{ "authors": [ "origami-z" ], "repo": "jpmorganchase/salt-ds", "url": "https://github.com/jpmorganchase/salt-ds/issues/2669", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
227171197
Cannot install plugin in ElasticSearch (elasticsearch-5.3.2) Hi, When I try to install the .ZIP on your project into elasticsearch-5.3.2, I get this error: ./elasticsearch-plugin install file:///apps/elasticsearch-5.3.2/bin/elasticsearch-arrayformat-1.4.0.0-plugin.zip -> Downloading file:///apps/elasticsearch-5.3.2/bin/elasticsearch-arrayformat-1.4.0.0-plugin.zip [=================================================] 100%   ERROR: elasticsearch directory is missing in the plugin zip I think that this is because site plugins are no longer witking with ES 5+. So how can I install your plugin on ES 5.3.2? Thanks, Jim Hi, I was able to get the plugin installed. I had to install on an older ES 2.3.5 instance and I had install it "manually" by creating a dir under the /plugins/arrayformat. In that arrayformat dir, I put the jar from the ZIP and I manually made a plugin-descriptor.properties, configured for "site=true". However, searches do not seem to work. For example, this search works: curl -XGET '192.168.0.142:9200/bank/account/_search?q=account_number:25&pretty' { "took" : 6, "timed_out" : false, "_shards" : { "total" : 5, "successful" : 5, "failed" : 0 }, "hits" : { "total" : 1, "max_score" : 5.5900564, "hits" : [ { "_index" : "bank", "_type" : "account", "_id" : "25", "_score" : 5.5900564, "_source" : { "account_number" : 25, "balance" : 40540, "firstname" : "Virginia", "lastname" : "Ayala", "age" : 39, "gender" : "F", "address" : "171 Putnam Avenue", "employer" : "Filodyne", "email" : "virginiaayala@filodyne.com", "city" : "Nicholson", "state" : "PA" } } ] } } but if I use arrayformat, I get: curl -XGET '192.168.0.142:9200/bank/account/_search_arrayformat?q=account_number:25&pretty' { "_index" : "bank", "_type" : "account", "_id" : "_search_arrayformat", "found" : false } What is wrong with this? Why isn't it showing the results? This plugin has been upgraded to ElasticSearch 7.16 and the repo with the update is here: https://github.com/ramsoft-inc/ElasticSearch-Array-Plugin
gharchive/issue
2017-05-08T20:53:43
2025-04-01T04:34:42.870142
{ "authors": [ "ohaya", "s2ramana" ], "repo": "jprante/elasticsearch-arrayformat", "url": "https://github.com/jprante/elasticsearch-arrayformat/issues/5", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
121490787
Array field with MySQL Hi, I have a problem with a MySQL query and a field typed as an array. My mapping is : "partner_teams_ids" : { "type": "long", "index": "not_analyzed", "include_in_all": false }, I want a subquery result to be "translated" to an array e.g. : SELECT x.a, (SELECT group_concat(y.b) from y where y.ext_id = x.id) as b from x; With MySQL I have to use the group_concat() but the field is then a String and the import fails to convert the String ("1,2,3,4,5)" to an array of Long. Any idea how to solve this problem ? Thanks ! Xavier I do not support MySQL with JDBC importer, group_concat is defined to return a string. ES will index longs for you as you defined in the mapping, even if you pass strings, so there is not a big problem. In most cases, you don't need constructs like group_concat, as JDBC importer is able to handle multi values for you. Hi, I think that it's not possible because : MySQL doesn't support subquery to return multi-values When I use a group_concat() and return a String, with comma or not, there is an error during the import: message [MapperParsingException[failed to parse [subjective_contact_ids]]; nested: NumberFormatException[For input string: "0 0 0"]; ] MapperParsingException[failed to parse [subjective_contact_ids]]; nested: NumberFormatException[For input string: "632021842, 0, 0"]; ] Taking your example, you can just do SELECT x.a as _id, y.b as b from x, y where x.id = y.ext_id; for JDBC importer to create a numeric array b, so there is no need for something like group_concat BTW MySQL can do subqueries http://dev.mysql.com/doc/refman/5.7/en/subqueries.html but that does not matter here. Whatever is the query format or driver, MySQL doesn't support to return multi-values for one field. One field => One value. That's why I have to use the group_concat(). Nevermind, I wrote my own importer and split the string when I need to. What you can do is to add field params to inform your code that a field must split the input and convert to another format. (Look at Sphinx configuration : http://sphinxsearch.com/docs/latest/conf-sql-attr-multi.html) JDBC importer supports multi-value select for all databases automatically, out of the box. The values can be string, integer, long, boolean etc. This step is called row merging https://github.com/jprante/elasticsearch-jdbc/blob/master/src/main/java/org/xbib/elasticsearch/jdbc/strategy/standard/StandardSource.java#L781-L814 and an innovation of JDBC importer to overcome JDBC deficiencies. No need to instruct a database to return groups. lists, or something. 这是来自QQ邮箱的假期自动回复邮件。您好,我最近正在休假中,无法亲自回复您的邮件。我将在假期结束后,尽快给您回复。
gharchive/issue
2015-12-10T13:58:28
2025-04-01T04:34:42.879040
{ "authors": [ "jprante", "wanghc85", "xavierfacq" ], "repo": "jprante/elasticsearch-jdbc", "url": "https://github.com/jprante/elasticsearch-jdbc/issues/720", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
859811929
Shouldn't EmployeeID references to ReportsTo? EmployeeID referring to EmployeeID seems weird. I tried it works, but shouldn't the foreign key be set to ReportsTo column? https://github.com/jpwhite3/northwind-SQLite3/blob/591cd3253c327b1eed7155c1fec57464565c0932/Northwind.Sqlite3.create.sql#L279 I checked the Microsoft SQL samples for this table and they also have it referred to ReportsTo. Have a look at microsoft/sql-server-samples. Correct foreignkeyconstraint should be: FOREIGN KEY ([ReportsTo]) REFERENCES [Employees] ([EmployeeID]) ON DELETE NO ACTION ON UPDATE NO ACTION Thanks for reporting the issue. Script has been updated.
gharchive/issue
2021-04-16T13:23:27
2025-04-01T04:34:42.885125
{ "authors": [ "jpwhite3", "metinsenturk" ], "repo": "jpwhite3/northwind-SQLite3", "url": "https://github.com/jpwhite3/northwind-SQLite3/issues/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1689512494
Python 3.11 compatibility We are currently using a private C-API of Python, which has been removed in Python 3.11. Namely the function _PyObject_GC_Malloc is now gone. We use it in native/python/pyjp_value.cpp: In function ‘PyObject* PyJPValue_alloc(PyTypeObject*, Py_ssize_t)’ to allocate a Python object with a some extra space for the wrapped Java type. There is a detailed discussion of the possible replacements/workarounds: https://discuss.python.org/t/equivalent-of-pyobject-gc-malloc-in-python-3-11/25919 However the problem there is more related to a complicated class hierarchy. But it could give us some hints. sorry, my bad. This was already fixed. Yes I did address this one, but I have to make another attempt in 3.12 when then add the replacement API.
gharchive/issue
2023-04-29T10:29:48
2025-04-01T04:34:42.888694
{ "authors": [ "Thrameos", "marscher" ], "repo": "jpype-project/jpype", "url": "https://github.com/jpype-project/jpype/issues/1129", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
3307174
LoadError: ... signer information does not match ... Hey, i have the latest gems installed and i'm getting a security exception saying that the signer information does not match. Here's my code / environment and the error. Is this a problem with your adapter jar? require 'rubygems' require 'arjdbc' ActiveRecord::Base.establish_connection( :adapter=> "jdbcmysql", :host => "localhost", :database=> "xxxx", :username => "xxxx", :password => "xxxx" ) I'm running Os-X 10.6.8, jruby 1.6.5 (ruby-1.8.7-p330) (2011-10-25 9dcd388) (Java HotSpot(TM) 64-Bit Server VM 1.6.0_29) [darwin-x86_64-java] *** LOCAL GEMS *** activemodel (3.2.1) activerecord (3.2.1) activerecord-jdbc-adapter (1.2.2) activerecord-jdbcmysql-adapter (1.2.2) activesupport (3.2.1) arel (3.0.0) builder (3.0.0) i18n (0.6.0) jdbc-mysql (5.1.13) multi_json (1.1.0) rake (0.9.2.2, 0.8.7) sources (0.0.1) tzinfo (0.3.31) java.lang.SecurityException: class "arjdbc.jdbc.RubyJdbcConnection$s$1$0$select_p"'s signer information does not match signer information of other classes in the same package at java.lang.ClassLoader.checkCerts(ClassLoader.java:806) at java.lang.ClassLoader.preDefineClass(ClassLoader.java:487) at java.lang.ClassLoader.defineClassCond(ClassLoader.java:625) at java.lang.ClassLoader.defineClass(ClassLoader.java:615) at org.jruby.util.JRubyClassLoader.defineClass(JRubyClassLoader.java:76) at org.jruby.internal.runtime.methods.InvocationMethodFactory.endClass(InvocationMethodFactory.java:1378) at org.jruby.internal.runtime.methods.InvocationMethodFactory.getAnnotatedMethodClass(InvocationMethodFactory.java:792) at org.jruby.internal.runtime.methods.InvocationMethodFactory.getAnnotatedMethod(InvocationMethodFactory.java:810) at org.jruby.RubyModule.defineAnnotatedMethod(RubyModule.java:711) at org.jruby.RubyModule.defineAnnotatedMethod(RubyModule.java:679) at org.jruby.anno.TypePopulator$DefaultTypePopulator.populate(TypePopulator.java:68) at org.jruby.RubyModule.defineAnnotatedMethodsIndividually(RubyModule.java:673) at org.jruby.RubyModule.defineAnnotatedMethods(RubyModule.java:561) at arjdbc.jdbc.RubyJdbcConnection.createJdbcConnectionClass(RubyJdbcConnection.java:90) at arjdbc.jdbc.AdapterJavaService.basicLoad(AdapterJavaService.java:53) at org.jruby.Ruby.loadExtension(Ruby.java:2586) at org.jruby.runtime.load.ClassExtensionLibrary.load(ClassExtensionLibrary.java:55) at org.jruby.runtime.load.LoadService.tryLoadingLibraryOrScript(LoadService.java:767) at org.jruby.runtime.load.LoadService.smartLoad(LoadService.java:334) at org.jruby.runtime.load.LoadService.require(LoadService.java:381) at org.jruby.runtime.load.LoadService.lockAndRequire(LoadService.java:306) at org.jruby.RubyKernel.requireCommon(RubyKernel.java:1055) at org.jruby.RubyKernel.require(RubyKernel.java:1038) at org.jruby.RubyKernel$s$1$0$require.call(RubyKernel$s$1$0$require.gen:65535) at org.jruby.internal.runtime.methods.JavaMethod$JavaMethodOneOrNBlock.call(JavaMethod.java:319) at org.jruby.internal.runtime.methods.AliasMethod.call(AliasMethod.java:61) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:167) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.chained_0_rescue_1$RUBY$SYNTHETIC__file__(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb:36) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:187) at org.jruby.runtime.callsite.SuperCallSite.callBlock(SuperCallSite.java:190) at org.jruby.runtime.callsite.SuperCallSite.call(SuperCallSite.java:197) at org.jruby.runtime.callsite.SuperCallSite.callVarargs(SuperCallSite.java:108) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.block_0$RUBY$file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:251) at rubyjit$require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F$block_0$RUBY$file.call(rubyjit$require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F$block_0$RUBY$file:65535) at org.jruby.runtime.CompiledBlock.yield(CompiledBlock.java:112) at org.jruby.runtime.CompiledBlock.yield(CompiledBlock.java:95) at org.jruby.runtime.CompiledBlock.yieldSpecific(CompiledBlock.java:75) at org.jruby.runtime.Block.yieldSpecific(Block.java:99) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.chained_0_rescue_1$RUBY$SYNTHETIC__file__(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:236) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:187) at org.jruby.runtime.callsite.CachingCallSite.callBlock(CachingCallSite.java:176) at org.jruby.runtime.callsite.CachingCallSite.callIter(CachingCallSite.java:187) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:251) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:167) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169) at org.jruby.ast.FCallOneArgNode.interpret(FCallOneArgNode.java:36) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.ast.RootNode.interpret(RootNode.java:129) at org.jruby.evaluator.ASTInterpreter.INTERPRET_ROOT(ASTInterpreter.java:120) at org.jruby.Ruby.runInterpreter(Ruby.java:720) at org.jruby.Ruby.loadFile(Ruby.java:2470) at org.jruby.runtime.load.ExternalScript.load(ExternalScript.java:66) at org.jruby.runtime.load.LoadService.tryLoadingLibraryOrScript(LoadService.java:767) at org.jruby.runtime.load.LoadService.smartLoad(LoadService.java:334) at org.jruby.runtime.load.LoadService.require(LoadService.java:381) at org.jruby.runtime.load.LoadService.lockAndRequire(LoadService.java:306) at org.jruby.RubyKernel.requireCommon(RubyKernel.java:1055) at org.jruby.RubyKernel.require(RubyKernel.java:1038) at org.jruby.RubyKernel$s$1$0$require.call(RubyKernel$s$1$0$require.gen:65535) at org.jruby.internal.runtime.methods.JavaMethod$JavaMethodOneOrNBlock.call(JavaMethod.java:319) at org.jruby.internal.runtime.methods.AliasMethod.call(AliasMethod.java:61) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:167) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.chained_0_rescue_1$RUBY$SYNTHETIC__file__(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb:36) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:187) at org.jruby.runtime.callsite.SuperCallSite.callBlock(SuperCallSite.java:190) at org.jruby.runtime.callsite.SuperCallSite.call(SuperCallSite.java:197) at org.jruby.runtime.callsite.SuperCallSite.callVarargs(SuperCallSite.java:108) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.block_0$RUBY$file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:251) at rubyjit$require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F$block_0$RUBY$file.call(rubyjit$require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F$block_0$RUBY$file:65535) at org.jruby.runtime.CompiledBlock.yield(CompiledBlock.java:112) at org.jruby.runtime.CompiledBlock.yield(CompiledBlock.java:95) at org.jruby.runtime.CompiledBlock.yieldSpecific(CompiledBlock.java:75) at org.jruby.runtime.Block.yieldSpecific(Block.java:99) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.chained_0_rescue_1$RUBY$SYNTHETIC__file__(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:236) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:187) at org.jruby.runtime.callsite.CachingCallSite.callBlock(CachingCallSite.java:176) at org.jruby.runtime.callsite.CachingCallSite.callIter(CachingCallSite.java:187) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:251) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:167) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169) at org.jruby.ast.FCallOneArgNode.interpret(FCallOneArgNode.java:36) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.ast.RootNode.interpret(RootNode.java:129) at org.jruby.evaluator.ASTInterpreter.INTERPRET_ROOT(ASTInterpreter.java:120) at org.jruby.Ruby.runInterpreter(Ruby.java:720) at org.jruby.Ruby.loadFile(Ruby.java:2470) at org.jruby.runtime.load.ExternalScript.load(ExternalScript.java:66) at org.jruby.runtime.load.LoadService.tryLoadingLibraryOrScript(LoadService.java:767) at org.jruby.runtime.load.LoadService.smartLoad(LoadService.java:334) at org.jruby.runtime.load.LoadService.require(LoadService.java:381) at org.jruby.runtime.load.LoadService.lockAndRequire(LoadService.java:306) at org.jruby.RubyKernel.requireCommon(RubyKernel.java:1055) at org.jruby.RubyKernel.require(RubyKernel.java:1038) at org.jruby.RubyKernel$s$1$0$require.call(RubyKernel$s$1$0$require.gen:65535) at org.jruby.internal.runtime.methods.JavaMethod$JavaMethodOneOrNBlock.call(JavaMethod.java:319) at org.jruby.internal.runtime.methods.AliasMethod.call(AliasMethod.java:61) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:167) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.chained_0_rescue_1$RUBY$SYNTHETIC__file__(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb:36) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:167) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169) at org.jruby.ast.FCallOneArgNode.interpret(FCallOneArgNode.java:36) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.ast.RootNode.interpret(RootNode.java:129) at org.jruby.evaluator.ASTInterpreter.INTERPRET_ROOT(ASTInterpreter.java:120) at org.jruby.Ruby.runInterpreter(Ruby.java:720) at org.jruby.Ruby.loadFile(Ruby.java:2470) at org.jruby.runtime.load.ExternalScript.load(ExternalScript.java:66) at org.jruby.runtime.load.LoadService.tryLoadingLibraryOrScript(LoadService.java:767) at org.jruby.runtime.load.LoadService.smartLoad(LoadService.java:334) at org.jruby.runtime.load.LoadService.require(LoadService.java:381) at org.jruby.runtime.load.LoadService.lockAndRequire(LoadService.java:306) at org.jruby.RubyKernel.requireCommon(RubyKernel.java:1055) at org.jruby.RubyKernel.require(RubyKernel.java:1038) at org.jruby.RubyKernel$s$1$0$require.call(RubyKernel$s$1$0$require.gen:65535) at org.jruby.internal.runtime.methods.JavaMethod$JavaMethodOneOrNBlock.call(JavaMethod.java:319) at org.jruby.internal.runtime.methods.AliasMethod.call(AliasMethod.java:61) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:167) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.chained_0_rescue_1$RUBY$SYNTHETIC__file__(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb:36) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:167) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169) at org.jruby.ast.FCallOneArgNode.interpret(FCallOneArgNode.java:36) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.ast.IfNode.interpret(IfNode.java:119) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.ast.RescueNode.executeBody(RescueNode.java:216) at org.jruby.ast.RescueNode.interpretWithJavaExceptions(RescueNode.java:120) at org.jruby.ast.RescueNode.interpret(RescueNode.java:110) at org.jruby.ast.BeginNode.interpret(BeginNode.java:83) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.IfNode.interpret(IfNode.java:117) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.ast.RootNode.interpret(RootNode.java:129) at org.jruby.evaluator.ASTInterpreter.INTERPRET_ROOT(ASTInterpreter.java:120) at org.jruby.Ruby.runInterpreter(Ruby.java:720) at org.jruby.Ruby.loadFile(Ruby.java:2470) at org.jruby.runtime.load.ExternalScript.load(ExternalScript.java:66) at org.jruby.runtime.load.LoadService.tryLoadingLibraryOrScript(LoadService.java:767) at org.jruby.runtime.load.LoadService.smartLoad(LoadService.java:334) at org.jruby.runtime.load.LoadService.require(LoadService.java:381) at org.jruby.runtime.load.LoadService.lockAndRequire(LoadService.java:306) at org.jruby.RubyKernel.requireCommon(RubyKernel.java:1055) at org.jruby.RubyKernel.require(RubyKernel.java:1038) at org.jruby.RubyKernel$s$1$0$require.call(RubyKernel$s$1$0$require.gen:65535) at org.jruby.internal.runtime.methods.JavaMethod$JavaMethodOneOrNBlock.call(JavaMethod.java:319) at org.jruby.internal.runtime.methods.AliasMethod.call(AliasMethod.java:61) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169) at org.jruby.ast.FCallOneArgNode.interpret(FCallOneArgNode.java:36) at org.jruby.ast.ReturnNode.interpret(ReturnNode.java:92) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.IfNode.interpret(IfNode.java:117) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.ast.RescueBodyNode.interpret(RescueBodyNode.java:108) at org.jruby.ast.RescueNode.handleException(RescueNode.java:181) at org.jruby.ast.RescueNode.interpretWithJavaExceptions(RescueNode.java:124) at org.jruby.ast.RescueNode.interpret(RescueNode.java:110) at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:75) at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:190) at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:179) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169) at org.jruby.ast.FCallOneArgNode.interpret(FCallOneArgNode.java:36) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.ast.RootNode.interpret(RootNode.java:129) at org.jruby.evaluator.ASTInterpreter.INTERPRET_ROOT(ASTInterpreter.java:120) at org.jruby.Ruby.runInterpreter(Ruby.java:720) at org.jruby.Ruby.loadFile(Ruby.java:2470) at org.jruby.runtime.load.ExternalScript.load(ExternalScript.java:66) at org.jruby.runtime.load.LoadService.load(LoadService.java:272) at org.jruby.RubyKernel.loadCommon(RubyKernel.java:1081) at org.jruby.RubyKernel.load(RubyKernel.java:1063) at org.jruby.RubyKernel$s$0$1$load.call(RubyKernel$s$0$1$load.gen:65535) at org.jruby.internal.runtime.methods.DynamicMethod.call(DynamicMethod.java:211) at org.jruby.internal.runtime.methods.DynamicMethod.call(DynamicMethod.java:207) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169) at org.jruby.ast.FCallOneArgNode.interpret(FCallOneArgNode.java:36) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.ast.IfNode.interpret(IfNode.java:117) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:75) at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:147) at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:163) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:292) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:135) at Library.Ruby.Gems.$1_dot_8.gems.ruby_minus_processing_minus_1_dot_0_dot_11.lib.ruby_minus_processing.runners.run.file(/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/ruby-processing/runners/run.rb:7) at Library.Ruby.Gems.$1_dot_8.gems.ruby_minus_processing_minus_1_dot_0_dot_11.lib.ruby_minus_processing.runners.run.load(/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/ruby-processing/runners/run.rb) at org.jruby.Ruby.runScript(Ruby.java:693) at org.jruby.Ruby.runScript(Ruby.java:686) at org.jruby.Ruby.runNormally(Ruby.java:593) at org.jruby.Ruby.runFromMain(Ruby.java:442) at org.jruby.Main.doRunFromMain(Main.java:321) at org.jruby.Main.internalRun(Main.java:241) at org.jruby.Main.run(Main.java:207) at org.jruby.Main.run(Main.java:191) at org.jruby.Main.main(Main.java:171) activerecord-jdbc-adapter requires ActiveRecord at runtime java.lang.SecurityException: class "arjdbc.jdbc.RubyJdbcConnection$s$1$0$select_p"'s signer information does not match signer information of other classes in the same package at java.lang.ClassLoader.checkCerts(ClassLoader.java:806) at java.lang.ClassLoader.preDefineClass(ClassLoader.java:487) at java.lang.ClassLoader.defineClassCond(ClassLoader.java:625) at java.lang.ClassLoader.defineClass(ClassLoader.java:615) at org.jruby.util.JRubyClassLoader.defineClass(JRubyClassLoader.java:76) at org.jruby.internal.runtime.methods.InvocationMethodFactory.endClass(InvocationMethodFactory.java:1378) at org.jruby.internal.runtime.methods.InvocationMethodFactory.getAnnotatedMethodClass(InvocationMethodFactory.java:792) at org.jruby.internal.runtime.methods.InvocationMethodFactory.getAnnotatedMethod(InvocationMethodFactory.java:810) at org.jruby.RubyModule.defineAnnotatedMethod(RubyModule.java:711) at org.jruby.RubyModule.defineAnnotatedMethod(RubyModule.java:679) at org.jruby.anno.TypePopulator$DefaultTypePopulator.populate(TypePopulator.java:68) at org.jruby.RubyModule.defineAnnotatedMethodsIndividually(RubyModule.java:673) at org.jruby.RubyModule.defineAnnotatedMethods(RubyModule.java:561) at arjdbc.jdbc.RubyJdbcConnection.createJdbcConnectionClass(RubyJdbcConnection.java:90) at arjdbc.jdbc.AdapterJavaService.basicLoad(AdapterJavaService.java:53) at org.jruby.Ruby.loadExtension(Ruby.java:2586) at org.jruby.runtime.load.ClassExtensionLibrary.load(ClassExtensionLibrary.java:55) at org.jruby.runtime.load.LoadService.tryLoadingLibraryOrScript(LoadService.java:767) at org.jruby.runtime.load.LoadService.smartLoad(LoadService.java:334) at org.jruby.runtime.load.LoadService.require(LoadService.java:381) at org.jruby.runtime.load.LoadService.lockAndRequire(LoadService.java:306) at org.jruby.RubyKernel.requireCommon(RubyKernel.java:1055) at org.jruby.RubyKernel.require(RubyKernel.java:1038) at org.jruby.RubyKernel$s$1$0$require.call(RubyKernel$s$1$0$require.gen:65535) at org.jruby.internal.runtime.methods.JavaMethod$JavaMethodOneOrNBlock.call(JavaMethod.java:319) at org.jruby.internal.runtime.methods.AliasMethod.call(AliasMethod.java:61) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:167) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.chained_0_rescue_1$RUBY$SYNTHETIC__file__(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb:36) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:187) at org.jruby.runtime.callsite.SuperCallSite.callBlock(SuperCallSite.java:190) at org.jruby.runtime.callsite.SuperCallSite.call(SuperCallSite.java:197) at org.jruby.runtime.callsite.SuperCallSite.callVarargs(SuperCallSite.java:108) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.block_0$RUBY$file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:251) at rubyjit$require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F$block_0$RUBY$file.call(rubyjit$require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F$block_0$RUBY$file:65535) at org.jruby.runtime.CompiledBlock.yield(CompiledBlock.java:112) at org.jruby.runtime.CompiledBlock.yield(CompiledBlock.java:95) at org.jruby.runtime.CompiledBlock.yieldSpecific(CompiledBlock.java:75) at org.jruby.runtime.Block.yieldSpecific(Block.java:99) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.chained_0_rescue_1$RUBY$SYNTHETIC__file__(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:236) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:187) at org.jruby.runtime.callsite.CachingCallSite.callBlock(CachingCallSite.java:176) at org.jruby.runtime.callsite.CachingCallSite.callIter(CachingCallSite.java:187) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:251) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:167) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169) at org.jruby.ast.FCallOneArgNode.interpret(FCallOneArgNode.java:36) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.ast.RootNode.interpret(RootNode.java:129) at org.jruby.evaluator.ASTInterpreter.INTERPRET_ROOT(ASTInterpreter.java:120) at org.jruby.Ruby.runInterpreter(Ruby.java:720) at org.jruby.Ruby.loadFile(Ruby.java:2470) at org.jruby.runtime.load.ExternalScript.load(ExternalScript.java:66) at org.jruby.runtime.load.LoadService.tryLoadingLibraryOrScript(LoadService.java:767) at org.jruby.runtime.load.LoadService.smartLoad(LoadService.java:334) at org.jruby.runtime.load.LoadService.require(LoadService.java:381) at org.jruby.runtime.load.LoadService.lockAndRequire(LoadService.java:306) at org.jruby.RubyKernel.requireCommon(RubyKernel.java:1055) at org.jruby.RubyKernel.require(RubyKernel.java:1038) at org.jruby.RubyKernel$s$1$0$require.call(RubyKernel$s$1$0$require.gen:65535) at org.jruby.internal.runtime.methods.JavaMethod$JavaMethodOneOrNBlock.call(JavaMethod.java:319) at org.jruby.internal.runtime.methods.AliasMethod.call(AliasMethod.java:61) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:167) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.chained_0_rescue_1$RUBY$SYNTHETIC__file__(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb:36) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:187) at org.jruby.runtime.callsite.SuperCallSite.callBlock(SuperCallSite.java:190) at org.jruby.runtime.callsite.SuperCallSite.call(SuperCallSite.java:197) at org.jruby.runtime.callsite.SuperCallSite.callVarargs(SuperCallSite.java:108) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.block_0$RUBY$file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:251) at rubyjit$require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F$block_0$RUBY$file.call(rubyjit$require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F$block_0$RUBY$file:65535) at org.jruby.runtime.CompiledBlock.yield(CompiledBlock.java:112) at org.jruby.runtime.CompiledBlock.yield(CompiledBlock.java:95) at org.jruby.runtime.CompiledBlock.yieldSpecific(CompiledBlock.java:75) at org.jruby.runtime.Block.yieldSpecific(Block.java:99) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.chained_0_rescue_1$RUBY$SYNTHETIC__file__(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:236) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:187) at org.jruby.runtime.callsite.CachingCallSite.callBlock(CachingCallSite.java:176) at org.jruby.runtime.callsite.CachingCallSite.callIter(CachingCallSite.java:187) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:251) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:167) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169) at org.jruby.ast.FCallOneArgNode.interpret(FCallOneArgNode.java:36) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.ast.RootNode.interpret(RootNode.java:129) at org.jruby.evaluator.ASTInterpreter.INTERPRET_ROOT(ASTInterpreter.java:120) at org.jruby.Ruby.runInterpreter(Ruby.java:720) at org.jruby.Ruby.loadFile(Ruby.java:2470) at org.jruby.runtime.load.ExternalScript.load(ExternalScript.java:66) at org.jruby.runtime.load.LoadService.tryLoadingLibraryOrScript(LoadService.java:767) at org.jruby.runtime.load.LoadService.smartLoad(LoadService.java:334) at org.jruby.runtime.load.LoadService.require(LoadService.java:381) at org.jruby.runtime.load.LoadService.lockAndRequire(LoadService.java:306) at org.jruby.RubyKernel.requireCommon(RubyKernel.java:1055) at org.jruby.RubyKernel.require(RubyKernel.java:1038) at org.jruby.RubyKernel$s$1$0$require.call(RubyKernel$s$1$0$require.gen:65535) at org.jruby.internal.runtime.methods.JavaMethod$JavaMethodOneOrNBlock.call(JavaMethod.java:319) at org.jruby.internal.runtime.methods.AliasMethod.call(AliasMethod.java:61) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:167) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.chained_0_rescue_1$RUBY$SYNTHETIC__file__(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb:36) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:187) at org.jruby.runtime.callsite.SuperCallSite.callBlock(SuperCallSite.java:190) at org.jruby.runtime.callsite.SuperCallSite.call(SuperCallSite.java:197) at org.jruby.runtime.callsite.SuperCallSite.callVarargs(SuperCallSite.java:108) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.block_0$RUBY$file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:251) at rubyjit$require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F$block_0$RUBY$file.call(rubyjit$require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F$block_0$RUBY$file:65535) at org.jruby.runtime.CompiledBlock.yield(CompiledBlock.java:112) at org.jruby.runtime.CompiledBlock.yield(CompiledBlock.java:95) at org.jruby.runtime.CompiledBlock.yieldSpecific(CompiledBlock.java:75) at org.jruby.runtime.Block.yieldSpecific(Block.java:99) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.chained_0_rescue_1$RUBY$SYNTHETIC__file__(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:236) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:187) at org.jruby.runtime.callsite.CachingCallSite.callBlock(CachingCallSite.java:176) at org.jruby.runtime.callsite.CachingCallSite.callIter(CachingCallSite.java:187) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:251) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:167) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169) at org.jruby.ast.FCallOneArgNode.interpret(FCallOneArgNode.java:36) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.ast.RootNode.interpret(RootNode.java:129) at org.jruby.evaluator.ASTInterpreter.INTERPRET_ROOT(ASTInterpreter.java:120) at org.jruby.Ruby.runInterpreter(Ruby.java:720) at org.jruby.Ruby.loadFile(Ruby.java:2470) at org.jruby.runtime.load.ExternalScript.load(ExternalScript.java:66) at org.jruby.runtime.load.LoadService.tryLoadingLibraryOrScript(LoadService.java:767) at org.jruby.runtime.load.LoadService.smartLoad(LoadService.java:334) at org.jruby.runtime.load.LoadService.require(LoadService.java:381) at org.jruby.runtime.load.LoadService.lockAndRequire(LoadService.java:306) at org.jruby.RubyKernel.requireCommon(RubyKernel.java:1055) at org.jruby.RubyKernel.require(RubyKernel.java:1038) at org.jruby.RubyKernel$s$1$0$require.call(RubyKernel$s$1$0$require.gen:65535) at org.jruby.internal.runtime.methods.JavaMethod$JavaMethodOneOrNBlock.call(JavaMethod.java:319) at org.jruby.internal.runtime.methods.AliasMethod.call(AliasMethod.java:61) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:167) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.chained_0_rescue_1$RUBY$SYNTHETIC__file__(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb:36) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:187) at org.jruby.runtime.callsite.SuperCallSite.callBlock(SuperCallSite.java:190) at org.jruby.runtime.callsite.SuperCallSite.call(SuperCallSite.java:197) at org.jruby.runtime.callsite.SuperCallSite.callVarargs(SuperCallSite.java:108) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.block_0$RUBY$file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:251) at rubyjit$require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F$block_0$RUBY$file.call(rubyjit$require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F$block_0$RUBY$file:65535) at org.jruby.runtime.CompiledBlock.yield(CompiledBlock.java:112) at org.jruby.runtime.CompiledBlock.yield(CompiledBlock.java:95) at org.jruby.runtime.CompiledBlock.yieldSpecific(CompiledBlock.java:75) at org.jruby.runtime.Block.yieldSpecific(Block.java:99) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.chained_0_rescue_1$RUBY$SYNTHETIC__file__(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:236) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:187) at org.jruby.runtime.callsite.CachingCallSite.callBlock(CachingCallSite.java:176) at org.jruby.runtime.callsite.CachingCallSite.callIter(CachingCallSite.java:187) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:251) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:167) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169) at org.jruby.ast.FCallOneArgNode.interpret(FCallOneArgNode.java:36) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.ast.RootNode.interpret(RootNode.java:129) at org.jruby.evaluator.ASTInterpreter.INTERPRET_ROOT(ASTInterpreter.java:120) at org.jruby.Ruby.runInterpreter(Ruby.java:720) at org.jruby.Ruby.loadFile(Ruby.java:2470) at org.jruby.runtime.load.ExternalScript.load(ExternalScript.java:66) at org.jruby.runtime.load.LoadService.tryLoadingLibraryOrScript(LoadService.java:767) at org.jruby.runtime.load.LoadService.smartLoad(LoadService.java:334) at org.jruby.runtime.load.LoadService.require(LoadService.java:381) at org.jruby.runtime.load.LoadService.lockAndRequire(LoadService.java:306) at org.jruby.RubyKernel.requireCommon(RubyKernel.java:1055) at org.jruby.RubyKernel.require(RubyKernel.java:1038) at org.jruby.RubyKernel$s$1$0$require.call(RubyKernel$s$1$0$require.gen:65535) at org.jruby.internal.runtime.methods.JavaMethod$JavaMethodOneOrNBlock.call(JavaMethod.java:319) at org.jruby.internal.runtime.methods.AliasMethod.call(AliasMethod.java:61) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.chained_0_rescue_1$RUBY$SYNTHETIC__file__(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb:36) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:187) at org.jruby.runtime.callsite.SuperCallSite.callBlock(SuperCallSite.java:190) at org.jruby.runtime.callsite.SuperCallSite.call(SuperCallSite.java:197) at org.jruby.runtime.callsite.SuperCallSite.callVarargs(SuperCallSite.java:108) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.block_0$RUBY$file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:251) at rubyjit$require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F$block_0$RUBY$file.call(rubyjit$require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F$block_0$RUBY$file:65535) at org.jruby.runtime.CompiledBlock.yield(CompiledBlock.java:112) at org.jruby.runtime.CompiledBlock.yield(CompiledBlock.java:95) at org.jruby.runtime.CompiledBlock.yieldSpecific(CompiledBlock.java:75) at org.jruby.runtime.Block.yieldSpecific(Block.java:99) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.chained_0_rescue_1$RUBY$SYNTHETIC__file__(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:236) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:187) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:322) at org.jruby.runtime.callsite.CachingCallSite.callBlock(CachingCallSite.java:178) at org.jruby.runtime.callsite.CachingCallSite.callIter(CachingCallSite.java:187) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:251) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:167) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169) at org.jruby.ast.FCallOneArgNode.interpret(FCallOneArgNode.java:36) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.RootNode.interpret(RootNode.java:129) at org.jruby.evaluator.ASTInterpreter.INTERPRET_ROOT(ASTInterpreter.java:120) at org.jruby.Ruby.runInterpreter(Ruby.java:720) at org.jruby.Ruby.loadFile(Ruby.java:2470) at org.jruby.runtime.load.ExternalScript.load(ExternalScript.java:66) at org.jruby.runtime.load.LoadService.tryLoadingLibraryOrScript(LoadService.java:767) at org.jruby.runtime.load.LoadService.smartLoad(LoadService.java:334) at org.jruby.runtime.load.LoadService.require(LoadService.java:381) at org.jruby.runtime.load.LoadService.lockAndRequire(LoadService.java:306) at org.jruby.RubyKernel.requireCommon(RubyKernel.java:1055) at org.jruby.RubyKernel.require(RubyKernel.java:1038) at org.jruby.RubyKernel$s$1$0$require.call(RubyKernel$s$1$0$require.gen:65535) at org.jruby.internal.runtime.methods.JavaMethod$JavaMethodOneOrNBlock.call(JavaMethod.java:319) at org.jruby.internal.runtime.methods.AliasMethod.call(AliasMethod.java:61) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.chained_1_rescue_line_56(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb:59) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.chained_0_rescue_1$RUBY$SYNTHETIC__file__(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb:55) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at rubyjit.require_FCD32B31AAEAA95645A351B9267520ED99453200.file(file:/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/core/jruby-complete.jar!/META-INF/jruby.home/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:187) at org.jruby.runtime.callsite.SuperCallSite.callBlock(SuperCallSite.java:190) at org.jruby.runtime.callsite.SuperCallSite.call(SuperCallSite.java:197) at org.jruby.runtime.callsite.SuperCallSite.callVarargs(SuperCallSite.java:108) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.block_0$RUBY$file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:251) at rubyjit$require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F$block_0$RUBY$file.call(rubyjit$require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F$block_0$RUBY$file:65535) at org.jruby.runtime.CompiledBlock.yield(CompiledBlock.java:112) at org.jruby.runtime.CompiledBlock.yield(CompiledBlock.java:95) at org.jruby.runtime.CompiledBlock.yieldSpecific(CompiledBlock.java:75) at org.jruby.runtime.Block.yieldSpecific(Block.java:99) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.chained_0_rescue_1$RUBY$SYNTHETIC__file__(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:236) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at rubyjit.load_dependency_5D6F56C3567075A267D6D6EB9252FEEACFB01169.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:187) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:322) at org.jruby.runtime.callsite.CachingCallSite.callBlock(CachingCallSite.java:178) at org.jruby.runtime.callsite.CachingCallSite.callIter(CachingCallSite.java:187) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb:251) at rubyjit.require_4B6D6FCE45594EFEA2491A08CD47BA7595C42B2F.file(/Users/fjenett/.gem/jruby/1.8/gems/activesupport-3.2.1/lib/active_support/dependencies.rb) at org.jruby.internal.runtime.methods.JittedMethod.call(JittedMethod.java:167) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169) at org.jruby.ast.FCallOneArgNode.interpret(FCallOneArgNode.java:36) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.RescueNode.executeBody(RescueNode.java:216) at org.jruby.ast.RescueNode.interpretWithJavaExceptions(RescueNode.java:120) at org.jruby.ast.RescueNode.interpret(RescueNode.java:110) at org.jruby.ast.BeginNode.interpret(BeginNode.java:83) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:75) at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:190) at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:179) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169) at org.jruby.ast.FCallOneArgNode.interpret(FCallOneArgNode.java:36) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.WhenOneArgNode.whenSlowTest(WhenOneArgNode.java:36) at org.jruby.ast.WhenOneArgNode.when(WhenOneArgNode.java:46) at org.jruby.ast.CaseNode.interpret(CaseNode.java:133) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:75) at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:147) at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:163) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:292) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:135) at org.jruby.ast.CallNoArgNode.interpret(CallNoArgNode.java:63) at org.jruby.ast.LocalAsgnNode.interpret(LocalAsgnNode.java:123) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:75) at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:190) at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:179) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169) at org.jruby.ast.CallOneArgNode.interpret(CallOneArgNode.java:57) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.ast.RootNode.interpret(RootNode.java:129) at org.jruby.evaluator.ASTInterpreter.INTERPRET_ROOT(ASTInterpreter.java:120) at org.jruby.Ruby.runInterpreter(Ruby.java:720) at org.jruby.Ruby.loadFile(Ruby.java:2470) at org.jruby.runtime.load.ExternalScript.load(ExternalScript.java:66) at org.jruby.runtime.load.LoadService.load(LoadService.java:272) at org.jruby.RubyKernel.loadCommon(RubyKernel.java:1081) at org.jruby.RubyKernel.load(RubyKernel.java:1063) at org.jruby.RubyKernel$s$0$1$load.call(RubyKernel$s$0$1$load.gen:65535) at org.jruby.internal.runtime.methods.DynamicMethod.call(DynamicMethod.java:211) at org.jruby.internal.runtime.methods.DynamicMethod.call(DynamicMethod.java:207) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169) at org.jruby.ast.FCallOneArgNode.interpret(FCallOneArgNode.java:36) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.ast.IfNode.interpret(IfNode.java:117) at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104) at org.jruby.ast.BlockNode.interpret(BlockNode.java:71) at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:75) at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:147) at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:163) at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:292) at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:135) at Library.Ruby.Gems.$1_dot_8.gems.ruby_minus_processing_minus_1_dot_0_dot_11.lib.ruby_minus_processing.runners.run.file(/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/ruby-processing/runners/run.rb:7) at Library.Ruby.Gems.$1_dot_8.gems.ruby_minus_processing_minus_1_dot_0_dot_11.lib.ruby_minus_processing.runners.run.load(/Library/Ruby/Gems/1.8/gems/ruby-processing-1.0.11/lib/ruby-processing/runners/run.rb) at org.jruby.Ruby.runScript(Ruby.java:693) at org.jruby.Ruby.runScript(Ruby.java:686) at org.jruby.Ruby.runNormally(Ruby.java:593) at org.jruby.Ruby.runFromMain(Ruby.java:442) at org.jruby.Main.doRunFromMain(Main.java:321) at org.jruby.Main.internalRun(Main.java:241) at org.jruby.Main.run(Main.java:207) at org.jruby.Main.run(Main.java:191) at org.jruby.Main.main(Main.java:171) LoadError: Please install the jdbcmysql adapter: gem install activerecord-jdbcmysql-adapter (load error: arjdbc/jdbc/adapter_java -- java.lang.RuntimeException: (LoadError) class "arjdbc.jdbc.RubyJdbcConnection$s$1$0$select_p"'s signer information does not match signer information of other classes in the same package) Sorry, Markdown makes it look a bit crazy .. Ah and i should mention this is running through the ruby-processing gem. We don't sign any jars -- not sure why you're seeing this. Is there something in your security policy that requires all jars to be signed? Checking with ruby-processing team .. https://github.com/jashkenas/ruby-processing/issues/41 Ok, solved. Problem is on my side ... i was running the vendor'ed JRuby that comes with ruby-processing. After installing a full JRuby and running ruby-processing with that everything works as expected.
gharchive/issue
2012-02-21T06:32:17
2025-04-01T04:34:43.073754
{ "authors": [ "fjenett", "nicksieger" ], "repo": "jruby/activerecord-jdbc-adapter", "url": "https://github.com/jruby/activerecord-jdbc-adapter/issues/175", "license": "bsd-2-clause", "license_type": "permissive", "license_source": "bigquery" }
2133132873
Significantly better performance by limiting left joins to max block id range Compared performance across various hosts Revised query to limit nested joins based on maximum potential block range given the request. Improved performance over 4x versus old query. Improves #8
gharchive/pull-request
2024-02-13T21:10:18
2025-04-01T04:34:43.077785
{ "authors": [ "jrwashburn" ], "repo": "jrwashburn/mina-payouts-data-provider", "url": "https://github.com/jrwashburn/mina-payouts-data-provider/pull/18", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
892159296
added getCompositeStatus helper function closes #35 Codecov Report Merging #40 (53212b0) into develop (bf23c7f) will not change coverage. The diff coverage is 100.00%. @@ Coverage Diff @@ ## develop #40 +/- ## ========================================= Coverage 100.00% 100.00% ========================================= Files 20 21 +1 Lines 243 248 +5 Branches 10 11 +1 ========================================= + Hits 243 248 +5 Impacted Files Coverage Δ src/helpers/getCompositeStatus.ts 100.00% <100.00%> (ø) Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update bf23c7f...53212b0. Read the comment docs.
gharchive/pull-request
2021-05-14T18:50:42
2025-04-01T04:34:43.127732
{ "authors": [ "codecov-commenter", "j-a-k-git" ], "repo": "js-thing/http-status-codes", "url": "https://github.com/js-thing/http-status-codes/pull/40", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2144044008
Stops with error "Too many requests from Identifier" For larger tenants, processing stops with this error message. Any ideas on how to tackle that? [...] implementing throttle success on batch 1445 of 20 requests success on batch 1446 of 20 requests success on batch 1448 of 20 requests success on batch 1449 of 20 requests success on batch 1447 of 20 requests batching 1451 of 1454 batching 1452 of 1454 batching 1453 of 1454 batching 1454 of 1454 success on batch 1451 of 20 requests success on batch 1450 of 20 requests success on batch 1454 of 7 requests success on batch 1452 of 20 requests success on batch 1453 of 20 requests 29067 3450 https://graph.microsoft.com/beta/directoryRoles/ { error: { code: 'TooManyRequests', message: 'Too many requests from Identifier:<id> under category:aadgraph.app.tenant_large. Please try again later.', innerError: { date: '2024-02-20T10:10:57' } } } Hi! Thx for reporting the issue. I had configured an rate limit that works quite well, so you would not necessarily run into these. That being said, I could check for re-run few retries for with slower batches for the responses that get 429. Alternatively I could expose the configuration parameters to change the rates Hi! Thanks, I tried several values for the throttleinMS. Initially it failed in the call to "await admins()" part. But got it working now. Thanks! Thx for all the feedback, I am gonna close this now, and thank you for your ability to navigate the internal params. I shall look back into exposing these params in the starting arguments :)
gharchive/issue
2024-02-20T10:26:15
2025-04-01T04:34:43.130160
{ "authors": [ "jsa2", "sqlpirate" ], "repo": "jsa2/AADAppAudit", "url": "https://github.com/jsa2/AADAppAudit/issues/9", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
651336471
Support other types of databases If support other database? For example:MySQL, DM(达梦) .. @wtujvk I'm not sure about other database products, but normal [Index] attribute usage will work fine another database products such as MySQL, I think. Have you been tried [Index] attribute of this package?
gharchive/issue
2020-07-06T08:08:50
2025-04-01T04:34:43.131576
{ "authors": [ "jsakamoto", "wtujvk" ], "repo": "jsakamoto/EntityFrameworkCore.IndexAttribute", "url": "https://github.com/jsakamoto/EntityFrameworkCore.IndexAttribute/issues/9", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
806147923
Fix top-xxx positions In example 5, the Top Left button was showing a notification in the top right and vice versa. This switches the positions around. Thank you! :)
gharchive/pull-request
2021-02-11T07:30:22
2025-04-01T04:34:43.133035
{ "authors": [ "C-Lodder", "jsanahuja" ], "repo": "jsanahuja/Notifierjs", "url": "https://github.com/jsanahuja/Notifierjs/pull/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1593933870
Multi line secret text? Is there a way to do a multi line secret text block? If i do this, it works: @secret[this works] If I do this, it doesn't work: @secret[this does not work] I will look into extending the regex to include multiple lines Due to the way foundry parses the text for HTML Enrichment multiline isn't possible at this time
gharchive/issue
2023-02-21T18:31:08
2025-04-01T04:34:43.138035
{ "authors": [ "jsavko", "kiloforce" ], "repo": "jsavko/inline-secrets", "url": "https://github.com/jsavko/inline-secrets/issues/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }