id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
93295585
|
Additional hardcoded table value in subquery
Forgot to also update the hardcoded AS locations in the subquery to also look at the Model table.
Thanks for this, I had just changed it when I accepted another pull request so this should all be good.
|
gharchive/pull-request
| 2015-07-06T14:41:08 |
2025-04-01T04:34:39.025762
|
{
"authors": [
"jackpopp",
"jonwurtzler"
],
"repo": "jackpopp/geodistance",
"url": "https://github.com/jackpopp/geodistance/pull/6",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1763852449
|
How to unbind \?
I tried unbind \ and unbind-key \ but not work.
And in my case, prefix + \ does not work for me, right-click works.
Either '' or without quotes as \
I've tried but still nothing changed
I just tried and this worked for me: tmux unbind \\ This disabled the default shortcut
I am running tmux 3.3a, If you run some other version, please let me know and I will investigate if this procedure does not work for that version
I got it.
My previous mistake is about the order.
Bind \ to split pane
Set plugins (I thought this step means import)
Then unbind (Canceled my own bind function)
And now changed to
unbind
bind
set plugin (Looks it works only on install, and will import at the very first before step 1)
|
gharchive/issue
| 2023-06-19T16:25:15 |
2025-04-01T04:34:39.048287
|
{
"authors": [
"Erimus-Koo",
"jaclu"
],
"repo": "jaclu/tmux-menus",
"url": "https://github.com/jaclu/tmux-menus/issues/33",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2503382286
|
AttributeError for ScoreCAM Class
Issue: AttributeError for ScoreCAM Class
Description
I encountered an AttributeError when using the ScoreCAM class from the pytorch_grad_cam package. The error message is as follows:
`AttributeError: 'ScoreCAM' object has no attribute 'device'
Code Context
In the ScoreCAM class, the line causing the error is:
activation_tensor = activation_tensor.to(self.device)
Expected Behavior
The ScoreCAM class should either:
Initialize self.device properly, or
Avoid using self.device if it is not applicable.
Possible Solution:
One potential fix would be to pass device as a param to the constructor
Steps to Reproduce:
Initialize a ScoreCAM object.
Call the object with
Hi,
I can't reproduce this.
ScoreCAM inherits from BaseCAM which does have a self.device attribute.
Could you please share more information - which version are you using ?
Could you please share more information - which version are you using ?
Version 1.4.8
This should be fixed in the latest releases. Could you please try upgrading the package ?
|
gharchive/issue
| 2024-09-03T17:14:38 |
2025-04-01T04:34:39.052618
|
{
"authors": [
"dwil2444",
"jacobgil"
],
"repo": "jacobgil/pytorch-grad-cam",
"url": "https://github.com/jacobgil/pytorch-grad-cam/issues/528",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
168399923
|
Add --forever to Slack notifications documentation
The Slack notifications documentation should mention the --forever flag, otherwise the integration will seem to fail shortly after installation.
Ah, very true. Thanks for this!
|
gharchive/pull-request
| 2016-07-29T19:45:57 |
2025-04-01T04:34:39.053678
|
{
"authors": [
"cdzombak",
"jacobmarshall"
],
"repo": "jacobmarshall/pokevision-cli",
"url": "https://github.com/jacobmarshall/pokevision-cli/pull/13",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1825265235
|
Add a config to allocate the read buffer when vectored reads are enabled
Currently, when vector reads are enabled, this FUSE library stops providing the pre-allocated read buffer. We would like to use vectored reads, but the pre-allocated read buffer is still very useful for us.
Our application prefetches data from the cloud, but sometimes the kernel will ask for data faster than we can prefetch it, in which case we need a buffer to read the data into from over the network. Without using vectored reads, we can make use of the pre-allocated read buffer. This is very efficient because the pre-allocated read buffer uses extra space from the buffer that reads operations from the kernel.
This change proposes adding another config that would allow the ReadFileOp.Dst buffer to be allocated for vectored reads. The file system implementation can choose to add data to ReadFileOp.Dst if it desires, and it must append ReadFileOp.Dst to ReadFileOp.Data if it wants the data to be included in the response to the kernel.
Hi @stapelberg , I would really appreciate it if you could take a look at this PR when you get a chance!
Currently, when vector reads are enabled, this FUSE library stops providing the pre-allocated read buffer. We would like to use vectored reads, but the pre-allocated read buffer is still very useful for us.
I though the whole point of using vectored reads is to avoid using library-provided buffers. Are you saying using vectored reads is still faster, even without the buffer management difference?
cc @vitalif who contributed vectored read support
@stapelberg The benefit of using vectored reads is that file system implementations are not forced to copy data into the library provided buffer.
In our case, we sometimes have the data in memory already and enabling vectored reads allows us to skip a data copy. But sometimes we don't have the data in memory, and if we enable vectored reads, we will need to allocate our own memory to hold the data. This is a bit wasteful and slow in comparison to reusing the library provided buffer, which has already been allocated by the library.
I'd also say it's a strange idea. Why can't you just keep a pool of allocated buffers if you want to avoid memory allocation cost? And in fact why do you want to avoid the allocation cost if you're anyway reading from the network? Reading from the network would be slower anyway. And you won't be able to reuse this buffer, it will only be used for 1 read...
Thanks for the reply @vitalif . It is not just about the memory allocation cost, its also about the total memory usage of our system. The library provided buffer memory has already been allocated, and it actually is not released. This library provided memory buffer is kept around and later reinserted into the freelists so it can be reused for the next operation. Also, this memory buffer is always at least 1 MB in size since it is allocated to be large enough to hold the largest possible message from the kernel.
Yes, we certainly can keep our own pool of 1 MB buffers, or just suffer the performance hit of frequently allocating and releasing memory buffers (which we have actually found to be non-negligible even with the network latency).
However, if we can simply reuse the library provided buffer, then we don't need to duplicate memory allocations, which reduces the overall memory usage and complexity of our system and increases performance.
Hi @stapelberg any more comments on this?
I’m willing to accept patches that result in jacobsa/fuse releasing memory earlier if possible, or to introduce size classes within the freelist to rightsize overall memory consumption.
But, the pull request in its current form seems too specific of a performance optimization to me.
Sharing buffers between the jacobsa/fuse package and its users only for performance reasons (as opposed to interchange data) increases complexity of the code and API surface too much for too little gain.
Therefore, I’ll close this PR for now. Do feel free to suggest different performance optimizations, though.
Thanks.
(Note for myself: I vaguely remembered Go code that assigned function arguments to local variables in order for the garbage collector to reclaim the associated memory more quickly, and I was wondering if we should pursue something like that here. However, that technique seems to be not required with Go 1.8+, so nevermind that.)
|
gharchive/pull-request
| 2023-07-27T21:41:04 |
2025-04-01T04:34:39.061278
|
{
"authors": [
"sbauersfeld",
"stapelberg",
"vitalif"
],
"repo": "jacobsa/fuse",
"url": "https://github.com/jacobsa/fuse/pull/148",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
113883464
|
Add I/O protection
Add series resistor to I/Os
Add (DNP) zener diods to I/Os
Not scheduled to be fixed in any revision.
|
gharchive/issue
| 2015-10-28T17:57:33 |
2025-04-01T04:34:39.070739
|
{
"authors": [
"erikwelsh",
"jadonk"
],
"repo": "jadonk/beaglebone-blue",
"url": "https://github.com/jadonk/beaglebone-blue/issues/3",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
572754305
|
Operator with namespace-only permissions hangs during reconciliation
Split from #931, caused by operator-framework/operator-sdk#2608.
When the operator doesn't have cluster-scope permissions, it hangs during the reconciliation with the following message happening every second:
E0228 10:28:58.020049 1 reflector.go:123] pkg/mod/k8s.io/client-go@v0.0.0-20191016111102-bec269661e48/tools/cache/reflector.go:96: Failed to list *v1.Secret: secrets is forbidden: User "system:serviceaccount:observability:jaeger-operator" cannot list resource "secrets" in API group "" at the cluster scope
Unless I set my Jaeger Operator up incorrectly, this seems to be an issue even when intended with cluster-scope permissions:
E0301 02:01:03.884054 1 reflector.go:123] pkg/mod/k8s.io/client-go@v0.0.0-20191016111102-bec269661e48/tools/cache/reflector.go:96: Failed to list *v1.Secret: secrets is forbidden: User "system:serviceaccount:observability:jaeger-operator" cannot list resource "secrets" in API group "" at the cluster scope
Looking at the cluster role spec here, I don't see where permissions are given for secrets (if its supposed to): https://raw.githubusercontent.com/jaegertracing/jaeger-operator/master/deploy/cluster_role.yaml
@ewohltman right, you are facing the problem that is being fixed as part of #936. This issue here is being addressed by #937: the lack of permissions will cause errors while reconciling the secrets, except that it won't block indefinitely.
@ewohltman right, you are facing the problem that is being fixed as part of #936. This issue here is being addressed by #937: the lack of permissions will cause errors while reconciling the secrets, except that it won't block indefinitely.
@jpkrohling Thanks for pointing me in the right direction!
|
gharchive/issue
| 2020-02-28T13:26:26 |
2025-04-01T04:34:39.077033
|
{
"authors": [
"ewohltman",
"jpkrohling"
],
"repo": "jaegertracing/jaeger-operator",
"url": "https://github.com/jaegertracing/jaeger-operator/issues/935",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
798652517
|
Divide build-binaries job to multiple jobs per platform
Signed-off-by: Ashmita Bohara ashmita.bohara152@gmail.com
Which problem is this PR solving?
Right now build-binaries is taking around 30 minutes to complete. This should do this in parallel.
Short description of the changes
Moving each platform binaries build into separate job.
One unfortunate side effect is that it explodes the number of jobs we need to configure as required for the branch to pass:
One unfortunate side effect is that it explodes the number of jobs we need to configure as required for the branch to pass:
I see. Hopefully we don't hit any limit imposed by github.
Yuri, I have one question: I remember you mentioning that jaegertracing org should get more credits as compared to personal open-source project. But from my experience I haven't seen any difference in the build times for a job in my jaeger's fork vs same build in jaeger.
I remember you mentioning that jaegertracing org should get more credits as compared to personal open-source project.
I am not sure about that. This page shows 2000 minutes for both personal and orgs. I am not sure if this is enforced yet, because I would've thought we easily exceed 2000 minutes.
It would be nice if GHA were able to automatically cancel running jobs for stale commits, we're burning through those a lot.
It would be nice if GHA were able to automatically cancel running jobs for stale commits, we're burning through those a lot.
I will look into it. Thinking if we can do something like this:
Create a custom GHA to kill all GHA.
Add the above custom action as a step if any of the task fails in the jobs which are marked as required.
I am not sure about the 2nd point. For the first one, there are already GHAs that kill other runs, but when I looked at few of them none were particularly great (may have changed since then)
|
gharchive/pull-request
| 2021-02-01T19:41:47 |
2025-04-01T04:34:39.090340
|
{
"authors": [
"Ashmita152",
"yurishkuro"
],
"repo": "jaegertracing/jaeger",
"url": "https://github.com/jaegertracing/jaeger/pull/2757",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
271321203
|
no such module Shout
mkdir ssh_shout
cd ssh_shout
swift package init --type executable
Then I added
.package(url: "https://github.com/jakeheis/Shout", from: "0.2.0")
in the Package.swift
Then I entered
swift package fetch
Then in Souces/ssh_shout/main.swift I entered import Shout, then
swift build
and I got "no such module Shout".
What can be the problem?
Thanks
Did you add Shout as a dependency of your main target?
let package = Package(
...
dependencies: [
.package(url: "https://github.com/jakeheis/Shout", from: "0.2.0")
],
targets: [
target(name: "MyTarget", dependencies: ["Shout"]),
]
Thanks very much! That's been solved, but the main thing still not works.
Error message:
Fatal error: Error raised at top level: Shout.LibSSH2Error.error(-16): file /BuildRoot/Library/Caches/com.apple.xbs/Sources/swiftlang/swiftlang-900.0.69.2/src/swift/stdlib/public/core/ErrorType.swift, line 187
Illegal instruction: 4
My main.swift:
import Shout
let session = try SSH.Session(host: "35.18.24.25")
try session.authenticate(username: "taqo4", privateKey: "~/dora", passphrase: "demo")
let (status, output) = try session.capture("pwd")
print("status: (status)")
print("output: (output)")
Thanks for the help!
|
gharchive/issue
| 2017-11-05T22:38:22 |
2025-04-01T04:34:39.116548
|
{
"authors": [
"jakeheis",
"taqios"
],
"repo": "jakeheis/Shout",
"url": "https://github.com/jakeheis/Shout/issues/6",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
300835268
|
Move Site Extension to nuget.org
Please see https://github.com/Azure/app-service-announcements/issues/87 for detailed announcement.
Important: if your site extension is still relevant, you need to do this before June 1st 2018. Please let me know if you have any questions.
Thanks!
Thanks David.
|
gharchive/issue
| 2018-02-27T23:36:58 |
2025-04-01T04:34:39.125644
|
{
"authors": [
"davidebbo",
"jakkaj"
],
"repo": "jakkaj/funcgraph",
"url": "https://github.com/jakkaj/funcgraph/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1228853842
|
Works with bevy_mod_picking?
The bevy world inspector is excellent and has helped me out of a few holes already. I would love to use the standard inspector with the bevy mod picking. Is there a way to hint to the inspector what it should be inspecting?
( https://docs.rs/bevy_mod_picking/latest/bevy_mod_picking/ )
Ah wait I see this:
https://github.com/jakobhellermann/bevy-inspector-egui/blob/main/examples/mouse_picking.rs.disabled
Let me see if that gets me going...
Yeah got that working.
|
gharchive/issue
| 2022-05-08T10:42:47 |
2025-04-01T04:34:39.133751
|
{
"authors": [
"gilescope"
],
"repo": "jakobhellermann/bevy-inspector-egui",
"url": "https://github.com/jakobhellermann/bevy-inspector-egui/issues/60",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2132731981
|
Allow more types of Icons as IconProps
Goal
I would like to use lucide (or any other) icon library for my icons in the expandIcon button variant.
When doing so tho, the type is not allowing it.
I managed to get around this by changing in the IconProps the Icon type to React.ReactNode and updating the Icon use from <Icon /> to {Icon}.
This allows me to use the Icon like this:
Problem
Now i have the problem that the expanding Animation for the Icon is not working anymore / the Icon is not showing up at all.
Could you help / update the code to allow other Icons to be animated too?
Resources
Reproduction Link: BirthdayyBot/dashboardV2
I don't think any of this is necessary.
Just tried using a lucide icon and it works fine as is.
I don't think any of this is necessary. Just tried using a lucide icon and it works fine as is.
Can you show me your Implementation? It wasn't working for me.
I don't think any of this is necessary. Just tried using a lucide icon and it works fine as is.
Can you show me your Implementation? It wasn't working for me.
import { ArrowRight } from "lucide-react";
Test
|
gharchive/issue
| 2024-02-13T16:33:41 |
2025-04-01T04:34:39.138467
|
{
"authors": [
"jakobhoeg",
"nikolaischunk"
],
"repo": "jakobhoeg/enhanced-button",
"url": "https://github.com/jakobhoeg/enhanced-button/issues/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2568349301
|
🛑 Repo1 is down
In a25ded9, Repo1 (https://repo1.dso.mil) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Repo1 is back up in e49aa61 after 8 minutes.
|
gharchive/issue
| 2024-10-05T22:28:14 |
2025-04-01T04:34:39.161918
|
{
"authors": [
"james-martinez"
],
"repo": "james-martinez/dso-mil",
"url": "https://github.com/james-martinez/dso-mil/issues/1566",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2600725999
|
🛑 Portal is down
In 7d85268, Portal (https://portal.cnap.dso.mil) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Portal is back up in 06a0038 after 6 minutes.
|
gharchive/issue
| 2024-10-20T17:30:49 |
2025-04-01T04:34:39.164218
|
{
"authors": [
"james-martinez"
],
"repo": "james-martinez/dso-mil",
"url": "https://github.com/james-martinez/dso-mil/issues/1701",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
280349294
|
use caret instead of tilde requirements, delete cargo lock
same as https://github.com/jameshurst/rust-metaflac/pull/5
thanks
IIRC, I was the one who comitted Cargo.lock to git because I thought it should always be to allow reproducible builds. TIL :)
Before we merge your PR, could you add Cargo.lock to .gitignore?
done :)
can you also publish this updates on cargo?
Done!
thanks!
|
gharchive/pull-request
| 2017-12-08T02:30:12 |
2025-04-01T04:34:39.172572
|
{
"authors": [
"jxs",
"polyfloyd"
],
"repo": "jameshurst/rust-id3",
"url": "https://github.com/jameshurst/rust-id3/pull/22",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
333066980
|
Execution failed for task app:transformClassesWithDexForDebug
How can i fix it ? when i run command "react-native run-android"
try running
cd android && gradlew clean
and then
react-native run-android
|
gharchive/issue
| 2018-06-17T14:29:11 |
2025-04-01T04:34:39.174072
|
{
"authors": [
"ddthanhdat",
"vinaygosain"
],
"repo": "jamesisaac/react-native-background-task",
"url": "https://github.com/jamesisaac/react-native-background-task/issues/45",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1167381557
|
system.notimplementedexception: This functionality is not implemented in the portable version of this assembly.
If you are creating an issue for a BUG please fill out this information. If you are asking a question or requesting a feature you can delete the sections below.
Failure to fill out this information will result in this issue being closed. If you post a full stack trace in a bug it will be closed, please post it to http://gist.github.com and then post the link here.
Bug Information
I have implemented InAppBilling plugin v4 and it is working fine in iOS but in Android I am getting "system.notimplementedexception: This functionality is not implemented in the portable version of this assembly." error. I have the plugin installed in my Android and iOS specific projects and tried with target versions Android 7,8.1,10 & 11 but no luck. @jamesmontemagno Could you please help me in this regard?
Version Number of Plugin: 4.0
Device Tested On: Android
Simulator Tested On:
Version of VS:
Version of Xamarin: Xamarin.mac: v8.0
Xamarin.iOS: v15.0
Xamarin.Android: v12.0
Versions of other things you are using:
Steps to reproduce the Behavior
Implement the InAppBilling code and test it on Android devices.
Expected Behavior
Should open the InAppBilling popup for Android
Actual Behavior
The pop up is not opening and throwing an error
Code snippet
Screenshots
What are you "Compiling" it against on Android? You must have it set to COMPILE against Android 10+.
Can you send me a sample project?
@jamesmontemagno Thank you for your quick response, I tried changing the compile against Android 10+ and it started working.
Also, is there any method or a way to check the purchases in this plugin like getPurchases(); ?
Yes, see documentation: https://jamesmontemagno.github.io/InAppBillingPlugin/CheckAndRestorePurchases.html
|
gharchive/issue
| 2022-03-12T19:48:53 |
2025-04-01T04:34:39.201516
|
{
"authors": [
"b-nishitgupta",
"jamesmontemagno"
],
"repo": "jamesmontemagno/InAppBillingPlugin",
"url": "https://github.com/jamesmontemagno/InAppBillingPlugin/issues/454",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
268816833
|
need to add workers to MIQ_WORKER_TYPES
to create a new provider and it's workers you need to add the classes to MIQ_WORKER_TYPES
@jrafanie do you know why we just cant do like RefreshWorker.leaf_subclasses but need to list all explicitly?
@durandom Because we can't autoload every provider class and it's dependencies. We need a better way to register workers instead of loading a whole provider
ok - @jameswnl so for now we'll have to add that to manual instructions for creating a new provider.
I dont think its worth automating that, although it should be easy (see how the Gemfile is modified by the generator)
This will be automated in https://github.com/ManageIQ/manageiq/pull/16416
PR merged.
|
gharchive/issue
| 2017-10-26T15:55:51 |
2025-04-01T04:34:39.214962
|
{
"authors": [
"durandom",
"jameswnl",
"jrafanie"
],
"repo": "jameswnl/manageiq-providers-dummy_provider",
"url": "https://github.com/jameswnl/manageiq-providers-dummy_provider/issues/8",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2169878267
|
Set up in custom dynamically scaled fonts
[ ] Add font files
[ ] allow to change pairings in interface on a native sample setting screen
[ ] allow changing on native Apple font text style level
[ ] map font file on the said scale
[ ] #11
[ ] accessible color modes?
Ressources
https://fonts.google.com/specimen/Ojuju
https://fonts.google.com/specimen/Josefin+Slab?classification=Display
https://fonts.google.com/specimen/Pixelify+Sans?query=pixel
https://fonts.google.com/specimen/Micro+5?query=micro+5&classification=Display
https://fonts.google.com/specimen/Jacquarda+Bastarda+9?query=jacquarda&classification=Display
https://fonts.google.com/specimen/Chewy?classification=Display
https://fonts.google.com/specimen/Pirata+One?classification=Display
https://fonts.google.com/specimen/Workbench?classification=Display
Chewy,Jacquarda_Bastarda_9,Josefin_Slab,Micro_5,Ojuju,etc.zip
Pairings screen features
[ ] editable text display for either body or header
[ ] apple text style selector
❓ What about color editing? Make a new issue?
❓ Are all these fonts "boldable" and if so, how can I add that field to the display font section programmatically?
❓ How do I legally appropriately cite that I'm using Google Fonts in my project?
|
gharchive/issue
| 2024-03-05T18:21:16 |
2025-04-01T04:34:39.236266
|
{
"authors": [
"jamie-brannan"
],
"repo": "jamie-brannan/galere",
"url": "https://github.com/jamie-brannan/galere/issues/10",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
29954828
|
2nd order linkwitzriley filters are phase inverted
The 2nd order Linkwitz-Riley Highpass and Lowpass filter are phase inverted to each other, which results in an unwanted sound coloration when summing their output. See patch and screen shot.
This might be a bug in the original filter design, I don't know.
----------begin_max5_patcher----------
752.3oc0WtsbZCCDF9ZySgFM8tRXrjwFnWk9bzISFisvnTaIWI4.jLgm8pC1
IjvISBmxMhQKqj92O1Uq34Ndvw74DID7Kve.ddO2wyyZxXvqdtGrHddRdrz5
FLgWTPXJXW22oHyUV6xRRhRTU.3S.ozISHBBKgz31DNSIoOQLthv87aLKxFa
L4+pEVUAkkST1CCsxpYwE1UC+sfFmCeycdkpw+lMoLVkLkxxtWn0jK3BQA8B
6B5GFpOJ.Jzbh.rdDbW8hno1smO9gavXnw1Kc5XF5dDIirp3pAIA99G.RPGS
jPYkUJPCXtdRRFXPPKAh+mAHtPlKS3kjkM5SxY7UiDYhfmmarDro.FuwvBs2
beLxEW8sQINHXk.qPKgjoBtCe9qGtngMp0clpEkD2tCgutK47rIBx+rx46Bb
pqB9ZvYvoGNLxL8QsVkzMK2S0yFoSaqU1eREZHx.r.rEhaoVAEsU9HoYL8oC
t63gjedoPRcpT6PR3oAIUEiIhCI9OzqOw6Hk.6Jdri691ST+sF9TS+B.bbLK
6shmxXgVaJh3dBKdbtqP7SvmBhTFmQVKmwTYVoevxBvOPWpBpZt02LNzROzv
MSuQ655liWgDiSkjkmxjIz95EisbneztxkFbRpjtLcpcA8WpWD5h0K5gdSn4
5Zzk.D3Vy4Bx4yJ0KpWNk82YT0SBZNYQOL31Wq2P999Maj1KRBuhoVEnsIsK
33cGFdzHK68ce3x7L+ProLu81Vq6QthbcBOklM8aFhqunqAwAt+X1VPbvAiX
6tXCzO7+ZsRwX+8bWxqDIMaZ88If2zRJQpnrXEkyVwmf24yTZZJgsZAYJUZZ
Skt89TsUMitpTSPKTi4oM57yqF4Dc1jSaxbhtpxbNepIpsYNWQx47QGTqzyv
ymdBaidFb1zSaprLu74bgm9sPOi9jxw0.Ktr7QhPVukVknehvCbgYZTW6TJy
M09hRnf7Hsw+AVKwBcuWktwakv0vd9vHXGy47Rm+C.cuWXE
-----------end_max5_patcher-----------
Do that mean that this is not an issue, and could be closed? Other filter designs should rethater be added as separate enhancement issues in the tracker.
I'm assigning to you @tap to close this issue if LinkwitzRiley 2nd order works as expected.
I would like to make sure this is documented, then we can close it.
Tim
On Mon, Mar 30, 2015 at 4:30 AM, Trond Lossuis notifications@github.com
wrote:
Do that mean that this is not an issue, and could be closed? Other filter
designs should rethater be added as separate enhancement issues in the
tracker.
I'm assigning to you @tap https://github.com/tap to close this issue if
LinkwitzRiley 2nd order works as expected.
—
Reply to this email directly or view it on GitHub
https://github.com/jamoma/JamomaCore/issues/280#issuecomment-87611649.
|
gharchive/issue
| 2014-03-22T05:11:31 |
2025-04-01T04:34:39.253071
|
{
"authors": [
"Nilson",
"lossius",
"tap"
],
"repo": "jamoma/JamomaCore",
"url": "https://github.com/jamoma/JamomaCore/issues/280",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2558029819
|
ci: The CI needs to package the cortex.llamacpp dependencies into the binary file by default for Cortex's integration into Jan
Given that: The current CI is publishing the binary file and installer to the release, but the binary does not include the llamacpp engine.
Expectation: Jan will pull the Cortex binary that has the llamacpp engine avx2 pre-integrated.
Approach:
Similar to the old approach like Nitro, ship multiple Cortex binaries including different llamacpp variants.
Modify Jan's CI to pull llamacpp engines from the cortex.llamacpp repo and the Cortex binary file from the cortex repo.
cc @dan-homebrew @louis-jan @vansangpfiev
@hiento09 is this a Sprint21 issue and is there a corresponding PR? Else we can shift it to Sprint22
@hiento09 is this a Sprint21 issue and is there a corresponding PR? Else we can shift it to Sprint22 cc @dan-homebrew
Let shift this issue to sprint 22 @gabrielle-ong , we have not finalized the approach yet
|
gharchive/issue
| 2024-10-01T02:28:34 |
2025-04-01T04:34:39.318012
|
{
"authors": [
"gabrielle-ong",
"hiento09"
],
"repo": "janhq/cortex.cpp",
"url": "https://github.com/janhq/cortex.cpp/issues/1369",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
99173742
|
API Information
I am trying to use $sails.get and was wondering where the API was documented.
I want to pass some information in the request.
+1
1.x is a wrapper for the sails.io client and as noted in the read me you can refer to the sails socket documention for the API. That documentation at least indicates what methods are available. 1.x wraps it so that you don't have to do all of the $apply() stuff and also provides a promise instead of accepting a callback.
2.x is a implementation of angular's $http that uses sockets to communicate with sails. You can refer to the angular $http documentation for methods (this is noted in the 2.x readme). Anything you can do with $http you can do with $sails in 2.x. If you can't do something that $http can do, file a defect.
Yeah, angular-sails should have it's own documentation rather than just referring to others in the readme...
Thanks Evan, I'll bare that in mind. If I find some time I'll submit a pull request with some more detailed updated docs.
|
gharchive/issue
| 2015-08-05T10:23:05 |
2025-04-01T04:34:39.321733
|
{
"authors": [
"TheSharpieOne",
"charlieryan",
"tommykennedy"
],
"repo": "janpantel/angular-sails",
"url": "https://github.com/janpantel/angular-sails/issues/82",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1769679642
|
🔌 Plugin: AAP (Ansible)
Goal
Programmatically populate the Backstage catalog with job templates from your AAP instance.
What problem does this solve?
Although there is a GPT to run a Job Template, there is no way to discover what Job Templates can be run, or what they do from within Backstage.
Use cases
Details to follow
Acceptance criteria
Details to follow
Issues in Epic
https://github.com/janus-idp/backstage-plugins/issues/528
https://github.com/janus-idp/backstage-plugins/issues/522
https://github.com/janus-idp/backstage-plugins/issues/545
https://github.com/janus-idp/backstage-plugins/issues/546
Could we make it possible to invoke a one-off run of the playbook from its entity page? I have a customer who seeks the ability to invoke a playbook from the Backstage UI.
This is related to https://github.com/janus-idp/backstage-plugins/issues/332 - if we could invoke a playbook from its entity page perhaps a Scaffolder action wouldn't be needed.
Could we make it possible to invoke a one-off run of the playbook from its entity page? I have a customer who seeks the ability to invoke a playbook from the Backstage UI.
This is related to #332 - if we could invoke a playbook from its entity page perhaps a Scaffolder action wouldn't be needed.
Also note that the template for running an Ansible job here doesn't directly invoke the job; rather it creates a Kubernetes CRD describing the job run and writes it to a "GitOps" repo. My ask is for a way to directly invoke the job without Kubernetes or GitOps, just Ansible.
This epic was scoped to list job templates in the software catalog and subsequently, we need to plan more features for the plugin for AAP.
@sonyccd / @Kasturi1820 / @cooktheryan FYI
there would need to be a host or container that would actually launch the playbook. With that host we would need to have credentials to connect to the object. This could be generally created but the ansible-playbook would have to run from a job pod which I'm not sure if that is completely supported from RH.
It may be worth talking with Christian Adams, Roger Lopez, or even someone like Sean Cavanaugh who would have insights with interacting with different components
Closing it as all linked tasks are in close state
|
gharchive/issue
| 2023-06-22T13:27:50 |
2025-04-01T04:34:39.337985
|
{
"authors": [
"cooktheryan",
"invincibleJai",
"joshgav",
"serenamarie125"
],
"repo": "janus-idp/backstage-plugins",
"url": "https://github.com/janus-idp/backstage-plugins/issues/481",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2126987659
|
[Test automation] Sign in using Google as authentication provider RHIDP-758
Description
[Test automation] Sign in using Google as authentication provider
Which issue(s) does this PR fix
https://issues.redhat.com/browse/RHIDP-758
PR acceptance criteria
Please make sure that the following steps are complete:
[ ] GitHub Actions are completed and successful
[ ] Unit Tests are updated and passing
[ ] E2E Tests are updated and passing
[ ] Documentation is updated if necessary (requirement for new features)
[ ] Add a screenshot if the change is UX/UI related
How to test changes / Special notes to the reviewer
/lgtm
/lgtm
/approve
/approve
/lgtm
|
gharchive/pull-request
| 2024-02-09T11:54:27 |
2025-04-01T04:34:39.341572
|
{
"authors": [
"gustavolira",
"josephca",
"kadel",
"subhashkhileri"
],
"repo": "janus-idp/backstage-showcase",
"url": "https://github.com/janus-idp/backstage-showcase/pull/958",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
92443460
|
option to hide node
Bitbucket: https://bitbucket.org/pypa/setuptools_scm/issue/16
Originally reported by: Ronny Pfannschmidt
Originally created at: 2015-02-18T13:42:55.041
make version construction pluggable
fixes #17
fixes #19
fixes #16
→ <<cset 08029d3bf341>>
Original comment by: Ronny Pfannschmidt
|
gharchive/issue
| 2015-07-01T17:40:51 |
2025-04-01T04:34:39.352111
|
{
"authors": [
"jaraco"
],
"repo": "jaraco/setuptools_scm",
"url": "https://github.com/jaraco/setuptools_scm/issues/35",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
329652469
|
Getting backpack working with Ava
Any advice on how to get around this?
Error
Seems to stem from Babel 6 (Backpack) & Babel 7 (Ava) incompatibility
Module build failed: Error: Plugin/Preset files are not allowed to export objects, only functions. In /...path.../node_modules/backpack-core/babel.js
Backpack Core versions tried
^0.7.0
^0.8.0-0
Relevant Resources
Plugin/Preset files are not allowed to export objects, only functions. · Issue #6808 · babel/babel · GitHub
This thread shows how to fix this error (ie. just upgrade babel from 6 -> 7)
Getting ready for Babel 7 · Issue #1598 · avajs/ava · GitHub
Relevant Ava plugins for Babel 7
Related Issues
Migration of Babel 7.x · Issue #106 · jaredpalmer/backpack · GitHub
Backpack thread about v0.8.0-0
Plugin/Preset files are not allowed to export objects, only functions
Found this duplicate after creating this issue
Config
.babelrc
{
"presets": [
"backpack-core/babel",
"@ava/stage-4",
"@ava/transform-test-files"
]
}
Looking for a hack-through
As suggested here, tried to see if forcing babel-preset-back from version 0.5.0 to 0.8.0-0 would resolve the issue. But still receiving the issue.
Control group
Sample repo of Ava working with Babel 7: https://github.com/servexyz/npm-starter
hello there, any update? how did you solve this?
|
gharchive/issue
| 2018-06-05T22:51:33 |
2025-04-01T04:34:39.364786
|
{
"authors": [
"alechp",
"sandeepDevJs"
],
"repo": "jaredpalmer/backpack",
"url": "https://github.com/jaredpalmer/backpack/issues/128",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
255122213
|
SCSS enhancement for local build
First off, thanks for all your work! Really love the ease of build using your framework!
I'd like to continue to use Jalpc for other projects and would like the capability of modifying SCSS file with live updating on the local server. I added my own custom.css file to the "static/assets/css" folder and added it to the file.conf.js file. Works great except that I am limited to CSS and that it doesn't offer live updating unless I'm missing something. When I make a change, I stop "bundle exec jekyll serve" (Windows 10, ugh...) and run "npm run build" to rebuild the minified CSS.
I really prefer using SCSS, which seems to be built in, but I haven't been able to get it to work on top of the serve command.
Requested Local Workflow:
Start jekyll serve/watch (using enhanced method)
_sass/_custom.scss - modify
CSS is compiled from SCSS (automatic)
CSS is minified into build (automatic)
Site updates live on local
Any help would be appreciated!
Thanks for this advice, SCSS is better for changing some styles with established styles, since Jekyll can auto generate static files to _site folder with command jekyll server [--watch] and Jekyll support SCSS, I think you can add your customised SCSS files like this example, but you need to add CSS files which compiled from SCSS files to HTML header and it can't be compressed by node package.
Later I will add this feature for others to customise Jalpc conveniently and make CSS compress work together with SCSS.
|
gharchive/issue
| 2017-09-04T22:03:54 |
2025-04-01T04:34:39.383502
|
{
"authors": [
"jarrekk",
"rosscyoga"
],
"repo": "jarrekk/Jalpc",
"url": "https://github.com/jarrekk/Jalpc/issues/85",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
119040442
|
Lost all saved project
Originally reported on Google Code with ID 298
What steps will reproduce the problem?
1.Try to load saved project from Drive
What is the expected output? What do you see instead?
See my saved projects
On what operating system, browser and browser version?
Seven
Please provide any additional information below.
Application version: 3.1.9
After the last update (Today) the Rest client dont found any saved project...
Reported by adriano.carolei on 2015-01-21 14:47:00
There's no much information here I can work with.
|
gharchive/issue
| 2015-11-26T13:00:34 |
2025-04-01T04:34:39.385050
|
{
"authors": [
"jarrodek"
],
"repo": "jarrodek/ChromeRestClient",
"url": "https://github.com/jarrodek/ChromeRestClient/issues/294",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
121990039
|
Use assert.raises instead of .throws
In some environments .throws() is a reserved word, so QUnit offers an alias .raises().
Using .raises seems safer, and also allows us to remove an eslint exception.
See the blue box about half way down this page for the documentation of this alias: http://api.qunitjs.com/throws/
We originally made the exact inverse of this change here, because QUnit deprecated raises: https://github.com/jashkenas/underscore/pull/1828
Looks like they've since brought it back: https://github.com/jquery/qunit/issues/663
Can you name a particular browser/version that has a problem with this?
The QUnit docs mention Closure Compiler, which is presumably not a problem for our tests. I think the main advantage here, is that is allows us to re-enable the eslint setting as a rule instead of a warning.
|
gharchive/pull-request
| 2015-12-14T07:33:21 |
2025-04-01T04:34:39.412088
|
{
"authors": [
"captbaritone",
"michaelficarra"
],
"repo": "jashkenas/underscore",
"url": "https://github.com/jashkenas/underscore/pull/2388",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
143143116
|
Fix debounce immediate
Fixes an issue with _.debounce when { immediate: true }. When
called twice within wait ms, we cleared the timeout. Because it is
truthy on the second run, it was cleared but timeout remained
truthy (on any later run, timeout remains truthy). Because timeout
was cleared, the later function is never run to null out timeout.
Anyways, we should always be setting a fresh timeout function when
{ immediate: true }, to prevent further calls to the debounced
function until the wait ms after the last call.
Fixes #2478, supersedes #2479. Thanks for the bug report @hanzichi.
Coverage increased (+0.1%) to 96.863% when pulling 882cb7fd1d848d6d4f33e3460d2bac6adf2d3f10 on jridgewell:debounce-immediate into 7d07d80ba8f4b1c8e2da13379cdb3886a8fca4c1 on jashkenas:master.
Fine with this change, but did that not pass before?
The old tests pass, but nothing tested calling the debounced function again after wait is over.
Coverage decreased (-0.004%) to 96.863% when pulling 96719954d086f8f77c5521698ac138f595c4cb9b on jridgewell:debounce-immediate into 669fb75e89587e36acc9bde453e1c7f00bc23b7e on jashkenas:master.
Is thee a way to tell in which version of underscore this is fixed?
@bent0b0x it was not in the released versions, it was fixed in the edge version, so the bug is not in any one of the released ones
|
gharchive/pull-request
| 2016-03-24T03:54:51 |
2025-04-01T04:34:39.417981
|
{
"authors": [
"bent0b0x",
"coveralls",
"hanzichi",
"jridgewell",
"megawac"
],
"repo": "jashkenas/underscore",
"url": "https://github.com/jashkenas/underscore/pull/2482",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
180492196
|
Assert: _.binding to a primitive returns a wrapped obj
Enhance out tests to more clearly document the behavior of binding a
primitive as the context of a function.
Note: The first commit just simplifies how these tests are written. The second commit actually changes what's being tested. See the individual commit messages for more details.
Coverage remained the same at 96.863% when pulling bfa74a2487ab0e5d5d98d3550012f33b3f8059ca on captbaritone:bind-primitive into a9432276b90bd23d3022deb89ac1ba6b10ee7495 on jashkenas:master.
|
gharchive/pull-request
| 2016-10-02T03:00:13 |
2025-04-01T04:34:39.420267
|
{
"authors": [
"captbaritone",
"coveralls"
],
"repo": "jashkenas/underscore",
"url": "https://github.com/jashkenas/underscore/pull/2595",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1403140332
|
Unexpected payload from device
I have 5 USB switch.
The device scan receives correct payload from four of them while for the last one I see "No response".
When I try to get the status from python I get Unexpected payload from device.
The device work perfectly with SmartLife and Amazon Alexa.
Any help please?
Thanks
ok solved, nevermind
Thanks @maxmonz69 - closing this.
|
gharchive/issue
| 2022-10-10T13:00:23 |
2025-04-01T04:34:39.441814
|
{
"authors": [
"jasonacox",
"maxmonz69"
],
"repo": "jasonacox/tinytuya",
"url": "https://github.com/jasonacox/tinytuya/issues/193",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
120720295
|
Understanding language problem
Since the last update Jasper is not supporting German language correctly with pocketsphinx.
English language so far is ok.
Jasper with German language:
pi@pi:~/jasper $ ./jasper.py --debug
*******************************************************
* JASPER - THE TALKING COMPUTER *
* (c) 2015 Shubhro Saha, Charlie Marsh & Jan Holthuis *
*******************************************************
DEBUG:client.diagnose:Checking network connection to server 'www.google.com'...
DEBUG:client.diagnose:Network connection working
DEBUG:__main__:Trying to read config file: '/home/pi/.jasper/profile.yml'
INFO:root:Using language 'de-DE'
INFO:root:audio_engine not specified in profile, using defaults.
DEBUG:root:Using Audio engine 'pyaudio'
DEBUG:root:Using STT engine 'sphinx'
DEBUG:root:Using passive STT engine 'sphinx'
DEBUG:root:Using TTS engine 'pico-tts'
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/speechhandler/birthday
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/speechhandler/birthday/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: birthday 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/speechhandler/mpdcontrol
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/speechhandler/mpdcontrol/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: mpdcontrol 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/speechhandler/gmail
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/speechhandler/gmail/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: gmail 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/speechhandler/weather
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/speechhandler/weather/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: weather 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/speechhandler/hackernews
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/speechhandler/hackernews/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: hn 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/speechhandler/news
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/speechhandler/news/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: news 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/speechhandler/notifications
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/speechhandler/notifications/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: notifications 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/speechhandler/joke
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/speechhandler/joke/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: joke 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/speechhandler/life
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/speechhandler/life/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: life 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/speechhandler/unclear
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/speechhandler/unclear/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: unclear 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/speechhandler/clock
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/speechhandler/clock/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: clock 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/stt/witai-stt
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/stt/witai-stt/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: witai-stt 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/stt/pocketsphinx-stt
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/stt/pocketsphinx-stt/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: sphinx 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/stt/google-stt
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/stt/google-stt/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: google 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/stt/att-stt
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/stt/att-stt/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: att-stt 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/stt/julius-stt
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/stt/julius-stt/plugin.info' parsed successfully!
DEBUG:client.diagnose:Checking executable 'julius'...
DEBUG:client.diagnose:Executable 'julius' not found
WARNING:client.pluginstore:Plugin at '/home/pi/jasper/plugins/stt/julius-stt' skipped! (Reason: Can't find julius executable)
Traceback (most recent call last):
File "/home/pi/jasper/client/pluginstore.py", line 155, in detect_plugins
plugin_info = self.parse_plugin(root)
File "/home/pi/jasper/client/pluginstore.py", line 183, in parse_plugin
self._categories_map.values())
File "/home/pi/jasper/client/pluginstore.py", line 51, in parse_plugin_class
("py", "r", imp.PKG_DIRECTORY))
File "/home/pi/jasper/plugins/stt/julius-stt/__init__.py", line 2, in <module>
from .julius import JuliusSTTPlugin
File "/home/pi/jasper/plugins/stt/julius-stt/julius.py", line 10, in <module>
raise ImportError("Can't find julius executable")
ImportError: Can't find julius executable
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/stt/kaldigstserver-stt
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/stt/kaldigstserver-stt/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: kaldigstserver-stt 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/tts/mary-tts
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/tts/mary-tts/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: mary-tts 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/tts/festival-tts
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/tts/festival-tts/plugin.info' parsed successfully!
DEBUG:client.diagnose:Checking executable 'text2wave'...
DEBUG:client.diagnose:Executable 'text2wave' not found
WARNING:client.pluginstore:Plugin at '/home/pi/jasper/plugins/tts/festival-tts' skipped! (Reason: Executables "text2wave" and/or "festival" not found!)
Traceback (most recent call last):
File "/home/pi/jasper/client/pluginstore.py", line 155, in detect_plugins
plugin_info = self.parse_plugin(root)
File "/home/pi/jasper/client/pluginstore.py", line 183, in parse_plugin
self._categories_map.values())
File "/home/pi/jasper/client/pluginstore.py", line 51, in parse_plugin_class
("py", "r", imp.PKG_DIRECTORY))
File "/home/pi/jasper/plugins/tts/festival-tts/__init__.py", line 2, in <module>
from .festival import FestivalTTSPlugin
File "/home/pi/jasper/plugins/tts/festival-tts/festival.py", line 10, in <module>
raise ImportError('Executables "text2wave" and/or "festival" not found!')
ImportError: Executables "text2wave" and/or "festival" not found!
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/tts/osx-tts
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/tts/osx-tts/plugin.info' parsed successfully!
WARNING:client.pluginstore:Plugin at '/home/pi/jasper/plugins/tts/osx-tts' skipped! (Reason: Invalid platform!)
Traceback (most recent call last):
File "/home/pi/jasper/client/pluginstore.py", line 155, in detect_plugins
plugin_info = self.parse_plugin(root)
File "/home/pi/jasper/client/pluginstore.py", line 183, in parse_plugin
self._categories_map.values())
File "/home/pi/jasper/client/pluginstore.py", line 51, in parse_plugin_class
("py", "r", imp.PKG_DIRECTORY))
File "/home/pi/jasper/plugins/tts/osx-tts/__init__.py", line 2, in <module>
from .osx import MacOSXTTSPlugin
File "/home/pi/jasper/plugins/tts/osx-tts/osx.py", line 13, in <module>
raise ImportError('Invalid platform!')
ImportError: Invalid platform!
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/tts/ivona-tts
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/tts/ivona-tts/plugin.info' parsed successfully!
WARNING:client.pluginstore:Plugin at '/home/pi/jasper/plugins/tts/ivona-tts' skipped! (Reason: No module named pyvona)
Traceback (most recent call last):
File "/home/pi/jasper/client/pluginstore.py", line 155, in detect_plugins
plugin_info = self.parse_plugin(root)
File "/home/pi/jasper/client/pluginstore.py", line 183, in parse_plugin
self._categories_map.values())
File "/home/pi/jasper/client/pluginstore.py", line 51, in parse_plugin_class
("py", "r", imp.PKG_DIRECTORY))
File "/home/pi/jasper/plugins/tts/ivona-tts/__init__.py", line 2, in <module>
from .ivona import IvonaTTSPlugin
File "/home/pi/jasper/plugins/tts/ivona-tts/ivona.py", line 4, in <module>
import pyvona
ImportError: No module named pyvona
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/tts/cereproc-tts
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/tts/cereproc-tts/plugin.info' parsed successfully!
WARNING:client.pluginstore:Plugin at '/home/pi/jasper/plugins/tts/cereproc-tts' skipped! (Reason: No module named suds)
Traceback (most recent call last):
File "/home/pi/jasper/client/pluginstore.py", line 155, in detect_plugins
plugin_info = self.parse_plugin(root)
File "/home/pi/jasper/client/pluginstore.py", line 183, in parse_plugin
self._categories_map.values())
File "/home/pi/jasper/client/pluginstore.py", line 51, in parse_plugin_class
("py", "r", imp.PKG_DIRECTORY))
File "/home/pi/jasper/plugins/tts/cereproc-tts/__init__.py", line 2, in <module>
from .cereproc import CereprocTTSPlugin
File "/home/pi/jasper/plugins/tts/cereproc-tts/cereproc.py", line 3, in <module>
import suds
ImportError: No module named suds
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/tts/flite-tts
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/tts/flite-tts/plugin.info' parsed successfully!
DEBUG:client.diagnose:Checking executable 'flite'...
DEBUG:client.diagnose:Executable 'flite' not found
WARNING:client.pluginstore:Plugin at '/home/pi/jasper/plugins/tts/flite-tts' skipped! (Reason: Executable 'flite' not found!)
Traceback (most recent call last):
File "/home/pi/jasper/client/pluginstore.py", line 155, in detect_plugins
plugin_info = self.parse_plugin(root)
File "/home/pi/jasper/client/pluginstore.py", line 183, in parse_plugin
self._categories_map.values())
File "/home/pi/jasper/client/pluginstore.py", line 51, in parse_plugin_class
("py", "r", imp.PKG_DIRECTORY))
File "/home/pi/jasper/plugins/tts/flite-tts/__init__.py", line 2, in <module>
from .flite import FliteTTSPlugin
File "/home/pi/jasper/plugins/tts/flite-tts/flite.py", line 12, in <module>
raise ImportError("Executable '%s' not found!" % EXECUTABLE)
ImportError: Executable 'flite' not found!
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/tts/pico-tts
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/tts/pico-tts/plugin.info' parsed successfully!
DEBUG:client.diagnose:Checking executable 'pico2wave'...
DEBUG:client.diagnose:Executable 'pico2wave' found: '/usr/bin/pico2wave'
DEBUG:client.pluginstore:Found valid plugin: pico-tts 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/tts/google-tts
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/tts/google-tts/plugin.info' parsed successfully!
WARNING:client.pluginstore:Plugin at '/home/pi/jasper/plugins/tts/google-tts' skipped! (Reason: No module named gtts)
Traceback (most recent call last):
File "/home/pi/jasper/client/pluginstore.py", line 155, in detect_plugins
plugin_info = self.parse_plugin(root)
File "/home/pi/jasper/client/pluginstore.py", line 183, in parse_plugin
self._categories_map.values())
File "/home/pi/jasper/client/pluginstore.py", line 51, in parse_plugin_class
("py", "r", imp.PKG_DIRECTORY))
File "/home/pi/jasper/plugins/tts/google-tts/__init__.py", line 2, in <module>
from .google import GoogleTTSPlugin
File "/home/pi/jasper/plugins/tts/google-tts/google.py", line 3, in <module>
import gtts
ImportError: No module named gtts
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/tts/espeak-tts
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/tts/espeak-tts/plugin.info' parsed successfully!
DEBUG:client.diagnose:Checking executable 'espeak'...
DEBUG:client.diagnose:Executable 'espeak' not found
WARNING:client.pluginstore:Plugin at '/home/pi/jasper/plugins/tts/espeak-tts' skipped! (Reason: espeak executable not found!)
Traceback (most recent call last):
File "/home/pi/jasper/client/pluginstore.py", line 155, in detect_plugins
plugin_info = self.parse_plugin(root)
File "/home/pi/jasper/client/pluginstore.py", line 183, in parse_plugin
self._categories_map.values())
File "/home/pi/jasper/client/pluginstore.py", line 51, in parse_plugin_class
("py", "r", imp.PKG_DIRECTORY))
File "/home/pi/jasper/plugins/tts/espeak-tts/__init__.py", line 2, in <module>
from .espeak import EspeakTTSPlugin
File "/home/pi/jasper/plugins/tts/espeak-tts/espeak.py", line 11, in <module>
raise ImportError("espeak executable not found!")
ImportError: espeak executable not found!
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/audioengine/alsa-ae
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/audioengine/alsa-ae/plugin.info' parsed successfully!
WARNING:client.pluginstore:Plugin at '/home/pi/jasper/plugins/audioengine/alsa-ae' skipped! (Reason: No module named alsaaudio)
Traceback (most recent call last):
File "/home/pi/jasper/client/pluginstore.py", line 155, in detect_plugins
plugin_info = self.parse_plugin(root)
File "/home/pi/jasper/client/pluginstore.py", line 183, in parse_plugin
self._categories_map.values())
File "/home/pi/jasper/client/pluginstore.py", line 51, in parse_plugin_class
("py", "r", imp.PKG_DIRECTORY))
File "/home/pi/jasper/plugins/audioengine/alsa-ae/__init__.py", line 2, in <module>
from .alsaaudioengine import AlsaAudioEnginePlugin
File "/home/pi/jasper/plugins/audioengine/alsa-ae/alsaaudioengine.py", line 4, in <module>
import alsaaudio
ImportError: No module named alsaaudio
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/audioengine/pyaudio-ae
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/audioengine/pyaudio-ae/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: pyaudio 1.0.0
INFO:pyaudio_1_0_0.pyaudioengine:Initializing PyAudio. ALSA/Jack error messages that pop up during this process are normal and can usually be safely ignored.
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.front.0:CARD=0'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM front
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.rear
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.center_lfe
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.side
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.surround51.0:CARD=0'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM surround21
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.surround51.0:CARD=0'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM surround21
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.surround40.0:CARD=0'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM surround40
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.surround51.0:CARD=0'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM surround41
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.surround51.0:CARD=0'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM surround50
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.surround51.0:CARD=0'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM surround51
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.surround71.0:CARD=0'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM surround71
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.iec958.0:CARD=0,AES0=4,AES1=130,AES2=0,AES3=2'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM iec958
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.iec958.0:CARD=0,AES0=4,AES1=130,AES2=0,AES3=2'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM spdif
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.iec958.0:CARD=0,AES0=4,AES1=130,AES2=0,AES3=2'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM spdif
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.hdmi
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.hdmi
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.modem
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.modem
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.phoneline
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.phoneline
Cannot connect to server socket err = No such file or directory
Cannot connect to server request channel
jack server is not running or cannot be started
INFO:pyaudio_1_0_0.pyaudioengine:Initialization of PyAudio engine finished
DEBUG:pyaudio_1_0_0.pyaudioengine:Found 6 PyAudio devices
DEBUG:pyaudio_1_0_0.pyaudioengine:Found 6 PyAudio devices
DEBUG:pyaudio_1_0_0.pyaudioengine:Found 6 PyAudio devices
DEBUG:pyaudio_1_0_0.pyaudioengine:Found 6 PyAudio devices
WARNING:__main__:Plugin 'hn' skipped! (Reason: Unsupported Language!)
Traceback (most recent call last):
File "./jasper.py", line 187, in __init__
plugin = info.plugin_class(info, self.config)
File "/home/pi/jasper/client/plugin.py", line 37, in __init__
self, self.info.translations, self.profile)
File "/home/pi/jasper/client/i18n.py", line 28, in __init__
self.__get_translations()
File "/home/pi/jasper/client/i18n.py", line 37, in __get_translations
raise ValueError('Unsupported Language!')
ValueError: Unsupported Language!
WARNING:sphinx_1_0_0.sphinxplugin:This STT plugin doesn't have multilanguage support!
DEBUG:client.vocabcompiler:compiled_revision is '6f7088b193d8c2b1494e4aedc19a8f1ee66d2300'
DEBUG:sphinx_1_0_0.sphinxplugin:Initializing PocketSphinx Decoder with hmm_dir '/home/pi/voxforge-de-r20141117/model_parameters/voxforge.cd_cont_3000'
WARNING:sphinx_1_0_0.sphinxplugin:This STT plugin doesn't have multilanguage support!
DEBUG:client.vocabcompiler:compiled_revision is 'bb74ae36d130ef20de710e3a77b43424b8fa774f'
DEBUG:sphinx_1_0_0.sphinxplugin:Initializing PocketSphinx Decoder with hmm_dir '/home/pi/voxforge-de-r20141117/model_parameters/voxforge.cd_cont_3000'
DEBUG:pico_tts_1_0_0.pico:Executing pico2wave -w /tmp/tmph0_zlq.wav -l de-DE 'Was kann ich für dich tun, Thomas?'
DEBUG:pyaudio_1_0_0.pyaudioengine:output stream opened on device 'default' (16000 Hz, 1 channel, 16 bit)
DEBUG:pyaudio_1_0_0.pyaudioengine:output stream closed on device 'default'
INFO:client.conversation:Starting to handle conversation with keyword 'JASPER'.
DEBUG:pyaudio_1_0_0.pyaudioengine:input stream opened on device 'default' (16000 Hz, 1 channel, 16 bit)
Jasper with English language:
DEBUG:client.diagnose:Executable 'julius' not found
WARNING:client.pluginstore:Plugin at '/home/pi/jasper/plugins/stt/julius-stt' skipped! (Reason: Can't find julius executable)
Traceback (most recent call last):
File "/home/pi/jasper/client/pluginstore.py", line 155, in detect_plugins
plugin_info = self.parse_plugin(root)
File "/home/pi/jasper/client/pluginstore.py", line 183, in parse_plugin
self._categories_map.values())
File "/home/pi/jasper/client/pluginstore.py", line 51, in parse_plugin_class
("py", "r", imp.PKG_DIRECTORY))
File "/home/pi/jasper/plugins/stt/julius-stt/__init__.py", line 2, in <module>
from .julius import JuliusSTTPlugin
File "/home/pi/jasper/plugins/stt/julius-stt/julius.py", line 10, in <module>
raise ImportError("Can't find julius executable")
ImportError: Can't find julius executable
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/stt/kaldigstserver-stt
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/stt/kaldigstserver-stt/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: kaldigstserver-stt 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/tts/mary-tts
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/tts/mary-tts/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: mary-tts 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/tts/festival-tts
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/tts/festival-tts/plugin.info' parsed successfully!
DEBUG:client.diagnose:Checking executable 'text2wave'...
DEBUG:client.diagnose:Executable 'text2wave' not found
WARNING:client.pluginstore:Plugin at '/home/pi/jasper/plugins/tts/festival-tts' skipped! (Reason: Executables "text2wave" and/or "festival" not found!)
Traceback (most recent call last):
File "/home/pi/jasper/client/pluginstore.py", line 155, in detect_plugins
plugin_info = self.parse_plugin(root)
File "/home/pi/jasper/client/pluginstore.py", line 183, in parse_plugin
self._categories_map.values())
File "/home/pi/jasper/client/pluginstore.py", line 51, in parse_plugin_class
("py", "r", imp.PKG_DIRECTORY))
File "/home/pi/jasper/plugins/tts/festival-tts/__init__.py", line 2, in <module>
from .festival import FestivalTTSPlugin
File "/home/pi/jasper/plugins/tts/festival-tts/festival.py", line 10, in <module>
raise ImportError('Executables "text2wave" and/or "festival" not found!')
ImportError: Executables "text2wave" and/or "festival" not found!
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/tts/osx-tts
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/tts/osx-tts/plugin.info' parsed successfully!
WARNING:client.pluginstore:Plugin at '/home/pi/jasper/plugins/tts/osx-tts' skipped! (Reason: Invalid platform!)
Traceback (most recent call last):
File "/home/pi/jasper/client/pluginstore.py", line 155, in detect_plugins
plugin_info = self.parse_plugin(root)
File "/home/pi/jasper/client/pluginstore.py", line 183, in parse_plugin
self._categories_map.values())
File "/home/pi/jasper/client/pluginstore.py", line 51, in parse_plugin_class
("py", "r", imp.PKG_DIRECTORY))
File "/home/pi/jasper/plugins/tts/osx-tts/__init__.py", line 2, in <module>
from .osx import MacOSXTTSPlugin
File "/home/pi/jasper/plugins/tts/osx-tts/osx.py", line 13, in <module>
raise ImportError('Invalid platform!')
ImportError: Invalid platform!
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/tts/ivona-tts
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/tts/ivona-tts/plugin.info' parsed successfully!
WARNING:client.pluginstore:Plugin at '/home/pi/jasper/plugins/tts/ivona-tts' skipped! (Reason: No module named pyvona)
Traceback (most recent call last):
File "/home/pi/jasper/client/pluginstore.py", line 155, in detect_plugins
plugin_info = self.parse_plugin(root)
File "/home/pi/jasper/client/pluginstore.py", line 183, in parse_plugin
self._categories_map.values())
File "/home/pi/jasper/client/pluginstore.py", line 51, in parse_plugin_class
("py", "r", imp.PKG_DIRECTORY))
File "/home/pi/jasper/plugins/tts/ivona-tts/__init__.py", line 2, in <module>
from .ivona import IvonaTTSPlugin
File "/home/pi/jasper/plugins/tts/ivona-tts/ivona.py", line 4, in <module>
import pyvona
ImportError: No module named pyvona
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/tts/cereproc-tts
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/tts/cereproc-tts/plugin.info' parsed successfully!
WARNING:client.pluginstore:Plugin at '/home/pi/jasper/plugins/tts/cereproc-tts' skipped! (Reason: No module named suds)
Traceback (most recent call last):
File "/home/pi/jasper/client/pluginstore.py", line 155, in detect_plugins
plugin_info = self.parse_plugin(root)
File "/home/pi/jasper/client/pluginstore.py", line 183, in parse_plugin
self._categories_map.values())
File "/home/pi/jasper/client/pluginstore.py", line 51, in parse_plugin_class
("py", "r", imp.PKG_DIRECTORY))
File "/home/pi/jasper/plugins/tts/cereproc-tts/__init__.py", line 2, in <module>
from .cereproc import CereprocTTSPlugin
File "/home/pi/jasper/plugins/tts/cereproc-tts/cereproc.py", line 3, in <module>
import suds
ImportError: No module named suds
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/tts/flite-tts
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/tts/flite-tts/plugin.info' parsed successfully!
DEBUG:client.diagnose:Checking executable 'flite'...
DEBUG:client.diagnose:Executable 'flite' not found
WARNING:client.pluginstore:Plugin at '/home/pi/jasper/plugins/tts/flite-tts' skipped! (Reason: Executable 'flite' not found!)
Traceback (most recent call last):
File "/home/pi/jasper/client/pluginstore.py", line 155, in detect_plugins
plugin_info = self.parse_plugin(root)
File "/home/pi/jasper/client/pluginstore.py", line 183, in parse_plugin
self._categories_map.values())
File "/home/pi/jasper/client/pluginstore.py", line 51, in parse_plugin_class
("py", "r", imp.PKG_DIRECTORY))
File "/home/pi/jasper/plugins/tts/flite-tts/__init__.py", line 2, in <module>
from .flite import FliteTTSPlugin
File "/home/pi/jasper/plugins/tts/flite-tts/flite.py", line 12, in <module>
raise ImportError("Executable '%s' not found!" % EXECUTABLE)
ImportError: Executable 'flite' not found!
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/tts/pico-tts
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/tts/pico-tts/plugin.info' parsed successfully!
DEBUG:client.diagnose:Checking executable 'pico2wave'...
DEBUG:client.diagnose:Executable 'pico2wave' found: '/usr/bin/pico2wave'
DEBUG:client.pluginstore:Found valid plugin: pico-tts 1.0.0
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/tts/google-tts
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/tts/google-tts/plugin.info' parsed successfully!
WARNING:client.pluginstore:Plugin at '/home/pi/jasper/plugins/tts/google-tts' skipped! (Reason: No module named gtts)
Traceback (most recent call last):
File "/home/pi/jasper/client/pluginstore.py", line 155, in detect_plugins
plugin_info = self.parse_plugin(root)
File "/home/pi/jasper/client/pluginstore.py", line 183, in parse_plugin
self._categories_map.values())
File "/home/pi/jasper/client/pluginstore.py", line 51, in parse_plugin_class
("py", "r", imp.PKG_DIRECTORY))
File "/home/pi/jasper/plugins/tts/google-tts/__init__.py", line 2, in <module>
from .google import GoogleTTSPlugin
File "/home/pi/jasper/plugins/tts/google-tts/google.py", line 3, in <module>
import gtts
ImportError: No module named gtts
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/tts/espeak-tts
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/tts/espeak-tts/plugin.info' parsed successfully!
DEBUG:client.diagnose:Checking executable 'espeak'...
DEBUG:client.diagnose:Executable 'espeak' not found
WARNING:client.pluginstore:Plugin at '/home/pi/jasper/plugins/tts/espeak-tts' skipped! (Reason: espeak executable not found!)
Traceback (most recent call last):
File "/home/pi/jasper/client/pluginstore.py", line 155, in detect_plugins
plugin_info = self.parse_plugin(root)
File "/home/pi/jasper/client/pluginstore.py", line 183, in parse_plugin
self._categories_map.values())
File "/home/pi/jasper/client/pluginstore.py", line 51, in parse_plugin_class
("py", "r", imp.PKG_DIRECTORY))
File "/home/pi/jasper/plugins/tts/espeak-tts/__init__.py", line 2, in <module>
from .espeak import EspeakTTSPlugin
File "/home/pi/jasper/plugins/tts/espeak-tts/espeak.py", line 11, in <module>
raise ImportError("espeak executable not found!")
ImportError: espeak executable not found!
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/audioengine/alsa-ae
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/audioengine/alsa-ae/plugin.info' parsed successfully!
WARNING:client.pluginstore:Plugin at '/home/pi/jasper/plugins/audioengine/alsa-ae' skipped! (Reason: No module named alsaaudio)
Traceback (most recent call last):
File "/home/pi/jasper/client/pluginstore.py", line 155, in detect_plugins
plugin_info = self.parse_plugin(root)
File "/home/pi/jasper/client/pluginstore.py", line 183, in parse_plugin
self._categories_map.values())
File "/home/pi/jasper/client/pluginstore.py", line 51, in parse_plugin_class
("py", "r", imp.PKG_DIRECTORY))
File "/home/pi/jasper/plugins/audioengine/alsa-ae/__init__.py", line 2, in <module>
from .alsaaudioengine import AlsaAudioEnginePlugin
File "/home/pi/jasper/plugins/audioengine/alsa-ae/alsaaudioengine.py", line 4, in <module>
import alsaaudio
ImportError: No module named alsaaudio
DEBUG:client.pluginstore:Found plugin candidate at: /home/pi/jasper/plugins/audioengine/pyaudio-ae
DEBUG:client.pluginstore:Plugin info file '/home/pi/jasper/plugins/audioengine/pyaudio-ae/plugin.info' parsed successfully!
DEBUG:client.pluginstore:Found valid plugin: pyaudio 1.0.0
INFO:pyaudio_1_0_0.pyaudioengine:Initializing PyAudio. ALSA/Jack error messages that pop up during this process are normal and can usually be safely ignored.
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.front.0:CARD=0'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM front
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.rear
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.center_lfe
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.side
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.surround51.0:CARD=0'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM surround21
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.surround51.0:CARD=0'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM surround21
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.surround40.0:CARD=0'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM surround40
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.surround51.0:CARD=0'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM surround41
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.surround51.0:CARD=0'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM surround50
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.surround51.0:CARD=0'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM surround51
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.surround71.0:CARD=0'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM surround71
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.iec958.0:CARD=0,AES0=4,AES1=130,AES2=0,AES3=2'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM iec958
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.iec958.0:CARD=0,AES0=4,AES1=130,AES2=0,AES3=2'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM spdif
ALSA lib confmisc.c:1286:(snd_func_refer) Unable to find definition 'cards.bcm2835.pcm.iec958.0:CARD=0,AES0=4,AES1=130,AES2=0,AES3=2'
ALSA lib conf.c:4259:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
ALSA lib conf.c:4738:(snd_config_expand) Evaluate error: No such file or directory
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM spdif
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.hdmi
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.hdmi
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.modem
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.modem
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.phoneline
ALSA lib pcm.c:2239:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.phoneline
Cannot connect to server socket err = No such file or directory
Cannot connect to server request channel
jack server is not running or cannot be started
INFO:pyaudio_1_0_0.pyaudioengine:Initialization of PyAudio engine finished
DEBUG:pyaudio_1_0_0.pyaudioengine:Found 6 PyAudio devices
DEBUG:pyaudio_1_0_0.pyaudioengine:Found 6 PyAudio devices
DEBUG:pyaudio_1_0_0.pyaudioengine:Found 6 PyAudio devices
DEBUG:pyaudio_1_0_0.pyaudioengine:Found 6 PyAudio devices
WARNING:sphinx_1_0_0.sphinxplugin:This STT plugin doesn't have multilanguage support!
DEBUG:client.vocabcompiler:Vocabulary dir '/home/pi/.jasper/vocabularies/en-US/sphinx/default' does not exist, creating...
INFO:client.vocabcompiler:Starting compilation...
DEBUG:sphinx_1_0_0.g2p:Using FST model: '/home/pi/phonetisaurus/g014b2b.fst'
DEBUG:sphinx_1_0_0.g2p:Will use the 3 best results.
DEBUG:sphinx_1_0_0.sphinxvocab:Languagemodel path: /home/pi/.jasper/vocabularies/en-US/sphinx/default/languagemodel
DEBUG:sphinx_1_0_0.sphinxvocab:Dictionary path: /home/pi/.jasper/vocabularies/en-US/sphinx/default/dictionary
DEBUG:sphinx_1_0_0.sphinxvocab:Compiling languagemodel...
DEBUG:sphinx_1_0_0.sphinxvocab:Creating vocab file: '/tmp/tmpohgwZy.vocab'
DEBUG:cmuclmtk:text2wfreq : Reading text from standard input...
DEBUG:cmuclmtk:text2wfreq : Done.
DEBUG:cmuclmtk:Command 'text2wfreq -hash 1000000 -verbosity 2' returned with exit code '0'.
DEBUG:cmuclmtk:wfreq2vocab : Will generate a vocabulary containing the most
DEBUG:cmuclmtk:frequent 20000 words. Reading wfreq stream from stdin...
DEBUG:cmuclmtk:wfreq2vocab : Done.
DEBUG:cmuclmtk:Command 'wfreq2vocab -verbosity 2 -records 1000000' returned with exit code '0'.
DEBUG:sphinx_1_0_0.sphinxvocab:Getting words from vocab file and removing it afterwards...
DEBUG:sphinx_1_0_0.sphinxvocab:Creating languagemodel file: '/home/pi/.jasper/vocabularies/en-US/sphinx/default/languagemodel'
DEBUG:cmuclmtk:text2idngram
DEBUG:cmuclmtk:Vocab : /tmp/tmpohgwZy.vocab
DEBUG:cmuclmtk:Output idngram : /tmp/tmp8IZDm0.idngram
DEBUG:cmuclmtk:N-gram buffer size : 100
DEBUG:cmuclmtk:Hash table size : 2000000
DEBUG:cmuclmtk:Temp directory : cmuclmtk-C1mWCk
DEBUG:cmuclmtk:Max open files : 20
DEBUG:cmuclmtk:FOF size : 10
DEBUG:cmuclmtk:n : 3
DEBUG:cmuclmtk:Initialising hash table...
DEBUG:cmuclmtk:Reading vocabulary...
DEBUG:cmuclmtk:Allocating memory for the n-gram buffer...
DEBUG:cmuclmtk:Reading text into the n-gram buffer...
DEBUG:cmuclmtk:20,000 n-grams processed for each ".", 1,000,000 for each line.
DEBUG:cmuclmtk:Sorting n-grams...
DEBUG:cmuclmtk:Writing sorted n-grams to temporary file cmuclmtk-C1mWCk/1
DEBUG:cmuclmtk:Merging 1 temporary files...
DEBUG:cmuclmtk:2-grams occurring: N times > N times Sug. -spec_num value
DEBUG:cmuclmtk:0 45 55
DEBUG:cmuclmtk:1 44 1 11
DEBUG:cmuclmtk:2 0 1 11
DEBUG:cmuclmtk:3 0 1 11
DEBUG:cmuclmtk:4 0 1 11
DEBUG:cmuclmtk:5 0 1 11
DEBUG:cmuclmtk:6 0 1 11
DEBUG:cmuclmtk:7 0 1 11
DEBUG:cmuclmtk:8 0 1 11
DEBUG:cmuclmtk:9 0 1 11
DEBUG:cmuclmtk:10 0 1 11
DEBUG:cmuclmtk:3-grams occurring: N times > N times Sug. -spec_num value
DEBUG:cmuclmtk:0 64 74
DEBUG:cmuclmtk:1 64 0 10
DEBUG:cmuclmtk:2 0 0 10
DEBUG:cmuclmtk:3 0 0 10
DEBUG:cmuclmtk:4 0 0 10
DEBUG:cmuclmtk:5 0 0 10
DEBUG:cmuclmtk:6 0 0 10
DEBUG:cmuclmtk:7 0 0 10
DEBUG:cmuclmtk:8 0 0 10
DEBUG:cmuclmtk:9 0 0 10
DEBUG:cmuclmtk:10 0 0 10
DEBUG:cmuclmtk:text2idngram : Done.
DEBUG:cmuclmtk:Command 'text2idngram -vocab /tmp/tmpohgwZy.vocab -idngram /tmp/tmp8IZDm0.idngram -buffer 100 -hash 2000000 -files 20 -verbosity 2 -n 3 -fof_size 10' returned with exit code '0'.
DEBUG:cmuclmtk:Warning : OOV fraction specified, but will not be used, since vocab type is not 2.
DEBUG:cmuclmtk:n : 3
DEBUG:cmuclmtk:Input file : /tmp/tmp8IZDm0.idngram (binary format)
DEBUG:cmuclmtk:Output files :
DEBUG:cmuclmtk:ARPA format : /home/pi/.jasper/vocabularies/en-US/sphinx/default/languagemodel
DEBUG:cmuclmtk:Vocabulary file : /tmp/tmpohgwZy.vocab
DEBUG:cmuclmtk:Cutoffs :
DEBUG:cmuclmtk:2-gram : 0 3-gram : 0
DEBUG:cmuclmtk:Vocabulary type : Open - type 1
DEBUG:cmuclmtk:Minimum unigram count : 0
DEBUG:cmuclmtk:Zeroton fraction : 1
DEBUG:cmuclmtk:Counts will be stored in two bytes.
DEBUG:cmuclmtk:Count table size : 65535
DEBUG:cmuclmtk:Discounting method : Good-Turing
DEBUG:cmuclmtk:Discounting ranges :
DEBUG:cmuclmtk:1-gram : 1 2-gram : 7 3-gram : 7
DEBUG:cmuclmtk:Memory allocation for tree structure :
DEBUG:cmuclmtk:Allocate 100 MB of memory, shared equally between all n-gram tables.
DEBUG:cmuclmtk:Back-off weight storage :
DEBUG:cmuclmtk:Back-off weights will be stored in four bytes.
DEBUG:cmuclmtk:Reading vocabulary.
DEBUG:cmuclmtk:read_wlist_into_siht: a list of 25 words was read from "/tmp/tmpohgwZy.vocab".
DEBUG:cmuclmtk:read_wlist_into_array: a list of 25 words was read from "/tmp/tmpohgwZy.vocab".
DEBUG:cmuclmtk:WARNING: <s> appears as a vocabulary item, but is not labelled as a
DEBUG:cmuclmtk:context cue.
DEBUG:cmuclmtk:Allocated space for 3571428 2-grams.
DEBUG:cmuclmtk:Allocated space for 8333333 3-grams.
DEBUG:cmuclmtk:table_size 26
DEBUG:cmuclmtk:Allocated 57142848 bytes to table for 2-grams.
DEBUG:cmuclmtk:Allocated (2+33333332) bytes to table for 3-grams.
DEBUG:cmuclmtk:Processing id n-gram file.
DEBUG:cmuclmtk:20,000 n-grams processed for each ".", 1,000,000 for each line.
DEBUG:cmuclmtk:Calculating discounted counts.
DEBUG:cmuclmtk:Warning : 1-gram : Discounting range is 1; setting P(zeroton)=P(singleton).
DEBUG:cmuclmtk:Discounted value : 0.91
DEBUG:cmuclmtk:Warning : 2-gram : GT statistics are out of range; lowering cutoff to 6.
DEBUG:cmuclmtk:Warning : 2-gram : GT statistics are out of range; lowering cutoff to 5.
DEBUG:cmuclmtk:Warning : 2-gram : GT statistics are out of range; lowering cutoff to 4.
DEBUG:cmuclmtk:Warning : 2-gram : GT statistics are out of range; lowering cutoff to 3.
DEBUG:cmuclmtk:Warning : 2-gram : GT statistics are out of range; lowering cutoff to 2.
DEBUG:cmuclmtk:Warning : 2-gram : GT statistics are out of range; lowering cutoff to 1.
DEBUG:cmuclmtk:Warning : 2-gram : Discounting range of 1 is equivalent to excluding
DEBUG:cmuclmtk:singletons.
DEBUG:cmuclmtk:Warning : 2-gram : GT statistics are out of range; lowering cutoff to 0.
DEBUG:cmuclmtk:Warning : 2-gram : Discounting is disabled.
DEBUG:cmuclmtk:Warning : 3-gram : GT statistics are out of range; lowering cutoff to 6.
DEBUG:cmuclmtk:Warning : 3-gram : GT statistics are out of range; lowering cutoff to 5.
DEBUG:cmuclmtk:Warning : 3-gram : GT statistics are out of range; lowering cutoff to 4.
DEBUG:cmuclmtk:Warning : 3-gram : GT statistics are out of range; lowering cutoff to 3.
DEBUG:cmuclmtk:Warning : 3-gram : GT statistics are out of range; lowering cutoff to 2.
DEBUG:cmuclmtk:Warning : 3-gram : GT statistics are out of range; lowering cutoff to 1.
DEBUG:cmuclmtk:Warning : 3-gram : Discounting range of 1 is equivalent to excluding
DEBUG:cmuclmtk:singletons.
DEBUG:cmuclmtk:Warning : 3-gram : GT statistics are out of range; lowering cutoff to 0.
DEBUG:cmuclmtk:Warning : 3-gram : Discounting is disabled.
DEBUG:cmuclmtk:Unigrams's discount mass is 0.0285326 (n1/N = 0.328125)
DEBUG:cmuclmtk:2 zerotons, P(zeroton) = 0.0142663 P(singleton) = 0.0142663
DEBUG:cmuclmtk:prob[UNK] = 0.0142663
DEBUG:cmuclmtk:Incrementing contexts...
DEBUG:cmuclmtk:Calculating back-off weights...
DEBUG:cmuclmtk:Writing out language model...
DEBUG:cmuclmtk:ARPA-style 3-gram will be written to /home/pi/.jasper/vocabularies/en-US/sphinx/default/languagemodel
DEBUG:cmuclmtk:idngram2lm : Done.
DEBUG:cmuclmtk:Command 'idngram2lm -idngram /tmp/tmp8IZDm0.idngram -vocab /tmp/tmpohgwZy.vocab -vocab_type 1 -oov_fraction 0.5 -min_unicount 0 -verbosity 2 -n 3 -arpa /home/pi/.jasper/vocabularies/en-US/sphinx/default/languagemodel -bin_input' returned with exit code '0'.
DEBUG:sphinx_1_0_0.sphinxvocab:Starting dictionary...
DEBUG:sphinx_1_0_0.sphinxvocab:Getting phonemes for 23 words...
DEBUG:sphinx_1_0_0.g2p:Converting 23 words to phonemes
DEBUG:sphinx_1_0_0.g2p:G2P conversion returned phonemes for 23 words
DEBUG:sphinx_1_0_0.sphinxvocab:Creating dict file: '/home/pi/.jasper/vocabularies/en-US/sphinx/default/dictionary'
INFO:client.vocabcompiler:Compilation done.
DEBUG:sphinx_1_0_0.sphinxplugin:Initializing PocketSphinx Decoder with hmm_dir '/home/pi/voxforge-de-r20141117/model_parameters/voxforge.cd_cont_3000'
WARNING:sphinx_1_0_0.sphinxplugin:This STT plugin doesn't have multilanguage support!
DEBUG:client.vocabcompiler:Vocabulary dir '/home/pi/.jasper/vocabularies/en-US/sphinx/keyword' does not exist, creating...
INFO:client.vocabcompiler:Starting compilation...
DEBUG:sphinx_1_0_0.g2p:Using FST model: '/home/pi/phonetisaurus/g014b2b.fst'
DEBUG:sphinx_1_0_0.g2p:Will use the 3 best results.
DEBUG:sphinx_1_0_0.sphinxvocab:Languagemodel path: /home/pi/.jasper/vocabularies/en-US/sphinx/keyword/languagemodel
DEBUG:sphinx_1_0_0.sphinxvocab:Dictionary path: /home/pi/.jasper/vocabularies/en-US/sphinx/keyword/dictionary
DEBUG:sphinx_1_0_0.sphinxvocab:Compiling languagemodel...
DEBUG:sphinx_1_0_0.sphinxvocab:Creating vocab file: '/tmp/tmpzFY3fn.vocab'
DEBUG:cmuclmtk:text2wfreq : Reading text from standard input...
DEBUG:cmuclmtk:text2wfreq : Done.
DEBUG:cmuclmtk:Command 'text2wfreq -hash 1000000 -verbosity 2' returned with exit code '0'.
DEBUG:cmuclmtk:wfreq2vocab : Will generate a vocabulary containing the most
DEBUG:cmuclmtk:frequent 20000 words. Reading wfreq stream from stdin...
DEBUG:cmuclmtk:wfreq2vocab : Done.
DEBUG:cmuclmtk:Command 'wfreq2vocab -verbosity 2 -records 1000000' returned with exit code '0'.
DEBUG:sphinx_1_0_0.sphinxvocab:Getting words from vocab file and removing it afterwards...
DEBUG:sphinx_1_0_0.sphinxvocab:Creating languagemodel file: '/home/pi/.jasper/vocabularies/en-US/sphinx/keyword/languagemodel'
DEBUG:cmuclmtk:text2idngram
DEBUG:cmuclmtk:Vocab : /tmp/tmpzFY3fn.vocab
DEBUG:cmuclmtk:Output idngram : /tmp/tmpVrExHA.idngram
DEBUG:cmuclmtk:N-gram buffer size : 100
DEBUG:cmuclmtk:Hash table size : 2000000
DEBUG:cmuclmtk:Temp directory : cmuclmtk-QOmJyz
DEBUG:cmuclmtk:Max open files : 20
DEBUG:cmuclmtk:FOF size : 10
DEBUG:cmuclmtk:n : 3
DEBUG:cmuclmtk:Initialising hash table...
DEBUG:cmuclmtk:Reading vocabulary...
DEBUG:cmuclmtk:Allocating memory for the n-gram buffer...
DEBUG:cmuclmtk:Reading text into the n-gram buffer...
DEBUG:cmuclmtk:20,000 n-grams processed for each ".", 1,000,000 for each line.
DEBUG:cmuclmtk:Sorting n-grams...
DEBUG:cmuclmtk:Writing sorted n-grams to temporary file cmuclmtk-QOmJyz/1
DEBUG:cmuclmtk:Merging 1 temporary files...
DEBUG:cmuclmtk:2-grams occurring: N times > N times Sug. -spec_num value
DEBUG:cmuclmtk:0 36 46
DEBUG:cmuclmtk:1 35 1 11
DEBUG:cmuclmtk:2 0 1 11
DEBUG:cmuclmtk:3 0 1 11
DEBUG:cmuclmtk:4 0 1 11
DEBUG:cmuclmtk:5 0 1 11
DEBUG:cmuclmtk:6 0 1 11
DEBUG:cmuclmtk:7 0 1 11
DEBUG:cmuclmtk:8 0 1 11
DEBUG:cmuclmtk:9 0 1 11
DEBUG:cmuclmtk:10 0 1 11
DEBUG:cmuclmtk:3-grams occurring: N times > N times Sug. -spec_num value
DEBUG:cmuclmtk:0 52 62
DEBUG:cmuclmtk:1 52 0 10
DEBUG:cmuclmtk:2 0 0 10
DEBUG:cmuclmtk:3 0 0 10
DEBUG:cmuclmtk:4 0 0 10
DEBUG:cmuclmtk:5 0 0 10
DEBUG:cmuclmtk:6 0 0 10
DEBUG:cmuclmtk:7 0 0 10
DEBUG:cmuclmtk:8 0 0 10
DEBUG:cmuclmtk:9 0 0 10
DEBUG:cmuclmtk:10 0 0 10
DEBUG:cmuclmtk:text2idngram : Done.
DEBUG:cmuclmtk:Command 'text2idngram -vocab /tmp/tmpzFY3fn.vocab -idngram /tmp/tmpVrExHA.idngram -buffer 100 -hash 2000000 -files 20 -verbosity 2 -n 3 -fof_size 10' returned with exit code '0'.
DEBUG:cmuclmtk:Warning : OOV fraction specified, but will not be used, since vocab type is not 2.
DEBUG:cmuclmtk:n : 3
DEBUG:cmuclmtk:Input file : /tmp/tmpVrExHA.idngram (binary format)
DEBUG:cmuclmtk:Output files :
DEBUG:cmuclmtk:ARPA format : /home/pi/.jasper/vocabularies/en-US/sphinx/keyword/languagemodel
DEBUG:cmuclmtk:Vocabulary file : /tmp/tmpzFY3fn.vocab
DEBUG:cmuclmtk:Cutoffs :
DEBUG:cmuclmtk:2-gram : 0 3-gram : 0
DEBUG:cmuclmtk:Vocabulary type : Open - type 1
DEBUG:cmuclmtk:Minimum unigram count : 0
DEBUG:cmuclmtk:Zeroton fraction : 1
DEBUG:cmuclmtk:Counts will be stored in two bytes.
DEBUG:cmuclmtk:Count table size : 65535
DEBUG:cmuclmtk:Discounting method : Good-Turing
DEBUG:cmuclmtk:Discounting ranges :
DEBUG:cmuclmtk:1-gram : 1 2-gram : 7 3-gram : 7
DEBUG:cmuclmtk:Memory allocation for tree structure :
DEBUG:cmuclmtk:Allocate 100 MB of memory, shared equally between all n-gram tables.
DEBUG:cmuclmtk:Back-off weight storage :
DEBUG:cmuclmtk:Back-off weights will be stored in four bytes.
DEBUG:cmuclmtk:Reading vocabulary.
DEBUG:cmuclmtk:read_wlist_into_siht: a list of 20 words was read from "/tmp/tmpzFY3fn.vocab".
DEBUG:cmuclmtk:read_wlist_into_array: a list of 20 words was read from "/tmp/tmpzFY3fn.vocab".
DEBUG:cmuclmtk:WARNING: <s> appears as a vocabulary item, but is not labelled as a
DEBUG:cmuclmtk:context cue.
DEBUG:cmuclmtk:Allocated space for 3571428 2-grams.
DEBUG:cmuclmtk:Allocated space for 8333333 3-grams.
DEBUG:cmuclmtk:table_size 21
DEBUG:cmuclmtk:Allocated 57142848 bytes to table for 2-grams.
DEBUG:cmuclmtk:Allocated (2+33333332) bytes to table for 3-grams.
DEBUG:cmuclmtk:Processing id n-gram file.
DEBUG:cmuclmtk:20,000 n-grams processed for each ".", 1,000,000 for each line.
DEBUG:cmuclmtk:Calculating discounted counts.
DEBUG:cmuclmtk:Warning : 1-gram : Discounting range is 1; setting P(zeroton)=P(singleton).
DEBUG:cmuclmtk:Discounted value : 0.89
DEBUG:cmuclmtk:Warning : 2-gram : GT statistics are out of range; lowering cutoff to 6.
DEBUG:cmuclmtk:Warning : 2-gram : GT statistics are out of range; lowering cutoff to 5.
DEBUG:cmuclmtk:Warning : 2-gram : GT statistics are out of range; lowering cutoff to 4.
DEBUG:cmuclmtk:Warning : 2-gram : GT statistics are out of range; lowering cutoff to 3.
DEBUG:cmuclmtk:Warning : 2-gram : GT statistics are out of range; lowering cutoff to 2.
DEBUG:cmuclmtk:Warning : 2-gram : GT statistics are out of range; lowering cutoff to 1.
DEBUG:cmuclmtk:Warning : 2-gram : Discounting range of 1 is equivalent to excluding
DEBUG:cmuclmtk:singletons.
DEBUG:cmuclmtk:Warning : 2-gram : GT statistics are out of range; lowering cutoff to 0.
DEBUG:cmuclmtk:Warning : 2-gram : Discounting is disabled.
DEBUG:cmuclmtk:Warning : 3-gram : GT statistics are out of range; lowering cutoff to 6.
DEBUG:cmuclmtk:Warning : 3-gram : GT statistics are out of range; lowering cutoff to 5.
DEBUG:cmuclmtk:Warning : 3-gram : GT statistics are out of range; lowering cutoff to 4.
DEBUG:cmuclmtk:Warning : 3-gram : GT statistics are out of range; lowering cutoff to 3.
DEBUG:cmuclmtk:Warning : 3-gram : GT statistics are out of range; lowering cutoff to 2.
DEBUG:cmuclmtk:Warning : 3-gram : GT statistics are out of range; lowering cutoff to 1.
DEBUG:cmuclmtk:Warning : 3-gram : Discounting range of 1 is equivalent to excluding
DEBUG:cmuclmtk:singletons.
DEBUG:cmuclmtk:Warning : 3-gram : GT statistics are out of range; lowering cutoff to 0.
DEBUG:cmuclmtk:Warning : 3-gram : Discounting is disabled.
DEBUG:cmuclmtk:Unigrams's discount mass is 0.034413 (n1/N = 0.326923)
DEBUG:cmuclmtk:2 zerotons, P(zeroton) = 0.0172065 P(singleton) = 0.0172065
DEBUG:cmuclmtk:P(zeroton) was reduced to 0.0172064774 (1.000 of P(singleton))
DEBUG:cmuclmtk:prob[UNK] = 0.0172065
DEBUG:cmuclmtk:Incrementing contexts...
DEBUG:cmuclmtk:Calculating back-off weights...
DEBUG:cmuclmtk:Writing out language model...
DEBUG:cmuclmtk:ARPA-style 3-gram will be written to /home/pi/.jasper/vocabularies/en-US/sphinx/keyword/languagemodel
DEBUG:cmuclmtk:idngram2lm : Done.
DEBUG:cmuclmtk:Command 'idngram2lm -idngram /tmp/tmpVrExHA.idngram -vocab /tmp/tmpzFY3fn.vocab -vocab_type 1 -oov_fraction 0.5 -min_unicount 0 -verbosity 2 -n 3 -arpa /home/pi/.jasper/vocabularies/en-US/sphinx/keyword/languagemodel -bin_input' returned with exit code '0'.
DEBUG:sphinx_1_0_0.sphinxvocab:Starting dictionary...
DEBUG:sphinx_1_0_0.sphinxvocab:Getting phonemes for 18 words...
DEBUG:sphinx_1_0_0.g2p:Converting 18 words to phonemes
DEBUG:sphinx_1_0_0.g2p:G2P conversion returned phonemes for 18 words
DEBUG:sphinx_1_0_0.sphinxvocab:Creating dict file: '/home/pi/.jasper/vocabularies/en-US/sphinx/keyword/dictionary'
INFO:client.vocabcompiler:Compilation done.
DEBUG:sphinx_1_0_0.sphinxplugin:Initializing PocketSphinx Decoder with hmm_dir '/home/pi/voxforge-de-r20141117/model_parameters/voxforge.cd_cont_3000'
DEBUG:pico_tts_1_0_0.pico:Executing pico2wave -w /tmp/tmpJXSxJO.wav -l en-US 'How can I be of service, Thomas?'
DEBUG:pyaudio_1_0_0.pyaudioengine:output stream opened on device 'default' (16000 Hz, 1 channel, 16 bit)
DEBUG:pyaudio_1_0_0.pyaudioengine:output stream closed on device 'default'
INFO:client.conversation:Starting to handle conversation with keyword 'JASPER'.
DEBUG:pyaudio_1_0_0.pyaudioengine:input stream opened on device 'default' (16000 Hz, 1 channel, 16 bit)
^CDEBUG:pyaudio_1_0_0.pyaudioengine:input stream closed on device 'default'
Traceback (most recent call last):
File "./jasper.py", line 278, in <module>
app.run()
File "./jasper.py", line 236, in run
self.conversation.handleForever()
File "/home/pi/jasper/client/conversation.py", line 42, in handleForever
input = self.mic.listen()
File "/home/pi/jasper/client/mic.py", line 137, in listen
self.wait_for_keyword(self._keyword)
File "/home/pi/jasper/client/mic.py", line 97, in wait_for_keyword
self._input_rate):
File "/home/pi/jasper/plugins/audioengine/pyaudio-ae/pyaudioengine.py", line 178, in record
frame = stream.read(chunksize)
File "/usr/lib/python2.7/dist-packages/pyaudio.py", line 605, in read
return pa.read_stream(self._stream, num_frames)
KeyboardInterrupt
pi@pi:~/jasper $
There is no error visible in your logs.
Maybe it is somewhere else cause since the last update the keyword jasper won't be recognized in German language. That was different before.
I have done:
cd ~/jasper
git fetch origin
git checkout -- .
git checkout origin/jasper-dev
git checkout origin/feat/multilanguage-support
The feat/multilanguage-support branch doesn't exist anymore, because it has been merged into jasper-dev in e3f2bf530057af596f245782b260a65db96a1658, so you need to use:
cd ~/jasper
git fetch origin
git checkout origin/jasper-dev
You can check that via:
$ git rev-parse --verify --short=7 HEAD
If you're using the latest version, it should print the same commit id like the uppermost entry on this page, i.e. 7d4d306 at the moment.
If that does not fix your issue, remove all compiled vocabularies to force recompilation:
$ rm -rf ~/.jasper/vocabularies
Then run jasper an check again. If the problem is still there, I need the output of:
$ cat ~/.jasper/vocabularies/de-DE/sphinx/keyword/dictionary
git rev-parse --verify --short=7 HEAD
the output is like you wrote: 7d4d306
It seems to be that jasper will not understand the German pocketsphinx anymore.
Before it was ok.
Here the output of
pi@pi:~/jasper $ cat ~/.jasper/vocabularies/de-DE/sphinx/keyword/dictionary
WELCHES W EH L CH AH Z
WELCHES(2) W EH L CH IY Z
WELCHES(3) W EH L CH AH S
WELCHER W EH L CH ER
WELCHER(2) W EH L K ER
WELCHER(3) V EH L CH ER
ABER EY B ER
ABER(2) AE B ER
ABER(3) AH B ER
WAREN W AO R AH N
WAREN(2) W EH R AH N
WAREN(3) W EH R N
DIE D IY
DIE(2) D AY
DIE(3) D AY IY
DER D ER
DER(2) D EH R
DER(3) D EY ER
IST IH S T
IST(2) AY S T
IST(3) AH S T
SIE S IY
SIE(2) EH S AY IY
SIE(3) S AY IY
SIND S IH N D
SIND(2) S AY N D
SIND(3) S IH N
MIT M IH T
MIT(2) M AH T
MIT(3) M IY T
BIN B IH N
BIN(2) B AH N
BIN(3) B IY N
WIRD W ER D
WIRD(2) W IH R D
WIRD(3) W AY R D
WAS W AA Z
WAS(2) W AH Z
WAS(3) W AA S
SAG S AE G
SAG(2) S AA G
SAG(3) S EY G
JA Y AH
JA(2) JH AH
JA(3) ZH AH
WESHALB W EH SH AO L B
WESHALB(2) W IH SH AO L B
WESHALB(3) W EH SH AE L B
SEID S AY D
SEID(2) S IY D
SEID(3) S AY T
ICH IH CH
ICH(2) IH HH
ICH(3) IH K
NEIN N IY N
NEIN(2) N AY N
NEIN(3) N EY N
DAS D AE S
DAS(2) D AH Z
DAS(3) D AH S
AN AE N
AN(2) AH N
AN(3) AA N
WER W ER
WER(2) W EH R
WER(3) W R
WEIL W AY L
WEIL(2) W IY L
WEIL(3) V AY L
JASPER JH AE S P ER
JASPER(2) Y AH S P ER
JASPER(3) Y AA S P ER
IN IH N
IN(2) AH N
IN(3) AE N
WIESO W IY S OW
WIESO(2) W AY S OW
WIESO(3) V IY S OW
ERST ER S T
ERST(2) EH R S T
ERST(3) EH R AH S T
WAR W AO R
WAR(2) W AA R
WAR(3) W ER
ES EH S
ES(2) IY Z
ES(3) AH S
ER ER
ER(2) EH R
ER(3) IH R
WIE W IY
WIE(2) V IY
WIE(3) W AY
WARUM W AO R AH M
WARUM(2) W EH R AH M
WARUM(3) W ER AH M
IHR IH R
IHR(2) IH
IHR(3) AY R
VON V AA N
VON(2) V AO N
VON(3) V AH N
WERDET W ER D EH T
WERDET(2) W ER D AH T
WERDET(3) W ER D IH T
JETZT JH EH T S T
JETZT(2) Y EH T S T
JETZT(3) JH EH T Z T
WERDEN W ER D AH N
WERDEN(2) W ER D EH N
WERDEN(3) W ER D N
ARBEIT AA R B AY T
ARBEIT(2) AA R B IY AH T
ARBEIT(3) AA R B AY T IY
WIR W ER
WIR(2) W IH R
WIR(3) W AY ER
pi@pi:~/jasper $
Looks normal. What about this?
$ cat ~/.jasper/vocabularies/de-DE/sphinx/keyword/languagemodel
It seems to be normal and runs for the first time. If I say Jasper once all looks ok but if I try to say Uhrzeit or something else Jasper will not understand. Before it was working.
Here the language model.
pi@pi:~ $ cat ~/.jasper/vocabularies/de-DE/sphinx/keyword/languagemodel
#############################################################################
## Copyright (c) 1996, Carnegie Mellon University, Cambridge University,
## Ronald Rosenfeld and Philip Clarkson
## Version 3, Copyright (c) 2006, Carnegie Mellon University
## Contributors includes Wen Xu, Ananlada Chotimongkol,
## David Huggins-Daines, Arthur Chan and Alan Black
#############################################################################
=============================================================================
=============== This file was produced by the CMU-Cambridge ===============
=============== Statistical Language Modeling Toolkit ===============
=============================================================================
This is a 3-gram language model, based on a vocabulary of 41 words,
which begins "</s>", "<s>", "ABER"...
This is an OPEN-vocabulary model (type 1)
(OOVs were mapped to UNK, which is treated as any other vocabulary word)
Good-Turing discounting was applied.
1-gram frequency of frequency : 38
2-gram frequency of frequency : 77 0 0 0 0 0 0
3-gram frequency of frequency : 115 0 0 0 0 0 0
1-gram discounting ratios : 0.95
2-gram discounting ratios :
3-gram discounting ratios :
This file is in the ARPA-standard format introduced by Doug Paul.
p(wd3|wd1,wd2)= if(trigram exists) p_3(wd1,wd2,wd3)
else if(bigram w1,w2 exists) bo_wt_2(w1,w2)*p(wd3|wd2)
else p(wd3|w2)
p(wd2|wd1)= if(bigram exists) p_2(wd1,wd2)
else bo_wt_1(wd1)*p_1(wd2)
All probs and back-off weights (bo_wt) are given in log10 form.
Data formats:
Beginning of data mark: \data\
ngram 1=nr # number of 1-grams
ngram 2=nr # number of 2-grams
ngram 3=nr # number of 3-grams
\1-grams:
p_1 wd_1 bo_wt_1
\2-grams:
p_2 wd_1 wd_2 bo_wt_2
\3-grams:
p_3 wd_1 wd_2 wd_3
end of data mark: \end\
\data\
ngram 1=42
ngram 2=78
ngram 3=115
\1-grams:
-2.0830 <UNK> 0.0000
-0.4809 </s> -1.4222
-0.4696 <s> -1.7287
-2.0830 ABER -0.3029
-2.0830 AN -0.3029
-2.0830 ARBEIT -0.3029
-2.0830 BIN -0.3029
-2.0830 DAS -0.3029
-2.0830 DER -0.3029
-2.0830 DIE -0.3029
-2.0830 ER -0.3029
-2.0830 ERST -0.3029
-2.0830 ES -0.3029
-2.0830 ICH -0.3029
-2.0830 IHR -0.3029
-2.0830 IN -0.3029
-2.0830 IST -0.3029
-2.0830 JA -0.3029
-2.0830 JASPER 0.0000
-2.0830 JETZT -0.3029
-2.0830 MIT -0.3029
-2.0830 NEIN -0.3029
-2.0830 SAG -0.3029
-2.0830 SEID -0.3029
-2.0830 SIE -0.3029
-2.0830 SIND -0.3029
-2.0830 VON -0.3029
-2.0830 WAR -0.3029
-2.0830 WAREN -0.3029
-2.0830 WARUM -0.3029
-2.0830 WAS -0.3029
-2.0830 WEIL -0.3029
-2.0830 WELCHER -0.3029
-2.0830 WELCHES -0.3029
-2.0830 WER -0.3029
-2.0830 WERDEN -0.3029
-2.0830 WERDET -0.3029
-2.0830 WESHALB -0.3029
-2.0830 WIE -0.3029
-2.0830 WIESO -0.3029
-2.0830 WIR -0.3029
-2.0830 WIRD -0.3029
\2-grams:
-0.0110 </s> <s> -0.1706
-1.5966 <s> ABER 0.1761
-1.5966 <s> AN 0.1761
-1.5966 <s> ARBEIT 0.1761
-1.5966 <s> BIN 0.1761
-1.5966 <s> DAS 0.1761
-1.5966 <s> DER 0.1761
-1.5966 <s> DIE 0.1761
-1.5966 <s> ER 0.1761
-1.5966 <s> ERST 0.1761
-1.5966 <s> ES 0.1761
-1.5966 <s> ICH 0.1761
-1.5966 <s> IHR 0.1761
-1.5966 <s> IN 0.1761
-1.5966 <s> IST 0.1761
-1.5966 <s> JA 0.1761
-1.5966 <s> JASPER -0.1268
-1.5966 <s> JETZT 0.1761
-1.5966 <s> MIT 0.1761
-1.5966 <s> NEIN 0.1761
-1.5966 <s> SAG 0.1761
-1.5966 <s> SEID 0.1761
-1.5966 <s> SIE 0.1761
-1.5966 <s> SIND 0.1761
-1.5966 <s> VON 0.1761
-1.5966 <s> WAR 0.1761
-1.5966 <s> WAREN 0.1761
-1.5966 <s> WARUM 0.1761
-1.5966 <s> WAS 0.1761
-1.5966 <s> WEIL 0.1761
-1.5966 <s> WELCHER 0.1761
-1.5966 <s> WELCHES 0.1761
-1.5966 <s> WER 0.1761
-1.5966 <s> WERDEN 0.1761
-1.5966 <s> WERDET 0.1761
-1.5966 <s> WESHALB 0.1761
-1.5966 <s> WIE 0.1761
-1.5966 <s> WIESO 0.1761
-1.5966 <s> WIR 0.1761
-1.5966 <s> WIRD 0.1761
-0.1761 ABER </s> 1.3010
-0.1761 AN </s> 1.3010
-0.1761 ARBEIT </s> 1.3010
-0.1761 BIN </s> 1.3010
-0.1761 DAS </s> 1.3010
-0.1761 DER </s> 1.3010
-0.1761 DIE </s> 1.3010
-0.1761 ER </s> 1.3010
-0.1761 ERST </s> 1.3010
-0.1761 ES </s> 1.3010
-0.1761 ICH </s> 1.3010
-0.1761 IHR </s> 1.3010
-0.1761 IN </s> 1.3010
-0.1761 IST </s> 1.3010
-0.1761 JA </s> 1.3010
-0.1761 JETZT </s> 1.3010
-0.1761 MIT </s> 1.3010
-0.1761 NEIN </s> 1.3010
-0.1761 SAG </s> 1.3010
-0.1761 SEID </s> 1.3010
-0.1761 SIE </s> 1.3010
-0.1761 SIND </s> 1.3010
-0.1761 VON </s> 1.3010
-0.1761 WAR </s> 1.3010
-0.1761 WAREN </s> 1.3010
-0.1761 WARUM </s> 1.3010
-0.1761 WAS </s> 1.3010
-0.1761 WEIL </s> 1.3010
-0.1761 WELCHER </s> 1.3010
-0.1761 WELCHES </s> 1.3010
-0.1761 WER </s> 1.3010
-0.1761 WERDEN </s> 1.3010
-0.1761 WERDET </s> 1.3010
-0.1761 WESHALB </s> 1.3010
-0.1761 WIE </s> 1.3010
-0.1761 WIESO </s> 1.3010
-0.1761 WIR </s> 1.3010
-0.1761 WIRD </s> 1.3010
\3-grams:
-1.5911 </s> <s> ABER
-1.5911 </s> <s> AN
-1.5911 </s> <s> ARBEIT
-1.5911 </s> <s> BIN
-1.5911 </s> <s> DAS
-1.5911 </s> <s> DER
-1.5911 </s> <s> DIE
-1.5911 </s> <s> ER
-1.5911 </s> <s> ERST
-1.5911 </s> <s> ES
-1.5911 </s> <s> ICH
-1.5911 </s> <s> IHR
-1.5911 </s> <s> IN
-1.5911 </s> <s> IST
-1.5911 </s> <s> JA
-1.5911 </s> <s> JASPER
-1.5911 </s> <s> JETZT
-1.5911 </s> <s> MIT
-1.5911 </s> <s> NEIN
-1.5911 </s> <s> SAG
-1.5911 </s> <s> SEID
-1.5911 </s> <s> SIE
-1.5911 </s> <s> SIND
-1.5911 </s> <s> VON
-1.5911 </s> <s> WAR
-1.5911 </s> <s> WAREN
-1.5911 </s> <s> WARUM
-1.5911 </s> <s> WAS
-1.5911 </s> <s> WEIL
-1.5911 </s> <s> WELCHER
-1.5911 </s> <s> WELCHES
-1.5911 </s> <s> WERDEN
-1.5911 </s> <s> WERDET
-1.5911 </s> <s> WESHALB
-1.5911 </s> <s> WIE
-1.5911 </s> <s> WIESO
-1.5911 </s> <s> WIR
-1.5911 </s> <s> WIRD
-0.3010 <s> ABER </s>
-0.3010 <s> AN </s>
-0.3010 <s> ARBEIT </s>
-0.3010 <s> BIN </s>
-0.3010 <s> DAS </s>
-0.3010 <s> DER </s>
-0.3010 <s> DIE </s>
-0.3010 <s> ER </s>
-0.3010 <s> ERST </s>
-0.3010 <s> ES </s>
-0.3010 <s> ICH </s>
-0.3010 <s> IHR </s>
-0.3010 <s> IN </s>
-0.3010 <s> IST </s>
-0.3010 <s> JA </s>
-0.3010 <s> JASPER </s>
-0.3010 <s> JETZT </s>
-0.3010 <s> MIT </s>
-0.3010 <s> NEIN </s>
-0.3010 <s> SAG </s>
-0.3010 <s> SEID </s>
-0.3010 <s> SIE </s>
-0.3010 <s> SIND </s>
-0.3010 <s> VON </s>
-0.3010 <s> WAR </s>
-0.3010 <s> WAREN </s>
-0.3010 <s> WARUM </s>
-0.3010 <s> WAS </s>
-0.3010 <s> WEIL </s>
-0.3010 <s> WELCHER </s>
-0.3010 <s> WELCHES </s>
-0.3010 <s> WER </s>
-0.3010 <s> WERDEN </s>
-0.3010 <s> WERDET </s>
-0.3010 <s> WESHALB </s>
-0.3010 <s> WIE </s>
-0.3010 <s> WIESO </s>
-0.3010 <s> WIR </s>
-0.3010 <s> WIRD </s>
-0.3010 ABER </s> <s>
-0.3010 AN </s> <s>
-0.3010 ARBEIT </s> <s>
-0.3010 BIN </s> <s>
-0.3010 DAS </s> <s>
-0.3010 DER </s> <s>
-0.3010 DIE </s> <s>
-0.3010 ER </s> <s>
-0.3010 ERST </s> <s>
-0.3010 ES </s> <s>
-0.3010 ICH </s> <s>
-0.3010 IHR </s> <s>
-0.3010 IN </s> <s>
-0.3010 IST </s> <s>
-0.3010 JA </s> <s>
-0.3010 JETZT </s> <s>
-0.3010 MIT </s> <s>
-0.3010 NEIN </s> <s>
-0.3010 SAG </s> <s>
-0.3010 SEID </s> <s>
-0.3010 SIE </s> <s>
-0.3010 SIND </s> <s>
-0.3010 VON </s> <s>
-0.3010 WAR </s> <s>
-0.3010 WAREN </s> <s>
-0.3010 WARUM </s> <s>
-0.3010 WAS </s> <s>
-0.3010 WEIL </s> <s>
-0.3010 WELCHER </s> <s>
-0.3010 WELCHES </s> <s>
-0.3010 WER </s> <s>
-0.3010 WERDEN </s> <s>
-0.3010 WERDET </s> <s>
-0.3010 WESHALB </s> <s>
-0.3010 WIE </s> <s>
-0.3010 WIESO </s> <s>
-0.3010 WIR </s> <s>
-0.3010 WIRD </s> <s>
\end\
pi@pi:~ $
This also seems fine. Does recognition work when running:
pocketsphinx_continuous -hmm /home/pi/voxforge-de-r20141117/model_parameters/voxforge.cd_cont_3000 -dict ~/.jasper/vocabularies/de-DE/sphinx/default/dictionary -lm ~/.jasper/vocabularies/de-DE/sphinx/default/languagemodel -inmic yes
It is doing something but not what am saying.
As I said it already was running before. SSH connection will not work with raspi cause of security.
Therefore another program has to be installed.
I have installed teamviewer on another laptop. That you can use to try to connect.
Teamfiewer ID = 538257853
Password = 8200
btw. English language is working.
Sorry, I don't have teamviewer installed. What FST model are you using?
Also, it would be helpful if you could post the output when if it's actually compiling the vocabulary, i.e. the output of:
$ rm -rf ~/.jasper/vocabularies
$ ./jasper.py --debug
Here is the file you like to get.
I put it in a separate file cause it is a bit larger then others
startup.txt
Hmm there is no problem at all during compilation. What does happen if you actually say something?
Three times I said Jasper till the beep,
then I said "Uhrzeit", then again "Jasper" then "Zeit", then "Jasper" then "Wie spät ist es"
Any time he said that jasper doesn't understand it.
Here again the file:
jasper.txt
Okay, then there is not program error at all.
Recognition doesn't work properly. because:
The acoustic model isn't very good → You could try to adapt/train it
The phones used in the FST model don't match the phones used in the acoustic model...
...and you're using a german acoustic model with an English FST-Model instead of a german one
The reason why you didn't notice that before was because there were only a few words in that model and Pocketsphinx kinda recognized them "by chance". I added more words to compile into that model in commit 841f41e, and now the bad recognition rate is much more noticeable.
Thus, I reverted that change. I also added X-SAMPA support, so that you can use a german FST model instead of an english one:
pocketsphinx:
hmm_dir: '/path/to/voxforge-de-r20141117/model_parameters/voxforge.cd_cont_3000'
fst_model: '/path/to/dict-xsampa.fst'
fst_model_alphabet: 'xsampa'
Got the dict-xsampa.fst for German and that voxforge model and added it. If am talking now to Jasper it recognizes nothing as I have seen in the debug output.
Here is a new output of jasper.
new.txt
Did you add fst_model_alphabet: 'xsampa'?
I added in my folder ..../de dict-xsampa.arpa ....corpus .....fst
Something else I haven't found.
No, i was referring to your profile.yml.
My profile.yml
language: 'de-DE'
carrier: ''
first_name: Thomas
gmail_password: ''
location: Ingolstadt
phone_number: ''
prefers_email: false
stt_engine: sphinx
pocketsphinx:
hmm_dir: '/home/pi/voxforge-de-r20141117/model_parameters/voxforge.cd_cont_3000'
fst_model: '/home/pi/de/dict-xsampa.fst'
fst_model_alphabet: 'xsampa'
# fst_model: '../phonetisaurus/g014b2b.fst' #optional
# hmm_dir: '/usr/local/share/pocketsphinx/model/hmm/en_US/hub4wsj_sc_8k' #optional
# hmm_dir: '/home/pi/pocketsphinx-5prealpha/test/data/tidigits/hmm'
# hmm_dir: '/home/pi/pocketsphinx-0.8/model/hmm/en_US/hub4wsj_sc_8k'
# hmm_dir: '/home/pi/voxforge-de-r20141117/model_parameters/voxforge.cd_cont_3000'
tts_engine: pico-tts
#pico-tts:
# language: 'de-DE'
timezone: Europe/Berlin
input_device: 'default'
output_device: 'default'
For testings I changed the language back to English.
As long as I am not using my own modules it seems to be all is working fine.
If I copy my modules in the folder ~/.jasper/plugin/speechhandler and start jasper then my modules are not working. After I start your command
./compile_translations.sh
my modules are working but then jasper for example hardly understand if I say TIME.
It recognize POWER or NEWS instead of time.
Hello
Hello
Hi G10DRAS,
may I ask which language model you are using cause in my case it will not work.
How does your profil.yml looks like?
I am trying to get the German language working but without many success.
Now I am playing with Simon and there is a possibility to train the acoustic Model and it will work.
Is there a possibility to convert that model that it can be used together with Jasper?
Regards
Thomas
Thomas
See the threadd below
https://groups.google.com/forum/#!searchin/jasper-support-forum/german/jasper-support-forum/0cvbkisUyOE/C-ufpsmzDwAJ
I have generated an German FST model which works with German Acoustic Model
(voxforge.cd_cont_3000 from
http://goofy.zamia.org/voxforge/de/voxforge-de-r20141117.tgz).
No code change required in Jasper.
Download it from
https://www.dropbox.com/s/vbkf8qjvsf67b66/voxforge-de.fst?dl=0
See how it works...........
On 23 January 2016 at 14:34, Thomas301263 <notifications@github.com
<javascript:_e(%7B%7D,'cvml','notifications@github.com');>> wrote:
Hi G10DRAS,
may I ask which language model you are using cause in my case it will not
work.
How does your profil.yml looks like?
I am trying to get the German language working but without many success.
Now I am playing with Simon and there is a possibility to train the
acoustic Model and it will work.
Is there a possibility to convert that model that it can be used together
with Jasper?
Regards
Thomas
—
Reply to this email directly or view it on GitHub
https://github.com/jasperproject/jasper-client/issues/403#issuecomment-174163630
.
In profile.yml specify the path to hmm dir and fst as usual.
On Saturday, 23 January 2016, Thomas301263 notifications@github.com wrote:
Hi G10DRAS,
may I ask which language model you are using cause in my case it will not
work.
How does your profil.yml looks like?
I am trying to get the German language working but without many success.
Now I am playing with Simon and there is a possibility to train the
acoustic Model and it will work.
Is there a possibility to convert that model that it can be used together
with Jasper?
Regards
Thomas
—
Reply to this email directly or view it on GitHub
https://github.com/jasperproject/jasper-client/issues/403#issuecomment-174163630
.
I have updated Jasper before and using that files you wrote to run Jasper in German language.
Jasper stops now since 5 hours on the same step:
DEBUG:sphinx_1_0_0.sphinxvocab:Getting phonemes for 39 words...
DEBUG:sphinx_1_0_0.g2p:Converting 39 words to phonemes
Testings with English language before was working.
Could you try previous version of Jasper (Jasper v1 or jasper-main) with german acoustic model and my FST. My FST model was compiled using best available dictionary which contains only 23000 german words. I am away from my RPI2 for next 2 days hence can not figure out where is the problem exactly. But i tested it in the past with almost 50 WORDS and my jasper (Jasper v1) works very well.
You are right with version v1 of Jasper it works with me as well till there were some updates and I try to find out now why this version is not working. v1 of Jasper has the problem of misunderstanding of my spoken words so I go further steps and try if there will be a solution.
I know German language is not supported that much then English that's why I am using Simon in parallel to train Simon till it works and later I will try to compile Simon *.sbm that I can use it for Jasper. Hope there is a way doing that.
I have the same issue with the dev-branch as mentioned by Thomas301263. Jasper V1 not tested.
g2p conversion stumbles over the words:
ES
ICH
IN
IST
SIE
leaving these words out and the g2p conversion is able to finish. It works fine with dict-xsampa.fst. Thus it seems to be a problem of g2p with the fst model of G10DRAS (https://www.dropbox.com/s/vbkf8qjvsf67b66/voxforge-de.fst?dl=0)
could you try below config with jasper dev-branch (comment out fst_model_alphabet: 'xsampa' line )
pocketsphinx:
hmm_dir: '/path/to/voxforge-de-r20141117/model_parameters/voxforge.cd_cont_3000'
fst_model: '/path/to/downloaded_dropbox.fst'
fst_model_alphabet: 'xsampa'
That's I already made without success.
@Thomas301263: if you like to get it running, go to:
./jasper/data/standard_phrases/de-DE.txt and delete the words mentioned above.
@G10DRAS: its not a problem with an option. Inline conversion of the phonemes is done after the g2p conversion call. You get the same result directly calling phonetisaurus:
phonetisaurus-g2p --model=/path/to/downloaded_dropbox.fst --input=SIE --nbest=3 is broken.
phonetisaurus-g2p --model=/path/to/downloaded_dropbox.fst --input=DAS --nbest=3 is working
and
phonetisaurus-g2p --model=/path/to/dict-xsempa.fst --input=SIE --nbest=3 is working
I have not yet realy understood what the content of *.fst is. This might be a problem with phonetisaurus or with the model.
I'm using phonetisaurus 0.7.8 - as mentioned in the install instructions of jasper - but I will check with a master build from github.
Ok What I did just now with Jasper on RPi2B.
I took 38 German Words (Keyword Phrases) + Use My German FST voxforge-de.fst + HMM voxforge.cd_cont_3000 + Pocketsphinx-5prealpha (Compiled and installed from src downloaded from Github) + Inbuilt Mic of Logicool c270 Webcam
ABER
AN
ARBEIT
BIN
DAS
DER
DIE
ER
ERST
ES
ICH
IHR
IN
IST
JA
JASPER
JETZT
MIT
NEIN
SAG
SEID
SIE
SIND
VON
WAR
WAREN
WARUM
WAS
WEIL
WELCHER
WELCHES
WERDEN
WERDET
WESHALB
WIE
WIESO
WIR
WIRD
I got result as follows (I can say results are 60% accurate, though I am not a native German speaker)
Transcribed:=======> ''
Transcribed:=======> 'WERDET'
Transcribed:=======> ''
Transcribed:=======> 'SIND SIE'
Transcribed:=======> ''
Transcribed:=======> 'WERDEN'
Transcribed:=======> ''
Transcribed:=======> 'WELCHES'
Transcribed:=======> 'WIR'
Transcribed:=======> 'WERDET'
Transcribed:=======> 'ER'
Transcribed:=======> 'ER'
Transcribed:=======> 'ER'
Transcribed:=======> 'NEIN'
Transcribed:=======> 'SEID'
Transcribed:=======> ''
Transcribed:=======> 'SEID SIE'
Transcribed:=======> 'JA'
Transcribed:=======> 'IST WIR'
Transcribed:=======> ''
Transcribed:=======> 'NEIN'
Transcribed:=======> 'NEIN'
Transcribed:=======> 'ER'
Transcribed:=======> ''
Transcribed:=======> 'IN'
Transcribed:=======> ''
Transcribed:=======> 'IST'
Transcribed:=======> 'SIE'
Transcribed:=======> 'SEID'
Transcribed:=======> ''
Transcribed:=======> ''
Transcribed:=======> 'WERDEN'
Transcribed:=======> 'JETZT'
Transcribed:=======> 'ES WAS'
Transcribed:=======> ''
Transcribed:=======> 'JETZT'
Transcribed:=======> 'WELCHER'
Transcribed:=======> 'ER'
Transcribed:=======> 'JASPER'
Transcribed:=======> ''
Transcribed:=======> 'IST'
Transcribed:=======> 'ER'
Transcribed:=======> ''
Transcribed:=======> 'NEIN'
Transcribed:=======> 'WEIL'
Transcribed:=======> ''
Transcribed:=======> 'WEIL'
Transcribed:=======> 'JA'
Transcribed:=======> 'WERDEN'
Transcribed:=======> ''
Transcribed:=======> 'IST'
Transcribed:=======> 'IST'
Transcribed:=======> 'ER'
Transcribed:=======> ''
Transcribed:=======> 'IN'
Transcribed:=======> ''
Transcribed:=======> 'WIE'
Transcribed:=======> 'WIR'
Transcribed:=======> ''
Transcribed:=======> 'MIT'
Transcribed:=======> 'WIESO'
Transcribed:=======> ''
Transcribed:=======> 'WELCHER'
Transcribed:=======> 'ER'
Transcribed:=======> 'IST'
Try installing Pocketsphinx-5prealpha and see if it improves recognization rate.
I tried your solution and you are right that the recognition rate is about 60%. That is to less for me cause Jasper should control my flat and if he do a lot of wrong things then maybe lights, power heating system, rollerblinds and others will not work correctly.
I am using now 2 Raspi's 2b with different solutions. One of them is Jasper and the other one is Simon. Raspi 2b together with Simon will work well and the recognition rate is about 95% cause I can train Simon as I want. That doesn't mean that I am not working with Jasper anymore but I am trying to get the best solution for my smart home controller which work perfect but control everything so there should be a very high recognition rate.
If you know how to Train / Adapt an Acoustic Model for CMUSphinx, you can try that, and use that Trained / Adapted Acoustic Model with Jasper. You will get a lot better accuracy with Jasper too.
See links below:
http://cmusphinx.sourceforge.net/wiki/tutorialam
http://cmusphinx.sourceforge.net/wiki/tutorialadapt
or at least try to generate mllr_matrix and use that with Pocketsphinx in Jasper.
For voice controlled home automation, it is too much to say
"TURN ON
another solution is use shortcut, like
SEVEN --- for Bedroom Light
TEN --- for Kichen Light
example command:
TURN ON SEVEN
SWITCH OFF TEN
HEATER ON SEVEN
Thanks I will try to do it that way you wrote but I like to get the voice command in another way.
I like to say "Strom Schlafzimmer einschalten" and then there should be a reaction. If there is a way to get the sphinx training from Simon to Jasper then it will be good. To train Simon is really simple and quick and if it can be transformed in a format for Jasper then it will be more then ok.
Now am trying your way of training.
I am not sure if you can use trained model from simon into jasper, could you attached a sample file?
There is another application I.L.A which do the same thing as Simon do i.e Train a Model on the fly.
https://sites.google.com/site/ilavoiceassistant/
I check the code but sorce code is not available for it (only binaries.).
Take alook if it works for you.
I will take a look and try. Thanks
Normally the file is basemodel.sbm but I cannot place it here so I made .txt at the end.
Just rename it and then you should have it.
basemodel.sbm.txt
attached sbm file is a compressed file like zip file, I opened it with 7-zip and it contains all Acoustic model files. Try uncompressed model with Jasper, It should work.
Now I tried your way and basically Jasper understood my words I am talking but the problem is if am using long phrases like "Jasper Rollo Wohnzimmer Fenster Ost schliessen" then Jasper struggle.
There should be a timer where pauses between words can be defined.
Now I am running Jasper and Simon on Raspi 2 and 3 in parallel and Simon is better if I am using long phrases. The main reason that I need long phrases is that I am trying to get something which support me and making things for me easier. That could not be done by short phrases.
At the end there should be a system for elder people which can use it supporting them in difficult situation.
Example:
Somebody is not able to take a phone cause he cannot move (maybe a stroke) but he is able to talk so he can say
"Computer I need emergency help"
The computer is connecting that person to somebody else by phone or by mail ..... etc.
telling them that there is a emergency.
@Thomas301263: looks like I'm trying to get something similar running - just curiosity, does your basemodel.sbm includes an updated trained model? I tried that too, but my basemodel remained untrained.
Concerning the recognition: Are you using a language model corresponding to your input?
@Thomas301263 Seems Simon is using JSGF file with PocketSphinx Decoder and not using LM file.
"-hmm", model.data(),
"-jsgf", grammar.data(),
"-dict", dict.data(),
"-samprate", samprate.data(),
Do some experiment using pocketsphinx_continue .
Use JSGF file in one test.
Use LM file in another test.
See the difference, if you feel JSGF file is better, configure your Jasper's PocketSphinx Decoder to use JSGF file and see how Jasper perform.
in stt.py you need to do something like below:
# psConfig.set_string('-lm', "</path/to/lm_file>")
psConfig.set_string('-jsgf', "</path/to/jsgf_file>")
@andweber
I have not used a base model and I was starting from scratch with Simon. I trained every word about 30-40 times build my phrases like sentences trained them as well and Simon is understanding me well. Then I took the .sbm file uncompress it and replaced the files from Jasper.
Now I will try the way G10DRAS suggest me and we will found a way.
@G10DRAS
Now there was time trying your suggestion. For me it seems to be that you are using Jasper V1.
I am using Jasper V2 and there is no stt.py file anymore.
I do not know where I have to set .jsgf or .lm files in that case. I know in the old version there was a
stt.py file.
the right spot is in plugins/stt/pocketsphinx-stt/sphinxplugin.py line 90ff
if self._pocketsphinx_v5:
# Pocketsphinx v5
config = pocketsphinx.Decoder.default_config()
config.set_string('-hmm', hmm_dir)
config.set_string('-jsgf', jsgf_path) # <--- replaced line with lm
config.set_string('-dict', dict_path)
config.set_string('-logfn', self._logfile)
self._decoder = pocketsphinx.Decoder(config)
else:
# Pocketsphinx v4 or sooner
self._decoder = pocketsphinx.Decoder(
hmm=hmm_dir, logfn=self._logfile, jsgf=jsgf_path, dict=dict_path)
`
For Pocketsphinx v4 I only guessed the code. Instead of jsgf_path insert the correct path for you.
@andweber
Thanks for that information. I tried it immediately but there is another error now and I do not know where to set it.
File "/home/pi/jasper/plugins/stt/pocketsphinx-stt/sphinxplugin.py", line 94, in __init__
config.set_string('-jsgf', jsgf_path)
NameError: global name 'jsgf_path' is not defined
|
gharchive/issue
| 2015-12-07T08:06:14 |
2025-04-01T04:34:39.585799
|
{
"authors": [
"G10DRAS",
"Holzhaus",
"Thomas301263",
"andweber"
],
"repo": "jasperproject/jasper-client",
"url": "https://github.com/jasperproject/jasper-client/issues/403",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
250236952
|
请指点,万分感激
您好,首先您搭的这个框架功能很齐全,但我克隆后,想在这个基础上为redux的connect方法配置装饰器,比如下面这种写法
@connect(
state => ({
num:state.app.num
}),
dispatch =>({
action: bindActionCreators(actionA, dispatch)
});
这个要如何配置,您方便更新下吗?
需要npm install --save-dev babel-preset-stage-1
然后将根目录下的.babelrc中的"stage-3" 修改为 “stage-1” 就可以使用修饰器了
@javaLuo 这个需要我自己配置吗? 还是说您已经配置好了 ,我可以直接用?
@javaLuo 根据您的提示,修改后,这样编辑,如下:
@connect(
state => ({
num: state.app.num,
}),
dispatch =>({
})
)
class Mytest extends React.Component {
constructor(props) {
super(props);
this.state = {
};
}
render() {
const {npm} = this.props;
console.log(num);
return (
6666
)
}
}
运行项目时提示我识别不了@
重新npm run dll 试试
@javaLuo 报这个错
C:\Users\Administrator\Desktop\react_luo\src\a_container\test\index.js
148:2 error Missing semicolon semi
156:16 error 'num' is missing in props validation react/prop-types
160:10 error Missing semicolon semi
✖ 3 problems (3 errors, 0 warnings)
额...不是吧,我这边测试可以识别@了
@javaLuo 我从新拷贝你的项目 然后按您的指示,报如下错误:
ERROR in ./src/a_container/test/index.js
C:\Users\Administrator\Desktop\react+redux\react_luo\src\a_container\test\index. js
143:1 error Parsing error: Unexpected character '@'
✖ 1 problem (1 error, 0 warnings)
@ ./src/route/index.js 25:14-50
@ ./src/app.js
@ multi webpack-dev-server/client?http://localhost:8888 webpack/hot/dev-server ./src/app
ERROR in ./src/a_container/test/index.js
Module build failed: SyntaxError: C:/Users/Administrator/Desktop/react+redux/rea ct_luo/src/a_container/test/index.js: Decorators are not officially supported ye t in 6.x pending a proposal update.
However, if you need to use them you can install the legacy decorators transform with:
npm install babel-plugin-transform-decorators-legacy --save-dev
and add the following line to your .babelrc file:
{
"plugins": ["transform-decorators-legacy"]
}
The repo url is: https://github.com/loganfsmyth/babel-plugin-transform-decorator s-legacy.
148 | })
149 | )
150 | class Mytest extends React.Component {
| ^
151 | constructor(props) {
152 | super(props);
153 | this.state = {
@ ./src/route/index.js 25:14-50
@ ./src/app.js
@ multi webpack-dev-server/client?http://localhost:8888 webpack/hot/dev-server ./src/app
然后我这样:
npm install babel-plugin-transform-decorators-legacy --save-dev
接着修改.babelrc
{
"presets": ["es2015", "stage-1", "react"],
"plugins": [
"transform-decorators-legacy",
["import", {
"libraryName": "antd",
"style": "css"
}
]
]
}
最后运行时还是报错
C:\Users\Administrator\Desktop\react+redux\react_luo\src\a_container\test\index.js
143:1 error Parsing error: Unexpected character '@'
✖ 1 problem (1 error, 0 warnings)
@ ./src/route/index.js 25:14-50
@ ./src/app.js
@ multi webpack-dev-server/client?http://localhost:8888 webpack/hot/dev-server ./src/app
我更新了代码,重新上传了。你再试试。
暂时取消了eslint代码检测。eslint现在的版本不支持decorator
@javaLuo 您好,现在能识别@了,但是运行后会报这个错,如下:
vendor.dll.js:34 Uncaught TypeError: Cannot read property 'location' of undefined
at new t (vendor.dll.js:34)
at d._constructComponentWithoutOwner (vendor.dll.js:43)
at d._constructComponent (vendor.dll.js:43)
at d.mountComponent (vendor.dll.js:43)
at Object.mountComponent (vendor.dll.js:20)
at d.performInitialMount (vendor.dll.js:43)
at d.mountComponent (vendor.dll.js:43)
at Object.mountComponent (vendor.dll.js:20)
at d.performInitialMount (vendor.dll.js:43)
at d.mountComponent (vendor.dll.js:43)
再次麻烦您给指导下了
重新拉一下代码
重新install
那是react-router版本不兼容
暂时没有升级到react-router4.0
|
gharchive/issue
| 2017-08-15T07:04:23 |
2025-04-01T04:34:39.611734
|
{
"authors": [
"javaLuo",
"wangjunwei910208"
],
"repo": "javaLuo/react-luo",
"url": "https://github.com/javaLuo/react-luo/issues/2",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1400120845
|
Unify groupId, artifactId and name of examples
I had quite some issues with duplicate coordinates while creating the aggregator poms.
This PR unifies the groupId to io.javalin.samples.javalin4 / io.javalin.samples.javalin5.
Names are unified to Javalin 4 ${artifactId} / Javalin 5 ${artifactId} and the artifactId is the same as the containing folder.
Thank you @Playacem !
|
gharchive/pull-request
| 2022-10-06T18:25:00 |
2025-04-01T04:34:39.620267
|
{
"authors": [
"Playacem",
"tipsy"
],
"repo": "javalin/javalin-samples",
"url": "https://github.com/javalin/javalin-samples/pull/6",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1332347826
|
Default path for Before/After handlers doesn't require trailing slashes
resolving issue #1638
Thanks @MazenAmria, could you add a test for the new behavior?
@tipsy Actually I've tried using before("*", handler) instead of before("*") since it would have the same effect as what I've changed in this PR but still the same problem. So unfortunately my changes didn't solve the issue.
Perhaps the issue is with the prefix method? I remember we added something specifically to address this case, but maybe it was limited to path params 🤔
The problem is with prefixPath method, it always adds a forward slash to the beginning of the provided path.
Do you think that we need the case where we just match /path/* without /path*? if there's no such cases, then we can modify the PathMatcher to ignore trailing slashes.
if there's no such cases, then we can modify the PathMatcher to ignore trailing slashes.
That's already a config option for that (which is true by default), so I'm not sure if that will help?
The problem is with prefixPath method, it always adds a forward slash to the beginning of the provided path.
We could provide a config option for this. Either by adding something to config.routing.prefixApiBuilderPaths, or perhaps better as an argument to the routes() call itself:
app.routes(ApiBuilderOption.DONT_PREFIX_PATHS) { // enum naming might need some work...
}
That's already a config option for that (which is true by default)
Yeah, I've seen that, but this for ignoring the request trailing slash, like requesting /path/ would match both /path and /path/ but requesting /path won't match /path/, just /path.
ApiBuilderOption.DONT_PREFIX_PATHS I think this option makes sense more. The default behavior would be ApiBuilderOption.DO_PREFIX_PATHS or something similar, and when user needs to include /path in his handler he would specify the DONT option.
Thanks @MazenAmria !
|
gharchive/pull-request
| 2022-08-08T20:14:15 |
2025-04-01T04:34:39.627266
|
{
"authors": [
"MazenAmria",
"tipsy"
],
"repo": "javalin/javalin",
"url": "https://github.com/javalin/javalin/pull/1639",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2078767187
|
🛑 Castello 110 is down
In 560d3a7, Castello 110 (http://159.65.204.200:7007/) was down:
HTTP code: 502
Response time: 18701 ms
Resolved: Castello 110 is back up in e31aa3c after 15 minutes.
|
gharchive/issue
| 2024-01-12T12:43:58 |
2025-04-01T04:34:39.656940
|
{
"authors": [
"javisu"
],
"repo": "javisu/monitor",
"url": "https://github.com/javisu/monitor/issues/10634",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1261175312
|
🛑 Colegio Malvar Piscina is down
In 5d7c4c6, Colegio Malvar Piscina (http://159.65.204.200:7041/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Colegio Malvar Piscina is back up in c9ebb22.
|
gharchive/issue
| 2022-06-05T22:21:32 |
2025-04-01T04:34:39.659296
|
{
"authors": [
"javisu"
],
"repo": "javisu/monitor",
"url": "https://github.com/javisu/monitor/issues/1973",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1296151317
|
🛑 Rodriguez Marin, 88 is down
In 95ee72e, Rodriguez Marin, 88 (http://159.65.204.200:7057/) was down:
HTTP code: 502
Response time: 18687 ms
Resolved: Rodriguez Marin, 88 is back up in fc09577.
|
gharchive/issue
| 2022-07-06T17:19:36 |
2025-04-01T04:34:39.661612
|
{
"authors": [
"javisu"
],
"repo": "javisu/monitor",
"url": "https://github.com/javisu/monitor/issues/2427",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1406473878
|
🛑 Antonio Acuña, 10 is down
In 25615b5, Antonio Acuña, 10 (http://159.65.204.200:7047/) was down:
HTTP code: 502
Response time: 18622 ms
Resolved: Antonio Acuña, 10 is back up in e0d5ae8.
|
gharchive/issue
| 2022-10-12T16:11:36 |
2025-04-01T04:34:39.664061
|
{
"authors": [
"javisu"
],
"repo": "javisu/monitor",
"url": "https://github.com/javisu/monitor/issues/3505",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1604757726
|
🛑 Colegio Malvar Piscina is down
In e1ccd37, Colegio Malvar Piscina (http://159.65.204.200:7041/) was down:
HTTP code: 503
Response time: 228 ms
Resolved: Colegio Malvar Piscina is back up in 4615fc8.
|
gharchive/issue
| 2023-03-01T11:07:17 |
2025-04-01T04:34:39.666574
|
{
"authors": [
"javisu"
],
"repo": "javisu/monitor",
"url": "https://github.com/javisu/monitor/issues/6595",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1694182219
|
🛑 Ofelia Nieto, 57 is down
In 2d2a339, Ofelia Nieto, 57 (http://159.65.204.200:7022/) was down:
HTTP code: 502
Response time: 18697 ms
Resolved: Ofelia Nieto, 57 is back up in d9f1435.
|
gharchive/issue
| 2023-05-03T14:25:11 |
2025-04-01T04:34:39.668878
|
{
"authors": [
"javisu"
],
"repo": "javisu/monitor",
"url": "https://github.com/javisu/monitor/issues/7546",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2551617848
|
Add .pylintrc file
The JAX project utilizes non-standard PEP 8 settings for indentation.
To validate Python files, we can leverage a .pylintrc configuration file.
For reference, the TensorFlow project has an existing .pylintrc file that outlines its settings.
This PR introduces a .pylintrc file for the JAX project, modeled after the configuration used in TensorFlow.
[FORMAT]
# Maximum number of characters on a single line.
max-line-length=80
# String used as indentation unit. This is usually " " (4 spaces) or "\t" (1
# tab).
indent-string=' '
Usage Example:
### check files in HEAD commit
git diff --name-only HEAD^ HEAD | xargs pylint
Hi - thanks for the PR. We already have pylint configuration in pyproject.toml, and I'd prefer to keep all such configs in a single location
Also, a while ago we moved off pylint and now use ruff for formatting. It's set up as a pre-commit hook.
Got it. Thank you for the info!
|
gharchive/pull-request
| 2024-09-26T22:18:53 |
2025-04-01T04:34:39.671557
|
{
"authors": [
"apivovarov",
"jakevdp"
],
"repo": "jax-ml/jax",
"url": "https://github.com/jax-ml/jax/pull/23959",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
294180924
|
README file explain more about the project
Give more description in README file
That was my first issue
|
gharchive/issue
| 2018-02-04T08:38:54 |
2025-04-01T04:34:39.672596
|
{
"authors": [
"jayakishore82"
],
"repo": "jayakishore82/treeHouse",
"url": "https://github.com/jayakishore82/treeHouse/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
902090267
|
Add voice inputs for disabled users
These inputs should be able to filled with voice inputs
fixed
|
gharchive/issue
| 2021-05-26T09:28:56 |
2025-04-01T04:34:39.673745
|
{
"authors": [
"rumeshmadhusanka"
],
"repo": "jayampathiadhikari/bmi-calculator",
"url": "https://github.com/jayampathiadhikari/bmi-calculator/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
707232576
|
Deploying to Heroku with pgbouncer
Hi,
I'm deploying a Django app to Heroku using the config function, so far it worked well.
now I'm trying to scale a bit, so I added pgbouncer to the app's Procfile.
web: daphne rivendell.asgi:application --port $PORT --bind 0.0.0.0 -v2
worker: python manage.py runworker channel_layer -v2
but this crashes the server with the error:
django.db.utils.OperationalError: server does not support SSL, but SSL was required
My settings:
DATABASES = {'default': dj_database_url.config(conn_max_age=600, ssl_require=True)}
I think this will be an issue with configuration, seeing as its now an old issue I am going to close. If anyone would like to add to this issue please do.
|
gharchive/issue
| 2020-09-23T10:14:22 |
2025-04-01T04:34:39.698992
|
{
"authors": [
"mattseymour",
"yovelcohen"
],
"repo": "jazzband/dj-database-url",
"url": "https://github.com/jazzband/dj-database-url/issues/139",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
1346078977
|
Add support for CONN_HEALTH_CHECKS
https://docs.djangoproject.com/en/4.1/ref/settings/#std-setting-CONN_HEALTH_CHECKS
This feature is present in release 1.1.0 available now on Pypi
|
gharchive/issue
| 2022-08-22T09:13:17 |
2025-04-01T04:34:39.700211
|
{
"authors": [
"lorddaedra",
"mattseymour"
],
"repo": "jazzband/dj-database-url",
"url": "https://github.com/jazzband/dj-database-url/issues/180",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
1862212420
|
data loss upgrading from 2.9.1 to 3.1.0 with database backend
Describe the problem
Upgrading from 2.9.1 to 3.1.0 and running python manage migrate, both migrations appear successful, but the data isn't transferred, nor is the old table dropped.
Steps to reproduce
Install version 2.9.1 and make a couple of sample settings, changing from their defaults.
Update to 3.1.0 and run python manage.py migrate
Inspect the DB table and notice the new constance_constance table created from 0001_initial is empty, and the old constance_config table still exists.
After some debugging, I think it concerns the SQL in 0002_migrate_from_old_table. Specifically, If I remove the try/except block, I get the following exception:
django.db.utils.ProgrammingError: (1064, "You have an error in your SQL syntax;
check the manual that corresponds to your MySQL server version for the right syntax
to use near 'key, value ) SELECT id, key, value FROM constance_config' at line 1")
But by escaping the column names (e.g. ( id, key, value ) → ( `id`, `key`, `value` )), I was able to get the 2nd migration to work (data transferred/old table dropped), leading me to believe it's because of SQL reserved words.
System configuration
Django version: 4.2.4
Python version: 3.10.6
Django-Constance version: 3.1.0
MySQL version: 8.0.18
It will be fixed soon https://github.com/jazzband/django-constance/pull/531
Sorry, my fault
No worries and thanks for your help!
It seems that fixing this error caused an error for me.
System configuration
Django version: 3.1
Python version: 3.9
Django-Constance version: 3.1.0
MySQL version: 8.0.22
Using connection = schema_editor.connection does not work in my case.
I have just experienced the very same issue with MySQL, fixed it on my own and later found this issue...
If anyone lands in here looking for the answer "why": this migration may be indeed fixed, but it is not yet released.
I am not sure if the fix works for me either, I had to run this sqls manually to fix it (state was already messed up a bit):
DELETE FROM constance_constance;
INSERT INTO constance_constance SELECT * FROM constance_config;
DROP TABLE constance_config;
On that note, would it be possible to generate a new release with the corrected migration file?
Our experience with upgrading from 2.9.1 to 3.1.0 was a bit shocking. The newly created table constance_constance was not filled with existing data. The old table constance_config was preserved (luckily). The incomplete transfer also caused the admin interface not saving things and showing a lot of defaults.
We've fixed it by running the following query in the database:
truncate table constance_constance;
insert into constance_constance select * from constance_config;
Hope this helps someone that is feeling the stress :)
To pick up on @mfisco's comment again:
#531 is essential as there is a high risk of data loss, but unfortunately it has not yet been released.
Would it be possible to generate a new release? @camilonova or @sergei-iurchenko
Or at least a warning in the README that there is a risk of data loss with version 3.1.0 would certainly not be amiss.
So just a suggestion, don't stress about it. Thanks a lot!
@camilonova please make a new release. I don`t have such permissions
|
gharchive/issue
| 2023-08-22T21:09:36 |
2025-04-01T04:34:39.710130
|
{
"authors": [
"Volody2006",
"jasisz",
"mfisco",
"schefDev",
"sergei-iurchenko",
"stitch"
],
"repo": "jazzband/django-constance",
"url": "https://github.com/jazzband/django-constance/issues/528",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2398844063
|
docs(changelog): bump to 5.0.0, add token warning
I've increased the version to major, as this kind of a big change asks for one.
Fixes https://github.com/jazzband/django-rest-knox/issues/357
/cc @johnraz
Thanks !
Bit confused by these warnings now. We upgraded to 4.2 a year ago, which broke existing tokens. Will upgrading to 5.0.0 break them again? Or was this warning just added twice to highlight it?
@jmsmkn I'm not sure, and I haven't had the time to test if they will actually break yet.
It might have been a miscommunication and the warning isn't real, but I'm not sure.
If you have some time could you test this and report back here?
It does indeed break them, caused by #272, bit of a pain considering the default is not to use that feature.
Looking through that PR:
Why was sha added? Seems unnecessary.
Why was max_length increased to include CONSTANTS.MAXIMUM_TOKEN_PREFIX_LENGTH? CONSTANTS.MAXIMUM_TOKEN_PREFIX_LENGTH is not considered when saving the token. The TOKEN_KEY_LENGTH was increased in that same PR, which causes the breakage of existing tokens. Can't old-style, shorter token keys be handled when authenticating and then updated to the new longer style?
Hi @jmsmkn @giovannicimolin This MR does not break it. I added tests at the time to ensure that, however the release does.
What breaks the token compatibility is: https://github.com/jazzband/django-rest-knox/commit/b02a1553baff849d0f3a2b83da5dda5bf793157c
I filed an issue here regarding this as well: https://github.com/jazzband/django-rest-knox/issues/357
b02a155 does not break anything as that is just adding a warning to the documentation.
I tested this by generating the tokens with 4.2.0 (which already includes the changes of the salt removal described by b02a155), then installed 5.0.1, migrated, and ran a request against that process.
Both 4.1.0 -> 4.2.0 and 4.2.0 -> 5.0.1 break the tokens.
@jmsmkn Thanks for taking the time to test this 👌🏻
@max-wittig I reviewed the PR and the test there is not testing that a token generated before your change will work after. In order to properly do that you would have to checkout the code at the previous version then at the next etc. So we cannot state strongly that this is not breaking it.
I do trust @jmsmkn carefully tested this manually and I believe there is an actual different breakage between 4.2.0 and 5.0.0.
Now, @jmsmkn if you feel things should be done differently, I think it would be more constructive to open a PR so we can discuss the code and your change proposal rather than hypothetically talking about what could / should have been done.
5.0.1 is out, there is no going back so let’s look ahead and try to avoid breaking changes in the future.
Cheers!
@max-wittig I reviewed the PR and the test there is not testing that a token generated before your change will work after. In order to properly do that you would have to checkout the code at the previous version then at the next etc. So we cannot state strongly that this is not breaking it.
Ahh I see. It just tested that it works when changing the prefix. I wasn't aware of the breakage that I've caused! I'm sorry!
5.0.1 is out, there is no going back so let’s look ahead and try to avoid breaking changes in the future.
Yes that is true. Maybe there needs to be a test in CI using tokens in main and comparing them with any additional PR to see if the tokens still work.
Yes that is true. Maybe there needs to be a test in CI using tokens in main and comparing them with any additional PR to see if the tokens still work.
Added in #362 for tokens from 4.2.0 without prefixes.
|
gharchive/pull-request
| 2024-07-09T18:11:17 |
2025-04-01T04:34:39.726165
|
{
"authors": [
"giovannicimolin",
"jmsmkn",
"johnraz",
"max-wittig"
],
"repo": "jazzband/django-rest-knox",
"url": "https://github.com/jazzband/django-rest-knox/pull/358",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
706609204
|
Is there an option to convert words to number?
Is there an option to convert words to number? (the reverse of number to words)
Not that I'm aware, but there is word2number. Perhaps that does what you need?
I think so, I will give it a try.
|
gharchive/issue
| 2020-09-22T18:42:40 |
2025-04-01T04:34:39.727804
|
{
"authors": [
"jaraco",
"ndvbd"
],
"repo": "jazzband/inflect",
"url": "https://github.com/jazzband/inflect/issues/114",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
247110179
|
Binary wheels for windows
Please consider providing binary wheels for windows. They can be auto built on appveyor. I was able to locally build current master for python 2.7 32bit by installing the "python compiler release" from Microsoft and using vcpkg-built libjpeg-turbo which is somewhat heavy weight. I also had to define snprintf to _snprintf to get it linked. I am not really sure it works yet.
This sounds like a nice thing to build into future CI improvements.
I'd happily review such a change but I'm unsure how to start doing it myself: contributions welcome!
|
gharchive/issue
| 2017-08-01T15:32:26 |
2025-04-01T04:34:39.734851
|
{
"authors": [
"cbm755",
"zwn"
],
"repo": "jbaiter/jpegtran-cffi",
"url": "https://github.com/jbaiter/jpegtran-cffi/issues/20",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
570659601
|
No DURATION without AT
The current version of remind doesn't allow DURATION to be specified without AT.
Also, while we're at it remove some extra spaces from the reminder output.
Woops, I screwed up the commits in this PR. Closing to clean up and resubmit a new one.
|
gharchive/pull-request
| 2020-02-25T15:40:21 |
2025-04-01T04:34:39.735873
|
{
"authors": [
"jikamens"
],
"repo": "jbalcorn/ical2rem",
"url": "https://github.com/jbalcorn/ical2rem/pull/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2679631129
|
BibTex Citeration of ECCV2024.
Great work! Could you provide the BiBTex Citeration format of ECCV2024? It is important for me to cite your paper. I only found the arXiv's format on Google Scholar.
Thank a lot for your reply!
|
gharchive/issue
| 2024-11-21T14:13:51 |
2025-04-01T04:34:39.771191
|
{
"authors": [
"Core9724"
],
"repo": "jbji/RepVF",
"url": "https://github.com/jbji/RepVF/issues/3",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
230581906
|
use version number instead of release in .travis.yml
release will change over time, but your REQUIRE file says this package supports julia 0.4 so it should continue to be tested
Coverage remained the same at 98.876% when pulling 09d7cc59fd800222616d54cfaf2fdce08fe64f17 on tkelman:travisver into 231befb1db6124b7005af4e1593da864b1bc167e on jbn:master.
Thanks, @tkelman!
|
gharchive/pull-request
| 2017-05-23T04:03:44 |
2025-04-01T04:34:39.773313
|
{
"authors": [
"coveralls",
"jbn",
"tkelman"
],
"repo": "jbn/WordNet.jl",
"url": "https://github.com/jbn/WordNet.jl/pull/6",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
359448244
|
"Cannot obtain java method" error
java.lang.IllegalStateException: Cannot obtain java method for: Add
--
at org.jboss.fuse.wsdl2rest.util.IllegalStateAssertion.assertNotNull(IllegalStateAssertion.java:50)
at org.jboss.fuse.wsdl2rest.impl.codegen.CamelContextGenerator.getJavaMethod(CamelContextGenerator.java:121)
at org.jboss.fuse.wsdl2rest.impl.codegen.CamelContextGenerator.addTypeMapping(CamelContextGenerator.java:94)
at org.jboss.fuse.wsdl2rest.impl.codegen.CamelContextGenerator.process(CamelContextGenerator.java:76)
at org.jboss.fuse.wsdl2rest.impl.Wsdl2Rest.process(Wsdl2Rest.java:92)
at org.fusesource.ide.wsdl2rest.ui.wizard.Wsdl2RestWizard.generate(Wsdl2RestWizard.java:323)
at org.fusesource.ide.wsdl2rest.ui.wizard.Wsdl2RestWizard.access$0(Wsdl2RestWizard.java:286)
at org.fusesource.ide.wsdl2rest.ui.wizard.Wsdl2RestWizard$1.run(Wsdl2RestWizard.java:112)
see more details on https://issues.jboss.org/browse/FUSETOOLS-3091 (includign the wsdl to reproduce)
Created PR - https://github.com/jboss-fuse/wsdl2rest/pull/63
All fixes merged
|
gharchive/issue
| 2018-09-12T12:02:58 |
2025-04-01T04:34:39.781861
|
{
"authors": [
"apupier",
"bfitzpat"
],
"repo": "jboss-fuse/wsdl2rest",
"url": "https://github.com/jboss-fuse/wsdl2rest/issues/60",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2553922813
|
🛑 https://notes.runtimeterror.dev is down
In 278cae4, https://notes.runtimeterror.dev (https://notes.runtimeterror.dev) was down:
HTTP code: 404
Response time: 245 ms
Resolved: https://notes.runtimeterror.dev is back up in 5a7243e after 1 hour, 7 minutes.
|
gharchive/issue
| 2024-09-28T00:47:35 |
2025-04-01T04:34:39.795830
|
{
"authors": [
"jbowdre"
],
"repo": "jbowdre/upptime",
"url": "https://github.com/jbowdre/upptime/issues/209",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1292671496
|
Simplify Token form
Description
The Token form/drawer needs to be simpler.
Screenshots
Link to thread: https://discord.com/channels/939317843059679252/970421807767236708/999120040840413234
|
gharchive/issue
| 2022-07-04T06:06:07 |
2025-04-01T04:34:39.820776
|
{
"authors": [
"aeolianeth",
"strath-m"
],
"repo": "jbx-protocol/juice-interface",
"url": "https://github.com/jbx-protocol/juice-interface/issues/1327",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
624886915
|
proxying mail.google.com
Hello All,
Is there a way of proxying mail.google.com ?
I was trying different combinations of settings : with and without custom locations, certificates, etc without any luck.
Thanks a lot for an awesome product!
What exactly are you trying to accomplish by proxying mail.google.com?
Thanks for your prompt reply.
Basically, I am trying to avoid gmail being blocked by government, employer, etc.
Not sure even if it is possible with proxying through nginx.
Regards!
I would consider trying Cloudfront for that purpose, would be faster internet access from anywhere.
I can't explain why off the top of my head but I'd expect google to have measures in place to prevent this kind of proxying, otherwise their Gsuite offering could be diminished.
Yeah, makes perfect sense.
Will keep on looking on the best way of doing this.
Again, thanks a lot for an awesome product.
Using it and love it!
|
gharchive/issue
| 2020-05-26T13:32:37 |
2025-04-01T04:34:39.824385
|
{
"authors": [
"Indemnity83",
"jc21",
"jdedev"
],
"repo": "jc21/nginx-proxy-manager",
"url": "https://github.com/jc21/nginx-proxy-manager/issues/425",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
187943241
|
Documentation Request: Sample Project
More than a sample Cakefile, I think a working sample project would be very illuminating for beginner Xcake users.
I'm trying to convert a project now, and although Xcake will make just fine, the resulting project file can't be parsed by Xcode. I assume this is due to an error in the Cakefile.
Of course, more complete Cakefile validation would help solve this issue as well, but I think it would also be helpful to have a working sample project to compare to.
A working sample project would help new users to see both the configuration and the results. Because the current sample Cakefile is just a fake Cakefile and not a whole sample project, I can't actually build it and see the output. As a result, I can't do the "sanity check" that my basic configuration is correct (and that Xcake is compatible with my current version of Xcode).
@divergio Thanks for raising this, I'll try to build a better example and get it working with all Xcode versions.
@divergio to be honest I think that having a whole sample project just does not make sense. The sample Cakefile(s) are pretty self-explanatory. This is the idea behind Xcake - you write easy to read declarative configuration file and Xcake takes care about generating whole Xcode project based on this.
After learning the sample files, If you do not understand what's going there - probably you need to learn more about how Xcode project file is organized first. Also, in my opinion, Xcake is not supposed to be used by beginners and/or for small projects with single developer. It's a tool for saving time for mid- and senior level developers when they work on a relatively big project with multiple collaborators. So probably you just don't need this tool yet? @jcampbell05 correct me if I'm wrong.
If you are still sure it's a tool you really need in your development process and you are trying to write Cakefile for your project, but project generated from your Cakefile does not work properly - try to comment almost everything in your Cakefile and keep there only very basic configurations. Generate the project and check if it works. The ideas is to make a very-very basic Cakefile first, make sure generated project opens properly by Xcode (even if it's empty yet). After you find minimum project configuration that generates functional project file - start uncommenting more configurations, line by line. After each addition - remove existing project file and regenerate project. At some point you'll find which exact line leads to a broken project file, that can not be opened by Xcode properly. At this point you most likely will get the idea what's wrong and how to fix it. Or you can come back here and ask us.
Hope that will help you resolve your issue. Cheers!
@maximkhatskevich Thanks for your reply.
Perhaps I wasn't clear, I'm not referring to beginner Xcode users, I'm referring to people trying out xcake for the first time or perhaps just evaluating xcake for use in their project. The relatively large project with multiple collaborators that I'm working on is exactly why I am interested in xcake (I'm so sick of digging through pbxproj files to resolve merge errors!).
I already resolved my issue getting a Cakefile to work using a trial-and-error approach similar to what you suggested. The reason I suggested a sample project is to help others avoid this step, or at least to have one working version they can play with and take apart to learn how xcake works.
Further, I think it's best to figure out how a Cakefile works and to test its functionality for your own project by trying it out on a smaller project first (i.e. don't jump straight into trying to convert your 1000 file project). A sample project provides a place for new/potential users to do this test.
To reiterate, the drawback of providing only a sample Cakefile is that it can't actually generate a working project file. The files it references aren’t there. It is documentation, not a demonstration.
@divergio okay, I see your point. I'll try to prepare something like this for you soon.
@divergio keep an eye on this pull request: https://github.com/jcampbell05/xcake/pull/127 Once it's merged, you'll find everything you need on master branch, also check ReadMe file for comments about samples.
|
gharchive/issue
| 2016-11-08T09:37:44 |
2025-04-01T04:34:39.843451
|
{
"authors": [
"divergio",
"jcampbell05",
"maximkhatskevich"
],
"repo": "jcampbell05/xcake",
"url": "https://github.com/jcampbell05/xcake/issues/118",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
641800245
|
add exclude properties
I need the exclude properties in combination with the includeallproperties attribute.
1.0.6 of the NuGet package was just uploaded to NuGet.org - danke!
|
gharchive/pull-request
| 2020-06-19T08:24:47 |
2025-04-01T04:34:39.845175
|
{
"authors": [
"chrisdecker1201",
"jcapellman"
],
"repo": "jcapellman/NLog.Targets.Fluentd",
"url": "https://github.com/jcapellman/NLog.Targets.Fluentd/pull/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1840666441
|
parent parent task
First of all, I thank you for your cooperation and quick response
in our project we have four level with subtask
root tracker :- tender
parent parent tracker: wallet
parent tracker: company
child tracker: Paid
your plugin help me to contral that and filier it between levels
if i need search which paid from any wallet its hard its can be solved by added parent parent
i hope if u can added
1- parent parent tracker
2- parent parent status
I am so thankful you to saved my life
@jcatrysse
I think we added
parent.parent_id
Get Outlook for Androidhttps://aka.ms/AAb9ysg
From: Jan Catrysse @.>
Sent: Tuesday, August 8, 2023 1:37:58 PM
To: jcatrysse/redmine_parent_child_filters @.>
Cc: ashrafalzyoud @.>; Author @.>
Subject: Re: [jcatrysse/redmine_parent_child_filters] parent parent task (Issue #3)
I will have a look, but probably not today… I'll keep you posted. Can this also be "a" parent filter as an extra? This wil anyway be a very slow and complicated filter :-)
I can imagine something like "a" parent status = x
Not sure if negative filters will be very useable.
—
Reply to this email directly, view it on GitHubhttps://github.com/jcatrysse/redmine_parent_child_filters/issues/3#issuecomment-1669365911, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AL7CU7YBGXPF3IZIRL5FJVDXUIJINANCNFSM6AAAAAA3H7OBT4.
You are receiving this because you authored the thread.Message ID: @.***>
I made you an update, but this is not on the second or third parent, it is on any parent.
Commit: ffe4d6f0dcdc5a12189cef286521d3e9f3d6c204
Already on master, so you can try and experiment a bit.
Checkout the new release! Release: 0.0.5 (56fe60eb096a2d209d55a87abad930455c673c15) ...
It was a long day and night, but I think I managed to make you something.
I added you some JavaScript to make the filter more intuitive, Disabling incompatible depths on multiselect filters.
Please reach out to me if the changes were useful.
|
gharchive/issue
| 2023-08-08T06:43:52 |
2025-04-01T04:34:39.853300
|
{
"authors": [
"ashrafalzyoud",
"jcatrysse"
],
"repo": "jcatrysse/redmine_parent_child_filters",
"url": "https://github.com/jcatrysse/redmine_parent_child_filters/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
961931290
|
How to compile for Unity IL2CPP
It compiles but getting runtime error while running on Android IL2CPP 64 bit. documentation mentions how to compile for mono.
`08-05 20:03:26.004 3817 3841 I Unity : SystemInfo CPU = ARM64 FP ASIMD AES, Cores = 8, Memory = 5500mb
08-05 20:03:26.004 3817 3841 I Unity : SystemInfo ARM big.LITTLE configuration: 1 big (mask: 0x80), 7 little (mask: 0x7f)
08-05 20:03:26.004 3817 3841 I Unity : ApplicationInfo com.DefaultCompany.Empty version 0.1 build 2de827ca-90f1-4fb6-91dd-44feeb7ef817
08-05 20:03:26.004 3817 3841 I Unity : Built from '2020.1/staging' branch, Version '2020.1.6f1 (fc477ca6df10)', Build type 'Release', Scripting Backend 'il2cpp', CPU 'arm64-v8a', Stripping 'Enabled'
08-05 20:03:30.627 3817 3841 E Unity : SocketException: mono-io-layer-error (113)
08-05 20:03:30.627 3817 3841 E Unity : at System.Net.Sockets.SocketAsyncResult.CheckIfThrowDelayedException () [0x00000] in <00000000000000000000000000000000>:0
08-05 20:03:30.627 3817 3841 E Unity : at System.Net.Sockets.TcpClient.EndConnect (System.IAsyncResult asyncResult) [0x00000] in <00000000000000000000000000000000>:0
08-05 20:03:30.627 3817 3841 E Unity : at WatsonTcp.WatsonTcpClient.Connect () [0x00000] in <00000000000000000000000000000000>:0
08-05 20:03:30.627 3817 3841 E Unity : at ClientManager.Start () [0x00000] in <00000000000000000000000000000000>:0
08-05 20:03:30.627 3817 3841 E Unity :
08-05 20:03:30.627 3817 3841 E Unity : (Filename: currently not available on il2cpp Line: -1)
`
Hi @Ahmed310 sorry I don't have the environment necessary to reproduce this. Cursory google searches indicate that there may be a problem with the IP/port on which you're trying to connect. Please share code for how you are instantiating the client and connecting to the server.
@jchristn you are right the ip on serverside was different then client side, now the connection is established but getting exception while sending message.
here is the sample code
` public Button btn;
WatsonTcpClient client;
void Start()
{
client = new WatsonTcpClient("192.168.10.7", 9000);
client.Events.ServerConnected += ServerConnected;
client.Events.ServerDisconnected += ServerDisconnected;
client.Events.MessageReceived += MessageReceived;
client.Connect();
btn.onClick.AddListener(onClickBtn);
}
void onClickBtn()
{
if (client != null && client.Connected)
{
client.Send(BitConverter.GetBytes(786));
client.Send(BitConverter.GetBytes(1010));
}
else
{
Debug.Log("Client is null or not connected");
if (client != null)
{
Debug.Log($"Oh Well {client.Connected}");
}
}
}`
Debugging results are as follow if I call the send function I get this exception
08-06 01:00:15.957 30780 30830 E Unity : NullReferenceException: Object reference not set to an instance of an object. 08-06 01:00:15.957 30780 30830 E Unity : at WatsonTcp.WatsonTcpClient.Send (System.Int64 contentLength, System.IO.Stream stream, System.Collections.Generic.Dictionary2[TKey,TValue] metadata) [0x00000] in <00000000000000000000000000000000>:0
08-06 01:00:15.957 30780 30830 E Unity : at WatsonTcp.WatsonTcpClient.Send (System.Byte[] data, System.Collections.Generic.Dictionary2[TKey,TValue] metadata, System.Int32 start) [0x00000] in <00000000000000000000000000000000>:0 08-06 01:00:15.957 30780 30830 E Unity : at ClientManager.onClickBtn () [0x00000] in <00000000000000000000000000000000>:0 08-06 01:00:15.957 30780 30830 E Unity : at UnityEngine.Events.UnityAction.Invoke () [0x00000] in <00000000000000000000000000000000>:0 08-06 01:00:15.957 30780 30830 E Unity : at UnityEngine.Events.UnityEvent.Invoke () [0x00000] in <00000000000000000000000000000000>:0 08-06 01:00:15.957 30780 30830 E Unity : at UnityEngine.EventSystems.ExecuteEvents+EventFunction1[T1].Invoke (T1 handler, UnityEngine.EventSystems.BaseEventData eventData) [0x00000] in <00000000000000000000000000000000>:0
08-06 01:00:15.957 30780 30830 E Unity : at UnityEngine.EventSystems.ExecuteEvents.Execute[T] (UnityEngine.GameObject target, U
`
The client itself is not null but the status Connected remains false
in logs it prints Oh Well false : meaning Connected is false
Can you set client.Events.ExceptionEncountered and paste the event argument contents here? Also please set client.Settings.Logger and paste any relevant log messages.
sure I will by tomorrow
After enabling logs & exceptions it turns out the Newtonsoft was throwing the exception as it was not compatible with il2cpp.
here are the steps I did to make it functional
1: Use source code of WatsonTCP rather dll
2: use il2cpp compatible Newtonsoft, I used this [ https://github.com/jilleJr/Newtonsoft.Json-for-Unity/tree/13.0.102 ]
Good to know. Thanks for sharing @Ahmed310. I may have to look into implementing custom serializers.
custom serializer will be even better as It will remove the dependency of newtonsoft
|
gharchive/issue
| 2021-08-05T15:14:17 |
2025-04-01T04:34:39.872374
|
{
"authors": [
"Ahmed310",
"jchristn"
],
"repo": "jchristn/WatsonTcp",
"url": "https://github.com/jchristn/WatsonTcp/issues/148",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1199532956
|
Refine commandline misusages
Refine diagnostic for misusage of commandline e,g, when user will not properly give arguments then print Help next to Panic of Rust.
fixed with #16
|
gharchive/issue
| 2022-04-11T07:47:40 |
2025-04-01T04:34:39.887091
|
{
"authors": [
"jczaja"
],
"repo": "jczaja/e-trade-tax-return-pl-helper",
"url": "https://github.com/jczaja/e-trade-tax-return-pl-helper/issues/11",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
1416923915
|
Merge all Proton products
merged already existing ProtonMail and ProtonVPN to Proton, added some subdomains
added subdomains for missing services ProtonDrive and ProtonCalender to Proton
changed difficulty of Proton to medium, because it requires selecting a reason and giving a short feedback
Thanks for your contributions!
|
gharchive/pull-request
| 2022-10-20T16:26:46 |
2025-04-01T04:34:39.948227
|
{
"authors": [
"Kefaku",
"tupaschoal"
],
"repo": "jdm-contrib/jdm",
"url": "https://github.com/jdm-contrib/jdm/pull/1388",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
3063845
|
Cache State Backends
Alright, here's my first crack at what @cyberdelia, @bryanveloso, and I were talking about in #88. Essentially, it's my solution to #74, #75, #81, and #88. In summary, ImageSpecs are no longer directly responsible for creating and deleting cached images. Instead they delegate the responsibility to "cache state backends."
Cache state backends must implement two methods: invalidate() and validate(). The default backend maintains the current behavior: invalidating an image deletes it, and validating it creates the file (if it doesn't already exist). This behavior is very safe (which is why I think it makes a good default) but, as @mkai noted in #81, it can be prohibitively slow for certain file storage backends. This new setup allows him to define a much more optimistic backend that, for example, could immediately regenerate cache images on invalidation and define validation as a no-op. Since images are generally infrequently invalidated and frequently validated, this system would probably work out much better for him.
Similarly, @davelowe could use a custom cache state backend to queue the creation of images via celery instead of having the processing of the image hold up his page loads.
API Changes
I ended up removing the pre_cache argument from the ImageSpec constructor. The main reason for this is that cache state backends provide a much more powerful mechanism for controlling this behavior. @bryanveloso's :+1: pushed me over the edge. I put in an exception to warn anybody who migrates.
Since the decision about which cache state backend to use will likely be tightly coupled to your choice of storage, I wanted to make it very similar to use. Therefore, there are two ways to set a cache state backend:
Pass an instance to the ImageSpec constructor (like the storage kwarg)
Set IMAGEKIT_DEFAULT_CACHE_STATE_BACKEND in settings.py (like DEFAULT_FILE_STORAGE)
In place of ikflush, we now have two shiny new management commands: ikcacheinvalidate and ikcachevalidate. (The old ikflush was dependent on pre_cache and that wasn't going to cut it anymore.) They work pretty much exactly how you'd expect. ikcachevalidate also accepts a --force flag which makes it behave like ikflush did when pre_cache was true; using this flag will cause each file to be invalidated and then immediately validated. (This is pretty much the same as running ikcacheinvalidate and then ikcachevalidate, except that would invalidate all files before validating any of them.)
Finally, IK metadata ("_ik") is now bound to each model instance. I doubt anybody who's not hacking around in the guts of IK will even notice this change, but it makes acquiring a list of bound filefields a little prettier.
So
This is another nontrivial change, so I figured I'd submit a pull request for us to discus. So let's hear your thoughts! And, if you have some free time after that, here's some other stuff to do:
Test it out (I haven't really tested much at all)
Write some cache state backends! It would be nice if we could bundle a few with IK. The optimistic one I mentioned earlier might be a candidate.
Docs, please!
First thought while playing with this: spec_file.invalidate() is called in both the post_save_handler and the post_delete_handler, but no context is passed. It'd be awfully nice if my custom cache state backend's invalidate method knew if the instance was in the process of being destroyed or not. I wouldn't want to set an image regeneration task in motion if it were unnecessary.
@davelowe Yeah, you're definitely right about the context. 011c0c2 adds a new method named clear() which is now called by the post_delete receiver instead of invalidate()—the distinction being that calling clear() is not a request for validation. In the case of the default backend, clear() and invalidate() happen to be the same, but they needn't be. I'm not thrilled with having a clear() and delete() method, but I don't want to deviate from the File API's delete behavior and at least "clear" seems less likely to cause confusion than, for example, "remove."
As far as passing the instance—it is passed! However, it's passed as a property of the file object (file.instance) and not an argument itself. Incidentally, @cyberdalia also wanted/wants it to be passed as an argument, but I have my reasons for disagreeing. I never articulated them in #88, so I'll do so now:
First, though—just so everybody reading is on the same page: the file object (which is passed to the cache state backend methods) carries a reference to the instance (as well as the field and plenty of other great information), so passing the instance as an argument would be redundant. This isn't to say that the redundancy doesn't add value—it does (clarity). But anything that you want to do that requires the instance, you can do.
So if I admit it adds clarity, why not just do it? The reason boils down to this: I want to keep the cache state backends as decoupled from the model layer as possible. Basically, I can foresee wanting to use a cache state backend separately from ImageSpec fields; specifically, I'm still mulling over the possibility/desirability of manipulating images via template tags (as discussed with @dmeehan in #78). If someone were to write a cache state backend for images in this context (or any other that doesn't involve models), it would be very strange to implement an interface that has a model "dependency" (albeit in the weak form of an always-None keyword argument). Of course, any backends that do rely on the presence of a model instance are not going to be very useful for those that have none, but I think that many (most?) will not. Also, not including the instance keyword argument should be a subtle hint that, if you can do it without the model, you probably should.
Switching topics completely, the cache state backend system has a feature which I haven't mentioned yet: an register_field callback. This method (if it exists) is called by contribute_to_class so—in the event that your cache backend does need to do things more complicated than validate and invalidate—you won't be resorted to hacking around. You said:
As it is, I really couldn't do anything in the cache state backend. I'd probably just remove file deletion and use post_save & post_delete handlers at the model level.
I think that this no longer applies (given that you actually do have access to the model instance), but if this were necessary, the register_field method would be the perfect place to add your receivers:
class DavesCacheStateBackend(object):
def register_field(self, model_class, field, attname):
post_delete.connect(...)
pre_save.connect(...)
post_save.connect(...)
That way everything is nicely contained in your backend.
Again, because the instance is available as a property on the field, my hunch is that this will rarely be necessary. However, I can think of at least one concrete application: comparing the image's filename at the time of model initialization to its filename when the model is saved to determine whether the image has changed (as @cyberdelia did in his original proposal). As I mentioned to him, I think the potential for false negatives make this technique unpalatable as a default implementation, but I love it as an opt-in alternative.
Well... I feel a bit sheepish that I didn't realize the instance was accessible from the file object. :) That's perfect, and I completely agree that passing instance separately couples it too closely to the model layer.
And wow, register_field just blows this all up! In a good way, of course. pre_save.connect(...) might really come in handy (especially if there's a reliable way to determine if the image changed... time to experiment).
Awesome, @matthewwithanm! The custom CacheStateBackend() is a wonderful addition to IK.
Thanks @davelowe!
Actually, if you can figure out a reliable way to determine if the image has changed, I'd like that to be in the models module. I couldn't think of any that were completely dependable (short of monkey-patching ImageField or providing a custom one) so I decided to live with the false positives in _post_save_handler. Of course, a way to avoid those would be ideal!
@davelowe: Once you get a handle of your Celery backend, I'd love to include it as a use-case example in the documentation.
I will give a try to this branch tomorrow. I hope to come back with some backends for it.
Seems a bit strange that you have to do a custom backend to have a pre_cache behavior isn't it ?
However, here is an filename change backend :
# -*- coding: utf-8 -*-
from django.db.models.signals import post_init
def _post_init_handler(sender, instance, **kwargs):
value = getattr(instance, _post_init_handler.field)
setattr(instance, "_%s_state" % _post_init_handler.field, value)
class FilenameCacheStateBackend(object):
def has_changed(self, file):
field_name = getattr(file.field, 'image_field', None)
return getattr(file.instance, "_%s_state" % field_name) != getattr(file.instance, field_name)
def is_invalid(self, file):
if not getattr(file, '_file', None):
# No file on object. Have to check storage.
return not file.storage.exists(file.name)
return False
def validate(self, file):
if self.is_invalid(file):
file.generate(save=True)
def invalidate(self, file):
if self.has_changed(file):
file.delete(save=False)
def clear(self, file):
file.delete(save=False)
def register_field(self, model_class, field, attname):
uid = "%s.%s" % (model_class.__module__, model_class.__name__)
_post_init_handler.field = field.image_field
post_init.connect(_post_init_handler, sender=model_class, dispatch_uid="%s_init" % uid)
I've made a more generic celery backend (with a pre_cache like behavior) : https://gist.github.com/1778947
@cyberdelia said:
Seems a bit strange that you have to do a custom backend to have a pre_cache behavior isn't it ?
Well, we could use a setting (or even resurrect the constructor property). The problem is that there's no guarantee that the backend would actually honor it. Besides, I don't think you'll have to use a custom one (we'll bundle one), just have to change the IMAGEKIT_DEFAULT_CACHE_STATE_BACKEND setting.
I like the FilenameCacheStateBackend but I think it should probably be a mixin. We don't want people to have to recreate this behavior over and over. So I think all we'll need is has_changed and the register_field. We could wrap invalidate in a meta class, but that might be too much magic.
@davelowe @cyberdelia I think both of your implementations might be assuming that there's only one source image per instance. This was the case in the pre-1.0 days (with the ImageModel class), but not anymore. I think we're going to have to iterate over the specs in our post save handlers to check them all.
As long as the backend dictate the pre_cache behavior that's ok. What if I want aFilenameCacheStateBackend that behave like pre_cache ? We need to bundle two version of every backend ? I might be missing something.
Feels like a FilenameCacheStateMixin would be a good thing, as long as we provide a full implementation.
My current needs are to listen to changes on some instance attributes and send task to celery, so maybe we could have a AttributesChangeMixin too.
And yes, my use of function attribute with _post_init_handlermight be problematic with multiple source :| But the celery one looks correct.
we're going to start testing this branch as well as we're storing on S3.
I dug a little more deeply into Django's innards tonight and I think I may have been overly paranoid about the whole maybe-the-filenames-won't-change thing. Long story short, I think we're safe comparing the "before and after" File objects to know when to invalidate. So no need for a FilenameCacheStateMixin—it's baked right into ImageSpec as of 1956e16. This should drastically reduce I/O operations, even when using the pessimistic backend.
I whipped up a NonValidatingCacheStateBackend (15b15af). At first I was going to call this OptimisticCacheStateBackend, but I think that would be underemphasizing its most important feature—namely, that it doesn't actually validate anything. This one's is going to cut your storage access down to nothing, but if a cached file were to go missing somehow, you could be in trouble. (In practice, this probably isn't a concern and you could also run a cron or something to guarantee the files are there.)
I also forked @cyberdelia's gist to demonstrate my own idea about how a Celery backend would work. It's mostly the same, however mine guarantees the existence of a file (i.e. it's "pessimistic") and invalidating doesn't delete the file. I added comments so read them to see why I did this. Of course, we could also create a NonValidatingCeleryCacheStateBackend (phew!)—there are a lot of ways you could tweak this to get exactly what you want as far as balancing speed and reliability.
I think that most other backends will probably take the form of extending PessimisticCacheStateBackend and overriding is_invalid and invalidate. For example, to use the database or a cache to store the state. But I guess we won't really know until we put it out there and people start tinkering.
What do you guys think about calling them image cache backends instead of cache state backends? I think "cache state" might get a little confusing once other caches enter the picture (like the one in my last post).
Any objections to me merging this? I've been hacking on IK all weekend and I have a CRAPLOAD of things ready to go in, but they're dependent on this and all the rebasing and conflict resolution is becoming a bit of a hassle.
OT: Let's keep IK 2 in alpha for at least the next week…
This flurry of activity has been leading up to an experimental templatetag feature (which I'd like in 2.0—it probably more than justifies a full revision bump itself). Once this and #95 are merged, I'm going to merge a boring refactoring branch that I have waiting which paves the way for the templatetags and then push the templatetag branch and open a request for that.
I'll merge it in right now.
|
gharchive/issue
| 2012-02-02T06:20:35 |
2025-04-01T04:34:39.975878
|
{
"authors": [
"bryanveloso",
"chrisdrackett",
"cyberdelia",
"davelowe",
"matthewwithanm"
],
"repo": "jdriscoll/django-imagekit",
"url": "https://github.com/jdriscoll/django-imagekit/issues/89",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
554284014
|
Really slow to create control
Hi,
I just came across your library. It's great and just what I need.
One wierd thing though; It takes about 3 seconds to create the control from a storyboard.
(on a fairly recent iPad air)
I'm working around this by just keeping a static version of my picker view controller - but it seems like an odd slowdown in a control that doesn't look like it should be super-complex
I'm using everything in with default settings
any suggestions?
thank you.
First of all, apologies for taking so long to respond. I was ill.
Second, I'm actually having difficulty reproducing your problem, which is annoying, as obviously you can see it, but as I can't I'm not sure what to do. However, I think it's something to do with the new-style SwiftUI constraints mechanism.
Looking at ColourPickerVC.storyboard, I note that it's currently set for the main View to use the new approach to layout (Translates Mask into Constraints), while the Picker uses the old approach (auto). It tried switching them both to auto and I started to get some second delay before the Picker first appeared. Changing back, that delay seemed to go away.
So, the problem may be simply that this is a pre-SwiftUI control, and I need to update it to take SwiftUI into account. Would that be an acceptable next step?
Thanks for getting back on this - I appreciate this.
Do you mean SwiftUI?
My demo project isn't using SwiftUI in any way. (It does have a couple of leftover import statements, but you can remove them as they're not actually used)
Constraints don't sound like a very likely problem. I'm manually fixing the height/width of the picker in the demo, so it seems unlikely that the solver would be having problems there...
Of course - a SwiftUI picker would be fantastic - but that's a separate thing :)
I seem to have become confused.
The reason I suspect constraints is that, if you set them up correctly, with the picker a fixed size (as you have) then the (fairly intensive) process of building the colour wheel itself (which involves computing and setting the colour of a bitmap pixel by pixel to get really clean graphics) will be done once and for all in your Storyboard / NIBFile. On the other hand, if the width of the thing is computed dynamically, it can take a very long time to do that once-off work.
Have a look at the example : the version on master uses a width based on the screen-size, and loading it on an iPad takes a long time. On the other hand, the version on branch fixed-width fixes the width at 335 pixels, and it loads more or less instantly.
Given its similarity to that you're doing, the only thing I can think of that might be causing a bit of delay is that the width of the pop-over view you're using may be dynamically computed. What happens if you replace it with a full-screen view, like mine?
That's really interesting.
if I turn off the presentation animation, then:
presenting fullscreen, the controller takes ~ 1.6seconds
presenting as a form, it takes 3.3, so about twice as long
digging down, if the view controller is a form, then layoutSubviews is called twice on ColorWheel
so updateImage is called twice.
I assume that this is the expensive function.
the frame hasn't changed between the two layout calls, a check on whether the frame has changed saves some time
(now 1.9 seconds)
var laidOutFrame:CGRect?
public override func layoutSubviews() {
print("layout: \(self.frame)")
super.layoutSubviews()
if self.frame != laidOutFrame {
mid=bounds.mid
radius=0.5*Swift.min(bounds.width,bounds.height)-Swift.max(0,borderWidth)
updateImage()
laidOutFrame = self.frame
}
}
But here's the weirdest Heisenbuggyness. If I run the code on the same iPad with the time profiler, then showing the controller takes 0.4 seconds.
This seems nuts - but it's really true. (testing on my iPad Air 2 (Model A1566))
I tried debugging with release-level optimisation turned on, and I don't see a speed change, so I don't think that's the issue.
It's very odd.
4x speed gain by turning on the time profiler (which should slow things down...)
That is all seriously weird.
You're right in identifying the slow bit. I can see a possible work-around for that: if I do some cacheing / checking, so if the size hasn't changed since the last layout, it doesn't bother to recompute, then that should work. I'll try that tomorrow.
As for why it does a double layout - thanks for that indicator. I'll have a dig to see what the triggers are. Might not got anywhere, but surely interesting to know.
And as the four-fold speed up when you're running Instruments? Er, WHAT? That shouldn't happen.
So I have two actions:
Put in checks / a cache so there control doesn't rebuild the wheel if it already has one the right size
Check calling conditions for the layout
I'll start in on (1) tomorrow.
awesome - thank you.
I have implemented a simple kludge that means that, though the layout method is still called twice (I don't seem to be able to do anything about the - presumably something weird deep in the guts of UIKit) - it only does the work of computing the wheel first time round.
It seems a bit faster, but I have yet to do a proper timing test. SO, before I release it, I just need to :
(1) check whether I have introduced any unanticipated new features
(2) hopefully come up with a slightly more elegant way of doing it
(3) rebuild the Pod and issue a new version
If you'd like to try out a version of it, I'm happy to oblige, but be warned, it isn't nicely packaged as a Pod (obviously) and I don't guarantee anything much by way of not doing weird, unexpected things.
thanks - I'm running from that branch now and it seems to be fine.
it certainly looks like you have stopped the double calculation
thanks.
I'd love to know why the profiler speeds things up. That feels like a compiler bug.
ok - the profiling vs debug thing turns out not to be mysterious at all.
profiling uses release configuration.
If I switch the run settings to use the release config, then I get the same timing without the profiler.
which is now a very manageable 0.5 seconds.
Thank you for working that out. Of course, it begs the question of why debug settings caused a double render, but, as I said above, I suspect that's something deep in the implementation of Storyboards that we'll never know.
I won't close the issue just yet, as I haven't quite finished testing, but as soon as I have, I will update the Pod and then close the issue.
to be clear - the double layout thing is true in debug and release.
It's just that in release, the work is much faster generating the control.
removing the double layout reduces the time to display from ~0.51s to ~0.45s
it's not a lot, but it isn't nothing either!
Good. Thank you. Issue closed, I think, with new version 5.0.1.
|
gharchive/issue
| 2020-01-23T16:58:06 |
2025-04-01T04:34:39.991873
|
{
"authors": [
"ConfusedVorlon",
"jdstmporter"
],
"repo": "jdstmporter/ColourWheel",
"url": "https://github.com/jdstmporter/ColourWheel/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1235724767
|
Error corriendo codigo docker run
Hola, me sale el siguiente error al correr el codigo docker run --rm -it -v "%PWD%":/workspace --name hadoop -p 8888:8888 -p 50070:50070 -p 8088:8088 jdvelasq/hadoop:2.8.5p
Unable to find image 'jdvelasq/hadoop:2.8.5p' locally
docker: Error response from daemon: manifest for jdvelasq/hadoop:2.8.5p not found: manifest unknown: manifest unknown.
Que debo hacer?
Este error parece porque Docker no está ejecutando o estás ejecutando los comandos como administrador (cuando Docker no lo requiere). si se sigue presentando el error, pega los pantallazos para que los validemos
|
gharchive/issue
| 2022-05-13T21:19:01 |
2025-04-01T04:34:39.994686
|
{
"authors": [
"MateoSDB",
"lvasquezve"
],
"repo": "jdvelasq/courses",
"url": "https://github.com/jdvelasq/courses/issues/27",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1380797163
|
fix: fish shell
This PR is replacing the here-document syntax with an install file due to fish shell not supporting the here-document syntax...
I've tested it on linux/fish and macOS.
resolves: #7
I changed it to just use bash -c can you test it does it still work on fish?
It does work with fish. You have a better solution!
|
gharchive/pull-request
| 2022-09-21T11:47:10 |
2025-04-01T04:34:40.001059
|
{
"authors": [
"daiyam",
"jeanp413"
],
"repo": "jeanp413/open-remote-ssh",
"url": "https://github.com/jeanp413/open-remote-ssh/pull/14",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1893394303
|
🛑 Distribuidora Navarrete is down
In e29044d, Distribuidora Navarrete (https://www.distribuidoranavarrete.com.pe) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Distribuidora Navarrete is back up in 9c73a6f after 10 minutes.
|
gharchive/issue
| 2023-09-12T22:27:24 |
2025-04-01T04:34:40.003458
|
{
"authors": [
"jeanp94"
],
"repo": "jeanp94/server",
"url": "https://github.com/jeanp94/server/issues/490",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
156084134
|
What is "Server returned weak ETag" ?
Couldn't read from client: Connection reset by peer
Couldn't read from client: Connection reset by peer
Couldn't read from client: Connection reset by peer
Discarded all objects, 2 chunks and 157 atoms left.
Couldn't read from client: Connection reset by peer
Server returned weak ETag -- ignored.
Server returned weak ETag -- ignored.
Server returned weak ETag -- ignored.
Unsupported Cache-Control directive stale-while-revalidate -- ignored.
Why polipo log output error above?
@Arnosir, please send questions to the mailing list. The tracker is for bug reports.
Couldn't read from client: Connection reset by peer
This means that the client closed the connection before it received all data we sent to it.
Discarded all objects, 2 chunks and 157 atoms left.
That means that after garbage-collecting all memory, Polipo was left with a small amount it couldn't free.
Server returned weak ETag -- ignored.
This means that the server sent a weak ETag, but Polipo doesn't support weak ETags. Polipo will not cache the page after it expires.
Unsupported Cache-Control directive stale-while-revalidate -- ignored.
This means that the server sent a cache control which Polipo doesn't understand. Polipo will ignore it.
|
gharchive/issue
| 2016-05-21T05:03:22 |
2025-04-01T04:34:40.039027
|
{
"authors": [
"Arnosir",
"jech"
],
"repo": "jech/polipo",
"url": "https://github.com/jech/polipo/issues/88",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1733629921
|
表单开发j-category-select placeholder不显示问题
版本号:
3.4.4
前端版本:vue3版?还是 vue2版?
vue3
问题描述:
使用表单开发时,弹框组件里的下拉框使用了分类字典,设置placeholder不显示的问题
截图&代码:
友情提示(为了提高issue处理效率):
未按格式要求发帖,会被直接删掉;
描述过于简单或模糊,导致无法处理的,会被直接删掉;
请自己初判问题描述是否清楚,是否方便我们调查处理;
针对问题请说明是Online在线功能(需说明用的主题模板),还是生成的代码功能;
sq
已修改,下版本发布
|
gharchive/issue
| 2023-05-31T08:13:06 |
2025-04-01T04:34:40.122145
|
{
"authors": [
"lsqGitHub716",
"wlqwcs",
"zhangdaiscott"
],
"repo": "jeecgboot/jeecg-boot",
"url": "https://github.com/jeecgboot/jeecg-boot/issues/5003",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2153560628
|
springboot3和springboot3_sas的区别是什么?
版本号:3.6.1
前端版本:vue3版
问题描述:区别在哪看?
截图&代码:
springboot3 shiro权限
springboot3_sas SAS权限(Spring Security)
|
gharchive/issue
| 2024-02-26T08:25:42 |
2025-04-01T04:34:40.123543
|
{
"authors": [
"t5752139",
"zhangdaiscott"
],
"repo": "jeecgboot/jeecg-boot",
"url": "https://github.com/jeecgboot/jeecg-boot/issues/5914",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2628488664
|
docs: Add third-party deployment tutorial, correct the spelling of "DNSPod"
What does this PR do?
Add third-party deployment tutorial, correct the spelling of "DNSPod"
Motivation
在国内有大部分用户都在使用宝塔面板管理服务器,因此增加使用宝塔面板部署的教程,可视化的部署方式可以帮助用户更加便捷的部署ddns-go
Dnspod正确拼写为DNSPod
Additional Notes
首页已经有说明了,没必要在写一个宝塔的
你好,可以的话可以加下我的企业微信,咱们沟通一下,我是宝塔面板官方项目入驻负责人
恸
@.***
------------------ 原始邮件 ------------------
发件人: @.>;
发送时间: 2024年11月1日(星期五) 下午3:29
收件人: @.>;
抄送: @.>; @.>;
主题: Re: [jeessy2/ddns-go] docs: Add third-party deployment tutorial, correct the spelling of "DNSPod" (PR #1305)
首页已经有说明了,没必要在写一个宝塔的
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you authored the thread.Message ID: @.***>
暂时不加。有什么可以沟通的可以email
|
gharchive/pull-request
| 2024-11-01T07:22:00 |
2025-04-01T04:34:40.137391
|
{
"authors": [
"bestlaw66",
"jeessy2"
],
"repo": "jeessy2/ddns-go",
"url": "https://github.com/jeessy2/ddns-go/pull/1305",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1023805852
|
当联合主键中有数据库AUTO_INCREMENT 字段时,无法自动填充
当联合主键中有数据库AUTO_INCREMENT 字段时,无法自动填充
经过测试,当联合主键中有数据库AUTO_INCREMENT 字段时,仍然可以通过@InsertFill自动填充。
|
gharchive/issue
| 2021-10-12T13:10:16 |
2025-04-01T04:34:40.155525
|
{
"authors": [
"anhuisunfei",
"jeffreyning"
],
"repo": "jeffreyning/mybatisplus-plus",
"url": "https://github.com/jeffreyning/mybatisplus-plus/issues/3",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1287686823
|
Typespec mismatch with HTTPoison.process_response/1
Running mix dialyzer returns a number of errors in Forcex.Api.Http and Forcex.Bulk around the override functions for process_response/1. HTTPoison changed the signature for that function several years ago, such that headers has been changed from a Map to a kwlist. This makes the function guards (used in both locations) no longer feasible.
These functions should be updated to work with headers as a keyword list.
Since dialyzer has a couple of distinct errors, I wanted to document them individually here. I think I have code that's sufficient to fix this error, so I can either extract it from my current branch as its own PR or include it in a big ball of fixups.
|
gharchive/issue
| 2022-06-28T17:46:52 |
2025-04-01T04:34:40.158931
|
{
"authors": [
"fastjames"
],
"repo": "jeffweiss/forcex",
"url": "https://github.com/jeffweiss/forcex/issues/52",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
896863077
|
closeOnRowBeginSwipe not working
Hey i'v got function component AppointmentsContainer that has SwipeListView.
let listRef = createRef();
<View>
<SwipeListView
keyExtractor={(item) => item.id.toString()}
data={appointments}
closeOnRowOpen
closeOnRowBeginSwipe
listViewRef={(ref) => (listRef = ref)}
onRowDidOpen={() => {
console.log("this doesn't fire");
}}
renderItem={(data, rowMap) => (
<Appointment
key={data.id}
appointment={data.item}
/>
)}
/>
</View>
Appointment.js:
const Appointment = React.forwardRef(
({ appointment }, listViewRef) => {
...
return (
<SwipeRow
leftOpenValue={listActionButtonWidth}
rightOpenValue={-(listActionButtonWidth * 2)}
stopLeftSwipe={listActionButtonWidth + 10}
stopRightSwipe={-(listActionButtonWidth * 2 + 10)}
friction={1000}
tension={500}
>
<View style={styles.rowBack}>
///Some action button
</View>
<View>
///some appointment info
</View>
</SwipeRow>
);
}
);
export default Appointment;
But when I open row the onRowDidOpen on SwipeListView doesn't fire
And when I begin swipe another row , the previuse row doesn't close
What am I doing wrong ?
99% of the time issues like this are related to improper use of keys or keyExtractor. Please double check that you're implementing one of those correctly with unique keys, thanks!
If someone is here because onRowDidOpen is not firing on SwipeRow then use it on SwipeListView instead as SwipeListView is overwriting the onRowDidOpen prop for the row.
I have provide key, keyExtractor and swipeKey for SwipeRow, but onRowDidOpen I got only swipe layout change, not key for close the row (
|
gharchive/issue
| 2021-05-20T13:50:29 |
2025-04-01T04:34:40.256922
|
{
"authors": [
"Eldar4Levi",
"code-by",
"ishan-cleartax",
"jemise111"
],
"repo": "jemise111/react-native-swipe-list-view",
"url": "https://github.com/jemise111/react-native-swipe-list-view/issues/553",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1390402608
|
JFrog plugin
Repository URL
https://github.com/jfrog/jfrog-plugin
New Repository Name
jfrog-plugin
Description
This plugin exposes the JFrog CLI as a tool. It installs the JFrog CLI and allows configuring the JFrog platform connection details.
GitHub users to have commit permission
@IL-Automation
@jfrog-ecosystem
Jenkins project users to have release permission
jfrog_ecosystem
Issue tracker
GitHub issues
Security audit, information and commands
The security team is auditing all the hosting requests, to ensure a better security by default.
This message informs you that the security team was notified about the request and will soon participate in this issue to assist.
The team is usually starting by a quick superficial audit and if it's not sufficient, they are planning a deeper audit.
Commands
Security team only:
/audit-ok => the audit is complete, the hosting can continue :tada:.
/audit-skip => the audit is not necessary, the hosting can continue :tada:.
/audit-required => the superficial audit was not sufficient, a deeper look is necessary :mag:.
/audit-findings => the audit reveals some issues that require corrections :pencil2:.
Anyone:
/audit-review => the findings from the audits were corrected, this command will ping the security team to review the findings :eyes:.
It's only applicable when the previous audit required changes.
Only one command can be requested per comment.
(automatically generated message)
/hosting re-check
/hosting re-check
/hosting re-check
/hosting re-check
I've updated your new repository name to remove jenkins as that's not allowed.
Jenkins CERT will have a deeper look. Tracked internally as JENSEC-1897.
/audit-required
@yahavi,
We've had a deeper look and have some feedback for you.
Security issues
Missing permission check in JFrogPlatformBuilder.java#L63.
Recommendation: https://www.jenkins.io/doc/developer/security/form-validation/#checking-permissions
Non-security issues (hence non-blocking for the hosting process)
Credentials.java#L25-L29 re-encrypts credentials every time configuration is saved. Normally to prevent that you'd stick to using Secret type for constructor arguments. I've checked usages of the Credentials class and I see it's a part of JFrogPlatformInstance.java and JFrogPlatformInstances are stored in JFrogPlatformBuilder.java#L39. Looking at jenkins.plugins.jfrog.configuration.JfrogPlatformBuilder.xml on disk I see that both actual encrypted credentials and credentialsId are stored for an instance:
<credentialsConfig>
<credentials>
<username>username</username>
<password>password</password>
<accessToken>accessToken</accessToken>
</credentials>
<credentialsId>jenkins-io</credentialsId>
</credentialsConfig>
I don't think you need both, storing only credentialsId should be enough. If you only store credentialsId and adjust your code to pull credentials from credential store when needed you can probably get rid of your Credentials class completely.
JfrogBuilder/global.jelly seems like a development leftover, it adds an empty section on Global Configuration page and does nothing else.
We'll let you fix the security issue and let us know when you're done. Non-security ones are up to you. In case of questions about the findings please feel free to ask.
/audit-findings
Thanks for looking into the plugin, @yaroslavafenkin!
We merged https://github.com/jfrog/jfrog-plugin/pull/7 to fix the security issue. We'll fix the other issues soon as well.
Could we please proceed now with the hosting process?
The fix looks alright to me.
I think https://github.com/jfrog/jfrog-plugin/pull/7/files#diff-74dd0585b30d5770a1e4047f48f035ae5beeaa33d81d425627150247c86dddceR92 is not needed though, doCheckPlatformUrl doesn't do anything special.
/audit-ok
Hey @yahavi,
the security audit passes, but I have a few minor nits I'd like to highlight:
https://github.com/jfrog/jfrog-plugin/blob/7c4da31a672243931591d57c5b979d19955962e1/pom.xml#L33 should be bom-2.346.x, given that's your baseline chosen.
https://github.com/jfrog/jfrog-plugin/blob/7c4da31a672243931591d57c5b979d19955962e1/pom.xml#L67-L69 instead of depending on commons-lang3 directly, you can depend on the corresponding Jenkins plugin: https://plugins.jenkins.io/commons-lang3-api/#dependencies
It ships the commons-lang3 api without the need to bundle it. The version is omitted, because it's shipped by the bom.
Your GAV coordinates picked don't align with the one you're actually using. I'd recommend you move your group id from jenkins.plugins to io.jenkins.plugins, as specified appropriately in https://github.com/jfrog/jfrog-plugin/blob/7c4da31a672243931591d57c5b979d19955962e1/pom.xml#L16, but referenced incorrectly in https://github.com/jfrog/jfrog-plugin/tree/main/src/main/java/jenkins/plugins/, package names and imports.
/hosting host
|
gharchive/issue
| 2022-09-29T07:39:10 |
2025-04-01T04:34:40.295783
|
{
"authors": [
"NotMyFault",
"eyalbe4",
"jenkins-cert-app",
"timja",
"yahavi",
"yaroslavafenkin"
],
"repo": "jenkins-infra/repository-permissions-updater",
"url": "https://github.com/jenkins-infra/repository-permissions-updater/issues/2826",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
706252158
|
Add typz as contributer to repo
Repo plugin
Submitter checklist for adding or changing permissions
@rsandell - needs to confirm
@Typz - me, to get permissions
Always
[x] Add link to plugin/component Git repository in description above
When adding new uploaders (this includes newly created permissions files)
[x] Make sure to @mention an existing maintainer to confirm the permissions request, if applicable
[x] Use the Jenkins community (LDAP) account name in the YAML file, not the GitHub account name
[x] Make sure to @mention the users being added so their GitHub account names are known if they require GitHub merge access (see below).
[x] All newly added users have logged in to Artifactory at least once
Reviewer checklist (not for requesters!)
[ ] Check this if newly added person also needs to be given merge permission to the GitHub repo (please @ the people/person with their GitHub username in this issue as well). If needed, it can be done using an IRC Bot command
[ ] Check that the $pluginId Developers team has Admin permissions while granting the access.
[ ] In the case of plugin adoption, ensure that the Jenkins Jira default assignee is either removed or changed to the new maintainer.
[ ] If security contacts are changed (this includes add/remove), ping the security officer (currently @daniel-beck) in this pull request. If an email contact is changed, wait for approval from the security officer.
There are IRC Bot commands for it
rsandell approved in mailing list
|
gharchive/pull-request
| 2020-09-22T10:17:10 |
2025-04-01T04:34:40.302005
|
{
"authors": [
"Typz",
"timja"
],
"repo": "jenkins-infra/repository-permissions-updater",
"url": "https://github.com/jenkins-infra/repository-permissions-updater/pull/1672",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
530907786
|
chore: bdd-spring-1575265719 to 0.0.1
chore: Promote bdd-spring-1575265719 to version 0.0.1
[APPROVALNOTIFIER] This PR is NOT APPROVED
This pull-request has been approved by:
To fully approve this pull request, please assign additional approvers.
We suggest the following additional approvers:
If they are not already assigned, you can assign the PR to them by writing /assign in a comment when ready.
The full list of commands accepted by this bot can be found here.
The pull request process is described here
Needs approval from an approver in each of these files:
OWNERS
Approvers can indicate their approval by writing /approve in a comment
Approvers can cancel approval by writing /approve cancel in a comment
|
gharchive/pull-request
| 2019-12-02T05:53:07 |
2025-04-01T04:34:40.323003
|
{
"authors": [
"jenkins-x-bot-test"
],
"repo": "jenkins-x-bot-test/environment-pr-486-18arc-staging",
"url": "https://github.com/jenkins-x-bot-test/environment-pr-486-18arc-staging/pull/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
562808069
|
chore: bdd-spring-1581366163 to 0.0.1
chore: Promote bdd-spring-1581366163 to version 0.0.1
[APPROVALNOTIFIER] This PR is NOT APPROVED
This pull-request has been approved by:
To fully approve this pull request, please assign additional approvers.
We suggest the following additional approvers:
If they are not already assigned, you can assign the PR to them by writing /assign in a comment when ready.
The full list of commands accepted by this bot can be found here.
The pull request process is described here
Needs approval from an approver in each of these files:
OWNERS
Approvers can indicate their approval by writing /approve in a comment
Approvers can cancel approval by writing /approve cancel in a comment
|
gharchive/pull-request
| 2020-02-10T20:27:25 |
2025-04-01T04:34:40.326626
|
{
"authors": [
"jenkins-x-bot-test"
],
"repo": "jenkins-x-bot-test/environment-pr-710-2arc-staging",
"url": "https://github.com/jenkins-x-bot-test/environment-pr-710-2arc-staging/pull/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
524904076
|
chore: bdd-spring-1574156727 to 0.0.1
chore: Promote bdd-spring-1574156727 to version 0.0.1
[APPROVALNOTIFIER] This PR is NOT APPROVED
This pull-request has been approved by:
To fully approve this pull request, please assign additional approvers.
We suggest the following additional approvers:
If they are not already assigned, you can assign the PR to them by writing /assign in a comment when ready.
The full list of commands accepted by this bot can be found here.
The pull request process is described here
Needs approval from an approver in each of these files:
OWNERS
Approvers can indicate their approval by writing /approve in a comment
Approvers can cancel approval by writing /approve cancel in a comment
|
gharchive/pull-request
| 2019-11-19T10:07:21 |
2025-04-01T04:34:40.330249
|
{
"authors": [
"jenkins-x-bot-test"
],
"repo": "jenkins-x-bot-test/environment-vs-pr-133-1arc-production",
"url": "https://github.com/jenkins-x-bot-test/environment-vs-pr-133-1arc-production/pull/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
535229480
|
Delete application bdd-spring-1575920231 from this environment
The command jx delete application was run by and it generated this Pull Request
[APPROVALNOTIFIER] This PR is NOT APPROVED
This pull-request has been approved by:
To fully approve this pull request, please assign additional approvers.
We suggest the following additional approvers:
If they are not already assigned, you can assign the PR to them by writing /assign in a comment when ready.
The full list of commands accepted by this bot can be found here.
The pull request process is described here
Needs approval from an approver in each of these files:
OWNERS
Approvers can indicate their approval by writing /approve in a comment
Approvers can cancel approval by writing /approve cancel in a comment
|
gharchive/pull-request
| 2019-12-09T19:55:02 |
2025-04-01T04:34:40.334040
|
{
"authors": [
"jenkins-x-bot-test"
],
"repo": "jenkins-x-bot-test/environment-vs-pr-200-63arc-staging",
"url": "https://github.com/jenkins-x-bot-test/environment-vs-pr-200-63arc-staging/pull/2",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
548149856
|
fix: turn on cert manager v1alpha2 by default
Getting the certs.newApi value overrided was...hard. So let's fix
this properly and just accept that anyone who installs jx-app-sso at
this point needs to be on a current cert manager setup.
Signed-off-by: Andrew Bayer andrew.bayer@gmail.com
/lgtm
|
gharchive/pull-request
| 2020-01-10T15:44:09 |
2025-04-01T04:34:40.335546
|
{
"authors": [
"abayer",
"pmuir"
],
"repo": "jenkins-x/dex",
"url": "https://github.com/jenkins-x/dex/pull/25",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
812563790
|
fix: upgrade to version 3.1.266
from: https://github.com/jenkins-x/jx-cli
[APPROVALNOTIFIER] This PR is NOT APPROVED
This pull-request has been approved by:
To complete the pull request process, please assign macox
You can assign the PR to them by writing /assign @macox in a comment when ready.
The full list of commands accepted by this bot can be found here.
Needs approval from an approver in each of these files:
OWNERS
Approvers can indicate their approval by writing /approve in a comment
Approvers can cancel approval by writing /approve cancel in a comment
@jenkins-x-bot-test: The following test failed, say /retest to rerun them all:
Test name
Commit
Details
Rerun command
kube
4251b4d8e701048d38f5194abef44584ce9e1331
link
/test kube
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the jenkins-x/lighthouse repository. I understand the commands that are listed here.
/retest
@jenkins-x-bot-test: The following test failed, say /retest to rerun them all:
Test name
Commit
Details
Rerun command
kube
4251b4d8e701048d38f5194abef44584ce9e1331
link
/test kube
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the jenkins-x/lighthouse repository. I understand the commands that are listed here.
@jenkins-x-bot-test: The following tests failed, say /retest to rerun them all:
Test name
Commit
Details
Rerun command
kube
4251b4d8e701048d38f5194abef44584ce9e1331
link
/test kube
gsm
4251b4d8e701048d38f5194abef44584ce9e1331
link
/test gsm
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the jenkins-x/lighthouse repository. I understand the commands that are listed here.
@jenkins-x-bot-test: The following tests failed, say /retest to rerun them all:
Test name
Commit
Details
Rerun command
kube
4251b4d8e701048d38f5194abef44584ce9e1331
link
/test kube
gsm
4251b4d8e701048d38f5194abef44584ce9e1331
link
/test gsm
vault
4251b4d8e701048d38f5194abef44584ce9e1331
link
/test vault
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the jenkins-x/lighthouse repository. I understand the commands that are listed here.
@jenkins-x-bot-test: The following tests failed, say /retest to rerun them all:
Test name
Commit
Details
Rerun command
kube
4251b4d8e701048d38f5194abef44584ce9e1331
link
/test kube
gsm
4251b4d8e701048d38f5194abef44584ce9e1331
link
/test gsm
vault
4251b4d8e701048d38f5194abef44584ce9e1331
link
/test vault
tls
4251b4d8e701048d38f5194abef44584ce9e1331
link
/test tls
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the jenkins-x/lighthouse repository. I understand the commands that are listed here.
@jenkins-x-bot-test: The following tests failed, say /retest to rerun them all:
Test name
Commit
Details
Rerun command
kube
4251b4d8e701048d38f5194abef44584ce9e1331
link
/test kube
gsm
4251b4d8e701048d38f5194abef44584ce9e1331
link
/test gsm
vault
4251b4d8e701048d38f5194abef44584ce9e1331
link
/test vault
tls
4251b4d8e701048d38f5194abef44584ce9e1331
link
/test tls
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the jenkins-x/lighthouse repository. I understand the commands that are listed here.
@jenkins-x-bot-test: The following tests failed, say /retest to rerun them all:
Test name
Commit
Details
Rerun command
tls
4251b4d8e701048d38f5194abef44584ce9e1331
link
/test tls
kube
c26a7e5074d088724c27e7fc31deba691244adda
link
/test kube
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the jenkins-x/lighthouse repository. I understand the commands that are listed here.
@jenkins-x-bot-test: The following tests failed, say /retest to rerun them all:
Test name
Commit
Details
Rerun command
kube
c26a7e5074d088724c27e7fc31deba691244adda
link
/test kube
tls
c26a7e5074d088724c27e7fc31deba691244adda
link
/test tls
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the jenkins-x/lighthouse repository. I understand the commands that are listed here.
@jenkins-x-bot-test: The following tests failed, say /retest to rerun them all:
Test name
Commit
Details
Rerun command
kube
c26a7e5074d088724c27e7fc31deba691244adda
link
/test kube
tls
c26a7e5074d088724c27e7fc31deba691244adda
link
/test tls
gsm
c26a7e5074d088724c27e7fc31deba691244adda
link
/test gsm
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the jenkins-x/lighthouse repository. I understand the commands that are listed here.
@jenkins-x-bot-test: The following tests failed, say /retest to rerun them all:
Test name
Commit
Details
Rerun command
kube
c26a7e5074d088724c27e7fc31deba691244adda
link
/test kube
tls
c26a7e5074d088724c27e7fc31deba691244adda
link
/test tls
gsm
c26a7e5074d088724c27e7fc31deba691244adda
link
/test gsm
vault
c26a7e5074d088724c27e7fc31deba691244adda
link
/test vault
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the jenkins-x/lighthouse repository. I understand the commands that are listed here.
@jenkins-x-bot-test: The following tests failed, say /retest to rerun them all:
Test name
Commit
Details
Rerun command
kube
c26a7e5074d088724c27e7fc31deba691244adda
link
/test kube
gsm
c26a7e5074d088724c27e7fc31deba691244adda
link
/test gsm
vault
c26a7e5074d088724c27e7fc31deba691244adda
link
/test vault
tls
c26a7e5074d088724c27e7fc31deba691244adda
link
/test tls
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the jenkins-x/lighthouse repository. I understand the commands that are listed here.
|
gharchive/pull-request
| 2021-02-20T09:32:00 |
2025-04-01T04:34:40.385577
|
{
"authors": [
"jenkins-x-bot-test",
"jstrachan"
],
"repo": "jenkins-x/jx3-versions",
"url": "https://github.com/jenkins-x/jx3-versions/pull/2162",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
677359716
|
Map unstable build to check failure
Fixes https://github.com/jenkinsci/checks-api-plugin/issues/11
I updated the description to include ‘fixes’ which means the issue will automatically be closed on merge
|
gharchive/pull-request
| 2020-08-12T03:46:27 |
2025-04-01T04:34:40.396832
|
{
"authors": [
"XiongKezhi",
"timja"
],
"repo": "jenkinsci/checks-api-plugin",
"url": "https://github.com/jenkinsci/checks-api-plugin/pull/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1309619874
|
Tag logs with timestamp, datadog.product and ci.pipeline.name
What does this PR do?
Adds the timestamp, datadog.product and ci.pipeline.name tags to logs, matching logs produced by other CIs.
I changed the way we read the pipeline name so it doesn't require git information, since that's not available when we create the logs DatadogWriter object.
Verification Process
After sending logs with this patch, the new fields should show up here:
Review checklist (to be filled by reviewers)
[ ] Feature or bug fix MUST have appropriate tests (unit, integration, etc...)
[ ] PR title must be written as a CHANGELOG entry (see why)
[ ] Files changes must correspond to the primary purpose of the PR as described in the title (small unrelated changes should have their own PR)
[ ] PR must have one changelog/ label attached. If applicable it should have the backward-incompatible label attached.
[ ] PR should not have do-not-merge/ label attached.
[ ] If Applicable, issue must have kind/ and severity/ labels attached at least.
We should look into it but I think this was already broken before this PR. If not feel free to revert.
yes it was already wrong 👍
|
gharchive/pull-request
| 2022-07-19T14:36:15 |
2025-04-01T04:34:40.410269
|
{
"authors": [
"AdrianLC",
"albertvaka"
],
"repo": "jenkinsci/datadog-plugin",
"url": "https://github.com/jenkinsci/datadog-plugin/pull/297",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
234227135
|
Fix #508 do not use latest alpine parent image
It breaks startup with UnsatisfiedLinkError
https://bugs.alpinelinux.org/issues/7372
@reviewbybees
This pull request originates from a CloudBees employee. At CloudBees, we require that all pull requests be reviewed by other CloudBees employees before we seek to have the change accepted. If you want to learn more about our process please see this explanation.
This is https://issues.jenkins-ci.org/browse/JENKINS-44777.
Will we be at risk of Java security issues if we stay on the already-old version?
Sorry, I should have checked duplicates first
https://issues.jenkins-ci.org/browse/JENKINS-44733 too.
The latest one doesn't work so how does it matter if it is more "secure"
We will need to update the FROM once it is fixed
|
gharchive/pull-request
| 2017-06-07T14:23:17 |
2025-04-01T04:34:40.416039
|
{
"authors": [
"carlossg",
"oleg-nenashev",
"reviewbybees"
],
"repo": "jenkinsci/docker",
"url": "https://github.com/jenkinsci/docker/pull/510",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
134061407
|
jenkins::cli::exec notify Class['jenkins::cli::reload'] is not always necessary and causes issues with running builds and pipelines
Here are the behaviors I have observed:
Creating 2 new build jobs one dependent on the other by "trigger parameterized build on other projects" or "Build after other projects are built". This results in the relationship being observed here:
Then I trigger a build which is visible from the pipeline view:
Then I manually invoked a puppet run:
[root@jenkins ~]# puppet agent --test
Info: Using configured environment 'test'
Info: Retrieving pluginfacts
Info: Retrieving plugin
Info: Loading facts
Info: Caching catalog for jenkins
Info: Applying configuration version 'af9c2adf1a6c5abe1ffaadc372d39bf147077ddb'
Notice: /Stage[main]/Jenkins_ci/Jenkins::User[steve]/Jenkins::Cli::Exec[create-jenkins-user-steve]/Exec[create-jenkins-user-steve]/returns: executed successfully
Info: /Stage[main]/Jenkins_ci/Jenkins::User[steve]/Jenkins::Cli::Exec[create-jenkins-user-steve]/Exec[create-jenkins-user-steve]: Scheduling refresh of Class[Jenkins::Cli::Reload]
Info: Class[Jenkins::Cli::Reload]: Scheduling refresh of Exec[reload-jenkins]
Notice: /Stage[main]/Jenkins::Cli::Reload/Exec[reload-jenkins]: Triggered 'refresh' from 1 events
Notice: Applied catalog in 14.90 seconds
Now reloading the pipeline view the second job is now missing:
I replicated this same behavior from within jenkins without running puppet but instead clicking "reload configuration from disk" I can conclude this is an issue with jenkins directly, however it is a constantly invoked because of the jenkins::user not being idempotent https://github.com/jenkinsci/puppet-jenkins/issues/454 and jenkins::cli::exec being very heavy handed with its reload configurations from disk. While true some functions do require reloading from disk, most of the Jenkins_CLI operations are performed without reloading being necessary, adding / modifying a user via jenkins::user being one of them. I have also seen this disrupt additional build steps such as hipchat notify complete within Jenkins.
Resolution to this issue in my opinion would be to parameterize the notify on the exec (and subsequently update each .pp as needed to invoke the reload with the notify) and to provide for a "safe" reload when no jobs are running.
Lastly reload of the jenkins service should be confined to when there are no build jobs running.
I have created a bug with Jenkins to track the problem with reload configuration from disk breaking running build pipelines: https://issues.jenkins-ci.org/browse/JENKINS-32984
@james-powis It might be worth taking a look at the native types we are hoping will replace the implementation of the defined types in the future: the https://github.com/jenkinsci/puppet-jenkins#experimental-types-and-providers
|
gharchive/issue
| 2016-02-16T18:37:52 |
2025-04-01T04:34:40.500112
|
{
"authors": [
"james-powis",
"jhoblitt"
],
"repo": "jenkinsci/puppet-jenkins",
"url": "https://github.com/jenkinsci/puppet-jenkins/issues/501",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
144122942
|
make jenkins_job type cloudbees-folder aware
If the resource name contains slashes, treat them as a
cloudbees-folder path and autorequire all parent folders.
Eg., foo/bar/myjob would autorequire the jenkins_job resources named
foo and foo/bar.
relates to #472
Merge when travis goes green, LGTM
@rtyler Great!
|
gharchive/pull-request
| 2016-03-29T00:52:02 |
2025-04-01T04:34:40.502242
|
{
"authors": [
"jhoblitt",
"rtyler"
],
"repo": "jenkinsci/puppet-jenkins",
"url": "https://github.com/jenkinsci/puppet-jenkins/pull/540",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1592556005
|
🛑 Wanted Gigs is down
In 723ddaa, Wanted Gigs (https://www.wanted.co.kr/gigs) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Wanted Gigs is back up in 2128e3a.
|
gharchive/issue
| 2023-02-20T23:17:21 |
2025-04-01T04:34:40.529007
|
{
"authors": [
"jeongsk"
],
"repo": "jeongsk/upptime",
"url": "https://github.com/jeongsk/upptime/issues/287",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
119545015
|
Support ES6 default named exports for Backbone-ES6
This subtle change takes into account how ES6 modules export default named modules and how they are consumed in the UMD loading code of Backbone.localStorage. backbone-es6 (https://github.com/typhonjs/backbone-es6) is a fork of Backbone using ES6. The ES6 module output export default backbone; found in ModuleRuntime.js (https://github.com/typhonjs/backbone-es6/blob/master/src/ModuleRuntime.js#L39) puts the actual reference under a default named key in the ES5 UMD loading code.
A complete TODOs demo using Backbone-ES6 and Backbone.localStorage can be found here:
http://js.demos.typhonrt.org/backbone-es6-localstorage-todos/
and repo of the demo:
https://github.com/typhonjs-demos/backbone-es6-localstorage-todos
The change in this PR adds a check for the existence of Backbone.default in the resolution of calling factory in the Backbone.localStorage UMD wrapper.
Hi @typhonrt, thanks for this, however I'm going to close it in favour of #207
Oh no worries.. Sounds good. At the time I didn't know what I do now. Back then I just rewrote local storage in ES6; I see #207 does this... Now I just use babel-plugin-add-module-exports with my ES6 modules especially everything released on NPM.
|
gharchive/pull-request
| 2015-11-30T17:53:21 |
2025-04-01T04:34:40.546511
|
{
"authors": [
"scott-w",
"typhonrt"
],
"repo": "jeromegn/Backbone.localStorage",
"url": "https://github.com/jeromegn/Backbone.localStorage/pull/194",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2353237448
|
Default to terser instead of uglifyjs
I've just read that uglify-js is actually no longer maintained. However, terser is, and it's recommended by elm-optimize-level-2 as a replacement. Maybe mkElmDerivation should use terser, then? Usage seems exactly the same as uglifyjs.
The reason I used uglify-js was because that is was that is what Evan uses in the minification guide.
I have actually updated the way mkElmDerivation works by default to make it less opinionated. Essentially, all this does now is fetch the elm dependencies and puts them in the right place. It's then up to the user to run whatever elm commands they want to.
I am happy to update to terser if you think that is a good idea, but it will be updating legacy functionality of this repository. Would you benefit from that?
Ahhh, I get what you're saying. It seems I'm still using an old version of mkElmDerivation that is more opinionated.
I will update mkElmDerivation then and get back to you.
|
gharchive/issue
| 2024-06-14T12:10:52 |
2025-04-01T04:34:40.557522
|
{
"authors": [
"jeslie0",
"pmiddend"
],
"repo": "jeslie0/mkElmDerivation",
"url": "https://github.com/jeslie0/mkElmDerivation/issues/13",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2590342197
|
spago.lock format changed to JSON
https://github.com/jeslie0/mkSpagoDerivation/blob/beeac0abd28514fc5b4c55d5e5e8e5d75adf7eaf/nix/buildFromLockFile.nix#L5
https://github.com/purescript/spago/pull/1280
Fabrizio said on the Discord today that “the format for the spago.lock is likely to change one more time before we call it stable.”
This looks like a really nice project @jeslie0 , good work.
Okay, I am happy to say that this has been done. I decided against keeping support for the YAML based lock files as Spago seems to change all lock files to JSON whenever it is invoked now.
Sweet, thanks man.
|
gharchive/issue
| 2024-10-16T01:30:32 |
2025-04-01T04:34:40.559875
|
{
"authors": [
"jamesdbrock",
"jeslie0"
],
"repo": "jeslie0/mkSpagoDerivation",
"url": "https://github.com/jeslie0/mkSpagoDerivation/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
464949668
|
High CPU usage by Docker for Mac when lazydocker is running idle
Describe the bug
When running lazydocker on my Mac (latest version), while no containers are running, Docker for Mac uses ~25-30% cpu.
To Reproduce
Steps to reproduce the behavior:
Start lazydocker
Open Activity Monitor in CPU tab
Find process com.docker.hyperkit
See CPU usage
Expected behaviour
CPU usage low (without lazydocker it's ~2%). Maybe small impact.
Screenshots
NA
Desktop (please complete the following information):
OS: Mac OS (latest as of July 7th 2019).
Lazydocker Version Version: commit=a536c09dab90fc9a4f4931285466d6126c8bc121, build date=2019-07-05T10:59:26Z, build source=binaryRelease, version=0.5.4, os=darwin, arch=amd64
The last commit id if you built project from sources (run : git rev-parse HEAD)
Additional context
I assume the load is due to continuous querying of the Docker API. Might be a good idea to allow setting the polling interval so I can choose to config a lower polling rate and reduce the load on Docker.
Update: upgrading to 0.5.5 improved a bit. Still CPU is > 20%
I'm not a mac user but this seems to be a common issue with docker on mac:
https://github.com/docker/for-mac/issues/3499
https://github.com/docker/for-mac/issues/1759
https://github.com/docker/for-mac/issues/2582
There are a few suggested solutions in those issues you could try,
From what i know docker seems to have a performance issue on macos with the file system. That will either result in really slow read/write speed or have high cpu usage when doing anything related to the file system.
The high CPU happens only when I start lazydocker. Without it, my Docker seem to be clam.
Yes that is probably because lazydocker gets information about the docker containers and volumes every 0.1 seconds for holding the UI up to date.
Makes sense. It would be great to be able to slow it down by config so to accept slower UI updates, but less load on the Docker daemon...
Agree we could add that to the lazydocker config for people with these kinds of problems.
I'll see if i can create a PR for your issue.
@eldada I've put up a pr (#127) to allow for a user set pull duration.
Awesome! Thanks for the work 😄
I just had this happen to me, I logged into my server this morning and the load avg was 95 and climbing, and when I was finally able to get 'top' to run, lazydocker had consumed well over 12GB of ram, as well as 118% of CPU. So perhaps there's a memory leak somewhere, not just the pull duration?
had consumed well over 12GB of ram
That doesn't sound good, The pr is still open so 0.2.4 won't fix your issue.
I'll see if i can get the pr merged in and you can take a look if the issue is fixed then.
|
gharchive/issue
| 2019-07-07T12:52:30 |
2025-04-01T04:34:40.572224
|
{
"authors": [
"eldada",
"goodwid",
"mjarkk"
],
"repo": "jesseduffield/lazydocker",
"url": "https://github.com/jesseduffield/lazydocker/issues/115",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1535270690
|
Add "networks" panel under "volumes"
Just a little improvement with adding "Networks" panel under "Volumes" one
Thanks for making this! Sorry I've been super busy lately but I'll give this a review when I get the chance
Tested locally and it works like a charm.
Looks like you need to regenerate cheatsheets though (using go run scripts/cheatsheet/main.go generate)
Thanks for making this @Tony-Sol !
Should we be seeing this in the brew tap? I don't see it currently.
|
gharchive/pull-request
| 2023-01-16T17:21:19 |
2025-04-01T04:34:40.574663
|
{
"authors": [
"Tony-Sol",
"jesseduffield",
"jessejoe"
],
"repo": "jesseduffield/lazydocker",
"url": "https://github.com/jesseduffield/lazydocker/pull/424",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
192295632
|
2.0.5 fails to install without bower
Hello,
2.0.5 fails to install if system or project doesn't have Bower installed. Should it be listed as dependency (it will be weird)?
> card@2.0.5 postinstall C:\Users\konst\Sites\transfers.do\node_modules\card
> bower install
'bower' is not recognized as an internal or external command,
operable program or batch file.
transfers.do@1.5.26 C:\Users\konst\Sites\transfers.do
`-- eslint-plugin-more@0.1.1 (git://github.com/webbylab/eslint-plugin-more.git#7f9d5ef1b4632cff36b44ebabe4660f7fad96974)
npm WARN optional SKIPPING OPTIONAL DEPENDENCY: fsevents@^1.0.0 (node_modules\chokidar\node_modules\fsevents):
npm WARN notsup SKIPPING OPTIONAL DEPENDENCY: Unsupported platform for fsevents@1.0.15: wanted {"os":"darwin","arch":"any"} (current: {"os":"win32","arch":"x64"})
npm ERR! Windows_NT 10.0.14971
npm ERR! argv "C:\\Program Files\\nodejs\\node.exe" "C:\\Users\\konst\\AppData\\Roaming\\npm\\node_modules\\npm\\bin\\npm-cli.js" "update"
npm ERR! node v7.2.0
npm ERR! npm v4.0.3
npm ERR! code ELIFECYCLE
npm ERR! card@2.0.5 postinstall: `bower install`
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the card@2.0.5 postinstall script 'bower install'.
npm ERR! Make sure you have the latest version of node.js and npm installed.
npm ERR! If you do, this is most likely a problem with the card package,
npm ERR! not with npm itself.
npm ERR! Tell the author that this fails on your system:
npm ERR! bower install
npm ERR! You can get information on how to open an issue for this project with:
npm ERR! npm bugs card
npm ERR! Or if that isn't available, you can get their info via:
npm ERR! npm owner ls card
npm ERR! There is likely additional logging output above.
npm ERR! Please include the following file with any support request:
npm ERR! C:\Users\konst\Sites\transfers.do\npm-debug.log
Thanks for the heads up! Fixed in 2.0.6.
|
gharchive/issue
| 2016-11-29T14:27:12 |
2025-04-01T04:34:40.581177
|
{
"authors": [
"jessepollak",
"tinovyatkin"
],
"repo": "jessepollak/card",
"url": "https://github.com/jessepollak/card/issues/298",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
168181209
|
Setting UICollectionViewCell.isAccessibilityElement to YES hides cell contents from UI Testing
New issue checklist
[x] I have read all of the README, documentation, and FAQ.
[x] I have reviewed the contributing guidelines. Confirmation: 💪😎👊
[x] I have searched existing issues and this is not a duplicate.
General information
Library version(s): 7.3.2
iOS version(s): 8.1, 9
Devices/Simulators affected: iPhone 6s
Reproducible in the demo project? (Yes/No): Yes
Related issues: https://github.com/jessesquires/JSQMessagesViewController/issues/298
Bug report
Hi! We're trying to do UI testing on our app, but we cannot access the contents of cells from the XCUIApplication querying methods. It seems to be related to collectionView:accessibilityForCell:indexPath:message: setting the isAccessibilityElement to YES
Expected behavior
XCUIQueryElements of cells can see contents of cells
Log, from a modified version where I commented out cell.isAccessibilityElement = YES; unfortunately I do not know the repercussions of doing this.
Find: Target Application 0x7fbb72c51760
Output: {
Application 0x7fbb71503240: {{0.0, 0.0}, {414.0, 736.0}}, label: 'JSQMessages'
}
↪︎Find: Descendants matching type Cell
Output: {
Cell 0x7fbb7153e630: traits: 146028888064, {{4.0, -99.0}, {406.0, 79.0}}, label: 'Jesse Squires: Welcome to JSQMessages: A messaging UI framework for iOS.'
Cell 0x7fbb71543c60: traits: 146028888064, {{4.0, -16.0}, {406.0, 123.0}}, label: 'Steve Wozniak: It is simple, elegant, and easy to use. There are super sweet default settings, but you can customize like crazy.'
Cell 0x7fbb71548ea0: traits: 146028888064, {{4.0, 111.0}, {406.0, 103.0}}, label: 'Jesse Squires: It even has data detectors. You can call me tonight. My cell number is 123-456-7890. My website is www.hexedbits.com.'
Cell 0x7fbb7154d960: traits: 146028888064, {{4.0, 218.0}, {406.0, 121.0}}, label: 'Jobs: JSQMessagesViewController is nearly an exact replica of the iOS Messages App. And perhaps, better.'
Cell 0x7fbb715523c0: traits: 146028888064, {{4.0, 343.0}, {406.0, 79.0}}, label: 'Tim Cook: It is unit-tested, free, open-source, and documented.'
t = 9.61s Find: Descendants matching type TextView
Cell 0x7fbb71556e20: traits: 146028888064, {{4.0, 426.0}, {406.0, 38.0}}, label: 'Jesse Squires: Now with media messages!'
Cell 0x7fbb7155b920: traits: 146028888064, {{4.0, 468.0}, {406.0, 170.0}}, label: 'Jesse Squires: media message'
Cell 0x7fbb7147b680: traits: 146028888064, {{4.0, 642.0}, {406.0, 40.0}}, label: 'Jesse Squires: media message'
}
↪︎Find: Descendants matching type TextView
Output: {
TextView 0x7fbb71541e50: traits: 140746078289984, {{88.0, -79.0}, {284.0, 59.0}}, value: Welcome to JSQMess...
TextView 0x7fbb715468d0: traits: 140746078289984, {{42.0, 4.0}, {317.0, 103.0}}, value: It is simple, eleg...
TextView 0x7fbb7154bae0: traits: 140746078289984, {{67.0, 111.0}, {305.0, 103.0}}, value: It even has data d...
TextView 0x7fbb71550580: traits: 140746078289984, {{42.0, 258.0}, {322.0, 81.0}}, value: JSQMessagesViewCon...
TextView 0x7fbb71554fc0: traits: 140746078289984, {{42.0, 363.0}, {303.0, 59.0}}, value: It is unit-tested,...
TextView 0x7fbb71559aa0: traits: 140746078289984, {{131.0, 426.0}, {241.0, 38.0}}, value: Now with media mes...
}
Actual behavior
XCUIQueryElements of cells show no child contents
from the demo project:
Find: Target Application 0x7ff19c08cd60
Output: {
Application 0x7ff19c4007e0: {{0.0, 0.0}, {414.0, 736.0}}, label: 'JSQMessages'
}
↪︎Find: Descendants matching type Cell
Output: {
Cell 0x7ff19c41eee0: traits: 146028888064, {{4.0, -99.0}, {406.0, 79.0}}, label: 'Jesse Squires: Welcome to JSQMessages: A messaging UI framework for iOS.'
Cell 0x7ff19c41f5d0: traits: 146028888064, {{4.0, -16.0}, {406.0, 123.0}}, label: 'Steve Wozniak: It is simple, elegant, and easy to use. There are super sweet default settings, but you can customize like crazy.'
Cell 0x7ff19c420380: traits: 146028888064, {{4.0, 111.0}, {406.0, 103.0}}, label: 'Jesse Squires: It even has data detectors. You can call me tonight. My cell number is 123-456-7890. My website is www.hexedbits.com.'
Cell 0x7ff19c420ae0: traits: 146028888064, {{4.0, 218.0}, {406.0, 121.0}}, label: 'Jobs: JSQMessagesViewController is nearly an exact replica of the iOS Messages App. And perhaps, better.'
Cell 0x7ff19c4211f0: traits: 146028888064, {{4.0, 343.0}, {406.0, 79.0}}, label: 'Tim Cook: It is unit-tested, free, open-source, and documented.'
Cell 0x7ff19c421930: traits: 146028888064, {{4.0, 426.0}, {406.0, 38.0}}, label: 'Jesse Squires: Now with media messages!'
Cell 0x7ff19c421fd0: traits: 146028888064, {{4.0, 468.0}, {406.0, 170.0}}, label: 'Jesse Squires: media message'
Cell 0x7ff19c422690: traits: 146028888064, {{4.0, 642.0}, {406.0, 40.0}}, label: 'Jesse Squires: media message'
}
↪︎Find: Descendants matching type TextView
Steps to reproduce
Break executing during a UI test with a JSQ message view listed
In the console, execute po XCUIApplication().cells.textViews
observe that only the cells are queryable, not the contents
Crash log? Screenshots? Videos? Sample project?
Related stack overflow: http://stackoverflow.com/questions/14821713/ios-ui-automation-element-finds-no-sub-elements/22394083#22394083
Thanks so much for the great report @mwgray ! 😄
@jessesquires I think we will need to add accessibilityIdentifier and accessibilityLabel to make it visible. You can assign to me actually. 😺
I have dug a little bit deeper and found out that we probably can not consider this as a feature. Here is what I found out after assigning JSQMessagesCollectionViewCell and the textView property each an accessbilityIdentifier.
(lldb) po jsqmessagescollectionviewcellindexpathrow2Cell.textViews
t = 44.22s Snapshot accessibility hierarchy for com.hexedbits.JSQMessages
t = 44.49s Find: Descendants matching type CollectionView
t = 44.49s Find: Descendants matching type Cell
Find: Target Application 0x7fbe50737d20
Output: {
Application 0x7fbe5052b6f0: {{0.0, 0.0}, {414.0, 736.0}}, label: 'JSQMessages'
}
↪︎Find: Descendants matching type CollectionView
Output: {
CollectionView 0x7fbe506ddd30: traits: 35192962023424, {{0.0, 0.0}, {414.0, 736.0}}
}
↪︎Find: Descendants matching type Cell
Output: {
Cell 0x7fbe506f0950: traits: 146028888064, {{4.0, -62.0}, {406.0, 77.0}}, identifier: 'JSQMessagesCollectionViewCellIndexPathRow0', label: 'Jesse Squires: Welcome to JSQMessages: A messaging UI framework for iOS.'
Cell 0x7fbe506ee0e0: traits: 146028888064, {{4.0, 19.0}, {406.0, 98.0}}, identifier: 'JSQMessagesCollectionViewCellIndexPathRow1', label: 'Steve Wozniak: It is simple, elegant, and easy to use. There are super sweet default settings, but you can customize like crazy.'
Cell 0x7fbe506b4700: traits: 146028888072, {{4.0, 121.0}, {406.0, 99.0}}, identifier: 'JSQMessagesCollectionViewCellIndexPathRow2', label: 'Jesse Squires: It even has data detectors. You can call me tonight. My cell number is 123-456-7890. My website is www.hexedbits.com.'
t = 44.49s Find: Elements matching predicate '"JSQMessagesCollectionViewCellIndexPathRow2" IN identifiers'
t = 44.49s Find: Descendants matching type TextView
Cell 0x7fbe506d9c90: traits: 146028888064, {{4.0, 224.0}, {406.0, 118.0}}, identifier: 'JSQMessagesCollectionViewCellIndexPathRow3', label: 'Jobs: JSQMessagesViewController is nearly an exact replica of the iOS Messages App. And perhaps, better.'
Cell 0x7fbe506d7d30: traits: 146028888064, {{4.0, 346.0}, {406.0, 77.0}}, identifier: 'JSQMessagesCollectionViewCellIndexPathRow4', label: 'Tim Cook: It is unit-tested, free, open-source, and documented.'
Cell 0x7fbe5054cc40: traits: 146028888064, {{4.0, 427.0}, {406.0, 37.0}}, identifier: 'JSQMessagesCollectionViewCellIndexPathRow5', label: 'Jesse Squires: Now with media messages!'
Cell 0x7fbe50544cf0: traits: 146028888064, {{4.0, 468.0}, {406.0, 170.0}}, identifier: 'JSQMessagesCollectionViewCellIndexPathRow6', label: 'Jesse Squires: media message'
Cell 0x7fbe5054ba60: traits: 146028888064, {{4.0, 642.0}, {406.0, 40.0}}, identifier: 'JSQMessagesCollectionViewCellIndexPathRow7', label: 'Jesse Squires: media message'
}
↪︎Find: Elements matching predicate '"JSQMessagesCollectionViewCellIndexPathRow2" IN identifiers'
Output: {
Cell 0x7fbe506b4700: traits: 146028888072, {{4.0, 121.0}, {406.0, 99.0}}, identifier: 'JSQMessagesCollectionViewCellIndexPathRow2', label: 'Jesse Squires: It even has data detectors. You can call me tonight. My cell number is 123-456-7890. My website is www.hexedbits.com.'
}
↪︎Find: Descendants matching type TextView
(lldb) po jsqmessagescollectionviewcellindexpathrow2Cell.textViews[@"JSQMessagesCollectionViewCellIndexPathRow2TextView"]
t = 79.86s Use cached accessibility hierarchy for com.hexedbits.JSQMessages
t = 79.87s Find: Descendants matching type CollectionView
t = 79.87s Find: Descendants matching type Cell
Query chain:
→Find: Target Application 0x7fbe50737d20
Output: {
Application 0x7fbe5052b6f0: {{0.0, 0.0}, {414.0, 736.0}}, label: 'JSQMessages'
}
↪︎Find: Descendants matching type CollectionView
Output: {
CollectionView 0x7fbe506ddd30: traits: 35192962023424, {{0.0, 0.0}, {414.0, 736.0}}
}
↪︎Find: Descendants matching type Cell
Output: {
Cell 0x7fbe506f0950: traits: 146028888064, {{4.0, -62.0}, {406.0, 77.0}}, identifier: 'JSQMessagesCollectionViewCellIndexPathRow0', label: 'Jesse Squires: Welcome to JSQMessages: A messaging UI framework for iOS.'
Cell 0x7fbe506ee0e0: traits: 146028888064, {{4.0, 19.0}, {406.0, 98.0}}, identifier: 'JSQMessagesCollectionViewCellIndexPathRow1', label: 'Steve Wozniak: It is simple, elegant, and easy to use. There are super sweet default settings, but you can customize like crazy.'
Cell 0x7fbe506b4700: traits: 146028888072, {{4.0, 121.0}, {406.0, 99.0}}, identifier: 'JSQMessagesCollectionViewCellIndexPathRow2', l t = 79.87s Find: Elements matching predicate '"JSQMessagesCollectionViewCellIndexPathRow2" IN identifiers'
t = 79.87s Find: Descendants matching type TextView
t = 79.87s Find: Elements matching predicate '"JSQMessagesCollectionViewCellIndexPathRow2TextView" IN identifiers'
t = 79.87s Use cached accessibility hierarchy for com.hexedbits.JSQMessages
t = 79.88s Find: Descendants matching type CollectionView
t = 79.88s Find: Descendants matching type Cell
t = 79.88s Find: Elements matching predicate '"JSQMessagesCollectionViewCellIndexPathRow2" IN identifiers'
t = 79.88s Find: Descendants matching type TextView
t = 79.88s Find: Elements matching predicate '"JSQMessagesCollectionViewCellIndexPathRow2TextView" IN identifiers'
abel: 'Jesse Squires: It even has data detectors. You can call me tonight. My cell number is 123-456-7890. My website is www.hexedbits.com.'
Cell 0x7fbe506d9c90: traits: 146028888064, {{4.0, 224.0}, {406.0, 118.0}}, identifier: 'JSQMessagesCollectionViewCellIndexPathRow3', label: 'Jobs: JSQMessagesViewController is nearly an exact replica of the iOS Messages App. And perhaps, better.'
Cell 0x7fbe506d7d30: traits: 146028888064, {{4.0, 346.0}, {406.0, 77.0}}, identifier: 'JSQMessagesCollectionViewCellIndexPathRow4', label: 'Tim Cook: It is unit-tested, free, open-source, and documented.'
Cell 0x7fbe5054cc40: traits: 146028888064, {{4.0, 427.0}, {406.0, 37.0}}, identifier: 'JSQMessagesCollectionViewCellIndexPathRow5', label: 'Jesse Squires: Now with media messages!'
Cell 0x7fbe50544cf0: traits: 146028888064, {{4.0, 468.0}, {406.0, 170.0}}, identifier: 'JSQMessagesCollectionViewCellIndexPathRow6', label: 'Jesse Squires: media message'
Cell 0x7fbe5054ba60: traits: 146028888064, {{4.0, 642.0}, {406.0, 40.0}}, identifier: 'JSQMessagesCollectionViewCellIndexPathRow7', label: 'Jesse Squires: media message'
}
↪︎Find: Elements matching predicate '"JSQMessagesCollectionViewCellIndexPathRow2" IN identifiers'
Output: {
Cell 0x7fbe506b4700: traits: 146028888072, {{4.0, 121.0}, {406.0, 99.0}}, identifier: 'JSQMessagesCollectionViewCellIndexPathRow2', label: 'Jesse Squires: It even has data detectors. You can call me tonight. My cell number is 123-456-7890. My website is www.hexedbits.com.'
}
↪︎Find: Descendants matching type TextView
↪︎Find: Elements matching predicate '"JSQMessagesCollectionViewCellIndexPathRow2TextView" IN identifiers'
Based on what I have found out, simply assigning accessibilityIdentifier to the collection cell does not actually help to make the sub-elements queryable. So, we suggest fellows, who do UI testing, to assign their custom identifiers for unveiling sub elements they want.
BTW, the identifier of textView property is assigned by combining the collection cell identifier with the concatenation of TextView. That's why it's queryable under this
po jsqmessagescollectionviewcellindexpathrow2Cell.textViews[@"JSQMessagesCollectionViewCellIndexPathRow2TextView"]
|
gharchive/issue
| 2016-07-28T19:51:03 |
2025-04-01T04:34:40.593660
|
{
"authors": [
"Lucashuang0802",
"jessesquires",
"mwgray"
],
"repo": "jessesquires/JSQMessagesViewController",
"url": "https://github.com/jessesquires/JSQMessagesViewController/issues/1768",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
76260887
|
Inputtoolbar disappears if I dismiss the view and than present it again
Hi, I am writing a small chat app, where I use the JSQMessageViewController for the main component of the chat view controller. Beyond that, I have another view, which contains a table showing the list of current chats, where I should open the chat by clicking the cells of the table.
The problem is that, opening the chat by clicking the cells of the table is fine at the first time; but after I dismiss that chat by clicking left top button of the navigation bar and try to reopen the chat by clicking the cell again, the problem happens, where the inputtoolbar doesn't show anymore.
Any idea on it?
are you using swift or Objective c for your project?
@jayisidoro I am using objective c. Thanks!
Same issue and I am using Swift
For me is fixed when change the modalTransitionStyle from default value .coverVertical to any other style e.g. .flipHorizontal
Hello everyone!
I'm sorry to inform the community that I'm officially deprecating this project. 😢 Please read my blog post for details:
http://www.jessesquires.com/blog/officially-deprecating-jsqmessagesviewcontroller/
Thus, I'm closing all issues and pull requests and making the necessary updates to formally deprecate the library. I'm sorry if this is unexpected or disappointing. Please know that this was an extremely difficult decision to make. I'd like to thank everyone here for contributing and making this project so great. It was a fun 4 years. 😊
Thanks for understanding,
— jsq
|
gharchive/issue
| 2015-05-14T07:01:43 |
2025-04-01T04:34:40.598757
|
{
"authors": [
"A1exandre",
"jayisidoro",
"jessesquires",
"kakashysen",
"zhangtemplar"
],
"repo": "jessesquires/JSQMessagesViewController",
"url": "https://github.com/jessesquires/JSQMessagesViewController/issues/982",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
520517213
|
Helm v3 compatible chart release
Is your feature request related to a problem? Please describe.
Helm v3 chart changes have already been made (#2286). I saw the comment that this is coming soon (https://github.com/jetstack/cert-manager/issues/1744#issuecomment-542790564). I'm adding this issue just to help follow the next version.
Describe the solution you'd like
New release that includes (#2286) CRD YAML in /crd directory for Helm v3
Describe alternatives you've considered
Workaround is create the CRDs manually from the manifest, wait for them to be established with kubeapi-server, then install the chart.
Additional context
Just a placeholder to follow the next release 😄
Environment details (if applicable):
Kubernetes version: Any relevant version. This is for Helm v3 compatibility
Cloud-provider/provisioner (e.g. GKE, kops AWS, etc): Any
cert-manager version (e.g. v0.4.0): Fix (#2286) was merged after v0.11.0
Install method: Helm v3 (most recent v3 pre-release is v3.0.0-rc.3)
/kind feature
I’m going to close this issue as there’s not really anything more actionable from us, as you said, the relevant PR has already merged 😀
We’re aiming to have 0.12 cut by the end of this week 🎉
Great looking forward to that. Especially as Helm 3.0.0 released Today 😸
|
gharchive/issue
| 2019-11-09T21:36:44 |
2025-04-01T04:34:40.653267
|
{
"authors": [
"munnerz",
"scottrigby"
],
"repo": "jetstack/cert-manager",
"url": "https://github.com/jetstack/cert-manager/issues/2344",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
590585590
|
Not able to generate a certificate because of 404 with GKE
Describe the bug:
I am trying to use cert-manager to issue a LetsEncrypt certificate but after the creation of a cm-acme-http-solver ingress instance, I started getting 404 both on my domain which was working and the challenge domain.
Here are the logs cert-manager pod gave:
I0330 20:30:58.433950 1 ingress.go:91] cert-manager/controller/challenges/http01/selfCheck/http01/ensureIngress "msg"="found one existing HTTP01 solver ingress" "dnsName"="example.domain.com" "related_resource_kind"="Ingress" "related_resource_name"="cm-acme-http-solver-6l5bv" "related_resource_namespace"="http-echo" "resource_kind"="Challenge" "resource_name"="example-domain-com-tls-2208256997-1752918605-1819454371" "resource_namespace"="http-echo" "type"="http-01"
E0330 20:30:58.452220 1 sync.go:184] cert-manager/controller/challenges "msg"="propagation check failed" "error"="wrong status code '404', expected '200'" "dnsName"="example.domain.com" "resource_kind"="Challenge" "resource_name"="example-domain-com-tls-2208256997-1752918605-1819454371" "resource_namespace"="http-echo" "type"="http-01"
The certificate which was created events:
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal GeneratedKey 40m cert-manager Generated a new private key
Normal Requested 40m cert-manager Created new CertificateRequest resource "example-domain-com-tls-2208256997"
The ingress YAML:
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
annotations:
cert-manager.io/issuer: letsencrypt-staging
kubernetes.io/ingress.class: nginx
name: http-echo-ingress
namespace: http-echo
spec:
backend:
serviceName: http-echo-service
servicePort: 5678
rules:
- host: domain.example.com
http:
paths:
- backend:
serviceName: http-echo-service
servicePort: 5678
path: /
tls:
- hosts:
- domain.example.com
secretName: domain-example-com-tls
status:
loadBalancer:
ingress:
- ip: xx.xx.xx.xx
The service YAML:
apiVersion: v1
kind: Service
metadata:
name: http-echo-service
namespace: http-echo
labels:
app: http-echo
spec:
type: NodePort
ports:
- name: http
port: 5678
protocol: TCP
selector:
app: http-echo
The issuer YAML:
apiVersion: cert-manager.io/v1alpha2
kind: Issuer
metadata:
name: letsencrypt-staging
spec:
acme:
# The ACME server URL
server: https://acme-staging-v02.api.letsencrypt.org/directory
email: email@email.com
privateKeySecretRef:
name: letsencrypt-staging
solvers:
- http01:
ingress:
class: nginx
Expected behavior:
Have the certificate marked as READY and correctly issued by LetsEncrypt
Steps to reproduce the bug:
Have a GKE cluster with cert-manager installed and an issuer for LE as the one presented above
Create a simple app deployment and service with the YAML below:
apiVersion: apps/v1
kind: Deployment
metadata:
labels:
app: http-echo
name: http-echo
namespace: http-echo
spec:
selector:
matchLabels:
app: http-echo
replicas: 1
strategy: {}
template:
metadata:
labels:
app: http-echo
spec:
containers:
- name: http-echo
image: hashicorp/http-echo
args: ["-text=\"It works!\""]
ports:
- containerPort: 5678
resources: {}
restartPolicy: Always
status: {}
----
apiVersion: v1
kind: Service
metadata:
name: http-echo-service
namespace: http-echo
labels:
app: http-echo
spec:
type: NodePort
ports:
- name: http
port: 5678
protocol: TCP
selector:
app: http-echo
Create an Ingress with cert-manager annotation commented
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
annotations:
# cert-manager.io/issuer: letsencrypt-staging
kubernetes.io/ingress.class: nginx
name: http-echo-ingress
namespace: http-echo
spec:
backend:
serviceName: http-echo-service
servicePort: 5678
rules:
- host: domain.example.com
http:
paths:
- backend:
serviceName: http-echo-service
servicePort: 5678
path: /
tls:
- hosts:
- domain.example.com
secretName: domain-example-com-tls
status:
loadBalancer:
ingress:
- ip: xx.xx.xx.xx
Try to access the service using the domain and everything works fine considering you already set your DNS properly
Uncomment the line on the annotations for the ingress and apply
cert-manager.io/issuer: letsencrypt-staging
Get 404 for the example.domain.com which was working, certificate created but never ready.
Anything else we need to know?:
Environment details::
Kubernetes version (e.g. v1.10.2):
Client Version: version.Info{Major:"1", Minor:"17", GitVersion:"v1.17.4", GitCommit:"8d8aa39598534325ad77120c120a22b3a990b5ea", GitTreeState:"clean", BuildDate:"2020-03-12T21:03:42Z", GoVersion:"go1.13.8", Compiler:"gc", Platform:"linux/amd64"}
Server Version: version.Info{Major:"1", Minor:"15+", GitVersion:"v1.15.9-gke.22", GitCommit:"9bae7fcacb520dfee658b26cc1a9643bf787dfc3", GitTreeState:"clean", BuildDate:"2020-02-27T18:43:37Z", GoVersion:"go1.12.12b4", Compiler:"gc", Platform:"linux/amd64"}
Cloud-provider/provisioner (e.g. GKE, kops AWS, etc): GKE
cert-manager version (e.g. v0.4.0): v0.14.1
Install method (e.g. helm or static manifests): static manifests
/kind bug
I have the same issue, I'm using GKE Ingress so my ClusterIssuer is different:
apiVersion: cert-manager.io/v1alpha2
kind: ClusterIssuer
metadata:
name: letsencrypt-prod
spec:
acme:
server: https://acme-v02.api.letsencrypt.org/directory
email: 'email@email.com.it'
privateKeySecretRef:
name: letsencrypt-prod
solvers:
- http01:
ingress:
class: gce
I think that #2781 is related to this issue
In my case, 2 ingress was created.
I try to use this annotation
acme.cert-manager.io/http01-edit-in-place: "true"
https://cert-manager.io/docs/usage/ingress/#supported-annotations
With this annotation only one ingress is created whit this spec:
spec:
rules:
- host: dominio.it
http:
paths:
- backend:
serviceName: cm-acme-http-solver-r9rb7
servicePort: 8089
path: /.well-known/acme-challenge/_j2KAsOUStfL9agubLN6wnrYpVpVlqE11RMfE4QQMIg
- backend:
serviceName: dominio-it-prestashop
servicePort: http
path: /
tls:
- hosts:
- dominio.it
secretName: dominio.it-tls
If I describe the ingress:
Name: domnio-it-prestashop
Namespace: verticalshops
Address:
Default backend: default-http-backend:80 (<error: endpoints "default-http-backend" not found>)
TLS:
domnio.it-tls terminates domnio.it
Rules:
Host Path Backends
---- ---- --------
domnio.it
/.well-known/acme-challenge/_j2KAsOUStfL9agubLN6wnrYpVpVlqE11RMfE4QQMIg cm-acme-http-solver-r9rb7:8089 (10.48.15.5:8089)
/ domnio-it-prestashop:http (10.48.13.9:80)
Annotations: acme.cert-manager.io/http01-edit-in-place: true
cert-manager.io/cluster-issuer: letsencrypt-prod
kubernetes.io/ingress.class: gce
kubernetes.io/tls-acme: true
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal CreateCertificate 7m16s cert-manager Successfully created Certificate "domnio.it-tls"
The cert-manager pod logs still get the same error:
E0410 01:48:18.012676 1 sync.go:184] cert-manager/controller/challenges "msg"="propagation check failed" "error"="wrong status code '404', expected '200'" "dnsName"="domnio.it" "resource_kind"="Challenge" "resource_name"="domnio.it-tls-1884924373-1288961440-649377576" "resource_namespace"="verticalshops" "type"="http-01"
I0410 01:48:18.012726 1 controller.go:144] cert-manager/controller/challenges "msg"="finished processing work item" "key"="verticalshops/domnio.it-tls-1884924373-1288961440-649377576"
Can you check in the Google Cloud console what the status of the new path is in the GCE loadbalancer?
@meyskens The HTTP load balancer service was disable in the cluster. I have enabled it and all works fine.
thanks
I couldn't debug further since I created a new cluster and started over, I used the same configuration and all worked well. Because of that, I will close the issue.
|
gharchive/issue
| 2020-03-30T20:48:17 |
2025-04-01T04:34:40.666278
|
{
"authors": [
"Nittarab",
"meyskens",
"vitorfhc"
],
"repo": "jetstack/cert-manager",
"url": "https://github.com/jetstack/cert-manager/issues/2767",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.