id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
214504715
|
fix daylight savings test issues
I was getting errors running my tests ever since DST: https://gist.github.com/dasl-/0d227303ec280644c7ead44e28629232
Looks like there is already precedent in the code for doing this sort of thing: https://github.com/dasl-/debezium/blob/031c4a15524c1de72510d60d64c87aa344435599/debezium-connector-mysql/src/test/java/io/debezium/connector/mysql/MySqlConnectorRegressionIT.java#L604-L604
@dasl- Thanks for fixing these! Would you mind logging a JIRA issue for this, and add some of the same details into the issue rather than linking to a gist (which at some point might not resolvable)? Thanks!
Integration tests on Travis are failing due to a failure in Travis infrastructure.
|
gharchive/pull-request
| 2017-03-15T19:45:30 |
2025-04-01T04:33:57.713068
|
{
"authors": [
"dasl-",
"rhauch"
],
"repo": "debezium/debezium",
"url": "https://github.com/debezium/debezium/pull/199",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
964406151
|
DBZ-3843: Add a transformer to unwrap the records with struct-wrapped fields
I opened this issue and described the background in the issue ticket. I think adding a separate SMT is the most non-invasive and effective approach to solve this problem. Please have a look.
Thanks
https://issues.redhat.com/browse/DBZ-3843
Hey @bingqinzhou, do you know why the Cassandra connector generates the struct style in the first place? I believe there's a good reason, but I don't quite remember, unfortunately. Or, @jgao54, if you happen to read this, can you share some insights? Thanks a lot!
@elakito, could you rebase and force-push this one? We're struggling a bit with intermittent test failures atm. and I'm trying to get them sorted out:
Vitess: should be fixed by now
Postgres: should be fixed by now
SQL Server: Still open, trracked via https://issues.redhat.com/browse/DBZ-2398
Server: : should be fixed by now
Hey @elakito, could you rebase and force-push this PR? Most/all of those test failures above should be addressed by now. Thx!
Hey @bingqinzhou, do you know why the Cassandra connector generates the struct style in the first place? I believe there's a good reason, but I don't quite remember, unfortunately. Or, @jgao54, if you happen to read this, can you share some insights? Thanks a lot!
From my understanding, the deletion_ts field is used for downstream processing to differentiate if column is just not included in an update operation or is actually deleted. I'm good with the change in this PR as long as it doesn't affect the current use case of Cassandra Connector.
yes. I'll rebase this PR.
Hey @elakito, coming back to this one, sorry for the long delay. So my main question/concern is: when using this SMT, how would a downstream consumer tell apart a column whose value wasn't modified from one which was set to null?
Hi @gunnarmorling, I had only insertion scenarios and didn't notice Cassandra-connector has a different behavior for the update or delete scenarios and Cassandra-connector is not using before and after to specify the changes as described in debezium's event structure documentation https://debezium.io/documentation/reference/transformations/event-flattening.html but only using after to describe the changes.
In that case, an SMT cannot be used to transform Cassandra-connecto'r records into debezium's current CDC envelope records using SMT. Is Cassandra-connector's record format also supported by other debezium components? Will this format be treated as a valid stable debezium CDC envelope (type 2) in the future? In that case, the downstream consumers can directly consume those records. Could you tell me whether Debezium team is planning to align these formats?
Thanks.
aki
Hi @gunnarmorling, have you had time to follow up on my question above? I now tend to think it is good to document the CDC envelope format that Cassandra-connector currently uses and name it a type-2 or something to differentiate from the standard CDC with before/after properties.
These two formats are not compatible, as converting from the standard debezium CDC format to this type-2 format is possible with a stateless transformation, but converting in the other direction is not. But if the format is documented by Debeizum project, one can develop downstream applications to consume this type-2 format more directly and efficiently. Furthermore, for some data sources, it may be easier to generate this type-2 format and there will be some additional source connectors. If we can go in this direction, we can close this PR and open an issue to follow up on documenting this type-2 format.
regards, aki
Hey @elakito, the format is already documented in the Cassandra connector docs (section Events). The Cassandra connector is the only one which uses this format, and I don't think it'd make too much sense to add support for it to the other connectors. So in that light, is there actually anything to be done here?
These two formats are not compatible, as converting from the standard debezium CDC format to this type-2 format is possible with a stateless transformation, but converting in the other direction is not.
Thinking about it, you actually can convert from "type-2" to the standard Debezium format. Specifically, the Postgres connector already supports a special marker value which will be emitted if the value of an unchanged column cannot be retrieved (a corner case for PG TOAST columns). So it should be possible to transform the type-2 format from Cassandra into the regular one, emitting the "__debezium_unavailable_value" value for all unchanged columns.
Would that help your case? If so, would you be willing to update this PR accordingly? I could see value in having such SMT provided by Debezium OOTB. WDYT?
@gunnarmorling Thank you for the documentation link to the type-2 format.
But regarding the transformation from type-2 to type-1 (the normal debezium's event format), I don't think it makes much sense. If we use some special marker to represent the missing values, we can only use such a marker for string or bytes based types but not for numeric types.
And if such transformation made sense, Cassandra-connector could have adopted it.
For datasource types like Cassandra, only the type-2 format makes sense and the downstream applications can be programmed to directly consume those type-2 records than to be programmed to consume the corresponding type-1 converted records that include the special markers.
So, I think we can close this PR and also the issue linked from this PR.
Ok, going to close this PR then. Thanks a lot nevertheless for the good discussion, it also helped me to get clarity about some things. We are also exploring the idea to have a Debezium-specific JDBC sink connector (see DBZ-3647), in which case we may add support for the "type 2" format for a full end-to-end flow.
Thank you. It was a good discussion to understand the problem and find a good solution.
|
gharchive/pull-request
| 2021-08-09T21:31:13 |
2025-04-01T04:33:57.725521
|
{
"authors": [
"bingqinzhou",
"elakito",
"gunnarmorling"
],
"repo": "debezium/debezium",
"url": "https://github.com/debezium/debezium/pull/2588",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1134103516
|
DBZ-4707 MySQL unparseable DDL - CREATE PROCEDURE
https://issues.redhat.com/browse/DBZ-4707
upstream pr: https://github.com/antlr/grammars-v4/pull/2509
Thanks, this LGTM. Though for the test it would suffice (or be better actually), to add this to the right SQL file in the parser module itself, as e.g. done in this commit: dd54a37. Could you change this accordingly?
Yes, put this procedure to ddl_create.sql
|
gharchive/pull-request
| 2022-02-12T14:06:56 |
2025-04-01T04:33:57.728727
|
{
"authors": [
"harveyyue"
],
"repo": "debezium/debezium",
"url": "https://github.com/debezium/debezium/pull/3233",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1993864766
|
Error : after the second argument: expected table, got nil
Error executing Lua callback: vim/shared.lua:0: after the second argument: expected table, got nil
stack traceback:
[C]: in function 'error'
vim/shared.lua: in function 'validate'
vim/shared.lua: in function 'tbl_deep_extend'
...y/telescope-undo.nvim/lua/telescope/_extensions/undo.lua:37: in function <...y/telescope-undo.nvim/lua/telescope/_extensions/undo.lua:36>
...share/nvim/lazy/telescope.nvim/lua/telescope/command.lua:193: in function 'run_command'
...share/nvim/lazy/telescope.nvim/lua/telescope/command.lua:259: in function 'load_command'
...ocal/share/nvim/lazy/telescope.nvim/plugin/telescope.lua:108: in function <...ocal/share/nvim/lazy/telescope.nvim/plugin/telescope.lua:107>
The entire Telescope conf: https://github.com/LamprosPitsillos/nixos-config/blob/main/home-manager/programs/nvim/lua/plugins/Telescope.lua
A snipet:
extensions = {
undo = {
use_delta = false,
use_custom_command = nil, -- setting this implies `use_delta = false`. Accepted format is: { "bash", "-c", "echo '$DIFF' | delta" }
side_by_side = false,
layout_strategy = "vertical",
layout_config = {
preview_height = 0.8,
},
diff_context_lines = vim.o.scrolloff,
entry_format = "state #$ID, $STAT, $TIME",
time_format = "",
mappings = {
i = {
-- IMPORTANT: Note that telescope-undo must be available when telescope is configured if
-- you want to replicate these defaults and use the following actions. This means
-- installing as a dependency of telescope in it's `requirements` and loading this
-- extension from there instead of having the separate plugin definition as outlined
-- above.
["<cr>"] = require("telescope-undo.actions").yank_additions,
["<S-cr>"] = require("telescope-undo.actions").yank_deletions,
["<C-cr>"] = require("telescope-undo.actions").restore,
},
},
},
ast_grep = {
command = {
......
Are you on the latest version? I fixed very similar error with the latest commit, that bug was in the latest version for only a few hours. In case this is a different thing (on my phone, can't check details right now) and you are on the latest version, please additionally provide how you invoked telescope-undo.
Yup, that's fixed on the latest version afaict from what you provided :+1:
Sorry but for me it's still an issue.
On the latest commit.
additionally provide how you invoked telescope-undo.
I invoke it using Telescope undo
It's not reproducible with a minimal config, this has to be something about your setup somewhere. I suggest you clear your cache/data dirs and rebootstrap that using your config, you probably somehow pulled in the bad version (sorry for having published that one for a few hours 💔). No idea why Lazy would show the current commit though 🤔 If it still fails after rebootstrapping, please make this minimal config break, let me know, and I'll take a look:
-- bend paths for this repro
local root = vim.fn.fnamemodify("./.repro", ":p")
for _, name in ipairs({ "config", "data", "state", "cache" }) do
vim.env[("XDG_%s_HOME"):format(name:upper())] = root .. "/" .. name
end
-- bootstrap lazy
local lazypath = root .. "/plugins/lazy.nvim"
if not vim.loop.fs_stat(lazypath) then
vim.fn.system({
"git",
"clone",
"--filter=blob:none",
"--single-branch",
"https://github.com/folke/lazy.nvim.git",
lazypath,
})
end
vim.opt.runtimepath:prepend(lazypath)
-- plugin specs
local plugins = {
{
"debugloop/telescope-undo.nvim",
lazy = true,
dependencies = {
{
"nvim-telescope/telescope.nvim",
dependencies = { "nvim-lua/plenary.nvim" },
},
},
keys = {
{
"<leader>u",
"<cmd>Telescope undo<cr>",
desc = "undo history",
},
},
opts = {
extensions = {
undo = {},
},
},
config = function(_, opts)
require("telescope").setup(opts)
require("telescope").load_extension("undo")
end,
},
}
-- let's go
require("lazy").setup(plugins, {
root = root .. "/plugins",
})
Test using nvim -u repro.lua testfile, make a quick undoable edit, and \u (default leader) or use the lua invoke you posted. Go from there, substitute your config for telescope(-undo)?, then add whatever else you might need to make it break, I'll have a look then 👍🏻
Best of luck finding the bad circumstances quickly 🍀
Lazy is showing the correct commit , navigating in to the file shows that the plugin is indeed at the latest commit.
I forked , and used my own fork from dir using lazy and tried to figure out why this is happening...
M.exports.undo = function(config)
vim.notify(vim.inspect(M.config))
vim.notify(vim.inspect(M))
vim.notify(vim.inspect(config))
config = vim.tbl_deep_extend("force", M.config, config or {})
if config.theme then
config = require("telescope.themes")["get_" .. config.theme](config)
end
require("telescope-undo").undo(config)
end
M.setup = function(extension_config, telescope_config)
M.config = vim.tbl_deep_extend("force", defaults, extension_config)
vim.notify("config set")
vim.notify(vim.inspect(M.config))
if M.config["side_by_side"] and not M.config["use_delta"] then
error("telescope_undo.nvim: setting side_by_side but not use_delta will have no effect")
end
end
return telescope.register_extension(M)
2023-11-16T00:19:41 INFO config set
2023-11-16T00:19:41 INFO {
diff_context_lines = 0,
entry_format = "state #$ID, $STAT, $TIME",
layout_config = {
preview_height = 0.8
},
layout_strategy = "vertical",
mappings = {
i = {
["<C-cr>"] = <function 1>,
["<C-r>"] = <function 1>,
["<C-y>"] = <function 2>,
["<S-cr>"] = <function 2>,
["<cr>"] = <function 3>
},
n = {
Y = <function 2>,
u = <function 1>,
y = <function 3>
}
},
saved_only = false,
side_by_side = false,
time_format = "",
use_delta = false
}
2023-11-16T00:19:41 INFO config set
2023-11-16T00:19:41 INFO {
diff_context_lines = 0,
entry_format = "state #$ID, $STAT, $TIME",
layout_config = {
preview_height = 0.8
},
layout_strategy = "vertical",
mappings = {
i = {
["<C-cr>"] = <function 1>,
["<C-r>"] = <function 1>,
["<C-y>"] = <function 2>,
["<S-cr>"] = <function 2>,
["<cr>"] = <function 3>
},
n = {
Y = <function 2>,
u = <function 1>,
y = <function 3>
}
},
saved_only = false,
side_by_side = false,
time_format = "",
use_delta = false
}
////// After calling `Telescope undo`
2023-11-16T00:20:02 INFO nil
2023-11-16T00:20:02 INFO {
formatter = <function 1>,
setup = <function 2>
}
2023-11-16T00:20:02 INFO {}
Thanks for your efforts any way , i will try and investigate some more and notify back with a pr if i find anything.
|
gharchive/issue
| 2023-11-15T01:41:29 |
2025-04-01T04:33:57.756067
|
{
"authors": [
"LamprosPitsillos",
"debugloop"
],
"repo": "debugloop/telescope-undo.nvim",
"url": "https://github.com/debugloop/telescope-undo.nvim/issues/41",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2329285104
|
NotFoundError: The object can not be found here. (adding a nc-root div causes error)
Describe the bug
Adding a <div id="nc-root"></div> to the page I want to load Decap causes an error
I am using SveltKit, and in routes/admin/+page.svelte I have the following code:
<svelte:head>
<meta charset="utf-8" />
<meta name="viewport" content="width=device-width, initial-scale=1.0" />
<meta name="robots" content="noindex" />
<script src="https://identity.netlify.com/v1/netlify-identity-widget.js"></script>
<title>OwC Management</title>
</svelte:head>
<div id="nc-root"></div>
{@html '<script src="https://unpkg.com/decap-cms@^3.0.0/dist/decap-cms.js"></script>'}
{@html '<script> CMS.registerPreviewStyle("styles/cms-preview.css") </script>'}
Loads fine without the div with nc-root id, but when I add the id, it give the below error (the error is in the placement of the div)
Error
There's been an error - please [open an issue on GitHub.](https://github.com/decaporg/decap-cms/issues/new?title=NotFoundError%3A+The+object+can+not+be+found+here.&body=%0A**Describe+the+bug**%0A%0A**To+Reproduce**%0A%0A**Expected+behavior**%0A%0A**Screenshots**%0A%0A**Applicable+Versions%3A**%0A+-+Decap+CMS+version%3A+%60decap-cms%403.1.10%60%0A+-+Git+provider%3A+%60github%60%0A+-+Browser+version%3A+%60Mozilla%2F5.0+%28Macintosh%3B+Intel+Mac+OS+X+10_15_7%29+AppleWebKit%2F605.1.15+%28KHTML%2C+like+Gecko%29+Version%2F17.5+Safari%2F605.1.15%60%0A%0A**CMS+configuration**%0A%60%60%60%0Abackend%3A%0A++name%3A+github%0A++repo%3A+LunarShadow%2Fowc-svelte%0A++branch%3A+main%0A++publish_mode%3A+editorial_workflow%0A++squash_merges%3A+true%0A++commit_messages%3A%0A++++create%3A+Created+%7B%7Bcollection%7D%7D+%22%7B%7Bslug%7D%7D%22%0A++++update%3A+Updated+%7B%7Bcollection%7D%7D+%E2%80%9C%7B%7Bslug%7D%7D%E2%80%9D%0A++++delete%3A+Deleted+%7B%7Bcollection%7D%7D+%E2%80%9C%7B%7Bslug%7D%7D%E2%80%9D%0Asite_url%3A+https%3A%2F%2Fovercomewithchrist.com%0Amedia_library%3A%0A++name%3A+cloudinary%0A++use_secure_url%3A+true%0A++config%3A%0A++++cloud_name%3A+dn7ubiddg%0A++++api_key%3A+866914922276626%0Adisplay_url%3A+%2F%0Alogo_url%3A+https%3A%2F%2Fres.cloudinary.com%2Fdn7ubiddg%2Fimage%2Fupload%2Fv1715370001%2Fbranding%2Ffull-color-logo-text.png%0Aslug%3A%0A++encoding%3A+ascii%0A++clean_accents%3A+true%0A++sanitize_replacement%3A+%22-%22%0Acollections%3A%0A++-+name%3A+page%0A++++label%3A+Page%0A++++folder%3A+src%2Fcontent%2Fpages%0A++++preview_path%3A+%2F%7B%7Bslug%7D%7D%0A++++create%3A+true%0A++++slug%3A+%22%7B%7Bslug%7D%7D%22%0A++++fields%3A%0A++++++-+label%3A+Layout%0A++++++++name%3A+layout%0A++++++++widget%3A+hidden%0A++++++++default%3A+page%0A++++++-+label%3A+Title%0A++++++++name%3A+title%0A++++++++widget%3A+string%0A++++++-+label%3A+Body%0A++++++++name%3A+body%0A++++++++widget%3A+markdown%0A++++publish%3A+true%0A++++type%3A+folder_based_collection%0A++++sortable_fields%3A%0A++++++-+commit_date%0A++++++-+title%0A++++++-+commit_author%0A++++view_filters%3A+%5B%5D%0A++++view_groups%3A+%5B%5D%0A++-+name%3A+post%0A++++label%3A+Posts%0A++++label_singular%3A+Post%0A++++folder%3A+src%2Fcontent%2Fposts%0A++++create%3A+true%0A++++slug%3A+%22%7B%7Bslug%7D%7D%22%0A++++preview_path%3A+posts%2F%7B%7Bslug%7D%7D%0A++++fields%3A%0A++++++-+label%3A+Layout%0A++++++++name%3A+layout%0A++++++++widget%3A+hidden%0A++++++++default%3A+post%0A++++++-+label%3A+Title%0A++++++++name%3A+title%0A++++++++widget%3A+string%0A++++++-+label%3A+Category%0A++++++++name%3A+category%0A++++++++widget%3A+select%0A++++++++options%3A%0A++++++++++-+Bible+Study%0A++++++++++-+Resources%0A++++++++++-+Testimony%0A++++++++++-+Salvation%0A++++++++++-+Prayer%0A++++++++++-+Creative%0A++++++-+label%3A+Featured+Image%0A++++++++name%3A+featuredImage%0A++++++++widget%3A+image%0A++++++-+label%3A+Video+Link%0A++++++++name%3A+videoLink%0A++++++++widget%3A+string%0A++++++++required%3A+false%0A++++++-+label%3A+Summary%0A++++++++name%3A+summary%0A++++++++widget%3A+text%0A++++++-+label%3A+Publish+Date%0A++++++++name%3A+date%0A++++++++widget%3A+datetime%0A++++++++date_format%3A+LL%0A++++++++format%3A+LL%0A++++++-+label%3A+Body%0A++++++++name%3A+body%0A++++++++widget%3A+markdown%0A++++++-+label%3A+Author%0A++++++++name%3A+author%0A++++++++widget%3A+string%0A++++++++default%3A+Cyntia+Seumo%0A++++publish%3A+true%0A++++type%3A+folder_based_collection%0A++++sortable_fields%3A%0A++++++-+commit_date%0A++++++-+title%0A++++++-+date%0A++++++-+author%0A++++++-+summary%0A++++view_filters%3A+%5B%5D%0A++++view_groups%3A+%5B%5D%0A++-+name%3A+product%0A++++label%3A+Products%0A++++label_singular%3A+Product%0A++++folder%3A+src%2Fcontent%2Fproducts%0A++++create%3A+true%0A++++slug%3A+%22%7B%7Bslug%7D%7D%22%0A++++fields%3A%0A++++++-+label%3A+Layout%0A++++++++name%3A+layout%0A++++++++default%3A+product%0A++++++-+label%3A+Name%0A++++++++name%3A+title%0A++++++++widget%3A+string%0A++++++-+label%3A+Image%0A++++++++name%3A+image%0A++++++++widget%3A+image%0A++++++-+label%3A+Price%0A++++++++name%3A+price%0A++++++++widget%3A+number%0A++++++++value_type%3A+float%0A++++++-+label%3A+Description%0A++++++++name%3A+description%0A++++++++widget%3A+text%0A++++++-+label%3A+Sizes%0A++++++++name%3A+sizes%0A++++++++widget%3A+list%0A++++++++allow_add%3A+true%0A++++++++fields%3A%0A++++++++++-+label%3A+Size%0A++++++++++++name%3A+size%0A++++++++++++widget%3A+string%0A++++++++++-+label%3A+Price%0A++++++++++++name%3A+price%0A++++++++++++widget%3A+number%0A++++++++++++value_type%3A+float%0A++++publish%3A+true%0A++++type%3A+folder_based_collection%0A++++sortable_fields%3A%0A++++++-+commit_date%0A++++++-+title%0A++++++-+commit_author%0A++++++-+description%0A++++view_filters%3A+%5B%5D%0A++++view_groups%3A+%5B%5D%0Apublish_mode%3A+simple%0Apublic_folder%3A+%2F%0AisFetching%3A+false%0Aerror%3A+null%0A%0A%60%60%60%0A%0A**Additional+context**%0A&labels=type%3A+bug)
Opening an issue pre-populates it with the error message and debugging data.
Please verify the information is correct and remove sensitive data if exists.
Details
removeChild@[native code] hl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2211743 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213007 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213405 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2214585 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213270 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2214585 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213270 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2214585 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2214585 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213405 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2214585 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2214585 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213405 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213405 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2214585 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213270 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2214585 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213270 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213270 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2214585 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213270 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2214585 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213270 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213270 gl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2213157 vl@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2214456 @https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2228336 xu@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2228891 ou@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2221390 x@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2607115 P@https://unpkg.com/decap-cms@3.1.10/dist/decap-cms.js:485:2607647
To Reproduce
add <div id="nc-root"></div> to svelte page where you want to load decap
Expected behavior
decap loads in the div as it normally would, but respects the layout set for the page.
Screenshots
with div:
without div:
Applicable Versions:
Decap CMS version: decap-cms@3.1.10
Git provider: github
Browser version: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/17.5 Safari/605.1.15
CMS configuration
backend:
name: github
repo: ...
branch: main
publish_mode: editorial_workflow
squash_merges: true
commit_messages:
create: Created {{collection}} "{{slug}}"
update: Updated {{collection}} “{{slug}}”
delete: Deleted {{collection}} “{{slug}}”
site_url: ...
media_library:
name: cloudinary
use_secure_url: true
config:
cloud_name: ....
api_key: ...
display_url: /
logo_url: .../branding/full-color-logo-text.png
slug:
encoding: ascii
clean_accents: true
sanitize_replacement: "-"
collections:
- name: page
label: Page
folder: src/content/pages
preview_path: /{{slug}}
create: true
slug: "{{slug}}"
fields:
- label: Layout
name: layout
widget: hidden
default: page
- label: Title
name: title
widget: string
- label: Body
name: body
widget: markdown
publish: true
type: folder_based_collection
sortable_fields:
- commit_date
- title
- commit_author
view_filters: []
view_groups: []
- name: post
label: Posts
label_singular: Post
folder: src/content/posts
create: true
slug: "{{slug}}"
preview_path: posts/{{slug}}
fields:
- label: Layout
name: layout
widget: hidden
default: post
- label: Title
name: title
widget: string
- label: Category
name: category
widget: select
options:
- Bible Study
- Resources
- Testimony
- Salvation
- Prayer
- Creative
- label: Featured Image
name: featuredImage
widget: image
- label: Video Link
name: videoLink
widget: string
required: false
- label: Summary
name: summary
widget: text
- label: Publish Date
name: date
widget: datetime
date_format: LL
format: LL
- label: Body
name: body
widget: markdown
- label: Author
name: author
widget: string
default: Cyntia Seumo
publish: true
type: folder_based_collection
sortable_fields:
- commit_date
- title
- date
- author
- summary
view_filters: []
view_groups: []
- name: product
label: Products
label_singular: Product
folder: src/content/products
create: true
slug: "{{slug}}"
fields:
- label: Layout
name: layout
default: product
- label: Name
name: title
widget: string
- label: Image
name: image
widget: image
- label: Price
name: price
widget: number
value_type: float
- label: Description
name: description
widget: text
- label: Sizes
name: sizes
widget: list
allow_add: true
fields:
- label: Size
name: size
widget: string
- label: Price
name: price
widget: number
value_type: float
publish: true
type: folder_based_collection
sortable_fields:
- commit_date
- title
- commit_author
- description
view_filters: []
view_groups: []
publish_mode: simple
public_folder: /
isFetching: false
error: null
Additional context
Side Note:
It seems Decap CMS doesn't load well with snipcart. It seems to refresh or something constantly and causes my browser to crash. Maybe I'm not doing something correctly, but figured I'd add it here as well.
Thanks
Taking a look at html of successfully built cms, it's seen that decap cms uses the same id nc-root.
Because you inserted another div with the same id, cms "detected" that app already exists and tried to use existing div. Hence the error.
You'll need to use some other id
I added the doc with that root according to the instructions on the documentation for having decap appear in a specific spot.
https://decapcms.org/docs/custom-mounting/
guess you guys don’t have a solution, which is really disappointing.
Taking a look at html of successfully built cms, it's seen that decap cms uses the same id nc-root.
Because you inserted another div with the same id, cms "detected" that app already exists and tried to use existing div. Hence the error. You'll need to use some other id
I added the div according to the instructions on the documentation for having decap custom mounted.
https://decapcms.org/docs/custom-mounting/
guess you guys don’t have a solution, which is really disappointing. Probably would be a good idea to update the documentation then.
@LunarShadow next time please include instructions url in original description, to avoid similar missunderstandings
@LunarShadow
{@html '<script defer src="https://unpkg.com/decap-cms@^3.0.0/dist/decap-cms.js"></script>'} solves the issue for me.
You need to execute decapCms script after document is ready. Otherwise document.getElementById() is not able to find element with proper id and inserts new one. Result is then having two divs with the same id.
@hip3r could you add this clarification to the docs, please?
I don't get an error, but also nothing loads when I add the defer.
code in admin.svelte
<svelte:head>
<meta name="robots" content="noindex" />
<script src="https://identity.netlify.com/v1/netlify-identity-widget.js"></script>
<title>OwC Management</title>
</svelte:head>
<div id="nc-root"></div>
{@html '<script defer src="https://unpkg.com/decap-cms@^3.0.0/dist/decap-cms.js"></script>'}
{@html '<script> CMS.registerPreviewStyle("styles/cms-preview.css") </script>'}
@hip3r can you investigate?
using provided config.yml and svelte code
<svelte:head>
<meta name="robots" content="noindex" />
<script src="https://identity.netlify.com/v1/netlify-identity-widget.js"></script>
<title>OwC Management</title>
</svelte:head>
<div id="nc-root"></div>
{@html '<script defer src="https://unpkg.com/decap-cms@^3.0.0/dist/decap-cms.js"></script>'}
succesfully shows cms
The difference is that I'm missing line {@html '<script> CMS.registerPreviewStyle("styles/cms-preview.css") </script>'}, because I don't have source file.
@LunarShadow if you change this line into {@html '<script defer> CMS.registerPreviewStyle("styles/cms-preview.css") </script>'}, does it help?
You must have error in console, since CMS.registerPreviewStyle is called before object CMS is created
if you change this line into {@html ''}, does it help?
Yes, this helped.
Thank you guys for helping me figure this out. I really appreciate your time and effort.
|
gharchive/issue
| 2024-06-01T18:44:10 |
2025-04-01T04:33:57.779969
|
{
"authors": [
"LunarShadow",
"hip3r",
"martinjagodic"
],
"repo": "decaporg/decap-cms",
"url": "https://github.com/decaporg/decap-cms/issues/7223",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2366669697
|
chore: enable forward plus
What does this PR change?
Enables the forward plus rendering path.
How to test the changes?
On both mac & windows everything should function as normal with no graphical differences between the existing versions.
Our Code Review Standards
https://github.com/decentraland/unity-renderer/blob/master/docs/code-review-standards.md
App crashes as soon as it is opened. It gets stuck here:
And these are the crash logs:
Player.log
|
gharchive/pull-request
| 2024-06-21T14:30:13 |
2025-04-01T04:33:57.789985
|
{
"authors": [
"LucasLioyQA",
"m3taphysics"
],
"repo": "decentraland/unity-explorer",
"url": "https://github.com/decentraland/unity-explorer/pull/1299",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1994033948
|
Send individual id to data4 diabetes registration
send individual id in the last name field
Resolved
|
gharchive/issue
| 2023-11-15T04:46:03 |
2025-04-01T04:33:57.790759
|
{
"authors": [
"josmilan"
],
"repo": "decentralised-dataexchange/data4diabetes-app",
"url": "https://github.com/decentralised-dataexchange/data4diabetes-app/issues/191",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1102499439
|
Signature verification methods should not panic if they return Result
Hey there. I was planning to use this lib for my project but am unable to if signature verification panics on invalid signature data. Since signature content can come from untrusted sources, I need to be able to gracefully handle invalid signature data.
https://docs.rs/did-key/0.1.0/src/did_key/ed25519.rs.html#180-186
The verify method for various keypairs returns a Result type, but the signature data uses .expect() instead of mapping the error so I can't handle that case.
Thanks!
Fixed with #27. Thanks for reporting this
|
gharchive/issue
| 2022-01-13T22:33:09 |
2025-04-01T04:33:57.792704
|
{
"authors": [
"tmarkovski",
"wellcaffeinated"
],
"repo": "decentralized-identity/did-key.rs",
"url": "https://github.com/decentralized-identity/did-key.rs/issues/26",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1362389493
|
WACI Pex
Adding Issue Credential and Present Proof 1.0, 2.0 and 3.0. The latest version with draft status with links to WACI PEx profile.
lgtm, but need to fix the failing check. Not sure what that is about
lgtm, but need to fix the failing check. Not sure what that is about
GitHub automated actions for rendering. I think @dhh1128 knows how to fix.
|
gharchive/pull-request
| 2022-09-05T19:01:02 |
2025-04-01T04:33:57.795099
|
{
"authors": [
"brianorwhatever",
"rodolfomiranda"
],
"repo": "decentralized-identity/didcomm.org",
"url": "https://github.com/decentralized-identity/didcomm.org/pull/58",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2701100690
|
Docs/ss.md
What does this PR do?
Secret Sharing Scheme 부분 작성하였습니다!
근데 image 출처는 따로 표시하는 방법이 있나요?
아니면 제가 이미지를 따로 만드는게 나을까요?
잘못된 부분이나 좀 더 명확한 내용이 필요한 부분 피드백 주시면 감사하겠습니다!
Home ❯ Privacy Enhancing Technologies (PET) ❯ Secret Sharing 이 부분 하나 만들었습니다!
Before reviewing
[x] Check for missing / redundant content
[x] Grammar Check
[x] Check if the content is accurate
Who can review?
Anyone in the community is free to review the PR once the tests have passed. Feel free to tag
members/contributors who may be interested in your PR.
@yijun-lee @healthykim @Cosmodude
secret sharing 맨 밑에 Index로 Shamir’s Secret Sharing Scheme, Linear Secret Sharing Scheme (LSSS) 분리해서 작성 완료 하였습니다!
해당 부분 내용 추가적으로 더 작성하였고, 수정 필요한 부분 알려주시면 감사하겠습니다!
@yijun-lee
|
gharchive/pull-request
| 2024-11-28T07:21:25 |
2025-04-01T04:33:57.823715
|
{
"authors": [
"hunjyeong"
],
"repo": "decipherhub/Cryptography",
"url": "https://github.com/decipherhub/Cryptography/pull/79",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1558783783
|
gives an error when copying to the sd card
Describe the bug
A clear and concise description of what the bug is.
To Reproduce
Steps to reproduce the behavior:
Go to '...'
Click on '....'
Scroll down to '....'
See error
Expected behavior
A clear and concise description of what you expected to happen.
Screenshots
If applicable, add screenshots or screen recordings to help explain your problem. If they are too private for this public space, feel free to send them by email.
System information and logs:
In the app, there are instructions in the About page > Bug Report section. After following them, paste here your system information and attach your logs.
Additional context
Add any other context about the problem here.
Package: deckers.thibault.aves
Aves version: 1.7.10-play
Aves build: 90
Flutter version: 3.3.10
Flutter channel: stable
Android version: 11
Android API: 30
Android build: RP1A.200720.011
Device: Xiaomi 2201117PG
Geocoder: ready
Mobile services: ready
System locales: ru_RU
Aves locale: system -> ru
Installer: com.android.vending
aves-logs-20230126_230225.txt
Thanks. The logs are weirdly empty :/
Can you write to your SD card with other apps? I'm asking because some SD cards are buggy and read-only.
Could you go to Settings > Privacy > Storage Access and take a screenshot?
Something like this:
If you already have an image or video on your SD card, could you open it in the viewer, then go to its Info page and take a screenshot?
Something like this:
I'm starting to think it's Xiaomi's fault because no file manager can copy files to the SD card except the pre-installed one
Thanks for checking. The format of the SD card volume ID (FE38BD1638BCCF41) looks weird to me. I'm expecting something like 1243-5678. Maybe that's confusing non-Xiaomi apps.
Also, as I've mentioned before, maybe the SD card is not mounted properly, which prevents regular apps from writing to it.
I'll close this issue, as it seems to be unrelated to Aves, but feel free to comment if you have additional insights.
I don't know how, but this gallery manages to copy files from me. Maybe you can find out how they do it.
Is that app open-source? Please share if you have a link to the app source.
Also, if you can please check my question 3. here.
it looks like this gallery is not open source unfortunately
The gallery app that works for you targets API 29 (Android 10), so it doesn't follow the requirements of the Play Store. Soon it will no longer be listed for download to new users.
Targeting older API allows that app to bypass the permission system of modern Android.
why do they complicate everything so much with these permissions that now because of this everything is not working well, what worked fine...
I don't want to give up just yet.
If you're willing to help further, I'll prepare a test build for you to try. I don't know what's wrong, so it will be trial and error... Let me know and I'll work on it.
ok, let's try to solve this problem
Nice, thanks.
In the meantime, could you try what the user described here?
I still don't understand where to go to do the same, Xiaomi is so confused
The App Manager app which that user refers to is not a standard Android or Xiaomi app. It's an open-source app with advanced features and insights for app management. You can install it to try what worked for that user.
I wanted to install this appmanager, but I don't find it in googleplay, and the code that I understand needs to be compiled on GitHub, if possible, give a link to the compiled version or tell me where to look for it on GitHub
You can install it with this apk. I found that one attached as an asset to the latest stable release in the releases page.
I've prepared a build with more detailed logs: here it is.
This will install an app called "Aves [Profile]". It won't interfere with the settings and preferences in the regular app "Aves".
Please do the following:
install the APK, launch it and wait for it to analyze everything,
close and relaunch the app,
try to copy 1 image from your main storage to the SD card (give permission when needed),
go to About > Bug Report and save the logs,
attach the logs here.
I tried appmanager, but my tab is empty
did as you asked
Package: deckers.thibault.aves.profile
Aves version: 1.7.10-play
Aves build: 90
Flutter version: unknown
Flutter channel: unknown
Android version: 11
Android API: 30
Android build: RP1A.200720.011
Device: Xiaomi 2201117PG
Geocoder: ready
Mobile services: ready
System locales: ru_RU
Aves locale: system -> ru
Installer: com.google.android.packageinstaller
aves-logs-20230127_230932.txt
I don't have root access on my phone
I tried appmanager, but my tab is empty
Same for me. Sorry i didn't know that root was required. I just saw this solution from that other user.
Thanks for the logs! There's something interesting, but I can't take a closer look at the moment. I'll try to get back to you in a few days.
Here's another build. It will replace "Aves [Profile]".
Please do the following:
install the APK, launch the app,
try to copy 1 image from your main storage to the SD card,
If it succeeds, please take a close look at the created file on your SD card. Can you open it? Does it have the same size as the original? Does it look ok?
If it still fails, please do the following:
3) go to About > Bug Report and save the logs,
4) attach the logs here.
Also, I have a question: with the normal Aves release, you said it fails when you copy a file to your SD card, but does it also fail when you edit (rotating, tagging, rating, etc.) a file already on your SD card?
I copied the file as you requested. The photo has opened and the dimensions have been preserved and the quality seems to have remained, but it will need to be viewed on a stationary computer. I also tried to make changes to the photo on the sd card on the original application and everything went well.
Ok, good. Let me know if anything is weird. Otherwise, I'll consider the issue fixed.
Thanks again for all the help!
Okay, I'll play around with copying photos, I hope everything will be ok. Is it possible to make a few small changes in your gallery?
While everything is working well, tomorrow I will compare the quality of the copied photo and the original. If you had tweaked a few more things in design and management, it would be super :)
Is it possible to make a few small changes in your gallery?
What do you mean? If you have feature requests, you can open issues, but please read this before.
|
gharchive/issue
| 2023-01-26T21:11:09 |
2025-04-01T04:33:57.849517
|
{
"authors": [
"borivit",
"deckerst"
],
"repo": "deckerst/aves",
"url": "https://github.com/deckerst/aves/issues/506",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
1135576260
|
sonoff Zigbee Stick as a router cannot be paired
I faked the latest router firmware on my Sonoff stick. but I am not able to pair it. is the zigbee stick supported as a router?
This is unrelated to docker. Best to ask in the forum or on discord.
|
gharchive/issue
| 2022-02-13T09:56:51 |
2025-04-01T04:33:57.900307
|
{
"authors": [
"New-three",
"senilio"
],
"repo": "deconz-community/deconz-docker",
"url": "https://github.com/deconz-community/deconz-docker/issues/85",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2763616601
|
🛑 Ristorante La Pagodina Numana is down
In b310d63, Ristorante La Pagodina Numana (https://lapagodina.it) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Ristorante La Pagodina Numana is back up in 341bc5f after 17 minutes.
|
gharchive/issue
| 2024-12-30T19:32:44 |
2025-04-01T04:33:57.918550
|
{
"authors": [
"dedotombo"
],
"repo": "dedotombo/upptime",
"url": "https://github.com/dedotombo/upptime/issues/235",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2031162295
|
Support for microsoft/table-transformer-structure-recognition-v1.1-all
Is there a way to use the enhanced TATR structure recognition (1.1)?
In the HF repo there's no pytorch_model.bin (only model.safetensors).
Enhancement 🚀
Support for microsoft/table-transformer-structure-recognition-v1.1-all
Motivation 💪
It is better at defining the table structure.
Alternatives ⚖️
It's basic version (1.0)
The model itself is the same as the first recognition model.
Differences are training procedure and the fact that padding is not necessary anymore when running in inference mode.
So you can take the original TATR checkpoint, convert the checkpoint using the older script here:
https://github.com/huggingface/transformers/blob/main/src/transformers/models/table_transformer/convert_table_transformer_to_hf.py
to create a .pt checkpoint.
You will then have to add the model to the ModelCatalog as described here.
I have it on my agenda, but this will definitely take some take.
|
gharchive/issue
| 2023-12-07T16:59:57 |
2025-04-01T04:33:57.942656
|
{
"authors": [
"JaMe76",
"ee-lang"
],
"repo": "deepdoctection/deepdoctection",
"url": "https://github.com/deepdoctection/deepdoctection/issues/273",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2724507793
|
The training dataset that used in Object Detection
Could you share the source of your training dataset?
this is our huggingface page: https://huggingface.co/deepghs
|
gharchive/issue
| 2024-12-07T09:37:07 |
2025-04-01T04:33:57.943824
|
{
"authors": [
"ainrichman",
"narugo1992"
],
"repo": "deepghs/imgutils",
"url": "https://github.com/deepghs/imgutils/issues/128",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2327587826
|
start create outside block for drunet
Checks to be done before submitting your PR
[ ] python3 -m pytest tests/ runs successfully.
[ ] black . runs successfully.
[ ] make html runs successfully (in the docs/ directory).
[ ] Updated docstrings related to the changes (as applicable).
[ ] Added an entry to the CHANGELOG.rst.
Closing this for now
|
gharchive/pull-request
| 2024-05-31T11:15:03 |
2025-04-01T04:33:57.950675
|
{
"authors": [
"aregnier-idris",
"matthieutrs"
],
"repo": "deepinv/deepinv",
"url": "https://github.com/deepinv/deepinv/pull/231",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
262615767
|
Windows CPU build in trouble
CUDA 7.5 having issues as well, but that's of course minor:
This is a cpu issue
[Stream 1 Build CPU] In file included from D:/jenkins/workspace/dl4j/all-multiplatform_windows-x86_64/libnd4j/stream1/libnd4j/blas/cpu/NativeOps.cpp:32:0:
[Stream 1 Build CPU] D:/jenkins/workspace/dl4j/all-multiplatform_windows-x86_64/libnd4j/stream1/libnd4j/blas/cpu/NDArray.cpp: In member function 'void nd4j::NDArray<T>::transposei() [with T = double]':
[Stream 1 Build CPU] D:/jenkins/workspace/dl4j/all-multiplatform_windows-x86_64/libnd4j/stream1/libnd4j/blas/cpu/NDArray.cpp:705:5: warning: 'newShapeBuffer' may be used uninitialized in this function [-Wmaybe-uninitialized]
[Stream 1 Build CPU] shape::doPermuteShapeBuffer(newShapeBuffer, rearrange);
[Stream 1 Build CPU] ^~~~~
[Stream 1 Build CPU] D:/jenkins/workspace/dl4j/all-multiplatform_windows-x86_64/libnd4j/stream1/libnd4j/blas/cpu/NDArray.cpp: In member function 'void nd4j::NDArray<T>::transposei() [with T = float16]':
[Stream 1 Build CPU] D:/jenkins/workspace/dl4j/all-multiplatform_windows-x86_64/libnd4j/stream1/libnd4j/blas/cpu/NDArray.cpp:705:5: warning: 'newShapeBuffer' may be used uninitialized in this function [-Wmaybe-uninitialized]
[Stream 1 Build CPU] shape::doPermuteShapeBuffer(newShapeBuffer, rearrange);
[Stream 1 Build CPU] ^~~~~
[Stream 1 Build CPU] D:/jenkins/workspace/dl4j/all-multiplatform_windows-x86_64/libnd4j/stream1/libnd4j/blas/cpu/NDArray.cpp: In member function 'void nd4j::NDArray<T>::transposei() [with T = float]':
[Stream 1 Build CPU] D:/jenkins/workspace/dl4j/all-multiplatform_windows-x86_64/libnd4j/stream1/libnd4j/blas/cpu/NDArray.cpp:705:5: warning: 'newShapeBuffer' may be used uninitialized in this function [-Wmaybe-uninitialized]
[Stream 1 Build CPU] shape::doPermuteShapeBuffer(newShapeBuffer, rearrange);
[Stream 1 Build CPU] ^~~~~
[Stream 1 Build CPU] In file included from D:/jenkins/workspace/dl4j/all-multiplatform_windows-x86_64/libnd4j/stream1/libnd4j/include/ops/declarable/generic/parity_ops.h:21:0,
[Stream 1 Build CPU] from D:/jenkins/workspace/dl4j/all-multiplatform_windows-x86_64/libnd4j/stream1/libnd4j/blas/cpu/NativeOps.cpp:36:
[Stream 1 Build CPU] D:/jenkins/workspace/dl4j/all-multiplatform_windows-x86_64/libnd4j/stream1/libnd4j/include/ops/declarable/generic/convo/convo_ops.h: In member function 'Nd4jStatus nd4j::ops::fullconv3d_bp<T>::validateAndExecute(nd4j::graph::Block<T>&) [with T = double]':
[Stream 1 Build CPU] D:/jenkins/workspace/dl4j/all-multiplatform_windows-x86_64/libnd4j/stream1/libnd4j/include/ops/declarable/generic/convo/convo_ops.h:1398:25: warning: 'gradColumns' may be used uninitialized in this function [-Wmaybe-uninitialized]
[Stream 1 Build CPU] NDArray<T> *gradColumns;
[Stream 1 Build CPU] ^~~~~~~~~~~
[Stream 1 Build CPU] D:/jenkins/workspace/dl4j/all-multiplatform_windows-x86_64/libnd4j/stream1/libnd4j/include/ops/declarable/generic/convo/convo_ops.h: In member function 'Nd4jStatus nd4j::ops::fullconv3d_bp<T>::validateAndExecute(nd4j::graph::Block<T>&) [with T = float]':
[Stream 1 Build CPU] D:/jenkins/workspace/dl4j/all-multiplatform_windows-x86_64/libnd4j/stream1/libnd4j/include/ops/declarable/generic/convo/convo_ops.h:1398:25: warning: 'gradColumns' may be used uninitialized in this function [-Wmaybe-uninitialized]
[Stream 1 Build CPU] D:/jenkins/workspace/dl4j/all-multiplatform_windows-x86_64/libnd4j/stream1/libnd4j/include/ops/declarable/generic/convo/convo_ops.h: In member function 'Nd4jStatus nd4j::ops::fullconv3d_bp<T>::validateAndExecute(nd4j::graph::Block<T>&) [with T = float16]':
[Stream 1 Build CPU] D:/jenkins/workspace/dl4j/all-multiplatform_windows-x86_64/libnd4j/stream1/libnd4j/include/ops/declarable/generic/convo/convo_ops.h:1398:25: warning: 'gradColumns' may be used uninitialized in this function [-Wmaybe-uninitialized]
[Stream 1 Build CPU] D:\jenkins_temp\ccQz3sEe.s: Assembler messages:
[Stream 1 Build CPU] D:\jenkins_temp\ccQz3sEe.s:1623362: Error: invalid register for .seh_savexmm
[Stream 1 Build CPU] make[2]: *** [blas/CMakeFiles/nd4jcpu.dir/build.make:66: blas/CMakeFiles/nd4jcpu.dir/cpu/NativeOps.cpp.obj] Error 1
[Stream 1 Build CPU] make[2]: Leaving directory '/d/jenkins/workspace/dl4j/all-multiplatform_windows-x86_64/libnd4j/stream1/libnd4j/blasbuild/cpu'
[Stream 1 Build CPU] make[1]: *** [CMakeFiles/Makefile2:89: blas/CMakeFiles/nd4jcpu.dir/all] Error 2
[Stream 1 Build CPU] make[1]: Leaving directory '/d/jenkins/workspace/dl4j/all-multiplatform_windows-x86_64/libnd4j/stream1/libnd4j/blasbuild/cpu'
[Stream 1 Build CPU] make: *** [Makefile:153: all] Error 2
Full log:
http://master-jenkins.skymind.io:8080/job/dl4j/job/all-multiplatform/2726/consoleFull
The Cuda 7.5 snippet:
[edit: removed to avoid confusion]
[Stream 1 Build CPU] D:\jenkins_temp\ccQz3sEe.s:1623362: Error: invalid register for .seh_savexmm
Please update gcc on msys.
We're dropping support for CUDA 7.5, so don't worry about that one :)
@saudet please read the beginning of the issue, it's about a cpu failure
Closed by https://github.com/deeplearning4j/pipelines/issues/14 https://github.com/deeplearning4j/pipelines/issues/12
|
gharchive/issue
| 2017-10-03T22:51:45 |
2025-04-01T04:33:57.977872
|
{
"authors": [
"huitseeker",
"raver119",
"saudet"
],
"repo": "deeplearning4j/libnd4j",
"url": "https://github.com/deeplearning4j/libnd4j/issues/491",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
744373036
|
Add OpenSpiel interface
This pull request is a follow up to https://github.com/deepmind/acme/issues/80 and adds the ability to run Acme’s DQN on OpenSpiel games. All changes and additions are contained in the acme/open_spiel directory. The example run_dqn.py runs Acme’s DQN on OpenSpiel’s tic-tac-toe. It’s set up to print out a full playthrough (policies included) every 1000 episodes making it easy to verify that the agent is in fact learning a good policy. The agents start playing sensibly after roughly 10k episodes, and by about 50k episodes pretty much every game in which an exploratory action is not taken results in a draw.
Though I’ve endeavored to keep the code clean and production ready, this PR is only intended to be a starting point. There are a number of TODOs. Some are only meant to call attention to important considerations, others are tasks which seemed better addressed after there’s been a chance for feedback and discussion.
Perhaps the most important design feature, and the one that stands to improve most from further thought, is the environment wrapper. In an effort to keep as much existing structure intact as possible, the wrapper is built around OpenSpiel’s python rl_environement - which notably is not a dm_env. An alternative approach is to wrap the OpenSpiel game directly as a dm_env. However, the dm_env doesn’t appear to have been explicitly designed with multiplayer games in mind. Perhaps someone who has worked on dm_env, or the related (and very useful) tree repo might have some insight into how best to do this.
I look forward to feedback and I’ll be happy to implement any suggested changes. You’ll notice that very few modifications to the DQN agent/learner were necessary. With this interface it should be straightforward to adapt other Acme algorithms to run on OpenSpiel games :-)
And to import OpenSpiel, add these two lines to Acme's virtual environment.
export PYTHONPATH=$PYTHONPATH:/<path_to_open_spiel>
export PYTHONPATH=$PYTHONPATH:/<path_to_open_spiel>/build/python
Take this all with a grain of salt, because I'm thinking out loud a little bit at the moment, but:
I don't know how much of a pain it would be to retroactively mark some of these files as copies (i.e. git cp). I'm thinking mostly with respect to the learner, but as I think this might be painful, could you highlight some of the changes made here just to center the discussion?
But from looking I think the main changes in the agent/actor/learner (please correct me if I'm wrong!) revolve around the legal actions. I've noted this above, and I wonder if we can push these directly as observations and let the network handle the legal actions. I think this will allow the learner to remain unchanged. Ideally the same could be done for actor/agent (putting aside the changes to epsilon greedy that are probably needed). But if this only really helps the learner that'd be great.
Ideally we could have something where we have an openspiel-specific Environment Loop, wrapper, and potentially a "Q-network wrapper" but the agent itself isn't changed (and we can peel back as much of that as is necessary if/when this ideal isn't possible 😄).
Ahh I see, yeah moving the legal actions mask into the observation and letting the network handle everything sounds like the right approach. Should streamline things quite a bit.
Tried to mark the new files as copies but couldn’t get the git diff to work, it kept marking everything as new additions. Either way I think you already identified the core issue. I’ll implement those changes with the goal of confining everything to the network, loop and wrapper. Thanks for the quick feedback!
Pushed a big update that's pretty close the ideal @mwhoffman described. Moving legal actions into the observation really was the way to go, simplified everything dramatically. All that’s needed now to run the OpenSpiel example is the wrapper, environment loop, custom networks, and the slightest of modifications to the DQN agent.
The issue with the DQN agent is the policy network. Currently the DQN agent creates the policy network by appending trfl’s epsilon_greedy() layer to the learning network. As far as I can tell, no matter what our custom network outputs this layer will select illegal actions. My workaround was to add the policy network as an optional argument. It offers more flexibility while not changing the default behavior of the DQN agent.
Now that OpenSpiel’s footprint is much smaller, I moved it out of its own directory and placed the files in the appropriate folders. The wrapper is < 150 lines, env loop < 250, and the networks are pretty simple, so this update gets us a lot closer to the finish line.
Also there's an issue with pytest failing, looks like it isn't installing openspiel.
@fastturtle thanks for the comments! Regarding pytest failing, OpenSpiel does not currently have a pip package for direct installation (though it's in the works). Given that, any advice on how to handle this? And tagging @lanctot, who knows all things about installing OpenSpiel :)
@fastturtle @jhtschultz I am planning for an OpenSpiel release in a few weeks, and have it in PyPi, but only by source distribution for now (which means it'll be 1-2 min install, and requires the machines have CMake, clang++, and python3-dev). Do you think that would solve it?
Hi! Wanted to check in on the status of this PR.
I saw commit https://github.com/deepmind/acme/commit/996c63db160a3daec376ec8fe5ed3b1b3a3d16c1 followed the approach taken here and added the ability to provide DQN's policy network directly. That's great because now this PR requires zero changes to Acme’s DQN.
Commit https://github.com/deepmind/acme/commit/996c63db160a3daec376ec8fe5ed3b1b3a3d16c1 also added NetworkWithMaskedEpsilonGreedy, which uses trfl.epsilon_greedy(). As documented here https://github.com/deepmind/trfl/issues/27 and https://github.com/deepmind/trfl/pull/28, there is a bug in trfl’s action masking which currently renders this network unusable. Once fixed, I could refactor the networks used in this PR to use NetworkWithMaskedEpsilonGreedy instead.
I’ve been implementing NFSP, a deep-RL algorithm for imperfect information games, using this Acme-OpenSpiel interface and the interface works quite well. If we can identify any remaining obstacles to importing this PR I'll be happy to work on them. Thanks!
Hi! Wanted to check in on the status of this PR.
I saw commit https://github.com/deepmind/acme/commit/996c63db160a3daec376ec8fe5ed3b1b3a3d16c1 followed the approach taken here and added the ability to provide DQN's policy network directly. That's great because now this PR requires zero changes to Acme’s DQN.
Commit https://github.com/deepmind/acme/commit/996c63db160a3daec376ec8fe5ed3b1b3a3d16c1 also added NetworkWithMaskedEpsilonGreedy, which uses trfl.epsilon_greedy(). As documented here https://github.com/deepmind/trfl/issues/27 and https://github.com/deepmind/trfl/pull/28, there is a bug in trfl’s action masking which currently renders this network unusable. Once fixed, I could refactor the networks used in this PR to use NetworkWithMaskedEpsilonGreedy instead.
I’ve been implementing NFSP, a deep-RL algorithm for imperfect information games, using this Acme-OpenSpiel interface and the interface works quite well. If we can identify any remaining obstacles to importing this PR I'll be happy to work on them. Thanks!
@fastturtle thanks for the comments! Regarding pytest failing, OpenSpiel does not currently have a pip package for direct installation (though it's in the works). Given that, any advice on how to handle this? And tagging @lanctot, who knows all things about installing OpenSpiel :)
I spoke to @fastturtle yesterday. OpenSpiel does have a pip now but it still requires a build from source, which requires specific things from your environment and could take a long time. So, until we have a binary wheel (at least a few months), we've agreed just disable tests that require/depend on OpenSpiel.
Please keep the code for these tests around, though. When it's merged, I will enable this test internally to ensure that it passes on our side. Thanks!
@fastturtle thanks for the comments! Regarding pytest failing, OpenSpiel does not currently have a pip package for direct installation (though it's in the works). Given that, any advice on how to handle this? And tagging @lanctot, who knows all things about installing OpenSpiel :)
I spoke to @fastturtle yesterday. OpenSpiel does have a pip now but it still requires a build from source, which requires specific things from your environment and could take a long time. So, until we have a binary wheel (at least a few months), we've agreed just disable tests that require/depend on OpenSpiel.
Please keep the code for these tests around, though. When it's merged, I will enable this test internally to ensure that it passes on our side. Thanks!
Alright all checks are passing! And I added two simple tests for the wrapper and environment loop. Following the approach used for other Acme tests, they are skipped when they run into an import error, so they don’t need to be disabled/reenabled @lanctot.
Small detail: I tried to follow the pattern used throughout Acme with __init__.py imports, but couldn’t avoid pytest failing if OpenSpiel imports were included in __init__.py. For the time being they’re commented out. Maybe there’s some other work around, perhaps it’s best to delete them altogether.
Alright all checks are passing! And I added two simple tests for the wrapper and environment loop. Following the approach used for other Acme tests, they are skipped when they run into an import error, so they don’t need to be disabled/reenabled @lanctot.
Small detail: I tried to follow the pattern used throughout Acme with __init__.py imports, but couldn’t avoid pytest failing if OpenSpiel imports were included in __init__.py. For the time being they’re commented out. Maybe there’s some other work around, perhaps it’s best to delete them altogether.
@fastturtle thanks for looking it over! Most of the changes are already in. I still need to do some further digging to fully explain the -np.inf threshold causing the priority error. Good call on your part to highlight that. I'll address the comments individually and let you know when everything is ready to go on my end.
@fastturtle thanks for looking it over! Most of the changes are already in. I still need to do some further digging to fully explain the -np.inf threshold causing the priority error. Good call on your part to highlight that. I'll address the comments individually and let you know when everything is ready to go on my end.
@jhtschultz sounds good, I'll keep an eye out for your comments.
@jhtschultz sounds good, I'll keep an eye out for your comments.
@fastturtle Done! Not sure why some pytests get stuck. Recent commits, including the previous one https://github.com/deepmind/acme/pull/87/commits/ff9b7addd5ee83c333f5c0725672673446f01477, have been consistently passing so I think everything is good to go. Let me know if there's anything else I can do.
@fastturtle Done! Not sure why some pytests get stuck. Recent commits, including the previous one https://github.com/deepmind/acme/pull/87/commits/ff9b7addd5ee83c333f5c0725672673446f01477, have been consistently passing so I think everything is good to go. Let me know if there's anything else I can do.
We exposed _use_observation as a property in OpenSpiel's rl_environment, and the wrapper here has been updated to use the property.
We exposed _use_observation as a property in OpenSpiel's rl_environment, and the wrapper here has been updated to use the property.
Thanks! Good catches, fixed 👍
Thanks! Good catches, fixed 👍
A few minor changes were made (mostly linter reasons!) and this was submitted in https://github.com/deepmind/acme/commit/7fa9b804e47fd1b1fbb20ed5e62275d822607c8b. So I'll close this down for now, and any further discussion can carry on outside this PR. Thanks for all the hard work!
|
gharchive/pull-request
| 2020-11-17T03:16:02 |
2025-04-01T04:33:58.001722
|
{
"authors": [
"fastturtle",
"jhtschultz",
"lanctot",
"mwhoffman"
],
"repo": "deepmind/acme",
"url": "https://github.com/deepmind/acme/pull/87",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1393861996
|
Import error when upgrading to jax 0.3.18
Hi,
jax version = 0.3.18
I am faced with the following error when importing dm_pix
I suspect it's due to the removal of ._src since jax 0.3.18 https://github.com/google/jax/releases/tag/jax-v0.3.18
----> [1](vscode-notebook-cell:/Users/asem/serket/test.ipynb#Y121sZmlsZQ%3D%3D?line=0) import dm_pix
File ~/miniforge3/envs/dev-jax15/lib/python3.10/site-packages/dm_pix/__init__.py:16, in <module>
1 # Copyright 2020 DeepMind Technologies Limited. All Rights Reserved.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
(...)
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14 """PIX public APIs."""
---> 16 from dm_pix._src import augment
17 from dm_pix._src import color_conversion
18 from dm_pix._src import depth_and_space
File ~/miniforge3/envs/dev-jax15/lib/python3.10/site-packages/dm_pix/_src/augment.py:25, in <module>
22 import functools
23 from typing import Callable, Sequence, Tuple, Union
---> 25 import chex
26 from dm_pix._src import color_conversion
27 from dm_pix._src import interpolation
File ~/miniforge3/envs/dev-jax15/lib/python3.10/site-packages/chex/__init__.py:17, in <module>
1 # Copyright 2020 DeepMind Technologies Limited. All Rights Reserved.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
(...)
13 # limitations under the License.
14 # ==============================================================================
15 """Chex: Testing made fun, in JAX!"""
---> 17 from chex._src.asserts import assert_axis_dimension
18 from chex._src.asserts import assert_axis_dimension_comparator
19 from chex._src.asserts import assert_axis_dimension_gt
File ~/miniforge3/envs/dev-jax15/lib/python3.10/site-packages/chex/_src/asserts.py:26, in <module>
23 import unittest
24 from unittest import mock
---> 26 from chex._src import asserts_internal as _ai
27 from chex._src import pytypes
28 import jax
File ~/miniforge3/envs/dev-jax15/lib/python3.10/site-packages/chex/_src/asserts_internal.py:32, in <module>
29 from typing import Any, Sequence, Union, Callable, Optional, Set, Tuple, Type
31 from absl import logging
---> 32 from chex._src import pytypes
33 import jax
34 import jax.numpy as jnp
File ~/miniforge3/envs/dev-jax15/lib/python3.10/site-packages/chex/_src/pytypes.py:44, in <module>
40 Device = jax.lib.xla_extension.Device
42 ArrayTree = Union[Array, Iterable['ArrayTree'], Mapping[Any, 'ArrayTree']]
---> 44 ArrayDType = jax._src.numpy.lax_numpy._ScalarMeta
AttributeError: module 'jax' has no attribute '_src'
Hi @ASEM000, seems like there's a little hiccup with the Jax version and Chex there. I'll ask if a new version of Chex is going to be released soon to fix this issue with Jax.
If you have pip installed pix it might be that Jax has been installe dtransitively by Chex. The best thing to do is actually to install Jax befrehand (and maybe even with CUDA) and then install everything else. Anyway, thanks for flagging!
Hi @ASEM000, the fix is already integrated into chex v0.1.5.
You might need to update your chex version to fix the error that you're getting.
|
gharchive/issue
| 2022-10-02T18:46:49 |
2025-04-01T04:33:58.006496
|
{
"authors": [
"ASEM000",
"claudiofantacci",
"hbq1"
],
"repo": "deepmind/dm_pix",
"url": "https://github.com/deepmind/dm_pix/issues/49",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
432026191
|
Create terraform lib in python for Kadet
Following #191 create a python lib for terraform's most common operations (in the same spirit of kube.libjsonnet) using the new kadet/python input type.
@uberspot I was reading over and interested in how would this look like and the use case. Can you share some info?
The idea behind it was to create a jsonnet library with most common resources that could be created with terraform but abstracted in a function. So that people can re-use them with ease.
But after thinking about this for a while, it might just make more sense to add more jsonnet terraform examples instead since that's what people would use.
I can add some example templates for google cloud resources by the end of this week and see if it helps. :)
|
gharchive/issue
| 2019-04-11T13:32:47 |
2025-04-01T04:33:58.008595
|
{
"authors": [
"Cyanopus",
"uberspot"
],
"repo": "deepmind/kapitan",
"url": "https://github.com/deepmind/kapitan/issues/271",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1783398384
|
Custom decorative objects
Hi,
I'm using mujoco for a project that moves a cuboid to the target area. I use a thin cylinder object as the receptacle/target area indicator. But I found no matter how thin the cylinder is, the cuboid will be stuck near the edge of the cylinder. My question is: is there any way to add custom decorative objects, which are not considered by the physics engine of mujoco, to my MJCF? If not, could you suggest possible solutions to my problem?
Yes, of course. Run the tutorial colab, this is explained in the last cell.
Thanks. And there is a more direct solution: set contype="2" conaffinity="2" in MJCF to avoid collision
Thanks. I found there is a more direct solution: set contype="2" conaffinity="2" in MJCF to avoid collision
This isn't correct. You should set both to zero.
Or just use sites instead of geoms. Sites are designed for this.
https://mujoco.readthedocs.io/en/latest/overview.html#kinematic-tree
Thanks, both using site and setting cons to zero work.
|
gharchive/issue
| 2023-07-01T03:01:37 |
2025-04-01T04:33:58.011903
|
{
"authors": [
"saran-t",
"yuvaltassa",
"zxwang96"
],
"repo": "deepmind/mujoco",
"url": "https://github.com/deepmind/mujoco/issues/962",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1868678335
|
Modify mjd_inverseFD for discrete inverse dynamics
Modifications to mjd_inverseFD to compute correct Jacobians when discrete inverse dynamics are used.
When mj_inverse is called with discrete inverse dynamics enabled, qacc is corrected to a continuous time representation. As a result, we need to reset qacc to its discrete representation prior calling mj_inverse.
Does it not make more sense to do this in mj_inverse?
I believe this is solved by f44c6575444fc393d74ba937ba432c51f55bba2e ?
|
gharchive/pull-request
| 2023-08-27T23:53:44 |
2025-04-01T04:33:58.013691
|
{
"authors": [
"thowell",
"yuvaltassa"
],
"repo": "deepmind/mujoco",
"url": "https://github.com/deepmind/mujoco/pull/1034",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
816640513
|
OpenSpiel 0.3.0 is now available!
For details see, https://github.com/deepmind/open_spiel/releases/tag/v0.3.0.
I mistakenly forgot to include the optional dependencies, so universal_poker and hanabi are not available from colab. Will fix in a few hours. Reopening https://github.com/deepmind/open_spiel/issues/512
Yay! Thanks everyone involved! And especially @lanctot for the precious care :)
Looking forward to Python games. I might use OpenSpiel in an AI intro course at the uni and make some example game like tic-tac-toe in python for the students :)
I mistakenly forgot to include the optional dependencies, so universal_poker and hanabi are not available from colab. Will fix in a few hours. Reopening #189
Ok pip package is now at version 0.3.1 and has hanabi and universal_poker back in them. Will issue another release, should fix colab.
... did I say 0.3.0? I meant 0.3.1 :)
https://github.com/deepmind/open_spiel/releases/tag/v0.3.1
|
gharchive/issue
| 2021-02-25T17:24:45 |
2025-04-01T04:33:58.017360
|
{
"authors": [
"lanctot",
"michalsustr"
],
"repo": "deepmind/open_spiel",
"url": "https://github.com/deepmind/open_spiel/issues/512",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1848628984
|
optax and tensorflow's Adam optimizer's setting.
Currently, optax.scale_by_adam should be equivalent to torch.optim.Adam. However, Tensorflow has a different implementation.
In short, if we change https://github.com/deepmind/optax/blob/cebdeff4a1922113a96c520e7a81b5bf79825b77/optax/_src/transform.py#L345-L348 to the following, then the adam optimizer would be the same as tensorflow's imlementation.
updates = jax.tree_util.tree_map(
lambda m, v: (jnp.sqrt(1- b2**count_inc) / (1-b1**count_inc)) * m / (jnp.sqrt(v + eps_root) + eps), mu, nu)
More context
Basically, PyTorch and optax's adam follow Algorithm 1 of the Kingma and Ba’s Adam paper (arxiv/1412.6980), but TensorFlow uses the formulation just before Section 2.1 of the paper and its epsilon referred to here is epsilon hat in the paper.
This was a relevant issue in my recent reproduction of openai's work in https://github.com/openai/lm-human-preferences. Long story short, below is an end-to-end experiment with torch's adam adam_pt and tensorlfow-style adam adam_tf. While the final performance (objective/scores) look the same, the learning curves are different in a non-trivial way. E.g., the torch adam version had a much higher clipfrac initially, causing a more initial significant update.
The "initial aggressive update" issue gets aggravated in larger models (e.g., gpt2-large). You can see that objective/kl had a spike with adam_tf, so this could be a reproducibility issue.
Desired solution
include a
def scale_by_adam_tfstyle(
b1: float = 0.9,
b2: float = 0.999,
eps: float = 1e-8,
eps_root: float = 0.0,
mu_dtype: Optional[chex.ArrayDType] = None,
) -> base.GradientTransformation:
mu_dtype = utils.canonicalize_dtype(mu_dtype)
def init_fn(params):
mu = jax.tree_util.tree_map( # First moment
lambda t: jnp.zeros_like(t, dtype=mu_dtype), params)
nu = jax.tree_util.tree_map(jnp.zeros_like, params) # Second moment
return ScaleByAdamState(count=jnp.zeros([], jnp.int32), mu=mu, nu=nu)
def update_fn(updates, state, params=None):
del params
mu = update_moment(updates, state.mu, b1, 1)
nu = update_moment_per_elem_norm(updates, state.nu, b2, 2)
count_inc = numerics.safe_int32_increment(state.count)
updates = jax.tree_util.tree_map(
lambda m, v: (jnp.sqrt(1- b2**count_inc) / (1-b1**count_inc)) * m / (jnp.sqrt(v + eps_root) + eps), mu, nu) #
mu = utils.cast_tree(mu, mu_dtype)
return updates, ScaleByAdamState(count=count_inc, mu=mu, nu=nu)
return base.GradientTransformation(init_fn, update_fn)
obviously this is bad naming, but I figure you'd have much better ideas :)
Thanks for raising this!
Rather than including an alternative adam version in optax, I would instead suggest we add a warning in the doc-string of the adam and scale_by_adam gradient transformations. Once they are aware, it's easy for people who need to reproduce old results to fork and modify the transformation.
E.g. we could add:
WARNING: PyTorch and optax's adam follow Algorithm 1 of the Kingma and Ba’s Adam paper, if reproducing old results note that TensorFlow used instead the formulation just before Section 2.1 of the paper.
Can you put together a short PR with this fix?
@mtthss with pleasure. Thanks for the suggestion and see https://github.com/deepmind/optax/pull/572.
P.S. big fan of your work and it made my day seeing your response :D
Hi @vwxyzjn @mtthss ,
Thanks Costa @vwxyzjn for the observation regarding Optax vs. Tensorflow Adam and for bringing this issue to my attention.
While the final performance (objective/scores) look the same, the learning curves are different in a non-trivial way. E.g., the torch adam version had a much higher clipfrac initially, causing a more initial significant update.
PyTorch and optax's adam follow Algorithm 1 of the Kingma and Ba’s Adam paper (arxiv/1412.6980), but TensorFlow uses the formulation just before Section 2.1 of the paper and its epsilon referred to here is epsilon hat in the paper.
Here are derivations that verify Costa's observations mentioned above. Following the notation of Kingma and Ba’s Adam paper, let's compare the update equations implemented by Optax and Tensorflow. We have the gradient update rules for optax/pytorch adam (Algorithm 1 of Kingma and Ba’s paper) and tensorflow adam (the formulation just before Section 2.1 of Kingma and Ba’s paper) as below.
\begin{equation}
\begin{aligned}\text{optax/pytorch adam :}\quad
\theta_t & =\theta_{t-1}-\alpha \cdot \hat{m}_t /\left(\sqrt{\hat{v}_t}+\varepsilon\right) \\
& =\theta_{t-1}- \alpha \underbrace{\left[m_t /\left(1-\beta_1^t\right)\right]}_{=\hat{m}_t} /\left[\sqrt{\underbrace{v_t /\left(1-\beta_2^t\right)}_{=\hat{v}_t} }+\varepsilon\right]\\
& =\theta_{t-1}- \alpha\left[m_t /\left(1-\beta_1^t\right)\right]\frac{\sqrt{1-\beta_2^t}}{\sqrt{v_t}+\color{blue}{\varepsilon \sqrt{1-\beta_2^t}}}
\end{aligned}
\end{equation}
\begin{equation} \begin{aligned}\text{tensorflow adam:}\quad \theta_t & =\theta_{t-1}-\alpha_t m_t /\left(\sqrt{v_t}+\hat{\varepsilon}\right) \\& =\theta_{t-1}-\underbrace{\left[\alpha \sqrt{1-\beta_2^t} /\left(1-\beta_1^t\right)\right]}_{=\alpha_t} m_t /\left(\sqrt{v_t}+\hat{\varepsilon}\right) \\& =\theta_{t-1}- \alpha\left[m_t /\left(1-\beta_1^t\right)\right] \frac{\sqrt{1-\beta_2^t}}{\sqrt{v_t}+\color{blue}{\hat{\varepsilon}}} \end{aligned}\end{equation}
The equations above highlight that the distinction between optax and tensorflow implementation is their normalization terms, $\color{blue}{\varepsilon \sqrt{1-\beta_2^t}}$ and $\color{blue}{\hat{\varepsilon}}$. The two versions are equivalent if we set $\hat{\varepsilon} =\varepsilon \sqrt{1-\beta_2^t}$ . However, in the optax and tensorflow APIs, we can only set $\varepsilon$ (optax and pytorch) and $\hat{\varepsilon}$ (tensorflow) via the eps argument, causing differences in their update equations.
So, what if we set $\varepsilon$ and $\hat{\varepsilon}$ to the same value, say, 1e-5? Then for tensorflow adam, the normalization term $\hat{\varepsilon} = \text{1e-5}$ is just a constant. But for optax/pytorch adam, the normalization term ${\varepsilon \sqrt{1-\beta_2^t}}$ changes over time. Importantly, the normalization term ${\varepsilon \sqrt{1-\beta_2^t}}$ is initially much smaller than 1e-5 when the timestep $t$ is small, and gradually approaches to 1e-5 as timesteps increase. The plot below compares these two normalization terms over timesteps:
The above figure shows that, if we set the same eps in optax adam and tensorflow adam, then optax/pytorch adam uses a much smaller normalization term than tensorflow-adam in the early phase of training. In other words, optax/pytorch adam goes for more aggressive gradient updates early in the training. This aligns nicely with Costa's empirical results shared earlier!
Tianlin
|
gharchive/issue
| 2023-08-13T15:04:00 |
2025-04-01T04:33:58.029282
|
{
"authors": [
"liutianlin0121",
"mtthss",
"vwxyzjn"
],
"repo": "deepmind/optax",
"url": "https://github.com/deepmind/optax/issues/571",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2009711406
|
tokenizer.model
Hey!
I'd like to work on implementing exllama support, but the tokenizer.model is missing.
https://discord.com/channels/1169871344037548062/1170016413428228168/1171019891839615077
https://discord.com/channels/1169871344037548062/1170016413428228168/1171040772229967952
A couple of weeks ago I asked for this in the discord, I was told that you are working on that
I would highly appreciate if you published the tokenizer.model :)
Thank you for your interest in DeepSeek Coder. Currently we can not provide a sentencepiece model (tokenizer.model) that achieves the same output as the existing HuggingFace Tokenizer. We are working hard to help the model quantization community directly support Huggingface's Tokenizer, and updates on our progress will be timely synced to https://github.com/deepseek-ai/DeepSeek-Coder#7-qa.
Thank you for your fast reply.
For exllamaV2, we are also working on it.
So, would it totally suffice if Exllamav2 also supported the HF tokenzier? Is the deepseek model architecture equivalent to Llama only except for the tokenizer?
Yes, we use the same architecture as Llama except for the tokenizer. We attempted to convert the HuggingFace Tokenizer into a sentencepiece model (tokenizer.model), but couldn't achieve the same output. Therefore, we have to try to enable Exllamav2 to support the HF tokenizer. The current issue is that Exllamav2 uses a feature (ignore special token) within the sentencepiece model, which the HF tokenizer doesn't have. We are actively working on a solution.
That's awesome, thank you for the work!!!
@SinanAkkoyun Hi, we just updated our readme about exllamav2, you could try out the forked version which supports huggingface tokenizer and we'd like to hear from your feedback.
Hi! Thank you so much, I will test your fork asap
@BingxuanWang I just tried to run the fork and the models only output nonsense
I even followed the exact prompt formatting, I am using GPTQ quants of TheBloke of the 6.7B and 1.3B model
python examples/chat.py -m../../models/deepseek/deepseek-coder-1.3b-instruct-GPTQ/ -mode deepseek
-- Model: ../../models/deepseek/deepseek-coder-1.3b-instruct-GPTQ/
-- Options: ['rope_scale 1.0', 'rope_alpha 1.0']
-- Loading model...
-- Loading tokenizer...
-- Prompt format: deepseek
-- System prompt:
You are an AI programming assistant, utilizing the Deepseek Coder model, developed by Deepseek Company, and you only answer questions related to computer science. For politically sensitive questions, security and privacy issues, and other non-computer science questions, you will refuse to answer.
User: How can I tell the time in python?
���ııiiiii iii iv i i ii io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io io
6.7B shows very similar behaviour, but often returns some kind of invisible token in an infinite loop
Turboderp told me that RoPe scale 4 is needed, now the model outputs english sentences. The 6.7B model works fine, the 1.3B quickly runs into repetition loops, but that seems to be a matter of quantization bitrate rather than pure inference
Please add the the chat_formats from here to the PR:
https://github.com/SinanAkkoyun/exllama-v2/tree/deepseek
Thank you for the work!
Wanted to check in on this - trying to compile for llama.cpp but missing tokenizer.model is hamstringing that effort.
@leedrake5 Please refer to https://github.com/deepseek-ai/DeepSeek-Coder#7-qa for the tokenizer.model issue
@BingxuanWang That worked perfectly, thanks!
|
gharchive/issue
| 2023-11-24T13:14:03 |
2025-04-01T04:33:58.083056
|
{
"authors": [
"BingxuanWang",
"SinanAkkoyun",
"guoday",
"leedrake5"
],
"repo": "deepseek-ai/DeepSeek-Coder",
"url": "https://github.com/deepseek-ai/DeepSeek-Coder/issues/50",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2289860131
|
using chat_with_website pipeline
After deploying to the hayhooks server the pipeline chat_with_website, I'm trying to use it.
Starting from your example for test_pipeline_01.yml I'm trying
curl -X 'POST' \ 'http://localhost:1416/chat_with_website' \ -H 'accept: application/json' \ -H 'Content-Type: application/json' \ -d '{ "converter": { "value": "what is haystack?" } }'
since converter is the name of the first element of the pipeline and th input is an HTML document (a simple string in my case)
this leads to the error 422 Unprocessable Entity
{"detail":[{"type":"missing","loc":["body","fetcher"],"msg":"Field required","input":{"converter":{"value":"what is haystack?"}}},{"type":"missing","loc":["body","llm"],"msg":"Field required","input":{"converter":{"value":"what is haystack?"}}},{"type":"missing","loc":["body","prompt"],"msg":"Field required","input":{"converter":{"value":"what is haystack?"}}}]}
What am I doing wrong?
Thanks.
With this pipeline (taken from Haystack's predefined templates and stripped from the {%raw%} blocks):
components:
converter:
init_parameters:
extractor_type: DefaultExtractor
type: haystack.components.converters.html.HTMLToDocument
fetcher:
init_parameters:
raise_on_failure: true
retry_attempts: 2
timeout: 3
user_agents:
- haystack/LinkContentFetcher/2.0.0b8
type: haystack.components.fetchers.link_content.LinkContentFetcher
llm:
init_parameters:
api_base_url: null
api_key:
env_vars:
- OPENAI_API_KEY
strict: true
type: env_var
generation_kwargs: {}
model: gpt-3.5-turbo
streaming_callback: null
system_prompt: null
type: haystack.components.generators.openai.OpenAIGenerator
prompt:
init_parameters:
template: |
"According to the contents of this website:
{% for document in documents %}
{{document.content}}
{% endfor %}
Answer the given question: {{query}}
Answer:
"
type: haystack.components.builders.prompt_builder.PromptBuilder
connections:
- receiver: converter.sources
sender: fetcher.streams
- receiver: prompt.documents
sender: converter.documents
- receiver: llm.prompt
sender: prompt.prompt
metadata: {}
This is the curl command I used:
curl -X 'POST' \
'http://localhost:1416/chat_with_website' \
-H 'accept: application/json' \
-H 'Content-Type: application/json' \
-d '{
"converter": {
"meta": {}
},
"fetcher": {
"urls": [
"https://haystack.deepset.ai"
]
},
"llm": {
"generation_kwargs": {}
},
"prompt": {
"query": "What is Haystack?"
}
}'
@masci thanks for the explanation.
|
gharchive/issue
| 2024-05-10T14:35:13 |
2025-04-01T04:33:58.087513
|
{
"authors": [
"masci",
"ssword62"
],
"repo": "deepset-ai/hayhooks",
"url": "https://github.com/deepset-ai/hayhooks/issues/15",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2486585606
|
feat: add NVIDIA NIM ranker support
Proposed Changes:
NVIDIA NIMs (hosted on https://build.nvidia.com and downloadable) support reranking models. This adds an NvidiaRanker component to access those models.
How did you test it?
[x] unit tests
[x] integration tests of hosted NIM
[x] integration tests of downloaded NIM
Notes for the reviewer
should the input documents be mutated to add a score?
the ranker endpoints are not standardized, so i introduced an endpoint arg to rank(). an alternative is to refactor all components to pass the full inference endpoint to the backend.
Checklist
I have read the contributors guidelines and the code of conduct
I have updated the related issue with new insights and changes
I added unit tests and updated the docstrings
I've used one of the conventional commit types for my PR title: fix:, feat:, build:, chore:, ci:, docs:, style:, refactor:, perf:, test:.
@anakin87 can you also have a look on this PR?
@mattf FYI - we avoid force-pushing to a branch once a review's been published. Otherwise, the review comments lose their anchors and it's difficult for the reviewer to diff changes since their previous review.
@mattf FYI - we avoid force-pushing to a branch once a review's been published. Otherwise, the review comments lose their anchors and it's difficult for the reviewer to diff changes since their previous review.
understood. it was one of the options the github presented to catch up to main when the pr was out of date. it may be possible to disable this in the project settings.
|
gharchive/pull-request
| 2024-08-26T11:03:41 |
2025-04-01T04:33:58.094036
|
{
"authors": [
"davidsbatista",
"mattf",
"shadeMe"
],
"repo": "deepset-ai/haystack-core-integrations",
"url": "https://github.com/deepset-ai/haystack-core-integrations/pull/1023",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1092699809
|
fix: update gitopia-mirror-action version to 0.3.0
https://github.com/deepsourcelabs/good-first-issue/runs/4564937725?check_suite_focus=true is failing due to :
`gitopia/gitopia-mirror-action@v0.1.2`, unable to find version `v0.1.2`
We Should Update this to: https://github.com/deepsourcelabs/good-first-issue/blob/master/.github/workflows/gitopia-mirror.yml#L20
v0.3.0
ℹ️ Repository information
The repository has:
[ ] At least three issues with the good first issue label.
[x] Detailed setup instructions for the project.
[x] CONTRIBUTING.md
[x] Actively maintained.
I'm not sure if the action is actively maintained. I couldn't find a release with that tag on the repository. We should remove the action entirely.
Yeah, Likewise.
It is good if @jai-deepsource can point out the thing, what was the cause of adding gitopia mirror. If it is all good now. We can remove this gitopia-mirror workflow
Hi. This is the old unmaintained version of the Gitopia mirror action which worked with our MVP. So, you can disable it for now. We will be releasing the new version soon.
@Siddhant-K-code Can you remove the Gitopia mirror action in this PR?
@Siddhant-K-code Can you remove the Gitopia mirror action in this PR?
Created #373 for that, since I deleted the previous fork, so was unable to update this PR, Sorry about that.
|
gharchive/pull-request
| 2022-01-03T17:49:10 |
2025-04-01T04:33:58.136416
|
{
"authors": [
"Siddhant-K-code",
"faza",
"jai-deepsource",
"sanket-deepsource"
],
"repo": "deepsourcelabs/good-first-issue",
"url": "https://github.com/deepsourcelabs/good-first-issue/pull/366",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
132935505
|
How to run this example?
Any idea on how to run this repo and its setup? Databases, config, etc..
I can't find it anywhere..
Please help. Thank you.
Oh dear. I'm afraid that was demo code for the participants of a workshop we gave a while ago. You can't really run this without a broker giving you direct access to their pricing api
@WolframHempel oww... okay, too bad I missed that workshop
Thanks for ur quick response =)
|
gharchive/issue
| 2016-02-11T10:27:11 |
2025-04-01T04:33:58.138626
|
{
"authors": [
"WolframHempel",
"incubus8"
],
"repo": "deepstreamIO/deepstream-rethinkdb-stocks-example",
"url": "https://github.com/deepstreamIO/deepstream-rethinkdb-stocks-example/issues/1",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
234128018
|
Best practices with single authoritative client on cluster
We have deployed a cluster of deepstream nodes, and, beside it, a cluster of our own servers where we have various types of authoritative deepstream clients that run on backend (matchmaker, gamemaster, whatnot).
However, logic entails that only a single of those backend clients would run at any time, others being marked inactive and activating in case of, say, node failure.
For instance, there's always a single active matchmaker in entire infrastructure, and there's only single active gamemaster per active game.
We're not sure what would be idiomatic way to run this on deepstream abstraction concepts. Any hints?
Hey,
It depends greatly on what the backend services do. We use micro-services and load distribution quite alot which means you don't have to run a single instance usually.
However, that being said, there are multiple other solutions depending on your desired output.
As a high level feature offer to solve this:
1- active providers to only allow one backend process to satisfy commands
2- presence to listen to other backend processes logged in, basically using it as a lock. You can then use deepstreams apis to dynamically provide/unprovide things like RPC endpoints.
There are a couple other solutions, but the two above are the patterns I see/recommend using most often
@synthmeat I hope this answers your question, if you have any follow up feel free to reopen this issue.
|
gharchive/issue
| 2017-06-07T08:17:18 |
2025-04-01T04:33:58.142188
|
{
"authors": [
"synthmeat",
"yasserf"
],
"repo": "deepstreamIO/deepstream.io",
"url": "https://github.com/deepstreamIO/deepstream.io/issues/714",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
250521949
|
Ability to change permissions.yml at runtime
Besides writing your own custom permission handler is it possible to have deepstream monitor permissions.yml for changes and reload permissions at runtime ?
Hey,
We have an internal API for allowing permissions to be reloaded during runtime, but have not yet exposed it. The reason behind that is because its hard to find the connections the permissions may effect for existing subscriptions. So while making rpcs, emitting events and updating or deleting records would immediately reflect the changes, all connections prior will either need to be force closed (very expensive) or revalidated (less expensive but very bug prone).
Happy to hear any suggestions on how we can potentially find a solution for this that is performant!
Cheers,
Yasser
I had a quick look at the config-permission-handler. Would something like this work
Still loading permissions as normal with the permissions.yml file, (as most permissions are fairly static do not need to be changed at runtime). Additionally I would have a deepstream list of other permissions objects in the same format as what comes out of configCompiler.compile. I would listen to that list for changes and when it changes add/remove/replace the changed permission in the _config object in the permission handler, and clear the permission cache. Not sure about what to do about existing subscriptions except publish an event to let the client know it has changed and should resubscribe. Anyway I understand the issues now around changing permissions run time.
@yasserf To solve your problem, have you tried using "smart" subscription objects?
When registering a subscription, not only store information about the subscribed user or the record id, etc, but also information about the version of the permissioning rule that was used to permit the subscription.
When the subscription means that the user now is going to receive a message, first check if the permit version is still up-to-date or not. If it is, everything carries on normally. Else, re-evaluate the permission using the new permissioning rule. If the user endures the change and is still permitted, then just update the version field recorded in the subscription to the newest one just used. Else, notify the user that they have lost their access and therefor the subscription and that their data is no longer guaranteed to be up to date -- and of course delete the subscription.
While that does make alot of sense, it sounds like quite a bit of work too, specially since it isn't isolated in a separate part of the application.
It also doesn't solve the issue with cross referencing, so if a subscription permission depends on the state of another record then we need to have a concept of watches which is quite expensive to maintain during runtime.
So currently valve doesn't recompute a cross reference after it changes?
Not for subscriptions no, everything else yes, because they are more crud like
@yasserf This might be a stupid question, but why can't valve be implemented using deepstream itself? It seems like a perfect use case to me and you are already optimizing deepstream for speed all the time, and, best of all, now this becomes a separate part that is much easier to manage and extend.
Its certainly not a stupid question, valve uses deepstream under the hood for record cross referencing. It's just quite a tough issue to get right (running computed rules), requires alot of work on client and server side and is not high on our priority list in comparison to many other features we want to get in. Hopefully one day though!
After all, all valve does is respond to events, perform RPCs and react to data changes.
Valve doesn't react to data changes, rules are only run functionally and do not contain state outside of a very small execution time.
I am back with a proposed solution, @yasserf. So, recently I've been playing with the following setup in Node.js for my deepstream server (conceptual pseudo code):
let main_server = new DeepstreamServer({port: 6020, allowRemoteConnections: true});
let helper_server = new DeepstreamServer({port: 6021, allowRemoteConnections: false});
helper_server.start();
helper_server.client = new DeepstreamClient({port: 6021});
await helper_server.login();
main_server.set('permissionHandler', {
isReady: true,
canPerformAction: function(username, message, callback, authData, revoke) {
let msg = decode_message(message);
if (msg.type === 'subscribe to record data') {
let permission_record_name = 'permission/'+msg.record.type //eg: 'permission/user'
let rule = await helper_server.client.getRecord(permission_record_name).whenReady;
let user = await helper_server.client.getRecord('user/'+username).whenReady;
let result = calculate_permission(rule, user, data); //data is data being read.
if (result) {
callback(null, true); //grant permission;
rule.subscribe((new_rule) => {
if (!calculate_permission(new_rule, user, data)) {
revoke();
rule.discard();
}
});
} else callback(null, false); //reject permission.
}
}
});
main_server.start();
Now you have fully reactive permissioning rules as deepstream records! This opens the door for so many exciting possibilities. My favorites are:
reusable permissioning rules. (eg: a "userIsAdmin" rule);
rule composition and referencing.
data write validation.
frequency limiting (how many reads and writes per unit of time is allowed).
Note that revoke is an artificial parameter, crucial to the process here, but not really implemented in deepstream. However, in this demonstrative case of subscribing to a record's data, the implementation should be as simple as emulating a client-side unsubscibe on the server (perhaps also with a courtesy notification to the client that their permission has expired). Notice also that calculate_permission depends on the data being read, but no subscription to that data is actually performed so that calculate_permission is called again after data changes, and then maybe the permission needs to be revoked.
This is just scratching the surface. Deepstream is powerful enough to implement any dynamic permissioning system based on permission records, permission events and RPCs. However, for a full fledged permissioning system (dare I say, valve 2.0?), the permission handler needs to be able to revoke permissions via a revoke callback, as well as being able to screen data before it is sent to the user so that the same record holding a user's profile picture and authentication password is screened and filtered appropriately for different users with different permissions. Perhaps GraphQL might be of use at this place.
I am already in the midst of creating my own permissioning system based on data types and sophisticated permissioning schemas, reusable rules, rules composition, data validation, and more. I might decide to opensource it and work on this together, but for now I am stuck at being not able to revoke or screen the data before it is sent to the client.
We will be implementing refreshing permissions soon using an admin api, but revoking individual subscriptions is still quite hard to do. Totally doable, but I don't think its worth the cost and maintenance vs killing the connection. Specially now with bulk subscriptions.
|
gharchive/issue
| 2017-08-16T06:17:26 |
2025-04-01T04:33:58.153184
|
{
"authors": [
"RashadSaleh",
"grendo",
"yasserf"
],
"repo": "deepstreamIO/deepstream.io",
"url": "https://github.com/deepstreamIO/deepstream.io/issues/787",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
272104385
|
Refresh Golden-Layout size at resize of a DIV container
I have Golden-Layout inside a DIV that can change its size INDEPENDENTLY by window resize.
If GL would size itself according to DIV container... I would not need anything but it's sized in px and listens for window resize not generic container resize... How to solve?
One way could be force refresh of GL size at a given event (not ideal because I have to manually force it but... anyway could be workaround). Any idea?
Anyone could help me?
Hello,
I am facing the same issue, using GL with angular.
The div is updated with angular content but GL is not updating its size accordingly.
When I trigger some re-sizing event (docking tabs for instance), GL size itself correctly.
Could anyone help for this ?
Thanks
|
gharchive/issue
| 2017-11-08T07:56:47 |
2025-04-01T04:33:58.156426
|
{
"authors": [
"adoukkali",
"ldetomi"
],
"repo": "deepstreamIO/golden-layout",
"url": "https://github.com/deepstreamIO/golden-layout/issues/371",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
604800237
|
[Bug] Found conflicts! when installing v1.5.1 over using Conda for Win 10 x64
Description
Installation fails for Deezer Spleeter v1.5.1 Conda-Forge environment over a clean Conda installation on Microsoft Windows 10 x64. Found conflicts in several packages.
Step to reproduce
From Anaconda prompt run as administrator:
conda config --add channels conda-forge
conda install -c conda-forge spleeter
Output
(base) PS C:\Deezer\spleeter> conda install -c conda-forge spleeter
Collecting package metadata (current_repodata.json): done
Solving environment: failed with initial frozen solve. Retrying with flexible solve.
Solving environment: failed with repodata from current_repodata.json, will retry with next repodata source.
Collecting package metadata (repodata.json): done
Solving environment: failed with initial frozen solve. Retrying with flexible solve.
Solving environment: |
Found conflicts! Looking for incompatible packages.
This can take several minutes. Press CTRL-C to abort.
[Several minutes of 'Examining'... Please see full log in post below]
UnsatisfiableError: The following specifications were found
to be incompatible with the existing python installation in your environment:
Specifications:
anaconda-navigator -> python[version='2.7.|3.5.|3.6.|3.4.']
anaconda-project -> python[version='3.4.*|>=3.8,<3.9.0a0']
anaconda==2020.02 -> python[version='3.6.10|3.8.1|3.7.6',build='h60c2a47_2|h5fd99cc_8_cpython|h9f7ef89_0']
autopep8 -> python[version='>=3.8,<3.9.0a0']
backports.functools_lru_cache -> python[version='2.7.|3.5.|3.6.|3.4.|>=2.7,<2.8.0a0']
bcrypt -> python[version='2.7.|3.5.|3.6.|3.6.9|3.4.|3.7.|3.8.',build='2_73_pypy|1_73_pypy|0_73_pypy']
bkcharts -> python=3.4
bleach -> python[version='>=3.8,<3.9.0a0']
cffi -> python[version='2.7.|3.5.|3.6.|3.6.9|3.6.9|>=2.7,<2.8.0a0|>=3.8,<3.9.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|3.4.',build='2_73_pypy|1_73_pypy|0_73_pypy']
conda-env -> python[version='2.7.|3.4.|3.5.*']
cryptography -> python[version='3.4.*|3.6.9|<=3.3',build='2_73_pypy|1_73_pypy|0_73_pypy']
curl -> python=2.7
diff-match-patch -> python[version='2.7.|3.4.|3.5.|3.6.']
gevent -> python[version='3.6.9|3.7.|3.8.',build='2_73_pypy|1_73_pypy|0_73_pypy']
h5py -> python[version='<3']
html5lib -> python[version='3.4.*|>=3.8,<3.9.0a0']
intervaltree -> python[version='2.7.|3.5.|3.6.*']
jinja2 -> python[version='>=3.8,<3.9.0a0']
joblib -> python[version='>=3.8,<3.9.0a0']
jupyterlab -> python[version='>=3.8,<3.9.0a0']
jupyterlab_server -> python[version='2.7.|3.5.|3.6.|>=2.7,<2.8.0a0|>=3.5,<3.6.0a0|>=3.8,<3.9.0a0|3.4.']
krb5 -> python=2.7
libcurl -> python=2.7
matplotlib -> python[version='2.7.|3.6.|<3']
mkl-service -> python[version='2.7.|3.5.|3.6.|3.4.']
more-itertools -> python=3.4
navigator-updater -> python[version='2.7.|3.5.|3.6.|>=3.8,<3.9.0a0|3.4.']
networkx -> python[version='>=3.8,<3.9.0a0']
numba -> python=3.4
openpyxl -> python[version='>=3.8,<3.9.0a0']
packaging -> python[version='>=3.8,<3.9.0a0']
partd -> python[version='>=3.8,<3.9.0a0']
path -> python[version='>=2.7,<2.8.0a0']
path.py -> python[version='>=3.8,<3.9.0a0']
pathtools -> python[version='2.7.|3.4.|3.5.|3.6.']
prometheus_client -> python[version='>=3.8,<3.9.0a0']
pygments -> python[version='>=3.8,<3.9.0a0']
pynacl -> python[version='2.7.|3.5.|3.6.|3.6.9|3.4.',build='2_73_pypy|1_73_pypy|0_73_pypy']
pytest-arraydiff -> python[version='2.7.|3.5.|3.6.|3.4.']
pytest-doctestplus -> python[version='2.7.|3.5.|3.6.|>=3.8,<3.9.0a0|3.4.']
pytest-openfiles -> python[version='2.7.|3.5.|3.6.|>=3.8,<3.9.0a0|3.4.']
pytest-remotedata -> python[version='2.7.|3.5.|3.6.|3.4.']
python-dateutil -> python[version='>=3.8,<3.9.0a0']
pywinpty -> python=3.4
qtawesome -> python[version='>=3.8,<3.9.0a0']
scikit-learn -> python[version='2.7.|3.5.|3.6.*']
scipy -> python[version='3.5.|3.6.']
seaborn -> python[version='>=3.8,<3.9.0a0']
sortedcollections -> python[version='>=3.8,<3.9.0a0']
sphinx -> python[version='>=3.8,<3.9.0a0']
sphinxcontrib-websupport -> python[version='>=3.8,<3.9.0a0']
testpath -> python[version='3.4.*|>=3.8,<3.9.0a0']
watchdog -> python[version='>=3.5,<3.6.0a0']
zict -> python[version='>=3.8,<3.9.0a0']
Your python: python=3.7
If python is on the left-most side of the chain, that's the version you've asked for.
When python appears to the right, that indicates that the thing on the left is somehow
not available for the python version you are constrained to. Note that conda will not
change your python version to a different minor version unless you explicitly specify
that.
The following specifications were found to be incompatible with each other:
Output in format: Requested package -> Available versions
Package sqlite conflicts for:
[BIG chuck of data like 'jsonschema -> python[version='>=2.7,<2.8.0a0'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.27.2,<4.0a0|>=3.30.1,<4.0a0|>=3.31.1,<4.0a0|>=3.29.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0']'. Please see full log in post below]
Package ecdsa conflicts for:
paramiko -> ecdsa[version='>=0.11,<2.0']
spyder -> paramiko -> ecdsa[version='>=0.11,<2.0']
Package icu conflicts for:
qt -> icu[version='>=58.1,<59.0a0|>=64.2,<65.0a0|>=58.2,<59.0a0']
anaconda==2020.02 -> icu==58.2=ha66f8fd_1
pyqt -> qt[version='>=5.12.5,<5.13.0a0'] -> icu[version='>=58.1,<59.0a0|>=64.2,<65.0a0|>=58.2,<59.0a0']
anaconda==2020.02 -> qt==5.9.7=vc14h73c81de_0 -> icu[version='>=58.2,<59.0a0']
Package autopep8 conflicts for:
python-language-server -> autopep8
spyder -> python-language-server[version='>=0.31.9,<0.32.0'] -> autopep8
anaconda==2020.02 -> python-language-server==0.31.7=py36_0 -> autopep8
anaconda==2020.02 -> autopep8==1.4.4=py_0
Package terminado conflicts for:
widgetsnbextension -> notebook[version='>=4.4.1'] -> terminado[version='>=0.8.1']
anaconda==2020.02 -> notebook==6.0.3=py36_0 -> terminado[version='>=0.8.1']
jupyterlab -> notebook[version='>=4.3.1'] -> terminado[version='>=0.8.1']
notebook -> terminado[version='>=0.8.1']
jupyterlab_server -> notebook[version='>=4.2.0'] -> terminado[version='>=0.8.1']
jupyter -> notebook -> terminado[version='>=0.8.1']
ipywidgets -> notebook -> terminado[version='>=0.8.1']
anaconda==2020.02 -> terminado==0.8.3[build='py38_0|py36_0|py37_0']
Package m2w64-gcc-libs-core conflicts for:
pywinpty -> m2w64-gcc-libs -> m2w64-gcc-libs-core
anaconda==2020.02 -> m2w64-gcc-libgfortran==5.3.0=6 -> m2w64-gcc-libs-core
blas -> m2w64-gcc-libs -> m2w64-gcc-libs-core
m2w64-gcc-libs -> m2w64-gcc-libs-core
anaconda==2020.02 -> m2w64-gcc-libs-core==5.3.0=7
m2w64-gcc-libgfortran -> m2w64-gcc-libs-core
Package pathtools conflicts for:
watchdog -> pathtools[version='>=0.1.1']
anaconda==2020.02 -> pathtools==0.1.2=py_1
anaconda==2020.02 -> watchdog==0.10.2=py36_0 -> pathtools[version='>=0.1.1']
spyder -> watchdog -> pathtools[version='>=0.1.1']
Package scikit-learn conflicts for:
anaconda==2020.02 -> scikit-learn==0.22.1[build='py36h6288b17_0|py38h6288b17_0|py37h6288b17_0']
spleeter -> librosa -> scikit-learn[version='>=0.14.0|>=0.14.0,!=0.19.0|>=0.19.1']
nltk -> scikit-learn
Package lazy-object-proxy conflicts for:
anaconda==2020.02 -> astroid==2.3.3=py36_0 -> lazy-object-proxy
astroid -> lazy-object-proxy=1.4
pylint -> astroid[version='>=2.3.0,<2.4'] -> lazy-object-proxy=1.4
anaconda==2020.02 -> lazy-object-proxy==1.4.3[build='py37he774522_0|py38he774522_0|py36he774522_0']
Package patsy conflicts for:
seaborn -> statsmodels[version='>=0.8.0'] -> patsy[version='>=0.4.0|>=0.5.1']
anaconda==2020.02 -> patsy==0.5.1[build='py38_0|py36_0|py37_0']
statsmodels -> patsy[version='>=0.4.0|>=0.5.1']
anaconda==2020.02 -> statsmodels==0.11.0=py36he774522_0 -> patsy[version='>=0.5.1']
seaborn -> patsy
Package backports.shutil_which conflicts for:
pywinpty -> backports.shutil_which
terminado -> pywinpty -> backports.shutil_which
Package alabaster conflicts for:
sphinx -> alabaster[version='>=0.7,<0.8']
anaconda==2020.02 -> sphinx==2.4.0=py_0 -> alabaster[version='>=0.7,<0.8']
anaconda==2020.02 -> alabaster==0.7.12[build='py_0|py36_0|py37_0']
numpydoc -> sphinx -> alabaster[version='>=0.7,<0.8']
spyder -> sphinx[version='>=0.6.6'] -> alabaster[version='>=0.7,<0.8']
Package jupyter_console conflicts for:
anaconda==2020.02 -> jupyter==1.0.0=py36_7 -> jupyter_console
jupyter -> jupyter_console
anaconda==2020.02 -> jupyter_console==6.1.0=py_0
Package lz4 conflicts for:
blosc -> zstd[version='>=1.3.7,<1.3.8.0a0'] -> lz4
libarchive -> zstd[version='>=1.3.7,<1.3.8.0a0'] -> lz4
libtiff -> zstd[version='>=1.3.7,<1.3.8.0a0'] -> lz4
zstd -> lz4
Package libxslt conflicts for:
anaconda==2020.02 -> lxml==4.5.0=py36h1350720_0 -> libxslt[version='>=1.1.33,<2.0a0']
anaconda==2020.02 -> libxslt==1.1.33=h579f668_0
lxml -> libxslt[version='>=1.1.32,<2.0a0|>=1.1.33,<2.0a0']
Package notebook conflicts for:
_ipyw_jlab_nb_ext_conf -> ipywidgets -> notebook[version='>=4.2.0|>=4.2|>=4.3|>=4.3.1|>=4.4.1']
jupyterlab -> notebook[version='>=4.2|>=4.3|>=4.3.1']
widgetsnbextension -> notebook[version='>=4.2.0|>=4.4.1']
anaconda==2020.02 -> notebook==6.0.3[build='py38_0|py36_0|py37_0']
jupyter -> notebook
jupyterlab -> jupyterlab_server[version='>=1.0.0,<2.0.0'] -> notebook[version='>=4.2.0']
ipywidgets -> notebook
ipywidgets -> widgetsnbextension[version='>=3.5.0,<3.6.0'] -> notebook[version='>=4.2.0|>=4.4.1']
jupyterlab_server -> notebook[version='>=4.2.0']
anaconda==2020.02 -> jupyter==1.0.0=py36_7 -> notebook[version='>=4.2.0|>=4.3.1|>=4.4.1']
Package statsmodels conflicts for:
anaconda==2020.02 -> statsmodels==0.11.0[build='py37he774522_0|py38he774522_0|py36he774522_0']
seaborn -> statsmodels[version='>=0.5.0|>=0.8.0']
anaconda==2020.02 -> seaborn==0.10.0=py_0 -> statsmodels[version='>=0.8.0']
Package pycparser conflicts for:
gevent -> cffi[version='>=1.11.5'] -> pycparser
pynacl -> cffi[version='>=1.4.1'] -> pycparser
bcrypt -> cffi[version='>=1.1'] -> pycparser
anaconda==2020.02 -> pycparser==2.19[build='py_0|py36_0|py37_0']
cffi -> pycparser
anaconda==2020.02 -> cffi==1.14.0=py36h7a1dbc1_0 -> pycparser
cryptography -> cffi[version='>=1.8,!=1.11.3'] -> pycparser
Package winpty conflicts for:
terminado -> pywinpty -> winpty[version='>=0.4.3,<1.0a0']
pywinpty -> winpty[version='>=0.4.3,<1.0a0']
anaconda==2020.02 -> winpty==0.4.3=4
anaconda==2020.02 -> pywinpty==0.5.7=py36_0 -> winpty
Package jdcal conflicts for:
anaconda==2020.02 -> jdcal==1.4.1=py_0
anaconda==2020.02 -> openpyxl==3.0.3=py_0 -> jdcal
openpyxl -> jdcal
Package intel-openmp conflicts for:
numpy-base -> mkl[version='>=2019.4,<2021.0a0'] -> intel-openmp
numexpr -> mkl[version='>=2019.4,<2021.0a0'] -> intel-openmp
scikit-learn -> mkl[version='>=2019.4,<2021.0a0'] -> intel-openmp
scipy -> mkl[version='>=2019.4,<2020.0a0'] -> intel-openmp
mkl_random -> mkl[version='>=2020.0,<2021.0a0'] -> intel-openmp
mkl-service -> mkl[version='>=2019.4,<2020.0a0'] -> intel-openmp
anaconda==2020.02 -> intel-openmp==2020.0=166
blas -> mkl -> intel-openmp
mkl_fft -> mkl[version='>=2019.4,<2020.0a0'] -> intel-openmp
mkl -> intel-openmp
numpy -> mkl[version='>=2019.4,<2021.0a0'] -> intel-openmp
anaconda==2020.02 -> mkl==2020.0=166 -> intel-openmp
Package snappy conflicts for:
blosc -> snappy[version='>=1.1.7,<2.0a0']
pytables -> blosc[version='>=1.16.3,<2.0a0'] -> snappy[version='>=1.1.7,<2.0a0']
anaconda==2020.02 -> blosc==1.16.3=h7bd577a_0 -> snappy[version='>=1.1.7,<2.0a0']
anaconda==2020.02 -> snappy==1.1.7=h777316e_3
Package prompt-toolkit conflicts for:
ipython -> prompt-toolkit[version='!=3.0.0,!=3.0.1,<3.1.0,>=2.0.0']
prompt_toolkit -> prompt-toolkit[version='>=3.0.4,<3.0.5.0a0|>=3.0.5,<3.0.6.0a0']
jupyter_console -> ipython -> prompt-toolkit[version='!=3.0.0,!=3.0.1,<3.1.0,>=2.0.0|>=3.0.5,<3.0.6.0a0|>=3.0.4,<3.0.5.0a0']
ipykernel -> ipython[version='>=5.0'] -> prompt-toolkit[version='!=3.0.0,!=3.0.1,<3.1.0,>=2.0.0']
ipywidgets -> ipython[version='>=4.0.0'] -> prompt-toolkit[version='!=3.0.0,!=3.0.1,<3.1.0,>=2.0.0']
ipython -> prompt_toolkit[version='>=2.0.0,<4,!=3.0.0,!=3.0.1'] -> prompt-toolkit[version='>=3.0.4,<3.0.5.0a0|>=3.0.5,<3.0.6.0a0']
Package bzip2 conflicts for:
conda-package-handling -> libarchive[version='>=3.3.3'] -> bzip2[version='>=1.0.6,<2.0a0|>=1.0.8,<2.0a0']
anaconda==2020.02 -> bzip2==1.0.8=he774522_0
python-libarchive-c -> libarchive -> bzip2[version='>=1.0.6,<2.0a0|>=1.0.8,<2.0a0']
libarchive -> bzip2[version='>=1.0.6,<2.0a0|>=1.0.8,<2.0a0']
conda-build -> libarchive -> bzip2[version='>=1.0.6,<2.0a0|>=1.0.8,<2.0a0']
anaconda==2020.02 -> libarchive==3.3.3=h0643e63_5 -> bzip2[version='>=1.0.6,<2.0a0|>=1.0.8,<2.0a0']
pytables -> bzip2[version='1.0.*|>=1.0.6,<2.0a0|>=1.0.8,<2.0a0']
Package python-simplegeneric conflicts for:
ipykernel -> ipython[version='>=4.0'] -> python-simplegeneric[version='>0.8']
ipython -> python-simplegeneric[version='>0.8']
ipywidgets -> ipython[version='>=4.0.0'] -> python-simplegeneric[version='>0.8']
jupyter_console -> ipython -> python-simplegeneric[version='>0.8']
Package wheel conflicts for:
python=3.7 -> pip -> wheel
anaconda==2020.02 -> wheel==0.34.2[build='py38_0|py36_0|py37_0']
pip -> wheel
bcrypt -> pip -> wheel
anaconda==2020.02 -> pip==20.0.2=py36_1 -> wheel
Package zipp conflicts for:
anaconda==2020.02 -> zipp==2.2.0=py_0
pytest -> importlib_metadata[version='>=0.12'] -> zipp[version='>=0.5']
jsonschema -> importlib_metadata -> zipp[version='>=0.3.2|>=0.5']
anaconda==2020.02 -> importlib_metadata==1.5.0=py36_0 -> zipp[version='>=0.5']
keyring -> importlib_metadata -> zipp[version='>=0.3.2|>=0.5']
importlib_metadata -> zipp[version='>=0.3.2|>=0.5']
path -> importlib_metadata[version='>=0.5'] -> zipp[version='>=0.3.2|>=0.5']
path.py -> importlib_metadata[version='>=0.5'] -> zipp[version='>=0.3.2|>=0.5']
pluggy -> importlib_metadata[version='>=0.12'] -> zipp[version='>=0.3.2|>=0.5']
spleeter -> importlib_resources -> zipp[version='>=0.4']
Package json5 conflicts for:
jupyterlab_server -> json5
anaconda==2020.02 -> jupyterlab_server==1.0.6=py_0 -> json5
anaconda==2020.02 -> json5==0.9.1=py_0
jupyterlab -> jupyterlab_server[version='>=1.0.0,<2.0.0'] -> json5
Package argh conflicts for:
spyder -> watchdog -> argh[version='>=0.24.1']
watchdog -> argh[version='>=0.24.1']
anaconda==2020.02 -> argh==0.26.2[build='py38_0|py36_0|py37_0']
anaconda==2020.02 -> watchdog==0.10.2=py36_0 -> argh[version='>=0.24.1']
Package cryptography-vectors conflicts for:
paramiko -> cryptography[version='>=1.5'] -> cryptography-vectors[version='2.3.|2.3.1.']
urllib3 -> cryptography[version='>=1.3.4'] -> cryptography-vectors[version='2.3.|2.3.1.']
pyopenssl -> cryptography[version='>=2.2.1'] -> cryptography-vectors[version='2.3.|2.3.1.']
cryptography -> cryptography-vectors[version='2.3.|2.3.1.']
Package joblib conflicts for:
scikit-learn -> joblib[version='>=0.11']
nltk -> scikit-learn -> joblib[version='>=0.11']
spleeter -> librosa -> joblib[version='>=0.12.0|>=0.7.0']
anaconda==2020.02 -> joblib==0.14.1=py_0
anaconda==2020.02 -> scikit-learn==0.22.1=py36h6288b17_0 -> joblib[version='>=0.11']
Package beautifulsoup4 conflicts for:
anaconda==2020.02 -> beautifulsoup4==4.8.2[build='py38_0|py36_0|py37_0']
conda-build -> beautifulsoup4
Package gmpy2 conflicts for:
mpmath -> gmpy2
sympy -> mpmath[version='>=0.19'] -> gmpy2
Package paramiko conflicts for:
spyder -> paramiko[version='>=2.4.0']
anaconda==2020.02 -> spyder==4.0.1=py36_0 -> paramiko
anaconda==2020.02 -> paramiko==2.7.1=py_0
Package jaraco.itertools conflicts for:
zipp -> jaraco.itertools
importlib_metadata -> zipp[version='>=0.5'] -> jaraco.itertools
Package widgetsnbextension conflicts for:
ipywidgets -> widgetsnbextension[version='>=1.2.2|>=1.2.3|>=1.2.6|>=1.2.6,<2|>=2.0.0|>=2.0.0,<3.0.0|>=3.0.0|>=3.0.0,<3.1.0|>=3.1.0,<3.2.0|>=3.2.0,<3.3.0|>=3.3.0,<3.4.0|>=3.4.0,<3.5.0|>=3.5.0,<3.6.0|>=3.2.0,<4.0.0|>=3.1.0,<4.0.0|>=3.0.0,<4.0.0|>=3.1.0,<4.0']
jupyter -> ipywidgets -> widgetsnbextension[version='>=1.2.2|>=1.2.3|>=1.2.6|>=1.2.6,<2|>=2.0.0|>=2.0.0,<3.0.0|>=3.0.0|>=3.0.0,<3.1.0|>=3.1.0,<3.2.0|>=3.2.0,<3.3.0|>=3.3.0,<3.4.0|>=3.4.0,<3.5.0|>=3.5.0,<3.6.0|>=3.2.0,<4.0.0|>=3.1.0,<4.0.0|>=3.0.0,<4.0.0|>=3.1.0,<4.0']
_ipyw_jlab_nb_ext_conf -> widgetsnbextension
anaconda==2020.02 -> ipywidgets==7.5.1=py_0 -> widgetsnbextension[version='>=3.5.0,<3.6.0']
_ipyw_jlab_nb_ext_conf -> ipywidgets -> widgetsnbextension[version='>=1.2.2|>=1.2.3|>=1.2.6|>=1.2.6,<2|>=2.0.0|>=2.0.0,<3.0.0|>=3.0.0|>=3.0.0,<3.1.0|>=3.1.0,<3.2.0|>=3.2.0,<3.3.0|>=3.3.0,<3.4.0|>=3.4.0,<3.5.0|>=3.5.0,<3.6.0|>=3.2.0,<4.0.0|>=3.1.0,<4.0.0|>=3.0.0,<4.0.0|>=3.1.0,<4.0']
anaconda==2020.02 -> widgetsnbextension==3.5.1[build='py38_0|py36_0|py37_0']
Package mpmath conflicts for:
sympy -> mpmath[version='>=0.19']
anaconda==2020.02 -> mpmath==1.1.0[build='py38_0|py36_0|py37_0']
anaconda==2020.02 -> sympy==1.5.1=py36_0 -> mpmath[version='>=0.19']
Package boto3 conflicts for:
dask -> distributed[version='>=1.16.0'] -> boto3
distributed -> boto3
Package qtpy conflicts for:
qtconsole -> qtpy
jupyter -> qtconsole -> qtpy
qtawesome -> qtpy
anaconda-navigator -> qtpy[version='>=1.2|>=1.4.1']
spyder -> qtpy[version='>=1.1|>=1.2.0|>=1.5.0']
spyder -> qtawesome[version='>=0.5.7'] -> qtpy
anaconda==2020.02 -> qtawesome==0.6.1=py_0 -> qtpy[version='>=1.5.0']
navigator-updater -> qtpy[version='>=1.2']
anaconda==2020.02 -> qtpy==1.9.0=py_0
Package qtconsole conflicts for:
anaconda==2020.02 -> qtconsole==4.6.0=py_1
jupyter -> qtconsole
anaconda==2020.02 -> jupyter==1.0.0=py36_7 -> qtconsole[version='>=4.6.0']
spyder -> qtconsole[version='>=4.2|>=4.6.0']
Package rope conflicts for:
anaconda==2020.02 -> rope==0.16.0=py_0
anaconda==2020.02 -> python-language-server==0.31.7=py36_0 -> rope[version='>=0.10.5']
spyder -> rope[version='0.9.4|>=0.10.5|>=0.9.4']
python-language-server -> rope[version='>=0.10.5']
Package cycler conflicts for:
bokeh -> matplotlib -> cycler[version='>=0.10']
matplotlib-base -> cycler[version='>=0.10']
anaconda==2020.02 -> matplotlib-base==3.1.3=py36h64f37c6_0 -> cycler[version='>=0.10']
anaconda==2020.02 -> cycler==0.10.0[build='py36h009560c_0|py38_0|py37_0']
scikit-image -> matplotlib-base[version='>=2.0.0'] -> cycler[version='>=0.10']
matplotlib -> cycler[version='>=0.10']
seaborn -> matplotlib-base[version='>=2.1.2'] -> cycler[version='>=0.10']
nltk -> matplotlib -> cycler[version='>=0.10']
Package pysocks conflicts for:
anaconda==2020.02 -> pysocks==1.7.1[build='py38_0|py36_0|py37_0']
urllib3 -> pysocks[version='>=1.5.6,<2.0,!=1.5.7']
anaconda==2020.02 -> urllib3==1.25.8=py36_0 -> pysocks[version='>=1.5.6,<2.0,!=1.5.7']
requests -> urllib3[version='>=1.21.1,<1.26,!=1.25.0,!=1.25.1'] -> pysocks[version='>=1.5.6,<2.0,!=1.5.7']
Package typing_extensions conflicts for:
distributed -> bokeh[version='>=0.12.3'] -> typing_extensions[version='>=3.7.4']
bokeh -> typing_extensions[version='>=3.7.4']
dask -> bokeh[version='>=1.0.0,!=2.0.0'] -> typing_extensions[version='>=3.7.4']
Package pyflakes conflicts for:
spyder -> pyflakes
spyder -> python-language-server[version='>=0.31.9,<0.32.0'] -> pyflakes[version='>=1.6.0|>=1.6.0,<2.2.0']
flake8 -> pyflakes[version='>=1.2.3,<1.3|>=1.2.3,<1.4|>=1.5.0,<1.6.0|>=1.5.0,<1.7.0|>=2.0.0,<2.1.0|>=2.1.0,<2.2.0']
anaconda==2020.02 -> flake8==3.7.9=py36_0 -> pyflakes[version='>=1.6.0|>=2.1.0,<2.2.0']
python-language-server -> flake8 -> pyflakes[version='>=1.2.3,<1.3|>=1.2.3,<1.4|>=1.5.0,<1.6.0|>=1.5.0,<1.7.0|>=2.0.0,<2.1.0|>=2.1.0,<2.2.0']
anaconda==2020.02 -> pyflakes==2.1.1[build='py38_0|py36_0|py37_0']
python-language-server -> pyflakes[version='>=1.6.0|>=1.6.0,<2.2.0']
Package sphinxcontrib-jsmath conflicts for:
anaconda==2020.02 -> sphinxcontrib-jsmath==1.0.1=py_0
anaconda==2020.02 -> sphinx==2.4.0=py_0 -> sphinxcontrib-jsmath
sphinx -> sphinxcontrib-jsmath
spyder -> sphinx[version='>=0.6.6'] -> sphinxcontrib-jsmath
numpydoc -> sphinx -> sphinxcontrib-jsmath
Package keyring conflicts for:
spyder -> keyring
anaconda==2020.02 -> keyring==21.1.0[build='py38_0|py36_0|py37_0']
anaconda==2020.02 -> spyder==4.0.1=py36_0 -> keyring
Package pexpect conflicts for:
spyder -> pexpect[version='>=4.4.0']
anaconda==2020.02 -> pexpect==4.8.0[build='py38_0|py36_0|py37_0']
anaconda==2020.02 -> spyder==4.0.1=py36_0 -> pexpect
Package curl conflicts for:
anaconda==2020.02 -> curl==7.68.0=h2a8f88b_0
pycurl -> curl[version='7.55.*|>=7.55.1,<8.0a0']
Package bleach conflicts for:
jupyter -> nbconvert -> bleach
anaconda==2020.02 -> nbconvert==5.6.1=py36_0 -> bleach
spyder -> nbconvert[version='>=4.0'] -> bleach
nbconvert -> bleach
notebook -> nbconvert -> bleach
anaconda==2020.02 -> bleach==3.1.0[build='py_0|py36_0|py37_0']
Package sphinxcontrib-devhelp conflicts for:
sphinx -> sphinxcontrib-devhelp
spyder -> sphinx[version='>=0.6.6'] -> sphinxcontrib-devhelp
numpydoc -> sphinx -> sphinxcontrib-devhelp
anaconda==2020.02 -> sphinxcontrib-devhelp==1.0.1=py_0
anaconda==2020.02 -> sphinx==2.4.0=py_0 -> sphinxcontrib-devhelp
Package heapdict conflicts for:
distributed -> zict[version='>=0.1.3'] -> heapdict
anaconda==2020.02 -> heapdict==1.0.1=py_0
zict -> heapdict
anaconda==2020.02 -> zict==1.0.0=py_0 -> heapdict
Package numpydoc conflicts for:
spyder -> numpydoc[version='>=0.6.0']
python-language-server -> jedi[version='>=0.12'] -> numpydoc
anaconda==2020.02 -> numpydoc==0.9.2=py_0
anaconda==2020.02 -> spyder==4.0.1=py36_0 -> numpydoc
ipython -> jedi[version='>=0.10'] -> numpydoc
jedi -> numpydoc
Package pympler conflicts for:
spyder -> pympler
hypothesis -> attrs[version='>=16.0.0'] -> pympler
attrs -> pympler
pytest -> attrs[version='>=17.2.0'] -> pympler
Package win_inet_pton conflicts for:
anaconda==2020.02 -> win_inet_pton==1.1.0[build='py38_0|py36_0|py37_0']
anaconda==2020.02 -> pysocks==1.7.1=py36_0 -> win_inet_pton
urllib3 -> pysocks[version='>=1.5.6,<2.0,!=1.5.7'] -> win_inet_pton
pysocks -> win_inet_pton
Package werkzeug conflicts for:
anaconda==2020.02 -> werkzeug==1.0.0=py_0
flask -> werkzeug[version='>=0.14|>=0.15|>=0.7']
anaconda==2020.02 -> flask==1.1.1=py_0 -> werkzeug[version='>=0.14']
Package et_xmlfile conflicts for:
openpyxl -> et_xmlfile
anaconda==2020.02 -> openpyxl==3.0.3=py_0 -> et_xmlfile
anaconda==2020.02 -> et_xmlfile==1.0.1[build='py38_0|py36h3d2d736_0|py37_0']
Package qdarkstyle conflicts for:
spyder -> qdarkstyle[version='>=2.7|>=2.8']
anaconda==2020.02 -> spyder==4.0.1=py36_0 -> qdarkstyle[version='>=2.7']
anaconda==2020.02 -> qdarkstyle==2.8=py_0
Package msgpack-python conflicts for:
dask -> distributed[version='>=2.14.0'] -> msgpack-python[version='<1.0.0|>=0.6.0']
anaconda==2020.02 -> msgpack-python==0.6.1[build='py37h74a9793_1|py36h74a9793_1|py38h74a9793_1']
pip -> cachecontrol -> msgpack-python
anaconda==2020.02 -> distributed==2.11.0=py36_0 -> msgpack-python[version='>=0.6.0']
distributed -> msgpack-python[version='<1.0.0|>=0.6.0']
Package python-libarchive-c conflicts for:
conda-package-handling -> python-libarchive-c
conda[version='>=4.8.3'] -> conda-package-handling[version='>=1.3.0'] -> python-libarchive-c
anaconda==2020.02 -> python-libarchive-c==2.8[build='py38_7|py36_13|py37_13']
conda-build -> python-libarchive-c
conda-verify -> conda-package-handling[version='>=1.0.4'] -> python-libarchive-c
Package libopenblas conflicts for:
numpy -> libblas[version='>=3.8.0,<4.0a0'] -> libopenblas[version='0.3.6|0.3.7|0.3.7|0.3.8|0.3.9',build='h29e5d5d_6|h29e5d5d_0|h29e5d5d_1|hcfe7411_0']
blas -> libblas==3.8.0=16_openblas -> libopenblas[version='0.3.6|0.3.6|0.3.6|0.3.6|0.3.7|0.3.7|0.3.8|0.3.9|0.3.7|0.3.7|0.3.7|0.3.7|0.3.7|0.3.7',build='h29e5d5d_4|hcfe7411_4|hcfe7411_5|hcfe7411_6|h29e5d5d_6|h29e5d5d_0|h29e5d5d_1|hcfe7411_0|hcfe7411_7|h29e5d5d_3|h29e5d5d_2|h29e5d5d_5|h29e5d5d_3']
scipy -> libblas[version='>=3.8.0,<4.0a0'] -> libopenblas[version='0.3.6|0.3.7|0.3.7|0.3.8|0.3.9',build='h29e5d5d_6|h29e5d5d_0|h29e5d5d_1|hcfe7411_0']
Package qtawesome conflicts for:
spyder -> qtawesome[version='>=0.4.1|>=0.5.7']
anaconda==2020.02 -> qtawesome==0.6.1=py_0
anaconda==2020.02 -> spyder==4.0.1=py36_0 -> qtawesome[version='>=0.5.7']
Package twisted conflicts for:
notebook -> prometheus_client -> twisted
prometheus_client -> twisted
Package jupyterlab_launcher conflicts for:
jupyterlab -> jupyterlab_launcher[version='<0.2|>=0.10.0,<0.11.0|>=0.11.2,<0.12.0|>=0.13.1,<0.14.0|>=0.6.0,<0.7.0|>=0.5.4,<0.6.0|>=0.5.4|>=0.4.0|>=0.3.0|>=0.2.3|>=0.11.0,<0.12.0']
_ipyw_jlab_nb_ext_conf -> jupyterlab -> jupyterlab_launcher[version='<0.2|>=0.10.0,<0.11.0|>=0.11.2,<0.12.0|>=0.13.1,<0.14.0|>=0.6.0,<0.7.0|>=0.5.4,<0.6.0|>=0.5.4|>=0.4.0|>=0.3.0|>=0.2.3|>=0.11.0,<0.12.0']
Package pycosat conflicts for:
anaconda==2020.02 -> pycosat==0.6.3[build='py37he774522_0|py38he774522_0|py36he774522_0']
conda[version='>=4.8.3'] -> pycosat[version='>=0.6.3']
conda-build -> conda[version='>=4.5'] -> pycosat[version='>=0.6.1|>=0.6.3']
Package pbr conflicts for:
pytables -> mock -> pbr[version='>=1.3']
mock -> pbr[version='>=1.3']
Package mistune conflicts for:
spyder -> nbconvert[version='>=4.0'] -> mistune[version='>0.6|>=0.7.4|>=0.8.1']
jupyter -> nbconvert -> mistune[version='>0.6|>=0.7.4|>=0.8.1']
anaconda==2020.02 -> mistune==0.8.4[build='py37he774522_0|py38he774522_1000|py36he774522_0']
notebook -> nbconvert -> mistune[version='>0.6|>=0.7.4|>=0.8.1']
anaconda==2020.02 -> nbconvert==5.6.1=py36_0 -> mistune[version='>=0.8.1']
nbconvert -> mistune[version='>0.6|>=0.7.4|>=0.8.1']
Package backports.tempfile conflicts for:
conda-build -> conda-verify -> backports.tempfile
conda-verify -> backports.tempfile
Package pywinpty conflicts for:
notebook -> terminado[version='>=0.8.1'] -> pywinpty
anaconda==2020.02 -> terminado==0.8.3=py36_0 -> pywinpty
terminado -> pywinpty
anaconda==2020.02 -> pywinpty==0.5.7[build='py38_0|py36_0|py37_0']
Package scandir conflicts for:
testpath -> pathlib2 -> scandir
ipython -> pathlib2 -> scandir
pickleshare -> pathlib2 -> scandir
conda-build -> scandir
pathlib2 -> scandir
pytest -> pathlib2[version='>=2.2.0'] -> scandir
importlib_metadata -> pathlib2 -> scandir
Package networkx conflicts for:
anaconda==2020.02 -> scikit-image==0.16.2=py36h47e9c7a_0 -> networkx[version='>=2.0']
anaconda==2020.02 -> networkx==2.4=py_0
scikit-image -> networkx[version='>=1.8,<2.0|>=1.8|>=2.0']
Package llvmlite conflicts for:
anaconda==2020.02 -> numba==0.48.0=py36h47e9c7a_0 -> llvmlite[version='>=0.31.0|>=0.31.0,<0.32.0a0']
anaconda==2020.02 -> llvmlite==0.31.0[build='py37ha925a31_0|py38ha925a31_0|py36ha925a31_0']
numba -> llvmlite[version='0.20.|0.21.|0.22.|0.23.|0.24.*|>=0.25.0,<0.26.0a0|>=0.26.0,<0.27.0a0|>=0.27.0,<0.28.0a0|>=0.28.0,<0.29.0a0|>=0.29.0,<0.30.0a0|>=0.29.0|>=0.30.0,<0.31.0a0|>=0.30.0|>=0.31.0,<0.32.0a0|>=0.32.0|>=0.32.0,<0.33.0a0|>=0.31.0']
Package mock conflicts for:
anaconda==2020.02 -> pytables==3.6.1=py36h1da0976_0 -> mock
anaconda==2020.02 -> mock==4.0.1=py_0
pytables -> mock
Package liblief conflicts for:
py-lief -> liblief==0.9.0[build='ha925a31_3|ha925a31_2|ha925a31_0']
anaconda==2020.02 -> liblief==0.9.0=ha925a31_2
conda-build -> py-lief -> liblief==0.9.0[build='ha925a31_3|ha925a31_2|ha925a31_0']
Package _low_priority conflicts for:
python-libarchive-c -> libarchive -> _low_priority
libarchive -> _low_priority
conda-build -> libarchive -> _low_priority
Package filelock conflicts for:
conda-build -> filelock
anaconda==2020.02 -> filelock==3.0.12=py_0
Package jupyterlab conflicts for:
_ipyw_jlab_nb_ext_conf -> jupyterlab
anaconda==2020.02 -> jupyterlab==1.2.6=pyhf63ae98_0
Package pkginfo conflicts for:
conda-build -> pkginfo
anaconda==2020.02 -> pkginfo==1.5.0.1[build='py38_0|py36_0|py37_0']
Package backports.weakref conflicts for:
backports.tempfile -> backports.weakref
conda-verify -> backports.tempfile -> backports.weakrefThe following specifications were found to be incompatible with your CUDA driver:
feature:/win-64::__cuda==10.2=0
feature:|@/win-64::__cuda==10.2=0
Your installed CUDA driver is: 10.2
Environment
OS: Microsoft Windows 10 v1909 x64
Installation type: Conda v4.8.3 (Anaconda v3-2020.02) / Python v3.7.6
RAM available: 64 GB
Hardware specifications: nVidia GeForce GTX 970 4GB / Intel Core i7-8700K @ 3.70 GHz / ASRock Z370 Taichi
Additional context
This issue is similar to #295 (might be, in fact, the same). I've been installing Deezer Spleeter since v1.4.4 without problems, but I had to use the 'python -m' workaround since v1.4.9. Antivirus was always deactivated, as per issue #252.
full error log.txt
was the same problem, was solved when i first run conda update --all
|
gharchive/issue
| 2020-04-22T14:24:28 |
2025-04-01T04:33:58.325697
|
{
"authors": [
"pmamieva",
"sashko3891"
],
"repo": "deezer/spleeter",
"url": "https://github.com/deezer/spleeter/issues/339",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1686807398
|
!spleeter separate -h not supported on Google Collab
This step on the Collab is returning the message
Usage: spleeter separate [OPTIONS] FILES...
Try 'spleeter separate --help' for help.
Error: no such option: -h
There is no comment and I'm not sure what this step was intending to do, but it prevents me from running the step by step example on Collab
They may have changed format.
spleeter separate --help but you dont need to run this line in the colab its just to show you the help menu.
You can skip it and go to the next code block.
|
gharchive/issue
| 2023-04-27T13:01:15 |
2025-04-01T04:33:58.329461
|
{
"authors": [
"dafinley",
"jackmattsonAI"
],
"repo": "deezer/spleeter",
"url": "https://github.com/deezer/spleeter/issues/842",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1100867138
|
Create Zarf Product Website
Outcome: We want to have a website dedicated solely to the open-source project Zarf. This website will contain all the details about Zarf functionality, how to contribute, contributors, partners, etc. DU website will have a high-level overview of Zarf and a link to the Zarf website.
Background:
Since Zarf is an open-source project, we know its ownership may change. Therefore, we would like a dedicated product website that is separate from the DU website.
Acceptance Criteria:
A single resource for developers to discover, learn, and contribute to Zarf.
A published website with a push goal of SEO landing us at the top of Google.
Notes:
Examples of great open-source websites (Flux, Argo)
Timeline: Mid May
Notes from last refinement session on high-level desired outcomes for the Zarf website:
Technical docs site
Training videos
Making it pretty
Good tutorials
Tie-in with Zarf UI
Closing this, future work will be tracked in that project: https://github.com/defenseunicorns/zarf-website/issues
|
gharchive/issue
| 2022-01-12T22:48:29 |
2025-04-01T04:33:58.334612
|
{
"authors": [
"RothAndrew",
"jeff-mccoy",
"mgrim2"
],
"repo": "defenseunicorns/zarf",
"url": "https://github.com/defenseunicorns/zarf/issues/230",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2669089202
|
GetAll, GetValues, GetKeys as gRPC server side stream.
// KeyValueService defines the gRPC service
service KeyValueService {
rpc Get(GetRequest) returns (KeyValue);
rpc GetAll(GetAllRequest) returns (GetAllResponse);
Have you thought about returning the list of values as a stream?
Originally posted by @louislef299 in https://github.com/defoeam/Herd/pull/38#discussion_r1846843322
https://grpc.io/docs/languages/go/basics/#server-side-streaming-rpc
|
gharchive/issue
| 2024-11-18T16:14:16 |
2025-04-01T04:33:58.362951
|
{
"authors": [
"defoeam",
"louislef299"
],
"repo": "defoeam/Herd",
"url": "https://github.com/defoeam/Herd/issues/39",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2470555696
|
Update go version in Dockerfile
Updated go version to 1.21 to match go.mod.
I included this fix in my new PR #25
|
gharchive/pull-request
| 2024-08-16T16:09:39 |
2025-04-01T04:33:58.363736
|
{
"authors": [
"defoeam",
"tkoivi"
],
"repo": "defoeam/kvs",
"url": "https://github.com/defoeam/kvs/pull/24",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
634799545
|
Make Email Template more readable
Before with 12 points the style guide was a bit intimidating and hard to read.
Merge formatting into one numbered point.
Add code block to [YOUR NAME] and friends.
Add periods to end bullet items.
I submitted a PR at #883
|
gharchive/issue
| 2020-06-08T17:23:51 |
2025-04-01T04:33:58.374096
|
{
"authors": [
"NickCrews"
],
"repo": "defund12/defund12.org",
"url": "https://github.com/defund12/defund12.org/issues/881",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
102440300
|
can not bootstrap v1.9.0 cluster
the only thing I ran is
make discovery-url and
./contrib/digitalocean/provision-do-cluster.sh ams3 etc 16GB deis-eu1
When sshing into deis-1 node It seems etcd is not running
journalctl -u etcd
Aug 21 18:32:21 deis-eu1-1 systemd[1]: Starting etcd2 container...
Aug 21 18:32:21 deis-eu1-1 systemd[1]: Started etcd2 container.
Aug 21 18:32:21 deis-eu1-1 docker[3728]: 2015/08/21 18:32:21 etcdmain: setting maximum number of CPUs to 1, total number of available CPUs is 8
Aug 21 18:32:21 deis-eu1-1 docker[3728]: 2015/08/21 18:32:21 etcdmain: listening for peers on http://0.0.0.0:2380
Aug 21 18:32:21 deis-eu1-1 docker[3728]: 2015/08/21 18:32:21 etcdmain: listening for peers on http://0.0.0.0:7001
Aug 21 18:32:21 deis-eu1-1 docker[3728]: 2015/08/21 18:32:21 etcdmain: listening for client requests on http://0.0.0.0:2379
Aug 21 18:32:21 deis-eu1-1 docker[3728]: 2015/08/21 18:32:21 etcdmain: listening for client requests on http://0.0.0.0:4001
Aug 21 18:32:22 deis-eu1-1 docker[3728]: 2015/08/21 18:32:22 etcdmain: stopping listening for client requests on http://0.0.0.0:4001
Aug 21 18:32:22 deis-eu1-1 docker[3728]: 2015/08/21 18:32:22 etcdmain: stopping listening for client requests on http://0.0.0.0:2379
Aug 21 18:32:22 deis-eu1-1 docker[3728]: 2015/08/21 18:32:22 etcdmain: stopping listening for peers on http://0.0.0.0:7001
Aug 21 18:32:22 deis-eu1-1 docker[3728]: 2015/08/21 18:32:22 etcdmain: stopping listening for peers on http://0.0.0.0:2380
Aug 21 18:32:22 deis-eu1-1 docker[3728]: 2015/08/21 18:32:22 etcdmain: member "d3707e031a2741998c395165bbce36f6" has previously registered with discovery service token (https://discovery.etcd.io/2c086b9947da609d
Aug 21 18:32:22 deis-eu1-1 docker[3728]: 2015/08/21 18:32:22 etcdmain: But etcd could not find vaild cluster configuration in the given data dir (/var/lib/etcd2).
Aug 21 18:32:22 deis-eu1-1 docker[3728]: 2015/08/21 18:32:22 etcdmain: Please check the given data dir path if the previous bootstrap succeeded
Aug 21 18:32:22 deis-eu1-1 docker[3728]: 2015/08/21 18:32:22 etcdmain: or use a new discovery token if the previous bootstrap failed.
Aug 21 18:32:22 deis-eu1-1 systemd[1]: etcd.service: Main process exited, code=exited, status=1/FAILURE
Aug 21 18:32:22 deis-eu1-1 systemd[1]: etcd.service: Unit entered failed state.
Aug 21 18:32:22 deis-eu1-1 systemd[1]: etcd.service: Failed with result 'exit-code'.
Aug 21 18:32:32 deis-eu1-1 systemd[1]: etcd.service: Service hold-off time over, scheduling restart.
And this over and over again
it starts a little bit different though
-- Logs begin at Fri 2015-08-21 18:17:49 UTC, end at Fri 2015-08-21 18:36:48 UTC. --
Aug 21 18:17:58 deis-eu1-1 systemd[1]: Starting etcd2 container...
Aug 21 18:18:09 deis-eu1-1 sh[660]: time="2015-08-21T18:18:09Z" level=fatal msg="Error response from daemon: v1 ping attempt failed with error: Get https://quay.io/v1/_ping: dial tcp: i/o timeout. If this privat
Aug 21 18:18:09 deis-eu1-1 systemd[1]: etcd.service: Control process exited, code=exited status=1
Aug 21 18:18:09 deis-eu1-1 systemd[1]: Failed to start etcd2 container.
Aug 21 18:18:09 deis-eu1-1 systemd[1]: etcd.service: Unit entered failed state.
Aug 21 18:18:09 deis-eu1-1 systemd[1]: etcd.service: Failed with result 'exit-code'.
Aug 21 18:18:19 deis-eu1-1 systemd[1]: etcd.service: Service hold-off time over, scheduling restart.
Aug 21 18:18:19 deis-eu1-1 systemd[1]: Starting etcd2 container...
Aug 21 18:18:20 deis-eu1-1 sh[682]: Pulling repository quay.io/coreos/etcd
Aug 21 18:18:21 deis-eu1-1 sh[682]: 58e6c0c28ea7: Pulling image (v2.1.1) from quay.io/coreos/etcd
Aug 21 18:18:21 deis-eu1-1 sh[682]: 58e6c0c28ea7: Pulling image (v2.1.1) from quay.io/coreos/etcd, endpoint: https://quay.io/v1/
Aug 21 18:18:22 deis-eu1-1 sh[682]: 58e6c0c28ea7: Pulling dependent layers
Aug 21 18:18:22 deis-eu1-1 sh[682]: 3d0e9f1ec11c: Pulling metadata
Aug 21 18:18:23 deis-eu1-1 sh[682]: 3d0e9f1ec11c: Pulling fs layer
Aug 21 18:18:25 deis-eu1-1 sh[682]: 3d0e9f1ec11c: Download complete
Aug 21 18:18:25 deis-eu1-1 sh[682]: 4b05e49a9ce6: Pulling metadata
Aug 21 18:18:26 deis-eu1-1 sh[682]: 4b05e49a9ce6: Pulling fs layer
Aug 21 18:18:30 deis-eu1-1 sh[682]: 4b05e49a9ce6: Download complete
Aug 21 18:18:30 deis-eu1-1 sh[682]: ea63d5f7cdda: Pulling metadata
Aug 21 18:18:31 deis-eu1-1 sh[682]: ea63d5f7cdda: Pulling fs layer
Aug 21 18:18:32 deis-eu1-1 sh[682]: ea63d5f7cdda: Download complete
Aug 21 18:18:32 deis-eu1-1 sh[682]: 58e6c0c28ea7: Pulling metadata
Aug 21 18:18:33 deis-eu1-1 sh[682]: 58e6c0c28ea7: Pulling fs layer
Aug 21 18:18:33 deis-eu1-1 sh[682]: 58e6c0c28ea7: Download complete
Aug 21 18:18:33 deis-eu1-1 sh[682]: 58e6c0c28ea7: Download complete
Aug 21 18:18:33 deis-eu1-1 sh[682]: Status: Downloaded newer image for quay.io/coreos/etcd:v2.1.1
Aug 21 18:18:33 deis-eu1-1 systemd[1]: Started etcd2 container.
Aug 21 18:18:33 deis-eu1-1 docker[723]: 2015/08/21 18:18:33 etcdmain: setting maximum number of CPUs to 1, total number of available CPUs is 8
Aug 21 18:18:33 deis-eu1-1 docker[723]: 2015/08/21 18:18:33 etcdmain: listening for peers on http://0.0.0.0:2380
Aug 21 18:18:33 deis-eu1-1 docker[723]: 2015/08/21 18:18:33 etcdmain: listening for peers on http://0.0.0.0:7001
Aug 21 18:18:33 deis-eu1-1 docker[723]: 2015/08/21 18:18:33 etcdmain: listening for client requests on http://0.0.0.0:2379
Aug 21 18:18:33 deis-eu1-1 docker[723]: 2015/08/21 18:18:33 etcdmain: listening for client requests on http://0.0.0.0:4001
Aug 21 18:18:35 deis-eu1-1 docker[723]: 2015/08/21 18:18:35 discovery: found self a76b8132681dea7a in the cluster
Aug 21 18:18:35 deis-eu1-1 docker[723]: 2015/08/21 18:18:35 discovery: found 1 peer(s), waiting for 2 more
Aug 21 18:19:35 deis-eu1-1 docker[723]: 2015/08/21 18:19:35 etcdmain: stopping listening for client requests on http://0.0.0.0:4001
Aug 21 18:19:35 deis-eu1-1 docker[723]: 2015/08/21 18:19:35 etcdmain: stopping listening for client requests on http://0.0.0.0:2379
Aug 21 18:19:35 deis-eu1-1 docker[723]: 2015/08/21 18:19:35 etcdmain: stopping listening for peers on http://0.0.0.0:7001
Aug 21 18:19:35 deis-eu1-1 docker[723]: 2015/08/21 18:19:35 etcdmain: stopping listening for peers on http://0.0.0.0:2380
Aug 21 18:19:35 deis-eu1-1 docker[723]: 2015/08/21 18:19:35 etcdmain: unexpected end of JSON input
Aug 21 18:19:35 deis-eu1-1 systemd[1]: etcd.service: Main process exited, code=exited, status=1/FAILURE
Aug 21 18:19:35 deis-eu1-1 systemd[1]: etcd.service: Unit entered failed state.
Aug 21 18:19:35 deis-eu1-1 systemd[1]: etcd.service: Failed with result 'exit-code'.
What is weird though is that on my 3rd node etcd is running correctly. Although obviously it can not peer with its friends
-- Logs begin at Fri 2015-08-21 18:21:09 UTC, end at Fri 2015-08-21 18:38:36 UTC. --
Aug 21 18:21:24 deis-eu1-3 systemd[1]: Starting etcd2 container...
Aug 21 18:21:36 deis-eu1-3 sh[666]: Pulling repository quay.io/coreos/etcd
Aug 21 18:21:37 deis-eu1-3 sh[666]: 58e6c0c28ea7: Pulling image (v2.1.1) from quay.io/coreos/etcd
Aug 21 18:21:37 deis-eu1-3 sh[666]: 58e6c0c28ea7: Pulling image (v2.1.1) from quay.io/coreos/etcd, endpoint: https://quay.io/v1/
Aug 21 18:21:37 deis-eu1-3 sh[666]: 58e6c0c28ea7: Pulling dependent layers
Aug 21 18:21:37 deis-eu1-3 sh[666]: 3d0e9f1ec11c: Pulling metadata
Aug 21 18:21:37 deis-eu1-3 sh[666]: 3d0e9f1ec11c: Pulling fs layer
Aug 21 18:21:40 deis-eu1-3 sh[666]: 3d0e9f1ec11c: Download complete
Aug 21 18:21:40 deis-eu1-3 sh[666]: 4b05e49a9ce6: Pulling metadata
Aug 21 18:21:41 deis-eu1-3 sh[666]: 4b05e49a9ce6: Pulling fs layer
Aug 21 18:21:44 deis-eu1-3 sh[666]: 4b05e49a9ce6: Download complete
Aug 21 18:21:44 deis-eu1-3 sh[666]: ea63d5f7cdda: Pulling metadata
Aug 21 18:21:45 deis-eu1-3 sh[666]: ea63d5f7cdda: Pulling fs layer
-- Logs begin at Fri 2015-08-21 18:21:09 UTC, end at Fri 2015-08-21 18:38:36 UTC. --
Aug 21 18:21:24 deis-eu1-3 systemd[1]: Starting etcd2 container...
Aug 21 18:21:36 deis-eu1-3 sh[666]: Pulling repository quay.io/coreos/etcd
Aug 21 18:21:37 deis-eu1-3 sh[666]: 58e6c0c28ea7: Pulling image (v2.1.1) from quay.io/coreos/etcd
Aug 21 18:21:37 deis-eu1-3 sh[666]: 58e6c0c28ea7: Pulling image (v2.1.1) from quay.io/coreos/etcd, endpoint: https://quay.io/v1/
Aug 21 18:21:37 deis-eu1-3 sh[666]: 58e6c0c28ea7: Pulling dependent layers
Aug 21 18:21:37 deis-eu1-3 sh[666]: 3d0e9f1ec11c: Pulling metadata
Aug 21 18:21:37 deis-eu1-3 sh[666]: 3d0e9f1ec11c: Pulling fs layer
Aug 21 18:21:40 deis-eu1-3 sh[666]: 3d0e9f1ec11c: Download complete
Aug 21 18:21:40 deis-eu1-3 sh[666]: 4b05e49a9ce6: Pulling metadata
Aug 21 18:21:41 deis-eu1-3 sh[666]: 4b05e49a9ce6: Pulling fs layer
Aug 21 18:21:44 deis-eu1-3 sh[666]: 4b05e49a9ce6: Download complete
Aug 21 18:21:44 deis-eu1-3 sh[666]: ea63d5f7cdda: Pulling metadata
Aug 21 18:21:45 deis-eu1-3 sh[666]: ea63d5f7cdda: Pulling fs layer
Aug 21 18:21:46 deis-eu1-3 sh[666]: ea63d5f7cdda: Download complete
Aug 21 18:21:46 deis-eu1-3 sh[666]: 58e6c0c28ea7: Pulling metadata
Aug 21 18:21:46 deis-eu1-3 sh[666]: 58e6c0c28ea7: Pulling fs layer
Aug 21 18:21:47 deis-eu1-3 sh[666]: 58e6c0c28ea7: Download complete
Aug 21 18:21:47 deis-eu1-3 sh[666]: 58e6c0c28ea7: Download complete
Aug 21 18:21:47 deis-eu1-3 sh[666]: Status: Downloaded newer image for quay.io/coreos/etcd:v2.1.1
Aug 21 18:21:47 deis-eu1-3 systemd[1]: Started etcd2 container.
Aug 21 18:21:47 deis-eu1-3 docker[717]: 2015/08/21 18:21:47 etcdmain: setting maximum number of CPUs to 1, total number of available CPUs is 8
Aug 21 18:21:47 deis-eu1-3 docker[717]: 2015/08/21 18:21:47 etcdmain: listening for peers on http://0.0.0.0:2380
Aug 21 18:21:47 deis-eu1-3 docker[717]: 2015/08/21 18:21:47 etcdmain: listening for peers on http://0.0.0.0:7001
Aug 21 18:21:47 deis-eu1-3 docker[717]: 2015/08/21 18:21:47 etcdmain: listening for client requests on http://0.0.0.0:2379
Aug 21 18:21:47 deis-eu1-3 docker[717]: 2015/08/21 18:21:47 etcdmain: listening for client requests on http://0.0.0.0:4001
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 discovery: found peer a76b8132681dea7a in the cluster
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 discovery: found peer f5eb1592e5613aa1 in the cluster
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 discovery: found self e7977e7c136c63d7 in the cluster
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 discovery: found 3 needed peer(s)
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 etcdserver: name = b1cc4207ac364d17ac78c39f4e77b45b
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 etcdserver: data dir = /var/lib/etcd2
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 etcdserver: member dir = /var/lib/etcd2/member
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 etcdserver: heartbeat = 400ms
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 etcdserver: election = 2000ms
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 etcdserver: snapshot count = 10000
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 etcdserver: discovery URL= https://discovery.etcd.io/2c086b9947da609d1d8792b6e78cbb86
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 etcdserver: advertise client URLs = http://10.133.22.107:2379,http://10.133.22.107:4001
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 etcdserver: initial advertise peer URLs = http://10.133.22.107:2380,http://10.133.22.107:7001
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 etcdserver: initial cluster = b1cc4207ac364d17ac78c39f4e77b45b=http://10.133.22.107:2380,b1cc4207ac364d17ac78c39f4e77b45b=http://10.133.22.107:7001
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 etcdserver: starting member e7977e7c136c63d7 in cluster f84abb03ef569926
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 raft: e7977e7c136c63d7 became follower at term 0
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 raft: newRaft e7977e7c136c63d7 [peers: [], term: 0, commit: 0, applied: 0, lastindex: 0, lastterm: 0]
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 raft: e7977e7c136c63d7 became follower at term 1
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 etcdserver: starting server... [version: 2.1.1, cluster version: to_be_decided]
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 etcdserver: added member a76b8132681dea7a [http://10.133.22.105:2380 http://10.133.22.105:7001] to cluster f84abb03ef569926
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 etcdserver: added local member e7977e7c136c63d7 [http://10.133.22.107:2380 http://10.133.22.107:7001] to cluster f84abb03ef569926
Aug 21 18:21:48 deis-eu1-3 docker[717]: 2015/08/21 18:21:48 etcdserver: added member f5eb1592e5613aa1 [http://10.133.22.106:2380 http://10.133.22.106:7001] to cluster f84abb03ef569926
Aug 21 18:21:51 deis-eu1-3 docker[717]: 2015/08/21 18:21:51 rafthttp: failed to dial f5eb1592e5613aa1 on stream Message (read tcp 10.133.22.106:2380: connection reset by peer)
Aug 21 18:21:51 deis-eu1-3 docker[717]: 2015/08/21 18:21:51 rafthttp: failed to dial f5eb1592e5613aa1 on stream MsgApp v2 (read tcp 10.133.22.106:2380: connection reset by peer)
Aug 21 18:21:51 deis-eu1-3 docker[717]: 2015/08/21 18:21:51 rafthttp: failed to dial f5eb1592e5613aa1 on stream Message (dial tcp 10.133.22.106:7001: connection refused)
Aug 21 18:21:51 deis-eu1-3 docker[717]: 2015/08/21 18:21:51 rafthttp: failed to dial f5eb1592e5613aa1 on stream MsgApp v2 (dial tcp 10.133.22.106:7001: connection refused)
Aug 21 18:21:51 deis-eu1-3 docker[717]: 2015/08/21 18:21:51 raft: e7977e7c136c63d7 is starting a new election at term 1
Aug 21 18:21:51 deis-eu1-3 docker[717]: 2015/08/21 18:21:51 raft: e7977e7c136c63d7 became candidate at term 2
Aug 21 18:21:51 deis-eu1-3 docker[717]: 2015/08/21 18:21:51 raft: e7977e7c136c63d7 received vote from e7977e7c136c63d7 at term 2
Aug 21 18:21:51 deis-eu1-3 docker[717]: 2015/08/21 18:21:51 raft: e7977e7c136c63d7 [logterm: 1, index: 3] sent vote request to f5eb1592e5613aa1 at term 2
Aug 21 18:21:51 deis-eu1-3 docker[717]: 2015/08/21 18:21:51 raft: e7977e7c136c63d7 [logterm: 1, index: 3] sent vote request to a76b8132681dea7a at term 2
Aug 21 18:21:51 deis-eu1-3 docker[717]: 2015/08/21 18:21:51 rafthttp: failed to write f5eb1592e5613aa1 on pipeline (dial tcp 10.133.22.106:2380: connection refused)
Aug 21 18:21:51 deis-eu1-3 docker[717]: 2015/08/21 18:21:51 rafthttp: failed to dial f5eb1592e5613aa1 on stream MsgApp v2 (dial tcp 10.133.22.106:2380: connection refused)
Aug 21 18:21:51 deis-eu1-3 docker[717]: 2015/08/21 18:21:51 rafthttp: failed to dial f5eb1592e5613aa1 on stream Message (dial tcp 10.133.22.106:2380: connection refused)
Aug 21 18:21:52 deis-eu1-3 docker[717]: 2015/08/21 18:21:52 rafthttp: failed to dial f5eb1592e5613aa1 on stream Message (dial tcp 10.133.22.106:7001: connection refused)
Aug 21 18:21:52 deis-eu1-3 docker[717]: 2015/08/21 18:21:52 rafthttp: failed to dial f5eb1592e5613aa1 on stream MsgApp v2 (dial tcp 10.133.22.106:7001: connection refused)
Aug 21 18:21:52 deis-eu1-3 docker[717]: 2015/08/21 18:21:52 rafthttp: failed to dial f5eb1592e5613aa1 on stream MsgApp v2 (dial tcp 10.133.22.106:2380: connection refused)
Aug 21 18:21:52 deis-eu1-3 docker[717]: 2015/08/21 18:21:52 rafthttp: failed to dial f5eb1592e5613aa1 on stream Message (dial tcp 10.133.22.106:2380: connection refused)
Aug 21 18:21:52 deis-eu1-3 docker[717]: 2015/08/21 18:21:52 rafthttp: failed to dial f5eb1592e5613aa1 on stream Message (dial tcp 10.133.22.106:7001: connection refused)
Aug 21 18:21:52 deis-eu1-3 docker[717]: 2015/08/21 18:21:52 rafthttp: failed to dial f5eb1592e5613aa1 on stream MsgApp v2 (dial tcp 10.133.22.106:7001: connection refused)
Aug 21 18:21:52 deis-eu1-3 docker[717]: 2015/08/21 18:21:52 rafthttp: failed to dial f5eb1592e5613aa1 on stream MsgApp v2 (dial tcp 10.133.22.106:2380: connection refused)
Aug 21 18:21:52 deis-eu1-3 docker[717]: 2015/08/21 18:21:52 rafthttp: failed to dial f5eb1592e5613aa1 on stream Message (dial tcp 10.133.22.106:2380: connection refused)
Even weirder is that I repeated this twice (destroyed, started over again with new discovery URL) and the same thing happened twice. Node 1 & 2 not working. Node 3 working but not able to connect to its peers...
This looks to be a clue:
Aug 21 18:32:22 deis-eu1-1 docker[3728]: 2015/08/21 18:32:22 etcdmain: member "d3707e031a2741998c395165bbce36f6" has previously registered with discovery service token (https://discovery.etcd.io/2c086b9947da609d
Aug 21 18:32:22 deis-eu1-1 docker[3728]: 2015/08/21 18:32:22 etcdmain: But etcd could not find vaild cluster configuration in the given data dir (/var/lib/etcd2).
So, a host with that IP already registered with that service. Are you sure you're redeploying with a fresh discovery URL each time?
yes I am sure...
Note that at first it works (kinda) but it fails on unexpected end of json.
[nathan@zeta deis]$ make discovery-url
sed -e "s,discovery #DISCOVERY_URL,discovery $(curl -s -w '\n' https://discovery.etcd.io/new?size=$DEIS_NUM_INSTANCES)," contrib/coreos/user-data.example > contrib/coreos/user-data
[nathan@zeta deis]$ cat contrib/coreos/user-data
...
[nathan@zeta deis]$ ./contrib/digitalocean/provision-do-cluster.sh ams3 230790 16GB deis-eu1
Provisioning deis-eu1-1...
Waiting for droplet to become available............................................Completed
You can connect to your Droplet via
188.166.67.57
Provisioning deis-eu1-2...
Waiting for droplet to become available............................................................^T....................................Completed
You can connect to your Droplet via
188.166.64.130
Provisioning deis-eu1-3...
Waiting for droplet to become available...........................................Completed
You can connect to your Droplet via
178.62.195.168
Your Deis cluster has successfully deployed to DigitalOcean.
Please continue to follow the instructions in the README.
[nathan@zeta deis]$
Aug 21 18:18:35 deis-eu1-1 docker[723]: 2015/08/21 18:18:35 discovery: found self a76b8132681dea7a in the cluster
Aug 21 18:18:35 deis-eu1-1 docker[723]: 2015/08/21 18:18:35 discovery: found 1 peer(s), waiting for 2 more
Aug 21 18:19:35 deis-eu1-1 docker[723]: 2015/08/21 18:19:35 etcdmain: stopping listening for client requests on http://0.0.0.0:4001
Aug 21 18:19:35 deis-eu1-1 docker[723]: 2015/08/21 18:19:35 etcdmain: stopping listening for client requests on http://0.0.0.0:2379
Aug 21 18:19:35 deis-eu1-1 docker[723]: 2015/08/21 18:19:35 etcdmain: stopping listening for peers on http://0.0.0.0:7001
Aug 21 18:19:35 deis-eu1-1 docker[723]: 2015/08/21 18:19:35 etcdmain: stopping listening for peers on http://0.0.0.0:2380
Aug 21 18:19:35 deis-eu1-1 docker[723]: 2015/08/21 18:19:35 etcdmain: unexpected end of JSON input
Aug 21 18:19:35 deis-eu1-1 systemd[1]: etcd.service: Main process exited, code=exited, status=1/FAILURE
I think the clue is in the unexpected end of JSON and the fact that is exactly 1 minute after startup.
I have the feeling that node 2 and 3 are not fast enough to join, and somehow etcd crashes and results in a forever failed state
I was able to reproduce the exact same issue using vagrant on a RAM-constrained env.
My machines on DO were running 16GB each, so memory does not seem to be the problem here
Looking at the other issue also on DO, might it be that DO is shipping with a CoreOS version that is somehow not compatible with deis 1.9/etcd2?
Looking at the other issue also on DO, might it be that DO is shipping with a CoreOS version that is somehow not compatible with deis 1.9/etcd2?
I think that's a reasonable assumption. I was testing #4315 today and had to manually start fleetd to get things to come up. I know a newer Docker is shipping that is known to cause issues (which is why we're still on 642.2.0), but I'm not sure what changes were made to etcd2.
I just ran #4315 twice in a row and had actually no problems....
I believe this might be more a timing issue in which droplets are created (if the 3rd one is too late to the party, number one might already have given up and went "home" in a broken state -- sorry for the bad analogy).
I''l try the old method again from current master and see if I still have a problem, and also see if I can fix it by not using --wait in the script
Yep can reliably reproduce on master / 1.9.0, but if I remove --wait from the original script (making it so that nodes boot up at almost teh same time) it works perfectly....
It seems etcd2 might have a problem bootstrapping when first 3 nodes don;t come up in approx the same time.
I guess this is a wider problem than DO, and in fact might also happenin with the new script if for some reason one node is slower to boot up etcd2 (eg delays in DO, or some reason slower in downloading the image)
I believe this might be coreos/etcd#3257
I believe this might be coreos/etcd#3257
If so, then it should be fixed by #4316.
It looks closely related to #4310.
I am now retesting with latest master which should have the fix as @mboersma said. DO is currently having problems booting up machines + network slowness in these machines, so ideal test conditions for this bug :)
Ill let you know if everything comes up
but so far looks good with
Aug 24 19:19:17 deis-eu3-1 docker[675]: 2015/08/24 19:19:17 discovery: found self e877b98c5c2bb701 in the cluster
Aug 24 19:19:17 deis-eu3-1 docker[675]: 2015/08/24 19:19:17 discovery: found 1 peer(s), waiting for 2 more
after at least one tineout on the first node
Ok came up like a charm without manual intervention
systemctl status flanneld.service is showing an error though, but might be nornal (and totally unrelated)
core@deis-eu3-2 ~ $ sudo systemctl status flanneld.service
● flanneld.service - Network fabric for containers
Loaded: loaded (/etc/systemd/system/flanneld.service; static; vendor preset: disabled)
Active: active (running) since Mon 2015-08-24 19:45:37 UTC; 2h 34min ago
Docs: https://github.com/coreos/flannel
Process: 1238 ExecStartPost=/usr/bin/docker run --net=host --rm -v /run:/run quay.io/coreos/flannel:${FLANNEL_VER} /opt/bin/mk-docker-opts.sh -d /run/flannel_docker_opts.env -i (code=exited, status=0/SUCCESS)
Process: 1167 ExecStartPre=/usr/bin/etcdctl mk /coreos.com/network/config {"Network":"10.244.0.0/16", "SubnetLen": 24, "SubnetMin":"10.244.0.0", "Backend": {"Type": "vxlan"}} (code=exited, status=4)
Process: 1164 ExecStartPre=/usr/bin/touch /run/flannel/options.env (code=exited, status=0/SUCCESS)
Process: 1161 ExecStartPre=/usr/bin/mkdir -p /run/flannel (code=exited, status=0/SUCCESS)
Process: 1158 ExecStartPre=/sbin/modprobe ip_tables (code=exited, status=0/SUCCESS)
Main PID: 1184 (sdnotify-proxy)
Memory: 4.9M
CPU: 269ms
CGroup: /system.slice/flanneld.service
├─1184 /usr/libexec/sdnotify-proxy /run/flannel/sd.sock /usr/bin/docker run --net=host --privile...
└─1187 /usr/bin/docker run --net=host --privileged=true --rm --volume=/run/flannel:/run/flannel ...
Aug 24 19:57:33 deis-eu3-2 sdnotify-proxy[1184]: go-etcd2015/08/24 19:57:33 DEBUG: send.request.to http... GET
Aug 24 19:57:33 deis-eu3-2 sdnotify-proxy[1184]: I0824 19:57:33.745520 00001 vxlan.go:232] Subnet added...0/24
Aug 24 19:57:33 deis-eu3-2 sdnotify-proxy[1184]: I0824 19:57:33.745592 00001 device.go:164] calling Nei...5:a2
Aug 24 19:57:33 deis-eu3-2 sdnotify-proxy[1184]: go-etcd2015/08/24 19:57:33 DEBUG: rawWatch /coreos.com...s []
Aug 24 19:57:33 deis-eu3-2 sdnotify-proxy[1184]: go-etcd2015/08/24 19:57:33 DEBUG: get /coreos.com/netw...379]
Aug 24 19:57:33 deis-eu3-2 sdnotify-proxy[1184]: go-etcd2015/08/24 19:57:33 DEBUG: Connecting to etcd: ...=521
Aug 24 19:57:33 deis-eu3-2 sdnotify-proxy[1184]: go-etcd2015/08/24 19:57:33 DEBUG: send.request.to http... GET
Aug 24 19:57:33 deis-eu3-2 sdnotify-proxy[1184]: go-etcd2015/08/24 19:57:33 DEBUG: recv.response.from
Aug 24 19:57:33 deis-eu3-2 sdnotify-proxy[1184]: go-etcd2015/08/24 19:57:33 DEBUG: recv.success
Aug 24 19:57:33 deis-eu3-2 sdnotify-proxy[1184]: go-etcd2015/08/24 19:57:33 DEBUG: recv.response.from
Hint: Some lines were ellipsized, use -l to show in full.
As reported this has been fixed in a subsequen trelease, so closing issue myself
|
gharchive/issue
| 2015-08-21T18:40:23 |
2025-04-01T04:33:58.396643
|
{
"authors": [
"MaxenceAdnot",
"carmstrong",
"krancour",
"mboersma",
"nathansamson"
],
"repo": "deis/deis",
"url": "https://github.com/deis/deis/issues/4311",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
920788633
|
Use timeouts on download operations in the client
Right now we don't have a timeout value for the client library. This means that if anything happens to the connection (or it is malformed), it can just hang (as happened in #148). We should set a default timeout and allow the user to specify a value when constructing the client. We also will need logic to handle a failed download. Basically, if timeout of X is reached, emit an error and remove the partially downloaded parcel.
After looking into this some more, there are only 3 timeout options available in reqwest
connect_timeout (time to wait for connection to the server)
timeout (connection + read + write, basically until the whole body is read out)
idle_pool_timeout (controls how long an idle connection is held on for, not for how long an active connection is idle)
We can't set timeout as some parcels could be large and it could be across a slow network, so there is no good default value we can pick.
So I am closing this as won't fix
|
gharchive/issue
| 2021-06-14T21:23:46 |
2025-04-01T04:33:58.400446
|
{
"authors": [
"thomastaylor312"
],
"repo": "deislabs/bindle",
"url": "https://github.com/deislabs/bindle/issues/156",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1677784709
|
Install Mulping (Windows/Pycharm)
Hey,
thank you for your work. Look very intresting and I will try it, but when I will install it with "pip install mulping" its say:
ERROR: Could not find a version that satisfies the requirement mulping (from versions: none)
ERROR: No matching distribution found for mulping
In Pycharm I can't find it. I have also tryed it manual but it also doesnt look really good... Someone have a hint for me or eventually I doesnt understand something wrong?
That is correct, mulping is not published as a Python package, you have to use pip only to install the requests module if you are on Windows and don't have it installed yet, by running
$ pip install requests
Then you can download mulping from this repository and run it normally with python
$ python3 ./mulping.py
If you want to have it as a global command I believe you can achieve it by placing it in one of the PATH tracked folders
Let me know if everything works fine :)
Oh, now I understand. Thank you <3
But I have to use this answer:
https://stackoverflow.com/questions/57131654/using-utf-8-encoding-chcp-65001-in-command-prompt-windows-powershell-window
because I become this error "utf-8' codec can t decode byte 0x81".
Don't know if this has something to do with your Project ;)
|
gharchive/issue
| 2023-04-21T04:36:01 |
2025-04-01T04:33:58.406964
|
{
"authors": [
"Codein99",
"deivshon"
],
"repo": "deivshon/mulping",
"url": "https://github.com/deivshon/mulping/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2598721388
|
Go module: Separate datastar core from implementation-specific dependencies.
Currently the datastar module in this repo pulls in a whole raft of dependencies (e.g. templ, gostar, go-chi, nats, etc) including all the stuff required for the example apps - 66 deps in total as of this writing.
For projects that wouldn't otherwise rely on any of these dependencies, that's a lot of code pulled in just to get access to the datastar-specific helpers for constructing SSE events.
Moving example applications to their own modules would help significantly, but I'd also suggest moving any template-/component-library specific code out of the main lib, and into their own separate modules so that users that choose other solutions don't have to pull in those dependency trees.
I don't see much good reason for depending on anything outside the stdlib in main module TBH - functions like BodySanitize(), BodyUnmarshal() and QueryStringUnmarshal() are probably better left to documentation, as they're all both trivial and need not necessarily be implemented using the libraries chosen here.
I mostly agree with you. There are some people that are going to contribute helpers for other languages. I think we'll be moving them into a subfolder anyways and it could have it's own go.mod. I will say that the deps for gostar/templ/chi are less of a concern that's to tree shaking.
It's true that a lot of the deps get shaken out, but templ/gostar etc end up included because of their render methods in the main module.
right, I :heart: templ and I made templ, i'm ok with these deps for rendering. still will get tree shook at the final binary level depending of what you use. But moving to it's own dir in helpers/go/datastar/datastar.go makes a lot of sense
|
gharchive/issue
| 2024-10-19T04:06:30 |
2025-04-01T04:33:58.413260
|
{
"authors": [
"delaneyj",
"pdf"
],
"repo": "delaneyj/datastar",
"url": "https://github.com/delaneyj/datastar/issues/162",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
166710045
|
Use of Undeclared type 'TimeRange'
`func audioPlayer(audioPlayer: AudioPlayer, didLoadRange range: TimeRange, forItem item: AudioItem) {
}`
This delegate method has a TimeRange type. Not sure what is the datatype to be declared in the ViewController.
You should use AudioPlayer.TimeRange
|
gharchive/issue
| 2016-07-21T00:07:36 |
2025-04-01T04:33:58.414595
|
{
"authors": [
"colonelsshanslanda",
"delannoyk"
],
"repo": "delannoyk/AudioPlayer",
"url": "https://github.com/delannoyk/AudioPlayer/issues/64",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
559709163
|
get_hull_area and sum are unused
The two functions: get_hull_area and sum seem to be unused.
They're used for test.
|
gharchive/issue
| 2020-02-04T13:09:31 |
2025-04-01T04:33:58.416535
|
{
"authors": [
"abellgithub",
"vchizhov"
],
"repo": "delfrrr/delaunator-cpp",
"url": "https://github.com/delfrrr/delaunator-cpp/issues/25",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
819070485
|
[stable/cluster-overprovisioner] Add service account to cluster-overprovisioner chart
Description
Add ability to use a non-default service account for the cluster-overprovisioner
Checklist
[X] Title of the PR starts with chart name (e.g. [stable/mychartname])
[x] Github actions are passing
Hi @joshuastern, thanks for the PR. Can you fix the docs? Details are here: https://github.com/deliveryhero/helm-charts#contributing
Also, I'm curious, why does cluster-overprovisioner need a serviceAccount? It just runs and does nothing, doesn't need to connect to the k8s API etc?
Hi @joshuastern, thanks for the PR. Can you fix the docs? Details are here: https://github.com/deliveryhero/helm-charts#contributing
Also, I'm curious, why does cluster-overprovisioner need a serviceAccount? It just runs and does nothing, doesn't need to connect to the k8s API etc?
Exactly, it does not need a service account token. That's why we intend to set the automountServiceAccountToken to false when installing the chart. I could set the default value for this to false, but it seems better to keep it consistent with the default kubernetes behavior.
Exactly, it does not need a service account token. That's why we intend to set the automountServiceAccountToken to false
I still don't get the point though. You just want to have the pods running without the token?
Exactly, it does not need a service account token. That's why we intend to set the automountServiceAccountToken to false
I still don't get the point though. You just want to have the pods running without the token? Therefore you create a new serviceAccount so that you can set automountServiceAccountToken to false?
Yup. In its current state, it runs with a token for the default service account.
@joshuastern the CI test for markdown lint has started failing for some reason. I will investigate.
I opened an issue here: https://github.com/avto-dev/markdown-lint/issues/13
|
gharchive/pull-request
| 2021-03-01T17:07:00 |
2025-04-01T04:33:58.430206
|
{
"authors": [
"joshuastern",
"max-rocket-internet"
],
"repo": "deliveryhero/helm-charts",
"url": "https://github.com/deliveryhero/helm-charts/pull/119",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1476744140
|
[stable/k8s-event-logger] Added securityContext and podSecurityContext
Description
Added the ability to add securityContext and podSecurityContext to the deployment
Checklist
[x] Title of the PR starts with chart name (e.g. [stable/mychartname])
[x] I have read the contribution instructions, bumped chart version and regenerated the docs
[x] Github actions are passing
@max-rocket-internet could you please take a look?
|
gharchive/pull-request
| 2022-12-05T14:52:55 |
2025-04-01T04:33:58.432529
|
{
"authors": [
"pierluigilenoci"
],
"repo": "deliveryhero/helm-charts",
"url": "https://github.com/deliveryhero/helm-charts/pull/419",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1612607416
|
Update goscaleio to v1.10.0
Description
Update goscaleio to v1.10.0
GitHub Issues
List the GitHub issues impacted by this PR:
GitHub Issue #
https://github.com/dell/csm/issues/583
Checklist:
[x] I have performed a self-review of my own code to ensure there are no formatting, vetting, linting, or security issues
[x] I have verified that new and existing unit tests pass locally with my changes
[x] I have not allowed coverage numbers to degenerate
[x] I have maintained at least 90% code coverage
[ ] I have inspected the Grafana dashboards to verify the data is displayed properly
[ ] I have commented my code, particularly in hard-to-understand areas
[ ] I have made corresponding changes to the documentation
[ ] I have added tests that prove my fix is effective or that my feature works
[ ] I have maintained backward compatibility
How Has This Been Tested?
Please describe the tests that you ran to verify your changes. Please also list any relevant details for your test configuration
[ ] make check test
❯ make check test
./scripts/check.sh ./cmd/... ./opentelemetry/... ./internal/...
=== Checking format...
=== Finished
=== Vetting...
=== Finished
=== Linting...
=== Finished
=== Running gosec...
=== Finished
go test -count=1 -cover -race -timeout 30s -short ./...
? github.com/dell/karavi-metrics-powerflex/cmd/metrics-powerflex [no test files]
? github.com/dell/karavi-metrics-powerflex/internal/k8s/mocks [no test files]
? github.com/dell/karavi-metrics-powerflex/internal/service/mocks [no test files]
? github.com/dell/karavi-metrics-powerflex/opentelemetry/exporters [no test files]
? github.com/dell/karavi-metrics-powerflex/opentelemetry/exporters/mocks [no test files]
ok github.com/dell/karavi-metrics-powerflex/internal/entrypoint 2.922s coverage: 90.2% of statements
ok github.com/dell/karavi-metrics-powerflex/internal/k8s 0.089s coverage: 91.9% of statements
ok github.com/dell/karavi-metrics-powerflex/internal/service 0.801s coverage: 94.1% of statements
Manual inspection of the GUI
I have verified that the dashboards show the data properly while generating I/O and storage resources
[ ] Yes
[x] No
run e2e test
run e2e test
|
gharchive/pull-request
| 2023-03-07T02:45:03 |
2025-04-01T04:33:58.438326
|
{
"authors": [
"baoy1"
],
"repo": "dell/karavi-metrics-powerflex",
"url": "https://github.com/dell/karavi-metrics-powerflex/pull/141",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
867820650
|
Updated installer and documentation for powerstore metrics
Description
Updated the installer and documentation to support the PowerStore metrics service. The installer will now detect which CSI Drivers are installed and only enable those metrics services.
Testing
Successfully tested the installer to deploy the PowerStore metrics service.
Issues
List the issues impacted by this PR:
Issue ID
https://github.com/dell/karavi-observability/issues/48
Checklist:
[x] I have performed a self-review of my own changes.
@tdawe: Thank you for submitting your pull request!
Please make sure you have followed our contributing guidelines. We will review it as soon as possible.
We want your feedback! If you have any questions or suggestions regarding our contributing process/workflow, please reach out to us at karavi@dell.com.
|
gharchive/pull-request
| 2021-04-26T15:19:09 |
2025-04-01T04:33:58.442152
|
{
"authors": [
"ioeicicdsvc",
"tdawe"
],
"repo": "dell/karavi-observability",
"url": "https://github.com/dell/karavi-observability/pull/49",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2061176428
|
Proposal to create logical plans for operations
Description
Propose further work that I'd like to perform regarding the creation reusable logical relations. Also helps with identifying relations we would need with substrait.
Delta Find Files
Purpose: Identify files that contain records that satisfy a predicate.
This relation will generate a record batch stream with a single column called path. path will then map to an Add action in the Delta table.
This relation will also maintain a list of files that satisfy the predicate which can be passed sideways to relations downstream.
Delta Scan
Purpose: Scan the Delta Table
Update DeltaScan to take an optional input stream that contains paths of files to be scanned. This will enable DeltaScan to consume output of DeltaFindFile.
Currently when using find files, we must wait for the entire operation to complete and then we build the scan. The change enables Delta Scan to start when the first candidate file is identified.
I think this will require some significant work since it will involve refactoring the current DeltaScan implementation.
Delta Write
Purpose: Write records to storage, conflict resolution, and commit creation
Takes an single input stream of data that matches that tables schema and creates Add actions for each new file.
Information can be passed sideways to include additional delta actions to add to the commit. E.G DeltaDelete can provide a stream of Remove actions.
Delta Delete
Purpose: Delete Records from the table.
Given a predicate delete records from the Delta table.
Delta Delete can take an optional stream of records and will output records that do NOT satisfy the predicate.
It will maintain a stream of Remove actions can be passed sideways to other operations downstream.
The input stream is optional since there are cases where delete determine which files to remove without a need for a scan. An optimization phase can help determine when this is the case.
Diagram
High level diagram of how these relation will connect.
┌───────────────────────┐
│ Delta Find Files │
│ │
│ Predicate: │
┌───┤ Version: │
│ │ │
│ └──────────┬────────────┘
│ │
│ ▼
Files │ ┌───────────────────────┐
Matched │ │ Delta Scan │
List │ │ │
│ │ Version: │
│ │ │
│ │ │
│ └──────────┬────────────┘
│ │
│ ▼
│ ┌───────────────────────┐
└──►│ Delta Delete │
│ │
│ Predicate: │
┌───┤ │
│ └──────────┬────────────┘
Remove │ │
Actions │ ▼
│ ┌───────────────────────┐
│ │ Delta Write │
└──►│ │
│ │
│ │
└───────────────────────┘
Converting the ReplaceWhere operation to a logical view can look something like this
┌───────────────────────┐
│ Delta Find Files │
│ │
│ Predicate: │
┌───┤ Version: │
│ │ │
│ └──────────┬────────────┘
│ │
│ ▼ ┌────────────────────────────┐
Files │ ┌───────────────────────┐ │ Data Source │
Matched │ │ Delta Scan │ │ │
List │ │ │ │ │
│ │ Version: │ └────────────┬───────────────┘
│ │ │ │
│ │ │ ▼
│ └──────────┬────────────┘ ┌────────────────────────────┐
│ │ │ Delta Constraint Check │
│ ▼ │ │
│ ┌───────────────────────┐ └────────────┬───────────────┘
└──►│ Delta Delete │ │
│ │ │
│ Predicate: │ │
┌───┤ │ │
│ └──────────┬────────────┘ │
Remove │ │ │
Actions │ └────────────────┐ ┌─────────────┘
│ ▼ ▼
│ ┌──────────────────────────────────────────────────────────┐
│ │ Union │
│ │ │
│ └─────────────────────────┬────────────────────────────────┘
│ │
│ ▼
│ ┌───────────────────────┐
│ │ Delta Write │
└──────────────────────┤ │
│ │
│ │
└───────────────────────┘
Use Case
Once we have logical plans for Update and Delete we can expose new Datafusion SQL statements for them
May help with reuse of Delete & Update other for logical plans.
Related Issue(s)
related #1921
related #1984
Great writeup @Blajda - I am a great fan of going the logical plan route for all our operations.
In another repo, I have been experimenting a bit with the end to end flow starting with parsing delta-specific SQL and will upstream some of these changes soon. Reading through your proposals, it seems the "find files" plan would essentially contain the logic from kernel and the concept of a Snapshot. Some of the work I have been doing around moving the state management to Arrow RecodBatch'es touches some of that I think.
Essentially the plan is to make out state less eager and hopefully improve processing performance along the way. The logic is as follows.
create a LogSegment - a list of commit and checkpoint files relevant for a given version.
On creation of a Snapshot replay the Log, to find the Protocol and Metadata actions ASAP - discard everything else.
I recently discovered some updates in Datafusion around NDJSON reading, that we should leverage.
on replay / scan - essentially whenever a file list is requested - "replay" (find relevant files based on log reconciliation rules / predicate
I think this would pay into the proposed DeltaFindFiles and DeltaScan operations and also support the described steam processing to start.
As a "corollary" of this work we also have some progress on a much improved parquet reader the supports selective reading of leaf columns (also for nested structs) as well as more fine-granulary casting the schema - i.e. support schema evolution.
Good news is, recently we have been consolidating the used APIs from the current DeltaTableState to become a snapshot. With the removal of parquet2 support in #1995 we will also greatly simplify log parsing by using the existing arrow integration in parquet crate, and eleminating the need to maintain manual parsing.
Implicity your proposal also moves us to keep data much longer as RecordBatches and avoid crating actual Add, Remove structs etc. which I strongly support as well :).
As you said, this will be a lot of work, but also generate a lot of impact!
As a "corollary" of this work we also have some progress on a much improved parquet reader the supports selective reading of leaf columns (also for nested structs) as well as more fine-granulary casting the schema - i.e. support schema evolution.
Is this work available anywhere? To start I want to update DeltaScans implement to manage parquet reading itself instead of depending on depending entirely on Datafusion's parquet reader. We have multiple bugs caused by non-compliant writers (spark) that write timestamp info using Int96. This can be resolved by having an 'adapter' to conform data to the correct type and can help with 'Big' datatypes that are also used in the ecosystem. The 'adapter' will also align with schema evolution since not every parquet file will have the same physical schema.
I'm aligned with changes being made for LogSegment and Snapshot. Something that might beneficial is to have some sort of local 'registry'/'cache' that given a physical name / 'table handle' and version returns a shared structure in memory to reduce our overall memory footprint and network calls. When building operators there are some cases were clone Snapshot to satisfy the borrow checker.
To clarify about operations needing to clone Snapshot. Operations are pinned to a particular version of the table and this data is read is multiple places however when the operation ends we return a table with a new version. The snapshot of the new table is cloned from the previous. Suppose each snapshot takes 10 units of memory we now may consume 20 units.
It would be neat if we can build some of 'linked list' where snapshots can share data and each version would just maintain a logical view of which segments they need.
So, part of #2095 I was suggesting was refactoring the builders to pass back their actions instead of doing everything in their own internal way. I'm in agreeance that moving to this method would help a lot for generalizing execution, but at a slightly higher level how do we think this would alter operation builders? Their public facing API obviously would not change, but internally they would pass back actions and an execution plan? Or does the execution plan also take care of actions and commit log updates?
@hntd187 The execution plan would also take care of the actions and commit log updates at this point. In the above diagrams this would have been done in the Delta write relation however we can consider splitting write and commit into two different relation if it makes it easier.
Well I suppose the logical plans don't need to care about the details of what has to happen with actions, but the physical ones will need to be able to pass that context along. I don't have a strong opinion about where the write and commit happen, just that you can compose the actions of various logical operators into a single planned write and commit. Hopefully that makes sense.
|
gharchive/issue
| 2023-12-31T22:01:42 |
2025-04-01T04:33:58.462919
|
{
"authors": [
"Blajda",
"hntd187",
"roeap"
],
"repo": "delta-io/delta-rs",
"url": "https://github.com/delta-io/delta-rs/issues/2006",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2515086715
|
[BUG]
Delta Error Log 09/09/2024 04:56:28 p. m.
File Name: C:\Users\Zero\AppData\Roaming\delta_core\error_logs\ERROR_LOG_82588152.9382.txt
HResult: -2147467259
Exception Data: System.Collections.ListDictionaryInternal
Causing Func: GetModuleInfos
Target Site: System.Diagnostics.ModuleInfo[] GetModuleInfos(Int32, Boolean)
Outer Exception Message: Los procesos de 32 bits no pueden tener acceso a los módulos de un proceso de 64 bits.
Outer Exception Source: System
Exception as String: System.ComponentModel.Win32Exception (0x80004005): Los procesos de 32 bits no pueden tener acceso a los módulos de un proceso de 64 bits.
en System.Diagnostics.NtProcessManager.GetModuleInfos(Int32 processId, Boolean firstModuleOnly)
en System.Diagnostics.NtProcessManager.GetFirstModuleInfo(Int32 processId)
en System.Diagnostics.Process.get_MainModule()
en Delta.XWindow.Inject()
en Delta.XWindow.Button_Click_7(Object sender, RoutedEventArgs e)
Hello @ZeroKyojuro are you using the Delta Lake project - https://delta.io/
This is the source code for delta lake project, and it does not look like from your error that it is related.
|
gharchive/issue
| 2024-09-09T23:13:48 |
2025-04-01T04:33:58.467907
|
{
"authors": [
"ZeroKyojuro",
"tdas"
],
"repo": "delta-io/delta",
"url": "https://github.com/delta-io/delta/issues/3659",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1720967807
|
make FixedPoint hashable and use for dictionaries
Addresses #428
Codecov Report
Patch coverage: 90.38% and project coverage change: +0.01 :tada:
Comparison is base (e122bf4) 86.74% compared to head (08af8e8) 86.75%.
:exclamation: Current head 08af8e8 differs from pull request most recent head d5a1683. Consider uploading reports for the commit d5a1683 to get more accurate results
Additional details and impacted files
@@ Coverage Diff @@
## main #451 +/- ##
==========================================
+ Coverage 86.74% 86.75% +0.01%
==========================================
Files 32 32
Lines 4216 4220 +4
==========================================
+ Hits 3657 3661 +4
Misses 559 559
Flag
Coverage Δ
unittests
86.75% <90.38%> (+0.01%)
:arrow_up:
Flags with carried forward coverage won't be shown. Click here to find out more.
Impacted Files
Coverage Δ
elfpy/agents/agent.py
92.61% <ø> (ø)
elfpy/agents/get_wallet_state.py
87.03% <ø> (ø)
elfpy/agents/policies/single_long.py
84.37% <ø> (ø)
elfpy/markets/hyperdrive/hyperdrive_actions.py
87.13% <77.77%> (ø)
elfpy/agents/wallet.py
93.58% <86.95%> (ø)
elfpy/markets/borrow.py
84.74% <100.00%> (ø)
elfpy/markets/hyperdrive/hyperdrive_market.py
84.70% <100.00%> (ø)
elfpy/math/fixed_point.py
92.18% <100.00%> (+0.10%)
:arrow_up:
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Do you have feedback about the report comment? Let us know in this issue.
|
gharchive/pull-request
| 2023-05-23T01:19:46 |
2025-04-01T04:33:58.528136
|
{
"authors": [
"codecov-commenter",
"dpaiton"
],
"repo": "delvtech/elf-simulations",
"url": "https://github.com/delvtech/elf-simulations/pull/451",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2213655909
|
Security Roadmap Blogpost #2
[ ] Announce the results of the final audit
[ ] Make a decision on the Hyperdrive Bug Bounty (@jrhea, @CPSTL, @jalextowle)
Talk to Cantina Bounties (need to record this call)
Worst-case scenario we use security@delv.tech
Closing this since this was completed.
|
gharchive/issue
| 2024-03-28T16:32:49 |
2025-04-01T04:33:58.530193
|
{
"authors": [
"jalextowle"
],
"repo": "delvtech/hyperdrive",
"url": "https://github.com/delvtech/hyperdrive/issues/898",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
183209735
|
Question about patterns
Good Afternoon, I have a question about Builder pattern.
As we see an example last lesson, when we invoke builder, we write something like this:
final House house = new House.Builder().setTitle("").setDoors("").setSize(0).build();
I don't understand a little bit what is the advantage of using Builder, when we can just create getters/setters and write almost the same:
House house = new House();
house.setTitle("");
house.setDoors("");
house.setSize(0);
house.build();
Or is it all about higher level abstraction?
Main idea of Builder Pattern is that you are not using your constructors directly, that will do a Builder class for you, also more suitable code style.
final House house = new House.Builder().setTitle("").setDoors("").setSize(0).build();
Good profit of this pattern is that every "set" method returning a House object, so this give us a
opportunity to run new method with help of just dot( outside the Builder class). More compact code and understandably. Just a guess.
If i wrong, correct me please.
Correct
|
gharchive/issue
| 2016-10-15T13:27:58 |
2025-04-01T04:33:58.565511
|
{
"authors": [
"AlexTheKing",
"alexdzeshko",
"chacha1337"
],
"repo": "deniotokiari/training-epam-2016",
"url": "https://github.com/deniotokiari/training-epam-2016/issues/129",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1226873518
|
Active Users Rename as Online
Online chart on admin dashboard should be renamed as online instead of active.
Text is changed.
|
gharchive/issue
| 2022-05-05T15:54:14 |
2025-04-01T04:33:58.569460
|
{
"authors": [
"denizdikbiyik"
],
"repo": "denizdikbiyik/SWE574Project",
"url": "https://github.com/denizdikbiyik/SWE574Project/issues/290",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
60185931
|
Check for PreferenceInjector.inject call at compile time
It would improve UX to check that the user has properly called PreferenceInjector.inject and PreferenceInjector.stopListening (if necessary) in the target class if there are annotations present.
This may not actually be possible, especially when dealing with inheritance.
It turns out this is possible using java's Compiler Tree API. It has been implemented using a technique from this stackoverflow answer
This feature seems to have some compatibility issues. It really needs to work perfectly as it can cause unfixable compilation errors when it functions incorrectly.
I'm going to remove this feature pending extensive testing on a range of use cases and build cases.
|
gharchive/issue
| 2015-03-07T02:30:36 |
2025-04-01T04:33:58.571829
|
{
"authors": [
"denley"
],
"repo": "denley/PreferenceInjector",
"url": "https://github.com/denley/PreferenceInjector/issues/19",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2447181039
|
add: iter method to iterate with params
Summary
Add an iter method that accepts arguments to allow iterating through results.
Currently, there is no way to iterate providing args to bind to.
Symbol[Iterator] works for ones without any args.
This PR exposes the method as iter that works both with and without bind values.
Why
Without this, there's no way to iterate on large results efficiently. The current all methods and other variants always have to return the full query, which is not an ideal way to scan through the DB.
Alternative approaches
The current Symbol[iterator] works for items without bind params, but any bind params will be reset on execution. Can just remove the resetting, but in that case, it breaks compatibility and older behavior, since it's the responsibility of the user of the prepared statement to ensure it's also reset and bound.
This also breaks backwards compatibility as bound statements are no longer iterable. This worked before
const [row] = db.prepare(`SELECT sqlite_version() AS ver WHERE ver > ?`).bind(0)
But now it throws with:
error: Uncaught (in promise) Error: Statement already bound to values
if (this.#bound) throw new Error("Statement already bound to values");
^
at Statement.#bindAll (https://jsr.io/@db/sqlite/0.12.0/src/statement.ts:376:28)
at Statement.iter (https://jsr.io/@db/sqlite/0.12.0/src/statement.ts:703:18)
at iter.next (<anonymous>)
at file:///<snipped>.ts:4:8
at eventLoopTick (ext:core/01_core.js:175:7)
Ohh, there's a little detail I missed here - it's always resetting binding values. I'll push a fix and make a new release
|
gharchive/pull-request
| 2024-08-04T15:24:52 |
2025-04-01T04:33:58.615590
|
{
"authors": [
"DjDeveloperr",
"pitekantrop",
"prasannavl"
],
"repo": "denodrivers/sqlite3",
"url": "https://github.com/denodrivers/sqlite3/pull/133",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
623506092
|
Images does not load when viewing Third Party Modules
Noticed a strange thing when looking through some of the modules.
While looking at ink, I see the images does not load or that the loading seems to be broken. Like you see in the picture below.
This is a problem with ink's README. I submitted a PR to fix it, and now the problem becomes #809 😅
Part of #747
|
gharchive/issue
| 2020-05-22T22:31:50 |
2025-04-01T04:33:58.649581
|
{
"authors": [
"kidonng",
"lucacasonato",
"timharek"
],
"repo": "denoland/deno_website2",
"url": "https://github.com/denoland/deno_website2/issues/720",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2411106201
|
feat: add custom element name validator
These changes add a isValidCustomElement method to validate a custom element created in the HTML document. It is implemented according to the HTML spec that can be found at: https://html.spec.whatwg.org/multipage/custom-elements.html#valid-custom-element-name
I have also added tests for all the cases I could think of, if any case is missing please let me know and I can add it.
I was wondering about the name staring with is. I am not sure if that is a common name prefix for a function instead of a property. Maybe we could borrow from regexp test or check from the element api?
@timreichen hmmmm, I'm more inclined to do'testCustomElementIsValid(). Would that be okay? I don't want to leave the name too open to interpretations. The is pattern is pretty common in Ruby tho, I'm not against keeping it, although in JS/TS it's more common for properties as you said.
testValidCustomElement also works for me, I think it's a very descriptive and nice name.
I would put name also in the function name. How about testCustomElementName() or checkCustomElementNameValidity()?
I'd prefer the name isValidCustomElementName, which clearly shows the function is a predicate, returning boolean value.
I was wondering about the name staring with is. I am not sure if that is a common name prefix for a function instead of a property.
The function name starting with is is common in std. There are many examples:
isLeap
isUtcLeap
isStatus
isInformationalStatus
isSuccessfulStatus
isRedirectStatus
isClientErrorStatus
isServerErrorStatus
isErrorStatus
isAbsolute
isGlob
isRange
isSemVer
isNil
The function name starting with test has issues as discussed in https://github.com/denoland/std/pull/4364. I think we should avoid testXxx naming convention in new APIs.
Sorry to take long to answer guys!
|
gharchive/pull-request
| 2024-07-16T13:10:51 |
2025-04-01T04:33:58.659341
|
{
"authors": [
"kt3k",
"luk3skyw4lker",
"timreichen"
],
"repo": "denoland/std",
"url": "https://github.com/denoland/std/pull/5456",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2358589933
|
fix bug, exprs is empty when category_id = 0
fix bug, exprs is empty when category_id = 0
Description
Related Issue
Type of Change
[ ] 📚 Examples / docs / tutorials / dependencies update
[x] 🔧 Bug fix (non-breaking change which fixes an issue)
[ ] 🥂 Improvement (non-breaking change which improves an existing feature)
[ ] 🚀 New feature (non-breaking change which adds functionality)
[ ] 💥 Breaking change (fix or feature that would cause existing functionality to change)
[ ] 🔐 Security fix
Checklist
[ ] I've read the CODE_OF_CONDUCT.md document.
[ ] I've read the CONTRIBUTING.md guide.
[ ] I've updated the code style using make codestyle.
[ ] I've written tests for all new methods and classes that I created.
[ ] I've written the docstring in Google format for all the methods and classes that I used.
Thanks for the bug fix. Did you run make test to ensure all unit tests pass.
|
gharchive/pull-request
| 2024-06-18T01:01:45 |
2025-04-01T04:33:58.678187
|
{
"authors": [
"x281737830",
"zhiheng-huang"
],
"repo": "denser-org/denser-retriever",
"url": "https://github.com/denser-org/denser-retriever/pull/11",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
386099111
|
Kubernetes namespace display can be turned off
Description
Adds a config flag to the kubecontext section to control whether namespace is displayed or not.
By default, original behavior is conserved and display is on.
By setting SPACESHIP_KUBECONTEXT_NAMESPACE_SHOW=false in .zshrc it is possible to get rid of this display in case namespace is not something one is interested in. This also avoids running kubectl config view --minify --output 'jsonpath={..namespace}' command.
Screenshot
I forgot to update the documentation, will be done in a moment
There was issues with publishing v3.8.1 earlier, Corresponding commit wasn't pushed because of branch protection. This can be merged as soon as it's resolved.
|
gharchive/pull-request
| 2018-11-30T09:12:20 |
2025-04-01T04:33:58.680881
|
{
"authors": [
"nomaed",
"salmanulfarzy"
],
"repo": "denysdovhan/spaceship-prompt",
"url": "https://github.com/denysdovhan/spaceship-prompt/pull/575",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2395075030
|
Dependabot not closing "old" PRs after proposing updates for newer versions of the same dependency
Is there an existing issue for this?
[X] I have searched the existing issues
Package ecosystem
pip and others
Package manager version
N/A
Language version
N/A
Manifest location and content before the Dependabot update
dependencies/python/checkov.txt
dependabot.yml content
https://github.com/super-linter/super-linter/blob/main/.github/dependabot.yml
Updated dependency
Happens with several dependencies.
Example: checkov
From:
What you expected to see, versus what you actually saw
Dependabot should close PRs that propose old versions of the same dependencies.
The example is about pip and python, but happens with other ecosystems as well.
This looks like a regression.
Thanks for your support!
Native package manager behavior
Not needed because the dependency updates are proposed as expected. The issue is that Dependabot doesn't close old PRs about the same dependency.
Images of the diff or a link to the PR, issue, or logs
https://github.com/super-linter/super-linter/pull/5848
https://github.com/super-linter/super-linter/pull/5831
https://github.com/super-linter/super-linter/pull/5803
https://github.com/super-linter/super-linter/pull/5782
Smallest manifest that reproduces the issue
version: 2
updates:
- package-ecosystem: "pip"
commit-message:
prefix: "deps(python)"
directory: "/dependencies/python/"
schedule:
interval: "weekly"
open-pull-requests-limit: 100
/cc @zkoppert
/cc @Hanse00
It doesn't appear in my case with maven. Maybe it's only related to python or with specific options
My config:
version: 2
updates:
- package-ecosystem: "maven"
directory: "/"
schedule:
interval: "weekly"
In our case, it also happens for npm, for example:
https://github.com/super-linter/super-linter/pull/5854
https://github.com/super-linter/super-linter/pull/5833
https://github.com/super-linter/super-linter/pull/5795
And for ruby:
https://github.com/super-linter/super-linter/pull/5827
https://github.com/super-linter/super-linter/pull/5778
Happening for me as well with pip and terraform
I think this issue is duplicate of #10103
I think this issue is duplicate of #10103
It's not. Here, the case is about not closing old PR, but well creating new one. The issue you mentionned is only about not creating new PR
I'm seeing the same with docker.
Facing the same issue on dependabot. Old PR's are not getting closed after proposing updates for newer versions of the same dependency.
Seeing the same issue for gradle, but only for one of two repositories:
Repo1 has build.gradle in the repository root. This repo is not affected by this bug.
Repo2 has build.gradle in a subdirectory. This repo is affected by this bug.
Maybe there's a connection?
This is fixed, but closing as a duplicate of https://github.com/dependabot/dependabot-core/issues/10103
This should be fixed now, I was able to reproduce the problem and verify the fix.
The Dependabot service had a bug preventing superseding but it only affected non-root directories when wildcards weren't in use (e.g. **/*).
Going forward, PRs should be superseded correctly. You're welcome to leave them open to verify yourself that it's solved.
Thanks for you patience!
@soeren-glimm-insurfox yep, you were right!
|
gharchive/issue
| 2024-07-08T09:06:07 |
2025-04-01T04:33:59.099701
|
{
"authors": [
"Elikill58",
"bryanhorstmann",
"ferrarimarco",
"jakecoffman",
"oburgosm",
"poojadaine",
"rileygriffith",
"soeren-glimm-insurfox"
],
"repo": "dependabot/dependabot-core",
"url": "https://github.com/dependabot/dependabot-core/issues/10169",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
98709385
|
cache:clear --no-warmup fails
All seems well until I get to deploy:vendors, using the default Symfony recipe I get this output
Generating optimized autoload files
Script Sensio\Bundle\DistributionBundle\Composer\ScriptHandler::clearCache handling the post-install-cmd event terminated with an exception
[RuntimeException]
An error occurred when executing the "'cache:clear --no-warmup'" command.
Exception trace:
() at /var/www/vhosts/deployer_test/releases/20150803094915/vendor/sensio/distribution-bundle/Sensio/Bundle/DistributionBundle/Composer/ScriptHandler.php
:437
Sensio\Bundle\DistributionBundle\Composer\ScriptHandler::executeCommand() at /var/www/vhosts/deployer_test/releases/20150803094915/vendor/sensio/distribu
tion-bundle/Sensio/Bundle/DistributionBundle/Composer/ScriptHandler.php:138
Sensio\Bundle\DistributionBundle\Composer\ScriptHandler::clearCache() at phar:///usr/local/bin/composer/src/Composer/EventDispatcher/EventDispatcher.php:198
Composer\EventDispatcher\EventDispatcher->executeEventPhpScript() at phar:///usr/local/bin/composer/src/Composer/EventDispatcher/EventDispatcher.php:166
Composer\EventDispatcher\EventDispatcher->doDispatch() at phar:///usr/local/bin/composer/src/Composer/EventDispatcher/EventDispatcher.php:91
Composer\EventDispatcher\EventDispatcher->dispatchScript() at phar:///usr/local/bin/composer/src/Composer/Installer.php:342
Composer\Installer->run() at phar:///usr/local/bin/composer/src/Composer/Command/InstallCommand.php:131
Composer\Command\InstallCommand->execute() at phar:///usr/local/bin/composer/vendor/symfony/console/Symfony/Component/Console/Command/Command.php:257
Symfony\Component\Console\Command\Command->run() at phar:///usr/local/bin/composer/vendor/symfony/console/Symfony/Component/Console/Application.php:874
Symfony\Component\Console\Application->doRunCommand() at phar:///usr/local/bin/composer/vendor/symfony/console/Symfony/Component/Console/Application.php:195
Symfony\Component\Console\Application->doRun() at phar:///usr/local/bin/composer/src/Composer/Console/Application.php:146
Composer\Console\Application->doRun() at phar:///usr/local/bin/composer/vendor/symfony/console/Symfony/Component/Console/Application.php:126
Symfony\Component\Console\Application->run() at phar:///usr/local/bin/composer/src/Composer/Console/Application.php:82
Composer\Console\Application->run() at phar:///usr/local/bin/composer/bin/composer:43
require() at /usr/local/bin/composer:25
install [--prefer-source] [--prefer-dist] [--dry-run] [--dev] [--no-dev] [--no-plugins] [--no-custom-installers] [--no-autoloader] [--no-scripts] [--no-progress] [-v|vv|vvv|--
verbose] [-o|--optimize-autoloader] [--ignore-platform-reqs] [packages1] ... [packagesN]
I can see this is an error, but there is little to suggest what the actual issue is?
Try to run command manually: ... cache:clear --no-warmup and see if it does well.
It works from my regular app. Which is in a different dir. I am just trying to get this deployed elsewhere before making the switch.
I think its an issue to do with the dir I am placing my 'deployed' code into.
If I cd to the dir where I am deploying my code to and run the following commands;
cache:clear --no-warmup
php cache:clear --no-warmup
php app/console cache:clear --no-warmup
They all fail. But I am unsure what prefix, if any, I should be trying with cache:clear --no-warmup.
Thanks for your time and help.
Look like you need to fix it to pass well.
Fix what?
The console works fine for my app, its just not working from inside the deployed code directory.
Ou, hhhh. What kink of ssh impl you use?
Looks like this
server('prod', '1.2.3.4', 22)
->user('root')
->identityFile()
->stage('production')
->env('deploy_path', '/var/www/vhosts/deployer_test');
I found a number of times this was issues with parameters.yml.
Composer, when run through Deployer.org, when deploying a new version with no existing parameters.yml will just get the defaults from parameters.yml.dist, unlike when running manually where it asks you for the parameter values.
If these are not valid for your environment (wrong db user/pass, etc) then composer will fail like this.
Deployer doesn't do well at passing composer's true error message back to the user so it can be a pain (especially on new deployments) when things go wrong on the composer stage. I've found myself always having to start a deployment and then ctrl-c when it gets to installing vendors stage. I then ssh in and run it manually (using the exact composer command deployer does) this usually gives me the true error message and I can fix it pretty quickly.
Case in point, right now I'm deploying a new deployment. I make a habit of never putting DB credentials in the git repo. I always configure parameters.yml on the server itself. Just got
Script Sensio\Bundle\DistributionBundle\Composer\ScriptHandler::clearCache handling the post-install-cmd event terminated with an exception
[RuntimeException]
An error occurred when executing the "'cache:clear --no-warmup'" command.
On deployment. Went into the folder on the server, ran composer install and got
[Doctrine\DBAL\Exception\DriverException]
An exception occured in driver: SQLSTATE[HY000] [2003] Can't connect to MySQL server on '127.0.0.1' (111)
So I think some improvements can be made there. Also perhaps an option in symfony recipe that will run composer interactively so that it asks you what it should put in parameters.yml when it creates it.
Wow, thanks a lot for that. But...
If I SSH into my box and then run composer install inside the deployer_test dir from Terminal then I get the following error;
Composer could not find a composer.json file in /var/www/vhosts/the.battle.directory/deployer_test
To initialize a project, please create a composer.json file as described in the https://getcomposer.org/ "Getting Started" section
If I cd to release/app and run composer install then I get the same error. Despite composer.json being present inside that dir.
OK I have it running composer install now from within my deployer dir and the error is
Could not open input file: app/console
My console file is in bin, not app, so in my deploy.php file I added
set('bin_dir', 'bin);
But that seems to have made no difference and its still referencing console in app.
This has now morphed into a different and more specific issue.... here https://github.com/deployphp/deployer/issues/425
I know this is closed but I think it still is an issue, @AlexStansfield response was exactly the problem for me.
This will presumably happen to everyone deploying Symfony without their db credentials in the parameters.yml.dist file (which is best practice). Plus like Alex said the issue isn't obvious from Deployer's error output, whereas Composer's is.
@mcwebb since opening this ticket I have used Deployer again on another project and ran into the same issue. The error output needs to be improved in order to fix this, because as you say its best practice to deploy in this manner.
Yeah I would love to have this fixed.
My to new servers now seems to follow this workflow:
Deploy once and fail
Edit parameters.yml
Deploy second time and succeed
Not ideal
You also get another failure because of the need to type YES for the RSA key thing.
I also had the RSA fingerprint accept issue when pulling from a new git server.
And @AlexStansfield current workflow is what I ended up doing too, though I think you could possibly accomplish it through environment variables.
I think at the very least these issues should be documented.
I had the same issue today...
post-install-cmd: Incenteev\ParameterHandler\ScriptHandler::buildParameters
post-install-cmd: Sensio\Bundle\DistributionBundle\Composer\ScriptHandler::buildBootstrap
post-install-cmd: Sensio\Bundle\DistributionBundle\Composer\ScriptHandler::clearCache
Script Sensio\Bundle\DistributionBundle\Composer\ScriptHandler::clearCache handling the post-install-cmd event terminated with an exception
[RuntimeException]
An error occurred when executing the "'cache:clear --no-warmup'" command:
PHP Fatal error: Cannot use $this as parameter in /var/www/html/wwwdev.za-charts.portal/releases/20160830192302/var/cache/prod/appProdProjectContainer.php on line 4213
.
Exception trace:
() at /var/www/html/wwwdev.za-charts.portal/releases/20160830192302/vendor/sensio/distribution-bundle/Composer/ScriptHandler.php:307
Sensio\Bundle\DistributionBundle\Composer\ScriptHandler::executeCommand() at /var/www/html/wwwdev.za-charts.portal/releases/20160830192302/vendor/sensio/distribution-bundle/Composer/ScriptHandler.ph
p:138
Sensio\Bundle\DistributionBundle\Composer\ScriptHandler::clearCache() at phar:///usr/local/bin/composer/src/Composer/EventDispatcher/EventDispatche
I managed to get something working by changing the composer parameters by adding
the following line in my deploy.php
// Environment vars
env('composer_options', 'update --verbose --prefer-dist --optimize-autoloader --no-interaction');
Everything runs fine now, I know it's not wise to use 'composer update' but it did the trick
It seems that error of cache:clear --no-warmup
run composer install => event post-install => run all command, function, method in post-install-cmd of your composer.json file => error of cache:clear --no-warmup
I think you should try using composer install with option --no-script
I still have this issue, so it hasn't been fixed. Why was this closed, @jakenoble?!
I fixed the issue I was having but there are many reasons you might get this error. Try running deployed with -vvv to see the real error.
Creating a task to upload the correct database configuration fixed it for
me. You can also remove the non-interactive flag from composer on one of
the tasks running previously so you get asked for those db credentials.
On 26 Sep 2016 9:03 a.m., "Jake Noble" notifications@github.com wrote:
I fixed the issue I was having but there are many reasons you might get
this error. Try running deployed with -vvv to see the real error.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
https://github.com/deployphp/deployer/issues/423#issuecomment-249504340,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AIrCpGi_pEvikMnXK5NcG7sDNEtzcdDIks5qt3w8gaJpZM4Fkd5k
.
|
gharchive/issue
| 2015-08-03T09:52:41 |
2025-04-01T04:33:59.134284
|
{
"authors": [
"AlexStansfield",
"elfet",
"jakenoble",
"mcwebb",
"migueldemoura",
"oanhnn",
"sonnybrilliant"
],
"repo": "deployphp/deployer",
"url": "https://github.com/deployphp/deployer/issues/423",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2265807604
|
DD#0000: feat: Update cache prefix to use stage if availble for multi…
… server deploys
[x] Bug fix #3790?
[ ] New feature?
[ ] BC breaks?
[ ] Tests added?
[x] Docs added?
Please, regenerate docs by running next command:
$ php bin/docgen
Hi @antonmedv not sure what's going on here but the docgen think I need to make an update to docs/magento2.md but i've already done that locally. Is there any special config I need to apply?
Cheers
@peterjaap
Not sure if relevant for this PR, but there are restrictions to the prefix.
Invalid id or tag 'staging.mydomain.com_129_GLOBAL__DICONFIG' : must use only [a-zA-Z0-9_]
So probably good to add a regex to check the output?
|
gharchive/pull-request
| 2024-04-26T13:36:15 |
2025-04-01T04:33:59.138528
|
{
"authors": [
"antonmedv",
"barryvdh",
"valguss"
],
"repo": "deployphp/deployer",
"url": "https://github.com/deployphp/deployer/pull/3822",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
905863777
|
옵션 UI 적용 및 커스텀 메뉴 추가 API 적용.
커스텀 메뉴 추가 API의 경우 Response값 받아오는 부분까진 아직 적용을 못했습니다.
문서랑 스웨거에서 제가 못찾는지 없더라구요 ㅎㅎ
옵션 선택에 대한 제약사항또한 아직 추가하지 못했어요.
우선 푸시하고 내일 다시 작업하도록 하겠습니다.
고생하셨습니다 👍👍👍👍👍
|
gharchive/pull-request
| 2021-05-28T18:31:40 |
2025-04-01T04:33:59.146738
|
{
"authors": [
"HanweeeeLee",
"mini0212"
],
"repo": "depromeet/myRecipick_iOS",
"url": "https://github.com/depromeet/myRecipick_iOS/pull/26",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1868013542
|
🛑 Derailed Home is down
In 7b19075, Derailed Home (https://derailed.one) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Derailed Home is back up in cfd9017 after 326 days, 8 hours, 20 minutes.
|
gharchive/issue
| 2023-08-26T10:21:30 |
2025-04-01T04:33:59.157314
|
{
"authors": [
"VincentRPS"
],
"repo": "derailed-org/derailed-oss-status",
"url": "https://github.com/derailed-org/derailed-oss-status/issues/2251",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1829985087
|
🛑 Derailed Home is down
In e40b599, Derailed Home (https://derailed.one) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Derailed Home is back up in 4adfc0e.
|
gharchive/issue
| 2023-07-31T20:32:36 |
2025-04-01T04:33:59.159597
|
{
"authors": [
"VincentRPS"
],
"repo": "derailed-org/derailed-oss-status",
"url": "https://github.com/derailed-org/derailed-oss-status/issues/665",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2450384155
|
Ameerul / FEQ-2557 Advertiser seeing his own real-name when going to another advertiser's profile.
Changed logic to show your nickname only in MyProfile
Updated test cases for AdvertiserName
https://github.com/user-attachments/assets/b1df0fe2-1c99-4ddb-bf40-05fa248eb3a6
Pull Request Test Coverage Report for Build 10263825448
Details
0 of 0 changed or added relevant lines in 0 files are covered.
No unchanged relevant lines lost coverage.
Overall coverage remained the same at 85.439%
Totals
Change from base Build 10260249169:
0.0%
Covered Lines:
5713
Relevant Lines:
6220
💛 - Coveralls
|
gharchive/pull-request
| 2024-08-06T09:20:14 |
2025-04-01T04:33:59.173107
|
{
"authors": [
"ameerul-deriv",
"coveralls"
],
"repo": "deriv-com/p2p",
"url": "https://github.com/deriv-com/p2p/pull/256",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1670070866
|
image rendering without playwright
docker-compose-viz currently embeds the minified mermaid.min.js script along with the playwright library (headless browser automation). Playwright is however very heavy and doesn't work well inside a docker image.
This issue is about looking for a replacement, that is, a way to still render mermaid diagrams offline and from Kotlin, but with a lighter tool.
why not the official cli ? https://github.com/mermaid-js/mermaid-cli
Hello there!
Just a heads up: the latest release supports the official mermaid-js as well, provided you have npm installed on your system. Let me know what you think.
|
gharchive/issue
| 2023-04-16T18:02:38 |
2025-04-01T04:33:59.175118
|
{
"authors": [
"derlin",
"mgaitan"
],
"repo": "derlin/docker-compose-viz-mermaid",
"url": "https://github.com/derlin/docker-compose-viz-mermaid/issues/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1025836911
|
Cannot autodetect Frigate camera name
Hello,
I have completed setting up Frigate as Docker on Unraid with Frigate NVR Proxy and so far things are working. I am now trying to add in the Lovelace card. I am getting an error that the card cannot bring in the camera entity. I am unsure if this is a Frigate issue that I should raise there or if the configuration I am using with Docker on Unraid with the Frigate Proxy to HA is an issue. I have not yet tried to use the HA Frigate add-on directly. I am looking for some input here where I might go next to resolve my issue. Any help guidance is appreciated. Let me know if more info is needed than below. Thanks in advance!
Error Message on Card Configuration:
“Cannot autodetect Frigate camera name, you need to either set camera_entity and / or frigate_camera_name. “
Configuration:
Frigate installed as Docker on Unraid. Home Assistant CORE as VM with Frigate NVR Proxy.
I have setup one camera for my pilot. Camera works in Frigate from the Docker Web UI and also comes through to HA via the NVR Proxy.
Config.yml:
Frigate viewer. I am able to access events, clips, and recordings within the NVR.
Hello once again,
The last issue with the spinning save button was my own doing when I had also been playing with my yaml mode settings on Lovelace. When moving that back to UI mode I was able to save the card fine.
From my perspective this resolves my issue.
If you at all are looking to work on the proxy NVR config at some point I would be happy to help with any testing there. For now I will work with the configuration you have provided. For now I will close my question and thanks again so much for your help!
Hi @dlWhitaker -- filed a separate bug to track the work for the NVR proxy.
I still don't understand how you somehow didn't have any camera entities (neither in the card editor dropdown, nor in Home Assistant itself). That suggests the integration may not have been correctly installed.
Nevertheless, glad you got it to work out. Follow along on that other bug to track progress on the NVR proxy work when I, or someone else, gets around to it in the backlog!
|
gharchive/issue
| 2021-10-14T01:24:26 |
2025-04-01T04:33:59.180125
|
{
"authors": [
"dermotduffy",
"dlWhitaker"
],
"repo": "dermotduffy/frigate-hass-card",
"url": "https://github.com/dermotduffy/frigate-hass-card/issues/126",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1675488245
|
8 High Severity Vulnerabilities
When running npm i the returned audit looks a bit grim.
If I attempt to update your project, I could break it.
Why it matters:
As an influencer buying decision, I am noting that you are offering sample projects that are not maintained.
This could be viewed as a strike against the adoption of the commercial-grade product because it implies the risk that your practices on the commercial-grade product may be weak.
When I am evaluating an auth provider I look for signs like this.
--> Recommendation --->
Limit your sample apps to a ‘certified’ set and hold them at the same quality bar you hold your commercial-grade product offering.
Details
npm WARN deprecated stable@0.1.8: Modern JS already guarantees Array#sort() is a stable sort, so this library is deprecated. See the compatibility table on MDN: https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Array/sort#browser_compatibility
npm WARN deprecated rollup-plugin-terser@7.0.2: This package has been deprecated and is no longer maintained. Please use @rollup/plugin-terser
npm WARN deprecated w3c-hr-time@1.0.2: Use your platform's native performance.now() and performance.timeOrigin.
npm WARN deprecated svgo@1.3.2: This SVGO version is no longer supported. Upgrade to v2.x.x.
added 1553 packages, and audited 1554 packages in 29s
227 packages are looking for funding
run npm fund for details
8 high severity vulnerabilities
To address issues that do not require attention, run:
npm audit fix
To address all issues (including breaking changes), run:
npm audit fix --force
thanks for letting us know, we updated the deps and configured dependabot to keep us up to date going forward
|
gharchive/issue
| 2023-04-19T19:23:54 |
2025-04-01T04:33:59.185661
|
{
"authors": [
"missulmer",
"omercnet"
],
"repo": "descope-sample-apps/b2c-retail-sample-app",
"url": "https://github.com/descope-sample-apps/b2c-retail-sample-app/issues/32",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
119889504
|
Label suggestion
I'd like to suggest first timers only label, as seen in https://github.com/pouchdb/pouchdb/labels/first timers only. Don't know if it's widely used, but it should be. =)
See #56
duplicate, please close
Closing in favor of #56
|
gharchive/issue
| 2015-12-02T08:15:06 |
2025-04-01T04:33:59.210829
|
{
"authors": [
"despo",
"kahlin",
"leotm",
"xanderyzwich"
],
"repo": "despo/issuehub.io",
"url": "https://github.com/despo/issuehub.io/issues/69",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2553940888
|
Populate partialFingerprints for results in SARIF reports
Expected Behavior
partialFingerprints is populated for results in SARIF reports.
Current Behavior
No fingerprint is generated.
Context
SARIF report should provide all data that's "required" by GitHub.
This attribute is useful:
to avoid seeing duplicate alerts
to accurately identify which results are the same across commits and branches
Related with #7661
This link is very interesting to know how to implement this https://docs.oasis-open.org/sarif/sarif/v2.1.0/errata01/os/sarif-v2.1.0-errata01-os-complete.html#_Toc141791182
To me this is the most important part:
It is difficult to devise an algorithm that constructs a truly stable fingerprint for a result. Fortunately, for practical purposes, the fingerprint does not need to be absolutely stable; it only needs to be stable enough to reduce the number of results that are erroneously reported as “new” to a low enough level that the development team can manage the erroneously reported results without too much effort.
We should be aware that it's impossible to create a bullet proof implementation for the partialFingerprint.
|
gharchive/issue
| 2024-09-28T01:31:47 |
2025-04-01T04:33:59.227291
|
{
"authors": [
"3flex",
"BraisGabin"
],
"repo": "detekt/detekt",
"url": "https://github.com/detekt/detekt/issues/7683",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2218547001
|
Installation error
Hi dev,
First of thank for creating such a web based beautiful tool. Im very curious to download the tools,
Actually I never used docker before, it little bit tough learn quickly, During installation I have so many doubt and error.
Could you guide how to download the tools that would be better.
Even though I followed as you instructed but still im facing issue.
Pls kindly help me on this.
Hi, sorry for getting back to you late. I've never encountered this error before, and after some research, it seems that the problem is caused by a bug from Docker on Windows. For further details, refer to this GitHub issue: https://github.com/docker/for-win/issues/13611.
Hi Dev,
I newly installed latest docker in my windows machine, but still I'm facing issue, is there specific docker version that you recommend.
Hi Dev,
How long it will take, for the last half hours it will be same 18/21
NPM error
Hi danger112233,
please try it again using the following command docker-compose build --no-cache
|
gharchive/issue
| 2024-04-01T15:51:57 |
2025-04-01T04:33:59.244078
|
{
"authors": [
"danger112233",
"dev-lu"
],
"repo": "dev-lu/osint_toolkit",
"url": "https://github.com/dev-lu/osint_toolkit/issues/31",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
374363091
|
Thias/puppet-sysctl has not be updated in over two years.
Thias/puppet-sysctl was last update Feb 5th 2016
The thias/puppet-sysctl module (https://github.com/thias/puppet-sysctl) has seen no update in a long time and @Thias does not seem to be active any more. The duritong/puppet-sysctl module (https://github.com/duritong/puppet-sysctl) is seeming to be updated more frequently and is implemented as an type/provider. Would you accept and a pull request that changes the dependency to duritong/puppet-sysctl instead of thias/puppet-sysctl?
Change would be to the metadata.json and change
sysctl
To
sysctl::value
@enemarke Sure we accept a PR for this, thanks in advance!
My plan was to do the change, but in the mean time the hercules-team/augeasproviders_sysctl module has made a new release and gotten the approved stamp in forge.puppet.com. If the change needs to be done I think that the change should be to the hercules-team module instead. I will do a bit more investigation before committing a pull req.
Fixed with #183
|
gharchive/issue
| 2018-10-26T12:32:36 |
2025-04-01T04:33:59.250603
|
{
"authors": [
"enemarke",
"mcgege"
],
"repo": "dev-sec/puppet-os-hardening",
"url": "https://github.com/dev-sec/puppet-os-hardening/issues/159",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
204090287
|
Add argument to determine how alert was closed
Issue Description
No way to determine how the alert was closed (timeout, press, pan)
Steps to Reproduce / Code Snippets
Open alert
Close it
Only arguments received to onClose are type, title, message
Expected Results
Event function should receive an additional prop for method - timeout, press or pan
Additional Information
React Native version: 0.40.0
react-native-dropdownalert version: 2.4.0
Platform(s) (iOS, Android, or both?): both
I like it as it's quite simple but this might beg the question if it needs to go a little further and make separate prop callback functions for each type. That way the receiver can do specific logic for whatever type it is. What you think?
I think that would be better and have done so :)
I was also looking for a way to attach data to the notification so I can know more about the context of the pop up (i"m using this to handle notifications in the foreground)
|
gharchive/issue
| 2017-01-30T18:24:13 |
2025-04-01T04:33:59.254343
|
{
"authors": [
"devBrian",
"morrislaptop"
],
"repo": "devBrian/react-native-dropdownalert",
"url": "https://github.com/devBrian/react-native-dropdownalert/issues/31",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1435509679
|
Github Action으로 npm package, github package 자동화되도록 설정하기
상기 작업을 시간 관계상 완료하지 못함.
npm package 퍼블리시는 CLI 환경에서 했을 때 완벽히 동작하며,
Github package는 Github Action을 통해 완벽히 동작함을 확인함.
다만 npm package 퍼블리시가 Github Action을 통해 올라가지 않는 문제가 있음.
추후 해결 필요.
npm package 퍼블리시 파이프라인은 개발 완료됨.
|
gharchive/issue
| 2022-11-04T03:52:15 |
2025-04-01T04:33:59.263594
|
{
"authors": [
"devcomfort"
],
"repo": "devcomfort/-devcomfort-text-transcoder",
"url": "https://github.com/devcomfort/-devcomfort-text-transcoder/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
432851869
|
每日一问-go与跨平台
很多时候我们发现,从底层rpc网络框架到上层业务逻辑,都可以用纯go或依赖库来实现,如果在windows办公环境下,我们不在需要像以前写c/c++时登录到专门的linux开发机或者本地弄一个vagrant环境,而是直接本地ide开撸,就像写java一样,甚至设置下GOOS和GOARCH真的能够交叉编译。显然go比c更像是跨平台的语言,但go并没有虚拟机,所以跨平台能力应该是不及java,那么go是跨平台的吗?
Mac 下编译 Linux 和 Windows 64位可执行程序
CGO_ENABLED=0 GOOS=linux GOARCH=amd64 go build main.go
CGO_ENABLED=0 GOOS=windows GOARCH=amd64 go build main.go
Linux 下编译 Mac 和 Windows 64位可执行程序
CGO_ENABLED=0 GOOS=darwin GOARCH=amd64 go build main.go
CGO_ENABLED=0 GOOS=windows GOARCH=amd64 go build main.go
Windows 下编译 Mac 和 Linux 64位可执行程序
SET CGO_ENABLED=0
SET GOOS=darwin
SET GOARCH=amd64
go build main.go
SET CGO_ENABLED=0
SET GOOS=linux
SET GOARCH=amd64
go build main.go
From: https://blog.csdn.net/panshiqu/article/details/53788067
:trident: 这俩感觉不咋冲突诶,各有各的好处吧,像楼上发的交叉编译解决方案就很好呢
|
gharchive/issue
| 2019-04-13T14:15:08 |
2025-04-01T04:33:59.270846
|
{
"authors": [
"LIYINGZHEN",
"guoruibiao",
"woofyzhao"
],
"repo": "developer-learning/reading-go",
"url": "https://github.com/developer-learning/reading-go/issues/347",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
230784266
|
build problem with sass/scss
Hello.
It is just not working.
I've tried .scss, .sass files, but builder fails with error message
Template execution failed: Error: Cannot find module './style' ...
.less files with (and without) -l option is working fine
Am I doing something wrong?
Hmm - there were some folks working on a scss template, I wonder if they ran into this as well?
@stasusov do you have style/index.scss file (or .sass)? If so you can just change your import from import style from './style' to import style from './style/index.scss'
Keep in mind that prerendering won't work currently with SASS and you'll have to run build like so: preact build --no-prerender or npm run build -- --no-prerender
@rkostrzewski I have faced with the same issues as described in #59, so this issue can be closed as duplicate.
Thanks @stasusov ! We appreciate the report anyway 😄
|
gharchive/issue
| 2017-05-23T17:21:11 |
2025-04-01T04:33:59.276474
|
{
"authors": [
"developit",
"lukeed",
"rkostrzewski",
"stasusov"
],
"repo": "developit/preact-cli",
"url": "https://github.com/developit/preact-cli/issues/49",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
386189132
|
How to add custom polyfill?
Hey I wanted to ask what's the best approach to add custom polyfill (like IntersectionObserver)? I believe it should be added to html template file similar to fetch polyfill. The only difference is that such import is done on application side, not preact-cli webpack configuration. Any ideas?
Sorry I'm not very familiar with webpack.
We're looking to expore the html template very soon, so you could add it there
|
gharchive/issue
| 2018-11-30T13:32:22 |
2025-04-01T04:33:59.277642
|
{
"authors": [
"ForsakenHarmony",
"lchimaru"
],
"repo": "developit/preact-cli",
"url": "https://github.com/developit/preact-cli/issues/676",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
232473125
|
Spaces instead of tabs
I know this is a minor thing, but could we somewhere on the roadmap have a small configuration option to use spaces instead of tabs?
For now good old sed: sed -i.bak -E $'s/\t/ /' fileName.js
how about prettier?
@developit @kristoferbaxter @lukeed do we want linters and prettier to be added?
I may be alone on this, but I believe this is user land. Once #56 makes its way in, devs can add in whatever they'd like. And in the meantime, prettier can be added now on top of any install.
Maybe we could include some recipes to illustrate how to achieve some of these things, but then this becomes yet another bit to maintain. What do you think?
@lukeed #56 is based for the project created with preact-cli.
whereas i was talking about adding prettier to the preact-cli codebase itself, to lint this repo, not the ones created with it.
Oh, lol. But I think the issue was opened about the generated output?
ohh @JMSantos94 we need to ask you abt this 😛
but anyways how about our project @lukeed ?
Haha 😜
As per our codebase, that's not up to me & I don't really have input there.
@lukeed @prateekbh Sorry if I wasn't clear enough. Yes, I am referring to the output project when you use preact create. Prettier is a nice option, but I agree with @lukeed on it being on the user land.
Cool 😄 Is it okay to close this issue then?
I guess if it is complicated to do something like:
preact create some-project --indentation spaces
... or not worth the time.
Right now files aren't processed by preact create, just moved.
That said, I'd like to have preact create [type] actually go to Github to fetch the project, and scaffold it out from a real repo. If we do that, we can have much more varied project types. We'd also have the opportunity to transform incoming files, either manually or via eslint --fix / prettier / etc. Vue-cli uses a lib called download-git-repo, or we could use Yeoman which does transformations already.
Thanks for the insight @developit!
Yeah, .editorconfig is pretty awesome.
As I mentioned before, this is not a big deal. Would be nice to keep it in somewhere though 😄
Leaving this here regarding
https://twitter.com/JedWatson/status/875491536003186690
Sent from my OnePlus ONEPLUS A3003 using FastHub
lol - correlation VS causation there, python is used for data science and that's a high paying career right now :P
+1 for editorconfig
|
gharchive/issue
| 2017-05-31T07:06:35 |
2025-04-01T04:33:59.286037
|
{
"authors": [
"JMSantos94",
"developit",
"jeanleonino",
"lukeed",
"prateekbh",
"reznord"
],
"repo": "developit/preact-cli",
"url": "https://github.com/developit/preact-cli/issues/79",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
196696155
|
No d.ts file.
There isn't a typings file currently for preact-redux, which is a shame.
However it's only a micro layer around react-redux, which already has a typing, so it should be fairly easy to fix.
PR's welcome :)
See #7
Merged in #20 !
|
gharchive/issue
| 2016-12-20T15:09:50 |
2025-04-01T04:33:59.288122
|
{
"authors": [
"developit",
"mseddon"
],
"repo": "developit/preact-redux",
"url": "https://github.com/developit/preact-redux/issues/6",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
311810577
|
preview creates n+1 tiles
label-maker preview -n 3 generates 4 example tiles for each class, not 3.
probable fix:
preview.py line 60
if n > number: -> if n >= number
We should probably add some documentation about this. It uses unittest. You can test a single file like:
python -m unittest test/unit/test_validate.py
or a folder with
python -m unittest discover -v -s test/unit
Full options here
|
gharchive/issue
| 2018-04-06T00:14:38 |
2025-04-01T04:33:59.292968
|
{
"authors": [
"drewbo",
"jreiberkyle"
],
"repo": "developmentseed/label-maker",
"url": "https://github.com/developmentseed/label-maker/issues/63",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2396799443
|
Why "Input being reprojected to EPSG:4326 CRS"?
I'm plotting multiple layers of a Map and I get a bunch of warning saying:
Input being reprojected to EPSG:4326 CRS . Why is that warning showing up? I think a more explanatory message would be helpful.
That warning is shown whenever the input data contains a CRS that is not EPSG:4326. I added that warning because I believed it was useful to be explicit about reprojecting user data. Lonboard currently is only able to render data in EPSG:4326 projection
Thanks, that's very helpful. I think having a warning that is slightly more descriptive would help. Something like...
"Input being reprojected to EPSG:4326 CRS. At the moment lonboard is only able to render data in EPSG:4326"
Happy to make a PR at some point this week if you think some version of the expanded warning message is useful.
|
gharchive/issue
| 2024-07-09T00:00:36 |
2025-04-01T04:33:59.294892
|
{
"authors": [
"kylebarron",
"ncclementi"
],
"repo": "developmentseed/lonboard",
"url": "https://github.com/developmentseed/lonboard/issues/566",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1903054110
|
Replace Alizer io/ioutil dependency
Which area/kind this issue is related to?
/area alizer
Issue Description
At the moment alizer is using the io/ioutil in order to read files. As it is mentioned:
"io/ioutil" has been deprecated since Go 1.19: As of Go 1.16, the same functionality is now provided
by package io or package os, and those implementations should be preferred in new code. See the
specific function documentation for details
As a result, we shall replace this dependency and start using the os library.
Acceptance Criteria
[ ] Remove the io/ioutils dependency and replace it with a non-deprecated one.
[ ] Enforce the package is not used in the future using linter (run the linter into CI)
I can work on this
@trader7 Thank you for offering your contribution! I have assigned you this issue, keep us up to date on your progress. I recommend reviewing CONTRIBUTING.md as well before starting.
When your changes are ready you can open a PR to main branch of devfile/alizer.
I submitted a pull request in Alizer repo with Fixes #1257
Branch is alizer-ioutil-dependency
I submitted a pull request in Alizer repo with Fixes #1257 Branch is alizer-ioutil-dependency
Awesome, I'll review your changes.
Partial fix applied: https://github.com/devfile/alizer/pull/30
Re-opening for the addition of the linter check, I will assign myself to work on this part.
/assign @rm3l
/unassign @michael-valdron
Going to look into this issue in the upcoming sprint (2249)
|
gharchive/issue
| 2023-09-19T13:56:39 |
2025-04-01T04:33:59.300486
|
{
"authors": [
"michael-valdron",
"rm3l",
"thepetk",
"trader7"
],
"repo": "devfile/api",
"url": "https://github.com/devfile/api/issues/1257",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1175832244
|
🛑 Checkout is down
In 89b5c4a, Checkout (https://checkout.sportclub.com.ar/paso2/total-mensual) was down:
HTTP code: 502
Response time: 389 ms
Resolved: Checkout is back up in 47911c6.
|
gharchive/issue
| 2022-03-21T19:16:43 |
2025-04-01T04:33:59.313371
|
{
"authors": [
"devmuvi"
],
"repo": "devmuvi/uptime-monitor",
"url": "https://github.com/devmuvi/uptime-monitor/issues/16",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2645535439
|
🛑 Message Board is down
In edf0ee1, Message Board (https://msg-board-server.onrender.com/ping) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Message Board is back up in 7f17855 after 13 minutes.
|
gharchive/issue
| 2024-11-09T03:21:29 |
2025-04-01T04:33:59.315706
|
{
"authors": [
"devnullsans"
],
"repo": "devnullsans/upptime-monitor",
"url": "https://github.com/devnullsans/upptime-monitor/issues/385",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2699350863
|
🛑 Photoz ICVJ is down
In c419bcd, Photoz ICVJ (https://photoz-icvj.onrender.com/ping) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Photoz ICVJ is back up in b5b5ebc after 6 minutes.
|
gharchive/issue
| 2024-11-27T17:44:43 |
2025-04-01T04:33:59.318042
|
{
"authors": [
"devnullsans"
],
"repo": "devnullsans/upptime-monitor",
"url": "https://github.com/devnullsans/upptime-monitor/issues/447",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
56730689
|
Not able to select a file on Cordova
Hey there,
I've just deployed my app to my android device by running the command
meteor run android-device --mobile-server http://mydomain
Dropzonejs is working on every browser on my mobile phone I've tried however when I deploy my app as above, it just doesn't show me any choice for uploading. Do you have any idea why is this happening?
Best,
Hey,
I can't think of any reason why it won't work. Can you provide more info, do you get any errors anywhere? Are you trying to render a Dropzone programatically or using the template helper?
Hi,
I'm using the template helper. However, I've figured out that this is a Cordova problem which dropzone.js itself is also suffering since Cordova doesn't support file input tags as it should. So I'm trying to find out how can I achieve this with another kind of thing (there are image pickers for cordova ios but I've failed to find a plugin for android so far.)
In the end, it is not directly about your meteor plugin if you are not thinking to fix this in your template helper and this issue can be closed.
Many thanks for your attention.
Ah, thank you for the feedback! :smiley:
|
gharchive/issue
| 2015-02-05T21:13:35 |
2025-04-01T04:33:59.321087
|
{
"authors": [
"devonbarrett",
"suhaboncukcu"
],
"repo": "devonbarrett/meteor-dropzone",
"url": "https://github.com/devonbarrett/meteor-dropzone/issues/14",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1366888713
|
Additional resource (CTF course)
I've found this great platform with free courses on computing topics, called Roppers Academy. They have a course on CTF, which contains several topics with some tasks - https://www.roppers.org/courses/ctf. I would like you to add it, but I'm not sure under which category it should be put
@devploit what do you think?
Added. Thanks for the contribution!
|
gharchive/issue
| 2022-09-08T19:28:44 |
2025-04-01T04:33:59.346641
|
{
"authors": [
"Globart1337",
"devploit"
],
"repo": "devploit/ctf-awesome-resources",
"url": "https://github.com/devploit/ctf-awesome-resources/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1761318642
|
⚠️ Mobitopup has degraded performance
In 980e623, Mobitopup (https://napcard.voz.tech) experienced degraded performance:
HTTP code: 200
Response time: 6420 ms
Resolved: Mobitopup performance has improved in 0198c20.
|
gharchive/issue
| 2023-06-16T20:49:05 |
2025-04-01T04:33:59.350431
|
{
"authors": [
"devtdq1701"
],
"repo": "devtdq1701/upptime-nextify",
"url": "https://github.com/devtdq1701/upptime-nextify/issues/547",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1615619782
|
[DOC]: Add interview question to JSON file
We are collecting often asked interview questions for Web Developers. Before adding questions and answers, please make sure if the question is already added or not. Use the search option to search your questions Ctlr+F/ cmd+F
path src/json/interview/
Topics:
i would like to participate in it
Assigned @yugborana
Anyone can work on this issue
I have a bunch of questions that I would like to add , Can I ? @devvsakib
sure @ArslanYM
Merged #223
PR: @ArslanYM
I'd like to help add some questions!
Assigned @Jrobbins973
I have added a question based on react
PR #240
I have added question based on typescript
PR #275
Hello, I would like to add a Javascript question. Kindly assign me
I have added a question.
PR: @david-mbugua #304
Merged #304
PR: @david-mbugua
Hey @devvsakib I am ok with the topics listed in the issue. Could you please assign me and I would like to contribute some frequently asked interview questions.
Assigned @HemanthPaila
Hey @devvsakib I am getting an error that will push code into the repo.
remote: Permission to devvsakib/power-the-web.git denied to HemanthPaila.
fatal: unable to access 'https://github.com/devvsakib/power-the-web.git/': The requested URL returned error: 403
you should add your forked repo instead my repo. remove git origin URL and set your forked repository url
ok @devvsakib.
@devvsakib I made a PR #348 could you please review and merge it?
Hello, I would like to add a Javascript question. Kindly assign me
Assigned @vim12345
|
gharchive/issue
| 2023-03-08T17:09:59 |
2025-04-01T04:33:59.363446
|
{
"authors": [
"ArslanYM",
"HemanthPaila",
"Jrobbins973",
"amelia2802",
"david-mbugua",
"devraushan",
"devvsakib",
"vim12345",
"yugborana"
],
"repo": "devvsakib/power-the-web",
"url": "https://github.com/devvsakib/power-the-web/issues/221",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
87286736
|
Possible bug when using query {.prop === $somthing && .prop !== $else}
git clone https://github.yandex-team.ru/gist/9acc9792f0bd7ba68f7b.git jspath-demo && cd jspath-demo && npm i jspath && node ./demo.js
Here is the code to demonstrate: https://github.yandex-team.ru/gist/arikon/9acc9792f0bd7ba68f7b
My fault. Thanks @dfilatov!
|
gharchive/issue
| 2015-06-11T09:55:40 |
2025-04-01T04:33:59.378002
|
{
"authors": [
"arikon"
],
"repo": "dfilatov/jspath",
"url": "https://github.com/dfilatov/jspath/issues/37",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1488433415
|
Typo await should be await*
https://github.com/dfinity/motoko/blob/c8c0917052eda95669dabd612834d66cba5a19ec/doc/md/language-manual.md?plain=1#L2249
Should this be "Expression await* <exp> evaluates ..."?
And a newline before the paragraph?
And in the subsequent sentences
"If r is trap, evaluation returns trap. Otherwise r is a delayed computation"
we only know if it is a trap when the computation is attempted, or? The word "otherwise" is confusing.
https://github.com/dfinity/motoko/blob/c8c0917052eda95669dabd612834d66cba5a19ec/doc/md/language-manual.md?plain=1#L2249
Should this be "Expression await* <exp> evaluates ..."? And a newline before the paragraph?
Probably, yes. I'll fix it.
And in the subsequent sentences "If r is trap, evaluation returns trap. Otherwise r is a delayed computation" we only know if it is a trap when the computation is attempted, or? The word "otherwise" is confusing.
No, I think that's ok.
await* (if rand() { Debug.trap "ohoh"} else f())
would be an example of immediately trapping, before evaluating the delayed compuation.
Thanks for the reports!
I don't understand the example of immediate trapping. Why does it trap immediately, what are the expressions that do!
|
gharchive/issue
| 2022-12-10T12:13:33 |
2025-04-01T04:33:59.382257
|
{
"authors": [
"crusso",
"timohanke"
],
"repo": "dfinity/motoko",
"url": "https://github.com/dfinity/motoko/issues/3636",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
471564054
|
HasErrors is still present on the excel output even it is excluded in property.
PowerShell Info:
Name Value
PSVersion 5.1.14393.
PSEdition Desktop
PSCompatibleVersions {1.0, 2.0,
BuildVersion 10.0.14393
CLRVersion 4.0.30319.
WSManStackVersion 3.0
PSRemotingProtocolVersion 2.3
SerializationVersion 1.1.0.1
I'm still getting the HasErrors field in excel.
Here is my PowerShell script:
Invoke-Sqlcmd -Query "select @@version" -Database master -ServerInstance localhost | Select-Object -Property * -ExcludeProperty RowError, RowState, Table, ItemArray, HasErrors | Export-Excel "D:\OutputFile.xlsx"
See screenshot of issue:
Export-Excel now supports DataTables, try this.
$DataTable = Invoke-Sqlcmd -Query "select @@Version" -Database master -ServerInstance localhost -OutputAs DataTables
Export-Excel "D:\OutputFile.xlsx" -InputObject $DataTable
That looks like you have run the same script more than once against the same file, and over written first column E, then D, then C, then B
Either delete the file before exporting or use -ClearSheet
if you remove the export is the right data coming out of the select.
@ili101 has also pointed out there is a better way when you have a datatable.
That looks like you have run the same script more than once against the same file, and over written first column E, then D, then C, then B
Either delete the file before exporting or use -ClearSheet
if you remove the export is the right data coming out of the select.
@ili101 has also pointed out there is a better way when you have a datatable.
Hope you can share your script using the -ClearSheet. Thank you.
Export-Excel now supports DataTables, try this.
$DataTable = Invoke-Sqlcmd -Query "select @@Version" -Database master -ServerInstance localhost -OutputAs DataTables
Export-Excel "D:\OutputFile.xlsx" -InputObject $DataTable
Exactly what I need. Thanks Mate!
Export-Excel "D:\OutputFile.xlsx" -clearsheet
or better
Export-Excel "D:\OutputFile.xlsx" -WorksheetName "Name" -ClearSheet
|
gharchive/issue
| 2019-07-23T08:46:46 |
2025-04-01T04:33:59.389006
|
{
"authors": [
"blhf",
"ili101",
"jhoneill"
],
"repo": "dfinke/ImportExcel",
"url": "https://github.com/dfinke/ImportExcel/issues/636",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1175076625
|
Trigger notification only when a change affects a filter
Describe the bug
I set up a filter on a web page monitor and the preview correctly shows me the part triggered by the filter in blue.
Every time the page has some part changed (even outside the filter) I receive a notification.
Version
v0.39.10
To Reproduce
Create a monitor for a web page
Create a filter
Expected behavior
Receive a notification only when the change affects the filter
Desktop (please complete the following information):
Debian docker
Additional context
Some days ago I set up the same page and the same filter, but the behaviour was correct and only changes falling into the filter rule triggered the notification.
not information information
Sorry, can you explain to me what information do you need (screenshot maybe?)
Thanks
|
gharchive/issue
| 2022-03-21T08:56:56 |
2025-04-01T04:33:59.648394
|
{
"authors": [
"dgtlmoon",
"jeanpaulrh"
],
"repo": "dgtlmoon/changedetection.io",
"url": "https://github.com/dgtlmoon/changedetection.io/issues/482",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2367001768
|
Improving the text to voice converter
Improving the text to voice converter
The UI is not aligned properly and text box is aligned right side and adding more styles to the page will make it more attractive and appealing.
To Reproduce
I would like to add more styles to the ui of the page
Expected behavior
A clear and concise description of what you expected to happen.
Screenshots
If applicable, add screenshots to help explain your problem.
Desktop (please complete the following information):
OS: [e.g. iOS]
Browser [e.g. chrome, safari]
Version [e.g. 22]
Smartphone (please complete the following information):
Device: [e.g. iPhone6]
OS: [e.g. iOS8.1]
Browser [e.g. stock browser, safari]
Version [e.g. 22]
Additional context
Add any other context about the problem here.
@dhairyagothi I AM CONTRIBUTING UNDER VSOC 24. I would like to work on this issue
okay go ahead @SuhainaFathimaM
|
gharchive/issue
| 2024-06-21T17:54:54 |
2025-04-01T04:33:59.655157
|
{
"authors": [
"SuhainaFathimaM",
"dhairyagothi"
],
"repo": "dhairyagothi/50_days_50_web_project",
"url": "https://github.com/dhairyagothi/50_days_50_web_project/issues/141",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
284256131
|
Configuration option for stand-alone maintainable repo
Type of Issue
[ ] Bug Report
[x] Feature Request
If I understand correctly, ng-packagr assumes one use-case, where you want to publish an npm module from some component you made that is part of a whole angular app.
The issue with this is, it's not maintainable by anyone else except yourself. Meaning, since the source files are not included in dist, no one else can submit a PR and improve my component.
I can go ahead and package the source files in the level above dist (and include dist), but now I have 2 duplicate package.json files, and only the one in dist has the correct path for main, plus a duplicate README. I'll need to tell my contributors to manually delete those, after they run yarn build and before they make a PR.
Can we get an option like "stand-alone": true which would not dupe all those assets?
It would also be great to have "name" so that it doesn't use the name in my root package.json.
Thank you!
Update
For now I did this under "scripts". Still would be nice not to have to though.
"build": "ng-packagr && rm dist/package.json && rm dist/README.md"
Hey so while you would only publish the fiat folder to npm you would keep all of the source code on github or something if you want it to be open source.
This is by design. The dist binaries should be distribution-ready. See Design Doc on Build Artefacts
@dherges Don't you think everyone would benefit more from a component you distribute via npm/open source if others could improve upon it? All of my repos have improved immeasurably from contributions.
@deebloo Thanks for the suggestion but I have never seen an npm module that is distributed separately from the github repo, they're always mirrors. That's the purpose of the dist folder; it keeps minified/modularized code separate from the unminified editable code.
@inorganik a lot of npm packages are distributed that way. Angular itself is an example
@deebloo Fair enough. Still, framework vs component
@inorganik angular material then. sorry I am not trying to pick, just trying to point out that there are examples
@deebloo ok ok, I stand corrected. :) Still, one package for contributing and distributing is convenient.
I don't know what I think. Right now, the convention is that the distributables are written to a dedicated dist folder. I don't consider the distributables to be sources. No one is capable of reading or understanding the binary js files.
Maybe one time, a pipeline will be exposed through the public API of ng-packagr and you will be able to register steps that will be run before or after a default build step or maybe you even replace/overwrite a build step. Up until then, ng-packagr && do-thing && do-other-thing is a proven pipeline customiation 😄
I think you misunderstood me. In other repos of mine, such as countUp.js, and many other repos such as Twitter Bootstrap, they are structured like this:
Repo
|
+-- myComponent.js
+-- package.json
+-- README.md
|
+-- dist
|
+-- myComponent.min.js
So, dist is still only for distributables, but the root holds editable files and other miscellany. The advantage is: developers can customize the component to be exactly how they want it, and repackage it themselves, or they can make PRs to improve what's published on npm.
npm install will put the entire repo in the node_modules folder, but only what's in dist is included in your bundle.
With ng-packagr, it's expecting you to only publish what's in the dist folder to npm. So in effect it is not by default truly open source. If you want it to be open source, you have to make a separate repo for it (as you pointed out), with separate package.json and README files.
My feature request was to simply omit the duplicate package.json and README files in the build step given a certain build option, to facilitate the creation of an open-source component and maintain only one package.
Hm, maybe we're talking past each other.
Imo, package.json is a binary artefact. There's main and module and typings properties. The values need to point to the generated binary js artefacts. Btw, what happens when setting dest: "." to the directory of the source package.json? Is the package.json overwritten?
|
gharchive/issue
| 2017-12-22T21:36:13 |
2025-04-01T04:33:59.679531
|
{
"authors": [
"deebloo",
"dherges",
"inorganik"
],
"repo": "dherges/ng-packagr",
"url": "https://github.com/dherges/ng-packagr/issues/440",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2630690616
|
Create monsalon.sql
adding the missing sql
thanks
|
gharchive/pull-request
| 2024-11-02T18:31:57 |
2025-04-01T04:33:59.706831
|
{
"authors": [
"diallohouleymatou",
"traorecheikh"
],
"repo": "diallohouleymatou/Gestion_Salon_Spa",
"url": "https://github.com/diallohouleymatou/Gestion_Salon_Spa/pull/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
333072448
|
Broadcasting ?
Imagine I have a histogram like this:
Hist(bin("x", 100, 0.0, 1.0), groupby("condition"))
and I have an array of x collected under condition="test". So I want to put these data into the histogram. To do that, I need to do something like this:
hist.fill(x=xarray, condition=["test"]*len(xarray))
I think it would be useful to adopt a simple "broadcasting" rule to allow this:
hist.fill(x=xarray, condition="test")
hist.fill(x=xarray, condition="'test'") (one more level of quoting) might already work. If not, it's related to the constants-handling bug (#23, I think).
No, that does not work and it generates this:
ValueError: array 'tt' has len 1 but other arrays have len 100
h = Hist(bin("a", 10, 0, 1), groupby("tt"))
h.fill(a=np.random.random((100,)), tt='"a"')
h.fill(a=np.random.random((100,)), tt='"b"')
h.stack("tt").area("a").to(canvas)
The ValueError you got was correct. The suggestion I have you was wrong. I meant to say
hist = Hist(bin("x", 100, 0.0, 1.0), groupby("'test'"))
hist.fill(x=xarray)
I put the single string in quotations in the wrong place. It's supposed to be a constant expression. Also, because of #23, I might need to fix constant expressions.
["test"]*len(xarray) would be very slow, but that's because it's creating a large Python list and converting it to a string-valued Numpy array before it ever enters histbook.
No, that is not the use case I meant.
I meant this:
h = Hist(bin("data", 10, 0, 1), groupby("condition"))
h.fill(data=np.random.random((100,)), condition="a")
h.fill(data=np.random.random((100,)), condition="b")
...
Real life example: I want to plot some distribution for a number of datasets. So I call fill() some number of times for dataset "a", then for dataset "b" and so on, and then I want to see stacked histogram with all the datasets together.
So I am not grouping by a constant. It is constant only for a single call to fill(). Next call to fill() will use another value.
I figured that out in the other issue. You want to use the Hist.group static method. (I'm trying to do all of this on a phone in a hotel room without wifi.)
I appreciate your help, Jim. Sorry you have to do this in such inhumane conditions :)
Maybe grouping would work too, but I got it figured out, implementing the "broadcasting" myself for now. Here is the fragment of real code which works:
h_all = Hist(hbin("NJets", 20, 0, 20), groupby("dataset"))
h_all_display = IPythonDisplay(
h_all
.stack("dataset")
.area("NJets", width=400)
)
for dataset_name in Datasets[:5]:
job = session.createJob(dataset_name, display=False)
job.addHistogram(h_all, ["NJets", "dataset"], constants={"dataset":dataset_name})
job.run()
runtime = job.TFinish - job.TStart
nevents = job.EventsProcessed
nworkers = len(job.WorkerAddresses)
h_all_display.update()
print "%s: %d events processed using %d workers in %.1f seconds, %.2e events/second" % (
dataset_name, nevents, nworkers, runtime, float(nevents)/runtime)
I don't see the fill in this. (Maybe that happens remotely?)
One of the complaints about Histogrammar was that everything that appeared in a final plot had to come from a single dataset, but it's common to combine data from different Monte Carlo and data samples. It seemed unnatural to have to create new fields in the data that were piecewise constant to emulate what the physicists had in mind: bringing together data from different places.
So histbook has methods for adding information (like "group") in addition to removing information (like "select", "rebin", and "project").
Yes, the actual filling happens remotely and once in a while the Histbook histogram gets picked and shipped back and then cleared.
Here is the implementation of the broadcasting at the remote:
...
# by this time my_dict contains "regular" data in ndarray's
if my_dict:
nitems = len(my_dict[my_dict.keys()[0]])
for dn, dv in self.Constants.items():
my_dict[dn] = [dv]*nitems
self.H.fill(**my_dict)
>>> from histbook import *
>>> h = Hist(bin("x", 5, 0, 5), groupby("condition"))
>>> h.pandas()
Empty DataFrame
Columns: [count(), err(count())]
Index: []
>>> h.fill(x=[1, 2, 3], condition="one")
>>> h.pandas()
count() err(count())
condition x
one [-inf, 0.0) 0.0 0.0
[0.0, 1.0) 0.0 0.0
[1.0, 2.0) 1.0 1.0
[2.0, 3.0) 1.0 1.0
[3.0, 4.0) 1.0 1.0
[4.0, 5.0) 0.0 0.0
[5.0, inf) 0.0 0.0
{NaN} 0.0 0.0
>>> h.fill(x=[4, 5], condition="two")
>>> h.pandas()
count() err(count())
condition x
one [-inf, 0.0) 0.0 0.0
[0.0, 1.0) 0.0 0.0
[1.0, 2.0) 1.0 1.0
[2.0, 3.0) 1.0 1.0
[3.0, 4.0) 1.0 1.0
[4.0, 5.0) 0.0 0.0
[5.0, inf) 0.0 0.0
{NaN} 0.0 0.0
two [-inf, 0.0) 0.0 0.0
[0.0, 1.0) 0.0 0.0
[1.0, 2.0) 0.0 0.0
[2.0, 3.0) 0.0 0.0
[3.0, 4.0) 0.0 0.0
[4.0, 5.0) 1.0 1.0
[5.0, inf) 1.0 1.0
{NaN} 0.0 0.0
Numerical values also get broadcasted:
>>> from histbook import *
>>> h = Hist(bin("x", 5, 0, 5), bin("y", 2, 0, 2, underflow=False, ove
rflow=False, nanflow=False))
>>> h.pandas()
count() err(count())
x y
[-inf, 0.0) [0.0, 1.0) 0.0 0.0
[1.0, 2.0) 0.0 0.0
[0.0, 1.0) [0.0, 1.0) 0.0 0.0
[1.0, 2.0) 0.0 0.0
[1.0, 2.0) [0.0, 1.0) 0.0 0.0
[1.0, 2.0) 0.0 0.0
[2.0, 3.0) [0.0, 1.0) 0.0 0.0
[1.0, 2.0) 0.0 0.0
[3.0, 4.0) [0.0, 1.0) 0.0 0.0
[1.0, 2.0) 0.0 0.0
[4.0, 5.0) [0.0, 1.0) 0.0 0.0
[1.0, 2.0) 0.0 0.0
[5.0, inf) [0.0, 1.0) 0.0 0.0
[1.0, 2.0) 0.0 0.0
{NaN} [0.0, 1.0) 0.0 0.0
[1.0, 2.0) 0.0 0.0
>>> h.fill(x=[1, 2, 3], y=0)
>>> h.pandas()
count() err(count())
x y
[-inf, 0.0) [0.0, 1.0) 0.0 0.0
[1.0, 2.0) 0.0 0.0
[0.0, 1.0) [0.0, 1.0) 0.0 0.0
[1.0, 2.0) 0.0 0.0
[1.0, 2.0) [0.0, 1.0) 1.0 1.0
[1.0, 2.0) 0.0 0.0
[2.0, 3.0) [0.0, 1.0) 1.0 1.0
[1.0, 2.0) 0.0 0.0
[3.0, 4.0) [0.0, 1.0) 1.0 1.0
[1.0, 2.0) 0.0 0.0
[4.0, 5.0) [0.0, 1.0) 0.0 0.0
[1.0, 2.0) 0.0 0.0
[5.0, inf) [0.0, 1.0) 0.0 0.0
[1.0, 2.0) 0.0 0.0
{NaN} [0.0, 1.0) 0.0 0.0
[1.0, 2.0) 0.0 0.0
>>> h.fill(x=[4, 5], y=1)
>>> h.pandas()
count() err(count())
x y
[-inf, 0.0) [0.0, 1.0) 0.0 0.0
[1.0, 2.0) 0.0 0.0
[0.0, 1.0) [0.0, 1.0) 0.0 0.0
[1.0, 2.0) 0.0 0.0
[1.0, 2.0) [0.0, 1.0) 1.0 1.0
[1.0, 2.0) 0.0 0.0
[2.0, 3.0) [0.0, 1.0) 1.0 1.0
[1.0, 2.0) 0.0 0.0
[3.0, 4.0) [0.0, 1.0) 1.0 1.0
[1.0, 2.0) 0.0 0.0
[4.0, 5.0) [0.0, 1.0) 0.0 0.0
[1.0, 2.0) 1.0 1.0
[5.0, inf) [0.0, 1.0) 0.0 0.0
[1.0, 2.0) 1.0 1.0
{NaN} [0.0, 1.0) 0.0 0.0
[1.0, 2.0) 0.0 0.0
I am trying to use broadcasting with groupby() and a string as the groupby axis value, and it is not working:
import histbook
from histbook import Hist, bin, groupby
import numpy as np
h = Hist(bin("NJets", 10, 0, 10), groupby("dataset"))
h.fill(dataset="dataset1", NJets=np.arange(100))
this fragment gives me a warning and an error:
/home/ivm/anaconda2/lib/python2.7/site-packages/numpy/core/numeric.py:301: FutureWarning: in the future, full(100, array('dataset1',
dtype='|S8')) will return an array of dtype('S8')
format(shape, fill_value, array(fill_value).dtype), FutureWarning)
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
<ipython-input-11-6c35a986016c> in <module>()
4
5 h = Hist(bin("NJets", 10, 0, 10), groupby("dataset"))
----> 6 h.fill(dataset="dataset1", NJets=np.arange(100))
7
8 h1 = Hist(bin("NJets", 10, 0, 10), bin("dataset", 100, 0, 100))
/home/ivm/anaconda2/lib/python2.7/site-packages/histbook-1.1.0-py2.7.egg/histbook/hist.pyc in fill(self, arrays, **more)
574
575 self._prefill()
--> 576 length = self._fill(arrays)
577 self._postfill(arrays, length)
578
/home/ivm/anaconda2/lib/python2.7/site-packages/histbook-1.1.0-py2.7.egg/histbook/hist.pyc in _fill(self, arrays)
151 array = numpy.array(array)
152 if array.shape == ():
--> 153 array = numpy.full(length, array)
154
155 if length != array.shape[0]:
/home/ivm/anaconda2/lib/python2.7/site-packages/numpy/core/numeric.pyc in full(shape, fill_value, dtype, order)
300 "in the future, full({0}, {1!r}) will return an array of {2!r}".
301 format(shape, fill_value, array(fill_value).dtype), FutureWarning)
--> 302 multiarray.copyto(a, fill_value, casting='unsafe')
303 return a
304
ValueError: could not convert string to float: dataset1
this works fine:
h1 = Hist(bin("NJets", 10, 0, 10), bin("dataset", 100, 0, 100))
h1.fill(dataset=13, NJets=np.arange(100))
my version is 1.2.0
Hmmm.
>>> from histbook import *
>>> import numpy as np
>>> h = Hist(bin("NJets", 10, 0, 10), groupby("dataset"))
>>> from histbook import *
>>> import numpy as np
>>> h = Hist(bin("NJets", 10, 0, 10), groupby("dataset"))
My version is 1.2.0 also. What's your Numpy version?
>>> histbook.__version__
'1.2.0'
>>> np.__version__
'1.14.4'
This might imply a higher minimum Numpy version: right now I say install_requires = ["numpy>=1.8.0"].
import numpy
numpy.version
'1.11.3'
I found Numpy 1.11.3 on lxplus and tried it there. Indeed:
>>> numpy.full(10, "hello")
/afs/cern.ch/user/p/pivarski/miniconda3/lib/python3.5/site-packages/numpy/core/numeric.py:301: FutureWarning: in the future, full(10, 'hello') will return an array of dtype('<U5')
format(shape, fill_value, array(fill_value).dtype), FutureWarning)
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/afs/cern.ch/user/p/pivarski/miniconda3/lib/python3.5/site-packages/numpy/core/numeric.py", line 302, in full
multiarray.copyto(a, fill_value, casting='unsafe')
ValueError: could not convert string to float: 'hello'
I'll keep the Numpy requirement minimum where it is and put a work-around in this bit of code. I think this just affects everywhere numpy.full is called with a string argument.
I upgraded numpy to 1.14.5 and now the error is different:
import histbook
from histbook import Hist, bin, groupby
import numpy as np
print np.__version__
h = Hist(bin("NJets", 10, 0, 10), groupby("dataset"))
h.fill("dataset1", NJets=np.arange(100))
1.14.5
---------------------------------------------------------------------------
TypeError Traceback (most recent call last)
<ipython-input-4-84a0ed9de926> in <module>()
6
7 h = Hist(bin("NJets", 10, 0, 10), groupby("dataset"))
----> 8 h.fill("dataset1", NJets=np.arange(100))
9
/home/ivm/anaconda2/lib/python2.7/site-packages/histbook-1.2.0-py2.7.egg/histbook/hist.pyc in fill(self, arrays, **more)
374
375 self._prefill()
--> 376 length = self._fill(arrays)
377 self._postfill(arrays, length)
378
/home/ivm/anaconda2/lib/python2.7/site-packages/histbook-1.2.0-py2.7.egg/histbook/fill.pyc in _fill(self, arrays)
118 else:
119 try:
--> 120 array = arrays[instruction.extern.value]
121 except KeyError:
122 if instruction.extern.value in histbook.expr.Expr.maybeconstants:
/home/ivm/anaconda2/lib/python2.7/site-packages/histbook-1.2.0-py2.7.egg/histbook/util/__init__.pyc in __getitem__(self, n)
38 return self._two[n] # and it has precedence
39 else:
---> 40 return self._one[n] # self._one might only have __getitem__
TypeError: string indices must be integers, not str
I just patched master so that it works as shown in my comment in Numpy 1.14.4 and 1.11.3.
But your problem is that you're missing a keyword argument:
h.fill("dataset1", NJets=np.arange(100))
instead of
h.fill(dataset="dataset1", NJets=np.arange(100))
It's trying to interpret the positional (non-keyword) argument as a dict of arrays, a Pandas DataFrame, or a Spark DataFrame. I wanted to be loose about it so that anything with df["column"] → array would work in that slot (i.e. duck typing). But if the "df" is a string, it applies square-bracket-string-close-square-bracket to what is actually a string, hence "string indices must be integers, not str."
|
gharchive/issue
| 2018-06-17T15:51:45 |
2025-04-01T04:33:59.734181
|
{
"authors": [
"imandr",
"jpivarski"
],
"repo": "diana-hep/histbook",
"url": "https://github.com/diana-hep/histbook/issues/24",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
356726088
|
使用cat-client-agent埋点插件导致内存溢出
使用cat-client-agent埋点插件会导致内存溢出,The class "org.aspectj.weaver.loadtime.Aj$WeaverContainer", loaded by "sun.misc.Launcher$AppClassLoader @ 0x6c001f0b0", occupies 2,468,554,024 (58.85%) bytes. The memory is accumulated in one instance of "java.util.HashMap$Node[]" loaded by "".
已解决。-Daj.weaving.loadersToSkip=com.alibaba.fastjson.util.ASMClassLoader
|
gharchive/issue
| 2018-09-04T09:28:05 |
2025-04-01T04:33:59.749243
|
{
"authors": [
"bulong0721"
],
"repo": "dianping/cat",
"url": "https://github.com/dianping/cat/issues/1302",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.