id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
2120825214
Improve net/bind to use SO_REUSEPORT and friends All the Listeners from the bind folder should have the option of being created with SO_REUSEPORT, so a Config should look like: // Config is the configuration for Bind() type Config struct { ....... Port uint16 // PortStrict tells us not to try other ports PortStrict bool // PortAttempts indicates how many times we will try finding a port PortAttempts int // Defaultport indicates the port to try on the first attempt if Port is zero DefaultPort uint16 // PortReuse indicates that the listener is able to reuse the port between threads/processes PortReuse bool ..... // ListenTCP is the helper to use to listen on TCP ports ListenTCP func(network string, laddr *net.TCPAddr) (*net.TCPListener, error) // ListenUDP is the helper to use to listen on UDP ports ListenUDP func(network string, laddr *net.UDPAddr) (*net.UDPConn, error) ....... } also default ListenTCP and ListenUDP should be created var listenConfig = net.ListenConfig{ Control: Control, } func ListenTCP(network, address string) (net.Listener, error) { return listenConfig.Listen(context.Background(), "tcp", address) } func ListenUDP(network, address string) (net.PacketConn, error) { return listenConfig.ListenPacket(context.Background(), "udp", address) } with Control function implemented for windows, linux, unix example for unix: import ( "syscall" "golang.org/x/sys/unix" ) func Control(network, address string, c syscall.RawConn) (err error) { controlErr := c.Control(func(fd uintptr) { err = unix.SetsockoptInt(int(fd), unix.SOL_SOCKET, unix.SO_REUSEADDR, 1) if err != nil { return } err = unix.SetsockoptInt(int(fd), unix.SOL_SOCKET, unix.SO_REUSEPORT, 1) }) if controlErr != nil { err = controlErr } return } func Control(network, address string, c syscall.RawConn) (err error) { controlErr := c.Control(func(fd uintptr) { err = unix.SetsockoptInt(int(fd), unix.SOL_SOCKET, unix.SO_REUSEADDR, 1) if err != nil { return } err = unix.SetsockoptInt(int(fd), unix.SOL_SOCKET, unix.SO_REUSEPORT, 1) }) if controlErr != nil { err = controlErr } return } why do you do the same setsockoptInt twice? @karasz Never mind, I just noticed they set different options. PR #143 should close this.
gharchive/issue
2024-02-06T13:47:57
2025-04-01T04:33:56.349975
{ "authors": [ "amery", "karasz" ], "repo": "darvaza-proxy/x", "url": "https://github.com/darvaza-proxy/x/issues/16", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
821006375
fix(OntologyResponderV2): Fix check when updating ontology label and comment (DSP-1390) resolves DSP-1390 Looks good to me! Could you also add some logic to allow an empty string for a comment? I don't believe comments are required and currently there is no way to remove a comment once one is provided. @mdelez Thanks! Could you also add some logic to allow an empty string for a comment? We don't allow empty strings in the triplestore, so I guess we have two options: Change the semantics of that route so that if you don't provide a comment, the existing comment is removed. Add a separate route for deleting the comment. Please open a separate YouTrack issue for this indicating which you prefer. separate Youtrack issue has been made https://dasch.myjetbrains.com/youtrack/issue/DSP-1397
gharchive/pull-request
2021-03-03T11:09:11
2025-04-01T04:33:56.371986
{ "authors": [ "benjamingeer", "mdelez" ], "repo": "dasch-swiss/dsp-api", "url": "https://github.com/dasch-swiss/dsp-api/pull/1826", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
977815447
arrows on edges Hi, Very nice library! I just wonder if it is possible to have arrows on the edges to indicate the flow? kr, Filip. Thanks for your feedback. The arrows on the edges are exactly the feature I'm working on right now! Here's a screenshot in development. It's a bit large modification, so it will take a few more days to release. Best regards, dash14. Hi @dash14 , Thx for the update! Again ... really nice and will written library. Hi @bzd2000, Thank you very much for your positive feedback! This motivates me a lot.
gharchive/issue
2021-08-24T07:57:01
2025-04-01T04:33:56.380118
{ "authors": [ "bzd2000", "dash14" ], "repo": "dash14/v-network-graph", "url": "https://github.com/dash14/v-network-graph/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1208516889
Add/remove operations failed to render Hi, I have been trying to reproduce the add/remove nodes and edges operations on my end, but when I click the buttons, no nodes/edges could be added or removed. I can tell that the buttons are working because after I select a node or an edge, the "remove" buttons would be enabled, which means that the functions should be partially working. So the lines which failed to work should be nodes[nodeId] = { name } for function addNode() and delete nodes[nodeId] for function removeNode(). The only differences between my code and the demo in the documentation is that I'm using a d3 force layout, and that my data is imported from a JSON file with vNG. I added a checkbox to enable/disable d3-force, but even after I disabled it and have the graph present in a simple layout, the add/remove operations still wouldn't work. Is there a possible cause why this might have happened? Thank you in advance! Hi @happylilem, Maybe, an object imported from JSON is not reactive, so it is not detected changes. After importing from JSON, try wrapping it with reactive() to make changes to the resulted object. Sample code is shown below: <script setup lang="ts"> import { reactive } from "vue"; import testData from "./test-data.json"; const data = reactive(testData); function addNode() { const number = Object.keys(data.nodes).length + 1; const nodeId = `node${number}`; const name = `Node ${number}`; data.nodes[nodeId] = { name }; } </script> <template> <div class="graph"> <v-network-graph :nodes="data.nodes" :edges="data.edges" :layouts="data.layouts" :configs="data.configs" /> <button @click="addNode">Add a node</button> </div> </template> <style> .graph { border: 1px solid #888; width: 600px; height: 400px; margin: 0 auto; } </style> test-data.json { "nodes": { "node1": { "name": "Node 1" }, "node2": { "name": "Node 2" }, "node3": { "name": "Node 3" }, "node4": { "name": "Node 4" } }, "edges": { "edge1": { "source": "node1", "target": "node2" }, "edge2": { "source": "node2", "target": "node3" }, "edge3": { "source": "node3", "target": "node4" } }, "layouts": { "nodes": { "node1": { "x": 0, "y": 0 }, "node2": { "x": 70, "y": 70 }, "node3": { "x": 140, "y": 0 }, "node4": { "x": 210, "y": 70 } } }, "configs": { "node": { "normal": { "radius": 20 } } } } I did const nodes: Nodes = reactive({ ...data.nodes }) const edges: Edges = reactive({ ...data.edges }) instead of const data = reactive(testData); I guess they function similarly? But I've tried either way and the addNode function still wouldn't work... My code looks like this: import * as vNG from "v-network-graph" import { Nodes, Edges } from "v-network-graph" import { reactive, ref, computed } from "vue" import data from "./test-data.json" const nodes: Nodes = reactive({ ...data.nodes }) const edges: Edges = reactive({ ...data.edges }) const nextNodeIndex = ref(Object.keys(nodes).length + 1) const nextEdgeIndex = ref(Object.keys(edges).length + 1) const selectedNodes = ref<string[]>([]) const selectedEdges = ref<string[]>([]) function addNode() { const nodeId = node${nextNodeIndex.value} const name = N${nextNodeIndex.value} nodes[nodeId] = { name } nextNodeIndex.value++ } Problem solved! It isn't because I didn't set the graph to be reactive, but that inside of <v-network-graph> tag, using :nodes="data.nodes" :edges="data.edges" :configs="data.configs" prevented the graph from rendering the changes because it's always using nodes from the original data file. So I took away the data. part and did: :nodes="nodes" :edges="edges" :configs="configs" For graphs with vNG, all features for the added Node are needed to make a new node, like this: function addNode() { const nodeId = nextNodeIndex.value nodes[nodeId] = { name: String(nodeId), size: 16, color: "blue", label: true } nextNodeIndex.value++ The only problem left is the same as described in issue 54, a node cannot be removed when it's connected to another node with an edge. I need to delete the edge first and then delete the node. Hi @happylilem, Thanks for your report. Glad you have it resolved. Please check the issue when deleting a node (#54), as it has been fixed in v0.5.12. I close this issue for now. If you have any other problems, please open a new issue.
gharchive/issue
2022-04-19T15:21:09
2025-04-01T04:33:56.390841
{ "authors": [ "dash14", "happylilem" ], "repo": "dash14/v-network-graph", "url": "https://github.com/dash14/v-network-graph/issues/52", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
709658975
contrib: Replace developer keys with list of pgp fingerprints See https://github.com/bitcoin/bitcoin/pull/11909 Is this a sensible change that we should enact as well? It seems to me to not provide much benefit to remove the keys from the repo, however they point out some potential benefit such as dealing with " Outdated keys. Unclear whether and when to replace by fresh copies. Unclear when to add a key of a new developer or Gitian builder. " Thoughts? IMHO I would include the PGP-keys fingerprints in the repository as part of the README.md and include the keys itself in the releases. This is basically how it is done right now. What is more relevant to me is that the keys are signed by each other. Right now I have the issue that I paused to use Dash for a couple of years and right now the PGP-key in use does not match the keys I have. So from the cryptographic perspective I can not trust the new "pasta"-key because it was not signed by the codablock-key (or some other key used for signing the older releases).
gharchive/issue
2020-09-27T03:37:24
2025-04-01T04:33:56.397462
{ "authors": [ "PastaPastaPasta", "SecTec" ], "repo": "dashpay/dash", "url": "https://github.com/dashpay/dash/issues/3739", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
995715341
Implement support for fsspec callbacks in put_file/get_file This patch adds support for fsspec callbacks. It also removes some load from the get_file: There is a completely redundant ls() call it is making on every get_file operation, which is very costful It also tries to do an isdir() check, which is also very costful (if not cached) when considering it is done for every get_file especially on Azure, since there are no directories per se. Resolves #276 CC: @TomAugspurger @hayesgb Thanks!
gharchive/pull-request
2021-09-14T08:02:58
2025-04-01T04:33:56.400023
{ "authors": [ "hayesgb", "isidentical" ], "repo": "dask/adlfs", "url": "https://github.com/dask/adlfs/pull/275", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
820789888
Setting explicit chunk sizes in dask.array.from_array() fails for 2D array: "Chunks do not add up to shape" What happened: Running dask.array.from_array() on a 2D numpy.ndarray and passing a specific list of chunk sizes to chunks, construction of the Dask Array failed with the following error: Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/Users/jlamb/miniconda3/lib/python3.8/site-packages/dask/array/core.py", line 3054, in from_array chunks = normalize_chunks( File "/Users/jlamb/miniconda3/lib/python3.8/site-packages/dask/array/core.py", line 2726, in normalize_chunks raise ValueError( ValueError: Chunks do not add up to shape. Got chunks=((67, 6), (33, 6)), shape=(100, 6) What you expected to happen: I expected that running da.from_array(X, chunks=((67, 6), (33, 6))) on a numpy array with shape (100, 6) would create a Dask Array with two chunks, where the first chunk had size (67, 6) and the second chunk had size (33, 6). I expected to be able to do this based on the documentation at https://docs.dask.org/en/latest/array-api.html#dask.array.from_array, which includes following in the list of valid values for the chunks argument to dask.array.from_array(): Explicit sizes of all blocks along all dimensions like ((1000, 1000, 500), (400, 400)). Minimal Complete Verifiable Example: import dask.array as da import numpy as np X = np.random.random((100, 6)) da.from_array( X, chunks=((67, 6), (33, 6)) ) Anything else we need to know?: I did search for other issues with this error message, and didn't find any that seemed like the same question. I don't know for sure if the error I hit is the root cause of https://github.com/dask/dask/issues/6709 or not, but it seems possible. I don't even know for sure if it's an error or if I just misinterpreted the documentation 😬 . I tried to find tests on the pattern "pass a list of specific chunk sizes to .from_array()", to see if those tests used that feature differently and maybe I was misunderstanding the docs. I ran git grep from_array dask/tests and saw that there don't appear to be any tests currently on that pattern. Environment: Dask version: dask: 2021.2.0 numpy: 1.20.1 Python version: 3.8.3 Operating System: MacOS Mojave (10.14.6) Install method (conda, pip, source): pip conda info active environment : None user config file : /Users/jlamb/.condarc populated config files : /Users/jlamb/.condarc conda version : 4.9.2 conda-build version : not installed python version : 3.8.3.final.0 virtual packages : __osx=10.14.6=0 __unix=0=0 __archspec=1=x86_64 base environment : /Users/jlamb/miniconda3 (writable) channel URLs : https://repo.anaconda.com/pkgs/main/osx-64 https://repo.anaconda.com/pkgs/main/noarch https://repo.anaconda.com/pkgs/r/osx-64 https://repo.anaconda.com/pkgs/r/noarch package cache : /Users/jlamb/miniconda3/pkgs /Users/jlamb/.conda/pkgs envs directories : /Users/jlamb/miniconda3/envs /Users/jlamb/.conda/envs platform : osx-64 user-agent : conda/4.9.2 requests/2.24.0 CPython/3.8.3 Darwin/18.7.0 OSX/10.14.6 UID:GID : 501:20 netrc file : None offline mode : False Thanks for your time and help with this. I think what you are looking for is: da.from_array(X, chunks=((67, 33), (6,))) oh I see, yep that worked, thank you! That wasn't clear to me from the documentation. Would you be open to a pull request that adds an example under https://github.com/dask/dask/blob/b73e7d0e187cf321f50b302e1f790aeb7528f11c/dask/array/core.py#L3158 and a unit test on this pattern in https://github.com/dask/dask/blob/b73e7d0e187cf321f50b302e1f790aeb7528f11c/dask/tests/test_base.py? Always open to more docs! I'll go ahead and close this, but feel free to open a docs PR :) Always open to more docs! I'll go ahead and close this, but feel free to open a docs PR :) thanks! I just opened https://github.com/dask/dask/pull/7330
gharchive/issue
2021-03-03T06:20:40
2025-04-01T04:33:56.440321
{ "authors": [ "jameslamb", "jsignell", "theXYZT" ], "repo": "dask/dask", "url": "https://github.com/dask/dask/issues/7310", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1527869217
upstream CI report failing It looks like our upstream CI job is failing during the step where it creates/updates an issue on GitHub (see this build for an example) Parsing logs ... Traceback (most recent call last): File "/home/runner/work/_actions/xarray-contrib/issue-from-pytest-log/v1.2.4/parse_logs.py", line 217, in <module> preformatted = [preformat_report(report) for report in failed] File "/home/runner/work/_actions/xarray-contrib/issue-from-pytest-log/v1.2.4/parse_logs.py", line 217, in <listcomp> preformatted = [preformat_report(report) for report in failed] File "/usr/share/miniconda3/envs/test-environment/lib/python3.10/functools.py", line 889, in wrapper return dispatch(args[0].__class__)(*args, **kw) File "/home/runner/work/_actions/xarray-contrib/issue-from-pytest-log/v1.2.4/parse_logs.py", line 89, in _ message = report.longrepr.chain[0][1].message AttributeError: 'str' object has no attribute 'chain' Note we're using the latest xarray-contrib/issue-from-pytest-log@v1.2.4 release. cc @keewis for visibility I just released version v1.2.5 of the action, which should fix this. Awesome, thanks for fixing so quickly @keewis. Updating to 1.2.5 in https://github.com/dask/dask/pull/9822
gharchive/issue
2023-01-10T19:03:35
2025-04-01T04:33:56.443419
{ "authors": [ "jrbourbeau", "keewis" ], "repo": "dask/dask", "url": "https://github.com/dask/dask/issues/9818", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
330299777
Add parquet file support [x] Tests added / passed [x] Passes flake8 dask Following suggestion by @martindurant on #3571 @martindurant I got a bit confused with the sequence of comments. But basically a tested now with a ParquetFile from S3 and it worked. Basically something like this: s3 = s3fs.S3FileSystem(key=*****, secret=****) s3_pf = fastparquet.ParquetFile('s3://***********/sample_parquet/', open_with=s3.open) dsk_df1 = dd.read_parquet('file_partitioned/', filters=[('fake_categorical1', '==', 'G'), ('fake_categorical2', '==', 13)]) dsk_df3 = dd.read_parquet(s3_pf, filters=[('fake_categorical1', '==', 'G'), ('fake_categorical2', '==', 13)]) assert_eq(dsk_df1, dsk_df3) This asserts to True. also when checking s3_pf.fn it does contain s3://... in the begining. So I guess this answer your initial comment. I can either: 1- assert if .open is default Local or is S3FS.open, because these are the ones I tried or 2- Do not check .open and assume any other will work as weel (the way it is in the latest commit) @martindurant I got a bit confused with the sequence of comments. But basically a tested now with a ParquetFile from S3 and it worked. Something like this: s3 = s3fs.S3FileSystem(key=*****, secret=****) s3_pf = fastparquet.ParquetFile('s3://***********/sample_parquet/', open_with=s3.open) dsk_df1 = dd.read_parquet('file_partitioned/', filters=[('fake_categorical1', '==', 'G'), ('fake_categorical2', '==', 13)]) dsk_df3 = dd.read_parquet(s3_pf, filters=[('fake_categorical1', '==', 'G'), ('fake_categorical2', '==', 13)]) assert_eq(dsk_df1, dsk_df3) This asserts to True. I also added a check to see if path contains protocol. So if instead, you do s3_pf = fastparquet.ParquetFile('***********/sample_parquet/', open_with=s3.open) it will raise an error in the line dd.read_parquet(s3_pf, .... Tests failed because of style. You should generally run flake8 for code in PRs. dask/dataframe/io/parquet.py:951:80: W291 trailing whitespace dask/dataframe/io/parquet.py:962:45: E126 continuation line over-indented for hanging indent dask/dataframe/io/parquet.py:965:49: E126 continuation line over-indented for hanging indent dask/dataframe/io/parquet.py:969:45: E126 continuation line over-indented for hanging indent dask/dataframe/io/parquet.py:971:49: E126 continuation line over-indented for hanging indent dask/dataframe/io/tests/test_parquet.py:827:21: E128 continuation line under-indented for visual indent dask/dataframe/io/tests/test_parquet.py:828:21: E128 continuation line under-indented for visual indent dask/dataframe/io/tests/test_parquet.py:846:1: E303 too many blank lines (4) @mrocklin , the test failure does not appear to me to be related, is this a known flake problem? __________________________ test_temporary_directory ___________________________ tmpdir = local('C:\\Users\\appveyor\\AppData\\Local\\Temp\\1\\pytest-of-appveyor\\pytest-0\\test_temporary_directory1') def test_temporary_directory(tmpdir): df = pd.DataFrame({'x': np.random.random(100), 'y': np.random.random(100), 'z': np.random.random(100)}) ddf = dd.from_pandas(df, npartitions=10, name='x', sort=False) with dask.config.set(temporary_directory=str(tmpdir), scheduler='processes'): ddf2 = ddf.set_index('x', shuffle='disk') ddf2.compute() > assert any(fn.endswith('.partd') for fn in os.listdir(str(tmpdir))) E assert False E + where False = any(<generator object test_temporary_directory.<locals>.<genexpr> at 0x00000097E0152D58>) dask\dataframe\tests\test_shuffle.py:671: AssertionError I haven't seen this failure before. I don't understand the Travis error. I ran flake8 on both files I modified and there are no issues... OK, that fixed it - either it was something intermittent, or somehow some state was left over. I have one final request (sorry): can you include in the test an attempt to use inconsistent arguments, such as engine='pyarrow', which should raise an error (use pytest.raises). @martindurant hopefully I understood what you meant and is now implemented @jcrist , getting error with test_orc_with_backend (py27, numpy 1.13), says requests should be installed. Has anything changed recently? getting error with test_orc_with_backend (py27, numpy 1.13), says requests should be installed. Has anything changed recently? If you look at the traceback in the tests, requests is installed but fails to import due to the docstring rewriting they're doing. This is because that part of the test matrix runs with optimize on (https://github.com/dask/dask/blob/master/.travis.yml#L33). For other libraries that have this issue we just skip those tests (https://github.com/dask/dask/blob/master/.travis.yml#L10). I'd add the http filesystem here as well. @jcrist @martindurant I've added the importorskip for requests on my (completely unrelated) PR here https://github.com/dask/dask/pull/3594 , waiting to see if the travis build succeeds @martindurant I think there's still something wrong and not related with my modifications Restarted the one failing build, it looks like maybe an intermittent network issue CondaHTTPError: HTTP 504 GATEWAY_TIMEOUT for url <https://conda.anaconda.org/conda-forge/linux-64/numpy-1.14.5-py36_blas_openblash24bf2e0_200.tar.bz2> Yep, everything looks good. I'm happy with these changes.
gharchive/pull-request
2018-06-07T14:31:51
2025-04-01T04:33:56.454013
{ "authors": [ "andrethrill", "jcrist", "martindurant", "mathewlee11", "mrocklin" ], "repo": "dask/dask", "url": "https://github.com/dask/dask/pull/3573", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1318362209
Add maintainer documentation page This PR documents some of our maintainer practices that folks use but have yet to be written down. This will hopefully help when onboarding new maintainers. Currently this focuses on best practices around merging PRs. I'd like to start with this and then iterate on this page in follow-up PRs. cc @jsignell @jacobtomlinson @pavithraes Also @ian-r-rose @fjetter @crusaderky (who are currently OOO) just a heads up I'm merging main, to avoid CI stalling again. I think there is value it wrapping this up and getting it in! @jrbourbeau should we just commit my last two suggestions and merge? I have no objections to this content (even if I have some personal disagreements about squash merges :wink: ), thanks for writing it up @jrbourbeau! I wonder if some additional text about CI would be helpful. Something along the lines of: Before merging a pull request, maintainers should make every effort to ensure that CI passes. Often this will require looking into the logs of a failed run to see what went wrong, and alerting the pull-request author. Ideally, no pull-request should be merged if there are CI failures, as broken CI in main can easily mask more problems with other PRs, and over the long haul a consistently broken CI is demoralizing for maintainers. However, in the real world, there are sometimes flaky tests, broken upstream dependencies, and failures that are otherwise obviously not related to the PR at hand. If that is the case, a maintainer may merge a PR with failing tests, but they should be prepared to follow up with any failures that result from such an unsafe operation. Thanks for the reviews all -- I'll go ahead and push up suggestions and then we should be good to merge this in and iterate @jrbourbeau Thank you so much for writing this up! The CI failure is unrelated, and I think merging main will fix it. :) Alright, I believe I've incorporated or responded to everyone's suggestions, so this should be ready for a (possibly) final review. Florian brought up a couple of points that, while totally valid, I think merit a separate conversation outside of this PR. Thanks all for the reviews / suggestions. Going to merge this in and we can keep iterating
gharchive/pull-request
2022-07-26T14:53:12
2025-04-01T04:33:56.459753
{ "authors": [ "ian-r-rose", "jrbourbeau", "jsignell", "ncclementi", "pavithraes" ], "repo": "dask/dask", "url": "https://github.com/dask/dask/pull/9309", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2357274237
Add search (Autocomplete) in environment/group dropdown in Share request modal Feature or Bugfix Feature Detail In the following view, instead of using a fixed list in the environment, team and consumption roles dropdowns; this PR introduces search capabilities as requested in #1012. There are still many other dropdowns to change, where we will be able to extract a common frontend component. This one however, is a bit particular so I implemented it separately. Relates #1012 Security Please answer the questions below briefly where applicable, or write N/A. Based on OWASP 10. Does this PR introduce or modify any input fields or queries - this includes fetching data from storage outside the application (e.g. a database, an S3 bucket)? Is the input sanitized? What precautions are you taking before deserializing the data you consume? Is injection prevented by parametrizing queries? Have you ensured no eval or similar functions are used? Does this PR introduce any functionality or component that requires authorization? How have you ensured it respects the existing AuthN/AuthZ mechanisms? Are you logging failed auth attempts? Are you using or adding any cryptographic features? Do you use a standard proven implementations? Are the used keys controlled by the customer? Where are they stored? Are you introducing any new policies/roles/users? Have you used the least-privilege principle? How? By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license. Tested locally: [ ] Happy path: new share request - select an environment, select a group, submit [ ] Happy path2: new share request - select an environment, select a group, select a consumption role, submit [ ] Happy path3: existing share request - select an environment, select a group, submit [ ] Happy path4: existing share request - select an environment, select a group, select a consumption role, submit [ ] Select environment, click on cross in environment ---> removes environment and group options [ ] Select environment, select group, click on cross in group ---> removes group and consumption role options [ ] Select environment, select group, click on cross in environment ---> removes environment and group and consumption role options [ ] Select environment, overwrite it with some random string --> throws validation error [ ] Select group overwrite it with some random string --> throws validation error [ ] Select comsumption, overwrite it with some random string --> throws validation error (DOES NOT OPEN REQUEST FOR GROUP)
gharchive/pull-request
2024-06-17T12:51:53
2025-04-01T04:33:56.508632
{ "authors": [ "dlpzx" ], "repo": "data-dot-all/dataall", "url": "https://github.com/data-dot-all/dataall/pull/1335", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2074400575
[Cherrypick][HotFix] Remove duplicate path when writing nested JSON array [Cherrypick] Remove duplicate path when writing nested JSON array Commit : ca5b35907d4c9f8cac80042ff24292657000b54d PR : #1350 Hotfix for bq sink json support , data was not formatted when writing nested json array. This was caused due to path having duplicate entry, which is removed in this fix. Version bumped to 0.23.1-SNAPSHOT
gharchive/pull-request
2024-01-10T13:39:14
2025-04-01T04:33:56.515071
{ "authors": [ "psainics" ], "repo": "data-integrations/google-cloud", "url": "https://github.com/data-integrations/google-cloud/pull/1354", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
961137410
CARDS-1264: Add a Pagination Test questionnaire to the test runmode Turned the Patient Information form into a paginated one @sdumitriu Please check the latest version of #715 . The test questionnaire is an excerpt from the current Patient information questionnaire with several questions as well as whole sections removed and (simple) conditions added for two of the remaining sections. There's also a nested section to test that only top-level sections become pages. There's of course room for improvement, such as more complex conditionals, but IMO it's "lighter" and better than this one. I don't see any pagination in the Pagination Test form. Something is wrong with DOB and Pedigree, Date of genetic diagnosis, Age of genetic diagnosis...
gharchive/pull-request
2021-08-04T22:15:31
2025-04-01T04:33:56.520554
{ "authors": [ "marta-", "sashaandjic", "sdumitriu" ], "repo": "data-team-uhn/cards", "url": "https://github.com/data-team-uhn/cards/pull/716", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
232123235
GPU integration Interface with cuda. Note we have a proof of concept for pop counting of vectors - would have to change it to do top k DICE-Sorensen index + threshold similar to our C implementation. Aha! Link: https://csiro.aha.io/features/ANONLINK-79 See proof of concept: https://github.com/hardbyte/anonlink-cuda
gharchive/issue
2017-05-30T01:41:59
2025-04-01T04:33:56.526216
{ "authors": [ "hardbyte" ], "repo": "data61/anonlink", "url": "https://github.com/data61/anonlink/issues/17", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1349738847
'append_new_columns' flag causes failures for 'append' incremental models Describe the bug When using the append_new_columns flag for incremental models with an append incremental strategy, the first dbt run adding new columns will fail. However, the next run and all runs after will succeed. The reason for this appears to be that new columns are successfully added, but the first insert statement after the table has been altered still uses the old columns (i.e., 3 instead of 4 columns). Note that this does not happen for the merge incremental strategy. Steps To Reproduce 1.) Configure an incremental model to with the append incremental_strategy and append_new_columns for on_schema_change, like so: {{ config( materialized = 'incremental' , incremental_strategy = 'append' , on_schema_change = 'append_new_columns' ) }} 2.) Execute an initial (full refresh) build of the model 3.) Add a new column 4.) Execute an incremental build of the model (it should fail) 5.) Execute an incremental build of the model again (it should succeed) Expected behavior This should succeed the first run after a schema change (new column), rather than fail the first time. Screenshots and log output Here's an example of what the error looks like: Cannot write to '[REDACTED].incremental_test', not enough data columns; target table has 4 column(s) but the inserted data has 3 column(s) Example of databricks query log showing the insert failure after the alter table, then succeeding on the next insert statement: System information The output of dbt --version: dbtenv info: Using dbt-databricks==1.1.0 (set by dbt_project.yml). Core: - installed: 1.1.1 - latest: 1.2.0 - Update available! Your version of dbt-core is out of date! You can find instructions for upgrading here: https://docs.getdbt.com/docs/installation Plugins: - databricks: 1.1.0 - Update available! - spark: 1.1.0 - Update available! At least one plugin is out of date or incompatible with dbt-core. You can find instructions for upgrading here: https://docs.getdbt.com/docs/installation (I realize this isn't the most up-to-date version, but I didn't see any closed issues related to this) The operating system you're using: macOS Monterey v12.0.1 The output of python --version: Python 3.8.13 Looks like this bug was recently raised over in the dbt-spark repo. Hopefully we can apply the same fix here.
gharchive/issue
2022-08-24T16:37:17
2025-04-01T04:33:56.555027
{ "authors": [ "TannerHopkins" ], "repo": "databricks/dbt-databricks", "url": "https://github.com/databricks/dbt-databricks/issues/162", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
634794547
Outline review Hi @vivekv73y & Sowmya :wave: Hope you're well and thanks a million for this brilliant outline. Everything's looking solid now! My only comment would be to add a Q&A session after each section - and if that means the content is a bit too long, we can always rid away from the final section on visualizing tweet location and do a part II 😄 Cheers, Adel Hi @adelnehme , Thanks very much for the quick review and kind comments. I have updated the README file to include a Q&A session at the end of each section. Best wishes Vivek & Sowmya
gharchive/issue
2020-06-08T17:15:25
2025-04-01T04:33:56.574576
{ "authors": [ "adelnehme", "vivekv73y" ], "repo": "datacamp/Brand-Analysis-using-Social-Media-Data-in-R-Live-Training", "url": "https://github.com/datacamp/Brand-Analysis-using-Social-Media-Data-in-R-Live-Training/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1739536563
Pydantic V2 support Pydantic will soon release their V2 rewrite. This library isn't compatible with it. This might break some users, depending on how they installed this library. Then somebody has installed the library with pip install dataclass-mapper[pydantic], it will already force the user to use a version < 2.0.0. However if somebody has installed it with pip install dataclass-mapper pydantic, it will (in the future) install Pydantic V2 and then crash. We should prevent that with some version in the library. And also we should support the new version. [x] Add a version check and add warnings [x] Support new Pydantic library Implemented with 1.8.0
gharchive/issue
2023-06-03T13:16:07
2025-04-01T04:33:56.580366
{ "authors": [ "jakobkogler" ], "repo": "dataclass-mapper/dataclass-mapper", "url": "https://github.com/dataclass-mapper/dataclass-mapper/issues/18", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2510977721
Correction for raw affiliation MIT/NASA Goddard, Cambridge, MA, USA Correction needed for raw affiliation MIT/NASA Goddard, Cambridge, MA, USA raw_affiliation_name: MIT/NASA Goddard, Cambridge, MA, USA new_rors: 0171mag52 previous_rors: 01cyfxe35 works_examples: W4401451184 contact: b4dee591ed953e5d303da46fe8fa6842:6d59f4c5 @ ourresearch.org This issue was accepted and ingested by the OpenAlex team on 2024-10-10. The new affiliations should be visible within the next 7 days.
gharchive/issue
2024-09-06T18:22:47
2025-04-01T04:33:56.595637
{ "authors": [ "dataesri" ], "repo": "dataesr/openalex-affiliations", "url": "https://github.com/dataesr/openalex-affiliations/issues/3728", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2618386186
Correction for raw affiliation Tabin Jean-Pierre, professeur de sociologie, Haute école de travail social et de la santé EESP et université de Lausanne (Suisse) Correction needed for raw affiliation Tabin Jean-Pierre, professeur de sociologie, Haute école de travail social et de la santé EESP et université de Lausanne (Suisse) raw_affiliation_name: Tabin Jean-Pierre, professeur de sociologie, Haute école de travail social et de la santé EESP et université de Lausanne (Suisse) new_rors: 0278ff426 previous_rors: 029005e08 works_examples: W1923614379;W1911583998 contact: d3acff210105c2446586893d491dae49:a78dcb7637e778a4943e7b2ae98cf63326 @ hetsl.ch This issue was accepted and ingested by the OpenAlex team on 2024-11-18. The new affiliations should be visible within the next 7 days.
gharchive/issue
2024-10-28T13:11:07
2025-04-01T04:33:56.598011
{ "authors": [ "dataesri" ], "repo": "dataesr/openalex-affiliations", "url": "https://github.com/dataesr/openalex-affiliations/issues/6492", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2622095352
Correction for raw affiliation LISAH, University of Montpellier, AgroParisTech, INRAE, Institut Agro Montpellier, IRD, Montpellier, France Correction needed for raw affiliation LISAH, University of Montpellier, AgroParisTech, INRAE, Institut Agro Montpellier, IRD, Montpellier, France raw_affiliation_name: LISAH, University of Montpellier, AgroParisTech, INRAE, Institut Agro Montpellier, IRD, Montpellier, France new_rors: 02kbmgc12;051escj72;05q3vnk25;03rnk6m14;003vg9w96;05deqk823 previous_rors: 02kbmgc12;051escj72;05q3vnk25;03rnk6m14;003vg9w96 works_examples: W4392582347 contact: f0ae19184db8f6158c707b32a8cf81a7:1a6001699e8a4667c70c2e8130ccf3b34171 @ inrae.fr This issue was accepted and ingested by the OpenAlex team on 2024-11-18. The new affiliations should be visible within the next 7 days.
gharchive/issue
2024-10-29T19:01:10
2025-04-01T04:33:56.600497
{ "authors": [ "dataesri" ], "repo": "dataesr/openalex-affiliations", "url": "https://github.com/dataesr/openalex-affiliations/issues/7040", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2623194526
Correction for raw affiliation CNES Toulouse France; IRSTEA/University of Montpellier TETIS Montpellier France Correction needed for raw affiliation CNES Toulouse France; IRSTEA/University of Montpellier TETIS Montpellier France raw_affiliation_name: CNES Toulouse France; IRSTEA/University of Montpellier TETIS Montpellier France new_rors: 051escj72;04h1h0y33;0458hw939 previous_rors: 051escj72;04h1h0y33 works_examples: W4234868018 contact: 5370739efdc675c06521f32f096bfb4a:94076e08e50c5b853a17da5f2dfacbdd4702 @ inrae.fr This issue was accepted and ingested by the OpenAlex team on 2024-12-19. The new affiliations should be visible within the next 7 days.
gharchive/issue
2024-10-30T07:55:11
2025-04-01T04:33:56.602741
{ "authors": [ "dataesri" ], "repo": "dataesr/openalex-affiliations", "url": "https://github.com/dataesr/openalex-affiliations/issues/7246", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2718880588
Correction for raw affiliation Western University Correction needed for raw affiliation Western University raw_affiliation_name: Western University new_rors: 02grkyz14 previous_rors: 05cgtjz78 works_examples: W4319790822;W4382723359;W4386076548;W4322489680;W4387999649;W4321614769;W4315478430;W4387029269;W4315640536;W4322490223;W4387266420;W4321501520;W4327544910;W4381189975;W4320488102;W4327545073;W4390046661;W4387385136;W4318954124;W4381799994;W4319458023;W4389629549;W4386249513;W4385900206;W4386972561;W4387708662;W4319068399;W4385658592;W4375930397;W4386723256;W4386249057;W4385334833;W4387708501;W4385492240;W4389002293;W4320482034;W3123401757;W4386417799;W4366320858;W4390113936;W4387215865;W4389317873;W4318957125;W4319311603;W4393459224;W4393573975;W4390574535;W4390298645;W4398894221;W4389135093;W4361271106;W4385805068;W4389920500;W4386249082;W4386376369;W4386997571;W4327941905;W4386050273;W4386127432;W4387732701;W4385839398;W4367856450;W4386210114;W4386568621;W4389822960;W4381189738;W4376850558;W4319083120;W4384928505;W4386250144;W4385224046;W4389947605;W4387384456;W4389078325;W4384026911;W4389615918;W4389353432;W4379509077;W4386495524;W4386862744;W4385616787;W4387957670;W4387821371;W4324123729;W4390404783;W4387940175;W4385163394;W4387161393;W4361828956;W4324266072;W4379392013;W4387017287;W4388520375;W4377088286;W4394231523;W4393614945;W4393690898;W4386844239;W4360591561;W4393802534;W4317894152;W4387300036;W4362002287;W4381415497;W4317894139;W4364295667;W4387698800;W4393576387;W4320014916;W4375928548;W4388727027;W4394025235;W4379285265;W4377240979;W4385885009;W4367296517;W4388647598;W4386928715;W4385340298;W4388638470;W4386492848;W4362592362;W4393620518;W4387565956;W4387918311;W4386348348;W4393785264;W4393687981;W4394043096;W4377246087;W4378696856;W4393468051;W4390366879;W4386733095;W4385896688;W4388895231;W4394045077;W4387957387;W4386200970;W4389924416;W4386973552;W4387985549;W4389170238;W4393765603;W4386536838;W4384165533;W4390059549;W4393603780;W4387968570;W4385412636;W4388239675;W4324265372;W4366960435;W4367296796;W4389884743;W4385338557;W4394383201;W4360828019;W4385653528;W4313893459;W4393680763;W4381295152;W4381952341;W4384572824;W4393579871;W4362685181;W4362500201;W4386392862;W4389894697;W4393544835;W4388519778;W4394043385;W4383567661;W4389934888;W4362648545;W4321767023;W4376473027;W4386297780;W4388304706;W4393679716;W4321763811;W4320011284;W4323896443;W4387091749;W4327779692;W4393430368;W4386317369;W4389356807;W4393537539;W4393720814;W4386241633;W4389095781;W4385844278;W4384659488;W4391640239;W4320864453;W4390574503;W4322733534;W4389561031;W4391639690;W4382362705;W4385412767;W4313579447;W4313681502;W4379382097;W4317952735;W4327892680;W4361991102;W4366982994;W4379207407;W4393622861;W4386641032;W4387252721;W4327573636;W4366401156;W4379466740;W4394039754;W4384693996;W4386459760;W4386470748;W4387222387;W4389050268;W4387830969;W4313681431;W4313681510;W4377021290;W4378374569;W4390958018;W4388186615;W4389263020;W4321380144;W4387466093;W4317516632;W4320723715;W4320725597;W4376637272;W4383620254;W4387129578;W4389833189;W4393536833;W4317856737;W4322756208;W4377141359;W4386172985;W4380362172;W4386397098;W4388564464;W4362500366;W4381333225;W4394335294;W4379398281;W4321501330;W4322756130;W4324045344;W4376107399;W4386004191;W4387693950;W4388896119;W4393454635;W4386367258;W4386705744;W4387986020;W4388143862;W4321243456;W4321445166;W4323351797;W4366827969;W4378506470;W4383956314;W4385344759;W4389334386;W4386368373;W4386870986;W4389514819;W4322756231;W4366528996;W4383896207;W4386641068;W4387246804;W4366703111;W4385992220;W4388996326;W4387906777;W4394519198;W4385435160;W4389493910;W4385246225;W4383876039;W4377002668;W4389324069;W4321484207;W4386698784;W4321353836;W4375861634;W4387009531;W4387095741;W4380151820;W4382721994;W4382996706;W4386948335;W4318714501;W4319863569;W4377832775;W4321763676;W4385992228;W4389327742;W4384806978;W4318384513;W4320925288;W4394215269;W4366150443;W4320737011;W4389353004;W4389108254;W4386275979;W4318715078;W4387501691;W4386169437;W4386241595;W4388778011;W4387547315;W4386768249;W4386103834;W4389614662;W4317698256;W4386901768;W4388579373;W4318339907;W4382457278;W4361771667;W4380324579;W4386451621;W4385589131;W4386247690;W4389462449;W4387970049;W4386291375;W4383219933;W4385194714;W4387094902;W4386469648;W4367296367;W4367304200;W4386242549;W4388626321;W4382704544;W4389557203;W4393890840;W4393895787;W4387501529;W4388579472;W4388162053;W4389671316;W4324378789;W4385341601;W4367296581;W4394104135;W4389743100;W4320010114;W4388579362;W4386915401;W4321435955;W4387009599;W4377089432;W4387938564;W4367305046;W4392008162;W4386868835;W4367050288;W4388846524;W4394385459;W4394146280;W4383960249;W4323830126;W4319083356;W4394523494;W4384523005;W4387665359;W4394181603;W4394099414;W4394325067;W4323020930;W4394412196;W4394102160;W4394089209;W4394303115;W4394324580;W4367308277;W4394373361;W4394414217;W4386708710 contact: f89c1accee93594440ec0d77d278eb4f:75541746d34809e7 @ uwo.ca This issue was accepted and ingested by the OpenAlex team on 2024-12-19. The new affiliations should be visible within the next 7 days.
gharchive/issue
2024-12-04T22:05:32
2025-04-01T04:33:56.609743
{ "authors": [ "dataesri" ], "repo": "dataesr/openalex-affiliations", "url": "https://github.com/dataesr/openalex-affiliations/issues/9469", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1215087946
rkyv support I'm trying to use openraft with 20k+ messages per second, and I seem to be running into quite a bit of overhead simply allocating/copying/serializing stuff. I'm currently using bincode, but would like to switch to rkyv as I believe it will do better. Unfortunately rkyv doesn't work with serde Serialize/Deserialize (due to fundamental differences around supporting zero-copy), and thus there would need to be support added into openraft for me to be able to use rkyv to serialize things like VoteRequest/AppendEntriesRequest/InstallSnapshotRequest. I think it would just be a matter of adding some different derives behind a feature flag for rkyv. Is there interest in this? If so I will try to make a PR. There is an ongoing PR to make serde an optional feature: #243 What you need looks similar to this? An optional feature that enables several rkyv derive? I do not have any experience on rkyv other than reading its API doc:) A PR is welcomed! :DDD Adding rkyv support would be a nice addition indeed. It's a very efficient format wrt (de)serialization speed which can easily become a bottleneck. IMHO the solution outlined in the comment https://github.com/datafuselabs/openraft/pull/243#pullrequestreview-901486343 in the currently open serde PR would be a very simple and good solution wrt maintainablity. Hi All, I'm looking at adding this feature and ran into a small issue. Is there any issue with moving away from using AnyError directly in some of the messages? An example: https://github.com/datafuselabs/openraft/blob/main/openraft/src/error.rs#L319 The issue being that AnyError is defined in the external crate so we can't easily add the rkyv traits to it Hi All, I'm looking at adding this feature and ran into a small issue. Is there any issue with moving away from using AnyError directly in some of the messages? An example: https://github.com/datafuselabs/openraft/blob/db047b978e08fdb6f23bc4c27f94d4bd36ab925b/openraft/src/error.rs#L314-L320 The issue being that AnyError is defined in the external crate so we can't easily add the rkyv traits to it (Edit) I see now that the AnyError crate is actually apart of drmindrmer's repositories. I guess my question becomes if rkyv support should be added there? (Edit 2) Looking more closely at this, it doesn't appear that the error's themselves would/should ever be sent over the network so it doesn't look like there would be any reason to add this serialization to them. I should have a pr for this sometime today. It is possible to add rkyv to AnyError. But AFAIK, only types that need to be persisted on disk need rkyv support. Errors in openraft are only used for transport. Does it have to be encoded in the rkyv format for transport? E.g., for `send_append_entries(): https://github.com/datafuselabs/openraft/blob/1202d1b059f7e07f0f3353b74f0b2795f063ab56/openraft/src/network.rs#L49-L52 The response and response errors are small structs, maybe it's possible to use some other serialization such as serde to encode responses?
gharchive/issue
2022-04-25T21:56:48
2025-04-01T04:33:56.628924
{ "authors": [ "MikaelCall", "djahandarie", "drmingdrmer", "zach-schoenberger" ], "repo": "datafuselabs/openraft", "url": "https://github.com/datafuselabs/openraft/issues/316", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1333497200
lineage_emitter_kafka.py connection refused Describe the bug When I tried to run lineage_emitter_kafka.py result is connection refused. Broker is located on port 29092, I tried with public IP/localhost but no change when I tried with default value "broker" it can't solve the naming says: "Failed to resolve 'broker:29092". Port is already open on firewall. Datahub runs with Docker and everything is located on Azure VM. To Reproduce Steps to reproduce the behavior: Download the script from here https://raw.githubusercontent.com/datahub-project/datahub/master/metadata-ingestion/examples/library/lineage_emitter_kafka.py Open file with nano and change the "broker" and "schema-registry" parts with yours Run script with python3 Expected behavior Script will run without error. Screenshots Desktop (please complete the following information): OS: Linux 20.04 Browser: Edge Version latest Additional context None. I changed the config and now it is working. Solution is changing schema_registry url and bootstrap with your publicIP.
gharchive/issue
2022-08-09T16:22:14
2025-04-01T04:33:56.637336
{ "authors": [ "maxmanus96" ], "repo": "datahub-project/datahub", "url": "https://github.com/datahub-project/datahub/issues/5596", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1487350466
feat(CI): Add python virtualenvs to github caches Checklist [ ] The PR conforms to DataHub's Contributing Guideline (particularly Commit Message Format) [ ] Links to related issues (if applicable) [ ] Tests for the changes have been added/updated (if applicable) [ ] Docs related to the changes have been added/updated (if applicable). If a new feature has been added a Usage Guide has been added for the same. [ ] For any breaking change/potential downtime/deprecation/big changes an entry has been made in Updating DataHub @szalai1 I was looking into this a bit more and I don't think this is the right way to implement caching. I think we might be able to use the example under "Caching projects that use setup.py" from https://github.com/actions/setup-python/blob/main/docs/advanced-usage.md#caching-packages, but not 100% that it'll work with our matrix setup since it doesn't let us set the cache key ourselves. More broadly, we want to cache pip's cache directory, but not the venv itself. The issue with caching the venv directory is that it can cause pip to leave outdated packages as-is instead of downloading and installing the newer version of the package. That means that a build without caching could differ from a build with the cache enabled. @hsheth2 thanks for review. so these are valid concerns. I think the main "problem" with setup-python is that it caches the downloads and not the installs, which is the majority of the time. A solution to not leaving outdated installs in venv would be: pip install -r requirments.txt --dry-run > would_be_installed.txt use hashFiles('./would_be_installed.txt') } for caching This way we can be sure that we invalidate the cached venv whenever a new package is available. @szalai1 yup that should work. One thing to be careful of is that we still always want to run the pip install -e . step. @hsheth2 Added dry run hashes to the venv cache name to trigger rebuild if anything would change. install -e . will run everytime, even if the venv is resotored. This saved 7mins for me locally for smoke-tests.
gharchive/pull-request
2022-12-09T19:41:08
2025-04-01T04:33:56.644002
{ "authors": [ "hsheth2", "szalai1" ], "repo": "datahub-project/datahub", "url": "https://github.com/datahub-project/datahub/pull/6725", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1112779340
dpkg --remove --ignore-depends=git-annex git-annex (if installed) before installing git-annex-standalone deb Was trying to update git annex to fresh build from datalad/git-annex but that failed due to $> src/datalad_installer.py git-annex -m datalad/git-annex ... About to run the following command as an administrator: dpkg -i '/home/yoh/.tmp/tmp4cg2zkax/git-annex-standalone_8.20211231+git126-gf7a3c1355-1~ndall+1_amd64.deb' Proceed? [y/a/n] y 2022-01-24T10:13:11-0500 [INFO ] datalad_installer Running: sudo dpkg -i '/home/yoh/.tmp/tmp4cg2zkax/git-annex-standalone_8.20211231+git126-gf7a3c1355-1~ndall+1_amd64.deb' [sudo] password for yoh: Selecting previously unselected package git-annex-standalone. dpkg: regarding .../git-annex-standalone_8.20211231+git126-gf7a3c1355-1~ndall+1_amd64.deb containing git-annex-standalone: git-annex-standalone conflicts with git-annex git-annex (version 8.20211123-1) is present and installed. I think the right way to handle this would be to check first if git-annex ( not git-annex-standlone) is installed, and if it is -- uninstall it explicitly first and then proceed to dpkg -i @yarikoptic Should this be done for just the datalad/git-annex and datalad/git-annex:tested methods or also the deb-url method? I assume this should not be done when installing a .deb into a specific target directory rather than system-wide, correct?
gharchive/issue
2022-01-24T15:15:53
2025-04-01T04:33:56.653083
{ "authors": [ "jwodder", "yarikoptic" ], "repo": "datalad/datalad-installer", "url": "https://github.com/datalad/datalad-installer/issues/99", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1000057581
Update tokio and futures crates Fix the vulnerable version of the tokio crate (RUSTSEC-2021-0072). Update the yanked version of the futures crate. [x] This PR has been added to CHANGELOG.md (at the top of the list); Codecov Report Merging #87 (8583318) into main (49dcb88) will increase coverage by 0.00%. The diff coverage is n/a. :exclamation: Current head 8583318 differs from pull request most recent head 3f9954a. Consider uploading reports for the commit 3f9954a to get more accurate results @@ Coverage Diff @@ ## main #87 +/- ## ======================================= Coverage 78.82% 78.83% ======================================= Files 49 49 Lines 2938 2939 +1 ======================================= + Hits 2316 2317 +1 Misses 622 622 Impacted Files Coverage Δ datanymizer_engine/src/utils.rs 100.00% <0.00%> (ø) Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update ed318cf...3f9954a. Read the comment docs.
gharchive/pull-request
2021-09-18T15:06:16
2025-04-01T04:33:56.680693
{ "authors": [ "codecov-commenter", "evgeniy-r" ], "repo": "datanymizer/datanymizer", "url": "https://github.com/datanymizer/datanymizer/pull/87", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2241436920
Animate UI Would be cool to have a animation for pressing the button Also add sound
gharchive/issue
2024-04-13T08:43:40
2025-04-01T04:33:56.700076
{ "authors": [ "Herkarl", "foodelevator" ], "repo": "datasektionen/darkmode", "url": "https://github.com/datasektionen/darkmode/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
340415689
JAVA-1900: Support virtual tables (aka system views) There are a couple of things here that I'm not sure about. Is removing the describe for virtual keyspaces/table really what we want to do? Are there applications upstream which might utilize it in some way. 2.Is having a bunch of unpopulated fields in the table/keyspace acceptable, or do we want another type? Is removing the describe for virtual keyspaces/table really what we want to do? Are there applications upstream which might utilize it in some way. I've given this a bit more thought. Returning an empty string is nice to automatically "ignore" the element if you're generating a script for the whole schema: StringBuilder sb = new StringBuilder(); for (KeyspaceMetadata keyspace : keyspaces) { sb.append(keyspace.describeWithChildren(true)); } However it doesn't work that well for an individual keyspace, because the empty string is not a valid CQL query: session.execute(keyspace.describeWithChildren(true)); // SyntaxError: line 0:-1 no viable alternative at input '<EOF>' So maybe we could instead have the describe methods throw UnsupportedOperationException, and add a Describable.isDescribable() method that you call beforehand to check if you can get a script for this element. Is having a bunch of unpopulated fields in the table/keyspace acceptable, or do we want another type? I think it's acceptable. We should just make sure that non-collection fields that can be unpopulated for virtual elements return Optional. As far as I can tell, the only such field is getId() for tables. PS - for my previous suggestion: So maybe we could instead have the describe methods throw UnsupportedOperationException, and add a Describable.isDescribable() method to check beforehand. This makes things more explicit, but on the other hand it adds a bit more ceremony when dealing with script generation. So it's open for debate, any thoughts?
gharchive/pull-request
2018-07-11T21:40:13
2025-04-01T04:33:56.706715
{ "authors": [ "GregBestland", "olim7t" ], "repo": "datastax/java-driver", "url": "https://github.com/datastax/java-driver/pull/1054", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2116117190
[#1908] improvement(docs): Upgrade the link of java docs from 0.3.1 to 0.4.0 What changes were proposed in this pull request? Change the link value of the Java docs from 0.3.1 to 0.4.0. Why are the changes needed? Version 0.4.0 will be released. We need to update the docs link accordingly. Fix: # (issue) Does this PR introduce any user-facing change? N/A. How was this patch tested? N/A. Is this #1098 correct? @yuqi1129 Is this #1098 correct? Sorry, my mistakes
gharchive/pull-request
2024-02-03T01:55:27
2025-04-01T04:33:56.712451
{ "authors": [ "jerryshao", "yuqi1129" ], "repo": "datastrato/gravitino", "url": "https://github.com/datastrato/gravitino/pull/1998", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2241461732
[MINOR] improvement(catalog-doris): Upgrade Doris CI image version to 0.1.3 What changes were proposed in this pull request? Change the Doris CI image version from 0.1.2 to 0.1.3 Why are the changes needed? 0.1.3 remove log from container stdout, which is better for CI framework Does this PR introduce any user-facing change? N/A How was this patch tested? N/A should wait for #2887
gharchive/pull-request
2024-04-13T09:33:03
2025-04-01T04:33:56.714787
{ "authors": [ "zhoukangcn" ], "repo": "datastrato/gravitino", "url": "https://github.com/datastrato/gravitino/pull/2922", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
583953427
Redesign query tool Make it more 'content page-like' with filterbar on top and breadcrumbs instead of sidebar. put description on top, only keep data an api tabs in table from fact view ticket: generate region breadcrumb with 'Home', 'Bundesland', 'Regierungsbezirk', 'Kreis', ('Gemeinde') generate attribute breadcrumb with 'Home', 'Attribute', 'Argument' generate teaser from attribute docs generate headline with attribute label and year generate table (region is fix, columns: years, argument) make table sortable build .csv download button (with nice filenames) generate check list of attributes / arguments for this topic change table on checked arguments This has been implemented apart from 'make table sortable'. We'll close this issue for now.
gharchive/issue
2020-03-18T19:27:05
2025-04-01T04:33:56.752540
{ "authors": [ "crijke", "sjockers" ], "repo": "datenguide/datenguide", "url": "https://github.com/datenguide/datenguide/issues/106", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1713417973
Announce MarkdownDB on Twitter Acceptance [ ] draft tweets thread [ ] publish @olayway looks good - ship early, ship often. Let's get this announce out asap 😄 @rufuspollock do we want to invite people to Discord? If so, which one? See last tweet in the thread. @rufuspollock nvmd, I think I'll just say star us on GH or sth and if we ever create a discord channel we can tweet about it too ;) Done https://twitter.com/datopian/status/1658818662600523782
gharchive/issue
2023-05-17T08:40:38
2025-04-01T04:33:56.763037
{ "authors": [ "olayway", "rufuspollock" ], "repo": "datopian/markdowndb", "url": "https://github.com/datopian/markdowndb/issues/18", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
812644464
Adding Akita to new angular 11 app - "af" not found in collection "@schematics/angular" Hello, I am submitting a bug I Have a new angular 11 project and I followed the docs and installed Akita by running: ng add @datorama/akita The process completes and the console says the the packages were installed. I can confirm this as I see them added to my packages.json file. I ran npm i just to be sure. The schematics are not getting picked up however. when I run ng g af core/core --plain I get the following error: An unhandled exception occurred: Schematic "af" not found in collection "@schematics/angular". See "C:\Users\jorda\AppData\Local\Temp\ng-97z847\angular-errors.log" for further details. However, I was able to get it to work using the following: ng g @datorama/akita:as core/core Maybe I am missing something but this does not match the docs for use with angular. Thanks okay, my bad I found the docs on angular schematics I ran the following: ng config cli.defaultCollection @datorama/akita and now the schematics are picked up. It might be worth linking to this from the angular architecture page However, in my angular.json file I had a "project" with a name that had a period in it ie: "myngapp.webui" this did not work and the above command failed, i had to remove the period form the project name in order to get it to work. I had to do the same for the "myngapp.webUI-e2e" project. I changed that to "myngappwebuie2e" anyways, great work, !
gharchive/issue
2021-02-20T15:39:59
2025-04-01T04:33:56.767327
{ "authors": [ "kingjordan" ], "repo": "datorama/akita", "url": "https://github.com/datorama/akita/issues/623", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1078947207
Region select in DM is not colored where? i don't see it This is in direct calls in upper right corner. That should be fixed now
gharchive/issue
2021-12-13T20:04:30
2025-04-01T04:33:56.771268
{ "authors": [ "datsfilipe", "markusgod" ], "repo": "datsfilipe/smooth-theme", "url": "https://github.com/datsfilipe/smooth-theme/issues/5", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
143800654
Update API Endpoint To provide better performances on reSmush.it, API Endpoint domain has changed Hi, As specified on http://www.resmush.it/, the API Endpoint will expire on September, 30 2016. You should modify endpoint if you want your plugin still working. Sorry I totally missed this. Thanks for keeping on the ball.
gharchive/pull-request
2016-03-27T11:35:06
2025-04-01T04:33:56.828395
{ "authors": [ "charlyie", "davgothic" ], "repo": "davgothic/SmushIt", "url": "https://github.com/davgothic/SmushIt/pull/3", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
875973296
Guided Scenario Builder Guided assistance feature to help inexperienced editors create new scenarios Likely an option that can be selected when creating a new scenario. Also should be able to turn this feature on or off within the editor interface BACKLOGGED: Ran out of time to implement; needed to focus on integrating with the backend team to set up new endpoints that worked with the new schema before deploying changes to the server. This is a quality of life issue. However, the customers have made it clear that this is a top-priority QoL feature.
gharchive/issue
2021-05-05T01:48:14
2025-04-01T04:33:56.829820
{ "authors": [ "erotondo" ], "repo": "david-fisher/320-S21-Track2", "url": "https://github.com/david-fisher/320-S21-Track2/issues/275", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1557068377
Add twitter-style link previews Right now, for posts that I imported from Twitter, I use their truncated display URLs (but not the t.co shortened URLs) as the display text for any links. This helps deal with links that are long/unruly, but it also makes it less easy to tell at a glance where the link goes. A great way to deal with this could be to add twitter-style link preview cards that use the link's OpenGraph data if present. Despite being written for Vue, this blog post at least details a good markup structure that I could use: https://ahmednagi.com/snippets/create-link-previews/ I think I can consider this done as of 81839647cf9bd1f6288fd9d8768630b98d28f3e1
gharchive/issue
2023-01-25T18:03:01
2025-04-01T04:33:56.856400
{ "authors": [ "davidcelis" ], "repo": "davidcelis/davidcel.is", "url": "https://github.com/davidcelis/davidcel.is/issues/11", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1804917313
Add completed Items to Search Currently, the search pool only contains items which were removed from the completed list before (or global items). It would be nice if the search results also contain items which are completed (but still shown at the bottom of the list). Selecting a search result should set it to uncompleted. This behaviour could be an optional setting. Added now in release 1.0.4. I did not make it a setting but changed the behavior to do only that.... Hopefully no one wants it the other way... Please test and validate if it works for you. Thank you. Looks good so far!
gharchive/issue
2023-07-14T13:47:42
2025-04-01T04:33:56.874005
{ "authors": [ "MariusSp", "davideshay" ], "repo": "davideshay/groceries", "url": "https://github.com/davideshay/groceries/issues/66", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
520740464
a path separator bug prevents ctypesgen running directly on windows I'm using ctypesgen without Cygwin, and it failed to generate output file due to the regular expression below didn't match any preamble py file. printer_python\printer.py: def get_preamble(...) m = re.search("preamble/(\d)_(\d).py$", fp) The slash(/) in the regular expression doesn't match any preamble file in Windows (preamble\2_5.py). verified it using latest source and got an error below: line 26, in get_preamble m = re.search(preamble_pattern, fp) File "C:\Python27_64\lib\re.py", line 146, in search return _compile(pattern, flags).search(string) File "C:\Python27_64\lib\re.py", line 251, in _compile raise error, v # invalid expression sre_constants.error: unbalanced parenthesis strange. Just tried this with py27 and py3, both successful. What is your version exactly? v2.7.15:ca079a3ea3 64-bits The code below seams work: m = re.search("preamble\\(\d)_(\d).py$", fp) I think this is enough because every 'fp' always starts with 'preamble': m = re.search("(\d)_(\d).py$", fp) That is a good point. I guess the windows path separator needs additional escaping anyway when used as a regular expression. Fixed (again)
gharchive/issue
2019-11-11T04:58:25
2025-04-01T04:33:56.944638
{ "authors": [ "AllenLius", "olsonse" ], "repo": "davidjamesca/ctypesgen", "url": "https://github.com/davidjamesca/ctypesgen/issues/75", "license": "bsd-2-clause", "license_type": "permissive", "license_source": "bigquery" }
221677300
The resizing stopped to work. Dear David J Bradshaw, first of all big thank you for this script. It worked great for me now for over 2 years. But all of the sudden it stopped working on sites where it used to work. It simply doesn't resize anymore. I will send you an error log, maybe this can point you in a direction, what could be going wrong. The iframe is set to 150px height, allthough the content is much larger. Any help is much appreciated! [iFrameSizer][Host page] IFrame scrolling disabled for etb iframeResizer.js:97 [iFrameSizer][Host page][init] Sending msg to iframe (etb:8:false:true:32:true:true:null:offset:null:null:0:true:parent) jquery-migrate.min.js?ver=1.4.1:2 JQMIGRATE: Migrate is installed, version 1.4.1 iframeResizer.js:97 [iFrameSizer][Host page][iFrame.onload] Sending msg to iframe (etb:8:false:true:32:true:true:null:offset:null:null:0:true:parent) iframeResizer.contentWindow.js:60 [iFrameSizer][etb] HTML & body height set to "auto" iframeResizer.contentWindow.js:60 [iFrameSizer][etb] Enable public methods iframeResizer.contentWindow.js:60 [iFrameSizer][etb] Added event listener: Animation Start iframeResizer.contentWindow.js:60 [iFrameSizer][etb] Added event listener: Animation Iteration iframeResizer.contentWindow.js:60 [iFrameSizer][etb] Added event listener: Animation End iframeResizer.contentWindow.js:48 The deviceorientation event is deprecated on insecure origins, and support will be removed in the future. You should consider switching your application to a secure origin, such as HTTPS. See https://goo.gl/rStTGz for more details. a @ iframeResizer.contentWindow.js:48 d @ iframeResizer.contentWindow.js:143 k @ iframeResizer.contentWindow.js:152 l @ iframeResizer.contentWindow.js:162 n @ iframeResizer.contentWindow.js:184 e @ iframeResizer.contentWindow.js:81 f @ iframeResizer.contentWindow.js:655 J @ iframeResizer.contentWindow.js:689 postMessage (async) m @ iframeResizer.js:418 (anonymous) @ iframeResizer.js:492 iframeResizer.contentWindow.js:60 [iFrameSizer][etb] Added event listener: Device Orientation Change iframeResizer.contentWindow.min.js:9 [iFrameSizer][etb] Added event listener: Transition End iframeResizer.contentWindow.min.js:9 [iFrameSizer][etb] Added event listener: Window Clicked iframeResizer.contentWindow.min.js:9 [iFrameSizer][etb] Enable MutationObserver iframeResizer.contentWindow.min.js:9 [iFrameSizer][etb] Setting up location.hash handlers iframeResizer.contentWindow.min.js:9 [iFrameSizer][etb] Trigger event lock on iframeResizer.contentWindow.min.js:9 [iFrameSizer][etb] Sending message to host page (etb:150:938:init) iframeResizer.min.js:8 [iFrameSizer][Host page] Received: [iFrameSizer]etb:150:938:init iframeResizer.min.js:8 [iFrameSizer][Host page] Checking connection is from: https://www.eticketablanca.com iframeResizer.min.js:8 [iFrameSizer][Host page] Checking height is in range 0-Infinity iframeResizer.min.js:8 [iFrameSizer][Host page] Checking width is in range 0-Infinity iframeResizer.min.js:8 [iFrameSizer][Host page] Requesting animation frame iframeResizer.min.js:8 [iFrameSizer][Host page] IFrame (etb) height set to 150px iframeResizer.contentWindow.min.js:9 [iFrameSizer][etb] Trigger event: Device Orientation Change iframeResizer.contentWindow.min.js:9 [iFrameSizer][etb] No change in size detected iframeResizer.contentWindow.min.js:9 [iFrameSizer][etb] Trigger event lock off iframeResizer.contentWindow.min.js:9 [iFrameSizer][etb] -- inactive-logout.js:42 Last Active on: 1492116082937 iframeResizer.min.js:8 [iFrameSizer][Host page][Window resize] Sending msg to iframe (resize) iframeResizer.contentWindow.min.js:9 [iFrameSizer][etb] Trigger event: Parent window resized iframeResizer.contentWindow.min.js:9 [iFrameSizer][etb] No change in size detected iframeResizer.min.js:8 [iFrameSizer][Host page][Window resize] Sending msg to iframe (resize) iframeResizer.contentWindow.min.js:9 [iFrameSizer][etb] Trigger event: Parent window resized iframeResizer.contentWindow.min.js:9 [iFrameSizer][etb] No change in size detected Most likely an issue with your CSS in the iFrame, try the different resze methods. Where exactly do I change the resize methods? And which options are there? RTFM Thanks David! I got it to work with 'lowestElement'. Thanks for your help., Much appreciated!
gharchive/issue
2017-04-13T20:45:34
2025-04-01T04:33:56.947997
{ "authors": [ "davidjbradshaw", "mrweix" ], "repo": "davidjbradshaw/iframe-resizer", "url": "https://github.com/davidjbradshaw/iframe-resizer/issues/484", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
385126470
Typescript: matchesState typed for StateValue, also works with State Bug or feature request? Maybe a bug? Description: The type signature for matchesState does not support all the functionality available via vanilla JS. The signature (in utils.ts) is export function matchesState( parentStateId: StateValue, childStateId: StateValue, delimiter: string = STATE_DELIMITER ) But the function implementation in practice also accepts a State object. You can see that in @carloslfu's example code for their useMachine library - if you look at the App function you'll see a line (in render(...)) {matchesState(machine.state, 'Off') ? 'Off' : 'On'} In this case machine.state is an instance of State. And this works fine. But if you convert this code to Typescript you'll need to change to using machine.state.value, or it won't compile. (Bug) Expected result: Either matchesState supports both State and StateValue objects in both TS and JS (preferred) OR matchesState doesn't accept State in JS, to make it consistent with the TS version. (Bug) Actual result: matchesState only supports StateValue in TS. (Bug) Potential fix: Change the signature of matchesState to accept either StateValue or State, if you think it's worthwhile to have this convenience (I think it probably is, especially as the function is named matchesState not matchesStateId or matchesStateValue. If you don't think this is a good change I'll suggest to @carloslfu that the example in the useMachine repo be changed to use machine.state.value. I tried to create a PR for this, but as a TS newbie I wasn't sure whether machineState should accept StateValue | StateInterface or StateValue | State. Note: you can (and should) use state.matches(...) to match against an actual state. See here: https://xstate.js.org/docs/guides/states.html#state-methods-and-getters matchesState in JavaScript will actually resolve State instances to its state value. This will be an error in TypeScript, which is expected - you should use state.value or state.matches(value) instead.
gharchive/issue
2018-11-28T06:52:24
2025-04-01T04:33:57.010158
{ "authors": [ "craigglennie", "davidkpiano" ], "repo": "davidkpiano/xstate", "url": "https://github.com/davidkpiano/xstate/issues/260", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2053308602
about ur code do u have pytorch-version of this code? no, just tensorflow
gharchive/issue
2023-12-22T03:03:58
2025-04-01T04:33:57.011184
{ "authors": [ "davidlainesv", "plumedhawk" ], "repo": "davidlainesv/SL-TSSI-DenseNet", "url": "https://github.com/davidlainesv/SL-TSSI-DenseNet/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
130831905
would it be possible to use a chrome extention in this service you are providing for proxies would it be possible to use a chrome extention in this service you are providing for proxies, the chrome extension is from browse sec heres the code require=(function e(t,n,r){function s(o,u){if(!n[o]){if(!t[o]){var a=typeof require=="function"&&require;if(!u&&a)return a(o,!0);if(i)return i(o,!0);var f=new Error("Cannot find module '"+o+"'");throw f.code="MODULE_NOT_FOUND",f}var l=n[o]={exports:{}};t[o][0].call(l.exports,function(e){var n=t[o][1][e];return s(n?n:e)},l,l.exports,e,t,n,r)}return n[o].exports}var i=typeof require=="function"&&require;for(var o=0;o<r.length;o++)s(r[o]);return s})({"./config":[function(require,module,exports){ module.exports = { name: "Browsec", browsec: { apiPrefix: "https://d1blmth2c5vbem.cloudfront.net/v1/", locationApiPrefix: "http://d2nib9hpvmumkf.cloudfront.net/v1/" }, ga: { enabled: true, extension_id: "omghfjlpggmjjaagoclmmobgdodcjboh", tracking_id: 'UA-43024042-1' }, proxy: { defaultCountry: "nl", settings: { version: 4, countries: { nl: { servers: [ { host: "nl1.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl2.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl3.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl4.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl5.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl6.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl7.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl8.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl9.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl10.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl11.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl12.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl13.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl14.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl15.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl16.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl17.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl18.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl19.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl20.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "nl22.postls.com", port: 443, alt_ports: [444, 8443] } ] }, sg: { servers: [ { host: "sg1.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "sg2.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "sg3.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "sg4.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "sg5.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "sg6.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "sg7.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "sg8.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "sg9.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "sg10.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "sg11.postls.com", port: 443, alt_ports: [444, 8443] } ] }, uk: { servers: [ { host: "uk1.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "uk2.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "uk3.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "uk4.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "uk5.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "uk6.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "uk7.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "uk8.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "uk9.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "uk10.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "uk11.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "uk12.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "uk13.postls.com", port: 443, alt_ports: [444, 8443] } ] }, us: { servers: [ { host: "us1.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us2.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us3.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us4.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us5.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us6.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us7.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us8.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us9.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us10.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us11.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us12.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us13.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us14.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us15.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us16.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us17.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us18.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us19.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us20.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us21.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us22.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us23.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us24.postls.com", port: 443, alt_ports: [444, 8443] }, { host: "us25.postls.com", port: 443, alt_ports: [444, 8443] } ] }, us_test: { hidden: true, servers: [ { host: "us-test-1.postls.com", port: 443, alt_ports: [444, 8443] } ] } } } } }; },{}],1:[function(require,module,exports){ /* Returns random int value between 0 (inclusive) and the specified value (exclusive) TODO(grig): more accurate randomInt */ function randomInt(max) { return Math.floor(Math.random() * max); } /* Returns random array element using weights. Element weight should be stored in 'weight' property of array element. If 'weight' property is absent then weight for this element is 1. */ function weightedRandom(array) { var map = []; var totalWeight = 0; for (var i = 0; i < array.length; i++) { var weight = array[i].weight || 1; map.push({ start : totalWeight, end : totalWeight + weight }); totalWeight += weight; } var random = randomInt(totalWeight); for (i = 0; i < map.length; i++) { if ((random >= map[i].start) && (random < map[i].end)) { return array[i]; } } } /* Clone array */ function clone(array) { return array.slice(0); } /* Return shuffled array without modifying original one TODO(grig): replace cloning with a new empty array */ function shuffle(array) { var arrayClone = clone(array); var currentIndex = arrayClone.length; // While there remain elements to shuffle... while (0 !== currentIndex) { // Pick a remaining element... var randomIndex = randomInt(currentIndex); currentIndex--; // And swap it with the current element. var temporaryValue = arrayClone[currentIndex]; arrayClone[currentIndex] = arrayClone[randomIndex]; arrayClone[randomIndex] = temporaryValue; } return arrayClone; } /* Return shuffled array with attention to element weights. Element weight should be stored in 'weight' property of array element. If 'weight' property is absent then weight for this element is 1. Original array is not modified. */ function weightedShuffle(array) { var arrayClone = clone(array); var result = []; for (var i = 0; i < array.length; i++) { var item = weightedRandom(arrayClone); result.push(item); arrayClone.splice(arrayClone.indexOf(item), 1); } return result; } exports.weightedShuffle = weightedShuffle; exports.shuffle = shuffle; },{}],2:[function(require,module,exports){ (function (global){ var $ = require('jquery'); require('./common'); var proxy = global.proxy = require('./proxy'); var scheduler = global.scheduler = require('./update_scheduler'); chrome.proxy.settings.onChange.addListener(proxy.onChange.bind(proxy)); var ui = global.ui = require('./ui'); var browsec = global.browsec = require('./browsec'); var ga = require('./ga'); browsec.init(); chrome.webRequest.onAuthRequired.addListener( function(details, callback) { try { console.group("onAuthRequired"); console.log(details); if (details.isProxy && details.realm == 'Browsec' && (details.challenger.host == 'postlm.com' || details.challenger.host.endsWith(".postlm.com") || details.challenger.host == 'postls.com' || details.challenger.host.endsWith(".postls.com"))) { var username = localStorage.userId || localStorage.email; var password = localStorage.password; // If we don't have username/password // OR // username is email and password is incorrect if (!username || !password || (username.indexOf('@') > 0 && checkAuthLimitExceeded(details))) { var result = browsec.signup(); if (result) { delete localStorage.email; localStorage.userId = username = result.user_id; localStorage.password = password = result.password; } else { proxy.clearProxySettings(); alert(chrome.i18n.getMessage("signup_error")); return; } } console.log("Authenticate with username '%s'", username); callback({authCredentials: {username: username, password: password}}); } else { // If auth request is not from browsec proxy, do not handle it. callback(); } } finally { console.groupEnd(); } }, {urls: ["<all_urls>"]}, ["asyncBlocking"] ); var statusLineRegexp = new RegExp("^HTTP/1.[01] 408"); chrome.webRequest.onHeadersReceived.addListener( function (details) { if (localStorage.connected && localStorage.connected != "off") { if (statusLineRegexp.test(details.statusLine)) { ga.trackEvent("extension", "http_error", details.statusLine); } } }, {urls: ["<all_urls>"]} ); function checkAuthLimitExceeded(details) { var timeFrameMs = 15000; var authRequestsLimit = 3; var authHistory = localStorage.authHistory ? JSON.parse(localStorage.authHistory) : {}; try { console.group("checkAuthLimitExceeded"); var hostAuthHistory = authHistory[details.challenger.host]; if (!hostAuthHistory || !hostAuthHistory.periodStart || !hostAuthHistory.authRequestsNum) { hostAuthHistory = authHistory[details.challenger.host] = {periodStart: details.timeStamp, authRequestsNum: 1}; console.log("No host auth history. Create record: %s", JSON.stringify(hostAuthHistory)); } else { console.log("Host history record already exists: %s", JSON.stringify(hostAuthHistory)); var periodStart = hostAuthHistory.periodStart; if (details.timeStamp - periodStart > timeFrameMs) { console.log("Time frame already finished. Create new row."); hostAuthHistory.periodStart = details.timeStamp; hostAuthHistory.authRequestsNum = 1; } else { console.log("We are still in tracked time frame. Increase auth requests counter"); hostAuthHistory.authRequestsNum++; if (hostAuthHistory.authRequestsNum >= authRequestsLimit) { console.log("Limit exceeded. Return true and clear host auth history."); delete hostAuthHistory.periodStart; delete hostAuthHistory.authRequestsNum; ga.trackEvent("auth_request_limit_exceeded", localStorage.userId || localStorage.email); return true; } } } return false; } finally { console.log("Save auth history: %O", authHistory); localStorage.authHistory = JSON.stringify(authHistory); console.groupEnd(); } } chrome.runtime.onInstalled.addListener(function(details) { console.log("chrome.runtime.onInstalled", details); ga.trackEvent("extension", details.reason, chrome.runtime.getManifest().version); if (details.reason == "install") { proxy.detectDefaultCountry(function() { proxy.setProxySettings(proxy.lastConnected(), function(result) { // Workaroung for Chrome bug $.ajax({ url: "http://www.google.com/favicon.ico", cache: false, complete: function(jqXHR, textStatus) { console.log("Complete test request: %s", textStatus); } }); }); }); } else if (details.reason == "update") { var previousVersion = details.previousVersion; var currentVersion = chrome.runtime.getManifest().version; if (previousVersion <= "2.2.99" && currentVersion >= "2.3.0") { // Upgrade from 2.2 and earlier up to 2.3 and later if (proxy.connected() === "de") { proxy.setProxySettings("uk"); } } else { if (proxy.connected()) { proxy.setProxySettings(proxy.lastConnected(), function() { ui.updateUiConnectionStatus(); }); } } } else { if (proxy.connected()) { proxy.setProxySettings(proxy.lastConnected(), function() { ui.updateUiConnectionStatus(); }); } } ui.updateUiConnectionStatus(); }); chrome.proxy.onProxyError.addListener(function (details) { proxy.checkProxySet(function(isConnected) { if (isConnected) { console.error("Proxy on error: " + JSON.stringify(details)); ga.trackEvent("extension", "proxy_error", JSON.stringify(details)); } }); }); chrome.runtime.onStartup.addListener(function() { console.log("chrome.runtime.onStartup"); delete localStorage.lastRedirectOnError; ui.updateUiConnectionStatus(); ga.trackEvent("extension", "start", chrome.runtime.getManifest().version); }); // TODO if/when we switch to event page, this should be changed scheduler.init(); }).call(this,typeof global !== "undefined" ? global : typeof self !== "undefined" ? self : typeof window !== "undefined" ? window : {}) },{"./browsec":3,"./common":4,"./ga":5,"./proxy":6,"./ui":8,"./update_scheduler":9,"jquery":"jquery"}],3:[function(require,module,exports){ var $ = require('jquery'); var config = require('./config').browsec; var ga = require('./ga'); var browsec = function() { var apiPrefix = config.apiPrefix || "https://d1blmth2c5vbem.cloudfront.net/v1/"; var locationApiPrefix = config.locationApiPrefix || "http://d1blmth2c5vbem.cloudfront.net/v1/"; return { init: function() { $.ajaxSetup({ tryCount : 0, retryLimit : 1, timeout: 10000, retry: function() { this.tryCount++; if (this.tryCount <= this.retryLimit) { //try again $.ajax(this); return true; } else { return false; } } }); }, checkCredentials : function (username, password) { try { console.group("checkCredentials"); var result = false; $.ajax({ type: "POST", url: apiPrefix + "auth", data: { email: username, password: password }, async: false, success: function(data, textStatus, jqXHR) { console.log("Response status: %d", data.status); if (data.status === 0) { result = true; } }, error: function(jqXHR, textStatus, errorThrown) { if (!this.retry()) { ga.trackEvent("check_credentials_error", textStatus, errorThrown); throw errorThrown; } } }); return result; } finally { console.groupEnd(); } }, clearCredentials : function () { delete localStorage.userId; delete localStorage.email; delete localStorage.password; }, ipInfo : function (callback) { var start; $.ajax({ type: "GET", url: locationApiPrefix + "location?locale=" + chrome.runtime.getManifest().current_locale, async: true, success: function(data, textStatus, jqXHR) { callback(data); }, error: function(jqXHR, textStatus, errorThrown) { this.retry(); }, beforeSend: function() { start = Date.now(); }, complete: function() { var duration = Date.now() - start; console.debug("Location request duration %d ms", duration); } }); }, signup : function() { try { console.group("Signing up new user"); var result = null; $.ajax({ type: "POST", url: apiPrefix + "signup", async: false, success: function(data, textStatus, jqXHR) { console.log("Response status: %d", data.status); if (data.status === 0) { result = data.result; } else { ga.trackEvent("error_signup", data); } }, error: function(jqXHR, textStatus, errorThrown) { if (!this.retry()) { ga.trackEvent("error_signup", textStatus, errorThrown); } } }); return result; } finally { console.groupEnd(); } }, servers: function(callback) { var canceled = false; var xhr = $.ajax({ type: "GET", url: apiPrefix + "servers", async: true, success: function(data, textStatus, jqXHR) { if (canceled) { if (typeof callback === 'function') { callback('canceled'); } return; } if (typeof callback === 'function') { callback(null, data); } }, error: function(jqXHR, textStatus, errorThrown) { if (canceled) { // Aborted request return textStatus == 'abort', errorThrown == 'abort'. // We might have checked for these statuses; however, once a task // has been canceled, any status should be ignored. console.log('servers', 'request canceled: status:', textStatus, 'error:', errorThrown); // may check for XHR if (typeof callback === 'function') { callback('canceled'); } return; } var error = {status: textStatus, error: errorThrown}; console.error("browsec.servers", JSON.stringify(error)); ga.trackEvent("error", "browsec.servers", JSON.stringify(error)); callback(error); } }); var task = { cancel: function() { canceled = true; xhr.abort(); } }; return task; } }; }(); module.exports = browsec; },{"./config":"./config","./ga":5,"jquery":"jquery"}],4:[function(require,module,exports){ require('./utils/format'); require('./utils/ends_with'); require('./utils/starts_with'); require('./utils/error_handler'); },{"./utils/ends_with":11,"./utils/error_handler":12,"./utils/format":13,"./utils/starts_with":14}],5:[function(require,module,exports){ (function (global){ var config = require('./config').ga; if (config.enabled) { // Google Analytics (function() { var ga = document.createElement('script'); ga.type = 'text/javascript'; ga.async = true; ga.src = 'https://ssl.google-analytics.com/ga.js'; var s = document.getElementsByTagName('script')[0]; s.parentNode.insertBefore(ga, s); })(); global._gaq = global._gaq || []; _gaq.push(['_setAccount', config.tracking_id]); _gaq.push(['_trackPageview']); var ga = { trackEvent : function (category, action, label, value, noninteraction) { // Track events only in production console.log("Track event: " + JSON.stringify([category, action, label, value, noninteraction])); if (chrome.i18n.getMessage("@@extension_id") === config.extension_id) { _gaq.push(['_trackEvent', category, action, label, value, noninteraction]); } } }; } else { var ga = { trackEvent: function() { console.log("ga.trackEvent", "[DISABLED]", arguments); } }; } module.exports = ga; }).call(this,typeof global !== "undefined" ? global : typeof self !== "undefined" ? self : typeof window !== "undefined" ? window : {}) },{"./config":"./config"}],6:[function(require,module,exports){ (function (global){ var ga = require('./ga'); var weightedShuffle = require('./array').weightedShuffle; var config = require('./config').proxy; var servers = require('./server_list'); var proxy = function() { var bypassList = ["google-analytics.com"]; var pacScript = 'function domainIs(host,domain){return host==domain||dnsDomainIs(host,"."+domain)}function FindProxyForURL(url,host){var config="{0}";var bypass={1};host=host.toLowerCase();if(isPlainHostName(host)||isInNetEx(host,"127.0.0.0/8")||isInNetEx(host,"10.0.0.0/8")||isInNetEx(host,"172.16.0.0/12")||isInNetEx(host,"192.168.0.0/16")||isInNetEx(host,"fc00::/7")||isInNetEx(host,"fe80::/10")){return"DIRECT"}for(var i=0;i<bypass.length;i++){if(domainIs(host,bypass[i])){return"DIRECT"}}return config}'; var defaultCountry = config.defaultCountry || "nl"; function getServerString(countryConfig, altPort) { var serverString = ""; var servers = weightedShuffle(countryConfig.servers); /* jshint -W004 */ for (var i = 0; i < servers.length; i++) { var server = servers[i]; var port = getServerPort(server, altPort); serverString += "HTTPS " + server.host + ":" + port + "; "; } if (countryConfig.backupServers) { var backupServers = weightedShuffle(countryConfig.backupServers); for (var i = 0; i < backupServers.length; i++) { var server = backupServers[i]; var port = getServerPort(server, altPort); serverString += "HTTPS " + server.host + ":" + port + "; "; } } /* jshint +W004 */ console.log("Srv string: " + serverString); return serverString; } function getServerPort(server, altPort) { if (altPort && Array.isArray(server.alt_ports) && server.alt_ports.indexOf(altPort) >= 0) { return altPort; } else { return server.port; } } return { countryCodes: function(callback, includeHidden) { if (typeof includeHidden === 'undefined') { includeHidden = this.isIncludeHidden(); } servers.get(function(err, settings) { if (err) { console.error(err); callback([]); return; } var countries = settings.countries; var countryCodes = Object.keys(countries).filter(function(countryCode) { return includeHidden || !countries[countryCode].hidden; }).filter(function(countryCode) { return Array.isArray(countries[countryCode].servers); }); callback(countryCodes); }); }, isIncludeHidden : function() { return localStorage.includeHidden && JSON.parse(localStorage.includeHidden); }, setIncludeHidden : function(value) { localStorage.includeHidden = JSON.stringify(value); }, isProxySet : function (config) { if (config.levelOfControl == "controlled_by_this_extension") if (config.value.mode == 'pac_script') return true; return false; }, checkProxySet : function (callback) { chrome.proxy.settings.get({incognito : false}, function (config) { var isConnected = this.isProxySet(config); callback(isConnected); }.bind(this)); }, setProxySettings : function (countryCode, callback) { chrome.proxy.settings.get({incognito : false}, function(config) { if (typeof config != 'undefined' && config.levelOfControl != "controllable_by_this_extension" && config.levelOfControl != "controlled_by_this_extension") { console.warn("Proxy settings levelOfControl is: " + config.levelOfControl); ga.trackEvent("extension", "alert", "error_unable_change_proxy_settings " + config.levelOfControl); alert(chrome.i18n.getMessage('error_unable_change_proxy_settings')); if (typeof callback == "function") { callback(false); } } else { var altPort = localStorage.altPort && Number(localStorage.altPort); servers.get(function(err, settings) { if (err) { callback(false); return; } var countries = settings.countries; var countryConfig = countries[countryCode]; if (!countryConfig) { console.warn("No servers for country: " + countryCode + ", rolling back to default country"); countryCode = defaultCountry; countryConfig = countries[countryCode]; } if (!countryConfig) { console.error("Failed to find any suitable servers, turning off"); delete localStorage.lastConnected; localStorage.connected = 'off'; callback(false); return; } var serverString = getServerString(countryConfig, altPort); var proxyConfig = { mode: "pac_script", pacScript: { data: pacScript.format(serverString, JSON.stringify(bypassList)) } }; chrome.proxy.settings.set({scope: 'regular', value: proxyConfig}, function() { localStorage.connected = localStorage.lastConnected = countryCode; if (typeof callback == "function") { callback(true); } }); }); } }.bind(this)); }, clearProxySettings : function (callback) { chrome.proxy.settings.clear({scope: 'regular'}, function() { localStorage.connected = "off"; if (typeof callback == "function") { callback(true); } }); }, logProxySettings : function () { chrome.proxy.settings.get({incognito : false}, function(config) { console.log(JSON.stringify(config)); }); }, onChange : function(details) { console.log("Proxy on change: " + JSON.stringify(details)); this.checkProxySet(function (isConnected) { ui.setUiConnectionStatus(isConnected); if (!isConnected) { localStorage.connected = "off"; } }); }, connected : function(localStorage) { localStorage = localStorage || global.localStorage; var status = localStorage.connected; if (status === "off") { return false; } else { return status; } }, lastConnected : function(localStorage) { localStorage = localStorage || global.localStorage; return localStorage.lastConnected || defaultCountry; }, detectDefaultCountry : function(callback) { browsec.ipInfo(function (location) { /* jshint laxbreak:true */ if (location.continent_code === "NA" || location.continent_code === "SA") { // If user isfrom North or South America send him to US server defaultCountry = "us"; } else if (location.country_code === "SG" // Singapore || location.country_code === "MY" // Malaysia || location.country_code === "ID" // Indonesia || location.country_code === "PH" // Philippines || location.country_code === "TH" // Thailand ) { defaultCountry = "sg"; } else if (location.country_code === "UK" || location.country_code === "IE") { defaultCountry = "uk"; } /* jshint laxbreak:false */ callback(); }); } }; }(); module.exports = proxy; }).call(this,typeof global !== "undefined" ? global : typeof self !== "undefined" ? self : typeof window !== "undefined" ? window : {}) },{"./array":1,"./config":"./config","./ga":5,"./server_list":7}],7:[function(require,module,exports){ /** Manages list of servers TODO(grig): transaction / last known good support @module ./server_list */ var ga = require('./ga'); var defaultConfig = require('./config').proxy.settings || {}; /* TODO(grig): test callback status */ exports.set = function set(value, callback) { var error = validate(value); if (error) { throw new ValidationError(error); } try { localStorage.currentConfig = JSON.stringify(value); if (typeof callback === 'function') { setTimeout(function() { callback(null); }, 0); } } catch(e) { if (typeof callback === 'function') { setTimeout(function() { callback(e); }, 0); } } }; function validate(settings) { if (typeof settings !== 'object' || settings === null) { return 'settings should be an object'; } if (!settings.hasOwnProperty('countries')) { return 'settings should have a "countries" property'; } var config = settings.countries; if (Object.keys(config).length === 0) { return "config should have at least one key (country)"; } for (var countryCode in config) { if (!config.hasOwnProperty(countryCode)) { continue; } var country = config[countryCode]; if (!Array.isArray(country.servers) && !Array.isArray(country.premium_servers)) { return "country must have either 'servers' or 'premium_servers' fields set to an array of servers"; } var error; if (Array.isArray(country.servers)) { error = validateServers(country.servers); if (error) { return error; } } if (Array.isArray(country.premium_servers)) { error = validateServers(country.premium_servers); if (error) { return error; } } } return null; } function validateServers(servers) { for (var i = 0; i < servers.length; i++) { var server = servers[i]; if (typeof server.host !== 'string') { return "server must have a 'host' property"; } if (!(typeof server.port === 'string' || typeof server.port === 'number')) { return "server must have a 'port' property"; } } return null; } function ValidationError(message) { this.name = 'ValidationError'; this.message = message || ''; this.stack = (new Error()).stack; } ValidationError.prototype = Object.create(Error.prototype); ValidationError.prototype.constructor = ValidationError; /* TODO(grig): test no countries in default config TODO(grig): test no proxy config TODO(grig): test errors / exports.get = function get(callback) { var value = localStorage.currentConfig; if (typeof value === 'string') { try { var currentConfig = JSON.parse(value); var error = validate(currentConfig); if (error) { throw new ValidationError(error); } setTimeout(function() { callback(null, currentConfig || defaultConfig); }, 0); } catch (e) { var version = 'n/a'; if (chrome.runtime && chrome.runtime.getManifest()) { version = chrome.runtime.getManifest().version; } var message = e.toString(); try { // find application-level file name and line number from the stack var stack = e.stack.split("\n"); for (var i = 0; i < stack.length; i++) { if (/ at ..js:.*/.test(stack[i])) { message = message + stack[i]; break; } } } catch (_) { // nothing } ga.trackEvent("error", version, message, 0, false); console.warn("invalid current configuration, fall back to default: ", e); delete localStorage.currentConfig; setTimeout(function() { callback(null, defaultConfig); }, 0); } } else { setTimeout(function() { callback(null, defaultConfig); }, 0); } }; exports.clear = function(callback) { delete localStorage.currentConfig; if (typeof callback === 'function') { setTimeout(callback, 0); } }; },{"./config":"./config","./ga":5}],8:[function(require,module,exports){ var proxy = require('./proxy'); var ui = { setUiConnectionStatus : function (connected) { if (connected) { chrome.browserAction.setIcon({path: 'images/icon-enabled.png'}); // chrome.browserAction.setTitle({title: chrome.i18n.getMessage('browser_action_active_title')}); } else { chrome.browserAction.setIcon({path: 'images/icon-disabled.png'}); // chrome.browserAction.setTitle({title: chrome.i18n.getMessage('browser_inaction_active_title')}); } }, updateUiConnectionStatus : function () { proxy.checkProxySet(function (isConnected) { this.setUiConnectionStatus(isConnected); }.bind(this)); } }; module.exports = ui; },{"./proxy":6}],9:[function(require,module,exports){ /** @module ./update_scheduler */ var update_task = require('./update_task'); var REFRESH_TIMEOUT = 636001000; var RETRY_TIMEOUT = 5601000; var retryTimeout = RETRY_TIMEOUT; var refreshTimeout = REFRESH_TIMEOUT; function run() { var task; var canceled = false; var timeoutId = setTimeout(function() { console.log("retry"); run(); // TODO schedule on next tick canceled = true; if (task) { task.cancel(); } }, retryTimeout); task = update_task.start(function(err) { if (err) { console.warn("Error updating severs, retrying in " + retryTimeout / (60 * 1000) + " minutes", err); // nothing; retry will be triggered anyway return; } if (canceled) { console.error("Attempting to complete an already-canceled update task"); return; } clearTimeout(timeoutId); setTimeout(function() { console.log("refresh"); // TODO(grig): possible parallel task executions? run(); }, refreshTimeout); }); } exports.init = function init(options) { console.log("init"); options = options || {}; retryTimeout = options.retryTimeout || retryTimeout; refreshTimeout = options.refreshTimeout || refreshTimeout; setTimeout(run, 0); }; },{"./update_task":10}],10:[function(require,module,exports){ /** @module ./update_task */ var server_list = require('./server_list'); var proxy = require('./proxy'); var browsec = require('./browsec'); exports.start = function update(callback) { console.log("update"); var task = function() { var canceled = false; var subtask; subtask = browsec.servers(function(err, servers) { if (err) { if (typeof callback === 'function') { callback(err); } return; } if (canceled) { if (typeof callback === 'function') { callback("canceled"); } return; } server_list.set(servers, function(err) { if (err) { if (typeof callback === 'function') { callback(err); } return; } if (canceled) { if (typeof callback === 'function') { callback("canceled"); } return; } if (proxy.connected(localStorage)) { // update proxy settings var currentCountry = proxy.lastConnected(localStorage); proxy.setProxySettings(currentCountry, function(result) { if (canceled) { if (typeof callback === 'function') { callback("canceled"); } return; } if (typeof callback === 'function') { if (result) { callback(); } else { callback(new Error("failed to update proxy settings")); } } }); } else { if (canceled) { if (typeof callback === 'function') { callback("canceled"); } return; } else { if (typeof callback === 'function') { callback(null); return; } } } }); }); var task = { cancel: function cancel() { canceled = true; if (subtask) { subtask.cancel(); } } }; return task; }(); return task; }; },{"./browsec":3,"./proxy":6,"./server_list":7}],11:[function(require,module,exports){ if (!String.prototype.endsWith) { String.prototype.endsWith = function(suffix) { return this.indexOf(suffix, this.length - suffix.length) !== -1; }; } },{}],12:[function(require,module,exports){ var ga = require('../ga'); // Error handler window.onerror = function (message, filename, lineno) { try { console.error("message: {0}\nfilename: {1}\nlineno: {2}".format(message, filename, lineno)); var version = 'n/a'; if (chrome.runtime.getManifest()) { version = chrome.runtime.getManifest().version; } ga.trackEvent("error", version, "{0} at {1}:{2}".format(message, filename, lineno), 0, false); } catch (e) { console.error(e); } return false; }; },{"../ga":5}],13:[function(require,module,exports){ // Helper functions function format(str, args) { return str.replace(/{(\d+)}/g, function(match, number) { return typeof args[number] != 'undefined' ? args[number] : match; }); } if (!String.prototype.format) { String.prototype.format = function() { return format(this, arguments); }; } module.exports = format; },{}],14:[function(require,module,exports){ if (!String.prototype.startsWith) { String.prototype.startsWith = function(prefix) { return this.indexOf(prefix) === 0; }; } },{}]},{},[2]); are you publishing unlicensed code as an issue? whats wrong with you people. show a little respect to people who worked on something.
gharchive/issue
2016-02-02T22:02:05
2025-04-01T04:33:57.151795
{ "authors": [ "akaashjain1", "davidmann4" ], "repo": "davidmann4/agario-feeder-bot", "url": "https://github.com/davidmann4/agario-feeder-bot/issues/124", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
660412567
Improve usage of "pieces" For things like a lime wedge or a cherry, we currently use the unit pieces. This makes the caption say something like "1 pieces of cherry", which sounds bizarre. To fix this, we should add a field to ingredients that indicates when it will always be a piece, and instead format it differently, to instead say something like "1 cherry". While working on this, we should also take into account plurals as well. For example, it should say "1 cherry" or "2 cherries". If the ingredient is a "piece" ingredient, we should store that into the backend, just like #9 wants to do with units. We should also probably restrict usage of custom ingredients to the user who created the ingredient. If an ingredient is common enough, we can make it a built-in ingredient. For now, we shouldn't redirect user content to a newly created built-in ingredient, but that is something we should consider.
gharchive/issue
2020-07-18T22:12:18
2025-04-01T04:33:57.166449
{ "authors": [ "davidov541" ], "repo": "davidov541/MixologyJournal", "url": "https://github.com/davidov541/MixologyJournal/issues/8", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
62220344
Copy project's steps Be able to duplicate an existing step in a project, as well as duplicate the step into other projects. Sort of like the Clone project, but finer grained. https://octopusdeploy.uservoice.com/forums/170787-general/suggestions/7166247-add-copy-feature-to-deployment-step-along-with-edi This would be smashing, especially duplicate the step into other projects :+1: @Milkybar100 can already to the first half (copy within a project), but to another project... How would you expect the UI to work? Ideally: Click the hamburger on a step in Process It has option "copy to project" Brings up drop down list with all projects Select project from list and step is duplicated there Manually reorder in new project Would that work? Probably not all that difficult... Just need to find the time :) If you're up for diving into some javascript, I can give you some pointers? With some pointers I'd be up for giving it a shot :) Okay, cool. Welcome to the fun :) To get going with dev, clone the repo, and in chrome extensions, click "Load Unpacked Extension", and point that to the src folder. If you've got the published chrome extension loaded, make sure you untick that. Every time you make a change to the javasript, you need to reload the extension on this page. A good thing to note while we are on this page is the 'background page' link, which enables you to see the chrome debug tools for the background worker. You'll probably want to add a new template similar to this one which has the source for the popup. You'll also want a function like this one (but much simpler) to get the projects and show the popup. Once its closed, it will need to send a message similar to this one via chrome.runtime.sendMessage, but with the added project id. That function will need to be injected into the page (so its in the scope of the angular app) like this one does. You'll need to modify the handleCloneStepRequest function to use the project id to get the project, then get the link to the deployment process out of that, and then modify this function to deal with a potentially different source and destination DeploymentProcess. Depends if you want to modify the existing Clone Step menu item, or add another... (I'm undecided, personally.). If you want modify the existing one, you'll just need to modify the cloneStep function to call your angular popup. If you want to add another menu item, you'll need to add a new menu item like it does here. There are a bunch of jasmine tests, specifically here and here. Hopefully they should make sense and be easy to extend. There is a slightly odd issue around the build server - make sure your tests pass on both phantomjs 1.9.8 and 2.x. Just looking at how much i've written here - I'm hoping I haven't scared you away :smile:. Its really not as scary as it sounds - everything you'll need to do has existing examples, so its just a case of adapting those to the new requirements. Shout if anything is unclear or you need assistance!
gharchive/issue
2015-03-16T21:58:15
2025-04-01T04:33:57.177622
{ "authors": [ "Milkybar100", "davidroberts63", "matt-richardson" ], "repo": "davidroberts63/OctoPygmy", "url": "https://github.com/davidroberts63/OctoPygmy/issues/11", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
588244550
Image not showing in editor if loaded from url Hi folks, I have some html content in a variable as string which contains some image as url. But In vue editor the content is not showing. Can you please help me out. var description = '<div><h2>This is heading</h2><p>This is paragraph.</p><p><img src="https://www.google.com/images/branding/googlelogo/1x/googlelogo_color_272x92dp.png"></p></div> Laravel mix:- 2.1.11 Vue2-editor:- 2.6.6 Do you embeding with image function ?? Editor.insertEmbed(cursorLocation, "image", url); resetUploader(); Examples https://github.com/davidroyer/vue2-editor/blob/master/src/App.vue https://github.com/davidroyer/vue2-editor#example---custom-image-handler This issue is stale because it has been open for 60 days with no activity. This issue was closed because it has been inactive for 7 days since being marked as stale.
gharchive/issue
2020-03-26T08:34:35
2025-04-01T04:33:57.180991
{ "authors": [ "davidroyer", "deepakbasnal", "papalardo" ], "repo": "davidroyer/vue2-editor", "url": "https://github.com/davidroyer/vue2-editor/issues/245", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
152922705
JS Stack Trace I'm hitting this issue. I think it's an issue with the .gyp stuff. Running on node 6. (node) v8::ObjectTemplate::Set() with non-primitive values is deprecated (node) and will stop working in the next major release. ==== JS stack trace ========================================= Security context: 0xffdfdec9fa9 <JS Object>#0# 1: .node [module.js:568] [pc=0x29d0fa3f4604] (this=0x10a9db3ee9f9 <an Object with map 0x10929e2ae641>#1#,module=0x294c25c7f449 <a Module with map 0x10929e2181b9>#2#,filename=0x294c25c7f421 <String[64]: /XXX/node_modules/geoip2/lib/node_mmdb.node>) 2: load [module.js:~447] [pc=0x29d0f9b260f6] (this=0x294c25c7f449 <a Module with map 0x10929e2181b9>#2#,filename=0x294c25c7f421 <String[64]: /XXX/node_modules/geoip2/lib/node_mmdb.node>) 3: tryModuleLoad(aka tryModuleLoad) [module.js:415] [pc=0x29d0f9738edd] (this=0xffdfde04189 <undefined>,module=0x294c25c7f449 <a Module with map 0x10929e2181b9>#2#,filename=0x294c25c7f421 <String[64]: /XXX/node_modules/geoip2/lib/node_mmdb.node>) 4: _load [module.js:~381] [pc=0x29d0f9f4bbad] (this=0x10a9db3eea79 <JS Function Module (SharedFunctionInfo 0x10a9db328801)>#3#,request=0x3453ce4b ==== C stack trace =============================== 1: v8::Template::Set(v8::Local<v8::Name>, v8::Local<v8::Data>, v8::PropertyAttribute) 2: init(v8::Local<v8::Object>) 3: node::DLOpen(v8::FunctionCallbackInfo<v8::Value> const&) 4: v8::internal::FunctionCallbackArguments::Call(void (*)(v8::FunctionCallbackInfo<v8::Value> const&)) 5: v8::internal::MaybeHandle<v8::internal::Object> v8::internal::(anonymous namespace)::HandleApiCallHelper<false>(v8::internal::Isolate*, v8::internal::(anonymous namespace)::BuiltinArguments<(v8::internal::BuiltinExtraArguments)1>) 6: v8::internal::Builtin_HandleApiCall(int, v8::internal::Object**, v8::internal::Isolate*) 7: 0x29d0f960961b (node) v8::ObjectTemplate::Set() with non-primitive values is deprecated (node) and will stop working in the next major release. ==== JS stack trace ========================================= Security context: 0xffdfdec9fa9 <JS Object>#0# 1: .node [module.js:568] [pc=0x29d0fa3f4604] (this=0x10a9db3ee9f9 <an Object with map 0x10929e2ae641>#1#,module=0x294c25c7f449 <a Module with map 0x10929e2181b9>#2#,filename=0x294c25c7f421 <String[64]: /XXX/node_modules/geoip2/lib/node_mmdb.node>) 2: load [module.js:~447] [pc=0x29d0f9b260f6] (this=0x294c25c7f449 <a Module with map 0x10929e2181b9>#2#,filename=0x294c25c7f421 <String[64]: /XXX/node_modules/geoip2/lib/node_mmdb.node>) 3: tryModuleLoad(aka tryModuleLoad) [module.js:415] [pc=0x29d0f9738edd] (this=0xffdfde04189 <undefined>,module=0x294c25c7f449 <a Module with map 0x10929e2181b9>#2#,filename=0x294c25c7f421 <String[64]: /XXX/node_modules/geoip2/lib/node_mmdb.node>) 4: _load [module.js:~381] [pc=0x29d0f9f4bbad] (this=0x10a9db3eea79 <JS Function Module (SharedFunctionInfo 0x10a9db328801)>#3#,request=0x3453ce4b 1: v8::Template::Set(v8::Local<v8::Name>, v8::Local<v8::Data>, v8::PropertyAttribute) 2: init(v8::Local<v8::Object>) 3: node::DLOpen(v8::FunctionCallbackInfo<v8::Value> const&) 4: v8::internal::FunctionCallbackArguments::Call(void (*)(v8::FunctionCallbackInfo<v8::Value> const&)) 5: v8::internal::MaybeHandle<v8::internal::Object> v8::internal::(anonymous namespace)::HandleApiCallHelper<false>(v8::internal::Isolate*, v8::internal::(anonymous namespace)::BuiltinArguments<(v8::internal::BuiltinExtraArguments)1>) 6: v8::internal::Builtin_HandleApiCall(int, v8::internal::Object**, v8::internal::Isolate*) 7: 0x29d0f960961b I can confirm that updating nan dependency to version 2.3.3 solves this problem. Fixed in 1.0.3
gharchive/issue
2016-05-04T03:32:54
2025-04-01T04:33:57.187235
{ "authors": [ "davidtsai", "eddiemoore", "pandrese" ], "repo": "davidtsai/node-geoip2", "url": "https://github.com/davidtsai/node-geoip2/issues/8", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1661494747
Code GPT in VSC seems broken Now in VSC, every prompt has an extra 4000 tokens tacked onto it, so every prompt is rejected as having too many tokens. When it says how many tokens the prompt is, its always 4000 higher (which is the max token count) than it should be. This issue just started happening a couple days ago. Everything worked as intended until then. Ive tried reinstalling, re-inputting my API (which I have tested elsewhere and does work), and other tricks, no luck. Other GPT extensions seem to work fine still, but this was the best one until it broke. Please lower the number of max tokens to 2500 and it should work. Sadly no. Just tried as suggested. Same result. It seems to calculate the real token value, add a flat 4K, then send the errorOn Apr 10, 2023, at 10:56 PM, Daniel Avila @.***> wrote: Please lower the number of max tokens to 2500 and it should work. —Reply to this email directly, view it on GitHub, or unsubscribe.You are receiving this because you authored the thread.Message ID: @.***> My mistake. I spoke too soon. After a restart, yes, that did fix it, sort of. Its still calculating tokens wrong or at least differently than it did before. Bummer. Can only do very very small code segments since it seems to add the input and output together now, as part of the same token limit. Essentially looks like it can do half as much code as before; but your right that it does work with small chunks of code and 2500 instead of 4K set. On Apr 10, 2023, at 11:15 PM, @.*** wrote:Sadly no. Just tried as suggested. Same result. It seems to calculate the real token value, add a flat 4K, then send the errorOn Apr 10, 2023, at 10:56 PM, Daniel Avila @.***> wrote: Please lower the number of max tokens to 2500 and it should work. —Reply to this email directly, view it on GitHub, or unsubscribe.You are receiving this because you authored the thread.Message ID: @.***>
gharchive/issue
2023-04-10T22:45:58
2025-04-01T04:33:57.191969
{ "authors": [ "Fowlplaychiken", "davila7" ], "repo": "davila7/code-gpt-docs", "url": "https://github.com/davila7/code-gpt-docs/issues/116", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
422708374
Correlation Tracker Unable to Start Expected Behavior I am trying to pass the opencv colour image as input to the tracker. This is a test code as of now. This will be changed to video processing. Commenting the Issue line, causes proper conversion between the opencv image to and from dlib image. Current Behavior Error occurs when i try to carry out tracking. Unable to start the tracker. Steps to Reproduce Code Section #include <dlib/image_processing.h> #include <dlib/gui_widgets.h> #include <dlib/image_io.h> #include <dlib/opencv.h> #include "dlib/matrix.h" #include "opencv2/core/core.hpp" #include "opencv2/highgui/highgui.hpp" #include "opencv2/imgproc/imgproc.hpp" #include "opencv2/imgcodecs/imgcodecs.hpp" #include "opencv2/videoio/videoio.hpp" #include "opencv2/objdetect/objdetect.hpp" #include "opencv2/core/utility.hpp" #include <opencv2/opencv.hpp> #include #include <stdio.h> #include #include <stdlib.h> #include <string.h> #include <assert.h> #include <time.h> using namespace dlib; using namespace cv; using namespace std; int main() { Mat frame= imread("sample.png"); dlib::cv_image<bgr_pixel> correct(frame); correlation_tracker tracker; /************************************ Issue Line / tracker.start_track(correct, dlib::centered_rect(dlib::point(93,110), 38, 86)); /***************************************************/ Mat temp= dlib::toMat(correct); namedWindow("Display"); imshow("Display" , temp); waitKey(0); } Building Command: g++ -std=c++11 sample.cpp -o sample_proj pkg-config --cflags --libs opencv dlib-1 Error: /tmp/ccfroENo.o: In function dlib::lapack::binding::gesvd(char, char, int, int, double*, int, double*, double*, int, double*, int, double*, int)': sample.cpp:(.text._ZN4dlib6lapack7binding5gesvdEcciiPdiS2_S2_iS2_iS2_i[_ZN4dlib6lapack7binding5gesvdEcciiPdiS2_S2_iS2_iS2_i]+0x91): undefined reference to dgesvd_' /tmp/ccfroENo.o: In function dlib::blas_bindings::cblas_gemm(dlib::blas_bindings::CBLAS_ORDER, dlib::blas_bindings::CBLAS_TRANSPOSE, dlib::blas_bindings::CBLAS_TRANSPOSE, int, int, int, double, double const*, int, double const*, int, double, double*, int)': sample.cpp:(.text._ZN4dlib13blas_bindings10cblas_gemmENS0_11CBLAS_ORDERENS0_15CBLAS_TRANSPOSEES2_iiidPKdiS4_idPdi[_ZN4dlib13blas_bindings10cblas_gemmENS0_11CBLAS_ORDERENS0_15CBLAS_TRANSPOSEES2_iiidPKdiS4_idPdi]+0x71): undefined reference to cblas_dgemm' collect2: error: ld returned 1 exit status Platform: Ubuntu 16.04.6 LTS Version dlib-19.10 from github Compiler: g++ COLLECT_GCC=g++ COLLECT_LTO_WRAPPER=/usr/lib/gcc/x86_64-linux-gnu/5/lto-wrapper Target: x86_64-linux-gnu Configured with: ../src/configure -v --with-pkgversion='Ubuntu 5.4.0-6ubuntu1~16.04.11' --with-bugurl=file:///usr/share/doc/gcc-5/README.Bugs --enable-languages=c,ada,c++,java,go,d,fortran,objc,obj-c++ --prefix=/usr --program-suffix=-5 --enable-shared --enable-linker-build-id --libexecdir=/usr/lib --without-included-gettext --enable-threads=posix --libdir=/usr/lib --enable-nls --with-sysroot=/ --enable-clocale=gnu --enable-libstdcxx-debug --enable-libstdcxx-time=yes --with-default-libstdcxx-abi=new --enable-gnu-unique-object --disable-vtable-verify --enable-libmpx --enable-plugin --with-system-zlib --disable-browser-plugin --enable-java-awt=gtk --enable-gtk-cairo --with-java-home=/usr/lib/jvm/java-1.5.0-gcj-5-amd64/jre --enable-java-home --with-jvm-root-dir=/usr/lib/jvm/java-1.5.0-gcj-5-amd64 --with-jvm-jar-dir=/usr/lib/jvm-exports/java-1.5.0-gcj-5-amd64 --with-arch-directory=amd64 --with-ecj-jar=/usr/share/java/eclipse-ecj.jar --enable-objc-gc --enable-multiarch --disable-werror --with-arch-32=i686 --with-abi=m64 --with-multilib-list=m32,m64,mx32 --enable-multilib --with-tune=generic --enable-checking=release --build=x86_64-linux-gnu --host=x86_64-linux-gnu --target=x86_64-linux-gnu Thread model: posix gcc version 5.4.0 20160609 (Ubuntu 5.4.0-6ubuntu1~16.04.11) Use cmake, it will link to things for you. But if you don't want to use cmake then read the instructions: http://dlib.net/compile.html. You in particular are getting linker errors for missing LAPACK symbols. Link to whatever LAPACK you compiled dlib against. Expected Behavior I am trying to pass the opencv colour image as input to the tracker. This is a test code as of now. This will be changed to video processing. Commenting the Issue line, causes proper conversion between the opencv image to and from dlib image. Current Behavior Error occurs when i try to carry out tracking. Unable to start the tracker. Steps to Reproduce Code Section #include <dlib/image_processing.h> #include <dlib/gui_widgets.h> #include <dlib/image_io.h> #include <dlib/opencv.h> #include "dlib/matrix.h" #include "opencv2/core/core.hpp" #include "opencv2/highgui/highgui.hpp" #include "opencv2/imgproc/imgproc.hpp" #include "opencv2/imgcodecs/imgcodecs.hpp" #include "opencv2/videoio/videoio.hpp" #include "opencv2/objdetect/objdetect.hpp" #include "opencv2/core/utility.hpp" #include <opencv2/opencv.hpp> #include #include <stdio.h> #include #include <stdlib.h> #include <string.h> #include <assert.h> #include <time.h> using namespace dlib; using namespace cv; using namespace std; int main() { Mat frame= imread("sample.png"); dlib::cv_image<bgr_pixel> correct(frame); correlation_tracker tracker; /************************************ Issue Line / tracker.start_track(correct, dlib::centered_rect(dlib::point(93,110), 38, 86)); /***************************************************/ Mat temp= dlib::toMat(correct); namedWindow("Display"); imshow("Display" , temp); waitKey(0); } Building Command: g++ -std=c++11 sample.cpp -o sample_proj pkg-config --cflags --libs opencv dlib-1 Error: /tmp/ccfroENo.o: In function dlib::lapack::binding::gesvd(char, char, int, int, double*, int, double*, double*, int, double*, int, double*, int)': sample.cpp:(.text._ZN4dlib6lapack7binding5gesvdEcciiPdiS2_S2_iS2_iS2_i[_ZN4dlib6lapack7binding5gesvdEcciiPdiS2_S2_iS2_iS2_i]+0x91): undefined reference to dgesvd_' /tmp/ccfroENo.o: In function dlib::blas_bindings::cblas_gemm(dlib::blas_bindings::CBLAS_ORDER, dlib::blas_bindings::CBLAS_TRANSPOSE, dlib::blas_bindings::CBLAS_TRANSPOSE, int, int, int, double, double const*, int, double const*, int, double, double*, int)': sample.cpp:(.text._ZN4dlib13blas_bindings10cblas_gemmENS0_11CBLAS_ORDERENS0_15CBLAS_TRANSPOSEES2_iiidPKdiS4_idPdi[_ZN4dlib13blas_bindings10cblas_gemmENS0_11CBLAS_ORDERENS0_15CBLAS_TRANSPOSEES2_iiidPKdiS4_idPdi]+0x71): undefined reference to cblas_dgemm' collect2: error: ld returned 1 exit status Platform: Ubuntu 16.04.6 LTS Version dlib-19.10 from github Compiler: g++ COLLECT_GCC=g++ COLLECT_LTO_WRAPPER=/usr/lib/gcc/x86_64-linux-gnu/5/lto-wrapper Target: x86_64-linux-gnu Configured with: ../src/configure -v --with-pkgversion='Ubuntu 5.4.0-6ubuntu1~16.04.11' --with-bugurl=file:///usr/share/doc/gcc-5/README.Bugs --enable-languages=c,ada,c++,java,go,d,fortran,objc,obj-c++ --prefix=/usr --program-suffix=-5 --enable-shared --enable-linker-build-id --libexecdir=/usr/lib --without-included-gettext --enable-threads=posix --libdir=/usr/lib --enable-nls --with-sysroot=/ --enable-clocale=gnu --enable-libstdcxx-debug --enable-libstdcxx-time=yes --with-default-libstdcxx-abi=new --enable-gnu-unique-object --disable-vtable-verify --enable-libmpx --enable-plugin --with-system-zlib --disable-browser-plugin --enable-java-awt=gtk --enable-gtk-cairo --with-java-home=/usr/lib/jvm/java-1.5.0-gcj-5-amd64/jre --enable-java-home --with-jvm-root-dir=/usr/lib/jvm/java-1.5.0-gcj-5-amd64 --with-jvm-jar-dir=/usr/lib/jvm-exports/java-1.5.0-gcj-5-amd64 --with-arch-directory=amd64 --with-ecj-jar=/usr/share/java/eclipse-ecj.jar --enable-objc-gc --enable-multiarch --disable-werror --with-arch-32=i686 --with-abi=m64 --with-multilib-list=m32,m64,mx32 --enable-multilib --with-tune=generic --enable-checking=release --build=x86_64-linux-gnu --host=x86_64-linux-gnu --target=x86_64-linux-gnu Thread model: posix gcc version 5.4.0 20160609 (Ubuntu 5.4.0-6ubuntu1~16.04.11) Meets the same problem and try to fix now. Use cmake, it will link to things for you. But if you don't want to use cmake then read the instructions: http://dlib.net/compile.html. You in particular are getting linker errors for missing LAPACK symbols. Link to whatever LAPACK you compiled dlib against. In my case there's LAPACK. cmake prints information like these: -- Searching for BLAS and LAPACK -- Searching for BLAS and LAPACK -- Found PkgConfig: /usr/bin/pkg-config (found version "0.29.1") -- Checking for module 'cblas' -- No package 'cblas' found -- Checking for module 'lapack' -- Found lapack, version 0.2.20+ds -- Looking for sys/types.h -- Looking for sys/types.h - found -- Looking for stdint.h -- Looking for stdint.h - found -- Looking for stddef.h -- Looking for stddef.h - found -- Check size of void* -- Check size of void* - done -- Found OpenBLAS library Use cmake, it will link to things for you. But if you don't want to use cmake then read the instructions: http://dlib.net/compile.html. You in particular are getting linker errors for missing LAPACK symbols. Link to whatever LAPACK you compiled dlib against. I used dlib in Qt. And I had link to LAPACK by "LIBS += -L/usr/lib/x86_64-linux-gnu/openblas -llapack"
gharchive/issue
2019-03-19T12:37:35
2025-04-01T04:33:57.224414
{ "authors": [ "balajiselvaraj1601", "davisking", "nizqsut" ], "repo": "davisking/dlib", "url": "https://github.com/davisking/dlib/issues/1698", "license": "BSL-1.0", "license_type": "permissive", "license_source": "github-api" }
434556160
quantum_computing_ex.cpp does not compile SHA1 d9af22e53697d3fe676231be4f3432c84c8b3872 On Debian 9.8 AMD 64-bits I got this compilation error when compiling examples (while I compiled successfully recently on Ubuntu 18.04 32-bits on older desktop that I do not have access right now so unknown SHA1). ../../src/init2.c:52: MPFR assertion failed: p >= 2 && p <= ((mpfr_prec_t)((mpfr_uprec_t)(~(mpfr_uprec_t)0)>>1)) /home/qq/dlib/examples/quantum_computing_ex.cpp: In function ‘void shor_decode(dlib::quantum_register&)’: /home/qq/dlib/examples/quantum_computing_ex.cpp:77:6: internal compiler error: Abandon void shor_decode ( ^~~~~~~~~~~ Please submit a full bug report, with preprocessed source if appropriate. See <file:///usr/share/doc/gcc-6/README.Bugs> for instructions. CMakeFiles/quantum_computing_ex.dir/build.make:62 : la recette pour la cible « CMakeFiles/quantum_computing_ex.dir/quantum_computing_ex.cpp.o » a échouée Retried with SHA1 8001b924e63ac920a9047610ece2e4e6266e8a4c no longer see this error. i have this error too......... when i assign my opencv dir cmake .. -G"Unix Makefiles" -DCMAKE_BUILD_TYPE=Release -DDLIB_USE_CUDA=ON -DUSE_SSE2_INSTRUCTIONS=ON -DUSE_SSE4_INSTRUCTIONS=ON -DOpenCV_CONFIG_PATH=/opt/opencv/share/OpenCV -DOpenCV_DIR=/opt/opencv/share/OpenCV its the dlib::rand rnd; - problem after comment all rnd build is ok
gharchive/issue
2019-04-18T01:54:44
2025-04-01T04:33:57.227754
{ "authors": [ "Lecrapouille", "fatalfeel" ], "repo": "davisking/dlib", "url": "https://github.com/davisking/dlib/issues/1734", "license": "BSL-1.0", "license_type": "permissive", "license_source": "github-api" }
1205478988
Deterministic versions of find_max_global and find_min_global Expected Behavior I'm looking for a way to use find_min_global in Python that is 100% deterministic/reproducible. Current Behavior I sometimes get slightly different results in different runs. Looking at the source code, it seems clear that find_max_global is not in fact deterministic, because it changes its behaviour based on the runtime of the target function, which depends on the architecture, whatever else is running at the same time and potentially the parameter values that are being evaluated. Specifically, it scales back the Monte Carlo sampling when that takes too much time compared to evaluating the target function, and ultimately resorts to random searching. Steps to Reproduce I've managed to reproduce this issue with a modified version of the example code: import dlib from math import sin,cos,pi,exp,sqrt def holder_table(x0,x1): # Do some extra work for i in range(100): i = i**2 return -abs(sin(x0)*cos(x1)*exp(abs(1-sqrt(x0*x0+x1*x1)/pi))) x,y = dlib.find_min_global(holder_table, [-10,-10], [10,10], 80) print("optimal inputs: {}".format(x)) print("optimal output: {}".format(y)) I mostly get: optimal inputs: [-8.055023486163353, -9.664590009526423] optimal output: -19.208502567886732 but sometimes one of these: optimal inputs: [-8.064969430503735, 9.662303063982197] optimal output: -19.20747862796451 optimal inputs: [7.628545844250482, 9.5718335077948] optimal output: -17.454451199912715 optimal inputs: [10.0, 10.0] optimal output: -15.140223856952055 You may not be able to reproduce this, as I'm currently running two intensive calculations in the background. I haven't been able to increase the chance of getting one of the alternative results by simply making the target function computationally more intensive. When I run the unmodified example code I always seem to get the same result. Version: 19.22.0 Where did you get dlib: conda-forge Platform: Ubuntu 20.04.4 LTS Compiler: N/A Question I guess what I am asking for is how the global_function_search object can be use to exactly replicate the behaviour of find_min_global, without scaling back the Monte Carlo sampling. (Also, I don't need the log conversion, since I handle that myself.) I've come up with the following: import dlib from math import sin,cos,pi,exp,sqrt def holder_table(x0,x1): return -abs(sin(x0)*cos(x1)*exp(abs(1-sqrt(x0*x0+x1*x1)/pi))) spec = dlib.function_spec(bound1=[-10, -10], bound2=[10, 10]) search = dlib.global_function_search(spec) for i in range(80): next = search.get_next_x() x0, x1 = next.x next.set(-holder_table(x0, x1)) x, y, _ = search.get_best_function_eval() print("optimal inputs: {}".format(list(x))) print("optimal output: {}".format(-y)) Which results in: optimal inputs: [8.055023473597158, 9.664590011114276] optimal output: -19.20850256788674 This is not exactly the same result as above, but it's very close, so may be due to rounding? Is there anything extra that find_max_global does that I've missed? (Longterm, it could also be nice for find_max_global and find_min_global to have a parameter controlling this behaviour.) Yeah that's what you do. global_function_search is the real tool here. find_max_global is just a convenience function so someone can "just do it" without thinking about anything. Like about if log scaling makes sense or if they should use faster but less smart sampling (e.g. it's pointless to spend 1s thinking about what point to sample if the objective function is so fast you can evaluate it a million times per second). So more sophisticated users should just use global_function_search directly :)
gharchive/issue
2022-04-15T10:38:54
2025-04-01T04:33:57.235077
{ "authors": [ "davisking", "oulenz" ], "repo": "davisking/dlib", "url": "https://github.com/davisking/dlib/issues/2567", "license": "BSL-1.0", "license_type": "permissive", "license_source": "github-api" }
971132344
Delete .hgtags After seeing commit https://github.com/davisking/dlib/commit/0a5d5a2c686ff5e69eed8b4b19b0501f29a24d1d we might want to get rid of this as well. Yeah, don't need this anymore either :)
gharchive/pull-request
2021-08-15T13:19:01
2025-04-01T04:33:57.236525
{ "authors": [ "arrufat", "davisking" ], "repo": "davisking/dlib", "url": "https://github.com/davisking/dlib/pull/2414", "license": "BSL-1.0", "license_type": "permissive", "license_source": "github-api" }
197556341
Simple language - Add smart completion We would need the grammar from the catalog https://issues.apache.org/jira/browse/CAMEL-10654 But until then we could hardcode the grammar in the IDEA plugin and then we could offer smart completion for simple language as well. This would be supported in the DSL where you use <simple> or simple("...") etc. You get validation error today, its good enough. We can look at this in the future if there is a proper grammar from camel, which I dont see around the corner
gharchive/issue
2016-12-26T08:22:49
2025-04-01T04:33:57.242455
{ "authors": [ "davsclaus" ], "repo": "davsclaus/camel-idea-plugin", "url": "https://github.com/davsclaus/camel-idea-plugin/issues/17", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2468567990
feat: multi-profile support for git providers Git providers support multiple identities for the same user. Description Please include a summary of the change or the feature being introduced. Include relevant motivation and context. List any dependencies that are required for this change. [ ] This change requires a documentation update [ ] I have made corresponding changes to the documentation This PR addresses issue #777 /claim #777 Screenshots If relevant, please add screenshots. Notes Please add any relevant notes if necessary. Would like to add any other case to this? You are free to suggest the best approach. I have nothing to add at this point. We need a way to figure out how to select a global access token if there are multiple for the same git provider. I suggest you try to find a tool that has a similar feature (handling multiple git provider tokens) and find inspiration there.
gharchive/pull-request
2024-08-15T17:27:25
2025-04-01T04:33:57.273416
{ "authors": [ "Tpuljak", "the-johnwick" ], "repo": "daytonaio/daytona", "url": "https://github.com/daytonaio/daytona/pull/925", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
203641678
"create new PRIVATE gist" creates anonymous gist after showing error message Repro steps: install the extension from VSC, reload VSC, run the command "Gist: create new PRIVATE gist", enter some name to the "Enter the gist description" box and press Enter, an error message "Sorry, git must be installed." is displayed. Press the "Close" button. a browser window is opened with a new anonymous gist. The gist contains the text from the currently opened (in VSC) file. What is wrong here: First of all the gist should not be anonymous (or it should be mentioned in the command name that it will be anonymous). It should not create a gist at all if an error occurred. It should check system requirements (like presence of git) before doing anything. These requirements should be mentioned in the installation instructions. It should be mentioned in the command name that the gist will be created from the text in the current window. Because otherwise it is too easy to blindly create anonymous gist with sensitive and private information that could be accidentally opened in VSC (as it was in my case). It's worth noting that anonymous gists cannot be deleted or edited which makes the overall user experience truly breathtaking. Extension version 0.5.2, VSC version 1.8.1, Windows 10 RS1 x64. I must clarify the last point. This whole mess happened because I expected that the "Gist: create new PRIVATE gist" will create a new document and will allow me to edit and save it. Because that's what the "New File" command does. The real action was more like "create anonymous gist from the current file" or "publish current file as anonymous gist". Private gists should create "Secret" gists from you current file. The way it works off the current file is in the README.md. The fact that on a failure it created an anonymous one Is a bug. I don't have time to look into it at the moment. closing in favor of https://github.com/kenhowardpdx/vscode-gist/issues/3
gharchive/issue
2017-01-27T13:58:07
2025-04-01T04:33:57.296492
{ "authors": [ "dbankier", "kenhowardpdx", "sekogan" ], "repo": "dbankier/vscode-gist", "url": "https://github.com/dbankier/vscode-gist/issues/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
726221401
Error while edit postgresql inet fields System information: Operating system (distribution) and version linux (ununtu 20.04 and debian sid) DBeaver version 7.2.3.202010191702 Connection specification: Database name and version postgresql 9.3+ Driver name Do you use tunnels or proxies (SSH, SOCKS, etc)? No Describe the problem you're observing: Cant edit (add or upgrade( any inet fields Запись экрана от 21.10.2020 09_35_39.zip Include any warning/errors/backtraces from the logs dbeaver-debug.log It should be fixed in scope of https://github.com/dbeaver/dbeaver/issues/9964#issuecomment-713126342 Please try early access version https://dbeaver.io/files/ea/ Can reproduce it in EA version 7.2.3.202010202034 Fixed in 7.2.5 verified
gharchive/issue
2020-10-21T07:33:48
2025-04-01T04:33:57.325024
{ "authors": [ "Kazun3500", "LonwoLonwo", "kseniiaguzeeva", "uslss" ], "repo": "dbeaver/dbeaver", "url": "https://github.com/dbeaver/dbeaver/issues/10122", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
863545204
Autocomplete does not appear if I write SQL using table synonyms with Oracle 12 database System information: Operating system (distribution) and version: Windows 10 Pro 20H2 DBeaver version Additional extensions: Office integration Connection specification: Database name and version Driver name: Oracle (JDBC) Do you use tunnels or proxies (SSH, SOCKS, etc)? NetScaler Access Gateway Describe the problem you're observing: On SQL Editor: If I use the table name directly on "FROM" clause, SQL Autosuggestion appears as expected If I use the table synonym on "FROM" clause, the SQL Autosuggestion does not appear anymore Steps to reproduce, if exist: Create a sample table: CREATE TABLE "BIDTMS"."BID_CALENDAR1" ( "CALENDAR_DATE" NUMBER(14,0), "CALENDAR_KEY" DATE )SEGMENT CREATION IMMEDIATE PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 NOCOMPRESS LOGGING STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645 PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT FLASH_CACHE DEFAULT CELL_FLASH_CACHE DEFAULT) TABLESPACE "BIDTMSDAT" ; Create the Synonym: On SQL editor: SELECT f.[hit ctrl + space => SQL suggestion appears] FROM table_name; SELECT f.[hit ctrl + space => Nothing happens] FROM synonym_name; thanks for the report Verified I tried to build Dbeaver from devel branch, in order to check if it works. Unfortunately, the problem is still there, exactly the same behavior. Hello @MatteoR23 As I can see on your images, you use schema identifier with synonym name. do you use another default schema in SQL Editor? It works for me without schema identifier for default schema: and for another schema, if global search is enabled Effectively, I try to access from another schema. Anyway, even if I check that setting, the problem is not resolved. Here some screenshots from my devel version of DbEaver (I cloned the git repo yesterday after 6 PM CEST tz): @Matvey16 could you please reopen this bug? The problem is not resolved. Verified Hello @MatteoR23 In scope with this ticket https://github.com/dbeaver/dbeaver/issues/11327 we added new connection setting for Oracle. Enabling this setting will help with solving the problem of auto-completion with fully qualified names.
gharchive/issue
2021-04-21T07:39:27
2025-04-01T04:33:57.338368
{ "authors": [ "LonwoLonwo", "MatteoR23", "Matvey16", "uslss" ], "repo": "dbeaver/dbeaver", "url": "https://github.com/dbeaver/dbeaver/issues/12190", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2025830761
db2 discov3er owner entity slow Description Hi I upgraded to a latest version a few weeks ago but am still having real issues specifically when querying / browsing federated objects. I always have to spin up aqua data studio in order to do that. It will just hang on discover owner entity and i have to kill dbeaver after 10 minutes. Any ideas? Version 23.2.4.202311070022 DBeaver Version Community Version 23.2.4.202311070022 Operating System ventura 13.4.1 Database and driver Db2 11 jdbc Steps to reproduce either i go to an object and expand the columns or indexes for example and it just spins for ages. then ill try and go and query syscat columns instread and it will hang on doscover owner entity. it does this for normal tables too Additional context No response What normally happens is it then has a connection timeout and then i can query the table. It seems related to the browsing of objects theres a chance this has improved since turning off show statistics info in the connection view Hello. Probably, it is connected to this ticket: https://github.com/dbeaver/dbeaver/issues/21216 Hello. We changed the query for statistics reading. Could you please check our Early Access version and check your performance issues in this version? It's been a while since no update here. If the issue is still actual please let me know and provide additional information.
gharchive/issue
2023-12-05T10:13:37
2025-04-01T04:33:57.343639
{ "authors": [ "E1izabeth", "LonwoLonwo", "nickhuge666" ], "repo": "dbeaver/dbeaver", "url": "https://github.com/dbeaver/dbeaver/issues/22099", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2668052775
Search Function Fails to Find Object Description The search function in the app is not able to find the table despite it is seen. Attaching the screen recording. In this case, I am trying to find BIGDATA_FUSE__AIRFLOW table. However, it is not able to find it. After locating it in Snowflake's own Web UI search under DATAPIPE_PROD_RAW_DB, I highlight it in the attached screen recording. DBeaver Version Version 24.2.5.202411171748 Operating System Windows 11 Enterprise Database and driver No response Steps to reproduce No response Additional context No response https://github.com/user-attachments/assets/a67b4401-4573-4f3a-be7f-11fd46fb26c3 Thanks for the issue. DBeaver search doesn't work for schemas right now. You can see the object types for search at the right side of the DB Metadata wizard Hi @E1izabeth , I am confused. Here I am not searching for a schema but a table. I just tested the search for other tables under the default schema, it is completely broken. Nothing comes up in the search results. We will try reproducing the issue. Thanks for clarification Right, sorry for the confusion. I am attaching another screen recording. Here, I am searching for FACT_MESSAGES whose full path is: ANALYTICS_PROD_DB.ANALYTICS.FACT_MESSAGES. As you can see in the recording, nothing comes up. https://github.com/user-attachments/assets/1d1c8e74-7304-463e-bfb0-ce86d9d8c7fb Does this help? Yes, now the issue is clear to me. Thanks a lot @ufuk-ergin-carbon, could you tell me how you open the Search dialog? With the Search button on the main toolbar or with the Ctrl+H shortcut? I'm trying to reproduce the issue with no success for now. I'm trying to notice the difference with your video. For me, the tree is always collapsed when the Search dialog is opened. Maybe the trick to reproduce is somewhere there @E1izabeth , I opening the dialog using the search button as I didn't know the keyboard shortcut for this. Now, I tried to replicate the issue using both methods, and the issue is still there. I am attaching a screenshot as to how the dialog opens. Thank you for quick reply Sure, you're welcome. I am attaching videos using both methods. https://github.com/user-attachments/assets/f958aa02-0c02-43d4-805a-3a2626125ee8 https://github.com/user-attachments/assets/e94784c1-9b16-485d-8185-793c68ba1436 Does this help? I figured out how to get an expanded tree in the Search dialog—you need to have only one connection in the project. However, it didn't help me reproduce the bug. Could you provide log files? They are located in the %APPDATA%\DBeaverData\workspace6\.metadata directory. In case your OS isn't Windows, here is our wiki page on this matter: [dbeaver.com/docs/wiki/Log-files/#log-files](http://dbeaver.com/docs/wiki/Log-files/#log-files Also, you can collect them to a zip archive automatically with Help->Collect Diagnostic Info @E1izabeth , I am attaching the log files. dbeaver-diagnostic-info-1732644339389.zip From the logs, I see the following warnings Can't create sql dialect for basic:SQL I'm unsure if it might be a reason, but we'll fix it. Also, from the previous issue #35169, I see that the select query with the table that you can't find using the Search dialog has an error saying that the table is not found. It means that DBeaver didn't get metadata about this table for some reason. Could you try writing a simple select like select * from FACT_MESSAGES and check if you have an error for this query in the SQL Editor? P.S. I'm talking about this indication in SQL Editor The issue that you referenced does not give error for the fact_messages but for SELECT * FROM DATAPIPE_PROD_RAW_DB.MONOLITH__AIRFLOW.SLOTS LIMIT 10; This is another issue that I reported where the query runs perfectly but there is syntax coloring like the one you see in the screenshot. As you see, there are 10 results in the results tab. But, I ran the following two select statements and there is no error for any. The results for the two came out just fine. select * from FACT_MESSAGES; SELECT * FROM ANALYTICS_PROD_DB.ANALYTICS.FACT_MESSAGES
gharchive/issue
2024-11-18T10:43:11
2025-04-01T04:33:57.357349
{ "authors": [ "E1izabeth", "ufuk-ergin-carbon" ], "repo": "dbeaver/dbeaver", "url": "https://github.com/dbeaver/dbeaver/issues/36263", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
440007328
SQL Editor / Set Active Connection doesn't do anything in 6.0.3 Neither does the Ctrl-9 shortcut. Windows 10 Enterprise version 1803. Ctrl+9 shortcut was fixed in #5767. I can't reproduce "doesn't do anything" part. If you choose another schema/connection in main toolbar it changes for the editor. Could you provide a bit more details? Sure. In 5.1.6 if I select the menu option or do Ctrl-9 the window pops up to select a different DB connection. This does not happen in 6.0.3. I see the fix and will await to test 6.0.4. The issue isn't reproducible on 6.0.4. If it is still reproducible for you, feel free to reopen the ticket.
gharchive/issue
2019-05-03T11:38:14
2025-04-01T04:33:57.360214
{ "authors": [ "bobf32", "serge-rider", "uslss" ], "repo": "dbeaver/dbeaver", "url": "https://github.com/dbeaver/dbeaver/issues/5846", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
575130504
Inline column editor for Database table editor Is your feature request related to a problem? Please describe. It is a bit annoying to have to go to "add column", click and bring up a column editor when the temptation is to just double click and add/edit columns. Describe the solution you'd like As stated above, would like to be able to have inline editor be available for adding/editing columns by double-clicking on column "line" thanks for suggestion I'm afraid I don't get it. Where exactly double click can be performed? Some screenshot would be very helpful. In the “table editor” (new or edit), I would like to be able to click on a column line in the “column grid editor” (see “red arrow”) and be able to edit columns directly, rather than having to go to bottom select “add / edit column”, and have to use the pop-up editor, just like the way “data editors” work with result sets. On Mar 5, 2020, at 3:27 AM, Serge Rider notifications@github.com wrote: I'm afraid I don't get it. Where exactly double click can be performed? Some screenshot would be very helpful. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/dbeaver/dbeaver/issues/8077?email_source=notifications&email_token=AIJ4HKF36IRLXFJOQ5AMU3DRF5PAZA5CNFSM4LA257X2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEN4GVAA#issuecomment-595094144, or unsubscribe https://github.com/notifications/unsubscribe-auth/AIJ4HKDEDD3M6KJGV4BKIHLRF5PAZANCNFSM4LA257XQ.
gharchive/issue
2020-03-04T04:33:17
2025-04-01T04:33:57.364607
{ "authors": [ "rotorboy", "serge-rider", "uslss" ], "repo": "dbeaver/dbeaver", "url": "https://github.com/dbeaver/dbeaver/issues/8077", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1211147884
#15755 Improve ERD routing Closes #15755 List of things to be done: [x] - implement Mikami-Tabuchi for orthogonal routing(Still needs path manipulation support) [x] - Change anchor from element to attributes, allowing to show path directly from the attributes. (Partly works not for all attribute types) Current progress This feature is disabled by default due to major changes in the program behavior, it can be enabled in the preferences. Attribute - attribute associations will be shown only if all attributes are shown. This was done to make all connections consistent.
gharchive/pull-request
2022-04-21T14:44:58
2025-04-01T04:33:57.366924
{ "authors": [ "Destrolaric" ], "repo": "dbeaver/dbeaver", "url": "https://github.com/dbeaver/dbeaver/pull/16272", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
740192153
Check streamstats sites against what's in ref_gages. https://test.streamstats.usgs.gov/docs/gagestatsservices/#/Stations/GET/Stations as a source for sites. Done in latest commit
gharchive/issue
2020-11-10T19:43:07
2025-04-01T04:33:57.385763
{ "authors": [ "dblodgett-usgs" ], "repo": "dblodgett-usgs/ref_gages", "url": "https://github.com/dblodgett-usgs/ref_gages/issues/12", "license": "CC0-1.0", "license_type": "permissive", "license_source": "github-api" }
1461715132
Feat/submit mutation Publish package to npmjs This PR exceeds the recommended size of 1200 lines. Please make sure you are NOT addressing multiple issues with one PR. Note this PR might be rejected due to its size. This PR exceeds the recommended size of 1200 lines. Please make sure you are NOT addressing multiple issues with one PR. Note this PR might be rejected due to its size.
gharchive/pull-request
2022-11-23T12:57:43
2025-04-01T04:33:57.394563
{ "authors": [ "db3fans", "xiyangjun" ], "repo": "dbpunk-labs/db3", "url": "https://github.com/dbpunk-labs/db3/pull/180", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2530979188
[Bug] Incorrect CTE name referenced when testing versioned incremental models in incremental mode Is this a new bug? [X] I believe this is a new bug [X] I have searched the existing issues, and I could not find an existing issue for this bug Current Behavior When defining a unit test for the incremental load on a versioned incremental model, the generated SQL is invalid. The name of the CTE "injected" at the top contains the version name, when this is replaced by the CTE name, but without the version name. e.g., with the versioned model my_model we get with __dbt__cte__my_model_v1 as ( ... ) select * from ... where my_date > (select max(my_date) from __dbt__cte__my_model) This is not valid SQL Expected Behavior this is replaced by the CTE name that includes the version number Steps To Reproduce create an incremental model add a version to the model add a unit test to the model with is_incremental: true and some rows for this run the unit test Relevant log output Not relevant Environment - dbt-adapters:1.6.1 Additional Context No response Reprex Create these files: models/my_model.sql {{ config( materialized='incremental' ) }} select 2 as id, 1 as event_time {% if is_incremental() %} where event_time >= (select coalesce(max(event_time), 0) from {{ this }} ) {% endif %} models/_properties.yml models: - name: my_model latest_version: 1 versions: - v: 1 unit_tests: - name: dbt_adapters_309 model: my_model overrides: macros: is_incremental: false given: - input: this rows: - {id: 2, event_time: 1} expect: rows: - {id: 2, event_time: 1} Then run these commands: dbt run --empty --full-refresh dbt build --select my_model ✅ Everything should run just fine. Then change is_incremental: false to be is_incremental: true. Then re-run this command: dbt build --select my_model 💥 It will fail with the following output: 18:48:05 Running with dbt=1.8.6 18:48:06 Registered adapter: duckdb=1.8.3 18:48:06 Found 1 model, 410 macros, 1 unit test 18:48:06 18:48:06 Concurrency: 1 threads (target='dev') 18:48:06 18:48:06 1 of 2 START unit_test my_model::dbt_adapters_309_v1 ........................... [RUN] 18:48:06 1 of 2 ERROR my_model::dbt_adapters_309_v1 ..................................... [ERROR in 0.17s] 18:48:06 2 of 2 SKIP relation feature_456.my_model ...................................... [SKIP] 18:48:06 18:48:06 Finished running 1 unit test, 1 incremental model in 0 hours 0 minutes and 0.39 seconds (0.39s). 18:48:06 18:48:06 Completed with 1 error and 0 warnings: 18:48:06 18:48:06 Runtime Error in unit_test dbt_adapters_309 (models/_properties.yml) An error occurred during execution of unit test 'dbt_adapters_309'. There may be an error in the unit test definition: check the data types. Runtime Error Catalog Error: Table with name __dbt__cte__my_model does not exist! Did you mean "feature_456.my_model_v1"? LINE 29: ...version": "1.8.6", "profile_name": "duckdb", "target_name": "dev", "node_id": "unit_test.my_project.my_model.dbt_adapters_309_v1"} */ create temporary table "dbt_adapters_309_v1__dbt_tmp20240918124806893647" as ( select * from ( with __dbt__cte__my_model_v1 as ( -- Fixture for my_model select cast(2 as INTEGER) as id, cast(1 as INTEGER) as event_time ) select 2 as id, 1 as event_time where event_time >= (select coalesce(max(event_time), 0) from __dbt__cte__my_model ) ^ 18:48:06 18:48:06 Done. PASS=0 WARN=0 ERROR=1 SKIP=1 TOTAL=2 As @b-per mentioned, the difference is __dbt__cte__my_model (actual) vs. __dbt__cte__my_model_v1 (expected) in the portion of the logic that is only executed in incremental mode (like from this example).
gharchive/issue
2024-09-17T12:05:24
2025-04-01T04:33:57.402198
{ "authors": [ "b-per", "dbeatty10" ], "repo": "dbt-labs/dbt-adapters", "url": "https://github.com/dbt-labs/dbt-adapters/issues/309", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2676450026
[Microbatch] Update default make_temp_relation macro to incorporate a batch specific identifier if available Desired Improvement Microbatch batches have been improved in core such that they may be run concurrently. When a batch is executed, the necessary data is first moved into a temp_relation in the data warehouse. The path for this temp_relation is <resource_identifier>__dbt_tmp. As it currently stands, the <resource_identifier>__dbt_tmp will be the same for each batch of a given microbatch model. If the batches are run concurrently, they may end up clobbering the temp_relation destination, which may lead to some wonkiness. As such, we need a way to ensure that each batch for a microbatch model gets a unique temp_relation path. Helpful Prior Art Similar to dbt-postgres: https://github.com/dbt-labs/dbt-postgres/blob/ae48e67dae6c1b00cda37ee9bdc61d3330506638/dbt/include/postgres/macros/adapters.sql#L149-L152 Hello, what a shame that I didn't find this issue before learning how the source code works, and found that my problem could be solved by https://github.com/dbt-labs/dbt-adapters/pull/361. I was currently starting the same PR. At least, I've learned a lot and waiting to this release. However, let me give you a little more context, because this solves my issue, but it's not related to microbatches. I'm currently implementing dbt with Airflow, and we are used to having multiples runs to ingest data in different partitions of BigQuery table. Thus, having a dbt model model_A, materialized as incremental with insert_overwrite strategy, launching multiples runs, let's say D-2 and D-1, for different partitions, will create a conflict with model_A__dbt_tmp table. And the dbt-bigquery don't leverage this situation and relies on the make_temp_relation on dbt-adapters. So, it's the exact same situation as microbatch but on a less granular scope. You'll probably know and understand better than me, but I wanted to share my situation.
gharchive/issue
2024-11-20T16:31:43
2025-04-01T04:33:57.407193
{ "authors": [ "MichelleArk", "vanAkim" ], "repo": "dbt-labs/dbt-adapters", "url": "https://github.com/dbt-labs/dbt-adapters/issues/360", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2373688307
[Bug] Incorrect column data type with incremental contracted model and varchar data_type Current Behavior A temporary table is created for incremental models when the table already exists. When this temporary table is created and one of the column contracted data_types is "character varying(1)" (size is not significant) the string size is lost in the creation of the temporary relation and instead the column is altered after the temporary table is created like: alter table "xxx"."xxxx"."xxx" add column "<column_name>__dbt_alter" character varying(256); The column is created with the wrong string size. Expected Behavior The column is created with the string size that's specified in the contract. Steps to recreate Create an incremental model, update it with additional rows (causing the creation of a temporary table) and observe the the wrong string size is used. Additional Context Reported internally by Lee Bond-Kennedy. I was able to reproduce this for both varchar(1) and character varying(1). In both cases running dbt-postgres 1.6.16 the following occurs on the incremental run. models: - name: my_first_dbt_model config: materialized: incremental on_schema_change: append_new_columns contract: enforced: true description: "A starter dbt model" columns: - name: id description: "The primary key for this table" data_type: int - name: vchar description: "Test varying char contract" data_type: character varying(1) - name: my_other_model config: materialized: incremental on_schema_change: append_new_columns contract: enforced: true columns: - name: id description: "The primary key for this table" data_type: int - name: vchar description: "Test varying char contract" data_type: varchar(1) ^[[0m10:17:46.720110 [debug] [Thread-1 (]: SQL status: SELECT 2 in 0.0 seconds ^[[0m10:17:46.720852 [debug] [Thread-1 (]: Changing col type from character varying(1) to character varying(256) in table database: "lee" schema: "lee" identifier: "my_first_dbt_model" ^[[0m10:17:46.724173 [debug] [Thread-1 (]: Using postgres connection "model.pgtest.my_first_dbt_model" ^[[0m10:17:46.724366 [debug] [Thread-1 (]: On model.pgtest.my_first_dbt_model: /* {"app": "dbt", "dbt_version": "1.6.16", "profile_name": "pgtest", "target_name": "dev", "node_id": "model.pgtest.my_first_dbt_model"} */ alter table "lee"."lee"."my_first_dbt_model" add column "vchar__dbt_alter" character varying(256); update "lee"."lee"."my_first_dbt_model" set "vchar__dbt_alter" = "vchar"; alter table "lee"."lee"."my_first_dbt_model" drop column "vchar" cascade; alter table "lee"."lee"."my_first_dbt_model" rename column "vchar__dbt_alter" to "vchar" ^[[0m10:21:12.614573 [debug] [Thread-1 (]: SQL status: SELECT 2 in 0.0 seconds ^[[0m10:21:12.615230 [debug] [Thread-1 (]: Changing col type from character varying(1) to character varying(256) in table database: "lee" schema: "lee" identifier: "my_other_model" ^[[0m10:21:12.618429 [debug] [Thread-1 (]: Using postgres connection "model.pgtest.my_other_model" ^[[0m10:21:12.618613 [debug] [Thread-1 (]: On model.pgtest.my_other_model: /* {"app": "dbt", "dbt_version": "1.6.16", "profile_name": "pgtest", "target_name": "dev", "node_id": "model.pgtest.my_other_model"} */ alter table "lee"."lee"."my_other_model" add column "vchar__dbt_alter" character varying(256); update "lee"."lee"."my_other_model" set "vchar__dbt_alter" = "vchar"; alter table "lee"."lee"."my_other_model" drop column "vchar" cascade; alter table "lee"."lee"."my_other_model" rename column "vchar__dbt_alter" to "vchar"
gharchive/issue
2024-06-25T21:26:37
2025-04-01T04:33:57.412246
{ "authors": [ "gshank", "lbk-fishtown" ], "repo": "dbt-labs/dbt-core", "url": "https://github.com/dbt-labs/dbt-core/issues/10362", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1565402922
[CT-1988] Model view: when available, display model owner in dbt-docs In the model view (/model/{model_id}) of dbt-docs, display the model owner (by name?) when available. If a model is part of a group, its owner information should be available as part of the group property. What is the owner? Do we want to add a field per owner type(name, email, slack, etc)? Have some kind of ranking? Use name is available, then email, then... and if none are available use the warehouse value. Exposures currently list both the owner name and the owner email and dynamically hide one if it's not available.
gharchive/issue
2023-02-01T04:15:33
2025-04-01T04:33:57.414825
{ "authors": [ "MichelleArk", "emmyoop" ], "repo": "dbt-labs/dbt-core", "url": "https://github.com/dbt-labs/dbt-core/issues/6821", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2089304189
[CT-3568] print / no_print in profiles.yml has no effect in dbt v1.7 Here's the relevant documentation about the print global config: https://docs.getdbt.com/reference/dbt-jinja-functions/print https://docs.getdbt.com/reference/global-configs/print-output I tried specifying print: False within profiles.yml as described in the docs, but it didn't have any effect. So maybe it was fully deprecated sometime between dbt v1.5 and v1.7 but we didn't yet update the docs? Or maybe it was accidentally stopped working during the click migration (in https://github.com/dbt-labs/dbt-core/pull/7086)? Originally posted by @dbeatty10 in https://github.com/dbt-labs/docs.getdbt.com/pull/4740#discussion_r1454071332 To do Try out the following within each of v1.5, v1.6, and v1.7. Figure out if it was fully deprecated or not and update the docs accordingly. Create this simple model file: models/my_model.sql {{ print("") }} {{ print("*** Parsing, compiling, or running: " ~ this) }} {{ print("") }} select 1 as id Add the following config to profiles.yml: config: print: False See if *** Parsing ... shows up in the logs or not: dbt parse See also: https://github.com/dbt-labs/dbt-core/issues/7036 (which also mentions print and would presumably resolve this issue).
gharchive/issue
2024-01-16T21:46:11
2025-04-01T04:33:57.420755
{ "authors": [ "dbeatty10" ], "repo": "dbt-labs/dbt-core", "url": "https://github.com/dbt-labs/dbt-core/issues/9407", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
675192369
incremental models on postgres are only fast if all CTEs are ephemeral Describe the bug When making a postgres-backed model incremental, you would expect the incremental loads of raw data to be faster. Instead, they get slower. However, once you set all the contribution CTEs (separate DBT files) to ephemeral, and make sure the CTEs are consumed in a row with one descendant only (i.e.: with a from b, b from c, c from d, d from source -- so that it can be inlined), you do get the performance benefits. That means: Views are performance barriers that prevent the optimizer from pushing down the {{is_incremental}} stuff to the underlying queries from which the views are selected. Maybe you could mention that in the documentation? Steps To Reproduce In as much detail as possible, please provide steps to reproduce the issue. Sample data that triggers the issue, example model code, etc is all very helpful here. Expected behavior A clear and concise description of what you expected to happen. Screenshots and log output If applicable, add screenshots or log output to help explain your problem. System information Which database are you using dbt with? [X] postgres [ ] redshift [ ] bigquery [ ] snowflake [ ] other (specify: ____________) The output of dbt --version: installed version: 0.17.1 latest version: 0.17.2 Your version of dbt is out of date! You can find instructions for upgrading here: https://docs.getdbt.com/docs/installation Plugins: - postgres: 0.17.1 - redshift: 0.17.1 - snowflake: 0.17.1 - bigquery: 0.17.1 The operating system you're using: $ uname -a Linux 21d117632cc9 5.3.0-55-generic fishtown-analytics/dbt#49-Ubuntu SMP Thu May 21 12:47:19 UTC 2020 x86_64 GNU/Linux i.e. the official docker container for python:latest The output of python --version: Python 3.8.4 Additional context Add any other context about the problem here. I agree, that's where I was expecting this to be documented. As an aside, what's also missing from that section is how to handle views. Do I put the is_incremental stuff in there, too? At least, that might be a way around the optimization fence problem. But it just doesn't feel right, from a functional programming point of view. What kind of missing additional context to you have in mind? The tricky thing about is_incremental() is that it checks four conditions, one of which is that the current model is materialized as incremental. For that reason, if you include is_incremental() in a view or ephemeral model, the condition will never resolve to true. I think the larger context here is: Incremental modeling is an attempt to optimize database performance. As such it depends significantly on the database you're using, and it relies heavily on your understanding of what that database's optimizer will and won't be able to do. Does this boil down to "if Postgres and incremental, then all models should be tables"? No, a view does not work. That is the point of this bug report. Postgres 12.4 from https://hub.docker.com/_/postgres My mistake in writing that above, I think this is what I meant: "If you're creating an incremental model on postgres, it should be selecting from (and filtering against) a table, not a view or an ephemeral model." Closing this issue as won't fix. If you believe I have closed this in error, then let me know and I willput it on the icebox instead of closing completely. @runleonarun I ran into this issue a few weeks ago and had no guidance from the dbt docs on why using incremental models wasn't improving my model runs in postgres at all. The docs should be edited with this information, it feels critical to new dbt users who are working in postgres
gharchive/issue
2020-08-07T15:08:52
2025-04-01T04:33:57.430803
{ "authors": [ "TjrGithub", "jtcohen6", "kyletl", "runleonarun" ], "repo": "dbt-labs/docs.getdbt.com", "url": "https://github.com/dbt-labs/docs.getdbt.com/issues/335", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2263589850
dbt quoting does not work as expected The quoting for msft sql not working. Dbt use double quotes when the config quoting is turned on but ms sql or tsql use square brackets. I need quoting for databases and table. Our table or database prefix is sometimes a number like 101_Customers . Example: dbt config: quoting: database: true schema: true identifier: true The dbt compilation result: SELECT * FROM "101_supplier"."core"."supplier" What I expected: SELECT * FROM [101_supplier].[core].[supplier] Question: Is there a way to configure the quoting for database, schema and identifier with square brackets. Or this issue a topic for the dbt core team? Additional context I found a core issue with similar problems: link The double quotes work fine in MS SQL. Did you try running the query? It works fine with tables that start with numbers. I don't have servers named like that to try. I think this issue can be closed as 'by design'. DBT handles database and table names beginning with numbers or contains spaces (ack!). Sql Server databases have Quoted Identifiers set to False by default but most drivers, including the Sql Server ODBC driver, set Quoted Identifiers are on by default, which is the ANSI standard. If OP is having issues with quoted identifiers it might be the driver he is using. Here is a good article describing Sql Server's quoted identifiers settings: https://www.sqlshack.com/set-quoted_identifier-settings-in-sql-server/ Issue 2986 referenced by OP is simply cleaning up terminology that probably evolved during the organic development of the project. This assumes the DBT configuration has quoting set to true (default). If it is set to false then tables beginning with numbers or contains spaces will fail. quoting: database: true schema: true identifier: true I tested this using the following database and tables: create database "123 testdb"; use "123 testdb"; create table "123 test" (pk int primary key, one int null, two varchar(100) null); insert "123 test" (pk, one, two) values (1, 1, 'one') drop table "123 test"; drop database "123 testdb"; And the following DBT model: {{ config(materialized='table') }} with source_data as ( select * from {{ source('123 testdb', '123 test') }} ) select * from source_data
gharchive/issue
2024-04-25T13:27:23
2025-04-01T04:33:57.437441
{ "authors": [ "FlorianVc", "bsarge88", "hernanparra" ], "repo": "dbt-msft/dbt-sqlserver", "url": "https://github.com/dbt-msft/dbt-sqlserver/issues/497", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
785464211
Cannot Authenticate with CLI I am trying to use the SQL Server adapter with the CLI authentication. My config is the following : dbt version: 0.18.1 dbt-mssql version: 0.18.1 python version: 3.7.4 os info: Windows-10-10.0.19041-SP0 Here is my profiles.yml my_profile: target: dev outputs: dev: type: sqlserver driver: 'ODBC Driver 17 for SQL Server' server: ip_of_my_server port: port_of_my_server database: db_of_my_server_that_is_not_master schema: my_schema authentication: CLI I run az login sucessfully You have logged in. Now let us find all the subscriptions to which you have access... But when I run dbt debug I encounter the following error : Connection: database: my_db schema: my_schema port: 1433 UID: None client_id: None authentication: CLI encrypt: False trust_cert: False Connection test: ERROR dbt was unable to connect to the specified database. The database returned the following error: >Database Error ('28000', "[28000] [Microsoft][ODBC Driver 17 for SQL Server][SQL Server]Login failed for user '<token-identified principal>'. (18456) (SQLDriverConnect); [28000] [Microsoft][ODBC Driver 17 for SQL Server][SQL Server]Login failed for user '<token-identified principal>'. (18456)") If I use : authentication: ActiveDirectoryInteractive user: me@github.com It works but I have to authenticate at every command executed which is not efficient. Do you have any idea of what I could be doing wrong ? Or how we can fix this ? @EatZeBaby thanksX1000 for the time to write up this issue. I have none of the normally-required follow up questions. @mikaelene and @NandanHegde15 are the SQL Server Server experts here, but I'll take a stab at this.... What version of SQL Server do you have? Where is it hosted? How is it connected to Azure? I'm am by no means an expert, but my understanding is that you can only use the Azure CLI if your SQL Server db is associated with the Azure tenant. The two ways that I think this can be done is either by: Using SQL Server Managed Instance -- a pseudo-platform-as-a-service Azure offering; or Using Active Directory Federation Services (ADFS) to link your Azure Active Directory to an on-premise ("classic") Active Directory @EatZeBaby thanksX1000 for the time to write up this issue. I have none of the normally-required follow up questions. @mikaelene and @NandanHegde15 are the SQL Server Server experts here, but I'll take a stab at this.... What version of SQL Server do you have? Where is it hosted? How is it connected to Azure? I'm am by no means an expert, but my understanding is that you can only use the Azure CLI if your SQL Server db is associated with the Azure tenant. The two ways that I think this can be done is either by: Using SQL Server Managed Instance -- a pseudo-platform-as-a-service Azure offering; or Using Active Directory Federation Services (ADFS) to link your Azure Active Directory to an on-premise ("classic") Active Directory I am using the following version (output of SELECT @@VERSION): Microsoft SQL Azure (RTM) - 12.0.2000.8 Oct 1 2020 18:48:35 Copyright (C) 2019 Microsoft Corporation The SQL Server is associated to the Azure tenant I am logging into. It is hosted on Azure and I have been able to connect with Azure Data Studio for example but with MFA only : From my understanding of our AD, we do not have ADFS. We have MFA turned on so anytime I have to log to Azure (including dbt debug with interactive mode), I have to use my phone to confirm the authentication. Among other properties I can think of, the SQL database configuration is : General Purpose: Serverless, Gen5, 1 vCore (means it's managed right ?) I am using the following version (output of SELECT @@VERSION): Microsoft SQL Azure (RTM) - 12.0.2000.8 Oct 1 2020 18:48:35 Copyright (C) 2019 Microsoft Corporation The SQL Server is associated to the Azure tenant I am logging into. It is hosted on Azure and I have been able to connect with Azure Data Studio for example but with MFA only : From my understanding of our AD, we do not have ADFS. We have MFA turned on so anytime I have to log to Azure (including dbt debug with interactive mode), I have to use my phone to confirm the authentication. Among other properties I can think of, the SQL database configuration is : General Purpose: Serverless, Gen5, 1 vCore (means it's managed right ?) Can you tell me what the integer SELECT SERVERPROPERTY('EngineEdition') returns? Can you tell me what the integer SELECT SERVERPROPERTY('EngineEdition') returns? Sure. It returns 5 Sure. It returns 5 ah ok. so you're on an Azure SQL. For some reason I thought you were on an on-prem SQL Server instance... let me research a little more ah ok. so you're on an Azure SQL. For some reason I thought you were on an on-prem SQL Server instance... let me research a little more I actually just made it work by running az logout and az login. I tried to reproduce the issue but couldn't. Sorry for the false alert. Hope it might help someone one day ! I actually just made it work by running az logout and az login. I tried to reproduce the issue but couldn't. Sorry for the false alert. Hope it might help someone one day !
gharchive/issue
2021-01-13T21:18:44
2025-04-01T04:33:57.450038
{ "authors": [ "EatZeBaby", "swanderz" ], "repo": "dbt-msft/dbt-sqlserver", "url": "https://github.com/dbt-msft/dbt-sqlserver/issues/82", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1083477867
running pip install dbt-synapse==0.18.1 gives success but the plugin isn't really installed Dbt debug doesn't find any adapter type synapse. @BeatrizSTavares thanks for reporting! have you gotten dbt-synapse to work before? does this only happen with 0.18.1? what version of Python are you using? @swanderz Yes I have gotten to work before, with version 0.21.0. The python version I'm using is 3.8.10. I saw here https://github.com/dbt-msft/dbt-synapse that it only has the support needed for version 0.18.0, so I needed to use this version Try version 0.21! The readme is out of date @swanderz I tried with 0.21 dbt version (synapse as well) and I can't reference one model from another, for instance referencing an ephemeral from a table model. The compiled code gives me an error about nested with's (CTE's). Ah! Ephemeral isn't supported w dbt-synapse unfortunately. Please vote on this idea so it may be supported! Ok, thank you!
gharchive/issue
2021-12-17T17:05:15
2025-04-01T04:33:57.454724
{ "authors": [ "BeatrizSTavares", "swanderz" ], "repo": "dbt-msft/dbt-synapse", "url": "https://github.com/dbt-msft/dbt-synapse/issues/73", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1776600956
Create CI checks for refactored CML the checks should verify the format, clippy warnings and run tests Regarding warnings, what should we do about the old code we brought over from IOHK's legacy chain-libs? We have it in crypto's chain_core / chain_crypto / impl_lockchain / typed_bytes. There are a bunch of unused warnings since we don't pub export the modules as they're mostly dependencies we use for our crypto code to work. Should we pub export it? Pros: Maybe useful to someone? Easiest way to get rid of those errors. Cons: Could confuse users with all those types. Or delete/rename to _* all the unused functions / types? Or put a clippy ignore in those mods? (also would be super easy and wouldn't add to our public API). I"m leaning to this at least for now as it's easy to undo if we change our minds and doesn't change the API at all and we generally never touch the code in those modules (more like included deps) anyway. Fixed by #250
gharchive/issue
2023-06-27T10:31:29
2025-04-01T04:33:57.469670
{ "authors": [ "gostkin", "rooooooooob" ], "repo": "dcSpark/cardano-multiplatform-lib", "url": "https://github.com/dcSpark/cardano-multiplatform-lib/issues/230", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2146874399
Support all.snippets Problem Statement Storing snippets for all filetypes is currently not easily possible? Ideas or possible solutions Support loading snippets/all.snippets with default priority for all filetypes Alternatives you have considered No response Isn't this the similar solution that is already implemented? Indeed - sorry for missing that.
gharchive/issue
2024-02-21T14:20:08
2025-04-01T04:33:57.471518
{ "authors": [ "Yozhylo", "muellerj" ], "repo": "dcampos/nvim-snippy", "url": "https://github.com/dcampos/nvim-snippy/issues/126", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2064396627
Original name doesn't show Hello, When I use disablePreview() the original file name appears but if I don't use it, the stored file name is used. I modified the LoadController like this and it seems to work. return Storage::disk($data->disk)->response($data->path, $data->filename); Thanks released it as v1.0.5
gharchive/issue
2024-01-03T17:29:27
2025-04-01T04:33:57.473556
{ "authors": [ "Keko-94", "milewski" ], "repo": "dcasia/nova-filepond", "url": "https://github.com/dcasia/nova-filepond/issues/43", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
172130615
Warnings on using Gadfly statement In this morning's interactive session I came across the following warnings, and was unable to use Gadfly. I've included a text file of the Julia session, as well as a cut and past of the output. Any assistance would be appreciated. Gadfly Warnings.txt julia> workspace() julia> Pkg.update() INFO: Updating METADATA... INFO: Updating Homebrew... INFO: Computing changes... INFO: No packages to install, update or remove julia> using Gadfly WARNING: Method definition readavailable(Main.Base.IOStream) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:927 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:936. WARNING: Method definition readavailable(Main.Base.AbstractIOBuffer{Array{UInt8, 1}}) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:928 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:937. WARNING: Method definition write(Main.Base.IO, Main.Base.IO) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:931 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:940. WARNING: Method definition write(AbstractString, Any...) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:920 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:929. WARNING: Method definition readlines(AbstractString) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:926 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:935. WARNING: Method definition readline(AbstractString) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:925 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:934. WARNING: Method definition cov(AbstractArray{T<:Any, 1}, AbstractArray{T<:Any, 1}, Bool) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:905 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:914. WARNING: Method definition cov(AbstractArray{T<:Any, 2}, AbstractArray{T<:Any, 2}, Integer) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:906 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:915. WARNING: Method definition cov(AbstractArray{T<:Any, 2}, AbstractArray{T<:Any, 2}, Integer, Bool) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:907 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:916. WARNING: Method definition cov(AbstractArray{T<:Any, 1}, Bool) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:902 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:911. WARNING: Method definition cov(AbstractArray{T<:Any, 2}, Integer) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:903 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:912. WARNING: Method definition cov(AbstractArray{T<:Any, 2}, Integer, Bool) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:904 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:913. WARNING: Method definition remotecall_wait(Function, Main.Base.LocalProcess, Any...) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:844 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:853. WARNING: Method definition remotecall_wait(Function, Main.Base.Worker, Any...) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:845 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:854. WARNING: Method definition remotecall_wait(Function, Integer, Any...) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:846 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:855. WARNING: Method definition precision(Type{Main.Base.MPFR.BigFloat}) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:968 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:977. WARNING: Method definition remotecall(Function, Main.Base.LocalProcess, Any...) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:844 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:853. WARNING: Method definition remotecall(Function, Main.Base.Worker, Any...) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:845 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:854. WARNING: Method definition remotecall(Function, Integer, Any...) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:846 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:855. WARNING: Method definition readuntil(AbstractString, Any...) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:924 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:933. WARNING: Method definition eachline(AbstractString) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:937 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:946. WARNING: Method definition cor(AbstractArray{T<:Any, 2}, AbstractArray{T<:Any, 2}, Integer) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:910 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:919. WARNING: Method definition cor(AbstractArray{T<:Any, 2}, Integer) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:909 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:918. WARNING: Method definition read(Main.Base.IO) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:917 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:926. WARNING: Method definition read(Main.Base.IO, Any) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:918 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:927. WARNING: Method definition read(AbstractString, Any...) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:922 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:931. WARNING: Method definition call(Type{Symbol}, Any...) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:400 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:408. WARNING: Method definition call(Type{Union{UTF8String, ASCIIString}}, Main.Base.AbstractIOBuffer{T<:AbstractArray{UInt8, 1}}) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:1278 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:1354. WARNING: Method definition call(Type{Union{UTF8String, ASCIIString}}, Main.Base.Cstring) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:1283 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:1359. WARNING: Method definition call(Type{Union{UTF8String, ASCIIString}}, Array{UInt8, 1}) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:1284 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:1360. WARNING: Method definition call(Type{Union{UTF8String, ASCIIString}}, Union{Ptr{UInt8}, Ptr{Int8}}) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:1285 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:1361. WARNING: Method definition call(Type{Union{UTF8String, ASCIIString}}, Union{Ptr{UInt8}, Ptr{Int8}}, Integer) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:1286 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:1362. WARNING: Method definition call(Type{Union{UTF8String, ASCIIString}}, AbstractString) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:1287 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:1363. WARNING: Method definition remotecall_fetch(Function, Main.Base.LocalProcess, Any...) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:844 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:853. WARNING: Method definition remotecall_fetch(Function, Main.Base.Worker, Any...) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:845 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:854. WARNING: Method definition remotecall_fetch(Function, Integer, Any...) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:846 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:855. WARNING: Method definition read!(AbstractString, Any) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:923 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:932. WARNING: Method definition remote_do(Function, Main.Base.LocalProcess, Any...) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:844 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:853. WARNING: Method definition remote_do(Function, Main.Base.Worker, Any...) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:845 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:854. WARNING: Method definition remote_do(Function, Integer, Any...) in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:846 overwritten in module Compat at /Users/Aaron/.julia/Compat/src/Compat.jl:855. INFO: Recompiling stale cache file /Users/Aaron/.julia/lib/v0.4/Gadfly.ji for module Gadfly. INFO: Recompiling stale cache file /Users/Aaron/.julia/lib/v0.4/ColorTypes.ji for module ColorTypes. INFO: Recompiling stale cache file /Users/Aaron/.julia/lib/v0.4/Colors.ji for module Colors. INFO: Recompiling stale cache file /Users/Aaron/.julia/lib/v0.4/Graphics.ji for module Graphics. INFO: Recompiling stale cache file /Users/Aaron/.julia/lib/v0.4/Compose.ji for module Compose. INFO: Recompiling stale cache file /Users/Aaron/.julia/lib/v0.4/Cairo.ji for module Cairo. WARNING: New definition +(AbstractArray{T<:Any, 2}, WoodburyMatrices.SymWoodbury) at /Users/Aaron/.julia/WoodburyMatrices/src/SymWoodburyMatrices.jl:106 is ambiguous with: +(DataArrays.DataArray, AbstractArray) at /Users/Aaron/.julia/DataArrays/src/operators.jl:276. To fix, define +(DataArrays.DataArray{T<:Any, 2}, WoodburyMatrices.SymWoodbury) before the new definition. WARNING: New definition +(AbstractArray{T<:Any, 2}, WoodburyMatrices.SymWoodbury) at /Users/Aaron/.julia/WoodburyMatrices/src/SymWoodburyMatrices.jl:106 is ambiguous with: +(DataArrays.AbstractDataArray, AbstractArray) at /Users/Aaron/.julia/DataArrays/src/operators.jl:300. To fix, define +(DataArrays.AbstractDataArray{T<:Any, 2}, WoodburyMatrices.SymWoodbury) before the new definition. WARNING: Method definition ctranspose(Function) in module Calculus at /Users/Aaron/.julia/Calculus/src/derivative.jl:26 overwritten in module Calculus at /Users/Aaron/.julia/Calculus/src/derivative.jl:26. WARNING: Method definition gradient(Function) in module Calculus at /Users/Aaron/.julia/Calculus/src/derivative.jl:22 overwritten in module Calculus at /Users/Aaron/.julia/Calculus/src/derivative.jl:22. WARNING: Method definition gradient(Function, Union{Array{#T<:Number, 1}, #T<:Number}) in module Calculus at /Users/Aaron/.julia/Calculus/src/derivative.jl:17 overwritten in module Calculus at /Users/Aaron/.julia/Calculus/src/derivative.jl:17. WARNING: Method definition gradient(Function, Union{Array{#T<:Number, 1}, #T<:Number}, Symbol) in module Calculus at /Users/Aaron/.julia/Calculus/src/derivative.jl:17 overwritten in module Calculus at /Users/Aaron/.julia/Calculus/src/derivative.jl:17. WARNING: Method definition gradient(Function, Symbol) in module Calculus at /Users/Aaron/.julia/Calculus/src/derivative.jl:22 overwritten in module Calculus at /Users/Aaron/.julia/Calculus/src/derivative.jl:22. WARNING: New definition +(AbstractArray{T<:Any, 2}, WoodburyMatrices.SymWoodbury) at /Users/Aaron/.julia/WoodburyMatrices/src/SymWoodburyMatrices.jl:106 is ambiguous with: +(DataArrays.DataArray, AbstractArray) at /Users/Aaron/.julia/DataArrays/src/operators.jl:276. To fix, define +(DataArrays.DataArray{T<:Any, 2}, WoodburyMatrices.SymWoodbury) before the new definition. WARNING: New definition +(AbstractArray{T<:Any, 2}, WoodburyMatrices.SymWoodbury) at /Users/Aaron/.julia/WoodburyMatrices/src/SymWoodburyMatrices.jl:106 is ambiguous with: +(DataArrays.AbstractDataArray, AbstractArray) at /Users/Aaron/.julia/DataArrays/src/operators.jl:300. To fix, define +(DataArrays.AbstractDataArray{T<:Any, 2}, WoodburyMatrices.SymWoodbury) before the new definition. julia> These warnings are due to Julia 0.4 throwing an error when an imported method is ambiguous with another method. This is fixed in the next version of Julia 0.5: https://github.com/JuliaLang/julia/pull/16125 It is unlikely that this will be backported, see https://github.com/timholy/WoodburyMatrices.jl/issues/13.
gharchive/issue
2016-08-19T13:25:14
2025-04-01T04:33:57.517144
{ "authors": [ "jacksaarden", "tlnagy" ], "repo": "dcjones/Gadfly.jl", "url": "https://github.com/dcjones/Gadfly.jl/issues/872", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
337214313
dcos-docker muti cluster Docker Env, Not in same machine, I want to create most masters and agents, I dont not how to do. In document, I know how to create in same machine, but Not in same machine? Yeah,@adamtheturtle I‘m using the DC/OS E2E CLI. I‘m using the dcos-docker create command. I want to use it for multiple machine clusters, not stand-alone. @xyjwsj Thank you for clarifying. That is not currently supported. What is your use case? Knowing this might help us know whether to prioritise this. In addition to what @adamtheturtle asked I would like to clarify that if your goal is to run DC/OS on multiple machines then canonical installation of DC/OS is the right way to go, and dcos-docker is not. dcos-docker (DC/OS E2E with the Docker backend) can be thought of as a convenient way to "simulate running a real DC/OS cluster on a single machine". Could support this mode? I need run DC/OS on multiple machines, how to do? There are multiple ways to install DC/OS on multiple nodes. See https://docs.mesosphere.com/1.11/installing/oss/. I do not think that DC/OS E2E (and DC/OS in Docker) is the way to go - it is not supported for this use case. I would follow the instructions in the above link. However - is there a particular reason you want to use DC/OS E2E? @xyjwsj I will close this - please re-open or open a new issue if you are still having trouble.
gharchive/issue
2018-06-30T13:18:00
2025-04-01T04:33:57.526068
{ "authors": [ "adamtheturtle", "jgehrcke", "xyjwsj" ], "repo": "dcos/dcos-e2e", "url": "https://github.com/dcos/dcos-e2e/issues/1045", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
458186037
fix: permission builder exception Update reactjs-components to resolve issue with permissions builder Closes DCOS-48704 Testing Create a user account Open the permissions builder for the new user Select Mesos -> Master -> Quota Select both read & Update Change Quota -> Executor Verify modal doesn't crash Trade-offs Dependencies Screenshots :tada: This PR is included in version 2.109.8 :tada: The release is available on GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2019-06-19T19:23:46
2025-04-01T04:33:57.530233
{ "authors": [ "TattdCodeMonkey", "mesosphere-ci" ], "repo": "dcos/dcos-ui", "url": "https://github.com/dcos/dcos-ui/pull/3939", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
166576735
DCOS-8722: Remove the id in edit mode from the jobs form add job name to title This will remove the id field in edit mode of the job form modal and add the job name to the title. Reviewer: @philipnrmn Kenny's comment aside, LGTM LGTM 💯 ✨
gharchive/pull-request
2016-07-20T13:07:40
2025-04-01T04:33:57.531853
{ "authors": [ "Poltergeist", "mesosphere-web", "orlandohohmeier", "philipnrmn" ], "repo": "dcos/dcos-ui", "url": "https://github.com/dcos/dcos-ui/pull/778", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
316579623
[DCOS-14199][1.11.1] Write strings to files atomically in pkgpanda.util High-level description This is a backport of https://github.com/dcos/dcos/pull/2730 which has a Ship It label This changes pkgpanda.util.write_string to avoid a race condition as described in DCOS-14199. I have also applied this change to some related functions as a precaution. I have stopped returning data from the functions as it appears that those return values are never used. This also adds a test for write_string to enshrine the permission requirements and basic functionality of write_string (thanks @mhrabovcin for help). That test passes on master and with the changes here. Corresponding DC/OS tickets (obligatory) These DC/OS JIRA ticket(s) must be updated (ideally closed) in the moment this PR lands: DCOS-14199 exhibitor.py try_shortcut() failed due to an empty file Checklist for all PRs [ ] Added a comprehensible changelog entry to CHANGES.md or explain why this is not a user-facing change: See https://github.com/dcos/dcos/pull/2730#issuecomment-380506106. [ ] Included a test which will fail if code is reverted but test is not. If there is no test please explain here: This code is already called by the test suite. We could test the race condition with something like the following with a timeout but I'd rather not. I'm happy to discuss this with a reviewer: # This will trigger the error from the issue if run by two processes in parallel. # Calling the new function in the loop will not. filename = 'foo.txt' while True: with open(filename, 'w+') as f: f.write('hello') with open(filename) as f: stripped = f.read().strip() assert stripped != '' [X] Read the DC/OS contributing guidelines [X] Followed relevant code rules Rules for Packages and Systemd @mesosphere-mergebot bump-ee @mesosphere-mergebot bump-ee @mesosphere-mergebot label Ready For Review I think that @orsenthil should review this as I'm not 100% sure about what goes on with branches against 1.11.1. How come this has a Ship It?
gharchive/pull-request
2018-04-22T13:34:41
2025-04-01T04:33:57.539358
{ "authors": [ "adamtheturtle" ], "repo": "dcos/dcos", "url": "https://github.com/dcos/dcos/pull/2789", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
349115856
[1.11] changelog: Add Java update note Adds missing release note for https://github.com/dcos/dcos/pull/3225. High-level description Update java to version 1.8.0_181 which includes security updates. Corresponding DC/OS tickets (obligatory) These DC/OS JIRA ticket(s) must be updated (ideally closed) in the moment this PR lands: DCOS_OSS-3932 [1.11] packages/java: Update java to version 1.8.0_181 Checklist for all PRs [ ] Added a comprehensible changelog entry to CHANGES.md or explain why this is not a user-facing change: [ ] Included a test which will fail if code is reverted but test is not. If there is no test please explain here: [ ] Read the DC/OS contributing guidelines [ ] Followed relevant code rules Rules for Packages and Systemd Checklist for component/package updates: If you are changing components or packages in DC/OS (e.g. you are bumping the sha or ref of anything underneath packages), then in addition to the above please also include: [ ] Change log from the last version integrated (this should be a link to commits for easy verification and review): example [ ] Test Results: [link to CI job test results for component] [ ] Code Coverage (if available): [link to code coverage report] @mesosphere-mergebot label Ready For Review @mesosphere-mergebot bump-ee This will conflict with two other PRs on the next train :) @orsenthil fortunately provided the outlook that he'll take care of resolving the merge conflict(s). Thanks!! Please run: @mesosphere-mergebot override-status teamcity/dcos/test/aws/cloudformation/simple https://jira.mesosphere.com/browse/DCOS_OSS-2115 mergebot/enterprise/build-status/aggregate check is missing even if EE PR has shipit label. Please run: @mesosphere-mergebot override-status mergebot/enterprise/build-status/aggregate https://jira.mesosphere.com/browse/DCOS-19905 @mesosphere-mergebot override-status mergebot/enterprise/build-status/aggregate https://jira.mesosphere.com/browse/DCOS-19905 Please run: @mesosphere-mergebot override-status mergebot/enterprise/build-status/aggregate https://jira.mesosphere.com Hm by the time I came here the status check was already green. @mesosphere-mergebot override-status teamcity/dcos/test/aws/cloudformation/simple https://jira.mesosphere.com/browse/DCOS_OSS-2115 @mesosphere-mergebot bump-ee
gharchive/pull-request
2018-08-09T12:56:54
2025-04-01T04:33:57.548939
{ "authors": [ "jgehrcke", "mhrabovcin" ], "repo": "dcos/dcos", "url": "https://github.com/dcos/dcos/pull/3248", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
624263652
add cache-header for dcos-version.json as during a SOAK update the UI did not consistently report the correct version of the cluster, we now let the UI know, that it must always revalidate dcos-version.json. Should this use no-cache to force re-validation? It seems must-revalidate would make sense if we were setting a non-zero cache time. no-cache The response may be stored by any cache, even if the response is normally non-cacheable. However, the stored response MUST always go through validation with the origin server first before using it. must-revalidate Indicates that once a resource becomes stale, caches must not use their stale copy without successful validation on the origin server. @jongiddy it is my understanding that if there's no other headers indicating how long it takes a file to become stale, it is immediately stale when must-revalidate is set. setting an additional cache-time or ETags (as in our case) enables browsers to actually make use of caching. please correct me if i'm wrong here. The only info I'm going on is the PR description - is there a JIRA as well? But it sounds like every request needs to go back to the server, whether the ETag matches or not. no-cache forces every request to go back to the server, no matter what expiration time is set or calculated, or whether the cached value is considered to be fresh or stale. With must-revalidate and no other cache headers, the client can use a heuristic to create an expiration time and, during that time, the cached value is fresh so the client does not contact the server. See https://tools.ietf.org/html/rfc7234#section-4.2.2 In either case, the ETag can allow the server to avoid returning the actual response (but in this case, the data is small and cheap enough that it saves very little). that makes a lot of sense. thanks for educating! 💜 @mesosphere-mergebot label Ready For Review @mesosphere-mergebot changelog-not-required insignificant external visibility @mesosphere-mergebot backport 2.1 @mesosphere-mergebot backport 2.0 @mesosphere-mergebot backport 1.13
gharchive/pull-request
2020-05-25T12:16:42
2025-04-01T04:33:57.555663
{ "authors": [ "jongiddy", "pierrebeitz" ], "repo": "dcos/dcos", "url": "https://github.com/dcos/dcos/pull/7311", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2520410174
Strategy to embed images There's a lot of visuals in rulebooks that'd be great to reference. Ignoring whether we can textual represent them, the marker parser we're using will actually transform text and extract images at the same time. I believe we could design a system that lets us easily reference those in outputs (and embed them). Yes, that's exactly copali for. Copali is used to embed the images. and direct semantically search the images using text. Also, the cool part is that you can then feed the retrieved image and text into a multi-modal LLMs like gpt-4o, gemini etc. they will answer you questions using both.
gharchive/issue
2024-09-11T18:05:50
2025-04-01T04:33:57.557270
{ "authors": [ "dcramer", "tian-yi" ], "repo": "dcramer/gamegame", "url": "https://github.com/dcramer/gamegame/issues/6", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1467900544
🛑 Danbs is down In b761d9e, Danbs (https://danbs.net) was down: HTTP code: 404 Response time: 155 ms Resolved: Danbs is back up in d6da346.
gharchive/issue
2022-11-29T11:29:02
2025-04-01T04:33:57.578224
{ "authors": [ "ddanbs" ], "repo": "ddanbs/upptime", "url": "https://github.com/ddanbs/upptime/issues/1195", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1964821150
🛑 Danbs is down In 709fb7c, Danbs (https://danbs.net) was down: HTTP code: 404 Response time: 148 ms Resolved: Danbs is back up in f5ba374 after 18 minutes.
gharchive/issue
2023-10-27T05:58:23
2025-04-01T04:33:57.580493
{ "authors": [ "ddanbs" ], "repo": "ddanbs/upptime", "url": "https://github.com/ddanbs/upptime/issues/12821", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2044494046
🛑 Danbs is down In 9aefa31, Danbs (https://danbs.net) was down: HTTP code: 404 Response time: 79 ms Resolved: Danbs is back up in a710264 after 6 minutes.
gharchive/issue
2023-12-15T23:35:25
2025-04-01T04:33:57.582901
{ "authors": [ "ddanbs" ], "repo": "ddanbs/upptime", "url": "https://github.com/ddanbs/upptime/issues/14662", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2196949144
🛑 Danbs is down In e3b8148, Danbs (https://danbs.net) was down: HTTP code: 404 Response time: 149 ms Resolved: Danbs is back up in 79b9cb3 after 13 minutes.
gharchive/issue
2024-03-20T08:33:40
2025-04-01T04:33:57.585119
{ "authors": [ "ddanbs" ], "repo": "ddanbs/upptime", "url": "https://github.com/ddanbs/upptime/issues/18264", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1513488944
🛑 Danbs is down In 2c73ff9, Danbs (https://danbs.net) was down: HTTP code: 404 Response time: 166 ms Resolved: Danbs is back up in 1991b95.
gharchive/issue
2022-12-29T08:07:14
2025-04-01T04:33:57.587319
{ "authors": [ "ddanbs" ], "repo": "ddanbs/upptime", "url": "https://github.com/ddanbs/upptime/issues/2194", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1661719397
🛑 Danbs is down In 614a94b, Danbs (https://danbs.net) was down: HTTP code: 404 Response time: 175 ms Resolved: Danbs is back up in 7fad613.
gharchive/issue
2023-04-11T03:42:30
2025-04-01T04:33:57.589520
{ "authors": [ "ddanbs" ], "repo": "ddanbs/upptime", "url": "https://github.com/ddanbs/upptime/issues/5628", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1665205794
🛑 Danbs is down In b14bc3e, Danbs (https://danbs.net) was down: HTTP code: 404 Response time: 121 ms Resolved: Danbs is back up in 16dcb84.
gharchive/issue
2023-04-12T20:30:37
2025-04-01T04:33:57.591733
{ "authors": [ "ddanbs" ], "repo": "ddanbs/upptime", "url": "https://github.com/ddanbs/upptime/issues/5685", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1444448818
🛑 Danbs is down In d08d2c1, Danbs (https://danbs.net) was down: HTTP code: 404 Response time: 208 ms Resolved: Danbs is back up in 8a26f12.
gharchive/issue
2022-11-10T19:29:53
2025-04-01T04:33:57.594092
{ "authors": [ "ddanbs" ], "repo": "ddanbs/upptime", "url": "https://github.com/ddanbs/upptime/issues/706", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1795436109
🛑 Danbs is down In f42cc85, Danbs (https://danbs.net) was down: HTTP code: 404 Response time: 3028 ms Resolved: Danbs is back up in b7d8816.
gharchive/issue
2023-07-09T14:56:01
2025-04-01T04:33:57.596456
{ "authors": [ "ddanbs" ], "repo": "ddanbs/upptime", "url": "https://github.com/ddanbs/upptime/issues/8845", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1817000753
🛑 Danbs is down In 900b1d6, Danbs (https://danbs.net) was down: HTTP code: 404 Response time: 144 ms Resolved: Danbs is back up in 0d15838.
gharchive/issue
2023-07-23T05:58:33
2025-04-01T04:33:57.598984
{ "authors": [ "ddanbs" ], "repo": "ddanbs/upptime", "url": "https://github.com/ddanbs/upptime/issues/9295", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2511661316
[TASK] Add blog post about using DDEV in GitLab CI The Issue People have been asking about this for years and it would be lovely for them to know about it. Anywhere would be great, but how about a guest blog on ddev.com? It's just markdown and super easy, https://github.com/ochorocho/ddev-gitlab-ci/issues/11 How This PR Solves The Issue Manual Testing Instructions Automated Testing Overview Related Issue Link(s) Release/Deployment Notes Thanks! Well, the featuredImage was ignored by my global .gitignore 🤦 it should be displayed now. Added/changed the text according to your suggestions. Thank you. Could you just say a bit about how testing is done or enabled in gitlab? Since I've never done that, I'm kind of lost at the very beginning here (and also in the project README). Which testing are your talking about? The bats tests for the container or actual e2e (end to end) tests? Well, i'm not sure if it makes sense to explain how to install ddev addons and set up playwright. It feels a bit out of scope. This and the README can be dramatically improved with a specific example. (I spent some time wondering if the build.sh was important to me, it obviously isn't. But it's the most prominent thing in the README.) So an example might be... "If you have a TYPO3 project checked into Gitlab and want to add testing that uses DDEV, ..." I've extended and restructured the README to be more clear what the important sections are. Important projects sometimes have a hard time living on in a personal repository like ochorocho. I invite you to move this to the ddev org if you're open to that. Alternately, move it into a TYPO3 org or b13 ? If you're interested in moving it, it could move sooner rather than later and then the article wouldn't need updating at that time. I'm happy to move it to the ddev org. FYI, i've moved the project to the ddev org :-) Sorry forgot to answer this one: Shouldn't the default be current DDEV stable version? Why all the references to really old DDEV versions, and why should one have to specify a version? So you're saying the image should have a latest tag too? You sure move fast!!!! Thanks! Which testing are your talking about? The bats tests for the container or actual e2e (end to end) tests? I was talking about using DDEV in other tests, not the tests for this add-on. People like me would just like to understand in context how this will be used. So you're saying the image should have a latest tag too? No, I'm probably saying that it should automatically use the current stable, unless another version is specified. This is all so great, I'll follow up with issues in the repo, will take one more look at the article here and we'll get it moving along. I imagine you already changed the article to change the URL and such. I imagine you already changed the article to change the URL and such. Not just yet. Will do later today or early tomorrow. :-) Instead of going back-and-forth I pushed one small commit and will pull it now. Thanks so much for this contribution and for maintaining it! @rfay thanks for all the support you provided on the way!
gharchive/pull-request
2024-09-07T14:03:38
2025-04-01T04:33:57.607957
{ "authors": [ "ochorocho", "rfay" ], "repo": "ddev/ddev.com", "url": "https://github.com/ddev/ddev.com/pull/243", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2466256007
Regression: Let's Encrypt not working with DDEV v1.23.4 Preliminary checklist [X] I am using the latest stable version of DDEV (see upgrade guide) [X] I have searched existing issues [X] I have checked the troubleshooting guide [X] I have run ddev debug test to include output below Output of ddev debug test Expand `ddev debug test` diagnostic information [COPY-PASTE HERE THE OUTPUT OF `ddev debug test`] Expected Behavior Let's Encrypt should work as described Actual Behavior I see this in docker logs -f ddev-router: 2024-08-14T15:50:48Z ERR github.com/traefik/traefik/v3/pkg/provider/acme/provider.go:396 > Unable to obtain ACME certificate for domains error="unable to generate a certificate for the domains [^d10\\.traefik\\.thefays\\.us$]: acme: error: 400 :: POST :: https://acme-staging-v02.api.letsencrypt.org/acme/new-order :: urn:ietf:params:acme:error:rejectedIdentifier :: Invalid identifiers requested :: Cannot issue for \"^d10\\\\.traefik\\\\.thefays\\\\.us$\": Domain name contains an invalid character" ACME CA=https://acme-staging-v02.api.letsencrypt.org/directory acmeCA=https://acme-staging-v02.api.letsencrypt.org/directory domains=["^d10\\.traefik\\.thefays\\.us$"] providerName=acme-tlsChallenge.acme routerName=d10-web-80-https@file rule="Host(^d10\.traefik\.thefays\.us$)" Reported in https://github.com/ddev/ddev/pull/6388#issuecomment-2288711456 Steps To Reproduce Set up Let's Encrypt, simple config.yaml like name: d10 type: drupal docroot: web php_version: "8.3" webserver_type: nginx-fpm xdebug_enabled: false additional_hostnames: [] additional_fqdns: [] database: type: mariadb version: "10.11" use_dns_when_possible: true composer_version: "2" web_environment: [] corepack_enable: false And with global settings like $ ddev config global |egrep "lets|tld" letsencrypt-email=email@example.com project-tld=traefik.thefays.us use-letsencrypt=true After start, docker logs -f ddev-router Anything else? As reported, changing the project .ddev/traefik/config/<project>.yaml to change Host() statements to remove regex characters in the hostnames fixes this. (And remove #ddev-generated from the top of the file), ddev restart In addition, with additional_hostnames set, the generated config in traefik comes out to be something like rule: Host(d10.traefik.thefays.us, one.traefik.thefays.us) but it should be Host(`d10.traefik.thefays.us`) || Host(`one.traefik.thefays.us`) Obviously you have been faster creating the issue, thanks again 👍 Just an idea, but I testet it in my case and it seems to work: Maybe one can use the HostRegexp Syntax also if letsencrypt is enabled, if domains are defined in the tls section, for your example: rule: "HostRegexp(`^d10\.traefik\.thefays\.us$`) || HostRegexp(`^one\.traefik\.thefays\.us$`)" tls: certResolver: "acme-tlsChallenge" domains: - main: "d10.traefik.thefays.us" sans: - "one.traefik.thefays.us" Please try out the HostRegexp in your setup. I don't see how it offers value when you have to explicitly specify the domains a second time. I hope you'll try out https://github.com/ddev/ddev/pull/6494 @h1nds1ght Don't forget that you have to move your existing edited config settings out of the way or delete them to test
gharchive/issue
2024-08-14T16:07:16
2025-04-01T04:33:57.617063
{ "authors": [ "h1nds1ght", "rfay" ], "repo": "ddev/ddev", "url": "https://github.com/ddev/ddev/issues/6486", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1624609154
Consider using session to provide services Clear and concise description of the problem Consider using session to provide services, because api-key is still a bit expensive Suggested solution I see that other projects generally provide paid plans (apikey) and free plans (session), so see if this project can also support it. Alternative No response Additional context No response Validations [X] Follow our Code of Conduct [X] Read the Contributing Guide. [X] Check that there isn't already an issue that reports the same bug to avoid creating a duplicate. Thanks for the advice, but this approach requires a third-party agent and can be relatively unreliable.
gharchive/issue
2023-03-15T02:26:20
2025-04-01T04:33:57.623946
{ "authors": [ "fangyuan99", "yzh990918" ], "repo": "ddiu8081/chatgpt-demo", "url": "https://github.com/ddiu8081/chatgpt-demo/issues/261", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1636682487
Keeping a constant system prompt Clear and concise description of the problem Can you show how to make the system role the same in the code? Instead of inputting it each time, there is just a generic system role built into the backend. Suggested solution Can you show how to make the system role the same in code? Instead of inputting it each time, there is just a generic system role built into the backend.the Alternative No response Additional context No response Validations [X] Follow our Code of Conduct [X] Read the Contributing Guide. [X] Check that there isn't already an issue that reports the same bug to avoid creating a duplicate. I suspect this may be a duplicate question. Does #290 solve your question?
gharchive/issue
2023-03-23T00:49:53
2025-04-01T04:33:57.626978
{ "authors": [ "ddiu8081", "krish-shahh" ], "repo": "ddiu8081/chatgpt-demo", "url": "https://github.com/ddiu8081/chatgpt-demo/issues/313", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1690699768
feat: public "getCroppedBitmap" method There are cases when we need to use that crop method beyond the use of an instance of a CropController. Well, if you think... :-) Hi @deakjahn! Oh it's even a fact: we need it in our app. The use case is: we crop an image the size of the screen in real time (with the CropController), which is good enough and fast enough in a first approach, but processing the full image takes too long and we do it later in a background task. And in that background task we don't have the CropController anymore, just the full file and the crop parameters. I did copy the code, temporarily, until you release 1.0.6 ;) I assumed that making the method public would make sense and couldn't jeopardy your package's integrity. Nudge, nudge, wink, wink. I know... :-) Nudge, nudge, wink, wink. I know... :-) Thank you for the 1.0.6, and thank you for the quote! Slightly related: https://www.youtube.com/watch?v=ifLqzLEB3E0
gharchive/pull-request
2023-05-01T11:25:40
2025-04-01T04:33:57.667269
{ "authors": [ "deakjahn", "monsieurtanuki" ], "repo": "deakjahn/crop_image", "url": "https://github.com/deakjahn/crop_image/pull/25", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }