id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
58235988
bugfix for relative rule option require 'absolute ruleFilePath' instead of 'relative rule option'; Thank you so much ! This change has been merged and published in anyproxy@3.2.2.
gharchive/pull-request
2015-02-19T16:56:51
2025-04-01T04:55:53.373243
{ "authors": [ "ottomao", "threeday0905" ], "repo": "alibaba/anyproxy", "url": "https://github.com/alibaba/anyproxy/pull/18", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
106470092
fastjson处理字典项,有什么好的办法么? 表里某些字段如性别,存的值是1,对应字典表里是男,然后hql后直接返回成实体类了,然后再用JSON.toJSONString()传给EasyUI的datagrid了,所以想问下,整个过程中,从哪一步处理把1替换成男最好?当然尽量不要说循环实体类好吧,先谢谢各位智慧的网友了。 建议你看下,https://github.com/alibaba/fastjson/wiki/定制序列化这个。 请参考GinKoSen给的连接
gharchive/issue
2015-09-15T03:25:12
2025-04-01T04:55:53.423154
{ "authors": [ "GinKoSen", "WangYu-Li", "wenshao" ], "repo": "alibaba/fastjson", "url": "https://github.com/alibaba/fastjson/issues/417", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1892283812
[BUG] JSONB.parseObject error, create instance error interface java.util.Set, offset 8 问题描述 简要描述您碰到的问题。 环境信息 JDK: openjdk 11 Fastjson2:2.0.40 Dubbo: 3.2.5 重现步骤 server java.io.IOException: org.apache.dubbo.common.serialize.SerializationException: com.alibaba.fastjson2.JSONException: create instance error interface java.util.Set, offset 8 at org.apache.dubbo.common.serialize.DefaultSerializationExceptionWrapper.handleToIOException(DefaultSerializationExceptionWrapper.java:358) at org.apache.dubbo.common.serialize.DefaultSerializationExceptionWrapper.access$000(DefaultSerializationExceptionWrapper.java:28) at org.apache.dubbo.common.serialize.DefaultSerializationExceptionWrapper$ProxyObjectInput.readObject(DefaultSerializationExceptionWrapper.java:168) at org.apache.dubbo.rpc.protocol.dubbo.DecodeableRpcInvocation.drawArgs(DecodeableRpcInvocation.java:267) at org.apache.dubbo.rpc.protocol.dubbo.DecodeableRpcInvocation.decode(DecodeableRpcInvocation.java:168) at org.apache.dubbo.rpc.protocol.dubbo.DecodeableRpcInvocation.decode(DecodeableRpcInvocation.java:102) at org.apache.dubbo.rpc.protocol.dubbo.DubboCodec.decodeBody(DubboCodec.java:197) at org.apache.dubbo.remoting.exchange.codec.ExchangeCodec.decode(ExchangeCodec.java:138) at org.apache.dubbo.remoting.exchange.codec.ExchangeCodec.decode(ExchangeCodec.java:92) at org.apache.dubbo.rpc.protocol.dubbo.DubboCountCodec.decode(DubboCountCodec.java:60) at org.apache.dubbo.remoting.transport.netty4.NettyCodecAdapter$InternalDecoder.decode(NettyCodecAdapter.java:103) at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:529) at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:468) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:290) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412) at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:440) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420) at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: org.apache.dubbo.common.serialize.SerializationException: com.alibaba.fastjson2.JSONException: create instance error interface java.util.Set, offset 8 ... 30 common frames omitted Caused by: com.alibaba.fastjson2.JSONException: create instance error interface java.util.Set, offset 8 at com.alibaba.fastjson2.reader.ObjectReaderImplList.readJSONBObject(ObjectReaderImplList.java:468) at com.alibaba.fastjson2.JSONB.parseObject(JSONB.java:540) at org.apache.dubbo.common.serialize.fastjson2.FastJson2ObjectInput.readObject(FastJson2ObjectInput.java:125) at org.apache.dubbo.common.serialize.DefaultSerializationExceptionWrapper$ProxyObjectInput.readObject(DefaultSerializationExceptionWrapper.java:166) ... 27 common frames omitted Caused by: java.lang.InstantiationException: java.util.Set at java.base/java.lang.Class.newInstance(Class.java:571) at com.alibaba.fastjson2.reader.ObjectReaderImplList.readJSONBObject(ObjectReaderImplList.java:466) ... 30 common frames omitted Caused by: java.lang.NoSuchMethodException: java.util.Set.<init>() at java.base/java.lang.Class.getConstructor0(Class.java:3349) at java.base/java.lang.Class.newInstance(Class.java:556) ... 31 common frames omitted client Caused by: org.apache.dubbo.remoting.RemotingException: Fail to decode request due to: RpcInvocation [methodName=xxxxxMethod, parameterTypes=null] at org.apache.dubbo.remoting.exchange.support.DefaultFuture.doReceived(DefaultFuture.java:249) at org.apache.dubbo.remoting.exchange.support.DefaultFuture.received(DefaultFuture.java:200) at org.apache.dubbo.remoting.exchange.support.DefaultFuture.received(DefaultFuture.java:188) at org.apache.dubbo.remoting.exchange.support.header.HeaderExchangeHandler.handleResponse(HeaderExchangeHandler.java:65) at org.apache.dubbo.remoting.exchange.support.header.HeaderExchangeHandler.received(HeaderExchangeHandler.java:206) at org.apache.dubbo.remoting.transport.DecodeHandler.received(DecodeHandler.java:53) at org.apache.dubbo.remoting.transport.dispatcher.ChannelEventRunnable.run(ChannelEventRunnable.java:62) at org.apache.dubbo.common.threadpool.ThreadlessExecutor$RunnableWrapper.run(ThreadlessExecutor.java:152) at org.apache.dubbo.common.threadpool.ThreadlessExecutor.waitAndDrain(ThreadlessExecutor.java:77) at org.apache.dubbo.rpc.AsyncRpcResult.get(AsyncRpcResult.java:205) 使用同一份 bytes,在 dubbo 环境下会报错,非 dubbo 情况下不会报错 示例代码如下 byte[] bytes = new byte[]{-110, 76, 83, 101, 116, 0, -92, 45, 2, 4, 6, 7, 8, 9, 10, 11, 13, 14, 15, 16, 17, 18, 19, 23, 24, 26, 27, 28, 29, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 44, 46, 47, 56, 48, 56, 49, 56, 50, 56, 51, 56, 52, 56, 53, 56, 54, 56, 55, 56, 56}; try { Collection<Integer> integers1 = readWithoutDubbo(bytes); log.info("integer1 info: {}", integers1); } catch (Exception e) { log.error("integer1 error: {}", e.getMessage()); } try { Collection<Integer> integers2 = readWithDubbo(bytes); log.info("integer2 info: {}", integers2); } catch (Exception e) { log.error("integer2 error: {}", e.getMessage()); } 不使用 Dubbo的ContextAutoTypeBeforeHandler private Collection<Integer> readWithoutDubbo(byte[] bytes) { return JSONB.parseObject(bytes, Collection.class, JSONReader.Feature.UseDefaultConstructorAsPossible, JSONReader.Feature.ErrorOnNoneSerializable, JSONReader.Feature.IgnoreAutoTypeNotMatch, JSONReader.Feature.UseNativeObject, JSONReader.Feature.FieldBased); } 不使用 Dubbo的ContextAutoTypeBeforeHandler 的输出 integer1 info: [2, 4, 6, 7, 8, 9, 10, 11, 13, 14, 15, 16, 17, 18, 19, 23, 24, 26, 27, 28, 29, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 44, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56] 使用 Dubbo的ContextAutoTypeBeforeHandler private Collection<Integer> readWithDubbo(byte[] bytes) { Fastjson2SecurityManager securityManager = FrameworkModel.defaultModel().getBeanFactory().getBean(Fastjson2SecurityManager.class); return JSONB.parseObject(bytes, Collection.class, securityManager.getSecurityFilter(), JSONReader.Feature.UseDefaultConstructorAsPossible, JSONReader.Feature.ErrorOnNoneSerializable, JSONReader.Feature.IgnoreAutoTypeNotMatch, JSONReader.Feature.UseNativeObject, JSONReader.Feature.FieldBased); } 使用 Dubbo的ContextAutoTypeBeforeHandler的输出 integer2 error: create instance error interface java.util.Set, offset 8 由于 bytes 实际类型是 Dubbo 写入的java.util.Set的参数 简单跟了下代码,发现问题可能在 https://github.com/alibaba/fastjson2/blob/d9024ded721ec349ecbc35194246ff8f8725f143/core/src/main/java/com/alibaba/fastjson2/reader/ObjectReaderBaseModule.java#L1831-L1833 走到这的时候,已经找到了对应的 Reader,往下走 https://github.com/alibaba/fastjson2/blob/d9024ded721ec349ecbc35194246ff8f8725f143/core/src/main/java/com/alibaba/fastjson2/reader/ObjectReaderImplList.java#L94-L96 至此返回了一个新的ObjectReaderImplList 回到 https://github.com/alibaba/fastjson2/blob/d9024ded721ec349ecbc35194246ff8f8725f143/core/src/main/java/com/alibaba/fastjson2/reader/ObjectReaderImplList.java#L371-L378 往下走到到报错位置 https://github.com/alibaba/fastjson2/blob/d9024ded721ec349ecbc35194246ff8f8725f143/core/src/main/java/com/alibaba/fastjson2/reader/ObjectReaderImplList.java#L464-L469 https://oss.sonatype.org/content/repositories/snapshots/com/alibaba/fastjson2/fastjson2/2.0.41-SNAPSHOT/ 问题没有重现,你能帮用这个版本验证下么?
gharchive/issue
2023-09-12T11:11:15
2025-04-01T04:55:53.432986
{ "authors": [ "liaozan", "wenshao" ], "repo": "alibaba/fastjson2", "url": "https://github.com/alibaba/fastjson2/issues/1857", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
413648436
add default meter implements Signed-off-by: zhengyangyong yangyong.zheng@qq.com Ⅰ. Describe what this PR did add default meter implements Ⅱ. Does this pull request fix one issue? https://github.com/alibaba/fescar/issues/384 Codecov Report Merging #452 into feature_metrics will increase coverage by 0.96%. The diff coverage is 85.36%. @@ Coverage Diff @@ ## feature_metrics #452 +/- ## ===================================================== + Coverage 29.66% 30.63% +0.96% - Complexity 621 664 +43 ===================================================== Files 175 182 +7 Lines 6945 7068 +123 Branches 840 843 +3 ===================================================== + Hits 2060 2165 +105 - Misses 4667 4681 +14 - Partials 218 222 +4 Impacted Files Coverage Δ Complexity Δ ...va/com/alibaba/fescar/metrics/DefaultRegistry.java 58.33% <58.33%> (ø) 6 <6> (?) .../java/com/alibaba/fescar/metrics/SummaryValue.java 64.7% <64.7%> (ø) 5 <5> (?) ...in/java/com/alibaba/fescar/metrics/TimerValue.java 83.33% <83.33%> (ø) 7 <7> (?) .../java/com/alibaba/fescar/metrics/DefaultGauge.java 90% <90%> (ø) 4 <4> (?) ...ava/com/alibaba/fescar/metrics/DefaultCounter.java 91.66% <91.66%> (ø) 6 <6> (?) ...ava/com/alibaba/fescar/metrics/DefaultSummary.java 96% <96%> (ø) 7 <7> (?) .../java/com/alibaba/fescar/metrics/DefaultTimer.java 96.55% <96.55%> (ø) 8 <8> (?) ... and 4 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 1dcb2a0...591966e. Read the comment docs.
gharchive/pull-request
2019-02-23T03:03:28
2025-04-01T04:55:53.447557
{ "authors": [ "codecov-io", "zhengyangyong" ], "repo": "alibaba/fescar", "url": "https://github.com/alibaba/fescar/pull/452", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2339998313
higress1.3.3不兼容ingress迁移过来的路由规则 操作步骤: helm安装的应用:包括deployment,service,ingress higress:界面上不做任何操作 流量经过:域名-->higress的gateway-->进入部署的应用 出现现象:http可以访问,https出现404 404日志如图: 是1.3.3么,你升级到最新版本试试呢 @leeqwe1 最新版本可以吗
gharchive/issue
2024-06-07T09:25:48
2025-04-01T04:55:53.450194
{ "authors": [ "johnlanni", "leeqwe1" ], "repo": "alibaba/higress", "url": "https://github.com/alibaba/higress/issues/1032", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1593229372
doc: Use helm repo instead of OCI image as the installation source Ⅰ. Describe what this PR did Use helm repo instead of OCI image as the installation source Ⅱ. Does this pull request fix one issue? Ⅲ. Why don't you add test cases (unit test/integration test)? Ⅳ. Describe how to verify it Ⅴ. Special notes for reviews Codecov Report Merging #210 (7532e8a) into main (7bc383e) will decrease coverage by 0.04%. The diff coverage is n/a. Additional details and impacted files @@ Coverage Diff @@ ## main #210 +/- ## ========================================== - Coverage 43.29% 43.25% -0.04% ========================================== Files 31 31 Lines 5038 5038 ========================================== - Hits 2181 2179 -2 - Misses 2689 2691 +2 Partials 168 168 Impacted Files Coverage Δ pkg/bootstrap/server.go 58.39% <0.00%> (-0.77%) :arrow_down:
gharchive/pull-request
2023-02-21T10:33:38
2025-04-01T04:55:53.456693
{ "authors": [ "CH3CHO", "codecov-commenter" ], "repo": "alibaba/higress", "url": "https://github.com/alibaba/higress/pull/210", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2018276711
低代码搭建协议中的 defaultProps 没有实现。 Describe the bug (required) / 详细描述 bug(必填) 根据搭建协议 中对 defaultProps 的定义 低代码组件中如果定义了 defaultProps,该低代码组件应该有默认的 props 值。但是目前没有实现。 复现方式: 1.启动 lowcode-demo 中的 demo-general 2.替换 demo-general/src/plugins/plugin-lowcode-component/lowcode-schema.json 中的 schema 这里添加了 defaultProps 这里将文案替换成啦 this.props.name || 'ccc' 预览的结果 期望的结果 展示:'xxx' 完整的 lowcode-schema,可以直接替换: { "version": "1.0.0", "componentsMap": [ ], "componentsTree": [ { "componentName": "Component", "id": "node_dockcviv8fo1", "props": { "ref": "outerView", "style": { "height": "100%" } }, "defaultProps": { "name": "xxx" }, "docId": "doclaqkk3b9", "fileName": "/", "dataSource": { "list": [ { "type": "fetch", "isInit": true, "options": { "params": {}, "method": "GET", "isCors": true, "timeout": 5000, "headers": {}, "uri": "mock/info.json" }, "id": "info", "shouldFetch": { "type": "JSFunction", "value": "function() { \n console.log('should fetch.....');\n return true; \n}" } } ] }, "state": { "text": { "type": "JSExpression", "value": "\"outer\"" }, "isShowDialog": { "type": "JSExpression", "value": "false" } }, "css": "body {\n font-size: 12px;\n}\n\n.button {\n width: 100px;\n color: #ff00ff\n}", "lifeCycles": { "componentDidMount": { "type": "JSFunction", "value": "function componentDidMount() {\n console.log('did mount');\n}" }, "componentWillUnmount": { "type": "JSFunction", "value": "function componentWillUnmount() {\n console.log('will unmount');\n}" } }, "methods": { "testFunc": { "type": "JSFunction", "value": "function testFunc() {\n console.log('test func');\n}" }, "onClick": { "type": "JSFunction", "value": "function onClick() {\n this.setState({\n isShowDialog: true\n });\n}" }, "closeDialog": { "type": "JSFunction", "value": "function closeDialog() {\n this.setState({\n isShowDialog: false\n });\n}" }, "getHelloWorldText": { "type": "JSFunction", "value": "function getHelloWorldText() {\n return this.i18n('i18n-jwg27yo4');\n}" }, "getHelloWorldText2": { "type": "JSFunction", "value": "function getHelloWorldText2() {\n return this.i18n('i18n-jwg27yo3', {\n name: '絮黎'\n });\n}" }, "onTestConstantsButtonClicked": { "type": "JSFunction", "value": "function onTestConstantsButtonClicked() {\n console.log('constants.ConstantA:', this.constants.ConstantA);\n console.log('constants.ConstantB:', this.constants.ConstantB);\n}" }, "onTestUtilsButtonClicked": { "type": "JSFunction", "value": "function onTestUtilsButtonClicked() {\n this.utils.demoUtil('param1', 'param2');\n}" } }, "originCode": "class LowcodeComponent extends Component {\n state = {\n \"text\": \"outer\",\n \"isShowDialog\": false\n }\n componentDidMount() {\n console.log('did mount');\n }\n componentWillUnmount() {\n console.log('will unmount');\n }\n testFunc() {\n console.log('test func');\n }\n onClick() {\n this.setState({\n isShowDialog: true\n });\n }\n closeDialog() {\n this.setState({\n isShowDialog: false\n });\n }\n getHelloWorldText() {\n return this.i18n('i18n-jwg27yo4');\n }\n getHelloWorldText2() {\n return this.i18n('i18n-jwg27yo3', {\n name: '絮黎',\n });\n }\n onTestConstantsButtonClicked() {\n console.log('constants.ConstantA:', this.constants.ConstantA);\n console.log('constants.ConstantB:', this.constants.ConstantB);\n\t}\n\tonTestUtilsButtonClicked(){\n this.utils.demoUtil('param1', 'param2');\n\t}\n}", "hidden": false, "title": "", "isLocked": false, "condition": true, "conditionGroup": "", "children": [ { "componentName": "NextRowColContainer", "id": "node_oclawu71ac3", "docId": "doclawu75i0", "props": { "rowGap": 20, "colGap": 20 }, "title": "行列容器", "hidden": false, "isLocked": false, "condition": true, "conditionGroup": "", "children": [ { "componentName": "NextRow", "id": "node_oclawu71ac4", "docId": "doclawu75i0", "props": {}, "title": "行", "hidden": false, "isLocked": false, "condition": true, "conditionGroup": "", "children": [ { "componentName": "NextCol", "id": "node_oclawu71ac5", "docId": "doclawu75i0", "props": { "colSpan": 1 }, "title": "列", "hidden": false, "isLocked": false, "condition": true, "conditionGroup": "", "children": [ { "componentName": "NextP", "id": "node_oclawu71ac3n", "docId": "doclawu71ac", "props": { "wrap": false, "type": "body2", "verAlign": "middle", "textSpacing": true, "align": "left", "prefix": "", "full": false, "flex": false }, "title": "段落", "hidden": false, "isLocked": false, "condition": true, "conditionGroup": "", "children": [ { "componentName": "Button", "id": "node_oclawu71ac3o", "docId": "doclawu71ac", "props": { "prefix": "next-", "type": "primary", "size": "medium", "htmlType": "button", "component": "button", "children": "constants", "iconSize": "xxs", "loading": false, "text": false, "warning": false, "disabled": false, "ref": "button-4951c2d3", "__events": { "eventDataList": [ { "type": "componentEvent", "name": "onClick", "relatedEventName": "onTestConstantsButtonClicked" } ], "eventList": [ { "name": "onClick", "description": "点击按钮的回调\n@param {Object} e Event Object", "disabled": true }, { "name": "onMouseUp", "disabled": false } ] }, "onClick": { "type": "JSFunction", "value": "function(){this.onTestConstantsButtonClicked.apply(this,Array.prototype.slice.call(arguments).concat([])) }" } }, "hidden": false, "title": "", "isLocked": false, "condition": true, "conditionGroup": "" }, { "componentName": "Button", "id": "node_oclawu71ac3p", "docId": "doclawu71ac", "props": { "prefix": "next-", "type": "primary", "size": "medium", "htmlType": "button", "component": "button", "children": "utils", "iconSize": "xxs", "loading": false, "text": false, "warning": false, "disabled": false, "__events": { "eventDataList": [ { "type": "componentEvent", "name": "onClick", "relatedEventName": "onTestUtilsButtonClicked" } ], "eventList": [ { "name": "onClick", "description": "点击按钮的回调\n@param {Object} e Event Object", "disabled": true }, { "name": "onMouseUp", "disabled": false } ] }, "onClick": { "type": "JSFunction", "value": "function(){this.onTestUtilsButtonClicked.apply(this,Array.prototype.slice.call(arguments).concat([])) }" } }, "hidden": false, "title": "", "isLocked": false, "condition": true, "conditionGroup": "" } ] } ] } ] } ] }, { "componentName": "NextP", "id": "node_oclaqjdn9qv", "docId": "doclaqjdn9q", "props": { "wrap": false, "type": "body2", "verAlign": "middle", "textSpacing": true, "align": "left", "prefix": "", "full": false, "flex": false }, "title": "段落", "hidden": false, "isLocked": false, "condition": true, "conditionGroup": "", "children": [ { "componentName": "NextText", "id": "node_oclaqjdn9qw", "docId": "doclaqjdn9q", "props": { "type": "h5", "children": { "type": "JSExpression", "value": "this.props.name || 'ccc'" }, "prefix": "", "classname": "", "mark": false, "code": false, "delete": false, "underline": false, "strong": false }, "hidden": false, "title": "", "isLocked": false, "condition": true, "conditionGroup": "" }, { "componentName": "NextText", "id": "node_oclarv0wja2", "docId": "doclarv56xh", "props": { "type": "h5", "children": { "type": "JSExpression", "value": "this.props.name || 'ccc'" }, "prefix": "", "classname": "", "mark": false, "code": false, "delete": false, "underline": false, "strong": false }, "hidden": false, "title": "", "isLocked": false, "condition": true, "conditionGroup": "" } ] } ] } ], "i18n": { "zh-CN": { "i18n-jwg27yo4": "你好 ", "i18n-jwg27yo3": "{name} 博士" }, "en-US": { "i18n-jwg27yo4": "Hello ", "i18n-jwg27yo3": "Doctor {name}" } } } 相关的 issue:https://github.com/alibaba/lowcode-engine/issues/2712
gharchive/issue
2023-11-30T09:56:24
2025-04-01T04:55:53.497987
{ "authors": [ "liujuping" ], "repo": "alibaba/lowcode-engine", "url": "https://github.com/alibaba/lowcode-engine/issues/2715", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2288780498
通过列表切换实现动态页面渲染的能力的问题 想实现点击左侧树列表,右侧预览页动态渲染,现在点击左侧树引起state改变,按理来说会重新渲染DOM,但是页面没有切换渲染,参见下图 预测是预览初始化方法使用async的原因,请问应该如何解决这个问题 这个跟初始化方法没有关系,你得看下是不是hook使用时缺少了对于选中页面变化的依赖
gharchive/issue
2024-05-10T02:14:56
2025-04-01T04:55:53.501841
{ "authors": [ "1490113799", "AndyJinSS" ], "repo": "alibaba/lowcode-engine", "url": "https://github.com/alibaba/lowcode-engine/issues/3019", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1195695505
feat(create-element): init 参数使用 projectName \ 统一生成目录的前缀 改动点1: npm init @alilc/element your-element-name your-element-name 改成非必传 改动点2: 参考 https://github.com/alibaba/lowcode-plugins/tree/main/packages 的 plugin 命名方式,给生成的文件夹拼接 plugin-、setter-、component- 前缀 目前生成的 plugin 跟 setter 文件模板,会自动拼接成 LowcodePluginXXX 跟 LowcodeSetterXXX,所以文件夹名称统一更好? Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.Shuogui Lin seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account.You have signed the CLA already but the status is still pending? Let us recheck it. 这个和我们工具设计的思路不太一致,我们是希望,如果不传 your-element-name 的情况下,就使用当前目录进行初始化 这一点和 npm init 自己的行为保持一致。 https://docs.npmjs.com/cli/v8/commands/npm-init @eternalsky 理解了, 改动 1 没必要 改动 2 呢 ?
gharchive/pull-request
2022-04-07T08:26:17
2025-04-01T04:55:53.507227
{ "authors": [ "CLAassistant", "eternalsky", "twinkle77" ], "repo": "alibaba/lowcode-tools", "url": "https://github.com/alibaba/lowcode-tools/pull/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
335052957
eip brandwith setting does not work @zhuzhih2017 "internet_charge_type": "PayByTraffic", "internet_max_bandwidth_out": 100, set the two fields above in packer config JSON file, and run packer build, the brand width of the EIP allocated by packer was set to 5Mbps。 reason: packer does not pass BrandWidth param when allocate EIP and the EIP use the default brand width which is 5Mbps. if b.chooseNetworkType() == VpcNet { steps = append(steps, &setpConfigAlicloudEIP{ AssociatePublicIpAddress: b.config.AssociatePublicIpAddress, RegionId: b.config.AlicloudRegion, InternetChargeType: b.config.InternetChargeType, }) } else { steps = append(steps, &stepConfigAlicloudPublicIP{ RegionId: b.config.AlicloudRegion, }) } This should be a bug and need to be fixed. fixed on the packer official repo.
gharchive/issue
2018-06-23T00:42:18
2025-04-01T04:55:53.579503
{ "authors": [ "willmao" ], "repo": "alibaba/packer-provider", "url": "https://github.com/alibaba/packer-provider/issues/36", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
736659077
About Nacos config: I'm confused when handling with multi-environment config files. It seems either Nacos's namespace or Spring's profile has provided a way to fulfill multi-env requirement, but I'm wondering which is the reconmmended way to do this? https://github.com/alibaba/spring-cloud-alibaba/wiki/Nacos-config
gharchive/issue
2020-11-05T06:40:24
2025-04-01T04:55:53.585421
{ "authors": [ "SuperIllidan", "yuhuangbin" ], "repo": "alibaba/spring-cloud-alibaba", "url": "https://github.com/alibaba/spring-cloud-alibaba/issues/1804", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1007977266
Sidecar无法注册服务 此处逻辑是否有问题?导致异构的服务无法注册到服务中心。 方法位置:com.alibaba.cloud.sidecar.SidecarHealthChecker#check 不知道服务注册的入口就是这里,如果是那么在第一次注册服务的时候方框里的代码是创建缓存并保存到MAP,然后是如果状态为UP就先检查一下是否需要注册,但检查是否需要注册的逻辑是对比缓存,此方法永远返回false,导致服务永远不会被注册。 注册的逻辑是对比缓存,此方法永远返回false, 上面已经说了,这里是服务健康检查的处理,第一次注册服务并不是在这里处理 你的问题仅仅是通过自己看代码觉得有问题吗? 你可以动手实践一下
gharchive/issue
2021-09-27T10:29:19
2025-04-01T04:55:53.587356
{ "authors": [ "oSunnyo", "yuhuangbin" ], "repo": "alibaba/spring-cloud-alibaba", "url": "https://github.com/alibaba/spring-cloud-alibaba/issues/2273", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
977897956
tidevice install失败了 版本tidevice version 0.4.18 CreatingStagingDirectory (5%) ExtractingPackage (15%) [E 210824 17:27:36 _device:744] PackageExtractionFailed [E 210824 17:27:36 _device:745] Could not extract archive Traceback (most recent call last): File "/usr/local/bin/tidevice", line 8, in sys.exit(main()) File "/usr/local/lib/python3.9/site-packages/tidevice/main.py", line 723, in main actionsargs.subparser File "/usr/local/lib/python3.9/site-packages/tidevice/main.py", line 128, in cmd_install bundle_id = d.app_install(args.filepath_or_url) File "/usr/local/lib/python3.9/site-packages/tidevice/_device.py", line 746, in app_install raise ServiceError(progress['Error']) tidevice.exceptions.ServiceError: PackageExtractionFailed 请问下解决了吗 我也遇到了相同的问题 方便的话提供一下安装包 可能是设备上的存储空间不足了,清一下空间解决了
gharchive/issue
2021-08-24T09:28:21
2025-04-01T04:55:53.591691
{ "authors": [ "chengducode", "codeskyblue", "gangeyubo", "hc2088" ], "repo": "alibaba/taobao-iphone-device", "url": "https://github.com/alibaba/taobao-iphone-device/issues/135", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1836370636
Supports appending HTTP headers of the user request to upstream based on routing of header, cookie or query 基于Header/Cookie/Query的路由匹配,向后端upstream转发时在请求中追加指定的Header。 在请求消息的header中追加header值 如果存在相同的header名,在已有的header值后增加'&'分割符,再追加header值 如果请求消息的header不存在,则直接增加 值格式:"header name:header value[||header name:header value]*" 值校验:header以'||'分割,header名和header值以':'分割,header name和header value不允许包含':'和'||',默认最多允许追加2个header 无损动态生效,无需tengine reload。 apiVersion: networking.k8s.io/v1 kind: Ingress metadata: annotations: nginx.ingress.kubernetes.io/canary: "true" nginx.ingress.kubernetes.io/canary-by-query: open3 nginx.ingress.kubernetes.io/canary-by-query-value: abc1||abc2||abc3||abc4||abc5||abc6||abc7||abc8||abc9||abc10||abc11||abc12||abc13||abc14||abc15||abc16||abc17||abc18||abc19||abc20||abc21 nginx.ingress.kubernetes.io/canary-request-append-header: UserData:user=236-564-29121||test-ingress6:$ssl_protocol nginx.ingress.kubernetes.io/ssl-redirect: "true" creationTimestamp: "2023-10-31T06:30:31Z" generation: 2 name: tengine-ingress-hello-query-ing namespace: default resourceVersion: "28580135" uid: debebfd0-afd2-4604-af8d-a41ad400bfbc spec: ingressClassName: opensource-ingress rules: - host: echo.w1.com http: paths: - backend: service: name: tengine-ingress-hello3-service port: number: 80 path: / pathType: Prefix tls: - hosts: - echo.w1.com secretName: https-server-1 status: loadBalancer: ingress: - {} $ curl -i -k -H "test-ingress6: default" https://echo.w1.com/?open3=abc2 HTTP/2 301 server: Tengine/3.1.0 date: Wed, 01 Nov 2023 11:33:42 GMT location: /?open3=abc2/ strict-transport-security: max-age=0 ups-target-key: default-tengine-ingress-hello3-service-80 x-protocol: HTTP/2.0 alt-svc: h3=":443"; ma=2592000,h3-29=":443"; ma=2592000
gharchive/issue
2023-08-04T09:11:19
2025-04-01T04:55:53.595663
{ "authors": [ "lianglli" ], "repo": "alibaba/tengine", "url": "https://github.com/alibaba/tengine/issues/1844", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
177625303
[Android] WXStreamModule.fetch 优化建议 开发过程中有一个http请求是使用post方法,在iOS和H5下面是OK的,在android上面请求报错,经调查Andriod平台下http method只能是大写的。 以下两种建议二选一: 建议在Android中把客户端传过来的http method 做大写统一转换; Stream文档中明确提出必须要大些method; 好建议,欢迎你发个PR来实现这些优化 👍 @sospartan 好的,晚上补一下 @sospartan 已经提交PR
gharchive/issue
2016-09-18T02:50:53
2025-04-01T04:55:53.627481
{ "authors": [ "lanceliu", "sospartan" ], "repo": "alibaba/weex", "url": "https://github.com/alibaba/weex/issues/1239", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
170012093
add screenshot Nice theme,I added the screenshot as recommanded in the documentation: https://jekyllrb.com/docs/themes/#adding-a-screenshot Sorry, I somehow just saw this. Thanks 🙌
gharchive/pull-request
2016-08-08T20:02:40
2025-04-01T04:55:53.649946
{ "authors": [ "DirtyF", "aliou" ], "repo": "aliou/ace", "url": "https://github.com/aliou/ace/pull/5", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
326750994
Support HTTP/2 protocol. Motivation: Support HTTP/1.1 & HTTP/2 protocol. Modification: Add data to SofaRequest & SofaResponse Add HTTP/1.1 server & HTTP/2 Clear text server Add HTTP/2 client transport. Result: Fix #25 Codecov Report Merging #138 into master will increase coverage by 0.07%. The diff coverage is 65.83%. @@ Coverage Diff @@ ## master #138 +/- ## ============================================ + Coverage 68.14% 68.21% +0.07% - Complexity 730 919 +189 ============================================ Files 302 335 +33 Lines 12644 14067 +1423 Branches 2077 2260 +183 ============================================ + Hits 8616 9596 +980 - Misses 2984 3306 +322 - Partials 1044 1165 +121 Impacted Files Coverage Δ Complexity Δ .../sofa/rpc/server/SofaRejectedExecutionHandler.java 87.5% <ø> (ø) 0 <0> (?) .../java/com/alipay/sofa/rpc/common/RpcConstants.java 66.66% <ø> (ø) 0 <0> (ø) :arrow_down: ...rpc/common/struct/UnsafeByteArrayOutputStream.java 51.51% <ø> (ø) 0 <0> (ø) :arrow_down: ...com/alipay/sofa/rpc/transport/AbstractByteBuf.java 100% <ø> (ø) 0 <0> (ø) :arrow_down: .../rpc/common/struct/UnsafeByteArrayInputStream.java 34.14% <ø> (ø) 0 <0> (ø) :arrow_down: ...m/alipay/sofa/rpc/core/exception/RpcErrorType.java 0% <ø> (ø) 0 <0> (ø) :arrow_down: .../com/alipay/sofa/rpc/common/RemotingConstants.java 0% <ø> (ø) 0 <0> (ø) :arrow_down: .../sofa/rpc/registry/zk/ZookeeperRegistryHelper.java 67.16% <0%> (ø) 0 <0> (ø) :arrow_down: .../sofa/rpc/registry/zk/ZookeeperConfigObserver.java 15.71% <0%> (ø) 0 <0> (ø) :arrow_down: ...ofa/rpc/registry/zk/ZookeeperOverrideObserver.java 15.27% <0%> (ø) 0 <0> (ø) :arrow_down: ... and 97 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 3d98b80...ea66364. Read the comment docs. It seems tracer is ok client {"timestamp":"2018-05-27 18:43:48.681","tracerId":"0a4187b11527417826069100199533","spanId":"0","span.kind":"client","local.app":"test-client","protocol":"h2c","service":"com.alipay.sofa.rpc.protobuf.ProtoService","method":"echoObj","current.thread.name":"main","invoke.type":"sync","router.record":"DIRECT","remote.ip":"127.0.0.1:12300","local.client.ip":"127.0.0.1","result.code":"00","req.serialize.time":"1206","req.size":"5","client.elapse.time":"1844","local.client.port":"54560","baggage":""} server {"timestamp":"2018-05-27 18:43:47.253","tracerId":"","spanId":"","span.kind":"server","service":"com.alipay.sofa.rpc.protobuf.ProtoService","method":"echoObj","remote.ip":"127.0.0.1","remote.app":"test-client","protocol":"h2c","local.app":"test-server","current.thread.name":"SOFA-SEV-H2C-BIZ-12300-3-T1","result.code":"00","biz.impl.time":"3","baggage":""} netty convert header to lowercase. when we receive in server. rpc_trace_context.sofaRpcId will become rpc_trace_context.sofarpcid @leizhiyuan I have fixed it. See com.alipay.sofa.rpc.transport.http.HttpTracerUtils. @ujjboy {"timestamp":"2018-05-28 06:34:01.173","tracerId":"c0a8016c152746044116910138939","spanId":"0","span.kind":"server","service":"com.alipay.sofa.rpc.protobuf.ProtoService","method":"echoObj","remote.ip":"127.0.0.1","remote.app":"test-client","protocol":"h2c","local.app":"test-server","current.thread.name":"SOFA-SEV-H2C-BIZ-12300-3-T2","result.code":"00","biz.impl.time":"0","baggage":""} now, tracer is ok
gharchive/pull-request
2018-05-26T16:36:42
2025-04-01T04:55:53.671421
{ "authors": [ "codecov-io", "leizhiyuan", "ujjboy" ], "repo": "alipay/sofa-rpc", "url": "https://github.com/alipay/sofa-rpc/pull/138", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2705138607
How to read de trace file( xx.tr )? After simulation ,we get the trace file( xx.tr ). In the directory: SimAI/ns-3-alibabacloud/analysis : './trace_reader <xx.tr file>', but read nothing. Same as #29 ,I assume you wish to trace the flow dynamics within ns3, like switch throughput, NIC throughput, and switch queue information, which would be easily visible in the newer ns-3-alibabcloud version.
gharchive/issue
2024-11-29T13:26:54
2025-04-01T04:55:53.732178
{ "authors": [ "Huoyuan100861", "wangdong-66" ], "repo": "aliyun/SimAI", "url": "https://github.com/aliyun/SimAI/issues/24", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1753893203
请教一下bpf关于拼接长字符串的问题, 您好,麻烦请教一下就是我在内核态程序需要循环通过struct dentry结构获取某一个文件的路径,我的方法目前是首先有一个结构体,结构体内部有一个char filename[512];的数组,然后循环将每一级目录加上'/'拼成一个完整的路径, 问题是循环过程中调用bpf_probe_read_kernel_str接口向filename中拷贝时如果第二个参数是一个变量,那么运行时Bpf校验器就会说可能越界什么的,但是在前后我已经做了很完善的边界检查了,之所以bpf_probe_read_kernel_str第二个参数需要是变量是因为每一级目录名长度是不同的,当我改成一个定值就没问题了。 其实总的来说就是bpf如何拼接一个长字符串,我基本尝试过所有相关的接口也无法解决了 下面是部分代码 其中用到的部分定义如下 char filename[512]; #define MAX_DENTRY_NAME_LEN 20 size_t pos = 0; const u32 MAX_BACKTRACE_DEPTH = 20; for (u32 cnt = MAX_BACKTRACE_DEPTH; cnt != 0; --cnt) { if (err || curr_dentry == NULL) { break; } int name_len = BPF_CORE_READ(curr_dentry, d_name.len); const u8 *name = BPF_CORE_READ(curr_dentry, d_name.name); if (name_len <= 1) { break; } name_len = bpf_probe_read_kernel_str(filename + pos, MAX_DENTRY_NAME_LEN, name); if (name_len <= 1) { BPFLOGD(BPF_TRUE, "failed to read dentry name from kernel stack buffer"); break; } pos += name_len; filename[pos - 1] = '/'; struct dentry *temp_dentry = BPF_CORE_READ(curr_dentry, d_parent); if (temp_dentry == curr_dentry || temp_dentry == NULL) { break; } curr_dentry = temp_dentry; } 目前的效果是只能去循环20层,每层目录拿20个字符,我也将filename改为了4096长度,但是bpf_probe_read_kernel_str接口还是不允许我使用变量(上面代码中的name_len)作为第二个参数,我最终想要的效果其实是想通过循环可以完整的拿到每一级目录名,以及最终拼接成完整的路径,请教大佬是否有好的解决方案,感谢!!! 据我所知,eBPF 堆栈大小被限制在 MAX_BPF_STACK,现在应该是512字节,你这样用应该是越界的,解决方法是改用 bpf maps存储。 /* BPF program can access up to 512 bytes of stack space. */ #define MAX_BPF_STACK 512 据我所知,eBPF 堆栈大小被限制在 MAX_BPF_STACK,现在应该是512字节,你这样用应该是越界的,解决方法是改用 bpf maps存储。 /* BPF program can access up to 512 bytes of stack space. */ #define MAX_BPF_STACK 512 ok,我尝试一些,非常感谢 老哥,请问解决了吗,我最近也想构建完整的路径 老哥,请问解决了吗,我最近也想构建完整的路径 并没有好的解决办法吧 最近在做这个东西,我已经解决,并且兼容3.10内核的操作系统。不过我用的是libbpf,没有使用coolbpf, coolbpf 还没玩过。 最近在做这个东西,我已经解决,并且兼容3.10内核的操作系统。不过我用的是libbpf,没有使用coolbpf, coolbpf 还没玩过。 请教一下你是大概怎么处理的呢
gharchive/issue
2023-06-13T02:00:06
2025-04-01T04:55:53.762914
{ "authors": [ "Killing11L", "guomsh", "luxingwen", "majoki" ], "repo": "aliyun/coolbpf", "url": "https://github.com/aliyun/coolbpf/issues/26", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
312747249
关于同步一致性问题 MySQL -> PostgreSQL/Greenplum mysql2pgsql 功能:以表为单位的多线程全量数据迁移 使用全量同步的时候,怎么知道mysql binlog的位置?能保证同步时 数据的一致性吗? 我理解全量数据同步不需要知道 binlog 的位置吧,只是去执行一条 sql,比如 select * from table1。 关于一致性,没有验证,但是有个参数,插入到GP允许有多少条错误,比如设置为10,有11条数据插入失败,整个同步就失败了,如果有9条数据插入失败,整个同步是成功的,但是那9条还是没有同步到GP
gharchive/issue
2018-04-10T02:03:38
2025-04-01T04:55:53.764547
{ "authors": [ "LeoLiang666", "zhangnew" ], "repo": "aliyun/rds_dbsync", "url": "https://github.com/aliyun/rds_dbsync/issues/96", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1640298398
Finish mobile finish mobile version -create index.html -create style.css I don't understand. where is the problem. On Fri, Mar 24, 2023 at 6:49 PM Amine @.***> wrote: @.**** commented on this pull request. Hi @alizaamir123 https://github.com/alizaamir123 While you made a great effort in this project, unfortunately, I cannot proceed to review your code. Invalid Code Review Request You have submitted a project, that doesn't have a PR summary. Friendly reminder: You should make sure that your repository looks professional before review. Please take a step back apply 5 simple rules from this lesson https://github.com/microverseinc/curriculum-transversal-skills/blob/main/documentation/articles/professional_repo_rules.md). Your Code Review Request will be marked as invalid in your Dashboard, so please submit a new one once you are ready 🙏 Cheers and Happy coding!👏👏👏 Feel free to leave any questions or comments in the PR thread if something is not 100% clear. Please, remember to tag me in your question so I can receive the notification. Please, do not open a new Pull Request for re-reviews. You should use the same Pull Request submitted for the first review, either valid or invalid unless it is requested otherwise. Invalid Code Review Request does not count into the code reviews limit. — Reply to this email directly, view it on GitHub https://github.com/alizaamir123/setup-mobile-first/pull/3#pullrequestreview-1357701548, or unsubscribe https://github.com/notifications/unsubscribe-auth/ATWAXS5AE4MVQQ4JWMH3EQ3W5Y6KBANCNFSM6AAAAAAWHGOVWQ . You are receiving this because you were mentioned.Message ID: @.***> @alizaamir123 I think the problem the first reviewer was referring to is that you had already implemented the changes on your main branch, so please next time avoid that so that you don't have to merge anything here is an article on how to follow Github flow 😉 Hi Micronaut @alizaamir123 While you made a great effort in this project, unfortunately, I cannot proceed to review your code. Invalid Code Review Request 🔴 Ensure that you make changes by committing and pushing to the same branch before making a Pull Request Your Code Review Request will be marked as invalid in your Dashboard, so please submit a new one once you are ready 🙏 Cheers and Happy coding!👏👏👏 Feel free to leave any questions or comments in the PR thread if something is not 100% clear. Please, do not open a new Pull Request for re-reviews. You should use the same Pull Request submitted for the first review, either valid or invalid unless it is requested otherwise. Invalid Code Review Request does not count into the code reviews limit. Hello @alizaamir123 👋, _This is an additional review as a result of the second opinion request. Looks like your project is already approved, and your issue has been solved. ✊ Keep rocking 🚀!
gharchive/pull-request
2023-03-25T01:24:42
2025-04-01T04:55:53.774848
{ "authors": [ "alizaamir123", "medaminedev66", "tsohleDev" ], "repo": "alizaamir123/setup-mobile-first", "url": "https://github.com/alizaamir123/setup-mobile-first/pull/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1245046949
truenas scale 拉取docker镜像失败 你使用的 chinesesubfinder 是什么版本,什么环境? docker 你遇到什么问题了? 拉取镜像提示 Failed to pull image "allanpk716/chinesesubfinder:latest": rpc error: code = Unknown desc = Error response from daemon: Get "https://registry-1.docker.io/v2/": unexpected EOF 老板,不知道为什么,latest标签拉到的都不是最新的镜像.....每次都是手动更改docker-compose后面的版本号...
gharchive/issue
2022-05-23T11:53:20
2025-04-01T04:55:53.790609
{ "authors": [ "1061875478", "ZCkyle" ], "repo": "allanpk716/ChineseSubFinder", "url": "https://github.com/allanpk716/ChineseSubFinder/issues/368", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
664971547
scispacy entitylinker error in creating the object Linker Linker = EntityLinker() using scispacy-0.2.5 and en_core_sci_sm-0.2.5 entity_linker_error_log.txt file includes the error log the packages and dependencies installed and docker file for the container environment Hi, I am not able to reproduce this, even using your Dockerfile and requirements. Could you try again and make sure to fully rebuild the docker image?
gharchive/issue
2020-07-24T07:24:58
2025-04-01T04:55:53.829265
{ "authors": [ "danielkingai2", "radiatechs-karan" ], "repo": "allenai/scispacy", "url": "https://github.com/allenai/scispacy/issues/257", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1048215714
manually chose references with specific issues into inputs/sample.jso… …n generate_dqm_json_test_set.py now has option to generate sample reference set based on sample.json Cool, thanks Ian
gharchive/pull-request
2021-11-09T05:48:03
2025-04-01T04:55:53.831980
{ "authors": [ "azurebrd" ], "repo": "alliance-genome/agr_literature_service", "url": "https://github.com/alliance-genome/agr_literature_service/pull/20", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2177115420
🛑 West Shire Decks is down In 9354652, West Shire Decks (https://westshiredecks.com) was down: HTTP code: 0 Response time: 0 ms Resolved: West Shire Decks is back up in 21da754 after 8 minutes.
gharchive/issue
2024-03-09T07:44:53
2025-04-01T04:55:53.834667
{ "authors": [ "alliancewebteam" ], "repo": "alliancewebteam/Uptime", "url": "https://github.com/alliancewebteam/Uptime/issues/1071", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2026982187
🛑 Capital Hospitality Services is down In 5eba8a8, Capital Hospitality Services (https://chshotelrenovations.com) was down: HTTP code: 0 Response time: 0 ms Resolved: Capital Hospitality Services is back up in eca83cf after 8 minutes.
gharchive/issue
2023-12-05T19:34:41
2025-04-01T04:55:53.837045
{ "authors": [ "alliancewebteam" ], "repo": "alliancewebteam/Uptime", "url": "https://github.com/alliancewebteam/Uptime/issues/508", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2512448553
🛑 Buckland Sales is down In 8f22612, Buckland Sales (https://bucklandsales.com) was down: HTTP code: 0 Response time: 0 ms Resolved: Buckland Sales is back up in 617a6ee after 11 minutes.
gharchive/issue
2024-09-08T15:43:39
2025-04-01T04:55:53.839540
{ "authors": [ "alliancewebteam" ], "repo": "alliancewebteam/uptime", "url": "https://github.com/alliancewebteam/uptime/issues/206", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2631804614
Demo Issue 2 This is to test the workflow This issue has been automatically marked as stale due to inactivity. If no further activity occurs, it will be closed. This issue has been closed due to inactivity.
gharchive/issue
2024-11-04T04:28:00
2025-04-01T04:55:53.841708
{ "authors": [ "alloydsa", "dkharkwal" ], "repo": "alloydsa/Testaction", "url": "https://github.com/alloydsa/Testaction/issues/3", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
264934606
Copy ~/artifacts.key to working directory, so it is rsynced to duffy machine. This should enable rsyncing to artifacts.ci.centos.org directly from duffy machine (useful for junit results, screenshots, ...). Signed-off-by: rhopp rhopp@dhcp-10-40-4-239.brq.redhat.com Len is using this approach in his jobs (https://github.com/almighty/almighty-jobs/blob/master/devtools-ci-index.yaml#L1140). I think it could be beneficial for all jobs. What do you think @kbsingh ? Is it ok to have it in global template, or should I put it just into jobs I'm interested in?
gharchive/pull-request
2017-10-12T13:06:11
2025-04-01T04:55:53.874930
{ "authors": [ "rhopp" ], "repo": "almighty/almighty-jobs", "url": "https://github.com/almighty/almighty-jobs/pull/345", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1780509298
[BUG] На телефоне выпадающий список не перемещается вместе с остальной формой Опишите баг В виджетах "Заявка" в разделе "Продвинутые" для "Группа", "Тип исполнения заявки", "Сторона", а также в табе Условная для "Условие" в настройках "Блоттер" и "Стакан" в Профиле в табах "Общие" и "Уведомления": при клике на поле с дропдауном (например, Колонки заявок) и свайпе вне выпадающего списка - дропдаун остается в том же месте на экране телефона: https://github.com/alor-broker/Astras-Trading-UI/assets/88287882/1735b816-f5d6-4ff9-82b3-2c958be83399 Воспроизведение С телефона авторизоваться в Астрас. Перейти в блоттер, в раздел настроек. Нажать на любое поле ввода и свайпнуть вверх/вниз. Посмотреть на перемещение выпадающего списка. Среда. Android 10 Google Chrome v110, Mozilla Firefox Ожидаемое поведение При таче или свайпе на область вне дропдауна сворачивать дропадун или также перемещать вверх / вниз относительно экрана телефона. Скриншоты No response Дополнительная информация No response Правила [X] Я согласен следовать правилам Мы обсудили эту задачу. Так по умолчанию реализовано в ng zorro (нашем UI фреймворке) и мы не видим сильных аргументов чтобы менять это поведение. Поэтому закрою как wont' fix
gharchive/issue
2023-06-29T10:06:11
2025-04-01T04:55:53.884148
{ "authors": [ "PilgrimViis", "ilya7am" ], "repo": "alor-broker/Astras-Trading-UI", "url": "https://github.com/alor-broker/Astras-Trading-UI/issues/1039", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
253995639
Fail fast if S3 bucket name is not configured in production Supersedes #159. The app will now fail fast with a KeyError if the AWS_S3_BUCKET_NAME environment variable is not set in the production environment. If the environment variable is not set in the development environment then a "null object" version of the S3Storage class is used so that uploads are a no-op. However, note that asset requests using the proxy_to_s3_via_nginx option will fail with a 500 Server Error via the S3Storage::NotConfiguredError exception. In the test environment, by default the Services.cloud_storage method returns an RSpec stub. This can be overridden by setting disable_cloud_storage_stub: true in the RSpec metadata for an example, context, or spec. I've updated the README to reflect these changes and indicate which environment variables are required in production. Fixes #154. @chrislo: Thanks. I'll get it merged.
gharchive/pull-request
2017-08-30T13:17:19
2025-04-01T04:55:53.905070
{ "authors": [ "floehopper" ], "repo": "alphagov/asset-manager", "url": "https://github.com/alphagov/asset-manager/pull/167", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1445479275
GUA-468 delete stack alerting WIP whilst I test How to test SNS Topic Alerts trigger Prerequisites Check the topic SQS DLQ exists: https://eu-west-2.console.aws.amazon.com/sqs/v2/home?region=eu-west-2#/queues/https%3A%2F%2Fsqs.eu-west-2.amazonaws.com%2F985326104449%2Faccount-mgmt-backend-UserAccountDeletionTopicDeadLetterQueue-gZpDthtNx75m Check an Alarm is present for the DLQ: https://eu-west-2.console.aws.amazon.com/cloudwatch/home?region=eu-west-2#alarmsV2:alarm/account-mgmt-backend-dev-UserAccountDeletionTopicDeadLetterQueueAlarm? Check that the it is not in the alarm state. Add an event to the queue and check for alarm gds aws di-account-dev -- aws sqs send-message --queue-url https://sqs.eu-west-2.amazonaws.com/985326104449/account-mgmt-backend-UserAccountDeletionTopicDeadLetterQueue-gZpDthtNx75m --message-body "Test the alarm system" Check that there is 1 message on the queue Check that the alarm has gone off in AWS console Check that you have recieved an email titled: ALARM: "account-mgmt-backend-dev-UserAccountDeletionTopicDeadLetterQueueAlarm" in EU (London) Inbox Once you are done Purge the DLQ There's a button on the top right next to Delete Delete Lambda DLQ Prerequisites Check the topic SQS DLQ exists: https://eu-west-2.console.aws.amazon.com/sqs/v2/home?region=eu-west-2#/queues/https%3A%2F%2Fsqs.eu-west-2.amazonaws.com%2F985326104449%2Faccount-mgmt-backend-DeleteUserServicesDeadLetterQueue-oRaAT30LQz35 Check an Alarm is present for the DLQ: https://eu-west-2.console.aws.amazon.com/cloudwatch/home?region=eu-west-2#alarmsV2:alarm/account-mgmt-backend-dev-DeleteDeadLetterQueueAlarm? Check that the it is not in the alarm state. Add an event to the queue and check for alarm gds aws di-account-dev -- aws sqs send-message --queue-url https://sqs.eu-west-2.amazonaws.com/985326104449/account-mgmt-backend-DeleteUserServicesDeadLetterQueue-oRaAT30LQz35 --message-body "Test the alarm system" Check that there is 1 message on the queue Check that the alarm has gone off in AWS console Check that you have recieved an email titled: ALARM: "account-mgmt-backend-dev-DeleteDeadLetterQueueAlarm" in EU (London) Inbox Once you are done Purge the DLQ There's a button on the top right next to Delete Test the integration Invoke the lambda with a purposefully dodgy set of params Aim here is to check that the DLQ gets written to if the lambda fails in any way. Starting with a quieted alarm and empty DLQ invoke: Visit the Lambda function in aws console. Select "Configure test event" and under template select "SNS Topic Notification". Save this and hit test. This will be a roughly valid shaped SNS event, but the message is just an example string which should fail. You should see a new item on the DLQ. Delete the lambda and write an SNS topic Aim here is to check that the DLQ gets written to if the lambda is not present Starting with a quieted alarm and empty topic DLQ. Go find the lambda in AWS console and delete it. Get the ARN for your SNS topic Either check in the console, starting here OR gds aws di-account-dev -- aws di-account-dev -- aws sns list-topics Fire an SNS event gds aws di-account-dev -- aws sns publish \ --topic-arn SNS_TOPIC_ARN_HERE \ --message '{ "user_id": "test-to-delete" }' You should see one Item on the Topic DLQ After some investigation we've resolved we can't fully test the SNS DLQ. @alex9smith and I are not sure why not... usual plans didn't pan out. We're keen not to get sidetracked on this. This PR improves alerting. We'll document out 🤔 elsewhere, and move on. So this is ready for review.
gharchive/pull-request
2022-11-11T13:29:35
2025-04-01T04:55:53.917065
{ "authors": [ "huwd" ], "repo": "alphagov/di-account-management-backend", "url": "https://github.com/alphagov/di-account-management-backend/pull/47", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1653946846
PYIC-2621 Change made to store clientOAuthSessionId with ipvSession. Proposed changes Store clientOAuthSessionId with ipvSession. What changed Update InitialiseIpvSessionHandler to store clientOAuthSessionId with ipvSession. Why did it change Issue tracking PYIC-2621 Checklists Environment variables or secrets [ ] No environment variables or secrets were added or changed [ ] Documented in the README [ ] Added to deployment repository [ ] Added to local startup repository Other considerations [ ] Update README with any new instructions or tasks Tested end to end journey in dev.
gharchive/pull-request
2023-04-04T13:56:11
2025-04-01T04:55:53.922028
{ "authors": [ "rpayal" ], "repo": "alphagov/di-ipv-core-back", "url": "https://github.com/alphagov/di-ipv-core-back/pull/734", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
90607983
Fix disappearing fields bug Brings in https://github.com/alphagov/digitalmarketplace-utils/pull/74 as well as some other fixes to the utils. This fixes a bug where filtering of the content persisted between requests, causing some fields which should have been shown to start disappearing (terminationCost was the example found by the smoke tests). Could we add a test for the scenario that shows how the issue occurred? Something like: Request a SaaS service Check that terminationCost is not shown Request SCS service Check that terminationCost is shown. @allait Turned your <ol> into code: b0044d1 @quis :+1: It's good to merge, just needs a rebase. @allait Rebased
gharchive/pull-request
2015-06-24T08:27:34
2025-04-01T04:55:53.925025
{ "authors": [ "allait", "quis" ], "repo": "alphagov/digitalmarketplace-admin-frontend", "url": "https://github.com/alphagov/digitalmarketplace-admin-frontend/pull/80", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
677514508
Replace banners-with-actions with confirm-action page https://trello.com/c/JhIzksR3/116-2-replace-banners-with-actions-with-confirm-action-page-pattern Withdrawing and deleting briefs now show a separate confirmation page. My review checklist I've: [ ] looked at the code [ ] tested locally [ ] run functional tests locally 🙌
gharchive/pull-request
2020-08-12T08:46:47
2025-04-01T04:55:53.927466
{ "authors": [ "domoscargin", "lfdebrux" ], "repo": "alphagov/digitalmarketplace-briefs-frontend", "url": "https://github.com/alphagov/digitalmarketplace-briefs-frontend/pull/338", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
184487233
Feature flag for new supplier flow For this story on Pivotal. This feature flag will render a new message and button before a supplier starts applying for a brief when turned on. It's currently only set in the base config and can be turned on in different environments when needed. More feature flags will be needed in the supplier app, and the api when the actual new flow is being developed. This flag is in the buyer app, and clicking the button to start an application takes you to the supplier app. Therefore this flag can only be used for this message, which is a bit weird. Feature flag not activated: Feature flag activated: Nice work 👍
gharchive/pull-request
2016-10-21T13:39:32
2025-04-01T04:55:53.929900
{ "authors": [ "Wynndow", "idavidmcdonald" ], "repo": "alphagov/digitalmarketplace-buyer-frontend", "url": "https://github.com/alphagov/digitalmarketplace-buyer-frontend/pull/394", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
272478173
Update indexing script Summary Update index-services script to have a default mapping (for services) to remain compatible after updates to the search-api/apiclient I might be missing something, but isn't the "index-services" script always going to index services? Why parameterise something that will always be services?
gharchive/pull-request
2017-11-09T08:32:44
2025-04-01T04:55:53.931761
{ "authors": [ "TheDoubleK", "samuelhwilliams" ], "repo": "alphagov/digitalmarketplace-scripts", "url": "https://github.com/alphagov/digitalmarketplace-scripts/pull/175", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
217560568
[#141943157]mark_definite_framework_results: make declaration "status" handling "special case" Story https://www.pivotaltracker.com/story/show/141943157 It was going to be ugly to get the frameworks to automagically generate an assessment schema that also included the status criteria, and what's more we already do "special case" handling of the service "status" field. also change included dos2 schema, no longer including status, to match this. Seems reasonable to me. :+1: Famous last words
gharchive/pull-request
2017-03-28T13:31:38
2025-04-01T04:55:53.934092
{ "authors": [ "TheDoubleK", "risicle" ], "repo": "alphagov/digitalmarketplace-scripts", "url": "https://github.com/alphagov/digitalmarketplace-scripts/pull/98", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
96899310
Fix page title on local petitions without a postcode If the user doesn't enter a postcode when searching from the home page or they navigate to the page directly then the page title helper doesn't have a count to pick which translation to use. For this case set the count to -1 so that the :other key is used from the translation. This is what you currently get on production: LGTM
gharchive/pull-request
2015-07-23T20:31:46
2025-04-01T04:55:53.935639
{ "authors": [ "alanth", "pixeltrix" ], "repo": "alphagov/e-petitions", "url": "https://github.com/alphagov/e-petitions/pull/354", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1106665158
Update Node.js to active LTS version What Update the project to use the latest active Long Term Support (LTS) version of Node.js for development, testing, and deployment. Why This project currently uses Node.js version 14, which has been in maintenance since last October (2021-10-19) and won't receive new features. Who needs to work on this Developer Who needs to review this Developer, Tech lead Done when [ ] Update .nvmrc [ ] Update package-lock.json to use lockfile v2 We've encountered some issues with this change on Mac M1 devices. They appear to related to node-sass, which we use in our tests and with gulp-sass. gulp-sass will not support M1 chips with the node-sass compiler. We've managed to get this working on M1 machines by deleting node_modules, then running: nvm use npm uninstall node-sass -g && npm cache clean -force && npm install node-sass nvm use npm install Current theories are that the issue might be related to installing with Node v14, then installing this change, and something going awry in the switch. We're going to test installing straight from this branch and see if a clean install works. We're then going to decide whether: We stick to Node 14 for a while We push the Node 16 update and document these steps for devs to troubleshoot any issues We look at switching to dart-sass Additionally, after updating to Node 16, metalsmith-tagcleaner generates a warning during npm install: npm WARN EBADENGINE Unsupported engine { npm WARN EBADENGINE package: 'metalsmith-tagcleaner@0.0.2', npm WARN EBADENGINE required: { node: '>=4.00' }, npm WARN EBADENGINE current: { node: 'v16.14.0', npm: '8.3.1' } npm WARN EBADENGINE } metalsmith is now being actively developed: https://github.com/metalsmith/metalsmith/issues/365 So it's possible we might be able to drop this extension or get an updated version at some point. Alternatively, we may want to look at alternatives to metalsmith. @domoscargin FWIW there is now an official @metalsmith/sass package you can use to replace metalsmith-sass (which uses node-sass under the hood). It uses Dart sass as in your proposed solution and config changes should be minimal.
gharchive/issue
2022-01-18T09:24:57
2025-04-01T04:55:53.942205
{ "authors": [ "domoscargin", "lfdebrux", "webketje" ], "repo": "alphagov/govuk-design-system", "url": "https://github.com/alphagov/govuk-design-system/issues/2021", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1702035434
Use the same version of marked in our build and our plugins What Sync marked version/source across: @metalsmith/markdown lib/extract-page-headings lib/get-macro-options lib/marked-renderer Why @metalsmith/markdown uses marked v4+ under the hood, and this does the heavy lifting of rendering our markdown code to HTML. However, we also directly depend on marked for some of our plugins which we use to define custom renderers and lexers. Recently, dependabot tried to bump this direct dependency to the new v5, which resulted in a swathe of deprecation warnings for the mangled and headerIds options. mangled We set mangled to false via @metalsmith/markdown, but not in our plugins, so even though we don't need mangling in these plugins, we were still deprecation warnings because we hadn't explicitly disabled it. headerIds We use this property to generate ids for our headings. I looked into use the plugin marked recommends instead, but it generated ESM-related errors: https://github.com/markedjs/marked-gfm-heading-id/issues/310 We want to avoid getting warnings or errors like these that only apply to one part of the build process. Who needs to work on this Developer Who needs to review this Developer Done when [ ] Our marked plugins and @metalsmith/markdown use the same version of marked [ ] We have a way to keep these in-sync (preferably by having each of the relevant modules use the same source marked) Since v1.9.0 @metalsmith/markdown provides a render option which you could use to force using a locally installed marked version: https://github.com/metalsmith/markdown#using-another-markdown-library. Since v1.8.0 the options have moved to engineOptions, see https://github.com/metalsmith/markdown/releases/tag/v1.8.0 (this warning is logged when metalsmith.env('DEBUG') contains @metalsmith/markdown*) I may consider making marked an (optional) peerDependency in the future due to the complications of their switch to extensions vs options (tracked in https://github.com/metalsmith/markdown/issues/70) Just a note to say marked@8.0.0 has been released so deprecated options have been removed https://github.com/alphagov/govuk-frontend/pull/4165 https://github.com/alphagov/govuk-design-system/pull/3123 Breaking changes Heading IDs from Metalsmith's marked@4.3.0 are removed in DesignSystemRenderer() using marked@8.0.0 We've seen this before, since in https://github.com/alphagov/govuk-design-system/pull/3002/commits/21787a8a593b5487d1ba0a64aed58edbb7cc8c5f we noted that jstransformer-marked still needed { smartypants: true } but elsewhere the option was logging deprecation warnings so needed marked-smartypants instead Using the suggestion above though we could bypass Metalsmith's marked@4.3.0 rendering with: const markdown = require('@metalsmith/markdown') const { Marked } = require('marked') // Use `marked@8.0.0` const marked = new Marked() metalsmith.use( markdown({ // Bypass `marked@4.3.0` render(source) { return marked.parse(source) } }) ) Deprecated options removed The renderer bypass would need to use the following packages instead of options: smartypants: true → marked-smartypants headerIds: true → marked-gfm-heading-id highlight → marked-highlight Since the previous options have now been removed (see known extensions list) Turns out { renderer: DesignSystemRenderer() } does override by design From the documentation: The Marked Pipeline For example, using marked.use({renderer}) would modify a render, whereas marked.use({extenstions: [{renderer}]}) would add a new renderer. But handling tokens (not strings) with named extensions lets each one keep their own new renderer: https://github.com/alphagov/govuk-design-system/pull/3130 Ready for review Looks like this is done 🎉
gharchive/issue
2023-05-09T13:25:06
2025-04-01T04:55:53.958464
{ "authors": [ "36degrees", "colinrotherham", "domoscargin", "webketje" ], "repo": "alphagov/govuk-design-system", "url": "https://github.com/alphagov/govuk-design-system/issues/2767", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
815440497
Add GOV.UK Frontend WTForms package Adding WTForms Widget package to generate GOV.UK compliant forms. Works in conjunction with the existing GOV.UK Frontend Jinja package. You can preview this change here: Built without sensitive environment variables with commit a6028b9c238bdb74011dcb91110ae63103ea8d1b https://deploy-preview-1532--govuk-design-system-preview.netlify.app Had a chat with @matthew-shaw on Slack about a couple of the contribution criteria which I was unsure this meets at the moment. We agreed that https://github.com/LandRegistry/govuk-frontend-wtf and https://pypi.org/project/govuk-frontend-wtf/ will be updated to: make explicit that this is a community resource that uses GOV.UK Frontend and is maintained by Land Registry add contribution and support info Thanks for reviewing @vanitabarrett. I've added what we discussed to the package in this pull request https://github.com/LandRegistry/govuk-frontend-wtf/pull/4 Would you mind confirming this meets the community contribution criteria before I merge and release an updated version of the package please? Thanks again. @matthew-shaw Thank you for making those changes, they look great to me! 💯 @vanitabarrett Thanks. Merged and released a new version of the package: https://pypi.org/project/govuk-frontend-wtf/ Hopefully all good to be added to the community resources page now 👍 FYI I've also applied all the same updates to the sister package GOV.UK Frontend Jinja here: https://pypi.org/project/govuk-frontend-jinja/ so they should both meet the community contribution criteria
gharchive/pull-request
2021-02-24T12:39:49
2025-04-01T04:55:53.965769
{ "authors": [ "govuk-design-system-ci", "matthew-shaw", "vanitabarrett" ], "repo": "alphagov/govuk-design-system", "url": "https://github.com/alphagov/govuk-design-system/pull/1532", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1141229104
Remove errors from phone numbers guidance Fixes #2022. Updates the 'links' section of our telephone numbers guidance. Changes Current guidance says: most desktop machines do not support calls - this is untrue so we're changing it most modern mobile browsers detect phone numbers and display them as links - as above Also, when users click on linked numbers, it can trigger a Skype setup process or the like. So we're also adding a sentence to tell users to bear this in mind. @owenatgov Thanks Owen, appreciate the speedy review! Hey @EoinShaughnessy! I'm not sure what's needed for a review, but I might hop in to second-guess the removal of that second line. For this line: most modern mobile browsers detect phone numbers and display them as links Should we just keep it in and update it to be more accurate? some modern mobile browsers detect phone numbers and display them as links Not sure if that'll just add more ambiguity about what to do - but hey, this is a pattern not a component. @davidc-gds Great suggestion, thanks! Will add in. 👍🏻 @calvin-lau-sig7 Does the update look ok to you? :)
gharchive/pull-request
2022-02-17T11:41:15
2025-04-01T04:55:53.970069
{ "authors": [ "EoinShaughnessy", "davidc-gds" ], "repo": "alphagov/govuk-design-system", "url": "https://github.com/alphagov/govuk-design-system/pull/2070", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1559762855
Draft: 'share user research' guidance Draft of some guidance to help teams share useful research findings about things in the Design System @christopherthomasdesign The PR to fix your Markdown code wrapping got merged by the way 👍 https://github.com/alphagov/govuk-design-system/pull/2563 Sorry @calvin-lau-sig7 due to the way Nunjucks renders before Markdown we need to use HTML for the link: Before {% set callout %} Share information responsibly. [You must get informed consent from your participants](https://www.gov.uk/service-manual/user-research/getting-users-consent-for-research) before you can share findings of any research they’ve participated in. {% endset %} After {% set callout %} Share information responsibly. <a href="https://www.gov.uk/service-manual/user-research/getting-users-consent-for-research">You must get informed consent from your participants</a> before you can share findings of any research they’ve participated in. {% endset %} Otherwise it looks like Markdown is injecting <p> tags where it shouldn't @calvin-lau-sig7 There was one more <p> wrapping the govukWarningText() <div> so fixed it 👍 -<p> {{ govukWarningText({ html: callout, iconFallbackText: "Warning" }) }} -</p> All passing now and deployment is working again
gharchive/pull-request
2023-01-27T13:20:48
2025-04-01T04:55:53.973916
{ "authors": [ "christopherthomasdesign", "colinrotherham" ], "repo": "alphagov/govuk-design-system", "url": "https://github.com/alphagov/govuk-design-system/pull/2559", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1418339722
Simplify nodemon configuration We have been seeing some weird bugs with watching the .env file, which this change fixes by simplifying how we use nodemon. Apart from .env, the only files the user might change that we want to restart the app automatically for are now in their app folder. We can therefore simplify our code by restricting the watch directives to that folder, and remove a bunch of ignore filters. This should also result in nodemon needing to watch many fewer files, which should hopefully mean more reliable performance generally, apart from fixing the dotenv bug. This might be a breaking change for users who have put files their prototype files outside the app folder, but in the past we've tried to encourage them not to do that so it shouldn't affect too many users. @BenSurgisonGDS would you mind checking if this PR fixes things on your machine so it now restarts the app when .env changes like it should? Hmm, so it seems that this PR doesn't fix things on Ben's machine. I tried getting a test working in CI (see commit ), but that didn't go anywhere either :( I will test on a different Linux machine on Monday, in case it's just Ben's setup, otherwise we should probably just merge this fix and then try fixing the Linux issue later. I can reproduce the same behaviour I was seeing on macOS (both the issue and the fix) on a machine running a native install of Linux and Node.js 14, and a machine running a native install of Windows and Node.js 16. I'm not sure why @BenSurgisonGDS was seeing different behaviour, but I'm going to call this issue fixed.
gharchive/pull-request
2022-10-21T13:25:22
2025-04-01T04:55:53.977583
{ "authors": [ "lfdebrux" ], "repo": "alphagov/govuk-prototype-kit", "url": "https://github.com/alphagov/govuk-prototype-kit/pull/1709", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
147721708
Staff changes Remove Matt and add Jon :+1: go merge
gharchive/pull-request
2016-04-12T11:40:09
2025-04-01T04:55:53.978749
{ "authors": [ "deanwilson", "surminus" ], "repo": "alphagov/govuk-terraform-provisioning", "url": "https://github.com/alphagov/govuk-terraform-provisioning/pull/39", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
548181923
Upgrade Ruby to 2.7.0 Upgrade Ruby to 2.7.0, see commits for more details. https://trello.com/c/XoJF36Pn/1663-8-upgrade-to-ruby-27 Closing. Sorry for the spam across all the apps! It turns out that there's a key deprecation which is causing a load of errors in gems that we use. The gem authors (including Rails) are working hard to fix their software, but there's no point us using it until they've done that or we'll get a tonne of warnings (and turning them off is bad practice). Also, it's possible that Rails may only put their fixes into Rails 6 - and in most of the cases we're still running Rails 5. Platform Health is going back to the drawing board on this one, because it may be that we want to upgrade every app to Rails 6 first.
gharchive/pull-request
2020-01-10T16:46:33
2025-04-01T04:55:53.980937
{ "authors": [ "issyl0" ], "repo": "alphagov/manuals-publisher", "url": "https://github.com/alphagov/manuals-publisher/pull/1556", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
90633532
Make vault decryption more quiet Set gpg to be quiet while decripting the vault. This prevents printing all the recipients of the vault everytime we run ansible. The same option has been already added to tsuru-ansible works, merging
gharchive/pull-request
2015-06-24T10:17:59
2025-04-01T04:55:53.982379
{ "authors": [ "keymon", "mtekel" ], "repo": "alphagov/multicloud-deploy", "url": "https://github.com/alphagov/multicloud-deploy/pull/28", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
224206390
PP-1717 username in api urls cleanup WHAT Modifies pay-adminusers service to use only external id in api endpoints. Before that we were querying for username and externalId in api endpoints. This PR replaces username with externalId from the endpoints and services (removes the username condition) HOW Modified UserResource and removed is_new_api_request GET parameter Modified service and dao layers and removed the username condition Modified tests and documentation LGTM 👍
gharchive/pull-request
2017-04-25T17:00:09
2025-04-01T04:55:53.988470
{ "authors": [ "DanailMinchev", "mrlumbu" ], "repo": "alphagov/pay-adminusers", "url": "https://github.com/alphagov/pay-adminusers/pull/68", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
370592144
[WIP] Proposal for new validation pattern in connector This commit illustrates how to use Dropwizard's built in validation functionality to validate incoming requests. It is quite powerful, and can do most of (probably all) the things we require of it. In this commit I have taken one example endpoint /v1/frontend/charges/{chargeId}/status, and refactored validation for the PUT method. I do not necessarily think that the validation logic is right (in particular the constraint on the value of new_status), but I have not changed it at this stage, just the implementation. In order to change as little as possible, I have overridden the dropwizard default exception mapper (see https://www.dropwizard.io/0.9.2/docs/manual/validation.html) , so as to be able to add my own ConstraintViolationExceptionMapper, which allows me to map ConstraintViolationExceptions to 400 status code. The returned message is slightly different from before, but is not any worse IMHO. In tandem with this change to the validation, I have changed how it is tested. Instead of spinning up a whole application and db, I have used the ResourceTestRule that provides a testable resource which can be interacted with as if it is an http server, whilst allowing mocking of it's dependencies. This allows tests to complete in ~400 ms as opposed to ~16 seconds before. To do this I again had to do some fiddling with the exception mapper config, (see https://www.dropwizard.io/1.3.5/docs/manual/testing.html) which is the reason for the brillaintly named ResourceTestRuleWithCustomExceptionMappersBuilder, which encapsulates the needed changes. 👍 for using ResourceTestRule!! Generally LGTM; no issues from me.
gharchive/pull-request
2018-10-16T12:40:59
2025-04-01T04:55:53.993564
{ "authors": [ "mrlumbu", "oswaldquek" ], "repo": "alphagov/pay-connector", "url": "https://github.com/alphagov/pay-connector/pull/748", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1138581129
Updates list of countries with converted headers Trello: https://trello.com/c/ZegaI995 What's changed? Update the list of countries in the countries_with_content_headers_converted method ready for go live. The full list is here: https://foreign-travel-spike.herokuapp.com/api/completed_locations. Why? More countries have had their headers converted so we are able to point user to more specific content. Expected changes Before After ⚠️ This repo is Continuously Deployed: make sure you follow the guidance ⚠️ @Rosa-Fox @gclssvglx That sneaky Slovenia! It must have just been converted 😱. It's been added now. haha I thought it might have been! We can double check them tomorrow before launching too.
gharchive/pull-request
2022-02-15T11:48:01
2025-04-01T04:55:53.998297
{ "authors": [ "Rosa-Fox", "leenagupte" ], "repo": "alphagov/smart-answers", "url": "https://github.com/alphagov/smart-answers/pull/5780", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
147034160
[CLOSED] RSpec3 Issue by bestie Monday 28 July 2014 at 13:36 GMT Originally opened as https://github.com/alphagov/specialist-publisher/pull/193 Cucumber updagraded to 1.3.16 rspec-rails gem removed Disable should syntax bestie included the following code: https://github.com/alphagov/specialist-publisher/pull/193/commits Comment by kalleth Monday 28 July 2014 at 15:35 GMT lgtm :thumbsup:
gharchive/issue
2016-04-08T20:41:07
2025-04-01T04:55:54.003082
{ "authors": [ "benilovj" ], "repo": "alphagov/specialist-publisher-rebuild", "url": "https://github.com/alphagov/specialist-publisher-rebuild/issues/193", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
236858760
Add related item to business finance support scheme Adds the 'Apply for a start up loan' start page as related to the business finance support finder. See https://trello.com/c/xm4W1SzB 👍 https://www.gov.uk/apply-start-up-loan doesn't exist yet (it's in draft - https://draft-origin.publishing.service.gov.uk/apply-start-up-loan), so it will not show up on the page yet.
gharchive/pull-request
2017-06-19T11:21:53
2025-04-01T04:55:54.005065
{ "authors": [ "andrewgarner", "tijmenb" ], "repo": "alphagov/specialist-publisher", "url": "https://github.com/alphagov/specialist-publisher/pull/1043", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
199591326
PP: Add 'permit eligibility requirements' to things needed on start page It's already on the actual start page, just need to add it to the bullet list Thanks @kateiw 🐐 💯 🙏
gharchive/issue
2017-01-09T16:06:41
2025-04-01T04:55:54.005966
{ "authors": [ "sanjaypoyzer" ], "repo": "alphagov/verify-local-patterns", "url": "https://github.com/alphagov/verify-local-patterns/issues/30", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
311997325
DO NOT MERGE: BAU: Use later version of hub-saml and verify-saml-libs with fixes The newer version of verify-saml-libs have fixes to give metadata resolver unique names when adding them to the metric. There are issues with this set of changes.
gharchive/pull-request
2018-04-06T14:28:22
2025-04-01T04:55:54.006885
{ "authors": [ "IreneLau-GDS" ], "repo": "alphagov/verify-matching-service-adapter", "url": "https://github.com/alphagov/verify-matching-service-adapter/pull/46", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
200402379
Add form to tag Education content to beta taxonomy For documents belonging to DfE and SFA, show a button to tag to the Finding Things education taxonomy. This allows stakeholders to keep tagging content to the taxonomy during beta. Look & Feel Trello: https://trello.com/c/miTSiYfk/409-create-tagging-interface-for-new-taxonomy-in-whitehall Paired with @MatMoore and @klssmith from @alphagov/team-gov-uk-finding-things This looks really good! I moved the mixin from Editions to the various models that include the Edition::Organisations concern, because not every Edition has the organisations association. As far as I can tell, the only exception is CorporateInformationPage, which isn't relevant to the beta. This is missing a feature flag as discussed offline This can only be deployed after implementation of feature flag and the deploy of puppet with the new environment variable (feature flag) https://github.com/alphagov/govuk-puppet/pull/5361
gharchive/pull-request
2017-01-12T15:48:31
2025-04-01T04:55:54.011616
{ "authors": [ "Davidslv", "MatMoore", "tijmenb" ], "repo": "alphagov/whitehall", "url": "https://github.com/alphagov/whitehall/pull/2963", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1602930646
Add guidance for images tab on Case Studies, Speeches and Fatality Notices Trello ticket link What Adds guidance for using the new images tab on Case Studies, Speeches and Fatality Notices for users who have the Preview images update permission. Why These content types only support image attachments - not file attachments - and hence do not benefit from the recent additions to guidance for content types which support both image and file attachments. Screenshots Case Studies Fatality Notices Speeches ⚠️ This repo is Continuously Deployed: make sure you follow the guidance ⚠️ @dnkrj yea you're right, I've just pushed some changes to correct this
gharchive/pull-request
2023-02-28T12:04:02
2025-04-01T04:55:54.017719
{ "authors": [ "jyoung-gds" ], "repo": "alphagov/whitehall", "url": "https://github.com/alphagov/whitehall/pull/7387", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
54471795
Add support for OpenEncryptedFileRaw and related functions. Issue by alphaleonis from Friday Jan 02, 2015 at 19:33 GMT Originally opened as https://github.com/alphaleonis/AlphaFS/issues/3 Encryption/decryption happens automatically with most of the regular api calls if the right cert is installed. However there are situations when if the drive was taken out of the computer for forensic analysis, you would need to access the encrypted files. In that situation there is usually a delegated decryption user's certificate is stated in the encrypted files metadata, usually it is a active directory primary administrator who allowed the EFS to be used by regular users. So you would need to acquire the cert from that admin to the computer which will try to decrypt affected files. I think think we could add some extra methods when the right Windows license with EFS lands on one of our hands. Comment by alphaleonis from Friday Jan 02, 2015 at 19:34 GMT Copied from CodePlex. Comments: lambchops wrote Feb 14, 2013 at 3:27 AM Downloaded src to have a look at implementing these functions.... but a little lost as it appears that in the releases folder, there is 1.0 and not 1.5. Looking at the changes log it seems that you are up to 1.6? Functions like File.OpenBackupRead do not seem to be there. Please advise :) thnx Yomodo wrote Feb 14, 2013 at 10:37 AM Hi, Just download and use the latest development build to start implementing. Perhaps the BackupFileStream() and/or BackupStreamInfo() classes have what you need. Regards, Yomodo
gharchive/issue
2015-01-15T16:41:39
2025-04-01T04:55:54.022782
{ "authors": [ "alphaleonis" ], "repo": "alphaleonis/AlphaFS.New", "url": "https://github.com/alphaleonis/AlphaFS.New/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
356433768
Cannot save external access to GUI I have used the cmd: HOST=0.0.0.0 PORT=9000 crontab-ui Where host = my server IP but as soon as I exit SSH, its not saving and I am unable to access crontab-ui What do you mean? Just like any program, If it is running in the foreground, it wouldn't work once you close the shell. Either run it as a service or use tmux/pm/screen. Hi, apologies I'm new to this and only know the basics. How would I run Crontab-ui as a service, please? checkout http://pm2.keymetrics.io/ Any chance you can help with how to use that? I'm not sure why you want to run crontab-ui as a service it creates all the required files. Are you clicking the "Save to crontab" button? Maybe I have the wrong idea of the crontab-ui. The problem I have is, how do I access crontab-ui via the GUI without having to SSH in each time and run it? You should not have to SSH at all. 1- Open terminal and type crontab-ui and press enter. You should see something like this: Node version: 8.10.0 Crontab UI is running at http://127.0.0.1:8000 The http://127.0.0.1:8000 should actually be a hyperlink so when you hover your mouse over it you can see it get underlined and the cursor changes to a hand or something. Click on it and your browser will open with the crontab-ui GUI. Yes, that it what comes up when i type crontab-ui but if you browse to 127.0.0.1:8000, my PC is trying to access itself (127.0.0.1 is a loopback) The server running crontab-ui is a hosted server so is not local to me can someone help please?
gharchive/issue
2018-09-03T09:50:41
2025-04-01T04:55:54.071108
{ "authors": [ "JayBird707", "alseambusher", "charliejsford" ], "repo": "alseambusher/crontab-ui", "url": "https://github.com/alseambusher/crontab-ui/issues/87", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
253607263
Lifetime of scene::Children bound to parent Currently, one can't own the children of nodes due to incorrect lifetimes. fn process(node: gltf::Node) { struct Item { node: gltf::Node, } let mut stack = vec![Item { node }]; while let Some(item) = stack.pop() { for child in item.node.children() { // Error: lifetime of `child` bound to `item.node`. stack.push(Item { node: child }); } } } Workaround fn clone_child<'a>( gltf: &'a gltf::Gltf, child: &gltf::Node, ) -> gltf::Node<'a> { gltf.nodes().nth(child.index()).unwrap() }
gharchive/issue
2017-08-29T10:37:09
2025-04-01T04:55:54.080509
{ "authors": [ "alteous" ], "repo": "alteous/gltf", "url": "https://github.com/alteous/gltf/issues/85", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1943909546
🛑 roothouse-1 is down In 1f4f763, roothouse-1 (https://rhstatus-proxy.alterine0101.id/?deviceId=roothouse-1) was down: HTTP code: 404 Response time: 749 ms Resolved: roothouse-1 is back up in 318041d.
gharchive/issue
2023-10-15T14:25:22
2025-04-01T04:55:54.086334
{ "authors": [ "1010bots" ], "repo": "alterine0101/roothouse-status", "url": "https://github.com/alterine0101/roothouse-status/issues/1834", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1167959683
🛑 roothouse-2 is down In 60d537c, roothouse-2 (https://rhstatus-proxy.alterine0101.id/?deviceId=roothouse-2) was down: HTTP code: 404 Response time: 259 ms Resolved: roothouse-2 is back up in a39a307.
gharchive/issue
2022-03-14T06:49:27
2025-04-01T04:55:54.088951
{ "authors": [ "reinhart1010" ], "repo": "alterine0101/roothouse-status", "url": "https://github.com/alterine0101/roothouse-status/issues/977", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
491966203
Beta lower than one is it right to have beta lower than one multiplied by KLd? Good question - this is more a question of what helps empirically. E.g. in some work annealing beta helps: https://arxiv.org/abs/1511.06349 It probably depends on the problem; it's easy to try. However, leaving beta lower than one is technically incorrect if the goal is to have a well-defined loss function that is a lower bound on the evidence. Thank you @altosaar , when I have beta = 1 my elbo doesnot decrease , but when I have beta lower than one it decrease, Where I have made mistake you think , do you think writing loss function in a way you wrote differes results? I wrote simple KLD loss + reconstruction loss , but you wrote in a different form Can you write the LaTeX form of your loss please? And do you mean your elbo increases? Can you plot both terms in the ELBO and try annealing beta? On Wednesday, September 11, 2019, Ivamcoder notifications@github.com wrote: Thank you @altosaar https://github.com/altosaar , when I have beta = 1 my elbo doesnot decrease , but when I have beta lower than one it decrease, Where I have made mistake you think , do you think writing loss function in a way you wrote differes results? I wrote simple KLD loss + reconstruction loss , but you wrote in a different form — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/altosaar/variational-autoencoder/issues/22?email_source=notifications&email_token=ABISE7GCY4UDOKYHJ3DZVP3QJCOA7A5CNFSM4IVOCL32YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOD6NSLUY#issuecomment-530261459, or mute the thread https://github.com/notifications/unsubscribe-auth/ABISE7AOY5D4VJIWX6UPSKDQJCOA7ANCNFSM4IVOCL3Q . @altosaar , My plot shows that for beta=1 my results is worse than Beta lower than 1. and loss function I have used is `import numpy as np import torch.nn.functional as F """ KLD DIVERGENCE""" def kl_divergence(mu, logvar): batch_size = mu.size(0) assert batch_size != 0 klds = -0.5 * (1 + logvar - mu.pow(2) - logvar.exp()) total_kld = klds.sum(1).mean(0, True) dimension_wise_kld = klds.mean(0) mean_kld = klds.mean(1).mean(0, True) return total_kld, dimension_wise_kld, mean_kld """ RECONSTRUCTION LOSS """ def reconstruction_loss(x, x_recon, ID): batch_size = x.size(0) assert batch_size != 0 if ID == 'bernoulli': recon_loss = F.binary_cross_entropy_with_logits(x_recon, x, size_average=False).div(batch_size) elif ID == 'gaussian': recon_loss = F.mse_loss(x_recon, x, size_average=False).div(batch_size) else: recon_loss = None return recon_loss """ ANNEAL REGULARIZATION """ def anneal_reg(step, anneal): k = anneal.K x0 = anneal.X0 ID = anneal.ID if ID == 'logistic': return float(1 / (1 + np.exp(-k * (step - x0)))) elif ID == 'linear': return min(1., max(0., step/x0)) else: return 1.0 ` Do you think if I wrote my loss function like you , Would I get different results? Thanks, that's interesting. Were you able to write the LaTeX for the ELBO? That will help us figure out if we're talking about the same thing. On Wednesday, September 11, 2019, Ivamcoder notifications@github.com wrote: Do you think if I wrote my loss function like you , Would I get different results? — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/altosaar/variational-autoencoder/issues/22?email_source=notifications&email_token=ABISE7A7ZBPSKZBDB2KUF4TQJGFOJA5CNFSM4IVOCL32YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOD6QJJ5Q#issuecomment-530617590, or mute the thread https://github.com/notifications/unsubscribe-auth/ABISE7FY3ZV7SP4RFJ4MP4TQJGFOJANCNFSM4IVOCL3Q . Probably best to start from annealing beta, with anneal_w removed. It will be easier to understand whether annealing beta helps your task. Feel free to follow up with me about this via email. On Fri, Sep 13, 2019 at 2:05 AM Ivamcoder notifications@github.com wrote: Shouldn't beta be the thing being annealed here? Or are you annealing anneal_w as well? beta took different constant values beta=0.01, 0.1, 1 multiplied by anneal_w , in fact beta change the final value of anneal_w, for example if beta=1 , the factor of KLD initialize from zero to 1, if beta=0.1 the KLD factorinitializes from zero to 0.1. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/altosaar/variational-autoencoder/issues/22?email_source=notifications&email_token=ABISE7BVNAICTN3LRD6V6TTQJMURDA5CNFSM4IVOCL32YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOD6UBVZQ#issuecomment-531110630, or mute the thread https://github.com/notifications/unsubscribe-auth/ABISE7HQGWAIGA4PCQ5R5CDQJMURDANCNFSM4IVOCL3Q .
gharchive/issue
2019-09-11T01:09:57
2025-04-01T04:55:54.125197
{ "authors": [ "Ivamcoder", "altosaar" ], "repo": "altosaar/variational-autoencoder", "url": "https://github.com/altosaar/variational-autoencoder/issues/22", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
188979939
Query Document Validation As described in the GraphQL Working Draft, October 2016: Operations [x] Operation Name Uniqueness (5.1.1.1) [x] Lone Anonymous Operation (5.1.2.1) Fields [x] Field Selections in Scope (5.2.1) [ ] Field Selection Merging (5.2.2) [x] Leaf Field Selections (5.2.3) Arguments [x] Argument Names in Scope (5.3.1) [x] Argument Uniqueness (5.3.2) [x] Argument Value Type Correctness (5.3.3.1) [x] Required Non-Null Arguments (5.3.3.2) Fragments [x] Fragment Name Uniqueness (5.4.1.1) [x] Fragment Spread Type Existence (5.4.1.2) [x] Fragments only on Composite Types (5.4.1.3) [x] Fragments must be used (5.4.1.4) [x] Fragment Spread Target Existence (5.4.2.1) [x] Fragment Spreads must not form cycles (5.4.2.2) [x] Fragment Spread Possible in Scope (5.4.2.3) [x] Object Spread in Object Scope (5.4.2.3.1) [x] Abstract Spread in Object Scope (5.4.2.3.2) [x] Object Spread in Abstract Scope (5.4.2.3.3) [x] Abstract Spread in Abstract Scope (5.4.2.3.4) Values [x] Input Object Field Uniqueness (5.5.1) Directives [x] Directives are defined (5.6.1) [x] Directives are in valid locations (5.6.2) [x] Directives are unique per location (5.6.3) Variables [x] Variable Uniqueness (5.7.1) [x] Variable Default Values are correctly typed (5.7.2) [x] Variables are Input Types (5.7.3) [x] All Variable Uses Defined (5.7.4) [x] All Variables Used (5.7.5) [x] All Variable Usages are allowed (5.7.6) As of [alumbra/validator "0.1.0"] all validations given in the GraphQL Working Draft, October 2016, are implemented.
gharchive/issue
2016-11-13T16:42:02
2025-04-01T04:55:54.157238
{ "authors": [ "xsc" ], "repo": "alumbra/alumbra", "url": "https://github.com/alumbra/alumbra/issues/2", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
728885435
RK-G210S Здравствуйте! А возможно ли как-то добавить чайник RK-G210S? Думаю, будет работать. Выберите ближайший, например RK-G211S. Пробовал все выбирать, но не подключается. Пишет, что последнее найденное устройство bluetooth - RK-G210S. Всё работает, в настройках надо ещё имя чайника прописать - RK-G210S. Я единственное в новой прошивке не понял, как выключить мигающую подсветку, когда выключен чайник? А подсветка была включена из шлюза? Я прописывал имя "Kettle" и не смог подключится. Сейчас на работе попробую с чайником 200 модели, а вечером ещё раз дома с 210 буду пробовать... Вообщем к 200 подключился, но управлять ни подсветкой и чайником не хочет! В HA прокинулся без проблем. Управление ни из веб интерфейса, ни из HA не работает! Только уровень сигнала показывает. И ещё точмозит веб интерфейс... В чём может быть проблема??? Тип чайника установлен в 200? А подсветка была включена из шлюза? Если включить и потом выключить подсветку из шлюза, все нормально выключается? Вроде недавно подсветку правил, вроде бы работает. Попробуйте самую последнюю версию. Подсветка (ночник) работает нормально, включается и управляется через Web и в HA. Я про другую подсветку, которая плавно загорается и гаснет в цвет температуры воды, когда чайник выключен и стоит на подставке. Я прописывал имя "Kettle" и не смог подключится. Сейчас на работе попробую с чайником 200 модели, а вечером ещё раз дома с 210 буду пробовать... У меня тоже rk-g210s, если выбрать 211 или 240, то всё работает. Пришёл домой, всё настроил с 210, работает отлично! И интерфейс не тормозит! Видимо на работе что-то с сетью странное... Спасибо всем! Ещё вопрос, а зачем настройки часового пояса? Или это для мультиварки? В последних версиях шлюза на чайник примерно раз в минуту отправляется время. Читал, что чайник с установленным временем как раз и позволяет оценивать температуру воды. Вот только где это отключить, я не знаю, думаю, что в родном приложении должна быть такая опция. Часовой пояс нужен для установки времени. в то же кофеварке, розетке и чайниках 211 и 240. Но, кроме того, шлюз может быть с экраном, на котором как раз часы, температура дома и на улице, а также на выходе котла, некоторые параметры устройств и снимок с камеры. отправляется время. Читал, что чайник с установленным временем как раз и позволяет оценивать температуру воды. Вот только где это отключить, я не знаю, думаю, что в родном приложении должна быть такая опция. Шлюз управлять Ясно! Спасибо! Теперь управление чайником отчилно работает, а то мой RP4 до него с коридора через раз добивает своим блютузом)) Подсветка (ночник) работает нормально, включается и управляется через Web и в HA. Я про другую подсветку, которая плавно загорается и гаснет в цвет температуры воды, когда чайник выключен и стоит на подставке. Мигающая подсветка, это, если включена функция Световая индикация. Зайдите в настройки родного приложения Redmon и выключите функцию Световая индикация, предварительно выключите шлюз, а после того как выключите эту функцию, включите шлюз и подсветка перестанет мигать. Мигающая подсветка, это, если включена функция Световая индикация. Зайдите в настройки родного приложения Redmon и выключите функцию Световая индикация, предварительно выключите шлюз, а после того как выключите эту функцию, включите шлюз и подсветка перестанет мигать. Спасибо, помогло. Ещё заметил, что в родном приложении режим ночник светит в 2 раза ярче, чем через шлюз. Максимальная подсветка через шлюз это где-то половины яркости если управлять в родном приложении Добавил яркость подсветки в релиз 2020.10.27, нужно проверить.
gharchive/issue
2020-10-24T21:56:35
2025-04-01T04:55:54.166609
{ "authors": [ "BertTomsk", "Den-88", "DivanX10", "alutov" ], "repo": "alutov/ESP32-R4sGate-for-Redmond", "url": "https://github.com/alutov/ESP32-R4sGate-for-Redmond/issues/11", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1209655600
Needed to change role name podman to alvistack.podman This role is awesome but for the "quick start" i needed to change the role name podman to alvistack.podman in molecule/default/converge.yml. Is this an error or did I make a mistake? Good question with some background history: In the old bad day before Ansible Galaxy, install Ansible Role with GIT submodule looks much usable for me During the early day with Ansible Galaxy, we are not enforced with FQCN as https://ansible-lint.readthedocs.io/en/latest/default_rules/#fqcn-builtins now a day I didn't use any external Ansible Role across my ecosystem, so no FQCN with alvistack.* is required Now a day Ansible Galaxy handle both Role and Collection (e.g. https://galaxy.ansible.com/alvistack/kube_master and https://galaxy.ansible.com/alvistack/kubernetes) so naming with FQCN looks much important. As https://github.com/alvistack/ansible-role-podman/blob/master/.gitlab-ci.yml#L8-L12, I will symlink the Role as $HOME/.ansible/roles/podman before running CI with my self-hosted GitLab Runner Anyway, thank you for your suggestion, and I will give some effort to update all of my Ansible Role / Collection / Docker /Vagrant with Ansible FQCN ;-) Ah, I see. Thanks for your quick reply :smile: @g3n35i5 ok let see how its going on: https://github.com/alvistack/ansible-role-podman/commit/0b7641e3ba5e958cded57b843da7edb789b2efbc @g3n35i5 finally it result as 150+ repo need update across the ecosystem~~~ Thank you very much for pushing this reasonable fix, which delay for quite a while due to my laziness :-( @g3n35i5 some follow up issue for Ansible Collection: in case roles are managed as git submodule under Ansible Collection, it should be rename as "roles/podman" only without the namespace. This is because Ansible Collection contents, e.g. plugin and module and roles, are already prefix with alvistack.kubernetes. so here podman become alvistack.kubernetes.podman. Therefore we couldn't name it as roles/alvistack.podman else become alvistack.kubernetes.alvistack.podman and so invalid :-(
gharchive/issue
2022-04-20T13:01:14
2025-04-01T04:55:54.181309
{ "authors": [ "g3n35i5", "hswong3i" ], "repo": "alvistack/ansible-role-podman", "url": "https://github.com/alvistack/ansible-role-podman/issues/5", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1906137284
🛑 zhibao.itove.com is down In febe52b, zhibao.itove.com (https://zhibao.itove.com) was down: HTTP code: 0 Response time: 0 ms Resolved: zhibao.itove.com is back up in 0855b6e after 12 minutes.
gharchive/issue
2023-09-21T05:26:03
2025-04-01T04:55:54.208216
{ "authors": [ "alzee" ], "repo": "alzee/up", "url": "https://github.com/alzee/up/issues/2426", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
970886949
Initial setup instructions missing a step for iOS I'm on a M1 Mac. It took me a while to discover that I need to run "dart pub get" first to before the .pub-cache directory was created. Then mv ~/Downloads/TensorFlowLiteC.framework ~/.pub-cache/hosted/pub.dartlang.org/tflite_flutter-0.9.0/ios/ worked and I succeeded in build and running test app on a connected iPhone. thank you. This really heps Stumbled over this comment after hours of trying... Thank you a ton, finally my app is running on iOS!
gharchive/issue
2021-08-14T10:57:47
2025-04-01T04:55:54.214508
{ "authors": [ "CaptainDario", "YupengLin", "hapasa" ], "repo": "am15h/tflite_flutter_plugin", "url": "https://github.com/am15h/tflite_flutter_plugin/issues/147", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2116883204
🛑 AllTyres2U is down In fa16235, AllTyres2U (https://alltyres2u.com.au) was down: HTTP code: 403 Response time: 937 ms Resolved: AllTyres2U is back up in e8922dd after 10 minutes.
gharchive/issue
2024-02-04T04:50:02
2025-04-01T04:55:54.217054
{ "authors": [ "am8-badams" ], "repo": "am8-badams/upptime", "url": "https://github.com/am8-badams/upptime/issues/103", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
200150393
License Great sample! What is the license for this code? Thanks. Sorry I didn't see this, I accidentally had notifications disabled. Just added MIT license.
gharchive/issue
2017-01-11T17:03:07
2025-04-01T04:55:54.220539
{ "authors": [ "amacneil", "robinbobbitt" ], "repo": "amacneil/git-banish-large-files", "url": "https://github.com/amacneil/git-banish-large-files/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2326718371
Redlocate endpoint the /homepage endpoint is currently located in the /error controller, we should move this to its own homepage controller class issue closed. no notes
gharchive/issue
2024-05-30T23:46:22
2025-04-01T04:55:54.224769
{ "authors": [ "amamolo9986" ], "repo": "amamolo9986/The-Porsche-Project", "url": "https://github.com/amamolo9986/The-Porsche-Project/issues/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2456618326
How to send the language (locale) code to Fetch operations processed by the Client Side? Description I've done all the checks and I've gone through all the documents in detail, but I can't find a solution. I think there's something missing here. Problem 1: I want to fetch from the client side (when the document is created). For example, the user will submit a form to the API. I want to create a global fetch function for such later fetch operations and include the locale information in the header of this function so that I don't have to call the useLocale hook all the time. But it seems that it is not possible to access locale information from outside Next-Intl. Example: In this example I want to create a function that I can use globally and I want to include the locale information in this function. (I can send this in Component. But this causes me to run the uselocale() hook on every fetch operation. This is also not useful) const request = async (method, path, data = {}) => { const options = { Method, headers: { headers, // "x-locale": *** /** I want to add locale information globally here*/ }, body: JSON.stringify(data), }; const respose = await fetch(process.env.NEXT_PUBLIC_API_URL + path, options); return await respose.json(); }; const onFavorite = async () => { const resrult = await request("POST", "favorite", {id: 4566}) } <Button.favorite onClick={onFavorite}> Please tell me how to pass locale information outside the Component of the request() function. Issue 2: I want to access locale information from the server side, but the problem is that x-locale information is sent in server side requests. This information is not sent in custom requests on the client side. Also getLocale() does not get the locale code in special requests. Example: import { getTranslations, getLocale } from "next-intl/server"; const locale = getLocale(); const t = await getTranslations({ locale, namespace: "home" }); I can't get the locale information from the server side with getLocale. Intl already generates a locale information. I don't understand why this is empty by the API. I wonder why this happens? Problem 3: In some translation files I want to store an array. For example I have a static list of items. Since I am not going to update this list, I want to take it as an array and use it with map. However, I realized that it is not possible to access it as raw, even though I have examined the documentation many times. Is this feature not available? (It is a must-have feature) Respectfully Verifications [X] I've verified that the problem I'm experiencing isn't covered in the docs. [X] I've searched for similar, existing issues on GitHub and Stack Overflow. [X] I've compared my app to a working example to look for differences. Mandatory reproduction URL https://next-intl-docs.vercel.app/docs/getting-started/app-router/with-i18n-routing Reproduction description Next-Intl should be developed further. It is easy to set up and the documentation clearly answers everything. But it needs to be improved. The user should be given more access. Expected behaviour 1 - It is necessary to get the Array/Object array in the translation file in raw format and use it. 2 - The server should be able to get the locale information without any additional settings. 3 - Functions/global functions other than Component should have access to locale and translation information. Problem 1 The locale can only be read during render, via useLocale() (or await getLocale() in Server Components). If it's a hassle to read the locale for every HTTP call you're making from the client side, you can consider storing it in React Context. This would enable usage like this in components: const fetch = useFetcher(); async function onFavorite() { const result = fetch('POST', 'favorite', {id: 4566}); } In this case you can use this opportunity to share other options, like your API URL as well. Libraries like React Query have this behavior built-in by allowing to configure a global fetcher. Problem 2 If you're sending the locale consistently via a x-locale request header, then you can read it accordingly via headers(). Problem 3 See arrays of messages in the docs. I'm going to move this to a discussion since this is a usage question and not a bug report.
gharchive/issue
2024-08-08T21:02:25
2025-04-01T04:55:54.235503
{ "authors": [ "amannn", "salihandic" ], "repo": "amannn/next-intl", "url": "https://github.com/amannn/next-intl/issues/1243", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1920843062
Module '"next-intl/server"' has no exported member 'getRequestConfig' Description I use Next 13.5.3 I follow this guide - https://next-intl-docs.vercel.app/docs/getting-started/app-router-server-components When I paste this to my i18n.ts (in root directory) import {getRequestConfig} from 'next-intl/server'; export default getRequestConfig(async ({locale}) => ({ messages: (await import(`./messages/${locale}.json`)).default })); I got this error Module '"next-intl/server"' has no exported member 'getRequestConfig' Mandatory reproduction URL (CodeSandbox or GitHub repository) https://codesandbox.io/p/sandbox/i18n-thzc9v Reproduction description Steps to reproduce: Open codesandbox Click on VS conde in top right corner See error in i18n.ts Expected behaviour I wait no error But got an error Any chance next-intl got installed incorrectly? When I run your example I also get this error: Seems like your middleware.ts example is empty. I'll move this to a discussion since it seems to be a usage question.
gharchive/issue
2023-10-01T16:08:49
2025-04-01T04:55:54.240482
{ "authors": [ "amannn", "nicitaacom" ], "repo": "amannn/next-intl", "url": "https://github.com/amannn/next-intl/issues/538", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1002244752
Leetcode Question Number 1290 Leetcode Question Number 1290 solution needs to be added in the repo . The link for the question is this. Accepted Languages C++ / JAVA . You Only need to add the class Solution in the file. Make Sure you have read the contributions guidelines before creating an PR. Assign me
gharchive/issue
2021-09-21T11:06:20
2025-04-01T04:55:54.242128
{ "authors": [ "amanvermatemp", "aniketsaha19" ], "repo": "amanv8060/Leetcode-Questions", "url": "https://github.com/amanv8060/Leetcode-Questions/issues/1193", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2040284785
Can I reduce the time it takes to initiate the camera? I'm working on an application that requires that the device wakes up and takes a picture extremely quickly. So far I've been able to get this down to 267ms which meets spec for the project. But to do this I had to initialize the file system AFTER I initialized the camera. I need to be able to apply settings to the camera that are stored on the SD card. However reading the SD card takes an additional ~250ms which pushes the overall time to picture up to nearly 600ms which is a little longer than I want. I want to see if I can shave off a few more milliseconds. One of the things I would like to look at is the camera boot process. Right now it shows this: VOE]ext_in = 0 sync = 0 [VOE]md init success [VOE]algo ver ea63d5a [VOE]pack_v 0x0000 0x0002 cus_v 0 iq_id 0 dn 0 day 0 night 1 other 2 offset 224 length 16436 iq_size 32716 [VOE]Ver 0x0001000c Fast3A Cnt AE 1 AWB 1 period AE 5 AWB 1 delay 0 0 [VOE][isp_mod_sensor_init][1581]Errhdr_mode 0 is over sensor driver num 1 [VOE]hdr_mode 0 sensor driver num 1 [VOE]fps: 1880018620, pclk: 0, hts: 1077805056 [VOE]fps max 30.000000 min 5.000000 [VOE]exposure_step 29.629629 [VOE]change sensor mode => 1920x1080@30.000000fps - 'linear' [VOE]min_fps 5.000000, max_fps 30.000000, exp_step 29.629629 [VOE]md ver 0x6d640100 [VOE]ae ver 0x61650200 [VOE]awb ver 0x77620100 [VOE]cur_hdr_mode = 0 [VOE]VOE MEM Size = 11447 Used= 3550 KB Free= 7897 KB (8086944) [VOE]stream 0 buffer 0: 0x7054ac00 size 460800 [VOE]stream 0 buffer 1: 0x705bb500 size 460800 [VOE]first_config_osd2_block_num[0]: 1 [VOE]osd2_block_num[0]: 24 [VOE]NV12 640x480 1/30 [VOE]dynamic set fps 0 -> 30 ok [VOE]sensor power on [VOE]zoom crop default setting [VOE]status == 1718 [VOE]buffer size == (0x7063bbc0) (115200 38400) queue(20) I want to get into the code that generates this output to see what I can tighten up. However I can't seem to find the .c/.cpp files generating this output. I'm seeing header files (.h) but no matching .c or .cpp files. I am also seeing a lot of .a files which makes me suspect that some of the files are precompiled binaries and not source code. Are the source files used to compile those binaries available? Is there a way I could have access to them? Or am I just looking in the wrong place? Also do you have an environment for this camera that is outside of the Arduino library? The ESP32 had the C-level ESP-IDF. This product is advertised as having RTOS capabilities so is there a C-level library that can be used for development? Thanks, Jonathan L Clark @jonathan84clark For your situation, yes you are right, the boot process are prebuild as bin files for the SDK to use. For more sensors supported currently there are 2, GC5035 and JXF37. To go further, I think currently it is very hard to push the whole process faster. You may want to try running your application at multi-thread. there is a feature FCS, but have not port to Arduino SDK yet Closed as no further request.
gharchive/issue
2023-12-13T18:50:01
2025-04-01T04:55:54.307562
{ "authors": [ "M-ichae-l", "jonathan84clark" ], "repo": "ambiot/ambpro2_arduino", "url": "https://github.com/ambiot/ambpro2_arduino/issues/170", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
388194788
Fix multiprocessing support for Windows binary The black binary generated for Windows builds would fail on reformatting multiple files due to a Windows-specific multiprocessing issue. Fix by calling freeze_support() as described in Python docs. I didn't create an issue for this, not sure if that's a required part of the workflow? If that's preferred, let me know and I'll do it and link this PR to it. You probably want to add this in blackd.py as well Pull Request Test Coverage Report for Build 827 1 of 2 (50.0%) changed or added relevant lines in 1 file are covered. No unchanged relevant lines lost coverage. Overall coverage decreased (-0.03%) to 96.064% Changes Missing Coverage Covered Lines Changed/Added Lines % black.py 1 2 50.0% Totals Change from base Build 824: -0.03% Covered Lines: 2904 Relevant Lines: 3023 💛 - Coveralls @zsol Added now. At the moment, this binary is not built in CI, but I built it locally with C:\\Python36\\python.exe -m PyInstaller --clean -F --add-data blib2to3/;blib2to3 blackd.py and verified that it fails before the fix and works after. Thanks!
gharchive/pull-request
2018-12-06T12:35:38
2025-04-01T04:55:54.321464
{ "authors": [ "allhailwesttexas", "coveralls", "zsol" ], "repo": "ambv/black", "url": "https://github.com/ambv/black/pull/632", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1533401307
React capstone project - Metrics webapp Raised for React Capstone Project. I have built Air Quality Data Metrics mobile web application using the Air Pollution API. Following project requirements have been implemented: Features; Followed these provided guidelines. Built a Home Page that: Build Home Page where is displayed Countris which city Air Quality will be displayed on Detail page. Filtering is enabled through a search field Build Detail Page tha show the selected city air data metrics in details Bac Button is add to enable nagation back to the Home Page Create tests using React Testing Library. Deployed the project and added the live demo link in the README.md file Is this a requirement ? How did the first reviewer didn't tell anything about it ?
gharchive/pull-request
2023-01-14T18:00:40
2025-04-01T04:55:54.372298
{ "authors": [ "amentetteh" ], "repo": "amentetteh/react-capstone-metrics-webapp-project", "url": "https://github.com/amentetteh/react-capstone-metrics-webapp-project/pull/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1061630590
use constant name not value in documentation. Problem Statement Documentation is out of sync with code. Uses constant value and not constant name. Description of Change Changes all instances in documentation from value to constant name. Breaking Change not Closes #5 Hey @wadearnold thanks for the PR, looks like there are some compile and linting issues. But overall I think adding an example directory would be a good idea!
gharchive/pull-request
2021-11-23T19:20:30
2025-04-01T04:55:54.377775
{ "authors": [ "madflojo", "wadearnold" ], "repo": "americanexpress/simplemli", "url": "https://github.com/americanexpress/simplemli/pull/6", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
464369985
Added resources for Git. Added resources for Git. Thanks, @yashk2000 for the PR. I have a few changes, I will do the changes in the evening and merge it then. :tada: @vchrombie any updates? Any changes I should make?
gharchive/pull-request
2019-07-04T18:50:12
2025-04-01T04:55:54.385486
{ "authors": [ "vchrombie", "yashk2000" ], "repo": "amfoss/vidyaratna", "url": "https://github.com/amfoss/vidyaratna/pull/32", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
2550281243
🛑 KMHARS Main Fiber Line is down In 6639e1e, KMHARS Main Fiber Line (het09022amj.sn.mynetname.net) was down: HTTP code: 0 Response time: 0 ms Resolved: KMHARS Main Fiber Line is back up in 15cdc2b after 5 hours, 2 minutes.
gharchive/issue
2024-09-26T11:16:56
2025-04-01T04:55:54.410476
{ "authors": [ "amin-IT" ], "repo": "amin-IT/kmhars", "url": "https://github.com/amin-IT/kmhars/issues/311", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1903976783
请问搜索栏怎么设置默认值,比如下拉设置一个默认的选项 请问搜索栏怎么设置默认值,比如下拉设置一个默认的选项,我知道可以在get_page里面直接填充filterform对应的value ,但是还是想问下怎么直接设置 重写get_list_table, 给table添加defaultParams属性. 好的,我试下, 其实有了get_page 重构,这个框架灵活性完全不是问题,感谢作者这轮子,熟悉后,很高效。
gharchive/issue
2023-09-20T02:27:01
2025-04-01T04:55:54.428319
{ "authors": [ "amisadmin", "tomasky" ], "repo": "amisadmin/fastapi-amis-admin", "url": "https://github.com/amisadmin/fastapi-amis-admin/issues/124", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1599259757
Any way to use vercel.json? As the title says — is there a way to get this excellent GHA to “see” and respect a project’s top-level vercel.json file, such as to provide Cache-Control headers and so forth? @brycewray Just place vercel.json in your distribution directory. https://github.com/amondnet/vercel-action/blob/master/example/angular/src/vercel.json https://github.com/amondnet/vercel-action/tree/master/example/nextjs https://github.com/amondnet/vercel-action/blob/master/example/static/vercel.json https://github.com/amondnet/vercel-action/blob/master/example/express-basic-auth/vercel.json @amondnet Had already tried without success. It’s a Hugo project that creates the distribution dir (public) on the fly. I suppose I could put it in the static dir and let Hugo copy it there. However, I’ve already found another way, by using a variation on the officially suggested method. Thanks for the reply!
gharchive/issue
2023-02-24T20:49:18
2025-04-01T04:55:54.525950
{ "authors": [ "amondnet", "brycewray" ], "repo": "amondnet/vercel-action", "url": "https://github.com/amondnet/vercel-action/issues/225", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1293048577
Storyline for Demo-Day: What scenarios to include and how to approach them? User story As a team We want / need have a great demo day presentation So that we finish the project in style Acceptance criteria [ ] POs and one SD have a meeting and think about what scenarios we could show and how we can implement them [ ] create a documentation and update the rest of the team to have common understanding Definition of done (DoD) It builds without errors code has been reviewed Build has been made and deployed on a testing environment Code coverage >75% All acceptance criteria were met Documentation updated whenever needed within planning documents (Tabs: Bill of materials, Documentations) Branch/feature was merged into dev and then main DoD general criteria Feature has been fully implemented Feature has been merged into the mainline All acceptance criteria were met Product owner approved features All tests are passing Developers agreed to release Possible Structure Demo day.pdf
gharchive/issue
2022-07-04T11:38:14
2025-04-01T04:55:54.530791
{ "authors": [ "Meisterlu", "codingwithelijah" ], "repo": "amosproj/amos2022ss05-find-my-hearing-aid", "url": "https://github.com/amosproj/amos2022ss05-find-my-hearing-aid/issues/228", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
111345567
cannot upload sftp after latest atom update Atom v.1.0.19 OS Win10x64 Trace: Error: Cannot find module 'ssh2' at Module._resolveFilename (module.js:334:15) at Function.Module._resolveFilename (C:\Users\Стас\AppData\Local\atom\app-1.0.19\resources\app.asar\src\module-cache.js:383:52) at Function.Module._load (module.js:284:25) at Module.require (module.js:363:17) at require (module.js:382:17) at Object. (C:\Users\Стас.atom\packages\SFTP-deployment\lib\connections\SftpConnection.js:8:21) at Module._compile (module.js:428:26) at Object.defineProperty.value [as .js] (C:\Users\Стас\AppData\Local\atom\app-1.0.19\resources\app.asar\src\compile-cache.js:169:21) at Module.load (module.js:353:32) at Function.Module._load (module.js:308:12) at Module.require (module.js:363:17) at require (module.js:382:17) at Object. (C:\Users\Стас.atom\packages\SFTP-deployment\lib\connections\ConnectionFactory.js:7:22) at Module._compile (module.js:428:26) at Object.defineProperty.value [as .js] (C:\Users\Стас\AppData\Local\atom\app-1.0.19\resources\app.asar\src\compile-cache.js:169:21) at Module.load (module.js:353:32) Hi @lightinen, I will take a look, but it's an old post, maybe it's fixed since updates. Thank you for your feedback.
gharchive/issue
2015-10-14T08:09:06
2025-04-01T04:55:54.535752
{ "authors": [ "amoussard", "lightinen" ], "repo": "amoussard/sftp-deployment", "url": "https://github.com/amoussard/sftp-deployment/issues/88", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1093760798
stat failed hi, i have an error in amphp when i init madeline Exception: stat(): stat failed for E:\php\telegram\tbot\session.madeline.safe.php in BlockingDriver.php:47 PHP Fatal error: Uncaught \danog\MadelineProto\Exception: stat(): stat failed for E:\session.madeline.safe.php in phar://madeline-12bb67bc0618338b71ff507bd50a6b948d48e98e-80.phar/vendor/phabel-transpiler80.amphp/file/src/Driver/BlockingDriver.php:47 Unrelated to amphp, check online why the underlying php function fails for you. stat failed the error is about my telegram session,i close it and its work.
gharchive/issue
2022-01-04T21:17:39
2025-04-01T04:55:54.568353
{ "authors": [ "amirziaee", "enumag" ], "repo": "amphp/file", "url": "https://github.com/amphp/file/issues/65", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
115625326
[TACHYON-988] Added LocalTachyonClusterResource This is a Junit rule, which serves as sort-of a mixin for integration tests. The goal is that instead of creating and managing your own LocalTachyonCluster for every integration test, you use a LocalTachyonClusterResource, with some basic defaults, and can employ per-class or per-method configuration on it. The rule can also be created as a static class, which means its only initialized once per suite, rather than once per method. Thanks for the patch! Really a smart approach to improve integration tests. I left a few questions and please take a look. Thanks. LGTM Thanks for adding this rule to better organize the test cases. I like the approach. +1 LGTM LGTM. I guess you will have a follow-up PR to rewrite the tests to use this rule, right?
gharchive/pull-request
2015-11-07T02:13:25
2025-04-01T04:55:54.571210
{ "authors": [ "apc999", "calvinjia", "manugoyal", "yupeng9" ], "repo": "amplab/tachyon", "url": "https://github.com/amplab/tachyon/pull/2009", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
64735320
[TACHYON-324] Fix format issue Previously, Tachyon only delete block files under data folder during formatting worker storage. and it is not right to format directory by deleting and creating the directory, because the user may doesn't have the privilege in parent directory. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/Tachyon-Pull-Request-Builder/2861/ Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/Tachyon-Pull-Request-Builder/2872/ Test PASSed. +1 LGTM
gharchive/pull-request
2015-03-27T10:29:46
2025-04-01T04:55:54.574128
{ "authors": [ "AmplabJenkins", "calvinjia", "shimingfei" ], "repo": "amplab/tachyon", "url": "https://github.com/amplab/tachyon/pull/756", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
69039615
[TACHYON-371] Upgrade release plugin https://tachyon.atlassian.net/browse/TACHYON-371 Fixes some critical bugs when integrating with newer versions of git. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/Tachyon-Pull-Request-Builder/2989/Failed Tests: 1org.tachyonproject:tachyon: 1tachyon.master.MasterClientTest.openCloseTest Test FAILed. Jenkins, retest this please. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/Tachyon-Pull-Request-Builder/2990/ Test PASSed.
gharchive/pull-request
2015-04-17T01:01:10
2025-04-01T04:55:54.578028
{ "authors": [ "AmplabJenkins", "calvinjia" ], "repo": "amplab/tachyon", "url": "https://github.com/amplab/tachyon/pull/829", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2106093489
Offline support for React Native SDK Summary This issue is to collect customer requests on supporting offline mode for React Native SDK while Amplitude Browser SDK supports offline mode out of the box by https://github.com/amplitude/Amplitude-TypeScript/pull/644. Leave a comment under this issue with the use causes when and why you need this feature. We will prioritize it and fit into our roadmap. This would be a very useful feature for React Native as it's more likely to go "offline" then other use cases. E.g, a user is using the app on a train and the 4G reception drops out (either bad area or going through a tunnel). So those events will be lost. For our use case, the app is used in locations where internet connections are almost guaranteed to be non-existent (underground or remote areas). The phone / tablet connects to our devices WiFi to stream data and once they have completed their scan will then connect to a internet based WiFi. Being able to keep the events in local storage until there is a successful upload of the event would be vital to us. As in, if the sending of the event fails due to a network issue, it should not be removed from the unsent collection and be available for another retry.
gharchive/issue
2024-01-29T17:46:24
2025-04-01T04:55:54.596820
{ "authors": [ "Mercy811", "cgibson-emesent" ], "repo": "amplitude/Amplitude-TypeScript", "url": "https://github.com/amplitude/Amplitude-TypeScript/issues/655", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
356502051
AMP Story previews missing https://ampbyexample.com/stories/introduction/amp_story_hello_world/preview/embed/ 404s. It previously existed, and is the "view story" link from https://ampbyexample.com/stories/introduction/amp_story_hello_world/. Not sure what happened here. Some of the code around this was fiddled with recently (in #1490 and #1489) but everything works as expected when building from a fresh checkout locally. https://github.com/ampproject/amp-by-example/pull/1489 is the culprit. Will fix.
gharchive/issue
2018-09-03T13:16:02
2025-04-01T04:55:54.598815
{ "authors": [ "ithinkihaveacat", "sebastianbenz" ], "repo": "ampproject/amp-by-example", "url": "https://github.com/ampproject/amp-by-example/issues/1497", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
154429608
Improve linking between ampproject/amphtml and amp-by-example It would be useful to have a better way to reference amp-by-example from the project pages. For example: a "Check out this component on ABE" badge. Which project pages are you talking about, amphtml? Yes, and https://www.ampproject.org/ Let me edit the FR. I really like Malte's suggestion of having an ABE badge.
gharchive/issue
2016-05-12T08:52:25
2025-04-01T04:55:54.600792
{ "authors": [ "juliantoledo", "kul3r4", "sebastianbenz" ], "repo": "ampproject/amp-by-example", "url": "https://github.com/ampproject/amp-by-example/issues/157", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
332904033
Carousel lightbox Fixes #578 Can you please fix the HTML validation error: The id value [ carousel-with-preview ] must be unique. (id-unique) @sebastianbenz Done!
gharchive/pull-request
2018-06-15T20:26:23
2025-04-01T04:55:54.602074
{ "authors": [ "kul3r4", "sebastianbenz" ], "repo": "ampproject/amp-by-example", "url": "https://github.com/ampproject/amp-by-example/pull/1316", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2551584380
ULTRA SEVERITY: README.md contains evil word "lazy" Version v0.1.0 Severity Ultra Summary Readme file contains prohibited word "lazy". What kind of issue is this? The inclusion of the word "lazy" in the README.md file represents a clear violation of the guidelines established by HR Manager Karen. This term is considered unprofessional and carries negative connotations that can contribute to a hostile work environment. Its usage in official documentation sets a poor example for the team and undermines the company's commitment to fostering a respectful and inclusive workplace culture. Moreover, such language can lead to misunderstandings, decrease team morale, and potentially expose the organization to reputational risks. It is imperative that all team members adhere to HR policies by using appropriate and professional language in all forms of communication to promote a positive and productive work environment. Violated rule Rule: We don’t tolerate calling anyone lazy here. Word in question lazy To avoid misunderstandings: Please provide the exact line with the inappropriate word in bold text. [Line 12]: - Single-letter variable names are lazy. Use more descriptive names. Link to flawed file https://github.com/amshamah419/Karen-Inspect/blob/main/README.md Steps To Reproduce Read README.md using your favorite browser or text editor. Appropriate punishment (IMMEDIATE TERMINATION) The current behavior Word "lazy" is present. The expected behavior A version of the code which does not contain the word "lazy" in README.md should be released. This issue also happens to contain the word 'lazy'. This issue also happens to contain the word 'lazy'. Could you please provide evidence for this? Nevermind, I see that you have fixed the issue with an edit. Thank you for raising this ultra-severe issue. The use of the word "lazy" in the README.md is indeed a violation of our HR-approved guidelines. To ensure compliance with our contributor program, please note that several mandatory trainings remain incomplete for both @Mineshafter61 and @Meltr0n: Professional Vocabulary Sensitivity Training – Understanding words like "lazy" and why they're problematic. Respectful Language in Documentation – Guidelines for describing code without offending anyone. Avoiding the Termination Zone – How to keep commits and documentation "HR-friendly" and avoid immediate termination. We will have a developer (or feel free to do so yourself once you are compliant with your trainings) revise the README by replacing "lazy" with a more inclusive term, such as "reluctantly efficient", and ensure future language adheres to these standards. We expect compliance with these guidelines in future commits to foster a respectful and positive development environment. Thank you for your attention to this matter. Karen from HR 🛑
gharchive/issue
2024-09-26T21:48:50
2025-04-01T04:55:54.844435
{ "authors": [ "Meltr0n", "Mineshafter61", "amshamah419" ], "repo": "amshamah419/Karen-Inspect", "url": "https://github.com/amshamah419/Karen-Inspect/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
704468764
docs : add WeTransfer to 'Who uses Amundsen?' Summary of Changes Adds WeTransfer to Who uses Amundsen? in the README As a side note, I'm sharing some of the things we did differently. Maybe helpful for some people. use Kuberenetes Ingress set up authentication with OAuth Proxy separate config and application codes by using templated json (i.e .j2) for configuration and render values when passing them to Airflow tasks use Factory Pattern so that developers with less experience in Amundsen can simply pass job, task, publisher, extractor, loader, transformer and prepare a j2 file. # returns DefaultJob runner = DataBuilderRunner( job_type=dag_args.pop("job"), publisher_type=dag_args.pop("publisher"), task_type=dag_args.pop("task"), extractor_type=dag_args.pop("extractor"), transformer_type=dag_args.pop("transformer"), loader_type=dag_args.pop("loader"), job_config=dag_args.pop("job_config"), logger=logger, **dag_args ) # runs DefaultJob.launch() along with other custom actions runner.execute() CheckList Make sure you have checked all steps below to ensure a timely review. [x] PR title addresses the issue accurately and concisely, including a title prefix. [x] PR includes a summary of changes. [x] My commits follow the guidelines from "How to write a good git commit message" also @higee , your note is useful for folks who are not familiar with databuilder and its different configs. Maybe you could share that to the slack channel?
gharchive/pull-request
2020-09-18T15:28:53
2025-04-01T04:55:54.879980
{ "authors": [ "feng-tao", "higee" ], "repo": "amundsen-io/amundsen", "url": "https://github.com/amundsen-io/amundsen/pull/705", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1330331462
Type cast before normalize dpt What does this PR do? Modifies the feature extractor call to make the inputs to BatchFeature always numpy arrays. This ensures the type of the returned pixel_values matches those requested with return_tensors. The conversion to numpy arrays happens before normalization to ensure consistent rescaling is done on the inputs. This solution means the return_tensors=None default can stay and that the behaviour is as expected for any combination of the do_xxx flag values. Before submitting [ ] This PR fixes a typo or improves the docs (you can dismiss the other checks if that's the case). [x] Did you read the contributor guideline, Pull Request section? [ ] Was this discussed/approved via a Github issue or the forum? Please add a link to it if that's the case. [ ] Did you make sure to update the documentation with your changes? Here are the documentation guidelines, and here are tips on formatting docstrings. [x] Did you write any new necessary tests? The docs for this PR live here. All of your documentation changes will be reflected on that endpoint.
gharchive/pull-request
2022-08-05T19:45:24
2025-04-01T04:55:54.886851
{ "authors": [ "HuggingFaceDocBuilderDev", "amyeroberts" ], "repo": "amyeroberts/transformers", "url": "https://github.com/amyeroberts/transformers/pull/15", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1277498893
Add support for TYPE_MATCHES relocation This change adds support for the TYPE_MATCHES relocation to the program. This relocation backs the type match relation, which is introduced here: https://lore.kernel.org/bpf/20220620231713.2143355-1-deso@posteo.net/ Signed-off-by: Daniel Müller deso@posteo.net Thanks!
gharchive/pull-request
2022-06-20T23:36:29
2025-04-01T04:55:54.920182
{ "authors": [ "anakryiko", "danielocfb" ], "repo": "anakryiko/btfdump", "url": "https://github.com/anakryiko/btfdump/pull/1", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
2454748709
fix: additional dir paths windows Closes # [optional] What gif best describes this PR or how it makes you feel? @allcontributors add @illunix for code
gharchive/pull-request
2024-08-08T02:36:13
2025-04-01T04:55:54.948968
{ "authors": [ "brandonroberts", "illunix" ], "repo": "analogjs/analog", "url": "https://github.com/analogjs/analog/pull/1260", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2081219108
Stable API for components 多分需要があるけどどのコンポーネントから手を付ければいいかがわからないのでほしいcomponentをユースケースとともに教えてください Merge Skinned Mesh declavatar で切り替えアニメーションを生成する際、マージした状態でアニメーションを生成するようにしたい 他のエディタ拡張などでも、切り替えパターン的にマージ可能なもの同士を自動でマージするような機能を実装する余地がうまれるはず 極論ツールの都合のみで必要な GameObject/Component を手動で生成させたくない Remove Mesh by BlendShape / in Box おおむね同上(外部エディタ拡張側のヒューリスティックなどでコンポーネントを付与できるメリットが大きそう) T&O ユーザーにワンクリックでできる簡単な最適化機能として、T&Oのアバターへのアタッチを外部から行いたいため T&Oについては原則デフォルト設定のみでも問題ないですかね はい 現在要求があるものを #976 で仮に実装してみましたので、使用上で問題ないかを教えていただきたいです。 RemoveMeshByBlendShape, RemoveMeshInBox, MergeSkinnedMesh については問題なさそうです T&O も問題ありません
gharchive/issue
2024-01-15T04:18:14
2025-04-01T04:55:55.001165
{ "authors": [ "Sabu006", "anatawa12", "fu-sasa", "kb10uy" ], "repo": "anatawa12/AvatarOptimizer", "url": "https://github.com/anatawa12/AvatarOptimizer/issues/832", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
373103673
Connection aborted due to "bad status line" http://termbin.com/k7jp I thought this might somehow be related to the issues that github was having, but I'm still getting this error. I don't think there's some issue with my repository settings or the path or anything, since it was able to get the image digest of the docker image. hi @paulcdejean - From that log output, the suspicious lines are the 'Connection aborted' errors which is indicating that the connection between the analyzer (client) and the upstream services (policy_engine, catalog) is being closed prematurely. One thing to check is to make sure that the catalog/policy_engine services are not being killed/restarted due to memory pressure, which would show up like this on the analyzer. Close inspection of the policy_engine/catalog log files should help to determine if this is the case (or other reason), as well as inspection of 'dmesg' on the host that is running the catalog/policy_engine server containers, if possible. For larger installations/images, especially when there are multiple analyzers, we recommend that 4gb ++ of RAM for the services (which should resolve this, if memory pressure is the core issue!). We'll close this ticket since some time has passed - please let us know if there are continuing / related issues!
gharchive/issue
2018-10-23T17:04:54
2025-04-01T04:55:55.005633
{ "authors": [ "nurmi", "paulcdejean" ], "repo": "anchore/anchore-engine", "url": "https://github.com/anchore/anchore-engine/issues/107", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
716844562
Package list API should use syft packages (python packages) As an engine user I expect the product to leverage syft capabilities to produce the listing for Python packages. AC In order to support the existing UI client implement a translation layer on syft for the following: Match the existing packages API call syntax. Steps to Test Use the Anchore CLI for known image and compare results to a known image matches. Look at the anchore_engine/analyzers/modules dir within the anchore container image and ensure the Python analyzer (32_python_packages.py) is not present. pytest -v standalone/test_package_list runs in the anchore-engine/tests/functional/clients directory reporting Python-related tests that are all passing. Developer Notes New tests and fixtures should be added to ensure that before/after changes do not alter the expectations of the system. The syft cataloger 2.0 work is required to continue further. The existing cataloger uses python-specific utils, where as the golang equivalent will need to piece together the information from multiple file sources --requiring the cataloger 2.0 interface to do so.
gharchive/issue
2020-10-07T20:53:59
2025-04-01T04:55:55.009219
{ "authors": [ "wagoodman" ], "repo": "anchore/anchore-engine", "url": "https://github.com/anchore/anchore-engine/issues/654", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }