id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
1845019449
Typo in docs Describe the bug The variants docs has a wrong property. It should be called compoundVariants according to the source code Your Example Website or App No response Steps to Reproduce the Bug or Issue Custom variants Expected behavior It should be called compoundVariants Screenshots or Videos No response Operating System Version All Browser Chrome Good catch thank you! Fix is ont its way https://github.com/nextui-org/nextui/commit/cef7235c4bb3c8c3938bee2496b19beb6df83d22
gharchive/issue
2023-08-10T11:48:39
2025-04-01T04:35:12.605900
{ "authors": [ "ccorreia64", "jrgarciadev" ], "repo": "nextui-org/nextui", "url": "https://github.com/nextui-org/nextui/issues/1343", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1980927590
[BUG] - Accordion error Error: Unknown element <[object Object]> in collection. NextUI Version 2.0.24 Describe the bug I use @nextui-org/accordion @ 2.0.24 version can't make it work! Here's the full code: import {Accordion, AccordionItem} from "@nextui-org/accordion" export default function Home() { const defaultContent = "Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat."; return( <Accordion> <AccordionItem key="1" aria-label="Accordion 1" title="Accordion 1"> {defaultContent} </AccordionItem> <AccordionItem key="2" aria-label="Accordion 2" title="Accordion 2"> {defaultContent} </AccordionItem> <AccordionItem key="3" aria-label="Accordion 3" title="Accordion 3"> {defaultContent} </AccordionItem> </Accordion>); } The code is copied from https://nextui.org/docs/components/accordion this is the full console error: - error node_modules/.pnpm/@react-stately+collections@3.10.2_react@18.2.0/node_modules/@react-stately/collections/dist/import.mjs (205:0) @ $eb2240fc39a57fa5$export$bf788dd355e3a401.getFullNode - error Error: Unknown element <[object Object]> in collection. at getFullNode.next (<anonymous>) at iterateCollection.next (<anonymous>) at Generator.next (<anonymous>) Your Example Website or App No response Steps to Reproduce the Bug or Issue pnpm i @nextui-org/accordion copy example code from "https://nextui.org/docs/components/accordion" pnpm dev you can get the error Expected behavior works as normal Screenshots or Videos No response Operating System Version macOS 13.4.1 (c) Browser Chrome Duplicated https://github.com/nextui-org/nextui/issues/1619
gharchive/issue
2023-11-07T09:36:44
2025-04-01T04:35:12.611276
{ "authors": [ "jrgarciadev", "universeroc" ], "repo": "nextui-org/nextui", "url": "https://github.com/nextui-org/nextui/issues/1912", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1279138512
Prettier? 🤔 Yeah we could have a config but I think that may be too opinionated? I personally just use the default that comes with the vscode extension. If someone wants prettier they can install/configure it themselves imo. What do you think? Make it optional ? When generating a project, it prompts whether to add a prettier config file and packages or not. Thought a bit about this and I don't think it meaningfully differs from my stance on "add Jotai/Zustand" You can just NPM install if you want prettier. No need for us to include. I think I let my desire for prettier get ahead of me 😂 I think that most people have prettier installed in their editor and if they have a preferred config then can do that on their own. Closing this now after discussing on discord.
gharchive/issue
2022-06-21T21:42:26
2025-04-01T04:35:12.633642
{ "authors": [ "TheoBr", "booleans-oss", "nexxeln" ], "repo": "nexxeln/create-t3-app", "url": "https://github.com/nexxeln/create-t3-app/issues/25", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2739789935
[★𝐕𝐈𝐑𝐀𝐋★] 𝐒𝐮𝐛𝐡𝐚𝐬𝐡𝐫𝐞𝐞 𝐒𝐚𝐡𝐮 Video 𝙻𝚎𝚊𝚔𝚎𝚍 𝙽𝚘𝚠 𝘖𝘳𝘪𝘨𝘪𝘯𝘢𝘭 𝖳𝗋𝖾𝗇𝖽𝗂𝗇𝗀 Clips LINK On Social Media X Twitter 20 sec ago [★𝐕𝐈𝐑𝐀𝐋★] 𝐒𝐮𝐛𝐡𝐚𝐬𝐡𝐫𝐞𝐞 𝐒𝐚𝐡𝐮 Video 𝙻𝚎𝚊𝚔𝚎𝚍 𝙽𝚘𝚠 𝘖𝘳𝘪𝘨𝘪𝘯𝘢𝘭 𝖳𝗋𝖾𝗇𝖽𝗂𝗇𝗀 Clips LINK On Social Media X Twitter .. .. .. .. .. .. 𝐒𝐮𝐛𝐡𝐚𝐬𝐡𝐫𝐞𝐞 𝐒𝐚𝐡𝐮, a young and talented digital creator, recently gained widespread attention on social media platforms with her viral video. The video quickly became a trending topic across various platforms, sparking a significant amount of discussion among viewers. As a rising star in the digital world, 𝐒𝐮𝐛𝐡𝐚𝐬𝐡𝐫𝐞𝐞 𝐒𝐚𝐡𝐮's creativity and content have captivated audiences, contributing to her growing popularity on platforms like X and Twitter. This viral moment has sparked conversations about the impact of digital content on social media trends. 𝐒𝐮𝐛𝐡𝐚𝐬𝐡𝐫𝐞𝐞 𝐒𝐚𝐡𝐮's video, while generating attention, highlights the growing influence of young content creators who are reshaping the landscape of online media. With her engaging presence, 𝐒𝐮𝐛𝐡𝐚𝐬𝐡𝐫𝐞𝐞 𝐒𝐚𝐡𝐮 continues to inspire others to explore the possibilities of digital creation and interaction on social platforms. . . . .. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . To Reproduce Additional Context Validation [X] I confirm this is a bug about nezha (Nezha Dashboard). [X] I have searched Issues and confirm this bug has been reported before. . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄) . . . ### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄) . . . . . .
gharchive/issue
2024-12-14T11:27:01
2025-04-01T04:35:12.755927
{ "authors": [ "ahidiqba", "lliansm", "wahidsarker", "ylmenendez" ], "repo": "nezhahq/nezha", "url": "https://github.com/nezhahq/nezha/issues/739", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1743786754
Improve multiqc report Just a couple of small improvements for the multiqc report. PR checklist [x] This comment contains a description of changes (with reason). [ ] If you've fixed a bug or added code that should be tested, add tests! [ ] If you've added a new tool - have you followed the pipeline conventions in the contribution docs [ ] If necessary, also make a PR on the nf-core/nanostring branch on the nf-core/test-datasets repository. [x] Make sure your code lints (nf-core lint). [x] Ensure the test suite passes (nextflow run . -profile test,docker --outdir <OUTDIR>). [ ] Usage Documentation in docs/usage.md is updated. [ ] Output Documentation in docs/output.md is updated. [ ] CHANGELOG.md is updated. [ ] README.md is updated (including new tool citations and authors/contributors). Maybe a CHANGELOG update to draw people';s attention the output might look different might nice, but MultiQC wise looks good 👍 (I'm assuming teh write_out_prepared_gex is related to that) Thank you!
gharchive/pull-request
2023-06-06T12:39:10
2025-04-01T04:35:12.786436
{ "authors": [ "christopher-mohr" ], "repo": "nf-core/nanostring", "url": "https://github.com/nf-core/nanostring/pull/23", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
209791084
Issue with notifications when user is outside wifi and was an invite Hi Should notifications for an invited user work outside of wifi or not? I only receive them when in wifi never over cellular. The 'owner' always receives then I am using iPad as the home hub Also based on a motion sensor That's how motion works
gharchive/issue
2017-02-23T15:17:42
2025-04-01T04:35:12.796820
{ "authors": [ "brownad" ], "repo": "nfarina/homebridge", "url": "https://github.com/nfarina/homebridge/issues/1161", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
113545661
Stock Count - The loading sign keeps running when validate stock count For some stock counts containing many items, when trying to validate, system cannot complete the process since the loading sign on validate button keeps running. stock count speed improved here: https://github.com/nfco/netforce/commit/e51e7cbe434ff7cf0b0c4ce58667d8a015046ffa
gharchive/issue
2015-10-27T09:51:45
2025-04-01T04:35:12.798082
{ "authors": [ "bankwirat", "nf-dj" ], "repo": "nfco/netforce", "url": "https://github.com/nfco/netforce/issues/23", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1583810073
[BUG] Term search uses collapse to hide extended info This creates issues with safari rendering, children will not get collapsed correctly, so parent box size stays the same. @Brilator Can you check if this issue still happens in Swate-standalone? I can only assume that Excel uses deprecated browser versions for add-ins even on mac, as this should be fully supported: https://developer.mozilla.org/en-US/docs/Web/CSS/visibility?retiredLocale=de#browser_compatibility still occurs in Safari, not in Firefox on Mac
gharchive/issue
2023-02-14T09:30:52
2025-04-01T04:35:12.800328
{ "authors": [ "Brilator", "Freymaurer" ], "repo": "nfdi4plants/Swate", "url": "https://github.com/nfdi4plants/Swate/issues/270", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1858257057
[BUG] <Roster file has bad ID for Jay Cutler in 2010 and no ID in 2011> Is there an existing issue for this? [X] I have searched the existing issues Have you installed the latest development version of the package(s) in question? [X] I have installed the latest development version of the package. What version of the package do you have? na direct pull Describe the bug In the roster file, Jay Cutler's 2010 record uses Rashied Davies ID's. Jay Cutler does not have a record for 2011. Both Jay Cutler and Rashid Davies have inconsistent draft data (different records say they were drafted by different teams in different rounds) Reprex import pandas as pd import numpy ## load roster files ##\ roster_url = 'https://github.com/nflverse/nflverse-data/releases/download/rosters' rosters = [] for season in range(2006,2018): ## pull roster for that season ## temp = pd.read_csv( '{0}/roster_{1}.csv?raw=true'.format( roster_url, season ), low_memory=False ) rosters.append(temp) ## combine rosters ## r = pd.concat(rosters) ## can see that jay cutler has wrong ID 2010, is missing 2011, and strange draft data ## r[ r['full_name'] == 'Jay Cutler' ][[ 'season','team','full_name','gsis_id', 'espn_id','pff_id','pfr_id','esb_id', 'entry_year','draft_club','draft_number' ]] ## if you inspect the wrong ID, you see it's Rashied Davies ## r[ r['gsis_id'] == '00-0023429' ][[ 'season','team','full_name','gsis_id', 'espn_id','pff_id','pfr_id','esb_id', 'entry_year','draft_club','draft_number' ]] ## to confirm it's not an issue with concatination of DFs, you can see issue is at the file ## ## level ## r2010 = pd.read_csv( '{0}/roster_{1}.csv?raw=true'.format( roster_url, 2010 ) ) r2010[ r2010['full_name'] == 'Jay Cutler' ][[ 'season','team','full_name','gsis_id', 'espn_id','pff_id','pfr_id','esb_id', 'entry_year','draft_club','draft_number' ]] Expected Behavior Expected behavior is that the roster file would have consistent information for both players and not have missing seasons nflverse_sitrep na did in python Screenshots No response Additional context No response This has been resolved: nflreadr::load_rosters(2010:2011) |> dplyr::filter(full_name == "Jay Cutler") |> dplyr::select(season, week, full_name, gsis_id) #> ── nflverse roster data ──────────────────────────────────────────────────────── #> ℹ Data updated: 2023-09-06 10:40:56 PDT #> # A tibble: 2 × 4 #> season week full_name gsis_id #> <int> <int> <chr> <chr> #> 1 2010 20 Jay Cutler 00-0024226 #> 2 2011 16 Jay Cutler 00-0024226
gharchive/issue
2023-08-20T20:18:53
2025-04-01T04:35:12.806261
{ "authors": [ "greerreNFL", "john-b-edwards" ], "repo": "nflverse/nflverse-data", "url": "https://github.com/nflverse/nflverse-data/issues/31", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
607109161
Hyper-V cluster prepare isn't working It's been a while since I've created a Hyper-V cluster and while working on the latest node templates #811, I noticed that deploying Hyper-V cluster is failing in the prepare step. The problem is that neon-cli is trying to run the Hyper-V/Get-NetAdapter cmdlet but this no longer seems to be present in the Hyper-V namespace. The command exists without the namespace. This used to work, so MSFT must have relocated the cmdlet or something. We need to backport this to a new 1.x release. This is fixed for v1.2.3 and v2.1.0
gharchive/issue
2020-04-26T20:44:16
2025-04-01T04:35:12.811087
{ "authors": [ "jefflill" ], "repo": "nforgeio/neonKUBE", "url": "https://github.com/nforgeio/neonKUBE/issues/812", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1126892391
chore(deps): update dependency @angular/platform-browser to v13 This PR contains the following updates: Package Type Update Change @angular/platform-browser dependencies major ~10.1.3 -> ~13.2.0 Release Notes angular/angular v13.2.1 Compare Source animations Commit Type Description 4644886aaf perf remove no longer needed CssKeyframes classes (#​44903) (#​44919) common Commit Type Description b4e4617807 fix include query parameters for open HTTP requests in verify (#​44917) compiler-cli Commit Type Description 0778e6f7d7 fix accept nullish coalescing operator for any and unknown types (#​44862) 07185f4ed1 fix enable nullish coalescing check only with strictNullChecks (#​44862) 4a5ad1793f fix ensure casing of logical paths is preserved (#​44798) core Commit Type Description 7ec482d9c2 fix Add back support for namespace URIs in createElement of dom renderer (#​44914) 250dc40a46 fix flush delayed scoping queue while setting up TestBed (#​44814) forms Commit Type Description 1aebbf8714 fix ensure OnPush ancestors are marked dirty when the promise resolves (#​44886) 6b7fffcbeb fix Update the typed forms migration schematic to find all files. (#​44881) Special Thanks Alan, Andrew Kushnir, Andrew Scott, Aristeidis Bampakos, Arjen, Daniel Díaz, David Shevitz, Doug Parker, Dylan Hunn, Esteban Gehring, George Kalpakas, Jessica Janiuk, JoostK, Juri Strumpflohner, Lee Robinson, Maarten Tibau, Paul Gschwendtner, Theodore Brown, arturovt, dario-piotrowicz, fru2, markostanimirovic and mgechev v13.2.0 Compare Source Deprecations The CachedResourceLoader and RESOURCE_CACHE_PROVIDER symbols were previously necessary in some cases to test AOT-compiled components with View Engine, but they are no longer needed since Ivy. The ComponentFactory and ComponentFactoryResolver classes are deprecated. Since Ivy, there is no need to resolve Component factories. Please use other APIs where you Component classes can be used directly (without resolving their factories). Since Ivy, the CompilerOptions.useJit and CompilerOptions.missingTranslation config options are unused, passing them has no effect. Commit Type Description 9c11183e74 docs deprecate CachedResourceLoader and RESOURCE_CACHE_PROVIDER symbols (#​44749) 9f12e7fea4 docs deprecate ComponentFactory and ComponentFactoryResolver symbols (#​44749) 4e95a316ce docs deprecate unused config options from the CompilerOptions interface (#​44749) compiler Commit Type Description a4ab6d6b72 feat add support for safe calls in templates (#​44580) abd1bc8039 fix correct spans when parsing bindings with comments (#​44785) ed67a074ce fix properly compile DI factories when coverage reporting is enabled (#​44732) compiler-cli Commit Type Description fa835b5a29 feat enable extended diagnostics by default (#​44712) 73424def13 feat provide the animations for DirectiveMeta (#​44630) fe3e4d6865 fix Handle ng-template with structural directive in indexer (#​44788) 7316e72ec5 fix properly index elements when on a template (#​44785) 100091ebf0 fix remove leftover _extendedTemplateDiagnostics requirements (#​44777) d2ae96f742 fix skip ExtendedTemplateCheckerImpl construction if there were configuration errors (#​44778) core Commit Type Description 5626b34264 fix consistently use namespace short name rather than URI (#​44766) 94bfcdd9de fix error if NgZone.isInAngularZone is called with a noop zone (#​44800) forms Commit Type Description 72092ebd26 feat Allow a FormControl to use initial value as default. (#​44434) f7aa937cac fix Make some minor fixups for forward-compatibility with typed forms. (#​44540) router Commit Type Description 5a4ddfd4f5 feat Allow symbol keys for Route data and resolve properties (#​44519) Special Thanks Alex Rickabaugh, Andrew Kushnir, Andrew Scott, Dario Piotrowicz, Derek Cormier, Doug Parker, Douglas Parker, Dylan Hunn, George Kalpakas, Jessica Janiuk, JoostK, Kristiyan Kostadinov, Martin Probst, Oleg Postoev, Stephanie Tuerk, Tim Bowersox, Wiley Marques, Yousaf Nawaz, dario-piotrowicz, iRealNirmal, ivanwonder and shejialuo v13.1.3 Compare Source animations Commit Type Description af0a152a2c fix apply setStyles to only rootTimelines (#​44515) compiler-cli Commit Type Description 626f3f230b perf reduce analysis work during incremental rebuilds (#​44731) ngcc Commit Type Description f9ca4d8499 fix support element accesses for export declarations (#​44669) Special Thanks Alan Agius, Andrew Kushnir, AnkitSharma-007, Daniel Díaz, Dmytro Mezhenskyi, Jessica Janiuk, Joey Perrott, JoostK, Ramesh Thiruchelvam, dario-piotrowicz, iRealNirmal and Łukasz Holeczek v13.1.2 Compare Source animations Commit Type Description abc217b28e fix retain triggers values for moved tracked list items (#​44578) compiler Commit Type Description 59eef29a6c fix correct spans when parsing bindings with comments (#​44678) compiler-cli Commit Type Description 08049fa23f fix enable narrowing of using type guard methods (#​44447) a26afce68c fix fix crash during type-checking of library builds (#​44587) 1e918b6f31 fix handle property reads of ThisReceiver in the indexer (#​44678) 63c8e56a3a fix incorrectly interpreting $any calls with a property read (#​44657) 60fb27f12d fix properly index elements (#​44678) language-service Commit Type Description f5addee488 fix revert the test files for Ivy (#​44528) Special Thanks Abdurrahman Abu-Hijleh, Adam Plumer, Alex Rickabaugh, AlirezaEbrahimkhani, Andrew Kushnir, Andrew Scott, Borja Paz Rodríguez, Chihab Otmani, Chris Mancini, Dario Piotrowicz, Doug Parker, George Kalpakas, Joey Perrott, JoostK, Kristiyan Kostadinov, Kyoz, Patrick Prakash, Paul Gschwendtner, Serhey Dolgushev, Yousaf Nawaz, Yuchao Wu, alkavats1, dario-piotrowicz, huangqing, ivanwonder, shejialuo, twerske, wszgrcy and zuckjet v13.1.1 Compare Source animations Commit Type Description bb1d4ff315 fix don't consume instructions for animateChild (#​44357) d8b6adb7bc fix should not invoke disabled child animations (#​37724) forms Commit Type Description bce108ab49 fix _reduceValue arrow function now has correct types. (#​44483) 998c1e63fe fix I indroduced a minor error in a previous PR: pendingValue is a value not a boolean flag. (#​44450) Special Thanks Aristeidis Bampakos, Dylan Hunn, George Kalpakas, JoostK, Kristiyan Kostadinov, Paul Gschwendtner, Spej, Yousaf Nawaz, dario-piotrowicz, faso-dev, jaybell and zuckjet v13.1.0 Compare Source Deprecations The downgradeModule function calls with NgModule factories are deprecated. Please use NgModule class based downgradeModule calls instead. common TestRequest from @angular/common/http/testing no longer accepts ErrorEvent when simulating XHR errors. Instead instances of ProgressEvent should be passed, matching with the native browser behavior. Commit Type Description dbc46d68b9 docs deprecate factory-based signature of the downgradeModule function (#​44090) common Commit Type Description 489cf42cd0 fix incorrect error type for XHR errors in TestRequest (#​36082) 13362972bb perf code size reduction of ngFor directive (#​44315) compiler Commit Type Description c85bcb0c63 feat reference ICU message IDs from their placeholders (#​43534) core Commit Type Description 5dff077d50 feat add migration to remove entryComponents (#​44308) e65a245a0b feat add migration to remove entryComponents (#​44322) d56e3f43a1 feat support TypeScript 4.5 (#​44164) http Commit Type Description d452b388bd feat add has() method to HttpContext class (#​43887) localize Commit Type Description d3cf222a81 feat support "associated message ids" for placeholders (#​43534) ngcc Commit Type Description 41265919aa fix correctly resolve UMD dependencies (#​44381) upgrade Commit Type Description 34f990986c feat support NgModule class as an argument of the downgradeModule function (#​43973) Special Thanks Alex Rickabaugh, Andrew Kushnir, Andrew Scott, Doug Parker, Dustin M. Eastway, Dylan Hunn, George Kalpakas, HyperLife1119, Jelle Bruisten, Jessica Janiuk, Joey Perrott, JoostK, Kristiyan Kostadinov, Markus Doggweiler, Paul Gschwendtner, Pei Wang, Pete Bacon Darwin and dario-piotrowicz v13.0.3 Compare Source compiler-cli Commit Type Description 6cdbfdbe6e fix downlevel transform incorrectly extracting constructor parameters for nested classes (#​44281) 305b76b45f fix interpret string concat calls (#​44167) core Commit Type Description 0ca5c5bd09 fix add missing info about a component in the "pipe could not be found" error message (#​44081) 907da3977a fix destroy hooks not set up for useClass provider using forwardRef (#​44281) bcd3b4959b fix support cyclic metadata in TestBed overrides (#​44215) forms Commit Type Description 96fedd249e fix make the FormControlStatus available as a public API (#​44183) language-service Commit Type Description cabc1786de fix Correctly parse inputs and selectors with dollar signs (#​44268) ngcc Commit Type Description b68994d20a fix correctly report error when collecting dependencies of UMD module (#​44245) 6f5c0c1515 fix ensure that ngcc does not write a lock-file into node_modules package directories (#​44228) bf5f734e9c fix support the UMD wrapper function format emitted by Webpack (#​44245) router Commit Type Description d265d0d241 fix prevent componentless routes from being detached (#​44240) Special Thanks Alan Agius, Alex Rickabaugh, Andrew Kushnir, Andrew Scott, Artur, Christian-E, David Shevitz, Doug Parker, Douglas Parker, Dylan Hunn, George Kalpakas, Jessica Janiuk, Joey Perrott, JoostK, Kristiyan Kostadinov, Marc Redemske, Paul Gschwendtner, Pei Wang, Pete Bacon Darwin, Ramesh Thiruchelvam, Ravi Chandra, Rohan Pednekar, Ruslan Usmanov, dario-piotrowicz, profanis and unknown v13.0.2 Compare Source This release contains various API docs improvements. Special Thanks Andrew Kushnir, Armen Vardanyan, Dylan Hunn, Joey Perrott, Martin von Gagern, Paul Gschwendtner, Pete Bacon Darwin, Ramesh Thiruchelvam, dario-piotrowicz and fusho-takahashi v13.0.1 Compare Source compiler Commit Type Description ee2031d9f4 fix ensure that partially compiled queries can handle forward references (#​44113) e5a960b159 fix generate correct code for safe method calls (#​44088) compiler-cli Commit Type Description dede29e4f3 fix ensure literal types are retained when strictNullInputTypes is disabled (#​38305) 04df3a0b92 fix handle pre-release versions when checking version (#​44109) core Commit Type Description 4c700b6244 fix do not use Function constructors in development mode to avoid CSP violations (#​43587) platform-browser Commit Type Description 30a27adf9a fix use correct parent in animation removeChild callback (#​44033) Special Thanks A. Singh, Alan Agius, Alex Rickabaugh, Andrew Kushnir, Andrew Scott, Aristeidis Bampakos, George Kalpakas, Joe Martin (Crowdstaffing), Joel Lefkowitz, Joey Perrott, JoostK, Kristiyan Kostadinov, Michael Urban, Paul Gschwendtner, Pavan Kumar Jadda, Pei Wang, Pete Bacon Darwin, Roman Frołow, dario-piotrowicz, iRealNirmal, ileil, kreuzerk, mgechev, profanis and raman v13.0.0 Compare Source Blog post "Angular v13 is now available". Breaking Changes common The behavior of the SpyLocation used by the RouterTestingModule has changed to match the behavior of browsers. It no longer emits a 'popstate' event when Location.go is called. In addition, simulateHashChange now triggers both a hashchange and a popstate event. Tests which use location.go and expect the changes to be picked up by the Router should likely change to simulateHashChange instead. Each test is different in what it attempts to assert so there is no single change that works for all tests. Each test using the SpyLocation to simulate browser URL changes should be evaluated on a case-by-case basis. core TypeScript versions older than 4.4.2 are no longer supported. NodeJS versions older than v12.20.0 are no longer supported due to the Angular packages using the NodeJS package exports feature with subpath patterns. The WrappedValue class can no longer be imported from @angular/core, which may result in compile errors or failures at runtime if outdated libraries are used that are still using WrappedValue. The usage of WrappedValue should be removed as no replacement is available. forms A new type called FormControlStatus has been introduced, which is a union of all possible status strings for form controls. AbstractControl.status has been narrowed from string to FormControlStatus, and statusChanges has been narrowed from Observable<any> to Observable<FormControlStatus>. Most applications should consume the new types seamlessly. Any breakage caused by this change is likely due to one of the following two problems: (1) the app is comparing AbstractControl.status against a string which is not a valid status; or, (2) the app is using statusChanges events as if they were something other than strings. router The default url serializer would previously drop everything after and including a question mark in query parameters. That is, for a navigation to /path?q=hello?&other=123, the query params would be parsed to just {q: 'hello'}. This is incorrect because the URI spec allows for question mark characers in query data. This change will now correctly parse the params for the above example to be {v: 'hello?', other: '123'}. Previously null and undefined inputs for routerLink were equivalent to empty string and there was no way to disable the link's navigation. In addition, the href is changed from a property HostBinding() to an attribute binding (HostBinding('attr.href')). The effect of this change is that DebugElement.properties['href'] will now return the href value returned by the native element which will be the full URL rather than the internal value of the RouterLink href property. When storing and retrieving a DetachedRouteHandle, the Router traverses the Route children in order to correctly allow storing a parent route when there are several possible child Route configs that can be stored. This allows a RouteReuseStrategy to store a parent Route and a child, while preserving the ability to change the child route while restoring the parent. Some implementations of RouteReuseStrategy will need to be updated to correctly store and retrieve the DetachedRouteHandle of descendants as well as the stored parent ActivatedRouteSnapshot. Previously, the Router would only store the parent, making it impossible to change descendant paths when a stored parent was retrieved. See #​20114. The router will no longer replace the browser URL when a new navigation cancels an ongoing navigation. This often causes URL flicker and was only in place to support some AngularJS hybrid applications. Hybrid applications which rely on the navigationId being present on initial navigations that were handled by the Angular router should instead subscribe to NavigationCancel events and perform the location.replaceState themselves to add navigationId to the Router state. In addition, tests which assert urlChanges on the SpyLocation may need to be adjusted to account for the replaceState which is no longer triggered. It is no longer possible to use Route.loadChildren using a string value. The following supporting classes were removed from @angular/core: NgModuleFactoryLoader SystemJsNgModuleFactoryLoader The @angular/router package no longer exports these symbols: SpyNgModuleFactoryLoader DeprecatedLoadChildren The signature of the setupTestingRouter function from @angular/core/testing has been changed to drop its NgModuleFactoryLoader parameter, as an argument for that parameter can no longer be created. service-worker The return type of SwUpdate#activateUpdate and SwUpdate#checkForUpdate changed to Promise<boolean>. Although unlikely, it is possible that this change will cause TypeScript type-checking to fail in some cases. If necessary, update your types to account for the new return type. Deprecations core Angular no longer requires component factories to dynamically create components. The factory-based signature of the ViewContainerRef.createComponent function is deprecated in favor of a different signature that allows passing component classes instead. The getModuleFactory function is deprecated in favor of the getNgModuleById one. With Ivy it's possible to work with NgModule classes directly, without retrieving corresponding factories, so the getNgModuleById should be used instead. Ivy made it possible to avoid the need to resolve Component and NgModule factories. Framework APIs allow to use Component and NgModule Types directly. As a result, the PlatformRef.bootstrapModuleFactory and a factory-based signature of the ApplicationRef.bootstrap method are now obsolete and are now deprecated. The PlatformRef.bootstrapModuleFactory calls can be replaced with PlatformRef.bootstrapModule ones. The ApplicationRef.bootstrap method allows to provide Component Type, so this can be used a replacement for the factory-based calls. In ViewEngine, JIT compilation required special providers (like Compiler, CompilerFactory, etc) to be injected in the app and corresponding methods to be invoked. With Ivy, JIT compilation takes place implicitly if the Component, NgModule, etc have not already been AOT compiled. Those special providers were made available in Ivy for backwards-compatibility with ViewEngine to make the transition to Ivy smoother. Since ViewEngine is deprecated and will soon be removed, those symbols are now deprecated as well: ModuleWithComponentFactories Compiler CompilerFactory JitCompilerFactory NgModuleFactory Important note: this deprecation doesn't affect JIT mode in Ivy (JIT remains available with Ivy). In Ivy, AOT summary files are unused in TestBed. Passing AOT summary files in TestBed has no effect, so the aotSummaries usage in TestBed is deprecated and will be removed in a future version of Angular. platform-server The renderModuleFactory symbol in @angular/platform-server is no longer necessary as of Angular v13. The renderModuleFactory calls can be replaced with renderModule. service-worker The SwUpdate#activated observable is deprecated. The SwUpdate#activated observable only emits values as a direct response to calling SwUpdate#activateUpdate() and was only useful for determining whether the call resulted in an update or not. Now, the return value of SwUpdate#activateUpdate() can be used to determine the outcome of the operation and therefore using SwUpdate#activated does not offer any benefit. The SwUpdate#availalbe observable is deprecated. The new SwUpdate#versionUpdates observable provides the same information and more. Therefore, it is possible to rebuild the same behavior as SwUpdate#availalbe using the events emitted by SwUpdate#versionUpdates and filtering for VersionReadyEvent events. As a result, the SwUpdate#availalbe observable is now redundant. Commit Type Description 747553dd68 docs deprecate ViewEngine-based renderModuleFactory (#​43757) bazel Commit Type Description 62d7005a52 feat add strict_templates and experimental_extended_template_diagnostics to ng_module() rule (#​43582) d977701a43 feat allow for custom conditions to be set in ng_package targets (#​43764) 4886585875 feat create transition for enabling partial compilation (#​43431) cd1b52483e feat expose esm2020 and es2020 conditions in APF package exports (#​43740) 49b82ae561 feat implement partial compilation APF v13 for ng_package rule (#​43431) 274cb38e0b feat switch prodmode output to ES2020 (#​43431) 73ac50c447 feat wire up partial compilation build setting in ng_module (#​43431) e0a72857cc fix construct a manifest file even when warnings are emitted (#​43582) dbe656d1e0 fix ngc-wrapped should not rely on linker for external workspaces (#​43690) common Commit Type Description adf4481211 feat add injection token for default date pipe timezone (#​43611) c6a93001eb fix synchronise location mock behavior with the navigators (#​41730) compiler Commit Type Description 14b492df26 fix do not error if $any is used inside a listener (#​43866) compiler-cli Commit Type Description bed121c34f feat inline resources when generating class metadata calls (#​43178) 263feba5c2 fix handle nullable expressions correctly in the nullish coalescing extended template diagnostic (#​43572) 8f7fdc59af fix not evaluating new signature for __spreadArray (#​43618) 426a3ecae7 fix updates ngc to pass the build when only warnings are emitted (#​43673) core Commit Type Description a3960846da feat add createNgModuleRef function to create NgModuleRef based on NgModule class (#​43580) fe1f6421d2 feat add getNgModuleById function to retrieve loaded NgModules by id (#​43580) 81c7eb813c feat add migration to opt out existing apps from new test module teardown behavior (#​43353) e57691c9c5 feat Add migration to update empty routerLinks in templates (#​43176) 7dccbdd27b feat add support for Types in ViewContainerRef.createComponent (#​43022) c14085e434 feat drop support for TypeScript 4.2 and 4.3 (#​43642) 94ba59bc9d feat enable test module teardown by default (#​43353) ea61ec2562 feat support TypeScript 4.4 (#​43281) e0a0d05d45 feat update node version support range to support v16 (#​43740) 7396021e4b fix avoid duplicating comments in TestBed teardown migration (#​43776) 7fd0428aae fix don't rethrow errors if test teardown has been disabled (#​43635) 66fb311d20 fix incorrect signature for initTestEnvironment (#​43615) 8ae99821d6 fix support InjectFlags argument in NodeInjector.get() (#​41592) 8878183521 perf remove support for the deprecated WrappedValue (#​43507) elements Commit Type Description a468213f34 fix remove ng-add schematic (#​43975) f544a53f5f fix remove incorrect @angular/platform-browser peer dependency (#​43975) forms Commit Type Description d9d8f950e9 feat allow disabling min/max validators dynamically (by setting the value to null) (#​42978) e49fc96ed3 feat Make Form Statuses use stricter types. (#​42952) language-service Commit Type Description b10d90bef6 feat Add method for retrieving the component template at the cursor location (#​43208) d5f9890c92 feat auto-apply optional chaining on nullable symbol (#​42995) 69957f72e2 feat provide snippets for attribute (#​43590) fc3b50e427 fix exclude the SafePropertyRead when applying the optional chaining (#​43321) migrations Commit Type Description 95a68c5dc3 fix account for CRLF characters in template migrations (#​44013) 77bd2538cb fix apply individual expression edits to preserve newline characters (#​43519) d849350c7b fix Ensure routerLink migration doesn't update unrelated files (#​43519) 2efc18e675 fix migration failed finding tsconfig file (#​43343) b6f2a55147 fix prevent migrations from updating external templates multiple times (#​44013) router Commit Type Description 4f3beffdbf feat emit activate/deactivate events when an outlet gets attached/detached (#​43333) faf9f5a3bc feat new output that would notify when link is activated (#​43280) 3c6b653089 feat Option to correctly restore history on failed navigation (#​43289) 784671597e fix Allow question marks in query param values (#​31187) 796da641f0 fix Do not modify parts of URL excluded from with 'eager' updates (#​43421) 772e08d14e fix fix Router's public API for canceledNavigationResolution (#​43842) ccb09b4558 fix null/undefined routerLink should disable navigation (#​43087) 9e039ca68b fix Only trigger router navigation on popstate events from Location subscription (#​43328) c5d0bd4966 fix Prevent URL flicker when new navigations cancel ongoing ones (#​43496) adc68b100b fix reuse route strategy fix (#​43791) 361273fad5 refactor remove support for loadChildren string syntax (#​43591) service-worker Commit Type Description 59225f5586 feat SwUpdate#activeUpdate and SwUpdate#checkForUpdate should have a meaningful outcome (#​43668) 0dc45446fe feat expose more version update events (#​43668) Special Thanks Ahmed Ayed, Alan Agius, Alex Rickabaugh, Andrew Kushnir, Andrew Scott, Bjarki, Charles Lyding, Dmitrij Kuba, Doug Parker, Dylan Hunn, George Kalpakas, Jessica Janiuk, Jochen Kraushaar, Joe Martin (Crowdstaffing), Joey Perrott, Jon Rimmer, JoostK, Kristiyan Kostadinov, Maximilian Köller, Paul Gschwendtner, Pei Wang, Pete Bacon Darwin, Tomasz Domański, Willy Schott, anandtiwary, dario-piotrowicz, iRealNirmal, ivanwonder, krzysztof-grzybek, mgechev and vthinkxie v12.2.16 Compare Source 12.2.16 (2022-01-27) ngcc Commit Type Description 460befd693 fix support element accesses for export declarations (#​44824) Special Thanks Andrew Kushnir and JoostK v12.2.15 Compare Source ngcc Commit Type Description b6554d75cd fix correctly resolve UMD dependencies (#​44382) Special Thanks George Kalpakas v12.2.14 Compare Source compiler Commit Type Description e3db0385b6 fix ensure that partially compiled queries can handle forward references (#​44124) ngcc Commit Type Description a8be244113 fix correctly report error when collecting dependencies of UMD module (#​44245) fc072935ee fix support the UMD wrapper function format emitted by Webpack (#​44245) Special Thanks George Kalpakas, Pete Bacon Darwin and iRealNirmal v12.2.13 Compare Source compiler-cli Commit Type Description 9a89db790f fix avoid broken references in .d.ts files due to @​internal markers (#​43965) core Commit Type Description 8f402c9d06 fix support InjectFlags argument in NodeInjector.get() (#​41592) Special Thanks Alan Agius, George Kalpakas, Jochen Kraushaar, Joe Martin (Crowdstaffing), JoostK and vthinkxie v12.2.12 Compare Source compiler-cli Commit Type Description 112557497c fix avoid broken references in .d.ts files due to @​internal markers (#​43527) ngcc Commit Type Description 067ae54d46 fix support alternate UMD layout when adding new imports (#​43931) Special Thanks Alan Agius, Andrew Kushnir, George Kalpakas, Jessica Janiuk, Joey Perrott, JoostK, Mladen Jakovljević, Virginia Dooley, amayer42, dirk diebel and ericcheng2005 v12.2.11 Compare Source ngcc Commit Type Description cab21cea7a fix support alternate wrapper function layout for UMD (#​43879) router Commit Type Description 58c11865ac fix Do not clear currentNavigation if already set to next one (#​43852) Special Thanks Alan Agius, Andrew Kushnir, Andrew Scott, David Shevitz, George Kalpakas, Joe Martin (Crowdstaffing), Natalia Venditto, Pete Bacon Darwin, Younes Jaaidi and dario-piotrowicz v12.2.10 Compare Source Special Thanks Alan Agius, Daniel Díaz, David Shevitz, Doug Parker, George Kalpakas, Joe Martin (Crowdstaffing), Tanguy Nodet, Thomas Turrell-Croft, dario-piotrowicz, hchiam, markostanimirovic and mgechev v12.2.9 Compare Source core Commit Type Description b4b441077a fix handle invalid constructor parameters in partial factory declarations (#​43619) router Commit Type Description 7f6050587d fix unset attachRef when router-outlet is destroyed to avoid mounting a destroyed component (#​43697) service-worker Commit Type Description c4ecc07838 fix make ngsw.json generation deterministic and correct (#​43679) Special Thanks Alan Agius, Daniel Díaz, George Kalpakas, JoostK, Kristiyan Kostadinov, Pete Bacon Darwin, Wey-Han Liaw, dario-piotrowicz, iRealNirmal, little-pinecone, mgechev, ultrasonicsoft and xiaohanxu-nick v12.2.8 Compare Source compiler-cli Commit Type Description c1338bf837 fix correctly interpret token arrays in @​Injectable deps (#​43226) language-service Commit Type Description c8f8d7d3b1 fix provide dom event completions (#​43299) ngcc Commit Type Description 69299f7d4d fix do not fail for packages which correspond with Object members (#​43589) service-worker Commit Type Description 3cf41354ae fix do not unassign clients from a broken version (#​43518) Special Thanks Adrien Crivelli, Alex Rickabaugh, Andrew Scott, Bobby Galli, Chris, Daniel Díaz, Dario Piotrowicz, George Kalpakas, Joe Martin (Crowdstaffing), JoostK, Pete Bacon Darwin, Rafael Santana, Raj Sekhar, Ricardo Chavarria, Teri Glover, Virginia Dooley, dario-piotrowicz, enisfr and wszgrcy v12.2.7 Compare Source common Commit Type Description 2bb4bf1468 fix titlecase pipe incorrectly handling numbers (#​43476) compiler Commit Type Description 9c8a1f8a71 fix include leading whitespace in source-spans of i18n messages (#​43132) compiler-cli Commit Type Description defb02f11e fix handle directives that refer to a namespaced class in a type parameter bound (#​43511) platform-browser Commit Type Description adc7c56ede fix improve error message for missing animation trigger (#​41356) Special Thanks Andrew Scott, Daniel Díaz, George Kalpakas, JoostK, Kristiyan Kostadinov, Mwiku, Pete Bacon Darwin, Teri Glover, Virginia Dooley, Xiaohanxu1996, dario-piotrowicz and kirjs v12.2.6 Compare Source animations Commit Type Description 141fde1632 fix emit pure annotations to static property initializers (#​43344) core Commit Type Description ca510c87c5 fix emit pure annotations to static property initializers (#​43344) router Commit Type Description [4034f252c9](https://togithub.com/angular/angular/commit/4034f252c9707dabd0 Configuration 📅 Schedule: At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 🔕 Ignore: Close this PR and you won't be reminded about this update again. [ ] If you want to rebase/retry this PR, click this checkbox. This PR has been generated by Renovate Bot. ⚠ Artifact update problem Renovate failed to update an artifact related to this branch. You probably do not want to merge this PR as-is. ♻ Renovate will retry this branch, including artifacts, only when one of the following happens: any of the package files in this branch needs updating, or the branch becomes conflicted, or you click the rebase/retry checkbox if found above, or you rename this PR's title to start with "rebase!" to trigger it manually The artifact failure details are included below: File name: bprr-ui/package-lock.json npm WARN old lockfile npm WARN old lockfile The package-lock.json file was created with an old version of npm, npm WARN old lockfile so supplemental metadata must be fetched from the registry. npm WARN old lockfile npm WARN old lockfile This is a one-time fix-up, please be patient... npm WARN old lockfile npm ERR! code ERESOLVE npm ERR! ERESOLVE unable to resolve dependency tree npm ERR! npm ERR! While resolving: bprr-ui@0.0.0 npm ERR! Found: @angular/common@10.1.6 npm ERR! node_modules/@angular/common npm ERR! @angular/common@"~10.1.3" from the root project npm ERR! npm ERR! Could not resolve dependency: npm ERR! peer @angular/common@"13.2.1" from @angular/platform-browser@13.2.1 npm ERR! node_modules/@angular/platform-browser npm ERR! @angular/platform-browser@"~13.2.0" from the root project npm ERR! npm ERR! Fix the upstream dependency conflict, or retry npm ERR! this command with --force, or --legacy-peer-deps npm ERR! to accept an incorrect (and potentially broken) dependency resolution. npm ERR! npm ERR! See /tmp/renovate/cache/others/npm/eresolve-report.txt for a full report. npm ERR! A complete log of this run can be found in: npm ERR! /tmp/renovate/cache/others/npm/_logs/2022-02-08T07_34_00_941Z-debug.log
gharchive/pull-request
2022-02-08T07:34:10
2025-04-01T04:35:13.247335
{ "authors": [ "ngeor" ], "repo": "ngeor/bitbucket-pr-report", "url": "https://github.com/ngeor/bitbucket-pr-report/pull/47", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
636180189
Add option to store UUID in the agent.conf, you need to set store_UUID = True for persistence to work on a container or demo environment (where the underlying instance might be swapped but you want it to report the same on the controller). We should add a variable to set this to true. This solves that issue. Current my workaround is this sudo sed -i.bak 's/store_uuid = False/store_uuid = True/' /etc/controller-agent/agent.conf just added this as default behavior whenever nginx_controller_hostname or nginx_controller_instance_name is defined. With the expectation that if those are not defined and thus get written to the agent.conf via the installer, the hostname might change. Instance_name will always remain fixed and is set at installation. However, nginx_controller_hostname could safely change.
gharchive/issue
2020-06-10T11:46:57
2025-04-01T04:35:13.261095
{ "authors": [ "brianehlert", "magicalyak" ], "repo": "nginxinc/ansible-role-nginx-controller-agent", "url": "https://github.com/nginxinc/ansible-role-nginx-controller-agent/issues/6", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1145032124
feat(cli): add ability to create a store inside a class PR Checklist Please check if your PR fulfills the following requirements: [x] The commit message follows our guidelines: https://github.com/ngneat/elf/blob/master/CONTRIBUTING.md#commit [x] Tests for the changes have been added (for bug fixes / features) [x] Docs have been added / updated (for bug fixes / features) PR Type What kind of change does this PR introduce? [ ] Bugfix [x] Feature [ ] Code style update (formatting, local variables) [ ] Refactoring (no functional changes, no api changes) [ ] Build related changes [ ] CI related changes [x] Documentation content changes [ ] Other... Please describe: What is the current behavior? Issue Number: #162 What is the new behavior? Users can choose whether a store should be created inside a class or not. Does this PR introduce a breaking change? [ ] Yes [x] No Other information @EricPoul what do you say about https://github.com/ngneat/elf/discussions/165 It can be useful for repoTemplate: "function" or for creating a store outside of a class. Inside the class, we create this function anyway. It still takes a lot of space so I'd move it to the private method. Anyway, I don't see this as a bother, especially when it generates not by me. One main con of creating a state separately from the store is the identical names of state and arguments of reducer functions in the class. const { `state`, config } = createState(); const store = new Store({ name: 'todos', state, config }); update(): void { store.update((`state`) => { ...`state`, some: 'new' }) }
gharchive/pull-request
2022-02-20T14:16:12
2025-04-01T04:35:13.285144
{ "authors": [ "EricPoul", "NetanelBasal" ], "repo": "ngneat/elf", "url": "https://github.com/ngneat/elf/pull/167", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2760339144
Lỗi liên quan đến việc chuyển một trường nhập liệu giữa hai trạng thái uncontrolled và controlled trong React Coi thử xem nó là gì và fix được không @ngodat0103 ok da fix xem pull resquest Có lỗi thì fix đi đã, ko được hãy tạo issua @nguyenthiyenly0407
gharchive/issue
2024-12-27T03:22:49
2025-04-01T04:35:13.289956
{ "authors": [ "VietDucc", "ngodat0103", "nguyenthiyenly0407" ], "repo": "ngodat0103/se347", "url": "https://github.com/ngodat0103/se347/issues/102", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
802434859
what's the callback url for OIDC auth? Hi. I'm wanting to set up powerdns admin to work with my orgs oidc provider but i need a callback url and i can't find it anywhere. Could some tell me what it should be? Thanks. It should be address of powerdnsadmin plus /oidc/authorized E.g. if you host pdnsadmin at https://dns.domain.com callback url should be https://dns.domain.com/oidc/authorized
gharchive/issue
2021-02-05T19:56:28
2025-04-01T04:35:13.291568
{ "authors": [ "Prototik", "nktech1135" ], "repo": "ngoduykhanh/PowerDNS-Admin", "url": "https://github.com/ngoduykhanh/PowerDNS-Admin/issues/881", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
285023211
Chuyển tiền và tiền thối Em chưa hiểu về giao dịch và tiền thối lắm, cho em hỏi vd A có 10KCoin, muốn chuyển B 3KCoin thì phải tạo 1 transaction trong đó output gồm 2 phần tử, một cái chứa địa chỉ của B 3KCoin, một cái chứa địa chỉ A 7KCoin phải ko ạ? Và referencedOutputHash là hash của giao dịch gần nhất mà A đã chuyển? A có 10K có nghĩa là A có tổng giá trị transaction ouput chưa sử dụng là 10K Khi A muốn chuyển B 3K thì A phải tìm một số transaction ouput để làm input (dĩ nhiên là chưa dùng) với tổng >= 3K nghĩa là nhiều transaction output (hash + index) chứ không nhất thiết phải là một. Sau đó để output ở transaction bạn tạo: B: 3K A: 7K A ở đây có thể là địa chỉ nào đó của A @nguyenkha cho em hỏi về trường hợp này: Ban đầu addrA và addrB, mỗi địa chỉ được hệ thống KCoin của thầy chuyển vào 10K ở transaction có hash là hash1 Bây giờ muốn chuyển 3K từ addrA vào addrB thì input là 2 phần tử A,B hay chỉ 1 phần tử A, output là [A: 7, B:3] hay [A:7, B:13] ? Input chỉ bao gồm A, Output gồm B (3K) và địa chỉ nhận tiền thối (7K) - không nhất thiết phải là A, mà là của người sở hữu địa chỉ A/địa chỉ của hệ thống của bạn => tùy cách bạn cài đặt hệ thống.
gharchive/issue
2017-12-29T02:29:22
2025-04-01T04:35:13.304266
{ "authors": [ "nguyenkha", "vukhoa30" ], "repo": "nguyenkha/kcoin-blockchain", "url": "https://github.com/nguyenkha/kcoin-blockchain/issues/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
521431435
docs(readme): add section on hot reloading of translations I added a section in the FAQ to explain how to apply hot reloading of translations in an application. Thanks to the explanation of @keradus in #874 Until then...
gharchive/pull-request
2019-11-12T09:32:30
2025-04-01T04:35:13.315376
{ "authors": [ "Odonno" ], "repo": "ngx-translate/core", "url": "https://github.com/ngx-translate/core/pull/1139", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
638406393
Add truncate and ftruncate This change is  here we go
gharchive/pull-request
2020-06-14T18:37:51
2025-04-01T04:35:13.316597
{ "authors": [ "blackheaven" ], "repo": "nh2/hatrace", "url": "https://github.com/nh2/hatrace/pull/81", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
463356969
isAllDay always appears in the popup creation form Version 1.12.1 Development Environment Windows 10.0 build 17134 XAMPP for Windows 5.6.34(Apache 2.4.33 + PHP 7.2.5 + MySQL 5.0.12) No any NodeJS environment. Current Behavior I found it when I trying to modify the example files to fit my requirements. If the "isAllDay" option set to false, the checkbox of "All Day" inside popup creation form seemed not to hide at all. So when someone checked it, everything would stop. Expected Behavior Since the popup creation form layout seemed to be declared inside "calendar.js", so it should be a bug when the "isAllDay" of the calendar was set to false but not removing the checkbox. @LZong-tw The isAllDay option is not an option to control the UI within the creation popup. It determines whether the schedule is an 'All day' check or not. We do not yet support the features you want. If you need it right now, modify the following files and use them. src/js/view/popup/scheduleCreationPopup.js src/js/view/template/popup/scheduleCreationPopup.hbs
gharchive/issue
2019-07-02T17:46:13
2025-04-01T04:35:13.319910
{ "authors": [ "LZong-tw", "jungeun-cho" ], "repo": "nhn/tui.calendar", "url": "https://github.com/nhn/tui.calendar/issues/342", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
307894501
feat: New Design Version v0.8.0 Development Environment Any Current Behavior Expected Behavior Apply new design. Finished
gharchive/issue
2018-03-23T03:56:16
2025-04-01T04:35:13.321897
{ "authors": [ "dongsik-yoo" ], "repo": "nhnent/tui.calendar", "url": "https://github.com/nhnent/tui.calendar/issues/31", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1717366333
examples: Fix nidaqmx mypy errors (except aio-related errors) What does this Pull Request accomplish? Fix non-aio-related mypy errors for nidaqmx examples. The currently released version of mypy-protobuf (3.4.0) doesn't support aio yet, but it looks like support will be available in a future release. Why should this Pull Request be merged? Demonstrate type-checking best practices (I hope). Prerequisite for updating the CI workflow to validate nidaqmx examples in https://github.com/ni/grpc-device/pull/931 What testing has been done? Ran validate_examples.py -p nidaqmx before and after. Before: examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:116: error: "None" has no attribute "GetErrorString" [attr-defined] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:126: error: Module has no attribute "aio" [attr-defined] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:131: error: Incompatible types in "await" (actual type "Union[GetTaskAttributeUInt32Response, Any]", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:131: error: Item "None" of "Optional[NiDAQmxStub]" has no attribute "GetTaskAttributeUInt32" [union-attr] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:140: error: Incompatible types in "await" (actual type "Union[GetNthTaskDeviceResponse, Any]", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:140: error: Item "None" of "Optional[NiDAQmxStub]" has no attribute "GetNthTaskDevice" [union-attr] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:145: error: Incompatible types in "await" (actual type "Union[GetDeviceAttributeInt32Response, Any]", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:145: error: Item "None" of "Optional[NiDAQmxStub]" has no attribute "GetDeviceAttributeInt32" [union-attr] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:170: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:174: error: Incompatible types in "await" (actual type "CreateAIVoltageChanResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:184: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:194: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:198: error: Incompatible types in "await" (actual type "CreateAOVoltageChanResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:207: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:217: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:221: error: Incompatible types in "await" (actual type "CreateAIVoltageChanResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:231: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:241: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:245: error: Incompatible types in "await" (actual type "CreateAOVoltageChanResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:254: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:269: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:276: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:283: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:290: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:303: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:310: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:317: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:328: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:335: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:342: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:354: error: Incompatible types in "await" (actual type "CfgDigEdgeStartTrigResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:361: error: Incompatible types in "await" (actual type "CfgDigEdgeStartTrigResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:368: error: Incompatible types in "await" (actual type "CfgDigEdgeStartTrigResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:380: error: Incompatible types in "await" (actual type "WriteAnalogF64Response", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:392: error: Incompatible types in "await" (actual type "WriteAnalogF64Response", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:415: error: Incompatible types in "await" (actual type "Tuple[Tuple[str, Union[str, bytes]], ...]", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:439: error: Incompatible types in "await" (actual type "Tuple[Tuple[str, Union[str, bytes]], ...]", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:442: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:446: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:450: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:455: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:462: error: "CallIterator[RegisterEveryNSamplesEventResponse]" has no attribute "__aiter__" (not async iterable) [attr-defined] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:464: error: Incompatible types in "await" (actual type "Union[ReadAnalogF64Response, Any]", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:464: error: Item "None" of "Optional[NiDAQmxStub]" has no attribute "ReadAnalogF64" [union-attr] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:477: error: Incompatible types in "await" (actual type "Union[ReadAnalogF64Response, Any]", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:477: error: Item "None" of "Optional[NiDAQmxStub]" has no attribute "ReadAnalogF64" [union-attr] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:511: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:515: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:519: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:523: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:531: error: Cannot determine type of "key" [has-type] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:532: error: Cannot determine type of "value" [has-type] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:550: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:553: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:556: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:559: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\digital-output.py:75: error: Incompatible types in assignment (expression has type "WriteDigitalU32Response", variable has type "CreateTaskResponse") [assignment] examples\nidaqmx\digital-output.py:93: error: Cannot determine type of "key" [has-type] examples\nidaqmx\digital-output.py:94: error: Cannot determine type of "value" [has-type] examples\nidaqmx\digital-input.py:75: error: Incompatible types in assignment (expression has type "ReadDigitalU32Response", variable has type "CreateTaskResponse") [assignment] examples\nidaqmx\digital-input.py:88: error: "CreateTaskResponse" has no attribute "read_array" [attr-defined] examples\nidaqmx\digital-input.py:92: error: Cannot determine type of "key" [has-type] examples\nidaqmx\digital-input.py:93: error: Cannot determine type of "value" [has-type] examples\nidaqmx\counter-output.py:92: error: Cannot determine type of "key" [has-type] examples\nidaqmx\counter-output.py:93: error: Cannot determine type of "value" [has-type] examples\nidaqmx\counter-input.py:83: error: Incompatible types in assignment (expression has type "ReadCounterScalarF64Response", variable has type "CreateTaskResponse") [assignment] examples\nidaqmx\counter-input.py:90: error: "CreateTaskResponse" has no attribute "value" [attr-defined] examples\nidaqmx\counter-input.py:94: error: Cannot determine type of "key" [has-type] examples\nidaqmx\counter-input.py:95: error: Cannot determine type of "value" [has-type] examples\nidaqmx\analog-output.py:96: error: Cannot determine type of "key" [has-type] examples\nidaqmx\analog-output.py:97: error: Cannot determine type of "value" [has-type] examples\nidaqmx\analog-input-every-n-samples-aio.py:54: error: Module has no attribute "aio" [attr-defined] examples\nidaqmx\analog-input-every-n-samples-aio.py:61: error: Item "None" of "Optional[NiDAQmxStub]" has no attribute "GetErrorString" [union-attr] examples\nidaqmx\analog-input-every-n-samples-aio.py:68: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:73: error: Incompatible types in "await" (actual type "CreateAIVoltageChanResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:86: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:106: error: Incompatible types in "await" (actual type "Tuple[Tuple[str, Union[str, bytes]], ...]", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:112: error: Incompatible types in "await" (actual type "Tuple[Tuple[str, Union[str, bytes]], ...]", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:114: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:117: error: Incompatible types in "await" (actual type "GetTaskAttributeUInt32Response", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:129: error: "CallIterator[RegisterEveryNSamplesEventResponse]" has no attribute "__aiter__" (not async iterable) [attr-defined] examples\nidaqmx\analog-input-every-n-samples-aio.py:131: error: Incompatible types in "await" (actual type "Union[ReadAnalogF64Response, Any]", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:131: error: Item "None" of "Optional[NiDAQmxStub]" has no attribute "ReadAnalogF64" [union-attr] examples\nidaqmx\analog-input-every-n-samples-aio.py:158: error: "CallIterator[RegisterDoneEventResponse]" has no attribute "__aiter__" (not async iterable) [attr-defined] examples\nidaqmx\analog-input-every-n-samples-aio.py:168: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:174: error: Cannot determine type of "key" [has-type] examples\nidaqmx\analog-input-every-n-samples-aio.py:175: error: Cannot determine type of "value" [has-type] examples\nidaqmx\analog-input-every-n-samples-aio.py:186: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples.py:63: error: Item "None" of "Optional[NiDAQmxStub]" has no attribute "GetErrorString" [union-attr] examples\nidaqmx\analog-input-every-n-samples.py:132: error: Item "None" of "Optional[NiDAQmxStub]" has no attribute "ReadAnalogF64" [union-attr] examples\nidaqmx\analog-input-every-n-samples.py:183: error: Cannot determine type of "key" [has-type] examples\nidaqmx\analog-input-every-n-samples.py:184: error: Cannot determine type of "value" [has-type] examples\nidaqmx\analog-input.py:93: error: Incompatible types in assignment (expression has type "GetTaskAttributeUInt32Response", variable has type "CreateTaskResponse") [assignment] examples\nidaqmx\analog-input.py:98: error: "CreateTaskResponse" has no attribute "value" [attr-defined] examples\nidaqmx\analog-input.py:100: error: Incompatible types in assignment (expression has type "ReadAnalogF64Response", variable has type "CreateTaskResponse") [assignment] examples\nidaqmx\analog-input.py:113: error: "CreateTaskResponse" has no attribute "read_array" [attr-defined] examples\nidaqmx\analog-input.py:114: error: "CreateTaskResponse" has no attribute "samps_per_chan_read" [attr-defined] examples\nidaqmx\analog-input.py:119: error: Cannot determine type of "key" [has-type] examples\nidaqmx\analog-input.py:120: error: Cannot determine type of "value" [has-type] Found 99 errors in 9 files (checked 9 source files) After: examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:127: error: Module has no attribute "aio" [attr-defined] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:132: error: Incompatible types in "await" (actual type "GetTaskAttributeUInt32Response", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:141: error: Incompatible types in "await" (actual type "GetNthTaskDeviceResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:146: error: Incompatible types in "await" (actual type "GetDeviceAttributeInt32Response", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:170: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:174: error: Incompatible types in "await" (actual type "CreateAIVoltageChanResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:184: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:194: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:198: error: Incompatible types in "await" (actual type "CreateAOVoltageChanResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:207: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:217: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:221: error: Incompatible types in "await" (actual type "CreateAIVoltageChanResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:231: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:241: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:245: error: Incompatible types in "await" (actual type "CreateAOVoltageChanResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:254: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:269: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:276: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:283: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:290: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:303: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:310: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:317: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:328: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:335: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:342: error: Incompatible types in "await" (actual type "SetTimingAttributeStringResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:354: error: Incompatible types in "await" (actual type "CfgDigEdgeStartTrigResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:361: error: Incompatible types in "await" (actual type "CfgDigEdgeStartTrigResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:368: error: Incompatible types in "await" (actual type "CfgDigEdgeStartTrigResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:380: error: Incompatible types in "await" (actual type "WriteAnalogF64Response", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:392: error: Incompatible types in "await" (actual type "WriteAnalogF64Response", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:415: error: Incompatible types in "await" (actual type "Tuple[Tuple[str, Union[str, bytes]], ...]", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:439: error: Incompatible types in "await" (actual type "Tuple[Tuple[str, Union[str, bytes]], ...]", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:442: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:446: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:450: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:455: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:463: error: "CallIterator[RegisterEveryNSamplesEventResponse]" has no attribute "__aiter__" (not async iterable) [attr-defined] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:464: error: Incompatible types in "await" (actual type "ReadAnalogF64Response", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:475: error: Incompatible types in "await" (actual type "ReadAnalogF64Response", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:508: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:512: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:516: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:520: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:527: error: "_Metadatum" object is not iterable [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:528: error: Cannot determine type of "key" [has-type] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:529: error: Cannot determine type of "value" [has-type] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:547: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:550: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:553: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\dsa-shared-timebase-and-trig-analog-input-and-output-aio.py:556: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:51: error: Module has no attribute "aio" [attr-defined] examples\nidaqmx\analog-input-every-n-samples-aio.py:66: error: Incompatible types in "await" (actual type "CreateTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:69: error: Incompatible types in "await" (actual type "CreateAIVoltageChanResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:82: error: Incompatible types in "await" (actual type "CfgSampClkTimingResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:102: error: Incompatible types in "await" (actual type "Tuple[Tuple[str, Union[str, bytes]], ...]", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:108: error: Incompatible types in "await" (actual type "Tuple[Tuple[str, Union[str, bytes]], ...]", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:110: error: Incompatible types in "await" (actual type "StartTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:113: error: Incompatible types in "await" (actual type "GetTaskAttributeUInt32Response", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:125: error: "CallIterator[RegisterEveryNSamplesEventResponse]" has no attribute "__aiter__" (not async iterable) [attr-defined] examples\nidaqmx\analog-input-every-n-samples-aio.py:126: error: Incompatible types in "await" (actual type "ReadAnalogF64Response", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:152: error: "CallIterator[RegisterDoneEventResponse]" has no attribute "__aiter__" (not async iterable) [attr-defined] examples\nidaqmx\analog-input-every-n-samples-aio.py:162: error: Incompatible types in "await" (actual type "StopTaskResponse", expected type "Awaitable[Any]") [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:167: error: "_Metadatum" object is not iterable [misc] examples\nidaqmx\analog-input-every-n-samples-aio.py:168: error: Cannot determine type of "key" [has-type] examples\nidaqmx\analog-input-every-n-samples-aio.py:169: error: Cannot determine type of "value" [has-type] examples\nidaqmx\analog-input-every-n-samples-aio.py:180: error: Incompatible types in "await" (actual type "ClearTaskResponse", expected type "Awaitable[Any]") [misc] Found 67 errors in 2 files (checked 9 source files)
gharchive/pull-request
2023-05-19T14:40:12
2025-04-01T04:35:13.352332
{ "authors": [ "bkeryan" ], "repo": "ni/grpc-device", "url": "https://github.com/ni/grpc-device/pull/930", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1058006854
"Scripting Tool" source to be relinked to "grpc-labview-support" source for dev purpose Currently, "grpc-labview-support" source is built and deployed under the vi.lib as "grpc-labview-support-release" , and the "scripting tool" is dependent on the released version of "grpc-labview-support-release". So, when the "scripting tool" is deployed into vi.lib as "scripting tool release", it will maintain the dependency from the vi.lib for "grpc-labview-support-release". This might make the release process simple. But the development process needs clear documentation so the developers know to treat the source as two different projects and make sure the "grpc-labview-support" changes are installed before making changes to the "scripting tool" source. During the last discussion with @ccifra , the decision was to make the "scripting tool" source dependent on the "grpc-labview-support" source, so that the development process is easier, and cifra will take careful action during the release in his end. @ccifra I assume there are a few folks who are making changes to the repo and this change might be another merge issue headache. Since, people are working through this, we could make this change before a release or when there is downtime in development? let me know when to plan this.
gharchive/issue
2021-11-19T00:42:30
2025-04-01T04:35:13.355628
{ "authors": [ "navinsubramani" ], "repo": "ni/grpc-labview", "url": "https://github.com/ni/grpc-labview/issues/40", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2730180481
Site tweaks to improve strict CSP hosting Pull Request 🤨 Rationale The reason for this PR was trying to evaluate Nimble-based apps hosting in Web App hosting under our strict CSP policy. This was done by exercising the nimble site build which covers the following: Angular, Blazor, Vite, Storybook. In doing so identified some changes that are either useful or minimally invasive as workarounds for WebApp hosting issues or improving the experience of Nimble site if hosted in Web App hosting. 👩‍💻 Implementation Vite: A minimal change to provide a relative base configuration for vite was needed. This causes vite to change from paths relative to root (/script.js) to relative paths in the current directory (./script.js). Reason is WebApps are not hosted in root but from a subdirectory path. Angular: A minimal change to disable inlineCritical styles is needed so that stylesheets included in angular.json load correctly (which seems to be something we figured out and inherited in all apps but doesn't seem to be documented anywhere 🤷 it's like a herd immunity inherited in copy-paste generations 💉). Reason is described in blog / we don't allow unsafe-inline. Blazor: During investigation found that Blazor does not like to serve from index.html urls and relies on path urls. You can actually see it on the current published site if you use a blazor index.html style url instead of a directory url. The page fails to load and has many console errors as Blazor does poor URL parsing / manipulation to load JS resources. It's probably worth creating an issue but I did not yet. I thought of a workaround for nimble site by specifying a base url manually of <base href="./" />. This is not compatible with our strict CSP setting and is ignored due to base-uri: 'none';'' (the OWASP strict policy example) causing the errors to continue. We potentially could switch our CSP to base-uri: 'self' but it's not clear to me what the security implications are and I'm not recommending it yet (see following). I also found that Blazor seems generally against the idea of serving off index.html vs the directory path as index.html resolution is not supported out of the box in the Router either. I thought of a Router workaround for index.html resolution by serving the same component from both the path url and the index.html url. It seems to work but I'm not aware of any other concerns relying on that workaround. Even with the above a Blazor 8 app using Nimble will still not run as some Blazor 8 template binding features require using eval and are only addressed in Blazor 9. Based on the above, while I'll propose the changes for Nimble's Blazor site page to support index.html based urls working, I'm not going to propose changes to Web App hosting strict CSP base-uri configuration to support the workaround described above. Instead, I think we should say Nimble Blazor WebApp hosting is contingent on Blazor 9 support in Nimble and fixing path serving in Web App hosting. I don't think we should recommend the workarounds I figured out above to Blazor WebApp devs as from the linked issues Blazor does not seem interested in supporting that pattern. Storybook: Depends on unsafe-inline (which we do not allow) and they seem resistant to the change. Need to reply on the issue with a convincing discussion, not sure if MDN and OWASP recommendations are sufficient for them. Did not provide a comment / create an issue. All: Updated links to point to index.html paths as it's minimally invasive (makes the URL bar uglier) but works around AzDo 2941644. This could be reverted in Nimble once the WebApp directory hosting issue is address. Updated each page to have a link to the parent page to make them easier to navigate between when hosted in the WebApp hosting iframe. 🧪 Testing Manual and via built storybook. Also example hosted on dev: Landing (vite) Wafer Performance (vite) Angular Blazor broken, see above Storybook expected to be broken, see above (can't explicitly test as deleted from dev package due to size upload bug) ✅ Checklist [x] I have updated the project documentation to reflect my changes or determined no changes are needed. @jattasNI @msmithNI @atmgrifter00 these are minor non-source impacting changes just to the site so bypassing but can address concerns in a follow-up
gharchive/pull-request
2024-12-10T13:49:46
2025-04-01T04:35:13.370020
{ "authors": [ "rajsite" ], "repo": "ni/nimble", "url": "https://github.com/ni/nimble/pull/2492", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1971571849
drag and drop functionality in tags to form a sentence Please add drag and drop functionality in tags to form a sentence and change the words. words as tags and dragged and drop to form a sentence. Thanks Sai I am not very clear about your needs, could you please provide more detailed instructions, such as GIFs, web pages and so on.
gharchive/issue
2023-11-01T02:16:52
2025-04-01T04:35:13.379112
{ "authors": [ "nicedouble", "saisaigraph" ], "repo": "nicedouble/StreamlitAntdComponents", "url": "https://github.com/nicedouble/StreamlitAntdComponents/issues/27", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2442086752
ytmapi_rs: Remove raw_query and processed_query functions from public API. Replace with more ergonomic alternatives without lifetime requirements. e.g deserialize_query that will return T = DeserializeOwned and string_query that will return pretty printed. Since RawResult and ProcessedResult need to be a part of the public API for AuthToken and ParseFrom traits, may not be worth doing this. Closing due to above comment
gharchive/issue
2024-08-01T10:26:22
2025-04-01T04:35:13.413998
{ "authors": [ "nick42d" ], "repo": "nick42d/youtui", "url": "https://github.com/nick42d/youtui/issues/112", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
203915694
Added Boss, Skill Points, and fixed some minor text Added some things and fixed some things. Planning on adding other types of monsters later. I'll check it out tomorrow and some things might need to change but that won't be until the story gets added. I'll also need to reorganize the files for the boss and other enemies. The "Gaben boss" might become an Easter egg/rare boss but I would like to thank you for your work. I'm currently stuck doing exams so it is hard for me to work on my projects. Just an FYI, you can also update the README.md with your new commands or I can do it for you later. Also just a tip, try to comment your code. We are all guilty of this and I'll comment it more once exams are over but it will be better for people who just saw this project to understand it. Sorry, pressed wrong button but if you can, try to change them. Alright
gharchive/pull-request
2017-01-30T02:12:04
2025-04-01T04:35:13.433689
{ "authors": [ "KingofKFCJamal", "nickthegamer5" ], "repo": "nickthegamer5/SteamBattleBot", "url": "https://github.com/nickthegamer5/SteamBattleBot/pull/5", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2049824271
Add prettier editor config for code consistency My aim initially was to simply investigate. When I saved a file I noticed my editor removed all semicolons as it is my personal preference. I decided it would be useful to prevent editors like mine from transforming files on save and keep code as consistent as possible for future contributors on this very useful project. Great shout - thanks!
gharchive/pull-request
2023-12-20T04:47:18
2025-04-01T04:35:13.436186
{ "authors": [ "josephdburdick", "nicoalbanese" ], "repo": "nicoalbanese/kirimase", "url": "https://github.com/nicoalbanese/kirimase/pull/110", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2186153265
Initialize db if it doesn't exist instead of throwing an error running docker-compose -f ./docker/docker-compose.yml run app gives me the following output: Traceback (most recent call last): File "/commandjobs/src/menu.py", line 429, in <module> curses.wrapper(main) File "/usr/local/lib/python3.12/curses/__init__.py", line 94, in wrapper return func(stdscr, *args, **kwds) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/commandjobs/src/menu.py", line 425, in main app = MenuApp(stdscr, logger) ^^^^^^^^^^^^^^^^^^^^^^^ File "/commandjobs/src/menu.py", line 54, in __init__ self.db_manager = DatabaseManager(self.db_path) # Specify the path ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/commandjobs/src/database_manager.py", line 6, in __init__ self.conn = sqlite3.connect(db_path) ^^^^^^^^^^^^^^^^^^^^^^^^ sqlite3.OperationalError: unable to open database file I can see the path to database file in the .env file, but that path does not exist by default. I created repo/job_listings.db in the base directory of the repo, but the error remains. here is what I did: mkdir repo touch repo/job_listings.db docker-compose -f docker/docker-compose.yml build docker-compose -f docker/docker-compose.yml run app Hello @hammadfauz, thank you for reporting this. Sorry about that. You are right, the app should just use a default db name and initialize it Recently did some changes on the paths and forgot to update the config/sample.env file with the proper ones Just pushed this PR that fixes it https://github.com/nicobrenner/commandjobs/pull/42 You could git pull on your local repo and copy the sample.env file again, or if you prefer, you can just manually remove the /repo/ part of the path of the BASE_RESUME and DB_PATH variables
gharchive/issue
2024-03-14T11:51:56
2025-04-01T04:35:13.439741
{ "authors": [ "hammadfauz", "nicobrenner" ], "repo": "nicobrenner/commandjobs", "url": "https://github.com/nicobrenner/commandjobs/issues/41", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2195914228
Packaging/windows About This PR adds makefiles to build Window MSI installers. The versioning is tied to the latest git-tag. Notably, this does not include the necessary updates to create patch files; Windows has a pretty serious disconnect between Installation packages and Patch packages. This means that if someone wants to upgrade, they would need to explicitly uninstall the existing lfs-s3 and then reinstall the new one. https://github.com/nicolas-graves/lfs-s3/issues/60 @Bekreth I was able to left a review with a single question, thanks for explaining everything. Thanks again @Bekreth! How about also adding a word in the README for Windows users about how to install/use it on Windows? It could be just one line in the Download & Install section.
gharchive/pull-request
2024-03-19T20:15:28
2025-04-01T04:35:13.445834
{ "authors": [ "Bekreth", "nicolas-graves" ], "repo": "nicolas-graves/lfs-s3", "url": "https://github.com/nicolas-graves/lfs-s3/pull/64", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
345372528
openApi Applying security Hello, Today I learned about defining securitySchemes (step 1 in https://swagger.io/docs/specification/authentication/): openApi = OpenApi().apply { ... components.securitySchemes["BearerAuth"] = mapOf( "type" to "http", "scheme" to "bearer" ) // paths["/corporations"]["security"] = Operation.create(...) ... } How can I apply security (step 2) to a path (e.g. "/corporations")? Thanks Hi in0rdr, currently you can't, but feel free to contribute with a pull request. regards Niels
gharchive/issue
2018-07-27T21:32:20
2025-04-01T04:35:13.457798
{ "authors": [ "in0rdr", "nielsfalk" ], "repo": "nielsfalk/ktor-swagger", "url": "https://github.com/nielsfalk/ktor-swagger/issues/13", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2075879415
Update models Summary update models LGTM
gharchive/pull-request
2024-01-11T06:40:48
2025-04-01T04:35:13.459202
{ "authors": [ "aokumasan", "fuku2014" ], "repo": "nifcloud/nifcloud-cli", "url": "https://github.com/nifcloud/nifcloud-cli/pull/32", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1823085869
NgHttpCachingBrowserStorage based storage does not work correctly When I try to configure ng-http-caching to use Local or Session storage, the cache is being populated, but the results are never served from the cache. A new HTTP request is always triggered. I believe its because the 'version' field is not being set when writing to the cache. https://github.com/nigrosimone/ng-http-caching/blob/72b896306d36abee0b28b2ec90ac51d225b7cd23/projects/ng-http-caching/src/lib/storage/ng-http-caching-browser-storage.ts#L131 Hi, thanks! I have fixed in 16.0.5 just published on NPM. Now, it works as expected? Thanks, it works now.
gharchive/issue
2023-07-26T20:07:51
2025-04-01T04:35:13.500630
{ "authors": [ "bryhemm", "nigrosimone" ], "repo": "nigrosimone/ng-http-caching", "url": "https://github.com/nigrosimone/ng-http-caching/issues/9", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
936217971
🛑 RFH Status is down In d1e8195, RFH Status (https://rfh-koeln.sciebo.de/) was down: HTTP code: 503 Response time: 801 ms Resolved: RFH Status is back up in 0abcc90.
gharchive/issue
2021-07-03T09:45:07
2025-04-01T04:35:13.528713
{ "authors": [ "niklasmtj" ], "repo": "niklasmtj/sciebo-check", "url": "https://github.com/niklasmtj/sciebo-check/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
185705442
fix(thread_local): Removes unnecessary semicolon Not sure if this is explicitly a rust 1.9.0 thing or not Oh, interesting, the semicolon was only added to the thread_local! pattern in 1.11.0. It's completely optional though, so your fix is fine. Maybe add 1.9.0 to .travis.yml? That way we always know the required minimum, and make sure that it works. This is the true minimum right now too, for std::panic support. I don't know if @nikomatsakis has a policy on rustc versions being a breaking change -- opinions vary and I'm in the camp that it is breaking. Either way, it's good to have that in CI so we know what works right now. @cuviper I'm not sure if I've formed a strong opinion, but I do expect users to basically keep up with stable --- that is to say, if there is a new stable feature, I wouldn't hesitate to add a dependency on it. But it seems reasonable to consider that a breaking change, at the same time. In any case, I always forget that ; isn't allowed in thread_local!.
gharchive/pull-request
2016-10-27T15:40:32
2025-04-01T04:35:13.531575
{ "authors": [ "cuviper", "durango", "nikomatsakis" ], "repo": "nikomatsakis/rayon", "url": "https://github.com/nikomatsakis/rayon/pull/126", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
182448167
readme update for npm I was reading https://github.com/nilbus/Backbone.dualStorage/pull/141 and realized it might not be clear how to add this via npm if it isnt registered with https://npmjs.com. This will add "Backbone.dualStorage": "github:nilbus/backbone.dualstorage" to your package.json's dependencies and install the node module. Thanks!
gharchive/pull-request
2016-10-12T06:47:59
2025-04-01T04:35:13.534902
{ "authors": [ "digitalvapor", "nilbus" ], "repo": "nilbus/Backbone.dualStorage", "url": "https://github.com/nilbus/Backbone.dualStorage/pull/153", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1311989313
Enable downstream dependency analysis with updated javadoc This PR enables Annotator to process downstream dependencies while making decisions for public methods with non-primitive return types. This PR is a follow up for #624 Before starting the main process, annotator will do the followings: Collects public method with non-primitive-return-type Collects regions in downstream dependencies that will potentially introduce a new error if a method in target module is annotated. Constructs the conflict graph and computes the effect in the collected regions for each method. Aggregates the result of analysis on each sub-module and makes a uniformed report. Stores the effect on submodule and will provide the information to main process while making decisions. To activate this feature, flags below must be passed to Annotator: -ddbc or --downstream-dependencies-build-command list of commands to run downstream dependencies separated by comma, -nlmlp or --nullaway-library-model-loader-path path to library model loader for NullAway. This PR also adds javadoc on huge segment of the code. @lazaroclapp @ketkarameya I removed the request for review on this PR since I haven't added the proper javadoc on this PR methods. Will request a review as soon as I add the javadocs on this :) Quick high-level question. Based on the description, this is focused on checking when making a return type @Nullable will introduce errors to a downstream target. What about the case of errors in a downstream target due to a parameter not being @Nullable? Isn't it possible we would want to introduce some @Nullable parameter annotations based on downstream targets? Quick high-level question. Based on the description, this is focused on checking when making a return type @Nullable will introduce errors to a downstream target. What about the case of errors in a downstream target due to a parameter not being @Nullable? Isn't it possible we would want to introduce some @Nullable parameter annotations based on downstream targets? @msridhar Hi Manu, sure that’s also a useful information from downstream dependencies. This PR is mostly focused on preparing the infrastructure and pipelines to enable this process. Once this lands, we can add features as such incrementally in followup PRs :) Sorry I'm on holiday until August 1. I can try to look after that For the future: for ease of review, a PR should not make a lot of unrelated changes. The ideal is one PR per feature and separate PRs for specific kinds of refactorings. This PR was both way too large and attempting to do too many different things for me to truly review it properly. This is my mistake as well, since I told you to go ahead with javadocs, but I thought you meant on the new feature for this PR, not on unrelated files. Ideally, it should have been one PR to implement downstream dependency analysis and then one or more separate PRs doing refactorings/documentation, I think. That said, no need to break it down at this point. I did what I could to review the full PR. In the end, I skimmed some of the refactorings+javadoc changes, particularly around index/ and trackers/ in the core. I did go over submodules/ and the changes to injector/ a bit more carefully, though. Other than that, one thing that definitely stands out to me, though, is that all test changes seem to be due to an unrelated refactoring. I see no tests for the new functionality. How feasible is to add a multi-target test as a sanity check here? (If not very easy, maybe on Wednesday we can discuss the state of any internal manual testing? And leave automated testing for a follow up PR?) Edit: I see some testing infra added, but it happened in the middle of my review pass, will check that on the next pass. At this point, please avoid changes within this PR that aren't needed to address the comments or to test/fix the downstream target analysis feature, @nimakarimipour (happy to see further refactors/documentation improvements, but on new PRs!) Hi @lazaroclapp @ketkarameya, sorry for the inconveniences this PR caused in the review process. I will definitely keep that in mind for future PRs. While I was adding javadoc to the code, in many cases, it was required to also add documents to the files that are not necessarily changed but used in the current class. This lead to documenting large segment of the code. Also during adding documents I understood some parts could have been written better and did the refactoring to speed up the review process which now I know I should have created multiple PRs with better order of processing. Will definitely do that for all future PRs. For now as you said, I will only address comments and keep all future changes (such as completing test infra) in future PRs. Please let me know if you think I should break this PR in future. @ketkarameya @lazaroclapp This PR is ready for another review. @lazaroclapp @ketkarameya Thank you for your comments, I will add follow up PRs to complete all required needs.
gharchive/pull-request
2022-07-20T21:53:46
2025-04-01T04:35:13.695334
{ "authors": [ "msridhar", "nimakarimipour" ], "repo": "nimakarimipour/NullAwayAnnotator", "url": "https://github.com/nimakarimipour/NullAwayAnnotator/pull/26", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
197451637
Dark theme with textfield textcolor = white Hi - just stumbled upon this and running it through some tests and so far love it! Very good job. I'm not a good dev can get by - but not good enough to pull, figure out and fix. So if you don't mind I'll just make some suggestion here as I go along and see if they make sense to you. I like your dark theme - good thinking there. I'm trying a textfield in it and the text is black so it is hard to see in a dark text field. @rjpalermo1 Good catch! I will fix this ASAP and let you know once it's fixed. Hi @rjpalermo1, Thanks again for bringing this to my attention. I've added a fix for this in V1.2.9. Cheers 👍
gharchive/issue
2016-12-24T06:48:48
2025-04-01T04:35:13.698101
{ "authors": [ "nimati", "rjpalermo1" ], "repo": "nimati/FCAlertView", "url": "https://github.com/nimati/FCAlertView/issues/35", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
408310396
Make all requests reloadable Requests which render their DOM themselves (instead of having it predefined in HTML) cannot be reloaded properly. The hash-targeted element is created after the browser tries to focus it. Requests with that problem: export, ...? Solved by #174
gharchive/issue
2019-02-08T20:24:23
2025-04-01T04:35:13.707049
{ "authors": [ "Bettelstab" ], "repo": "nimiq/keyguard-next", "url": "https://github.com/nimiq/keyguard-next/issues/172", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
267215645
cowboy_clock crashes during lookup of rfc1123 I get below error log intermittently, possibly during high load in the system. 2017-10-20 04:56:50.535 [error] <0.720.0> Lager event handler error_logger_lager_h exited with reason {'EXIT',{{case_clause,[http,<0.4193.0>,1,<0.4194.0>,badarg,[{ets,lookup_element,[cowboy_clock,rfc1123,2],[]},{cowboy_clock,rfc1123,0,[{file,"/drone/src/code.xaptum.xyz/xaptum-development/xmx/_build/default/lib/cowboy/src/cowboy_clock.erl"},{line,54}]},{cowboy_req,response_headers,2,[{file,"/drone/src/code.xaptum.xyz/xaptum-development/xmx/_build/default/lib/cowboy/src/cowboy_req.erl"},{line,778}]},{cowboy_req,do_reply,4,[{file,"/drone/src/code.xaptum.xyz/xaptum-development/xmx/_build/defau..."},...]},...]]},...}} 2017-10-20 04:56:50.588 [error] <0.4197.0>@cowboy_clock:rfc1123:54 CRASH REPORT Process <0.4197.0> with 0 neighbours crashed with reason: bad argument in call to ets:lookup_element(cowboy_clock, rfc1123, 2) in cowboy_clock:rfc1123/0 line 54 2017-10-20 04:56:50.588 [error] <0.720.0> Lager event handler error_logger_lager_h exited with reason {'EXIT',{{case_clause,[http,<0.4196.0>,1,<0.4197.0>,badarg,[{ets,lookup_element,[cowboy_clock,rfc1123,2],[]},{cowboy_clock,rfc1123,0,[{file,"/drone/src/code.xaptum.xyz/xaptum-development/xmx/_build/default/lib/cowboy/src/cowboy_clock.erl"},{line,54}]},{cowboy_req,response_headers,2,[{file,"/drone/src/code.xaptum.xyz/xaptum-development/xmx/_build/default/lib/cowboy/src/cowboy_req.erl"},{line,778}]},{cowboy_req,do_reply,4,[{file,"/drone/src/code.xaptum.xyz/xaptum-development/xmx/_build/defau..."},...]},...]]},...}}``` Is there anything else? Because I think this could happen for one of two cases: the cowboy_clock process has crashed and is being restarted; the ets table is temporarily gone there is a bug in ets and the writes are not atomic Would be good if you can confirm whether cowboy_clock restarted (I guess the pid number can give a good hint) and if it has, see if you can find the crash reason. Hi Irina, I have the same issue when I did not start cowboy as application. The cowboy_clock process is started as worker by cowboy_sup that is started by cowboy_app. I hope it will help in your debugging. Regards, Alexei. Hi Alexei, Thanks so much! I have cowboy app running for sure as it's working for me all along. However, I saw this error again and this time I was mindful of what you said above and realized that this happened while my server was being restarted (obviously along with all the other apps restarting with it -- so looks like something is still trying to use cowboy when it's already stopped/stopping and that's why I get this error). I can't be sure now about the other times I was seeing it, maybe it was always this situation. I'll keep an eye on it to see if I can reproduce when the server isn't restarting, but I suspect now that this was probably always the case. Cheers, Irina. On Sun, Oct 22, 2017 at 11:08 PM, alekras notifications@github.com wrote: Hi Irina, I have the same issue when I did not start cowboy as application. The cowboy_clock process is started as worker by cowboy_sup that is started by cowboy_app. I hope it will help in your debugging. Regards, Alexei. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/ninenines/cowboy/issues/1228#issuecomment-338542884, or mute the thread https://github.com/notifications/unsubscribe-auth/ALcepvOrNw0OFyi69vsi0D-kk2XtcWABks5svBFFgaJpZM4QA15J . I can see that happening if Ranch supervises the connections because those will still run while Cowboy is being restarted, and if Cowboy is not there then cowboy_clock isn't either, so the ets table doesn't exist (anymore). I suppose a graceful way to deal with that is to catch the error and then compute the value directly if the ets table is gone. Hi Loïc, I've been watching our logs and seems like it does happen upon server exit. So I guess it isn't really an issue. Thanks so much for your help, Irina. On Mon, Oct 23, 2017 at 11:01 AM, Loïc Hoguin notifications@github.com wrote: I can see that happening if Ranch supervises the connections because those will still run while Cowboy is being restarted, and if Cowboy is not there then cowboy_clock isn't either, so the ets table doesn't exist (anymore). I suppose a graceful way to deal with that is to catch the error and then compute the value directly if the ets table is gone. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/ninenines/cowboy/issues/1228#issuecomment-338708338, or mute the thread https://github.com/notifications/unsubscribe-auth/ALcepieu9wPpm-5c-0AXFCSEfrsYKDcmks5svLhSgaJpZM4QA15J . Cheers, good to know! I will still do the proposed change in my previous comment so that these logs don't appear anymore and to not have connections get killed because of this. Fixed locally, this will be part of 2.1. Thanks! Awesome! Thanks so much Loïc!
gharchive/issue
2017-10-20T15:35:00
2025-04-01T04:35:13.720101
{ "authors": [ "alekras", "essen", "iguberman" ], "repo": "ninenines/cowboy", "url": "https://github.com/ninenines/cowboy/issues/1228", "license": "isc", "license_type": "permissive", "license_source": "bigquery" }
526573969
Required has_body missing for response with code = 400 Here is an error message from log: Log access failed for: [400, #{<<"connection">> => <<"close">>,<<"content-length">> => <<"0">>}, #{headers => #{<<"accept">> => <<"image/gif, i"...>>,<<"accept-chars"...>> => <<"iso-8859-1,*"...>>,<<"accept-langu"...>> => <<"en">>,<<"cache-co"...>> => <<"no-cache">>,...},method => <<"GET">>,path => <<"/acute-cp/">>,peer => {{10756,18944,...},48951},...}] with: error:function_clause stacktrace: in call to cowboy_req:has_body(#{headers => #{<<"accept">> => <<"image/gif, image/x-xbitm...">...) at /home/jenkins/jenkins-agent/jenkins-agent/workspace/ey_private_fistful-server_master/_build/default/lib/cowboy/src/cowboy_req.erl:466 called from cowboy_access_log_h:get_request_body_length/1 at /home/jenkins/jenkins-agent/jenkins-agent/workspace/ey_private_fistful-server_master/_build/default/lib/cowboy_access_log/src/cowboy_access_log_h.erl:122 called from cowboy_access_log_h:prepare_meta/3 at /home/jenkins/jenkins-agent/jenkins-agent/workspace/ey_private_fistful-server_master/_build/default/lib/cowboy_access_log... It look like there is no has_body attribute in request, possibly because where is a problem with request and it goes thru cowboy_http:error_terminate/4. What do you mean by required? If it goes through error_terminate then it calls early_error which gives you a PartialReq :: map() and not a cowboy_req:req(). In other words you get the values Cowboy figured out up until that point. I mean it required by type specification here and by code here. So if I get PartialReq :: map() how can I detect it? You are under the impression that early_error receives the Req, it does not, see https://ninenines.eu/docs/en/cowboy/2.7/manual/cowboy_stream/#_callbacks PartialReq - cowboy_req:req(), except all fields are optional It doesn't really matter if there's a body or not as far as Cowboy is concerned because it will just close the connection without attempting to read more. Sometimes it has not read enough to be able to know there is or is not a request body. I see, thanks!
gharchive/issue
2019-11-21T12:48:44
2025-04-01T04:35:13.725318
{ "authors": [ "essen", "kpy3" ], "repo": "ninenines/cowboy", "url": "https://github.com/ninenines/cowboy/issues/1417", "license": "isc", "license_type": "permissive", "license_source": "bigquery" }
295945129
Is this project dead? @hgschmie @electrum pinging you since Github shows you as the only contributors. I see no PRs were merged (nor commented on) since 2016 Build is trying to use java versions no longer supported on travis All in all it looks like there are no active maintainers. I'd like to help here if possible (e.g. by becoming a contributor) instead of making a hard fork Are you interested? wow, haven't checked that repo in a long time. Yes, it is. See https://github.com/basepom/dependency-versions-check-maven-plugin I know about basepom one but for some reason, that I no longer remember, I found something in this project useful that wasn't there. (I was wondering if I should also wait the customary 4 years before replying but meh :D) sorry. I don't really check the old ning repos anymore. basepom is more up-to-date. :-) I would close and archive those if I could but I lost my admin rights for the ning repos and it seems no one is left that can do this.
gharchive/issue
2018-02-09T17:16:24
2025-04-01T04:35:13.730884
{ "authors": [ "hgschmie", "jakub-bochenski" ], "repo": "ning/maven-dependency-versions-check-plugin", "url": "https://github.com/ning/maven-dependency-versions-check-plugin/issues/17", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
243847179
Router When i run the server on windows with a router and connect to myselfI get a bunch of "couldn't reroute" errors regarding the Ping packet. I'm a bit late sorry, but the good news is that I have a repro! I will push a fix soon
gharchive/issue
2017-07-18T20:50:41
2025-04-01T04:35:13.731896
{ "authors": [ "Nickito12", "ningirsu" ], "repo": "ningirsu/stepmania-server", "url": "https://github.com/ningirsu/stepmania-server/issues/40", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
206229038
ENH: Speed up rendering by sharing data We have a lot of data that could be shared (vertex locations), and we should do it: http://docs.enthought.com/mayavi/mayavi/tips.html#sharing-the-same-data-between-scenes I'm working on this at #191
gharchive/issue
2017-02-08T15:23:15
2025-04-01T04:35:13.747308
{ "authors": [ "Eric89GXL", "christianbrodbeck" ], "repo": "nipy/PySurfer", "url": "https://github.com/nipy/PySurfer/issues/177", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
2435274556
AVIF detection is not working when a file is represented as an ArrayBuffer Hey, thanks for the great library! I found an issue when detecting avif files. When work with an ArrayBuffer, isAvifStringIncluded utility returns false for an avif file, hence isAVIF function also returns false. getFileChunk preliminary converts ArrayBuffer into Uint8Array, in which numbers are stored in decimal format, therefore there is no need to parse them as hexadecimal here https://github.com/nir11/file-type-checker/blob/main/src/utils/index.ts#L158. This implementation of the function would be sufficient: export function isAvifStringIncluded(fileChunk: Array<number>): boolean { const signature = fileChunk .slice(4, 12) .map(hex => String.fromCharCode(hex)) .join('') return signature === 'ftypavif' } Hey, great catch! I’ve fixed it now. Thanks!
gharchive/issue
2024-07-29T12:22:48
2025-04-01T04:35:13.756087
{ "authors": [ "nir11", "prplx" ], "repo": "nir11/file-type-checker", "url": "https://github.com/nir11/file-type-checker/issues/13", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2228966090
[Question] What are the different Connection Strategies? Hey! Would it be possible to get a brief summary of the different Connection Strategies and when to use which. Found it in etc/aawgd.env
gharchive/issue
2024-04-06T00:05:54
2025-04-01T04:35:13.760140
{ "authors": [ "NabeelUppel" ], "repo": "nisargjhaveri/WirelessAndroidAutoDongle", "url": "https://github.com/nisargjhaveri/WirelessAndroidAutoDongle/issues/100", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2248444420
如何设置代理服务器 是要挂vpn一类的吗 手机上直接开全局科学上网
gharchive/issue
2024-04-17T14:29:16
2025-04-01T04:35:13.772763
{ "authors": [ "YANYANG1234", "niuhuan" ], "repo": "niuhuan/pikapika", "url": "https://github.com/niuhuan/pikapika/issues/274", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
666978385
Switch CDN from "unpkg.com" to "cdn.jsdelivr.net" In Febrary 2020 (#1022), docsify has switch its CDN from unpkg.com to cdn.jsdelivr.net. As a main plugin, it would be nice if docsify-edit-on-github can be hosted on cdn.jsdelivr.net ; more coherent with other docsify plugins. @axel3rd enabled,thx for suggestion. https://cdn.jsdelivr.net/npm/docsify-edit-on-github
gharchive/issue
2020-07-28T10:53:59
2025-04-01T04:35:13.805512
{ "authors": [ "axel3rd", "njleonzhang" ], "repo": "njleonzhang/docsify-edit-on-github", "url": "https://github.com/njleonzhang/docsify-edit-on-github/issues/12", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1802290456
資料共有フォーマットを作成 資料共有フォーマットを作成し完成したのでプルリクエストを送らせてもらいました。 レビューよろしくお願いします。 レイアウト完璧です! ありがとうございます マージします!
gharchive/pull-request
2023-07-13T06:31:36
2025-04-01T04:35:13.816515
{ "authors": [ "GodHexagon", "yu-za66" ], "repo": "nkc-ug/PSI", "url": "https://github.com/nkc-ug/PSI/pull/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
126656666
nothing happens when running bootprint swagger When I run bootprint swagger http://petstore.swagger.io/v2/swagger.json doc, I get the following output: Loading bootprint-swagger 0.13.1 Loading bootprint-json-schema 0.8.4 Loading bootprint-base 0.6.3 But nothing happens after that. No files were put in the doc directory. Hi there, have you made a fresh install or have you updated your version? Have a look at bootprint#12. This issue has appeared 2 days ago and should be resolved by now... You may have to clear the npm-cache, remove bootprint and reinstall to get the current version of deep-aplus (@1.0.2) That was it, thanks!
gharchive/issue
2016-01-14T13:40:08
2025-04-01T04:35:13.821810
{ "authors": [ "bendavis78", "nknapp" ], "repo": "nknapp/bootprint-swagger", "url": "https://github.com/nknapp/bootprint-swagger/issues/49", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
366930555
semaphore wait is causing stall in app There is few possible configurations that will cause app stall because of semaphore.wait. Most often is read/write characteristic from within notificationCallback. This is caused because notification/indication procedure is waiting for confirmation, but in mean time we want to request read/write from peer device and 2 semaphores block each other. Workaround to this situation is to create new task from within notificationCallback that will request read/write. This task will be executed right after notificationCallback ends. We have to remember that we cant in this case to call vTaskDelay/delay after creating such task, because this will cause semaphores blocking. I am working now on few upgrades and bugfixes but this is also on my list TODO. Yeah... There is a take and a wait for the notify. I commented this out but having stack overflow issues so an not sure if that helped. On Thu, Oct 4, 2018, 11:26 PM chegewara notifications@github.com wrote: Its hard to say what kind of confirmation it is, it can be confirmation from bt stack that notification has been send. The fact is that event CONF is triggered. — You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/nkolban/esp32-snippets/issues/673#issuecomment-427235270, or mute the thread https://github.com/notifications/unsubscribe-auth/AAmH_fbqaXPJvEiEN1VSbMiCJg-0ys5Iks5uhtFYgaJpZM4XIu-x . Ok, with some more tests i can for sure say that this semaphore. wait is not suppose to be in here: https://github.com/nkolban/esp32-snippets/blob/master/cpp_utils/BLECharacteristic.cpp#L572 It not only causes issues with stal, it also causing other issues. I am testing now enhancement with one esp32 server and 3 esp32 clients connected. Each client writes to server characteristic and when characteristic is write then server sends notification to every connected client. I have 2xesp32 that are writing every 250ms and one esp32 writing every 500ms and without that semaphore.wait i have no issue, but with that line some notifications are dropped. I (738524) SampleClient: 8--> 7460 E (738524) SampleClient: Notify callback for characteristic beb5483e-36e1-4688-b7f5-ea07361b26a8 of data Time since boot: 430.216617 length 27 E (738614) SampleClient: Notify callback for characteristic beb5483e-36e1-4688-b7f5-ea07361b26a8 of data Time since boot com3: 738.438042 length 32 E (738614) SampleClient: Notify callback for characteristic beb5483e-36e1-4688-b7f5-ea07361b26a8 of data Time since boot com9: 513.238381 length 32 I (738774) SampleClient: 8--> 7460 E (738854) SampleClient: Notify callback for characteristic beb5483e-36e1-4688-b7f5-ea07361b26a8 of data Time since boot com3: 738.688038 length 32 E (738854) SampleClient: Notify callback for characteristic beb5483e-36e1-4688-b7f5-ea07361b26a8 of data Time since boot com9: 513.488417 length 32 E (739004) SampleClient: Notify callback for characteristic beb5483e-36e1-4688-b7f5-ea07361b26a8 of data Time since boot: 430.71851788417 length 27 I` Implementing a keyboard with a notification going out on every key press, I've noticed the .notify() takes quite a long time. In fact so long that it introduces a noticeable delay after a key has been pressed. Rapidly hitting a key 3 times in a row only registers 2 keypresses with bluetooth host. Getting rid of that semaphore wait seems to have fixed the problem (discovered that on my own by trial and error, then found this thread – means I am not crazy?). This is happening on a single-core device with lots of other tasks running. @toxuin I have 2 questions, do you have input->notify() in separate task? and what is your Tick rate in menuconfig->FreeRTOS? @chegewara yes, it's a task that does very little apart from inputCharacteristic->setValue(report, sizeof(report)); inputCharacteristic->notify();``` Tick rate is default(?) 100hz. @toxuin Now with tick rate 100, do you have vTaskDelay() in it and if yes do you have port_TICK_PERIOD_MS to convert ticks to ms? Also try to create notify task with higher priority than other tasks if its possible. I can send message from my example code without any delays between characters. Bluetooth task in my code waits on a queue receive, so no task delays. Thanks, Che! I will set the priority and see if that helps. Just to confirm: is it ok to remove that semaphore wait on notify?.. You can try, i didnt encounter any issues when i commented it out. discussion moved here https://github.com/nkolban/esp32-snippets/issues/681
gharchive/issue
2018-10-04T19:11:53
2025-04-01T04:35:13.833089
{ "authors": [ "chegewara", "toxuin", "wegunterjr" ], "repo": "nkolban/esp32-snippets", "url": "https://github.com/nkolban/esp32-snippets/issues/673", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2137919755
[機能リクエスト]クリップボードからの画像添付 よくスクリーンショットをSNSに画像として添付することが多いので、画像をクリップボードからのペーストで添付できると助かります。 @bee7813993 さん、機能リクエストありがとうございます!こちらのIssueは Issue formの整備により移行させていただきました! 状況の確認を行いたい場合は、#2 へ Subscribeしていただければ幸いです。 上記対応により、本Issueはクローズさせていただきます。
gharchive/issue
2024-02-16T05:49:43
2025-04-01T04:35:13.834550
{ "authors": [ "bee7813993", "nkte8" ], "repo": "nkte8/skyshare", "url": "https://github.com/nkte8/skyshare/issues/1", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
366460056
When key is not found print the key name into error too e.g. when I do std::string result = jsonObject["hello"]; and "hello" is not found it just does an assert and I don't know what key is being looked up. It would be nice to also print out the key name. Duplicate #932.
gharchive/issue
2018-10-03T18:04:21
2025-04-01T04:35:13.841039
{ "authors": [ "graph", "nlohmann" ], "repo": "nlohmann/json", "url": "https://github.com/nlohmann/json/issues/1273", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
776585649
no matching function for call to ‘nlohmann::basic_json<>::basic_json()’ Hi Team, I am trying to convert json to managed object and in this regard, I am using the to_json and from_json functions provided by nlohmann/json. When I write a sample program, I am able to see no issues with this APIs but when I integrate with the actual project, observing the below error when i call get_to with int as datatype. no matching function for call to ‘nlohmann::basic_json<>::basic_json()’ CPP file (MTop.cpp) void MTop::to_json(json& j, const MTop& mo) { j = json{{"OClass", mOClass}}; } void MTop::from_json(const json& j, MTop& mo) { mo.mOClass = j.at("OClass").get_to< int >(); ---> This line is showing compilation error. } Headerfile (MTop.hpp) #include <MObjectCls.h> #include <nlohmann/json.hpp> using json = nlohmann::json; class MTop : public MObject { public: virtual void to_json(json& j, const MTop& mo); virtual void from_json(const json& j, MTop& mo); private: MObjectCls mOClass; }; Headerfile (MObjectCls.h) class MObjectCls { public: MObjectCls(int value); }; Since there to_json and from_json are used by the object of this class only, these two functions are not maintained under any namespace. if the from_json( ) logic is modified as below - j.at("OClass").get_to(mo.mOClass); MTop.cpp --> error: no matching function for call to ‘nlohmann::basic_json<>::basic_json(<brace-enclosed initializer list>)’ error: no matching function for call to ‘nlohmann::basic_json<>::get_to(MObjectCls&) const’ ../include/nlohmann/json.hpp:3017:28: error: no type named ‘type’ in ‘struct std::enable_if<false, int>’ ../include/nlohmann/json.hpp:3030:25: error: no type named ‘type’ in ‘struct std::enable_if<false, int>’ Please guide me what could be the issue. Note: Since the header and cpp files are big, I have shared only the contents which are of focus here. Deepak The functions from_json and to_json must be defined in the same namespace as the type. They must not be member functions. Please try moving them out of MTop. About "error: no matching function for call to ‘nlohmann::basic_json<>::get_to(MObjectCls&) const’" You misused get_to. It should be either int value = j.get<int>(); or int value; j.get_to( value ); First one in your case. @nlohmann Thanks for your response. I have modified the files as below - MTop.cpp #include <nlohmann/json.hpp> using json = nlohmann::json; namespace MTop_ns { typedef enum { STATE_LOCKED = 0, STATE_UNLOCKED = 1, STATE_SHUTTINGDOWN = 2 } MAdminStateEnum; NLOHMANN_JSON_SERIALIZE_ENUM ( MAdminStateEnum, { {STATE_LOCKED, "locked"}, {STATE_UNLOCKED, "unlocked"}, {STATE_SHUTTINGDOWN, "shutdown"}, }); struct MTop000 { int mOInstId; std::string mOInstName; MAdminStateEnum mAdminState; }; void to_json(json& j, const MTop000& mo) { j = json{{"mOInstId", mo.mOInstId}, {"mOInstName", mo.mOInstName}, }; } void from_json(const json& j, MTop000& mo) { j.at("mOInstId").get_to(mo.mOInstId); j.at("mOInstName").get_to(mo.mOInstName); } } MTop.hpp #include <MObjectCls.h> class MTop : public MObject { public: private: MObjectCls mOClass; }; Now, I don't see any compilation errors as the from_json and to_json are not part of the member functions. I would be accessing this as below in other file - MTop_ns::MTop000 mT = {"10, "mTopClass", STATE_LOCKED"}; json j = mT; Since I have multiple managed objects, there will be many class files. If I had to write to_json/from_json for all those, then I need to do the same for all right ? Is there any other way to do it ? Deepak Yes, you have to define it for every type. Maybe this article helps: https://json.nlohmann.me/features/arbitrary_types/#simplify-your-life-with-macros @nlohmann Thanks for the reference. Looking into that, I made changes to my code and here is the change - Headerfile (MTop.hpp) #include <MObjectCls.h> #include <nlohmann/json.hpp> using json = nlohmann::json; namespace MTop_ns { class MTop : public MObject { public: NLOHMANN_DEFINE_TYPE_INTRUSIVE(MTop, moClass) private: MObjectCls mOClass; }; } Headerfile (MObjectCls.h) class MObjectCls { public: MObjectCls( ); MObjectCls(int value); virtual ~MObjectCls( ); }; WIth this, I am observing below errors - MTop.h-> error: no match for ‘operator=’ (operand types are ‘nlohmann::basic_json<>::value_type’ {aka ‘nlohmann::basic_json<>’} and ‘const MibSccpInactivityTestTimer’) ../include/nlohmann/detail/macro_scope.hpp:268:81: error: no matching function for call to ‘nlohmann::basic_json<>::get_to(MibAdministrativeState&) const’ ../include/nlohmann/json.hpp:3017:28: error: no type named ‘type’ in ‘struct std::enable_if<false, int>’ If I modify as below, then I don't see any error. Headerfile (MTop.hpp) #include <MObjectCls.h> #include <nlohmann/json.hpp> using json = nlohmann::json; namespace MTop_ns { class MTop : public MObject { public: NLOHMANN_DEFINE_TYPE_INTRUSIVE(MTop, moClass) private: int mOClass; }; } This is a nested class where we need to access parameters from base class.. Should this macro be called inside all the inherited classes with the namespace ? Deepak Yes. Yes. @nlohmann Thanks for your continuous support. According to the document, we see that if the macro NLOHMANN_DEFINE_TYPE_INTRUSIVE is used, there is no need to define the macro within the namespace. Currently, I am following the same. Is this correct ? NLOHMANN_DEFINE_TYPE_NON_INTRUSIVE(name, member1, member2, ...) is to be defined inside of the namespace of the class/struct to create code for. NLOHMANN_DEFINE_TYPE_INTRUSIVE(name, member1, member2, ...) is to be defined inside of the class/struct to create code for. This macro can also access private members. #include <MObjectCls.h> #include <nlohmann/json.hpp> using json = nlohmann::json; class MTop : public MObject { public: NLOHMANN_DEFINE_TYPE_INTRUSIVE(MTop, moClass) private: int mOClass; }; MTop.h: error: invalid use of non-static data member ‘MTop::mOClass’ ../include/nlohmann/detail/macro_scope.hpp:276:130: error: ‘NLOHMANN_JSON_TO’ was not declared in this scope Are these errors related with not using namespace ? Deepak @nlohmann Thanks for your continuous support. According to the document, we see that if the macro NLOHMANN_DEFINE_TYPE_INTRUSIVE is used, there is no need to define the macro within the namespace. Currently, I am following the same. Is this correct ? NLOHMANN_DEFINE_TYPE_NON_INTRUSIVE(name, member1, member2, ...) is to be defined inside of the namespace of the class/struct to create code for. NLOHMANN_DEFINE_TYPE_INTRUSIVE(name, member1, member2, ...) is to be defined inside of the class/struct to create code for. This macro can also access private members. #include <MObjectCls.h> #include <nlohmann/json.hpp> using json = nlohmann::json; class MTop : public MObject { public: NLOHMANN_DEFINE_TYPE_INTRUSIVE(MTop, moClass) private: int mOClass; }; MTop.h: error: invalid use of non-static data member ‘MTop::mOClass’ ../include/nlohmann/detail/macro_scope.hpp:276:130: error: ‘NLOHMANN_JSON_TO’ was not declared in this scope Are these errors related with not using namespace ? Deepak Looks like a typo. In NLOHMANN_DEFINE_TYPE_INTRUSIVE, you write moClass (lower-case o), but the member is called mOClass (upper-case O). Looks like a typo. In NLOHMANN_DEFINE_TYPE_INTRUSIVE, you write moClass (lower-case o), but the member is called mOClass (upper-case O). @nlohmann My apology. The code which was pasted had this typo error. In the code what is being used, the variable passed is mOClass only. NLOHMANN_DEFINE_TYPE_INTRUSIVE(MTop, mOClass) Need your inputs on my below query - According to the document, we see that if the macro NLOHMANN_DEFINE_TYPE_INTRUSIVE is used, there is no need to define the macro within the namespace. Currently, I am following the same. Is this correct ? Deepak @nlohmann My apology. The code which was pasted had this typo error. In the code what is being used, the variable passed is mOClass only. NLOHMANN_DEFINE_TYPE_INTRUSIVE(MTop, mOClass) Need your inputs on my below query - According to the document, we see that if the macro NLOHMANN_DEFINE_TYPE_INTRUSIVE is used, there is no need to define the macro within the namespace. Currently, I am following the same. Is this correct ? Deepak What compiler are you using? Looks like there's a problem with the macro expansion. What compiler are you using? Looks like there's a problem with the macro expansion. Can you try replacing that macro with this? friend void to_json(nlohmann::json& nlohmann_json_j, const Type& nlohmann_json_t) { NLOHMANN_JSON_TO(mOClass) } friend void from_json(const nlohmann::json& nlohmann_json_j, Type& nlohmann_json_t) { NLOHMANN_JSON_FROM(mOClass) } If that doesn't work, then try this: friend void to_json(nlohmann::json& nlohmann_json_j, const Type& nlohmann_json_t) { nlohmann_json_j["mOClass"] = nlohmann_json_t.mOClass; } friend void from_json(const nlohmann::json& nlohmann_json_j, Type& nlohmann_json_t) { nlohmann_json_j.at("mOClass").get_to(nlohmann_json_t.mOClass); } This will at least narrow down the failure. Can you try replacing that macro with this? friend void to_json(nlohmann::json& nlohmann_json_j, const Type& nlohmann_json_t) { NLOHMANN_JSON_TO(mOClass) } friend void from_json(const nlohmann::json& nlohmann_json_j, Type& nlohmann_json_t) { NLOHMANN_JSON_FROM(mOClass) } If that doesn't work, then try this: friend void to_json(nlohmann::json& nlohmann_json_j, const Type& nlohmann_json_t) { nlohmann_json_j["mOClass"] = nlohmann_json_t.mOClass; } friend void from_json(const nlohmann::json& nlohmann_json_j, Type& nlohmann_json_t) { nlohmann_json_j.at("mOClass").get_to(nlohmann_json_t.mOClass); } This will at least narrow down the failure. @gregmarr Thanks for your inputs. Will modify as suggested and check. Deepak @gregmarr Thanks for your inputs. Will modify as suggested and check. Deepak @gregmarr Forgot to share the compiler details. I am using WSL2 (running Ubuntu 20.4) $ gcc --version gcc (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0 Copyright (C) 2019 Free Software Foundation, Inc. This is free software; see the source for copying conditions. There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. $ uname -a Linux IN-9KHT253 4.19.128-microsoft-standard #1 SMP Tue Jun 23 12:58:10 UTC 2020 x86_64 x86_64 x86_64 GNU/Linux @gregmarr Forgot to share the compiler details. I am using WSL2 (running Ubuntu 20.4) $ gcc --version gcc (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0 Copyright (C) 2019 Free Software Foundation, Inc. This is free software; see the source for copying conditions. There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. $ uname -a Linux IN-9KHT253 4.19.128-microsoft-standard #1 SMP Tue Jun 23 12:58:10 UTC 2020 x86_64 x86_64 x86_64 GNU/Linux @gregmarr I tried your suggestion but end up getting different error #include <MObjectCls.h> #include <nlohmann/json.hpp> using json = nlohmann::json; class MTop : public MObject { public: friend void to_json(nlohmann::json& nlohmann_json_j, const Type& nlohmann_json_t) { nlohmann_json_j["mOClass"] = nlohmann_json_t.mOClass; } friend void from_json(const nlohmann::json& nlohmann_json_j, Type& nlohmann_json_t) { nlohmann_json_j.at("mOClass").get_to(nlohmann_json_t.mOClass); } private: int mOClass; }; Below error is observed - error: ‘Type’ does not name a type error: ‘Type’ has not been declared mOClass is integer. Its Unsigned integer. It should be supported by JSON API right ? Deepak @gregmarr I tried your suggestion but end up getting different error #include <MObjectCls.h> #include <nlohmann/json.hpp> using json = nlohmann::json; class MTop : public MObject { public: friend void to_json(nlohmann::json& nlohmann_json_j, const Type& nlohmann_json_t) { nlohmann_json_j["mOClass"] = nlohmann_json_t.mOClass; } friend void from_json(const nlohmann::json& nlohmann_json_j, Type& nlohmann_json_t) { nlohmann_json_j.at("mOClass").get_to(nlohmann_json_t.mOClass); } private: int mOClass; }; Below error is observed - error: ‘Type’ does not name a type error: ‘Type’ has not been declared mOClass is integer. Its Unsigned integer. It should be supported by JSON API right ? Deepak Right, I forgot to mention to change Type to the type of mOClass. Right, I forgot to mention to change Type to the type of mOClass. @gregmarr Thanks for the clarification. This resolved the compilation error.. Hopefully, I should be able to integrate successfully. Thanks for your support. @gregmarr Thanks for the clarification. This resolved the compilation error.. Hopefully, I should be able to integrate successfully. Thanks for your support.
gharchive/issue
2020-12-30T18:38:10
2025-04-01T04:35:13.882137
{ "authors": [ "YarikTH", "deepakd82", "gregmarr", "nlohmann" ], "repo": "nlohmann/json", "url": "https://github.com/nlohmann/json/issues/2559", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
138709671
Initial commit Initial work on MPIM closing =)
gharchive/pull-request
2016-03-05T19:10:59
2025-04-01T04:35:13.883551
{ "authors": [ "james-lawrence", "nlopes" ], "repo": "nlopes/slack", "url": "https://github.com/nlopes/slack/pull/62", "license": "bsd-2-clause", "license_type": "permissive", "license_source": "bigquery" }
801219685
Add 1 second wait before storing another annotation Up to an hour and still running. not sure why - Edit: Completed, running, but this did work. @thomasyu888 I would like to understand better the cause of the issue so we can find a reliable solution. @thomasyu888 I would like to understand better the cause of the issue so we can find a reliable solution. @tschaffter My guess of why this worked is that I am waiting a second before doing another rest call. All of the code that actually returns the response in the server is abstracted away and is part of connexion (unmodifiable). My experiment of adding a sleep in the controller itself caused the same RemoteDisconnect issue, which led me to think that the client could be making too many calls too fast. Ive been trying to search more about this issue, but I can't seem to find more than what I've already found. @thomasyu888 A better approach would be to block/loop over the response until the status code 201 is received. This way we always wait enough time instead of assuming that the request will be completed within 1 second (response time depends on hardware and connection bandwidth). The ideal solution would be to send the requests asynchronously. In JavaScript/TypeScript, I would use mergeMap to send a bunch of requests at once, which is one instruction that would then return once all the requests have completed. The uWSGI server of the data node is configured to process one request at a time but this could be changed in a later revision to decrease processing time (e.g. pulling clinical notes concurrently, sending requests to tools concurrently, etc.). Could you give a try to the first solution proposed (wait for 201 response)? @tschaffter Thanks for the resources! I took a look at where I would even look for a response == 201, the issue is the api_instance.create_annotation(... function returns the json object already, so I would have to look deeper in the code. Upon investigate, they seemed to have generated request functions within the classes which is great but also, there are more response codes than just 201. I'm trying something different. Apparently I can add a async_req=True command to use asynchronous HTTP calls. lets see if that works.
gharchive/pull-request
2021-02-04T12:01:24
2025-04-01T04:35:13.888531
{ "authors": [ "thomasyu888", "tschaffter" ], "repo": "nlpsandbox/nlpsandbox-client", "url": "https://github.com/nlpsandbox/nlpsandbox-client/pull/93", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
906390100
Review Rules one more time before launch https://www.synapse.org/#!Synapse:syn22277124/wiki/604831 Doc link can be found here. Main feedback: **Organization and Structure ** Lots of good content here. Tried to group relevant bullet points together. (1) Registering for NLP Sandbox: more tactical rules that affect developer workflow (2) Publication Considerations: concerning publications (3) Service Agreement: less tactical rules, more the obligatory legal jargon for scenarios of conflict or malicious intent Consolidating some sections Wondering if we can consolidate some of the below sections, comments in line Terms of Use Terms of Use currently says, not sure if this is a placeholder for a template we need from governance? “BE SURE TO CONTACT GOVERNANCE ABOUT THIS PROJECT” A point of consideration for the future: per Meg Doerr's readability guidelines, I used Microsoft word to get some readability statistics. The NLP Sandbox rules are currently readable for a college student or graduate. If we want to make them readable to a wider audience of Americans, we may want to do a deeper dive on breaking down some of the language. However, given that there is a lot of obligatory legal language here, it may be challenging. No action required, just a consideration for the future.
gharchive/issue
2021-05-29T06:24:33
2025-04-01T04:35:13.892933
{ "authors": [ "jiaxinmachine88", "tschaffter" ], "repo": "nlpsandbox/nlpsandbox.io", "url": "https://github.com/nlpsandbox/nlpsandbox.io/issues/92", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
213584967
Stanford segmenter nltk Could not find SLF4J in your classpath I set up a nltk and stanford environment, nltk and stanford jars has downloaded, the program with nltk was ok, but I had a trouble with stanford segmenter. just make a simple program via stanford segmenter, I got a error is Could not find SLF4J in your classpath, although I had export all jars include slf4j-api.jar. Detail as follows Python3.5 NLTK, 3.2.2, Standford jars 3.7 OS: Centos Environment variable: export JAVA_HOME=/usr/java/jdk1.8.0_60 export NLTK_DATA=/opt/nltk_data export STANFORD_SEGMENTER_PATH=/opt/stanford/stanford-segmenter-3.7 export CLASSPATH=$CLASSPATH:$STANFORD_SEGMENTER_PATH/stanford-segmenter.jar export STANFORD_POSTAGGER_PATH=/opt/stanford/stanford-postagger-full-2016-10-31 export CLASSPATH=$CLASSPATH:$STANFORD_POSTAGGER_PATH/stanford-postagger.jar export STANFORD_NER_PATH=/opt/stanford/stanford-ner-2016-10-31 export CLASSPATH=$CLASSPATH:$STANFORD_NER_PATH/stanford-ner.jar export STANFORD_MODELS=$STANFORD_NER_PATH/classifiers:$STANFORD_POSTAGGER_PATH/models export STANFORD_PARSER_PATH=/opt/stanford/stanford-parser-full-2016-10-31 export CLASSPATH=$CLASSPATH:$STANFORD_PARSER_PATH/stanford-parser.jar:$STANFORD_PARSER_PATH/stanford-parser-3.6.0-models.jar:$STANFORD_PARSER_PATH/slf4j-api.jar:$STANFORD_PARSER_PATH/ejml-0.23.jar export STANFORD_CORENLP_PATH=/opt/stanford/stanford-corenlp-full-2016-10-31 export CLASSPATH=$CLASSPATH:$STANFORD_CORENLP_PATH/stanford-corenlp-3.7.0.jar:$STANFORD_CORENLP_PATH/stanford-corenlp-3.7.0-models.jar:$STANFORD_CORENLP_PATH/javax.json.jar:$STANFORD_CORENLP_PATH/joda-time.jar:$STANFORD_CORENLP_PATH/jollyday.jar:$STANFORD_CORENLP_PATH/protobuf.jar:$STANFORD_CORENLP_PATH/slf4j-simple.jar:$STANFORD_CORENLP_PATH/xom.jar export STANFORD_CORENLP=$STANFORD_CORENLP_PATH The program as follows: from nltk.tokenize import StanfordSegmenter >>> segmenter = StanfordSegmenter( path_to_sihan_corpora_dict="/opt/stanford/stanford-segmenter-3.7/data/", path_to_model="/opt/stanford/stanford-segmenter-3.7/data/pku.gz", path_to_dict="/opt/stanford/stanford-segmenter-3.7/data/dict-chris6.ser.gz" ) >>> res = segmenter.segment(u"这是斯坦福中文分词器测试") The error as follows: Exception in thread "main" java.lang.ExceptionInInitializerError at edu.stanford.nlp.ie.AbstractSequenceClassifier.<clinit>(AbstractSequenceClassifier.java:88) Caused by: java.lang.IllegalStateException: Could not find SLF4J in your classpath at edu.stanford.nlp.util.logging.RedwoodConfiguration$Handlers.lambda$static$530(RedwoodConfiguration.java:190) at edu.stanford.nlp.util.logging.RedwoodConfiguration$Handlers$7.buildChain(RedwoodConfiguration.java:309) at edu.stanford.nlp.util.logging.RedwoodConfiguration$Handlers$7.apply(RedwoodConfiguration.java:318) at edu.stanford.nlp.util.logging.RedwoodConfiguration.lambda$handlers$535(RedwoodConfiguration.java:363) at edu.stanford.nlp.util.logging.RedwoodConfiguration.apply(RedwoodConfiguration.java:41) at edu.stanford.nlp.util.logging.Redwood.<clinit>(Redwood.java:609) ... 1 more Caused by: edu.stanford.nlp.util.MetaClass$ClassCreationException: java.lang.ClassNotFoundException: edu.stanford.nlp.util.logging.SLF4JHandler at edu.stanford.nlp.util.MetaClass.createFactory(MetaClass.java:364) at edu.stanford.nlp.util.MetaClass.createInstance(MetaClass.java:381) at edu.stanford.nlp.util.logging.RedwoodConfiguration$Handlers.lambda$static$530(RedwoodConfiguration.java:186) ... 6 more Caused by: java.lang.ClassNotFoundException: edu.stanford.nlp.util.logging.SLF4JHandler at java.net.URLClassLoader.findClass(URLClassLoader.java:381) at java.lang.ClassLoader.loadClass(ClassLoader.java:424) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331) at java.lang.ClassLoader.loadClass(ClassLoader.java:357) at java.lang.Class.forName0(Native Method) at java.lang.Class.forName(Class.java:264) at edu.stanford.nlp.util.MetaClass$ClassFactory.construct(MetaClass.java:135) at edu.stanford.nlp.util.MetaClass$ClassFactory.<init>(MetaClass.java:202) at edu.stanford.nlp.util.MetaClass$ClassFactory.<init>(MetaClass.java:69) at edu.stanford.nlp.util.MetaClass.createFactory(MetaClass.java:360) ... 8 more Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/usr/local/python3/lib/python3.5/site-packages/nltk/tokenize/stanford_segmenter.py", line 96, in segment return self.segment_sents([tokens]) File "/usr/local/python3/lib/python3.5/site-packages/nltk/tokenize/stanford_segmenter.py", line 123, in segment_sents stdout = self._execute(cmd) File "/usr/local/python3/lib/python3.5/site-packages/nltk/tokenize/stanford_segmenter.py", line 143, in _execute cmd,classpath=self._stanford_jar, stdout=PIPE, stderr=PIPE) File "/usr/local/python3/lib/python3.5/site-packages/nltk/internals.py", line 134, in java raise OSError('Java command failed : ' + str(cmd)) OSError: Java command failed : ['/usr/java/jdk1.8.0_60/bin/java', '-mx2g', '-cp', '/opt/stanford/stanford-segmenter-3.7/stanford-segmenter.jar:/opt/stanford/stanford-parser-full-2016-10-31/slf4j-api.jar', 'edu.stanford.nlp.ie.crf.CRFClassifier', '-sighanCorporaDict', '/opt/stanford/stanford-segmenter-3.7/data/', '-textFile', '/tmp/tmpkttpldl6', '-sighanPostProcessing', 'true', '-keepAllWhitespaces', 'false', '-loadClassifier', '/opt/stanford/stanford-segmenter-3.7/data/pku.gz', '-serDictionary', '/opt/stanford/stanford-segmenter-3.7/data/dict-chris6.ser.gz', '-inputEncoding', 'UTF-8'] FYI - An workaround I used is to set path_to_slf4j to sth that does not include slf4j, e.g. path_to_slf4j = "/home/chen/Downloads/stanford-segmenter-2016-10-31/stanford-segmenter-3.7.0.jar" This should be resolved in our new interface to the Stanford Tools. If not, please reopen.
gharchive/issue
2017-03-12T06:11:13
2025-04-01T04:35:13.897567
{ "authors": [ "benbenguo", "cchunsheng", "stevenbird" ], "repo": "nltk/nltk", "url": "https://github.com/nltk/nltk/issues/1652", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
120471371
Update the default download URL to use https:// The index file that the downloader uses is currently served from GitHub Pages (via http://www.nltk.org), but while GitHub Pages doesn't directly offer https://, they provide an alternate way to easily download files securely over https:// via their raw.githubusercontent.com FQDN. This helps protect the index and the file URLs listed in it from being MITM'd and modified to something malicious and just helps with ensuring file integrity. The other half of this is https://github.com/nltk/nltk_data/pull/40. Great idea, thanks @reedloden.
gharchive/pull-request
2015-12-04T19:37:43
2025-04-01T04:35:13.900501
{ "authors": [ "reedloden", "stevenbird" ], "repo": "nltk/nltk", "url": "https://github.com/nltk/nltk/pull/1220", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
193268218
Fixed missing parenthesis on MASI distance found by @drevicko Fixed missing parenthesis found by @drevicko There was a missing parenthesis on my MASI code, see https://github.com/DrDub/nltk/commit/e8bbbd4cd4cc7cb6414cb78959b0641ef22ba9f9#commitcomment-18577279 for details. Thanks to @drevicko for the analysis and for pointing this out. This patch also fixed the doctests. Thanks @drdub
gharchive/pull-request
2016-12-03T05:48:15
2025-04-01T04:35:13.902512
{ "authors": [ "DrDub", "stevenbird" ], "repo": "nltk/nltk", "url": "https://github.com/nltk/nltk/pull/1535", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
273232027
RAN integration It would show the power of staart to integrate it with RAN would be interested in this!
gharchive/issue
2017-11-12T13:52:27
2025-04-01T04:35:13.907048
{ "authors": [ "nmaro", "sw-yx" ], "repo": "nmaro/staart", "url": "https://github.com/nmaro/staart/issues/15", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1622785609
meteva 为啥要设置绘图字体? 引入 meteva 包后发现 plt.rcParams['font.sans-serif'] 被设置成 simhei ,但是我是 ubuntu 服务器,已经设置好 matplotlib 的字体参数,为什么要重写matplolib字体,是有什么逻辑上的要求么? 绘图函数中使用了plt.rcParams['font.sans-serif']=['SimHei'] 语句来设置中文字体。这不是强制性的要求,如果需要更换字体,在调用 meteva中的绘图函数功能前,可以先运行 plt.rcParams['font.sans-serif'] = [其它字体参数]。
gharchive/issue
2023-03-14T05:58:02
2025-04-01T04:35:13.908182
{ "authors": [ "liucouhua", "qinglonger" ], "repo": "nmcdev/meteva", "url": "https://github.com/nmcdev/meteva/issues/150", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1374898779
Installation challenges; add node.js to required packages? Hi all, I installed ipygoldenlayout today and didn't already have node.js installed. Maybe we should add to the README that you may need to conda install -c conda-forge nodejs before you install the labextension? Also, I'm getting the following error when I run the labextension install command: Node v18.9.0 Yarn configuration loaded. > /opt/miniconda3/bin/npm pack ipygoldenlayout npm notice npm notice 📦 ipygoldenlayout@0.1.0 npm notice === Tarball Contents === npm notice 1.5kB LICENSE.txt npm notice 752B README.md npm notice 700.9kB dist/index.js npm notice 754B lib/extension.js npm notice 361B lib/index.js npm notice 1.1kB lib/plugin.js npm notice 533B lib/version.js npm notice 5.1kB lib/widget.js npm notice 2.8kB package.json npm notice === Tarball Details === npm notice name: ipygoldenlayout npm notice version: 0.1.0 npm notice filename: ipygoldenlayout-0.1.0.tgz npm notice package size: 205.6 kB npm notice unpacked size: 713.8 kB npm notice shasum: 2b0b5b6c782524848fdd46d25a056a611b7d6f5e npm notice integrity: sha512-OwsrKUV8mTpm9[...]KwKZff+tdos5g== npm notice total files: 9 npm notice ipygoldenlayout-0.1.0.tgz Node v18.9.0 Yarn configuration loaded. Building jupyterlab assets (production, minimized) > node /opt/miniconda3/lib/python3.9/site-packages/jupyterlab/staging/yarn.js install --non-interactive yarn install v1.21.1 [1/5] Validating package.json... [2/5] Resolving packages... success Already up-to-date. Done in 0.16s. > node /opt/miniconda3/lib/python3.9/site-packages/jupyterlab/staging/yarn.js yarn-deduplicate -s fewer --fail yarn run v1.21.1 $ /opt/miniconda3/share/jupyter/lab/staging/node_modules/.bin/yarn-deduplicate -s fewer --fail Done in 0.22s. > node /opt/miniconda3/lib/python3.9/site-packages/jupyterlab/staging/yarn.js run build:prod:minimize yarn run v1.21.1 $ webpack --config webpack.prod.minimize.config.js [webpack-cli] ModuleNotFoundError: Module not found: Error: Can't resolve './custom.css' in '/opt/miniconda3/share/jupyter/lab/staging/node_modules/ipygoldenlayout/lib' at /opt/miniconda3/share/jupyter/lab/staging/node_modules/webpack/lib/Compilation.js:1768:28 at /opt/miniconda3/share/jupyter/lab/staging/node_modules/webpack/lib/NormalModuleFactory.js:742:13 at eval (eval at create (/opt/miniconda3/share/jupyter/lab/staging/node_modules/tapable/lib/HookCodeFactory.js:33:10), <anonymous>:10:1) at /opt/miniconda3/share/jupyter/lab/staging/node_modules/webpack/lib/NormalModuleFactory.js:276:22 at eval (eval at create (/opt/miniconda3/share/jupyter/lab/staging/node_modules/tapable/lib/HookCodeFactory.js:33:10), <anonymous>:9:1) at /opt/miniconda3/share/jupyter/lab/staging/node_modules/webpack/lib/NormalModuleFactory.js:405:22 at /opt/miniconda3/share/jupyter/lab/staging/node_modules/webpack/lib/NormalModuleFactory.js:119:11 at /opt/miniconda3/share/jupyter/lab/staging/node_modules/webpack/lib/NormalModuleFactory.js:650:24 at /opt/miniconda3/share/jupyter/lab/staging/node_modules/webpack/lib/NormalModuleFactory.js:813:8 at /opt/miniconda3/share/jupyter/lab/staging/node_modules/webpack/lib/NormalModuleFactory.js:933:5 resolve './custom.css' in '/opt/miniconda3/share/jupyter/lab/staging/node_modules/ipygoldenlayout/lib' using description file: /opt/miniconda3/share/jupyter/lab/staging/node_modules/ipygoldenlayout/package.json (relative path: ./lib) Field 'browser' doesn't contain a valid alias configuration using description file: /opt/miniconda3/share/jupyter/lab/staging/node_modules/ipygoldenlayout/package.json (relative path: ./lib/custom.css) no extension Field 'browser' doesn't contain a valid alias configuration /opt/miniconda3/share/jupyter/lab/staging/node_modules/ipygoldenlayout/lib/custom.css doesn't exist .js Field 'browser' doesn't contain a valid alias configuration /opt/miniconda3/share/jupyter/lab/staging/node_modules/ipygoldenlayout/lib/custom.css.js doesn't exist .json Field 'browser' doesn't contain a valid alias configuration /opt/miniconda3/share/jupyter/lab/staging/node_modules/ipygoldenlayout/lib/custom.css.json doesn't exist .wasm Field 'browser' doesn't contain a valid alias configuration /opt/miniconda3/share/jupyter/lab/staging/node_modules/ipygoldenlayout/lib/custom.css.wasm doesn't exist as directory /opt/miniconda3/share/jupyter/lab/staging/node_modules/ipygoldenlayout/lib/custom.css doesn't exist error Command failed with exit code 2. info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command. JupyterLab failed to build Are you trying to install ipygoldenlayout to JupyterLab>=4? Probably ipygoldenlayout does not support a version >=4 of JupyterLab, yet.
gharchive/issue
2022-09-15T18:05:42
2025-04-01T04:35:13.912576
{ "authors": [ "bmorris3", "jhgoebbert" ], "repo": "nmearl/ipygoldenlayout", "url": "https://github.com/nmearl/ipygoldenlayout/issues/35", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1161910763
document Instability in initial F near Fmsy There is a non-monotonic relationship that can interfere with initial F estimation if the option to allow spawner-recruitment to operate inside initial year is used and if the input initial catch is near MSY. In the attached file there is a profile on initial F bracketing Fmsy. It shows, quite logically, that values of F greater or less than Fmsy give initial catch < MSY and that only init F = Fmsy produces an equilibrium catch equal to MSY. So there are two pathological problems: if initial catch is near MSY (of course, SS has not yet calculated Fmsy yet), then the search for init F will encounter the non-monotonic situation if initial catch is > MSY, then the mismatch between observed and estimated initial catch cannot be reduced by increasing init F further. However, it might be possible for SS to increase R0 or other parameters in order to increase MSY indirectly. original issue https://github.com/nmfs-stock-synthesis/stock-synthesis/issues/18 @Rick-Methot-NOAA it seems like this is detailed on page 33 of the user manual (also section 6.3.0.2 in the html file under Benchmark Calculations) when the Fishing Mortality info was updated. Is my understanding of that info correct? We need to add the text from Kathryn's comment (which I wrote) to section 8.7.1 regarding initial fishing mortality. The topic is already addressed in 8.6.0.1, so the 8.7.1 text can say something like: "If the initial equilibrium catch is near MSY, then a logical inconsistency may occur as documented in section 8.6.0.1."
gharchive/issue
2022-03-07T20:52:10
2025-04-01T04:35:13.922929
{ "authors": [ "Rick-Methot-NOAA", "e-perl-NOAA", "k-doering-NOAA" ], "repo": "nmfs-ost/ss3-doc", "url": "https://github.com/nmfs-ost/ss3-doc/issues/98", "license": "CC0-1.0", "license_type": "permissive", "license_source": "github-api" }
1750079595
Number of sent messages quota: Quota.Application Readiness checklist [x] I added/updated unit tests. [ ] I added/updated integration tests. [x] I ensured that the PR title is good enough for the changelog. [x] I labeled the PR. With the latest fixes to MetricStatus, we are back to the Transport tests tripping on that one test that fails most times, i.e., the other tests are now passing. @tnotheis I have merged the target branch with main as well.
gharchive/pull-request
2023-06-09T15:12:10
2025-04-01T04:35:13.926104
{ "authors": [ "Dannyps" ], "repo": "nmshd/backbone", "url": "https://github.com/nmshd/backbone/pull/160", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1977438004
Compability for jam and profile pages. The jam pages and profile pages don't have some of the color present in project pages #1 Description container for: Project page .formatted_description Profile page .user_profile Jam page .jam_content Devlog .post_body Impossible
gharchive/issue
2023-11-04T17:26:45
2025-04-01T04:35:13.940386
{ "authors": [ "nndda" ], "repo": "nndda/pitch-css", "url": "https://github.com/nndda/pitch-css/issues/3", "license": "CC0-1.0", "license_type": "permissive", "license_source": "github-api" }
115390138
Multiple Charts on a single canvas I think a useful feature would be the ability to overlay charts on a single canvas. I know this would be difficult though this could provide a very useful level of functionality. +1 What's wrong with using multiple canvases? Well I would like to try and overlay certain graphs (i.e a doughnut chart over a pie graph, to allow for 2 sets of the data to be correlated, or a column graph with a line chart) and allow to show correlation or save room on the page On 10 November 2015 at 01:58, Jake notifications@github.com wrote: What's wrong with using multiple canvases? — Reply to this email directly or view it on GitHub https://github.com/nnnick/Chart.js/issues/1624#issuecomment-155086251. @Maraket combo charts between bars and libes are supported in v2. The relevant sample file shows you how to accomplish this. Closing since v2 supports multiple datasets in pie/doughnut charts. We are not going to support multiple charts on a single canvas at this time.
gharchive/issue
2015-11-05T22:30:40
2025-04-01T04:35:13.947221
{ "authors": [ "Maraket", "etimberg", "fercreek", "jakesyl" ], "repo": "nnnick/Chart.js", "url": "https://github.com/nnnick/Chart.js/issues/1624", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1391766563
[Version] update version to 1.8.3 Release 1.8.2 for Tizen 7.0 M2. Start implementation for next release (v1.8.3). Signed-off-by: Jaeyun jy1210.jung@samsung.com :memo: TAOS-CI Version: 1.5.20200925. Thank you for submitting PR #217. Please a submit 1commit/1PR (one commit per one PR) policy to get comments quickly from reviewers. Your PR must pass all verificiation processes of cibot before starting a review process from reviewers. If you are new member to join this project, please read manuals in documentation folder and wiki page. In order to monitor a progress status of your PR in more detail, visit http://ci.nnstreamer.ai/.
gharchive/pull-request
2022-09-30T03:26:22
2025-04-01T04:35:13.958199
{ "authors": [ "jaeyun-jung", "taos-ci" ], "repo": "nnstreamer/api", "url": "https://github.com/nnstreamer/api/pull/217", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
998947553
[resnet/unittest] Add unittest for resnet This patch adds resnet18 unittest model generation with genModelTests. Further, the input data range is changed from 0 to x from -x to x as relu based models work better with 0 to x data range to prevent loss of information. This patch adds resnet models unittest for resnet18. The verification has been done offline for 2 iterations for the output of all layers with precision of 1.1e-4. Derivaitves and gradients have higher error because of relu: when some value is close to 0, it can be positive or negative with some error (of the order of e-7). Although this error is way within the error limit. however, this exacerbates the error in backwarding where derivatives (which are significant in values) can flow if the relu value was over 0, and not flow if under zero. This is manageable in smaller models but difficult to avoid in unittests for larger models. Other bug fixes in this patch: max error reported by unittest_nntrainer_models has been fixed error reporting now includes layer type as well ModelTestOption MINIMUM has been renamed to NO_THROW_RUN Resnet unittest is disabled as the golden data exceeds 70MB. This will be enabled when the golden data can be generated at runtime. Signed-off-by: Parichay Kapoor pk.kapoor@samsung.com :memo: TAOS-CI Version: 1.5.20200925. Thank you for submitting PR #1578. Please a submit 1commit/1PR (one commit per one PR) policy to get comments quickly from reviewers. Your PR must pass all verificiation processes of cibot before starting a review process from reviewers. If you are new member to join this project, please read manuals in documentation folder and wiki page. In order to monitor a progress status of your PR in more detail, visit http://nnstreamer.mooo.com/. :octocat: cibot: @kparichay, A builder checker could not be completed because one of the checkers is not completed. In order to find out a reason, please go to http://nnstreamer.mooo.com/nntrainer/ci/repo-workers/pr-checker/1578-202109231315220.69390988349915-c29de3dcac3d6dd3d4f8b936bac7a2b365d3b31d/.
gharchive/pull-request
2021-09-17T05:08:46
2025-04-01T04:35:13.964326
{ "authors": [ "kparichay", "taos-ci" ], "repo": "nnstreamer/nntrainer", "url": "https://github.com/nnstreamer/nntrainer/pull/1578", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2305265518
[WIP] [Wait for #2584] [ TEST ] add torch input and output test data for mixed precision In this PR This PR adds torch mixed precision golden data generation and input and output for the test. . some fixes to test. Resolves: Self evaluation: Build test: [X]Passed [ ]Failed [ ]Skipped Run test: [X]Passed [ ]Failed [ ]Skipped Signed-off-by: jijoong.moon jijoong.moon@samsung.com :memo: TAOS-CI Version: 1.5.20200925. Thank you for submitting PR #2591. Please a submit 1commit/1PR (one commit per one PR) policy to get comments quickly from reviewers. Your PR must pass all verificiation processes of cibot before starting a review process from reviewers. If you are new member to join this project, please read manuals in documentation folder and wiki page. In order to monitor a progress status of your PR in more detail, visit http://ci.nnstreamer.ai/. :octocat: cibot: @jijoongmoon, test/include/nntrainer_test_util.h does not include Doxygen tags such as @file @brief @author @bug. You must include the Doxygen tags in the source code. Please refer to a Doxygen manual at http://github.com/nnstreamer/TAOS-CI/blob/main/ci/doc/doxygen-documentation.md :octocat: cibot: @jijoongmoon, A builder checker could not be completed because one of the checkers is not completed. In order to find out a reason, please go to http://ci.nnstreamer.ai/nntrainer/ci/repo-workers/pr-checker/2591-202405201604200.90595698356628-66d13bcda6fb4d33dc335e99842756eb04ec4ec6/. :octocat: cibot: @jijoongmoon, test/include/nntrainer_test_util.h does not include Doxygen tags such as @file @brief @author @bug. You must include the Doxygen tags in the source code. Please refer to a Doxygen manual at http://github.com/nnstreamer/TAOS-CI/blob/main/ci/doc/doxygen-documentation.md :octocat: cibot: @jijoongmoon, A builder checker could not be completed because one of the checkers is not completed. In order to find out a reason, please go to http://ci.nnstreamer.ai/nntrainer/ci/repo-workers/pr-checker/2591-202405220946440.50683903694153-be6c1eede9268cb8bf348eee51c0b0c0046e6ac3/. :octocat: cibot: @jijoongmoon, test/include/nntrainer_test_util.h does not include Doxygen tags such as @file @brief @author @bug. You must include the Doxygen tags in the source code. Please refer to a Doxygen manual at http://github.com/nnstreamer/TAOS-CI/blob/main/ci/doc/doxygen-documentation.md :octocat: cibot: @jijoongmoon, A builder checker could not be completed because one of the checkers is not completed. In order to find out a reason, please go to http://ci.nnstreamer.ai/nntrainer/ci/repo-workers/pr-checker/2591-202405221527370.5227530002594-728fb074979c591cd4b8d6328fb18db5b020c5ed/. :octocat: cibot: @jijoongmoon, test/include/nntrainer_test_util.h does not include Doxygen tags such as @file @brief @author @bug. You must include the Doxygen tags in the source code. Please refer to a Doxygen manual at http://github.com/nnstreamer/TAOS-CI/blob/main/ci/doc/doxygen-documentation.md :octocat: cibot: @jijoongmoon, A builder checker could not be completed because one of the checkers is not completed. In order to find out a reason, please go to http://ci.nnstreamer.ai/nntrainer/ci/repo-workers/pr-checker/2591-202405221623260.88832211494446-4f547971ce01761e435f2c6cd9d3a804f3199c92/. :octocat: cibot: @jijoongmoon, A builder checker could not be completed because one of the checkers is not completed. In order to find out a reason, please go to http://ci.nnstreamer.ai/nntrainer/ci/repo-workers/pr-checker/2591-202405221719140.62206411361694-2d603edd465885797ab17968529b538f96f3c0a2/. :octocat: cibot: @jijoongmoon, A builder checker could not be completed because one of the checkers is not completed. In order to find out a reason, please go to http://ci.nnstreamer.ai/nntrainer/ci/repo-workers/pr-checker/2591-202405230753120.3925359249115-6c293942e0d1aadb7c3a2368c2f2e9136c3a228e/. closed by #2663
gharchive/pull-request
2024-05-20T07:04:17
2025-04-01T04:35:13.978882
{ "authors": [ "jijoongmoon", "taos-ci" ], "repo": "nnstreamer/nntrainer", "url": "https://github.com/nnstreamer/nntrainer/pull/2591", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
506932503
Tests are failing on older python versions For python version 2.6, 3.2, 3.3 travis is failing with this error message: Downloading archive: https://storage.googleapis.com/travis-ci-language-archives/python/binaries/ubuntu/16.04/x86_64/python-2.6.tar.bz2 0.14s$ curl -sSf -o python-2.6.tar.bz2 ${archive_url} curl: (22) The requested URL returned error: 404 Not Found Unable to download 2.6 archive. The archive may not exist. Please consider a different version. Your fix was merged (I managed to confuse github at the same time though, so it marked the PR as closed). Thank you for helping out! Your fix was merged (I managed to confuse github at the same time though, so it marked the PR as closed). Thank you for helping out!
gharchive/issue
2019-10-15T00:11:51
2025-04-01T04:35:13.982226
{ "authors": [ "masell", "noahmorrison" ], "repo": "noahmorrison/chevron", "url": "https://github.com/noahmorrison/chevron/issues/61", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
54027583
khashier displays this transaction incorrectly See http://khashier.com/tx/24acea41498b2f387ff751997f8dc90a4551f7b638e7cfc00750a9966f3bc018 It says the fee is -3.9961 when it was really 0.0039. It says it has 839 outputs when it really has 838. It lists one of the outputs twice, with two different 'spent' links: http://khashier.com/tx/5d46a5d2596e41b3f42c161a362532ce82a54c599b2fb5e75e153aa941acb553 http://khashier.com/tx/53a63ff503e39486cafc279d4ae3feabbbaf539c396404c1fff4ad70512244dd The 'output' link in both of the above points at http://khashier.com/tx/24acea41498b2f387ff751997f8dc90a4551f7b638e7cfc00750a9966f3bc018#o465 The output in question staked twice, but one of them was orphaned. The bug is more obvious here, where the fee is off by around 2000 CLAMs: http://khashier.com/tx/3423091ad3e4dbb17822baeeae773dc9741995a3db9b9ce77732167f8e9a1913 I just noticed it again: http://khashier.com/tx/f1cad229a6fbe5765e854ac194f892049c9af8528d93d7c64145326a9d27b22d
gharchive/issue
2015-01-12T07:45:39
2025-04-01T04:35:13.990972
{ "authors": [ "dooglus" ], "repo": "nochowderforyou/clams", "url": "https://github.com/nochowderforyou/clams/issues/130", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1838270578
[SDXL] Add SDXL pipeline to SHARK -- This commit adds SDXL pipeline to SHARK. Signed-off-by: Abhishek Varma abhishek@nod-labs.com Unrelated CI failure of arith.maxf issue. CC: @monorimet Merging for now, CI issue is just stale artifacts on one of the runners.
gharchive/pull-request
2023-08-06T16:37:06
2025-04-01T04:35:14.005804
{ "authors": [ "Abhishek-Varma", "monorimet" ], "repo": "nod-ai/SHARK", "url": "https://github.com/nod-ai/SHARK/pull/1731", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
427708324
Payload from HttpNode changed in 0.20 when posting a FormData What are the steps to reproduce? I have a Angular app that POST a manually build FormData to a HttpNode in NodeRed with a file as Base64 encoded. Something like this. uploadImaged(dadosImaged: UploadImaged): Observable<HttpEvent<Imaged>> { const url = `${environment.NODE_RED_URL}/uploadImaged`; let body = new FormData(); body.set('codTipoDoc', `${dadosImaged.codTipoDoc}`); body.set('descricao', dadosImaged.descricao); body.set('file', dadosImaged.file); body.set('filename', dadosImaged.filename); body.set('indices', JSON.stringify(dadosImaged.indices)); const req = new HttpRequest('POST', url, body, { reportProgress: true }); return this.http.request<any>(req).pipe(catchError(this.errorHandler)); } What happens? Until version 0.19.X the PAYLOAD of HttpNode was show as a buffer or a raw string in debug node. In version 0.20.X when POSTing to HttpNode an exception is raised. TypeError: invalid media type. Activating the flag Accept file uploads? the payload is parsed but I need the buffer version to make an HttpRequest to another backend. Example: Evidence Image/imgur What do you expect to happen? Payload should be a buffer formdata like in version 0.19 Please tell us about your environment: [x] Node-RED version: 0.20.X [x] node.js version: 8.15.1 [x] npm version: 6.4.1 [x] Platform/OS: Linux Mint [x] Browser: Any Fixed in master - will be in 0.20.4 this week
gharchive/issue
2019-04-01T13:33:50
2025-04-01T04:35:14.024478
{ "authors": [ "knolleary", "marcioluis" ], "repo": "node-red/node-red", "url": "https://github.com/node-red/node-red/issues/2122", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
359616940
Error code: 48001, message: api unauthorized 请问以下是什么问题?我该如何解决? “Error code: 48001, message: api unauthorized” 若还需要什么详情资料,请通知我。谢谢! The WeChat API Call encountered the following exception: Traceback (most recent call last): File "C:\STB\STAN_MG_API\WeChat\wc_follower_total.py", line 30, in extract data = client.user.get_followers() File "C:\STB\STAN_MG_API\WeChat\wechatpy\client\api\user.py", line 60, in get_followers params=params File "C:\STB\STAN_MG_API\WeChat\wechatpy\client\api\base.py", line 13, in _get return self._client.get(url, **kwargs) File "C:\STB\STAN_MG_API\WeChat\wechatpy\client\base.py", line 182, in get **kwargs File "C:\STB\STAN_MG_API\WeChat\wechatpy\client\base.py", line 115, in _request res, method, url, result_processor, **kwargs File "C:\STB\STAN_MG_API\WeChat\wechatpy\client\base.py", line 173, in _handle_result response=res wechatpy.exceptions.WeChatClientException: Error code: 48001, message: api unauthorized hint: [FNYrYa0206k455!] The repo just used for Node.js users.
gharchive/issue
2018-09-12T19:19:19
2025-04-01T04:35:14.031871
{ "authors": [ "JacksonTian", "STANAPI" ], "repo": "node-webot/wechat-api", "url": "https://github.com/node-webot/wechat-api/issues/253", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
146541005
Demo doesn't work No icons, no fonts, no anything: Try later, or another demo instance, it will refresh after 1 hour.
gharchive/issue
2016-04-07T08:07:33
2025-04-01T04:35:14.046929
{ "authors": [ "matusszabo", "wzup" ], "repo": "nodee-apps/cms", "url": "https://github.com/nodee-apps/cms/issues/7", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
109114356
Add wrapped object factory test Similar to the example at: https://nodejs.org/api/addons.html#addons_factory_of_wrapped_objects Thanks, looks good apart from future-proofing. I pushed some changes but please do not merge it yet! It fails when I try to run it on node 0.10.40, will investigate. Closing for now, maybe will raise a new PR. The problems should now be fixed, I tried it on Node 0.10.40 and 4.1.1 (was OK on Node 0.12.7 and 4.0.0 in the past). I added another change where the wrapped object can return another wrapped object (I need this functionality for my own project), and rebased on the latest master changes. Looks good. Test failure on travis seems to be a network error downloading something. Travis aargh @#$% Thanks for looking at this. I pushed a couple more changes: Small fix to test/js/wrappedobjectfactory-test.js: --- a/test/js/wrappedobjectfactory-test.js +++ b/test/js/wrappedobjectfactory-test.js @@ -10,7 +10,7 @@ const test = require('tap').test , testRoot = require('path').resolve(__dirname, '..') , bindings = require('bindings')({ module_root: testRoot, bindings: 'wrappedobjectfactory' }); -test('objectwraphandle', function (t) { +test('wrappedobjectfactory', function (t) { t.plan(4); var obj = bindings.newFactoryObjectInstance(123); Also fix test/cpp/objectwraphandle.cpp to move constructor object into a private static inline function Merged as 4392aa9df745960d553a47e61ae58d8bb811753c bff133ab771eef56315bb2f498a0d95777246f1f 3a4a0ae15bcf35e177d35087063146151c56449f Thanks for merging my changes!
gharchive/pull-request
2015-09-30T15:18:11
2025-04-01T04:35:14.202984
{ "authors": [ "brodybits", "kkoopa" ], "repo": "nodejs/nan", "url": "https://github.com/nodejs/nan/pull/478", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
624685008
gyp: fix XCode CLT version detection on Catalina Since Catalina XCode Command Line Tools don't show up among installed packages, which breaks version detection in gyp/pylib/gyp/xcode_emulation.py. The workaround is to remove the CLT installation directory, and run xcode-select --install which reinstalls an older version of CLT. This older version will eventually be upgraded by Software Update, which breaks gyp again. Checklist [x] commit message follows commit guidelines Description of change This patch adds a fallback mechanism to detect newer CLT versions based on software update history. After every CLT upgrade on 10.15, I have to reinstall Command Line Tools. Please merge and release this, soon. Since we vendor in gyp from https://github.com/nodejs/gyp-next this should really land there. Can we delete the code from this repository that we vendor in from some other repository? we could use git submodules but they're such a developer-hostile solution that it's usually better to just copy the code I think we just need to improve our documentation and get more pro-active with pointing people over there when they open a PR here. The code's the same so moving the PR shouldn't be too difficult since you can just copy your edited files. In times of Github Actions, a bot would help here. If changes are in specific paths, just post a comment about the situation. Unfortunately GitHub Actions running on pull requests from forks do not have permissions to post comments. Unless you are using a own personal access token.
gharchive/pull-request
2020-05-26T08:09:12
2025-04-01T04:35:14.212190
{ "authors": [ "cclauss", "jkroepke", "richardlau", "rvagg", "szekelyisz" ], "repo": "nodejs/node-gyp", "url": "https://github.com/nodejs/node-gyp/pull/2141", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1167149020
Test Issue created at Sat, 12 Mar 2022 02:23:47 GMT Test issue body Sat, 12 Mar 2022 02:23:47 GMT Comment on issue at Sat, 12 Mar 2022 02:23:49 GMT
gharchive/issue
2022-03-12T02:23:48
2025-04-01T04:35:14.487029
{ "authors": [ "nodemationqa" ], "repo": "nodemationqa/nodeQA", "url": "https://github.com/nodemationqa/nodeQA/issues/570", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
810518559
feat(enricher): Adds the rh-metering flag for adding Red Hat Runtimes metering to your application. The new flag is --rh-metering. can also be used as --rh-metering.component-version=NODE_VERSION This is currently an undocumented feature since it is more for those using Red Hat Runtimes build of Node.js Coverage decreased (-0.3%) to 98.299% when pulling 30d26e79c049dd1aa4e54f0824f3d4fc70972c5f on lholmquist:NODE-880-meterting-metadata into 089809e0fe7c8b95be1c6334175dd87b937075b6 on nodeshift:master.
gharchive/pull-request
2021-02-17T20:35:29
2025-04-01T04:35:14.495465
{ "authors": [ "coveralls", "lholmquist" ], "repo": "nodeshift/nodeshift", "url": "https://github.com/nodeshift/nodeshift/pull/543", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
857834213
1.2.10: Startup Crash @[java.lang.VerifyError: class mythicbotany.data.BlockTagProvider overrides final method] Hi. After upgrading to 1.2.10 I get the following crash at startup: java.lang.VerifyError: class mythicbotany.data.BlockTagProvider overrides final method io.github.noeppi_noeppi.libx.data.provider.BlockTagProviderBase.func_200432_c()V The full crash report can be found here: https://pastebin.com/pSZiay8g I have not changed anything else. If there is any further information you need to help track this down, please do let me know. Cheers. You should not use LibX alpha builds to play the game. Will be fixed soon. For now just downgrade LibX to the latest release version. You should not use LibX alpha builds to play the game. I do so mainly for testing purposes as it often helps to pick up bugs. I do not use them in my master pack, just in the testing branch.
gharchive/issue
2021-04-14T12:09:54
2025-04-01T04:35:14.498700
{ "authors": [ "noeppi-noeppi", "sciguyryan" ], "repo": "noeppi-noeppi/MythicBotany", "url": "https://github.com/noeppi-noeppi/MythicBotany/issues/31", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2450932013
refactor views to auto-generate permission_required ref app.test.abstract.views.AllViews Reason: Reduce what is required to create a view. !43 @pytest.mark.skip(reason='write test') def test_view_index_attribute_missing_permission_required(self): """ Attribute missing Test Ensure that `permission_required` attribute is not defined within the view. this can be done by mocking the inherited class with the `permission_required` attribute set to a value that if it changed would be considered defined in the created view. ## Why? This attribute can be dynamically added based of of the view name along with attributes `model._meta.model_name` and `str(__class__.__name__).lower()`. Additional test: - ensure that the attribute does get automagically created. - ensure that the classes name is one of add, change, delete, display or index. """ Requirements [ ] Remove permission_required attribute from all views [ ] ensure class names for views are correctly set. added 5m of time spent marked this issue as related to #74 marked this issue as related to #15
gharchive/issue
2024-07-20T13:28:29
2025-04-01T04:35:14.503726
{ "authors": [ "jon-nfc" ], "repo": "nofusscomputing/centurion_erp", "url": "https://github.com/nofusscomputing/centurion_erp/issues/144", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1682061457
chore: Use pollster as async executor This drops the direct tokio dependency (it still exists inside reqwest) and uses pollster for block_on. Their package readme has a good "why" for itself. I was trying to understand if we can block_on when compiling to JS with wasm-bindgen and found this comment: https://github.com/async-rs/async-std/issues/913#issuecomment-1296282060 which recommended pollster. I tried it in a demo wasm-pack project I have an it seems to block on an async function, so I think we can layer JS promises and it should work 🤞 Converting to draft because it seems that Condvar::wait throws when interop'ing with JS promises. Learned a lot here but closing this. We might use pollster to block_on async functions in something like nargo_cli.
gharchive/pull-request
2023-04-24T21:20:00
2025-04-01T04:35:14.506340
{ "authors": [ "phated" ], "repo": "noir-lang/aztec_backend", "url": "https://github.com/noir-lang/aztec_backend/pull/135", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2706128704
chore: test against noir 1.0.0-beta.0 Description Problem* Resolves Summary* We're looking to flip 1.0.0-beta.0 to an official release so we should test compatibility with it. Additional Context PR Checklist* [x] I have tested the changes locally. [x] I have formatted the changes with Prettier and/or cargo fmt on default settings. i hadn't seen this pr, however, since we're doing a few changes in parallel to this (documentation, fixing one of my mistakes, etcetc), this change will be handled in another pr (#74), so i'm closing this for simplicity.
gharchive/pull-request
2024-11-29T20:42:56
2025-04-01T04:35:14.509003
{ "authors": [ "TomAFrench", "jtriley-eth" ], "repo": "noir-lang/noir-bignum", "url": "https://github.com/noir-lang/noir-bignum/pull/63", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1687471085
chore: Add AI codereviewer Related issue(s) Resolves # Description This adds codeball as a PR reviewer Summary of changes Dependency additions / changes Test additions / changes Checklist [ ] I have tested the changes locally. [ ] I have formatted the changes with Prettier and/or cargo fmt with default settings. [ ] I have linked this PR to the issue(s) that it resolves. [ ] I have reviewed the changes on GitHub, line by line. [ ] I have ensured all changes are covered in the description. Documentation needs [ ] This PR requires documentation updates when merged. Additional context Okay thought this was gonna be a nice version of codereview.gpt -- it only applies a label :(
gharchive/pull-request
2023-04-27T19:59:55
2025-04-01T04:35:14.513200
{ "authors": [ "kevaundray" ], "repo": "noir-lang/noir", "url": "https://github.com/noir-lang/noir/pull/1242", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1701671909
chore: fix installation link in readme Related issue(s) Fixes https://github.com/noir-lang/noir/issues/1089#issuecomment-1539010915 (again) Description Summary of changes Dependency additions / changes Test additions / changes Checklist [x] I have tested the changes locally. [x] I have formatted the changes with Prettier and/or cargo fmt with default settings. [x] I have linked this PR to the issue(s) that it resolves. [x] I have reviewed the changes on GitHub, line by line. [x] I have ensured all changes are covered in the description. Documentation needs [ ] This PR requires documentation updates when merged. Additional context (I think you'll need approval from someone on the Noir team :) ) (I think you'll need approval from someone on the Noir team :) ) Ah wasn't aware you guys couldn't as we can approve in Aztec org.
gharchive/pull-request
2023-05-09T09:09:40
2025-04-01T04:35:14.517859
{ "authors": [ "TomAFrench", "iAmMichaelConnor" ], "repo": "noir-lang/noir", "url": "https://github.com/noir-lang/noir/pull/1326", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1308959606
スマホや iPad でシーンの下部が画面外にはみ出る 予期しない動作の内容 | Describe the unexpected behavior スマホや iPad でシーンの下部 2〜5 % ほどが画面外にはみ出る 再現方法 | To Reproduce 発生環境(モバイルの場合) | Smartphone (please complete the following information): デバイス | Device: Android, iPad OS: Android, iOS バージョン | Version: - ブラウザ (Web 版) | Browser: - 備考 | Additional context Reported by @Reputeless Commented by @su8ru on Siv3D Slack #app の height: 100vh 指定が原因かもしれません。 モバイルでの 100vh 指定はいろいろ問題を抱えており、js で window.innerHeight を参照する回避が一般的だと思います。 Commented by @nokotan on Siv3D Slack 該当箇所確認できました Siv3D プロジェクトテンプレートに付属している html テンプレートに以下の処理を追加してみています appContainer.style.height = `${window.innerHeight}px`;
gharchive/issue
2022-07-19T04:25:23
2025-04-01T04:35:14.530052
{ "authors": [ "nokotan" ], "repo": "nokotan/OpenSiv3D", "url": "https://github.com/nokotan/OpenSiv3D/issues/21", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
328708253
Modernize codebase Drop Bower support Add ES module support Use StandardJS Update dependencies Also add tests to get us to 100% code coverage
gharchive/pull-request
2018-06-02T03:15:23
2025-04-01T04:35:14.533011
{ "authors": [ "nolanlawson" ], "repo": "nolanlawson/chord-magic", "url": "https://github.com/nolanlawson/chord-magic/pull/5", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1148548517
Popping behavior Check that this is really a bug [X] I confirm Reproduction link https://atroposjs.com/ Bug description When a multi-component Atropos collection with a common eventsEl is moving, some of the elements will "pop" instead of smoothly animating. This manifests differently in different browsers. On Chrome the popping is with the scale of the element. On Safari the popping seems to be with the glare. https://user-images.githubusercontent.com/157106/155403478-689bf63c-f65c-4967-ab44-2102c40dd683.mov https://user-images.githubusercontent.com/157106/155403501-544484ec-9116-43c8-9834-6dd77706e8d9.mov Expected Behavior No response Actual Behavior No response Atropos version 1.0.2 Platform/Target and Browser Versions macOS: Chrome, Safari Validations [X] Follow our Code of Conduct [X] Read the docs. [X] Check that there isn't already an issue that request the same feature to avoid creating a duplicate. [X] Make sure this is a Atropos issue and not a framework-specific issue Would you like to open a PR for this bug? [ ] I'm willing to open a PR Is there maybe already a solution or possible quickfix for this ?
gharchive/issue
2022-02-23T20:33:14
2025-04-01T04:35:14.538300
{ "authors": [ "kylemcdonald", "triiickstaa" ], "repo": "nolimits4web/atropos", "url": "https://github.com/nolimits4web/atropos/issues/19", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
755151109
Navigation Arrows not showing on IE This is a (multiple allowed): [x] bug [ ] enhancement [ ] feature-discussion (RFC) Swiper Version: EXACT RELEASE VERSION OR COMMIT HASH, HERE. Platform/Target and Browser Versions: PLATFORM CLIENT YOU ARE TARGETING SUCH AS macOS, Windows, CORDOVA, IOS, ANDROID, CHROME, ETC. Live Link or JSFiddle/Codepen or website with isssue: PREFERABLY (IF YOU WANT YOUR ISSUE TO BE RESOLVED ASAP). What you did I couldn't see the arrows. Expected Behavior See Arrows Actual Behavior Arrows only visible on Firefox and Chrome P.S. Remember, an issue is not the place to ask questions. You can use Stack Overflow for that. Before you open an issue, please check if a similar issue already exists or has been closed before. IE is not supported by Swiper
gharchive/issue
2020-12-02T10:35:32
2025-04-01T04:35:14.542874
{ "authors": [ "EchoChen012", "nolimits4web" ], "repo": "nolimits4web/swiper", "url": "https://github.com/nolimits4web/swiper/issues/3996", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2302203413
Closing a modal changes active slide and breaks loop mode Check that this is really a bug [X] I confirm Reproduction link https://codesandbox.io/p/sandbox/swiper-element-boostrap-modal-x5535x Bug description I have a swiper that shows multiple slides per view, and when clicking the anchor inside the slide, a boostrap modal is triggered. However, when clicking a slide that is not the central one and closing the modal, the active slide is changed to the one that triggered the modal. Also, if it is the last one, loop mode completely breaks and no next slide is shown. The boostrap modal is not inside the swiper-container. Here's a demo video from the codesandbox I provided: screen-capture.webm Expected Behavior Closing a modal should not change the current active slide. Actual Behavior No response Swiper version 11.1.3 Platform/Target and Browser Versions Windows 10, Chrome 124 Validations [X] Follow our Code of Conduct [X] Read the docs. [X] Check that there isn't already an issue that request the same feature to avoid creating a duplicate. [X] Make sure this is a Swiper issue and not a framework-specific issue Would you like to open a PR for this bug? [ ] I'm willing to open a PR Thanks for the quick fix!! Any estimated date for the release?
gharchive/issue
2024-05-17T09:09:51
2025-04-01T04:35:14.548329
{ "authors": [ "mdmontesinos" ], "repo": "nolimits4web/swiper", "url": "https://github.com/nolimits4web/swiper/issues/7540", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2506913541
TypeError when Swiper instance is destroyed right after zooming Check that this is really a bug [X] I confirm Reproduction link https://codesandbox.io/p/devbox/swiper-react-zoom-pqy2gq Bug description On a project we have an image gallery build on top of Swiper. We have the zoom functionality enabled. On mobile we have a functionality in place that checks whether or not you have swiped the image in the vertical direction (gallery set to horizontal mode) and we close the gallery when this happens. However, both things - zooming + swiping (onTouchStart - onTouchEnd) - can happen at the same time. This means there are cases in which the swiper gallery is closed right after the zoom happens. During zooming, there is a setTimeout call with callback that invokes the allowTouchMove https://github.com/nolimits4web/swiper/blob/d0b6abd74805398ac3cbf41aeebf141a805ec64b/src/modules/zoom/zoom.mjs#L249 However, the asynchronous nature of setTimeout in our use case makes it happen after the instance is destroyed. Which leads to TypeErrors: ../../node_modules/swiper/modules/zoom.mjs in allowTouchMove at line 216:1 TypeError Cannot set properties of undefined (setting 'preventTouchMoveFromPointerMove') Reproduction link note: Our setup to reproduce this issue is quite complex and unreliable, but I think the issue is hopefully described/obvious enough. If not, I can try to setup a proper repro sandbox. Expected Behavior No errors are thrown after the Swiper instance is destroyed right after zooming. Actual Behavior Unreliably the error is thrown after the Swiper instance is destroyed right after zooming. Swiper version 11.1.4 Platform/Target and Browser Versions Android Chrome 128 Validations [X] Follow our Code of Conduct [X] Read the docs. [X] Check that there isn't already an issue that request the same feature to avoid creating a duplicate. [X] Make sure this is a Swiper issue and not a framework-specific issue Would you like to open a PR for this bug? [X] I'm willing to open a PR Fixed by https://github.com/nolimits4web/swiper/pull/7723
gharchive/issue
2024-09-05T06:45:28
2025-04-01T04:35:14.554807
{ "authors": [ "Heptonius", "nolimits4web" ], "repo": "nolimits4web/swiper", "url": "https://github.com/nolimits4web/swiper/issues/7722", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2038949947
🛑 f.cz is down In ab5e208, f.cz (https://f.cz/nodeinfo/2.0) was down: HTTP code: 0 Response time: 0 ms Resolved: f.cz is back up in 371375e after 5 minutes.
gharchive/issue
2023-12-13T04:50:42
2025-04-01T04:35:14.558007
{ "authors": [ "matejdivecky" ], "repo": "nolog-it/mastodon-uptime", "url": "https://github.com/nolog-it/mastodon-uptime/issues/447", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
103137290
change julia REQUIRE to 0.3 since Colors has no 0.2-compatible tagged versions at this time ref https://github.com/nolta/Winston.jl/pull/235/files#r37931603 Also noticed in #237. But it would be fine to merge this separately. Ah, whoops. Saw that you had done the same thing for ProfileView so held off there, but missed this one. Not too critical I think if bigger changes are in the works.
gharchive/pull-request
2015-08-25T23:22:19
2025-04-01T04:35:14.559548
{ "authors": [ "timholy", "tkelman" ], "repo": "nolta/Winston.jl", "url": "https://github.com/nolta/Winston.jl/pull/238", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
63718497
Install takes ages Hi, Is there any way to optimize this tool? We use it on Travis-CI and install takes about 160 seconds, which is a while for single tool. 154.00s $ gem install cupertino --no-ri --no-rdoc Fetching: highline-1.6.21.gem (100%) Successfully installed highline-1.6.21 Fetching: commander-4.2.1.gem (100%) Successfully installed commander-4.2.1 Fetching: terminal-table-1.4.5.gem (100%) Successfully installed terminal-table-1.4.5 Fetching: term-ansicolor-1.0.7.gem (100%) Successfully installed term-ansicolor-1.0.7 Fetching: net-http-digest_auth-1.4.gem (100%) Successfully installed net-http-digest_auth-1.4 Fetching: net-http-persistent-2.9.4.gem (100%) Successfully installed net-http-persistent-2.9.4 Fetching: mime-types-1.25.1.gem (100%) Successfully installed mime-types-1.25.1 Fetching: mini_portile-0.6.2.gem (100%) Successfully installed mini_portile-0.6.2 Fetching: nokogiri-1.6.6.2.gem (100%) Building native extensions. This could take a while... Successfully installed nokogiri-1.6.6.2 Fetching: ntlm-http-0.1.1.gem (100%) Successfully installed ntlm-http-0.1.1 Fetching: webrobots-0.1.1.gem (100%) Successfully installed webrobots-0.1.1 Fetching: unf_ext-0.0.6.gem (100%) Building native extensions. This could take a while... Successfully installed unf_ext-0.0.6 Fetching: unf-0.1.4.gem (100%) Successfully installed unf-0.1.4 Fetching: domain_name-0.5.23.gem (100%) Successfully installed domain_name-0.5.23 Fetching: mechanize-2.5.1.gem (100%) Successfully installed mechanize-2.5.1 Fetching: security-0.1.3.gem (100%) Successfully installed security-0.1.3 Fetching: certified-1.0.0.gem (100%) IMPORTANT: Remember to use the included executable `certifed-update` regularly to keep your certificate bundle up to date. Successfully installed certified-1.0.0 Fetching: cupertino-1.3.0.gem (100%) Successfully installed cupertino-1.3.0 18 gems installed Most of the time was taken by "Building native extensions. This could take a while...". That's just the price of doing business in Ruby. I'd entertain any PRs looking to speed things up, but this is not a priority.
gharchive/issue
2015-03-23T13:02:37
2025-04-01T04:35:14.571068
{ "authors": [ "AndrewDryga", "mattt" ], "repo": "nomad/cupertino", "url": "https://github.com/nomad/cupertino/issues/194", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
53592878
build error for Xcode6.1.1 I got a following issue.. xcodebuild MyApp.xcworkspace 2015-01-07 11:52:13.902 xcodebuild[3989:42121] [MT] DVTAssertions: Warning in /SourceCache/IDEFrameworks/IDEFrameworks-6611/IDEFoundation/SourceControl/Model/IDESourceControlManager.m:423 Details: Error Domain=com.apple.dt.IDESourceControlErrorDomain Code=-1 "Missing extension: public.vcs.subversion" UserInfo=0x7fd084558ad0 {NSLocalizedDescription=Missing extension: public.vcs.subversion} Object: <IDESourceControlManager: 0x7fd0840d7e40> Method: -loadRepositories Thread: <NSThread: 0x7fd082c35910>{number = 1, name = main} Please file a bug at http://bugreport.apple.com with this warning message and any useful information you can provide. ** BUILD FAILED ** The following build commands failed: Check dependencies (1 failure) ** ARCHIVE FAILED ** sorry.. duplicated #176
gharchive/issue
2015-01-07T03:15:41
2025-04-01T04:35:14.574473
{ "authors": [ "tikidunpon" ], "repo": "nomad/shenzhen", "url": "https://github.com/nomad/shenzhen/issues/177", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1079277805
请问正常使用中突然报错重开显示TUNNEL - music.163.com:443 如图 同问,今天上午正常用,中午也遇到这情况了 同问 现在无法正常播放 https://github.com/nondanee/UnblockNeteaseMusic/issues/85 按照这个可以解决 #85按照这个可以解决 你好,linux客户端生成证书替换后,还是报错 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 MITM > clientlog.music.163.com #85按照这个可以解决 你好,linux客户端生成证书替换后,还是报错 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 MITM > clientlog.music.163.com 添加HTTPS代理端口了吗 #85 按照这个可以解决 你好,linux客户端生成证书替换后,还是报错 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 MITM > clientlog.music.163.com 添加HTTPS代理端口号码 没有,之前我是源代码部署的 只开启了一个http端口 node app.js -p 33123 #85 按照这个可以解决 你好,linux客户端生成证书替换后,还是报错 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 MITM > clientlog.music.163.com 添加HTTPS代理端口号码 没有,之前我是源代码部署的 只开启了一个http端口 node app.js -p 33123 添加HTTPS代理试试 node app.js -p 33123:33124 这样的 #85 按照这个可以解决 你好,linux客户端生成证书替换后,还是报错 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 MITM > clientlog.music.163.com 添加HTTPS代理端口号码 没有,之前我是源代码部署的 只开启了一个http端口 node app.js -p 33123 添加HTTPS代理试试 node app.js -p 33123:33124 这样的 请问一下,两个端口的话 网易代理配置 怎么设置呢 网易客户端只有 服务器+端口 #85 按照这个可以解决 你好,linux客户端生成证书替换后,还是报错 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 MITM > clientlog.music.163.com 添加HTTPS代理端口号码 没有,之前我是源代码部署的 只开启了一个http端口 node app.js -p 33123 添加HTTPS代理试试 node app.js -p 33123:33124 这样的 请问一下,两个端口的话 网易代理配置 怎么设置呢 网易客户端只有 服务器+端口 网易客户端还按原来的配置就行 网易客户端还按原来的配置就行 直接走http的端口33123吗 不用配置https 33124? 网易客户端还按原来的配置就行 直接走http的端口33123吗 不用配置https 33124? 对 #85 按照这个可以解决 你好,linux 客户端生成证书替换后,还是报错 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 MITM > clientlog.music.163.com 添加 HTTPS 代理端口号码 没有,之前我是源代码部署的 只开启了一个 http 端口 node app.js -p 33123 添加 HTTPS 代理试试 node app.js -p 33123:33124 这样的 请问一下,node用这个代码可以使用,但pm2为什么就不可以(我需要让它在window后台运行) 代码: pm2 start app.js -p 8080:8081 Log: address: '::', 0|app | port: 8080 0|app | } 0|app | Error: listen EADDRINUSE: address already in use :::8080 0|app | at Server.setupListenHandle [as _listen2] (net.js:1318:16) 0|app | at listenInCluster (net.js:1366:12) 0|app | at Server.listen (net.js:1452:7) 0|app | at D:\Program\UnblockNeteaseMusic-0.25.3\src\app.js:83:27 0|app | at processTicksAndRejections (internal/process/task_queues.js:93:5) { 0|app | code: 'EADDRINUSE', 0|app | errno: -4091, 0|app | syscall: 'listen', 0|app | address: '::', 0|app | port: 8080 0|app | } 我的今天不能用了,都显示网络不给力哦,请检查你的网络设置,关键是我docker和windows里面运行的都是这个错误了。估计是接口改了还是怎么地 问题已解决,感谢大佬们 网易客户端还原来的配置就行 直接走http的端口33123吗不用配置https 33124? 对 3Q,按照您的方法,已解决!!! #85 按照这个可以解决 你好,linux客户端生成证书替换后,还是报错 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 TUNNEL - music.163.com:443 MITM > clientlog.music.163.com 添加HTTPS代理端口号码 没有,之前我是源代码部署的 只开启了一个http端口 node app.js -p 33123 添加HTTPS代理试试 node app.js -p 33123:33124 这样的 现在不行,网络连接错误,还有什么办法吗
gharchive/issue
2021-12-14T03:50:39
2025-04-01T04:35:14.612395
{ "authors": [ "ClankySun10936", "GenBUPT", "TDSofTM", "XianyuOne", "cg-llc", "jingouwangzi", "kk748548942", "silveryvoid" ], "repo": "nondanee/UnblockNeteaseMusic", "url": "https://github.com/nondanee/UnblockNeteaseMusic/issues/795", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1852622783
Encoding error when writing proposed_changes.txt file The bug When 'iambic plan' command is run on updated yaml file(s) the changes should be written to a proposed_changes.txt file. But 'UnicodeEncodeError' exception occurs. Error Traceback is added at the end for reference. To Reproduce Run 'iambic plan' command on updated yaml file(s) Expected behavior Should write proposed_changes.txt file without any error Desktop OS: Windows 10 Version 19045.3324 Python 3.10.11 Error Traceback 2023/08/16 11:54:54 [error ] Unhandled Exception error=UnicodeEncodeError('charmap', 'IAMbic Summary\r\n\r\nChange Detection\r\n\r\n* 1 Update actions.\r\n* 1 templates with changes.\r\n* 1 accounts affected.\r\n\r\nIAMbic Change Details\r\nUpdate\r\n└── resources/aws/iam/ma naged_policy/redacted redacted/ssm-parameter-store-access.yaml\r\n └── redacted redacted - (redacted)\r\n └── aws:policy_document // arn:aws:iam::redacted:policy/ssm-parameter-store-access\r\n └── policy _document:\r\n -iterable_item_removed\r\n {"root['Statement'][0]['Action'][1]": 'ssm:DeleteParameter', "root['Statement'][0]['Action'][5]": 'ssm:DeleteParameters'}\r\n', 145, 148, 'chara cter maps to ') exception=UnicodeEncodeError stacktrace= File "", line 198, in _run_module_as_main File "", line 88, in run_code File "redacted\venv\Scripts\iambic.exe_main.py", line 7, in ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "redacted\click\core.py", line 783, in invoke return __callback(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "redacted\iambic\main.py", line 360, in plan run_plan(templates, repo_dir=repo_dir) File "redacted\iambic\main.py", line 400, in run_plan output_proposed_changes(template_changes) File "redacted\iambic\main.py", line 54, in output_proposed_changes file_render_resource_changes(output_path, template_changes) File "redacted\iambic\output\text.py", line 25, in file_render_resource_changes f.write(rendered_data) File "redacted\Python\Python311\Lib\encodings\cp1252.py", line 19, in encode return codecs.charmap_encode(input,self.errors,encoding_table)[0] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Pull Request #576 Fix is included in iambic-core 0.11.48, available in PyPi
gharchive/issue
2023-08-16T06:51:58
2025-04-01T04:35:14.643247
{ "authors": [ "0xAJX", "smoy" ], "repo": "noqdev/iambic", "url": "https://github.com/noqdev/iambic/issues/575", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
93614414
Add null-check when updating or destroying an instance. Hey Jun, (not sure if the best practice was to do another pull request) but here it is. So I've found the issue is with the scroll bar being a direct child of an angular ui-router element. I have added in the console.error() as you suggested. But using console isn't supported by all browsers, (IE) especially, and not sure it's good to use here. I'd suggest removing it, imo. Thanks, Adam As you suggested, I just remove the line throwing an error and make it return early. Pls refer to 6c642d8a478ce231aad98aee7f6cd541cac816d2. Thanks!
gharchive/pull-request
2015-07-07T19:54:36
2025-04-01T04:35:14.650890
{ "authors": [ "adammiles10", "noraesae" ], "repo": "noraesae/perfect-scrollbar", "url": "https://github.com/noraesae/perfect-scrollbar/pull/359", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1466604244
feat: configure GoRelease to publish to homebrew Scope: configured GoReleaser to publish to homebrew Test: I didn't want to push a new tag and trigger a regular release so I could only test my changes by doing a local-only release with the command: goreleaser release --snapshot --rm-dist this generated dist/fman.rb formula file that would be published to homebrew-tab repo Thanks
gharchive/pull-request
2022-11-28T15:15:08
2025-04-01T04:35:14.654576
{ "authors": [ "nore-dev", "tomekz" ], "repo": "nore-dev/fman", "url": "https://github.com/nore-dev/fman/pull/61", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2345930314
boards: norik: Add support for Norik Octopus SoM & IO-Board Add support for Norik Octopus SoM & IO-Board based on nRF9160 SiP. Changed Norik systems d.o.o. to Norik Systems as requested. @FPlohl you still have one rename left in "boards/norik/octopus_som/board.cmake" to be consistent. @pfiser fixed remaining inconsistencies. Clear to open PR to main branch? yes, @FPlohl go ahead and open PR on upstream zephyr repo. Fingers crossed :)
gharchive/pull-request
2024-06-11T09:41:39
2025-04-01T04:35:14.659264
{ "authors": [ "FPlohl", "pfiser" ], "repo": "norikdoo/zephyr", "url": "https://github.com/norikdoo/zephyr/pull/2", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }