id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
820553410
build(typescript): update typescript dependencie to fix npx tsc issue reference: https://stackoverflow.com/questions/45790182/typescript-error-ts5014-unexpected-token-u-in-json-at-position-0 #186 This don't fix #186 for me
gharchive/pull-request
2021-03-03T02:08:07
2025-04-01T06:40:11.920335
{ "authors": [ "prescindivel" ], "repo": "react-native-text-input-mask/react-native-text-input-mask", "url": "https://github.com/react-native-text-input-mask/react-native-text-input-mask/pull/187", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1211930608
Minimum React Native version support I would like to only support RN versions from the past year. The thinking is that if you are not upgrading your RN app for more than a year, you are also probably not upgrading core dependencies like this one. This would put us at RN v0.64. This means publishing patches to older major versions if the latest has more recent requirements. For example, if we decide that v7 only supports RN 0.68 with new architecture, we will continue to support v6 for RN versions up to a year old. Thoughts? I agree it can be problematic to maintain old react-native version. In all RN modules I saw, they said they don't support of version ... Now with turbo module and new architecture, I am not sure of the exact impact in terms of dependancies. maybe @douglowder can give some clues react-native-maps already does something similar, even more strict actually: https://github.com/react-native-maps/react-native-maps#compatibility
gharchive/issue
2022-04-22T07:22:58
2025-04-01T06:40:11.932136
{ "authors": [ "Adnan-Bacic", "freeboub", "hueniverse" ], "repo": "react-native-video/react-native-video", "url": "https://github.com/react-native-video/react-native-video/issues/2655", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1339133662
not displaying properly on android Bug ##Android Environment info react-native version "react-native": "^0.64.2", "react-native-video": "^5.2.0", taking time to display video in flatlist with images to be followed in: https://github.com/react-native-video/react-native-video/issues/2668
gharchive/issue
2022-08-15T15:19:55
2025-04-01T06:40:11.934478
{ "authors": [ "freeboub", "ubaid-wp" ], "repo": "react-native-video/react-native-video", "url": "https://github.com/react-native-video/react-native-video/issues/2814", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1841114174
Stopping momentum scroll triggers click event on Android #863 Same issue, but not fixed yet react-native version: 0.71.6 react-native-webview version: 11.26.0 I just tried to reproduce this and I see exactly the same behavior in Chrome. If you see different behavior between the WebView and Chrome, please share a sample repro.
gharchive/issue
2023-08-08T11:22:53
2025-04-01T06:40:11.936151
{ "authors": [ "TheAlmightyBob", "nj0034" ], "repo": "react-native-webview/react-native-webview", "url": "https://github.com/react-native-webview/react-native-webview/issues/3090", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
386340499
You can't navigate back from last tab to first on iOS, if the tab has auto width Current Behavior When I give a width: 'auto' or width: null, flex: 1 to the tabStyle, I can't navigate back from last page if there is not enough tab to overflow the screen. Expected Behavior Either to be able to scroll the tabbar or to not move the selected tab item to the left side of the screen like on android. How to reproduce In the demo, on iOS select the Settings2 menu and try to navigate back to Home menu. A device which is big enough either on portrait or landscape to show all tabs. Your Environment software version react-navigation 2.18.2 react-native 0.57 node 10.13.0 npm or yarn yarn @satya164 - not sure what the expected behavior is for top tabs or if this is supported. @Xyzor it might be helpful if you share a mockup or description of what you're trying to accomplish My main goal is to use flex: 1 on tabStyle, instead of fixed width, because with fixed width the tab's label either break into new lines or some tabs are unnecessarily wide. On android the demo is working because when i select a tab, it doesn't move to the left. The width depends on the width of the tab bar. Dynamic width for the tab items is not supported because it's every complicated to implement. I'm not sure what you're trying to achieve, but you probably need to customize the tab bar itself or use a custom tab bar. @satya164 Thanks for the info.
gharchive/issue
2018-11-29T15:06:01
2025-04-01T06:40:11.941488
{ "authors": [ "Xyzor", "brentvatne", "satya164" ], "repo": "react-navigation/react-navigation-tabs", "url": "https://github.com/react-navigation/react-navigation-tabs/issues/73", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
585788576
Direct link not building hierarchy as desired Context/flow If it helps, this is what I'm trying to accomplish: Venues shows a list of venues Clicking on one of the venues takes you to VenueDetails At the venue you can add stuf to your cart, and open up your cart, which takes you to OrderDetails I need a way for the url to remember I took this route so I can go back down the chain in case the user refreshes or opens the link in a new tab. Description I've set up my linking as follows: export default function (containerRef) { return useLinking(containerRef, { prefixes: [Linking.makeUrl('/')], config: { Root: { initialRouteName: 'Venues', path: '', screens: { Venues: '', VenueDetails: { path: 'venue/:slug', parse: { slug: String }, }, OrderDetails: { path: 'venue/:slug/order-overview', parse: { slug: String }, }, } } } }) } What I want is to open OrderDetails when I navigate to /venue/some-slug/order-overview. If I revert the order in the useLinking object (so: OrderDetails, VenueDetails and then Venue) this works, but it doesn't see that VenueDetails is the previous page. How can I tell react-navigation/linking that going "back" from OrderDetails means it should go to VenueDetails? And that going back from VenueDetails means it should go to Venue? Can I define a parent somehow? Maybe I need to put it in the url somehow? Note: it works fine when starting at the home page and doing the normal navigation flow. It's only when I open up a nested page directly that this fails. Is this repo still being worked on by the way? The last release was a while ago. I'd like to help but I don't know where to start. I believe this repo deprecated and moved on react-navigation this file is for web https://github.com/react-navigation/react-navigation/blob/master/packages/native/src/useLinking.tsx
gharchive/issue
2020-03-22T19:03:37
2025-04-01T06:40:11.992223
{ "authors": [ "RWOverdijk", "soroushm" ], "repo": "react-navigation/web", "url": "https://github.com/react-navigation/web/issues/50", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1662592497
PrismLight using results in regular [object Object] output (but only in production) Describe the bug I'm using import { PrismLight as SyntaxHighlighter } from 'react-syntax-highlighter'; which results in regular [object Object] [object Object] ... output (around 7 of 10 refreshs of my webpage) for my code blocks (but only in production). Notice: import { Prism as SyntaxHighlighter } from 'react-syntax-highlighter'; has the same problem. Every 4th to 5th refresh, it's working correctly, otherwise it's rendering [object Object] ... To Reproduce Steps to reproduce the behavior: This is my implementation, notice that the processedCode output of the process function is always correct (also in production). But even without this process function, the problem persists. import { PrismLight as SyntaxHighlighter } from 'react-syntax-highlighter'; import tsx from 'react-syntax-highlighter/dist/esm/languages/prism/tsx'; import typescript from 'react-syntax-highlighter/dist/esm/languages/prism/typescript'; import { vs, vscDarkPlus } from 'react-syntax-highlighter/dist/esm/styles/prism'; SyntaxHighlighter.registerLanguage('tsx', tsx); SyntaxHighlighter.registerLanguage('typescript', typescript); interface ICodeBlock { code: string; language: 'typescript' | 'tsx'; } const process = (code = '') => { let skippedLeadingEmptyLines = false; let lastNonEmptyLineIndex = 0; let minRawStringIndentation = Number.MAX_SAFE_INTEGER; let numberOfRemovedLines = 0; const processNonEmptyLine = (line: string, index: number) => { // keep track of the index of the last non-empty line lastNonEmptyLineIndex = index - numberOfRemovedLines; // determine the minimum indentation level minRawStringIndentation = Math.min(minRawStringIndentation, Math.max(0, line.search(/\S/))); // return the processed line return [line.trimEnd()]; }; // split code into lines const codeLines = code.split('\n'); // remove empty lines, and process non-empty lines const nonEmptyLinesAtStart = codeLines.flatMap((line, index) => { if (!skippedLeadingEmptyLines) { if (line.match(/^\s*$/)) { numberOfRemovedLines += 1; return []; } skippedLeadingEmptyLines = true; return processNonEmptyLine(line, index); } if (line.match(/^\s*$/)) return ['']; return processNonEmptyLine(line, index); }); const nonEmptyLinesStartAndEnd = nonEmptyLinesAtStart.slice(0, lastNonEmptyLineIndex + 1); // If there are no non-empty lines, return an empty string if (nonEmptyLinesStartAndEnd.length === 0) return ''; const nonRawStringIndentationLines = minRawStringIndentation !== 0 ? nonEmptyLinesStartAndEnd.map((line) => line.substring(minRawStringIndentation)) : nonEmptyLinesStartAndEnd; return nonRawStringIndentationLines.join('\n'); }; export const CodeBlock = ({ code, language }: ICodeBlock) => { const { isLight } = useThemeContext(); const processedCode = process(code); const theme = isLight ? vs : vscDarkPlus; return ( <pre className={classes.pre}> <SyntaxHighlighter language={language} style={theme}> {processedCode} </SyntaxHighlighter> </pre> ); }; It's working when I do a bad workaround to force reloading of the component: // ... useEffect(() => { const timerId = setTimeout(() => setIsReloaded(true), 0); return () => clearTimeout(timerId); }, []); return isReloaded ? ( <pre className={classes.pre}> <SyntaxHighlighter language={language} style={theme}> {processedCode} </SyntaxHighlighter> </pre> ) : null; Expected behavior Output my code string (it's a simple code string of a React Component) instead of [object Object] in some cases. Oddly enough in my dev environment it's always working fine, only in production is the rendering issue with [object Object] Screenshots Desktop (please complete the following information): Browser firefox Version 111.0.1 (64-Bit) I'm getting this issue on a non-local environment as well We see this same issue intermittently with PrismAsyncLight as well. For others with this issue, using @fenkx's fork fixes the issue, the easiest way to switch is just change this line in your package.json: "react-syntax-highlighter": "npm:@fengkx/react-syntax-highlighter@15.6.1" I have the same issue but I guess this package is not maintained anymore 😕 I had the same problem, but the above scenario didn't solve my problem. +1 May I add my opinion here? I'm experiencing the same issue. Hopefully, it has been resolved, but I want to share this for anyone who might be facing the same problem. I think the following procedure is causing the issue: Open a page with target="_blank" (in this case, the user only opens the tab without switching to it). Once the tab is loaded and the user switches to it, the code is displayed as [Object object]. Therefore, I have made it re-render when the user switches tabs. Below is a part of the code I wrote. function CodeBlock({/* ... */}) { const [codeBlock, setCodeBlock] = useState<ReactElement | null>(null); useEffect(() => { renderCodeBlock(); document.addEventListener('visibilitychange', renderCodeBlock); return () => { setCodeBlock(() => null); document.removeEventListener('visibilitychange', renderCodeBlock); }; }, []); function renderCodeBlock() { setCodeBlock(() => ( <Box component={SyntaxHighlighter} language={language} showLineNumbers style={atomDark} > {code.trim()} </Box> )); } return ( <Box> {codeBlock === null ? ( <CircularProgress /> ) : ( codeBlock )} </Box> ); } Please note that this text has been translated and might contain some unnatural parts. This happens to me as well, what can I do about this? I'm experiencing this issue as well. I was able to work around it by adding this to my index.tsx file before I render my React app: import { PrismAsyncLight as SyntaxHighlighter } from 'react-syntax-highlighter' ReactDOM.createRoot(document.createElement('div')).render( <SyntaxHighlighter language="" children={''} /> ) Can confirm, the workaround by jeremyckahn works with SSR and production. No more problems when using hydrateRoot.
gharchive/issue
2023-04-11T14:16:04
2025-04-01T06:40:12.007273
{ "authors": [ "C5H8NNaO4", "FujiwaraChoki", "MatchuPitchu", "ShaofeiZi", "bodinsamuel", "humphd", "jeremyckahn", "jtsorlinis", "kkn1125", "lritter79" ], "repo": "react-syntax-highlighter/react-syntax-highlighter", "url": "https://github.com/react-syntax-highlighter/react-syntax-highlighter/issues/513", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
373564533
Vertical scrollbar hiding last column data Describe the bug Vertical scrollbar hiding last column data if we have horizontal bar too and last column width is reduced. To Reproduce Any react table with 3 or more columns, increase the width of middle column and reduce the width of last column , when you scroll right to end you will find one or two characters are hiding under vertical scroll bar. Expected behavior Vertical scrollbar should be post last column width should not hide the details showing in last column Codesandbox! you can use https://codesandbox.io/s/o5np5p0nrz codesand box and increase width of columns adding screen shot too. Screenshots Desktop (please complete the following information): Windows / Crome Browser - chrome Version 68.0.3440.106 RT wasn't designed for inline-scrolling, hence the pagination, so unfortunately, there is no official support for this use-case. You're more than welcome to explore and ask in the forum though! https://spectrum.chat/react-table @Mulli can you suggest how to do ? looking for few workarounds
gharchive/issue
2018-10-24T16:17:01
2025-04-01T06:40:12.012480
{ "authors": [ "Mukesh-Bhootra", "tannerlinsley" ], "repo": "react-tools/react-table", "url": "https://github.com/react-tools/react-table/issues/1167", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
541303755
Errors after downloading JavaScript bundle Warning: React.createElement: type is invalid -- expected a string (for built-in components) or a class/function (for composite components) but got: %s.%s%s, undefined, You likely forgot to export your component from the file it's defined in, or you might have mixed up default and named imports. Check your code at Button.js:38., in Button (at Welcome.js:144) in RCTView (at Block.js:155) in Block (at Welcome.js:143) in RCTView (at Block.js:155) in Block (at Welcome.js:129) in Welcome (at SceneView.js:9) in SceneView (at StackViewLayout.tsx:899) in RCTView (at StackViewLayout.tsx:898) in RCTView (at StackViewLayout.tsx:897) in RCTView (at createAnimatedComponent.js:151) in AnimatedComponent (at StackViewCard.tsx:106) in RCTView (at createAnimatedComponent.js:151) in AnimatedComponent (at screens.native.js:71) in Screen (at StackViewCard.tsx:93) in Card (at createPointerEventsContainer.tsx:95) in Container (at StackViewLayout.tsx:985) in RCTView (at screens.native.js:101) in ScreenContainer (at StackViewLayout.tsx:394) in RCTView (at createAnimatedComponent.js:151) in AnimatedComponent (at StackViewLayout.tsx:384) in PanGestureHandler (at StackViewLayout.tsx:377) in StackViewLayout (at withOrientation.js:30) in withOrientation (at StackView.tsx:104) in RCTView (at Transitioner.tsx:267) in Transitioner (at StackView.tsx:41) in StackView (at createNavigator.js:80) in Navigator (at createKeyboardAwareNavigator.js:12) in KeyboardAwareNavigator (at createAppContainer.js:430) in NavigationContainer (at App.js:62) in RCTView (at Block.js:155) in Block (at App.js:61) in App (at withExpoRoot.js:26) in RootErrorBoundary (at withExpoRoot.js:25) in ExpoRoot (at renderApplication.js:40) in RCTView (at AppContainer.js:101) in RCTView (at AppContainer.js:119) in AppContainer (at renderApplication.js:39) ` Also this one: Element type is invalid: expected a string (for built-in components) or a class/function (for composite components) but got: undefined. You likely forgot to export your component from the file it's defined in, or you might have mixed up default and named imports. Check the render method of `Button`. @litehacker hey, do you version of Expo and React-Native you'r using? any word on this bug? I got the same bug "expo": "^36.0.0", "react": "16.9.0", @devbrunopaula I think this is because the current code base in based on an older expo version. I'll update the code base to the latest versions.
gharchive/issue
2019-12-21T09:00:58
2025-04-01T06:40:12.015331
{ "authors": [ "devbrunopaula", "hetmann", "litehacker" ], "repo": "react-ui-kit/dribbble2react", "url": "https://github.com/react-ui-kit/dribbble2react/issues/50", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1411397637
added logo component and make logo component spin Fixes Issue PR to fix issue #84 Changes proposed added a logo component in the component directory and make logo spin Requirements Create a component named Logo in the components folder that implements the screenshots below. It should accept a size as prop for the different sizes on the header and footer. The logo should rotate similar to how it does in a fresh create react app installation as shown on this GIF. Obtain the logo from the Figma design. Use THIS VIDEO as a guide Acceptance Criteria [x] The implementation should match the design. Screenshots Note to reviewers The Logo spin animation was performed using the tailwind css animation class. ok i will do that now
gharchive/pull-request
2022-10-17T11:21:42
2025-04-01T06:40:12.027727
{ "authors": [ "Ezekiel8807" ], "repo": "reactdeveloperske/reactdevske-website", "url": "https://github.com/reactdeveloperske/reactdevske-website/pull/89", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
267512645
Memory leak leads to heap overflow 14 Oct 15:06:57 - Received: :nova.esper.net PONG nova.esper.net :115367 <--- Last few GCs ---> [2539:0x4299a10] 1791041298 ms: Mark-sweep 1325.7 (1350.2) -> 1323.7 (1350.2) MB, 718.2 / 0.1 ms allocation failure GC in old space requested [2539:0x4299a10] 1791042276 ms: Mark-sweep 1323.7 (1350.2) -> 1323.6 (1350.2) MB, 977.4 / 0.0 ms allocation failure GC in old space requested [2539:0x4299a10] 1791043255 ms: Mark-sweep 1323.6 (1350.2) -> 1323.6 (1350.2) MB, 979.5 / 0.0 ms last resort [2539:0x4299a10] 1791044225 ms: Mark-sweep 1323.6 (1350.2) -> 1323.3 (1350.2) MB, 969.6 / 0.0 ms last resort <--- JS stacktrace ---> ==== JS stack trace ========================================= Security context: 0x2a90f6a29891 2: heartbeat [/home/sfnet-discord-bot/node-v8.1.2-linux-x64/lib/node_modules/discord-irc/node_modules/discord.js/src/client/websocket/WebSocketConnection.js:~407] [pc=0x1141bd0b5fab](this=0x3bc473b2d581 <an EventEmitter with map 0x263359c3e3d1>,time=0x222777302311 ) 3: arguments adaptor frame: 0->1 4: _onTimeout [/home/sfnet-discord-bot/node-v8.1.2-linux-x64/lib/node_modu... FATAL ERROR: CALL_AND_RETRY_LAST Allocation failed - JavaScript heap out of memory 1: node::Abort() [node] 2: 0x13647ec [node] 3: v8::Utils::ReportOOMFailure(char const*, bool) [node] 4: v8::internal::V8::FatalProcessOutOfMemory(char const*, bool) [node] 5: v8::internal::Factory::NewUninitializedFixedArray(int) [node] 6: 0xe90ca3 [node] 7: v8::internal::Runtime_GrowArrayElements(int, v8::internal::Object**, v8::internal::Isolate*) [node] 8: 0x1141bc30437d Aborted I don't understand much of what's written here. From the stack trace, it looks like… the very last… call… was from discord.js? I don't think that helps track the problem down, though. Do you have any more information on the problem? I've never noticed it, so I wonder if you're doing something weird in your setup (are you on a low-memory machine?). No, I was running into the Node v8 default heap limit. There's not enough info here, but you can probably find it fairly quickly if you try pulling in heapdump from npm and take some snapshots as the bot is running: https://blog.risingstack.com/finding-a-memory-leak-in-node-js/ I'll look into it a bit more at some point, but my fix for now was to just stick discord-irc into a systemd unit so it autorestarts when this happens. Have there been any updates on this? I'm having what appears to be a similar issue as seen in the following log: error Command failed with exit code 134. Aborted 10: 0x15a8c4bc3b67 9: 0x8e62c6 [/nodejs/bin/node] 8: node::StringBytes::Encode(v8::Isolate*, char const*, unsigned long, node::encoding, v8::Local<v8::Value>*) [/nodejs/bin/node] 7: v8::String::NewFromUtf8(v8::Isolate*, char const*, v8::NewStringType, int) [/nodejs/bin/node] 6: v8::internal::Factory::NewStringFromUtf8(v8::internal::Vector<char const>, v8::internal::PretenureFlag) [/nodejs/bin/node] 5: v8::internal::Factory::NewRawTwoByteString(int, v8::internal::PretenureFlag) [/nodejs/bin/node] 4: v8::internal::V8::FatalProcessOutOfMemory(char const*, bool) [/nodejs/bin/node] 3: v8::Utils::ReportOOMFailure(char const*, bool) [/nodejs/bin/node] 2: 0x8ccf9c [/nodejs/bin/node] 1: node::Abort() [/nodejs/bin/node] FATAL ERROR: CALL_AND_RETRY_LAST Allocation failed - JavaScript heap out of memory 3: unpack [/app/node_modules/discord.js/src/client/websocket/WebSocketConnection.js:~170] [pc=0x15a8c49070bd](this=0x29ae8a484649 <EventEmitter map =… 2: arguments adaptor frame: 0->3 1: toString [buffer.js:~609] [pc=0x15a8c495b041](this=0x253f2b47cc99 <Uint8Array map = 0x32dd30a43a71>,encoding=0x282dd32822d1 <undefined>,start=0x282dd32822d1 <undefined>,end=0x282dd32822d1 <undefined>) Security context: 0x2af5574a58b9 <JSObject> ==== JS stack trace ========================================= <— JS stacktrace —> [27:0x3001070] 70137123 ms: Mark-sweep 1292.3 (1470.9) -> 1292.3 (1470.4) MB, 2666.6 / 0.0 ms last resort GC in old space requested [27:0x3001070] 70134456 ms: Mark-sweep 1292.3 (1511.9) -> 1292.3 (1470.9) MB, 2829.5 / 0.0 ms last resort GC in old space requested [27:0x3001070] 70131626 ms: Mark-sweep 1292.4 (1509.9) -> 1292.3 (1511.9) MB, 2763.2 / 0.0 ms allocation failure GC in old space requested <— Last few GCs —> Looking through the discord.js issues related to memory leaks [0] it seems like they cache a lot of things without actually clearing the cache. It doesn't seem like this is something they intend to fix: https://github.com/discordjs/discord.js/issues/1409#issuecomment-433648959 It's on the other hand possible to disable some of these caches [1] so maybe that's what we want to do? [0] https://github.com/discordjs/discord.js/search?q=memory+leak&type=Issues [1] https://github.com/discordjs/discord.js/pull/2883 @ekmartin Have you had any luck with disabling the cache? It looks like the fork was abandoned due to unexpected behavior when disabling stores. https://github.com/discordjs/discord.js/pull/2883#issuecomment-433649130 Our Discord Bot's heap can grow to 2gigs in a couple of hours resulting in a crash. And we have under 50 Discord groups using the bot, so this type of rapid memory expansion seems odd, given there are some bots with pretty massive usage. Thanks for pointing me to those threads. 👍 Is this still an issue? I'm looking into updating discord.js, but I don't know if they've made any improvements here.
gharchive/issue
2017-10-23T00:34:04
2025-04-01T06:40:12.040803
{ "authors": [ "Throne3d", "ekmartin", "pkovac", "sethlesky" ], "repo": "reactiflux/discord-irc", "url": "https://github.com/reactiflux/discord-irc/issues/330", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
675634764
Upgrade discord.js to version 12 Also updates test stubs. WIP: [ ] Manually test most main cases (pings of users + roles, emoji use, multi-channeling, join messages) [x] Investigate hanging tests (looks to be due to Bot Events) Coverage increased (+0.03%) to 96.884% when pulling 6a986152eddf95f1c1d67ea67daf28e687fd3010 on upgrade/discordjs-12 into ee2d70fac12860b9a78043b7f121979929842493 on master.
gharchive/pull-request
2020-08-09T04:34:21
2025-04-01T06:40:12.043596
{ "authors": [ "Throne3d", "coveralls" ], "repo": "reactiflux/discord-irc", "url": "https://github.com/reactiflux/discord-irc/pull/555", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
666064080
Add Maintenance Page Feature Name Add Coming Soon and Maintenance Pages Summary description Provide two a pages for stores which are either (1) not yet launched or (2) undergoing maintenance. In both cases provide the user an experience that allows the user to take some action which continues to grow the business (e.g. a launch email sign-up form). Rationale for why this feature is necessary It's standard fare for e-commerce sites. The launch page is expecially helpful for allowing store operators users to log in and view the store in production but before it's launched. Expected use cases Prospective customers may sign-up for email notification/newsletter prior to launch. Store operators may log-in from Coming Soon if they've been given the credentials by the admin. Admins may demo a production-ready product to stakeholders and gather feedback. Shoppers receive a more professional experience when the site is down. Here's an example Coming Soon page I mocked up using Chakra UI: Thanks for taking the time to file this issue and preparing a mockup @balibebas! The example storefront, as the name says, is just an example implementation meant to be customized. I don't really consider such a screen a priority at all as it is easy to add one based on individual requirements, although I also wouldn't consider it to be an issue to showcase this functionality if someone wants to do a PR. The issue I see is, that this would either involve some more work or give a false sense of security. To offer a proper maintenance mode, there would me adjustments on the API to maybe store this on the shop document, some admin ui to toggle this and then it could be used during getStaticProps in the storefront to present such a maintenance screen. But the API would actually need to take that into consideration and deny acess to most of the queries apart from shop for non-authorized users. Also, as we use static generation were possible, still allowing storefront access to some authenticated users could be another challenge and probably be solved by using nextjs preview mode. Honestly I wouln't expect what I just described to be implemented anytime soon, although I guess such a maintenance mode would be a welcome addition if someone wants to provide PRs for that. Wonder what @mikemurray and @focusaurus think about this. A really quick and simple solution instead could be to just add some MAINTENANCE_MODE env variable and accompanying pages + redirects, if others agree that this provides any value to the way they work with the storefront. Similar to how the IOU payment example works for the API plugins I feel it would be good to keep a custom landing page of some sort present so this rather typical use case will always be considered as refactoring of the app and docs occur. Given not everyone needs a fancy landing page providing some sort of experience for API outages or errors would be nice. So having a maintenance page which appears when one clicks a product during an outage would be useful for most uses. Closing due to lack of activity
gharchive/issue
2020-07-27T07:39:08
2025-04-01T06:40:12.052440
{ "authors": [ "balibebas", "janus-reith" ], "repo": "reactioncommerce/example-storefront", "url": "https://github.com/reactioncommerce/example-storefront/issues/708", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
317004222
Add segment compatible analytics event tracking to the Product Grid We need to start tracking ecommerce analytics events in our starterkit. Initially we'll be tracking all of the Segment V2 Ecommerce Events. You can read the docs for that here: https://segment.com/docs/spec/ecommerce/v2/ Start by add documentation for the events tracked to our event tracking/analytics documentation, explaining the properties tracked and any mapping decisions that were made. For our current implementation of the Product Grid, we'll need to track two events: Product List Viewed Product Clicked @spencern The product price may be a range, should I take the min or max or neither? I'd take the min if a range exists maybe create a constant variable for this that we can permit customization of in the .env file later as well?
gharchive/issue
2018-04-23T22:22:25
2025-04-01T06:40:12.064025
{ "authors": [ "mikemurray", "spencern" ], "repo": "reactioncommerce/reaction-next-starterkit", "url": "https://github.com/reactioncommerce/reaction-next-starterkit/issues/56", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
128632318
Sync repo with latest from st2-chatops-aliases master This PR syncs a few stray commits from st2-chatops-aliases master branch committed there after creating this cloned repo. This is the first step in shuttering the st2-chatops-aliases forked community repo and using this repo exclusively for our Stackstorm setups. Boxed seems to have been using the use-existing-dyn-inventory branch on this repo. GMR has been using the st2-chatops-aliases repo master branch. With this sync up we can switch GMR over to this repo and delete the st2-chatops-aliases repo. We can decide how to handle Boxed's branch separately from this PR. fwiw, the actions in Boxed branch will probably be moved into the boxed-infrastructure repo. they are highly customized in a way I'm not sure we can abstract out. Looks good to me.
gharchive/pull-request
2016-01-25T20:35:52
2025-04-01T06:40:12.066341
{ "authors": [ "davekonopka", "jmound", "sairez" ], "repo": "reactiveops/st2-pack-omnia", "url": "https://github.com/reactiveops/st2-pack-omnia/pull/2", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
106141358
Refactor to ES6 Either simply write the entire source in ES6 and require node 4 in the package.json or implement transpilation with Bable. This has been done in my recent clean up in https://github.com/reactivepod/fido/pull/4 via require('babel/register');.
gharchive/issue
2015-09-12T08:33:49
2025-04-01T06:40:12.067683
{ "authors": [ "hglattergotz", "kahlil" ], "repo": "reactivepod/fido", "url": "https://github.com/reactivepod/fido/issues/1", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
252757039
fix: remove duplicate ComponentModelTypeConverter What kind of change does this PR introduce? (Bug fix, feature, docs update, ...) fix What is the current behavior? (You can also link to an open issue here) The ReactiveUI.Wpf project includes ComponentModelTypeConverter even though it's already in ReactiveUI. What is the new behavior (if this is a feature change)? Remove ComponentModelTypeConverter from ReactiveUI.WPF. Also, fix comments in the copy inside ReactiveUI because it actually has nothing to do with WPF. What might this PR break? Nothing realistic. Please check if the PR fulfills these requirements [ ] The commit follows our guidelines: https://github.com/reactiveui/reactiveui#contribute [ ] Tests for the changes have been added (for bug fixes / features) [ ] Docs have been added / updated (for bug fixes / features) Other information: @kentcb, Thanks for having already signed the Contribution License Agreement. Your agreement was validated by .NET Foundation. We will now review your pull request. Thanks, .NET Foundation Pull Request Bot Changes Unknown when pulling 88a668b9490a0b9214b726cc1dc5d17e7eda0df7 on kentcb:component-model-type-converter into ** on reactiveui:develop**.
gharchive/pull-request
2017-08-24T23:14:50
2025-04-01T06:40:12.076931
{ "authors": [ "coveralls", "dnfclas", "kentcb" ], "repo": "reactiveui/ReactiveUI", "url": "https://github.com/reactiveui/ReactiveUI/pull/1433", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
234117220
Farewell For all the details check out my Medium post. The tl;dr is basically that I’m overwhelmed and out of time so I’m turning everything over to @diasbruno as I step away from React Modal. Thanks for everything and goodbye! Thank you so much for your time managing react-modal. @claydiffrient Awesome work!
gharchive/issue
2017-06-07T07:31:29
2025-04-01T06:40:12.078875
{ "authors": [ "claydiffrient", "diasbruno", "jamesjjk" ], "repo": "reactjs/react-modal", "url": "https://github.com/reactjs/react-modal/issues/396", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
211387267
[chore] added missing babel transformer plugin. Running webpack --config webpack.config.js fails due to missing babel's spread transformer. ERROR in ./examples/basic/app.js Module build failed: SyntaxError: Unexpected token (16:20) 14 | 15 | openModal: function() { > 16 | this.setState({ ...this.state, modalIsOpen: true }); | ^ 17 | }, 18 | 19 | closeModal: function() { Acceptance Checklist: [x] All commits have been squashed to one. [x] The commit message follows the guidelines in CONTRIBUTING.md. [x] Documentation (README.md) and examples have been updated as needed. [x] If this is a code change, a spec testing the functionality has been added. [x] If the commit message has [changed] or [removed], there is an upgrade path above. This will no longer be necessary.
gharchive/pull-request
2017-03-02T13:13:21
2025-04-01T06:40:12.081759
{ "authors": [ "diasbruno" ], "repo": "reactjs/react-modal", "url": "https://github.com/reactjs/react-modal/pull/341", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1835022900
🛑 Bellshade API is down In 25eff64, Bellshade API (https://api.bellshade.org) was down: HTTP code: 0 Response time: 0 ms Resolved: Bellshade API is back up in f1b4c07.
gharchive/issue
2023-08-03T13:01:49
2025-04-01T06:40:12.084212
{ "authors": [ "reacto11mecha" ], "repo": "reacto11mecha/uptime-status", "url": "https://github.com/reacto11mecha/uptime-status/issues/486", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1587698785
[#2PlaysAMonth]: Image Gallery - Create a responsive image gallery by using the free Unsplash API First thing, PLEASE READ THIS: ReactPlay Code Review Checklist Description The project contains use of Unsplash API, which is fetched using axios and then populated in the Photo Component. Following things have been used in this project: useState useEffect React Form Component Refactoring Material UI icon Axios npm module CSS Grid Infinite scroll functionality(This feature is still not showing the react play app but is working on local file system. I am working on it) The app is responsive and redirects to the individual images and profiles when clicked on it. Fixes #910 Type of change [ ] New feature (non-breaking change which adds functionality) How Has This Been Tested? To see working of the app repeat the following Go to the app. Some images are loaded on the screen by default Search for any keyword. The screen will update with new images. Hover over any image, the cursor will turn to pointer. On clicking any image, the page will re direct to that image. On hovering, the profile name, likes will come in front of the image. On clicking the profile photo, the profile of the image contributor will open The page on scrolling down should show more images. Checklist: [ ] I have performed a self-review of my own code [ ] I have commented my code, particularly in hard-to-understand areas [ ] I have made corresponding changes to the documentation [ ] My changes generate no new warnings [ ] Any dependent changes have been merged and published in downstream modules Screenshots or example output @siddhantsiddh15 , Thanks for the PR. I would request you change the PR title to "[#2PlaysAMonth]: Image Gallery - Create a responsive image gallery by using the free Unsplash API" and the second thing is to link your PR with the issue by adding the below line in the PR description. Example: Fixes #910 Greetings, I have updated the pull request name and linked the issue with the issue number Thanks @siddhantsiddh15 , Please format and lint the code by following this guide. https://github.com/reactplay/react-play#format-the-code Greetings Have formatted the code as per the guidelines as mentioned here https://github.com/reactplay/react-play#format-the-code and then pushed the code in the branch. Thanks @siddhantsiddh15 Kindly resolve the merge conflict These errors are not in my edited files, what can I do to run the react play? @siddhantsiddh15 Catch me up on Discord today to close it. Link to the video as I was facing repeated issue in creating an account on Stack Stream. Thankyou for merging my branch into main. Can I close this pull request now? Inspect I have kicked off a build. Please check if it is successful and test if things are fine. @siddhantsiddh15 here is the preview build.. I see the changes are breaking styles. Please take a look https://react-play-git-fork-siddhantsiddh15-unsplash-8f3010-reactplayio.vercel.app/ Have recorded the video here . The website is responsive and is working correctly. @siddhantsiddh15 almost there.. please add a cover image Also edit your play from localhost and add the stream recording. Ping where when you done, will merge it. @siddhantsiddh15 let us know when done Have added the cover image. Have added the cover image. Thanks! The cover image should be in KBs, please reduce the size. Also confirm that your demo recording has been added to the play by editing it. Greetings, I have not added the demo recording in the Play. I have updated the size of the cover.png to 13 kb. Regards Greetings, I have not added the demo recording in the Play. I have updated the size of the cover.png to 135 kb. Regards Can you please add the demo recording too.. then its al done. Cannot do it. Having difficulty. Can we skip the recording portion? Cannot do it. Having difficulty. Can we skip the recording portion? Ok, no worries, no pressure. It's still valuable to get your work in. I am just curious about what kind of issues you are facing. If you can post about it in our Discord, I may try the resolution.. In fact, you can add the recording after merge too...(before 5th March) I am having slow internet connection due to the place I have travelled to, recording is a big file to upload. Hey @siddhantsiddh15 , this play looks cool. I will be waiting for the video link to updated before merging it to production branch Okay, will update it by tonight. Hey @siddhantsiddh15 , this play looks cool. I will be waiting for the video link to updated before merging it to production branch I have uploaded here the updated video of the play. I have checked the responsiveness and working of my play. The delay from my side was unwanted, I have uploaded the video as soon as I got good internet connection. Well the video should be on https://stack-stream.com/ I can get this play merged if everything is ok however please record a stackstream video before EOD @atapas you need to unblock in order to merge this PR @siddhantsiddh15 please record demo on stackstream
gharchive/pull-request
2023-02-16T13:40:19
2025-04-01T06:40:12.162112
{ "authors": [ "Sachin-chaurasiya", "atapas", "koustov", "siddhantsiddh15" ], "repo": "reactplay/react-play", "url": "https://github.com/reactplay/react-play/pull/974", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
209195477
Reactstrap and React-router 4.0.0-beta.6 - active Issue description components: Navlink Steps to reproduce issue I'm using Reactstrap and React-router 4.0.0-beta.6 in the educational project that is located on gitlab with custom domain - http://harazd.org/. According to Reactstrap docs: that's the way I should use active navlink import { NavLink } from 'reactstrap' ... <NavLink href="#" active = true >Link< /NavLink> According to React-router v4 docs: import { NavLink } from 'react-router-dom' ... <NavLink to="/about" activeClassName="active">About</NavLink> So how should I do implement navlink active state and use react-router? The simplest answer is that you can't use reactstrap navlink active prop when passing in react-router navlink. And that's fine because it's just 1 class active and react-router NavLink can take care of that with the activeClassName prop. To use both, you'll need to rename one of those and use the tag prop in reactstrap NavLink. import { NavLink } from 'reactstrap'; import { NavLink as RRNavLink } from 'react-router-dom'; <NavLink to="/about" activeClassName="active" tag={RRNavLink}>About</NavLink> Thanks for a quick solution (You can paste your answer here: http://stackoverflow.com/questions/42372179/reactstrap-and-react-router-4-0-0-beta-6-active-navlink). And one more thing : ... - root path is always active for some reason.. The root path is always active because '/' is in every path. I am not too fimilar with react-router-dom and it's NavLink, it looks like it has an exact prop which may be what you are looking for. Here's my router code, so I do use exact for root path. ` <BrowserRouter> <div className='app'> <Switch> <Route exact path='/' component={Landing} /> <Route path='/products' component={Products} /> <Route path='/services' component={Services} /> <Route path='/price' component={Price} /> <Route path='/contacts' component={Contacts} /> <Route component={NoMatch} /> </Switch> </div> </BrowserRouter>` @TheSharpieOne yes adding exact to root Navlink helped) Previously you were using <NavLink> from react-router-dom. From that file, it looks like it creates the Route and the Link for you in one go. react-router v4 has a ton of changes and I am not familiar with it enough to know how to use it compared to previous version so I can only offer some limited help with it. @Y-Taras @TheSharpieOne can you give me example code? a have try it but doesnt work so far. @dhanyn10 here's a link to my working example https://gitlab.com/ytaras/pinobeton2/blob/master/js/Navigation.js Hi everyone, Is there anyway to use NavLink from reactstrap alone to style the active link? Thanks @Y-Taras I can confirm your solution works here too. Thanks so much - you saved me many hours of pain. @piavgh if you just want to use reactstrap's NavLink alone (without react-router-dom) you can use the active prop: import { NavLink } from 'reactstrap'; <NavLink to="/about" active>About</NavLink> I think you can add something like : <NavLink to="/about" active={window.location.hash === '/about'}>About</NavLink> This works for me: <NavItem active={window.location.pathname === "/thing"}> <NavLink href="/thing">Things</NavLink> </NavItem>
gharchive/issue
2017-02-21T16:19:21
2025-04-01T06:40:12.173408
{ "authors": [ "JonSilver", "TheSharpieOne", "Y-Taras", "dhanyn10", "eddywashere", "nathfy", "piavgh", "softmixt" ], "repo": "reactstrap/reactstrap", "url": "https://github.com/reactstrap/reactstrap/issues/336", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
616470240
Proposal for environment context This is a proposal for a new environment context which takes precedence of the config file. We had the need for a context which can be configured for automation without touching files. Happy to hear your feedback. Hmm instead of going through all that trouble of creating and mapping config variables, you could also create an environment variable ESQUE_CONFIG_YAML that can hold the whole esque config as yaml string. That would be more dynamic, future proof and would only require minimal changes. Hm. I get your point, in general. But that would only get rid of the need to write a file, not to create a YAML in general. :D You wrote: without touching files. So I thought, that was your problem :smile: Well, although I don't see how setting 15 environment variables is easier than creating a yaml string, I'm always a fan of "letting the user choose". So I'm not entirely against it. I'm just afraid that we might have to rename, add and/or remove some of the variables while our config evolves. We do have a migration mechanism for file based configs but not for environment variables. Do you think a separate command to add a section to the config would help you? Something like esque config add-context foo --bootstrap-servers broker1,broker2 --schema-registry registry ... I think it's easier to keep a command stable than it is to keep the environment variables the same over time. Guess I should have been clearer. Sorry for that. :D I think you're right with your approach of not wanting to make it unnecessarily complicated and I think the idea of just adding another command is good. I'll see with what I can come up with over the weekend. 😊
gharchive/pull-request
2020-05-12T08:37:43
2025-04-01T06:40:12.200888
{ "authors": [ "hfjn", "swenzel" ], "repo": "real-digital/esque", "url": "https://github.com/real-digital/esque/pull/148", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
98426054
C API For easiest integration with Go (via cgo), Erlang, and Rust (via FFI), a C API (not C++) is needed. This is a simple wrapper around C++. .NET users will also be very happy to use Aeron via P/Invoke! Do you have any ETA? A C API will instantly make Aeron as ubiquitous as ZMQ, native ports like #35 are not needed as much as a simple API. This is still a plan. However, have not had time to get to it. Any Update over the C API Currently, a C API is planned for the media driver. And would like to extend that API to cover the Aeron client API. However, no timeline is set for that yet. Closing until someone is willing to sponsor the work. Now the C API are able to be happen:) Preview in 1.28.X releases. Should be feature complete from 1.30.0.
gharchive/issue
2015-07-31T16:34:09
2025-04-01T06:40:12.204299
{ "authors": [ "Hemanth-Eduru", "buybackoff", "lygstate", "mjpt777", "tmontgomery" ], "repo": "real-logic/Aeron", "url": "https://github.com/real-logic/Aeron/issues/154", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
68196476
Constrain Mailing list form in Sidebar Is it possible to make the mailing list form only 300px wide, as the ads are? There's a custom CSS field in the plugin settings. I entered .et_bloom { width: 300px; } but no result Where did you enter that CSS? It's not a good idea to edit the stylesheets directly, as they are replaced with each new build of the theme. Changes that are not part of that build would be overwritten. Do you need provisions for manually inserting css? There's a CSS field within the plugin settings. In other words, when designing the look of the plugin, there are a host of options for customizing the colors and text fields. Then there is a "Custom CSS" field. On Fri, Apr 17, 2015 at 3:29 PM, Dylan Kinnett notifications@github.com wrote: Where did you enter that CSS? It's not a good idea to edit the stylesheets directly, as they are replaced with each new build of the theme. Changes that are not part of that build would be overwritten. Do you need provisions for manually inserting css? — Reply to this email directly or view it on GitHub https://github.com/real-pants/Real-Pants/issues/81#issuecomment-94058258 . Can this be closed? Dimensions look great now.
gharchive/issue
2015-04-13T21:17:02
2025-04-01T06:40:12.213406
{ "authors": [ "dylan-k", "realpants" ], "repo": "real-pants/Real-Pants", "url": "https://github.com/real-pants/Real-Pants/issues/81", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2467695136
Crash on trying to run file on sparse3 Crash encountered when running file on sparse3 from ufs-big test image Backtrace with RUST_LOG=trace and RUST_BACKTRACE=full: https://gist.github.com/casept/4f9f18b75458f4801746d3597e29c2d1 What system were you running it on? Can you please show us uname -a? Fedora 40, uname: Linux l13 6.10.3-200.fc40.x86_64 #1 SMP PREEMPT_DYNAMIC Mon Aug 5 14:30:00 UTC 2024 x86_64 GNU/Linux This is now fixed with #63.
gharchive/issue
2024-08-15T09:05:42
2025-04-01T06:40:12.232282
{ "authors": [ "asomers", "casept", "realchonk" ], "repo": "realchonk/fuse-ufs", "url": "https://github.com/realchonk/fuse-ufs/issues/54", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
2113528855
Fix Unit Tests and Remove Usage of Unsigned Byte Remove UByte Array usages since it was experimental Remove Roboelectric Update Unit Tests with Android mocking Base64 Tests successfully run in the terminal Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
gharchive/pull-request
2024-02-01T21:36:28
2025-04-01T06:40:12.243363
{ "authors": [ "CLAassistant", "realeyes-mike-patterson" ], "repo": "realeyes-media/scte35-decoder-multiplatform", "url": "https://github.com/realeyes-media/scte35-decoder-multiplatform/pull/10", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
242872720
Clarify error message I just wanted to clarify the error message that shows up when the library can't find the .so/.dll The previous error message implied that the error had something to do with the vulkan version (false, as far as I can tell) and also didn't really indicate that the library just may not have been in the loading path. It's a really small thing, but I thought I'd submit a PR anyways Also, random question. Why is there so much code duplication between vulkan.template.py and init.py? Hello @pipsqueaker. Thank you very much for your pull request. You are right about the error message, your version is a lot better. Nevertheless, you don't need to update the __init__.py file. This file is automatically generated from the vulkan.template.py which is a jinja2 template. When you run the generator script, it will do it for you. So what you can do is only update the template file and then I will regenerate the module. Thanks a lot for the contribution. @realitix Alright, just messed with my history a bit so that only vulkan.template.py is edited Thanks @pipsqueaker for the contribution !
gharchive/pull-request
2017-07-14T01:25:17
2025-04-01T06:40:12.249121
{ "authors": [ "pipsqueaker", "realitix" ], "repo": "realitix/vulkan", "url": "https://github.com/realitix/vulkan/pull/12", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
331549234
Could you publish it in Scala 2.12 ? jwt-scala_2.11 and jwt-scala_2.10 are already published, but we want to use your library in 2.12. Could you publish this library in Scala 2.12 ? forked and published for Scala 2.12 and 2.13 https://github.com/xuwei-k/jwt-scala/tree/v1.4.0 https://repo1.maven.org/maven2/com/github/xuwei-k/jwt-scala_2.12/ https://repo1.maven.org/maven2/com/github/xuwei-k/jwt-scala_2.13/
gharchive/issue
2018-06-12T11:34:22
2025-04-01T06:40:12.256024
{ "authors": [ "3tty0n", "xuwei-k" ], "repo": "reallylabs/jwt-scala", "url": "https://github.com/reallylabs/jwt-scala/issues/16", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
241852671
Change object store to allow sync user auth URL to be modified With this PR, getting a user with a different auth URL than the one it was originally given updates the URL, instead of throwing an exception. This fixes the issue (commonly encountered during development) when a user is opened with one URL, persisted, the server is moved or SSL is enabled, and an attempt to log in the user with the new address is made. This does not fix the corner case where two different ROSes have users with the same user ID, and the user wishes to be logged into both simultaneously, but this case wasn't supported before, nor is it supported by any of the other subsystem code. Did we get clarity on whether identity is expected to be unique across different servers? If that's not guaranteed, is this change safe? It's complicated. The default implementation of identity is a UUID. However, there's the possibility of plugging a module into ROS that allows you to specify arbitrary identities instead of using the built-in system, so a conflict is theoretically possible. Even if it is, though, using Realms across multiple servers concurrently is something we intentionally de-emphasized when designing v1 of ROS, so I don't think it's a supported use case anyways. It's something that Realm Browser inherently needs to support. Back to the drawing board, then.
gharchive/pull-request
2017-07-10T21:16:16
2025-04-01T06:40:12.332956
{ "authors": [ "austinzheng", "bdash" ], "repo": "realm/realm-object-store", "url": "https://github.com/realm/realm-object-store/pull/494", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
501618519
Switch to using ocaml-ci-scripts This seems like a more reliable CI until we can use ocaml-ci. This will allow us to easily add revdeps builds btw which could prove useful once we want to stabilize mdx's API and avoid breaking users' tests! Opam has a make built-in variable that opam lint suggests we use instead of the raw "make". Now that we use the ci-scripts, it's properly picked up and triggers a build failure if we don't have a compliant opam file. Ok I've read https://github.com/realworldocaml/mdx/pull/185#discussion_r330673627 which answers my question. However this is quite fragile as it will break if you run this in a duniverse setting (with a toplevel call to dune runtest). What do you mean? Do you mean that it will break if we vendor mdx in a duniverse? Aliases aren't resolved within the duniverse so it won't run the tests in this case. I agree it is a bit fragile and I wish there was a way to tell dune about the ocaml-mdx -> ocaml-mdx-test binary dependency but unfortunately there isn't any atm. I tried a couple things and they ended up suffering from the same race condition. Also when using mdx in duniverse mode, ocaml-mdx rule adds a (package mdx) dependency to the generated rules, thus solving the dependecny issue. We can't do that outside a duniverse because (package ...) deps only work for local packages, not opam ones. There's one last thing I can try which is to add an explicit dependency on the install alias for all runtest aliases that rely on ocaml-mdx but that is a bit tedious and doesn't solve the issue in a generic way either but at least running dune runtest should work then. Let's merge that as it already fix the opam test runs and I'll try that solution in a separate PR! Just had a quick look and the (alias install) dependency propably won't work as we generate most of the rules for runtest. I just mean that we should report that issue upstream as it should be fixed properly at one point :-)
gharchive/pull-request
2019-10-02T17:14:40
2025-04-01T06:40:12.360460
{ "authors": [ "NathanReb", "samoht" ], "repo": "realworldocaml/mdx", "url": "https://github.com/realworldocaml/mdx/pull/187", "license": "ISC", "license_type": "permissive", "license_source": "github-api" }
382682448
cli: reana-dev git-log New command git-log showing information about commits. Signed-off-by: Tibor Simko tibor.simko@cern.ch Example output: $ reana-dev git-log [reana-workflow-engine-yadage] git log -n 5 --graph --decorate ... * (upstream/pr/91) 34b1c8e tasks: add stop_workflow from reana_commons, Diego Rodriguez, 6 hours ago * (HEAD -> master, tag: v0.4.0, upstream/pr/90, upstream/master, origin/master, origin/HEAD) 11305f8 release: v0.4.0, Dinos Kousidis, 2 weeks ago * (upstream/pr/88) 720819b installation: upgrade REANA-Commons, Diego Rodriguez, 5 weeks ago * 8f13434 publisher: update to kombu producer, Diego Rodriguez, 5 weeks ago * (upstream/pr/89) 32251ab installation: bump reana-commons include pkg data, Diego Rodriguez, 6 weeks ago [reana-workflow-engine-serial] git log -n 5 --graph --decorate ... * (upstream/pr/52) e77af4e tasks: use stop_workflow from reana-commons, Diego Rodriguez, 5 hours ago * ada286d tasks: introduce a stop_workflow task, Diego Rodriguez, 21 hours ago * 522d3f1 config: disable task prefetching by process, Diego Rodriguez, 27 hours ago * 0e114d1 config: disable broker pool, Diego Rodriguez, 28 hours ago * 08f28d2 tasks: add revoke handler, Diego Rodriguez, 7 days ago [reana-job-controller] git log -n 5 --graph --decorate ... * (upstream/pr/98) d660142 api: return a dict on k8s_instantiate_job, Diego Rodriguez, 21 hours ago * e4a18d1 global: use flask run to start application, Diego Rodriguez, 5 days ago * c6b1578 tests: make application testable, Jan Okraska, 8 days ago * 82fc7bc api: add delete job endpoint, Diego Rodriguez, 8 days ago * (HEAD -> master, tag: v0.4.0, upstream/pr/97, upstream/master, origin/master, origin/HEAD) 527cc1b release: v0.4.0, Tibor Simko, 2 weeks ago [pytest-reana] git log -n 5 --graph --decorate ... * (upstream/pr/31) ff2f2df fixtures: renaming of operational parameters, Dinos Kousidis, 23 hours ago * (HEAD -> master, upstream/master, origin/master, origin/HEAD) 4ceb1e7 release: v0.5.0.dev201811191, Diego Rodriguez, 24 hours ago | * (upstream/pr/30) 70b1f6e release: v0.5.0.dev20181119.1, Diego Rodriguez, 24 hours ago |/ * d0d1920 fixtures: fix expose yadage workflow fixture, Diego Rodriguez, 24 hours ago * (upstream/pr/29) ce31f8a release: v0.5.0.dev20181119, Dinos Kousidis, 27 hours ago [reana-workflow-monitor] git log -n 5 --graph --decorate ... * (HEAD -> master, tag: v0.4.0, upstream/pr/26, upstream/master, origin/master, origin/HEAD) c242d4f release: v0.4.0, Dinos Kousidis, 2 weeks ago * (upstream/pr/25) b8d4d1c global: license change to MIT License, Tibor Simko, 6 weeks ago * (upstream/pr/24) 924ef8d docs: new logo, panel verbiage and links, Tibor Simko, 5 months ago * (upstream/pr/23) c884a9b docs: author ORCID links, Tibor Simko, 7 months ago * (tag: v0.2.0, upstream/pr/22) 8f5b219 release: v0.2.0, Dinos Kousidis, 7 months ago [reana-server] git log -n 5 --graph --decorate ... * (HEAD -> master, upstream/pr/109, upstream/master, origin/master, origin/HEAD) 698a34e installation: fix pytest-reana dependency version, Tibor Simko, 3 hours ago * (upstream/pr/104) a2cb517 api: new rest api endpoint which returns wf params, Rokas Maciulaitis, 24 hours ago | * (upstream/pr/108) a92b154 api: automatic openapi specs passing to reana-commons, Rokas Maciulaitis, 4 days ago |/ | * (upstream/pr/107) cd155f9 api: automatic openapi specs passing to reana-commons, Rokas Maciulaitis, 4 days ago | * 068441c api: new rest api endpoint which returns wf params, Rokas Maciulaitis, 5 days ago |/ [reana-message-broker] git log -n 5 --graph --decorate ... * (HEAD -> master, tag: v0.4.0, upstream/pr/16, upstream/master, origin/master, origin/HEAD) 7f18a21 release: v0.4.0, Tibor Simko, 2 weeks ago * (upstream/pr/15, upstream/license-change) 1da1398 global: license change to MIT License, Tibor Simko, 6 weeks ago * (upstream/pr/14, upstream/docs-logo-panel-links) 9d0bee3 docs: new logo, panel verbiage and links, Tibor Simko, 5 months ago * (upstream/pr/13, upstream/docs-authors-orcid) 6661833 docs: author ORCID links, Tibor Simko, 7 months ago * (tag: v0.2.0, upstream/pr/12) 1f7b82d release: v0.2.0, Tibor Simko, 7 months ago [reana-workflow-controller] git log -n 5 --graph --decorate ... * (upstream/pr/141) c1f00aa tests: utilities, Dinos Kousidis, 3 hours ago * efca3e4 rest: set_workflow_status parameters description, Dinos Kousidis, 4 hours ago * 120db5c rest: allow deletion of already deleted workflows, Dinos Kousidis, 4 hours ago * a9721db tests: workspace deletion, Dinos Kousidis, 5 hours ago * 74f9e53 rest: allow access to deleted workflows, Dinos Kousidis, 5 hours ago [reana-workflow-engine-cwl] git log -n 5 --graph --decorate ... * (upstream/pr/70) c92165a tasks: add stop_workflow task from reana_commons, Diego Rodriguez, 6 hours ago * (HEAD -> master, tag: v0.4.0, upstream/pr/68, upstream/master, origin/master, origin/HEAD) 7b4b7d5 release: v0.4.0, Dinos Kousidis, 2 weeks ago * (upstream/pr/66) e6cc7f9 installation: upgrade REANA-Commons, Diego Rodriguez, 5 weeks ago * a5bb2ee publisher: use Kombu publisher, Diego Rodriguez, 5 weeks ago * (upstream/pr/67) 696d526 installation: bump reana-commons version, Diego Rodriguez, 6 weeks ago [reana-commons] git log -n 5 --graph --decorate ... * (upstream/pr/66) 1c6ccb5 tasks: introduce common task to stop workflows, Diego Rodriguez, 5 hours ago * cd856de api: update openapi specs, Jan Okraska, 20 hours ago | * (HEAD -> master, upstream/pr/65, upstream/master, origin/master, origin/HEAD) e720eaa release: 0.5.0.dev20181116, Dinos Kousidis, 26 hours ago | * 6cdb587 installation: bump pytest-reana, Dinos Kousidis, 27 hours ago | * (upstream/pr/63) 9746ec4 api: new rest api endpoint, Rokas Maciulaitis, 4 days ago |/ [reana-db] git log -n 5 --graph --decorate ... * (HEAD -> master, upstream/pr/21, upstream/master, origin/master, origin/HEAD, installation-version) 0b0854b release: v0.5.0.dev20181116, Dinos Kousidis, 4 days ago * (upstream/pr/19) 4a5ebdf models: renaming operational parameters, Rokas Maciulaitis, 4 days ago | * (upstream/pr/20) 7f3621e models: addition of stopped WorkflowStatus, Diego Rodriguez, 6 days ago |/ * (upstream/pr/15) 3e7a812 models: addition of deleted WorkflowStatus, Dinos Kousidis, 12 days ago * (tag: v0.4.0, upstream/pr/17) e2bb3e6 release: v0.4.0, Tibor Simko, 2 weeks ago
gharchive/pull-request
2018-11-20T14:15:09
2025-04-01T06:40:12.388300
{ "authors": [ "tiborsimko" ], "repo": "reanahub/reana", "url": "https://github.com/reanahub/reana/pull/115", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
505310723
Playlist management: errors in cropping, appending or pasting I have issues with the playlist manager Once a playlist is made, it execute it finely But once I chose for Crop or Append or Paste, the final result is totally incorrect. Items are missing, or partially copied and pasted. I attached a .zip file containing a video that shows you the details. I would like to know if this is an issue or it's me badly operating, and what do you suggest for the best. Reaper-Playlist-Issue.zip Interesting, can you share the project RPP? Are any media items grouped? Interesting, can you share that TEST.RPP project? Are any media items grouped? Sorry for the late reply About the media items: not any group Yes sure, the TEST.RPP is here in attachment in the ZIP file TEST.zip I cannot reproduce the bug with SWS v2.10.0. "Paste playlist at edit cursor" behaves as expected here with that project. There is a possibility some REAPER setting is interfering. Can you share your reaper.ini as well? I can reproduce if I have Preferences (-> Project) -> Media Items Defaults: Overlap and crossfade items when splitting (length is set to 0:00.010 here). enabled when pasting here. I have that pref assigned to the toolbar button shown in the gif. https://i.imgur.com/PeL633Z.gif Fixed in this build: sws-2.11.0.0-Windows-x64-e12769f5.exe. Yep it works perfectly. Thanks a million! P.S. testing it closely I did notice that when you Crop project on new tab, the zones are replicated but not into the new position but they are referred to the original project. Thus it will be necessary to manually move or recreate the zones. Did you get the same result? "Crop project to playlist" and "Crop project to playlist (new project tab)" should produce the same output.
gharchive/issue
2019-10-10T14:30:14
2025-04-01T06:40:12.395369
{ "authors": [ "cfillion", "nofishonfriday", "tormyvancool" ], "repo": "reaper-oss/sws", "url": "https://github.com/reaper-oss/sws/issues/1204", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1262268220
For Code Review Check on Javascript best practices Done .
gharchive/pull-request
2022-06-06T19:14:49
2025-04-01T06:40:12.416845
{ "authors": [ "rebel216" ], "repo": "rebel216/To-Do-List-PR", "url": "https://github.com/rebel216/To-Do-List-PR/pull/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
107601150
Calling treetagger with untokenized test? Hi, the example on the tt4j homepage shows tokenized input being handled by treetagger. How to I give a untokenized text to treetagger to both POS tag and tokenize? Thanks Shyam See https://reckart.github.io/tt4j/tokenizer.html
gharchive/issue
2015-09-21T21:22:54
2025-04-01T06:40:12.449880
{ "authors": [ "reckart", "shyamupa" ], "repo": "reckart/tt4j", "url": "https://github.com/reckart/tt4j/issues/23", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2458966962
level 3 pull request not merged yet @sanjay-kv please review and merge this pull request https://github.com/Rakesh9100/CalcDiverse/pull/1820 This is the pull request i had created @sanjay-kv I resolved the conflicts please review and merge when you are free PR is getting reviewed one by one, you can check the PR before accepting the points request from the contributors @sanjay-kv You can remove the point label from here
gharchive/pull-request
2024-08-10T07:48:52
2025-04-01T06:40:12.457934
{ "authors": [ "Rakesh9100", "aditya-bhaumik" ], "repo": "recodehive/resume-pitch", "url": "https://github.com/recodehive/resume-pitch/pull/49", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2643658209
Simplified arrow middleware definition Issue In https://github.com/oleksandr-danylchenko/text-annotator-js/commit/ce7014d6f74d929488cb0c277f7015a5abc90b0e#r148871559 I spotted that the arrow middleware definition can be simplified to its default form. That should be safe for the floating itself because the arrow is a "Data Middleware" that only populates the context with the positioning props, but doesn't change the behavior. Demo https://github.com/user-attachments/assets/ab1dab97-7e4e-41e5-8737-66a55e218372 Tested the same thing in Annotorious - yep, works :-) Thanks!
gharchive/pull-request
2024-11-08T10:49:31
2025-04-01T06:40:12.460441
{ "authors": [ "oleksandr-danylchenko", "rsimon" ], "repo": "recogito/text-annotator-js", "url": "https://github.com/recogito/text-annotator-js/pull/177", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1180344326
Utilize GPT3 embedding / classification API for more automated bulk labelling Keyword-based bulk labeling using Rubrix Rules is still too slow if I were to label thousands of texts for classification - at maximum, I can label 20 at a time since keyword match is noisy with regard to ground-truth label. I have played with GPT3 embeddings, where applying UMAP on texts results in clear clusters with semantically similar texts in clusters. If one can create Rubrix Rules instead based on a selected area on a scatterplot of text embeddings, then one can label hundreds of texts at once, and can easily finish labeling 50,000 data points in a day with high quality. And that'd be a game-changer. Adding on to that, I also played with GPT3 classification API, where it gives pretty accurate few-shot classifications. And if that is added to the pipeline, it'd expedite labeling even further. Making it possible to label thousands of texts at once. I have the GPT3 embedding + classification pipeline done in a notebook, is there plan on Rubrix's side to look into utilizing GPT3? Hey @whaowhao Thank you for bringing this up! We are working on a tutorial in which we show how you can use Epoxy to achieve the same goal you are mentioning. The workflow would be something like this: Come up with a few rules trying to cover semantic diverse records Provide the weak label matrix and your embeddings of choice to Epoxy (with some thresholds) Get back an enhanced weak label matrix Hopefully, we can share it with you soon, we would love to get your feedback. Thanks @whaowhao ! Adding to what @dcfidalgo mentions, we've also been discussing the ability of "labelling-by-drawing" based on a 2-D embedding-based representation (UMAP or otherwise). This will be in the roadmap but not immediately. Meanwhile I'd be really interested on collaboration/supporting you with experiments with GPT3, would you be interested? Hi @dvsrepo yeah let's chat
gharchive/issue
2022-03-25T04:54:36
2025-04-01T06:40:12.464970
{ "authors": [ "dcfidalgo", "dvsrepo", "whaowhao" ], "repo": "recognai/rubrix", "url": "https://github.com/recognai/rubrix/issues/1309", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
301786422
valuesAreSet field always is false. Hi, I'm trying to create a RecommendItemsToUser request and set the returnProperties to true, but the valuesAreSet field in the recommendation response always comes as false. When i try to execute recommendation.getValues() it always throws an IllegalStateException. I checked the response body from the Api call and it does not contain the valuesAreSet property in the payload. Recombee Api Response Payload: {"recommId": "8b904849-19ad-47d9-90f2-b024ca67726f", "recomms": [{"values": {"manufacturer_code": 1}, "id": "123"}]} Request: final RecommendItemsToUser cf = new RecommendItemsToUser("CF", 5) .setReturnProperties(true); I confirm the bug, we will fix it ASAP Thanks @OndraFiedler It is fixed in https://github.com/recombee/java-api-client/releases/tag/v2.0.1. The new version has been pushed to the central repository, but it can take some time (up to few hours) before it is available. Thanks for reporting!
gharchive/issue
2018-03-02T14:36:42
2025-04-01T06:40:12.468701
{ "authors": [ "OndraFiedler", "alepinzon" ], "repo": "recombee/java-api-client", "url": "https://github.com/recombee/java-api-client/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1412090532
docker-compose not recognising boolean values When running docker-compose up -d the following error is observed: ERROR: The Compose file './docker-compose.yml' is invalid because: services.keycloak.environment.KC_HTTP_ENABLED contains false, which is an invalid type, it should be a string, number, or a null Remediation is by encapsulating any boolean values in single quotes so they are treated as a string as noted here Behaviour seen in commit 57d1665d323048c9feaf34a9fabe6cc454ec4ccc Running Docker version 20.10.19, build d85ef84 on Ubuntu 22.04.1 Just saw that a similar pull request has already been responded to
gharchive/pull-request
2022-10-17T19:21:31
2025-04-01T06:40:12.490828
{ "authors": [ "chssn" ], "repo": "reconmap/reconmap", "url": "https://github.com/reconmap/reconmap/pull/123", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
524424299
[references] Tweets with ❤️️ https://twitter.com/davert/status/1243285793709457411 https://github.com/shopsys/shopsys/pull/1392#issuecomment-555106644 https://twitter.com/afilina/status/1229406983486394369 https://twitter.com/davert/status/1243285793709457411 https://twitter.com/phpinsights/status/1248596452730953732 https://twitter.com/MrDanack/status/1248607716018282497 https://twitter.com/VotrubaT/status/1253631226163990529 https://twitter.com/rectorphp/status/1253432659075710977 https://twitter.com/gaborhojtsy/status/1256246616710885381 https://twitter.com/VotrubaT/status/1259164120148054016 https://twitter.com/utopicabsurd/status/1271075970712326144 https://twitter.com/LBajsarowicz/status/1272947900016967683 - https://github.com/magento/magento2/pull/27946/files https://twitter.com/LBajsarowicz/status/1273024967349866497 https://twitter.com/jrrtgimli/status/1275332001566986241 https://twitter.com/VotrubaT/status/1275863668878688257 https://twitter.com/lulco/status/1342198519009046528 https://twitter.com/dantleech/status/1342213179334676490 https://tweetdeck.twitter.com/ https://twitter.com/VotrubaT/status/1343551843759255552 https://twitter.com/rectorphp/status/1352758017410936833 https://twitter.com/rectorphp/status/1352758017410936833 https://twitter.com/Ruud_/status/1356256886417461249 https://twitter.com/tomasnorre/status/1359934323206615040 https://www.reddit.com/r/PHP/comments/lw5kvz/code_analysis_tools_to_help_with_upgrading_an_app/gpfhde7/?utm_source=reddit&utm_medium=web2x&context=3 https://www.reddit.com/r/PHP/comments/lw5kvz/code_analysis_tools_to_help_with_upgrading_an_app/gpfem49/?utm_source=reddit&utm_medium=web2x&context=3 https://github.com/sabbelasichon/typo3-rector/issues/2094#issuecomment-805320205 https://twitter.com/vojtabiberle/status/1379127544067321872 https://twitter.com/el_stoffel/status/1384865872959975425 https://twitter.com/KPikaza/status/1386791253275299842 https://twitter.com/rsinnbeck/status/1394188533888339969 https://twitter.com/althaus_it/status/1394235695611396097 https://www.facebook.com/christophe.avonture/posts/10225263983294673?notif_id=1624954306250211&notif_t=page_tag&ref=notif https://twitter.com/sarfraznawaz/status/1437335510213615617 https://twitter.com/Chemaclass/status/1459616944127201285 https://twitter.com/alex_s_/status/1464358800006299650 https://twitter.com/Guikingone/status/1465944064126300161 https://twitter.com/srogier/status/1473289580564647937 https://twitter.com/unagiunag/status/1478941430085459968 https://twitter.com/amakdessi/status/1479110174589603842 https://twitter.com/unagiunag/status/1494832218614935552 https://twitter.com/_stefanzweifel/status/1521922469111771136 https://twitter.com/Chemaclass/status/1531622088276156418 https://twitter.com/VotrubaT/status/1537142504839159808 https://twitter.com/Guikingone/status/1543927696274690049 https://twitter.com/tmotyl/status/1557397895388725249 https://twitter.com/brendt_gd/status/1559829464359600128 https://www.reddit.com/r/PHP/comments/wqnqs9/rector_reaches_10_million_downloads/ https://github.com/rectorphp/rector-src/pull/2778#issuecomment-1218534498 https://twitter.com/VotrubaT/status/1595734036487540736 https://twitter.com/tenmajkl/status/1596612085810876416 https://mobile.twitter.com/markusstaab/status/1592572730934980608 https://twitter.com/JustSteveKing/status/1666745322310172673 https://twitter.com/jan_linhart/status/1737386735430951208 https://twitter.com/aminetiyal/status/1746910353714204952 https://twitter.com/enunomaduro/status/1754929560112472521 https://twitter.com/ramageek/status/1754894293016391935 https://twitter.com/SlyFireFox/status/1754871230228758735 https://twitter.com/nuernberger_me/status/1754813770365337969 https://twitter.com/PatricioOnCode/status/1754860469171462399 https://github.com/rectorphp/rector/discussions/8848
gharchive/issue
2019-11-18T15:03:25
2025-04-01T06:40:12.517398
{ "authors": [ "TomasVotruba" ], "repo": "rectorphp/getrector-com", "url": "https://github.com/rectorphp/getrector-com/issues/70", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2125473133
[backport]: from 2.6 to 2.7 (#173) fix(trustyai): prometheus rules for probe update(trusty): prometheus to use job instead of instance name for record rules this is missed to get into main before 2.7 branch out
gharchive/pull-request
2024-02-08T15:36:04
2025-04-01T06:40:12.547395
{ "authors": [ "zdtsw" ], "repo": "red-hat-data-services/rhods-operator", "url": "https://github.com/red-hat-data-services/rhods-operator/pull/190", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2371990998
Add support for MCG in client console /lgtm
gharchive/pull-request
2024-06-25T08:08:01
2025-04-01T06:40:12.562036
{ "authors": [ "SanjalKatiyar", "bipuladh" ], "repo": "red-hat-storage/odf-console", "url": "https://github.com/red-hat-storage/odf-console/pull/1452", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
430020512
Added async and sync client wrappers. Added new async test mechanism. See #196. This PR improves async/sync APIs. Design decision I choose public inheritance and void function()=delete approach. Why don't use composition and forwarding? This approach requires many forwarding functions. If the function parameter would change, many part of codes need to change. It is difficult to maintain. Code image (details are omitted intentionally) // existing classes class client : public endpoint {}; // new classes by composition class sync_client { public: void publish(...) { c_->publish(...): } private: std::shared_ptr<client> c_; }; Why don't use private inheritance and using? Class endpoint use std::enable_shared_from_this. // existing classes class endpoint : public std::enable_shared_from_this<endpoint> {}; class client : public endpoint {}; // new classes by private inheritance class sync_client : private client { public: using client::publish; }; shared_from_this() throws bad_weak_ptr exception on runtime. Why use void function_name() = delete ? =delete only checks function name. So I choose the most simple signature and return type. Consider if I use the complete function signature and return type for =delete, if the function has overloads, which one should be chosen? I think there is no appropriate answer. // existing classes class endpoint : public std::enable_shared_from_this<endpoint> {}; class client : public endpoint {}; // new classes by public inheritance class sync_client : public client { public: void async_publish() = delete; // always no parameter and returns void }; Why don't you create asnyc_client and sync_client as the base classes of client? It requires big design change. I don't have good design concept.
gharchive/pull-request
2019-04-06T10:29:22
2025-04-01T06:40:12.606596
{ "authors": [ "redboltz" ], "repo": "redboltz/mqtt_cpp", "url": "https://github.com/redboltz/mqtt_cpp/pull/208", "license": "BSL-1.0", "license_type": "permissive", "license_source": "github-api" }
1763809983
Appstudio update test-component-pac-zvdm Pipelines as Code configuration proposal To start the PipelineRun, add a new comment with content /ok-to-test For more detailed information about running a PipelineRun, please refer to Pipelines as Code documentation Running the PipelineRun To customize the proposed PipelineRuns after merge, please refer to Build Pipeline customization Pipelines as Code CI/test-component-pac-zvdm-on-pull-request has successfully validated your commit. StatusDurationName ✅ Succeeded 7 seconds init ✅ Succeeded 12 seconds clone-repository ✅ Succeeded 29 seconds build-container ✅ Succeeded 20 seconds inspect-image ✅ Succeeded 1 minute deprecated-base-image-check ✅ Succeeded 3 minutes clair-scan ✅ Succeeded 52 seconds clamav-scan ✅ Succeeded 17 seconds sbom-json-check ✅ Succeeded 17 seconds label-check ✅ Succeeded 8 seconds show-sbom ✅ Succeeded 9 seconds show-summary Pipelines as Code CI/test-component-pac-zvdm-on-pull-request has failed. StatusDurationName --- --- init Pipelines as Code CI/test-component-pac-zvdm-on-pull-request has successfully validated your commit. StatusDurationName ✅ Succeeded 7 seconds init ✅ Succeeded 21 seconds clone-repository ✅ Succeeded 25 seconds build-container ✅ Succeeded 12 seconds deprecated-base-image-check ✅ Succeeded 12 seconds inspect-image ✅ Succeeded 11 seconds clair-scan ✅ Succeeded 37 seconds clamav-scan ✅ Succeeded 11 seconds sbom-json-check ✅ Succeeded 26 seconds label-check ✅ Succeeded 6 seconds show-sbom ✅ Succeeded 5 seconds show-summary Pipelines as Code CI/test-component-pac-zvdm-on-pull-request has successfully validated your commit. StatusDurationName ✅ Succeeded 7 seconds init ✅ Succeeded 21 seconds clone-repository ✅ Succeeded 25 seconds build-container ✅ Succeeded 12 seconds deprecated-base-image-check ✅ Succeeded 12 seconds inspect-image ✅ Succeeded 11 seconds clair-scan ✅ Succeeded 37 seconds clamav-scan ✅ Succeeded 11 seconds sbom-json-check ✅ Succeeded 26 seconds label-check ✅ Succeeded 6 seconds show-sbom ✅ Succeeded 5 seconds show-summary
gharchive/pull-request
2023-06-19T15:56:12
2025-04-01T06:40:12.660957
{ "authors": [ "rhtap-qe-bots" ], "repo": "redhat-appstudio-qe/devfile-sample-hello-world", "url": "https://github.com/redhat-appstudio-qe/devfile-sample-hello-world/pull/11000", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1518587243
Appstudio update test-component-pac-kjnh Pipelines as Code configuration proposal Pipelines as Code CI/test-component-pac-kjnh-on-pull-request has successfully validated your commit. StatusDurationName ✅ Succeeded 9 seconds appstudio-init ✅ Succeeded 9 seconds clone-repository ✅ Succeeded 18 seconds appstudio-configure-build ✅ Succeeded 19 seconds sast-snyk-check ✅ Succeeded 43 seconds build-container ✅ Succeeded 20 seconds sanity-inspect-image ✅ Succeeded 17 seconds deprecated-base-image-check ✅ Succeeded 49 seconds clamav-scan ✅ Succeeded 17 seconds clair-scan ✅ Succeeded 23 seconds sbom-json-check ✅ Succeeded 12 seconds sanity-label-check ✅ Succeeded 11 seconds sanity-optional-label-check ✅ Succeeded 8 seconds show-summary Pipelines as Code CI/test-component-pac-kjnh-on-pull-request has successfully validated your commit. StatusDurationName ✅ Succeeded 9 seconds appstudio-init ✅ Succeeded 9 seconds clone-repository ✅ Succeeded 18 seconds appstudio-configure-build ✅ Succeeded 19 seconds sast-snyk-check ✅ Succeeded 43 seconds build-container ✅ Succeeded 20 seconds sanity-inspect-image ✅ Succeeded 17 seconds clair-scan ✅ Succeeded 49 seconds clamav-scan ✅ Succeeded 17 seconds deprecated-base-image-check ✅ Succeeded 23 seconds sbom-json-check ✅ Succeeded 11 seconds sanity-optional-label-check ✅ Succeeded 12 seconds sanity-label-check ✅ Succeeded 8 seconds show-summary Pipelines as Code CI/test-component-pac-kjnh-on-pull-request has successfully validated your commit. StatusDurationName ✅ Succeeded 8 seconds appstudio-init ✅ Succeeded 15 seconds clone-repository ✅ Succeeded 13 seconds sast-snyk-check ✅ Succeeded 14 seconds appstudio-configure-build ✅ Succeeded 45 seconds build-container ✅ Succeeded 24 seconds sanity-inspect-image ✅ Succeeded 22 seconds deprecated-base-image-check ✅ Succeeded 47 seconds clamav-scan ✅ Succeeded 17 seconds clair-scan ✅ Succeeded 27 seconds sbom-json-check ✅ Succeeded 13 seconds sanity-optional-label-check ✅ Succeeded 11 seconds sanity-label-check ✅ Succeeded 6 seconds show-summary Pipelines as Code CI/test-component-pac-kjnh-on-pull-request has successfully validated your commit. StatusDurationName ✅ Succeeded 8 seconds appstudio-init ✅ Succeeded 15 seconds clone-repository ✅ Succeeded 14 seconds appstudio-configure-build ✅ Succeeded 13 seconds sast-snyk-check ✅ Succeeded 45 seconds build-container ✅ Succeeded 22 seconds deprecated-base-image-check ✅ Succeeded 24 seconds sanity-inspect-image ✅ Succeeded 47 seconds clamav-scan ✅ Succeeded 17 seconds clair-scan ✅ Succeeded 27 seconds sbom-json-check ✅ Succeeded 13 seconds sanity-optional-label-check ✅ Succeeded 11 seconds sanity-label-check ✅ Succeeded 6 seconds show-summary
gharchive/pull-request
2023-01-04T09:03:19
2025-04-01T06:40:12.693448
{ "authors": [ "jkopriva" ], "repo": "redhat-appstudio-qe/devfile-sample-hello-world", "url": "https://github.com/redhat-appstudio-qe/devfile-sample-hello-world/pull/1141", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1595725735
Appstudio update mvp-test-component Pipelines as Code configuration proposal Pipelines as Code CI/mvp-test-component-on-pull-request has successfully validated your commit. StatusDurationName ✅ Succeeded 10 seconds init ✅ Succeeded 18 seconds clone-repository ✅ Succeeded 1 minute build-container ✅ Succeeded 12 seconds sanity-inspect-image ✅ Succeeded 11 seconds deprecated-base-image-check ✅ Succeeded 58 seconds clamav-scan ✅ Succeeded 16 seconds clair-scan ✅ Succeeded 9 seconds sbom-json-check ✅ Succeeded 17 seconds sanity-label-check ✅ Succeeded 15 seconds sanity-optional-label-check ✅ Succeeded 6 seconds show-summary
gharchive/pull-request
2023-02-22T19:43:12
2025-04-01T06:40:12.703049
{ "authors": [ "redhat-appstudio-qe-bot2" ], "repo": "redhat-appstudio-qe/hacbs-test-project", "url": "https://github.com/redhat-appstudio-qe/hacbs-test-project/pull/32", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1318566968
Added ppc64le support This PR will add multi-architecture support to the application-service operator image /retest
gharchive/pull-request
2022-07-26T17:44:32
2025-04-01T06:40:12.704203
{ "authors": [ "mayurwaghmode" ], "repo": "redhat-appstudio/application-service", "url": "https://github.com/redhat-appstudio/application-service/pull/156", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2387910778
KFLUXINFRA-651: Adding Instance Types for Multi-Platform Builds Adding two types of instances Higher Memory (1:4 ratio for cpu: memory) Higher CPU (1:2 ratio for cpu: memory) Note Some users may require more memory and less CPU, making Memory Optimized Instances the better choice for them. On the other hand, some users may need more CPU power but less memory, in which case CPU-optimized instances would be a good fit. Therefore, a combination of memory-optimized and CPU-optimized instances is essential to cater to different user needs. In the label naming, m refers to memory optimised and c refers to compute optimised. Memory Optimised (not aws terminology, their memory optimized class starts from r) Multi-Platform Label Instance Type Architecture CPU Memory(GB) linux-mlarge/amd64 m5a.large AMD64 2 8 linux-mxlarge/amd64 m6a.xlarge AMD64 4 16 linux-m2xlarge/amd64 m6a.2xlarge AMD64 8 32 linux-m4xlarge/amd64 m6a.4xlarge AMD64 16 64 linux-m8xlarge/amd64 m6a.8xlarge AMD64 32 128 linux-mlarge/arm64 m6g.large ARM64 2 8 linux-mxlarge/arm64 m6g.xlarge ARM64 4 16 linux-m2xlarge/arm64 m6g.2xlarge ARM64 8 32 linux-m4xlarge/arm64 m6g.4xlarge ARM64 16 64 linux-m8xlarge/arm64 m6g.8xlarge ARM64 32 128 CPU Optimised (again not aws terminology) Multi-Platform Label Instance Type Architecture CPU Memory(GB) linux-clarge/amd64 c6a.xlarge AMD64 4 8 linux-cxlarge/amd64 c6a.2xlarge AMD64 8 16 linux-c2xlarge/amd64 c6a.4xlarge AMD64 16 32 linux-c4xlarge/amd64 c6a.8xlarge AMD64 32 64 linux-clarge/arm64 c6g.xlarge ARM64 4 8 linux-cxlarge/arm64 c6g.2xlarge ARM64 8 16 linux-c2xlarge/arm64 c6g.4xlarge ARM64 16 32 linux-c4xlarge/arm64 c6g.8xlarge ARM64 32 64 Suggestion We can minimize code duplication by utilizing default values for common parameters like imageId, subnetId, region, etc, unless specified otherwise. One thought: we will have quay.io with AWS internal communication right? otherwise how about using enhanced network enabled instances (with *n) for better push pull speeds? One thought: we will have quay.io with AWS internal communication right? otherwise how about using enhanced network enabled instances (with *n) for better push pull speeds? ENA is enabled for the instance types we are using m6a.xxxxx, m6g.xxxxx, c6a.xxxxx and c6g.xxxxx. Reference AWS Document -- https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/enhanced-networking-ena.html /lgtm
gharchive/pull-request
2024-07-03T07:39:20
2025-04-01T06:40:12.721748
{ "authors": [ "bamachrn", "manish-jangra" ], "repo": "redhat-appstudio/infra-deployments", "url": "https://github.com/redhat-appstudio/infra-deployments/pull/3997", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2752668445
Promote release-service from staging to production Included PRs: https://github.com/konflux-ci/release-service/pull/638 https://github.com/konflux-ci/release-service/pull/637 https://github.com/konflux-ci/release-service/pull/636 https://github.com/konflux-ci/release-service/pull/635 https://github.com/konflux-ci/release-service/pull/633 https://github.com/konflux-ci/release-service/pull/630 https://github.com/konflux-ci/release-service/pull/629 https://github.com/konflux-ci/release-service/pull/632 https://github.com/konflux-ci/release-service/pull/628 https://github.com/konflux-ci/release-service/pull/627 https://github.com/konflux-ci/release-service/pull/626 https://github.com/konflux-ci/release-service/pull/622 /lgtm
gharchive/pull-request
2024-12-20T12:43:26
2025-04-01T06:40:12.726992
{ "authors": [ "johnbieren", "mmalina" ], "repo": "redhat-appstudio/infra-deployments", "url": "https://github.com/redhat-appstudio/infra-deployments/pull/5199", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1358402092
fix: increase timeout for e2e Why Currently there's a 1h (default) timeout for running e2e. Since the number of tests is growing and takes more time to finish, we are sometimes hitting the limit. If that happens, tests get interrupted and a CI check is marked as failed. This is a temporary solution until we enable parallel e2e test run for infra-deployments (which was already enabled for e2e-tests repo) /lgtm Ran 87 of 92 Specs in 4424.652 seconds SUCCESS! -- 87 Passed | 0 Failed | 0 Pending | 5 Skipped
gharchive/pull-request
2022-09-01T07:44:33
2025-04-01T06:40:12.728576
{ "authors": [ "psturc", "tkdchen" ], "repo": "redhat-appstudio/infra-deployments", "url": "https://github.com/redhat-appstudio/infra-deployments/pull/684", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
981505573
Add auth-option to authenticate as githubapp Fixes #112. Adds support for authenticating as github-app, which offers: more narrow-grained permissions (avoiding PATs) increased security, using a private key increased API-quota The github-client is now wired up/facilitated through the go-github-app library, which offers caching, metrics etc. Note: Draft PR - not yet tested, for feedback. go-githubapp now has convenience function to configure the transport. I think this is good to go, but have not yet been able to test it. That should be it. I've updated the docs and tested it in my 4.6.x cluster, using app-based auth: group-sync-operator-controller-manager-54d5874d76-kh5cq manager 2021-09-06T17:31:04.501Z INFO controllers.GroupSync Beginning Sync {"groupsync": "group-sync-operator/github-groupsync", "Provider": "github"} group-sync-operator-controller-manager-54d5874d76-kh5cq manager 2021-09-06T17:31:06.354Z INFO controllers.GroupSync Sync Completed Successfully {"groupsync": "group-sync-operator/github-groupsync", "Provider": "github", "Groups Created or Updated": 1} CC @sabre1041 @raffaelespazzoli @sabre1041 I think all points have been covered now - thanks for the input!
gharchive/pull-request
2021-08-27T18:48:01
2025-04-01T06:40:12.731904
{ "authors": [ "davidkarlsen" ], "repo": "redhat-cop/group-sync-operator", "url": "https://github.com/redhat-cop/group-sync-operator/pull/118", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
989044632
Run go mod tidy on each pull request and fail the build if there are changes We will be outputting the vendor directory in the midstream repo, and this will prevent any possible conflicts. Actually this is not needed: https://github.blog/changelog/2020-10-19-dependabot-go-mod-tidy-and-vendor-support/
gharchive/issue
2021-09-06T10:54:31
2025-04-01T06:40:12.736214
{ "authors": [ "craicoverflow" ], "repo": "redhat-developer/app-services-cli", "url": "https://github.com/redhat-developer/app-services-cli/issues/1029", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1183258588
fix: add server side backed up metrics Motivation This change uses backend compiled metrics for consumer groups. This have been in production for quite a while and have been extensively tested. Verification Get Kafka Create new topic test Create kcat.properties based of the kcat guide kcat -b <yourhostname -F ./kcat.properties -P -t test` run rhoas kafka consumer-group list run rhoas kafka consumer-group describe I have used latest version of SDK. We need to wait to make sure that it is currently deployed to production @rkpattnaik780 FYI @mikeedgar API is working find and was tested end to end
gharchive/pull-request
2022-03-28T10:57:12
2025-04-01T06:40:12.739325
{ "authors": [ "wtrocki" ], "repo": "redhat-developer/app-services-cli", "url": "https://github.com/redhat-developer/app-services-cli/pull/1493", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
863700015
Provide simplified ability to switch between stagging and production environments We can add env variable into OLM as example (can be empty) and then scripts can patch it (as discussed with @b1zzu) I will verify if that really works and document this in contributing docs Added env variable directly to the OLM
gharchive/issue
2021-04-21T10:06:53
2025-04-01T06:40:12.740482
{ "authors": [ "wtrocki" ], "repo": "redhat-developer/app-services-operator", "url": "https://github.com/redhat-developer/app-services-operator/issues/189", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2109837968
doc: OpenShift argocd CLI client command reference documentation What type of PR is this? /kind documentation What does this PR do / why we need it: Have you updated the necessary documentation? [ ] Documentation update is required by this PR. [ ] Documentation has been updated. Which issue(s) this PR fixes: Fixes #? Test acceptance criteria: [ ] Unit Test [ ] E2E Test How to test changes / Special notes to the reviewer: @anandf How about the updates to these sections? Update these existing sections: Creating an application by using the oc tool https://docs.openshift.com/gitops/1.11/argocd_applications/deploying-a-spring-boot-application-with-argo-cd.html#creating-an-application-by-using-the-oc-tool_deploying-a-spring-boot-application-with-argo-cd https://docs.openshift.com/gitops/1.11/declarative_clusterconfig/configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations.html#creating-an-application-by-using-the-oc-tool_configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations Synchronizing your application with your Git repository https://docs.openshift.com/gitops/1.11/declarative_clusterconfig/configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations.html#synchronizing-your-application-application-with-your-git-repository_configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations @anandf Can we add some content for these sections, if they are relevant at all? About the OpenShift argo cd CLI (I think this is a nice to have section) Logging in to the OpenShift argo cd CLI using a web browser usage instructions - information that must go as admonitions such as Note, Important, Tip, caution, or warning. @anandf How about the updates to these sections? Update these existing sections: Creating an application by using the oc tool https://docs.openshift.com/gitops/1.11/argocd_applications/deploying-a-spring-boot-application-with-argo-cd.html#creating-an-application-by-using-the-oc-tool_deploying-a-spring-boot-application-with-argo-cd https://docs.openshift.com/gitops/1.11/declarative_clusterconfig/configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations.html#creating-an-application-by-using-the-oc-tool_configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations Synchronizing your application with your Git repository https://docs.openshift.com/gitops/1.11/declarative_clusterconfig/configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations.html#synchronizing-your-application-application-with-your-git-repository_configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations Added sections for creating and syncing app using CLI in both normal and core modes. @anandf Can we add some content for these sections, if they are relevant at all? Logging in to the OpenShift argo cd CLI using a web browser -> This is not applicable. One has to login via the CLI itself. Its not possible to login via web browser like its possible for oc login
gharchive/pull-request
2024-01-31T11:00:38
2025-04-01T06:40:12.750950
{ "authors": [ "Srivaralakshmi", "anandf" ], "repo": "redhat-developer/gitops-operator", "url": "https://github.com/redhat-developer/gitops-operator/pull/650", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1253785178
odo dev add flag to run non-default build command. User Story As an odo user I want to be able to execute alternative build command instead of the default one So that I can run my with different options (different configuration, flags, runtime versions etc..) without modifying the default command. Example: Go devfile has the following default build command. - exec: commandLine: GOCACHE=${PROJECT_SOURCE}/.cache go build main.go component: runtime group: isDefault: true kind: build workingDir: ${PROJECT_SOURCE} id: build In some situations, I want to pass arguments to linker to for example set string value of a variable. This can be done adding -ldflags="-X github.com/redhat-developer/odo/pkg/segment.writeKey=foo". Currently I have to edit the commandLine in order to be able to execute build with different arguments. Instead of that it would be nice to be able to add extra run command like this: - exec: commandLine: GOCACHE=${PROJECT_SOURCE}/.cache go build -ldflags="-X github.com/redhat-developer/odo/pkg/segment.writeKey=foo" main.go component: runtime group: isDefault: false kind: build workingDir: ${PROJECT_SOURCE} id: build-with-key And then simply execute odo dev --build-command build-with-key, when I want to switch back to using the default profile, I stop running odo dev command and start it again, but this time without --build-command flag. Acceptance Criteria [ ] odo dev should have --build-command flag that controls what command is used to build application. /kind user-story TODO (to not forget): As commented out in [1]: we should try to harmonize the behaviour between commands in pkg/libdevfile/libdevfile.go. Deploy is using getDefaultCommand, when Build/Test/Run is using getCommandAssociatedToGroup [1] https://github.com/redhat-developer/odo/pull/5768#discussion_r895480523
gharchive/issue
2022-05-31T12:18:45
2025-04-01T06:40:12.759820
{ "authors": [ "kadel", "rm3l" ], "repo": "redhat-developer/odo", "url": "https://github.com/redhat-developer/odo/issues/5776", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
322855232
fix release scripts - 'odo version' output changed 😞 forgot to change one place similar to #451
gharchive/pull-request
2018-05-14T14:59:48
2025-04-01T06:40:12.760897
{ "authors": [ "kadel" ], "repo": "redhat-developer/odo", "url": "https://github.com/redhat-developer/odo/pull/452", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1214188266
Provide tooltips on icon As a user, the icons themselves are not going to be sufficient to tell me what actions are behind each buttons. It would be nicer to provide tooltips behind the button, so that the user can discover what's the button will do, without having to click on it.
gharchive/issue
2022-04-25T08:56:52
2025-04-01T06:40:12.762474
{ "authors": [ "fbricon", "slemeur" ], "repo": "redhat-developer/openshift-dd-ext", "url": "https://github.com/redhat-developer/openshift-dd-ext/issues/21", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2550180431
RHIDP-3377: Adding cross-links to the Installation titles IMPORTANT: Do Not Merge - To be merged by Docs Team Only Version(s): 1.2, 1.3 Issue: RHIDP-3377 Reviews: [x] Docs review: @hmanwani-rh /cherry-pick release-1.3 /cherry-pick 1.2.x
gharchive/pull-request
2024-09-26T10:28:44
2025-04-01T06:40:12.770318
{ "authors": [ "hmanwani-rh", "pabel-rh" ], "repo": "redhat-developer/red-hat-developers-documentation-rhdh", "url": "https://github.com/redhat-developer/red-hat-developers-documentation-rhdh/pull/538", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2652518906
Added discover content I’m creating this PR because the About RHDH title is already published on docs.redhat.com, but currently, it doesn’t contain any content. For now, I’m adding content to the page, which can be refined and improved later. cc @jmagak Closing this PR, since the title is now unpublished in Pantheon.
gharchive/pull-request
2024-11-12T15:26:55
2025-04-01T06:40:12.771856
{ "authors": [ "hmanwani-rh" ], "repo": "redhat-developer/red-hat-developers-documentation-rhdh", "url": "https://github.com/redhat-developer/red-hat-developers-documentation-rhdh/pull/698", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
160856652
Trigger a first build upon fling completion Create a new Build resource from the generated BuildConfig resource in the OpenShift project. Would adding this trigger (https://github.com/tnozicka/openshift-templates/blob/master/pipeline-template.yaml#L34) be enough? 6d85abfc3440d379e7135325f7005bbf4b80e57c
gharchive/issue
2016-06-17T10:17:34
2025-04-01T06:40:12.824579
{ "authors": [ "ALRubinger", "tnozicka", "xcoulon" ], "repo": "redhat-kontinuity/catapult", "url": "https://github.com/redhat-kontinuity/catapult/issues/128", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1053032322
operator vfunction-server-operator (v2.3.594) New operator bundle Name: vfunction-server-operator Version: v2.3.594 Certification project: 5f64dc1e0a4d06443d2818ad Test result URL: https://catalog.redhat.com/api/containers/v1/projects/certification/test-results/id/61917043d115ef379a598e5e Test logs URL: https://catalog.redhat.com/api/containers/v1/projects/certification/artifacts/id/61917042d6b39d004292bad1 Hi, I'm sure 100% that my ci tests passed okay (even 3 times...) on my local OpenShift cluster with the most updated operator-ci-pipeline.yml and tasks from the operator-pipelines repo. I did have some issues with timeout exceeding on both ScorecardBasicSpecCheck & ScorecardOlmSuiteCheck before I updated. (I couldn't access the test result URL nor the logs URL from above to see the exact errors for this run - to be sure they are the same, though) Before starting the long journey of opening a support case for this, I just wanted to make sure that the errors are not on your end because the tests cluster is not updated correctly? Could you please verify that tests can pass at all now for ScorecardBasicSpecCheck & ScorecardOlmSuiteCheck on your testing cluster? Thanks. Now I could also see in the logs that both ScorecardBasicSpecCheck and ScorecardOlmSuiteCheck failed on the 30 seconds timeout (as I expected and mentioned in my previous comment). As I also mentioned - I had this error too until I pulled a new version from operator-pipelines and applied all pipelines and tasks on the OpenShift cluster. From the last Perflight.log: time="2021-11-15T13:58:35Z" level=info msg="running check: ScorecardBasicSpecCheck" time="2021-11-15T13:58:35Z" level=debug msg="Running operator-sdk scorecard check for quay.io/operator-pipeline-prod/vfunction-server-operator:v2.3.594" time="2021-11-15T13:58:35Z" level=debug msg="--selector=[test=basic-check-spec-test]" time="2021-11-15T13:58:35Z" level=trace msg="running scorecard with the following invocation[operator-sdk scorecard --output json --selector=test=basic-check-spec-test --kubeconfig /tmp/kubeconfig-3421170286 --namespace default --service-account default --config /tmp/scorecard-test-config-3472466720.yaml --verbose /tmp/preflight-3179944778/fs]" time="2021-11-15T13:59:05Z" level=error msg="stdout: " time="2021-11-15T13:59:05Z" level=error msg="stderr: time="2021-11-15T13:58:35Z" level=debug msg="Debug logging is set"\nError: error running tests context deadline exceeded\nUsage:\n operator-sdk scorecard [flags]\n\nFlags:\n -c, --config string path to scorecard config file\n -h, --help help for scorecard\n --kubeconfig string kubeconfig path\n -L, --list Option to enable listing which tests are run\n -n, --namespace string namespace to run the test images in\n -o, --output string Output format for results. Valid values: text, json, xunit (default "text")\n -l, --selector string label selector to determine which tests are run\n -s, --service-account string Service account to use for tests (default "default")\n -x, --skip-cleanup Disable resource cleanup after tests are run\n -b, --storage-image string Storage image to be used by the Scorecard pod (default "docker.io/library/busybox@sha256:c71cb4f7e8ececaffb34037c2637dc86820e4185100e18b4d02d613a9bd772af")\n -t, --test-output string Test output directory. (default "test-output")\n -u, --untar-image string Untar image to be used by the Scorecard pod (default "registry.access.redhat.com/ubi8@sha256:910f6bc0b5ae9b555eb91b88d28d568099b060088616eba2867b07ab6ea457c7")\n -w, --wait-time duration seconds to wait for tests to complete. Example: 35s (default 30s)\n\nGlobal Flags:\n --plugins strings plugin keys to be used for this subcommand execution\n --verbose Enable verbose logging\n\ntime="2021-11-15T13:59:05Z" level=fatal msg="error running tests context deadline exceeded"\n" time="2021-11-15T13:59:05Z" level=info msg="check completed: ScorecardBasicSpecCheck" ERROR="failed to run operator-sdk scorecard: exit status 1" result="failed to run operator-sdk scorecard: exit status 1" time="2021-11-15T13:59:05Z" level=info msg="running check: ScorecardOlmSuiteCheck" time="2021-11-15T13:59:05Z" level=debug msg="Running operator-sdk scorecard Check for quay.io/operator-pipeline-prod/vfunction-server-operator:v2.3.594" time="2021-11-15T13:59:05Z" level=debug msg="--selector=[suite=olm]" time="2021-11-15T13:59:05Z" level=trace msg="running scorecard with the following invocation[operator-sdk scorecard --output json --selector=suite=olm --kubeconfig /tmp/kubeconfig-3421170286 --namespace default --service-account default --config /tmp/scorecard-test-config-2813129432.yaml --verbose /tmp/preflight-3179944778/fs]" time="2021-11-15T13:59:36Z" level=error msg="stdout: " time="2021-11-15T13:59:36Z" level=error msg="stderr: time="2021-11-15T13:59:05Z" level=debug msg="Debug logging is set"\nError: error running tests context deadline exceeded\nUsage:\n operator-sdk scorecard [flags]\n\nFlags:\n -c, --config string path to scorecard config file\n -h, --help help for scorecard\n --kubeconfig string kubeconfig path\n -L, --list Option to enable listing which tests are run\n -n, --namespace string namespace to run the test images in\n -o, --output string Output format for results. Valid values: text, json, xunit (default "text")\n -l, --selector string label selector to determine which tests are run\n -s, --service-account string Service account to use for tests (default "default")\n -x, --skip-cleanup Disable resource cleanup after tests are run\n -b, --storage-image string Storage image to be used by the Scorecard pod (default "docker.io/library/busybox@sha256:c71cb4f7e8ececaffb34037c2637dc86820e4185100e18b4d02d613a9bd772af")\n -t, --test-output string Test output directory. (default "test-output")\n -u, --untar-image string Untar image to be used by the Scorecard pod (default "registry.access.redhat.com/ubi8@sha256:910f6bc0b5ae9b555eb91b88d28d568099b060088616eba2867b07ab6ea457c7")\n -w, --wait-time duration seconds to wait for tests to complete. Example: 35s (default 30s)\n\nGlobal Flags:\n --plugins strings plugin keys to be used for this subcommand execution\n --verbose Enable verbose logging\n\ntime="2021-11-15T13:59:36Z" level=fatal msg="error running tests context deadline exceeded"\n" time="2021-11-15T13:59:36Z" level=info msg="check completed: ScorecardOlmSuiteCheck" ERROR="failed to run operator-sdk scorecard: exit status 1" result="failed to run operator-sdk scorecard: exit status 1" Thank you. We are looking into this on our side. Hi, Any news regarding this? Thanks.
gharchive/pull-request
2021-11-14T20:23:45
2025-04-01T06:40:12.842182
{ "authors": [ "tonytcampbell", "vfunction" ], "repo": "redhat-openshift-ecosystem/certified-operators", "url": "https://github.com/redhat-openshift-ecosystem/certified-operators/pull/119", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1305416246
operator neuvector-community-operator (1.3.5) Signed-off-by: selvamt94 sthangaraj@neuvector.com Thanks submitting your Operator. Please check below list before you create your Pull Request. New Submissions [x] Are you familiar with our contribution guidelines? [x] Have you packaged and deployed your Operator for Operator Framework? [x] Have you tested your Operator with all Custom Resource Definitions? [x] Have you tested your Operator in all supported installation modes? [x] Have you considered whether you want use semantic versioning order? [x] Is your submission signed? [x] Is operator icon set? Updates to existing Operators [x] Did you create a ci.yaml file according to the update instructions? [x] Is your new CSV pointing to the previous version with the replaces property if you chose replaces-mode via the updateGraph property in ci.yaml? [x] Is your new CSV referenced in the appropriate channel defined in the package.yaml or annotations.yaml ? [x] Have you tested an update to your Operator when deployed via OLM? [x] Is your submission signed? Your submission should not [x] Modify more than one operator [x] Modify an Operator you don't own [x] Rename an operator - please remove and add with a different name instead [x] Modify any files outside the above mentioned folders [x] Contain more than one commit. Please squash your commits. Operator Description must contain (in order) [x] Description about the managed Application and where to find more information [x] Features and capabilities of your Operator and how to use it [x] Any manual steps about potential pre-requisites for using your Operator Operator Metadata should contain [x] Human readable name and 1-liner description about your Operator [x] Valid category name1 [x] One of the pre-defined capability levels2 [x] Links to the maintainer, source code and documentation [x] Example templates for all Custom Resource Definitions intended to be used [x] A quadratic logo Remember that you can preview your CSV here. -- 1 If you feel your Operator does not fit any of the pre-defined categories, file an issue against this repo and explain your need 2 For more information see here /merge possible /merge possible
gharchive/pull-request
2022-07-14T23:46:52
2025-04-01T06:40:12.855418
{ "authors": [ "framework-automation", "selvamt94" ], "repo": "redhat-openshift-ecosystem/community-operators-prod", "url": "https://github.com/redhat-openshift-ecosystem/community-operators-prod/pull/1414", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1435942856
operator cert-manager (1.10.1) Signed-off-by: Richard Wall richard.wall@jetstack.io /hold Sanity checking the changes between 1.10.0 and 1.10.1 ✦ $ git diff --no-index 1.10.0 1.10.1 diff --git a/1.10.0/manifests/cert-manager.clusterserviceversion.yaml b/1.10.1/manifests/cert-manager.clusterserviceversion.yaml index 4f298cca..402923d3 100644 --- a/1.10.0/manifests/cert-manager.clusterserviceversion.yaml +++ b/1.10.1/manifests/cert-manager.clusterserviceversion.yaml @@ -68,7 +68,8 @@ metadata: capabilities: Full Lifecycle categories: Security containerImage: quay.io/jetstack/cert-manager-controller:v1.10.0 - createdAt: '2022-11-02T14:25:05' + createdAt: '2022-11-04T10:22:12' + olm.skipRange: '>=1.10.0 <1.10.1' operators.operatorframework.io/builder: operator-sdk-v1.25.0 operators.operatorframework.io/internal-objects: |- [ @@ -83,7 +84,7 @@ metadata: operatorframework.io/arch.arm64: supported operatorframework.io/arch.ppc64le: supported operatorframework.io/arch.s390x: supported - name: cert-manager.v1.10.0 + name: cert-manager.v1.10.1 namespace: placeholder spec: apiservicedefinitions: {} @@ -745,9 +746,9 @@ spec: - --secure-port=10250 - --dynamic-serving-ca-secret-namespace=$(POD_NAMESPACE) - --dynamic-serving-ca-secret-name=cert-manager-webhook-ca - - --dynamic-serving-dns-names=cert-manager-webhook - - --dynamic-serving-dns-names=cert-manager-webhook.$(POD_NAMESPACE) - - --dynamic-serving-dns-names=cert-manager-webhook.$(POD_NAMESPACE).svc + - --dynamic-serving-dns-names=cert-manager-webhook-service.$(POD_NAMESPACE).svc + - --tls-cert-file=/apiserver.local.config/certificates/apiserver.crt + - --tls-private-key-file=/apiserver.local.config/certificates/apiserver.key env: - name: POD_NAMESPACE valueFrom: @@ -879,7 +880,7 @@ spec: provider: name: The cert-manager maintainers url: https://cert-manager.io/ - version: 1.10.0 + version: 1.10.1 webhookdefinitions: - admissionReviewVersions: - v1 /merge possible /merge possible /hold cancel /merge possible
gharchive/pull-request
2022-11-04T11:37:36
2025-04-01T06:40:12.858776
{ "authors": [ "framework-automation", "wallrj" ], "repo": "redhat-openshift-ecosystem/community-operators-prod", "url": "https://github.com/redhat-openshift-ecosystem/community-operators-prod/pull/1819", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1947502468
operator pulp-operator (1.0.0-beta.2) Thanks for submitting your Operator. Please check the below list before you create your Pull Request. New Submissions [ ] Are you familiar with our contribution guidelines? [ ] Have you packaged and deployed your Operator for Operator Framework? [ ] Have you tested your Operator with all Custom Resource Definitions? [ ] Have you tested your Operator in all supported installation modes? [ ] Have you considered whether you want to use semantic versioning order? [ ] Is your submission signed? [ ] Is operator icon set? Updates to existing Operators [ ] Did you create a ci.yaml file according to the update instructions? [ ] Is your new CSV pointing to the previous version with the replaces property if you chose replaces-mode via the updateGraph property in ci.yaml? [ ] Is your new CSV referenced in the appropriate channel defined in the package.yaml or annotations.yaml ? [ ] Have you tested an update to your Operator when deployed via OLM? [ ] Is your submission signed? Your submission should not [ ] Modify more than one operator [ ] Modify an Operator you don't own [ ] Rename an operator - please remove and add with a different name instead [ ] Modify any files outside the above mentioned folders [ ] Contain more than one commit. Please squash your commits. Operator Description must contain (in order) [ ] Description of the managed Application and where to find more information [ ] Features and capabilities of your Operator and how to use it [ ] Any manual steps about potential pre-requisites for using your Operator Operator Metadata should contain [ ] Human readable name and 1-liner description about your Operator [ ] Valid category name1 [ ] One of the pre-defined capability levels2 [ ] Links to the maintainer, source code and documentation [ ] Example templates for all Custom Resource Definitions intended to be used [ ] A quadratic logo Remember that you can preview your CSV here. -- 1 If you feel your Operator does not fit any of the pre-defined categories, file an issue against this repo and explain your need 2 For more information see here /merge possible
gharchive/pull-request
2023-10-17T13:57:57
2025-04-01T06:40:12.871475
{ "authors": [ "framework-automation", "git-hyagi" ], "repo": "redhat-openshift-ecosystem/community-operators-prod", "url": "https://github.com/redhat-openshift-ecosystem/community-operators-prod/pull/3465", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1988387321
operator [CI] konveyor-operator Just switching to use semver-mode. /merge possible /merge possible
gharchive/pull-request
2023-11-10T20:59:55
2025-04-01T06:40:12.872605
{ "authors": [ "djzager", "framework-automation" ], "repo": "redhat-openshift-ecosystem/community-operators-prod", "url": "https://github.com/redhat-openshift-ecosystem/community-operators-prod/pull/3595", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2046624214
operator jaeger (1.52.0) Thanks for submitting your Operator. Please check the below list before you create your Pull Request. New Submissions [ ] Are you familiar with our contribution guidelines? [ ] Are you familiar with our operator pipeline? [ ] Have you tested your Operator with all Custom Resource Definitions packaging? [ ] Have you tested your Operator in all supported installation modes? [ ] Have you considered whether you want to use semantic versioning order? [ ] Is your submission signed? [ ] Is operator icon set? Your submission should not [ ] Add more than one operator bundle per PR [ ] Modify any operator [ ] Rename an operator [ ] Modify any files outside the above mentioned folders [ ] Contain more than one commit. Please squash your commits. Operator Description must contain (in order) [ ] Description of the managed Application and where to find more information [ ] Features and capabilities of your Operator and how to use it [ ] Any manual steps about potential pre-requisites for using your Operator Operator Metadata should contain [ ] Human readable name and 1-liner description about your Operator [ ] Valid category name1 [ ] One of the pre-defined capability levels2 [ ] Links to the maintainer, source code and documentation [ ] Example templates for all Custom Resource Definitions intended to be used [ ] A quadratic logo Remember that you can preview your CSV here. -- 1 If you feel your Operator does not fit any of the pre-defined categories, file an issue against this repo and explain your need 2 For more information see here /pipeline restart community-hosted-pipelin /pipeline restart community-hosted-pipeline
gharchive/pull-request
2023-12-18T13:04:06
2025-04-01T06:40:12.882079
{ "authors": [ "iblancasa" ], "repo": "redhat-openshift-ecosystem/community-operators-prod", "url": "https://github.com/redhat-openshift-ecosystem/community-operators-prod/pull/3757", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1803228341
[DRAFT]: Manifests to install preGA sync operator from custom operator catalog TODO: Understand where to ubicate the added extra-manifests for day2 installation as a policy. We need to figure out how to add those manifests as Policies or PoliciesGenTemplate (PGT) objects referenced from the SiteConfig.
gharchive/pull-request
2023-07-13T15:04:52
2025-04-01T06:40:12.884794
{ "authors": [ "jnunyez" ], "repo": "redhat-partner-solutions/vse-carslab-hub", "url": "https://github.com/redhat-partner-solutions/vse-carslab-hub/pull/85", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1404110242
add gcp image list command to CLI Hey @major @FKolwa , I am a bit confused right now. We assign the registered github secret secrets.GOOGLE_APPLICATION_CREDENTIALS to an environment variable GCP_APP_CREDENTIALS, but we never use it. In the get_google_images function we create an compute_v1.ImagesClient client, is he supposed to use those credentials? I went to the google docs, but it also didn't really helped me. :) Can you give me some hints? :sweat_smile: @miyunari Yup that is correct! It isn't necessary to reference the ENV in code. In fact you won't find any of the other secrets either (like AWS_ACCESS_KEY_ID). Most (if not all) cloud provider CLIs authenticate using local configurations that are loaded into ENVs at runtime. Usually you would use you user credentials to sign into Google Cloud but with an automated service like this a service account can be used. The fact that "GOOGLE_APPLICATION_CREDENTIALS" is set for the workflow step means that the gcloud cli will be able to read it and use it to authenticate with the cloud provider API. I'll leave you a link in case you want to read more about Google Application Default Credentials (ADC): https://cloud.google.com/docs/authentication/provide-credentials-adc#local-key A small hint though: The env needs to be mapped to "GOOGLE_APPLICATION_CREDENTIALS" in the workflow context as well for it to properly work. Hope this helps a little! Thank you @FKolwa ! That was exactly the information I was looking for :smile: But now, there is another issue :sweat_smile: Unfortunately it's now unclear to me, how to test my changes. I got this reference from major, but I don't see the correlation: https://github.com/redhatcloudx/rhelocator/blob/b97967e77354d0331f0f7bc4a607c00b4b1eea16/tests/test_cli.py#L77-L103 @miyunari Haha yes this is a bit confusing tbh! The gcloud implementation is pretty rough at this point. get_google_images currently queries all images in the rhel-cloud project and returns everything that isn't deprecated. In the scope of this ticket, my requirement for an end to end test would be query the correct api endpoint by calling get_google_images by using 'runner.invoke' parse the json data confirm that all images that are returned do not contain the first level key status with the value 'DEPRECATED' In this case you can copy / paste most of what @major wrote for the azure test! For the offline test copy the structure of the e2e test you just wrote create a new mock for the google images in conftest (you can take a look at the AWS mockups. You need to create a new list of mocked images in a json format and create a new fixture that is passed to you offline test). Now for the tricky part: How do you know what data structure to expect from the google API? Well if you call gcloud and query for projects within 'rhel-cloud' you will receive something like this: { "architecture": "X86_64", "archiveSizeBytes": "4184623872", "creationTimestamp": "2022-09-20T16:32:45.492-07:00", "description": "Red Hat, Red Hat Enterprise Linux, 9, x86_64 built on 20220920, supports Shielded VM features", "diskSizeGb": "20", "family": "rhel-9", "guestOsFeatures": [ { "type": "UEFI_COMPATIBLE" }, { "type": "VIRTIO_SCSI_MULTIQUEUE" }, { "type": "SEV_CAPABLE" }, { "type": "GVNIC" } ], "id": "2043557223711896434", "kind": "compute#image", "labelFingerprint": "42WmSpB8rSM=", "licenseCodes": [ "7883559014960410759" ], "licenses": [ "https://www.googleapis.com/compute/beta/projects/rhel-cloud/global/licenses/rhel-9-server" ], "name": "rhel-9-v20220920", "rawDisk": { "containerType": "TAR", "source": "" }, "rolloutOverride": { "defaultRolloutTime": "2022-09-25T15:32:42Z", "locationRolloutPolicies": { "zones/asia-east1-a": "2022-09-22T04:32:42Z", .... "zones/us-west4-c": "2022-09-25T04:32:42Z" } }, "selfLink": "https://www.googleapis.com/compute/beta/projects/rhel-cloud/global/images/rhel-9-v20220920", "sourceType": "RAW", "status": "READY", "storageLocations": [ "eu", "asia", "us" ] } At this point we don't extract any specific information from this returned data (like we do for AWS) and this is not within the scope of your ticket so feel free to create a minimal test mockup version of this data structure that only contains something like { "status": "READY" } Oh boy, that was my mistake. I must have been looking at two things at the same time and put the wrong variable name in the actions workflow. 🤦🏻‍♂️ @major I think you were right :smile: . I shortened the variable name, because I thought we have to store it in config.py and use it somewhere :woman_facepalming: Oh, I see I have some merge conflicts, will try to resolve
gharchive/pull-request
2022-10-11T07:03:43
2025-04-01T06:40:12.898352
{ "authors": [ "FKolwa", "miyunari" ], "repo": "redhatcloudx/rhelocator", "url": "https://github.com/redhatcloudx/rhelocator/pull/37", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1797902420
feat: use uber-go/mock instead of golang/mock Resolves #297 Thanks @dlsrb6342!
gharchive/pull-request
2023-07-11T01:19:42
2025-04-01T06:40:12.910411
{ "authors": [ "dlsrb6342", "rueian" ], "repo": "redis/rueidis", "url": "https://github.com/redis/rueidis/pull/298", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
310341544
Feature Request: Add possibility to make new post with custom date-time I want to move many entries from my old blog to Hubzilla. I would like to set original date for this posts. Is it possible to add a field for setting a specific date of publication during editing post? I think we could send an invite once while creating the chatroom. We have a similar functionality for wikis and files already. We could ad a switch labeled Create a status post for this chat This is possible at post creation time using the 'Delayed Posting' feature. You can set the post date (either forwards or backwards) during post creation. This is not available when editing an existing post because the posts would have already been delivered. The feature requires an account level of 2; which you can find on your settings/account page; followed by activating the Delayed Posting feature at settings/features -> Post Composition Features The feature requires an account level of 2 Thanks, i changed account level and found it, it works good. I'll close this issue.
gharchive/issue
2018-04-01T17:55:02
2025-04-01T06:40:12.914714
{ "authors": [ "git-marijus", "zlaxy", "zotlabs" ], "repo": "redmatrix/hubzilla", "url": "https://github.com/redmatrix/hubzilla/issues/1033", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
717729354
update docs The previous documentation is misleading, as OpenSSL 1.1.0 did not change the fact that contextSetVerificationMode is necessary for OpenSSL to check anything about the certificate. Also, I added a link to the (new) function enabling the OpenSSL built-in hostname validation instead of referring to a PDF. Hi @amesgen, thanks for the PR. You absolutely right, contextSetVerificationMode has to be called even with OpenSSL >= 1.1.0. When I changed that part of the documentation a few weeks ago I wasn't careful enough. My only concern with the new phrasing that you propose is that I think it would be better to keep the note about old (<1.1.0) versions of OpenSSL not verifying automatically the peer's hostname, even with contextSetVerificationMode. How about if we added the following at the end of your version: In addition to that, if you are using a version of OpenSSL older than 1.1.0, you also need to verify that the peer's hostname matches with one specified in the certificate. You can find information on how to do that here. Newer versions of OpenSSL (>=1.1.0) perform that check automatically. or something like that. Hmm, I am no expert in OpenSSL, but I think that, even in OpenSSL >= 1.1.0, setting contextSetVerificationMode to VerifyPeer only checks that there is a valid certificate chain (how should OpenSSL know about the correct hostname? It is not passed as an argument to contextSetVerificationMode.). Hostname validation has to be enabled in addition to this, either manually or with the "new" functions added in OpenSSL 1.1.0 and 1.0.2, which I recently added to HsOpenSSL: enableHostnameValidation and vshabanov/HsOpenSSL#52, see https://wiki.openssl.org/index.php/Hostname_validation.
gharchive/pull-request
2020-10-08T23:07:53
2025-04-01T06:40:12.925479
{ "authors": [ "amesgen", "redneb" ], "repo": "redneb/HsOpenSSL-x509-system", "url": "https://github.com/redneb/HsOpenSSL-x509-system/pull/3", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
697055509
Added screenshots for some examples Context: Adds screenshots of examples. I didn't manage to get the same window theme as seen in the calculator. I took the pictures on win 10, I hope it's fine (at least it proves the lib is cross-platform :) Fixes #318 Thank you. I think the basic ideas was to embed the images like the screenshot on the README. Would be nice if you can adjust it. Images looks very very strange because there is not new line before and they seems to be aligned randomly Yes, I see now. I have checked the result in Visual Studio with a markdown plugin. Seems like GitHub and the plugin render the markdown test differently.
gharchive/pull-request
2020-09-09T18:41:42
2025-04-01T06:40:12.929094
{ "authors": [ "FloVanGH", "kivimango", "qarmin" ], "repo": "redox-os/orbtk", "url": "https://github.com/redox-os/orbtk/pull/331", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1927292706
[Bug] ... Platform Device: Phone, Laptop, Desktop, TV... windows 10 edge browser OS: Windows, Android, iOS... Browser: Chrome, Safari, Kiwi... Browser Version: Better xCloud Version: Describe the bug ...doesnt work on gta oline, it says the session has expired Screenshots/Videos If applicable, add screenshots/videos to help explain your problem. Report it to xCloud
gharchive/issue
2023-10-05T03:01:28
2025-04-01T06:40:13.018716
{ "authors": [ "Nachhito", "redphx" ], "repo": "redphx/better-xcloud", "url": "https://github.com/redphx/better-xcloud/issues/143", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
425289256
module is not defined When building with webpack 5, this line gives a ReferenceError for module: https://github.com/erikras/redux-form/blob/f209539ea65e1b46534e701c635df31c100c3259/src/util/isHotReloading.js#L3 I'm not sure how to fix this, normally I would just put a if (typeof module === 'undefined') return false above that line, but after the webpack build it just gets transformed into if (false) {} so maybe this is a webpack 5 bug. Update: Simply removing the castModule and using module directly makes it work properly. So it's a matter of the TypeScript workaround making it not work with Webpack 5. Another fix is prefixing module with globalThis. I don't know Flow enough to create a merge request that passes the type system. But using patch-package I could simply change this line to get my app working with Vite. var castModule = globalThis.module Another way I got it working was to define module: undefined in vite.config. However this might cause side-effects.
gharchive/issue
2019-03-26T08:44:27
2025-04-01T06:40:13.023757
{ "authors": [ "sbaechler", "wmertens" ], "repo": "redux-form/redux-form", "url": "https://github.com/redux-form/redux-form/issues/4412", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
753886108
Export all flow type aliases from root We need to be able to use all of them in userland, absolutely any type alias declared anywhere in the flow type defs. Making wrapper components is a huge hassle without these, because it's difficult to type props for our wrapper components. import type { FieldInputProps, FieldArrayInputProps, Validator /* etc */ } from 'redux-form' I'll try to make a PR for this soon
gharchive/issue
2020-11-30T23:47:32
2025-04-01T06:40:13.024998
{ "authors": [ "jedwards1211" ], "repo": "redux-form/redux-form", "url": "https://github.com/redux-form/redux-form/issues/4703", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1741960230
mapStateToProps should not be called with functional component rendering a connected class component What version of React, ReactDOM/React Native, Redux, and React Redux are you using? React: 18.2.0 ReactDOM: 18.2.0 Redux: 4.2.1 React Redux: 8.0.5 What is the current behavior? Description: Depending on a parent component being a functionnal or class component, children connected component do not see their mapStateToProps function called in the same way Reproduction: https://codesandbox.io/s/beautiful-gianmarco-wo67u0 Click on toggle and check the console logging we can see a call to mapStateToProps in the class component child, when the parent component is a functional component ([APP1]) but not when the parent is a class component ([APP2]) Here is the logging, by app, when toggling off connected children (conditional rendering): With a functional component as parent: [APP1] CLASS CHILD MAP STATE TO PROPS [APP1] NORMAL CHILD SELECTOR x2 [APP1] NORMAL PARENT SELECTOR [APP1] RENDER PARENT {showChild: false} [APP1] NORMAL PARENT SELECTOR [APP1] RENDER PARENT {showChild: false} With a class component as parent: [APP2] CLASS PARENT MAP STATE TO PROPS x2 [APP2] RENDER PARENT {showChild: false} [APP2] NORMAL CHILD SELECTOR Note: FYI this led to an issue in our code when refactoring a component, with uncaught error in the mapStateToProps function of a child of that component, because suddenly it was being called with unexpected state and trying to get state on undefined keys. Maybe all of this is a normal consequence of the inherent difference in React rendering functional or class component, but I still wanted to document this in an issue for the record, and see if you have any comment. Thanks! What is the expected behavior? mapStateToProps should be called consistently (or should it not?) Which browser and OS are affected by this issue? Chrome OSX Did this work in previous versions of React Redux? [ ] Yes I'm not sure what you're trying to describe here, tbh. At my first guess (just reading the description), I assume that the difference in rendering is not a bug, but simply a question of which components are re-rendering and which are not, per https://blog.isquaredsoftware.com/2020/05/blogged-answers-a-mostly-complete-guide-to-react-rendering-behavior/#standard-render-behavior . I have tried to explain as accurately as possible and provided a codesandbox, but if it is unclear or need more details, I'll be happy to provide them. What I'm pointing is that the two examples in that demo are rendering exactly the same thing (supposedly) but mapState is called on one side and not on the other, which was unexpected and led to a bug, so I'm reporting on that. @pandaiolo yeah, afraid I don't have time to look at this right now (busy with day job stuff). But, my immediate guess is that: When the parent is a class component, it avoids re-rendering the child, and the child's mapState is thus not called When the parent is a function component, it does re-render the child, and thus the child's mapState is called which sounds like a combination of how React itself works in general for rendering rather than a particular bug in React-Redux. I'll try to look into this at some point, but not sure when. Sure, thank you!
gharchive/issue
2023-06-05T14:56:46
2025-04-01T06:40:13.039989
{ "authors": [ "markerikson", "pandaiolo" ], "repo": "reduxjs/react-redux", "url": "https://github.com/reduxjs/react-redux/issues/2024", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
438610644
Update ThreePrinciples.md in the state explanation read-only is totally missed, need focused few words on it. Strongly applies when programmer codes, suppose to write getter methods not setters, in simple words for beginners. Hi. Appreciate the PR, but I don't see a need to change the wording at this time. Thanks anyway!
gharchive/pull-request
2019-04-30T05:38:38
2025-04-01T06:40:13.041727
{ "authors": [ "kambleaa007", "markerikson" ], "repo": "reduxjs/redux", "url": "https://github.com/reduxjs/redux/pull/3416", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
490573105
Move the examples to a separate repo I'd like to move the examples out of the repo. They add a lot of time to the build process and generally don't change significantly. They also don't require the source copy of Redux, so it's not like they are a canary for changes we make on master. Moving them to their own home means it's more reasonable to accept new additions, which is something we've been hesitant to do for a long while (years!). I've set them up in a new repo: reduxjs/examples Is this a terrible idea? Anything we should do besides updating the docs and creating a new README in the examples path (which is temporarily gone at the moment). Also of note: Typical build time today: 3:30-4:10 This PR build time: 1:10 That's a 3-3.5x faster build time. Thoughts and concerns: The approach shown here and in the other repo appears to blow away all the accumulated Git history. Could we at least do some git filter-branch work to extract the examples folder history as the starting point of the new repo, so that we can maintain that history? We have a bunch of links in the docs, and there's probably also external links: Yeah, we'd need to do a bunch of fixing up for our docs. External links would result in a 404. Can we at least leave placeholder README files that point to the new location? What additional examples would you want to have in the future repo? Out of curiosity, what about the examples is slowing down the build process? There are so many of them and they all require separate npm ci commands. But why are we even building / testing them at all as part of the main lib? Where are they getting kicked off? https://github.com/reduxjs/redux/blob/master/.travis.yml#L5-L6 Oh. Well, at a minimum, we should remove those lines from Travis. No reason to be checking those on every commit. Redid the examples repo with history intact: https://github.com/reduxjs/examples I'm still really not seeing a point to this. If we're concerned about Travis build times for some reason, let's just configure Travis to not worry about the examples on every build or something. I tried to figure out how to conditionally build certain examples based on the changed files using Travis' environment variables, but unfortunately it seems difficult to do this consistently. For example TRAVIS_COMMIT_HASH doesn't only show the commits from the PR which would cause false positives where examples are build when they shouldn't be, and if multiple commits are pushed outside of a pull request TRAVIS_COMMIT would only show the latest commit which could cause examples to break without being rerun in CI. I think the safest way to make our builds faster is to remove the examples (like with this pull quest, but possibly replacing the existing files with links to the new repository) and move the example scripts to the other repository. Alternatively, we might be able to speed up our multiple package installations using monorepos with yarn workspaces or lerna, but I don't have experience with them. We never did move these out, and tbh I'd like to keep them in here. Closing.
gharchive/pull-request
2019-09-07T01:01:10
2025-04-01T06:40:13.050332
{ "authors": [ "markerikson", "nickmccurdy", "timdorr" ], "repo": "reduxjs/redux", "url": "https://github.com/reduxjs/redux/pull/3565", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
568106110
Improve the example in GettingStarted. name: "New/Updated Documentation Content" about: Adding a new docs page, or updating content in an existing docs page Hello, Thanks so much for this great project. I've just started leaning redux, and found a little bit of difficulty in the GettingStarted example. I hope you will like it and please let me know if there are rooms for improvement in my PR. PR Type update an existing page. Checklist [ ] Is there an existing issue for this PR? link issue here [x] Have the files been linted and formatted? What docs page is being added or updated? Section:Introduction Page:Getting Started with Redux For Updating Existing Content What updates should be made to the page? The import part is a bit difficult for newbies to setup correctly. Actually I failed to make this example work. $ node example.js (node:20108) ExperimentalWarning: The ESM module loader is experimental. file:///private/tmp/my-redux/example.js:1 import { createStore } from 'redux' ^^^^^^^^^^^ SyntaxError: The requested module 'redux' does not provide an export named 'createStore' at ModuleJob._instantiate (internal/modules/esm/module_job.js:92:21) at async ModuleJob.run (internal/modules/esm/module_job.js:107:20) at async Loader.import (internal/modules/esm/loader.js:164:24) I know it's not a problem in Redux, but ideally example code should work without effort. Do these updates change any of the assumptions or target audience? If so, how do they change? Target audience: a newbie to overall ecosystem of JavaScript (JS/npm/Redux) Assumptions: They can create an html and a js file by hand, and open the html by their browser. Expected Results: They'll feel more confident that they can move forward by the first success of the working example code. Thanks, but we're not looking to create a working example here per se, just some example code of how a Redux store might be written. If you click over to the Examples page in the next section, we do have a working example of this exact code.
gharchive/pull-request
2020-02-20T07:32:55
2025-04-01T06:40:13.056669
{ "authors": [ "DQNEO", "timdorr" ], "repo": "reduxjs/redux", "url": "https://github.com/reduxjs/redux/pull/3709", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
618598023
ReactJS - setState causes viewer to re-render every time When using Resium with ReactJS and adding an ImageLayer, it is causing the viewer to refresh every time any setState is called. How can I use setState to update certain parts of the page without re-rendering the map? Example below and screenshot of the issue ` class LiveMap extends React.PureComponent { viewerRef = React.createRef(); state = { name: "test" }; handleSelectedEntityChanged = (event) => { console.log("selected: " + event); this.setState({ ...this.state, name: this.state.name === "test" ? "testt" : "test" }); }; render() { const defaultImageryProvider = new UrlTemplateImageryProvider({ url : "https://cartodb-basemaps-{s}.global.ssl.fastly.net/dark_all/{z}/{x}/{y}.png" }); return ( <Viewer full ref={this.viewerRef} onSelectedEntityChange={this.handleSelectedEntityChanged}> <ImageryLayer imageryProvider={defaultImageryProvider} /> <Entity name={this.state.name} position={Cartesian3.fromDegrees(61, 130, 100)} point={{ pixelSize: 10 }} /> </Viewer> ); } }; ` You are constructing a new imagryProvider on every render. It looks like it does not depend on any props or state though, so you should be able to just build it once inside the constructor. I am getting similar functionality with useEffect. I have an API call in my useEffect function which gets called every 5 seconds. With the data from the API I can set some state variables using "setMyVar()". This then updates the state variables and redraws a point in the Viewer. The weird thing is, this updates resets the clock. I think this is because useEffect re-renders the view with every call. I am not sure if this is how this should work or if I am using state incorrectly with Resium. Try to move new UrlTemplateImageryProvider(...) outside the render function. Before: render() { const defaultImageryProvider = new UrlTemplateImageryProvider({ url : "https://cartodb-basemaps-{s}.global.ssl.fastly.net/dark_all/{z}/{x}/{y}.png" }); After: defaultImageryProvider = new UrlTemplateImageryProvider({ url : "https://cartodb-basemaps-{s}.global.ssl.fastly.net/dark_all/{z}/{x}/{y}.png" }); render() { // ... <ImageryLayer imageryProvider={this.defaultImageryProvider} /> I think I have the same issue. When I update variable in parent that holds Entities as children, they all get redrawn Same as you guys, when state of my component with Viewer is changed (with useState() or useEffect() hooks) - ImageLayer is reloaded. Is there some way how to fix this? imageProvider prop of ImageryLayer component is read-only and is not changeable because ImageProvider of ImageryLayer of Cesium API is also read-only. If the imageProvider is static, you can use outside const variable or useMemo hook as described in this page. If you want to change the imageProvider dynamically, you cannot avoid regenerating the ImageryLayer. If you want to completely suppress warnings, key props can be used: const imageProvider = useMemo(() => new UrlTemplateImageryProvider({ url }), [url]); return <ImageryLayer key={url} imageryProvider={imageProvider} />; We are also considering suppressing the warnings. Either way, please understand that the imageProvider is read-only.
gharchive/issue
2020-05-14T23:35:31
2025-04-01T06:40:13.070307
{ "authors": [ "Chingili", "benwiles1", "digennnj", "mhaque02", "rot1024", "vlkpa" ], "repo": "reearth/resium", "url": "https://github.com/reearth/resium/issues/372", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
143764757
Suppressed the result of expressions at the end of test blocks Fix #34, #47 Converted the optional expr at the end of test blocks to a statement so that test blocks can end with an unsuppressed expression (including loop statements). A bit of a mess (lots of cloning) because of ownership issues. Suggestions are welcome. @reem Thoughts? Hey there, I volunteered to help out @reem with this crate. The code is probably not the prettiest ever as you said but it would be great to have this fixed so I'd be willing to merge this. Could you also rebase this on the latest master so that we are completely sure that it doesn't break any of the existing tests? Thanks for the changes. 👍
gharchive/pull-request
2016-03-27T00:50:29
2025-04-01T06:40:13.105906
{ "authors": [ "cramertj", "ujh" ], "repo": "reem/stainless", "url": "https://github.com/reem/stainless/pull/54", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
184636201
Fix for rustc 1.14.0-nightly (f09420685 2016-10-20). Fixes #71. See https://github.com/rust-lang/rust/commit/029dceedb9719ae5dbdbf2c033c920017e3d786e. 👍 Thank you.
gharchive/pull-request
2016-10-22T15:24:00
2025-04-01T06:40:13.107325
{ "authors": [ "ujh", "utkarshkukreti" ], "repo": "reem/stainless", "url": "https://github.com/reem/stainless/pull/72", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
247791176
Rails5 adds rails5 compatibility via upstream merge of collectiveideas master. lgtm ??
gharchive/pull-request
2017-08-03T18:12:41
2025-04-01T06:40:13.108111
{ "authors": [ "meesterdude", "nhance" ], "repo": "reenhanced/audited", "url": "https://github.com/reenhanced/audited/pull/4", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
52580886
remote forms doesn't work after upload:complete event When you try to submit a form with remote: true jquery-ujs checks if it's had any non blank file inputs https://github.com/rails/jquery-ujs/blob/master/src/rails.js#L422, so the remote option is ignore and the form is submitted with a normal POST. My solution was to modify the if in this line https://github.com/elabs/refile/blob/master/app/assets/javascripts/refile.js#L37 to: if((xhr.status >= 200 && xhr.status < 300) || xhr.status === 304) { var id = input.getAttribute("data-id") || JSON.parse(xhr.responseText).id; input.previousSibling.value = id; input.removeAttribute("name"); input.value = ''; // <-- included this dispatchEvent("upload:success", xhr.responseText); } This make the input file empty, so for jquery-ujs all the file fields are blank and the form is submitted remotely. I will be happy to make a pull request if this approach is considered correctly. I'm the same problem =\ This gem https://github.com/JangoSteve/remotipart resolved the problem with upload file with remote: true Is there any way to get this to work properly? I tried @cerdiogenes code and it didn't work. @danieldocki I looked at the remotipart gem. I wouldn't use it. It is looking very long in the tooth and somewhat neglected. I'm paranoid about adding too much magic, especially when it involves adding new middleware to my stack. Seems like overkill anyway to work around this issue as it triggers submitting forms through a hidden iframe, which is not longer needed thanks to refile! Keep it simple. We just need some way to tell Rails UJS to not count refile handled file inputs when determining whether to abort the AJAX form submission. I poked around through the remotipart gem to see how it works. I found it relies on handling the ajax:aborted:file event fired by Rails UJS when it aborts due to non-empty file inputs. I was able to come up with the event handler below which tells Rails UJS to proceed with submitting the form via Ajax if all the non-empty file inputs are associated with refile. I used the data-reference attribute to detect the presence of refile handling. @jnicklas: is that a safe bet? $(document).on('ajax:aborted:file', 'form', function(){ var form = $(this); // if all non-blank file inputs are handled by refile, // then resume var nonBlankFileInputs = $.rails.nonBlankInputs(form, $.rails.fileInputSelector); if (nonBlankFileInputs.filter(':not([data-reference])').length === 0) { $.rails.handleRemote(form); return false; } else { return true; } }); @masonhale I think that's a fine workaround. Someone (not me ;)) should really patch rails_ujs to do the right thing though. @jnicklas I was thinking about how to patch rails_ujs, and it occurs to me that it needs some way to know which file input fields should be excluded from its check of non-blank file inputs. I would propose that it looks for some specific attribute on the input as a flag to indicate it should be ignored. Something like data-rails-ajax-ignored=true. This way an application author could manually add the attribute to exclude it from the file input check (and potentially the separate required field input check), and rails-friendly file upload libraries like refile could potentially add the ignore attribute automatically to help make things work automatically. Any thoughts? No, the reason for exclusion should be that the field is disabled. Disabled fields aren't submitted by browsers, rails_ujs shouldn't care that there's a disabled field with a file attached in the form. @jnicklas thanks for the clarification. I didn't realize Refile disabled fields and that would be a good method to filter file input fields. Question: when does this disabling happen? Is it possible the rails_ujs script could fire before Refile sets the field to disabled? @jnicklas Looking more closely at how the refile.js operates, I notice it removes the name of the input field after the file is processed here. I don't see where it sets the input file to disabled. Am I missing something? It makes me wonder if ignoring file inputs with no name attribute would work as well or better than ignoring disabled fields? (To be clear, I'm asking for clarification because I intend to submit a PR to fix rails_ujs. Thanks for the help). PS: Thanks also for all the work you've put into Refile. I think it is really great work and I am very thankful you've put in the effort and hard-won experience to make it. Thank you! @masonhale uuhh I was remembering it wrong. We do indeed remove the name to not submit the field. Browsers don't submit it, so rails_ujs should not accout for it, but it does. Ok, based on what I've learned by working on a Rails UJS patch, I've come up with a better/simpler monkey-patch to get Refile and Rails UJS working together. This avoids having to catch the aborted file event. Much cleaner. It simply tweaks the fileInputSelect UJS uses to make sure the input has a name. (Interestingly, Rails UJS already has a similar check in place for required fields). // run this anytime after rails / jquery-ujs is loaded $.rails.fileInputSelector = 'input[name][type=file]'; @jnicklas actually does Rails UJS not include disabled inputs since v. 1.1.0 (https://github.com/rails/jquery-ujs/pull/430). Could we patch refile to disable the input instead of just removing the name attribute? @jokklan no this does not work. Disabling the input not only has a visual effect we do not want, but it also prevents the user from selecting another file for upload. It's really a simple fix in Rails UJS, which they should make anyway, someone should really submit a patch. It's not going to be me though, since I don't actually ever use rails_ujs ;) I just submitted a pull request to jquery-ujs for this issue: https://github.com/rails/jquery-ujs/pull/464 I meant to do it before the holidays, but got busy/distracted and didn't get back to it. You're a hero @masonhale! :smile: @masonhale :metal: @danieldocki @jnicklas Can you do a favor and comment on the jquery-ujs pull request I opened to draw more attention to it? https://github.com/rails/jquery-ujs/pull/464 @masonhale I added a comment. It's not been that long though, OSS moves slowly ;) They'll get to it eventually.
gharchive/issue
2014-12-21T00:09:13
2025-04-01T06:40:13.131672
{ "authors": [ "JoshuaNovak919", "cerdiogenes", "danieldocki", "jnicklas", "jokklan", "masonhale" ], "repo": "refile/refile", "url": "https://github.com/refile/refile/issues/58", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1244285152
Add collapse-duplicated-reviews feature Fixes #5394 Test URLs https://github.com/refined-github/sandbox/pull/10#pullrequestreview-958560681 Screenshot https://user-images.githubusercontent.com/46634000/169702068-8bfe4f63-09b6-45c1-b2a1-0dee4d717f86.mp4 I'd rather collapse the whole thread into the title (so it's visible) and link the title to the real comment. With the current PR, reacting to a comment means uncollapsing it and then clicking the permalink. Let's skip one step. I'd rather collapse the whole thread into the title By "title", do you mean the text next to the "review event"? But what happens if the review is made up of two duplicated comments, each responding to a different review thread? By "title", do you mean the text next to the "review event"? I mean this: instead of this: The latter is not actionable and it's out of context. That comment should not appear there, but the review should still keep the link to the real review comment, as shown in the first screenshot. By title I mean filename This is already what happens, see video above. We could avoid showing the "Show"/"Hide" text & icon, but since we hide something GitHub shows by default we should probably make it obvious it can be expanded. Unless you want to drop the review thread completely? and link the title to the real comment +1 @fregante wdyt about this: Instead of using the file name we make it clear this is a link to another comment (plus it's one less select because we don't need to grab the <summary>). Closing until you have time to pick it up. This feature is just: observe('collapsedReview', review => { if (isDuplicate(review)) { const link = select('a', review) details.replaceWith(link) link.classList.add(...details.classList, ...otherClasses) } })
gharchive/pull-request
2022-05-22T15:12:23
2025-04-01T06:40:13.138920
{ "authors": [ "cheap-glitch", "fregante" ], "repo": "refined-github/refined-github", "url": "https://github.com/refined-github/refined-github/pull/5656", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2729750494
🛑 SmartProtection is down In 1be32fd, SmartProtection (https://smartprotection.com) was down: HTTP code: 404 Response time: 297 ms Resolved: SmartProtection is back up in 7003f29 after 10 minutes.
gharchive/issue
2024-12-10T10:47:22
2025-04-01T06:40:13.178737
{ "authors": [ "reinaldoleon" ], "repo": "reinaldoleon/monitoring", "url": "https://github.com/reinaldoleon/monitoring/issues/333", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
560126621
ppo save expert demo hi, how am i supposed to save expert demo in ppo main? PPO is a method of reinforcement learning. However app, maxent and gail are all inverse reinforcement learning method. Due to the emergence of policy-based inverse reinforcement learning algorithms, you can use PPO with any inverse reinforcement learning algorithm to complete the training. References: Ng A Y, Russell S J. Algorithms for inverse reinforcement learning[C]//Icml. 2000, 1: 2. Ho J, Gupta J, Ermon S. Model-free imitation learning with policy optimization[C]//International Conference on Machine Learning. PMLR, 2016: 2760-2769.
gharchive/issue
2020-02-05T04:03:47
2025-04-01T06:40:13.180485
{ "authors": [ "francisduan", "gitouni" ], "repo": "reinforcement-learning-kr/lets-do-irl", "url": "https://github.com/reinforcement-learning-kr/lets-do-irl/issues/6", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1896122711
🛑 BINUSMAYA Praktikum (SLC Bluejack) is down In 49bbdda, BINUSMAYA Praktikum (SLC Bluejack) (https://bluejack.binus.ac.id/binusmayalab/) was down: HTTP code: 502 Response time: 1224 ms Resolved: BINUSMAYA Praktikum (SLC Bluejack) is back up in 1070ff7.
gharchive/issue
2023-09-14T09:13:05
2025-04-01T06:40:13.183030
{ "authors": [ "1010bots" ], "repo": "reinhart1010/binusmayadown", "url": "https://github.com/reinhart1010/binusmayadown/issues/2921", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2286834682
🛑 GreatNusa.com is down In 12ab2d7, GreatNusa.com (https://greatnusa.com) was down: HTTP code: 403 Response time: 120 ms Resolved: GreatNusa.com is back up in 76719e8.
gharchive/issue
2024-05-09T03:46:58
2025-04-01T06:40:13.186333
{ "authors": [ "1010bots" ], "repo": "reinhart1010/binusmayadown", "url": "https://github.com/reinhart1010/binusmayadown/issues/6058", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2329251001
🛑 GreatNusa.com is down In 3f36256, GreatNusa.com (https://greatnusa.com) was down: HTTP code: 403 Response time: 118 ms Resolved: GreatNusa.com is back up in 114ac4a.
gharchive/issue
2024-06-01T17:26:21
2025-04-01T06:40:13.189357
{ "authors": [ "1010bots" ], "repo": "reinhart1010/binusmayadown", "url": "https://github.com/reinhart1010/binusmayadown/issues/6788", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1683207136
Client: Data export not respecting archived parameter It seems the client.export_study_data endpoint does not adequately respect the archived parameter, returning archived reports when it is set to False. For now this means we can't filter on archived at the API, leading to a problem when exporting study data. Currently we extract all data and then filter it ourselves. Call has been opened. # DATA ANALYSIS def export_to_dataframe(self, archived=False) -> dict: """Exports all data from a study into a dict of dataframes for statistical analysis.""" # TODO: change this to the correct archived, # fails now because the parameter does not seem to be handled correctly server side self.map_data(archived=True) dataframes = { "Study": self.__export_study_data(archived), "Surveys": self.__export_survey_data(archived), "Reports": self.__export_report_data(archived), } return dataframes This issue will be fixed with the next CDMS maintenance release (2023.2.2.0), which is planned for end of June.
gharchive/issue
2023-04-25T13:44:34
2025-04-01T06:40:13.191403
{ "authors": [ "reiniervlinschoten", "slknijnenburg" ], "repo": "reiniervlinschoten/castoredc_api", "url": "https://github.com/reiniervlinschoten/castoredc_api/issues/89", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1119934882
Decide what to do about auto-updates See: https://github.com/actions/runner/issues/1396#issuecomment-983783264 also related to #16 This was transitively addressed via latest "version" and this PR: #39. Closing.
gharchive/issue
2022-01-31T21:09:29
2025-04-01T06:40:13.203992
{ "authors": [ "ravwojdyla" ], "repo": "related-sciences/gce-github-runner", "url": "https://github.com/related-sciences/gce-github-runner/issues/17", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1265857014
Update serializers to include new Locality data After adding new models (Locality, LocalityStack and BibItemLocality) (see https://github.com/relaton/relaton-py/issues/18) to accommodate more DOI data, we want to include them when serialising a BibliographiItem. @stefanomunarini, so doi2ietf (doilit) used to serialize this into <seriesInfo>—but it suddenly turned out to be wrong (violates XML specification). I stated the task as “reimplement the same handling as in doi2ietf/doilit”, so you have done what was requested—the mistake was at planning stage. Apparently, we are supposed to use <refcontent>, based on: Jean’s comment https://github.com/ietf-ribose/bibxml-service/issues/228#issuecomment-1177699864 The spec on <refcontent> (cc @ronaldtse) Note that <refcontent> data model is apparently plain text, so in the serializer we might need to format the entire locality hierarchy as one human-readable string. It’s up to you how to attack this, but some considerations: It’s probably worth encapsulating it in a separate, testable function that takes locality info and outputs a string (of course, if warranted, the logic can also be split further into smaller functions). This function doesn’t have to cover all possible locality data from the start. We can increase coverage as we go, using existing samples as minimum baseline. If we can write this nicely without extra requirements, that’d be ideal. However, if this is a problem (f-strings don’t have conditionals, after all…), to prevent this logic from becoming a mess it may be worth adding a lightweight templating library to package requirements. I don’t really know a good one (Jinja2 seems overkill?), so let me know if you want to do this.
gharchive/issue
2022-06-09T09:31:42
2025-04-01T06:40:13.209169
{ "authors": [ "stefanomunarini", "strogonoff" ], "repo": "relaton/relaton-py", "url": "https://github.com/relaton/relaton-py/issues/22", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1817566663
two big bugs! no support for "break" and inline loop Describe the bug* I am confront with two bugs, key word "break" in "for loop" and inline loop are not supported in hot reload run. It output the error info: "Process finished with exit code -1073741819 (0xC0000005)". here is the test code: site=[0,2,2,2] site=[ii for ii in site if ii!=0] print('test') for ii in site: if ii==2: break print('finished') when in debug mode, this code can run well in the first time, when something changed, such as print('somethin others'), it will quit directly when hot reload is triggered, and output the upper error code. Desktop or remote (please complete the following information):** OS: [Windows10, debian9] Reloadium package version: [1.2] Editor: [PyCharm] Python Version: [3.9 and 3.10] Python Architecture: [ 64bit] Run mode: [Debug] Additional context Add any other context about the problem here. Hi @jialay Good catch! Thanks for opening this issue. I've managed to fix the break issue but cannot reproduce the "inline loop" problem. Do you mean the line comprehension in the second line? Also could you provide the full version of your python interpreter (eg 3.9.8 )? Hi, dkrystki, thank you for the great tool reloadium! yes, the inline loop is just show as the second line. I tried it with python 3.9.12( debian 9, remote debug mode), and 3.10.0(windows 10, local mode), both of the them can't support hot reload。 such as the following code: site=[0,2,2,2] site=[ii for ii in site if ii!=0] print('test') print('finished') debug with a breakpoint at line "print('finished')", any change of the code will interrupte the debug.
gharchive/issue
2023-07-24T04:07:27
2025-04-01T06:40:13.222783
{ "authors": [ "dkrystki", "jialay" ], "repo": "reloadware/reloadium", "url": "https://github.com/reloadware/reloadium/issues/158", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1079511351
[BUG] ScrollView组件里slot无效 bug 描述 ScrollView组件slot无效 期望结果 如下截图所示 复现代码 可复现问题的代码仓库 关键代码 <View className="placeholder">placeholder</View> <ScrollView scrollY refresherEnabled refresherDefaultStyle="none"> <View className="refresher" slot="refresher"> refresher </View> <View className="body">body</View> </ScrollView> 版本信息: remax 版本: 2.9.1 开发环境 win10,微信开发者工具 其他信息 微信原生可以 remax不可以 这个有进展吗,遇到同样的问题 这个有进展吗,遇到同样的问题 没有,当时我是打算搞个下拉刷新的loading。最后放弃了,直接用小程序原生的整页下拉刷新了。 bug 描述 ScrollView组件slot无效 期望结果 如下截图所示 复现代码 可复现问题的代码仓库 关键代码 <View className="placeholder">placeholder</View> <ScrollView scrollY refresherEnabled refresherDefaultStyle="none"> <View className="refresher" slot="refresher"> refresher </View> <View className="body">body</View> </ScrollView> 版本信息: remax 版本: 2.9.1 开发环境 win10,微信开发者工具 其他信息 微信原生可以 remax不可以 目前不支持 slot 属性
gharchive/issue
2021-12-14T09:25:11
2025-04-01T06:40:13.242072
{ "authors": [ "fhyoga", "ramen-curator", "watsonhaw5566" ], "repo": "remaxjs/remax", "url": "https://github.com/remaxjs/remax/issues/1787", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
527875225
Fix/missing macros fix https://github.com/remaxjs/remax/issues/382#issuecomment-558000230 修复 babel-plugin-macros 找不到的错误 要改一下,要从开发者的角度去写 修复 babel-plugin-macros 找不到的错误 要改一下,要从开发者的角度去写 康康
gharchive/pull-request
2019-11-25T06:23:18
2025-04-01T06:40:13.244454
{ "authors": [ "Darmody", "yesmeck" ], "repo": "remaxjs/remax", "url": "https://github.com/remaxjs/remax/pull/406", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1977070674
Create an example of an app with nested routes I'm trying to organize my routes using route folders, but I can't seem to get it to work. I came to the examples repo to find a good example, but none of the examples show how to use route folders. Could someone who knows how Remix route folders works create an example? Thanks so much. Hi @elliotaplant, This is a topic extensively covered by the documentation: https://remix.run/docs/en/main/file-conventions/routes#nested-routes I suggest opening a Q&A conversation in the remix repository if you need guidance, showing us how you setup things and what's not working for you? I'd be happy to help you!
gharchive/issue
2023-11-04T00:16:27
2025-04-01T06:40:13.251334
{ "authors": [ "elliotaplant", "machour" ], "repo": "remix-run/examples", "url": "https://github.com/remix-run/examples/issues/376", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1062977694
chore: fix ESLint errors & warnings Fixes #625 Fixes #805 Thanks!
gharchive/pull-request
2021-11-24T22:51:02
2025-04-01T06:40:13.270038
{ "authors": [ "MichaelDeBoey", "kentcdodds" ], "repo": "remix-run/remix", "url": "https://github.com/remix-run/remix/pull/611", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }