id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
820553410
|
build(typescript): update typescript dependencie to fix npx tsc issue
reference:
https://stackoverflow.com/questions/45790182/typescript-error-ts5014-unexpected-token-u-in-json-at-position-0
#186
This don't fix #186 for me
|
gharchive/pull-request
| 2021-03-03T02:08:07 |
2025-04-01T06:40:11.920335
|
{
"authors": [
"prescindivel"
],
"repo": "react-native-text-input-mask/react-native-text-input-mask",
"url": "https://github.com/react-native-text-input-mask/react-native-text-input-mask/pull/187",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1211930608
|
Minimum React Native version support
I would like to only support RN versions from the past year. The thinking is that if you are not upgrading your RN app for more than a year, you are also probably not upgrading core dependencies like this one. This would put us at RN v0.64.
This means publishing patches to older major versions if the latest has more recent requirements. For example, if we decide that v7 only supports RN 0.68 with new architecture, we will continue to support v6 for RN versions up to a year old.
Thoughts?
I agree it can be problematic to maintain old react-native version.
In all RN modules I saw, they said they don't support of version ...
Now with turbo module and new architecture, I am not sure of the exact impact in terms of dependancies.
maybe @douglowder can give some clues
react-native-maps already does something similar, even more strict actually: https://github.com/react-native-maps/react-native-maps#compatibility
|
gharchive/issue
| 2022-04-22T07:22:58 |
2025-04-01T06:40:11.932136
|
{
"authors": [
"Adnan-Bacic",
"freeboub",
"hueniverse"
],
"repo": "react-native-video/react-native-video",
"url": "https://github.com/react-native-video/react-native-video/issues/2655",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1339133662
|
not displaying properly on android
Bug
##Android
Environment info
react-native version
"react-native": "^0.64.2",
"react-native-video": "^5.2.0",
taking time to display video in flatlist with images
to be followed in: https://github.com/react-native-video/react-native-video/issues/2668
|
gharchive/issue
| 2022-08-15T15:19:55 |
2025-04-01T06:40:11.934478
|
{
"authors": [
"freeboub",
"ubaid-wp"
],
"repo": "react-native-video/react-native-video",
"url": "https://github.com/react-native-video/react-native-video/issues/2814",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1841114174
|
Stopping momentum scroll triggers click event on Android
#863
Same issue, but not fixed yet
react-native version: 0.71.6
react-native-webview version: 11.26.0
I just tried to reproduce this and I see exactly the same behavior in Chrome.
If you see different behavior between the WebView and Chrome, please share a sample repro.
|
gharchive/issue
| 2023-08-08T11:22:53 |
2025-04-01T06:40:11.936151
|
{
"authors": [
"TheAlmightyBob",
"nj0034"
],
"repo": "react-native-webview/react-native-webview",
"url": "https://github.com/react-native-webview/react-native-webview/issues/3090",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
386340499
|
You can't navigate back from last tab to first on iOS, if the tab has auto width
Current Behavior
When I give a width: 'auto' or width: null, flex: 1 to the tabStyle, I can't navigate back from last page if there is not enough tab to overflow the screen.
Expected Behavior
Either to be able to scroll the tabbar or to not move the selected tab item to the left side of the screen like on android.
How to reproduce
In the demo, on iOS select the Settings2 menu and try to navigate back to Home menu.
A device which is big enough either on portrait or landscape to show all tabs.
Your Environment
software
version
react-navigation
2.18.2
react-native
0.57
node
10.13.0
npm or yarn
yarn
@satya164 - not sure what the expected behavior is for top tabs or if this is supported. @Xyzor it might be helpful if you share a mockup or description of what you're trying to accomplish
My main goal is to use flex: 1 on tabStyle, instead of fixed width, because with fixed width the tab's label either break into new lines or some tabs are unnecessarily wide.
On android the demo is working because when i select a tab, it doesn't move to the left.
The width depends on the width of the tab bar. Dynamic width for the tab items is not supported because it's every complicated to implement. I'm not sure what you're trying to achieve, but you probably need to customize the tab bar itself or use a custom tab bar.
@satya164 Thanks for the info.
|
gharchive/issue
| 2018-11-29T15:06:01 |
2025-04-01T06:40:11.941488
|
{
"authors": [
"Xyzor",
"brentvatne",
"satya164"
],
"repo": "react-navigation/react-navigation-tabs",
"url": "https://github.com/react-navigation/react-navigation-tabs/issues/73",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
585788576
|
Direct link not building hierarchy as desired
Context/flow
If it helps, this is what I'm trying to accomplish:
Venues shows a list of venues
Clicking on one of the venues takes you to VenueDetails
At the venue you can add stuf to your cart, and open up your cart, which takes you to OrderDetails
I need a way for the url to remember I took this route so I can go back down the chain in case the user refreshes or opens the link in a new tab.
Description
I've set up my linking as follows:
export default function (containerRef) {
return useLinking(containerRef, {
prefixes: [Linking.makeUrl('/')],
config: {
Root: {
initialRouteName: 'Venues',
path: '',
screens: {
Venues: '',
VenueDetails: {
path: 'venue/:slug',
parse: { slug: String },
},
OrderDetails: {
path: 'venue/:slug/order-overview',
parse: { slug: String },
},
}
}
}
})
}
What I want is to open OrderDetails when I navigate to /venue/some-slug/order-overview.
If I revert the order in the useLinking object (so: OrderDetails, VenueDetails and then Venue) this works, but it doesn't see that VenueDetails is the previous page.
How can I tell react-navigation/linking that going "back" from OrderDetails means it should go to VenueDetails? And that going back from VenueDetails means it should go to Venue? Can I define a parent somehow? Maybe I need to put it in the url somehow?
Note: it works fine when starting at the home page and doing the normal navigation flow. It's only when I open up a nested page directly that this fails.
Is this repo still being worked on by the way? The last release was a while ago. I'd like to help but I don't know where to start.
I believe this repo deprecated and moved on react-navigation
this file is for web
https://github.com/react-navigation/react-navigation/blob/master/packages/native/src/useLinking.tsx
|
gharchive/issue
| 2020-03-22T19:03:37 |
2025-04-01T06:40:11.992223
|
{
"authors": [
"RWOverdijk",
"soroushm"
],
"repo": "react-navigation/web",
"url": "https://github.com/react-navigation/web/issues/50",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1662592497
|
PrismLight using results in regular [object Object] output (but only in production)
Describe the bug
I'm using import { PrismLight as SyntaxHighlighter } from 'react-syntax-highlighter'; which results in regular [object Object] [object Object] ... output (around 7 of 10 refreshs of my webpage) for my code blocks (but only in production). Notice: import { Prism as SyntaxHighlighter } from 'react-syntax-highlighter'; has the same problem. Every 4th to 5th refresh, it's working correctly, otherwise it's rendering [object Object] ...
To Reproduce
Steps to reproduce the behavior: This is my implementation, notice that the processedCode output of the process function is always correct (also in production). But even without this process function, the problem persists.
import { PrismLight as SyntaxHighlighter } from 'react-syntax-highlighter';
import tsx from 'react-syntax-highlighter/dist/esm/languages/prism/tsx';
import typescript from 'react-syntax-highlighter/dist/esm/languages/prism/typescript';
import { vs, vscDarkPlus } from 'react-syntax-highlighter/dist/esm/styles/prism';
SyntaxHighlighter.registerLanguage('tsx', tsx);
SyntaxHighlighter.registerLanguage('typescript', typescript);
interface ICodeBlock {
code: string;
language: 'typescript' | 'tsx';
}
const process = (code = '') => {
let skippedLeadingEmptyLines = false;
let lastNonEmptyLineIndex = 0;
let minRawStringIndentation = Number.MAX_SAFE_INTEGER;
let numberOfRemovedLines = 0;
const processNonEmptyLine = (line: string, index: number) => {
// keep track of the index of the last non-empty line
lastNonEmptyLineIndex = index - numberOfRemovedLines;
// determine the minimum indentation level
minRawStringIndentation = Math.min(minRawStringIndentation, Math.max(0, line.search(/\S/)));
// return the processed line
return [line.trimEnd()];
};
// split code into lines
const codeLines = code.split('\n');
// remove empty lines, and process non-empty lines
const nonEmptyLinesAtStart = codeLines.flatMap((line, index) => {
if (!skippedLeadingEmptyLines) {
if (line.match(/^\s*$/)) {
numberOfRemovedLines += 1;
return [];
}
skippedLeadingEmptyLines = true;
return processNonEmptyLine(line, index);
}
if (line.match(/^\s*$/)) return [''];
return processNonEmptyLine(line, index);
});
const nonEmptyLinesStartAndEnd = nonEmptyLinesAtStart.slice(0, lastNonEmptyLineIndex + 1);
// If there are no non-empty lines, return an empty string
if (nonEmptyLinesStartAndEnd.length === 0) return '';
const nonRawStringIndentationLines =
minRawStringIndentation !== 0
? nonEmptyLinesStartAndEnd.map((line) => line.substring(minRawStringIndentation))
: nonEmptyLinesStartAndEnd;
return nonRawStringIndentationLines.join('\n');
};
export const CodeBlock = ({ code, language }: ICodeBlock) => {
const { isLight } = useThemeContext();
const processedCode = process(code);
const theme = isLight ? vs : vscDarkPlus;
return (
<pre className={classes.pre}>
<SyntaxHighlighter language={language} style={theme}>
{processedCode}
</SyntaxHighlighter>
</pre>
);
};
It's working when I do a bad workaround to force reloading of the component:
// ...
useEffect(() => {
const timerId = setTimeout(() => setIsReloaded(true), 0);
return () => clearTimeout(timerId);
}, []);
return isReloaded ? (
<pre className={classes.pre}>
<SyntaxHighlighter language={language} style={theme}>
{processedCode}
</SyntaxHighlighter>
</pre>
) : null;
Expected behavior
Output my code string (it's a simple code string of a React Component) instead of [object Object] in some cases. Oddly enough in my dev environment it's always working fine, only in production is the rendering issue with [object Object]
Screenshots
Desktop (please complete the following information):
Browser firefox
Version 111.0.1 (64-Bit)
I'm getting this issue on a non-local environment as well
We see this same issue intermittently with PrismAsyncLight as well.
For others with this issue, using @fenkx's fork fixes the issue, the easiest way to switch is just change this line in your package.json:
"react-syntax-highlighter": "npm:@fengkx/react-syntax-highlighter@15.6.1"
I have the same issue but I guess this package is not maintained anymore 😕
I had the same problem, but the above scenario didn't solve my problem.
+1
May I add my opinion here? I'm experiencing the same issue. Hopefully, it has been resolved, but I want to share this for anyone who might be facing the same problem.
I think the following procedure is causing the issue:
Open a page with target="_blank" (in this case, the user only opens the tab without switching to it).
Once the tab is loaded and the user switches to it, the code is displayed as [Object object].
Therefore, I have made it re-render when the user switches tabs.
Below is a part of the code I wrote.
function CodeBlock({/* ... */}) {
const [codeBlock, setCodeBlock] = useState<ReactElement | null>(null);
useEffect(() => {
renderCodeBlock();
document.addEventListener('visibilitychange', renderCodeBlock);
return () => {
setCodeBlock(() => null);
document.removeEventListener('visibilitychange', renderCodeBlock);
};
}, []);
function renderCodeBlock() {
setCodeBlock(() => (
<Box
component={SyntaxHighlighter}
language={language}
showLineNumbers
style={atomDark}
>
{code.trim()}
</Box>
));
}
return (
<Box>
{codeBlock === null ? (
<CircularProgress />
) : (
codeBlock
)}
</Box>
);
}
Please note that this text has been translated and might contain some unnatural parts.
This happens to me as well, what can I do about this?
I'm experiencing this issue as well. I was able to work around it by adding this to my index.tsx file before I render my React app:
import { PrismAsyncLight as SyntaxHighlighter } from 'react-syntax-highlighter'
ReactDOM.createRoot(document.createElement('div')).render(
<SyntaxHighlighter language="" children={''} />
)
Can confirm, the workaround by jeremyckahn works with SSR and production. No more problems when using hydrateRoot.
|
gharchive/issue
| 2023-04-11T14:16:04 |
2025-04-01T06:40:12.007273
|
{
"authors": [
"C5H8NNaO4",
"FujiwaraChoki",
"MatchuPitchu",
"ShaofeiZi",
"bodinsamuel",
"humphd",
"jeremyckahn",
"jtsorlinis",
"kkn1125",
"lritter79"
],
"repo": "react-syntax-highlighter/react-syntax-highlighter",
"url": "https://github.com/react-syntax-highlighter/react-syntax-highlighter/issues/513",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
373564533
|
Vertical scrollbar hiding last column data
Describe the bug
Vertical scrollbar hiding last column data if we have horizontal bar too and last column width is reduced.
To Reproduce
Any react table with 3 or more columns, increase the width of middle column and reduce the width of last column , when you scroll right to end you will find one or two characters are hiding under vertical scroll bar.
Expected behavior
Vertical scrollbar should be post last column width should not hide the details showing in last column
Codesandbox!
you can use https://codesandbox.io/s/o5np5p0nrz codesand box and increase width of columns adding screen shot too.
Screenshots
Desktop (please complete the following information):
Windows / Crome
Browser - chrome
Version 68.0.3440.106
RT wasn't designed for inline-scrolling, hence the pagination, so unfortunately, there is no official support for this use-case. You're more than welcome to explore and ask in the forum though! https://spectrum.chat/react-table
@Mulli can you suggest how to do ? looking for few workarounds
|
gharchive/issue
| 2018-10-24T16:17:01 |
2025-04-01T06:40:12.012480
|
{
"authors": [
"Mukesh-Bhootra",
"tannerlinsley"
],
"repo": "react-tools/react-table",
"url": "https://github.com/react-tools/react-table/issues/1167",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
541303755
|
Errors after downloading JavaScript bundle
Warning: React.createElement: type is invalid -- expected a string (for built-in components) or a class/function (for composite components) but got: %s.%s%s, undefined, You likely forgot to export your component from the file it's defined in, or you might have mixed up default and named imports.
Check your code at Button.js:38.,
in Button (at Welcome.js:144)
in RCTView (at Block.js:155)
in Block (at Welcome.js:143)
in RCTView (at Block.js:155)
in Block (at Welcome.js:129)
in Welcome (at SceneView.js:9)
in SceneView (at StackViewLayout.tsx:899)
in RCTView (at StackViewLayout.tsx:898)
in RCTView (at StackViewLayout.tsx:897)
in RCTView (at createAnimatedComponent.js:151)
in AnimatedComponent (at StackViewCard.tsx:106)
in RCTView (at createAnimatedComponent.js:151)
in AnimatedComponent (at screens.native.js:71)
in Screen (at StackViewCard.tsx:93)
in Card (at createPointerEventsContainer.tsx:95)
in Container (at StackViewLayout.tsx:985)
in RCTView (at screens.native.js:101)
in ScreenContainer (at StackViewLayout.tsx:394)
in RCTView (at createAnimatedComponent.js:151)
in AnimatedComponent (at StackViewLayout.tsx:384)
in PanGestureHandler (at StackViewLayout.tsx:377)
in StackViewLayout (at withOrientation.js:30)
in withOrientation (at StackView.tsx:104)
in RCTView (at Transitioner.tsx:267)
in Transitioner (at StackView.tsx:41)
in StackView (at createNavigator.js:80)
in Navigator (at createKeyboardAwareNavigator.js:12)
in KeyboardAwareNavigator (at createAppContainer.js:430)
in NavigationContainer (at App.js:62)
in RCTView (at Block.js:155)
in Block (at App.js:61)
in App (at withExpoRoot.js:26)
in RootErrorBoundary (at withExpoRoot.js:25)
in ExpoRoot (at renderApplication.js:40)
in RCTView (at AppContainer.js:101)
in RCTView (at AppContainer.js:119)
in AppContainer (at renderApplication.js:39)
`
Also this one:
Element type is invalid: expected a string (for built-in components) or a class/function (for composite components) but got: undefined. You likely forgot to export your component from the file it's defined in, or you might have mixed up default and named imports.
Check the render method of `Button`.
@litehacker hey, do you version of Expo and React-Native you'r using?
any word on this bug? I got the same bug
"expo": "^36.0.0",
"react": "16.9.0",
@devbrunopaula I think this is because the current code base in based on an older expo version. I'll update the code base to the latest versions.
|
gharchive/issue
| 2019-12-21T09:00:58 |
2025-04-01T06:40:12.015331
|
{
"authors": [
"devbrunopaula",
"hetmann",
"litehacker"
],
"repo": "react-ui-kit/dribbble2react",
"url": "https://github.com/react-ui-kit/dribbble2react/issues/50",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1411397637
|
added logo component and make logo component spin
Fixes Issue
PR to fix issue #84
Changes proposed
added a logo component in the component directory and make logo spin
Requirements
Create a component named Logo in the components folder that implements the screenshots below.
It should accept a size as prop for the different sizes on the header and footer.
The logo should rotate similar to how it does in a fresh create react app installation as shown on this GIF.
Obtain the logo from the Figma design. Use THIS VIDEO as a guide
Acceptance Criteria
[x] The implementation should match the design.
Screenshots
Note to reviewers
The Logo spin animation was performed using the tailwind css animation class.
ok i will do that now
|
gharchive/pull-request
| 2022-10-17T11:21:42 |
2025-04-01T06:40:12.027727
|
{
"authors": [
"Ezekiel8807"
],
"repo": "reactdeveloperske/reactdevske-website",
"url": "https://github.com/reactdeveloperske/reactdevske-website/pull/89",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
267512645
|
Memory leak leads to heap overflow
14 Oct 15:06:57 - Received: :nova.esper.net PONG nova.esper.net :115367
<--- Last few GCs --->
[2539:0x4299a10] 1791041298 ms: Mark-sweep 1325.7 (1350.2) -> 1323.7 (1350.2) MB, 718.2 / 0.1 ms allocation failure GC in old space requested
[2539:0x4299a10] 1791042276 ms: Mark-sweep 1323.7 (1350.2) -> 1323.6 (1350.2) MB, 977.4 / 0.0 ms allocation failure GC in old space requested
[2539:0x4299a10] 1791043255 ms: Mark-sweep 1323.6 (1350.2) -> 1323.6 (1350.2) MB, 979.5 / 0.0 ms last resort
[2539:0x4299a10] 1791044225 ms: Mark-sweep 1323.6 (1350.2) -> 1323.3 (1350.2) MB, 969.6 / 0.0 ms last resort
<--- JS stacktrace --->
==== JS stack trace =========================================
Security context: 0x2a90f6a29891
2: heartbeat [/home/sfnet-discord-bot/node-v8.1.2-linux-x64/lib/node_modules/discord-irc/node_modules/discord.js/src/client/websocket/WebSocketConnection.js:~407] [pc=0x1141bd0b5fab](this=0x3bc473b2d581 <an EventEmitter with map 0x263359c3e3d1>,time=0x222777302311 )
3: arguments adaptor frame: 0->1
4: _onTimeout [/home/sfnet-discord-bot/node-v8.1.2-linux-x64/lib/node_modu...
FATAL ERROR: CALL_AND_RETRY_LAST Allocation failed - JavaScript heap out of memory
1: node::Abort() [node]
2: 0x13647ec [node]
3: v8::Utils::ReportOOMFailure(char const*, bool) [node]
4: v8::internal::V8::FatalProcessOutOfMemory(char const*, bool) [node]
5: v8::internal::Factory::NewUninitializedFixedArray(int) [node]
6: 0xe90ca3 [node]
7: v8::internal::Runtime_GrowArrayElements(int, v8::internal::Object**, v8::internal::Isolate*) [node]
8: 0x1141bc30437d
Aborted
I don't understand much of what's written here. From the stack trace, it looks like… the very last… call… was from discord.js? I don't think that helps track the problem down, though.
Do you have any more information on the problem? I've never noticed it, so I wonder if you're doing something weird in your setup (are you on a low-memory machine?).
No, I was running into the Node v8 default heap limit. There's not enough info here, but you can probably find it fairly quickly if you try pulling in heapdump from npm and take some snapshots as the bot is running:
https://blog.risingstack.com/finding-a-memory-leak-in-node-js/
I'll look into it a bit more at some point, but my fix for now was to just stick discord-irc into a systemd unit so it autorestarts when this happens.
Have there been any updates on this?
I'm having what appears to be a similar issue as seen in the following log:
error Command failed with exit code 134.
Aborted
10: 0x15a8c4bc3b67
9: 0x8e62c6 [/nodejs/bin/node]
8: node::StringBytes::Encode(v8::Isolate*, char const*, unsigned long, node::encoding, v8::Local<v8::Value>*) [/nodejs/bin/node]
7: v8::String::NewFromUtf8(v8::Isolate*, char const*, v8::NewStringType, int) [/nodejs/bin/node]
6: v8::internal::Factory::NewStringFromUtf8(v8::internal::Vector<char const>, v8::internal::PretenureFlag) [/nodejs/bin/node]
5: v8::internal::Factory::NewRawTwoByteString(int, v8::internal::PretenureFlag) [/nodejs/bin/node]
4: v8::internal::V8::FatalProcessOutOfMemory(char const*, bool) [/nodejs/bin/node]
3: v8::Utils::ReportOOMFailure(char const*, bool) [/nodejs/bin/node]
2: 0x8ccf9c [/nodejs/bin/node]
1: node::Abort() [/nodejs/bin/node]
FATAL ERROR: CALL_AND_RETRY_LAST Allocation failed - JavaScript heap out of memory
3: unpack [/app/node_modules/discord.js/src/client/websocket/WebSocketConnection.js:~170] [pc=0x15a8c49070bd](this=0x29ae8a484649 <EventEmitter map =…
2: arguments adaptor frame: 0->3
1: toString [buffer.js:~609] [pc=0x15a8c495b041](this=0x253f2b47cc99 <Uint8Array map = 0x32dd30a43a71>,encoding=0x282dd32822d1 <undefined>,start=0x282dd32822d1 <undefined>,end=0x282dd32822d1 <undefined>)
Security context: 0x2af5574a58b9 <JSObject>
==== JS stack trace =========================================
<— JS stacktrace —>
[27:0x3001070] 70137123 ms: Mark-sweep 1292.3 (1470.9) -> 1292.3 (1470.4) MB, 2666.6 / 0.0 ms last resort GC in old space requested
[27:0x3001070] 70134456 ms: Mark-sweep 1292.3 (1511.9) -> 1292.3 (1470.9) MB, 2829.5 / 0.0 ms last resort GC in old space requested
[27:0x3001070] 70131626 ms: Mark-sweep 1292.4 (1509.9) -> 1292.3 (1511.9) MB, 2763.2 / 0.0 ms allocation failure GC in old space requested
<— Last few GCs —>
Looking through the discord.js issues related to memory leaks [0] it seems like they cache a lot of things without actually clearing the cache. It doesn't seem like this is something they intend to fix: https://github.com/discordjs/discord.js/issues/1409#issuecomment-433648959
It's on the other hand possible to disable some of these caches [1] so maybe that's what we want to do?
[0] https://github.com/discordjs/discord.js/search?q=memory+leak&type=Issues
[1] https://github.com/discordjs/discord.js/pull/2883
@ekmartin Have you had any luck with disabling the cache?
It looks like the fork was abandoned due to unexpected behavior when disabling stores. https://github.com/discordjs/discord.js/pull/2883#issuecomment-433649130
Our Discord Bot's heap can grow to 2gigs in a couple of hours resulting in a crash. And we have under 50 Discord groups using the bot, so this type of rapid memory expansion seems odd, given there are some bots with pretty massive usage.
Thanks for pointing me to those threads. 👍
Is this still an issue? I'm looking into updating discord.js, but I don't know if they've made any improvements here.
|
gharchive/issue
| 2017-10-23T00:34:04 |
2025-04-01T06:40:12.040803
|
{
"authors": [
"Throne3d",
"ekmartin",
"pkovac",
"sethlesky"
],
"repo": "reactiflux/discord-irc",
"url": "https://github.com/reactiflux/discord-irc/issues/330",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
675634764
|
Upgrade discord.js to version 12
Also updates test stubs.
WIP:
[ ] Manually test most main cases (pings of users + roles, emoji use, multi-channeling, join messages)
[x] Investigate hanging tests (looks to be due to Bot Events)
Coverage increased (+0.03%) to 96.884% when pulling 6a986152eddf95f1c1d67ea67daf28e687fd3010 on upgrade/discordjs-12 into ee2d70fac12860b9a78043b7f121979929842493 on master.
|
gharchive/pull-request
| 2020-08-09T04:34:21 |
2025-04-01T06:40:12.043596
|
{
"authors": [
"Throne3d",
"coveralls"
],
"repo": "reactiflux/discord-irc",
"url": "https://github.com/reactiflux/discord-irc/pull/555",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
666064080
|
Add Maintenance Page
Feature Name
Add Coming Soon and Maintenance Pages
Summary description
Provide two a pages for stores which are either (1) not yet launched or (2) undergoing maintenance. In both cases provide the user an experience that allows the user to take some action which continues to grow the business (e.g. a launch email sign-up form).
Rationale for why this feature is necessary
It's standard fare for e-commerce sites. The launch page is expecially helpful for allowing store operators users to log in and view the store in production but before it's launched.
Expected use cases
Prospective customers may sign-up for email notification/newsletter prior to launch.
Store operators may log-in from Coming Soon if they've been given the credentials by the admin.
Admins may demo a production-ready product to stakeholders and gather feedback.
Shoppers receive a more professional experience when the site is down.
Here's an example Coming Soon page I mocked up using Chakra UI:
Thanks for taking the time to file this issue and preparing a mockup @balibebas!
The example storefront, as the name says, is just an example implementation meant to be customized.
I don't really consider such a screen a priority at all as it is easy to add one based on individual requirements, although I also wouldn't consider it to be an issue to showcase this functionality if someone wants to do a PR.
The issue I see is, that this would either involve some more work or give a false sense of security. To offer a proper maintenance mode, there would me adjustments on the API to maybe store this on the shop document, some admin ui to toggle this and then it could be used during getStaticProps in the storefront to present such a maintenance screen. But the API would actually need to take that into consideration and deny acess to most of the queries apart from shop for non-authorized users.
Also, as we use static generation were possible, still allowing storefront access to some authenticated users could be another challenge and probably be solved by using nextjs preview mode.
Honestly I wouln't expect what I just described to be implemented anytime soon, although I guess such a maintenance mode would be a welcome addition if someone wants to provide PRs for that.
Wonder what @mikemurray and @focusaurus think about this.
A really quick and simple solution instead could be to just add some MAINTENANCE_MODE env variable and accompanying pages + redirects, if others agree that this provides any value to the way they work with the storefront.
Similar to how the IOU payment example works for the API plugins I feel it would be good to keep a custom landing page of some sort present so this rather typical use case will always be considered as refactoring of the app and docs occur. Given not everyone needs a fancy landing page providing some sort of experience for API outages or errors would be nice. So having a maintenance page which appears when one clicks a product during an outage would be useful for most uses.
Closing due to lack of activity
|
gharchive/issue
| 2020-07-27T07:39:08 |
2025-04-01T06:40:12.052440
|
{
"authors": [
"balibebas",
"janus-reith"
],
"repo": "reactioncommerce/example-storefront",
"url": "https://github.com/reactioncommerce/example-storefront/issues/708",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
317004222
|
Add segment compatible analytics event tracking to the Product Grid
We need to start tracking ecommerce analytics events in our starterkit. Initially we'll be tracking all of the Segment V2 Ecommerce Events. You can read the docs for that here: https://segment.com/docs/spec/ecommerce/v2/
Start by add documentation for the events tracked to our event tracking/analytics documentation, explaining the properties tracked and any mapping decisions that were made.
For our current implementation of the Product Grid, we'll need to track two events:
Product List Viewed
Product Clicked
@spencern The product price may be a range, should I take the min or max or neither?
I'd take the min if a range exists
maybe create a constant variable for this that we can permit customization of in the .env file later as well?
|
gharchive/issue
| 2018-04-23T22:22:25 |
2025-04-01T06:40:12.064025
|
{
"authors": [
"mikemurray",
"spencern"
],
"repo": "reactioncommerce/reaction-next-starterkit",
"url": "https://github.com/reactioncommerce/reaction-next-starterkit/issues/56",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
128632318
|
Sync repo with latest from st2-chatops-aliases master
This PR syncs a few stray commits from st2-chatops-aliases master branch committed there after creating this cloned repo. This is the first step in shuttering the st2-chatops-aliases forked community repo and using this repo exclusively for our Stackstorm setups.
Boxed seems to have been using the use-existing-dyn-inventory branch on this repo. GMR has been using the st2-chatops-aliases repo master branch. With this sync up we can switch GMR over to this repo and delete the st2-chatops-aliases repo. We can decide how to handle Boxed's branch separately from this PR.
fwiw, the actions in Boxed branch will probably be moved into the boxed-infrastructure repo. they are highly customized in a way I'm not sure we can abstract out.
Looks good to me.
|
gharchive/pull-request
| 2016-01-25T20:35:52 |
2025-04-01T06:40:12.066341
|
{
"authors": [
"davekonopka",
"jmound",
"sairez"
],
"repo": "reactiveops/st2-pack-omnia",
"url": "https://github.com/reactiveops/st2-pack-omnia/pull/2",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
106141358
|
Refactor to ES6
Either simply write the entire source in ES6 and require node 4 in the package.json or implement transpilation with Bable.
This has been done in my recent clean up in https://github.com/reactivepod/fido/pull/4 via require('babel/register');.
|
gharchive/issue
| 2015-09-12T08:33:49 |
2025-04-01T06:40:12.067683
|
{
"authors": [
"hglattergotz",
"kahlil"
],
"repo": "reactivepod/fido",
"url": "https://github.com/reactivepod/fido/issues/1",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
252757039
|
fix: remove duplicate ComponentModelTypeConverter
What kind of change does this PR introduce? (Bug fix, feature, docs update, ...)
fix
What is the current behavior? (You can also link to an open issue here)
The ReactiveUI.Wpf project includes ComponentModelTypeConverter even though it's already in ReactiveUI.
What is the new behavior (if this is a feature change)?
Remove ComponentModelTypeConverter from ReactiveUI.WPF. Also, fix comments in the copy inside ReactiveUI because it actually has nothing to do with WPF.
What might this PR break?
Nothing realistic.
Please check if the PR fulfills these requirements
[ ] The commit follows our guidelines: https://github.com/reactiveui/reactiveui#contribute
[ ] Tests for the changes have been added (for bug fixes / features)
[ ] Docs have been added / updated (for bug fixes / features)
Other information:
@kentcb,
Thanks for having already signed the Contribution License Agreement. Your agreement was validated by .NET Foundation. We will now review your pull request.
Thanks,
.NET Foundation Pull Request Bot
Changes Unknown when pulling 88a668b9490a0b9214b726cc1dc5d17e7eda0df7 on kentcb:component-model-type-converter into ** on reactiveui:develop**.
|
gharchive/pull-request
| 2017-08-24T23:14:50 |
2025-04-01T06:40:12.076931
|
{
"authors": [
"coveralls",
"dnfclas",
"kentcb"
],
"repo": "reactiveui/ReactiveUI",
"url": "https://github.com/reactiveui/ReactiveUI/pull/1433",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
234117220
|
Farewell
For all the details check out my Medium post.
The tl;dr is basically that I’m overwhelmed and out of time so I’m turning everything over to @diasbruno as I step away from React Modal.
Thanks for everything and goodbye!
Thank you so much for your time managing react-modal.
@claydiffrient Awesome work!
|
gharchive/issue
| 2017-06-07T07:31:29 |
2025-04-01T06:40:12.078875
|
{
"authors": [
"claydiffrient",
"diasbruno",
"jamesjjk"
],
"repo": "reactjs/react-modal",
"url": "https://github.com/reactjs/react-modal/issues/396",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
211387267
|
[chore] added missing babel transformer plugin.
Running webpack --config webpack.config.js fails due to missing babel's spread transformer.
ERROR in ./examples/basic/app.js
Module build failed: SyntaxError: Unexpected token (16:20)
14 |
15 | openModal: function() {
> 16 | this.setState({ ...this.state, modalIsOpen: true });
| ^
17 | },
18 |
19 | closeModal: function() {
Acceptance Checklist:
[x] All commits have been squashed to one.
[x] The commit message follows the guidelines in CONTRIBUTING.md.
[x] Documentation (README.md) and examples have been updated as needed.
[x] If this is a code change, a spec testing the functionality has been added.
[x] If the commit message has [changed] or [removed], there is an upgrade path above.
This will no longer be necessary.
|
gharchive/pull-request
| 2017-03-02T13:13:21 |
2025-04-01T06:40:12.081759
|
{
"authors": [
"diasbruno"
],
"repo": "reactjs/react-modal",
"url": "https://github.com/reactjs/react-modal/pull/341",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1835022900
|
🛑 Bellshade API is down
In 25eff64, Bellshade API (https://api.bellshade.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Bellshade API is back up in f1b4c07.
|
gharchive/issue
| 2023-08-03T13:01:49 |
2025-04-01T06:40:12.084212
|
{
"authors": [
"reacto11mecha"
],
"repo": "reacto11mecha/uptime-status",
"url": "https://github.com/reacto11mecha/uptime-status/issues/486",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1587698785
|
[#2PlaysAMonth]: Image Gallery - Create a responsive image gallery by using the free Unsplash API
First thing, PLEASE READ THIS: ReactPlay Code Review Checklist
Description
The project contains use of Unsplash API, which is fetched using axios and then populated in the Photo Component.
Following things have been used in this project:
useState
useEffect
React Form
Component Refactoring
Material UI icon
Axios npm module
CSS Grid
Infinite scroll functionality(This feature is still not showing the react play app but is working on local file system. I am working on it)
The app is responsive and redirects to the individual images and profiles when clicked on it.
Fixes #910
Type of change
[ ] New feature (non-breaking change which adds functionality)
How Has This Been Tested?
To see working of the app repeat the following
Go to the app. Some images are loaded on the screen by default
Search for any keyword. The screen will update with new images.
Hover over any image, the cursor will turn to pointer.
On clicking any image, the page will re direct to that image.
On hovering, the profile name, likes will come in front of the image. On clicking the profile photo, the profile of the image contributor will open
The page on scrolling down should show more images.
Checklist:
[ ] I have performed a self-review of my own code
[ ] I have commented my code, particularly in hard-to-understand areas
[ ] I have made corresponding changes to the documentation
[ ] My changes generate no new warnings
[ ] Any dependent changes have been merged and published in downstream modules
Screenshots or example output
@siddhantsiddh15 , Thanks for the PR.
I would request you change the PR title to "[#2PlaysAMonth]: Image Gallery - Create a responsive image gallery by using the free Unsplash API"
and the second thing is to link your PR with the issue by adding the below line in the PR description.
Example:
Fixes #910
Greetings,
I have updated the pull request name and linked the issue with the issue number
Thanks
@siddhantsiddh15 , Please format and lint the code by following this guide.
https://github.com/reactplay/react-play#format-the-code
Greetings
Have formatted the code as per the guidelines as mentioned here https://github.com/reactplay/react-play#format-the-code and then pushed the code in the branch.
Thanks
@siddhantsiddh15 Kindly resolve the merge conflict
These errors are not in my edited files, what can I do to run the react play?
@siddhantsiddh15 Catch me up on Discord today to close it.
Link to the video as I was facing repeated issue in creating an account on Stack Stream.
Thankyou for merging my branch into main. Can I close this pull request now?
Inspect
I have kicked off a build. Please check if it is successful and test if things are fine.
@siddhantsiddh15 here is the preview build.. I see the changes are breaking styles. Please take a look
https://react-play-git-fork-siddhantsiddh15-unsplash-8f3010-reactplayio.vercel.app/
Have recorded the video here . The website is responsive and is working correctly.
@siddhantsiddh15 almost there..
please add a cover image
Also edit your play from localhost and add the stream recording.
Ping where when you done, will merge it.
@siddhantsiddh15 let us know when done
Have added the cover image.
Have added the cover image.
Thanks!
The cover image should be in KBs, please reduce the size.
Also confirm that your demo recording has been added to the play by editing it.
Greetings,
I have not added the demo recording in the Play. I have updated the size of the cover.png to 13 kb.
Regards
Greetings,
I have not added the demo recording in the Play. I have updated the size of the cover.png to 135 kb.
Regards
Can you please add the demo recording too.. then its al done.
Cannot do it. Having difficulty. Can we skip the recording portion?
Cannot do it. Having difficulty. Can we skip the recording portion?
Ok, no worries, no pressure. It's still valuable to get your work in.
I am just curious about what kind of issues you are facing. If you can post about it in our Discord, I may try the resolution.. In fact, you can add the recording after merge too...(before 5th March)
I am having slow internet connection due to the place I have travelled to, recording is a big file to upload.
Hey @siddhantsiddh15 , this play looks cool. I will be waiting for the video link to updated before merging it to production branch
Okay, will update it by tonight.
Hey @siddhantsiddh15 , this play looks cool. I will be waiting for the video link to updated before merging it to production branch
I have uploaded here the updated video of the play. I have checked the responsiveness and working of my play. The delay from my side was unwanted, I have uploaded the video as soon as I got good internet connection.
Well the video should be on https://stack-stream.com/
I can get this play merged if everything is ok however please record a stackstream video before EOD
@atapas you need to unblock in order to merge this PR
@siddhantsiddh15 please record demo on stackstream
|
gharchive/pull-request
| 2023-02-16T13:40:19 |
2025-04-01T06:40:12.162112
|
{
"authors": [
"Sachin-chaurasiya",
"atapas",
"koustov",
"siddhantsiddh15"
],
"repo": "reactplay/react-play",
"url": "https://github.com/reactplay/react-play/pull/974",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
209195477
|
Reactstrap and React-router 4.0.0-beta.6 - active
Issue description
components: Navlink
Steps to reproduce issue
I'm using Reactstrap and React-router 4.0.0-beta.6 in the educational project that is located on gitlab with custom domain - http://harazd.org/.
According to Reactstrap docs: that's the way I should use active navlink
import { NavLink } from 'reactstrap'
...
<NavLink href="#" active = true >Link< /NavLink>
According to React-router v4 docs:
import { NavLink } from 'react-router-dom'
...
<NavLink to="/about" activeClassName="active">About</NavLink>
So how should I do implement navlink active state and use react-router?
The simplest answer is that you can't use reactstrap navlink active prop when passing in react-router navlink. And that's fine because it's just 1 class active and react-router NavLink can take care of that with the activeClassName prop.
To use both, you'll need to rename one of those and use the tag prop in reactstrap NavLink.
import { NavLink } from 'reactstrap';
import { NavLink as RRNavLink } from 'react-router-dom';
<NavLink to="/about" activeClassName="active" tag={RRNavLink}>About</NavLink>
Thanks for a quick solution (You can paste your answer here: http://stackoverflow.com/questions/42372179/reactstrap-and-react-router-4-0-0-beta-6-active-navlink). And one more thing : ... - root path is always active for some reason..
The root path is always active because '/' is in every path. I am not too fimilar with react-router-dom and it's NavLink, it looks like it has an exact prop which may be what you are looking for.
Here's my router code, so I do use exact for root path.
` <BrowserRouter>
<div className='app'>
<Switch>
<Route exact path='/' component={Landing} />
<Route path='/products' component={Products} />
<Route path='/services' component={Services} />
<Route path='/price' component={Price} />
<Route path='/contacts' component={Contacts} />
<Route component={NoMatch} />
</Switch>
</div>
</BrowserRouter>`
@TheSharpieOne yes adding exact to root Navlink helped)
Previously you were using <NavLink> from react-router-dom. From that file, it looks like it creates the Route and the Link for you in one go.
react-router v4 has a ton of changes and I am not familiar with it enough to know how to use it compared to previous version so I can only offer some limited help with it.
@Y-Taras @TheSharpieOne can you give me example code? a have try it but doesnt work so far.
@dhanyn10 here's a link to my working example
https://gitlab.com/ytaras/pinobeton2/blob/master/js/Navigation.js
Hi everyone,
Is there anyway to use NavLink from reactstrap alone to style the active link?
Thanks
@Y-Taras I can confirm your solution works here too. Thanks so much - you saved me many hours of pain.
@piavgh if you just want to use reactstrap's NavLink alone (without react-router-dom) you can use the active prop:
import { NavLink } from 'reactstrap';
<NavLink to="/about" active>About</NavLink>
I think you can add something like :
<NavLink to="/about" active={window.location.hash === '/about'}>About</NavLink>
This works for me:
<NavItem active={window.location.pathname === "/thing"}>
<NavLink href="/thing">Things</NavLink>
</NavItem>
|
gharchive/issue
| 2017-02-21T16:19:21 |
2025-04-01T06:40:12.173408
|
{
"authors": [
"JonSilver",
"TheSharpieOne",
"Y-Taras",
"dhanyn10",
"eddywashere",
"nathfy",
"piavgh",
"softmixt"
],
"repo": "reactstrap/reactstrap",
"url": "https://github.com/reactstrap/reactstrap/issues/336",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
616470240
|
Proposal for environment context
This is a proposal for a new environment context which takes precedence of the config file.
We had the need for a context which can be configured for automation without touching files.
Happy to hear your feedback.
Hmm instead of going through all that trouble of creating and mapping config variables, you could also create an environment variable ESQUE_CONFIG_YAML that can hold the whole esque config as yaml string. That would be more dynamic, future proof and would only require minimal changes.
Hm. I get your point, in general. But that would only get rid of the need to write a file, not to create a YAML in general. :D
You wrote:
without touching files.
So I thought, that was your problem :smile:
Well, although I don't see how setting 15 environment variables is easier than creating a yaml string, I'm always a fan of "letting the user choose". So I'm not entirely against it.
I'm just afraid that we might have to rename, add and/or remove some of the variables while our config evolves. We do have a migration mechanism for file based configs but not for environment variables. Do you think a separate command to add a section to the config would help you?
Something like esque config add-context foo --bootstrap-servers broker1,broker2 --schema-registry registry ...
I think it's easier to keep a command stable than it is to keep the environment variables the same over time.
Guess I should have been clearer. Sorry for that. :D
I think you're right with your approach of not wanting to make it unnecessarily complicated and I think the idea of just adding another command is good. I'll see with what I can come up with over the weekend. 😊
|
gharchive/pull-request
| 2020-05-12T08:37:43 |
2025-04-01T06:40:12.200888
|
{
"authors": [
"hfjn",
"swenzel"
],
"repo": "real-digital/esque",
"url": "https://github.com/real-digital/esque/pull/148",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
98426054
|
C API
For easiest integration with Go (via cgo), Erlang, and Rust (via FFI), a C API (not C++) is needed. This is a simple wrapper around C++.
.NET users will also be very happy to use Aeron via P/Invoke! Do you have any ETA? A C API will instantly make Aeron as ubiquitous as ZMQ, native ports like #35 are not needed as much as a simple API.
This is still a plan. However, have not had time to get to it.
Any Update over the C API
Currently, a C API is planned for the media driver. And would like to extend that API to cover the Aeron client API. However, no timeline is set for that yet.
Closing until someone is willing to sponsor the work.
Now the C API are able to be happen:)
Preview in 1.28.X releases.
Should be feature complete from 1.30.0.
|
gharchive/issue
| 2015-07-31T16:34:09 |
2025-04-01T06:40:12.204299
|
{
"authors": [
"Hemanth-Eduru",
"buybackoff",
"lygstate",
"mjpt777",
"tmontgomery"
],
"repo": "real-logic/Aeron",
"url": "https://github.com/real-logic/Aeron/issues/154",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
68196476
|
Constrain Mailing list form in Sidebar
Is it possible to make the mailing list form only 300px wide, as the ads are?
There's a custom CSS field in the plugin settings. I entered
.et_bloom { width: 300px; }
but no result
Where did you enter that CSS? It's not a good idea to edit the stylesheets directly, as they are replaced with each new build of the theme. Changes that are not part of that build would be overwritten.
Do you need provisions for manually inserting css?
There's a CSS field within the plugin settings. In other words, when
designing the look of the plugin, there are a host of options for
customizing the colors and text fields. Then there is a "Custom CSS" field.
On Fri, Apr 17, 2015 at 3:29 PM, Dylan Kinnett notifications@github.com
wrote:
Where did you enter that CSS? It's not a good idea to edit the stylesheets
directly, as they are replaced with each new build of the theme. Changes
that are not part of that build would be overwritten.
Do you need provisions for manually inserting css?
—
Reply to this email directly or view it on GitHub
https://github.com/real-pants/Real-Pants/issues/81#issuecomment-94058258
.
Can this be closed? Dimensions look great now.
|
gharchive/issue
| 2015-04-13T21:17:02 |
2025-04-01T06:40:12.213406
|
{
"authors": [
"dylan-k",
"realpants"
],
"repo": "real-pants/Real-Pants",
"url": "https://github.com/real-pants/Real-Pants/issues/81",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2467695136
|
Crash on trying to run file on sparse3
Crash encountered when running file on sparse3 from ufs-big test image
Backtrace with RUST_LOG=trace and RUST_BACKTRACE=full: https://gist.github.com/casept/4f9f18b75458f4801746d3597e29c2d1
What system were you running it on? Can you please show us uname -a?
Fedora 40, uname:
Linux l13 6.10.3-200.fc40.x86_64 #1 SMP PREEMPT_DYNAMIC Mon Aug 5 14:30:00 UTC 2024 x86_64 GNU/Linux
This is now fixed with #63.
|
gharchive/issue
| 2024-08-15T09:05:42 |
2025-04-01T06:40:12.232282
|
{
"authors": [
"asomers",
"casept",
"realchonk"
],
"repo": "realchonk/fuse-ufs",
"url": "https://github.com/realchonk/fuse-ufs/issues/54",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2113528855
|
Fix Unit Tests and Remove Usage of Unsigned Byte
Remove UByte Array usages since it was experimental
Remove Roboelectric
Update Unit Tests with Android mocking Base64
Tests successfully run in the terminal
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
|
gharchive/pull-request
| 2024-02-01T21:36:28 |
2025-04-01T06:40:12.243363
|
{
"authors": [
"CLAassistant",
"realeyes-mike-patterson"
],
"repo": "realeyes-media/scte35-decoder-multiplatform",
"url": "https://github.com/realeyes-media/scte35-decoder-multiplatform/pull/10",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
242872720
|
Clarify error message
I just wanted to clarify the error message that shows up when the library can't find the .so/.dll
The previous error message implied that the error had something to do with the vulkan version (false, as far as I can tell) and also didn't really indicate that the library just may not have been in the loading path. It's a really small thing, but I thought I'd submit a PR anyways
Also, random question. Why is there so much code duplication between vulkan.template.py and init.py?
Hello @pipsqueaker.
Thank you very much for your pull request. You are right about the error message, your version is a lot better. Nevertheless, you don't need to update the __init__.py file. This file is automatically generated from the vulkan.template.py which is a jinja2 template. When you run the generator script, it will do it for you. So what you can do is only update the template file and then I will regenerate the module.
Thanks a lot for the contribution.
@realitix Alright, just messed with my history a bit so that only vulkan.template.py is edited
Thanks @pipsqueaker for the contribution !
|
gharchive/pull-request
| 2017-07-14T01:25:17 |
2025-04-01T06:40:12.249121
|
{
"authors": [
"pipsqueaker",
"realitix"
],
"repo": "realitix/vulkan",
"url": "https://github.com/realitix/vulkan/pull/12",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
331549234
|
Could you publish it in Scala 2.12 ?
jwt-scala_2.11 and jwt-scala_2.10 are already published, but we want to use your library in 2.12.
Could you publish this library in Scala 2.12 ?
forked and published for Scala 2.12 and 2.13
https://github.com/xuwei-k/jwt-scala/tree/v1.4.0
https://repo1.maven.org/maven2/com/github/xuwei-k/jwt-scala_2.12/
https://repo1.maven.org/maven2/com/github/xuwei-k/jwt-scala_2.13/
|
gharchive/issue
| 2018-06-12T11:34:22 |
2025-04-01T06:40:12.256024
|
{
"authors": [
"3tty0n",
"xuwei-k"
],
"repo": "reallylabs/jwt-scala",
"url": "https://github.com/reallylabs/jwt-scala/issues/16",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
241852671
|
Change object store to allow sync user auth URL to be modified
With this PR, getting a user with a different auth URL than the one it was originally given updates the URL, instead of throwing an exception.
This fixes the issue (commonly encountered during development) when a user is opened with one URL, persisted, the server is moved or SSL is enabled, and an attempt to log in the user with the new address is made.
This does not fix the corner case where two different ROSes have users with the same user ID, and the user wishes to be logged into both simultaneously, but this case wasn't supported before, nor is it supported by any of the other subsystem code.
Did we get clarity on whether identity is expected to be unique across different servers? If that's not guaranteed, is this change safe?
It's complicated.
The default implementation of identity is a UUID. However, there's the possibility of plugging a module into ROS that allows you to specify arbitrary identities instead of using the built-in system, so a conflict is theoretically possible. Even if it is, though, using Realms across multiple servers concurrently is something we intentionally de-emphasized when designing v1 of ROS, so I don't think it's a supported use case anyways.
It's something that Realm Browser inherently needs to support.
Back to the drawing board, then.
|
gharchive/pull-request
| 2017-07-10T21:16:16 |
2025-04-01T06:40:12.332956
|
{
"authors": [
"austinzheng",
"bdash"
],
"repo": "realm/realm-object-store",
"url": "https://github.com/realm/realm-object-store/pull/494",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
501618519
|
Switch to using ocaml-ci-scripts
This seems like a more reliable CI until we can use ocaml-ci.
This will allow us to easily add revdeps builds btw which could prove useful once we want to stabilize mdx's API and avoid breaking users' tests!
Opam has a make built-in variable that opam lint suggests we use instead of the raw "make".
Now that we use the ci-scripts, it's properly picked up and triggers a build failure if we don't have a compliant opam file.
Ok I've read https://github.com/realworldocaml/mdx/pull/185#discussion_r330673627 which answers my question. However this is quite fragile as it will break if you run this in a duniverse setting (with a toplevel call to dune runtest).
What do you mean? Do you mean that it will break if we vendor mdx in a duniverse? Aliases aren't resolved within the duniverse so it won't run the tests in this case.
I agree it is a bit fragile and I wish there was a way to tell dune about the ocaml-mdx -> ocaml-mdx-test binary dependency but unfortunately there isn't any atm. I tried a couple things and they ended up suffering from the same race condition.
Also when using mdx in duniverse mode, ocaml-mdx rule adds a (package mdx) dependency to the generated rules, thus solving the dependecny issue. We can't do that outside a duniverse because (package ...) deps only work for local packages, not opam ones.
There's one last thing I can try which is to add an explicit dependency on the install alias for all runtest aliases that rely on ocaml-mdx but that is a bit tedious and doesn't solve the issue in a generic way either but at least running dune runtest should work then. Let's merge that as it already fix the opam test runs and I'll try that solution in a separate PR!
Just had a quick look and the (alias install) dependency propably won't work as we generate most of the rules for runtest.
I just mean that we should report that issue upstream as it should be fixed properly at one point :-)
|
gharchive/pull-request
| 2019-10-02T17:14:40 |
2025-04-01T06:40:12.360460
|
{
"authors": [
"NathanReb",
"samoht"
],
"repo": "realworldocaml/mdx",
"url": "https://github.com/realworldocaml/mdx/pull/187",
"license": "ISC",
"license_type": "permissive",
"license_source": "github-api"
}
|
382682448
|
cli: reana-dev git-log
New command git-log showing information about commits.
Signed-off-by: Tibor Simko tibor.simko@cern.ch
Example output:
$ reana-dev git-log
[reana-workflow-engine-yadage] git log -n 5 --graph --decorate ...
* (upstream/pr/91) 34b1c8e tasks: add stop_workflow from reana_commons, Diego Rodriguez, 6 hours ago
* (HEAD -> master, tag: v0.4.0, upstream/pr/90, upstream/master, origin/master, origin/HEAD) 11305f8 release: v0.4.0, Dinos Kousidis, 2 weeks ago
* (upstream/pr/88) 720819b installation: upgrade REANA-Commons, Diego Rodriguez, 5 weeks ago
* 8f13434 publisher: update to kombu producer, Diego Rodriguez, 5 weeks ago
* (upstream/pr/89) 32251ab installation: bump reana-commons include pkg data, Diego Rodriguez, 6 weeks ago
[reana-workflow-engine-serial] git log -n 5 --graph --decorate ...
* (upstream/pr/52) e77af4e tasks: use stop_workflow from reana-commons, Diego Rodriguez, 5 hours ago
* ada286d tasks: introduce a stop_workflow task, Diego Rodriguez, 21 hours ago
* 522d3f1 config: disable task prefetching by process, Diego Rodriguez, 27 hours ago
* 0e114d1 config: disable broker pool, Diego Rodriguez, 28 hours ago
* 08f28d2 tasks: add revoke handler, Diego Rodriguez, 7 days ago
[reana-job-controller] git log -n 5 --graph --decorate ...
* (upstream/pr/98) d660142 api: return a dict on k8s_instantiate_job, Diego Rodriguez, 21 hours ago
* e4a18d1 global: use flask run to start application, Diego Rodriguez, 5 days ago
* c6b1578 tests: make application testable, Jan Okraska, 8 days ago
* 82fc7bc api: add delete job endpoint, Diego Rodriguez, 8 days ago
* (HEAD -> master, tag: v0.4.0, upstream/pr/97, upstream/master, origin/master, origin/HEAD) 527cc1b release: v0.4.0, Tibor Simko, 2 weeks ago
[pytest-reana] git log -n 5 --graph --decorate ...
* (upstream/pr/31) ff2f2df fixtures: renaming of operational parameters, Dinos Kousidis, 23 hours ago
* (HEAD -> master, upstream/master, origin/master, origin/HEAD) 4ceb1e7 release: v0.5.0.dev201811191, Diego Rodriguez, 24 hours ago
| * (upstream/pr/30) 70b1f6e release: v0.5.0.dev20181119.1, Diego Rodriguez, 24 hours ago
|/
* d0d1920 fixtures: fix expose yadage workflow fixture, Diego Rodriguez, 24 hours ago
* (upstream/pr/29) ce31f8a release: v0.5.0.dev20181119, Dinos Kousidis, 27 hours ago
[reana-workflow-monitor] git log -n 5 --graph --decorate ...
* (HEAD -> master, tag: v0.4.0, upstream/pr/26, upstream/master, origin/master, origin/HEAD) c242d4f release: v0.4.0, Dinos Kousidis, 2 weeks ago
* (upstream/pr/25) b8d4d1c global: license change to MIT License, Tibor Simko, 6 weeks ago
* (upstream/pr/24) 924ef8d docs: new logo, panel verbiage and links, Tibor Simko, 5 months ago
* (upstream/pr/23) c884a9b docs: author ORCID links, Tibor Simko, 7 months ago
* (tag: v0.2.0, upstream/pr/22) 8f5b219 release: v0.2.0, Dinos Kousidis, 7 months ago
[reana-server] git log -n 5 --graph --decorate ...
* (HEAD -> master, upstream/pr/109, upstream/master, origin/master, origin/HEAD) 698a34e installation: fix pytest-reana dependency version, Tibor Simko, 3 hours ago
* (upstream/pr/104) a2cb517 api: new rest api endpoint which returns wf params, Rokas Maciulaitis, 24 hours ago
| * (upstream/pr/108) a92b154 api: automatic openapi specs passing to reana-commons, Rokas Maciulaitis, 4 days ago
|/
| * (upstream/pr/107) cd155f9 api: automatic openapi specs passing to reana-commons, Rokas Maciulaitis, 4 days ago
| * 068441c api: new rest api endpoint which returns wf params, Rokas Maciulaitis, 5 days ago
|/
[reana-message-broker] git log -n 5 --graph --decorate ...
* (HEAD -> master, tag: v0.4.0, upstream/pr/16, upstream/master, origin/master, origin/HEAD) 7f18a21 release: v0.4.0, Tibor Simko, 2 weeks ago
* (upstream/pr/15, upstream/license-change) 1da1398 global: license change to MIT License, Tibor Simko, 6 weeks ago
* (upstream/pr/14, upstream/docs-logo-panel-links) 9d0bee3 docs: new logo, panel verbiage and links, Tibor Simko, 5 months ago
* (upstream/pr/13, upstream/docs-authors-orcid) 6661833 docs: author ORCID links, Tibor Simko, 7 months ago
* (tag: v0.2.0, upstream/pr/12) 1f7b82d release: v0.2.0, Tibor Simko, 7 months ago
[reana-workflow-controller] git log -n 5 --graph --decorate ...
* (upstream/pr/141) c1f00aa tests: utilities, Dinos Kousidis, 3 hours ago
* efca3e4 rest: set_workflow_status parameters description, Dinos Kousidis, 4 hours ago
* 120db5c rest: allow deletion of already deleted workflows, Dinos Kousidis, 4 hours ago
* a9721db tests: workspace deletion, Dinos Kousidis, 5 hours ago
* 74f9e53 rest: allow access to deleted workflows, Dinos Kousidis, 5 hours ago
[reana-workflow-engine-cwl] git log -n 5 --graph --decorate ...
* (upstream/pr/70) c92165a tasks: add stop_workflow task from reana_commons, Diego Rodriguez, 6 hours ago
* (HEAD -> master, tag: v0.4.0, upstream/pr/68, upstream/master, origin/master, origin/HEAD) 7b4b7d5 release: v0.4.0, Dinos Kousidis, 2 weeks ago
* (upstream/pr/66) e6cc7f9 installation: upgrade REANA-Commons, Diego Rodriguez, 5 weeks ago
* a5bb2ee publisher: use Kombu publisher, Diego Rodriguez, 5 weeks ago
* (upstream/pr/67) 696d526 installation: bump reana-commons version, Diego Rodriguez, 6 weeks ago
[reana-commons] git log -n 5 --graph --decorate ...
* (upstream/pr/66) 1c6ccb5 tasks: introduce common task to stop workflows, Diego Rodriguez, 5 hours ago
* cd856de api: update openapi specs, Jan Okraska, 20 hours ago
| * (HEAD -> master, upstream/pr/65, upstream/master, origin/master, origin/HEAD) e720eaa release: 0.5.0.dev20181116, Dinos Kousidis, 26 hours ago
| * 6cdb587 installation: bump pytest-reana, Dinos Kousidis, 27 hours ago
| * (upstream/pr/63) 9746ec4 api: new rest api endpoint, Rokas Maciulaitis, 4 days ago
|/
[reana-db] git log -n 5 --graph --decorate ...
* (HEAD -> master, upstream/pr/21, upstream/master, origin/master, origin/HEAD, installation-version) 0b0854b release: v0.5.0.dev20181116, Dinos Kousidis, 4 days ago
* (upstream/pr/19) 4a5ebdf models: renaming operational parameters, Rokas Maciulaitis, 4 days ago
| * (upstream/pr/20) 7f3621e models: addition of stopped WorkflowStatus, Diego Rodriguez, 6 days ago
|/
* (upstream/pr/15) 3e7a812 models: addition of deleted WorkflowStatus, Dinos Kousidis, 12 days ago
* (tag: v0.4.0, upstream/pr/17) e2bb3e6 release: v0.4.0, Tibor Simko, 2 weeks ago
|
gharchive/pull-request
| 2018-11-20T14:15:09 |
2025-04-01T06:40:12.388300
|
{
"authors": [
"tiborsimko"
],
"repo": "reanahub/reana",
"url": "https://github.com/reanahub/reana/pull/115",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
505310723
|
Playlist management: errors in cropping, appending or pasting
I have issues with the playlist manager
Once a playlist is made, it execute it finely
But once I chose for Crop or Append or Paste, the final result is totally incorrect.
Items are missing, or partially copied and pasted.
I attached a .zip file containing a video that shows you the details.
I would like to know if this is an issue or it's me badly operating, and what do you suggest for the best.
Reaper-Playlist-Issue.zip
Interesting, can you share the project RPP? Are any media items grouped?
Interesting, can you share that TEST.RPP project? Are any media items grouped?
Sorry for the late reply
About the media items: not any group
Yes sure, the TEST.RPP is here in attachment in the ZIP file
TEST.zip
I cannot reproduce the bug with SWS v2.10.0. "Paste playlist at edit cursor" behaves as expected here with that project.
There is a possibility some REAPER setting is interfering. Can you share your reaper.ini as well?
I can reproduce if I have Preferences (-> Project) -> Media Items Defaults: Overlap and crossfade items when splitting (length is set to 0:00.010 here). enabled when pasting here.
I have that pref assigned to the toolbar button shown in the gif.
https://i.imgur.com/PeL633Z.gif
Fixed in this build: sws-2.11.0.0-Windows-x64-e12769f5.exe.
Yep it works perfectly. Thanks a million!
P.S. testing it closely I did notice that when you Crop project on new tab, the zones are replicated but not into the new position but they are referred to the original project.
Thus it will be necessary to manually move or recreate the zones.
Did you get the same result?
"Crop project to playlist" and "Crop project to playlist (new project tab)" should produce the same output.
|
gharchive/issue
| 2019-10-10T14:30:14 |
2025-04-01T06:40:12.395369
|
{
"authors": [
"cfillion",
"nofishonfriday",
"tormyvancool"
],
"repo": "reaper-oss/sws",
"url": "https://github.com/reaper-oss/sws/issues/1204",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1262268220
|
For Code Review
Check on Javascript best practices
Done .
|
gharchive/pull-request
| 2022-06-06T19:14:49 |
2025-04-01T06:40:12.416845
|
{
"authors": [
"rebel216"
],
"repo": "rebel216/To-Do-List-PR",
"url": "https://github.com/rebel216/To-Do-List-PR/pull/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
107601150
|
Calling treetagger with untokenized test?
Hi,
the example on the tt4j homepage shows tokenized input being handled by treetagger. How to I give a untokenized text to treetagger to both POS tag and tokenize?
Thanks
Shyam
See https://reckart.github.io/tt4j/tokenizer.html
|
gharchive/issue
| 2015-09-21T21:22:54 |
2025-04-01T06:40:12.449880
|
{
"authors": [
"reckart",
"shyamupa"
],
"repo": "reckart/tt4j",
"url": "https://github.com/reckart/tt4j/issues/23",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2458966962
|
level 3 pull request not merged yet
@sanjay-kv please review and merge this pull request
https://github.com/Rakesh9100/CalcDiverse/pull/1820
This is the pull request i had created
@sanjay-kv I resolved the conflicts please review and merge when you are free
PR is getting reviewed one by one, you can check the PR before accepting the points request from the contributors @sanjay-kv You can remove the point label from here
|
gharchive/pull-request
| 2024-08-10T07:48:52 |
2025-04-01T06:40:12.457934
|
{
"authors": [
"Rakesh9100",
"aditya-bhaumik"
],
"repo": "recodehive/resume-pitch",
"url": "https://github.com/recodehive/resume-pitch/pull/49",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2643658209
|
Simplified arrow middleware definition
Issue
In https://github.com/oleksandr-danylchenko/text-annotator-js/commit/ce7014d6f74d929488cb0c277f7015a5abc90b0e#r148871559 I spotted that the arrow middleware definition can be simplified to its default form. That should be safe for the floating itself because the arrow is a "Data Middleware" that only populates the context with the positioning props, but doesn't change the behavior.
Demo
https://github.com/user-attachments/assets/ab1dab97-7e4e-41e5-8737-66a55e218372
Tested the same thing in Annotorious - yep, works :-) Thanks!
|
gharchive/pull-request
| 2024-11-08T10:49:31 |
2025-04-01T06:40:12.460441
|
{
"authors": [
"oleksandr-danylchenko",
"rsimon"
],
"repo": "recogito/text-annotator-js",
"url": "https://github.com/recogito/text-annotator-js/pull/177",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
1180344326
|
Utilize GPT3 embedding / classification API for more automated bulk labelling
Keyword-based bulk labeling using Rubrix Rules is still too slow if I were to label thousands of texts for classification - at maximum, I can label 20 at a time since keyword match is noisy with regard to ground-truth label.
I have played with GPT3 embeddings, where applying UMAP on texts results in clear clusters with semantically similar texts in clusters. If one can create Rubrix Rules instead based on a selected area on a scatterplot of text embeddings, then one can label hundreds of texts at once, and can easily finish labeling 50,000 data points in a day with high quality. And that'd be a game-changer.
Adding on to that, I also played with GPT3 classification API, where it gives pretty accurate few-shot classifications. And if that is added to the pipeline, it'd expedite labeling even further. Making it possible to label thousands of texts at once.
I have the GPT3 embedding + classification pipeline done in a notebook, is there plan on Rubrix's side to look into utilizing GPT3?
Hey @whaowhao Thank you for bringing this up! We are working on a tutorial in which we show how you can use Epoxy to achieve the same goal you are mentioning. The workflow would be something like this:
Come up with a few rules trying to cover semantic diverse records
Provide the weak label matrix and your embeddings of choice to Epoxy (with some thresholds)
Get back an enhanced weak label matrix
Hopefully, we can share it with you soon, we would love to get your feedback.
Thanks @whaowhao ! Adding to what @dcfidalgo mentions, we've also been discussing the ability of "labelling-by-drawing" based on a 2-D embedding-based representation (UMAP or otherwise). This will be in the roadmap but not immediately.
Meanwhile I'd be really interested on collaboration/supporting you with experiments with GPT3, would you be interested?
Hi @dvsrepo yeah let's chat
|
gharchive/issue
| 2022-03-25T04:54:36 |
2025-04-01T06:40:12.464970
|
{
"authors": [
"dcfidalgo",
"dvsrepo",
"whaowhao"
],
"repo": "recognai/rubrix",
"url": "https://github.com/recognai/rubrix/issues/1309",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
301786422
|
valuesAreSet field always is false.
Hi, I'm trying to create a RecommendItemsToUser request and set the returnProperties to true, but the valuesAreSet field in the recommendation response always comes as false.
When i try to execute recommendation.getValues() it always throws an IllegalStateException.
I checked the response body from the Api call and it does not contain the valuesAreSet property in the payload.
Recombee Api Response Payload:
{"recommId": "8b904849-19ad-47d9-90f2-b024ca67726f", "recomms": [{"values": {"manufacturer_code": 1}, "id": "123"}]}
Request:
final RecommendItemsToUser cf = new RecommendItemsToUser("CF", 5)
.setReturnProperties(true);
I confirm the bug, we will fix it ASAP
Thanks @OndraFiedler
It is fixed in https://github.com/recombee/java-api-client/releases/tag/v2.0.1.
The new version has been pushed to the central repository, but it can take some time (up to few hours) before it is available.
Thanks for reporting!
|
gharchive/issue
| 2018-03-02T14:36:42 |
2025-04-01T06:40:12.468701
|
{
"authors": [
"OndraFiedler",
"alepinzon"
],
"repo": "recombee/java-api-client",
"url": "https://github.com/recombee/java-api-client/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1412090532
|
docker-compose not recognising boolean values
When running docker-compose up -d the following error is observed:
ERROR: The Compose file './docker-compose.yml' is invalid because:
services.keycloak.environment.KC_HTTP_ENABLED contains false, which is an invalid type, it should be a string, number, or a null
Remediation is by encapsulating any boolean values in single quotes so they are treated as a string as noted here
Behaviour seen in commit 57d1665d323048c9feaf34a9fabe6cc454ec4ccc
Running Docker version 20.10.19, build d85ef84 on Ubuntu 22.04.1
Just saw that a similar pull request has already been responded to
|
gharchive/pull-request
| 2022-10-17T19:21:31 |
2025-04-01T06:40:12.490828
|
{
"authors": [
"chssn"
],
"repo": "reconmap/reconmap",
"url": "https://github.com/reconmap/reconmap/pull/123",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
524424299
|
[references] Tweets with ❤️️
https://twitter.com/davert/status/1243285793709457411
https://github.com/shopsys/shopsys/pull/1392#issuecomment-555106644
https://twitter.com/afilina/status/1229406983486394369
https://twitter.com/davert/status/1243285793709457411
https://twitter.com/phpinsights/status/1248596452730953732
https://twitter.com/MrDanack/status/1248607716018282497
https://twitter.com/VotrubaT/status/1253631226163990529
https://twitter.com/rectorphp/status/1253432659075710977
https://twitter.com/gaborhojtsy/status/1256246616710885381
https://twitter.com/VotrubaT/status/1259164120148054016
https://twitter.com/utopicabsurd/status/1271075970712326144
https://twitter.com/LBajsarowicz/status/1272947900016967683
- https://github.com/magento/magento2/pull/27946/files
https://twitter.com/LBajsarowicz/status/1273024967349866497
https://twitter.com/jrrtgimli/status/1275332001566986241
https://twitter.com/VotrubaT/status/1275863668878688257
https://twitter.com/lulco/status/1342198519009046528
https://twitter.com/dantleech/status/1342213179334676490
https://tweetdeck.twitter.com/
https://twitter.com/VotrubaT/status/1343551843759255552
https://twitter.com/rectorphp/status/1352758017410936833
https://twitter.com/rectorphp/status/1352758017410936833
https://twitter.com/Ruud_/status/1356256886417461249
https://twitter.com/tomasnorre/status/1359934323206615040
https://www.reddit.com/r/PHP/comments/lw5kvz/code_analysis_tools_to_help_with_upgrading_an_app/gpfhde7/?utm_source=reddit&utm_medium=web2x&context=3
https://www.reddit.com/r/PHP/comments/lw5kvz/code_analysis_tools_to_help_with_upgrading_an_app/gpfem49/?utm_source=reddit&utm_medium=web2x&context=3
https://github.com/sabbelasichon/typo3-rector/issues/2094#issuecomment-805320205
https://twitter.com/vojtabiberle/status/1379127544067321872
https://twitter.com/el_stoffel/status/1384865872959975425
https://twitter.com/KPikaza/status/1386791253275299842
https://twitter.com/rsinnbeck/status/1394188533888339969
https://twitter.com/althaus_it/status/1394235695611396097
https://www.facebook.com/christophe.avonture/posts/10225263983294673?notif_id=1624954306250211¬if_t=page_tag&ref=notif
https://twitter.com/sarfraznawaz/status/1437335510213615617
https://twitter.com/Chemaclass/status/1459616944127201285
https://twitter.com/alex_s_/status/1464358800006299650
https://twitter.com/Guikingone/status/1465944064126300161
https://twitter.com/srogier/status/1473289580564647937
https://twitter.com/unagiunag/status/1478941430085459968
https://twitter.com/amakdessi/status/1479110174589603842
https://twitter.com/unagiunag/status/1494832218614935552
https://twitter.com/_stefanzweifel/status/1521922469111771136
https://twitter.com/Chemaclass/status/1531622088276156418
https://twitter.com/VotrubaT/status/1537142504839159808
https://twitter.com/Guikingone/status/1543927696274690049
https://twitter.com/tmotyl/status/1557397895388725249
https://twitter.com/brendt_gd/status/1559829464359600128
https://www.reddit.com/r/PHP/comments/wqnqs9/rector_reaches_10_million_downloads/
https://github.com/rectorphp/rector-src/pull/2778#issuecomment-1218534498
https://twitter.com/VotrubaT/status/1595734036487540736
https://twitter.com/tenmajkl/status/1596612085810876416
https://mobile.twitter.com/markusstaab/status/1592572730934980608
https://twitter.com/JustSteveKing/status/1666745322310172673
https://twitter.com/jan_linhart/status/1737386735430951208
https://twitter.com/aminetiyal/status/1746910353714204952
https://twitter.com/enunomaduro/status/1754929560112472521
https://twitter.com/ramageek/status/1754894293016391935
https://twitter.com/SlyFireFox/status/1754871230228758735
https://twitter.com/nuernberger_me/status/1754813770365337969
https://twitter.com/PatricioOnCode/status/1754860469171462399
https://github.com/rectorphp/rector/discussions/8848
|
gharchive/issue
| 2019-11-18T15:03:25 |
2025-04-01T06:40:12.517398
|
{
"authors": [
"TomasVotruba"
],
"repo": "rectorphp/getrector-com",
"url": "https://github.com/rectorphp/getrector-com/issues/70",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2125473133
|
[backport]: from 2.6 to 2.7 (#173)
fix(trustyai): prometheus rules for probe
update(trusty): prometheus to use job instead of instance name for record rules
this is missed to get into main before 2.7 branch out
|
gharchive/pull-request
| 2024-02-08T15:36:04 |
2025-04-01T06:40:12.547395
|
{
"authors": [
"zdtsw"
],
"repo": "red-hat-data-services/rhods-operator",
"url": "https://github.com/red-hat-data-services/rhods-operator/pull/190",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2371990998
|
Add support for MCG in client console
/lgtm
|
gharchive/pull-request
| 2024-06-25T08:08:01 |
2025-04-01T06:40:12.562036
|
{
"authors": [
"SanjalKatiyar",
"bipuladh"
],
"repo": "red-hat-storage/odf-console",
"url": "https://github.com/red-hat-storage/odf-console/pull/1452",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
430020512
|
Added async and sync client wrappers.
Added new async test mechanism.
See #196. This PR improves async/sync APIs.
Design decision
I choose public inheritance and void function()=delete approach.
Why don't use composition and forwarding?
This approach requires many forwarding functions. If the function parameter would change, many part of codes need to change. It is difficult to maintain.
Code image (details are omitted intentionally)
// existing classes
class client : public endpoint {};
// new classes by composition
class sync_client {
public:
void publish(...) {
c_->publish(...):
}
private:
std::shared_ptr<client> c_;
};
Why don't use private inheritance and using?
Class endpoint use std::enable_shared_from_this.
// existing classes
class endpoint : public std::enable_shared_from_this<endpoint> {};
class client : public endpoint {};
// new classes by private inheritance
class sync_client : private client {
public:
using client::publish;
};
shared_from_this() throws bad_weak_ptr exception on runtime.
Why use void function_name() = delete ?
=delete only checks function name. So I choose the most simple signature and return type. Consider if I use the complete function signature and return type for =delete, if the function has overloads, which one should be chosen? I think there is no appropriate answer.
// existing classes
class endpoint : public std::enable_shared_from_this<endpoint> {};
class client : public endpoint {};
// new classes by public inheritance
class sync_client : public client {
public:
void async_publish() = delete; // always no parameter and returns void
};
Why don't you create asnyc_client and sync_client as the base classes of client?
It requires big design change.
I don't have good design concept.
|
gharchive/pull-request
| 2019-04-06T10:29:22 |
2025-04-01T06:40:12.606596
|
{
"authors": [
"redboltz"
],
"repo": "redboltz/mqtt_cpp",
"url": "https://github.com/redboltz/mqtt_cpp/pull/208",
"license": "BSL-1.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1763809983
|
Appstudio update test-component-pac-zvdm
Pipelines as Code configuration proposal
To start the PipelineRun, add a new comment with content /ok-to-test
For more detailed information about running a PipelineRun, please refer to Pipelines as Code documentation Running the PipelineRun
To customize the proposed PipelineRuns after merge, please refer to Build Pipeline customization
Pipelines as Code CI/test-component-pac-zvdm-on-pull-request has successfully validated your commit.
StatusDurationName
✅ Succeeded
7 seconds
init
✅ Succeeded
12 seconds
clone-repository
✅ Succeeded
29 seconds
build-container
✅ Succeeded
20 seconds
inspect-image
✅ Succeeded
1 minute
deprecated-base-image-check
✅ Succeeded
3 minutes
clair-scan
✅ Succeeded
52 seconds
clamav-scan
✅ Succeeded
17 seconds
sbom-json-check
✅ Succeeded
17 seconds
label-check
✅ Succeeded
8 seconds
show-sbom
✅ Succeeded
9 seconds
show-summary
Pipelines as Code CI/test-component-pac-zvdm-on-pull-request has failed.
StatusDurationName
---
---
init
Pipelines as Code CI/test-component-pac-zvdm-on-pull-request has successfully validated your commit.
StatusDurationName
✅ Succeeded
7 seconds
init
✅ Succeeded
21 seconds
clone-repository
✅ Succeeded
25 seconds
build-container
✅ Succeeded
12 seconds
deprecated-base-image-check
✅ Succeeded
12 seconds
inspect-image
✅ Succeeded
11 seconds
clair-scan
✅ Succeeded
37 seconds
clamav-scan
✅ Succeeded
11 seconds
sbom-json-check
✅ Succeeded
26 seconds
label-check
✅ Succeeded
6 seconds
show-sbom
✅ Succeeded
5 seconds
show-summary
Pipelines as Code CI/test-component-pac-zvdm-on-pull-request has successfully validated your commit.
StatusDurationName
✅ Succeeded
7 seconds
init
✅ Succeeded
21 seconds
clone-repository
✅ Succeeded
25 seconds
build-container
✅ Succeeded
12 seconds
deprecated-base-image-check
✅ Succeeded
12 seconds
inspect-image
✅ Succeeded
11 seconds
clair-scan
✅ Succeeded
37 seconds
clamav-scan
✅ Succeeded
11 seconds
sbom-json-check
✅ Succeeded
26 seconds
label-check
✅ Succeeded
6 seconds
show-sbom
✅ Succeeded
5 seconds
show-summary
|
gharchive/pull-request
| 2023-06-19T15:56:12 |
2025-04-01T06:40:12.660957
|
{
"authors": [
"rhtap-qe-bots"
],
"repo": "redhat-appstudio-qe/devfile-sample-hello-world",
"url": "https://github.com/redhat-appstudio-qe/devfile-sample-hello-world/pull/11000",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1518587243
|
Appstudio update test-component-pac-kjnh
Pipelines as Code configuration proposal
Pipelines as Code CI/test-component-pac-kjnh-on-pull-request has successfully validated your commit.
StatusDurationName
✅ Succeeded
9 seconds
appstudio-init
✅ Succeeded
9 seconds
clone-repository
✅ Succeeded
18 seconds
appstudio-configure-build
✅ Succeeded
19 seconds
sast-snyk-check
✅ Succeeded
43 seconds
build-container
✅ Succeeded
20 seconds
sanity-inspect-image
✅ Succeeded
17 seconds
deprecated-base-image-check
✅ Succeeded
49 seconds
clamav-scan
✅ Succeeded
17 seconds
clair-scan
✅ Succeeded
23 seconds
sbom-json-check
✅ Succeeded
12 seconds
sanity-label-check
✅ Succeeded
11 seconds
sanity-optional-label-check
✅ Succeeded
8 seconds
show-summary
Pipelines as Code CI/test-component-pac-kjnh-on-pull-request has successfully validated your commit.
StatusDurationName
✅ Succeeded
9 seconds
appstudio-init
✅ Succeeded
9 seconds
clone-repository
✅ Succeeded
18 seconds
appstudio-configure-build
✅ Succeeded
19 seconds
sast-snyk-check
✅ Succeeded
43 seconds
build-container
✅ Succeeded
20 seconds
sanity-inspect-image
✅ Succeeded
17 seconds
clair-scan
✅ Succeeded
49 seconds
clamav-scan
✅ Succeeded
17 seconds
deprecated-base-image-check
✅ Succeeded
23 seconds
sbom-json-check
✅ Succeeded
11 seconds
sanity-optional-label-check
✅ Succeeded
12 seconds
sanity-label-check
✅ Succeeded
8 seconds
show-summary
Pipelines as Code CI/test-component-pac-kjnh-on-pull-request has successfully validated your commit.
StatusDurationName
✅ Succeeded
8 seconds
appstudio-init
✅ Succeeded
15 seconds
clone-repository
✅ Succeeded
13 seconds
sast-snyk-check
✅ Succeeded
14 seconds
appstudio-configure-build
✅ Succeeded
45 seconds
build-container
✅ Succeeded
24 seconds
sanity-inspect-image
✅ Succeeded
22 seconds
deprecated-base-image-check
✅ Succeeded
47 seconds
clamav-scan
✅ Succeeded
17 seconds
clair-scan
✅ Succeeded
27 seconds
sbom-json-check
✅ Succeeded
13 seconds
sanity-optional-label-check
✅ Succeeded
11 seconds
sanity-label-check
✅ Succeeded
6 seconds
show-summary
Pipelines as Code CI/test-component-pac-kjnh-on-pull-request has successfully validated your commit.
StatusDurationName
✅ Succeeded
8 seconds
appstudio-init
✅ Succeeded
15 seconds
clone-repository
✅ Succeeded
14 seconds
appstudio-configure-build
✅ Succeeded
13 seconds
sast-snyk-check
✅ Succeeded
45 seconds
build-container
✅ Succeeded
22 seconds
deprecated-base-image-check
✅ Succeeded
24 seconds
sanity-inspect-image
✅ Succeeded
47 seconds
clamav-scan
✅ Succeeded
17 seconds
clair-scan
✅ Succeeded
27 seconds
sbom-json-check
✅ Succeeded
13 seconds
sanity-optional-label-check
✅ Succeeded
11 seconds
sanity-label-check
✅ Succeeded
6 seconds
show-summary
|
gharchive/pull-request
| 2023-01-04T09:03:19 |
2025-04-01T06:40:12.693448
|
{
"authors": [
"jkopriva"
],
"repo": "redhat-appstudio-qe/devfile-sample-hello-world",
"url": "https://github.com/redhat-appstudio-qe/devfile-sample-hello-world/pull/1141",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1595725735
|
Appstudio update mvp-test-component
Pipelines as Code configuration proposal
Pipelines as Code CI/mvp-test-component-on-pull-request has successfully validated your commit.
StatusDurationName
✅ Succeeded
10 seconds
init
✅ Succeeded
18 seconds
clone-repository
✅ Succeeded
1 minute
build-container
✅ Succeeded
12 seconds
sanity-inspect-image
✅ Succeeded
11 seconds
deprecated-base-image-check
✅ Succeeded
58 seconds
clamav-scan
✅ Succeeded
16 seconds
clair-scan
✅ Succeeded
9 seconds
sbom-json-check
✅ Succeeded
17 seconds
sanity-label-check
✅ Succeeded
15 seconds
sanity-optional-label-check
✅ Succeeded
6 seconds
show-summary
|
gharchive/pull-request
| 2023-02-22T19:43:12 |
2025-04-01T06:40:12.703049
|
{
"authors": [
"redhat-appstudio-qe-bot2"
],
"repo": "redhat-appstudio-qe/hacbs-test-project",
"url": "https://github.com/redhat-appstudio-qe/hacbs-test-project/pull/32",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1318566968
|
Added ppc64le support
This PR will add multi-architecture support to the application-service operator image
/retest
|
gharchive/pull-request
| 2022-07-26T17:44:32 |
2025-04-01T06:40:12.704203
|
{
"authors": [
"mayurwaghmode"
],
"repo": "redhat-appstudio/application-service",
"url": "https://github.com/redhat-appstudio/application-service/pull/156",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2387910778
|
KFLUXINFRA-651: Adding Instance Types for Multi-Platform Builds
Adding two types of instances
Higher Memory (1:4 ratio for cpu: memory)
Higher CPU (1:2 ratio for cpu: memory)
Note Some users may require more memory and less CPU, making Memory Optimized Instances the better choice for them. On the other hand, some users may need more CPU power but less memory, in which case CPU-optimized instances would be a good fit. Therefore, a combination of memory-optimized and CPU-optimized instances is essential to cater to different user needs. In the label naming, m refers to memory optimised and c refers to compute optimised.
Memory Optimised (not aws terminology, their memory optimized class starts from r)
Multi-Platform Label
Instance Type
Architecture
CPU
Memory(GB)
linux-mlarge/amd64
m5a.large
AMD64
2
8
linux-mxlarge/amd64
m6a.xlarge
AMD64
4
16
linux-m2xlarge/amd64
m6a.2xlarge
AMD64
8
32
linux-m4xlarge/amd64
m6a.4xlarge
AMD64
16
64
linux-m8xlarge/amd64
m6a.8xlarge
AMD64
32
128
linux-mlarge/arm64
m6g.large
ARM64
2
8
linux-mxlarge/arm64
m6g.xlarge
ARM64
4
16
linux-m2xlarge/arm64
m6g.2xlarge
ARM64
8
32
linux-m4xlarge/arm64
m6g.4xlarge
ARM64
16
64
linux-m8xlarge/arm64
m6g.8xlarge
ARM64
32
128
CPU Optimised (again not aws terminology)
Multi-Platform Label
Instance Type
Architecture
CPU
Memory(GB)
linux-clarge/amd64
c6a.xlarge
AMD64
4
8
linux-cxlarge/amd64
c6a.2xlarge
AMD64
8
16
linux-c2xlarge/amd64
c6a.4xlarge
AMD64
16
32
linux-c4xlarge/amd64
c6a.8xlarge
AMD64
32
64
linux-clarge/arm64
c6g.xlarge
ARM64
4
8
linux-cxlarge/arm64
c6g.2xlarge
ARM64
8
16
linux-c2xlarge/arm64
c6g.4xlarge
ARM64
16
32
linux-c4xlarge/arm64
c6g.8xlarge
ARM64
32
64
Suggestion We can minimize code duplication by utilizing default values for common parameters like imageId, subnetId, region, etc, unless specified otherwise.
One thought: we will have quay.io with AWS internal communication right? otherwise how about using enhanced network enabled instances (with *n) for better push pull speeds?
One thought: we will have quay.io with AWS internal communication right? otherwise how about using enhanced network enabled instances (with *n) for better push pull speeds?
ENA is enabled for the instance types we are using m6a.xxxxx, m6g.xxxxx, c6a.xxxxx and c6g.xxxxx. Reference AWS Document -- https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/enhanced-networking-ena.html
/lgtm
|
gharchive/pull-request
| 2024-07-03T07:39:20 |
2025-04-01T06:40:12.721748
|
{
"authors": [
"bamachrn",
"manish-jangra"
],
"repo": "redhat-appstudio/infra-deployments",
"url": "https://github.com/redhat-appstudio/infra-deployments/pull/3997",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2752668445
|
Promote release-service from staging to production
Included PRs:
https://github.com/konflux-ci/release-service/pull/638
https://github.com/konflux-ci/release-service/pull/637
https://github.com/konflux-ci/release-service/pull/636
https://github.com/konflux-ci/release-service/pull/635
https://github.com/konflux-ci/release-service/pull/633
https://github.com/konflux-ci/release-service/pull/630
https://github.com/konflux-ci/release-service/pull/629
https://github.com/konflux-ci/release-service/pull/632
https://github.com/konflux-ci/release-service/pull/628
https://github.com/konflux-ci/release-service/pull/627
https://github.com/konflux-ci/release-service/pull/626
https://github.com/konflux-ci/release-service/pull/622
/lgtm
|
gharchive/pull-request
| 2024-12-20T12:43:26 |
2025-04-01T06:40:12.726992
|
{
"authors": [
"johnbieren",
"mmalina"
],
"repo": "redhat-appstudio/infra-deployments",
"url": "https://github.com/redhat-appstudio/infra-deployments/pull/5199",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1358402092
|
fix: increase timeout for e2e
Why
Currently there's a 1h (default) timeout for running e2e. Since the number of tests is growing and takes more time to finish, we are sometimes hitting the limit. If that happens, tests get interrupted and a CI check is marked as failed.
This is a temporary solution until we enable parallel e2e test run for infra-deployments (which was already enabled for e2e-tests repo)
/lgtm
Ran 87 of 92 Specs in 4424.652 seconds
SUCCESS! -- 87 Passed | 0 Failed | 0 Pending | 5 Skipped
|
gharchive/pull-request
| 2022-09-01T07:44:33 |
2025-04-01T06:40:12.728576
|
{
"authors": [
"psturc",
"tkdchen"
],
"repo": "redhat-appstudio/infra-deployments",
"url": "https://github.com/redhat-appstudio/infra-deployments/pull/684",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
981505573
|
Add auth-option to authenticate as githubapp
Fixes #112.
Adds support for authenticating as github-app, which offers:
more narrow-grained permissions (avoiding PATs)
increased security, using a private key
increased API-quota
The github-client is now wired up/facilitated through the go-github-app library, which offers caching, metrics etc.
Note: Draft PR - not yet tested, for feedback.
go-githubapp now has convenience function to configure the transport.
I think this is good to go, but have not yet been able to test it.
That should be it. I've updated the docs and tested it in my 4.6.x cluster, using app-based auth:
group-sync-operator-controller-manager-54d5874d76-kh5cq manager 2021-09-06T17:31:04.501Z INFO controllers.GroupSync Beginning Sync {"groupsync": "group-sync-operator/github-groupsync", "Provider": "github"}
group-sync-operator-controller-manager-54d5874d76-kh5cq manager 2021-09-06T17:31:06.354Z INFO controllers.GroupSync Sync Completed Successfully {"groupsync": "group-sync-operator/github-groupsync", "Provider": "github", "Groups Created or Updated": 1}
CC @sabre1041 @raffaelespazzoli
@sabre1041 I think all points have been covered now - thanks for the input!
|
gharchive/pull-request
| 2021-08-27T18:48:01 |
2025-04-01T06:40:12.731904
|
{
"authors": [
"davidkarlsen"
],
"repo": "redhat-cop/group-sync-operator",
"url": "https://github.com/redhat-cop/group-sync-operator/pull/118",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
989044632
|
Run go mod tidy on each pull request and fail the build if there are changes
We will be outputting the vendor directory in the midstream repo, and this will prevent any possible conflicts.
Actually this is not needed: https://github.blog/changelog/2020-10-19-dependabot-go-mod-tidy-and-vendor-support/
|
gharchive/issue
| 2021-09-06T10:54:31 |
2025-04-01T06:40:12.736214
|
{
"authors": [
"craicoverflow"
],
"repo": "redhat-developer/app-services-cli",
"url": "https://github.com/redhat-developer/app-services-cli/issues/1029",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1183258588
|
fix: add server side backed up metrics
Motivation
This change uses backend compiled metrics for consumer groups.
This have been in production for quite a while and have been extensively tested.
Verification
Get Kafka
Create new topic test
Create kcat.properties based of the kcat guide
kcat -b <yourhostname -F ./kcat.properties -P -t test`
run rhoas kafka consumer-group list
run rhoas kafka consumer-group describe
I have used latest version of SDK. We need to wait to make sure that it is currently deployed to production
@rkpattnaik780
FYI @mikeedgar API is working find and was tested end to end
|
gharchive/pull-request
| 2022-03-28T10:57:12 |
2025-04-01T06:40:12.739325
|
{
"authors": [
"wtrocki"
],
"repo": "redhat-developer/app-services-cli",
"url": "https://github.com/redhat-developer/app-services-cli/pull/1493",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
863700015
|
Provide simplified ability to switch between stagging and production environments
We can add env variable into OLM as example (can be empty) and then scripts can patch it (as discussed with @b1zzu)
I will verify if that really works and document this in contributing docs
Added env variable directly to the OLM
|
gharchive/issue
| 2021-04-21T10:06:53 |
2025-04-01T06:40:12.740482
|
{
"authors": [
"wtrocki"
],
"repo": "redhat-developer/app-services-operator",
"url": "https://github.com/redhat-developer/app-services-operator/issues/189",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2109837968
|
doc: OpenShift argocd CLI client command reference documentation
What type of PR is this?
/kind documentation
What does this PR do / why we need it:
Have you updated the necessary documentation?
[ ] Documentation update is required by this PR.
[ ] Documentation has been updated.
Which issue(s) this PR fixes:
Fixes #?
Test acceptance criteria:
[ ] Unit Test
[ ] E2E Test
How to test changes / Special notes to the reviewer:
@anandf How about the updates to these sections?
Update these existing sections:
Creating an application by using the oc tool
https://docs.openshift.com/gitops/1.11/argocd_applications/deploying-a-spring-boot-application-with-argo-cd.html#creating-an-application-by-using-the-oc-tool_deploying-a-spring-boot-application-with-argo-cd
https://docs.openshift.com/gitops/1.11/declarative_clusterconfig/configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations.html#creating-an-application-by-using-the-oc-tool_configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations
Synchronizing your application with your Git repository
https://docs.openshift.com/gitops/1.11/declarative_clusterconfig/configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations.html#synchronizing-your-application-application-with-your-git-repository_configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations
@anandf Can we add some content for these sections, if they are relevant at all?
About the OpenShift argo cd CLI (I think this is a nice to have section)
Logging in to the OpenShift argo cd CLI using a web browser
usage instructions - information that must go as admonitions such as Note, Important, Tip, caution, or warning.
@anandf How about the updates to these sections?
Update these existing sections:
Creating an application by using the oc tool
https://docs.openshift.com/gitops/1.11/argocd_applications/deploying-a-spring-boot-application-with-argo-cd.html#creating-an-application-by-using-the-oc-tool_deploying-a-spring-boot-application-with-argo-cd
https://docs.openshift.com/gitops/1.11/declarative_clusterconfig/configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations.html#creating-an-application-by-using-the-oc-tool_configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations
Synchronizing your application with your Git repository
https://docs.openshift.com/gitops/1.11/declarative_clusterconfig/configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations.html#synchronizing-your-application-application-with-your-git-repository_configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations
Added sections for creating and syncing app using CLI in both normal and core modes.
@anandf Can we add some content for these sections, if they are relevant at all?
Logging in to the OpenShift argo cd CLI using a web browser -> This is not applicable. One has to login via the CLI itself. Its not possible to login via web browser like its possible for oc login
|
gharchive/pull-request
| 2024-01-31T11:00:38 |
2025-04-01T06:40:12.750950
|
{
"authors": [
"Srivaralakshmi",
"anandf"
],
"repo": "redhat-developer/gitops-operator",
"url": "https://github.com/redhat-developer/gitops-operator/pull/650",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1253785178
|
odo dev add flag to run non-default build command.
User Story
As an odo user
I want to be able to execute alternative build command instead of the default one
So that I can run my with different options (different configuration, flags, runtime versions etc..) without modifying the default command.
Example:
Go devfile has the following default build command.
- exec:
commandLine: GOCACHE=${PROJECT_SOURCE}/.cache go build main.go
component: runtime
group:
isDefault: true
kind: build
workingDir: ${PROJECT_SOURCE}
id: build
In some situations, I want to pass arguments to linker to for example set string value of a variable.
This can be done adding -ldflags="-X github.com/redhat-developer/odo/pkg/segment.writeKey=foo".
Currently I have to edit the commandLine in order to be able to execute build with different arguments.
Instead of that it would be nice to be able to add extra run command like this:
- exec:
commandLine: GOCACHE=${PROJECT_SOURCE}/.cache go build -ldflags="-X github.com/redhat-developer/odo/pkg/segment.writeKey=foo" main.go
component: runtime
group:
isDefault: false
kind: build
workingDir: ${PROJECT_SOURCE}
id: build-with-key
And then simply execute odo dev --build-command build-with-key, when I want to switch back to using the default profile, I stop running odo dev command and start it again, but this time without --build-command flag.
Acceptance Criteria
[ ] odo dev should have --build-command flag that controls what command is used to build application.
/kind user-story
TODO (to not forget):
As commented out in [1]:
we should try to harmonize the behaviour between commands in pkg/libdevfile/libdevfile.go. Deploy is using getDefaultCommand, when Build/Test/Run is using getCommandAssociatedToGroup
[1] https://github.com/redhat-developer/odo/pull/5768#discussion_r895480523
|
gharchive/issue
| 2022-05-31T12:18:45 |
2025-04-01T06:40:12.759820
|
{
"authors": [
"kadel",
"rm3l"
],
"repo": "redhat-developer/odo",
"url": "https://github.com/redhat-developer/odo/issues/5776",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
322855232
|
fix release scripts - 'odo version' output changed
😞
forgot to change one place
similar to #451
|
gharchive/pull-request
| 2018-05-14T14:59:48 |
2025-04-01T06:40:12.760897
|
{
"authors": [
"kadel"
],
"repo": "redhat-developer/odo",
"url": "https://github.com/redhat-developer/odo/pull/452",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1214188266
|
Provide tooltips on icon
As a user, the icons themselves are not going to be sufficient to tell me what actions are behind each buttons.
It would be nicer to provide tooltips behind the button, so that the user can discover what's the button will do, without having to click on it.
|
gharchive/issue
| 2022-04-25T08:56:52 |
2025-04-01T06:40:12.762474
|
{
"authors": [
"fbricon",
"slemeur"
],
"repo": "redhat-developer/openshift-dd-ext",
"url": "https://github.com/redhat-developer/openshift-dd-ext/issues/21",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2550180431
|
RHIDP-3377: Adding cross-links to the Installation titles
IMPORTANT: Do Not Merge - To be merged by Docs Team Only
Version(s): 1.2, 1.3
Issue: RHIDP-3377
Reviews:
[x] Docs review: @hmanwani-rh
/cherry-pick release-1.3
/cherry-pick 1.2.x
|
gharchive/pull-request
| 2024-09-26T10:28:44 |
2025-04-01T06:40:12.770318
|
{
"authors": [
"hmanwani-rh",
"pabel-rh"
],
"repo": "redhat-developer/red-hat-developers-documentation-rhdh",
"url": "https://github.com/redhat-developer/red-hat-developers-documentation-rhdh/pull/538",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2652518906
|
Added discover content
I’m creating this PR because the About RHDH title is already published on docs.redhat.com, but currently, it doesn’t contain any content. For now, I’m adding content to the page, which can be refined and improved later.
cc @jmagak
Closing this PR, since the title is now unpublished in Pantheon.
|
gharchive/pull-request
| 2024-11-12T15:26:55 |
2025-04-01T06:40:12.771856
|
{
"authors": [
"hmanwani-rh"
],
"repo": "redhat-developer/red-hat-developers-documentation-rhdh",
"url": "https://github.com/redhat-developer/red-hat-developers-documentation-rhdh/pull/698",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
160856652
|
Trigger a first build upon fling completion
Create a new Build resource from the generated BuildConfig resource in the OpenShift project.
Would adding this trigger (https://github.com/tnozicka/openshift-templates/blob/master/pipeline-template.yaml#L34) be enough?
6d85abfc3440d379e7135325f7005bbf4b80e57c
|
gharchive/issue
| 2016-06-17T10:17:34 |
2025-04-01T06:40:12.824579
|
{
"authors": [
"ALRubinger",
"tnozicka",
"xcoulon"
],
"repo": "redhat-kontinuity/catapult",
"url": "https://github.com/redhat-kontinuity/catapult/issues/128",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1053032322
|
operator vfunction-server-operator (v2.3.594)
New operator bundle
Name: vfunction-server-operator
Version: v2.3.594
Certification project: 5f64dc1e0a4d06443d2818ad
Test result URL: https://catalog.redhat.com/api/containers/v1/projects/certification/test-results/id/61917043d115ef379a598e5e
Test logs URL: https://catalog.redhat.com/api/containers/v1/projects/certification/artifacts/id/61917042d6b39d004292bad1
Hi,
I'm sure 100% that my ci tests passed okay (even 3 times...) on my local OpenShift cluster with the most updated operator-ci-pipeline.yml and tasks from the operator-pipelines repo.
I did have some issues with timeout exceeding on both ScorecardBasicSpecCheck & ScorecardOlmSuiteCheck before I updated. (I couldn't access the test result URL nor the logs URL from above to see the exact errors for this run - to be sure they are the same, though)
Before starting the long journey of opening a support case for this, I just wanted to make sure that the errors are not on your end because the tests cluster is not updated correctly? Could you please verify that tests can pass at all now for ScorecardBasicSpecCheck & ScorecardOlmSuiteCheck on your testing cluster?
Thanks.
Now I could also see in the logs that both ScorecardBasicSpecCheck and ScorecardOlmSuiteCheck failed on the 30 seconds timeout (as I expected and mentioned in my previous comment).
As I also mentioned - I had this error too until I pulled a new version from operator-pipelines and applied all pipelines and tasks on the OpenShift cluster.
From the last Perflight.log:
time="2021-11-15T13:58:35Z" level=info msg="running check: ScorecardBasicSpecCheck"
time="2021-11-15T13:58:35Z" level=debug msg="Running operator-sdk scorecard check for quay.io/operator-pipeline-prod/vfunction-server-operator:v2.3.594"
time="2021-11-15T13:58:35Z" level=debug msg="--selector=[test=basic-check-spec-test]"
time="2021-11-15T13:58:35Z" level=trace msg="running scorecard with the following invocation[operator-sdk scorecard --output json --selector=test=basic-check-spec-test --kubeconfig /tmp/kubeconfig-3421170286 --namespace default --service-account default --config /tmp/scorecard-test-config-3472466720.yaml --verbose /tmp/preflight-3179944778/fs]"
time="2021-11-15T13:59:05Z" level=error msg="stdout: "
time="2021-11-15T13:59:05Z" level=error msg="stderr: time="2021-11-15T13:58:35Z" level=debug msg="Debug logging is set"\nError: error running tests context deadline exceeded\nUsage:\n operator-sdk scorecard [flags]\n\nFlags:\n -c, --config string path to scorecard config file\n -h, --help help for scorecard\n --kubeconfig string kubeconfig path\n -L, --list Option to enable listing which tests are run\n -n, --namespace string namespace to run the test images in\n -o, --output string Output format for results. Valid values: text, json, xunit (default "text")\n -l, --selector string label selector to determine which tests are run\n -s, --service-account string Service account to use for tests (default "default")\n -x, --skip-cleanup Disable resource cleanup after tests are run\n -b, --storage-image string Storage image to be used by the Scorecard pod (default "docker.io/library/busybox@sha256:c71cb4f7e8ececaffb34037c2637dc86820e4185100e18b4d02d613a9bd772af")\n -t, --test-output string Test output directory. (default "test-output")\n -u, --untar-image string Untar image to be used by the Scorecard pod (default "registry.access.redhat.com/ubi8@sha256:910f6bc0b5ae9b555eb91b88d28d568099b060088616eba2867b07ab6ea457c7")\n -w, --wait-time duration seconds to wait for tests to complete. Example: 35s (default 30s)\n\nGlobal Flags:\n --plugins strings plugin keys to be used for this subcommand execution\n --verbose Enable verbose logging\n\ntime="2021-11-15T13:59:05Z" level=fatal msg="error running tests context deadline exceeded"\n"
time="2021-11-15T13:59:05Z" level=info msg="check completed: ScorecardBasicSpecCheck" ERROR="failed to run operator-sdk scorecard: exit status 1" result="failed to run operator-sdk scorecard: exit status 1"
time="2021-11-15T13:59:05Z" level=info msg="running check: ScorecardOlmSuiteCheck"
time="2021-11-15T13:59:05Z" level=debug msg="Running operator-sdk scorecard Check for quay.io/operator-pipeline-prod/vfunction-server-operator:v2.3.594"
time="2021-11-15T13:59:05Z" level=debug msg="--selector=[suite=olm]"
time="2021-11-15T13:59:05Z" level=trace msg="running scorecard with the following invocation[operator-sdk scorecard --output json --selector=suite=olm --kubeconfig /tmp/kubeconfig-3421170286 --namespace default --service-account default --config /tmp/scorecard-test-config-2813129432.yaml --verbose /tmp/preflight-3179944778/fs]"
time="2021-11-15T13:59:36Z" level=error msg="stdout: "
time="2021-11-15T13:59:36Z" level=error msg="stderr: time="2021-11-15T13:59:05Z" level=debug msg="Debug logging is set"\nError: error running tests context deadline exceeded\nUsage:\n operator-sdk scorecard [flags]\n\nFlags:\n -c, --config string path to scorecard config file\n -h, --help help for scorecard\n --kubeconfig string kubeconfig path\n -L, --list Option to enable listing which tests are run\n -n, --namespace string namespace to run the test images in\n -o, --output string Output format for results. Valid values: text, json, xunit (default "text")\n -l, --selector string label selector to determine which tests are run\n -s, --service-account string Service account to use for tests (default "default")\n -x, --skip-cleanup Disable resource cleanup after tests are run\n -b, --storage-image string Storage image to be used by the Scorecard pod (default "docker.io/library/busybox@sha256:c71cb4f7e8ececaffb34037c2637dc86820e4185100e18b4d02d613a9bd772af")\n -t, --test-output string Test output directory. (default "test-output")\n -u, --untar-image string Untar image to be used by the Scorecard pod (default "registry.access.redhat.com/ubi8@sha256:910f6bc0b5ae9b555eb91b88d28d568099b060088616eba2867b07ab6ea457c7")\n -w, --wait-time duration seconds to wait for tests to complete. Example: 35s (default 30s)\n\nGlobal Flags:\n --plugins strings plugin keys to be used for this subcommand execution\n --verbose Enable verbose logging\n\ntime="2021-11-15T13:59:36Z" level=fatal msg="error running tests context deadline exceeded"\n"
time="2021-11-15T13:59:36Z" level=info msg="check completed: ScorecardOlmSuiteCheck" ERROR="failed to run operator-sdk scorecard: exit status 1" result="failed to run operator-sdk scorecard: exit status 1"
Thank you. We are looking into this on our side.
Hi,
Any news regarding this?
Thanks.
|
gharchive/pull-request
| 2021-11-14T20:23:45 |
2025-04-01T06:40:12.842182
|
{
"authors": [
"tonytcampbell",
"vfunction"
],
"repo": "redhat-openshift-ecosystem/certified-operators",
"url": "https://github.com/redhat-openshift-ecosystem/certified-operators/pull/119",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1305416246
|
operator neuvector-community-operator (1.3.5)
Signed-off-by: selvamt94 sthangaraj@neuvector.com
Thanks submitting your Operator. Please check below list before you create your Pull Request.
New Submissions
[x] Are you familiar with our contribution guidelines?
[x] Have you packaged and deployed your Operator for Operator Framework?
[x] Have you tested your Operator with all Custom Resource Definitions?
[x] Have you tested your Operator in all supported installation modes?
[x] Have you considered whether you want use semantic versioning order?
[x] Is your submission signed?
[x] Is operator icon set?
Updates to existing Operators
[x] Did you create a ci.yaml file according to the update instructions?
[x] Is your new CSV pointing to the previous version with the replaces property if you chose replaces-mode via the updateGraph property in ci.yaml?
[x] Is your new CSV referenced in the appropriate channel defined in the package.yaml or annotations.yaml ?
[x] Have you tested an update to your Operator when deployed via OLM?
[x] Is your submission signed?
Your submission should not
[x] Modify more than one operator
[x] Modify an Operator you don't own
[x] Rename an operator - please remove and add with a different name instead
[x] Modify any files outside the above mentioned folders
[x] Contain more than one commit. Please squash your commits.
Operator Description must contain (in order)
[x] Description about the managed Application and where to find more information
[x] Features and capabilities of your Operator and how to use it
[x] Any manual steps about potential pre-requisites for using your Operator
Operator Metadata should contain
[x] Human readable name and 1-liner description about your Operator
[x] Valid category name1
[x] One of the pre-defined capability levels2
[x] Links to the maintainer, source code and documentation
[x] Example templates for all Custom Resource Definitions intended to be used
[x] A quadratic logo
Remember that you can preview your CSV here.
--
1 If you feel your Operator does not fit any of the pre-defined categories, file an issue against this repo and explain your need
2 For more information see here
/merge possible
/merge possible
|
gharchive/pull-request
| 2022-07-14T23:46:52 |
2025-04-01T06:40:12.855418
|
{
"authors": [
"framework-automation",
"selvamt94"
],
"repo": "redhat-openshift-ecosystem/community-operators-prod",
"url": "https://github.com/redhat-openshift-ecosystem/community-operators-prod/pull/1414",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1435942856
|
operator cert-manager (1.10.1)
Signed-off-by: Richard Wall richard.wall@jetstack.io
/hold
Sanity checking the changes between 1.10.0 and 1.10.1
✦ $ git diff --no-index 1.10.0 1.10.1
diff --git a/1.10.0/manifests/cert-manager.clusterserviceversion.yaml b/1.10.1/manifests/cert-manager.clusterserviceversion.yaml
index 4f298cca..402923d3 100644
--- a/1.10.0/manifests/cert-manager.clusterserviceversion.yaml
+++ b/1.10.1/manifests/cert-manager.clusterserviceversion.yaml
@@ -68,7 +68,8 @@ metadata:
capabilities: Full Lifecycle
categories: Security
containerImage: quay.io/jetstack/cert-manager-controller:v1.10.0
- createdAt: '2022-11-02T14:25:05'
+ createdAt: '2022-11-04T10:22:12'
+ olm.skipRange: '>=1.10.0 <1.10.1'
operators.operatorframework.io/builder: operator-sdk-v1.25.0
operators.operatorframework.io/internal-objects: |-
[
@@ -83,7 +84,7 @@ metadata:
operatorframework.io/arch.arm64: supported
operatorframework.io/arch.ppc64le: supported
operatorframework.io/arch.s390x: supported
- name: cert-manager.v1.10.0
+ name: cert-manager.v1.10.1
namespace: placeholder
spec:
apiservicedefinitions: {}
@@ -745,9 +746,9 @@ spec:
- --secure-port=10250
- --dynamic-serving-ca-secret-namespace=$(POD_NAMESPACE)
- --dynamic-serving-ca-secret-name=cert-manager-webhook-ca
- - --dynamic-serving-dns-names=cert-manager-webhook
- - --dynamic-serving-dns-names=cert-manager-webhook.$(POD_NAMESPACE)
- - --dynamic-serving-dns-names=cert-manager-webhook.$(POD_NAMESPACE).svc
+ - --dynamic-serving-dns-names=cert-manager-webhook-service.$(POD_NAMESPACE).svc
+ - --tls-cert-file=/apiserver.local.config/certificates/apiserver.crt
+ - --tls-private-key-file=/apiserver.local.config/certificates/apiserver.key
env:
- name: POD_NAMESPACE
valueFrom:
@@ -879,7 +880,7 @@ spec:
provider:
name: The cert-manager maintainers
url: https://cert-manager.io/
- version: 1.10.0
+ version: 1.10.1
webhookdefinitions:
- admissionReviewVersions:
- v1
/merge possible
/merge possible
/hold cancel
/merge possible
|
gharchive/pull-request
| 2022-11-04T11:37:36 |
2025-04-01T06:40:12.858776
|
{
"authors": [
"framework-automation",
"wallrj"
],
"repo": "redhat-openshift-ecosystem/community-operators-prod",
"url": "https://github.com/redhat-openshift-ecosystem/community-operators-prod/pull/1819",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1947502468
|
operator pulp-operator (1.0.0-beta.2)
Thanks for submitting your Operator. Please check the below list before you create your Pull Request.
New Submissions
[ ] Are you familiar with our contribution guidelines?
[ ] Have you packaged and deployed your Operator for Operator Framework?
[ ] Have you tested your Operator with all Custom Resource Definitions?
[ ] Have you tested your Operator in all supported installation modes?
[ ] Have you considered whether you want to use semantic versioning order?
[ ] Is your submission signed?
[ ] Is operator icon set?
Updates to existing Operators
[ ] Did you create a ci.yaml file according to the update instructions?
[ ] Is your new CSV pointing to the previous version with the replaces property if you chose replaces-mode via the updateGraph property in ci.yaml?
[ ] Is your new CSV referenced in the appropriate channel defined in the package.yaml or annotations.yaml ?
[ ] Have you tested an update to your Operator when deployed via OLM?
[ ] Is your submission signed?
Your submission should not
[ ] Modify more than one operator
[ ] Modify an Operator you don't own
[ ] Rename an operator - please remove and add with a different name instead
[ ] Modify any files outside the above mentioned folders
[ ] Contain more than one commit. Please squash your commits.
Operator Description must contain (in order)
[ ] Description of the managed Application and where to find more information
[ ] Features and capabilities of your Operator and how to use it
[ ] Any manual steps about potential pre-requisites for using your Operator
Operator Metadata should contain
[ ] Human readable name and 1-liner description about your Operator
[ ] Valid category name1
[ ] One of the pre-defined capability levels2
[ ] Links to the maintainer, source code and documentation
[ ] Example templates for all Custom Resource Definitions intended to be used
[ ] A quadratic logo
Remember that you can preview your CSV here.
--
1 If you feel your Operator does not fit any of the pre-defined categories, file an issue against this repo and explain your need
2 For more information see here
/merge possible
|
gharchive/pull-request
| 2023-10-17T13:57:57 |
2025-04-01T06:40:12.871475
|
{
"authors": [
"framework-automation",
"git-hyagi"
],
"repo": "redhat-openshift-ecosystem/community-operators-prod",
"url": "https://github.com/redhat-openshift-ecosystem/community-operators-prod/pull/3465",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1988387321
|
operator [CI] konveyor-operator
Just switching to use semver-mode.
/merge possible
/merge possible
|
gharchive/pull-request
| 2023-11-10T20:59:55 |
2025-04-01T06:40:12.872605
|
{
"authors": [
"djzager",
"framework-automation"
],
"repo": "redhat-openshift-ecosystem/community-operators-prod",
"url": "https://github.com/redhat-openshift-ecosystem/community-operators-prod/pull/3595",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2046624214
|
operator jaeger (1.52.0)
Thanks for submitting your Operator. Please check the below list before you create your Pull Request.
New Submissions
[ ] Are you familiar with our contribution guidelines?
[ ] Are you familiar with our operator pipeline?
[ ] Have you tested your Operator with all Custom Resource Definitions packaging?
[ ] Have you tested your Operator in all supported installation modes?
[ ] Have you considered whether you want to use semantic versioning order?
[ ] Is your submission signed?
[ ] Is operator icon set?
Your submission should not
[ ] Add more than one operator bundle per PR
[ ] Modify any operator
[ ] Rename an operator
[ ] Modify any files outside the above mentioned folders
[ ] Contain more than one commit. Please squash your commits.
Operator Description must contain (in order)
[ ] Description of the managed Application and where to find more information
[ ] Features and capabilities of your Operator and how to use it
[ ] Any manual steps about potential pre-requisites for using your Operator
Operator Metadata should contain
[ ] Human readable name and 1-liner description about your Operator
[ ] Valid category name1
[ ] One of the pre-defined capability levels2
[ ] Links to the maintainer, source code and documentation
[ ] Example templates for all Custom Resource Definitions intended to be used
[ ] A quadratic logo
Remember that you can preview your CSV here.
--
1 If you feel your Operator does not fit any of the pre-defined categories, file an issue against this repo and explain your need
2 For more information see here
/pipeline restart community-hosted-pipelin
/pipeline restart community-hosted-pipeline
|
gharchive/pull-request
| 2023-12-18T13:04:06 |
2025-04-01T06:40:12.882079
|
{
"authors": [
"iblancasa"
],
"repo": "redhat-openshift-ecosystem/community-operators-prod",
"url": "https://github.com/redhat-openshift-ecosystem/community-operators-prod/pull/3757",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1803228341
|
[DRAFT]: Manifests to install preGA sync operator from custom operator catalog
TODO: Understand where to ubicate the added extra-manifests for day2 installation as a policy.
We need to figure out how to add those manifests as Policies or PoliciesGenTemplate (PGT) objects referenced from the SiteConfig.
|
gharchive/pull-request
| 2023-07-13T15:04:52 |
2025-04-01T06:40:12.884794
|
{
"authors": [
"jnunyez"
],
"repo": "redhat-partner-solutions/vse-carslab-hub",
"url": "https://github.com/redhat-partner-solutions/vse-carslab-hub/pull/85",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1404110242
|
add gcp image list command to CLI
Hey @major @FKolwa ,
I am a bit confused right now. We assign the registered github secret secrets.GOOGLE_APPLICATION_CREDENTIALS to an environment variable GCP_APP_CREDENTIALS, but we never use it.
In the get_google_images function we create an compute_v1.ImagesClient client, is he supposed to use those credentials?
I went to the google docs, but it also didn't really helped me. :)
Can you give me some hints? :sweat_smile:
@miyunari Yup that is correct! It isn't necessary to reference the ENV in code.
In fact you won't find any of the other secrets either (like AWS_ACCESS_KEY_ID).
Most (if not all) cloud provider CLIs authenticate using local configurations that are loaded into ENVs at runtime.
Usually you would use you user credentials to sign into Google Cloud but with an automated service like this a service account can be used.
The fact that "GOOGLE_APPLICATION_CREDENTIALS" is set for the workflow step means that the gcloud cli will be able to read it and use it to authenticate with the cloud provider API. I'll leave you a link in case you want to read more about Google Application Default Credentials (ADC): https://cloud.google.com/docs/authentication/provide-credentials-adc#local-key
A small hint though: The env needs to be mapped to "GOOGLE_APPLICATION_CREDENTIALS" in the workflow context as well for it to properly work.
Hope this helps a little!
Thank you @FKolwa ! That was exactly the information I was looking for :smile:
But now, there is another issue :sweat_smile: Unfortunately it's now unclear to me, how to test my changes.
I got this reference from major, but I don't see the correlation:
https://github.com/redhatcloudx/rhelocator/blob/b97967e77354d0331f0f7bc4a607c00b4b1eea16/tests/test_cli.py#L77-L103
@miyunari Haha yes this is a bit confusing tbh!
The gcloud implementation is pretty rough at this point. get_google_images currently queries all images in the rhel-cloud project and returns everything that isn't deprecated.
In the scope of this ticket, my requirement for an end to end test would be
query the correct api endpoint by calling get_google_images by using 'runner.invoke'
parse the json data
confirm that all images that are returned do not contain the first level key status with the value 'DEPRECATED'
In this case you can copy / paste most of what @major wrote for the azure test!
For the offline test
copy the structure of the e2e test you just wrote
create a new mock for the google images in conftest (you can take a look at the AWS mockups. You need to create a new list of mocked images in a json format and create a new fixture that is passed to you offline test).
Now for the tricky part: How do you know what data structure to expect from the google API?
Well if you call gcloud and query for projects within 'rhel-cloud' you will receive something like this:
{ "architecture": "X86_64", "archiveSizeBytes": "4184623872", "creationTimestamp": "2022-09-20T16:32:45.492-07:00", "description": "Red Hat, Red Hat Enterprise Linux, 9, x86_64 built on 20220920, supports Shielded VM features", "diskSizeGb": "20", "family": "rhel-9", "guestOsFeatures": [ { "type": "UEFI_COMPATIBLE" }, { "type": "VIRTIO_SCSI_MULTIQUEUE" }, { "type": "SEV_CAPABLE" }, { "type": "GVNIC" } ], "id": "2043557223711896434", "kind": "compute#image", "labelFingerprint": "42WmSpB8rSM=", "licenseCodes": [ "7883559014960410759" ], "licenses": [ "https://www.googleapis.com/compute/beta/projects/rhel-cloud/global/licenses/rhel-9-server" ], "name": "rhel-9-v20220920", "rawDisk": { "containerType": "TAR", "source": "" }, "rolloutOverride": { "defaultRolloutTime": "2022-09-25T15:32:42Z", "locationRolloutPolicies": { "zones/asia-east1-a": "2022-09-22T04:32:42Z", .... "zones/us-west4-c": "2022-09-25T04:32:42Z" } }, "selfLink": "https://www.googleapis.com/compute/beta/projects/rhel-cloud/global/images/rhel-9-v20220920", "sourceType": "RAW", "status": "READY", "storageLocations": [ "eu", "asia", "us" ] }
At this point we don't extract any specific information from this returned data (like we do for AWS) and this is not within the scope of your ticket so feel free to create a minimal test mockup version of this data structure that only contains something like
{ "status": "READY" }
Oh boy, that was my mistake. I must have been looking at two things at the same time and put the wrong variable name in the actions workflow. 🤦🏻♂️
@major I think you were right :smile: . I shortened the variable name, because I thought we have to store it in config.py and use it somewhere :woman_facepalming:
Oh, I see I have some merge conflicts, will try to resolve
|
gharchive/pull-request
| 2022-10-11T07:03:43 |
2025-04-01T06:40:12.898352
|
{
"authors": [
"FKolwa",
"miyunari"
],
"repo": "redhatcloudx/rhelocator",
"url": "https://github.com/redhatcloudx/rhelocator/pull/37",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1797902420
|
feat: use uber-go/mock instead of golang/mock
Resolves #297
Thanks @dlsrb6342!
|
gharchive/pull-request
| 2023-07-11T01:19:42 |
2025-04-01T06:40:12.910411
|
{
"authors": [
"dlsrb6342",
"rueian"
],
"repo": "redis/rueidis",
"url": "https://github.com/redis/rueidis/pull/298",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
310341544
|
Feature Request: Add possibility to make new post with custom date-time
I want to move many entries from my old blog to Hubzilla. I would like to set original date for this posts.
Is it possible to add a field for setting a specific date of publication during editing post?
I think we could send an invite once while creating the chatroom. We have a similar functionality for wikis and files already. We could ad a switch labeled Create a status post for this chat
This is possible at post creation time using the 'Delayed Posting' feature. You can set the post date (either forwards or backwards) during post creation. This is not available when editing an existing post because the posts would have already been delivered.
The feature requires an account level of 2; which you can find on your settings/account page; followed by activating the Delayed Posting feature at settings/features -> Post Composition Features
The feature requires an account level of 2
Thanks, i changed account level and found it, it works good.
I'll close this issue.
|
gharchive/issue
| 2018-04-01T17:55:02 |
2025-04-01T06:40:12.914714
|
{
"authors": [
"git-marijus",
"zlaxy",
"zotlabs"
],
"repo": "redmatrix/hubzilla",
"url": "https://github.com/redmatrix/hubzilla/issues/1033",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
717729354
|
update docs
The previous documentation is misleading, as OpenSSL 1.1.0 did not change the fact that contextSetVerificationMode is necessary for OpenSSL to check anything about the certificate. Also, I added a link to the (new) function enabling the OpenSSL built-in hostname validation instead of referring to a PDF.
Hi @amesgen, thanks for the PR.
You absolutely right, contextSetVerificationMode has to be called even with OpenSSL >= 1.1.0. When I changed that part of the documentation a few weeks ago I wasn't careful enough.
My only concern with the new phrasing that you propose is that I think it would be better to keep the note about old (<1.1.0) versions of OpenSSL not verifying automatically the peer's hostname, even with contextSetVerificationMode. How about if we added the following at the end of your version:
In addition to that, if you are using a version of OpenSSL older than 1.1.0, you also need to verify that the peer's hostname matches with one specified in the certificate. You can find information on how to do that here. Newer versions of OpenSSL (>=1.1.0) perform that check automatically.
or something like that.
Hmm, I am no expert in OpenSSL, but I think that, even in OpenSSL >= 1.1.0, setting contextSetVerificationMode to VerifyPeer only checks that there is a valid certificate chain (how should OpenSSL know about the correct hostname? It is not passed as an argument to contextSetVerificationMode.).
Hostname validation has to be enabled in addition to this, either manually or with the "new" functions added in OpenSSL 1.1.0 and 1.0.2, which I recently added to HsOpenSSL: enableHostnameValidation and vshabanov/HsOpenSSL#52, see https://wiki.openssl.org/index.php/Hostname_validation.
|
gharchive/pull-request
| 2020-10-08T23:07:53 |
2025-04-01T06:40:12.925479
|
{
"authors": [
"amesgen",
"redneb"
],
"repo": "redneb/HsOpenSSL-x509-system",
"url": "https://github.com/redneb/HsOpenSSL-x509-system/pull/3",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
697055509
|
Added screenshots for some examples
Context:
Adds screenshots of examples.
I didn't manage to get the same window theme as seen in the calculator.
I took the pictures on win 10, I hope it's fine (at least it proves the lib is cross-platform :)
Fixes #318
Thank you. I think the basic ideas was to embed the images like the screenshot on the README. Would be nice if you can adjust it.
Images looks very very strange because there is not new line before and they seems to be aligned randomly
Yes, I see now.
I have checked the result in Visual Studio with a markdown plugin.
Seems like GitHub and the plugin render the markdown test differently.
|
gharchive/pull-request
| 2020-09-09T18:41:42 |
2025-04-01T06:40:12.929094
|
{
"authors": [
"FloVanGH",
"kivimango",
"qarmin"
],
"repo": "redox-os/orbtk",
"url": "https://github.com/redox-os/orbtk/pull/331",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1927292706
|
[Bug] ...
Platform
Device: Phone, Laptop, Desktop, TV... windows 10 edge browser
OS: Windows, Android, iOS...
Browser: Chrome, Safari, Kiwi...
Browser Version:
Better xCloud Version:
Describe the bug
...doesnt work on gta oline, it says the session has expired
Screenshots/Videos
If applicable, add screenshots/videos to help explain your problem.
Report it to xCloud
|
gharchive/issue
| 2023-10-05T03:01:28 |
2025-04-01T06:40:13.018716
|
{
"authors": [
"Nachhito",
"redphx"
],
"repo": "redphx/better-xcloud",
"url": "https://github.com/redphx/better-xcloud/issues/143",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
425289256
|
module is not defined
When building with webpack 5, this line gives a ReferenceError for module: https://github.com/erikras/redux-form/blob/f209539ea65e1b46534e701c635df31c100c3259/src/util/isHotReloading.js#L3
I'm not sure how to fix this, normally I would just put a if (typeof module === 'undefined') return false above that line, but after the webpack build it just gets transformed into if (false) {} so maybe this is a webpack 5 bug.
Update: Simply removing the castModule and using module directly makes it work properly. So it's a matter of the TypeScript workaround making it not work with Webpack 5.
Another fix is prefixing module with globalThis. I don't know Flow enough to create a merge request that passes the type system. But using patch-package I could simply change this line to get my app working with Vite.
var castModule = globalThis.module
Another way I got it working was to define
module: undefined
in vite.config. However this might cause side-effects.
|
gharchive/issue
| 2019-03-26T08:44:27 |
2025-04-01T06:40:13.023757
|
{
"authors": [
"sbaechler",
"wmertens"
],
"repo": "redux-form/redux-form",
"url": "https://github.com/redux-form/redux-form/issues/4412",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
753886108
|
Export all flow type aliases from root
We need to be able to use all of them in userland, absolutely any type alias declared anywhere in the flow type defs. Making wrapper components is a huge hassle without these, because it's difficult to type props for our wrapper components.
import type { FieldInputProps, FieldArrayInputProps, Validator /* etc */ } from 'redux-form'
I'll try to make a PR for this soon
|
gharchive/issue
| 2020-11-30T23:47:32 |
2025-04-01T06:40:13.024998
|
{
"authors": [
"jedwards1211"
],
"repo": "redux-form/redux-form",
"url": "https://github.com/redux-form/redux-form/issues/4703",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1741960230
|
mapStateToProps should not be called with functional component rendering a connected class component
What version of React, ReactDOM/React Native, Redux, and React Redux are you using?
React: 18.2.0
ReactDOM: 18.2.0
Redux: 4.2.1
React Redux: 8.0.5
What is the current behavior?
Description:
Depending on a parent component being a functionnal or class component, children connected component do not see their mapStateToProps function called in the same way
Reproduction:
https://codesandbox.io/s/beautiful-gianmarco-wo67u0
Click on toggle and check the console logging
we can see a call to mapStateToProps in the class component child, when the parent component is a functional component ([APP1]) but not when the parent is a class component ([APP2])
Here is the logging, by app, when toggling off connected children (conditional rendering):
With a functional component as parent:
[APP1] CLASS CHILD MAP STATE TO PROPS
[APP1] NORMAL CHILD SELECTOR
x2 [APP1] NORMAL PARENT SELECTOR
[APP1] RENDER PARENT {showChild: false}
[APP1] NORMAL PARENT SELECTOR
[APP1] RENDER PARENT {showChild: false}
With a class component as parent:
[APP2] CLASS PARENT MAP STATE TO PROPS
x2 [APP2] RENDER PARENT {showChild: false}
[APP2] NORMAL CHILD SELECTOR
Note:
FYI this led to an issue in our code when refactoring a component, with uncaught error in the mapStateToProps function of a child of that component, because suddenly it was being called with unexpected state and trying to get state on undefined keys.
Maybe all of this is a normal consequence of the inherent difference in React rendering functional or class component, but I still wanted to document this in an issue for the record, and see if you have any comment.
Thanks!
What is the expected behavior?
mapStateToProps should be called consistently (or should it not?)
Which browser and OS are affected by this issue?
Chrome OSX
Did this work in previous versions of React Redux?
[ ] Yes
I'm not sure what you're trying to describe here, tbh.
At my first guess (just reading the description), I assume that the difference in rendering is not a bug, but simply a question of which components are re-rendering and which are not, per https://blog.isquaredsoftware.com/2020/05/blogged-answers-a-mostly-complete-guide-to-react-rendering-behavior/#standard-render-behavior .
I have tried to explain as accurately as possible and provided a codesandbox, but if it is unclear or need more details, I'll be happy to provide them.
What I'm pointing is that the two examples in that demo are rendering exactly the same thing (supposedly) but mapState is called on one side and not on the other, which was unexpected and led to a bug, so I'm reporting on that.
@pandaiolo yeah, afraid I don't have time to look at this right now (busy with day job stuff).
But, my immediate guess is that:
When the parent is a class component, it avoids re-rendering the child, and the child's mapState is thus not called
When the parent is a function component, it does re-render the child, and thus the child's mapState is called
which sounds like a combination of how React itself works in general for rendering rather than a particular bug in React-Redux.
I'll try to look into this at some point, but not sure when.
Sure, thank you!
|
gharchive/issue
| 2023-06-05T14:56:46 |
2025-04-01T06:40:13.039989
|
{
"authors": [
"markerikson",
"pandaiolo"
],
"repo": "reduxjs/react-redux",
"url": "https://github.com/reduxjs/react-redux/issues/2024",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
438610644
|
Update ThreePrinciples.md
in the state explanation read-only is totally missed, need focused few words on it.
Strongly applies when programmer codes, suppose to write getter methods not setters, in simple words for beginners.
Hi. Appreciate the PR, but I don't see a need to change the wording at this time. Thanks anyway!
|
gharchive/pull-request
| 2019-04-30T05:38:38 |
2025-04-01T06:40:13.041727
|
{
"authors": [
"kambleaa007",
"markerikson"
],
"repo": "reduxjs/redux",
"url": "https://github.com/reduxjs/redux/pull/3416",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
490573105
|
Move the examples to a separate repo
I'd like to move the examples out of the repo. They add a lot of time to the build process and generally don't change significantly. They also don't require the source copy of Redux, so it's not like they are a canary for changes we make on master.
Moving them to their own home means it's more reasonable to accept new additions, which is something we've been hesitant to do for a long while (years!).
I've set them up in a new repo: reduxjs/examples
Is this a terrible idea? Anything we should do besides updating the docs and creating a new README in the examples path (which is temporarily gone at the moment).
Also of note:
Typical build time today: 3:30-4:10
This PR build time: 1:10
That's a 3-3.5x faster build time.
Thoughts and concerns:
The approach shown here and in the other repo appears to blow away all the accumulated Git history. Could we at least do some git filter-branch work to extract the examples folder history as the starting point of the new repo, so that we can maintain that history?
We have a bunch of links in the docs, and there's probably also external links:
Yeah, we'd need to do a bunch of fixing up for our docs.
External links would result in a 404. Can we at least leave placeholder README files that point to the new location?
What additional examples would you want to have in the future repo?
Out of curiosity, what about the examples is slowing down the build process?
There are so many of them and they all require separate npm ci commands.
But why are we even building / testing them at all as part of the main lib? Where are they getting kicked off?
https://github.com/reduxjs/redux/blob/master/.travis.yml#L5-L6
Oh. Well, at a minimum, we should remove those lines from Travis. No reason to be checking those on every commit.
Redid the examples repo with history intact: https://github.com/reduxjs/examples
I'm still really not seeing a point to this. If we're concerned about Travis build times for some reason, let's just configure Travis to not worry about the examples on every build or something.
I tried to figure out how to conditionally build certain examples based on the changed files using Travis' environment variables, but unfortunately it seems difficult to do this consistently. For example TRAVIS_COMMIT_HASH doesn't only show the commits from the PR which would cause false positives where examples are build when they shouldn't be, and if multiple commits are pushed outside of a pull request TRAVIS_COMMIT would only show the latest commit which could cause examples to break without being rerun in CI.
I think the safest way to make our builds faster is to remove the examples (like with this pull quest, but possibly replacing the existing files with links to the new repository) and move the example scripts to the other repository. Alternatively, we might be able to speed up our multiple package installations using monorepos with yarn workspaces or lerna, but I don't have experience with them.
We never did move these out, and tbh I'd like to keep them in here. Closing.
|
gharchive/pull-request
| 2019-09-07T01:01:10 |
2025-04-01T06:40:13.050332
|
{
"authors": [
"markerikson",
"nickmccurdy",
"timdorr"
],
"repo": "reduxjs/redux",
"url": "https://github.com/reduxjs/redux/pull/3565",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
568106110
|
Improve the example in GettingStarted.
name: "New/Updated Documentation Content"
about: Adding a new docs page, or updating content in an existing docs page
Hello,
Thanks so much for this great project. I've just started leaning redux, and found a little bit of difficulty in the GettingStarted example.
I hope you will like it and please let me know if there are rooms for improvement in my PR.
PR Type
update an existing page.
Checklist
[ ] Is there an existing issue for this PR?
link issue here
[x] Have the files been linted and formatted?
What docs page is being added or updated?
Section:Introduction
Page:Getting Started with Redux
For Updating Existing Content
What updates should be made to the page?
The import part is a bit difficult for newbies to setup correctly.
Actually I failed to make this example work.
$ node example.js
(node:20108) ExperimentalWarning: The ESM module loader is experimental.
file:///private/tmp/my-redux/example.js:1
import { createStore } from 'redux'
^^^^^^^^^^^
SyntaxError: The requested module 'redux' does not provide an export named 'createStore'
at ModuleJob._instantiate (internal/modules/esm/module_job.js:92:21)
at async ModuleJob.run (internal/modules/esm/module_job.js:107:20)
at async Loader.import (internal/modules/esm/loader.js:164:24)
I know it's not a problem in Redux, but ideally example code should work without effort.
Do these updates change any of the assumptions or target audience? If so, how do they change?
Target audience:
a newbie to overall ecosystem of JavaScript (JS/npm/Redux)
Assumptions:
They can create an html and a js file by hand, and open the html by their browser.
Expected Results:
They'll feel more confident that they can move forward by the first success of the working example code.
Thanks, but we're not looking to create a working example here per se, just some example code of how a Redux store might be written. If you click over to the Examples page in the next section, we do have a working example of this exact code.
|
gharchive/pull-request
| 2020-02-20T07:32:55 |
2025-04-01T06:40:13.056669
|
{
"authors": [
"DQNEO",
"timdorr"
],
"repo": "reduxjs/redux",
"url": "https://github.com/reduxjs/redux/pull/3709",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
618598023
|
ReactJS - setState causes viewer to re-render every time
When using Resium with ReactJS and adding an ImageLayer, it is causing the viewer to refresh every time any setState is called. How can I use setState to update certain parts of the page without re-rendering the map?
Example below and screenshot of the issue
`
class LiveMap extends React.PureComponent {
viewerRef = React.createRef();
state = {
name: "test"
};
handleSelectedEntityChanged = (event) => {
console.log("selected: " + event);
this.setState({
...this.state,
name: this.state.name === "test" ? "testt" : "test"
});
};
render() {
const defaultImageryProvider = new UrlTemplateImageryProvider({
url : "https://cartodb-basemaps-{s}.global.ssl.fastly.net/dark_all/{z}/{x}/{y}.png"
});
return (
<Viewer
full
ref={this.viewerRef}
onSelectedEntityChange={this.handleSelectedEntityChanged}>
<ImageryLayer imageryProvider={defaultImageryProvider} />
<Entity
name={this.state.name}
position={Cartesian3.fromDegrees(61, 130, 100)}
point={{ pixelSize: 10 }} />
</Viewer>
);
}
};
`
You are constructing a new imagryProvider on every render. It looks like it does not depend on any props or state though, so you should be able to just build it once inside the constructor.
I am getting similar functionality with useEffect. I have an API call in my useEffect function which gets called every 5 seconds. With the data from the API I can set some state variables using "setMyVar()". This then updates the state variables and redraws a point in the Viewer. The weird thing is, this updates resets the clock.
I think this is because useEffect re-renders the view with every call. I am not sure if this is how this should work or if I am using state incorrectly with Resium.
Try to move new UrlTemplateImageryProvider(...) outside the render function.
Before:
render() {
const defaultImageryProvider = new UrlTemplateImageryProvider({
url : "https://cartodb-basemaps-{s}.global.ssl.fastly.net/dark_all/{z}/{x}/{y}.png"
});
After:
defaultImageryProvider = new UrlTemplateImageryProvider({
url : "https://cartodb-basemaps-{s}.global.ssl.fastly.net/dark_all/{z}/{x}/{y}.png"
});
render() {
// ...
<ImageryLayer imageryProvider={this.defaultImageryProvider} />
I think I have the same issue. When I update variable in parent that holds Entities as children, they all get redrawn
Same as you guys, when state of my component with Viewer is changed (with useState() or useEffect() hooks) - ImageLayer is reloaded. Is there some way how to fix this?
imageProvider prop of ImageryLayer component is read-only and is not changeable because ImageProvider of ImageryLayer of Cesium API is also read-only.
If the imageProvider is static, you can use outside const variable or useMemo hook as described in this page.
If you want to change the imageProvider dynamically, you cannot avoid regenerating the ImageryLayer. If you want to completely suppress warnings, key props can be used:
const imageProvider = useMemo(() => new UrlTemplateImageryProvider({ url }), [url]);
return <ImageryLayer key={url} imageryProvider={imageProvider} />;
We are also considering suppressing the warnings. Either way, please understand that the imageProvider is read-only.
|
gharchive/issue
| 2020-05-14T23:35:31 |
2025-04-01T06:40:13.070307
|
{
"authors": [
"Chingili",
"benwiles1",
"digennnj",
"mhaque02",
"rot1024",
"vlkpa"
],
"repo": "reearth/resium",
"url": "https://github.com/reearth/resium/issues/372",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
143764757
|
Suppressed the result of expressions at the end of test blocks
Fix #34, #47
Converted the optional expr at the end of test blocks to a statement so that test blocks can end with an unsuppressed expression (including loop statements).
A bit of a mess (lots of cloning) because of ownership issues. Suggestions are welcome.
@reem Thoughts?
Hey there, I volunteered to help out @reem with this crate. The code is probably not the prettiest ever as you said but it would be great to have this fixed so I'd be willing to merge this.
Could you also rebase this on the latest master so that we are completely sure that it doesn't break any of the existing tests?
Thanks for the changes. 👍
|
gharchive/pull-request
| 2016-03-27T00:50:29 |
2025-04-01T06:40:13.105906
|
{
"authors": [
"cramertj",
"ujh"
],
"repo": "reem/stainless",
"url": "https://github.com/reem/stainless/pull/54",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
184636201
|
Fix for rustc 1.14.0-nightly (f09420685 2016-10-20).
Fixes #71.
See https://github.com/rust-lang/rust/commit/029dceedb9719ae5dbdbf2c033c920017e3d786e.
👍 Thank you.
|
gharchive/pull-request
| 2016-10-22T15:24:00 |
2025-04-01T06:40:13.107325
|
{
"authors": [
"ujh",
"utkarshkukreti"
],
"repo": "reem/stainless",
"url": "https://github.com/reem/stainless/pull/72",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
247791176
|
Rails5
adds rails5 compatibility via upstream merge of collectiveideas master.
lgtm ??
|
gharchive/pull-request
| 2017-08-03T18:12:41 |
2025-04-01T06:40:13.108111
|
{
"authors": [
"meesterdude",
"nhance"
],
"repo": "reenhanced/audited",
"url": "https://github.com/reenhanced/audited/pull/4",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
52580886
|
remote forms doesn't work after upload:complete event
When you try to submit a form with remote: true jquery-ujs checks if it's had any non blank file inputs https://github.com/rails/jquery-ujs/blob/master/src/rails.js#L422, so the remote option is ignore and the form is submitted with a normal POST.
My solution was to modify the if in this line https://github.com/elabs/refile/blob/master/app/assets/javascripts/refile.js#L37 to:
if((xhr.status >= 200 && xhr.status < 300) || xhr.status === 304) {
var id = input.getAttribute("data-id") || JSON.parse(xhr.responseText).id;
input.previousSibling.value = id;
input.removeAttribute("name");
input.value = ''; // <-- included this
dispatchEvent("upload:success", xhr.responseText);
}
This make the input file empty, so for jquery-ujs all the file fields are blank and the form is submitted remotely.
I will be happy to make a pull request if this approach is considered correctly.
I'm the same problem =\
This gem https://github.com/JangoSteve/remotipart resolved the problem with upload file with remote: true
Is there any way to get this to work properly? I tried @cerdiogenes code and it didn't work.
@danieldocki I looked at the remotipart gem. I wouldn't use it. It is looking very long in the tooth and somewhat neglected. I'm paranoid about adding too much magic, especially when it involves adding new middleware to my stack. Seems like overkill anyway to work around this issue as it triggers submitting forms through a hidden iframe, which is not longer needed thanks to refile! Keep it simple.
We just need some way to tell Rails UJS to not count refile handled file inputs when determining whether to abort the AJAX form submission.
I poked around through the remotipart gem to see how it works. I found it relies on handling the ajax:aborted:file event fired by Rails UJS when it aborts due to non-empty file inputs.
I was able to come up with the event handler below which tells Rails UJS to proceed with submitting the form via Ajax if all the non-empty file inputs are associated with refile.
I used the data-reference attribute to detect the presence of refile handling. @jnicklas: is that a safe bet?
$(document).on('ajax:aborted:file', 'form', function(){
var form = $(this);
// if all non-blank file inputs are handled by refile,
// then resume
var nonBlankFileInputs = $.rails.nonBlankInputs(form, $.rails.fileInputSelector);
if (nonBlankFileInputs.filter(':not([data-reference])').length === 0) {
$.rails.handleRemote(form);
return false;
} else {
return true;
}
});
@masonhale I think that's a fine workaround. Someone (not me ;)) should really patch rails_ujs to do the right thing though.
@jnicklas I was thinking about how to patch rails_ujs, and it occurs to me that it needs some way to know which file input fields should be excluded from its check of non-blank file inputs. I would propose that it looks for some specific attribute on the input as a flag to indicate it should be ignored. Something like data-rails-ajax-ignored=true. This way an application author could manually add the attribute to exclude it from the file input check (and potentially the separate required field input check), and rails-friendly file upload libraries like refile could potentially add the ignore attribute automatically to help make things work automatically. Any thoughts?
No, the reason for exclusion should be that the field is disabled. Disabled fields aren't submitted by browsers, rails_ujs shouldn't care that there's a disabled field with a file attached in the form.
@jnicklas thanks for the clarification. I didn't realize Refile disabled fields and that would be a good method to filter file input fields. Question: when does this disabling happen? Is it possible the rails_ujs script could fire before Refile sets the field to disabled?
@jnicklas Looking more closely at how the refile.js operates, I notice it removes the name of the input field after the file is processed here. I don't see where it sets the input file to disabled. Am I missing something? It makes me wonder if ignoring file inputs with no name attribute would work as well or better than ignoring disabled fields? (To be clear, I'm asking for clarification because I intend to submit a PR to fix rails_ujs. Thanks for the help).
PS: Thanks also for all the work you've put into Refile. I think it is really great work and I am very thankful you've put in the effort and hard-won experience to make it. Thank you!
@masonhale uuhh I was remembering it wrong. We do indeed remove the name to not submit the field. Browsers don't submit it, so rails_ujs should not accout for it, but it does.
Ok, based on what I've learned by working on a Rails UJS patch, I've come up with a better/simpler monkey-patch to get Refile and Rails UJS working together. This avoids having to catch the aborted file event. Much cleaner. It simply tweaks the fileInputSelect UJS uses to make sure the input has a name. (Interestingly, Rails UJS already has a similar check in place for required fields).
// run this anytime after rails / jquery-ujs is loaded
$.rails.fileInputSelector = 'input[name][type=file]';
@jnicklas actually does Rails UJS not include disabled inputs since v. 1.1.0 (https://github.com/rails/jquery-ujs/pull/430). Could we patch refile to disable the input instead of just removing the name attribute?
@jokklan no this does not work. Disabling the input not only has a visual effect we do not want, but it also prevents the user from selecting another file for upload.
It's really a simple fix in Rails UJS, which they should make anyway, someone should really submit a patch. It's not going to be me though, since I don't actually ever use rails_ujs ;)
I just submitted a pull request to jquery-ujs for this issue: https://github.com/rails/jquery-ujs/pull/464
I meant to do it before the holidays, but got busy/distracted and didn't get back to it.
You're a hero @masonhale! :smile:
@masonhale :metal:
@danieldocki @jnicklas Can you do a favor and comment on the jquery-ujs pull request I opened to draw more attention to it? https://github.com/rails/jquery-ujs/pull/464
@masonhale I added a comment. It's not been that long though, OSS moves slowly ;) They'll get to it eventually.
|
gharchive/issue
| 2014-12-21T00:09:13 |
2025-04-01T06:40:13.131672
|
{
"authors": [
"JoshuaNovak919",
"cerdiogenes",
"danieldocki",
"jnicklas",
"jokklan",
"masonhale"
],
"repo": "refile/refile",
"url": "https://github.com/refile/refile/issues/58",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1244285152
|
Add collapse-duplicated-reviews feature
Fixes #5394
Test URLs
https://github.com/refined-github/sandbox/pull/10#pullrequestreview-958560681
Screenshot
https://user-images.githubusercontent.com/46634000/169702068-8bfe4f63-09b6-45c1-b2a1-0dee4d717f86.mp4
I'd rather collapse the whole thread into the title (so it's visible) and link the title to the real comment.
With the current PR, reacting to a comment means uncollapsing it and then clicking the permalink. Let's skip one step.
I'd rather collapse the whole thread into the title
By "title", do you mean the text next to the "review event"? But what happens if the review is made up of two duplicated comments, each responding to a different review thread?
By "title", do you mean the text next to the "review event"?
I mean this:
instead of this:
The latter is not actionable and it's out of context. That comment should not appear there, but the review should still keep the link to the real review comment, as shown in the first screenshot.
By title I mean filename
This is already what happens, see video above. We could avoid showing the "Show"/"Hide" text & icon, but since we hide something GitHub shows by default we should probably make it obvious it can be expanded.
Unless you want to drop the review thread completely?
and link the title to the real comment
+1
@fregante wdyt about this:
Instead of using the file name we make it clear this is a link to another comment (plus it's one less select because we don't need to grab the <summary>).
Closing until you have time to pick it up. This feature is just:
observe('collapsedReview', review => {
if (isDuplicate(review)) {
const link = select('a', review)
details.replaceWith(link)
link.classList.add(...details.classList, ...otherClasses)
}
})
|
gharchive/pull-request
| 2022-05-22T15:12:23 |
2025-04-01T06:40:13.138920
|
{
"authors": [
"cheap-glitch",
"fregante"
],
"repo": "refined-github/refined-github",
"url": "https://github.com/refined-github/refined-github/pull/5656",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2729750494
|
🛑 SmartProtection is down
In 1be32fd, SmartProtection (https://smartprotection.com) was down:
HTTP code: 404
Response time: 297 ms
Resolved: SmartProtection is back up in 7003f29 after 10 minutes.
|
gharchive/issue
| 2024-12-10T10:47:22 |
2025-04-01T06:40:13.178737
|
{
"authors": [
"reinaldoleon"
],
"repo": "reinaldoleon/monitoring",
"url": "https://github.com/reinaldoleon/monitoring/issues/333",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
560126621
|
ppo save expert demo
hi, how am i supposed to save expert demo in ppo main?
PPO is a method of reinforcement learning. However app, maxent and gail are all inverse reinforcement learning method. Due to the emergence of policy-based inverse reinforcement learning algorithms, you can use PPO with any inverse reinforcement learning algorithm to complete the training.
References:
Ng A Y, Russell S J. Algorithms for inverse reinforcement learning[C]//Icml. 2000, 1: 2.
Ho J, Gupta J, Ermon S. Model-free imitation learning with policy optimization[C]//International Conference on Machine Learning. PMLR, 2016: 2760-2769.
|
gharchive/issue
| 2020-02-05T04:03:47 |
2025-04-01T06:40:13.180485
|
{
"authors": [
"francisduan",
"gitouni"
],
"repo": "reinforcement-learning-kr/lets-do-irl",
"url": "https://github.com/reinforcement-learning-kr/lets-do-irl/issues/6",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1896122711
|
🛑 BINUSMAYA Praktikum (SLC Bluejack) is down
In 49bbdda, BINUSMAYA Praktikum (SLC Bluejack) (https://bluejack.binus.ac.id/binusmayalab/) was down:
HTTP code: 502
Response time: 1224 ms
Resolved: BINUSMAYA Praktikum (SLC Bluejack) is back up in 1070ff7.
|
gharchive/issue
| 2023-09-14T09:13:05 |
2025-04-01T06:40:13.183030
|
{
"authors": [
"1010bots"
],
"repo": "reinhart1010/binusmayadown",
"url": "https://github.com/reinhart1010/binusmayadown/issues/2921",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2286834682
|
🛑 GreatNusa.com is down
In 12ab2d7, GreatNusa.com (https://greatnusa.com) was down:
HTTP code: 403
Response time: 120 ms
Resolved: GreatNusa.com is back up in 76719e8.
|
gharchive/issue
| 2024-05-09T03:46:58 |
2025-04-01T06:40:13.186333
|
{
"authors": [
"1010bots"
],
"repo": "reinhart1010/binusmayadown",
"url": "https://github.com/reinhart1010/binusmayadown/issues/6058",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2329251001
|
🛑 GreatNusa.com is down
In 3f36256, GreatNusa.com (https://greatnusa.com) was down:
HTTP code: 403
Response time: 118 ms
Resolved: GreatNusa.com is back up in 114ac4a.
|
gharchive/issue
| 2024-06-01T17:26:21 |
2025-04-01T06:40:13.189357
|
{
"authors": [
"1010bots"
],
"repo": "reinhart1010/binusmayadown",
"url": "https://github.com/reinhart1010/binusmayadown/issues/6788",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1683207136
|
Client: Data export not respecting archived parameter
It seems the client.export_study_data endpoint does not adequately respect the archived parameter, returning archived reports when it is set to False. For now this means we can't filter on archived at the API, leading to a problem when exporting study data.
Currently we extract all data and then filter it ourselves.
Call has been opened.
# DATA ANALYSIS
def export_to_dataframe(self, archived=False) -> dict:
"""Exports all data from a study into a dict of dataframes for statistical analysis."""
# TODO: change this to the correct archived,
# fails now because the parameter does not seem to be handled correctly server side
self.map_data(archived=True)
dataframes = {
"Study": self.__export_study_data(archived),
"Surveys": self.__export_survey_data(archived),
"Reports": self.__export_report_data(archived),
}
return dataframes
This issue will be fixed with the next CDMS maintenance release (2023.2.2.0), which is planned for end of June.
|
gharchive/issue
| 2023-04-25T13:44:34 |
2025-04-01T06:40:13.191403
|
{
"authors": [
"reiniervlinschoten",
"slknijnenburg"
],
"repo": "reiniervlinschoten/castoredc_api",
"url": "https://github.com/reiniervlinschoten/castoredc_api/issues/89",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1119934882
|
Decide what to do about auto-updates
See: https://github.com/actions/runner/issues/1396#issuecomment-983783264 also related to #16
This was transitively addressed via latest "version" and this PR: #39. Closing.
|
gharchive/issue
| 2022-01-31T21:09:29 |
2025-04-01T06:40:13.203992
|
{
"authors": [
"ravwojdyla"
],
"repo": "related-sciences/gce-github-runner",
"url": "https://github.com/related-sciences/gce-github-runner/issues/17",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1265857014
|
Update serializers to include new Locality data
After adding new models (Locality, LocalityStack and BibItemLocality) (see https://github.com/relaton/relaton-py/issues/18) to accommodate more DOI data, we want to include them when serialising a BibliographiItem.
@stefanomunarini, so doi2ietf (doilit) used to serialize this into <seriesInfo>—but it suddenly turned out to be wrong (violates XML specification). I stated the task as “reimplement the same handling as in doi2ietf/doilit”, so you have done what was requested—the mistake was at planning stage.
Apparently, we are supposed to use <refcontent>, based on:
Jean’s comment https://github.com/ietf-ribose/bibxml-service/issues/228#issuecomment-1177699864
The spec on <refcontent>
(cc @ronaldtse)
Note that <refcontent> data model is apparently plain text, so in the serializer we might need to format the entire locality hierarchy as one human-readable string.
It’s up to you how to attack this, but some considerations:
It’s probably worth encapsulating it in a separate, testable function that takes locality info and outputs a string (of course, if warranted, the logic can also be split further into smaller functions).
This function doesn’t have to cover all possible locality data from the start. We can increase coverage as we go, using existing samples as minimum baseline.
If we can write this nicely without extra requirements, that’d be ideal.
However, if this is a problem (f-strings don’t have conditionals, after all…), to prevent this logic from becoming a mess it may be worth adding a lightweight templating library to package requirements. I don’t really know a good one (Jinja2 seems overkill?), so let me know if you want to do this.
|
gharchive/issue
| 2022-06-09T09:31:42 |
2025-04-01T06:40:13.209169
|
{
"authors": [
"stefanomunarini",
"strogonoff"
],
"repo": "relaton/relaton-py",
"url": "https://github.com/relaton/relaton-py/issues/22",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1817566663
|
two big bugs! no support for "break" and inline loop
Describe the bug*
I am confront with two bugs, key word "break" in "for loop" and inline loop are not supported in hot reload run. It output the error info: "Process finished with exit code -1073741819 (0xC0000005)".
here is the test code:
site=[0,2,2,2]
site=[ii for ii in site if ii!=0]
print('test')
for ii in site:
if ii==2:
break
print('finished')
when in debug mode, this code can run well in the first time, when something changed, such as print('somethin others'), it will quit directly when hot reload is triggered, and output the upper error code.
Desktop or remote (please complete the following information):**
OS: [Windows10, debian9]
Reloadium package version: [1.2]
Editor: [PyCharm]
Python Version: [3.9 and 3.10]
Python Architecture: [ 64bit]
Run mode: [Debug]
Additional context
Add any other context about the problem here.
Hi @jialay
Good catch! Thanks for opening this issue.
I've managed to fix the break issue but cannot reproduce the "inline loop" problem.
Do you mean the line comprehension in the second line?
Also could you provide the full version of your python interpreter (eg 3.9.8 )?
Hi, dkrystki, thank you for the great tool reloadium!
yes, the inline loop is just show as the second line. I tried it with python 3.9.12( debian 9, remote debug mode), and 3.10.0(windows 10, local mode), both of the them can't support hot reload。
such as the following code:
site=[0,2,2,2]
site=[ii for ii in site if ii!=0]
print('test')
print('finished')
debug with a breakpoint at line "print('finished')", any change of the code will interrupte the debug.
|
gharchive/issue
| 2023-07-24T04:07:27 |
2025-04-01T06:40:13.222783
|
{
"authors": [
"dkrystki",
"jialay"
],
"repo": "reloadware/reloadium",
"url": "https://github.com/reloadware/reloadium/issues/158",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1079511351
|
[BUG] ScrollView组件里slot无效
bug 描述
ScrollView组件slot无效
期望结果
如下截图所示
复现代码
可复现问题的代码仓库
关键代码
<View className="placeholder">placeholder</View>
<ScrollView scrollY refresherEnabled refresherDefaultStyle="none">
<View className="refresher" slot="refresher">
refresher
</View>
<View className="body">body</View>
</ScrollView>
版本信息:
remax 版本: 2.9.1
开发环境 win10,微信开发者工具
其他信息
微信原生可以
remax不可以
这个有进展吗,遇到同样的问题
这个有进展吗,遇到同样的问题
没有,当时我是打算搞个下拉刷新的loading。最后放弃了,直接用小程序原生的整页下拉刷新了。
bug 描述 ScrollView组件slot无效
期望结果 如下截图所示
复现代码 可复现问题的代码仓库 关键代码
<View className="placeholder">placeholder</View>
<ScrollView scrollY refresherEnabled refresherDefaultStyle="none">
<View className="refresher" slot="refresher">
refresher
</View>
<View className="body">body</View>
</ScrollView>
版本信息:
remax 版本: 2.9.1
开发环境 win10,微信开发者工具
其他信息
微信原生可以
remax不可以
目前不支持 slot 属性
|
gharchive/issue
| 2021-12-14T09:25:11 |
2025-04-01T06:40:13.242072
|
{
"authors": [
"fhyoga",
"ramen-curator",
"watsonhaw5566"
],
"repo": "remaxjs/remax",
"url": "https://github.com/remaxjs/remax/issues/1787",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
527875225
|
Fix/missing macros
fix https://github.com/remaxjs/remax/issues/382#issuecomment-558000230
修复 babel-plugin-macros 找不到的错误
要改一下,要从开发者的角度去写
修复 babel-plugin-macros 找不到的错误
要改一下,要从开发者的角度去写
康康
|
gharchive/pull-request
| 2019-11-25T06:23:18 |
2025-04-01T06:40:13.244454
|
{
"authors": [
"Darmody",
"yesmeck"
],
"repo": "remaxjs/remax",
"url": "https://github.com/remaxjs/remax/pull/406",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1977070674
|
Create an example of an app with nested routes
I'm trying to organize my routes using route folders, but I can't seem to get it to work. I came to the examples repo to find a good example, but none of the examples show how to use route folders. Could someone who knows how Remix route folders works create an example? Thanks so much.
Hi @elliotaplant,
This is a topic extensively covered by the documentation: https://remix.run/docs/en/main/file-conventions/routes#nested-routes
I suggest opening a Q&A conversation in the remix repository if you need guidance, showing us how you setup things and what's not working for you? I'd be happy to help you!
|
gharchive/issue
| 2023-11-04T00:16:27 |
2025-04-01T06:40:13.251334
|
{
"authors": [
"elliotaplant",
"machour"
],
"repo": "remix-run/examples",
"url": "https://github.com/remix-run/examples/issues/376",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1062977694
|
chore: fix ESLint errors & warnings
Fixes #625
Fixes #805
Thanks!
|
gharchive/pull-request
| 2021-11-24T22:51:02 |
2025-04-01T06:40:13.270038
|
{
"authors": [
"MichaelDeBoey",
"kentcdodds"
],
"repo": "remix-run/remix",
"url": "https://github.com/remix-run/remix/pull/611",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.