id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
596764319
|
Exception: Latex error converting to dvi.
Hi. When I installed Manim, I first tried to render an animation to see if it works. I first issued the command below to render the Square to Circle example, and it did well.
python3 ./manim.py example_scenes.py SquareToCircle -pl
But, whenever I try to render the other examples provided, it always makes the same error.
Exception: Latex error converting to dvi. See log output above or the log file: media/Tex/664e80e8d8d98482.log
Command issued : sudo -H python3 -m manim example_scenes.py OpeningManimExample -pl
This is the complete console log :
Media will be written to ./media/. You can change this behavior with the --media_dir flag.
latex: warning: running with administrator privileges
Sorry, but latex did not succeed.
The log file hopefully contains the information to get MiKTeX going again:
/var/root/Library/Application Support/MiKTeX/texmfs/data/miktex/log/latex.log
Traceback (most recent call last):
File "/Users/mac/Downloads/manim-master/manimlib/extract_scene.py", line 155, in main
scene = SceneClass(**scene_kwargs)
File "/Users/mac/Downloads/manim-master/manimlib/scene/scene.py", line 53, in __init__
self.construct()
File "example_scenes.py", line 20, in construct
title = TextMobject("This is some \\LaTeX")
File "/Users/mac/Downloads/manim-master/manimlib/mobject/svg/tex_mobject.py", line 147, in __init__
SingleStringTexMobject.__init__(
File "/Users/mac/Downloads/manim-master/manimlib/mobject/svg/tex_mobject.py", line 42, in __init__
file_name = tex_to_svg_file(
File "/Users/mac/Downloads/manim-master/manimlib/utils/tex_file_writing.py", line 20, in tex_to_svg_file
dvi_file = tex_to_dvi(tex_file)
File "/Users/mac/Downloads/manim-master/manimlib/utils/tex_file_writing.py", line 70, in tex_to_dvi
raise Exception(
Exception: Latex error converting to dvi. See log output above or the log file: media/Tex/664e80e8d8d98482.log
I tried to search on google for any help, but all the that I read was about the same issue but on Windows, and I'm on mac. However, I continued searching and found this on StackOverflow :
That answer is no longer useful now because Manim has already solved that problem. So this may be due to two things: To give us a hint of what is wrong, go to the manimlib folder with the terminal and run the following:
latex tex_template.tex
We distinguish the two possible errors:
Recognized the latex command but the compilation could not be completed because you are missing libraries, to solve this you can install the missing packages indicated by the terminal.
That LaTeX is not recognized as a command, in this case surely the PATH variable is not set correctly, here is the solution.
So I went to the manimlib folder and issued the following : latex tex_template.tex. And judging by the console log, it doesn't seem like its the second option because the latex command was recognized, but there is no further information about what's missing. This is the complete console log :
MacBook-Pro-de-MAC:manimlib mac$ latex tex_template.tex
log4cxx: setFile(/Users/mac/Library/Application Support/MiKTeX/texmfs/data/miktex/log/latex.log,true) call failed.
log4cxx: IO Exception : status code = 13
log4cxx: No output stream or file set for the appender named [RollingLogFile].
This is pdfTeX, Version 3.14159265-2.6-1.40.20 (MiKTeX 2.9.7250)
Sorry, but latex did not succeed.
The log file hopefully contains the information to get MiKTeX going again:
/Users/mac/Library/Application Support/MiKTeX/texmfs/data/miktex/log/latex.log
MacBook-Pro-de-MAC:manimlib mac$ latex tex_template.tex
log4cxx: setFile(/Users/mac/Library/Application Support/MiKTeX/texmfs/data/miktex/log/latex.log,true) call failed.
log4cxx: IO Exception : status code = 13
log4cxx: No output stream or file set for the appender named [RollingLogFile].
This is pdfTeX, Version 3.14159265-2.6-1.40.20 (MiKTeX 2.9.7250)
Sorry, but latex did not succeed.
The log file hopefully contains the information to get MiKTeX going again:
/Users/mac/Library/Application Support/MiKTeX/texmfs/data/miktex/log/latex.log
MacBook-Pro-de-MAC:manimlib mac$ latex tex_template.tex
This is pdfTeX, Version 3.14159265-2.6-1.40.20 (MiKTeX 2.9.7250)
Sorry, but latex did not succeed.
The log file hopefully contains the information to get MiKTeX going again:
/Users/mac/Library/Application Support/MiKTeX/texmfs/data/miktex/log/latex.log
So I went to see what's on the latex.log file, and found this :
2020-01-28 21:28:52,077+0100 INFO texworks - starting: MiKTeX TeXworks 0.6.3 (MiKTeX 2.9.7250)
2020-04-08 17:31:28,130+0100 INFO latex - starting with command line: latex tex_template.tex
2020-04-08 17:31:28,135+0100 INFO latex - allowing known shell commands
2020-04-08 17:31:28,135+0100 INFO latex - enabling input (output) from (to) processes
2020-04-08 17:31:28,151+0100 INFO latex - going to create file: latex.fmt
2020-04-08 17:31:29,772+0100 ERROR latex - /usr/local/bin/initexmf did not succeed; exitCode: 1
2020-04-08 17:31:29,772+0100 ERROR latex - output:
2020-04-08 17:31:29,772+0100 ERROR latex - log4cxx: setFile(/Users/mac/Library/Application Support/MiKTeX/texmfs/data/miktex/log/initexmf.log,true) call failed.
log4cxx: IO Exception : status code = 13
log4cxx: No output stream or file set for the appender named [RollingLogFile].
Sorry, but "MiKTeX Configuration Utility" did not succeed.
The log file hopefully contains the information to get MiKTeX going again:
/Users/mac/Library/Application Support/MiKTeX/texmfs/data/miktex/log/initexmf.log
2020-04-08 17:31:29,774+0100 FATAL latex - The memory dump file could not be found.
2020-04-08 17:31:29,774+0100 INFO latex - finishing with exit code 1
But I have no clue what to do. So please Help me. And also, I'm sorry for my probably bad english, I'm french so, be kind ^^
Got some other information on the initexmf.logfile :
020-04-08 19:28:22,695+0100 INFO initexmf - starting: MiKTeX Configuration Utility 2.9.7259 (MiKTeX 2.9.7250)
2020-04-08 19:28:22,697+0100 INFO initexmf - Operating on the private (per-user) MiKTeX setup
2020-04-08 19:28:22,697+0100 INFO initexmf - starting with command line: initexmf --dump-by-name=latex --engine=pdftex
2020-04-08 19:28:22,699+0100 INFO initexmf - running: miktex-makefmt --engine=pdftex --dest-name=latex --no-dump latex.ini --engine-option=-tcx=cp227.tcx --miktex-disable-maintenance
2020-04-08 19:28:23,144+0100 FATAL initexmf - Permission denied: path="/Users/mac/Library/Application Support/MiKTeX/texmfs/data/miktex/log/miktex-makefmt_2020-04-08-192823.out"
只需要对texlive进行更新即可
tlmgr update --self --all --reinstall-forcibly-removed
@xvetisky did it worked.
No
I had this error, I think it has to do with some network issues, not sure
MacBook-Pro-de-MAC:~ mac$ tlmgr update --self --all --reinstall-forcibly-removed
/Library/TeX/texbin/tlmgr: unexpected return value from verify_checksum: -5
MacBook-Pro-de-MAC:~ mac$
@liyiliuxingyu
@xvetisky try reinstalling Latex.
I had a same problem with you. And fix that by adding package
\usepackage{ucs} in tex file tex_template (manimlib/tex_template). I hope that help you fix this.
@xvetisky were you able to try @tai9856's suggestion?
Latex error converting to dvi has more than one thread in this repo. I tried every option which is mentioned here on git and the final solution was to use other distribution then MiKTeX (https://www.latex-project.org/get/), concretely TeX Live. After uninstall of MiKTeX, install Tex Live and restart my computer, everything works fine.
@DaliborCimr Hmm lets try it...
@leotrs Yes, but it didn't work. Currently trying to reinstall LaTeX.
You need to add LaTeX to the path.
For me, this is was the path. C:\Program Files\MiKTeX\miktex\bin\x64
Ok so for one MAC OS Solution. I'm currently using Catalina, and in some of the documentation within the Latext directory I found the following pdf. (See screenshot)
So from what I've understood, MACTEXT doesn't install TextLive. It has to be installed separately, so I did that by going to the link I pointed out. This is the first time I installed Latext and my frustration was enormous seeing that TextLive required us to install around 4000+ packages (btw u'll be given an option to install via a program with a lion and another via terminal. Choose the Terminal option). TextLive took around 7 GB of my storage, I was oblivious, I search on google why people in academia accept this bull$%t, why do they even like Latex?,
I was around 1000/4998 pkg., it was installing packages with the most mundane names, like "bib.archelogy or "learn_latex101" and shit like that...Anyhow, I opened a new shell and I decided to try again the command with just 1001/4998...packages installed into my computer.
python3 -m manim example_scenes.py WriteStuff -pl
And it worked, it fucking works! I decided to terminate the installation of TextLive and I hope this shit doesn't cause me any more problems in the near future.
I hope this can help someone and wish you all a great path of MANIM Learning. This is my first GitHub post and English is my second language. (Sorry for my Grammatical errors)
Cheers from Bolivia.
p.s. given that TextLive has eaten around 7 GB of my storage I'll try to delete these mundane packages little by little hoping that it doesn't affect my workflow with MANIM, I believe is just one file that is helping us fix the main issue that's why I'll try to search for that file and just keep that one/ones in my computer and delete the rest.
Cheers
Ok so for one MAC OS Solution. I'm currently using Catalina, and in some of the documentation within the Latext directory I found the following pdf. (See screenshot)
So from what I've understood, MACTEXT doesn't install TextLive. It has to be installed separately, so I did that by going to the link I pointed out. This is the first time I installed Latext and my frustration was enormous seeing that TextLive required us to install around 4000+ packages (btw u'll be given an option to install via a program with a lion and another via terminal. Choose the Terminal option). TextLive took around 7 GB of my storage, I was oblivious, I search on google why people in academia accept this bull$%t, why do they even like Latex?,
I was around 1000/4998 pkg., it was installing packages with the most mundane names, like "bib.archelogy or "learn_latex101" and shit like that...Anyhow, I opened a new shell and I decided to try again the command with just 1001/4998...packages installed into my computer.
python3 -m manim example_scenes.py WriteStuff -pl
And it worked, it fucking works! I decided to terminate the installation of TextLive and I hope this shit doesn't cause me any more problems in the near future.
I hope this can help someone and wish you all a great path of MANIM Learning. This is my first GitHub post and English is my second language. (Sorry for my Grammatical errors)
Cheers from Bolivia.
p.s. given that TextLive has eaten around 7 GB of my storage I'll try to delete these mundane packages little by little hoping that it doesn't affect my workflow with MANIM, I believe is just one file that is helping us fix the main issue that's why I'll try to search for that file and just keep that one/ones in my computer and delete the rest.
Cheers
|
gharchive/issue
| 2020-04-08T18:12:43 |
2025-04-01T04:54:39.800244
|
{
"authors": [
"AtaTrkgl",
"DaliborCimr",
"Mathoholic",
"dipamsen",
"leotrs",
"liyiliuxingyu",
"tai9856",
"viccctv26",
"xvetisky"
],
"repo": "3b1b/manim",
"url": "https://github.com/3b1b/manim/issues/969",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1374843728
|
Question
Can schools directly patch this or is this left up to google to fix it.
I hope it never gets patched
@paperoll this is a issue google has to fix and judging by google's past patches, this is not gonna be patched for a long time
pls star repo if this helped
|
gharchive/issue
| 2022-09-15T17:13:28 |
2025-04-01T04:54:39.807861
|
{
"authors": [
"3kh0",
"SupremeRubisco",
"paperoll"
],
"repo": "3kh0/ext-remover",
"url": "https://github.com/3kh0/ext-remover/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
890501409
|
modify README
remove descriptions about pretrained models & add detail steps for using the codes of audio and audiovisual models
approved by myself
|
gharchive/pull-request
| 2021-05-12T21:25:54 |
2025-04-01T04:54:39.823377
|
{
"authors": [
"winston-lin-wei-cheng"
],
"repo": "3loi/MSP_Face",
"url": "https://github.com/3loi/MSP_Face/pull/6",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2340748546
|
[ADD] Add SVG File preview support (maybe using cairosvg?)
I think I'd fine a good a idea to make the plugin more complete and add an ability to preview images maybe using the python module cairosvg (like how it is suggested they use it in the plugin benlubas/molten-nvim) in order to convert the SVG to png into a temp buffer and then let 3rd/image take over from there with the tricks it knows
That'd make the idea of an image previewer a little more complete.
Have you tried preview svg images ?
It's working well on my config
There is opened issues related to this topic
How? did you do it with the default config?
when I
a = require("image").from_file("my_image.svg")
it Error me:
E5108: Error executing lua ...ng/.local/share/nvim/lazy/image.nvim/lua/image/image.lua:300: attempt to index local 'magick_image' (a nil value)
stack traceback:
...ng/.local/share/nvim/lazy/image.nvim/lua/image/image.lua:300: in function 'from_file'
[string ":lua"]:1: in main chunk
and no mention about SVG files in the 3rd/image/README.md
SVGs work fine for me as well, you're not getting any other kind of errors?
Also does it work for you with other formats?
Hi OK so I was playing around with svg(xml) file manipulation and generation and some of the test images I created do display something on libreDraw and now with vifm preview or 3rd/image and I notices it's when no viewBox is specified for a path shape element or when no width, height the image wont display in 3rd/image
Maybe some mechanism of temporarly "injecting" parameters those based on calculated elements bounding box and some default screen preview pixel size
We should add a flag for that, I think it's more common to want to edit the SVG as text than to view it.
True!
The crazy thing could be to be able to visualise the SVG under the last line of code while editing it and visualise the changes.
That's such a cool idea <3
|
gharchive/issue
| 2024-06-07T15:50:16 |
2025-04-01T04:54:39.828817
|
{
"authors": [
"3rd",
"MikeLemo1",
"maxencetholomier",
"t1gu1"
],
"repo": "3rd/image.nvim",
"url": "https://github.com/3rd/image.nvim/issues/177",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
870195890
|
Use #[path] instead of include!() to "include" generated files
Add instructions on how to start redis for integration tests on mac os x
remove a trailing '/' from a link in readme that is warned on byy .md linter
I'll just do a new clean PR
|
gharchive/pull-request
| 2021-04-28T17:44:01 |
2025-04-01T04:54:39.830546
|
{
"authors": [
"andrewdavidmackenzie"
],
"repo": "3scale-labs/limitador",
"url": "https://github.com/3scale-labs/limitador/pull/35",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1397578524
|
osd 4.12 support
Following notice that OSD 4.12 will require k8s 1.25 we need to ensure that anything on this list is addressed before 4.12 release: https://kubernetes.io/docs/reference/using-api/deprecation-guide/#v1-25
Keep open until a new release with these changes is published
@MStokluska image quay.io/3scale/marin3r:v0.11.0-alpha.1 is available for testing. It includes the changes to address api deprecations in k8s 1.25.
Release v0.11.0 is out.
|
gharchive/issue
| 2022-10-05T10:29:20 |
2025-04-01T04:54:39.833690
|
{
"authors": [
"MStokluska",
"roivaz"
],
"repo": "3scale-ops/marin3r",
"url": "https://github.com/3scale-ops/marin3r/issues/141",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
583782305
|
backend: Handle time period elapses
During the flushing process, if we have a reporting failure we need to
...
Codecov Report
Merging #129 into caching will decrease coverage by 1.82%.
The diff coverage is 51.38%.
@@ Coverage Diff @@
## caching #129 +/- ##
===========================================
- Coverage 90.04% 88.21% -1.83%
===========================================
Files 12 12
Lines 994 1035 +41
===========================================
+ Hits 895 913 +18
- Misses 77 97 +20
- Partials 22 25 +3
Impacted Files
Coverage Δ
pkg/threescale/backend/helpers.go
76.47% <45.45%> (-23.53%)
:arrow_down:
pkg/threescale/backend/backend.go
92.22% <60.71%> (+0.25%)
:arrow_up:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 340fede...59b922e. Read the comment docs.
|
gharchive/pull-request
| 2020-03-18T14:47:50 |
2025-04-01T04:54:39.841415
|
{
"authors": [
"PhilipGough",
"codecov-io"
],
"repo": "3scale/3scale-istio-adapter",
"url": "https://github.com/3scale/3scale-istio-adapter/pull/129",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
142198097
|
Friendly amendments to pull #15
Hi @grempe, here are some amendments to the changes you pushed in pull request #15.
If these are ok with you i will merge them into the master branch. Or feel free to make some suggestions for additional changes.
Thank you again for your work on this. Kind regards.
Hi, glad you are still there. I guess its intentional that you are trying to completely obfuscate your identity and association with this lib? ;-)
Regarding Pow. It would not work for me with a clean install on OS X El Capitan. There are a lot of comments on several Pow issues related to this:
https://github.com/basecamp/pow/issues/452
https://github.com/basecamp/pow/issues/517
https://github.com/basecamp/pow/pull/521
The last commit to Pow repo was in 2014. It is fair to consider it abandoned and non-functional due to significant changes by Apple in the OS X releases since 2014.
https://github.com/basecamp/pow/commits/master
Caddy is a super simple replacement web server. It really doesn't matter what you use I suppose, maybe even a simple python command to run a local server. But please don't leave to Pow instructions in place.
e.g.
http://www.linuxjournal.com/content/tech-tip-really-simple-http-server-python
According to the browserify site it should be installed as a global since it provides a system wide binary:
http://browserify.org/#install
http://stackoverflow.com/questions/35992104/why-is-it-necessary-to-install-browserify-twice-to-bundle
docco should also be installed global according to their docs (again with the rule of binary's being installed globally, and packages you will require being installed locally):
http://jashkenas.github.io/docco/
Thanks again for responding and looking carefully.
Glenn
Thanks for pointing out the abandon-ed-ness of Pow. i didn’t know about any of that stuff. i will phase it out of this project to guard against anymore confusion.
Rather than add install instructions and config for Caddy, i decided to add the 'http-server' package to the development dependencies in the master branch (https://github.com/45678/miniLockLib/commit/ede5997726470f528dd195a225616d9300730f3c) so that everything required to run the tests is available after npm install (hopefully at least).
To start up the default webserver to facilitate the window tests type npm run webserver. You still use Caddy if you prefer it, but you may need to re-define the window tests address. The default address is http://localhost:45678/tests.html. It can be re-defined with the .window_tests_address file.
thanks for bringing up the issue about installing packages globally @grempe.
i see that docco and browserify recommend a global install but i don’t want to recommend it for the miniLockLib setup procedure. they both work good for me as local development dependencies and i would prefer to keep the setup procedure uncomplicated. i just want to git clone and npm install and then get to work.
of course, if that flow is problematic for you i'm willing to reconsider. for now i'm going to leave it unmentioned in the README i think.
Sounds good. I was not particularly tied to Caddy. Just suggesting a working replacement. Your solution sounds fine. I'll try it out.
Also fine for docco and browserify, as long as they work just installed locally. :-)
G
|
gharchive/pull-request
| 2016-03-20T19:50:53 |
2025-04-01T04:54:39.882758
|
{
"authors": [
"45678",
"grempe"
],
"repo": "45678/miniLockLib",
"url": "https://github.com/45678/miniLockLib/pull/17",
"license": "unlicense",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1296957751
|
Add setter to HasX interfaces where possible
Pull Request
[x] I have checked for similar PRs.
[x] I have read the contributing guidelines.
Changes
[x] Code
[ ] Documentation
[x] Other: JAXB Binding Declaration
Description
This MR adds setters to HasX interfaces where it was missing and possible to allow more abstraction.
Extended interfaces:
HasDescription
HasPart
HasRelatedAssembly
Also changed HasPart to be generic in order to allow adding the setter to the interface. Needed to adjust the JAXB Binding Declaration file for this to have multiple bindings instead of one.
Reverted 7d45586 due to breaking changes. Will be done in a separate PR.
|
gharchive/pull-request
| 2022-07-07T07:30:55 |
2025-04-01T04:54:39.890729
|
{
"authors": [
"DManstrator"
],
"repo": "4Soft-de/kbl-model",
"url": "https://github.com/4Soft-de/kbl-model/pull/76",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2402903872
|
Making smooth transition on "Back To Top" button
On clicking the "Back to Top" button, the transition to the page's top should be smooth. I am attaching a video as an example.
https://github.com/4darsh-Dev/CogniGaurd/assets/153659573/58d9ae57-2c4e-4147-a19c-f11cb680f79a
assigned to you @AkshitGarg24
Hey @AkshitGarg24 ! Thank you so much for your raising the issue💗
It’s all yours, you can come anytime again and make some contributions! 🚀
Alone, we can do little, but together we can do so much! 😇
GSSoC 24 has been completed Finally'
It was a great experience working with you all
Thanks💗 for your valuable contributions!
PA nomination has been started, Do fill out the forms soon. Share your experiences and let's connect on socials
|
gharchive/issue
| 2024-07-11T11:06:09 |
2025-04-01T04:54:39.908853
|
{
"authors": [
"4darsh-Dev",
"AkshitGarg24"
],
"repo": "4darsh-Dev/CogniGaurd",
"url": "https://github.com/4darsh-Dev/CogniGaurd/issues/143",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2108823917
|
erreur
Quelle suite d'actions a mené à cette erreur ?
Sans ces informations, il m'est impossible de résoudre ce bug
|
gharchive/issue
| 2024-01-30T21:34:59 |
2025-04-01T04:54:39.981769
|
{
"authors": [
"5cover",
"papychat"
],
"repo": "5cover/WinClean",
"url": "https://github.com/5cover/WinClean/issues/14",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
228729821
|
Connection with WTPs and Controller
Is it possible to connect the WTPs with the CPP and the CPP with the legacy switch ,if the CPP has 1 ethernet interface?
we can open some VLANs and therefore have 4 virtual interfaces over one physical on the CPP
In principle yes. I never tried to use it with VLANs.
On 15 May 2017 at 16:17, Georgios Tsiatsios notifications@github.com
wrote:
Is it possible to connect the WTPs with the CPP and the CPP with the
legacy switch ,if the CPP has 1 ethernet interface?
we can open some VLANs and therefore have 4 virtual interfaces over one
physical on the node
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
https://github.com/5g-empower/empower-lvnf-agent/issues/5, or mute the
thread
https://github.com/notifications/unsubscribe-auth/AAdzTLujAHiXwnGSr6K4ilb4LqUK1jR_ks5r6F5xgaJpZM4NbMrs
.
|
gharchive/issue
| 2017-05-15T14:17:21 |
2025-04-01T04:54:39.995382
|
{
"authors": [
"Giwrgosyea",
"rriggio"
],
"repo": "5g-empower/empower-lvnf-agent",
"url": "https://github.com/5g-empower/empower-lvnf-agent/issues/5",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2160537662
|
新功能 主码流子码流切换功能不起作用
根据日志和数据库来看,已经更新过了,但是播放之后看起来似乎没有效果
这个功能应该是给 GB28181-2022版使用的,GB28181-2016不支持子码流切换
摄像头的平台配置上看,他有选项的
摄像头的平台配置上看,他有选项的
明白了,看了下我的海康摄像头只有2016版,谢谢
|
gharchive/issue
| 2024-02-29T06:55:25 |
2025-04-01T04:54:40.001398
|
{
"authors": [
"luhengyu1",
"mengxing2017"
],
"repo": "648540858/wvp-GB28181-pro",
"url": "https://github.com/648540858/wvp-GB28181-pro/issues/1334",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1766627887
|
jwt登录的安全BUG
环境信息:
最新版本
描述错误
可以伪造admin用户进行jwt登录
截图
这里写死了公私钥,很多人部署的时候就不会改掉,然后导致别人拿这个公私钥加密token,进行伪造访问
回头做到配置里
|
gharchive/issue
| 2023-06-21T03:48:18 |
2025-04-01T04:54:40.003732
|
{
"authors": [
"648540858",
"liangfj"
],
"repo": "648540858/wvp-GB28181-pro",
"url": "https://github.com/648540858/wvp-GB28181-pro/issues/905",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1909846474
|
Update README.md
added some minor text corrections
Thanks for the contribution! ⚡
|
gharchive/pull-request
| 2023-09-23T12:08:53 |
2025-04-01T04:54:40.008968
|
{
"authors": [
"76creates",
"tacurran"
],
"repo": "76creates/terraform-provider-oryketo",
"url": "https://github.com/76creates/terraform-provider-oryketo/pull/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2551506421
|
Rename LICENSE to Whatsapp web
Whatsapp web
Uploading WhatsappWebToGo-master (2).zip…
Is it me or is this unwanted?
|
gharchive/pull-request
| 2024-09-26T20:48:20 |
2025-04-01T04:54:40.057222
|
{
"authors": [
"Minionguyjpro",
"roni1177"
],
"repo": "92lleo/WhatsappWebToGo",
"url": "https://github.com/92lleo/WhatsappWebToGo/pull/125",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
184689454
|
FirstStep: Add credentials-file for GitHub data
I've added a credentials-file for github username & password to avoid typing data over and over again
I know it isn't a caching, but i think it's a first step and the idea behind is similar - avoid typing data over and over again
connects to #13
Need changes with the current implementation
#13
@codingCookie we have a conflict here :)
@charithsoori
I'm still implementing to have both possibilities - It's a little bit busy at the moment - i try to do it as soon as possible...sorry :/
|
gharchive/pull-request
| 2016-10-23T13:18:08 |
2025-04-01T04:54:40.068954
|
{
"authors": [
"charithsoori",
"codingCookie",
"rajikaimal"
],
"repo": "99xt/github-manager-cli",
"url": "https://github.com/99xt/github-manager-cli/pull/21",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1483180554
|
Try to replace food database
Input
Food database with rating
Output
Look and check
Hi Hải,
Tớ ssh vào máy ảo ubuntu trên azure thông qua cổng 8081 thì bị lỗi này là sao nhỉ ?
BR
Quang
Hi Hải,
Hiện tại task này :
Done :
import file csv - 160 dòng chứa các món ăn và thành phần và bảng collector_foodlog trong sqlite3
tạo 1 app trong back-end foodgeek với tên gọi "ninehealth_logger" để check việc log và ghi nhận từ front-end vào back-end thông qua urls và view
Will do :
Phần front-end dự kiến sẽ tạo 1 cái view có mấy trường input simple bằng html
Thực hiện input thêm đồ ăn và dĩnh dưỡng vào form này để kiểm tra xem hàm log chạy và lưu dữ liệu vào db có giống movie hay không (trong module collectors)
Trao đổi thêm để chạy các bước recommend tiếp theo sau khi food đã được log vào hệ thống
Issue :
tạm thời chưa có issue gì. Tuy nhiên task này định là hôm nay là xong tuy nhiên có 1 vài việc gia đình chen ngang nên cố gắng tớ sẽ hoàn thành vào khoảng 14 tức 28/12/2022
có khả năng sẽ mất thêm thời gian vào 1 task cũng đang chuẩn bị chạy với anh Công trong tuần này đó là : lấy code của các màn hình chính đã thiết kế + thiết kế thêm màn hình signUp để đăng kí và lấy thông tin về thể trạng của người dùng
Khả năng sqlite3 không đủ để tải nên tớ đang muốn chuyển sang postGreSQL
BR
Quang
Hello @quangvv9Life,
Thanks Quang đã chia sẻ plan các thứ @@.
Cơ mà tớ đọc qua thấy cũng hơi khó hiểu :|
Sorry tớ chưa dành thời gian đọc kỹ =.=
Thực sự tớ cần nhìn tài liệu hay code thôi =.=
Kiểu như code review và design review ý =.=
Quang cứ tạo và share file hay code ý cho tớ là okay!!!
Ở trên OneDrive của M365 ý.
Gửi qua chat tớ cũng sợ, hơi ít vào 😢. Sorry Quang.
Thanks Quang!!!
Hi Hải @9health ,
Thank Hải đã chia sẻ ngắn gọn và dễ hiểu lại các kiến thức trong cuốn PRS nhé. T thấy đọc lại dạng summary cũng giúp tiếp cận và nắm được công nghệ này nhanh hơn đấy !!!
Sorry Hải hôm qua tớ có chạy task này bằng folder ...work/foodgeek/ và có cài đặt 1 package tên là djangorestframework
Rõ ràng là tớ đã activate venv nằm trong folder foodgeek rồi và thực hiện lệnh cài package trên mà không hiểu sao tớ sang screen của moviegeek lại thấy nó update cả version django và cài thêm package này vào của Hải :-? khó hiểu thật
Package djangorestframework là mục đích là :
Cài 1 cái apps mới trong django tên là ninehealth_logger
Apps này sẽ có models.py với class tên là FoodLog sẽ có những trường như : FoodId, name, ingrename, ingreid, totalliked, avgrating, v.v.v . Cái class này sẽ tương ứng vs bảng ninehealth_logger_foodlog trong DB SQLite3
Phần urls.py thì tớ sẽ route nó thông qua URL : .../api/ để xem toàn bộ thông tin theo dạng api json
Phần views.py thì tớ không tạo thêm front-end là 1 template html cho nó nữa mà dùng luôn api để phần tương tác vs DB thông qua phương thức HTTP vs api thôi
Tuy nhiên như tớ có nói ở trên là khi cài ở venv của foodgeek thì không hiểu sao nó lại ăn cả vào moviegeek vì tớ có vào kiểm tra pip freeze thì thấy moviegeek đã thành như sau:
Django==3.2.16
djangorestframework==3.14.0
Đối chiếu lại với requirement.txt của tác giả thì thấy đang chạy version thấp hơn :
Django==2.2.27
và djangorestframework thì không có
Nên suy ra do tớ chạy cái lệnh cài thêm nên nó ăn sang cả của Hải, dẫn tới hiện tại vào trang 8080 thì nó đang bị lỗi như thế này :
Tớ có tìm hiểu lỗi này ở trên mạng thì ra được bài sau :
Đại loại là do upgrade lên version cao hơn nên nó bị ảnh hưởng gì đó đến đống staticfiles
Vậy solution tớ nghĩ là :
Downgrade django version của moviegeek của Hải về version của tác giả
Sửa đoạn sau trong các file templates html {% load static from staticfiles %} anymore. You should just write {% load static %}
Hải xem và cho ý kiến nhé !!!!
BR
Quang
Thanks @quangvv9Life đã report nhé @@.
May quá sáng nay tớ chả động gì tới cái MovieGEEKs website hự @@
Chiều nay thì fix xong rồi @@.
Quang dùng package djangorestframework mà thấy tiện thì cứ dùng nhé!!!
Còn không tớ thấy với cái hiện tại của tác giả chắc cũng okay @@.
Nếu mà bị conflict version thì hạ version xuống thấp hơn thui!!!
Uhm còn tớ đọc như thế kiểu vừa đọc vừa trải nghiệm, vừa ngẫm nghĩ xem cái thuật toán này hay ho thế nào thui @@.
Chứ cứ đọc ào ào bỏ qua mấy cái hay phí lắm T_T.
Thôi mình không biết ăn ngon thì biết thưởng thức thuật toán vậy T_T
Mặc dù thuật toán tớ cũng dốt đặc hự. Xưa tớ viết được mỗi hàm sort thôi chứ cao cấp hơn như kiểu linked list hay cái gì gì đó tớ chịu thui O__O.
Thanks Quang!!!
Hi Hải @9health ,
Công nhận là ngồi đọc và cùng thảo luận về các thuật toán này cũng hay thật, mà cái phần hay nhất là khi nó chạy mà trả ra được kết quả mong muốn í
Tớ nghĩ đợt này tớ với Hải cùng quyết tâm làm chủ được mấy cái thuật toán này vì sau này nó có tính ứng dụng rất rộng vào các mảng khác nữa đấy !!!
Vì hiện tại gói nativebase startup+ họ cũng cung cấp cho mình 1 code mà chạy được ở cả 3 môi trường : web, android và ios thì có thể mình sẽ lên 1 kế hoạch build dần 1 cái web front end và cũng dùng chung back end với cái app ios và android Hải ạ
Web này sẽ với mục đích marketing cho app ios và android (SEO) và dùng để unlock level 3 của microsoft nữa
Quan trọng là mình nắm vững các thuật toán và biết ứng dụng thuật toán nào phù hợp nhất vào bài toán hiện tại là giải quyết vấn đề béo phì bằng gợi ý chế độ dinh dưỡng và tập luyện hợp lý cho người dùng là được rồi !!!
Cám ơn Hải đã đọc và hướng dẫn lại tớ, tớ thấy mình cùng đọc và trao đổi và suy ngẫm về kiến trúc này sẽ giúp đẩy nhanh được tiến độ hơn nữa !!! Tớ sẽ tận dụng tối đa thời gian để đọc lại cho kịp với tốc độ của Hải nhé !!!
Ngoài ra còn nên document lại kiến trúc theo hướng nghĩ của mình nữa giống như vẽ lại cái sườn để áp dụng linh hoạt được cho các bài toán cụ thể khác sau này. Có thể tận dụng các buổi trao đổi để ghi chép lại luôn
Cám ơn Hải đã trao đổi và dành thời gian !!! Chúc Hải cuối tuần vui vẻ và năm mới bình an và nhiều may mắn !!! Chúc năm sau 9reatOdin phát hành được một sản phẩm hoàn thiện tốt !!!
BR
Quang
Hello @quangvv9Life
Thanks Quang đã cùng cùng tớ tạo các bài test nho nhỏ cho model mà Quang đã tạo nhé @@.
Cuối cùng cũng có 1 cái gì đó chạy được và làm được tương đối đúng ý Quang @@@@.
Dưới đây là 1 vài điểm tớ note lại hoặc thêm vào O_O.
Về Network Latency
Cái này tớ thấy có lẽ là cần 1 thử nghiệm dài nên đã tạo 1 issues khác ở đây
#30
Cũng do ban đầu tớ mải setup thất nhanh nên dùng command line từ ví dụ trên Microsoft Learn để tạo máy ảo =.=
Ban đầu tớ nghĩ sẽ không okay lắm vì test thử ping có vẻ cao.
Cơ mà bật lên rồi và vào web app thì thấy cũng ổn =.=
Lúc tớ config, cài packages toàn vào bằng SSH và dùng screen hết =.=
Cơ mà sang Quang gặp vấn đề, haizzz.
Tớ nghĩ thui thì Quang cứ thử từng solution tớ đưa ra trong ý xem cái nào được :|.
Tớ hi vọng khi dùng Cloudflare WARP (1 dạng VPN) sẽ okay. Nếu không Cloudflare Zerotrust đảm bảo sẽ rất okay =.= vì dùng cho industry mà.
Quang cũng có thể mày mò setup vì account hiện tại có quyền root. Dùng sudo là install được cái gọi là client hay là server gì đó lên máy Ubuntu 18.04!!!
Về thời gian làm task này
Lần trước tớ gặp Quang cách đây 2 tuần cũng vào Thứ 4 thì Quang mới nghĩ ra task đấy và tớ cũng ủng hộ.
Sau 1 tuần cũng có 1 cái database model tương đối.
Và tới tuần này thì đã database model hoàn chỉnh hơn và chạy được với QuerySet của Django!!!
Thanks Quang nhé.
Vì Quang còn làm nhiều tasks khác như là UI/UX, setup Microsoft 365, đọc comment của tớ (rất dài...), level của Quang tớ nghĩ cũng là beginner thôi.
Tớ thấy nên update thời gian cho task này là 1 tuần có okay không Quang @@@@@@.
Thực ra thời gian làm có thể vào Linux server và count được thời gian...
ninehealth@LinuxVM15:~$ last | less
nineheal pts/0 116.101.89.XXX Thu Jan 5 07:47 still logged in
nineheal pts/0 116.101.89.XXX Thu Jan 5 07:43 - 07:46 (00:02)
nineheal pts/0 116.101.89.XXX Thu Jan 5 07:13 - 07:28 (00:14)
nineheal pts/0 116.101.89.XXX Thu Jan 5 06:43 - 06:44 (00:00)
nineheal pts/0 116.101.89.XXX Thu Jan 5 06:39 - 06:43 (00:03)
nineheal pts/9 :pts/13:S.3 Wed Jan 4 14:35 - 18:20 (03:44)
nineheal pts/6 :pts/13:S.0 Wed Jan 4 14:35 - 18:20 (03:44)
nineheal pts/7 :pts/13:S.1 Wed Jan 4 14:35 - 18:20 (03:44)
nineheal pts/8 :pts/13:S.2 Wed Jan 4 14:35 - 18:20 (03:44)
nineheal pts/13 123.24.150.XXX Wed Jan 4 14:35 - 18:20 (03:44)
nineheal pts/13 123.24.150.XXX Wed Jan 4 14:34 - 14:34 (00:00)
nineheal pts/13 123.24.150.XXX Wed Jan 4 14:33 - 14:33 (00:00)
nineheal pts/13 123.24.150.XXX Wed Jan 4 14:32 - 14:33 (00:00)
nineheal pts/13 123.24.150.XXX Wed Jan 4 14:32 - 14:32 (00:00)
nineheal pts/7 :pts/0:S.1 Wed Jan 4 14:30 - 14:35 (00:04)
nineheal pts/8 :pts/0:S.2 Wed Jan 4 14:30 - 14:35 (00:04)
nineheal pts/6 :pts/0:S.0 Wed Jan 4 14:30 - 14:35 (00:04)
nineheal pts/9 :pts/0:S.3 Wed Jan 4 14:30 - 14:35 (00:04)
nineheal pts/0 123.24.150.XXX Wed Jan 4 14:30 - 14:39 (00:09)
nineheal pts/0 116.101.89.XXX Wed Jan 4 10:40 - 11:01 (00:20)
nineheal pts/0 116.101.89.XXX Wed Jan 4 10:15 - 10:17 (00:01)
nineheal pts/7 :pts/0:S.1 Wed Jan 4 07:27 - 07:27 (00:00)
nineheal pts/8 :pts/0:S.2 Wed Jan 4 07:27 - 07:27 (00:00)
nineheal pts/6 :pts/0:S.0 Wed Jan 4 07:27 - 07:27 (00:00)
nineheal pts/9 :pts/0:S.3 Wed Jan 4 07:27 - 07:27 (00:00)
nineheal pts/0 116.101.89.XXX Wed Jan 4 07:26 - 07:27 (00:00)
nineheal pts/5 :pts/0:S.0 Wed Jan 4 07:26 - 07:26 (00:00)
nineheal pts/10 :pts/0:S.1 Wed Jan 4 07:26 - 07:26 (00:00)
nineheal pts/11 :pts/0:S.2 Wed Jan 4 07:26 - 07:26 (00:00)
nineheal pts/12 :pts/0:S.3 Wed Jan 4 07:26 - 07:26 (00:00)
nineheal pts/16 :pts/0:S.3 Wed Jan 4 07:25 - 07:25 (00:00)
nineheal pts/15 :pts/0:S.2 Wed Jan 4 07:25 - 07:25 (00:00)
nineheal pts/14 :pts/0:S.1 Wed Jan 4 07:25 - 07:25 (00:00)
nineheal pts/13 :pts/0:S.0 Wed Jan 4 07:25 - 07:25 (00:00)
nineheal pts/12 :pts/0:S.3 Wed Jan 4 07:00 - 07:25 (00:24)
nineheal pts/11 :pts/0:S.2 Wed Jan 4 07:00 - 07:25 (00:24)
nineheal pts/10 :pts/0:S.1 Wed Jan 4 07:00 - 07:25 (00:24)
nineheal pts/5 :pts/0:S.0 Wed Jan 4 07:00 - 07:25 (00:24)
nineheal pts/12 :pts/0:S.3 Wed Jan 4 07:00 - 07:00 (00:00)
nineheal pts/11 :pts/0:S.2 Wed Jan 4 07:00 - 07:00 (00:00)
nineheal pts/10 :pts/0:S.1 Wed Jan 4 07:00 - 07:00 (00:00)
nineheal pts/5 :pts/0:S.0 Wed Jan 4 07:00 - 07:00 (00:00)
nineheal pts/12 :pts/0:S.3 Wed Jan 4 07:00 - 07:00 (00:00)
nineheal pts/11 :pts/0:S.2 Wed Jan 4 07:00 - 07:00 (00:00)
nineheal pts/10 :pts/0:S.1 Wed Jan 4 07:00 - 07:00 (00:00)
nineheal pts/5 :pts/0:S.0 Wed Jan 4 07:00 - 07:00 (00:00)
nineheal pts/0 116.101.89.XXX Wed Jan 4 06:58 - 07:26 (00:28)
webserve pts/2 :pts/0:S.1 Wed Jan 4 05:59 - 06:11 (00:11)
webserve pts/3 :pts/0:S.2 Wed Jan 4 05:59 - 06:11 (00:11)
webserve pts/4 :pts/0:S.3 Wed Jan 4 05:59 - 06:11 (00:11)
webserve pts/1 :pts/0:S.0 Wed Jan 4 05:59 - 06:11 (00:11)
webserve pts/1 :pts/0:S.0 Wed Jan 4 05:58 - 05:59 (00:00)
webserve pts/2 :pts/0:S.1 Wed Jan 4 05:57 - 05:59 (00:01)
webserve pts/4 :pts/0:S.3 Wed Jan 4 05:57 - 05:59 (00:01)
webserve pts/3 :pts/0:S.2 Wed Jan 4 05:57 - 05:59 (00:01)
webserve pts/1 :pts/0:S.0 Wed Jan 4 05:57 - 05:58 (00:01)
webserve pts/4 :pts/0:S.3 Wed Jan 4 05:45 - 05:56 (00:10)
webserve pts/3 :pts/0:S.2 Wed Jan 4 05:45 - 05:56 (00:10)
webserve pts/2 :pts/0:S.1 Wed Jan 4 05:45 - 05:56 (00:10)
webserve pts/1 :pts/0:S.0 Wed Jan 4 05:45 - 05:56 (00:10)
webserve pts/1 :pts/0:S.0 Wed Jan 4 05:39 - 05:39 (00:00)
nineheal pts/2 :pts/0:S.1 Wed Jan 4 05:17 - 05:28 (00:10)
nineheal pts/4 :pts/0:S.3 Wed Jan 4 05:17 - 05:28 (00:10)
nineheal pts/3 :pts/0:S.2 Wed Jan 4 05:17 - 05:28 (00:10)
nineheal pts/5 :pts/0:S.4 Wed Jan 4 05:17 - 05:20 (00:02)
nineheal pts/1 :pts/0:S.0 Wed Jan 4 05:17 - 05:28 (00:10)
nineheal pts/2 :pts/0:S.1 Wed Jan 4 05:17 - 05:17 (00:00)
nineheal pts/4 :pts/0:S.3 Wed Jan 4 05:17 - 05:17 (00:00)
nineheal pts/3 :pts/0:S.2 Wed Jan 4 05:17 - 05:17 (00:00)
nineheal pts/1 :pts/0:S.0 Wed Jan 4 05:17 - 05:17 (00:00)
nineheal pts/5 :pts/0:S.4 Wed Jan 4 05:17 - 05:17 (00:00)
nineheal pts/2 :pts/0:S.1 Wed Jan 4 05:15 - 05:17 (00:02)
nineheal pts/4 :pts/0:S.3 Wed Jan 4 05:15 - 05:17 (00:02)
nineheal pts/3 :pts/0:S.2 Wed Jan 4 05:15 - 05:17 (00:02)
nineheal pts/5 :pts/0:S.4 Wed Jan 4 05:15 - 05:17 (00:02)
nineheal pts/1 :pts/0:S.0 Wed Jan 4 05:15 - 05:17 (00:02)
nineheal pts/0 116.101.89.XXX Wed Jan 4 05:12 - 06:11 (00:58)
nineheal pts/8 :pts/0:S.2 Wed Jan 4 05:11 - 05:11 (00:00)
nineheal pts/7 :pts/0:S.1 Wed Jan 4 05:11 - 05:11 (00:00)
nineheal pts/9 :pts/0:S.3 Wed Jan 4 05:11 - 05:11 (00:00)
nineheal pts/6 :pts/0:S.0 Wed Jan 4 05:11 - 05:11 (00:00)
nineheal pts/2 :pts/0:S.1 Wed Jan 4 05:10 - 05:10 (00:00)
nineheal pts/4 :pts/0:S.3 Wed Jan 4 05:10 - 05:10 (00:00)
nineheal pts/3 :pts/0:S.2 Wed Jan 4 05:10 - 05:10 (00:00)
nineheal pts/5 :pts/0:S.4 Wed Jan 4 05:10 - 05:10 (00:00)
nineheal pts/1 :pts/0:S.0 Wed Jan 4 05:10 - 05:10 (00:00)
nineheal pts/0 116.101.89.XXX Wed Jan 4 05:04 - 05:12 (00:08)
nineheal pts/9 :pts/0:S.3 Tue Jan 3 22:01 - 22:32 (00:30)
nineheal pts/6 :pts/0:S.0 Tue Jan 3 22:01 - 22:32 (00:30)
nineheal pts/7 :pts/0:S.1 Tue Jan 3 22:01 - 22:32 (00:30)
nineheal pts/8 :pts/0:S.2 Tue Jan 3 22:01 - 22:32 (00:30)
nineheal pts/0 123.24.150.XXX Tue Jan 3 22:01 - 22:32 (00:30)
nineheal pts/2 :pts/0:S.1 Tue Jan 3 21:35 - 21:35 (00:00)
nineheal pts/4 :pts/0:S.3 Tue Jan 3 21:35 - 21:35 (00:00)
nineheal pts/3 :pts/0:S.2 Tue Jan 3 21:35 - 21:35 (00:00)
nineheal pts/5 :pts/0:S.4 Tue Jan 3 21:35 - 21:35 (00:00)
Tớ chỉ copy lại trong vòng 2 ngày đổ lại thui @@.
Haizzz sao mà lắm thế nhỉ =.=
Do screen tạo ra nhiều sessions quá haizzz
Làm cho cái last này cũng bị tràn luôn chứ đúng ra là sẽ record từ lúc tớ tạo máy ảo này đấy haizzzz.
Nếu mà Quang muốn chính xác hơn có thể tham khảo issues tớ viết về Linux Administration để tạo 1 account mới cho Quang, permission folder của Quang nữa =.=
Cơ mà Quang làm lại chắc mất 1 ngày đấy =.=
Tùy Quang sắp thôi nhưng mà sau kiểu gì chắc cũng phải gặp hix.
Nếu mà dùng cái trên grep, cut để tạo thành file Excel rồi xử lý chắc sẽ tính ra đúng đấy cơ mà thui vậy =.= Ước lượng gần đúng thui vì thời gian giãn ra cũng là 1 cách để xem mình làm đúng hay không chứ liên tục thì code có khi loằng bà ngoằng =.=
Có lẽ là tớ sẽ tắt bớt đi cái screen session nào không dùng mất =.= Nhiều cái thừa ghê...
Mỗi cái session đấy lại tạo ra 1 Bash shell thread, lại tốn CPU và RAM nữa @@, mà ít thui @@.
Chắc là cứ cuối ngày tắt hết screen session đi ngoại trừ cái MovieGEEKs web app đang chạy thui @@@@@@@.
Còn bình thường thì sẽ tạo screen mới hết.
Cái gì đang code trong ý bằng Vim thì cũng sẽ save lại thành Vim session hết rồi vào và restore sau ...
Về open-source hay không
Cái này thì qua tớ lúc đi trên đường có chat với mấy bạn cũ của tớ để lấy contact của một bạn có nhiều xiền ở Thụy Sĩ ý để giúp bạn có $15K kia bị mất thì tớ mới nghĩ ra là các bạn ý cũng tương đối có thời gian =.=
Biết đâu các bạn ý thích nhảy vào giúp mình hix =.=
Biết đâu thôi.
Vì thế tớ nghĩ ra thêm 1 lý do là cần open-source hết mọi thứ hay là không.
Biết đâu có lẽ có 1 cao thủ nhảy vào giúp...
Dù gì open-source mọi thứ mọi người cũng sẽ hiểu kỹ khả năng của mình đến đâu và giúp được cái gì...
Nói thật với Quang chứ tớ thấy mấy cái máy ảo, docker, K8s và 1 đống thứ linh tinh khác học và làm mãi không hết =.=
Có cao thủ chỉ cho 1 ít cũng tốt hơn hix =.= Giúp 1 ít càng tốt.
Cái GitHub này còn use case về collaborate để viết phần mềm từ IoT thì phải là 1 ông farmer đang đi máy cày trên ruộng chạy phần mềm báo lỗi và 1 người ngồi ở Mỹ hay châu Phi fix lỗi đấy mà =.=
Tớ không nhớ video đấy ở đâu nữa với không search lại được hic =.=
Già quá rồi, hết teen rồi.
Quang thì có nói sẽ có 1 ai đó skill cao cao vào copy phát mình mất đi IP của mình =.=
Cơ mà nói thật phần mềm individual làm tớ thấy cũng không được quá nhiều đâu.
Mà như mình đọc sách, có document rồi mà hiểu code còn vật vã =.=
Tất nhiên bảo là giỏi thì hiểu nhanh hơn cơ mà lúc code lại thì tớ nghĩ vẫn phải qua đủ flow design, code, test mới đưa vào production được chứ bảo 1 ngày copy lại chứ phải 1 tháng tới 3 tháng =.=
Trả xiền cho 1 ông dev copy hay hơn hay là 1 ông dev còn 1 đống feature trong công ty còn chưa dev xong hay hơn =.=
Nên là Microsoft mới open-source Terminal để mọi người vào report hay là cái đống .NET ý =.= Lượng code nhiều quá =.=
Mọi người cùng vào report, cùng check code...
Cái mà lợi thế của mình ở đây là mình hiểu cái mình code thôi =.= Họ hiểu và dùng thì okay cơ mà không hiểu thì quay sang hỏi mình thôi.
Mà nói thật hiểu được cái Quang nói cũng khó lắm =.= Tớ ở Nhật mãi mới tìm được cách nói dễ hiểu đấy...
Vậy nên code Quang viết chắc còn khó hiểu hơn...
Nên là lo bị copy không sao đâu =.=
Copy cũng tốt nhưng mà mình nên ghi license là họ phải đề cập đến việc sử dụng code của mình là được rồi.
Mình biết họ copy rồi thì mình phải improve làm sao cho trở nên tốt hơn ý =.=
Nói chung copy code và ghi rõ license thì okay. Không ghi license mới là có vấn đề.
Còn copy quyển sách hay mua đi bán lại 1 quyển sách thì lại là 1 chuyện khác =.= Tớ cũng chưa hiểu lắm cái đấy =.=
Model của phần mềm là licensing hay gọi là li-xăng =.= Mình cấp phép để họ dùng phần mềm của mình thôi chứ không có bản hẳn kiểu như là thuốc đánh răng hay là laptop hay cái gì cầm được @@@@@@@.
Cái này có lẽ nên học thêm 1 khóa nữa về Copyright...
Và cũng nên tạo ra 1 Wiki page nữa hix. Topic này quá rộng...
Về commit lên Git
Đấy thì mong Quang là consider việc cứ open-source ban đầu nhé @@.
Sau khi okay rùi thì commit lên thôi.
Quang cứ để nguyên code đang chạy được như thế, đừng sửa gì và commit lên 1 branch nhé.
Branch name có lẽ nên theo 1 format nào đấy như là
[user][YYYYMMDD][branch_name]
[user][NN][branch_name]
[YYYYMMDD]_[branch_name]
Lý do YYYYMMDD (YearMonthDate) là để sau này sort cho dễ hix.
Tùy cách Quang chọn cái nào phù hợp thôi =.=
Rồi tớ sẽ làm theo.
Mục đích của việc này là làm sao cho sau này tìm lại branch dễ hơn thôi vì sẽ có rất nhiều feature develop, rất nhiều test cần viết.
Sắp xếp theo thời gian cũng là 1 cách tìm lại dễ hơn =.=
Format này tớ tự nghĩ ra thôi chứ ở công ty tớ remote onsite sang bên Sing họ cũng có 1 kiểu đặt tên hao hao thế vì cả dự án họ là 1 cái git to đùng và khổng lồ, hàng trăm branches ý!!!
Và sau này từ những branch này sẽ merge lại vào master nữa nhé nên phía dưới tớ cũng sẽ có 1 ít comment về code Quang nữa để làm sao merge không bị conflict, chạy được cả MovieGEEKs và FoodGEEKs hay HealthGEEKs
Merge vào master thì để tớ làm nhé =.= Quang làm mà toang chết tớ hự hự.
Và việc này là làm đầu tiên trước khi muốn sửa cái gì tiếp nhé =.=
Ban đầu thì sẽ hơi lâu thôi nhưng sau quen flow rồi sẽ quen dần và nhanh hơn hix.
Ngày đầu tớ commit vào cái git to đùng khổng lồ ở công ty bên Sing kia kìa tớ mất 1 ngày đấy để làm 1 phát là ăn luôn =.=
Chuẩn bị hết comment viết ra 1 file, kiểm tra xem files nào commit lên, có thiếu không.
Chứ sai là toang huhu
Về cấu trúc code, tên API, tên folder
Nói thật mấy cái này có lẽ đối với Quang chắc là hơi bị trừu tượng và chưa thể nắm được ngay @@.
Tớ thì khá hơn một xíu vì tớ cũng ít ra có thời gian kiểu code hay đọc code nhiều hơn Quang thôi, có lẽ thế @@.
Và làm việc trong môi trường nước ngoài và Nhật Bản nữa.
Và được mọi người cho thời gian để đọc/hiểu/ngẫm nữa.
Thành ra là cảm nhận về code, cảm nhận về cấu trúc folder các thứ của tớ cũng khác @@
Tớ thấy hay và tớ cũng về nhà tự code và bắt chước lại ý @@.
Và thấy nó cũng khá là hiệu quả với nhiều dự án khác nhau.
Tớ cũng cảm nhận được y như thế lúc đọc code bác này.
Tuy phần test chưa hoàn hảo như tớ nghĩ cơ mà khả năng là sẽ improve lên được @@.
Về đặt tên API
Cái này qua tớ với Quang cùng nhìn rồi đấy @@.
ÂPI của họ có đúng 1 từ thôi, của Quang tận 3 từ =.=
Quang thử nghĩ xem có từ tiếng Anh nào 1 từ mà bao hàm hết không @@
Về folder test
Qua tớ với Quang ngồi hì hục debug tại sao lệnh python -m test.food_create không chạy @@.
Hóa ra là do thiếu blank file __init__.py.
Cơ mà lúc về nhà tớ lại nghĩ là tác giả tại sao lại cố tình không để file đấy vào đấy.
Có lẽ tác giả không muốn chạy phải không?
Nếu mình để file đấy vào thì có sao không nếu đọc Chương 9 sau này (Chương 9 Quang có search là dùng các file Python trong folder test đấy).
Tớ nghĩ là mình nên tạo 1 folder khác kiểu như là test_food và move file food_create.py vào đấy =.= Tất nhiên có copy file __init__.py
Và remove __init__.py ở folder test kia đi để cho chắc chắn chạy về sau không sao.
Như tớ nói ở trên là làm sao merge vào master không có vấn đề gì ấy @@.
Về cấu trúc code, database
Tớ thấy Quang có tạo ra 1 app là ninehealth_logger và database model trong ý cũng có user_id, content_id, ...
Tớ cũng chưa hiểu ý của Quang lắm là sau này thế nào nhưng mà hiện tại thì có những rows như thế trong Log app rùi @@.
Vậy thì liệu có cần trong ninehealth_logger không T_T.
Do là Quang với tớ cũng mới vào vào chưa đọc kỹ thành ra chưa hiểu kỹ từng folder 1 để làm gì @@.
Giờ tớ với Quang cũng hiểu hòm hòm rùi thì tớ thấy khả năng là tạo 1 app về food và sau đó dev rồi merge và master mà MovieGEEKs vẫn chạy okay là hoàn toàn có khả năng @@.
Quang cứ thử nghĩ xem hự hự.
Làm sao để merge rồi mà vẫn chạy ý T_T.
Sorry Quang vì tớ nhận xét hơi bị tiểu tiết 1 xíu T_T.
Cơ mà đấy là xuất phát từ cảm nhận của tớ và thấy nó sao sao ý.
Tạm thời thử thì không sao nhưng sau này chưa biết có vấn đề không hự.
Tớ thấy 2 ý đầu sửa được nhưng ý thứ 3 thì chắc cần sửa nhiều, migrate nhiều các thứ :(.
Tùy xem là Quang chọn cái nào trước cái nào sau, cái nào commit trước, cái nào commit sau :(.
Cơ mà nếu được thì cứ commit tất lên rồi thực hiện mấy cái sửa đó sau, mặc dù hơi khó chịu :( Hix
Sau này nếu chỉ để chạy API liên quan đến food thì mình vẫn có thể comment (hay xóa) hết đống MovieGEEKs mà không có vấn đề gì ý Quang!!! Tớ nghĩ có thể làm được O_O.
Về test API
Cái này thì có lẽ là chuyên môn của tớ rồi vì tớ đã làm và thử nhiều môi trường =.=
Cơ mà tất nhiên cũng bị conflict với 1 số người khác và cũng phải theo họ.
Do là môi trường các anh ý architect và làm hay các bạn ý architect và làm nên mình theo thôi hix.
Thực sự tớ vẫn thích môi trường test các bác Nhật hơn vì nó đơn giản, dễ hiểu, dễ mở rộng ra với những use case khác, số dòng code ngắn.
Còn môi trường các anh hay các bạn làm loằng ngoằng lắm, regular expression loạn cả lên :(
Mặc dù đúng là ở công ty cũ tớ có dậy về Unit Test các thứ thật nhưng mà áp dụng vào chạy nó cứ sao sao ý :( Lý thuyết mà. Áp dụng rồi là không thay đổi được :(.
Đấy thì như hôm qua tớ có demo với Quang là làm sao để test 1 database model chẳng hạn.
Tại sao test cái đấy trước khi test API vì là mình test từ cái nhỏ tới cái to hay là từ vòng tròn bé ra vòng tròn to thì có gì sai ở vòng tròn to hay cái to mình sẽ debug dễ hơn =.=
Chứ cứ đùng đùng chạy test API liên quan tới HTTP các thứ hay một vài thứ khác lúc lỗi là không biết do đâu :(
Do là ông design cái app đấy hay là do ông integrate cái đấy vào cả cái MovieGEEKs web app huhu :(
Bên tớ gọi đấy là Chiến lượt verifiy (Verification Strategy). Cái này có lẽ tớ chắc cũng phải viết riêng 1 bài ở Wiki hix.
Thì tạm thời nhìn thấy code là như thế. Câu hỏi How trả lời được rồi. Câu hỏi What sẽ là test cái gì đây.
Bên bọn tớ có 1 cái gọi là Verification Plan hay Test plan cũng được sẽ liệt kê hết các test case ra trong ý, tên là gì, miêu tả thế nào, kết quả OK hay NG dựa vào report lúc chạy test.
Cái đấy thì viết bằng Excel cơ mà tớ nói thật mình làm với nhau trên GitHub rồi thì Quang cứ viết bằng file .md hay Markdown cũng được. Có gì tớ vào review và sửa cùng O_O.
Như hôm qua mình có thể gọi là Design Review cũng okay. Cũng là 1 milestone quan trọng hix T_T.
Làm xong verification list và Review cũng là 1 milestone quan trọng.
Code hết đống test case đấy trong verification list là 1 milestone quan trọng.
Chạy hết và tất cả test PASSED hay OK cũng là 1 milestone quan trọng.
Cơ mà Quang mới thì có thế cứ viết code trước đi. Viết tầm 3-4 cái tests rùi quay lại làm tài liệu cũng không sao =.=
Để riêng 1 folder kiểu như docs/test/food_unit_test_plan.md chẳng hạn ý =.= Kiểu kiểu vậy. Tên Quang cứ tự đặt nhé.
1 cái nữa về test như tớ nói hôm qua là giả sử sau này Quang update database model hay 1 function nào đó liệu có làm break đi các function cũ không hay các function khác =.=
Để trả lời được câu hỏi đấy chỉ có thể chạy lại test case và so sánh lại với golden result, expected result =.=
Như tớ nói với Quang ở bên hardware tớ thì thường sẽ có 1 C-model đi kèm và generate ra expected result đấy. Code design bọn tớ viết bằng Verilog. Cho input là register setting và data vào cùng model đấy và so sánh kết quả. Nếu khác chứng tỏ là có vấn đề ở đâu rồi =.= Thường là ở Verilog thui vì Verilog code nói chung là khó, dễ sai hơn.
Vậy thì không có C-model thì làm sao, test kiểu gì. Có 1 cách là mình sẽ output ra 1 file .txt nào đó và kiểm tra bằng mắt xem là kết quả như thế có okay không, có đúng với cái mình mong muốn (expected) không. Nếu đúng thì sẽ coi cái dấy là expected result và copy vào folder EXP
Sẽ có 1 script để so sánh kết quả log file từ RESULT folder chạy với code mới và EXP log file đã được kiểm tra bằng mắt.
Nếu đúng chứng tỏ code mình sửa không vấn đề gì
Nếu sai chứng tỏ code mình break cái gì rồi hoặc là expected result cần update lại
Tớ chỉ miêu tả lại như vậy thôi. Có gì sẽ demo dần cho Quang xem =.=
Nhìn đống phần mềm tớ thấy nói thật là quá nhiều test cases =.=
Chắc là đống CRUD Quang có khi chỉ chọn ra vài cái test chạy đúng là được chứ test kỹ corner case như bọn tớ vất vả lắm hự.
Thường hay có bug nhất là ở các giá trị max và min ấy T_T.
Mỗi 1 bạn verification engineer tự có plan riêng, có các viết test case riêng miễn sao cuối cùng verify rằng cái module đấy không còn bug ý O_O.
Thui tớ cứ nói trước vậy thôi :((( Hi vọng là Quang cứ đọc đã rồi áp dụng sau hix.
Về chiến lược test
Cái này thì nói thật mỗi người có một kiểu làm khác nhau cơ mà tớ thấy mình cũng nên lượng sức T_T.
Có 1 ví dụ là xưa các bác có 2 chiến thuật là đánh nhanh, thắng nhanh và đánh chắc, thắng chắc trong chiến dịch Điện Biên Phủ năm 1954 ấy.
Nói thật hồi đấy lực lượng quân mình cũng mạnh hơn so với năm 1945 rồi và còn được nước bạn giúp đỡ.
Cơ mà hồi năm 1945, 1946 lực lượng mỏng, chiến thuật các bác là gì. Đánh chắc, thắng chắc để bảo toàn lực lượng.
Vì ít vũ khí, đạn dược nên là mình chọn những chỗ điểm yếu của địch thôi để đánh vào O_O.
Chứ nếu mà cứ chọn chỗ bốt địch nào to và lao vào đánh thì chỉ có toang lực lượng, chả còn người nữa =.=
Đấy là hồi đầu mới thành lập Quân đội.
Còn tới chiến dịch ĐBP thì lúc ý cũng ở đẳng cấp khác nhưng mà quân Pháp cũng ở đẳng cấp khác =.=
Công sự như ở đấy là theo kiểu rất là kiên có hơn nhiều. Chiến tranh Thế giới Thứ 2 Đức có áp dụng để chống quân Liên Xô bằng kiểu đấy.
Cơ mà sau 10 năm thì nói chung vũ khí xịn hơn, có máy bay các thứ nữa, địa hình ở ĐBP cũng khác nữa nên nói thật là cách đánh cũng khác mặc dù lực lượng mình mạnh hơn rồi.
ĐBP thì có sân bay Mường Thanh là thả đồ tiếp tế xuống nhưng mà quân Pháp lại không bảo vệ kỹ nhất sân bay đấy mà lại bảo vệ trụ sở của tướng Đờ-cát.
Quân mình thì tổng cộng có 3 chiến dịch tấn công và cũng đưa pháo các thứ vào.
Nói thật pháo hồi ý chỉ là tính toán trên giấy thôi chứ cũng chưa thử bắn đâu =.=
Vậy mà chiến dịch thứ nhất đánh vào 1 phía gần sân bay Mường Thanh và thắng hồi giữa tháng 3 =.= Công nhận các bác plan khiếp thật =.= Quá là chuẩn.
Bác Giáp còn quan tâm cho bộ đội có nước nóng, được viết thư cho gia đình, còn cho cả giải trí văn công các thứ nữa ý =.= Nên là ai cũng máu chiến đấu.
Và nói thật thì cũng nhiều gương hi sinh thật nhưng mà không quá nhiều như chống Mỹ sau này =.= Tớ nghĩ thế. Lúc nào các bác ý cũng đặt ra là chiến thắng và với con số thương vong ít nhất.
Đấy là theo chiến thuật đánh chắc, thắng chắc đấy =.=
Chứ đánh nhanh, thắng nhanh chắc là toang với giờ Việt Nam chắc là chia đôi rùi =.=
Đấy thì tớ lấy ví dụ đấy có lẽ cũng không giống lắm tình trạng của mình bây giờ.
Cơ mà nói thật là tớ nghĩ mình cũng nên như thế Quang à =.=
Cứ bình tĩnh đi dần từng bước 1, từ level cao đến thấp thôi
Có nghĩa là mình sẽ đi theo hướng là
Nắm được QuerySet đã. Làm sao mà master được kiểu lấy ra 5 ingredients từ bảng SQL gốc của Quang hồi xưa ý =.= Chắc là viết sẽ mất 2-3 dòng gì đó
Lấy ra recommendation là Top kiểu nấu nhanh nhất, nấu ít ingredients nhất chẳng hạn và nhiều cái Top khác hay filter khác. Một dạng non-personalized nhưng mà tớ thấy cũng khá là hay ho =.= Nhất là với những người mà không biết mình thích gì như tớ =.= (quyển sách PRS cũng bảo thế)
Qua được 2 cái đấy rồi thì mới tính đến personalized recommendation. Hoặc là Quang có thể bắt đầu với cả công thức tính điểm của Quang cũng okay!!! 1 dạng personalized recommendation.
Qua tiếp 3 cái trên thì có thể bắt đầu dùng association rule ở Chương 5 cơ mà thực sự cũng đang hơi khó hiểu huhu. Thanks Quang đã đọc cùng tớ!!!
Tớ hi vọng là mình sẽ có nhiều công thức, nhiều cách tính điểm, nhiều cách implementation để không chỉ mình có thể dùng mà nhiều người khác như là bác sỹ có thể vào dùng chẳng hạn =.=
Mọi người chỉ cần biết 1 chút code là customize được (nghe khó nhỉ haizzz)
Nó giống như là Microsoft Lists có thể làm được vô vàn thứ chỉ bằng cách sửa file JSON =.=
https://pnp.github.io/List-Formatting/viewsamples/
Về 1 vài thứ khác
Hix sáng nay tớ dậy đầu tớ cứ bị kiểu cuốn vào mấy cái troubleshootings nên chả ngồi code hay đọc thêm được gì cả =.=
File này tớ viết độ trong khoảng 1 tiếng hay 1.5 tiếng gì đó hix.
Về quyển sách IQ hôm qua tớ kể với Quang thì ở đây @@
https://www.amazon.com/Test-Your-IQ-Explanation-Including/dp/0140249621
Sách của Nhà xuất bản Penguin (con chim cánh cụt)
Tớ không nhớ tên sách tiếng Việt luôn hix =.= May mà nhớ keyword là Penguin nên mới search lại ra được...
Có cái Trí tuệ lỏng và Trí tuệ rắn xưa chị Chuyên gia Tâm lý có nói 1 lần rồi mà chưa giải thích rõ. Trong ý có hết...
Với cả hóa ra IQ test cũng liên quan tới cả tâm lý học nữa, là 1 thước đo tâm lý psychometrics hix. Bất ngờ thật T_T.
Lúc mà test IQ cũng sẽ có kiểu test strength (sức mạnh) và speed (tốc độ). Cơ mà hướng speed thì họ hay theo hơn vì bài dễ và cho thật nhiều câu =.=
Thực ra ai làm mà nhanh thì cũng chứng tỏ rằng strength cao mà =.=
Còn IQ cũng hơi bị khác với cả sáng tạo =.=
Câu hỏi IQ là chỉ có các đáp án.
Câu hỏi sáng tạo thì mỗi người trả lời 1 kiểu, rất nhiều câu trả lời @@. Cơ mà nếu dùng xác suất thống kê thì có thể đưa ra là câu trả lời nào nhiều nhất, ít nhất @@.
Có lẽ những người mà trả lời ít nhất thì đầu óc cũng là có 1 cái sáng tạo gì đó đặc biệt @@
Và IQ thì cũng liên quan tới sóng não nữa (EEG) @@. Đoạn đấy dịch tiếng Việt cũng cứ hơi sao sao nên tớ cũng thấy nó hơi bị ngược. Nghĩa là sóng não người IQ cao và IQ thấp sẽ khác nhau ở cái số điểm lồi lõm ý @@. Theo như sách họ bảo thì là sóng não người IQ cao sẽ lồi lõm nhiều hơn vì họ nghĩ nhiều hơn khi làm 1 bài test chẳng hạn, người IQ thấp ngược lại =.= Họ giải thích thêm trong não nối với nhau thế nào nào, người nào mà kiểu tín hiệu bị gặp trắc trở suy nghĩ cũng sẽ khó khăn ý, nghĩ mãi không thông chẳng hạn @@. Còn nếu tín hiệu tốt thì suy nghĩ phát ra luôn ^_^. Vậy đó.
IQ thì chỉ tăng tới 16 tuổi thôi rồi chững lại.
IQ cao không đồng nghĩa với sống hạnh phúc hoặc là thành công :( Cái ý là tiền đề cho social intelligence thui huhu.
Ngoài ra thì còn 1 cái nữa là Quang có nói với tớ thử trên máy ở nhà (máy Windows hay Ubuntu???) mà không được ý @@.
Quang có thể commit hết cái đống Quang sửa lên 1 branch và đặt tên được không @@.
Nếu có thêm log file càng tốt.
Vì sau này kiểu gì mình cũng phải migrate lên Django phiên bản mới hơn để cập nhật hơn và tránh lỗi security ý @@@@@@@.
Bài viết này có 5K từ, 27K ký tự, dự là đọc trong vòng 20 phút!!!
Cảm ơn Quang rất là nhiều @__@
Tớ nghĩ task này sau khi Quang commit xong là đóng lại (closed) được rồi ^_^
Và Quang có thể tự mở ra 1 task mới ^_^ tùy theo ý Quang như các hướng tớ viết ở trên...
Lúc commit có thêm dấu #[issue_number] là nó tự động reference vào đây đấy, hay phết.
Cảm ơn Quang rất nhiều!!!!!!!!!!
À còn 1 cái về ý nữa là data mình lấy trên mạng về. Cái này tớ cũng chưa tìm hiểu kỹ. Có lẽ cần tạo 1 task mới hay 1 Wiki mới hix.
Đại ý là tác giả lúc distribute cái MovieGEEKs này trên GitHub thì không cho luôn database của TMDB mà kiểu lại còn viết 1 đoạn script để lấy dữ liệu về chạy tổng cộng hết 5 tiếng.
Có lẽ để dữ liệu đi kèm repo là vi phạm cái gì đó @@.
Tớ nghĩ là dữ liệu Cooky hay là Cookpad sau này mình cũng sẽ có script đi kèm để ai muốn download về và tự nghịch database thì nghịch!!!
Biết đâu có cao thủ nào đó thích nghịch thì sao!!!
Như vậy cũng giống MovieGEEKs tránh bị nói là để dữ liệu có copyright hay gì đó ở repo!!!
Thực ra script của em gì tớ quên tên mất rồi mà giúp Quang lấy dữ liệu Cooky về ý cũng giống như bác ý lấy dữ liệu TMDB về và để trong SQLite3 hay PostGreSQL!!!
Thanks Quang nhiều!!!!!!
À có thêm 1 cái nữa nếu Quang login vào trang GitHub và để ý góc dưới cùng bên phải thấy GitHub cũng recommend các repository liên quan tới ngôn ngữ của mình!!!
Tớ thấy và cũng đánh sao (starred) luôn :D.
Quang thích thử vào phần repository được đánh ưa thích của 9Health nhé ^^
|
gharchive/issue
| 2022-12-08T00:16:10 |
2025-04-01T04:54:40.140755
|
{
"authors": [
"9health",
"quangvv9Life"
],
"repo": "9health/moviegeek",
"url": "https://github.com/9health/moviegeek/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2193007467
|
🛑 instagram is down
In fc8f493, instagram (https://www.instagram.com/9renpoto/) was down:
HTTP code: 429
Response time: 336 ms
Resolved: instagram is back up in 87250c8 after 8 minutes.
|
gharchive/issue
| 2024-03-18T18:49:44 |
2025-04-01T04:54:40.160745
|
{
"authors": [
"9renpoto"
],
"repo": "9renpoto/upptime",
"url": "https://github.com/9renpoto/upptime/issues/1790",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2367466647
|
🛑 instagram is down
In 574cb6d, instagram (https://www.instagram.com/9renpoto/) was down:
HTTP code: 429
Response time: 199 ms
Resolved: instagram is back up in 6c979f2 after 13 minutes.
|
gharchive/issue
| 2024-06-22T02:24:45 |
2025-04-01T04:54:40.163028
|
{
"authors": [
"9renpoto"
],
"repo": "9renpoto/upptime",
"url": "https://github.com/9renpoto/upptime/issues/2400",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2469469762
|
🛑 threads is down
In b05bdee, threads (https://www.threads.net/@9renpoto) was down:
HTTP code: 429
Response time: 398 ms
Resolved: threads is back up in e0f80fc after 9 minutes.
|
gharchive/issue
| 2024-08-16T04:53:20 |
2025-04-01T04:54:40.165485
|
{
"authors": [
"9renpoto"
],
"repo": "9renpoto/upptime",
"url": "https://github.com/9renpoto/upptime/issues/2987",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1857649809
|
🛑 instagram is down
In 1a81314, instagram (https://www.instagram.com/9renpoto/) was down:
HTTP code: 429
Response time: 756 ms
Resolved: instagram is back up in a43e390 after 6 days, 8 hours, 45 minutes.
|
gharchive/issue
| 2023-08-19T09:25:34 |
2025-04-01T04:54:40.168232
|
{
"authors": [
"9renpoto"
],
"repo": "9renpoto/upptime",
"url": "https://github.com/9renpoto/upptime/issues/708",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
332395349
|
Explanation of Heroku
If possible, could a concise explanation of what heroku is and what parts are online/offline be made? Most folks are going to associate an app on a Google Chrome tab as connected to the internet and uploading files means uploading to the internet.
A quick note in the help banner or on this Github page would go a long way toward easing that confusion.
So Heroku doeasn't actually have anything to do with the app, and a casual user of the app shouldn't ultimately know that it exists (let alone that they're using it). However, a little exposition on the offline/online nature of the app may be a good addition. I'll write a wiki page about it.
https://github.com/AABoyles/WebMicrobeTrace/wiki/Offline-Usage
|
gharchive/issue
| 2018-06-14T13:08:12 |
2025-04-01T04:54:40.193773
|
{
"authors": [
"AABoyles",
"CopeSc"
],
"repo": "AABoyles/WebMicrobeTrace",
"url": "https://github.com/AABoyles/WebMicrobeTrace/issues/53",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
340134792
|
你好 我不懂一个类型的变量
我没看过这种变量, 请问该怎么生成
dataSet(@[@7.0, @6.9, @9.5, @14.5, @18.2, @21.5, @25.2, @26.5, @23.3, @18.3, @13.9, @9.6]),
@[@] 没见过这种 谢谢
兄弟你是 iOS 开发新手吧?这个里面的@符号其实就是一个语法糖啊.
@[]这个是快速声明一个不可变数组 NSArray 的意思.相当于
NSArray *array = [NSArray array];
或者是
NSArray *array = [[NSArray alloc]init];
又因为 OC 中的数组只能够存储对象,不能够存储基本的数据类型比如布尔值,整形数字,浮点数字等,所以想要在数组中存储这些基础的数据类型数字,就必须要将其转化为 OC 对象.
@7.0也是语法糖,这是一个将 OC 当中的 float 类型转化为 NSNumber 的一个语法糖,相当于是
float floatNumber = 7.0;
NSNumber *myNumber = [NSNumber numberWithFloat:floatNumber];
|
gharchive/issue
| 2018-07-11T08:08:12 |
2025-04-01T04:54:40.196091
|
{
"authors": [
"AAChartModel",
"mozhanmeng"
],
"repo": "AAChartModel/AAChartKit",
"url": "https://github.com/AAChartModel/AAChartKit/issues/358",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
679668756
|
Pago para quem desenvolver DBSeller
Boa noite pessoal,
Alguém interessado em desenvolver o provedor DBSeller ?
Quem estiver com tempo e quiser devenvolver, me avise que eu pago.
Dados de uma cidade do provedor:
[4321600]
Nome=Tramandai
UF=RS
Provedor=DBSeller
NomeURL_H=http://hom.nfse.tramandai.rs.gov.br
NomeURL_P=http://nfse.tramandai.rs.gov.br
@rftd Tem interesse ?
Eu acabei de finalizar um refactory que tava promovendo no componente, eu posso dar uma olhada neste provedor.
@transis2 , se o Rafael conseguir encaixar na agenda dele, pode contratar...
Fiz isso no provedor do RJ (se não me engano)... me entregou no prazo e praticamente quase todo funcionando (eu fiz os testes com o certificado do cliente, anotei os problemas e na rodada seguinte, tudo funcionando.)...
Recomendo!!!
@rftd, quanto tu cobraria ?
@transis2 entra no gitter
|
gharchive/issue
| 2020-08-16T01:15:13 |
2025-04-01T04:54:40.227220
|
{
"authors": [
"rafd75",
"rftd",
"transis2"
],
"repo": "ACBrNet/ACBr.Net.NFSe",
"url": "https://github.com/ACBrNet/ACBr.Net.NFSe/issues/109",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2449788351
|
37 update input paths pre-industrial
This pull request updates the pre-industrial configuration with the new input paths using the restructured directory.
Note that the md5 hash for pftlookup_csiro_v16_17tiles_wtlnds.csv changed as there are slight differences between the ~access/payu and the vk83 versions of the file which but the differences don't appear to be used by the model.
I've had a go at putting in those changes, and also swapped from the experiments to the new configuration directory. Everything looks like it's working when running a simulation with these changes.
Good point! I've just added the new module path in.
|
gharchive/pull-request
| 2024-08-06T02:14:19 |
2025-04-01T04:54:40.229456
|
{
"authors": [
"blimlim"
],
"repo": "ACCESS-NRI/access-esm1.5-configs",
"url": "https://github.com/ACCESS-NRI/access-esm1.5-configs/pull/38",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
272758699
|
Nightly
Fix for database using the wrong file issue, adds mpas*_in files, updates install steps.
All tests passing.
|
gharchive/pull-request
| 2017-11-09T23:00:34 |
2025-04-01T04:54:40.258423
|
{
"authors": [
"sterlingbaldwin"
],
"repo": "ACME-Climate/acme_processflow",
"url": "https://github.com/ACME-Climate/acme_processflow/pull/63",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
849586190
|
Renaming the attendance page and displaying subjects and labs which student has on the particular day.
Deliverables:
[ ] Rename the attendance page to Today.
[ ] Display only those subjects and labs which a student has on that day.
Screenshots
I would like to work on this issue. @Sanyam2000-dot
@Sanyam2000-dot @ADRE9 The default cards doesn't contains any days values, it would be better if someone could add edit those values, it would be easier for me to work.
Actually there is a days feature. But since for the first time the user gets template subjects , it is empty. As soon as the user starts editing subject it becomes necessary to add the days to it. And then you can show it in the today's page. We are imagining a page where a new registered user will be introduced to time table so taht the user can easily add and change time table. If you want to work on it before. Talk to @kunaljain0212 or @Sanyam2000-dot , they will help you to create an issue and and work on it.
@ADRE9 Ok thanks I got your point
@ADRE9 @Sanyam2000-dot check if this is the needed functionality.
https://user-images.githubusercontent.com/16490253/113484574-ad7cba00-94c6-11eb-940d-aedb337701d9.mp4
Yes Exactly ! Change the Route from "attendance" to "today". And Classes and Labs to Today's Classes and Today's Labs respectively.
@ADRE9 Ok sure
@ADRE9 should the Today Page contain the default cards?
No only what's today.
@ADRE9 how this looks?
And as per Route I think we should make another issue if we want to update route to '/today' as it will need a lot of changes to many files. Currently it is routed to '/' and I think it's better to keep that.
No by Route I meant the Link component that is now labelled as Attendance.
@ADRE9 Can you please help me with this part, I can't seem to get it.
@ADRE9 ok I got what you're trying to say, working on that
@ADRE9 @Sanyam2000-dot kindly review the PR.
Looks good to me. @ADRE9 please review it.
|
gharchive/issue
| 2021-04-03T05:56:14 |
2025-04-01T04:54:40.280531
|
{
"authors": [
"ADRE9",
"Artistic18",
"Sanyam2000-dot"
],
"repo": "ADRE9/bunk-manager-mern",
"url": "https://github.com/ADRE9/bunk-manager-mern/issues/102",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1056562489
|
yield on wait, no wait before command, SYNC timing, add setGain()
ESP32 has 2 cores and can use multiple tasks, waiting in a nop loop is not ideal
waitDRDY blocks at second command if first command does not lead to a DRDY
Longer delay after SYNC command: 24 cycles according to datasheet, not 4
Wrote a small test program for setGain()
#include <Arduino.h>
#include <ADS1256.h>
#include <SPI.h>
// Initialize ADS1256 object
ADS1256 adc(7.68, 2.5, false); // clockSpeed in Mhz, VREF in volt, if use RESET PIN
const uint8_t chan = 2; // channel to use
uint8_t gain = 0; // first gain value
volatile bool ready = false; // set to true on DRDY
// Interrupt service routine triggered by DRDY
void IRAM_ATTR readyIsr() {
ready = true;
}
// Wait for DRDY (or return immediately if not fromNowOn and ready is already true)
// Return true if DRDY and false if timeout without DRDY
bool readyWait( uint32_t timeout_ms = 2000, bool fromNowOn = true ) {
if( fromNowOn ) ready = false;
uint32_t start = millis();
while( !ready ) {
if( millis() - start > timeout_ms ) return false; // wait timed out
yield();
}
ready = false;
return true;
}
// Set new gain on key press 0..7 on serial console
// Return true if valid key pressed
bool handleKeypress( uint8_t &gain ) {
bool rc = false; // no gain change
while( Serial.available() ) {
int ch = Serial.read();
if( ch >= '0' && ch <= '7' ) {
gain = ch - '0';
rc = true; // gain has changed
}
}
return rc;
}
// Called on fatal errors: print message and do endless loop
void fatal( const char *msg ) {
Serial.println(msg);
while( true );
}
// Trigger self calibration and wait for it to finish
void calibrate() {
adc.sendCommand(ADS1256_CMD_SELFCAL);
if( !readyWait() ) {
fatal("Calibration failed!");
}
}
// Called at starup.
// Bring the chip into a state where valid adc readings can be done
void setup() {
Serial.begin(115200);
Serial.println("\nADS1256 gain test " __FILE__ " " __TIMESTAMP__);
attachInterrupt(pinDRDY, readyIsr, FALLING);
adc.sendCommand(ADS1256_CMD_RESET);
if( !readyWait() ) {
fatal("Reset failed!");
}
adc.setChannel(chan);
adc.begin(ADS1256_DRATE_100SPS, gain, true);
}
// Called after setup() over and over again
// Try to read new gain value from serial console
// If a new gain is requested set it with a following calibration
// Read an adc value, convert it to a voltage and print it
void loop() {
if( handleKeypress(gain) ) {
adc.setGain(gain);
calibrate();
Serial.print("\ngain: ");
Serial.println(gain);
}
Serial.printf("\r%.6f Volt ", adc.readCurrentChannel());
delay(100);
}
Result is a bit strange/disappointing: the voltage differs depending on the gain level although in range for all of them.
Also the values drift quite a bit even after a new self calibration.
The actual voltage value according to my multimeter should be 3.3 mV and stable (+/- 0.1 mV) .
ADS1256 gain test src/main.cpp Thu Nov 25 01:05:45 2021
0.003504 Volt
gain: 0
0.003495 Volt
gain: 1
0.003465 Volt
gain: 2
0.002375 Volt
gain: 3
0.001624 Volt
gain: 4
-0.001489 Volt
gain: 5
-0.017101 Volt
gain: 6
-0.017548 Volt
gain: 7
-0.008815 Volt
gain: 0
0.001982 Volt
Buffer on or off makes no difference. I tried both.
I checked your conversion formula with the datasheet and found no error. Any ideas?
|
gharchive/pull-request
| 2021-11-17T19:48:29 |
2025-04-01T04:54:40.284123
|
{
"authors": [
"joba-1"
],
"repo": "ADS1xxx-Series-ADC-Libraries/ADS1255-ADS1256",
"url": "https://github.com/ADS1xxx-Series-ADC-Libraries/ADS1255-ADS1256/pull/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2633338598
|
Very easy to bypass
console.log = () => {};
console.table = () => {};
Function.prototype.constructor = () => {};
setInterval = () => {};
for (let i=0;i <1e6; i++) {
clearInterval(i);
}
delete Navigator.prototype.userAgent
And a lot more.
The only people likely to open the developer tools intentionally are those who know exactly what they’re doing, which might make this method less effective in practice, even if it technically works as designed.
The only case I can imagine is comical one:
@ Open devtool
@ Sees it has devtool protection
@ Nah, I am too lazy to go any further
However, I really appreciate the impressive amount of research that’s gone into ensuring compatibility across nearly all browsers. Thanks for all the hard work!
By the way, here another way to stuck the browser without ability to debugg it during crash trigger:
(async function f() {f(), f()})();
Yes, this library merely adds a little bit of difficulty for the debuggers. If there's a higher demand for anti-debugging, you can try WebAssembly or JVM, or implement the core logic on the server side.
@AEPKILL, my focus is more on strengthening the system—using non-overridable methods and similar techniques. I’m researching this extensively and working to counter script-based attacks, which are the most common. I’ve developed anti-cheat measures for several browser games, and while I haven’t yet implemented my strongest concept, it's beyond any existing protections out there.
Maybe I will show later some of my tricks and how they can be used to protect this library from almost any "script based attacks"
@doctor8296 Your ideas are very interesting. I'm particularly intrigued by the unredefinable methods you mentioned. Can you share some more details about how they work?
@AEPKILL sure, but I'd prefer to talk privately about it, because some of protection methods us not public yet
@doctor8296 Sure, I understand your need for privacy. I think email would be a great way to have this discussion. Please send your thoughts and details to my email address at [a@aepkill.com]. Looking forward to hearing from you soon!
@doctor8296 It's almost 11 pm local time. I'm going to bed now. Sorry I might not reply to your emails and messages promptly.
|
gharchive/issue
| 2024-11-04T16:56:58 |
2025-04-01T04:54:40.290646
|
{
"authors": [
"AEPKILL",
"doctor8296"
],
"repo": "AEPKILL/devtools-detector",
"url": "https://github.com/AEPKILL/devtools-detector/issues/76",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
177627582
|
Miss imageResponseSerializer in UIImageView+AFNetworking with version 3.1.0
Do you remove this property?
@property (nonatomic, strong) id <AFURLResponseSerialization> imageResponseSerializer;
See AFImageDownloader. That class now controls the session that downloads images.
|
gharchive/issue
| 2016-09-18T03:53:04 |
2025-04-01T04:54:40.311498
|
{
"authors": [
"iTofu",
"kcharwood"
],
"repo": "AFNetworking/AFNetworking",
"url": "https://github.com/AFNetworking/AFNetworking/issues/3685",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
614469627
|
iOS not support uiwebview。please update
iOS not support uiwebview。please update
Use v4 instead.
Version 4.0 has removed uiwebview related code
Apple is rejecting apps using UIWebView instead of WKWebView. So if AFNetworking 2.x is still being used (which it is) this issue needs to be resolved.
AFNetworking 2 and 3 are no longer supported. If you want a supported version, use AFNetworking 4.
|
gharchive/issue
| 2020-05-08T03:02:14 |
2025-04-01T04:54:40.312987
|
{
"authors": [
"ElfSundae",
"TongFangyuan",
"elaget",
"jshier",
"zengjialong"
],
"repo": "AFNetworking/AFNetworking",
"url": "https://github.com/AFNetworking/AFNetworking/issues/4566",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2636686101
|
Establish Working CONTRIBUTING
The author is not satisfied with the current state of CONTRIBUTING.md and wants to update it.
Contributing Guidelines
The author(s) of this repository appreciate external feedback and contributions.
To improve this feedback process, we urge you to read and adhere to the
following guidelines:
Provide clear, evidenced arguments for your changes, where possible this will aid future onlookers in understanding why certain features came to be the way they are. Additionally, this facilitates the author(s) in learning why and how they've erred. Finally, taking the time to make your arguments clearer can prevent the creation of feedback that is misguided, incorrect, or inadequate relative to other possible changes.
[additional guidelines pending]
The updated contents should be general enough such that they can be still effective across most of the author's repositories without further edit. To some degree, the author wants CONTRIBUTING.md to reflect the author's preferred norms around code-development and communication.
The following are example CONTRIBUTING files to work from:
https://github.com/Yelp/detect-secrets/blob/master/CONTRIBUTING.md
https://github.com/python/mypy/blob/master/CONTRIBUTING.md
The template CONTRIBUTING can be included as a dropdown in the template CONTRIBUTING.
Notes and considerations for the content of CONTRIBUTING should be stored in
https://github.com/AFg6K7h4fhy2/AFg6K7h4fhy2/tree/main/decisions
|
gharchive/issue
| 2024-11-05T23:11:09 |
2025-04-01T04:54:40.318764
|
{
"authors": [
"AFg6K7h4fhy2"
],
"repo": "AFg6K7h4fhy2/AFg6K7h4fhy2-Template",
"url": "https://github.com/AFg6K7h4fhy2/AFg6K7h4fhy2-Template/issues/8",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1856772737
|
Lora locations
@kshitijkg
The only files modified are megatron/training.py, and megatron/models/adapter.py
|
gharchive/pull-request
| 2023-08-18T13:57:44 |
2025-04-01T04:54:40.321317
|
{
"authors": [
"daniel-z-kaplan"
],
"repo": "AGI-Collective/multimodal",
"url": "https://github.com/AGI-Collective/multimodal/pull/45",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1393461136
|
Output trained RL models for use in other projects
Wondering if there is a way to output trained RL models from FinRL for the purpose of my own research?
Personally, I feel FinRL is too encapsulated, and there seems no way to output the trained models with fine-tuned weights and hyper-parameters without making changes inside Stable-baselines3?
Anyone can help?
Thanks.
Yes, you can use pytorch to save the model.
Yes, you can use pytorch to save the model.
Is there such an example in the tutorials or somewhere? Thanx.
Hi @julian-z
You can refer to this link https://stable-baselines3.readthedocs.io/en/master/guide/save_format.html to save and load model in SB3
|
gharchive/issue
| 2022-10-01T16:15:35 |
2025-04-01T04:54:40.335850
|
{
"authors": [
"Athe-kunal",
"XiaoYangLiu-FinRL",
"julian-z"
],
"repo": "AI4Finance-Foundation/FinRL",
"url": "https://github.com/AI4Finance-Foundation/FinRL/issues/727",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
540028140
|
[BUG] Quotes in the AI response sometimes prematurely trigger the end of the response.
🐛 Bug Report
Describe the bug
Quotes in the AI response sometimes trigger the end of the response.
How to Reproduce
Play through with a lot of Dialog: say "blah blah blah".
This leads to responses with a lot of dialog, and situations where quotes cause the end of the AI response.
"She slaps you. You say "
Expected behavior
The response should reach a reasonable and logical end, not stopping in the middle.
Additional context
Colab or local machine?
Both
master or develop branch?
Master
(add any other context about the problem here)
This seems likely to be a training problem where the model was probably given clipped data in it's training where the quotes were stripped but it could be a parsing error where you're failing to parse the output correctly and are stripping important quotes and dialog. The training issue seems more likely since this problem is not consistent, where sometimes the quotes actually are retained.
Looks like same issue as #90: [BUG] Dialog items not show.
Agreed, closing for duplication with #90
|
gharchive/issue
| 2019-12-19T02:01:48 |
2025-04-01T04:54:40.339955
|
{
"authors": [
"bryanlyon",
"snarfblam"
],
"repo": "AIDungeon/AIDungeon",
"url": "https://github.com/AIDungeon/AIDungeon/issues/200",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2111619944
|
Add simplified chart image for top of report
[x] For reports with too many clusters for legible bubble graphs, or for applications where populations want a simplified view, we should show a less interactive chart.
[x] For now, this should be the default view. Chart should use updated color scheme from #116 (ignore colors in mocks below). Margins and text sizes in the mocks below are final.
[x] 1. Minimum viable version: bar chart with topics & descriptions
[x] - Bar graph shows the topic title and (brief) description for each topic. Title and description should each be a maximum of two lines (add ellipsis if it continues past two lines).
[x] - Bars should include text showing number of claims in the bar (bold white text, left-aligned)
[x] - Bar chart should have a single, vertical, unlabeled axis (no horizontal axis)
[x] 2. Target version: hover behavior
[x] When any part of a topic is hovered over (title, description, or bar area):
[x] - that entire row of the bar graph should be highlighted in a 10% opacity version of that bar's color
[x] - the text "Click to view topic" should appear in small italic text below the bar. To ensure there is room for this, each topic's text should have a min-height set (see mock)
[ ] 3. Nice-to-have: subtopics on hover
[ ] On hover, a topic should expand to include a list of subtopics in that cluster.
[ ] - "Subtopics" heading should show the total number of subtopics
[ ] - Subtopics should be limited to one line of text (elided)
[ ] - Subtopics should be delimited by commas and three extra spaces, for legibility
@lightningorb I'm hoping we can use an existing bar chart library for a lot of this; if you find one that comes close to supporting these designs but isn't quite there, lmk and I can try to revise designs to make it work
Calling this done! We can revisit subtopics on hover if subtopics aren't discoverable enough via sidebar and per-topic graphs, but I don't expect that to be the case
|
gharchive/issue
| 2024-02-01T05:58:27 |
2025-04-01T04:54:40.389667
|
{
"authors": [
"colleenm"
],
"repo": "AIObjectives/talk-to-the-city-reports",
"url": "https://github.com/AIObjectives/talk-to-the-city-reports/issues/115",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1770894139
|
💡 [REQUEST] - Enhance Contact Us form
[x] I am from GSSoC'23.
[x] I would like to be assigned this task.
Code of Conduct
By submitting this issue, you agree to follow our Code of Conduct.
[x] I follow the Contributing Guidelines of this project.
[x] I have checked the current issues for duplicate problems.
[x] I have read the readme of this repository.
Already assigned issue.
|
gharchive/issue
| 2023-06-23T06:57:26 |
2025-04-01T04:54:40.398308
|
{
"authors": [
"AKD-01",
"MisterMickey"
],
"repo": "AKD-01/blogweet",
"url": "https://github.com/AKD-01/blogweet/issues/480",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2280087665
|
Failed to open database '/data/data/com.baidu.searchbox/database/pushinfo.db'.
应用名:百度
测试手机:华为p20 Android 10
(1)
Failed to open database '/data/data/com.baidu.searchbox/database/pushinfo.db'.
android.database.sqlite.SQLiteCantOpenDatabaseException: unknown error (Sqlite code 14 SQLITE_CANTOPEN): Could not open database, (OS error - 13:Permission denied)
(2)
E Failed inserting a file into the cache
com.facebook.common.file.FileUtils$CreateDirectoryException: /storage/emulated/0/Android/data/com.baidu.searchbox/cache/image_cache
应用名:百度 测试手机:华为p20 Android 10
(1) Failed to open database '/data/data/com.baidu.searchbox/database/pushinfo.db'. android.database.sqlite.SQLiteCantOpenDatabaseException: unknown error (Sqlite code 14 SQLITE_CANTOPEN): Could not open database, (OS error - 13:Permission denied)
(2) E Failed inserting a file into the cache com.facebook.common.file.FileUtils$CreateDirectoryException: /storage/emulated/0/Android/data/com.baidu.searchbox/cache/image_cache
did you check the permisson ?
NewBlackbox 没有hook native c/c++ io 函数,native c/c++ open(),lstat()等没有对路径进行重定向
NewBlackbox does not hook native c/c++ IO function, native c/c++ open(), lstat() etc. do not redirect paths.
Test code
Vfs.h
#ifndef VFS_H
#define VFS_H
#include "Log.h"
#define SHADOWHOOK_LIBC_PATH "/apex/com.android.runtime/lib64/bionic/libc.so"
#define SHADOW_HOOK_DEF(ret, func, ...) \
void * func##_orig = NULL; \
void * func##_stub = NULL; \
ret func##_proxy(__VA_ARGS__)
#define SHADOW_HOOK_SYMBOL(func) \
func##_stub = shadowhook_hook_sym_name(SHADOWHOOK_LIBC_PATH, #func, (void *)func##_proxy, &func##_orig); \
if(func##_stub == NULL) { \
ALOGD("VfsHook Failed -------------- %s", #func); \
}else{ \
ALOGD("VfsHook Success-------------- %s %p", #func, func##_orig); \
}
namespace Vfs {
void init();
}
#endif //VFS_H
Vfs.cpp
#include "Vfs.h"
#include "IO.h"
#include "shadowhook.h"
SHADOW_HOOK_DEF(int, open, const char *pathname, int flags, int mode) {
SHADOWHOOK_STACK_SCOPE();
const char * redirectPath = IO::redirectPath(pathname);
ALOGD("Vfs open -------------- %s ---- %s", pathname , redirectPath);
return SHADOWHOOK_CALL_PREV(open_proxy, redirectPath, flags, mode);
}
void Vfs::init(){
if(shadowhook_init(SHADOWHOOK_MODE_SHARED, false) == 0 ){
SHADOW_HOOK_SYMBOL(open);
}
}
Log
NativeCore: Vfs open -------------- /data/data/com.baidu.searchbox/database/pushinfo.db ---- /data/user/0/com.vspace/blackbox/data/user/0/com.baidu.searchbox/database/pushinfo.db
NewBlackbox 没有hook native c/c++ io 函数,native c/c++ open(),lstat()等没有对路径进行重定向
NewBlackbox does not hook native c/c++ IO function, native c/c++ open(), lstat() etc. do not redirect paths.
是这个问题吗?
|
gharchive/issue
| 2024-05-06T05:53:05 |
2025-04-01T04:54:40.425830
|
{
"authors": [
"ALEX5402",
"shihangfresh"
],
"repo": "ALEX5402/NewBlackbox",
"url": "https://github.com/ALEX5402/NewBlackbox/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
916613388
|
Add smooth transition on hover
Feature request-
In the testimonials section, when each box is hovered upon, the transition is kind of harsh
Describe the solution you'd like
I would like to add a smooth transition to resolve this jerky motion.
/assign
Note:
If you want to work on an issue, you should check if it has already been assigned to anyone. If the issue is free you can comment /assign to get the issue assigned to you.
If you are raising a new issue and want to work on it then also you should comment /assign under the issue to get it auto assigned.
Please refrain from adding labels to your issue/pull-request on your own. It is the job of the Project Admin and the Mentors to review your issue/pull-request and add labels accordingly.
/reviewed
You want to increase the time of the animation right ?
Also, you should comment the keyword below the issue to get it assigned, I'll assign it this time. You can keep it in mind for the next time 👍
Yes, precisely. I am working on this issue under LGM SOC'21...
Yes, precisely. I am working on this issue under LGM SOC'21...can you please add the label?
Yes, the project admins have not received the official list of tags to be added for lgm soc, official list is imp. because tags can be case sensitive. I'll add the tags when I get it from the organizers, meanwhile you can work on this 👍
|
gharchive/issue
| 2021-06-09T19:39:00 |
2025-04-01T04:54:40.432642
|
{
"authors": [
"Pranav016",
"supzi-del"
],
"repo": "ALPHAVIO/WordNook",
"url": "https://github.com/ALPHAVIO/WordNook/issues/1000",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1041904613
|
chat 초대하기 검색 버그
버그 내용
검색 input을 비워놓은 상태로 검색 버튼을 누르면 아래와 같이 됨
버그 스샷
확인했습니다. 예외처리할게요
|
gharchive/issue
| 2021-11-02T05:42:35 |
2025-04-01T04:54:40.434165
|
{
"authors": [
"AMATEURTOSS",
"DomMorello"
],
"repo": "AMATEURTOSS/ft_transcendence",
"url": "https://github.com/AMATEURTOSS/ft_transcendence/issues/306",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2046948897
|
[bug] Test release keystore which is generated via Github Action
WIP: [bug] Test release keystore which is generated via Github Action
Closing this ticket as the investigation with signing key was completed.
|
gharchive/pull-request
| 2023-12-18T15:45:27 |
2025-04-01T04:54:40.500545
|
{
"authors": [
"tuancoltech"
],
"repo": "ARK-Builders/ARK-Memo",
"url": "https://github.com/ARK-Builders/ARK-Memo/pull/41",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
228189208
|
OpenCL Error in examples
I build the library with OpenCL support using the stub library, and tested the example cl_convolution on a Samsung Galaxy S4.
At the line:
CLScheduler::get().default_init();
The app crashed with the following error:
terminate called after throwing an instance of 'cl::Error'
what(): empty
Further testing:
If I comment out the CLScheduler, the program runs until
CLConvolution3x3 conv3x3;
I where it gives the error:
ERROR in const arm_compute::Program &arm_compute::CLKernelLibrary::load_program(const std::string &) const src/core/CL/CLKernelLibrary.cpp:553: Kernel file .convolution3x3.cl does not exist. No such file or directory
The path to libopencl.so is included in the phone&s LD_LIBRARY_PATH. Is there anything else that I have missed?
Samsung Galaxy S4 has a Qualcomm GPU so not sure what the OpenCL situation is for this platform.
default_init() just creates an OpenCL Context and Command Queue and the cl::Error get thrown by the OpenCL C++11 wrapper (cl2.hpp), which means it's an error returned by the OpenCL runtime at initialisation time. Not sure what it could be I'm afraid, we haven't seen it on any of the devices we've tested the library on
Thanks, I started testing on a HTC_M9pw and the first error went away. So I assume that was a device problem.
However, I still get an error at:
conv3x3.configure(&src, &tmp, gaussian3x3, 0, BorderMode::UNDEFINED);
The error message reads:
ERROR in const arm_compute::Program &arm_compute::CLKernelLibrary::load_program(const std::string &) const src/core/CL/CLKernelLibrary.cpp:553: Kernel file ./cl_kernels/convolution3x3.cl does not exist. No such file or directory
My build statement is:
CXX=clang++ CC=clang scons Werror=1 -j8 debug=1 neon=0 opencl=1 os=android arch=armv7a
You either need to upload the CL kernels from src/core/CL/cl_kernels to your platform or to recompile the library with embed_kernels=1 (then the OpenCL kernels will directly be embedded inside the library's binary)
Thanks, I added embed_kernels=1 and rebuilt the library, and the missing kernel file error disappeared. However, the same line
conv3x3.configure(&src, &tmp, gaussian3x3, 0, BorderMode::UNDEFINED);
gives me the error
error: unknown argument: '-cl-arm-non-uniform-work-group-size'
terminate called after throwing an instance of 'cl::Error'
what(): clCreateKernel
While searching for a way to fix this, I found the following statement in a Khronos web page.
Enabling non-uniform work-groups requires a kernel to be compiled with the -cl-std=CL2.0 flag and without the -cl-uniform-work-group-size flag.
Is this relevant to the error? If so, how do I build the kernels for CL2.0?
That's because the OpenCL driver you are using doesn't support the -cl-arm-non-uniform-work-group-size flag (Which means your GPU is probably not an ARM GPU), if your GPU supports CL 2.0 you can try to replace -cl-arm-non-uniform-work-group-size by -cl-std=CL2.0` here
Many thanks, that fixed the problem. In my case, it was -cl-std=CL1.1.
@mo-morikawa , hi, when you face the problem "default_init()" and "conv3x3.configure(&src, &tmp, gaussian3x3, 0, BorderMode::UNDEFINED);" is the executable program or the apk in the Andriod ?
When use the demo of "cl_convolution" executable program is successful on mate9. But when change the code to the Shared Library for apk ,I face the same problem to yours.
thank you.
I got the error when it was running in the device as an executable. So far, I haven't tried making an apk. Will try that some time, since I am going to need it later.
Hi,
I am new to the ARM compute library. I build the library with flags opencl=1 and neon=1 by cross-compiling for android.
I tested running graph_alexnet and other graphs without any problem.
Now I want to run it with --target==CL, but I will get the error:
ERROR in generate_build_options src/core/CL/CLCompileContext.cpp:265: Non uniform workgroup size is not supported!! No such file or directory
!!!!!!!!!!!!!!!!!!!!!!!!!!!
My platform is Khadas vim3, with Mali G52 GPU (Bifrost family). when I debug it, see that GPU is detected as Midgard!
I do not know exactly what .so files should find and use, or how I can see the version of OpenCL. I know that this GPU supports OpenCL 1.2 and 2.0.
I appreciate it if somebody could help me.
Best,
Ehsan
|
gharchive/issue
| 2017-05-12T05:09:37 |
2025-04-01T04:54:40.513556
|
{
"authors": [
"AnthonyARM",
"Ehsan-aghapour",
"mo-morikawa",
"zhangbaokun"
],
"repo": "ARM-software/ComputeLibrary",
"url": "https://github.com/ARM-software/ComputeLibrary/issues/104",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
867784236
|
RTOS Examples
For starters, as someone who has been following WebAssembly since 2017, this is extremely exciting and novel work. 👏
Are there plans to add examples being incorporated with a modern RTOS? I see CMSIS is already used in some places and I'd imagine Arm would like to have a demonstration with mbed. I've been involved with Zephyr lately and I can envision Bento working quite well with west, Zephyr's meta tool.
Funny how we end up down the same back alley's @beriberikix
|
gharchive/issue
| 2021-04-26T14:44:46 |
2025-04-01T04:54:40.516262
|
{
"authors": [
"0Grit",
"beriberikix"
],
"repo": "ARM-software/bento-linker",
"url": "https://github.com/ARM-software/bento-linker/issues/1",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
540373522
|
tests: staging: sched_android: Ensure frequencies are sorted
Cpufreq's scaling_available_frequencies sysfs file does not guarantee
any particular order for the frequencies. As such, on devices where
they're not sorted in increasing order, the schedtune frequency test
will fail to compute the right target frequency, leading to test
failures.
Fix this by sorting the frequencies when necessary in sched_android.py.
Signed-off-by: Quentin Perret qperret@google.com
@qperret FYI that PR will ensure that provided list of CPUs and frequencies are sorted, so it will raise an exception if devlib/the user is not doing it:
https://github.com/ARM-software/lisa/pull/1272
As commented, it's probably better to sort the output of devlib functions so that everyone benefits from it rather than just fixing it in PlatformInfo.
|
gharchive/pull-request
| 2019-12-19T15:12:20 |
2025-04-01T04:54:40.519136
|
{
"authors": [
"douglas-raillard-arm",
"qperret"
],
"repo": "ARM-software/lisa",
"url": "https://github.com/ARM-software/lisa/pull/1262",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
390980588
|
Broken link for combine script
The build instruction here explains:
Use this script to combine the bootloader with application python tools/combine_bootloader_with_app.py -a {application.bin} -b {bootloader.bin} --app-offset {application-start-address} --header-offset {firmware_metadata_header_address} -o {combined.bin}.
But the link is broken because of obsoleting in the mbed-cloud-client-example 2.0.0 release.
The example uses Mbed OS bootloader binaries and the new Mbed OS feature FEATURE_BOOTLOADER. This feature makes the combine_bootloader_with_app.py script obsolete. We have removed the obsolete script and old bootloader binaries from the application repository.
The repository was deprecated and it's being archived now. We are closing all issues/pull requests.
|
gharchive/issue
| 2018-12-14T06:17:13 |
2025-04-01T04:54:40.530738
|
{
"authors": [
"0xc0170",
"toyowata"
],
"repo": "ARMmbed/mbed-bootloader-extended",
"url": "https://github.com/ARMmbed/mbed-bootloader-extended/issues/2",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
246075245
|
Backport 1.3 - Rename time and index parameter to avoid name conflict.
This is the backport of #886 to mbed TLS 1.3.
Reviewed, approved and merged into the private mbedtls-1.3 branch, pending release.
|
gharchive/pull-request
| 2017-07-27T15:16:48 |
2025-04-01T04:54:40.697314
|
{
"authors": [
"hanno-arm",
"sbutcher-arm"
],
"repo": "ARMmbed/mbedtls",
"url": "https://github.com/ARMmbed/mbedtls/pull/1031",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
686209890
|
Feature/import cnzz
feat: 在所有页面中都引入cnzz统计
feat: 完善统计脚本的页面覆盖度
|
gharchive/pull-request
| 2020-08-26T10:33:37 |
2025-04-01T04:54:40.723140
|
{
"authors": [
"ATQQ"
],
"repo": "ATQQ/sugar-blog",
"url": "https://github.com/ATQQ/sugar-blog/pull/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
504087601
|
As an admin, I want to add remarks to a student page.
... so that I can refer back at a later time if needed.
Refactored and under approve appeal command now
|
gharchive/issue
| 2019-10-08T14:41:40 |
2025-04-01T04:54:40.774060
|
{
"authors": [
"MistyRainforest",
"chensu2436"
],
"repo": "AY1920S1-CS2103-T11-2/main",
"url": "https://github.com/AY1920S1-CS2103-T11-2/main/issues/54",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
729010737
|
Change print format
Closes #196
Codecov Report
Merging #198 into master will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #198 +/- ##
=========================================
Coverage 82.59% 82.59%
Complexity 1061 1061
=========================================
Files 151 151
Lines 3160 3160
Branches 379 379
=========================================
Hits 2610 2610
Misses 457 457
Partials 93 93
Impacted Files
Coverage Δ
Complexity Δ
...t/logic/commands/attendance/StatisticsCommand.java
91.17% <ø> (ø)
11.00 <0.00> (ø)
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update ad00e41...1e36e50. Read the comment docs.
|
gharchive/pull-request
| 2020-10-25T12:31:21 |
2025-04-01T04:54:40.784097
|
{
"authors": [
"codecov-io",
"samlsm"
],
"repo": "AY2021S1-CS2103T-T10-4/tp",
"url": "https://github.com/AY2021S1-CS2103T-T10-4/tp/pull/198",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1021830375
|
Update JavaDoc and resolve Java coding style violation
Fix #62
Update JavaDoc for isName() method in StringUtil.java
Use Egyptian style for curly bracket
Codecov Report
Merging #62 (e504db1) into master (c8f6c04) will not change coverage.
The diff coverage is 100.00%.
@@ Coverage Diff @@
## master #62 +/- ##
=========================================
Coverage 68.71% 68.71%
Complexity 424 424
=========================================
Files 75 75
Lines 1397 1397
Branches 156 156
=========================================
Hits 960 960
Misses 387 387
Partials 50 50
Impacted Files
Coverage Δ
...in/java/seedu/address/commons/util/StringUtil.java
95.45% <ø> (ø)
...rc/main/java/seedu/address/model/ModelManager.java
95.83% <100.00%> (ø)
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update c8f6c04...e504db1. Read the comment docs.
|
gharchive/pull-request
| 2021-10-09T23:16:12 |
2025-04-01T04:54:40.820955
|
{
"authors": [
"DesmondTo",
"codecov-commenter"
],
"repo": "AY2122S1-CS2103T-W10-4/tp",
"url": "https://github.com/AY2122S1-CS2103T-W10-4/tp/pull/62",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1040088535
|
[PE-D] Edit Command: not all tags cleared
Steps to reproduce: edit 1 -t
is INCOMPLETE / COMPLETE a tag?
Labels: severity.High type.FunctionalityBug
original: tsinyee/ped#9
I guess we should explicitly state that every person would have a completion status tag in UG. Documentation error
Resolved by #152
|
gharchive/issue
| 2021-10-30T08:33:29 |
2025-04-01T04:54:40.823614
|
{
"authors": [
"lhw-1",
"nus-pe-script",
"vigneshsankariyer1234567890"
],
"repo": "AY2122S1-CS2103T-W15-4/tp",
"url": "https://github.com/AY2122S1-CS2103T-W15-4/tp/issues/133",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1021937505
|
Add Ui logic for weekly panel
Move daily boxes into WeeklyPanel.fxml
Codecov Report
Merging #89 (500c938) into master (83ecbdb) will decrease coverage by 1.55%.
The diff coverage is 0.00%.
@@ Coverage Diff @@
## master #89 +/- ##
============================================
- Coverage 71.14% 69.58% -1.56%
Complexity 399 399
============================================
Files 71 73 +2
Lines 1251 1279 +28
Branches 128 128
============================================
Hits 890 890
- Misses 329 357 +28
Partials 32 32
Impacted Files
Coverage Δ
src/main/java/seedu/unify/ui/DailyPanel.java
0.00% <0.00%> (ø)
src/main/java/seedu/unify/ui/MainWindow.java
0.00% <0.00%> (ø)
src/main/java/seedu/unify/ui/TaskCard.java
0.00% <0.00%> (ø)
src/main/java/seedu/unify/ui/WeeklyPanel.java
0.00% <0.00%> (ø)
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 83ecbdb...500c938. Read the comment docs.
LGTM
|
gharchive/pull-request
| 2021-10-10T10:02:14 |
2025-04-01T04:54:40.835146
|
{
"authors": [
"KelvinSoo",
"codecov-commenter",
"lwj1711"
],
"repo": "AY2122S1-CS2103T-W17-4/tp",
"url": "https://github.com/AY2122S1-CS2103T-W17-4/tp/pull/89",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1160628528
|
Updated ReadMe CI Status link to match with our project's CI
Changed the CI Status link to ours.
Codecov Report
Merging #40 (4642437) into master (d3e617b) will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #40 +/- ##
=========================================
Coverage 72.15% 72.15%
Complexity 399 399
=========================================
Files 70 70
Lines 1232 1232
Branches 125 125
=========================================
Hits 889 889
Misses 311 311
Partials 32 32
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update d3e617b...4642437. Read the comment docs.
|
gharchive/pull-request
| 2022-03-06T14:43:21 |
2025-04-01T04:54:40.841840
|
{
"authors": [
"alfredkohhh",
"codecov-commenter"
],
"repo": "AY2122S2-CS2103-F11-1/tp",
"url": "https://github.com/AY2122S2-CS2103-F11-1/tp/pull/40",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1151328844
|
Clear data branch
Two new commands are added, which are "deletemodule" and "clearmodules".
deletemodule : Delete specified modules for someone
deletemodule 5 t/CS1231S t/CS3230
clearmodules : Clear all modules for someone
clearmodules 5
Added regex expression to ensure module codes are correct.
2-3 letters prefix followed by 4 digits and 1 optional letter.
[a-zA-Z]{2,3}\d{4}[a-zA-Z]?
Was lazy to change all tags to modules in variable names. Maybe after midterms.
LGTM. Might need to do some refactoring after this has been merged since new fields are added
Codecov Report
Merging #16 (a43fc93) into master (17dc973) will increase coverage by 0.67%.
The diff coverage is 83.90%.
@@ Coverage Diff @@
## master #16 +/- ##
============================================
+ Coverage 72.31% 72.99% +0.67%
- Complexity 437 463 +26
============================================
Files 76 80 +4
Lines 1333 1418 +85
Branches 139 148 +9
============================================
+ Hits 964 1035 +71
- Misses 331 335 +4
- Partials 38 48 +10
Impacted Files
Coverage Δ
...va/seedu/address/logic/commands/CommandResult.java
88.23% <ø> (ø)
.../seedu/address/logic/parser/AddressBookParser.java
76.19% <0.00%> (-8.03%)
:arrow_down:
...du/address/logic/commands/DeleteModuleCommand.java
77.41% <77.41%> (ø)
...ddress/logic/parser/DeleteModuleCommandParser.java
82.35% <82.35%> (ø)
...du/address/logic/commands/ClearModulesCommand.java
91.30% <91.30%> (ø)
...address/logic/parser/ClearModuleCommandParser.java
100.00% <100.00%> (ø)
src/main/java/seedu/address/model/AddressBook.java
93.33% <100.00%> (+1.66%)
:arrow_up:
src/main/java/seedu/address/model/tag/Tag.java
90.00% <100.00%> (ø)
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 17dc973...a43fc93. Read the comment docs.
|
gharchive/pull-request
| 2022-02-26T06:39:43 |
2025-04-01T04:54:40.859038
|
{
"authors": [
"bakano98",
"codecov-commenter",
"lawwm"
],
"repo": "AY2122S2-CS2103T-T13-4/tp",
"url": "https://github.com/AY2122S2-CS2103T-T13-4/tp/pull/16",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1436896023
|
matchbuyer: should match properties that are below their budget range too
and also for matchprops: should match buyers whose price range is above the property price too
Will not be fixed, instead will explain intended behavior in the UG
|
gharchive/issue
| 2022-11-05T08:09:04 |
2025-04-01T04:54:40.860505
|
{
"authors": [
"ChryslineLim",
"riccqi"
],
"repo": "AY2223S1-CS2103T-F12-1/tp",
"url": "https://github.com/AY2223S1-CS2103T-F12-1/tp/issues/240",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1427606109
|
[PE-D][Tester D] Syntax of user commands are inconsistent
The interest-related commands are in camel case where as the module-related commands are delimited with white space. E.g., addInt vs mod add.
Moving forward, perhaps the team could standardize to the same format for all commands such that it is easier for users to remember and use them.
Labels: type.FeatureFlaw severity.VeryLow
original: shwene/ped#2
@carriezhengjr We can possibly add an explanation why it is designed this way (ie. any commands relating to student information are done through add ..., findInt and addInt but we deliberately use the syntax mod <command word> to distinguish Student commands and Module commands. (Should we place this in UG or DG?)
@carriezhengjr We can possibly add an explanation why it is designed this way (ie. any commands relating to student information are done through add ..., findInt and addInt but we deliberately use the syntax mod <command word> to distinguish Student commands and Module commands. (Should we place this in UG or DG?)
Maybe we can put in UG, before all mod commands, as a Tip.
@carriezhengjr We can possibly add an explanation why it is designed this way (ie. any commands relating to student information are done through add ..., findInt and addInt but we deliberately use the syntax mod <command word> to distinguish Student commands and Module commands. (Should we place this in UG or DG?)
Maybe we can put in UG, before all mod commands, as a Tip.
Yes
|
gharchive/issue
| 2022-10-28T18:04:28 |
2025-04-01T04:54:40.865535
|
{
"authors": [
"carriezhengjr",
"nus-pe-script",
"sltsheryl"
],
"repo": "AY2223S1-CS2103T-T11-4/tp",
"url": "https://github.com/AY2223S1-CS2103T-T11-4/tp/issues/178",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1376997611
|
Add tjanenggerkevin PPP
Fixes #57
Codecov Report
Merging #58 (4a8e2e8) into master (d9c88f6) will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #58 +/- ##
=========================================
Coverage 72.15% 72.15%
Complexity 399 399
=========================================
Files 70 70
Lines 1232 1232
Branches 125 125
=========================================
Hits 889 889
Misses 311 311
Partials 32 32
:mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more
|
gharchive/pull-request
| 2022-09-18T09:31:41 |
2025-04-01T04:54:40.868749
|
{
"authors": [
"Nephelite",
"codecov-commenter"
],
"repo": "AY2223S1-CS2103T-T13-4/tp",
"url": "https://github.com/AY2223S1-CS2103T-T13-4/tp/pull/58",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1437920633
|
Ppp update shawn
Fixed UML diagrams
Codecov Report
Merging #273 (b4a0649) into master (4d9f856) will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #273 +/- ##
=========================================
Coverage 65.21% 65.21%
Complexity 604 604
=========================================
Files 93 93
Lines 2199 2199
Branches 271 271
=========================================
Hits 1434 1434
Misses 687 687
Partials 78 78
:mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more
|
gharchive/pull-request
| 2022-11-07T07:58:07 |
2025-04-01T04:54:40.871989
|
{
"authors": [
"Kok-je",
"codecov-commenter"
],
"repo": "AY2223S1-CS2103T-T15-1/tp",
"url": "https://github.com/AY2223S1-CS2103T-T15-1/tp/pull/273",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1431094411
|
Revert #206
Check on forum first
Approved by Prof Damith
|
gharchive/issue
| 2022-11-01T08:57:17 |
2025-04-01T04:54:40.872891
|
{
"authors": [
"waynezsy"
],
"repo": "AY2223S1-CS2103T-W13-3/tp",
"url": "https://github.com/AY2223S1-CS2103T-W13-3/tp/issues/294",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1393378789
|
Add team field to Person
Changes
Add team field to Person
Fixes #20
Codecov Report
Base: 72.15% // Head: 70.72% // Decreases project coverage by -1.43% :warning:
Coverage data is based on head (6569d31) compared to base (103c409).
Patch coverage: 0.00% of modified lines in pull request are covered.
Additional details and impacted files
@@ Coverage Diff @@
## master #21 +/- ##
============================================
- Coverage 72.15% 70.72% -1.44%
Complexity 399 399
============================================
Files 70 71 +1
Lines 1232 1257 +25
Branches 125 129 +4
============================================
Hits 889 889
- Misses 311 336 +25
Partials 32 32
Impacted Files
Coverage Δ
src/main/java/seedu/address/model/team/Team.java
0.00% <0.00%> (ø)
Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.
:umbrella: View full report at Codecov.
:loudspeaker: Do you have feedback about the report comment? Let us know in this issue.
Good adherence to coding standards. Looks good to merge. Would be nice to see a delete_member command and a visual effect in the UI upon adding a member to the team.
The delete_member command will be done in a separate PR. I think I won't add a visual effect to the so early, perhaps only when the Team model is implemented properly
|
gharchive/pull-request
| 2022-10-01T12:06:38 |
2025-04-01T04:54:40.881176
|
{
"authors": [
"codecov-commenter",
"marcuspang"
],
"repo": "AY2223S1-CS2103T-W13-4/tp",
"url": "https://github.com/AY2223S1-CS2103T-W13-4/tp/pull/21",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1383237632
|
As a purchasing manager, I can rename my inventory items
... so that I can update items with an incorrect name.
... so that I can update items with an incorrect name.
Closed by #151 during refactoring of Person to Item. This is already a command available in AB3.
|
gharchive/issue
| 2022-09-23T03:04:00 |
2025-04-01T04:54:40.882514
|
{
"authors": [
"Ferusel",
"yixiann"
],
"repo": "AY2223S1-CS2103T-W16-2/tp",
"url": "https://github.com/AY2223S1-CS2103T-W16-2/tp/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1976941755
|
[PE-D][Tester D] Being able to add multiple interviews at the same time slot
I was adding interviews and had not realise that i had clashes with my different types of interview for the same company. I think having an warning message to tell me i have clashing interviews would be great :)
Labels: type.FunctionalityBug severity.High
original: wasjoe1/ped#9
Can include interviews at the same time slots
Thank you for the feedback. But as discussed by the team, we think it is more logical to allow user to add interviews with the timeslot as it does not necessarily mean the user must attend the interview just by recording it
|
gharchive/issue
| 2023-11-03T21:16:00 |
2025-04-01T04:54:40.884864
|
{
"authors": [
"evanyan13",
"nus-se-script",
"peiran18"
],
"repo": "AY2324S1-CS2103T-W12-3/tp",
"url": "https://github.com/AY2324S1-CS2103T-W12-3/tp/issues/158",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1976824644
|
[PE-D][Tester B] edit-friend allows name to be replaced by numbers
UG indicates that name must be a string but the app allowed integers to replace name
Labels: severity.Low type.FunctionalityBug
original: imkwokyong/ped#5
Closing this issue, since add-friend also allows for numerical names to be used. (ie alphanumeric characters and spaces)
|
gharchive/issue
| 2023-11-03T19:40:42 |
2025-04-01T04:54:40.886706
|
{
"authors": [
"Elijah5399",
"nus-pe-script"
],
"repo": "AY2324S1-CS2103T-W16-4/tp",
"url": "https://github.com/AY2324S1-CS2103T-W16-4/tp/issues/130",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2192353329
|
Add company name feature into AddressBook
Add Company Name feature into AddressBook.
Use cn/ as parameter
Should only be up to 100 characters
resolved conflicts
LGTM :)
|
gharchive/pull-request
| 2024-03-18T14:18:27 |
2025-04-01T04:54:40.890187
|
{
"authors": [
"Lalelulilulela",
"dabzpengu"
],
"repo": "AY2324S2-CS2103T-T08-3/tp",
"url": "https://github.com/AY2324S2-CS2103T-T08-3/tp/pull/30",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2228867789
|
[PE-D][Tester A] marking feature moves position of loan in display
Initial state:
Input: mark loan 3
New state:
Expected: No change in index/position of loan in list display
Actual: the loan with amount 666 moves to the first position.
This can confuse the user, more so if the list is long
Labels: severity.Medium type.FunctionalityBug
original: Murugan-Maniish/ped#14
Stability in sorting by end date
Fix by sort by creation? Not exactly a bug anyways
|
gharchive/issue
| 2024-04-05T22:04:32 |
2025-04-01T04:54:40.893085
|
{
"authors": [
"Joseph31416",
"nus-pe-script"
],
"repo": "AY2324S2-CS2103T-W13-1/tp",
"url": "https://github.com/AY2324S2-CS2103T-W13-1/tp/issues/182",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2644978061
|
[PE-D][Tester D] Edit command with incorrect index returns inconsistent error message depending on whether the index is too high or too low
As seen above, when edit command is used with index 0(the list is one-indexed), I get an invalid command format message, whereas when an index such as 6 that is too high(the list has 5 people) is used, I get an invalid person index instead. This could slow down a user's ability to understand what is wrong if they accidentally key in index 0.
Labels: severity.Low type.FunctionalityBug
original: naythee169/ped#5
Can implement this fix
The error are not wrong so we wont be changing this.
|
gharchive/issue
| 2024-11-08T19:30:38 |
2025-04-01T04:54:40.895921
|
{
"authors": [
"DesSnowy",
"Incogdino",
"nus-se-script"
],
"repo": "AY2425S1-CS2103T-T11-2/tp",
"url": "https://github.com/AY2425S1-CS2103T-T11-2/tp/issues/307",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2627371328
|
Update view-wed command in UserGuide.md
Fixes #160
resolved merge conflict for this
|
gharchive/pull-request
| 2024-10-31T16:59:01 |
2025-04-01T04:54:40.896976
|
{
"authors": [
"jonushzw",
"siriousguy"
],
"repo": "AY2425S1-CS2103T-W13-4/tp",
"url": "https://github.com/AY2425S1-CS2103T-W13-4/tp/pull/161",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1146801105
|
Update PAT when it expires
I want to be able to replace the PAT for a specific organization, for e.g. when I'm regenerating the PAT or when it expires.
My suggestion is to allow Connect-AZDOPS to update an existing connection. Either by default or using something like a -Force switch.
|
gharchive/issue
| 2022-02-22T11:38:02 |
2025-04-01T04:54:40.897946
|
{
"authors": [
"PalmEmanuel",
"egullbrandsson"
],
"repo": "AZDOPS/AZDevOPS",
"url": "https://github.com/AZDOPS/AZDevOPS/issues/21",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1510051919
|
Documentation update
I think instead of saying it like in the above screenshot we could say it like this:
"It works by fetching the user's recent tweets from the Twitter API"
so that it's easy to understand
If approved I would love to make this change.
If approved I would love to make this change.
Yes sir go ahead
If approved I would love to make this change.
Yes sir go ahead
made a pull request
|
gharchive/issue
| 2022-12-24T11:08:40 |
2025-04-01T04:54:40.900461
|
{
"authors": [
"Aadarsh805",
"Ninad-Patil"
],
"repo": "Aadarsh805/TweetSage",
"url": "https://github.com/Aadarsh805/TweetSage/issues/83",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1013228542
|
Fix privacy policy link in error reporting window
The window says at the bottom "click here to view privacy policy" but the link doesn't actually go anywhere
Hi @Taikelenn I would like to take this up
|
gharchive/issue
| 2021-10-01T11:22:30 |
2025-04-01T04:54:40.901555
|
{
"authors": [
"Taikelenn",
"tulikavijay"
],
"repo": "Aalto-LeTech/aplus-courses",
"url": "https://github.com/Aalto-LeTech/aplus-courses/issues/767",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
413460500
|
Unsupported image type, must be 8bit gray or RGB image.
/home/tango/usr/local/torch/install/bin/luajit: ./facedetection_dlib.lua:21: Python error: opaque ref: call
RuntimeError: Unsupported image type, must be 8bit gray or RGB image.
either open the image in an image editor and save it as an RGB image, or
modify the code to accommodate such images.
Data_Science writes:
/home/tango/usr/local/torch/install/bin/luajit: ./facedetection_dlib.lua:21: Python error: opaque ref: call
RuntimeError: Unsupported image type, must be 8bit gray or RGB image.
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub, or mute the thread.
--
Aaron Jackson - M6PIU
http://aaronsplace.co.uk/
But this issue rises with the images provided by you...
Could you please help me
This issue comes with every other images. Not able to solve
You might need to upgrade Pillow
|
gharchive/issue
| 2019-02-22T15:43:56 |
2025-04-01T04:54:40.914307
|
{
"authors": [
"AaronJackson",
"NarenBabuR"
],
"repo": "AaronJackson/vrn",
"url": "https://github.com/AaronJackson/vrn/issues/119",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
154399235
|
NoSuchFieldError with OptiNotFine
java.lang.NoSuchFieldError: field_78286_d
at stellarapi.lib.gui.basicmodel.WrappedFontRenderer.func_78266_a(WrappedFontRenderer.java:39)
at net.minecraft.client.gui.FontRenderer.func_78278_a(FontRenderer.java:264)
at net.minecraft.client.gui.FontRenderer.func_78255_a(FontRenderer.java:523)
at net.minecraft.client.gui.FontRenderer.func_78258_a(FontRenderer.java:636)
at net.minecraft.client.gui.FontRenderer.func_85187_a(FontRenderer.java:390)
at net.minecraft.client.gui.FontRenderer.func_78276_b(FontRenderer.java:369)
at stellarapi.lib.gui.basicmodel.ModelFont.renderModel(ModelFont.java:65)
at stellarapi.lib.gui.GuiRenderer.render(GuiRenderer.java:123)
at stellarapi.lib.gui.text.GuiTextInternal.render(GuiTextInternal.java:269)
at stellarapi.lib.gui.text.GuiTextField.render(GuiTextField.java:63)
at stellarapi.lib.gui.list.GuiHasFixedList.render(GuiHasFixedList.java:134)
at stellarapi.lib.gui.simple.GuiSimpleSpacingElement.render(GuiSimpleSpacingElement.java:64)
at stellarapi.lib.gui.GuiContent.render(GuiContent.java:55)
at stellarium.client.overlay.clock.OverlayClock.render(OverlayClock.java:222)
at stellarapi.feature.gui.overlay.OverlayContainer.render(OverlayContainer.java:123)
at stellarapi.feature.gui.overlay.OverlayHandler.renderGameOverlay(OverlayHandler.java:40)
at stellarapi.StellarAPIClientForgeEventHook.renderGameOverlay(StellarAPIClientForgeEventHook.java:109)
at cpw.mods.fml.common.eventhandler.ASMEventHandler_29_StellarAPIClientForgeEventHook_renderGameOverlay_Post.invoke(.dynamic)
at cpw.mods.fml.common.eventhandler.ASMEventHandler.invoke(ASMEventHandler.java:54)
at cpw.mods.fml.common.eventhandler.EventBus.post(EventBus.java:140)
at net.minecraftforge.client.GuiIngameForge.post(GuiIngameForge.java:899)
at net.minecraftforge.client.GuiIngameForge.func_73830_a(GuiIngameForge.java:176)
at net.minecraft.client.renderer.EntityRenderer.func_78480_b(EntityRenderer.java:1362)
at net.minecraft.client.Minecraft.func_71411_J(Minecraft.java:1001)
at net.minecraft.client.Minecraft.func_99999_d(Minecraft.java:898)
at net.minecraft.client.main.Main.main(SourceFile:148)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at net.minecraft.launchwrapper.Launch.launch(Launch.java:135)
at net.minecraft.launchwrapper.Launch.main(Launch.java:28)
A detailed walkthrough of the error, its code path and all known details is as follows:
---------------------------------------------------------------------------------------
-- Head --
Stacktrace:
at stellarapi.lib.gui.basicmodel.WrappedFontRenderer.func_78266_a(WrappedFontRenderer.java:39)
at net.minecraft.client.gui.FontRenderer.func_78278_a(FontRenderer.java:264)
at net.minecraft.client.gui.FontRenderer.func_78255_a(FontRenderer.java:523)
at net.minecraft.client.gui.FontRenderer.func_78258_a(FontRenderer.java:636)
at net.minecraft.client.gui.FontRenderer.func_85187_a(FontRenderer.java:390)
at net.minecraft.client.gui.FontRenderer.func_78276_b(FontRenderer.java:369)
at stellarapi.lib.gui.basicmodel.ModelFont.renderModel(ModelFont.java:65)
at stellarapi.lib.gui.GuiRenderer.render(GuiRenderer.java:123)
at stellarapi.lib.gui.text.GuiTextInternal.render(GuiTextInternal.java:269)
at stellarapi.lib.gui.text.GuiTextField.render(GuiTextField.java:63)
at stellarapi.lib.gui.list.GuiHasFixedList.render(GuiHasFixedList.java:134)
at stellarapi.lib.gui.simple.GuiSimpleSpacingElement.render(GuiSimpleSpacingElement.java:64)
at stellarapi.lib.gui.GuiContent.render(GuiContent.java:55)
at stellarium.client.overlay.clock.OverlayClock.render(OverlayClock.java:222)
at stellarapi.feature.gui.overlay.OverlayContainer.render(OverlayContainer.java:123)
at stellarapi.feature.gui.overlay.OverlayHandler.renderGameOverlay(OverlayHandler.java:40)
at stellarapi.StellarAPIClientForgeEventHook.renderGameOverlay(StellarAPIClientForgeEventHook.java:109)
at cpw.mods.fml.common.eventhandler.ASMEventHandler_29_StellarAPIClientForgeEventHook_renderGameOverlay_Post.invoke(.dynamic)
at cpw.mods.fml.common.eventhandler.ASMEventHandler.invoke(ASMEventHandler.java:54)
at cpw.mods.fml.common.eventhandler.EventBus.post(EventBus.java:140)
at net.minecraftforge.client.GuiIngameForge.post(GuiIngameForge.java:899)
at net.minecraftforge.client.GuiIngameForge.func_73830_a(GuiIngameForge.java:176)
Installed last versions of StellarAPI and StellarSky for 1.7.10. Of other mods only OptiFIne Ultra D4.
Forge 1614.
Thanks, but it cause other error... Maybe i'm incorrectly compiled source code, but i have crash on start:
cpw.mods.fml.common.LoaderException: java.lang.AbstractMethodError
at cpw.mods.fml.common.LoadController.transition(LoadController.java:163)
at cpw.mods.fml.common.Loader.preinitializeMods(Loader.java:559)
at cpw.mods.fml.client.FMLClientHandler.beginMinecraftLoading(FMLClientHandler.java:243)
at net.minecraft.client.Minecraft.func_71384_a(Minecraft.java:480)
at net.minecraft.client.Minecraft.func_99999_d(Minecraft.java:878)
at net.minecraft.client.main.Main.main(SourceFile:148)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at net.minecraft.launchwrapper.Launch.launch(Launch.java:135)
at net.minecraft.launchwrapper.Launch.main(Launch.java:28)
Caused by: java.lang.AbstractMethodError
at stellarapi.api.gui.overlay.OverlayRegistry$RegistryDelegate.<init>(OverlayRegistry.java:65)
at stellarapi.api.gui.overlay.OverlayRegistry$RegistryDelegate.<init>(OverlayRegistry.java:51)
at stellarapi.api.gui.overlay.OverlayRegistry.registerOverlay(OverlayRegistry.java:36)
at stellarium.ClientProxy.preInit(ClientProxy.java:54)
at stellarium.StellarSky.preInit(StellarSky.java:63)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at cpw.mods.fml.common.FMLModContainer.handleModStateEvent(FMLModContainer.java:532)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at com.google.common.eventbus.EventSubscriber.handleEvent(EventSubscriber.java:74)
at com.google.common.eventbus.SynchronizedEventSubscriber.handleEvent(SynchronizedEventSubscriber.java:47)
at com.google.common.eventbus.EventBus.dispatch(EventBus.java:322)
at com.google.common.eventbus.EventBus.dispatchQueuedEvents(EventBus.java:304)
at com.google.common.eventbus.EventBus.post(EventBus.java:275)
at cpw.mods.fml.common.LoadController.sendEventToModContainer(LoadController.java:212)
at cpw.mods.fml.common.LoadController.propogateStateMessage(LoadController.java:190)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at com.google.common.eventbus.EventSubscriber.handleEvent(EventSubscriber.java:74)
at com.google.common.eventbus.SynchronizedEventSubscriber.handleEvent(SynchronizedEventSubscriber.java:47)
at com.google.common.eventbus.EventBus.dispatch(EventBus.java:322)
at com.google.common.eventbus.EventBus.dispatchQueuedEvents(EventBus.java:304)
at com.google.common.eventbus.EventBus.post(EventBus.java:275)
at cpw.mods.fml.common.LoadController.distributeStateMessage(LoadController.java:119)
at cpw.mods.fml.common.Loader.preinitializeMods(Loader.java:556)
... 10 more
I think something had been corrupted in the workspace, since that is something that can't happen.
Please wait for the release on Curse.
|
gharchive/issue
| 2016-05-12T05:21:02 |
2025-04-01T04:54:40.919583
|
{
"authors": [
"Abastro",
"Ghostlyr"
],
"repo": "Abastro/StellarAPI",
"url": "https://github.com/Abastro/StellarAPI/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2009402824
|
Margin Combineren in een regel
oplossing
img h1 h2 h3 p {
margin: 10px;
}
Opgelost!
|
gharchive/issue
| 2023-11-24T09:36:11 |
2025-04-01T04:54:40.939755
|
{
"authors": [
"Abeeryu"
],
"repo": "Abeeryu/look-and-feel-custom-properties",
"url": "https://github.com/Abeeryu/look-and-feel-custom-properties/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1551120926
|
Use unsafe_writes when writing PIMT logfile
Use unsafe_writes when writing PIMT logfile
Version 0.3.2
@ablbot rebase
Hmm, in spite of this fix, the error still occurs. :confused:
|
gharchive/pull-request
| 2023-01-20T16:16:58 |
2025-04-01T04:54:40.942884
|
{
"authors": [
"nre-ableton"
],
"repo": "Ableton/ansible-role-jenkins-jcasc",
"url": "https://github.com/Ableton/ansible-role-jenkins-jcasc/pull/267",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1857045722
|
Graph Editor: Implicit cast from "vec3" to "vec2" - using texcoord both as vector3 and vector2
This seems to be related to texcoord 0 being used for the normalmap as well as the basecolor - I'm still investigating this and will try to reproduce it with a fresh material but for the sake of documentation I'll open an issue here now 😅
<?xml version="1.0"?>
<materialx name="" version="1.38" >
<nodegraph name="needle_shadergraph" >
<!-- MainTex -->
<input name="_MainTex" type="filename" value="Tile_Albedo.jpg" />
<dot name="dot__MainTex" type="filename" >
<input name="in" type="filename" interfacename="_MainTex" />
</dot>
<texcoord name="uv" type="vector3" >
<input name="index" type="integer" value="0" />
</texcoord>
<convert name="convert_to_vector2" type="vector2" >
<input name="in" type="vector3" nodename="uv" />
</convert>
<image name="tex" type="color4" >
<input name="file" type="filename" nodename="dot__MainTex" />
<input name="texcoord" type="vector2" nodename="convert_to_vector2" />
</image>
<convert name="convert_to_color3" type="color3" >
<input name="in" type="color4" nodename="tex" />
</convert>
<output name="out" type="color3" nodename="convert_to_color3" />
<!-- Normal -->
<input name="_Normal" type="filename" value="Tile_Normal.jpg" />
<dot name="dot__Normal" type="filename" >
<input name="in" type="filename" interfacename="_Normal" />
</dot>
<texcoord name="uv_2" type="vector2" >
<input name="index" type="integer" value="0" />
</texcoord>
<image name="tex_1" type="color4" >
<input name="file" type="filename" nodename="dot__Normal" />
<input name="texcoord" type="vector2" nodename="uv_2" />
</image>
<swizzle name="swizzle_0_tex_1" type="vector3" >
<input name="in" type="color4" nodename="tex_1" />
<input name="channels" type="string" value="rgb" />
</swizzle>
<output name="out_1" type="vector3" nodename="normalmap" />
<normalmap name="normalmap" type="vector3" >
<input name="in" type="vector3" nodename="swizzle_0_tex_1" />
</normalmap>
<constant name="specular_roughness" type="float" >
<input name="value" type="float" value="0.0" />
</constant>
<output name="out_2" type="float" nodename="invert" />
<invert name="invert" type="float" >
<input name="in" type="float" nodename="specular_roughness" />
</invert>
</nodegraph>
<standard_surface name="needle_standard_surface" type="surfaceshader" >
<!-- BaseColor -->
<input name="base_color" type="color3" nodegraph="needle_shadergraph" output="out" />
<!-- NormalTS -->
<input name="normal" type="vector3" nodegraph="needle_shadergraph" output="out_1" />
<!-- Smoothness -->
<input name="specular_roughness" type="float" nodegraph="needle_shadergraph" output="out_2" />
</standard_surface>
<surfacematerial name="Default" type="material" >
<input name="surfaceshader" type="surfaceshader" nodename="needle_standard_surface" />
</surfacematerial>
</materialx>
Textures
The material compiles when I remove one of the texcoord usages (both use channel 0)
Here is a minimal material that reproduces the issue above
<?xml version="1.0"?>
<materialx version="1.38" colorspace="lin_rec709">
<standard_surface name="SR_marble1" type="surfaceshader" xpos="6.159420" ypos="-0.568965">
<input name="base" type="float" value="1" />
<input name="base_color" type="color3" nodename="image_color3" />
<input name="specular_roughness" type="float" value="0.1" />
<input name="subsurface" type="float" value="0.4" />
<input name="subsurface_color" type="color3" nodename="image_color4" />
</standard_surface>
<surfacematerial name="Marble_3D" type="material" xpos="8.695652" ypos="0.000000">
<input name="surfaceshader" type="surfaceshader" nodename="SR_marble1" />
</surfacematerial>
<image name="image_color3" type="color3" xpos="3.528986" ypos="-1.241379">
<input name="file" type="filename" value="Tile_Albedo.jpg" />
<input name="texcoord" type="vector2" nodename="texcoord_vector2" />
</image>
<texcoord name="texcoord_vector2" type="vector2" xpos="1.666667" ypos="-1.413793" />
<texcoord name="texcoord_vector3" type="vector3" xpos="0.275362" ypos="1.077586" />
<image name="image_color4" type="color3" xpos="3.485507" ypos="1.905172">
<input name="file" type="filename" value="Tile_Albedo.jpg" />
<input name="texcoord" type="vector2" nodename="swizzle_vector3_vector2" />
</image>
<swizzle name="swizzle_vector3_vector2" type="vector2" xpos="1.637681" ypos="1.767241">
<input name="in" type="vector3" nodename="texcoord_vector3" />
<input name="channels" type="string" value="xy" />
</swizzle>
</materialx>
This is the generated GLSL code snippet:
void main()
{
in VertexData
{
vec3 normalWorld;
vec3 tangentWorld;
vec2 texcoord_0;
vec3 positionWorld;
} vd;
vec3 geomprop_Nworld_out1 = normalize(vd.normalWorld);
vec3 geomprop_Tworld_out1 = normalize(vd.tangentWorld);
vec2 texcoord_vector2_out = vd.texcoord_0;
vec3 texcoord_vector3_out = vd.texcoord_0;
vec3 image_color3_out = vec3(0.0);
Seem we are declaring one stream input (vec2) and then skipping adding in the second (vec3) and then reusing the same stream route. Hence there is a vec2 to vec3 cast.
Adding @niklasharrysson, for thoughts on this.
I'm guessing that a type check is not being performed when determingin whether to create a geomstry stream ShaderNode?
BTW @marwie, The MaterialX Viewer has more diagnostic functionality as it allows you to do things like dump out the GLSL code (G key) if case you find this useful.
As suggested offline to avoid this declare unique stream inputs which are type vec3 as a workaround for now.
Hi @madmann91,
Thanks a lot for taking a look at this!
I think the choices are:
As you suggest, always keep vec3 and add code to extract a vec2 from vec3 if the lookup is vec2. (Basically you want to avoid explicit casts again). I like the idea to widen only when necessary.
Keep both a vec2 and a vec3. This means no code generator changes are required but there needs to be a way to avoid 2 streams having the same name. The published naming convention indicates a renderer can bind to a stream with name of the form: <stream_type>_<stream_number>. All I can think of is to add an additional qualifier to get something like texcoord_0:2, texcoord_0:3. Binding code needs to be updated.
I'm leaning towards 1 since it does not affect integrators.
In this case vec3->vec2 conversion can be done by:
1a. inserting additional shader code on lookups,
1b. replace/insert a vec3-to-vec2 conversion ShaderNode instead of a texcoord ShaderNode . ShaderNodes only exist for code generation and generators do insert additional nodes as necessary including geometry nodes.
Option 1b allows you to not worry about each shading language's syntax and I think it should be pretty robust.
There may be a 1c. but this is all that comes to mind currently :)
If this is okay to work on, it would great if you look at this for dev days. Of course there will be folks around to help out in this area :)
|
gharchive/issue
| 2023-08-18T17:22:07 |
2025-04-01T04:54:40.963450
|
{
"authors": [
"kwokcb",
"marwie"
],
"repo": "AcademySoftwareFoundation/MaterialX",
"url": "https://github.com/AcademySoftwareFoundation/MaterialX/issues/1464",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1266789019
|
variable R21 in mx_fresnel_airy (glsl) is unused. bug?
There is a variable called R21 in mx_fresnel_airy() (libraries/pbrlib/genglsl/lib/mx_microfacet_specular.glsl, line 360) that is unused.
Is this a bug?
Or is it safe to simply comment it out?
cheers!
Great catch, @BrianSharpe, and that same unused variable seems to be present in Laurent Belcour's original code. The link has changed since @niklasharrysson developed the GLSL version for MaterialX, and can now be found here:
https://belcour.github.io/blog/research/publication/2017/05/01/brdf-thin-film.html
@BrianSharpe If you have the bandwidth, feel free to remove this line from the GLSL code, and we should likely update the link to Laurent Belcour's paper to the new location above.
done (https://github.com/AcademySoftwareFoundation/MaterialX/pull/989)
cheers!
|
gharchive/issue
| 2022-06-10T00:05:12 |
2025-04-01T04:54:40.966779
|
{
"authors": [
"BrianSharpe",
"jstone-lucasfilm"
],
"repo": "AcademySoftwareFoundation/MaterialX",
"url": "https://github.com/AcademySoftwareFoundation/MaterialX/issues/986",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
627952713
|
Build error with python 3.7
FreeBSD 12.1 amd64.
With DOCS option OFF build fine with python 3.7 and python 2.7.
But with DOCS option ON build fine with python 2.7 only. With 3.7 I got error and search didn't help.
Build log:
-- The C compiler identification is Clang 8.0.1
-- The CXX compiler identification is Clang 8.0.1
-- Check for working C compiler: /usr/bin/cc
-- Check for working C compiler: /usr/bin/cc - works
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Detecting C compile features
-- Detecting C compile features - done
-- Check for working CXX compiler: /usr/bin/c++
-- Check for working CXX compiler: /usr/bin/c++ - works
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Setting Build Type to: Release
-- Setting Namespace to: OpenColorIO
-- Exec prefix not specified, defaulting to /usr/local
-- Use Boost Ptr: OFF
-- Setting python bin to: /usr/local/bin/python3.7
Python library: PYTHON_LIBRARY-NOTFOUND
-- Setting EXTDIST_BINPATH: /tmp/work/usr/ports/graphics/opencolorio/work/.build/ext/dist/bin
-- Setting EXTDIST_PYTHONPATH: /tmp/work/usr/ports/graphics/opencolorio/work/.build/ext/dist/lib/python3.7/site-packages
-- Found TinyXML: /usr/local/lib/libtinyxml.so
-- TinyXML version: 2.6.2
-- External TinyXML will be used.
-- Found PkgConfig: pkgconf (found version "1.6.3")
-- Found yaml-cpp: /usr/local/lib/libyaml-cpp03.so
-- Generate Documentation: true
-- Create sphinx conf.py from conf.py.in
-- Copying doc to staging area
-- Copy extra doc files to staging area
-- Extracting .rst files from C++ headers
-- SSE Optimizations: ON
-- Could NOT find Truelight (missing: Truelight_INCLUDE_DIR Truelight_LIBRARIES Truelight_LIBRARY_DIR)
-- Not building truelight transform support. Add the flag -D TRUELIGHT_INSTALL_PATH=... or set the TRUELIGHT_ROOT environment variable
-- Create OpenColorABI.h from OpenColorABI.h.in
-- Setting OCIO SOVERSION to: 1
-- Create OpenColorIO.pc from OpenColorIO.pc.in
-- Build Unit Tests: ON
-- Create ocio_core_tests.sh.in from ocio_core_tests.sh
-- Disable build of apps. See cmake options : OCIO_BUILD_APPS and OCIO_BUILD_SHARED/OCIO_BUILD_STATIC (requiered)
-- Python library to include 'lib' prefix: OFF
-- Python 3.7 okay (UCS: ucs4), will build the Python bindings against /usr/local/include/python3.7m
-- Python variant path is lib/python3.7/site-packages
-- Found PythonLibs: /usr/local/lib/libpython3.7m.so (found version "3.7.7")
PYTHON_VARIANT_PATH: lib/python3.7/site-packages
-- Configuring done
-- Generating done
CMake Warning:
Manually-specified variables were not used by the project:
BOOST_PYTHON_SUFFIX
CMAKE_CXX_FLAGS_DEBUG
CMAKE_C_FLAGS_DEBUG
CMAKE_C_FLAGS_RELEASE
CMAKE_VERBOSE_MAKEFILE
THREADS_HAVE_PTHREAD_ARG
-- Build files have been written to: /tmp/work/usr/ports/graphics/opencolorio/work/.build
===> Building for opencolorio-1.1.1_1
[1/200] cd /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs && /usr/local/bin/cmake -E make_directory /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools && /usr/local/bin/cmake -E make_directory /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools && /usr/local/bin/cmake -E make_directory /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix && /usr/local/bin/cmake -E make_directory /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/tmp && /usr/local/bin/cmake -E make_directory /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools-stamp && /usr/local/bin/cmake -E make_directory /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src && /usr/local/bin/cmake -E make_directory /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools-stamp && /usr/local/bin/cmake -E touch /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools-stamp/setuptools-mkdir
[2/200] cd /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src && /usr/local/bin/cmake -P /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools-stamp/verify-setuptools.cmake && /usr/local/bin/cmake -P /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools-stamp/extract-setuptools.cmake && /usr/local/bin/cmake -E touch /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools-stamp/setuptools-download
CMake Warning at setuptools-stamp/verify-setuptools.cmake:15 (message):
File will not be verified since no URL_HASH specified
-- extracting...
src='/tmp/work/usr/ports/graphics/opencolorio/work/OpenColorIO-1.1.1/ext/setuptools-1.1.6.tar.gz'
dst='/tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools'
-- extracting... [tar xfz]
-- extracting... [analysis]
-- extracting... [rename]
-- extracting... [clean up]
-- extracting... done
[3/200] cd /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs && /usr/local/bin/cmake -E echo_append && /usr/local/bin/cmake -E touch /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools-stamp/setuptools-update
[4/200] cd /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs && /usr/local/bin/cmake -E echo_append && /usr/local/bin/cmake -E touch /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools-stamp/setuptools-patch
[5/200] cd /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools && /usr/local/bin/cmake -E make_directory /tmp/work/usr/ports/graphics/opencolorio/work/.build/ext/dist/lib/python3.7/site-packages && /usr/local/bin/cmake -E touch /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools-stamp/setuptools-configure
[6/200] cd /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools && PYTHONPATH=/tmp/work/usr/ports/graphics/opencolorio/work/.build/ext/dist/lib/python3.7/site-packages: /usr/local/bin/python3.7 setup.py build && /usr/local/bin/cmake -E touch /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools-stamp/setuptools-build
FAILED: docs/setuptools-prefix/src/setuptools-stamp/setuptools-build
cd /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools && PYTHONPATH=/tmp/work/usr/ports/graphics/opencolorio/work/.build/ext/dist/lib/python3.7/site-packages: /usr/local/bin/python3.7 setup.py build && /usr/local/bin/cmake -E touch /tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools-stamp/setuptools-build
Traceback (most recent call last):
File "setup.py", line 17, in <module>
exec(init_file.read(), command_ns)
File "<string>", line 8, in <module>
File "/tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools/setuptools/__init__.py", line 11, in <module>
from setuptools.extension import Extension
File "/tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools/setuptools/extension.py", line 5, in <module>
from setuptools.dist import _get_unpatched
File "/tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools/setuptools/dist.py", line 16, in <module>
import pkg_resources
File "/tmp/work/usr/ports/graphics/opencolorio/work/.build/docs/setuptools-prefix/src/setuptools/pkg_resources.py", line 1435, in <module>
register_loader_type(importlib_bootstrap.SourceFileLoader, DefaultProvider)
AttributeError: module 'importlib._bootstrap' has no attribute 'SourceFileLoader'
[7/200] /usr/bin/c++ -DOpenColorIO_EXPORTS -DUSE_SSE -I/tmp/work/usr/ports/graphics/opencolorio/work/OpenColorIO-1.1.1/export -Iexport -I/tmp/work/usr/ports/graphics/opencolorio/work/OpenColorIO-1.1.1/ext/oiio/src/include -Iext/dist/include -O2 -pipe -march=core2 -fstack-protector-strong -isystem /usr/local/include -fno-strict-aliasing -Wno-deprecated-register -isystem /usr/local/include -msse2 -O2 -pipe -march=core2 -fstack-protector-strong -isystem /usr/local/include -fno-strict-aliasing -Wno-deprecated-register -isystem /usr/local/include -fPIC -DOLDYAML -fPIC -fvisibility=hidden -Werror -std=c++11 -MD -MT src/core/CMakeFiles/OpenColorIO.dir/AllocationOp.cpp.o -MF src/core/CMakeFiles/OpenColorIO.dir/AllocationOp.cpp.o.d -o src/core/CMakeFiles/OpenColorIO.dir/AllocationOp.cpp.o -c /tmp/work/usr/ports/graphics/opencolorio/work/OpenColorIO-1.1.1/src/core/AllocationOp.cpp
ninja: build stopped: subcommand failed.
Hi @VVD
OCIO 1.x does not support Python 3. But the coming OCIO v2 supports both Python versions.
That's a known limitation of OCIO v1 so I close the defect.
Please feel free to reopen it if needed.
But it build and work with python 3.7 without DOCS…
When v2 will be released?
OCIOv2 feature complete is planned for this summer.
Note: As any open source project we are open to contributions from the community. So, you can definitively submit a pull request to fix the Python 3 support in OCIOv1 if that's a blocker for you. Refer to CONTRIBUTING.md for details.
Thanks. I'm not python developer and I'll wait v2.
Is master support python 3 already?
All FreeBSD ports with mandatory dependency from python 2 will be removed on 2020-12-31 if not fixed:
https://bugs.freebsd.org/bugzilla/show_bug.cgi?id=249695
There are issue already: https://bugs.freebsd.org/bugzilla/show_bug.cgi?id=249534
|
gharchive/issue
| 2020-05-31T12:23:29 |
2025-04-01T04:54:40.974205
|
{
"authors": [
"VVD",
"hodoulp"
],
"repo": "AcademySoftwareFoundation/OpenColorIO",
"url": "https://github.com/AcademySoftwareFoundation/OpenColorIO/issues/1023",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2279307461
|
build: llvm 18 compat issue - include libclangAPINotes
Fixes #1809
@etsach are you able to try this change on your slide and let us know if it works?
It seems safe, and doesn't break anything. But then again, I could never get it to fail. Maybe because you were using llvm 18.0 and I was using 18.1 and maybe they fixed something on their end in between?
Hi,
This fixes the link error, but then I have another serie of link errors:
undefined reference to clang::SourceMgrAdapter::~SourceMgrAdapter()' undefined reference to clang::SourceMgrAdapter::handleDiag(llvm::SMDiagnostic const&, void*)'
etc...
And I can't find a library that fix them.
Definitely something else causing an issue.
Le dim. 5 mai 2024 à 05:48, Larry Gritz @.***> a écrit :
@etsach https://github.com/etsach are you able to try this change on
your slide and let us know if it works?
It seems safe, and doesn't break anything. But then again, I could never
get it to fail. Maybe because you were using llvm 18.0 and I was using 18.1
and maybe they fixed something on their end in between?
—
Reply to this email directly, view it on GitHub
https://github.com/AcademySoftwareFoundation/OpenShadingLanguage/pull/1812#issuecomment-2094570297,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/ACXIPYMDXBVO4JHQY2VSWZTZAWTZHAVCNFSM6AAAAABHHLUVYOVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDAOJUGU3TAMRZG4
.
You are receiving this because you were mentioned.Message ID:
<AcademySoftwareFoundation/OpenShadingLanguage/pull/1812/c2094570297@
github.com>
I forgot, this is with LLVM 18.1.3
Le lun. 6 mai 2024 à 22:24, Etienne Sandré-Chardonnal <
@.***> a écrit :
Hi,
This fixes the link error, but then I have another serie of link errors:
undefined reference to clang::SourceMgrAdapter::~SourceMgrAdapter()' undefined reference to clang::SourceMgrAdapter::handleDiag(llvm::SMDiagnostic const&, void*)'
etc...
And I can't find a library that fix them.
Definitely something else causing an issue.
Le dim. 5 mai 2024 à 05:48, Larry Gritz @.***> a
écrit :
@etsach https://github.com/etsach are you able to try this change on
your slide and let us know if it works?
It seems safe, and doesn't break anything. But then again, I could never
get it to fail. Maybe because you were using llvm 18.0 and I was using 18.1
and maybe they fixed something on their end in between?
—
Reply to this email directly, view it on GitHub
https://github.com/AcademySoftwareFoundation/OpenShadingLanguage/pull/1812#issuecomment-2094570297,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/ACXIPYMDXBVO4JHQY2VSWZTZAWTZHAVCNFSM6AAAAABHHLUVYOVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDAOJUGU3TAMRZG4
.
You are receiving this because you were mentioned.Message ID:
<AcademySoftwareFoundation/OpenShadingLanguage/pull/1812/c2094570297@
github.com>
@etasch, any comments? Should I merge what I have here and then we can try to find any remaining issues? Or are you worried that these changes are not correct?
This has languished for a while without a reply, but I think it's safe. Merging.
|
gharchive/pull-request
| 2024-05-05T03:46:16 |
2025-04-01T04:54:40.986173
|
{
"authors": [
"etsach",
"lgritz"
],
"repo": "AcademySoftwareFoundation/OpenShadingLanguage",
"url": "https://github.com/AcademySoftwareFoundation/OpenShadingLanguage/pull/1812",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
929625405
|
Clean up library VERSION and SOVERSION
Reduce confusion between "VERSION", "REVISION", and "SOVERSION":
Label the internal variable IMATH_LIBTOOL_* to indicate their purpose.
Use terminology closer to the libtool description
Add comment documenting the library update process
Signed-off-by: Cary Phillips cary@ilm.com
I'm bringing this up for the exact reason you cite, I strongly suspect we're not updating these numbers properly, because it's so confusing. I'm certainly open to other ideas or solutions.
I think you want the patch to be reflected in the age, not current, because it's ok to substitute a new patch release, right?
I believe that's true, I'm just wondering if that's a real use case? If someone is releasing a new ubuntu, it's unlikely they're going to cleverly avoid building openexr. If it's you or I building for our own not /usr/local directory, why are we doing that? If it's to rev an app like Blender, recompiling exr is 1% of 1% of the build time. So what I'm questioning is whether age is a pedantic correctness thing as opposed to a practical correctness. If it's merely pedantic, I'm all for all sails to the wind, let's do the thing that requires the least brain power, and only bump one number, because right now I'm feeling like maybe it's only an issue for PedanticKitty :) ... I am looking to be educated here, I feel like we've been coddling this problem along for almost plural decades now and still confused about it.
I don't disagree, as far as my personal workflow and priorities. I don't know if anybody else really cares.
I will note that if we wrap the whole thing up in current, we can never make a substitution that breaks (that's the single most important constraint), and although it doesn't allow for some possibly correct substitutions, it's no worse than the situation people would be in if they were linking statically.
Raising this again, as we need to resolve it before the 3.1 release.
In spite of how much I'd like to (a) not think about this again, and (b) not make mistakes, I'm not sure the automatic formula conforms to the libtool instructions: https://www.gnu.org/software/libtool/manual/html_node/Updating-version-info.html, which also say: "Never try to set the interface numbers so that they correspond to the release number of your package. This is an abuse that only fosters misunderstanding of the purpose of library versions." Since the project has traditionally followed this library versioning policy, I'd prefer to stick with it, and just be deliberate.
The 3.0.5 release has library version 29.0.0. The 3.1 release make a few minor internal fixes (most have been patched into 3.0.5) and adds the OpenEXRCore library. So:
Start with version information of ‘0:0:0’ for each libtool library.
Update the version information only immediately before a public release of your software.
If the library source code has changed at all since the last update, then increment revision (‘c:r:a’ becomes ‘c:r+1:a’).
yes -> 29.1.0
If any interfaces have been added, removed, or changed since the last update, increment current, and set revision to 0.
yes -> 30.0.0
If any interfaces have been added since the last public release, then increment age.
yes -> 30.0.1
If any interfaces have been removed or changed since the last public release, then set age to 0.
no -> 30.0.1
Sound right?
Since we have an entirely different library name for every minor release (libOpenEXR-3.1.so vs libOpenEXR-3.2.so), do we start at 0:0:0 again for every one of those? Or does 3.2 pick up at where 3.1 left off?
I'm not sure your examples are correct. I don't think you can ever have a x.0.1 because age says how far BACK in revision is considered compatible, and there is no revision prior to 0.
This is so hard to get right.
We also install libOpenEXR.so in the chain of symlinks to libOpenEXR-3_0.29.0.0, and I think that's where the "drop-in" policy figures in, so I don't think the -3_1 suffix resets things.
And as pointed out in the new issue, drop-in of arbitrary OpenEXR libs into a cascade of symlinks such that one can simply link libOpenEXR.so doesn't work anyway because the version is burned into the ABI.
I'd like propose that we go with this PR as is for now. It doesn't change behavior or policy, it simply names variables in a slightly less confusing way. We approved similar changes for Imath, which went into 3.1. I don't think we should hold up the 3.1 release of OpenEXR in hopes of resolving this any more effectively, although I'm open to ideas.
Can someone approve the review so I can merge it?
|
gharchive/pull-request
| 2021-06-24T21:28:23 |
2025-04-01T04:54:40.998754
|
{
"authors": [
"cary-ilm",
"lgritz",
"meshula"
],
"repo": "AcademySoftwareFoundation/openexr",
"url": "https://github.com/AcademySoftwareFoundation/openexr/pull/1069",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2329876832
|
Media queries
I've been trying to port Xess to gcss as an exercise to see what I think about it and I've run into an issue, I can't properly express media queries like this:
@media only screen and (max-device-width: 736px) {
main {
padding: 0;
}
}
What's the gcss way to do this?
Hi @Xe
Ive added a basic example of how you could do it. The end goal with any of it is you just need to wrap styles with something like the media string to the writer. In short there isn't a way and im keen not to impose it as that will keep the repo free of problems.
But here is an example:
package main
import (
"github.com/AccentDesign/gcss"
"github.com/AccentDesign/gcss/props"
"io"
"os"
)
type (
Styles []gcss.Style
Media struct {
Query string
Styles Styles
}
Stylesheet struct {
Styles Styles
Medias []Media
}
)
// WriteCSS writes the CSS for the media query to the writer
func (m Media) WriteCSS(w io.Writer) error {
if _, err := io.WriteString(w, m.Query); err != nil {
return err
}
if _, err := io.WriteString(w, "{"); err != nil {
return err
}
for _, style := range m.Styles {
if err := style.CSS(w); err != nil {
return err
}
}
if _, err := io.WriteString(w, "}"); err != nil {
return err
}
return nil
}
// WriteCSS writes the CSS for the stylesheet to the writer
func (ss Stylesheet) WriteCSS(w io.Writer) error {
// Write the base styles first
for _, style := range ss.Styles {
if err := style.CSS(w); err != nil {
return err
}
}
// Write the media queries next
for _, media := range ss.Medias {
if err := media.WriteCSS(w); err != nil {
return err
}
}
return nil
}
var (
base = Styles{
{
Selector: "body",
Props: gcss.Props{
Margin: props.UnitRaw(0),
},
},
}
screen736 = Media{
Query: "@media only screen and (max-device-width: 736px)",
Styles: Styles{
{
Selector: "main",
Props: gcss.Props{
Padding: props.UnitRaw(0),
},
},
},
}
stylesheet = Stylesheet{
Styles: base,
Medias: []Media{screen736},
}
)
// This is just a basic idea of how you could structure your CSS
// the goal hear is just to wrap the css how you wish with what ever you wish
// construct your stylesheet to suit your needs.
// The end goal is just to call CSS on each style with the object to write to.
func main() {
file, err := os.Create("media.css")
if err != nil {
panic(err)
}
defer file.Close()
if err := stylesheet.WriteCSS(file); err != nil {
panic(err)
}
}
hope this helps in some way
PS. when you get commit ur happy with feel free to pr it to the examples would love to see it :)
Sorry to wade in here but I've been evaluating gcss - couldn't this also be solved (and also enable nested styles) by just including a []Styles child field in Style?
hi @jamiec7919,
yeh there are a multitude of ways really, what ever suits the needs best. I did have a play around with a sort of starting point after this issue at https://github.com/AccentDesign/gcss-starter as a "could it actually generate a more complete basic stylesheet to cover the basics" and "does it actually need anything adding that can serve this kind of use more specifically".
I did toy with the nested idea, but thought I would wait and chew it over. The only thought off the cuff would be validity of the selectors and whether it is actually a media query or standard selectors.
i'm not sure where the lib is going at present. Still trying to work that out. it feels like we should either do it properly (aka more complete). The bare basics. Or nothing :)
Stu
|
gharchive/issue
| 2024-06-02T21:43:31 |
2025-04-01T04:54:41.004578
|
{
"authors": [
"Xe",
"jamiec7919",
"stuartaccent"
],
"repo": "AccentDesign/gcss",
"url": "https://github.com/AccentDesign/gcss/issues/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2110842576
|
Carousel syncing does not work when using 'responsive' setting
I have carousels that are synced using the following example
Using this example https://codepen.io/A360/pen/BaKwaGa
But I'm trying to get the thumbnail nav to become horizontal for smaller screens, I'm doing this via the 'responsive' setting, but it seems that once the breakpoint happens the syncing is broken (even if i drag the window back to a larger size).
e.g.
on https://codepen.io/A360/pen/BaKwaGa
If you add the responsive setting to the '.thumbnail-slider' e.g.
responsive: [ { breakpoint: 1000, settings: { vertical: false, } } ],
This will switch the slider to horizontal as expected but clicking on the thumbnail slider buttons do not change the main slider, syncing between the two is broken.
note* that example will look weird as the CSS in that example doesn't accommodate for that behavior, but the clicking behavior should still work I believe
I just realize that example is doing carousel syncing manually not using the settings, so its probably an issue with that
|
gharchive/issue
| 2024-01-31T19:47:28 |
2025-04-01T04:54:41.008285
|
{
"authors": [
"Pubudu-Basnayaka-COS"
],
"repo": "Accessible360/accessible-slick",
"url": "https://github.com/Accessible360/accessible-slick/issues/98",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
591031645
|
AT2-464 - Standalone service without vasco - [do not merge]
Following the pattern used in assess-api, this allows the static service to run securely as a standalone service. The vasco repo is still referenced, but only as a library.
It can't be merge to master yet since all services must be upgraded simulataneously. (They require AWS Load Balancer changes)
Following the pattern used in assess-api, this allows the static service to run securely as a standalone service. The vasco repo is still referenced, but only as a library.
It can't be merge to master yet since all services must be upgraded simulataneously. (They require AWS Load Balancer changes)
Is there a way to avoid this big bang approach so that each service can be done one at a time?
Following the pattern used in assess-api, this allows the static service to run securely as a standalone service. The vasco repo is still referenced, but only as a library.
It can't be merged to master yet since all services must be upgraded simultaneously. (They require AWS Load Balancer changes)
Is there a way to avoid this big bang approach so that each service can be done one at a time? If so, what is the cost?
Since this change decouples each service from vasco, the AT2 application itself fails until the AWS load balancer's listeners are reconfigured to not route to vasco, but rather route directly to the services. This is all due to vasco doing both routing and service discovery. I suppose it might possible, but I'd have to do some investigation. The biggest risk is that QT development is interrupted since QT calls the AT2 learnosity-mirror service to retrieve Learnosity data. I'll give it some thought. Thanks!
|
gharchive/pull-request
| 2020-03-31T11:33:37 |
2025-04-01T04:54:41.015606
|
{
"authors": [
"StevenPyle",
"glewis-ANet"
],
"repo": "AchievementNetwork/static",
"url": "https://github.com/AchievementNetwork/static/pull/55",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2464721134
|
-repack 之后微信提示清理缓存,repack没用
repack之后,微信打开小程序直接就告诉我清理缓存,应该是检测到修改了直接不让运行,有什么方法可以处理这个问题吗
repack之后,微信打开小程序直接就告诉我清理缓存,应该是检测到修改了直接不让运行,有什么方法可以处理这个问题吗
可以提供截图吗
|
gharchive/issue
| 2024-08-14T01:48:08 |
2025-04-01T04:54:41.023529
|
{
"authors": [
"Ackites",
"Bre8gljm20sendaWofqeoj8biu6odhouse"
],
"repo": "Ackites/KillWxapkg",
"url": "https://github.com/Ackites/KillWxapkg/issues/24",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
362218143
|
update Activiti/activiti-build to master
UpdateBot pushed version changes from the source code in repository: Activiti/activiti-build ref: master
UpdateBot commands:
updatebot push --ref master https://github.com/Activiti/activiti-build.git
Thank you for your submission, we really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.jenkins-x-bot seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account.You have signed the CLA already but the status is still pending? Let us recheck it.
|
gharchive/pull-request
| 2018-09-20T14:35:34 |
2025-04-01T04:54:41.035626
|
{
"authors": [
"CLAassistant",
"jx-activiti-cloud"
],
"repo": "Activiti/activiti-api",
"url": "https://github.com/Activiti/activiti-api/pull/9",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
588411878
|
fix(versions): update Activiti/activiti-cloud-modeling versions
UpdateBot pushed maven dependency: org.activiti.cloud.modeling:activiti-cloud-modeling-dependencies to: 7.1.425
UpdateBot commands:
updatebot push-version --kind maven org.activiti.cloud.modeling:activiti-cloud-modeling-dependencies 7.1.425 org.activiti.cloud.audit:activiti-cloud-audit-dependencies 7.1.425 org.activiti.cloud.api:activiti-cloud-api-dependencies 7.1.425 org.activiti.cloud.build:activiti-cloud-parent 7.1.425 org.activiti.cloud.build:activiti-cloud-dependencies-parent 7.1.425 org.activiti.cloud.connector:activiti-cloud-connectors-dependencies 7.1.425 org.activiti.cloud.messages:activiti-cloud-messages-dependencies 7.1.425 org.activiti.cloud.modeling:activiti-cloud-modeling-dependencies 7.1.425 org.activiti.cloud.notifications.graphql:activiti-cloud-notifications-graphql-dependencies 7.1.425 org.activiti.cloud.query:activiti-cloud-query-dependencies 7.1.425 org.activiti.cloud.rb:activiti-cloud-runtime-bundle-dependencies 7.1.425 org.activiti.cloud.acc:activiti-cloud-acceptance-tests-dependencies 7.1.425 org.activiti.cloud.common:activiti-cloud-service-common-dependencies 7.1.425
|
gharchive/pull-request
| 2020-03-26T13:14:35 |
2025-04-01T04:54:41.037501
|
{
"authors": [
"jx-activiti-cloud"
],
"repo": "Activiti/activiti-cloud-modeling",
"url": "https://github.com/Activiti/activiti-cloud-modeling/pull/807",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
368098765
|
update org.activiti.cloud.build:activiti-cloud-parent to 7.0.30
UpdateBot pushed maven dependency: org.activiti.cloud.build:activiti-cloud-parent to: 7.0.30
UpdateBot commands:
updatebot push-version --kind maven org.activiti.cloud.build:activiti-cloud-parent 7.0.30 org.activiti.cloud.build:activiti-cloud-dependencies-parent 7.0.30
UpdateBot commands:
updatebot push-version --kind maven org.activiti.cloud.build:activiti-cloud-parent 7.0.31 org.activiti.cloud.build:activiti-cloud-dependencies-parent 7.0.31
|
gharchive/pull-request
| 2018-10-09T08:37:20 |
2025-04-01T04:54:41.039636
|
{
"authors": [
"jx-activiti-cloud"
],
"repo": "Activiti/activiti-cloud-service-common",
"url": "https://github.com/Activiti/activiti-cloud-service-common/pull/70",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
494194174
|
update org.activiti.cloud.dependencies:activiti-cloud-dependencies to 7.1.138
UpdateBot pushed maven dependency: org.activiti.cloud.dependencies:activiti-cloud-dependencies to: 7.1.138
UpdateBot commands:
updatebot push-version --kind maven org.activiti.cloud.dependencies:activiti-cloud-dependencies 7.1.138 --merge false
|
gharchive/pull-request
| 2019-09-16T17:59:40 |
2025-04-01T04:54:41.041559
|
{
"authors": [
"jx-activiti-cloud"
],
"repo": "Activiti/example-runtime-bundle",
"url": "https://github.com/Activiti/example-runtime-bundle/pull/339",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
513982564
|
update org.activiti.cloud.dependencies:activiti-cloud-dependencies to 7.1.169
UpdateBot pushed maven dependency: org.activiti.cloud.dependencies:activiti-cloud-dependencies to: 7.1.169
UpdateBot commands:
updatebot push-version --kind maven org.activiti.cloud.dependencies:activiti-cloud-dependencies 7.1.169 --merge false
|
gharchive/pull-request
| 2019-10-29T14:56:54 |
2025-04-01T04:54:41.043169
|
{
"authors": [
"jx-activiti-cloud"
],
"repo": "Activiti/ttc-connectors-dummytwitter",
"url": "https://github.com/Activiti/ttc-connectors-dummytwitter/pull/244",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
45500745
|
validates_uniqueness_of_without_deleted undefined in Rails 4.2
More of an FYI I suppose since 4.2 is not official yet, but the above method is undefined both on master and on 0.5.0.beta1 when used as instructed in the README
Any update on this @zzak ?
@allaire there is a patch, maybe you can point your Gemfile to @mvz's fix-build branch?
https://github.com/mvz/acts_as_paranoid/tree/fix-build
If you try it please report any issues to #26, thank you!
+1
This should be fixed in the last release, please check it. <3
Please try 0.5.0, this should be fixed now
|
gharchive/issue
| 2014-10-10T16:50:05 |
2025-04-01T04:54:41.055734
|
{
"authors": [
"allaire",
"suan",
"vaughanj10",
"zzak"
],
"repo": "ActsAsParanoid/acts_as_paranoid",
"url": "https://github.com/ActsAsParanoid/acts_as_paranoid/issues/20",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
936281605
|
[Breath of the Wild] FPS++ - High Refresh Rate Causes Excessive Bokoblin/Moblin Ragdolling
Describe the issue
I'm playing the game at 144hz using the FPS++ mod. I have noticed that the ragdolling of enemies increases greatly. For example, when I get a headshot on a bokoblin, they go absolutely flying, several times further than expected. I would assume this issue has to do with the speed of the engine being accelerated due to the higher FPS.
To Reproduce
Steps to reproduce the behavior:
Enable a higher refresh rate in FPS++ (it is extremely noticeable past 100FPS)
Cause a bokoblin to go flying due to a critical hit of some sort (I recommend a headshot with an arrow)
The bokoblin will be yeeted across the environment instead of flying backwards a few feet.
Expected behavior
I would expect the knockback/ragdolling to behave similarly to how it does in 30fps.
Desktop (please complete the following information):
OS: Windows
GPU: Nvidia
Renderer: Vulkan
Version: 1.2.170
This is now fixed with the new major graphic pack update, thanks to the work of @Exzap!
|
gharchive/issue
| 2021-07-03T15:58:32 |
2025-04-01T04:54:41.059705
|
{
"authors": [
"Crementif",
"dimanaum"
],
"repo": "ActualMandM/cemu_graphic_packs",
"url": "https://github.com/ActualMandM/cemu_graphic_packs/issues/526",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
946678997
|
“models/: Contains the neural networks used. Also contains the trained models for each task.”
I did not see the trained model in this folder, could you elaborate more on how to play with different examples shown in paper?
Hi @Capri2014, so sorry about that! We had .gitignore'ed the trained models by accident. It has now been added to the models/folder. The instructions have also been updated with examples on how to run the scripts.
|
gharchive/issue
| 2021-07-17T01:34:13 |
2025-04-01T04:54:41.119625
|
{
"authors": [
"Capri2014",
"LeeYiyuan"
],
"repo": "AdaCompNUS/magic",
"url": "https://github.com/AdaCompNUS/magic/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
939466466
|
Issues flagged by Cloud Manager - Clientlib versioning
Required Information
[ ] AEM Version, including Service Packs, Cumulative Fix Packs, etc: 6.5.5
[ ] ACS AEM Commons Version: 4.11.2
[ ] Reproducible on Latest? yes/no Yes
Expected Behavior
As per versioned clientlib doc https://adobe-consulting-services.github.io/acs-aem-commons/features/versioned-clientlibs/index.html, we created config at Level 4... /apps/myapp/config/rewriter/versioned-clientlibs. With that we are seeing a major bug in Scan reports from Cloudmanager
ConfigAndInstallShouldOnlyContainOsgiNodes -- Node /apps/myapp/config is an OSGi config or install path which contains non-OSGi-related children that will not be visible to non-administrative users.
Actual Behavior
There shouldnt be Cloud Manager issue with versioned clientlib config
Steps to Reproduce
Install ACS commons package
Configure versioned clienlibs at /apps/myapp/config/rewriter/versioned-clientlibs and run CM pipeline
Links
N.A.
@viveksachdeva The newer AEM Archetype splits configs into another package (ui.confg) from the code (ui.apps) ... These might be part of the "cloud ready" recent changes ?
https://experienceleague.adobe.com/docs/experience-manager-core-components/using/developing/archetype/using.html?lang=en
@viveksachdeva mm - interesting. Just a quick question - any reason you arent using OOTB client lib versioning? Highly recommend you do so over ACS Commons, unless you have a good reason.
@justinedelson any insights on this? The rewriter config is a sling:Folder under the /apps/.../config node ... CM is unhappy that the node is not an OSGi config (ie. sling:OsgiConfig/.cfg.json).
Pretty sure this is a false positive in the CM build. Configuration pipelines must be configured in that path: https://sling.apache.org/documentation/bundles/output-rewriting-pipelines-org-apache-sling-rewriter.html#configuring-a-processor-1
@kwin that was my expectation as well. JE might be busy - ill log an issue with CM team internally and ask them to review the rule.
@davidjgonzalez @kwin I'd never say never, but this was actually an issue 2 years ago that I don't think has regressed.
FWIW, the content structured used in testing this specific case is:
/apps
/testco
/config
/com.day.cq.rewriter.linkchecker.impl.LinkCheckerImpl.xml (sling:OsgiConfig)
/rewriter (sling:Folder)
/pdf (nt:unstructured)
i.e. /apps/testco/config/rewriter/.content.xml declares the node as sling:Folder and /apps/testco/config/rewriter/pdf/.content.xml declares that node as nt:unstructured.
I see that the ACS AEM Commons docs say to use a sling:Folder node for the configuration. This, however, does not have an impact (just validated this as well).
@viveksachdeva If you can provide the execution ID, I can probably look up some more detail. We could also move this to an AEM support issue if you would be more comfortable with that.
This test project does not generate any issues.
I have connected with out PS to relay this information as this could have some client info. thanks!
Hi,
We too are facing same issue. Could you please let me know how it was resolved?
Thanks,
Shilpa
|
gharchive/issue
| 2021-07-08T04:51:26 |
2025-04-01T04:54:41.251399
|
{
"authors": [
"ShilpaAmbi",
"by-4x1",
"davidjgonzalez",
"justinedelson",
"kwin",
"viveksachdeva"
],
"repo": "Adobe-Consulting-Services/acs-aem-commons",
"url": "https://github.com/Adobe-Consulting-Services/acs-aem-commons/issues/2637",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
644641392
|
Update buildJDK id
Fix the run java version step
https://github.com/AdoptOpenJDK/build-jdk/runs/801659565?check_suite_focus=true
Signed-off-by: Sophia Guo sophiag@ca.ibm.com
Hotspot builds now ok, but java -version step still failing for OpenJ9 @sophia-guo
|
gharchive/pull-request
| 2020-06-24T14:03:12 |
2025-04-01T04:54:41.266102
|
{
"authors": [
"smlambert",
"sophia-guo"
],
"repo": "AdoptOpenJDK/build-jdk",
"url": "https://github.com/AdoptOpenJDK/build-jdk/pull/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
434259296
|
how should i install this in my custom docker environment?
Hello,
trying installing jdk 8 on my docker machine. Running this:
git clone https://github.com/AdoptOpenJDK/openjdk-build
cd openjdk-build
./makejdk-any-platform.sh --docker jdk8u
i am getting the following error:
Starting ./makejdk-any-platform.sh to configure, build (Adopt)OpenJDK binary
Parsing opt: --docker
Possible opt arg: jdk8u
Working dir is ./build/
[debug] COPY_MACOSX_FREE_FONT_LIB_FOR_JDK_FLAG=true
[debug] COPY_MACOSX_FREE_FONT_LIB_FOR_JRE_FLAG=true
JDK Image folder name: j2sdk-image
JRE Image folder name: j2re-image
Searching for JDK_BOOT_DIR
readlink: missing operand
Try 'readlink --help' for more information.
dirname: missing operand
Try 'dirname --help' for more information.
dirname: missing operand
Try 'dirname --help' for more information.
any suggestion?
thanks
Searching the repo for readlink gives only this as result:
https://github.com/AdoptOpenJDK/openjdk-build/blob/95dc19f9b5a9ecc02addae45312d047ad154cfe8/sbin/common/common.sh#L129
Install a JDK on your system, so which javac suceeeds and building works again (Just tried that, had the same problem).
What O/S are you running? what happens when you run:
which javac and subsequently readlink -f $(which javac)?
Hi @fede-green . Have you tried Martijn's comment above?
This was resolved some time ago (we enhanced the detection)
|
gharchive/issue
| 2019-04-17T12:37:20 |
2025-04-01T04:54:41.269938
|
{
"authors": [
"M-Davies",
"MeFisto94",
"fede-green",
"karianna"
],
"repo": "AdoptOpenJDK/openjdk-build",
"url": "https://github.com/AdoptOpenJDK/openjdk-build/issues/1043",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
603365798
|
remove zulu-7 from the jdk8 build
zulu-7 no longer appears to be needed or used during the build.
(Perhaps it was used previously as part of a bootstrap?)
The PR also installs ca-certificates, which is still needed and was
being pulled in by software-properties-common.
I realize that this doesn't change much. This saves one apt-get update during the build process and avoids unnecessary traffic to Azul's repo, so perhaps speeds up the build just a tad. But I was curious about "why we can't have nice things..."
@tmancill - how does the docker build bootstrap itself without the Java 7?
@tmancill - how does the docker build bootstrap itself without the Java 7?
@karianna I was puzzling over this as well when I started looking into it. Ostensibly the assumption is that we're already successfully bootstrapped and can retrieve jdk8 from api.adoptopenjdk.net in that Dockerfile here; these lines changed in https://github.com/AdoptOpenJDK/openjdk-build/pull/1278.
To test, I cleaned my environment with docker system prune -a to ensure that a container with a JDK wasn't being reused from cache, and then built with ./makejdk-any-platform.sh --docker jdk8u:
...
Step 4/24 : RUN mkdir -p /openjdk/target
---> Running in 4b5869d461f2
Removing intermediate container 4b5869d461f2
---> a06a1507ac08
Step 5/24 : RUN wget 'https://api.adoptopenjdk.net/v2/binary/releases/openjdk8?openjdk_impl=hotspot&os=linux&arch=x64&release=latest&type=jdk' -O jdk8.tar.gz
---> Running in 44d266989564
--2020-04-20 18:36:59-- https://api.adoptopenjdk.net/v2/binary/releases/openjdk8?openjdk_impl=hotspot&os=linux&arch=x64&release=latest&type=jdk
Resolving api.adoptopenjdk.net (api.adoptopenjdk.net)... 104.17.158.60, 104.17.159.60
Connecting to api.adoptopenjdk.net (api.adoptopenjdk.net)|104.17.158.60|:443... connected.
HTTP request sent, awaiting response... 302 Found
Location: https://github.com/AdoptOpenJDK/openjdk8-binaries/releases/download/jdk8u252-b09/OpenJDK8U-jdk_x64_linux_hotspot_8u252b09.tar.gz [following]
...
And then later on in openjdk_container:latest:
...
Searching for JDK_BOOT_DIR
Guessing JDK_BOOT_DIR: /usr/lib/jvm/jdk8
If this is incorrect explicitly configure JDK_BOOT_DIR
Boot dir set to /usr/lib/jvm/jdk8
Running gradle with /usr/lib/jvm/jdk8
...
So we're currently building jdk8 with itself. Should we break the recursion?
While building with JDK8 works, we generally use the previous version, but if we've already got it pulling the adoptopenjdk8 then this seems like a reasonable change ... thoughts @karianna ?
I'm actually going to be a pian here and state that we should stay with the Zulu 7 as the bootstrap JDK. We should ideally always build with version -1 (as recommended by the upstream guides).
Makes sense to me. Closing this PR. I'll have another look to see what it will take to build with 7 have Java 8 to run Gradle.
|
gharchive/pull-request
| 2020-04-20T16:24:16 |
2025-04-01T04:54:41.275920
|
{
"authors": [
"karianna",
"sxa",
"tmancill"
],
"repo": "AdoptOpenJDK/openjdk-build",
"url": "https://github.com/AdoptOpenJDK/openjdk-build/pull/1682",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
740843578
|
System unavailable: test-osuosl-aix71-ppc64-1 insufficient disk space
System tests are failing on test-osuosl-aix71-ppc64-1 because there is <3Gb available at the start of the tests (which is checked at the start of each test): https://ci.adoptopenjdk.net/job/Test_openjdk11_hs_sanity.system_ppc64_aix/48/consoleFull
17:13:31 ===============================================
17:13:31 Running test CLLoadTest_0 ...
17:13:31 ===============================================
17:13:31 CLLoadTest_0 Start Time: Tue Nov 10 17:13:31 2020 Epoch Time (ms): 1605028411166
17:13:31 Nothing to be done for setup.
17:13:31 variation: NoOptions
17:13:31 JVM_OPTIONS:
17:13:31 { itercnt=1; \
17:13:31 mkdir -p "/home/jenkins/workspace/Test_openjdk11_hs_sanity.system_ppc64_aix/openjdk-tests/TKG/../TKG/test_output_16050238357263/CLLoadTest_0"; \
17:13:31 cd "/home/jenkins/workspace/Test_openjdk11_hs_sanity.system_ppc64_aix/openjdk-tests/TKG/../TKG/test_output_16050238357263/CLLoadTest_0"; \
17:13:31 perl /home/jenkins/workspace/Test_openjdk11_hs_sanity.system_ppc64_aix/openjdk-tests/TKG/../../jvmtest/system/modularity/..//stf/stf.core/scripts/stf.pl -test-root="/home/jenkins/workspace/Test_openjdk11_hs_sanity.system_ppc64_aix/openjdk-tests/TKG/../../jvmtest/system/modularity/..//stf;/home/jenkins/workspace/Test_openjdk11_hs_sanity.system_ppc64_aix/openjdk-tests/TKG/../../jvmtest/system/modularity/..//openjdk-systemtest""" -systemtest-prereqs="/home/jenkins/workspace/Test_openjdk11_hs_sanity.system_ppc64_aix/openjdk-tests/TKG/../../jvmtest/system/modularity/..//systemtest_prereqs" -java-args=" " -results-root="/home/jenkins/workspace/Test_openjdk11_hs_sanity.system_ppc64_aix/openjdk-tests/TKG/../TKG/test_output_16050238357263/CLLoadTest_0" -test=CLLoadTest; \
17:13:31 if [ $? -eq 0 ] ; then echo ""; echo "CLLoadTest_0""_PASSED"; echo ""; cd /home/jenkins/workspace/Test_openjdk11_hs_sanity.system_ppc64_aix/openjdk-tests/TKG/..; else echo ""; echo "CLLoadTest_0""_FAILED"; echo ""; fi; } 2>&1 | tee -a "/home/jenkins/workspace/Test_openjdk11_hs_sanity.system_ppc64_aix/openjdk-tests/TKG/../TKG/test_output_16050238357263/TestTargetResult";
17:13:31 STF 17:13:31.421 - ========================= S T F =========================
17:13:31 systemtest-prereqs has been processed, and set to: /home/jenkins/workspace/Test_openjdk11_hs_sanity.system_ppc64_aix/jvmtest/system/systemtest_prereqsRetrieving amount of free space on drive containing /home/jenkins/workspace/Test_openjdk11_hs_sanity.system_ppc64_aix/openjdk-tests/TKG/../TKG/test_output_16050238357263/CLLoadTest_0
17:13:31 There is 3025 Mb free
17:13:31
17:13:31 Test machine has only 3025 Mb free on drive containing /home/jenkins/workspace/Test_openjdk11_hs_sanity.system_ppc64_aix/openjdk-tests/TKG/../TKG/test_output_16050238357263/CLLoadTest_0.
17:13:31
17:13:31 There must be at least 3Gb (3072Mb) free to be sure of capturing diagnostics
17:13:31 files in the event of a test failure.
17:13:31
17:13:31 Exiting.
17:13:31
17:13:31 CLLoadTest_0_FAILED
Seems to have about 15Gb free just now, so closing. Feel free to reopen if it recurs
|
gharchive/issue
| 2020-11-11T15:13:33 |
2025-04-01T04:54:41.278768
|
{
"authors": [
"lumpfish",
"sxa"
],
"repo": "AdoptOpenJDK/openjdk-infrastructure",
"url": "https://github.com/AdoptOpenJDK/openjdk-infrastructure/issues/1676",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
294998003
|
Initial attempt at formatting hostnames/inventory
So I have used some of the fantastic code written at https://github.com/nodejs/build to setup an inventory based system. The key thing to this is that the backend python does some pretty clever stuff. You can see the inventory here.
Example:
hosts:
- build:
- cloudcone:
ubuntu1604-x64-1: {ip: 173.82.219.221}
Becomes build-cloudcone-ubuntu1604-x64-1
And this in turn means that the ansible variables are a bit more useful:
ok: [build-cloudcone-ubuntu1604-x64-1] => {
"msg": [
"inventory_hostname: build-cloudcone-ubuntu1604-x64-1 ",
"ansible_hostname: build-cloudcone-x64-ubuntu-16-04-1",
"ansible_ssh_private_key_file: ~/.ssh/id_rsa",
"ansible_fqdn: build-cloudcone-x64-ubuntu-16-04-1.cloudcone.com",
"ansible_default_ipv4.address: 173.82.219.221",
"ansible_os_family: Debian ",
"ansible_distribution: Ubuntu ",
"ansible_distribution_major_version: 16 ",
"ansible_architecture: x86_64 ",
"ansible_processor_vcpus: 4 ",
"ansible_processor_cores: 1 ",
"Jenkins_Username: jenkins ",
"Superuser_Account: Enabled",
"Vendor_File: ***Undefined***",
"Nagios_Plugins: Enabled ",
"Nagios_Monitoring: Enabled",
"Nagios_Master_IP: 78.47.239.96"
]
}
This should then allow us to standardise the hostnames which was part of the problem but can also be used as the hostname for the nagios master updates etc.
To Run:
ansible-playbook playbooks/AdoptOpenJDK_Linux_Playbook/main.yml --limit "build-cloudcone*"
macOS requires you to run:
export PYTHONPATH=$(pip2 show pyyaml | grep Location | awk '{print $2}')
(thanks @gibfahn for working that one out)
I have managed to get AWX to pick up this inventory style so I am going to go ahead and merge the PR and update AWX to pull from our inventory.yml. I am then keen to go through nagios and remove most of the unix hosts as as lot of them will have the wrong name and then we can let the playbook, auto set the new hostname
|
gharchive/pull-request
| 2018-02-07T03:50:48 |
2025-04-01T04:54:41.283123
|
{
"authors": [
"gdams"
],
"repo": "AdoptOpenJDK/openjdk-infrastructure",
"url": "https://github.com/AdoptOpenJDK/openjdk-infrastructure/pull/181",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
816903210
|
JVM crash on ARM64
Summary
JVM inside Docker container is crashing in the C1 Compiler Thread after around 5-10 of running on AArch64 Amazon Linux 2 host
Steps to reproduce
Not sure what to provide here. This happens on one of three containers using the same JVM version, same Spring Boot version, same everything only the class name in current compilation ask is only used on container that is crashing. I'd be happy to provide this class privately. Let me know what else I can provide
Expected results
JVM does not crash
Actual results
Hard crash
Triaging info
Java version:
openjdk version "11.0.10" 2021-01-19
OpenJDK Runtime Environment AdoptOpenJDK (build 11.0.10+9)
OpenJDK 64-Bit Server VM AdoptOpenJDK (build 11.0.10+9, mixed mode)
What is your operating system and platform?
Container is based on arm64v8/adoptopenjdk:11-jdk-hotspot
Host: AArch64 Amazon Linux 2 - 4.14.219-161.340.amzn2.aarch64 #1 SMP Thu Feb 4 05:54:27 UTC 2021 aarch64 aarch64 aarch64 GNU/Linux
How did you install Java?
arm64v8/adoptopenjdk:11-jdk-hotspot container comes with it
Did it work before?
Switched to Arm64
Did you test with other Java versions?
yes, java 11.0.10 openj9 works fine; only hotspot having issues
hs_err_pid7.log
replay_pid7.log
No rush on this, I'm using the OpenJ9 build currently. This was more an FYI ticket.
Gets fix in April: https://bugs.openjdk.java.net/browse/JDK-8247766. Nightly builds should have it.
Might need a backport to 8.
Good enough for me.
|
gharchive/issue
| 2021-02-25T23:53:50 |
2025-04-01T04:54:41.289031
|
{
"authors": [
"aahlenst",
"markathomas"
],
"repo": "AdoptOpenJDK/openjdk-support",
"url": "https://github.com/AdoptOpenJDK/openjdk-support/issues/266",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
736560774
|
cmdLineTester_SCCommandLineOptionTests_1 fail (jdk8/j9)
https://github.com/AdoptOpenJDK/run-aqa/runs/1354982566?check_suite_focus=true#step:4:50735
Testing: Run !printallcallsites
Test start time: 2020/11/04 21:36:14 Coordinated Universal Time
Running command: /opt/hostedtoolcache/jdk-8-openj9/1.0.0/x64/bin/jdmpview -core j9core.dmp
Time spent starting: 2 milliseconds
Time spent executing: 689 milliseconds
Test result: FAILED
Output from test:
[OUT] DTFJView version 4.29.5, using DTFJ version 1.12.29003
[OUT] Loading image from DTFJ...
[OUT]
[OUT] Could not load dump file and/or could not load XML file: null
[OUT] For a list of commands, type "help"; for how to use "help", type "help help"
[OUT] > DDR is not enabled for this core file, '!' commands are disabled
[OUT] >
https://github.com/AdoptOpenJDK/run-aqa/runs/1354982566?check_suite_focus=true#step:4:53130
[OUT] nameOption2: Expected to find the cache name
[OUT] TEST PASSED
[ERR] JVMSHRC806I Compressed references persistent shared cache "runner" has been destroyed. Use option -Xnocompressedrefs if you want to destroy a non-compressed references cache.
Looks like cmdLineTester_callsitedbgddrext_openj9 passed in recent running. https://github.com/AdoptOpenJDK/run-aqa/runs/1476255594?check_suite_focus=true#step:4:33468. Not sure if it's intermittent or fixed.
Updating this issue to remove cmdLineTester_callsitedbgddrext_openj9 from the heading. That test was failing consistently a few weeks ago, but has not failed for a while. I do not see an obvious PR/fix that would change behaviour, but we can raise it as a separate issue if its turned into something intermittent.
Update cmdLineTester_SCCommandLineOptionTests_1 failures on linux with jdk8/11/15 ( Original one is the failure of cmdLineTester_callsitedbgddrext_openj9):
Testing: nameOption2
Test start time: 2020/12/02 20:42:49 Coordinated Universal Time
Running command: sh nameOption2.sh /opt/hostedtoolcache/jdk-8-openj9/1.0.0/x64/bin
Time spent starting: 1 milliseconds
Time spent executing: 272 milliseconds
Test result: FAILED
[OUT] Hello, world!
[OUT] TEST PASSED
[OUT] nameOption2: TEST FAILED
[OUT] nameOption2: Expected to find the cache name
[OUT] TEST PASSED
[ERR] JVMSHRC806I Compressed references persistent shared cache "runner" has been destroyed. Use option -Xnocompressedrefs if you want to destroy a non-compressed references cache.
Success condition was found: [Output match: TEST PASSED]
Failure condition was found: [Output match: TEST FAILED]
Failure condition was not found: [Output match: Error:]
Failure condition was not found: [Output match: Unhandled Exception]
Failure condition was not found: [Output match: Exception:]
...
Testing: nameOption4
Test start time: 2020/12/02 20:42:50 Coordinated Universal Time
Running command: sh nameOption4.sh /opt/hostedtoolcache/jdk-8-openj9/1.0.0/x64/bin
Time spent starting: 8 milliseconds
Time spent executing: 276 milliseconds
Test result: FAILED
[OUT] Hello, world!
[OUT] TEST PASSED
[OUT] nameOption4: TEST FAILED
[OUT] nameOption4: Expected to find the cache name
[OUT] TEST PASSED
[ERR] JVMSHRC806I Compressed references persistent shared cache "_runner_docker" has been destroyed. Use option -Xnocompressedrefs if you want to destroy a non-compressed references cache.
Success condition was found: [Output match: TEST PASSED]
Failure condition was found: [Output match: TEST FAILED]
Failure condition was not found: [Output match: Error:]
Failure condition was not found: [Output match: Unhandled Exception]
Failure condition was not found: [Output match: Exception:]
cmdLineTester_callsitedbgddrext_openj9 reopened in #50
Test nameOption2 and nameOption4 get the username from the environment variable LOGNAME
export TESTUSER=$LOGNAME
to match the shared Cache generated by NAME=_%u.
In github runner ubuntu environment _%u is runner and $LOGNAME turns out be empty ( github runner macos $LOGNAME is runner, so test passes with macos).
This actually should also affect test nameOption, which set DEFAULT_CACHE_NAME="sharedcc_$TESTUSER" . However as TESTUSER is empty so grep "$DEFAULT_CACHE_NAME" $TESTSCRIPT.out will return 0. Kind of luck, test is actually fragile.
Next step would need to figure out why $LOGNAME is empty on ubuntu runner, and be runner on macos runner. Is this expected?
LOGNAME is not in the Ubuntu GitHub Actions Environment Variables List. Instead RUNNER_USER is the environment for the user runner. There’s always a chance that the list will get outdated as soon as a new container is configured. We will put a warning if it happens and update our fixes.
|
gharchive/issue
| 2020-11-05T02:34:13 |
2025-04-01T04:54:41.303716
|
{
"authors": [
"smlambert",
"sophia-guo"
],
"repo": "AdoptOpenJDK/run-aqa",
"url": "https://github.com/AdoptOpenJDK/run-aqa/issues/39",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.