id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
1522863484
🛑 TROJAN 🇺🇸 United States USO 1 is down In 8557070, TROJAN 🇺🇸 United States USO 1 (https://ust-2.opensvr.net/) was down: HTTP code: 0 Response time: 0 ms Resolved: TROJAN 🇺🇸 United States USO 1 is back up in a40ded8.
gharchive/issue
2023-01-06T16:50:59
2025-04-01T04:35:44.486402
{ "authors": [ "roosterkid" ], "repo": "roosterkid/opentunnel-status-server", "url": "https://github.com/roosterkid/opentunnel-status-server/issues/15963", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1627771038
🛑 TROJAN 🇺🇸 United States USO 1 is down In d881a91, TROJAN 🇺🇸 United States USO 1 (https://ust-2.opensvr.net/) was down: HTTP code: 0 Response time: 0 ms Resolved: TROJAN 🇺🇸 United States USO 1 is back up in 1881246.
gharchive/issue
2023-03-16T15:42:35
2025-04-01T04:35:44.488844
{ "authors": [ "roosterkid" ], "repo": "roosterkid/opentunnel-status-server", "url": "https://github.com/roosterkid/opentunnel-status-server/issues/20282", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2035433464
🛑 SSH 🇮🇩 Indonesia IDA 1 is down In 53630c0, SSH 🇮🇩 Indonesia IDA 1 (http://ids-2.optnl.com:8080/) was down: HTTP code: 0 Response time: 0 ms Resolved: SSH 🇮🇩 Indonesia IDA 1 is back up in f884f0a.
gharchive/issue
2023-12-11T11:14:31
2025-04-01T04:35:44.491364
{ "authors": [ "roosterkid" ], "repo": "roosterkid/opentunnel-status-server", "url": "https://github.com/roosterkid/opentunnel-status-server/issues/22132", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2052061905
🛑 XRAY 🇬🇧 United Kingdom UKO 1 is down In 4d1f5ec, XRAY 🇬🇧 United Kingdom UKO 1 (https://ukx-1.openv2ray.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: XRAY 🇬🇧 United Kingdom UKO 1 is back up in 7db231c.
gharchive/issue
2023-12-21T09:37:28
2025-04-01T04:35:44.493791
{ "authors": [ "roosterkid" ], "repo": "roosterkid/opentunnel-status-server", "url": "https://github.com/roosterkid/opentunnel-status-server/issues/22893", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2056619355
🛑 OVPN 🇨🇦 Canada CAO 1 is down In 4c85fb8, OVPN 🇨🇦 Canada CAO 1 (http://cao-1.optnl.com:8080/) was down: HTTP code: 0 Response time: 0 ms Resolved: OVPN 🇨🇦 Canada CAO 1 is back up in e2915af.
gharchive/issue
2023-12-26T18:30:41
2025-04-01T04:35:44.496392
{ "authors": [ "roosterkid" ], "repo": "roosterkid/opentunnel-status-server", "url": "https://github.com/roosterkid/opentunnel-status-server/issues/23260", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2165866152
🛑 V2RAY 🇫🇷 France FR 1 is down In af758e0, V2RAY 🇫🇷 France FR 1 (https://frv-1.openv2ray.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: V2RAY 🇫🇷 France FR 1 is back up in 3921b3d.
gharchive/issue
2024-03-04T03:53:53
2025-04-01T04:35:44.498806
{ "authors": [ "roosterkid" ], "repo": "roosterkid/opentunnel-status-server", "url": "https://github.com/roosterkid/opentunnel-status-server/issues/28242", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2177351576
🛑 V2RAY 🇫🇷 France FR 1 is down In 150ba2d, V2RAY 🇫🇷 France FR 1 (https://frv-1.openv2ray.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: V2RAY 🇫🇷 France FR 1 is back up in 8621b4e.
gharchive/issue
2024-03-09T18:40:45
2025-04-01T04:35:44.501177
{ "authors": [ "roosterkid" ], "repo": "roosterkid/opentunnel-status-server", "url": "https://github.com/roosterkid/opentunnel-status-server/issues/28584", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1048637314
🛑 TROJAN 🇺🇸 United States USF 1 is down In 4bf17e4, TROJAN 🇺🇸 United States USF 1 (https://ust-1.opensvr.net/) was down: HTTP code: 0 Response time: 0 ms Resolved: TROJAN 🇺🇸 United States USF 1 is back up in 419799d.
gharchive/issue
2021-11-09T13:56:43
2025-04-01T04:35:44.502706
{ "authors": [ "roosterkid" ], "repo": "roosterkid/opentunnel-status-server", "url": "https://github.com/roosterkid/opentunnel-status-server/issues/3677", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1094085902
🛑 TROJAN 🇮🇩 Indonesia IDJ 1 is down In 02c4121, TROJAN 🇮🇩 Indonesia IDJ 1 (https://idt-1.opensvr.net/) was down: HTTP code: 0 Response time: 0 ms Resolved: TROJAN 🇮🇩 Indonesia IDJ 1 is back up in 94dcde6.
gharchive/issue
2022-01-05T08:08:57
2025-04-01T04:35:44.504302
{ "authors": [ "roosterkid" ], "repo": "roosterkid/opentunnel-status-server", "url": "https://github.com/roosterkid/opentunnel-status-server/issues/4693", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1187631289
🛑 TROJAN 🇸🇬 Singapore SGP 1 is down In 862c136, TROJAN 🇸🇬 Singapore SGP 1 (https://sgt-2.opensvr.net/) was down: HTTP code: 0 Response time: 0 ms Resolved: TROJAN 🇸🇬 Singapore SGP 1 is back up in 899af59.
gharchive/issue
2022-03-31T07:22:50
2025-04-01T04:35:44.506375
{ "authors": [ "roosterkid" ], "repo": "roosterkid/opentunnel-status-server", "url": "https://github.com/roosterkid/opentunnel-status-server/issues/6284", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1190765737
🛑 TROJAN 🇸🇬 Singapore SGA 1 is down In 3e383f2, TROJAN 🇸🇬 Singapore SGA 1 (https://sgt-3.opensvr.net/) was down: HTTP code: 0 Response time: 0 ms Resolved: TROJAN 🇸🇬 Singapore SGA 1 is back up in 6b0640c.
gharchive/issue
2022-04-02T23:45:20
2025-04-01T04:35:44.508010
{ "authors": [ "roosterkid" ], "repo": "roosterkid/opentunnel-status-server", "url": "https://github.com/roosterkid/opentunnel-status-server/issues/6324", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1410393715
🛑 TROJAN 🇺🇸 United States USF 1 is down In 3647f94, TROJAN 🇺🇸 United States USF 1 (https://ust-1.opensvr.net/) was down: HTTP code: 0 Response time: 0 ms Resolved: TROJAN 🇺🇸 United States USF 1 is back up in e659065.
gharchive/issue
2022-10-16T06:49:19
2025-04-01T04:35:44.509599
{ "authors": [ "roosterkid" ], "repo": "roosterkid/opentunnel-status-server", "url": "https://github.com/roosterkid/opentunnel-status-server/issues/9502", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1291917489
RootlyIncidentsPage & RootlyOverviewCard not exported I've followed the steps to install the rootly plugin yarn add @rootly/backstage-plugin but when trying to add RootlyIncidentsPage and RootlyOverviewCard , and runing yarn tsc I get the errors : Module '"@rootly/backstage-plugin"' has no exported member 'RootlyIncidentsPage'. Cannot find module '@rootly/backstage-plugin/src/components/Entity' or its corresponding type declarations. I've tried this on my companies internal backstage, but also I created a new backstage application from scratch to test this out with the same issue. Not sure if it's relevant, but I'm using node lts/gallium Thanks @kwent , I already have those two lines in my code, as I copied them over from the readme documentation. I have deleted them and paste them again ( this time from your comment), but it still didn't work. As far as I can tell, the npmjs package for rootly doesn't export those two objects, and that is why this doesn't work. This is what I see when I go through my node_modules for @rootly : import * as _backstage_core_plugin_api from '@backstage/core-plugin-api'; declare const rootlyPlugin: _backstage_core_plugin_api.BackstagePlugin<{ root: _backstage_core_plugin_api.RouteRef<undefined>; }, {}>; declare const RootlyPage: () => JSX.Element; export { RootlyPage, rootlyPlugin }; I'm also attaching a screenshot for sake of clarity :) The version of the rootly backstage pluging that I seem to be using is "@rootly/backstage-plugin": "^0.1.2" Hi @avidelatm, could you try 0.1.7 with import { RootlyIncidentsPage } from '@rootly/backstage-plugin'; import { RootlyOverviewCard } from '@rootly/backstage-plugin'; Regards I am having the same issue, seems like the types declarations are missing in /dist/index.d.ts
gharchive/issue
2022-07-02T00:15:43
2025-04-01T04:35:44.548936
{ "authors": [ "avidelatm", "drankou", "kwent" ], "repo": "rootlyhq/backstage-plugin", "url": "https://github.com/rootlyhq/backstage-plugin/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2629653948
Remove iron workflow and update readme Due after the next patch release on November 15th. @mergifyio backport humble
gharchive/pull-request
2024-11-01T18:41:59
2025-04-01T04:35:44.598968
{ "authors": [ "christophfroehlich" ], "repo": "ros-controls/kinematics_interface", "url": "https://github.com/ros-controls/kinematics_interface/pull/90", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2640884706
[JTC] Fix the JTC length_error exceptions in the tests Fix the exception in JTC tests from time to time [INFO] [1730930292.973865185] [test_joint_trajectory_controller]: Using 'splines' interpolation method. [INFO] [1730930292.974352669] [test_joint_trajectory_controller]: Action status changes will be monitored at 20.00 Hz. terminate called after throwing an instance of 'std::length_error' what(): basic_string::_M_create https://github.com/ros-controls/ros2_controllers/actions/runs/11712509515/job/32623343057?pr=1297 Thanks for the fix :) should we backport this? May be yes! Do we know why this is suddenly an issue? what has changed? Do we know why this is suddenly an issue? what has changed? We are just exiting the test in active state and the LoanedInterfaces are still available and it is throwing this exception upon their destruction(this happens only time to time). I think it should happen in other distros as well, but I don't know why it is not happening.
gharchive/pull-request
2024-11-07T12:29:39
2025-04-01T04:35:44.602536
{ "authors": [ "christophfroehlich", "saikishor" ], "repo": "ros-controls/ros2_controllers", "url": "https://github.com/ros-controls/ros2_controllers/pull/1360", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
431411808
Unused min/max_payload parameters This package's launch files expose min_payload and max_payload arguments, which are used to set equally named parameters, but those parameters are not used anywhere. Those were supported in the master version (which got it from ur_driver), but apparently @Zagitta didn't implement the checks: https://github.com/ros-industrial/ur_modern_driver/blob/77fa08ae9c846344310d3b50824a7affdc3eda47/include/ur_modern_driver/ros/io_service.h#L65-L70 I added help wanted as this could be a good first issue for someone to tackle. Did you make any progress here @ben-greenberg? I have made progress but some further work and testing is still needed. I will be continuing my work on this. I'm going to unassign this @ben-greenberg. You're welcome to keep working on it of course.
gharchive/issue
2019-04-10T09:57:15
2025-04-01T04:35:44.629215
{ "authors": [ "ben-greenberg", "gavanderhoorn", "miguelprada" ], "repo": "ros-industrial/ur_modern_driver", "url": "https://github.com/ros-industrial/ur_modern_driver/issues/289", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
497267410
fanuc: add Melodic source entries As per subject. Thanks for the suggestion @tfoote.
gharchive/pull-request
2019-09-23T18:53:23
2025-04-01T04:35:44.680154
{ "authors": [ "gavanderhoorn" ], "repo": "ros/rosdistro", "url": "https://github.com/ros/rosdistro/pull/22338", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1994179289
fields2cover: 1.2.1-1 in 'humble/distribution.yaml' [bloom] Increasing version of package(s) in repository fields2cover to 1.2.1-1: upstream repository: https://github.com/Fields2Cover/fields2cover.git release repository: https://github.com/Fields2Cover/fields2cover-release.git distro file: humble/distribution.yaml bloom version: 0.10.7 previous version for package: null @Gonzalo-Mier friendly ping I don't understand why or how to solve this error. @mjcarroll do you have any insight? Thank you Typically an entry will have a source, doc and release tag, something like: filters: doc: type: git url: https://github.com/ros/filters.git version: ros2 release: tags: release: release/humble/{package}/{version} url: https://github.com/ros2-gbp/filters-release.git version: 2.1.0-3 source: test_pull_requests: true type: git url: https://github.com/ros/filters.git version: ros2 status: maintained Can you update your pull request to at least include source, but maybe additionally doc? Done. @mjcarroll Thank you for your support!
gharchive/pull-request
2023-11-15T07:08:07
2025-04-01T04:35:44.684546
{ "authors": [ "Gonzalo-Mier", "mjcarroll" ], "repo": "ros/rosdistro", "url": "https://github.com/ros/rosdistro/pull/38990", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2655201865
feetech_ros2_driver: 0.1.0-1 in 'humble/distribution.yaml' [bloom] Increasing version of package(s) in repository feetech_ros2_driver to 0.1.0-1: upstream repository: https://github.com/JafarAbdi/feetech_ros2_driver.git release repository: https://github.com/JafarAbdi/feetech_ros2_driver-release.git distro file: humble/distribution.yaml bloom version: 0.12.0 previous version for package: null feetech_ros2_driver * Add feetech ros2 driver * Contributors: Jafar Uruç See the comment in https://github.com/ros/rosdistro/pull/43490#pullrequestreview-2443491346 , which will probably necessitate changes here. https://github.com/ros/rosdistro/pull/43993 replaces this
gharchive/pull-request
2024-11-13T11:38:49
2025-04-01T04:35:44.688563
{ "authors": [ "JafarAbdi", "clalancette" ], "repo": "ros/rosdistro", "url": "https://github.com/ros/rosdistro/pull/43488", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
105364721
repository_name: 0.0.2-0 in 'indigo/distribution.yaml' [bloom] Increasing version of package(s) in repository repository_name to 0.0.2-0: upstream repository: https://gitlab.uni-koblenz.de/robbie/homer_android_speech.git release repository: https://gitlab.uni-koblenz.de/robbie/homer_android_speech.git distro file: indigo/distribution.yaml bloom version: 0.5.20 previous version for package: null android_speech_pkg * big bang * Contributors: Raphael Memmesheimer repository_name is a generic name Changed to homer_android_speech. Don't know where the name came from. Found it. Copy Paste error from the bloom tutorial.
gharchive/pull-request
2015-09-08T12:04:40
2025-04-01T04:35:44.692141
{ "authors": [ "airglow", "vrabaud" ], "repo": "ros/rosdistro", "url": "https://github.com/ros/rosdistro/pull/9405", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
134445235
fix mega_genmsg to compile while trying to build executables for the Native POSIX portion of the FreeRTPS demo, I found that the messages generated by mega_genmsg.py for rcl_interfaces didn't compile. This fixes it. This isn't part of the binary packaging right? +1 Indeed. look, I can even start a CI job for this, except it won't run any tests. http://ci.ros2.org/job/freertps/19/
gharchive/pull-request
2016-02-18T00:38:05
2025-04-01T04:35:44.719783
{ "authors": [ "jacquelinekay", "wjwwood" ], "repo": "ros2/freertps", "url": "https://github.com/ros2/freertps/pull/30", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
252804460
js files have content undefined Hi, when I try uglify, it overwrites all .js files, but they contet is only one single word "undefined". stripped debug log: cordova -d prepare No scripts found for hook "before_prepare". Checking config.xml and package.json for saved platforms that haven't been added to the project Config.xml and package.json platforms are the same. No pkg.json modification. Package.json and config.xml platforms are different. Updating config.xml with most current list of platforms. Failed to require PlatformApi instance for platform "browser". Using polyfill instead. Checking config.xml for saved plugins that haven't been added to the project Checking for any plugins added to the project that have not been installed in browser platform No differences found between plugins added to project and installed in browser platform. Continuing... Generating config.xml from defaults for platform "browser" Executing script found in hooks directory for hook "after_prepare": hooks\after_prepare\uglify.js Running command: "C:\Program Files\nodejs\node.exe" D:\NetBeansProjects\ReturnReporting\hooks\after_prepare\uglify.js D:\NetBeansProjects\ReturnReporting processFiles err: Error: ENOENT: no such file or directory, scandir 'D:\NetBeansProjects\ReturnReporting\platforms\browser\www\module' uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\controller\BaseController.js uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\controller\Home.controller.js uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\controller\Login.controller.js uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\controller\Settings.controller.js minifying css file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\css\style.css uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\Component.js encountered a .xml file, not compressing it uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\confighelper.js uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\controller\App.controller.js uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\cordova.js uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\exec.js encountered a .html file, not compressing it encountered a .json file, not compressing it full debug log: uglify_debug.txt uglify installed from npm { name: 'cordova-uglify', description: 'Cordova hook that allows you to uglify or minify your apps JavaScript and CSS.', 'dist-tags': { latest: '0.2.8' }, versions: [ '0.0.1', '0.0.2', '0.0.3', '0.0.4', '0.0.5', '0.0.6', '0.0.7', '0.0.8', '0.0.9', '0.1.0', '0.1.1', '0.1.2', '0.1.3', '0.1.4', '0.1.5', '0.1.6', '0.1.7', '0.1.8', '0.1.9', '0.2.0', '0.2.1', '0.2.2', '0.2.3', '0.2.4', '0.2.5', '0.2.6', '0.2.7', '0.2.8' ], maintainers: [ 'rossmartin <rmartin311@gmail.com>' ], time: { modified: '2017-01-19T15:51:35.841Z', created: '2014-06-18T03:56:05.284Z', '0.0.1': '2014-06-18T03:56:05.284Z', '0.0.2': '2014-06-18T05:35:27.107Z', '0.0.3': '2014-06-18T05:42:27.368Z', '0.0.4': '2014-06-18T05:49:43.994Z', '0.0.5': '2014-06-18T05:52:58.761Z', '0.0.6': '2014-06-19T05:21:03.197Z', '0.0.7': '2014-06-22T22:23:42.514Z', '0.0.8': '2014-06-23T00:22:46.231Z', '0.0.9': '2014-11-12T05:09:29.435Z', '0.1.0': '2014-11-12T05:14:26.594Z', '0.1.1': '2014-11-12T05:24:53.626Z', '0.1.2': '2014-12-19T15:56:15.972Z', '0.1.3': '2014-12-19T16:12:12.345Z', '0.1.4': '2015-01-12T22:09:19.752Z', '0.1.5': '2015-04-09T20:55:45.840Z', '0.1.6': '2015-06-22T14:47:17.993Z', '0.1.7': '2015-07-11T02:08:10.764Z', '0.1.8': '2015-07-11T02:31:39.325Z', '0.1.9': '2015-08-03T21:01:21.157Z', '0.2.0': '2015-09-30T13:38:50.084Z', '0.2.1': '2015-10-03T16:06:54.283Z', '0.2.2': '2015-12-13T22:54:54.170Z', '0.2.3': '2015-12-16T01:08:51.901Z', '0.2.4': '2016-07-15T15:12:03.110Z', '0.2.5': '2016-09-11T17:32:27.083Z', '0.2.6': '2016-10-24T14:35:28.131Z', '0.2.7': '2016-11-27T04:43:01.635Z', '0.2.8': '2017-01-19T15:51:35.841Z' }, homepage: 'https://github.com/rossmartin/cordova-uglify', keywords: [ 'cordova', 'uglify', 'minify', 'hook', 'hooks' ], author: 'Ross Martin', bugs: { url: 'https://github.com/rossmartin/cordova-uglify/issues' }, license: 'MIT', readmeFilename: 'README.md', repository: { type: 'git', url: 'git+https://github.com/rossmartin/cordova-uglify.git' }, users: { okmogwai: true, batosai: true, cnring18: true, gdevise: true, ebbm: true }, version: '0.2.8', dependencies: { 'clean-css': '3.4.8', 'ng-annotate': '0.15.4', shelljs: '^0.7.0', 'uglify-js': '2.4.14' }, scripts: { postinstall: 'node scripts/install.js', postuninstall: 'node scripts/uninstall.js' }, gitHead: '395775651d1b3a5f3625e4fa3d3ec2ca823664b1', dist: { shasum: '5aa9f69797da4c9d51c2f5184f4dcf0510d1b327', tarball: 'https://registry.npmjs.org/cordova-uglify/-/cordova-uglify-0.2.8.tgz' }, directories: {} } Sorry for the late response. Can you please provide the contents of your uglify-config.json? I think this issue is isolated to Windows. Hello, this is my used config: { "alwaysRun": true, "recursiveFolderSearch": true, "foldersToProcess": [ "js", "css", "img", "build" ], "uglifyJsOptions": { "compress": { "drop_console": true }, "fromString": true, "mangle": false }, "cleanCssOptions": { "noAdvanced": true, "keepSpecialComments": 0 } } It is odd that it is modifying files outside of your foldersToProcess - that is how it appears from your debug file. This behavior doesn't happen on OSX or Linux - I'll have to see if I can reproduce your issue on Windows. I'm not sure this is the same problem, but it may help: I had the same problem, ie .js files with 'undefined' content. After debugging uglify.js, the following line result = UglifyJS.minify (res.src, hookConfig.uglifyJsOptions); return a "result.code" undefined an overview before correction: result = UglifyJS.minify(res.src, hookConfig.uglifyJsOptions); console.log('result.code ' + result.code); fs.writeFileSync(file, result.code, 'utf8'); // overwrite the original unminified file break;` Correction of the line (removed parameter "hookConfig.uglifyJsOptions" ,fast correction, there is surely better) result = UglifyJS.minify(res.src); console.log('result.code ' + result.code); fs.writeFileSync(file, result.code, 'utf8'); // overwrite the original unminified file break; Why? Updates with npm have pushed new version (uglify and other), the uglifyJsOptions parameters did not match the version of UglifyJs ps: no more time to spend on this project, it works! @jeromeXoo Thank you for providing this info. I had some time this evening to look into this issue and it appears that the last PR introduced a few problems. I thought I had tested it after I merged but apparently I didn't have the latest commit. I've fixed the issue and published a new version (0.2.9) to npm. Let me know if you are still having issues. Hi, it crashes on first file: cordova -d prepare No scripts found for hook "before_prepare". Checking config.xml and package.json for saved platforms that haven't been added to the project Config.xml and package.json platforms are the same. No pkg.json modification. Package.json and config.xml platforms are different. Updating config.xml with most current list of platforms. Failed to require PlatformApi instance for platform "browser". Using polyfill instead. Checking config.xml for saved plugins that haven't been added to the project Checking for any plugins added to the project that have not been installed in browser platform No differences found between plugins added to project and installed in browser platform. Continuing... Generating config.xml from defaults for platform "browser" Executing script found in hooks directory for hook "after_prepare": hooks\after_prepare\uglify.js Running command: "C:\Program Files\nodejs\node.exe" D:\NetBeansProjects\ReturnReporting\hooks\after_prepare\uglify.js D:\NetBeansProjects\ReturnReporting minifying css file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\css\style.css uglifying js file D:\NetBeansProjects\ReturnReporting\platforms\browser\www\controller\App.controller.js fs.js:584 return binding.open(pathModule._makeLong(path), stringToFlags(flags), mode); ^ Error: ENOENT: no such file or directory, open 'D:\NetBeansProjects\ReturnReporting\sap.ui.define([ 'sap\ui\returnreporting\controller\BaseController', ... ... at Error (native) at Object.fs.openSync (fs.js:584:18) at Object.fs.readFileSync (fs.js:431:33) at addFile (D:\NetBeansProjects\ReturnReporting\node_modules\uglify-js\tools\node.js:70:22) at D:\NetBeansProjects\ReturnReporting\node_modules\uglify-js\tools\node.js:81:17 at Array.forEach (native) at Object.exports.minify (D:\NetBeansProjects\ReturnReporting\node_modules\uglify-js\tools\node.js:79:26) at compress (D:\NetBeansProjects\ReturnReporting\hooks\after_prepare\uglify.js:127:25) at D:\NetBeansProjects\ReturnReporting\hooks\after_prepare\uglify.js:94:11 at FSReqWrap.oncomplete (fs.js:82:15) Command finished with error code 1: C:\Program Files\nodejs\node.exe D:\NetBeansProjects\ReturnReporting\hooks\after_prepare\uglify.js,D:\NetBeansProjects\ReturnReporting Error: Hook failed with error code 1: D:\NetBeansProjects\ReturnReporting\hooks\after_prepare\uglify.js my config: { "alwaysRun": true, "recursiveFolderSearch": true, "foldersToProcess": [ "css", "controller", "model", "view" ], "uglifyJsOptions": { "compress": { "drop_console": true }, "mangle": false, "output": { "code": true } }, "cleanCssOptions": { "noAdvanced": true, "keepSpecialComments": 0 } } cordova uglify installed today by npm D:\NetBeansProjects\ReturnReporting>npm show cordova-uglify { name: 'cordova-uglify', description: 'Cordova hook that allows you to uglify or minify your apps JavaScript and CSS.', 'dist-tags': { latest: '0.2.9' }, versions: [ '0.0.1', '0.0.2', '0.0.3', '0.0.4', '0.0.5', '0.0.6', '0.0.7', '0.0.8', '0.0.9', '0.1.0', '0.1.1', '0.1.2', '0.1.3', '0.1.4', '0.1.5', '0.1.6', '0.1.7', '0.1.8', '0.1.9', '0.2.0', '0.2.1', '0.2.2', '0.2.3', '0.2.4', '0.2.5', '0.2.6', '0.2.7', '0.2.8', '0.2.9' ], maintainers: [ 'rossmartin <rmartin311@gmail.com>' ], time: { modified: '2017-09-22T02:55:35.360Z', created: '2014-06-18T03:56:05.284Z', '0.0.1': '2014-06-18T03:56:05.284Z', '0.0.2': '2014-06-18T05:35:27.107Z', '0.0.3': '2014-06-18T05:42:27.368Z', '0.0.4': '2014-06-18T05:49:43.994Z', '0.0.5': '2014-06-18T05:52:58.761Z', '0.0.6': '2014-06-19T05:21:03.197Z', '0.0.7': '2014-06-22T22:23:42.514Z', '0.0.8': '2014-06-23T00:22:46.231Z', '0.0.9': '2014-11-12T05:09:29.435Z', '0.1.0': '2014-11-12T05:14:26.594Z', '0.1.1': '2014-11-12T05:24:53.626Z', '0.1.2': '2014-12-19T15:56:15.972Z', '0.1.3': '2014-12-19T16:12:12.345Z', '0.1.4': '2015-01-12T22:09:19.752Z', '0.1.5': '2015-04-09T20:55:45.840Z', '0.1.6': '2015-06-22T14:47:17.993Z', '0.1.7': '2015-07-11T02:08:10.764Z', '0.1.8': '2015-07-11T02:31:39.325Z', '0.1.9': '2015-08-03T21:01:21.157Z', 10:42 '0.2.0': '2015-09-30T13:38:50.084Z', '0.2.1': '2015-10-03T16:06:54.283Z', '0.2.2': '2015-12-13T22:54:54.170Z', '0.2.3': '2015-12-16T01:08:51.901Z', '0.2.4': '2016-07-15T15:12:03.110Z', '0.2.5': '2016-09-11T17:32:27.083Z', '0.2.6': '2016-10-24T14:35:28.131Z', '0.2.7': '2016-11-27T04:43:01.635Z', '0.2.8': '2017-01-19T15:51:35.841Z', '0.2.9': '2017-09-22T02:55:35.360Z' }, homepage: 'https://github.com/rossmartin/cordova-uglify', keywords: [ 'cordova', 'uglify', 'minify', 'hook', 'hooks' ], author: 'Ross Martin', bugs: { url: 'https://github.com/rossmartin/cordova-uglify/issues' }, license: 'MIT', readmeFilename: 'README.md', repository: { type: 'git', url: 'git+https://github.com/rossmartin/cordova-uglify.git' }, users: { okmogwai: true, cnring18: true, gdevise: true, ebbm: true }, version: '0.2.9', dependencies: { 'clean-css': '3.4.8', 'ng-annotate': '0.15.4', shelljs: '^0.7.0', 'uglify-js': '3.1.1' }, scripts: { postinstall: 'node scripts/install.js', postuninstall: 'node scripts/uninstall.js' }, gitHead: 'e101c6e9f673e43f2431d2a238c4fdbf301870b2', dist: { shasum: 'af25bcd17b1687d91d60856ad317590901eb3565', tarball: 'https://registry.npmjs.org/cordova-uglify/-/cordova-uglify-0.2.9.tgz' }, directories: {} }``` @rchovan Thanks for reporting the issue. I have a theatre PC with Windows on it I'll try it this weekend. I'm not experiencing any issues on Ubuntu or Mac OS. I wonder if the issue is because the project is on a separate drive than node/windows. @rossmartin I'm experiencing the same issue where the JS output files only contain "undefined". I am running the latest version of cordova-uglify 0.3.0 as reported by npm and node v8.9.1 on Windows 10. @koga73 Thanks for letting me know. Unfortunately my theatre PC that had Windows 10 on it died! Bad motherboard or power supply I think but I don't plan on reviving it. Recently I installed 2 Windows VMs to an external HDD on my Macbook for my day job to test IE11 issues. I'll have to use a VM to test this out and get it fixed. I haven't forgotten about this just haven't had time - it's frustrating because it works fine on Linux and Mac OS. @rossmartin Did some digging. Turns out your plugin is fine. Issue is we had an old uglify-config.json. Specifically this line was causing it: "uglifyJsOptions": { "fromString": true } @koga73 Thanks a bunch for letting me know. I wonder if this is the issue you are having @rchovan ? I could update the install script to overwrite the uglify-config.json but I think a better solution is to just document it in the readme and output after installing the package. I don't want anyone to lose their config on update. I have this error too. Using default options from README doesn't help. Hi all, started experiencing the same issue. What is generating the empty file in my case is the following code: BackgroundGeolocation.on('location', (location) => { // handle your locations here // to perform long running operation on iOS // you need to create background task BackgroundGeolocation.startTask(taskKey => { // execute long running task // eg. ajax post location // IMPORTANT: task has to be ended by endTask BackgroundGeolocation.endTask(taskKey); }); }); If I remove it then it works properly Running uglijs manually, I get BackgroundGeolocation.on('location', (position) => { ERROR: Unexpected token: operator (>) How could I fix it considering is legitimate js code? Thanks I'm trying https://github.com/mishoo/UglifyJS2/tree/harmony And it seems to solve the issue. In uglify hook I changed this line: var UglifyJS = require(path.join(dependencyPath, 'uglify-js')); to var UglifyJS = require(path.join(dependencyPath, 'uglify-es')); So far seems ok, will test more extensively @luigi37 I think you may have revealed why others are experiencing this same issue. uglify-js does not work with ES6. I think most transpile their "modern" JS to ES5 before feeding it to uglify-js. It appears that uglify-es might be the way to go moving forward to handle minifying code if it is ES6. There is also this interesting compress option with uglify-es - ecma (default: 5) -- Pass 6 or greater to enable compress options that will transform ES5 code into smaller ES6+ equivalent forms. Thanks for bringing this up @luigi37. I'm going to give it a go and see how it works out. If it works with the old and new stuff I'll publish a new version today. Cool, thanks! @luigi37 I published a new version (0.3.1) that uses uglify-es. I tried the latest code with some old projects that were plain ES5 and a new one with ES6 - I had no issues. I'll be leaving this issue open for now. @luigi37 and anyone else, please see #39 about the decision to abandon uglify-es in https://github.com/rossmartin/cordova-uglify/commit/eeb41a851fbf23b44439485885394d9e243fce9c (v0.3.2) Make sure your cordova project's code is transpiled to ES5 (I think you should even if you don't use this hook). @rossmartin thank you. I'm not familiar with ES6 enough to confirm if transpiling back to ES5 can affect the functionality. Could you please confirm? If so, I'm happy with your solution. Thanks a lot! Luigi @luigi37 if you use babel to transpile your ES6 to ES5 it won't affect functionality. Not all mobile device webviews support ES6 so it would be much safer to transpile. Thank you. Is there a way to include the transpile into cordova-uglify to avoid multiple steps? Thanks Luigi @luigi37 That would be a nice feature to have if you're using something like framework 7 & react without using react-scripts. Those that use Ionic it won't be of any benefit. I'll think about it - this may be a better candidate for a separate hook. Thanks Ross, it would be much appreciated. Specifically I'm using Cordova with Framework7... Anyway it's a specific cordova plugin which I use which brought me to the ES6 thing... Luigi ----------------------------------- Luigi Paiella luigi@pajella.com luigi@pajella.com**----------------------------------- 2018-04-15 18:25 GMT+02:00 Ross Martin notifications@github.com: @luigi37 https://github.com/luigi37 That would be a nice feature to have if you're using something like framework 7 & react without using react-scripts. Those that use Ionic it won't be of any benefit. I'll think about it - this may be a better candidate for a separate hook. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/rossmartin/cordova-uglify/issues/35#issuecomment-381419013, or mute the thread https://github.com/notifications/unsubscribe-auth/ABYb8VG7FS5SzsF9CSELk0fACMVn3isDks5to3R5gaJpZM4PCQEX . Thanks Ross, it would be much appreciated. Specifically I'm using Cordova with Framework7... Anyway it's a specific cordova plugin which I use which brought me to the ES6 thing... Luigi @rossmartin Hello! The same issue. Works great with uglify-es but with uglify-js one file always has undefined content. This same issue is happening to me too, however its only happening to three files, all three files are 9 bytes and all say undefined. I am running an Ionic v1 app, so know Ionic v2 or ES issues to worry about. Just straight old AngularJS and JS. I'm having the same issue, 1 file only has "undefined" and I'm not on Windows, please read more in #46
gharchive/issue
2017-08-25T05:48:56
2025-04-01T04:35:44.840641
{ "authors": [ "iliakonnov", "jeromeXoo", "koga73", "luigi37", "rchovan", "rodrigograca31", "rolinger", "rossmartin", "zankevich" ], "repo": "rossmartin/cordova-uglify", "url": "https://github.com/rossmartin/cordova-uglify/issues/35", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
118394506
server rendering error I am getting error when trying to render on the server: Warning: React can't find the root component node for data-reactid value .ba9ic960ao.1.2. If you're seeing this message, it probably means that you've loaded two copies of React on the page. At this time, only a single copy of React can be loaded at a time. TypeError: Cannot read property 'firstChild' of undefined at Object.ReactMount.findComponentRoot ...(and so on) (Here I created repo to reproduce this issue: https://github.com/olessavluk/r5bug ) I'm using node v5.0.0 & npm v3.4.1. And I have only one veriosn of React installed, because find ./ -name 'react' returns only one result .//node_modules/react. So I think problem in this file https://github.com/router5/react-router5/blob/master/modules/routeNode.js#L21. Because you are adding listener in constructor and removing it in componentWillUnmout, but unmount never gets executed on the server. I am quite inexperienced with universal applications so thank you for raising potential issues. I cloned and ran your example, the error goes away by removing the input element in Timer component. Not sure what that means. I have created new branch (logall) and put logs everywhere to see what is actually going on. As you can see, listener executes setState after rendering is complete and this actually causes an error. :tada: I think I managed to fix it. The problem is not with componentWillUnmount but with suff in component constructor (componentWillMount) which should be in componentDidMount. I will commit a fix and release it later tonight. You are talking about this ?) - https://github.com/olessavluk/react-router5/commit/12aad1612002f7d5380c60370e9dbeaffa55376e Yes, exactly. I also moved the check for the listeners plugin to component did mount so you are not forced to use it on the server side. 1.0.1 has been published Thank you very much for raising that bug and taking the time to create a repo. I really appreciate. :+1: Linking it to other similar issues, in case that helps others. RickWong/react-isomorphic-starterkit#77 denvned/isomorphic-relay-router#5
gharchive/issue
2015-11-23T14:39:28
2025-04-01T04:35:44.892908
{ "authors": [ "olessavluk", "troch" ], "repo": "router5/react-router5", "url": "https://github.com/router5/react-router5/issues/11", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
4447111
support COPY commands EM versions of be get_copy_data and put_copy_data are missing. I'm really interested by this feature. I was trying to use the "copy_table" method from sequel (via em-pg-sequel) in order to create a fast export to CSV and I was wondering why it was working great in the console but not in EM (ruby process hanging at 100% CPU)... I willing to help but I have no idea where to start and how difficult it coud be...
gharchive/issue
2012-05-06T22:49:57
2025-04-01T04:35:44.894454
{ "authors": [ "flop", "royaltm" ], "repo": "royaltm/ruby-em-pg-client", "url": "https://github.com/royaltm/ruby-em-pg-client/issues/4", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
192587020
Added GridFS support JAVARS-23 OK moved all implementation code into the internal package. The following internal classes are now public: MongoClientImpl - so can be constructed from MongoClients MongoDatabaseImpl - so can be accessed via GridFSBuckets GridFSBucketImpl - so can be constructed from GridFSBuckets GridFSAsyncStreamHelper - so can be accessed via AsyncStreamHelper If you haven't already, make sure you add an exclude for the internal package when generating javadoc.
gharchive/pull-request
2016-11-30T14:58:15
2025-04-01T04:35:44.903902
{ "authors": [ "jyemin", "rozza" ], "repo": "rozza/mongo-java-driver-reactivestreams", "url": "https://github.com/rozza/mongo-java-driver-reactivestreams/pull/7", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1694480392
🛑 Jay Industries is down In a78c3ec, Jay Industries (https://jaybrassindustries.com/) was down: HTTP code: 404 Response time: 17748 ms Resolved: Jay Industries is back up in 66ebce4.
gharchive/issue
2023-05-03T17:13:54
2025-04-01T04:35:44.906502
{ "authors": [ "rpharaniya" ], "repo": "rpharaniya/websites-uptime-monitor", "url": "https://github.com/rpharaniya/websites-uptime-monitor/issues/1036", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2102087986
🛑 Jay Industries is down In 1dd54f5, Jay Industries (https://jaybrassindustries.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Jay Industries is back up in 3336b9d after 5 minutes.
gharchive/issue
2024-01-26T11:47:53
2025-04-01T04:35:44.908879
{ "authors": [ "rpharaniya" ], "repo": "rpharaniya/websites-uptime-monitor", "url": "https://github.com/rpharaniya/websites-uptime-monitor/issues/1734", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2447220554
🛑 Silico Rubber Polymers is down In 6d437c2, Silico Rubber Polymers (https://silicorubberpolymers.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Silico Rubber Polymers is back up in 4fc285e after 15 minutes.
gharchive/issue
2024-08-04T17:07:02
2025-04-01T04:35:44.911240
{ "authors": [ "rpharaniya" ], "repo": "rpharaniya/websites-uptime-monitor", "url": "https://github.com/rpharaniya/websites-uptime-monitor/issues/2035", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1175018936
How to train the model? Can you please tell me how to train the model onmy own dataset ? ok,thanks.
gharchive/issue
2022-03-21T08:02:45
2025-04-01T04:35:44.957111
{ "authors": [ "zhw2020913" ], "repo": "rpm1412/LUS_Video_Summarization", "url": "https://github.com/rpm1412/LUS_Video_Summarization/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
833667577
Build error when .NET 6 preview is installed. Any suggestions to correct this is appreciated. We get the following error when Visual Studio 2019 Preview is installed on our workstation. We tried using global.json to limit to only released versions, but no go. Any thoughts on how to fix this? global.json { "sdk": { "allowPrerelease": false } } Change your framework target to a supported one, like "net5.0" Get Outlook for Androidhttps://aka.ms/AAb9ysg From: Richard Crane @.> Sent: Wednesday, March 17, 2021 12:21:39 PM To: rr-wfm/MSBuild.Sdk.SqlProj @.> Cc: Subscribed @.***> Subject: [rr-wfm/MSBuild.Sdk.SqlProj] Build error when .NET 6 preview is installed. Any suggestions to correct this is appreciated. (#125) We get the following error when Visual Studio 2019 Preview is installed on our workstation. We tried using global.json to limit to only released versions, but no go. Any thoughts on how to fix this? global.json { "sdk": { "allowPrerelease": false } } [image]https://user-images.githubusercontent.com/15870558/111459469-16a3b580-86f1-11eb-9827-5042fda3c946.png — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHubhttps://github.com/rr-wfm/MSBuild.Sdk.SqlProj/issues/125, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AA7Z3YZRGCSPUBPSAG7TQ7LTECGEHANCNFSM4ZKMP4KQ. @ChiefInnovator I would expect that global.json to fix this issue. Did you restart Visual Studio after adding the global.json? I believe it doesn't immediately take effect. If that doesn't work, could you share the output of dotnet --info running from the command line in the folder of your project? I had a similar issue but with "netcoreapp5.0" Adding the global.json worked for me. Example of the content: { "sdk": { "version": "3.1.407" } } Find out more here about global.json @ratynchukD As of version 1.6.0 there should be support for .NET 5 so it would work with any SDK version 5.0. If that is not the case we would like to understand why it doesn't work for you. We don't support .NET 6 previews just yet. Since .NET 6 is still early in the preview phase I don't expect many people will have an issue with this, but please upvote this issue if you feel this is important to you. @jmezach This needs to be discussed in my team since we are using the 1.2.0 version. I'm new to the project and might be the first who faced with the issue. Regarding .NET 6 preview - I don't feel this is important to my team or me. Thank you! I don't think there's much more we can do here so I'll go ahead and close this. If you feel there's still an issue here feel free to re-open.
gharchive/issue
2021-03-17T11:21:20
2025-04-01T04:35:44.978670
{ "authors": [ "ChiefInnovator", "ErikEJ", "jmezach", "ratynchukD" ], "repo": "rr-wfm/MSBuild.Sdk.SqlProj", "url": "https://github.com/rr-wfm/MSBuild.Sdk.SqlProj/issues/125", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
645959307
Stack Overflow when use Consumer in dialog Future<void> _showDialogAddGhiChu(BuildContext context) async { return showDialog<void>( context: context, barrierDismissible: false, // user must tap button! builder: (_) { return Consumer((ct, read) { final clickDate = read(mainStateNotifier).clickDate; return AddNote( date: clickDate, context: ct, ); //magic ^_^ }); }, ); } error Another exception was thrown: Stack Overflow ════════ Exception caught by widgets library ═══════════════════════════════════════════════════════ The following StackOverflowError was thrown building Consumer(dirty, dependencies: [ProviderStateOwnerScope], state: _ConsumerState#34eeb): Stack Overflow The relevant error-causing widget was: Consumer file:///Users/tbm98/dev/flutter_app/student-social/lib/presentation/screens/main/main_page.dart:132:16 When the exception was thrown, this was the stack: #0 MainState.getClickDate (package:studentsocial/presentation/screens/main/main_state.dart:35:3) #1 MainState.getClickDate (package:studentsocial/presentation/screens/main/main_state.dart:35:32) #2 MainState.getClickDate (package:studentsocial/presentation/screens/main/main_state.dart:35:32) #3 MainState.getClickDate (package:studentsocial/presentation/screens/main/main_state.dart:35:32) #4 MainState.getClickDate (package:studentsocial/presentation/screens/main/main_state.dart:35:32) ... ════════════════════════════════════════════════════════════════════════════════════════════════════ sorry. my getter called itself
gharchive/issue
2020-06-26T02:18:42
2025-04-01T04:35:44.998973
{ "authors": [ "tbm98" ], "repo": "rrousselGit/river_pod", "url": "https://github.com/rrousselGit/river_pod/issues/15", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1590822366
Warn against AsyncNotifier.update misuses class Example extends AsyncNotifier<int> { ... void fn() { state = AsyncLoading(); // KO, "update" will likely never complete update((data) => ...); } void fn() { update((data) { state = AsyncLoading(); // OK ... }); } void fn() { update((unused) => ...); // the parameter should be used. Otherwise use AsyncValue.guard } } @rrousselGit I am trying to update the state of a FamilyAsyncNotifier partially, I just want to know whether I am doing it right. Also in this case of updating a asyncvalue partially, I want to know what's differences in using whenData, AsyncGuard.value and update method . It would be a immense help, if you provide some clarity on this. update((state) async { List students = await ref.read(collegeProvider).getCollegeById(id: this.arg); return state.copyWith(students: students); }); With mutations, this is really low-priority. Closing
gharchive/issue
2023-02-19T19:29:07
2025-04-01T04:35:45.001282
{ "authors": [ "rrousselGit", "yknevenky" ], "repo": "rrousselGit/riverpod", "url": "https://github.com/rrousselGit/riverpod/issues/2205", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
803344574
rmfuse on OSX Hi, I really love the idea of rmfuse and I would love to use it on my Mac. It seems that that pyfuse3 is not supported on OSX: https://github.com/osxfuse/osxfuse/issues/390 https://github.com/libfuse/pyfuse3/issues/29 This is the error I get when doing pip install rmfuse Collecting rmfuse Using cached rmfuse-0.1.1-py3-none-any.whl (8.7 kB) Collecting rmcl<0.4.0,>=0.3.0 Using cached rmcl-0.3.0-py3-none-any.whl (16 kB) Collecting rmrl<0.2.0,>=0.1.2 Using cached rmrl-0.1.2-py3-none-any.whl (518 kB) Collecting pyfuse3<4.0.0,>=3.2.0 Using cached pyfuse3-3.2.0.tar.gz (465 kB) ERROR: Command errored out with exit status 2: command: /Users/schneider/anaconda3/bin/python3.8 -c 'import sys, setuptools, tokenize; sys.argv[0] = '"'"'/private/var/folders/d_/755ltxq50klbwlljp6rsqgzc0000gn/T/pip-install-9z2bjy84/pyfuse3/setup.py'"'"'; __file__='"'"'/private/var/folders/d_/755ltxq50klbwlljp6rsqgzc0000gn/T/pip-install-9z2bjy84/pyfuse3/setup.py'"'"';f=getattr(tokenize, '"'"'open'"'"', open)(__file__);code=f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' egg_info --egg-base /private/var/folders/d_/755ltxq50klbwlljp6rsqgzc0000gn/T/pip-pip-egg-info-9ju778vx cwd: /private/var/folders/d_/755ltxq50klbwlljp6rsqgzc0000gn/T/pip-install-9z2bjy84/pyfuse3/ Complete output (4 lines): Package fuse3 was not found in the pkg-config search path. Perhaps you should add the directory containing `fuse3.pc' to the PKG_CONFIG_PATH environment variable No package 'fuse3' found ---------------------------------------- ERROR: Command errored out with exit status 2: python setup.py egg_info Check the logs for full command output. Urgh. I was afraid of something like this. I was hoping that a version number > 3.0 on MacFuse indicated that it supported FUSE3, but apparently not. From those issues, I don't hold much hope of MacFuse ever making it to FUSE3. I don't think there's anything I'm doing that requires FUSE3 specifically; that requirement is coming from the fact that I chose to use pyfuse3, which seemed to be the most active and best documented library. There were a bunch of other bindings I found out there, but the only other one I seriously considered was python-fuse. This is for FUSE2, but I don't know whether it works with MacFuse. If someone would like to check and let me know, we could see if converting to the python-fuse library makes sense. (A simple test would be to pip install fuse-python and then see if this example works.) All that said, this particular error was because the pyfuse3 module needs to be compiled during installation, and it couldn't find the fuse3 headers. The fuse-python module is the same way, so I don't know if this would also hit the same problem. Yes I think libfuse2 should work with osx. I tested the example script that you provided (after I successfully installed fuse-python). I get an error, fuse: no mount point Traceback (most recent call last): File "hello.py", line 91, in <module> main() File "hello.py", line 88, in main server.main() File "/Users/schneider/anaconda3/lib/python3.8/site-packages/fuse.py", line 754, in main main(**d) fuse.FuseError: filesystem initialization failed which seems to be related to the example script, rather than fuse itself. However, since import fuse from fuse import Fuse works. I have no doubt that using libfuse2 would fix the problem for OSX. However, I suppose that libfuse3 is more stable.... Its a pity that macfuse did not update libfuse.... Thanks for checking. It appears that pyfuse3 forked off from python-llfuse some time ago. That project is now marked as undeveloped, but the API is still pretty similar. It has some hints on building solutions compatible with both libraries. I'll take a look and see if that's feasible here. Good news -- it seems to work. See #6 for a branch that implements support for llfuse. If you could try it out and let me know if it works with MacFuse, it'd be much appreciated. You'll need llfuse (pip install llfuse). If you're working from git, you can just pull the branch from #6. If not, you should be able to pip install the attached sdist file. rmfuse-0.1.2a0.tar.gz That sounds awesome! I will test it today! It installed successfully! These steps were necessary: brew install macfuse One needs to allow a kernel extension and reboot. brew install pkg-config I messed up the onetime code and now get the error: FileNotFoundError: [Errno 2] No such file or directory: '/Users/schneider/.cache/rmfuse' Seems to be unrelated to rmfuse. That is a bug in RMfuse--it stores file metadata in a cache database to avoid hitting the API too often. But I didn't make sure the ~/.cache directory existed before trying to create the db file. I'll get a fix in for that soon, but in the meantime, you ought to be able to fix this with a mkdir ~/.cache on the command line. thanks for getting back! I started rmfuse: (base) schneider@Aarons-MBP ~ % rmfuse remarkable (base) schneider@Aarons-MBP ~ % cd remarkable (base) schneider@Aarons-MBP . % ls ls: .: Function not implemented (base) schneider@Aarons-MBP . % open . 2021-02-14 12:52:14.104 open[1833:36366] CFURLCopyResourcePropertyForKey failed because it was passed an URL which has no scheme No application knows how to open . (Error Domain=NSOSStatusErrorDomain Code=-10814 "kLSApplicationNotFoundErr: E.g. no application claims the file" UserInfo={_LSLine=1483, _LSFunction=runEvaluator}). (base) schneider@Aarons-MBP . % cp ~/Desktop/Chapter_3.pdf ~/remarkable cp: /Users/schneider/remarkable: Function not implemented Is there a way to renew the authentication token? I tried to remove the access on my.remarkable.com and I tried to freshly install rmfuse. But now I can't link it again. (base) schneider@Aarons-MBP ~ % rmfuse remarkable Traceback (most recent call last): File "/usr/local/bin/rmfuse", line 8, in <module> sys.exit(main()) File "/usr/local/lib/python3.9/site-packages/rmfuse/fuse.py", line 452, in main fuse.main(workers=1) File "src/fuse_api.pxi", line 327, in llfuse.main File "src/handlers.pxi", line 80, in llfuse.fuse_getattr File "src/handlers.pxi", line 81, in llfuse.fuse_getattr File "/usr/local/lib/python3.9/site-packages/rmfuse/fuselib.py", line 25, in decorated return trio.run(runner) File "/usr/local/lib/python3.9/site-packages/trio/_core/_run.py", line 1932, in run raise runner.main_task_outcome.error File "/usr/local/lib/python3.9/site-packages/rmfuse/fuselib.py", line 24, in runner return await afunc(*args, **kw) File "/usr/local/lib/python3.9/site-packages/rmfuse/fuse.py", line 226, in getattr return await self._getattr(inode, ctx) File "/usr/local/lib/python3.9/site-packages/rmfuse/fuse.py", line 195, in _getattr item = await self.get_by_id(self.get_id(inode)) File "/usr/local/lib/python3.9/site-packages/rmfuse/fuse.py", line 139, in get_by_id return await Item.get_by_id(id_) File "/usr/local/lib/python3.9/site-packages/rmcl/items.py", line 46, in get_by_id return await (await api.get_client()).get_by_id(id_) File "/usr/local/lib/python3.9/site-packages/rmcl/api.py", line 245, in get_by_id await self.update_items() File "/usr/local/lib/python3.9/site-packages/rmcl/api.py", line 212, in update_items response = await self.request('GET', '/document-storage/json/2/docs') File "/usr/local/lib/python3.9/site-packages/rmcl/api.py", line 117, in request await self.renew_token() File "/usr/local/lib/python3.9/site-packages/rmcl/api.py", line 208, in renew_token raise AuthError("Can't renew token: {e}".format( rmcl.exceptions.AuthError: Can't renew token: 401 Second question first: the tokens are stored in ~/.config/rmcl/config.json. Delete that file, and you'll be back to pristine, as far as the API is concerned. Then you can get a new one-time code and get access again. As for the first problem: it appears that your system is trying to call a filesystem method that isn't implemented in RMfuse yet. (There are quite a few, and I implemented them as I found they were needed in testing.) It's not clear to me which one it is, but if you run rmfuse with the -vv output, you'll get a whole lot of debugging information, including each function being called. If you can capture that output and post it here, it'll help narrow down what the problem is. Thanks! Thanks! Here we go: (rmfuse) schneider@Aarons-MacBook-Pro ~ % rmfuse -vv remarkable FUSE library version: 2.9.9 unique: 2, opcode: INIT (26), nodeid: 0, insize: 56, pid: 5768 INIT: 7.19 flags=0xee000008 max_readahead=0x00100000 INIT: 7.19 flags=0x00000020 max_readahead=0x00100000 max_write=0x02000000 max_background=0 congestion_threshold=0 unique: 2, success, outsize: 40 unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 0 unique: 2, error: -78 (Function not implemented), outsize: 16 unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5768 unique: 3, error: -78 (Function not implemented), outsize: 16 unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 92 unique: 2, error: -78 (Function not implemented), outsize: 16 unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 130 unique: 3, error: -78 (Function not implemented), outsize: 16 unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 130 unique: 2, error: -78 (Function not implemented), outsize: 16 unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 130 unique: 3, error: -78 (Function not implemented), outsize: 16 unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 130 unique: 2, error: -78 (Function not implemented), outsize: 16 unique: 3, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 89 unique: 3, success, outsize: 136 unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5340 unique: 2, error: -78 (Function not implemented), outsize: 16 unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 516 unique: 3, error: -78 (Function not implemented), outsize: 16 unique: 4, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 4145 unique: 4, error: -78 (Function not implemented), outsize: 16 unique: 5, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 5, error: -78 (Function not implemented), outsize: 16 unique: 6, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 141 unique: 6, error: -78 (Function not implemented), outsize: 16 unique: 7, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 549 unique: 7, error: -78 (Function not implemented), outsize: 16 unique: 8, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 425 unique: 8, error: -78 (Function not implemented), outsize: 16 unique: 9, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 487 unique: 9, error: -78 (Function not implemented), outsize: 16 unique: 10, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 467 unique: 10, error: -78 (Function not implemented), outsize: 16 unique: 11, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 405 unique: 11, error: -78 (Function not implemented), outsize: 16 unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 407 unique: 2, error: -78 (Function not implemented), outsize: 16 unique: 12, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 545 unique: 12, error: -78 (Function not implemented), outsize: 16 unique: 13, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 442 unique: 13, error: -78 (Function not implemented), outsize: 16 unique: 14, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 531 unique: 14, error: -78 (Function not implemented), outsize: 16 unique: 15, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 15, error: -78 (Function not implemented), outsize: 16 unique: 16, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 89 unique: 16, error: -78 (Function not implemented), outsize: 16 unique: 17, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 387 unique: 17, error: -78 (Function not implemented), outsize: 16 unique: 18, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 18, error: -78 (Function not implemented), outsize: 16 unique: 19, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 505 unique: 19, error: -78 (Function not implemented), outsize: 16 unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 391 unique: 3, error: -78 (Function not implemented), outsize: 16 unique: 4, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5755 unique: 4, error: -78 (Function not implemented), outsize: 16 unique: 5, opcode: GETXATTR (22), nodeid: 1, insize: 77, pid: 408 unique: 5, error: -78 (Function not implemented), outsize: 16 unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 6, error: -2 (No such file or directory), outsize: 16 unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 387 unique: 7, error: -2 (No such file or directory), outsize: 16 unique: 8, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 8, error: -78 (Function not implemented), outsize: 16 unique: 9, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 9, error: -78 (Function not implemented), outsize: 16 unique: 10, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 10, error: -78 (Function not implemented), outsize: 16 unique: 11, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 11, error: -78 (Function not implemented), outsize: 16 unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 2, error: -78 (Function not implemented), outsize: 16 unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 387 unique: 12, error: -2 (No such file or directory), outsize: 16 unique: 13, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 13, error: -78 (Function not implemented), outsize: 16 unique: 14, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 387 unique: 14, error: -78 (Function not implemented), outsize: 16 unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 15, error: -2 (No such file or directory), outsize: 16 unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 16, error: -2 (No such file or directory), outsize: 16 unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 17, error: -2 (No such file or directory), outsize: 16 unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 18, error: -2 (No such file or directory), outsize: 16 unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 19, error: -2 (No such file or directory), outsize: 16 unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5771 unique: 3, error: -78 (Function not implemented), outsize: 16 unique: 4, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5771 unique: 4, error: -78 (Function not implemented), outsize: 16 unique: 5, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5771 unique: 5, error: -78 (Function not implemented), outsize: 16 unique: 6, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5771 unique: 6, error: -78 (Function not implemented), outsize: 16 unique: 7, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5774 unique: 7, error: -78 (Function not implemented), outsize: 16 unique: 8, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5774 unique: 8, error: -78 (Function not implemented), outsize: 16 unique: 9, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5774 unique: 9, error: -78 (Function not implemented), outsize: 16 unique: 10, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5774 unique: 10, error: -78 (Function not implemented), outsize: 16 unique: 11, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5776 unique: 11, error: -78 (Function not implemented), outsize: 16 unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5776 unique: 2, error: -78 (Function not implemented), outsize: 16 unique: 12, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5776 unique: 12, error: -78 (Function not implemented), outsize: 16 unique: 13, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5776 unique: 13, error: -78 (Function not implemented), outsize: 16 unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 5777 unique: 14, error: -2 (No such file or directory), outsize: 16 unique: 15, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777 unique: 15, error: -78 (Function not implemented), outsize: 16 unique: 16, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777 unique: 16, error: -78 (Function not implemented), outsize: 16 unique: 17, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777 unique: 17, error: -78 (Function not implemented), outsize: 16 unique: 18, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777 unique: 18, error: -78 (Function not implemented), outsize: 16 unique: 19, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777 unique: 19, error: -78 (Function not implemented), outsize: 16 unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 5777 unique: 3, error: -2 (No such file or directory), outsize: 16 unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 5777 unique: 4, error: -2 (No such file or directory), outsize: 16 unique: 5, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777 unique: 5, error: -78 (Function not implemented), outsize: 16 unique: 6, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777 unique: 6, error: -78 (Function not implemented), outsize: 16 unique: 7, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777 unique: 7, error: -78 (Function not implemented), outsize: 16 unique: 8, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777 unique: 8, error: -78 (Function not implemented), outsize: 16 unique: 9, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 5777 unique: 9, error: -78 (Function not implemented), outsize: 16 unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 442 unique: 10, error: -2 (No such file or directory), outsize: 16 I see a lot of failed calls to statfs, which hadn't been implemented. I've added a simple implementation in #6. Can you give it a try. Updated sdist file attached, if that helps. rmfuse-0.1.2a0.tar.gz Awesome! it works for copying stuff from the fuse drive to the local machine. But right now its giving me an error, when I try to copy a pdf to the fuse drive. Here is a new log: (rmfuse) schneider@Aarons-MacBook-Pro ~ % rmfuse remarkable -vv FUSE library version: 2.9.9 unique: 2, opcode: INIT (26), nodeid: 0, insize: 56, pid: 65083 INIT: 7.19 flags=0xee000008 max_readahead=0x00100000 INIT: 7.19 flags=0x00000020 max_readahead=0x00100000 max_write=0x02000000 max_background=0 congestion_threshold=0 unique: 2, success, outsize: 40 unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 65083 unique: 3, success, outsize: 96 unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 0 unique: 2, success, outsize: 96 unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 92 unique: 3, success, outsize: 96 unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 130 unique: 2, success, outsize: 96 unique: 3, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 89 unique: 3, success, outsize: 136 unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 130 unique: 2, success, outsize: 96 unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 130 unique: 3, success, outsize: 96 unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 2, success, outsize: 96 unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 3, success, outsize: 96 unique: 4, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 405 unique: 4, success, outsize: 96 unique: 5, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 5, success, outsize: 96 unique: 6, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 407 unique: 6, success, outsize: 96 unique: 7, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 442 unique: 7, success, outsize: 96 unique: 8, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 467 unique: 8, success, outsize: 96 unique: 9, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 425 unique: 9, success, outsize: 96 unique: 10, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 141 unique: 10, success, outsize: 96 unique: 11, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 8572 unique: 11, success, outsize: 96 unique: 12, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 545 unique: 12, success, outsize: 96 unique: 13, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 91827 unique: 13, success, outsize: 96 unique: 14, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 516 unique: 14, success, outsize: 96 unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 549 unique: 2, success, outsize: 96 unique: 15, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 89 unique: 15, success, outsize: 96 unique: 16, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 8568 unique: 16, success, outsize: 96 unique: 17, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 8568 unique: 17, success, outsize: 96 unique: 18, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 64910 unique: 18, success, outsize: 96 unique: 19, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 391 unique: 19, success, outsize: 96 unique: 20, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 64914 unique: 20, success, outsize: 96 unique: 21, opcode: GETXATTR (22), nodeid: 1, insize: 77, pid: 408 unique: 21, error: -78 (Function not implemented), outsize: 16 unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 7591 unique: 3, success, outsize: 96 unique: 4, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 7570 unique: 4, success, outsize: 96 unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 5, error: -2 (No such file or directory), outsize: 16 unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 387 unique: 6, error: -2 (No such file or directory), outsize: 16 unique: 7, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 7, success, outsize: 96 unique: 8, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 8, success, outsize: 96 unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 387 unique: 9, error: -2 (No such file or directory), outsize: 16 unique: 10, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 10, success, outsize: 96 unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 442 unique: 11, error: -2 (No such file or directory), outsize: 16 unique: 12, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 12, success, outsize: 96 unique: 13, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 13, success, outsize: 96 unique: 14, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 387 unique: 14, success, outsize: 96 unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 2, success, outsize: 96 unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 15, error: -2 (No such file or directory), outsize: 16 unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 16, error: -2 (No such file or directory), outsize: 16 unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 17, error: -2 (No such file or directory), outsize: 16 unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 18, error: -2 (No such file or directory), outsize: 16 unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 19, error: -2 (No such file or directory), outsize: 16 unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 20, error: -2 (No such file or directory), outsize: 16 unique: 21, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 21, error: -2 (No such file or directory), outsize: 16 unique: 3, opcode: OPENDIR (27), nodeid: 1, insize: 48, pid: 408 unique: 3, success, outsize: 32 unique: 4, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 4, success, outsize: 576 unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 46, pid: 408 unique: 5, success, outsize: 160 unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 6, error: -2 (No such file or directory), outsize: 16 unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 7, error: -2 (No such file or directory), outsize: 16 unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 8, error: -2 (No such file or directory), outsize: 16 unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 9, error: -2 (No such file or directory), outsize: 16 unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 47, pid: 408 unique: 10, success, outsize: 160 unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 11, error: -2 (No such file or directory), outsize: 16 unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408 unique: 12, error: -2 (No such file or directory), outsize: 16 unique: 13, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408 unique: 13, success, outsize: 136 unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 46, pid: 408 unique: 14, success, outsize: 160 unique: 2, opcode: CREATE (35), nodeid: 1, insize: 66, pid: 408 unique: 2, success, outsize: 176 unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 15, error: -2 (No such file or directory), outsize: 16 unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408 unique: 16, error: -2 (No such file or directory), outsize: 16 unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 17, success, outsize: 160 unique: 18, opcode: FLUSH (25), nodeid: 16, insize: 64, pid: 408 unique: 18, error: -78 (Function not implemented), outsize: 16 unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 19, error: -2 (No such file or directory), outsize: 16 unique: 20, opcode: RELEASE (18), nodeid: 16, insize: 64, pid: 408 ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file unique: 20, error: -5 (Input/output error), outsize: 16 unique: 21, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408 unique: 21, success, outsize: 160 unique: 3, opcode: GETATTR (3), nodeid: 16, insize: 56, pid: 408 unique: 3, success, outsize: 136 unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 4, error: -2 (No such file or directory), outsize: 16 unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408 unique: 5, error: -2 (No such file or directory), outsize: 16 unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 6, error: -2 (No such file or directory), outsize: 16 unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408 unique: 7, error: -2 (No such file or directory), outsize: 16 unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 57, pid: 408 unique: 8, success, outsize: 160 unique: 9, opcode: SETXATTR (21), nodeid: 16, insize: 109, pid: 408 unique: 9, error: -78 (Function not implemented), outsize: 16 unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408 unique: 10, error: -2 (No such file or directory), outsize: 16 unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408 unique: 11, error: -2 (No such file or directory), outsize: 16 unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 46, pid: 408 unique: 12, success, outsize: 160 unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 13, error: -2 (No such file or directory), outsize: 16 unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408 unique: 14, error: -2 (No such file or directory), outsize: 16 unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 109, pid: 408 unique: 2, success, outsize: 160 unique: 15, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408 unique: 15, success, outsize: 136 unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 16, error: -2 (No such file or directory), outsize: 16 unique: 17, opcode: CREATE (35), nodeid: 1, insize: 68, pid: 408 unique: 17, success, outsize: 176 unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 18, error: -2 (No such file or directory), outsize: 16 unique: 19, opcode: WRITE (16), nodeid: 17, insize: 4176, pid: 408 unique: 19, success, outsize: 24 unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408 unique: 20, success, outsize: 160 unique: 21, opcode: GETATTR (3), nodeid: 17, insize: 56, pid: 408 unique: 21, success, outsize: 136 unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 3, error: -2 (No such file or directory), outsize: 16 unique: 4, opcode: RELEASE (18), nodeid: 17, insize: 64, pid: 408 ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file unique: 4, error: -5 (Input/output error), outsize: 16 unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408 unique: 5, success, outsize: 160 unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 6, error: -2 (No such file or directory), outsize: 16 unique: 7, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408 unique: 7, success, outsize: 136 unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 8, success, outsize: 160 unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408 unique: 9, error: -2 (No such file or directory), outsize: 16 unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 10, error: -2 (No such file or directory), outsize: 16 unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 11, success, outsize: 160 unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 12, error: -2 (No such file or directory), outsize: 16 unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 46, pid: 408 unique: 13, success, outsize: 160 unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 14, error: -2 (No such file or directory), outsize: 16 unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408 unique: 2, success, outsize: 160 unique: 15, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 15, success, outsize: 96 unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 16, error: -2 (No such file or directory), outsize: 16 unique: 17, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 17, success, outsize: 16 unique: 18, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 18, success, outsize: 16 unique: 19, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 19, success, outsize: 16 unique: 20, opcode: RELEASEDIR (29), nodeid: 1, insize: 64, pid: 408 unique: 20, error: -78 (Function not implemented), outsize: 16 unique: 21, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408 unique: 21, success, outsize: 136 unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 3, success, outsize: 96 unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 4, error: -2 (No such file or directory), outsize: 16 unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 5, error: -2 (No such file or directory), outsize: 16 unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 6, error: -2 (No such file or directory), outsize: 16 unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 7, error: -2 (No such file or directory), outsize: 16 unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408 unique: 8, error: -2 (No such file or directory), outsize: 16 unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408 unique: 9, error: -2 (No such file or directory), outsize: 16 unique: 10, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690 unique: 10, success, outsize: 96 unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408 unique: 11, error: -2 (No such file or directory), outsize: 16 unique: 12, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690 unique: 12, success, outsize: 96 unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 13, error: -2 (No such file or directory), outsize: 16 unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 14, error: -2 (No such file or directory), outsize: 16 unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 62, pid: 9690 unique: 2, error: -2 (No such file or directory), outsize: 16 unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 15, error: -2 (No such file or directory), outsize: 16 unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 57, pid: 9690 unique: 16, error: -2 (No such file or directory), outsize: 16 unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 17, error: -2 (No such file or directory), outsize: 16 unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 18, error: -2 (No such file or directory), outsize: 16 unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 19, error: -2 (No such file or directory), outsize: 16 unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 20, error: -2 (No such file or directory), outsize: 16 unique: 21, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 21, error: -2 (No such file or directory), outsize: 16 unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 3, error: -2 (No such file or directory), outsize: 16 unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 4, error: -2 (No such file or directory), outsize: 16 unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 5, error: -2 (No such file or directory), outsize: 16 unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 6, error: -2 (No such file or directory), outsize: 16 unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 7, error: -2 (No such file or directory), outsize: 16 unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 8, error: -2 (No such file or directory), outsize: 16 unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 9, error: -2 (No such file or directory), outsize: 16 unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 10, error: -2 (No such file or directory), outsize: 16 unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 11, error: -2 (No such file or directory), outsize: 16 unique: 12, opcode: OPENDIR (27), nodeid: 1, insize: 48, pid: 408 unique: 12, success, outsize: 32 unique: 13, opcode: RELEASEDIR (29), nodeid: 1, insize: 64, pid: 408 unique: 13, error: -78 (Function not implemented), outsize: 16 unique: 14, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408 unique: 14, success, outsize: 136 unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 2, error: -2 (No such file or directory), outsize: 16 unique: 15, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690 unique: 15, success, outsize: 96 unique: 16, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690 unique: 16, success, outsize: 96 unique: 17, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690 unique: 17, success, outsize: 96 unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 18, error: -2 (No such file or directory), outsize: 16 unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 9690 unique: 19, error: -2 (No such file or directory), outsize: 16 unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 9690 unique: 20, error: -2 (No such file or directory), outsize: 16 unique: 21, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690 unique: 21, success, outsize: 96 unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 3, error: -2 (No such file or directory), outsize: 16 unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 9690 unique: 4, error: -2 (No such file or directory), outsize: 16 unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 9690 unique: 5, error: -2 (No such file or directory), outsize: 16 unique: 6, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690 unique: 6, success, outsize: 96 unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 7, error: -2 (No such file or directory), outsize: 16 unique: 8, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690 unique: 8, success, outsize: 96 unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 62, pid: 9690 unique: 9, error: -2 (No such file or directory), outsize: 16 unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 10, error: -2 (No such file or directory), outsize: 16 unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 9690 unique: 11, error: -2 (No such file or directory), outsize: 16 unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 9690 unique: 12, error: -2 (No such file or directory), outsize: 16 unique: 13, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690 unique: 13, success, outsize: 96 unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 57, pid: 9690 unique: 14, error: -2 (No such file or directory), outsize: 16 unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 2, error: -2 (No such file or directory), outsize: 16 unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 9690 unique: 15, error: -2 (No such file or directory), outsize: 16 unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 16, error: -2 (No such file or directory), outsize: 16 unique: 17, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690 unique: 17, success, outsize: 96 unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 18, error: -2 (No such file or directory), outsize: 16 unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 9690 unique: 19, error: -2 (No such file or directory), outsize: 16 unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 20, error: -2 (No such file or directory), outsize: 16 unique: 21, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690 unique: 21, success, outsize: 96 unique: 3, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690 unique: 3, success, outsize: 96 unique: 4, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690 unique: 4, success, outsize: 96 unique: 5, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690 unique: 5, success, outsize: 96 unique: 6, opcode: OPEN (14), nodeid: 7, insize: 48, pid: 64903 unique: 6, success, outsize: 32 unique: 7, opcode: SETATTR (4), nodeid: 7, insize: 168, pid: 64903 unique: 7, success, outsize: 136 unique: 8, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690 unique: 8, success, outsize: 96 unique: 9, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 INFO:rmcl.items:Rendering <Document "Quick sheets">: 10.0% INFO:rmcl.items:Rendering <Document "Quick sheets">: 20.0% INFO:rmcl.items:Rendering <Document "Quick sheets">: 30.0% INFO:rmcl.items:Rendering <Document "Quick sheets">: 40.0% INFO:rmcl.items:Rendering <Document "Quick sheets">: 50.0% INFO:rmcl.items:Rendering <Document "Quick sheets">: 60.0% INFO:rmcl.items:Rendering <Document "Quick sheets">: 70.0% INFO:rmcl.items:Rendering <Document "Quick sheets">: 80.0% INFO:rmcl.items:Rendering <Document "Quick sheets">: 90.0% INFO:rmcl.items:Rendering <Document "Quick sheets">: 100.0% INFO:rmrl.render:exported pdf unique: 9, success, outsize: 4112 unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 9690 unique: 10, error: -2 (No such file or directory), outsize: 16 unique: 9, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 9, success, outsize: 4112 unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 9690 unique: 11, error: -2 (No such file or directory), outsize: 16 unique: 9, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 9, success, outsize: 4112 unique: 12, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690 unique: 12, success, outsize: 96 unique: 9, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 9, success, outsize: 4112 unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 9690 unique: 13, error: -2 (No such file or directory), outsize: 16 unique: 9, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 9, success, outsize: 4112 unique: 14, opcode: OPEN (14), nodeid: 6, insize: 48, pid: 9690 unique: 14, success, outsize: 32 unique: 9, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 9, success, outsize: 4112 unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 9690 unique: 2, error: -2 (No such file or directory), outsize: 16 unique: 9, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 9, success, outsize: 4112 unique: 15, opcode: RELEASE (18), nodeid: 6, insize: 64, pid: 9690 unique: 15, success, outsize: 16 unique: 9, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 9, success, outsize: 4112 unique: 16, opcode: GETATTR (3), nodeid: 6, insize: 56, pid: 9690 unique: 16, success, outsize: 136 unique: 17, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 17, success, outsize: 4112 unique: 18, opcode: OPEN (14), nodeid: 6, insize: 48, pid: 9690 unique: 18, success, outsize: 32 unique: 17, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 17, success, outsize: 4112 unique: 19, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 9690 unique: 19, success, outsize: 96 unique: 17, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 17, success, outsize: 4112 unique: 20, opcode: RELEASE (18), nodeid: 6, insize: 64, pid: 9690 unique: 20, success, outsize: 16 unique: 17, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 17, success, outsize: 4112 unique: 17, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 17, success, outsize: 2618 unique: 21, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 21, success, outsize: 4112 unique: 21, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 21, success, outsize: 4112 unique: 21, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 21, success, outsize: 4112 unique: 21, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 21, success, outsize: 4112 unique: 3, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 3, success, outsize: 4112 unique: 3, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 3, success, outsize: 4112 unique: 3, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 3, success, outsize: 4112 unique: 3, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 3, success, outsize: 4112 unique: 4, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 4, success, outsize: 4112 unique: 4, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 4, success, outsize: 4112 unique: 5, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 5, success, outsize: 4112 unique: 5, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 5, success, outsize: 4112 unique: 5, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 5, success, outsize: 4112 unique: 5, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 5, success, outsize: 4112 unique: 5, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 5, success, outsize: 4112 unique: 5, opcode: READ (15), nodeid: 7, insize: 80, pid: 64903 unique: 5, success, outsize: 4112 unique: 6, opcode: GETATTR (3), nodeid: 6, insize: 56, pid: 64903 unique: 6, success, outsize: 136 unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408 unique: 7, error: -2 (No such file or directory), outsize: 16 unique: 8, opcode: OPEN (14), nodeid: 6, insize: 48, pid: 64903 unique: 8, success, outsize: 32 unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408 unique: 10, error: -2 (No such file or directory), outsize: 16 unique: 11, opcode: SETATTR (4), nodeid: 6, insize: 168, pid: 64903 unique: 11, success, outsize: 136 unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408 unique: 12, error: -2 (No such file or directory), outsize: 16 unique: 13, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 INFO:rmcl.items:Rendering <Document "todo">: 10.0% INFO:rmcl.items:Rendering <Document "todo">: 20.0% INFO:rmcl.items:Rendering <Document "todo">: 30.0% INFO:rmcl.items:Rendering <Document "todo">: 40.0% INFO:rmcl.items:Rendering <Document "todo">: 50.0% INFO:rmcl.items:Rendering <Document "todo">: 60.0% INFO:rmcl.items:Rendering <Document "todo">: 70.0% INFO:rmcl.items:Rendering <Document "todo">: 80.0% INFO:rmcl.items:Rendering <Document "todo">: 90.0% INFO:rmcl.items:Rendering <Document "todo">: 100.0% INFO:rmrl.render:exported pdf unique: 13, success, outsize: 4112 unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408 unique: 14, error: -2 (No such file or directory), outsize: 16 unique: 13, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 13, success, outsize: 4112 unique: 13, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 13, success, outsize: 4112 unique: 13, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 13, success, outsize: 4112 unique: 13, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 13, success, outsize: 4112 unique: 13, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 13, success, outsize: 4112 unique: 13, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 13, success, outsize: 4112 unique: 13, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 13, success, outsize: 4112 unique: 2, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 2, success, outsize: 4112 unique: 2, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 2, success, outsize: 4112 unique: 2, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 2, success, outsize: 4112 unique: 2, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 2, success, outsize: 4112 unique: 2, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 2, success, outsize: 1225 unique: 15, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 15, success, outsize: 4112 unique: 9, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 9, success, outsize: 4112 unique: 9, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 9, success, outsize: 4112 unique: 9, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 9, success, outsize: 4112 unique: 9, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 9, success, outsize: 4112 unique: 16, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 16, success, outsize: 4112 unique: 16, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 16, success, outsize: 4112 unique: 16, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 16, success, outsize: 4112 unique: 16, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 16, success, outsize: 4112 unique: 16, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 16, success, outsize: 4112 unique: 16, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 16, success, outsize: 4112 unique: 18, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 18, success, outsize: 4112 unique: 18, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 18, success, outsize: 4112 unique: 18, opcode: READ (15), nodeid: 6, insize: 80, pid: 64903 unique: 18, success, outsize: 4112 unique: 19, opcode: RELEASE (18), nodeid: 6, insize: 64, pid: 64903 unique: 19, success, outsize: 16 unique: 20, opcode: GETATTR (3), nodeid: 6, insize: 56, pid: 408 unique: 20, success, outsize: 136 unique: 17, opcode: RELEASE (18), nodeid: 7, insize: 64, pid: 64903 unique: 17, success, outsize: 16 unique: 21, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 21, error: -2 (No such file or directory), outsize: 16 unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 3, error: -2 (No such file or directory), outsize: 16 unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 4, error: -2 (No such file or directory), outsize: 16 unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 5, error: -2 (No such file or directory), outsize: 16 unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 6, error: -2 (No such file or directory), outsize: 16 unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 7, error: -2 (No such file or directory), outsize: 16 unique: 8, opcode: OPENDIR (27), nodeid: 1, insize: 48, pid: 408 unique: 8, success, outsize: 32 unique: 10, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 10, success, outsize: 576 unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 11, error: -2 (No such file or directory), outsize: 16 unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 12, error: -2 (No such file or directory), outsize: 16 unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408 unique: 14, error: -2 (No such file or directory), outsize: 16 unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 13, error: -2 (No such file or directory), outsize: 16 unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 2, error: -2 (No such file or directory), outsize: 16 unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 15, error: -2 (No such file or directory), outsize: 16 unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 9, error: -2 (No such file or directory), outsize: 16 unique: 16, opcode: GETATTR (3), nodeid: 7, insize: 56, pid: 408 unique: 16, success, outsize: 136 unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408 unique: 18, error: -2 (No such file or directory), outsize: 16 unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408 unique: 19, error: -2 (No such file or directory), outsize: 16 unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 20, error: -2 (No such file or directory), outsize: 16 unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 17, error: -2 (No such file or directory), outsize: 16 unique: 21, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 21, error: -2 (No such file or directory), outsize: 16 unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 3, error: -2 (No such file or directory), outsize: 16 unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 4, error: -2 (No such file or directory), outsize: 16 unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408 unique: 5, error: -2 (No such file or directory), outsize: 16 unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 6, error: -2 (No such file or directory), outsize: 16 unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 7, error: -2 (No such file or directory), outsize: 16 unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 8, error: -2 (No such file or directory), outsize: 16 unique: 10, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 10, success, outsize: 16 unique: 11, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 11, success, outsize: 16 unique: 12, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 12, success, outsize: 16 unique: 14, opcode: RELEASEDIR (29), nodeid: 1, insize: 64, pid: 408 unique: 14, error: -78 (Function not implemented), outsize: 16 unique: 13, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408 unique: 13, success, outsize: 136 unique: 2, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 2, success, outsize: 96 unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 15, error: -2 (No such file or directory), outsize: 16 unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 9, error: -2 (No such file or directory), outsize: 16 unique: 16, opcode: OPENDIR (27), nodeid: 1, insize: 48, pid: 408 unique: 16, success, outsize: 32 unique: 18, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 18, success, outsize: 576 unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 19, error: -2 (No such file or directory), outsize: 16 unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 20, error: -2 (No such file or directory), outsize: 16 unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408 unique: 17, error: -2 (No such file or directory), outsize: 16 unique: 21, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 21, error: -2 (No such file or directory), outsize: 16 unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 3, error: -2 (No such file or directory), outsize: 16 unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 4, error: -2 (No such file or directory), outsize: 16 unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 5, error: -2 (No such file or directory), outsize: 16 unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408 unique: 6, error: -2 (No such file or directory), outsize: 16 unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408 unique: 7, error: -2 (No such file or directory), outsize: 16 unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 8, error: -2 (No such file or directory), outsize: 16 unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 10, error: -2 (No such file or directory), outsize: 16 unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 11, error: -2 (No such file or directory), outsize: 16 unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 12, error: -2 (No such file or directory), outsize: 16 unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 14, error: -2 (No such file or directory), outsize: 16 unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408 unique: 13, error: -2 (No such file or directory), outsize: 16 unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 2, error: -2 (No such file or directory), outsize: 16 unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 15, error: -2 (No such file or directory), outsize: 16 unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 9, error: -2 (No such file or directory), outsize: 16 unique: 16, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 16, success, outsize: 16 unique: 18, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 18, success, outsize: 16 unique: 19, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 19, success, outsize: 16 unique: 20, opcode: RELEASEDIR (29), nodeid: 1, insize: 64, pid: 408 unique: 20, error: -78 (Function not implemented), outsize: 16 unique: 17, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408 unique: 17, success, outsize: 136 unique: 21, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 21, success, outsize: 96 unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 3, error: -2 (No such file or directory), outsize: 16 unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 408 unique: 4, error: -2 (No such file or directory), outsize: 16 unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 408 unique: 5, error: -2 (No such file or directory), outsize: 16 unique: 6, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 425 unique: 6, error: -2 (No such file or directory), outsize: 16 unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 425 unique: 7, error: -2 (No such file or directory), outsize: 16 unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 425 unique: 8, error: -2 (No such file or directory), outsize: 16 unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 425 unique: 10, error: -2 (No such file or directory), outsize: 16 unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 425 unique: 11, error: -2 (No such file or directory), outsize: 16 unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 408 unique: 12, error: -2 (No such file or directory), outsize: 16 unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 408 unique: 14, error: -2 (No such file or directory), outsize: 16 unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 408 unique: 13, error: -2 (No such file or directory), outsize: 16 unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 408 unique: 2, error: -2 (No such file or directory), outsize: 16 unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 408 unique: 15, error: -2 (No such file or directory), outsize: 16 unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 408 unique: 9, error: -2 (No such file or directory), outsize: 16 unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 65086 unique: 16, error: -2 (No such file or directory), outsize: 16 unique: 18, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 65086 unique: 18, success, outsize: 96 unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 95, pid: 65086 unique: 19, error: -2 (No such file or directory), outsize: 16 unique: 20, opcode: CREATE (35), nodeid: 1, insize: 111, pid: 65086 unique: 20, success, outsize: 176 unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 97, pid: 65086 unique: 17, error: -2 (No such file or directory), outsize: 16 unique: 21, opcode: RELEASE (18), nodeid: 18, insize: 64, pid: 65086 ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file unique: 21, error: -5 (Input/output error), outsize: 16 unique: 3, opcode: GETATTR (3), nodeid: 18, insize: 56, pid: 65086 unique: 3, success, outsize: 136 unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 97, pid: 65086 unique: 4, error: -2 (No such file or directory), outsize: 16 unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 97, pid: 65086 unique: 5, error: -2 (No such file or directory), outsize: 16 unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 97, pid: 65086 unique: 7, error: -2 (No such file or directory), outsize: 16 unique: 8, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 65086 unique: 8, success, outsize: 136 unique: 10, opcode: CREATE (35), nodeid: 1, insize: 113, pid: 65086 unique: 10, success, outsize: 176 unique: 11, opcode: WRITE (16), nodeid: 19, insize: 4176, pid: 65086 unique: 11, success, outsize: 24 unique: 12, opcode: GETATTR (3), nodeid: 19, insize: 56, pid: 65086 unique: 12, success, outsize: 136 unique: 14, opcode: RELEASE (18), nodeid: 19, insize: 64, pid: 65086 ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file unique: 14, error: -5 (Input/output error), outsize: 16 unique: 13, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 65 unique: 13, success, outsize: 136 unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 2, error: -2 (No such file or directory), outsize: 16 unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 15, error: -2 (No such file or directory), outsize: 16 unique: 9, opcode: OPENDIR (27), nodeid: 1, insize: 48, pid: 408 unique: 9, success, outsize: 32 unique: 16, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 16, success, outsize: 576 unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 18, error: -2 (No such file or directory), outsize: 16 unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 19, error: -2 (No such file or directory), outsize: 16 unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408 unique: 20, error: -2 (No such file or directory), outsize: 16 unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 17, error: -2 (No such file or directory), outsize: 16 unique: 21, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 21, error: -2 (No such file or directory), outsize: 16 unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 3, error: -2 (No such file or directory), outsize: 16 unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 4, error: -2 (No such file or directory), outsize: 16 unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408 unique: 5, error: -2 (No such file or directory), outsize: 16 unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408 unique: 7, error: -2 (No such file or directory), outsize: 16 unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 8, error: -2 (No such file or directory), outsize: 16 unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 10, error: -2 (No such file or directory), outsize: 16 unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 11, error: -2 (No such file or directory), outsize: 16 unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 12, error: -2 (No such file or directory), outsize: 16 unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 14, error: -2 (No such file or directory), outsize: 16 unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408 unique: 13, error: -2 (No such file or directory), outsize: 16 unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 2, error: -2 (No such file or directory), outsize: 16 unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 15, error: -2 (No such file or directory), outsize: 16 unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 9, error: -2 (No such file or directory), outsize: 16 unique: 16, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 16, success, outsize: 16 unique: 18, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 18, success, outsize: 16 unique: 19, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 19, success, outsize: 16 unique: 20, opcode: RELEASEDIR (29), nodeid: 1, insize: 64, pid: 408 unique: 20, error: -78 (Function not implemented), outsize: 16 unique: 17, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408 unique: 17, success, outsize: 136 unique: 21, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 21, success, outsize: 96 unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 3, error: -2 (No such file or directory), outsize: 16 unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 4, error: -2 (No such file or directory), outsize: 16 unique: 5, opcode: OPENDIR (27), nodeid: 1, insize: 48, pid: 408 unique: 5, success, outsize: 32 unique: 7, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 7, success, outsize: 576 unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 8, error: -2 (No such file or directory), outsize: 16 unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 10, error: -2 (No such file or directory), outsize: 16 unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408 unique: 11, error: -2 (No such file or directory), outsize: 16 unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 12, error: -2 (No such file or directory), outsize: 16 unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 14, error: -2 (No such file or directory), outsize: 16 unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 13, error: -2 (No such file or directory), outsize: 16 unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 2, error: -2 (No such file or directory), outsize: 16 unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408 unique: 15, error: -2 (No such file or directory), outsize: 16 unique: 9, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408 unique: 9, error: -2 (No such file or directory), outsize: 16 unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 16, error: -2 (No such file or directory), outsize: 16 unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 18, error: -2 (No such file or directory), outsize: 16 unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 19, error: -2 (No such file or directory), outsize: 16 unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 20, error: -2 (No such file or directory), outsize: 16 unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 17, error: -2 (No such file or directory), outsize: 16 unique: 21, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408 unique: 21, error: -2 (No such file or directory), outsize: 16 unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 3, error: -2 (No such file or directory), outsize: 16 unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 4, error: -2 (No such file or directory), outsize: 16 unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 5, error: -2 (No such file or directory), outsize: 16 unique: 7, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 7, success, outsize: 16 unique: 8, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 8, success, outsize: 16 unique: 10, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 10, success, outsize: 16 unique: 11, opcode: RELEASEDIR (29), nodeid: 1, insize: 64, pid: 408 unique: 11, error: -78 (Function not implemented), outsize: 16 unique: 12, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408 unique: 12, success, outsize: 136 unique: 14, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 14, success, outsize: 96 unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 13, error: -2 (No such file or directory), outsize: 16 unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 44, pid: 408 unique: 2, error: -2 (No such file or directory), outsize: 16 unique: 15, opcode: OPENDIR (27), nodeid: 1, insize: 48, pid: 408 unique: 15, success, outsize: 32 unique: 9, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 9, success, outsize: 576 unique: 16, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 16, error: -2 (No such file or directory), outsize: 16 unique: 18, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 18, error: -2 (No such file or directory), outsize: 16 unique: 19, opcode: LOOKUP (1), nodeid: 1, insize: 49, pid: 408 unique: 19, error: -2 (No such file or directory), outsize: 16 unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 20, error: -2 (No such file or directory), outsize: 16 unique: 17, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 17, error: -2 (No such file or directory), outsize: 16 unique: 21, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 21, error: -2 (No such file or directory), outsize: 16 unique: 3, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 3, error: -2 (No such file or directory), outsize: 16 unique: 4, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408 unique: 4, error: -2 (No such file or directory), outsize: 16 unique: 5, opcode: LOOKUP (1), nodeid: 1, insize: 59, pid: 408 unique: 5, error: -2 (No such file or directory), outsize: 16 unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 7, error: -2 (No such file or directory), outsize: 16 unique: 8, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 8, error: -2 (No such file or directory), outsize: 16 unique: 10, opcode: LOOKUP (1), nodeid: 1, insize: 111, pid: 408 unique: 10, error: -2 (No such file or directory), outsize: 16 unique: 11, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 11, error: -2 (No such file or directory), outsize: 16 unique: 12, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 12, error: -2 (No such file or directory), outsize: 16 unique: 14, opcode: LOOKUP (1), nodeid: 1, insize: 52, pid: 408 unique: 14, error: -2 (No such file or directory), outsize: 16 unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 50, pid: 408 unique: 13, error: -2 (No such file or directory), outsize: 16 unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 48, pid: 408 unique: 2, error: -2 (No such file or directory), outsize: 16 unique: 15, opcode: LOOKUP (1), nodeid: 1, insize: 51, pid: 408 unique: 15, error: -2 (No such file or directory), outsize: 16 unique: 9, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 9, success, outsize: 16 unique: 16, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 16, success, outsize: 16 unique: 18, opcode: READDIR (28), nodeid: 1, insize: 80, pid: 408 unique: 18, success, outsize: 16 unique: 19, opcode: RELEASEDIR (29), nodeid: 1, insize: 64, pid: 408 unique: 19, error: -78 (Function not implemented), outsize: 16 unique: 20, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 408 unique: 20, success, outsize: 136 unique: 17, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 408 unique: 17, success, outsize: 96 The error seems to be that RMfuse cannot determine that the file you're trying to upload is a PDF or EPUB file. Which type is it? (Note that if it's not one of these, the upload should fail.) If you're able to send me the file, I can take a look. If not, just the beginning of the file would be enough for me to see if RMfuse should be able to recognize it. You can run head -c 100 file.pdf > file.fragment, for example, to get the first 100 bytes of file.pdf for me. Hi, I tested it with the same file that I downloaded from the remarkable. It was definitely a pdf file. Here is again the python error trace when I try to cp: (rmfuse) schneider@Aarons-MacBook-Pro ~ % cp ~/Desktop/debug.pdf remarkable cp: remarkable/debug.pdf: fchmod failed: Input/output error Traceback (most recent call last): File "/Users/schneider/anaconda3/envs/rmfuse/bin/rmfuse", line 8, in <module> sys.exit(main()) File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmfuse/fuse.py", line 470, in main fuse.main(workers=1) File "src/fuse_api.pxi", line 327, in llfuse.main File "src/handlers.pxi", line 136, in llfuse.fuse_setattr File "src/handlers.pxi", line 137, in llfuse.fuse_setattr File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmfuse/fuselib.py", line 25, in decorated return trio.run(runner) File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/trio/_core/_run.py", line 1932, in run raise runner.main_task_outcome.error File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmfuse/fuselib.py", line 24, in runner return await afunc(*args, **kw) File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmfuse/fuse.py", line 232, in setattr log.debug(f'setattr called on {await self.get_by_id(self.get_id(inode))!r}') File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmfuse/fuse.py", line 139, in get_by_id return await Item.get_by_id(id_) File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmcl/items.py", line 46, in get_by_id return await (await api.get_client()).get_by_id(id_) File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmcl/api.py", line 247, in get_by_id return self.by_id[id_] KeyError: 'c369d328-db7f-4673-8588-ac0305b8fbad' cp: remarkable/debug.pdf: Device not configured [1] + exit 1 rmfuse remarkable cp: /Users/schneider/Desktop/debug.pdf: could not copy extended attributes to remarkable/debug.pdf: Device not configured Here is another error which I get upon renaming a file: (rmfuse) schneider@Aarons-MacBook-Pro remarkable % mv How\ to\ Debug\ Fortran\ code\ with\ GDB\ -\ Fortran\ Debugging\ -\ Undo.io.pdf debug.pdf mv: rename How to Debug Fortran code with GDB - Fortran Debugging - Undo.io.pdf to debug.pdf: Input/output error (rmfuse) schneider@Aarons-MacBook-Pro remarkable % Traceback (most recent call last): File "/Users/schneider/anaconda3/envs/rmfuse/bin/rmfuse", line 8, in <module> sys.exit(main()) File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmfuse/fuse.py", line 470, in main fuse.main(workers=1) File "src/fuse_api.pxi", line 327, in llfuse.main File "src/handlers.pxi", line 263, in llfuse.fuse_rename File "src/handlers.pxi", line 264, in llfuse.fuse_rename File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmfuse/fuselib.py", line 25, in decorated return trio.run(runner) File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/trio/_core/_run.py", line 1932, in run raise runner.main_task_outcome.error File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmfuse/fuselib.py", line 24, in runner return await afunc(*args, **kw) TypeError: rename() missing 1 required positional argument: 'ctx' Here is the debug.pdf that I copied from the rmcloud to my desktop. And which I tried to upload back to the rm. On the rm this file is called How\ to\ Debug\ Fortran\ code\ with\ GDB\ -\ Fortran\ Debugging\ -\ Undo.io.pdf debug.pdf Thanks! Thanks. I think I missed the tracebacks into the logs above. The good new is, both of these are simple bugs. I've pushed fixes; give it a try. (I swear we're getting closer, even if it doesn't seem that way!) rmfuse-0.1.2a0.tar.gz Thanks again for debugging! We are getting there! Renaming and deleting files seems to work now. Moving files (either by shell command or by finder) does not yet work. It often boils down to these errors: RELEASEDIR and LOOKUP. unique: 17, opcode: RELEASEDIR (29), nodeid: 1, insize: 64, pid: 434 unique: 17, error: -78 (Function not implemented), outsize: 16 unique: 18, opcode: GETATTR (3), nodeid: 1, insize: 56, pid: 434 unique: 18, success, outsize: 136 unique: 19, opcode: STATFS (17), nodeid: 1, insize: 40, pid: 434 unique: 19, success, outsize: 96 unique: 2, opcode: LOOKUP (1), nodeid: 1, insize: 60, pid: 65 unique: 2, error: -2 (No such file or directory), outsize: 16 Both give always errors. I'm not convinced that those errors are a problem. The LOOKUP is probably result of something on your system looking for a file that doesn't exist, perhaps some metadata directory. The RELEASEDIR isn't implemented, since RMfuse doesn't have anything to do to release a directory. But that shouldn't cause problems. (We can try adding a stub if we can't figure out anything else, but I suspect it wouldn't really help.) What's the failure mode of moving the file? Does it crash? If so, please provide the traceback. Does it hang? How long did you wait? Does nothing happen? Does this happen when moving files within the RMfuse mount point, moving a file from your system to RMfuse, or moving a file from RMfuse to your system? sorry for the poor Bugreport. Let me try to improve on your questions: It oddly depends on the file (and its size). I tested four files: ms2.pdf (a paper generated with pdflatex), size=961KB rmfuse_test.pdf (a random pdf printout of this issue CMD + P), size=200KB Molliere_thesis.pdf (some pdf of a phd thesis, probably also created with pdflatex), size=30.3MB SA_Chap2_1.pdf (a lecture presentation), size=16.2MB What's the failure mode of moving the file? In the file explorer (hereafter finder) I always get this error (regardless of the file): The operation could not be completed because at least one required object was not found (error -43). Copying using the terminal succeeds for the smaller documents (ms2.pdf and rmfuse_test.pdf). In terminal that is moving the file: cp: Desktop/paper_2/ms2.pdf: could not copy extended attributes to /Users/schneider/remarkable/ms2.pdf: Attribute not found In terminal that is in control of rmfuse: Error: Not a PDF or EPUB file However the two larger documents do not succeed: (base) schneider@Aarons-MacBook-Pro Desktop % cp Molliere_thesis.pdf ~/remarkable cp: Molliere_thesis.pdf: could not copy extended attributes to /Users/schneider/remarkable/Molliere_thesis.pdf: Operation not supported in rmfuse window: ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file ERROR:root:Attempt to get non-existant Item 699b20b3-6826-4625-b22a-600354ee4e38 Does it crash? If so, please provide the traceback. Does it hang? rmfuse does not crash nor hang. How long did you wait? a couple of minutes. I also checked with rmapi if the file was copied but just not listed. Does this happen when moving files within the RMfuse mount point, moving a file from your system to RMfuse, or moving a file from RMfuse to your system? Moving a file inside rmfuse and moving a file from rmfuse to system gives no error. Only moving files from local to rmfuse. Thanks, that's very informative. I just don't quite know what it all means. :) The errors about the extended attributes are expected -- RMfuse doesn't implement them right now. (It might be possible to stuff them into the document metadata at some point.) But it is weird that the error on setting them is different in the two cases. If they get annoying, we could add an option to RMfuse that would pretend to set the attributes, which should get rid of that message. I'm not sure what's going wrong with the "Not a PDF or EPUB file" error. My best guess is that there's some metadata file that the Mac is trying to create / copy along with the PDF file. The fact that it's happening in both successes and failures suggests that it's not a fatal problem. The worrisome bit is the "Attempt to get non-existent Item". That means something is going wrong in the upload process. I wonder if we have a race condition -- something is trying to do something to a file while it's in the process of being uploaded. This would slip in a gap of awareness of RMfuse -- it can't find the document in the set of editable new-created documents, since we closed it out to upload. But it also doesn't see it in the list of documents from the Cloud, because the upload hasn't finished. (This would explain why it's only happening to larger files.) I've built a new version with some more debugging output and an attempt to handle the above case. Please give it a try (with -v) and let me know what happens. One trick I've used for keeping track of what's happening: If you press 'Enter' in the rmfuse terminal between events you're triggering in another terminal, you can better associate the debugging lines with specific actions. rmfuse-0.1.2a0.tar.gz Thanks for the explainations! Copying Molliere_thesis or ms2 to rmfuse using finder (still same error code -43 in finder): DEBUG:rmfuse.fuse:Created Molliere_thesis for b'Molliere_thesis.pdf', with inode 16 and ID 9dc8cf3f-17ec-40e0-80e1-f4b61ea41289 DEBUG:rmfuse.fuse:Releasing inode 16 ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was Molliere_thesis) ERROR:root:Attempt to get non-existent Item 9dc8cf3f-17ec-40e0-80e1-f4b61ea41289 The same from command line (only the last couple lines...): (...) DEBUG:rmfuse.fuse:Writing to 18 at 30257152, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30261248, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30265344, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30269440, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30273536, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30277632, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30281728, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30285824, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30289920, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30294016, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30298112, length 755 DEBUG:rmfuse.fuse:setattr called on <Document "Molliere_thesis"> DEBUG:rmfuse.fuse: Attempting to set mode to 33188 DEBUG:rmfuse.fuse: No changes made DEBUG:rmfuse.fuse:setattr called on <Document "Molliere_thesis"> DEBUG:rmfuse.fuse: Attempting to set mode to 33188 DEBUG:rmfuse.fuse: No changes made DEBUG:rmfuse.fuse:Created ._Molliere_thesis for b'._Molliere_thesis.pdf', with inode 19 and ID 870c59c6-47aa-42a1-8599-291f5e2815e9 DEBUG:rmfuse.fuse:Writing to 19 at 0, length 4096 DEBUG:rmfuse.fuse:Releasing inode 19 ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was ._Molliere_thesis) DEBUG:rmfuse.fuse:Releasing inode 18 ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was Molliere_thesis) ERROR:root:Attempt to get non-existent Item 4a4313ac-4c86-493d-9f5b-2378666aa553 In contrast here we have ms2.pdf which works: DEBUG:rmfuse.fuse:Created ms2 for b'ms2.pdf', with inode 20 and ID 6e376b45-4c4b-4e4b-a5b1-34e9679da343 DEBUG:rmfuse.fuse:setattr called on <Document "ms2"> DEBUG:rmfuse.fuse: Attempting to set mode to 33152 DEBUG:rmfuse.fuse: No changes made DEBUG:rmfuse.fuse:Writing to 20 at 0, length 4096 DEBUG:rmfuse.fuse:Writing to 20 at 4096, length 4096 DEBUG:rmfuse.fuse:Writing to 20 at 8192, length 4096 (...) DEBUG:rmfuse.fuse:Writing to 20 at 950272, length 4096 DEBUG:rmfuse.fuse:Writing to 20 at 954368, length 4096 DEBUG:rmfuse.fuse:Writing to 20 at 958464, length 2067 DEBUG:rmfuse.fuse:setattr called on <Document "ms2"> DEBUG:rmfuse.fuse: Attempting to set mode to 33188 DEBUG:rmfuse.fuse: No changes made DEBUG:rmfuse.fuse:setattr called on <Document "ms2"> DEBUG:rmfuse.fuse: Attempting to set mode to 33188 DEBUG:rmfuse.fuse: No changes made DEBUG:rmfuse.fuse:Created ._ms2 for b'._ms2.pdf', with inode 21 and ID 539b5949-8baa-4336-9c60-5febbb25dfe3 DEBUG:rmfuse.fuse:Writing to 21 at 0, length 4096 DEBUG:rmfuse.fuse:Releasing inode 21 ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was ._ms2) DEBUG:rmfuse.fuse:Releasing inode 20 DEBUG:rmcl.items:Getting details for <Document "ms2"> DEBUG:rmcl.items:Details for <Document "ms2">: type pdf, size 960531 ERROR:root:Attempt to get non-existent Item 539b5949-8baa-4336-9c60-5febbb25dfe3 From what I can see: The difference in cp from terminal between Molliere and ms2 boils down to these two lines missing for Molliere_thesis: DEBUG:rmcl.items:Getting details for <Document "ms2"> DEBUG:rmcl.items:Details for <Document "ms2">: type pdf, size 960531 The difference between a finder copy and a terminal copy seems to be everything in between the first line and the last few lines. Thanks. I feel like we're inching towards understanding. The thing that has me worried is this: ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was Molliere_thesis) This means that RMfuse couldn't figure out that the file is a PDF document. Therefore, it couldn't upload the file, and that's why nothing's working. It looks like there may be two different failures, unfortunately. With finder, it appears that no data was written to the file. I'm not sure why that is. (Maybe finder planned on re-opening the file to write the contents later?) But it makes sense that it failed. But from the command line, we can see a lot of data did get written, so I'm not sure what went wrong there. RMfuse assumes that a PDF file starts with '%PDF'. Can you check whether this is the case for your problematic files? (The 'head' command is useful for reading the first n bytes of a file.) If it's not, can you include the first couple of KB of those files so I can take a look at what's going on there? On Sun, Feb 21, 2021 at 11:23 PM Aaron David Schneider < notifications@github.com> wrote: Thanks for the explainations! Copying Molliere_thesis or ms2 to rmfuse using finder (still same error code -43 in finder): DEBUG:rmfuse.fuse:Created Molliere_thesis for b'Molliere_thesis.pdf', with inode 16 and ID 9dc8cf3f-17ec-40e0-80e1-f4b61ea41289 DEBUG:rmfuse.fuse:Releasing inode 16 ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was Molliere_thesis) ERROR:root:Attempt to get non-existent Item 9dc8cf3f-17ec-40e0-80e1-f4b61ea41289 The same from command line (only the last couple lines...): (...) DEBUG:rmfuse.fuse:Writing to 18 at 30257152, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30261248, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30265344, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30269440, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30273536, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30277632, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30281728, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30285824, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30289920, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30294016, length 4096 DEBUG:rmfuse.fuse:Writing to 18 at 30298112, length 755 DEBUG:rmfuse.fuse:setattr called on <Document "Molliere_thesis"> DEBUG:rmfuse.fuse: Attempting to set mode to 33188 DEBUG:rmfuse.fuse: No changes made DEBUG:rmfuse.fuse:setattr called on <Document "Molliere_thesis"> DEBUG:rmfuse.fuse: Attempting to set mode to 33188 DEBUG:rmfuse.fuse: No changes made DEBUG:rmfuse.fuse:Created ._Molliere_thesis for b'._Molliere_thesis.pdf', with inode 19 and ID 870c59c6-47aa-42a1-8599-291f5e2815e9 DEBUG:rmfuse.fuse:Writing to 19 at 0, length 4096 DEBUG:rmfuse.fuse:Releasing inode 19 ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was ._Molliere_thesis) DEBUG:rmfuse.fuse:Releasing inode 18 ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was Molliere_thesis) ERROR:root:Attempt to get non-existent Item 4a4313ac-4c86-493d-9f5b-2378666aa553 In contrast here we have ms2.pdf which works: DEBUG:rmfuse.fuse:Created ms2 for b'ms2.pdf', with inode 20 and ID 6e376b45-4c4b-4e4b-a5b1-34e9679da343 DEBUG:rmfuse.fuse:setattr called on <Document "ms2"> DEBUG:rmfuse.fuse: Attempting to set mode to 33152 DEBUG:rmfuse.fuse: No changes made DEBUG:rmfuse.fuse:Writing to 20 at 0, length 4096 DEBUG:rmfuse.fuse:Writing to 20 at 4096, length 4096 DEBUG:rmfuse.fuse:Writing to 20 at 8192, length 4096 (...) DEBUG:rmfuse.fuse:Writing to 20 at 950272, length 4096 DEBUG:rmfuse.fuse:Writing to 20 at 954368, length 4096 DEBUG:rmfuse.fuse:Writing to 20 at 958464, length 2067 DEBUG:rmfuse.fuse:setattr called on <Document "ms2"> DEBUG:rmfuse.fuse: Attempting to set mode to 33188 DEBUG:rmfuse.fuse: No changes made DEBUG:rmfuse.fuse:setattr called on <Document "ms2"> DEBUG:rmfuse.fuse: Attempting to set mode to 33188 DEBUG:rmfuse.fuse: No changes made DEBUG:rmfuse.fuse:Created ._ms2 for b'._ms2.pdf', with inode 21 and ID 539b5949-8baa-4336-9c60-5febbb25dfe3 DEBUG:rmfuse.fuse:Writing to 21 at 0, length 4096 DEBUG:rmfuse.fuse:Releasing inode 21 ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was ._ms2) DEBUG:rmfuse.fuse:Releasing inode 20 DEBUG:rmcl.items:Getting details for <Document "ms2"> DEBUG:rmcl.items:Details for <Document "ms2">: type pdf, size 960531 ERROR:root:Attempt to get non-existent Item 539b5949-8baa-4336-9c60-5febbb25dfe3 From what I can see: The difference in cp from terminal between Molliere and ms2 boils down to these two lines missing for Molliere_thesis: DEBUG:rmcl.items:Getting details for <Document "ms2"> DEBUG:rmcl.items:Details for <Document "ms2">: type pdf, size 960531 The difference between a finder copy and a terminal copy seems to be everything in between the first line and the last few lines. — You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/rschroll/rmfuse/issues/2#issuecomment-783151912, or unsubscribe https://github.com/notifications/unsubscribe-auth/AACZ2HZQYXJG545Q2WXCKMLTAIA7DANCNFSM4XISMUSA . Thanks for your quick reply! Here is the head output of Molliere thesis: %PDF-1.3 %���� 1 0 obj <</CreationDate 8963 0 R /Creator 8962 0 R /ModDate 8963 0 R /Producer 8961 0 R >> endobj 2 0 obj <</Count 13/First 7488 0 R /Last 7489 0 R >> endobj 3 0 obj <</Contents 5 0 R /MediaBox[ 0 0 595.276 841.89]/Parent 4 0 R /Resources 7 0 R /Rotate 0/Type/Page>> This would also be a large test pdf (20 MB) - found by googling large test pdf : https://www.hq.nasa.gov/alsj/a17/A17_FlightPlan.pdf We could use this file as a benchmark for this issue since this one fails too and I can share it since its an open document. The bad (or maybe good?) news is that that NASA PDF file works just fine on my test system.  This suggests to me that the Mac is doing something I didn't anticipate when writing the file. Can you check through the 'Writing to X at Y, length Z' messages?  Does Y start at zero and always increase?  Or does it jump around.  (On my system, it increases steadily, but if it doesn't I can imagine that it might break things. Though I don't see how it can break the beginning of the file.) The only other thought I have is to look at the actual values RMfuse finds in the file.  If you're willing to get your hands dirty, try adding a line to rmfuse/fuse.py.  Right after the line log.error(f'Error: Not a PDF or EPUB file (file was {document.name})') Try adding log.error(f'File header: {data[:1024]}') This will print out the first KB of the file as RMfuse got it, and we can see if it's different from what it should be. If you're not comfortable mucking about like this, I can create an instrumented version with this and other debug output later. I just tested your second idea with the Nasa file. Upon (base) schneider@Aarons-MacBook-Pro ~ % cp Downloads/A17_FlightPlan.pdf ~/remarkable cp: Downloads/A17_FlightPlan.pdf: could not copy extended attributes to /Users/schneider/remarkable/A17_FlightPlan.pdf: Operation not supported I get the following output in the rmfuse window: ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was ._A17_FlightPlan) ERROR:rmfuse.fuse:File header: b'\x00\x05\x16\x07\x00\x02\x00\x00Mac OS X \x00\x02\x00\x00\x00\t\x00\x00\x002\x00\x00\x0e\xb0\x00\x00\x00\x02\x00\x00\x0e\xe2\x00\x00\x01\x1e\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00ATTR\x00\x00\x00\x12\x00\x00\x0e\xe2\x00\x00\x00x\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00' ERROR:rmfuse.fuse:Error: Not a PDF or EPUB file (file was A17_FlightPlan) ERROR:rmfuse.fuse:File header: b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00' ERROR:root:Attempt to get non-existent Item 0e77f5b9-e23f-423b-9fa9-a38a522c8b59 Yeah, it looks like the file it is trying to upload is full of null bytes. That ain't right. (As we saw above, the PDF file should start with '%PDF'.) Unfortunately, I have no idea why that would be the case. Perhaps the OS decided it would fill the file with zeros first, and then write the actual content? But I don't know why that would be the case, or why that would only happen with large files. Another possibility would be a bug with llfuse or Macfuse. But again, I don't know why that would only affect large files. I'll have to puzzle over this for a while. In happier news, those ._ files are apparently for filesystem metadata. It appears you can turn that off with a special fuse option. If you add fuse_options.add('noappledouble') around here, it might stop complaining about those files. But I don't think it has any effect on the real problem here. No hurry 👌 In happier news, those ._ files are apparently for filesystem metadata. It appears you can turn that off with a special fuse option. If you add fuse_options.add('noappledouble') around here, it might stop complaining about those files. But I don't think it has any effect on the real problem here. I can confirm. No brilliant thoughts so far. Just to confirm: Other than uploading large files, things seem to be working pretty well? I'm thinking that maybe I should merge #6 into master and cut a new release, with the hope that additional users will help us figure out what's going wrong. How does that sit with you? I've added MacOS install directions, based on your descriptions here. Can you check that I got the correct? https://github.com/rschroll/rmfuse/blob/release-0.2.0/README.md#macos Note that if you want to try this out yourself, you'll need to use the --pre option to install a prerelease version of rmfuse with pip. Thanks! yes that sounds good! The new version doesn't start on osx. (rmfuse) schneider@Aarons-MacBook-Pro ~ % rmfuse ~/remarkable -v DEBUG:llfuse:Initializing llfuse DEBUG:llfuse:Calling fuse_mount DEBUG:llfuse:Calling fuse_lowlevel_new fuse: unknown option `nodoubleapple' Traceback (most recent call last): File "/Users/schneider/anaconda3/envs/rmfuse/bin/rmfuse", line 8, in <module> sys.exit(main()) File "/Users/schneider/anaconda3/envs/rmfuse/lib/python3.9/site-packages/rmfuse/fuse.py", line 563, in main fuse.init(fs, options.mountpoint, fuse_options) File "src/fuse_api.pxi", line 253, in llfuse.init RuntimeError: fuse_lowlevel_new() failed The reason seems to be the typo: noappledouble instead of nodoubleapple. Also: I can't install it the way you suggest it: (rmfuse) schneider@Aarons-MacBook-Pro ~ % pip install rmfuse[llfuse] zsh: no matches found: rmfuse[llfuse] I suppose pip install llfuse rmfuse should work instead? On Sun, Feb 28, 2021 at 10:55 PM Aaron David Schneider < notifications@github.com> wrote: The new version doesn't start on osx. The reason seems to be the typo: noappledouble instead of nodoubleapple. D'oh! I can't install it the way you suggest it: (rmfuse) schneider@Aarons-MacBook-Pro ~ % pip install rmfuse[llfuse] zsh: no matches found: rmfuse[llfuse] You'll need to add the --pre option, to look for pre-release versions in pip. I'll close this issue for now, and we can keep working on the large file upload problem in #18.
gharchive/issue
2021-02-08T08:50:58
2025-04-01T04:35:45.158210
{ "authors": [ "AaronDavidSchneider", "rschroll" ], "repo": "rschroll/rmfuse", "url": "https://github.com/rschroll/rmfuse/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
109598598
fixed your link from http://www.codecur.io/us/songdemo to http://codecur.io/us/projects/song-demo i'm not not mad if you don't accept this merge req. It was just faster than opening a new issue and describe what i mean. Awesome project, didn't expect to see a low-pass filter at 4:20am by a humble bundle link ;) I just misunderstood. I though you meant the link in the repo description, but I see now that the link in the readme also needed to be updated. Thanks for the pull request!
gharchive/pull-request
2015-10-03T02:45:06
2025-04-01T04:35:45.176030
{ "authors": [ "AiyionPrime", "rserota" ], "repo": "rserota/wad", "url": "https://github.com/rserota/wad/pull/48", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2022658524
Reboot activity categorizations Right now there's too much present under "Contemporary" to find what you're looking for. Once TFS drops, I'll revisit the need for this. Many activities will be vaulted, etc.
gharchive/issue
2023-12-03T18:31:21
2025-04-01T04:35:45.252788
{ "authors": [ "rslifka" ], "repo": "rslifka/vault-zero", "url": "https://github.com/rslifka/vault-zero/issues/27", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
955106870
metadata is malformed how to use RSocketTCPClient#requestResponse? Steps to Reproduce import { RSocketClient, JsonSerializers } from "rsocket-core" import RSocketTCPClient from "rsocket-tcp-client" const route = "login" const MIME_TYPE = "application/json" const client = new RSocketClient({ serializers: JsonSerializers, setup: { keepAlive: 10000, lifetime: 86400000, dataMimeType: MIME_TYPE, metadataMimeType: MIME_TYPE, }, transport: new RSocketTCPClient({ host: "47.101.196.18", port: 7000, }), }) client.connect().subscribe({ onError: error => console.error(error), onSubscribe: cancel => { console.log("onSubscribe") console.log(cancel) }, onComplete: rsocket => { rsocket.requestResponse({ data: JSON.stringify({ username: "admin", password: "123456", }), metadata: route, }).subscribe({ onComplete: data => console.log(data), onError: e => console.error(e), }) console.log("onComplete") } }) My Environment RSocket version(s) used: "rsocket-core": "^0.0.27", "rsocket-tcp-client": "^0.0.27" node version:v14.17.3 Hi @yuri-li, Can you provide some more details about what issues you are encountering? In which way is the observed result differing from your expectations? Is there an error thrown that you do not expect? If so, can you be more specific? Hi @yuri-li, Can you provide some more details about what issues you are encountering? In which way is the observed result differing from your expectations? Is there an error thrown that you do not expect? If so, can you be more specific? hi, @viglucci , I am a Java development engineer, good at cloud native. I am afraid of misunderstanding, so the server (springboot.jar) is deployed on the cloud server, the host is real, and the front-end code(typescript) I gave can be run directly. I ran the back-end code (kotlin coroutines+rsocket+kotest) first, and the test code is as follows: @viglucci hi, I found the solution The front-end code after fixing the bug is as follows: import { RSocketClient, BufferEncoders, encodeAndAddWellKnownMetadata, MESSAGE_RSOCKET_COMPOSITE_METADATA, MESSAGE_RSOCKET_ROUTING, encodeRoute } from "rsocket-core" import RSocketTCPClient from "rsocket-tcp-client" const route = "login" const client = new RSocketClient({ setup: { keepAlive: 10000, lifetime: 86400000, dataMimeType: "application/json", metadataMimeType: MESSAGE_RSOCKET_COMPOSITE_METADATA.string, }, transport: new RSocketTCPClient({ host: "47.101.196.18", port: 7000, }, BufferEncoders), }) client.connect().subscribe({ onError: error => console.error(error), onComplete: rsocket => { const routeMetadata = encodeRoute(route) const metadata = encodeAndAddWellKnownMetadata( Buffer.alloc(0), MESSAGE_RSOCKET_ROUTING, routeMetadata ) rsocket.requestResponse({ data: Buffer.from(JSON.stringify({ username: "admin", password: "123456", })), metadata, }).subscribe({ onError: e => console.error(e), onComplete: payload => console.log((payload.data as Buffer).toString()), }) console.log("onComplete") } }) Hi @yuri-li, Thanks for following up and providing the solution you found. I believe this issue could be avoided if we had more extensive documentation around composite metadata, as well as more centralized and extensive examples. I'll keep this in mind as we improve the documentation and APIs moving forward. I'll close this issue since you've come to a solution. @yuri-li you have incorrect client configurations. Please see these examples to see how to utilize composite metadata -> https://github.com/rsocket/rsocket-js/blob/master/packages/rsocket-examples/src/CompositeMetadataExample.js Also, here is a working js -> spring-boot-rsocket example https://github.com/OlegDokuka/rsocket-crosslanguage-example
gharchive/issue
2021-07-28T17:58:01
2025-04-01T04:35:45.280495
{ "authors": [ "OlegDokuka", "viglucci", "yuri-li" ], "repo": "rsocket/rsocket-js", "url": "https://github.com/rsocket/rsocket-js/issues/153", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
56555526
Add content to homepage of website Explain the project This is taken care of.
gharchive/issue
2015-02-04T17:02:22
2025-04-01T04:35:45.281704
{ "authors": [ "jenningsanderson", "rsoden" ], "repo": "rsoden/osm-history", "url": "https://github.com/rsoden/osm-history/issues/6", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
324737
Netbeans is using "--runner" option when executing 'rake spec' This problem is described here: http://www.ruby-forum.com/topic/217187 Netbeans 7.0.1 has a Ruby/Rails Plugin and the issue still exists
gharchive/issue
2010-09-20T15:24:59
2025-04-01T04:35:45.282961
{ "authors": [ "PizzaPill", "bmcardoso" ], "repo": "rspec/rspec-core", "url": "https://github.com/rspec/rspec-core/issues/150", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
99131823
'block in shellsplit' Error Getting a random error when trying to run rake: /Users/GMAC/.rvm/rubies/ruby-2.2.2/bin/ruby -I/Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib:/Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-support-3.3.0/lib /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/exe/rspec --pattern spec/\*\*\{,/\*/\*\*\}/\*_spec.rb /Users/GMAC/.rvm/rubies/ruby-2.2.2/lib/ruby/2.2.0/shellwords.rb:75:in `block in shellsplit': Unmatched double quote: "\" Press ? for help" (ArgumentError) from /Users/GMAC/.rvm/rubies/ruby-2.2.2/lib/ruby/2.2.0/shellwords.rb:73:in `scan' from /Users/GMAC/.rvm/rubies/ruby-2.2.2/lib/ruby/2.2.0/shellwords.rb:73:in `shellsplit' from /Users/GMAC/.rvm/rubies/ruby-2.2.2/lib/ruby/2.2.0/shellwords.rb:189:in `shellsplit' from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/flat_map.rb:7:in `each' from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/flat_map.rb:7:in `flat_map' from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/flat_map.rb:7:in `flat_map' from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/configuration_options.rb:164:in `args_from_options_file' from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/configuration_options.rb:151:in `options_from' from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/configuration_options.rb:147:in `global_options' from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/configuration_options.rb:118:in `file_options' from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/configuration_options.rb:43:in `organize_options' from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/configuration_options.rb:13:in `initialize' from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/runner.rb:62:in `new' from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/runner.rb:62:in `run' from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib/rspec/core/runner.rb:41:in `invoke' from /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/exe/rspec:4:in `<main>' /Users/GMAC/.rvm/rubies/ruby-2.2.2/bin/ruby -I/Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/lib:/Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-support-3.3.0/lib /Users/GMAC/.rvm/gems/ruby-2.2.2@myapp/gems/rspec-core-3.3.2/exe/rspec --pattern spec/\*\*\{,/\*/\*\*\}/\*_spec.rb failed Any ideas? Can you paste the contents of your ~/.rspec file? Thanks for leading me in the right direction. Apparently created a .rspec file in my root that had my NERDTree menu pasted into it (thats what I get for trying to learn Vim). Deleted the .rspec and everything works again!
gharchive/issue
2015-08-05T06:08:51
2025-04-01T04:35:45.285067
{ "authors": [ "gxespino", "myronmarston" ], "repo": "rspec/rspec-core", "url": "https://github.com/rspec/rspec-core/issues/2053", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
208948468
Altering captured reference damages spark results. If you use a variable in dplyr::mutate() against a sparklyr data source the lazy eval captures references to user variables. Changing values of those variables implicitly changes the mutate and changes the values seen in the sparklyr result (which is itself a query). This can be worked around by dropping in dplyr::compute() but it seems like it can produce a lot of incorrect calculations. Below is a small example and a lot information on the versions of everything beeing run. I am assuming the is a sparklyr issue as the query views are failrly different than a number of other dplyr structures, but it could be a dplyr issue. OSX 10.11.6. Spark installed as described at http://spark.rstudio.com library('sparklyr') spark_install(version = "2.0.0") library('dplyr') # # Attaching package: 'dplyr' # The following objects are masked from 'package:stats': # # filter, lag # The following objects are masked from 'package:base': # # intersect, setdiff, setequal, union library('sparklyr') R.Version()$version.string # [1] "R version 3.3.2 (2016-10-31)" packageVersion('dplyr') # [1] '0.5.0' packageVersion('sparklyr') # [1] '0.5.2' my_db <- sparklyr::spark_connect(version='2.0.0', master = "local") class(my_db) # [1] "spark_connection" "spark_shell_connection" "DBIConnection" my_db$spark_home # [1] "/Users/johnmount/Library/Caches/spark/spark-2.0.0-bin-hadoop2.7" print(my_db) # $master # [1] "local[4]" # # $method # [1] "shell" # # $app_name # [1] "sparklyr" # # $config # $config$sparklyr.cores.local # [1] 4 # # $config$spark.sql.shuffle.partitions.local # [1] 4 # # $config$spark.env.SPARK_LOCAL_IP.local # [1] "127.0.0.1" # # $config$sparklyr.csv.embedded # [1] "^1.*" # # $config$`sparklyr.shell.driver-class-path` # [1] "" # # attr(,"config") # [1] "default" # attr(,"file") # [1] "/Library/Frameworks/R.framework/Versions/3.3/Resources/library/sparklyr/conf/config-template.yml" # # $spark_home # [1] "/Users/johnmount/Library/Caches/spark/spark-2.0.0-bin-hadoop2.7" # # $backend # description class mode text opened # "->localhost:58964" "sockconn" "wb" "binary" "opened" # can read can write # "yes" "yes" # # $monitor # description class mode text opened # "->localhost:58961" "sockconn" "rb" "binary" "opened" # can read can write # "yes" "yes" # # $output_file # [1] "/var/folders/7q/h_jp2vj131g5799gfnpzhdp80000gn/T//Rtmpxkn3Al/filefeb634813b11_spark.log" # # $spark_context # <jobj[5]> # class org.apache.spark.SparkContext # org.apache.spark.SparkContext@73d0c0e5 # # $java_context # <jobj[6]> # class org.apache.spark.api.java.JavaSparkContext # org.apache.spark.api.java.JavaSparkContext@4e527f99 # # $hive_context # <jobj[9]> # class org.apache.spark.sql.SparkSession # org.apache.spark.sql.SparkSession@1346ee7 # # attr(,"class") # [1] "spark_connection" "spark_shell_connection" "DBIConnection" Expected outcome: s1 has the same value Observed outcome: changing varaible v changes s1 column. support <- copy_to(my_db, data.frame(year=2005:2010), 'support') v <- 0 s1 <- dplyr::mutate(support,count=v) print(s1) # print 1 # Source: query [6 x 2] # Database: spark connection master=local[4] app=sparklyr local=TRUE # # year count # <int> <dbl> # 1 2005 0 # 2 2006 0 # 3 2007 0 # 4 2008 0 # 5 2009 0 # 6 2010 0 # s1 <- dplyr::compute(s1) # likely work-around v <- '' print(s1) # print 2 # Source: query [6 x 2] # Database: spark connection master=local[4] app=sparklyr local=TRUE # # year count # <int> <chr> # 1 2005 # 2 2006 # 3 2007 # 4 2008 # 5 2009 # 6 2010 Notice s1 changed its value (like due to lazy evaluation and having captured a reference to v). version # _ # platform x86_64-apple-darwin13.4.0 # arch x86_64 # os darwin13.4.0 # system x86_64, darwin13.4.0 # status # major 3 # minor 3.2 # year 2016 # month 10 # day 31 # svn rev 71607 # language R # version.string R version 3.3.2 (2016-10-31) # nickname Sincere Pumpkin Patch Thanks for the reproducible bug report! It looks to me like this is a bug in dplyr -- the lazy evaluation performed doesn't capture the value of v when the query is generated, and so every time the query is 'rendered' it looks up what the current value of v is. Sounds good, I have copied the content over to https://github.com/hadley/dplyr/issues/2455 . Look like that it this is known in https://github.com/hadley/dplyr/issues/2370 and fixed in the dev versio of dplyr.
gharchive/issue
2017-02-20T17:56:41
2025-04-01T04:35:45.329636
{ "authors": [ "JohnMount", "kevinushey" ], "repo": "rstudio/sparklyr", "url": "https://github.com/rstudio/sparklyr/issues/503", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
57287533
PDF build failing despite output being written, and PDF cannot be downloaded My PDF build is marked as "Failed" besides output being written in the end. It says in the log: Output written on pypet.pdf (254 pages, 2253138 bytes). However, when I check the documentation (http://pypet.readthedocs.org/en/latest/) and select to download the PDF, an older version of the documentation from last year is downloaded. Which is also weird, because the build corresponding to the old version is also marked as "Failed". The failing was due to a Latex Error too deeply nested. I added the following to the sphinx latex_elements and 'preamble' option to solve this: \usepackage{enumitem} \setlistdepth{99} Downloading an older version was due to caching issues with my browser, sorry :-D So can be closed :-)
gharchive/issue
2015-02-11T08:14:36
2025-04-01T04:35:45.370617
{ "authors": [ "SmokinCaterpillar" ], "repo": "rtfd/readthedocs.org", "url": "https://github.com/rtfd/readthedocs.org/issues/1153", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
67393989
Translation projects are not found When I try to add a translation to a RTD project under Admin -> Translations I get the error: Project https://github.com/tobias47n9e/innsbruck-stereographic/tree/master/docs_int/de was not found I tried around with other URLs to see if I made a mistake, but at this point I think that the web-form has a bug. The RTD-project is here: https://readthedocs.org/dashboard/innsbruck-stereographic This isn't a bug, but it could definitely be done better. The form isn't clear about what it's asking for. To link a translation, enter the project name as it exists in read the docs -- ie. 'innsbruck-stereographic' Thank you. That turned out to be rather easy. I now also figured out how to assign different relative directories to different conf.py files. Going multilingual is not as hard as I thought it would be.
gharchive/issue
2015-04-09T16:03:30
2025-04-01T04:35:45.373356
{ "authors": [ "agjohnson", "tobias47n9e" ], "repo": "rtfd/readthedocs.org", "url": "https://github.com/rtfd/readthedocs.org/issues/1220", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
199577545
Which branch for development Hi, i'd like to add a new builder for doxygen documentation (but native, not with breath). Since there are a lot of branches like real/relcorp which a far ahead of master, i'd like to know, which branch to choose for development. Thanks in advance! Oli Thanks for your interest. We likely wouldn't include a doxygen builder in code, but we'd be happy to make it easy to implement as a third party builder, with a good API in RTD. We'd also be happy to list the extension in our docs! :) The main site is deployed from rel, which shouldn't be ahead of master. Note that there are a few sphinx extensions for integrating doxygen content, notably https://github.com/michaeljones/breathe, you might want to check that out before starting from scratch. Should I close this issue? The original question was answered (regarding to the branch) :grin: Do you want to continue the discussion about how to it here? If so, we should change the title of the issue since it will be confusing. We should probably document how we do deployments and that the rel branch is in production.
gharchive/issue
2017-01-09T15:20:26
2025-04-01T04:35:45.376531
{ "authors": [ "alyjak", "ericholscher", "humitos", "zabeloliver" ], "repo": "rtfd/readthedocs.org", "url": "https://github.com/rtfd/readthedocs.org/issues/2583", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
538099426
App Args: Add tests, fix parsing of LitInt in application arguments It appears that input.parse::<LitBool>() consumes part of the token stream, so the subsequent call to input.parse::<LitInt>() does not behave as expected. This PR has one possible fix, but there may be a better fix using fork or step (see https://docs.rs/syn/1.0.11/syn/parse/struct.ParseBuffer.html#method.fork) Add tests. Excellent! Thank you. bors r+ Excellent! Thank you. bors r+ Great 👍 Bors appears to have failed. What happened? @richardeoin ah, sorry. I forgot to change the state of the review from "needs changes" to "approved" before sending this PR to bors that's why it didn't get processed by bors.
gharchive/pull-request
2019-12-15T20:12:02
2025-04-01T04:35:45.379807
{ "authors": [ "japaric", "richardeoin" ], "repo": "rtfm-rs/rtfm-syntax", "url": "https://github.com/rtfm-rs/rtfm-syntax/pull/21", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
230514185
Support for fallback values w/ guaranteed order This looks hacky... and it is! :-) But I believe it's the simplest way to support fallback values without having to make major changes in the core and affect performance. Fixes #6 Don't worry. I won't be sad if you decline this. :-P But it clearly works. I don't love the idea of the property names in the value string, but I think this actually pretty robust. I hadn't thought of this approach before, cool stuff! 👍 There's a bit of a tradeoff here in terms of reduced de-duplication, but since this change fully guarantees order I think it's more robust implementation than what exists currently. Also, in practice, it seems that value fallbacks are often merely vendor prefixes, so this would have the benefit of collapsing these into a single atomic class (further compressing the CSS) sort of like https://github.com/rtsao/styletron/issues/104 Since this is related to https://github.com/rtsao/styletron/issues/104 I'd want to make sure this wouldn't cause any problems down the line. I'm also open to updating the core to facilitating this in a first-class fashion, since admittedly this is pretty hacky and is essentially an abuse of the interface and implementation details. What about something like: styletron.injectRawDeclaration( 'color:red', ':hover', '@media (max-width: 400px)' ); styletron.injectRawDeclaration( 'color:red;color:blue', ':hover', '@media (max-width: 400px)' ); The internal cache structure would need to be changed from: { color: {red: 'a'}, pseudos: { ':hover': { color: {red: 'b'} } } } to: { 'color:red': 'a', pseudos: { ':hover': { 'color:red': 'b' } } } I'm not sure what the performance implications would be, but in my original benchmarking the former with the original injectDeclaration was faster because there's never any string concatenation performed. My intuition is that this new cache format would be faster in the case of new styles, but slower in the case of existing styles. Probably worth doing some benchmarking. I don't love the idea of the property names in the value string LOL, you're funny. You surly would be high if you'd love it. ;-) Sure, an injectRawDeclaration is a good idea. But do you mean it should complete the original injectDeclaration with a separate cache or replace it? But do you mean it should complete the original injectDeclaration with a separate cache or replace it? I think the new cache format could just replace the original, then both injectDeclaration and injectRawDeclaration could shared the same cache. Additionally, I think at this point it might be worth deprecating the injectDeclaration method, then removing it from the next major version. Nobody really interacts with this low-level API directly so I don't see much value in having two different methods -- and since injectRawDeclaration is lower-level, we might as well support that one only. I've added injectRawDeclaration to: styletron-core styletron-server styletron-client And injectStyle now uses injectRawDeclaration instead with full support of fallback values. Merged with #145
gharchive/pull-request
2017-05-22T20:45:25
2025-04-01T04:35:45.401641
{ "authors": [ "faddee", "rtsao" ], "repo": "rtsao/styletron", "url": "https://github.com/rtsao/styletron/pull/133", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1688274380
Do i need to write a windows driver to be able to talk to my custom USB HID device? Im using an ESP32S2 which has native usb support and i want to be able to send reports between a rust program and my custom usb device but so far the only way i can read data from the device is on linux when i give sudo permissions to the device path. In windows i can open the device but when i send a report the rust program cant read it at all. so far the only way i can read data from the device is on linux when i give sudo permissions to the device path Not sure about the Windows question, but on Linux this is usually handled with an udev rule, e.g.: ACTION=="add|change", SUBSYSTEMS=="usb", ATTRS{idVendor}=="04d9", ATTRS{idProduct}=="a052", MODE:="0666" or ACTION=="add|change", SUBSYSTEMS=="usb", ATTRS{idVendor}=="04d9", ATTRS{idProduct}=="a052", GROUP:="uucp" If I think about it, that CO2 meter sample used to work fine for me on Windows, so it might be something specific to your device. Where could i find that code sample?
gharchive/issue
2023-04-28T10:02:33
2025-04-01T04:35:45.427198
{ "authors": [ "Aggeloz", "lnicola" ], "repo": "ruabmbua/hidapi-rs", "url": "https://github.com/ruabmbua/hidapi-rs/issues/112", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
739831733
RSpec/FactoryBot/CreateList fails with "undefined method `to_sym' for nil:NilClass" The Rspec/FactoryBot/CreateList fails with undefined method to_sym for nil:NilClass. It fails for both styles (create_list, n_times). Mention the following information in the issue report: 1.2.0 (using Parser 2.7.2.0, rubocop-ast 1.1.1, running on jruby 2.5.7 java) The following versions are used: JRuby 9.2.13.0 rubocop 1.2.0 rubocop-rspec 1.42.0 Backtrace: org/jruby/RubyKernel.java:1009:in `load' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/bin/rubocop:23:in `<main>' org/jruby/RubyKernel.java:1009:in `load' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/cli/exec.rb:63:in `kernel_load' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/cli/exec.rb:28:in `run' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/cli.rb:476:in `exec' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/vendor/thor/lib/thor/command.rb:27:in `run' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/vendor/thor/lib/thor/invocation.rb:127:in `invoke_command' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/vendor/thor/lib/thor.rb:399:in `dispatch' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/cli.rb:30:in `dispatch' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/vendor/thor/lib/thor/base.rb:476:in `start' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/cli.rb:24:in `start' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/exe/bundle:46:in `block in <main>' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/lib/bundler/friendly_errors.rb:123:in `with_friendly_errors' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/bundler-2.1.4/exe/bundle:34:in `<main>' org/jruby/RubyKernel.java:1009:in `load' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/bin/bundle:23:in `<main>' An error occurred while RSpec/FactoryBot/CreateList cop was inspecting /Users/someuser/dev/someproject/spec/support/helper.rb:12:17. undefined method `to_sym' for nil:NilClass /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-1.2.0/lib/rubocop/cop/mixin/configurable_enforced_style.rb:69:in `style' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-rspec-1.42.0/lib/rubocop/cop/rspec/factory_bot/create_list.rb:61:in `on_send' org/jruby/RubyKernel.java:1960:in `public_send' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-1.2.0/lib/rubocop/cop/commissioner.rb:100:in `block in trigger_responding_cops' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-1.2.0/lib/rubocop/cop/commissioner.rb:166:in `with_cop_error_handling' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-1.2.0/lib/rubocop/cop/commissioner.rb:99:in `block in trigger_responding_cops' org/jruby/RubyArray.java:1809:in `each' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-1.2.0/lib/rubocop/cop/commissioner.rb:98:in `trigger_responding_cops' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-1.2.0/lib/rubocop/cop/commissioner.rb:69:in `on_send' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-ast-1.1.1/lib/rubocop/ast/traversal.rb:158:in `block in on_send' org/jruby/RubyArray.java:1809:in `each' org/jruby/RubyEnumerable.java:1258:in `each_with_index' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-ast-1.1.1/lib/rubocop/ast/traversal.rb:155:in `on_send' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-1.2.0/lib/rubocop/cop/commissioner.rb:71:in `on_send' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-ast-1.1.1/lib/rubocop/ast/traversal.rb:153:in `on_def' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-1.2.0/lib/rubocop/cop/commissioner.rb:71:in `on_def' /Users/someuser/.rbenv/versions/jruby-9.2.13.0/lib/ruby/gems/shared/gems/rubocop-ast-1.1.1/lib/rubocop/ast/traversal.rb:136:in `block in on_dstr' Line 69 from the rubocop mixin ConfigurableEnforcedStyle and surroundings: def style @style ||= begin s = cop_config[style_parameter_name].to_sym raise "Unknown style #{s} selected!" unless supported_styles.include?(s) s end end ... def style_parameter_name 'EnforcedStyle' end => Looks like cop_config[style_parameter_name] is nil. I think the issue is, that the cop_config only consists of: { "Enabled" => true } Working example: { "Description" => "Use Unix-style line endings.", "StyleGuide" => "#crlf", "Enabled" => true, "VersionAdded" => "0.49", "EnforcedStyle" => "native", "SupportedStyles" => ["native", "lf", "crlf"] }" rubocop-rspec 1.42 is not compatible with RuboCop 1.2. Please update to a recently released rubocop-rspec 2.0. @pirj thank you for pointing my to the 2.0 version. Somehow missed it.
gharchive/issue
2020-11-10T11:33:30
2025-04-01T04:35:45.456234
{ "authors": [ "ThomasKoppensteiner", "pirj" ], "repo": "rubocop-hq/rubocop-rspec", "url": "https://github.com/rubocop-hq/rubocop-rspec/issues/1086", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
623634184
Fix a false positive for Lint/SuppressedException This PR fixes a false positive for Lint/SuppressedException when empty rescue with comment in def. % cat example.rb def foo do_something rescue # noop end % bundle exec rubocop --only Lint/SuppressedException (snip) Inspecting 1 file W Offenses: example.rb:3:1: W: Lint/SuppressedException: Do not suppress exceptions. rescue ^^^^^^ 1 file inspected, 1 offense detected The above code should not be warned because AllowComments: true by defualt. Before submitting the PR make sure the following are checked: [x] Wrote good commit messages. [x] Commit message starts with [Fix #issue-number] (if the related issue exists). [x] Feature branch is up-to-date with master (if not - rebase it). [x] Squashed related commits together. [x] Added tests. [x] Added an entry to the Changelog if the new code introduces user-observable changes. See changelog entry format. [x] The PR relates to only one subject with a clear title and description in grammatically correct, complete sentences. [x] Run bundle exec rake default. It executes all tests and RuboCop for itself, and generates the documentation. Looking forward to this 🎉 Can you make sure that inline comments are also ok? (See #7999) E.g.: def foo bar rescue # ignore end Certainly! Also the following code seems to be able to do the same. begin bar rescue # ignore end I think that these cases can be extended separately from this PR.
gharchive/pull-request
2020-05-23T10:57:21
2025-04-01T04:35:45.462291
{ "authors": [ "koic", "marcandre" ], "repo": "rubocop-hq/rubocop", "url": "https://github.com/rubocop-hq/rubocop/pull/8017", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
713391809
Fix #8820 autocorrection for Style/IfWithSemicolon This change fixes autocorrection for Style/IfWithSemicolon when elsif present, which was reported on #8820 In case of one of more elsif conditions present, the autocorrector will create a full if, elsif, else structure. Before submitting the PR make sure the following are checked: [x] Wrote good commit messages. [x] Commit message starts with [Fix #issue-number] (if the related issue exists). [x] Feature branch is up-to-date with master (if not - rebase it). [x] Squashed related commits together. [x] Added tests. [x] Added an entry to the Changelog if the new code introduces user-observable changes. See changelog entry format. [x] The PR relates to only one subject with a clear title and description in grammatically correct, complete sentences. [x] Run bundle exec rake default. It executes all tests and RuboCop for itself, and generates the documentation. @adrian-rivera ping :-) @dvandersluis Would you be interested in taking over this PR and driving it over the finish line? Sure! I'll take a look in the morning @bbatsov I don't have access to push to the existing branch, should I create a new one on my fork and open a new PR? @bbatsov this one should be closed now 😄 Thanks!
gharchive/pull-request
2020-10-02T06:45:08
2025-04-01T04:35:45.467828
{ "authors": [ "adrian-rivera", "bbatsov", "dvandersluis" ], "repo": "rubocop-hq/rubocop", "url": "https://github.com/rubocop-hq/rubocop/pull/8831", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2026522980
Make the Rubocop::Minitest::Test suite re-usable by gem consumers I've been writing custom cops for various jobs for a while now, but I've never had a consistent way to write tests for products that use minitest as the test suite (officially Rubocop only provides rspec helpers) , I started to look around to see what this Gem did and figured it would make sense to make rubocop-minitest provide a minitest class for others that write custom cops for their own code bases. This does a couple of things Makes a new Rubocop::Minitest::Test class that includes the code from assert_offense (I don't feel like it needs to be a module), and it also uses declarative testing for people to use. I moved the "real" test suites in this gem to use the new base class, and kept the code blocks being tested as Minitest::Test I tested this locally in an app of mine and it works great! So it can allow a consistent way for people to write custom cops and easily test them! Before submitting the PR make sure the following are checked: [x] The PR relates to only one subject with a clear title and description in grammatically correct, complete sentences. [x] Wrote good commit messages. [-] Commit message starts with [Fix #issue-number] (if the related issue exists). [x] Feature branch is up-to-date with master (if not - rebase it). [x] Squashed related commits together. [x] Added tests. [x] Ran bundle exec rake default. It executes all tests and runs RuboCop on its own code. [x] Added an entry (file) to the changelog folder named {change_type}_{change_description}.md if the new code introduces user-observable changes. See changelog entry format for details. This makes a lot of sense to me and I'd love to use it. If this is being exposed like this then I think it needs some documetation on how to pass global and cop specific config to a test. See https://github.com/rubocop/rubocop-minitest/blob/b446022ea09b3f5558df9c0106c9e714c6fc1ec5/test/rubocop/cop/minitest/global_expectations_test.rb#L504-L513 and https://github.com/rubocop/rubocop-minitest/blob/b446022ea09b3f5558df9c0106c9e714c6fc1ec5/test/rubocop/cop/minitest/multiple_assertions_test.rb#L339-L343 @Earlopain Updated the casing of RuboCop, updated the sample docs (both inline and README.md), handling the nil case for @cop Squashed commits and confirmed CI is passing locally @koic 👋 Could I ask for a review? @koic it's been a while, but I'd still like to get this merged It's been a while since I last looked at this and this may be a dumb question to only come now but can you explain why this is needed? Specifically I think there already exists what you are looking for: https://github.com/rubocop/rubocop-minitest/issues/155 and https://github.com/rubocop/rubocop-minitest/pull/156 https://github.com/rubocop/rubocop-minitest/blob/master/lib/rubocop/minitest/support.rb There may be concern about it polluting the main test class but even then you can just make your own and mix that module in yourself.
gharchive/pull-request
2023-12-05T15:50:29
2025-04-01T04:35:45.477476
{ "authors": [ "DougEdey", "Earlopain", "exterm" ], "repo": "rubocop/rubocop-minitest", "url": "https://github.com/rubocop/rubocop-minitest/pull/278", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1594533112
Add new InternalAffairs/ProcessedSourceBufferName cop Follow up https://github.com/rubocop/rubocop-minitest/pull/217#discussion_r1063097686. This PR adds new InternalAffairs/ProcessedSourceBufferName cop. It enforces the use of processed_source.file_path instead of processed_source.buffer.name. # bad processed_source.buffer.name # good processed_source.file_path Before submitting the PR make sure the following are checked: [x] The PR relates to only one subject with a clear title and description in grammatically correct, complete sentences. [x] Wrote good commit messages. [ ] Commit message starts with [Fix #issue-number] (if the related issue exists). [x] Feature branch is up-to-date with master (if not - rebase it). [x] Squashed related commits together. [x] Added tests. [x] Ran bundle exec rake default. It executes all tests and runs RuboCop on its own code. [ ] Added an entry (file) to the changelog folder named {change_type}_{change_description}.md if the new code introduces user-observable changes. See changelog entry format for details. Looks good. Thanks!
gharchive/pull-request
2023-02-22T06:25:46
2025-04-01T04:35:45.482651
{ "authors": [ "bbatsov", "koic" ], "repo": "rubocop/rubocop", "url": "https://github.com/rubocop/rubocop/pull/11606", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
427335871
Expire/Revalidate a Partner A diaper bank needs to be able to expire a partner and get them to revalidate their information. On the partner show page there should be some sort of "expire partner" or similar button which when pushed, sets the partner into needs recertification. It will send a request to the partnerbase app setting their status to recertification required and preventing them from requesting more diapers. After the partner updates any new information they can submit for approval which changes their status on the diaperbank side ot awaiting review and it will follow as expected. Hi @seanmarcia, @chaserx and I will work on this. Thank you for describing the issue so clearly. Would you mind answering my clarification questions below please? Requested changes A diaper bank will be able to expire a partner who is currently approved (not a partner who is pending or awaiting approval)? Are we happy with the words "expire partner" for the button or is there anything else you would prefer? Are we happy with the status name "recertification_required" or would you prefer something else? Once the partner's status has been changed to "recertification_required", they will still be able to see their historical diaper requests, but the button "Create new diaper request" will not have any functionality. Does this sound right to you? Potential additional functionalities Would the partner, which has received the status "recertification_required" need some kind of alert (an email perhaps) telling them this? Are there other things that should trigger the "recertification_required" status in addition to the Diaper Bank choosing this? Eg something we might want to automate. I'm not sure of the exact reasons why recertification would be required, but wondering if there's something we already track which makes sense to trigger this status. Is there any particular field that each partner requiring recertification will be required to update in their application? How will they know this? Again, I'm not sure of the exact use case so apologies if this question or any of the above are redundant. Thank you very much 😃 Hey Amina! I'm excited you are continuing to contribute! Let me try and answer your questions! A diaper bank will be able to expire a partner who is currently approved (not a partner who is pending or awaiting approval)? Correct! Are we happy with the words "expire partner" for the button or is there anything else you would prefer? Nice catch! I think "expire partner" probably conveys the wrong message. How about "re-certify partner" Are we happy with the status name "recertification_required" or would you prefer something else? Sounds great! Once the partner's status has been changed to "recertification_required", they will still be able to see their historical diaper requests, but the button "Create new diaper request" will not have any functionality. Does this sound right to you? Yep, that sounds accurate! Potential additional functionalities 5) Would the partner, which has received the status "recertification_required" need some kind of alert (an email perhaps) telling them this? This is a super good idea! I think there are a few ways this could be handled and I'm not sure if I have a preference. You could have the partner app send them an email letting them know that they need to do a routine recertification of their details so the diaper bank is up to date. You could make a big brightly coloured banner that would show up when they log in. Or both :) 6) Are there other things that should trigger the "recertification_required" status in addition to the Diaper Bank choosing this? Eg something we might want to automate. I'm not sure of the exact reasons why recertification would be required, but wondering if there's something we already track which makes sense to trigger this status. This is also an excellent idea! The idea behind this is that diaper banks want up to date information on their partners for reporting purposes. It may make sense to have a field and flag in the organization page where they can set a yearly(?) partner recertification date? -- I'd definitely want to check with some of the the diaper bank stakeholders before going forward with this. 7) Is there any particular field that each partner requiring recertification will be required to update in their application? How will they know this? Again, I'm not sure of the exact use case so apologies if this question or any of the above are redundant. Definitely not redundant! I love the questions and thought you've put into all of this! I don't think a partner will necessarily need to update any fields, if all their organization details haven't changed they can probably just click the button asking to be recertified. Thanks again for all the amazing work you've done and continue to do! Hey @Nirvikalpa108 and @chaserx Is this completed? My apologies if you've been waiting on a review! I'll get to it shortly if it is ready! (Here and on the Partner side!) Hi @seanmarcia yeah. It fell of my radar for a bit, but. Yes, I think that it's ready for review. @mdworken has weighed. I think I've addressed their concerns.
gharchive/issue
2019-03-30T22:43:38
2025-04-01T04:35:45.524926
{ "authors": [ "Nirvikalpa108", "chaserx", "seanmarcia" ], "repo": "rubyforgood/diaper", "url": "https://github.com/rubyforgood/diaper/issues/773", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1644733012
3429 reinviting accepted partners Checklist: Checked, although I didn't add documentation Resolves #3429 Description This makes it so that if someone has accepted an invitation to the organization, a new one is not sent out when clicking on "Create User" in the partner coworker screen. The new behavior is being redirected to the coworker screen and a message appearing saying "(name) has already joined the organization Pros: does intended behavior Cons: added complexity and an additional check for the user before the invitation is sent no dependencies Type of change New feature (non-breaking change which adds functionality) This change requires a documentation update (I'm leaving this because I'm not sure it doesn't) How Has This Been Tested? Send an invitation to a new user In a separate browser, open the invitation link and fill out the forms to sign in to the new account (this Is what counts as "accepting" the invitation, and should maybe change?) In the original browser, invite the same person again (the email is what matters for identification) verify that the email isn't sent and a message says "(name) has already joined the organization" Screenshots @cielf What are the next steps from here? Thanks for the help thus far. @cielf would this still be an issue with the changes in #3372 ? I think this might conflict with that. Hmm, the commit with the test didn't get pushed apparently. Before I update the pull request, what simplifications did you have in mind? I've found that all of the logic that is different from before the fix is necessary for the fix to work. Hmm, the commit with the test didn't get pushed apparently. Before I update the pull request, what simplifications did you have in mind? I've found that all of the logic that is different from before the fix is necessary for the fix to work. Hey Technical, I am just a fellow contributor! The conversation was quiet so I figured I'd drop my opinion. There's probably a good way to remove the nested if statement, but I only quickly skimmed the code. Good luck! @technic-void my question was more for @cielf because the concept of "re-inviting a partner user" has changed somewhat in the most recent update of a different PR - I'm not sure this PR is actually needed if the other one is merged. Hmm. I think they are two different things. 3429 says "If they've already accepted an invitation to this partner, you shouldn't be inviting them again." Whereas with 3372, we're talking about a new partner/user relationship. @cielf but in the other PR, we no longer send invites if the user already exists in the DB (regardless of if they belong to the current partner or a different partner, or an organization). So wouldn't that cover this? See also the comments on the other PR Adding this to next week's planning discussion in case we aren't both on the same page, and know it in the meantime. I just checked the behaviour on main, and it does appear that this has been overtaken by other work. Thank you very much for the contribution, @technic-void ! We're going to close it, though.
gharchive/pull-request
2023-03-28T22:20:06
2025-04-01T04:35:45.533405
{ "authors": [ "cielf", "dorner", "lokisk1155", "technic-void" ], "repo": "rubyforgood/human-essentials", "url": "https://github.com/rubyforgood/human-essentials/pull/3507", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2421330688
874 Add strong_migrations gem 🔗 Issue Resolves https://github.com/rubyforgood/pet-rescue/issues/874 ✍️ Description Added strong_migrations gem 📷 Screenshots/Demos You know what, I will review a couple PRs now that are close to the finish line. I recall one has migrations. I will let you know once that is done/merged, so we can update the config start_after date in this PR. Just saves people having to re-write migrations in PRs for now. Re-commenting here because in the past people have not been notified if I request changes... @egemen-dev I merged two PRs with migrations. Can you please update the config file to use the latest migration on main branch as the start point? @kasugaijin updated the start_after value with the latest migration timestamp from the main branch.
gharchive/pull-request
2024-07-21T10:13:42
2025-04-01T04:35:45.537810
{ "authors": [ "egemen-dev", "kasugaijin" ], "repo": "rubyforgood/pet-rescue", "url": "https://github.com/rubyforgood/pet-rescue/pull/883", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
392460284
How can I produce the test bitcoin? I want to do a trade test between my 2 accounts. But I don't have bitcoin and money. How can I produce the test bitcoin? @UfukAntep You need to deposit BTC through faucet to this accounts.
gharchive/issue
2018-12-19T06:35:25
2025-04-01T04:35:45.579509
{ "authors": [ "UfukAntep", "mnaichuk" ], "repo": "rubykube/peatio", "url": "https://github.com/rubykube/peatio/issues/1909", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2249461829
Arrows don't work as expected in Firefox/Archlinux Arrows UP/DOWN/LEFT/RIGHT, don't Move cell selection box. Instead, they change cell to edit mode. Browser: Firefox 124.0.1 OS: Archlinux, Gnome (45 & 46) In Chromium works as expected. Cheers. hi, thanks for reporting this. i dont have a copy of arch with me to test this, will check if this is the same on firefox on windows/mac. would it be possible for you to attach a screenrecording from https://ruilisi.github.io/fortune-sheet-demo/?path=/story/features--basic showing this behaviour? upgraded to 19.4 and works correctly
gharchive/issue
2024-04-17T23:47:33
2025-04-01T04:35:45.634578
{ "authors": [ "pezr", "sanchit3008" ], "repo": "ruilisi/fortune-sheet", "url": "https://github.com/ruilisi/fortune-sheet/issues/537", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2286928541
create llama@latest community template nextjs-llama3 using create-llama@0.1.1 1 week before it was working now i received errors when calling replicate api, no changes to template just adding api keys and running using "llamaindex": "0.3.8" using llamaindex 0.2.10 Had same problem with other class SimpleDirectoryReader. Attempted import error: 'SimpleDirectoryReader' is not exported from 'llamaindex' (imported as 'SimpleDirectoryReader'). Did i do something wrong with the import? I tried these but same results: import { SimpleDirectoryReader } from "llamaindex/readers/SimpleDirectoryReader"; import { SimpleDocumentStore, VectorStoreIndex ,SimpleDirectoryReader} from "llamaindex"; @jte0711 This is happening with llamaindex 0.2.10?
gharchive/issue
2024-05-09T05:35:15
2025-04-01T04:35:45.640911
{ "authors": [ "jozemario", "jte0711", "marcusschiesser" ], "repo": "run-llama/create-llama", "url": "https://github.com/run-llama/create-llama/issues/78", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
975673035
Quality of Life for Mouse Click Counter Plugin Could it be possible to add an option in the settings for the Mouse Click Counter plugin that it resets upon every Login? It would be nice to see how many clicks you do in 1 session. I know you can manually reset it for this but having it automated would be nice. In addition to the above, could you take it one step further to export the number of clicks per session into a notepad? Something like; Login: 8/20/21 / 1400-1634 / 42124 Date / Time / Clicks The Mouse Click Counter plugin is a hub plugin, and not maintained by the RuneLite team. Please open this request on their repository instead.
gharchive/issue
2021-08-20T14:41:35
2025-04-01T04:35:45.660745
{ "authors": [ "Nightfirecat", "soccerplayr" ], "repo": "runelite/runelite", "url": "https://github.com/runelite/runelite/issues/14047", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
327869002
Show what needs repairing and what doors needs closing in Pest control Also some other optimisations would be cool to see. Like if something needs repairing, what doors need closing, what portals are weak against, total points, maybe even set a goal point amount and see your progress towards it. Please let me know if these need to be separate issues. Can I get this assigned? Also, many people think closing doors wastes time. I'm up for discussion on this. It can easily be added because I already have the functionality to highlight all repairables. Oh and because the points will be visible when in the minigame and on the main island I don't see why a goal feature would fit here. It might be better to add goals to something like #7912
gharchive/issue
2018-05-30T19:33:10
2025-04-01T04:35:45.662370
{ "authors": [ "RickertdeH", "yanikore" ], "repo": "runelite/runelite", "url": "https://github.com/runelite/runelite/issues/3500", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
338846256
Suggestion It'd be cool if you could add a way to reset the max default zoom option when expand outer zoom is enabled. I like being able to zoom out more, but also like to return to the normal default zoom limit without having to turn off the zoom plugin. Sounds like a duplicate of #919.
gharchive/issue
2018-07-06T08:14:12
2025-04-01T04:35:45.663319
{ "authors": [ "Nightfirecat", "Runescaper123" ], "repo": "runelite/runelite", "url": "https://github.com/runelite/runelite/issues/4197", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
302164804
Add Superantipoison to Timer Plugin Noticed today that superantipoison is not supported and it would be nice to have it added please. Resolved in #1159
gharchive/issue
2018-03-05T04:00:05
2025-04-01T04:35:45.664117
{ "authors": [ "Caia-OSRS", "deathbeam" ], "repo": "runelite/runelite", "url": "https://github.com/runelite/runelite/issues/827", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
466085230
mta: fix Telekinetic solver not moving The bug happened when the Telekinetic projectile existed on the same game tick as the guardian moving. Closes #3946 AFAICT this is due to a race with the projectiles being culled at the time the npc update moves the guardian. I can find no reason why this logic existed in the first place.
gharchive/pull-request
2019-07-10T03:56:04
2025-04-01T04:35:45.665358
{ "authors": [ "Adam-", "raiyni" ], "repo": "runelite/runelite", "url": "https://github.com/runelite/runelite/pull/9328", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
2188433512
Usages of klabel{}(_) attribute remain When working on https://github.com/runtimeverification/k/issues/4035, I noticed that one test exercising the legacy Haskell backend failed after making the mechanical klabel(_), symbol refactorings implied in that issue: https://github.com/runtimeverification/k/tree/master/k-distribution/tests/regression-new/issue-1676-koreBytes The error from the backend was: $ make issue-1676-koreBytes set -o pipefail; (cat 1.test.in 2>/dev/null || true) | /Users/brucecollie/code/k/k-distribution/bin/krun 1.test --output kore --no-exc-wrap --definition ./test-kompiled | diff - 1.test.out kore-exec: [147057] Error (ErrorException): Expecting hook 'BYTES.int2bytes' to reduce concrete pattern: /* T Fn D Spa */ LblInt2Bytes'LParUndsCommUndsCommUndsRParUnds'BYTES-HOOKED'Unds'Bytes'Unds'Int'Unds'Int'Unds'Endianness{}( /* T Fn D Sfa Cl */ \dv{SortInt{}}("0"), /* T Fn D Sfa Cl */ \dv{SortInt{}}("0"), /* T Fn D Sfa Cl */ LbllittleEndianBytes{}() ) CallStack (from HasCallStack): error, called at src/Kore/Rewrite/Axiom/EvaluationStrategy.hs:193:26 in kore-0.60.0.0-DddjtXjEFiY3CtblfXDv9H:Kore.Rewrite.Axiom.EvaluationStrategy Expecting hook 'BYTES.int2bytes' to reduce concrete pattern: /* T Fn D Spa */ LblInt2Bytes'LParUndsCommUndsCommUndsRParUnds'BYTES-HOOKED'Unds'Bytes'Unds'Int'Unds'Int'Unds'Endianness{}( /* T Fn D Sfa Cl */ \dv{SortInt{}}("0"), /* T Fn D Sfa Cl */ \dv{SortInt{}}("0"), /* T Fn D Sfa Cl */ LbllittleEndianBytes{}() ) CallStack (from HasCallStack): error, called at src/Kore/Rewrite/Axiom/EvaluationStrategy.hs:193:26 in kore-0.60.0.0-DddjtXjEFiY3CtblfXDv9H:Kore.Rewrite.Axiom.EvaluationStrategy Created bug report: kore-exec.tar.gz [Error] krun: kore-exec ./test-kompiled/definition.kore --module TEST --pattern /var/folders/8r/bzvz46gd0r5_fr11jt23fzyw0000gn/T/.krun-2024-03-15-12-34-50-bXtH0QYDVx/tmp.in.2uQgOtQkNg --output /var/folders/8r/bzvz46gd0r5_fr11jt23fzyw0000gn/T/.krun-2024-03-15-12-34-50-bXtH0QYDVx/result.kore [Error] krun: Backend crashed during rewriting with exit code 1 0a1,16 > /* T Fn D Sfa Cl */ > Lbl'-LT-'generatedTop'-GT-'{}( > /* T Fn D Sfa Cl */ > Lbl'-LT-'k'-GT-'{}( > /* T Fn D Sfa Cl */ > kseq{}( > /* T Fn D Sfa Cli */ > /* Inj: */ inj{SortBytes{}, SortKItem{}}( > /* T Fn D Sfa Cl */ \dv{SortBytes{}}("") > ), > /* T Fn D Sfa Cl */ dotk{}() > ) > ), > /* T Fn D Sfa Cl */ > Lbl'-LT-'generatedCounter'-GT-'{}(/* T Fn D Sfa Cl */ \dv{SortInt{}}("0")) > ) make: *** [1.test] Error 1 Some digging into the code reveals that this is because the backend treats Endianness and Signedness terms as special-cased builtins, but because these are not K hooked symbols, the backend needs to inspect the klabel{}(_) attribute to check for special-casing. I have opened a draft fix in https://github.com/runtimeverification/haskell-backend/pull/3741 that simply looks for symbol'Kywd' instead of klabel in these cases (the only such places klabel is used). With this fix applied, I can pass the K regression test suite. However, there is now a dependency loop between the frontend and backend that we need to resolve to merge the fix. If there's a way to first look at klabel, then symbol'Kywd' if there's no klabel, then I think that would address the issue. We would then be able to update the K dependency for the integration tests and immediately remove the backwards compatibility. As an update, I opened this Frankenstein PR: https://github.com/runtimeverification/haskell-backend/pull/3744 It demonstrates that if https://github.com/runtimeverification/k/pull/4045 is merged along with https://github.com/runtimeverification/haskell-backend/pull/3741, that the integration test suite here is all successful. We still need to figure out exactly how to close this loop though.
gharchive/issue
2024-03-15T12:43:09
2025-04-01T04:35:45.695618
{ "authors": [ "Baltoli" ], "repo": "runtimeverification/haskell-backend", "url": "https://github.com/runtimeverification/haskell-backend/issues/3742", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2179140573
cheatcodes.md: use buf instead of bufStrict in freshUInt closes: https://github.com/runtimeverification/evm-semantics/issues/2038 It does save us a simplification step, since that #bufStrict will reduce to #buf. I wonder if we could drop [preserves-definedness] now...
gharchive/pull-request
2024-03-11T13:29:11
2025-04-01T04:35:45.697450
{ "authors": [ "PetarMax", "anvacaru" ], "repo": "runtimeverification/kontrol", "url": "https://github.com/runtimeverification/kontrol/pull/430", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
529335776
Docs for the actual API? The docs only seem to explain how to use the example UI. I imagine most people are only interested in the provided UI as an example. What they will really want to do is apply those techniques to their own apps. I'm in the process of going through the code in RunwayHub to figure out where the useful stuff happens. It would be really helpful if this project contained a minimal example without all the cruft. Maybe a Monobehaviour that took a Texture2D input and renderered to another Texture2D? It's a bit annoying to have to deconstruct a fairly complex example just to do the most basic thing and a minimal example would be a real time saver for others. Wow. Six months and no reply?
gharchive/issue
2019-11-27T13:08:31
2025-04-01T04:35:45.705799
{ "authors": [ "andybak" ], "repo": "runwayml/RunwayML-for-Unity", "url": "https://github.com/runwayml/RunwayML-for-Unity/issues/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
422965366
Pr.wyrm I've followed this guide step by step carefully. Yes I've tested this package remotely and have excluded all broken builds. Here is the links to the Travis/AppVeyor with status "All passed": https://ci.appveyor.com/project/a-sevin/hunter/builds/23203978 https://travis-ci.org/a-sevin/hunter/builds/508609964 I have submitted CI configs to https://github.com/ingenue/hunter targeting pkg.template branch, see this merged pull request https://github.com/ingenue/hunter/pull/348 Testing: https://ci.appveyor.com/project/ingenue/hunter/builds/23221977 https://travis-ci.org/ingenue/hunter/builds/508924280?utm_source=github_status&utm_medium=notification update wyrm package to add pybind11 dependency https://travis-ci.org/a-sevin/hunter/builds/508964942 https://ci.appveyor.com/project/a-sevin/hunter/builds/23224246 Testing: https://ci.appveyor.com/project/ingenue/hunter/builds/23224721 https://travis-ci.org/ingenue/hunter/builds/508972654?utm_source=github_status&utm_medium=notification Uploading: https://ci.appveyor.com/project/ingenue/hunter/builds/23226072 https://travis-ci.org/ingenue/hunter/builds/508991900?utm_source=github_status&utm_medium=notification Released: https://github.com/ruslo/hunter/releases/tag/v0.23.141
gharchive/pull-request
2019-03-19T21:36:28
2025-04-01T04:35:45.735522
{ "authors": [ "a-sevin", "ruslo" ], "repo": "ruslo/hunter", "url": "https://github.com/ruslo/hunter/pull/1790", "license": "bsd-2-clause", "license_type": "permissive", "license_source": "bigquery" }
453600110
Update 'ethash' to v0.5.0 I've followed this guide step by step carefully. [Yes] I've tested this package remotely and have excluded all broken builds. Here is the links to the Travis/AppVeyor with status "All passed": https://ci.appveyor.com/project/chfast/hunter/builds/25124748 https://travis-ci.org/chfast/hunter/builds/542848354 Testing: https://ci.appveyor.com/project/ingenue/hunter/builds/25151795 https://travis-ci.org/ingenue/hunter/builds/543424429?utm_source=github_status&utm_medium=notification Released: https://github.com/ruslo/hunter/releases/tag/v0.23.190
gharchive/pull-request
2019-06-07T16:37:49
2025-04-01T04:35:45.740094
{ "authors": [ "chfast", "ruslo" ], "repo": "ruslo/hunter", "url": "https://github.com/ruslo/hunter/pull/1893", "license": "bsd-2-clause", "license_type": "permissive", "license_source": "bigquery" }
1408488835
Submit PR per: https://github.com/microsoft/vscode/issues/78013 Wondering if you submitted the PR: https://github.com/microsoft/vscode/issues/78013 Thanks. I never got around to it, as I've been using Jetbrains stuff over VS Code lately. Maybe someday, or feel free to take a stab at it. Alas, I don't have the expertise. Thanks for the very quick response.
gharchive/issue
2022-10-13T22:17:57
2025-04-01T04:35:45.744500
{ "authors": [ "HarjitSi", "russelldavis" ], "repo": "russelldavis/vscode-better-pageup-pagedown", "url": "https://github.com/russelldavis/vscode-better-pageup-pagedown/issues/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2473035106
private_key_debug_is_obfuscated test is in the wrong crate This test belongs to secp256k1, not bitcoin. The current situation is fragile and annoying. We can probably delete the tests if https://github.com/rust-bitcoin/rust-secp256k1/pull/722 merges. can i take on this issue ? We can probably delete the tests if rust-bitcoin/rust-secp256k1#722 merges. It looks like the PR has been merged, so now it's just a matter of removing private_key_debug_is_obfuscated? Yep, let's do it. Looks like this is can be closed - fixed by #3517 Thanks man, its rad that you notice these things.
gharchive/issue
2024-08-19T10:38:33
2025-04-01T04:35:45.754558
{ "authors": [ "Kixunil", "apoelstra", "diogocanut", "mahmudsudo", "shinghim", "tcharding" ], "repo": "rust-bitcoin/rust-bitcoin", "url": "https://github.com/rust-bitcoin/rust-bitcoin/issues/3186", "license": "CC0-1.0", "license_type": "permissive", "license_source": "github-api" }
2244571368
Allow deprecated Params field I'm not sure why I haven't see this before during the whole test cycle but while running cargo kani --only-codegen we get a bunch of warnings of form: warning: use of deprecated field consensus::params::Params::pow_limit We deprecated the pow_limit field but still set it (obviously) in const structs - just shoosh the warning. Found while investigating the current kani CI failures. Rebased to pick up the kani fix on master. cross test fail is network related, I"ll re-run it once the pipeline finishes.
gharchive/pull-request
2024-04-15T20:37:44
2025-04-01T04:35:45.756576
{ "authors": [ "tcharding" ], "repo": "rust-bitcoin/rust-bitcoin", "url": "https://github.com/rust-bitcoin/rust-bitcoin/pull/2687", "license": "CC0-1.0", "license_type": "permissive", "license_source": "github-api" }
787590844
FreeBSD: add libssp This is required to build some native projects which is build with -fstack-protector when linked. Also update sqlite3 version (quarterly is updated) Needs a rebase. bors r+
gharchive/pull-request
2021-01-17T01:26:44
2025-04-01T04:35:45.768745
{ "authors": [ "junhochoi", "reitermarkus", "svenstaro" ], "repo": "rust-embedded/cross", "url": "https://github.com/rust-embedded/cross/pull/511", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
704881938
WIP: First steps rewrite A rewrite of chapter 04 and partially chapter 05. I'm already uploading this now since the cargo-embed GDB stub is blowing up on my machine when trying to debug this chip so I'll probably have to spend some time debugging, however during that time my current work can already be reviewed so feel free to take a look at it! :warning: Warning :warning: Pull requests are usually filed against the master branch for this repo, but this one is against rewrite. Please double check that you specified the right target! I left a few notes asking for resources to link to in chapter 5 in the code, would be nice if someone could come up with some suggestions! The PR now contains everything that is required for ch04 and 05 and should be ready for review. Once review of the last parts passes the only thing we'd need is a nice resource for LED matrices and could actually merge this \o/
gharchive/pull-request
2020-09-19T13:21:50
2025-04-01T04:35:45.771084
{ "authors": [ "hargoniX", "rust-highfive" ], "repo": "rust-embedded/discovery", "url": "https://github.com/rust-embedded/discovery/pull/271", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
370750536
Make digital traits fallible This makes the digital traits fallible as discussed in #97, #95, #100 and #41 implemented as a clean break. I'm fine with merging this, but there's discussion in #92 about releasing a last 0.2.x release before merging breaking changes. Closing in favor of #108 as agreed on #100.
gharchive/pull-request
2018-10-16T18:52:10
2025-04-01T04:35:45.772514
{ "authors": [ "eldruin", "hannobraun" ], "repo": "rust-embedded/embedded-hal", "url": "https://github.com/rust-embedded/embedded-hal/pull/105", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
304730516
Using Deref in traits Currently the guidelines say "Only smart pointers implement Deref and DerefMut (C-DEREF)". In this PR in rand we are considering using this trick: impl<R: RngCore + ?Sized, T: DerefMut<Target = R>> RngCore for T { ... } With the comment This isn't abusing Deref for inheritance like the anti-pattern you listed; it's implementing a trait for anything Deref-able to a type implementing the trait. Would this be a good exception to the rule? That RngCore impl does not implement Deref or DerefMut for anything that is not a smart pointer, so the C-DEREF guideline has no bearing. But as a downside you should consider that adding this impl would make RngCore-related error messages much worse. Blanket impls (meaning for T where T is a type parameter) are pretty unusual and should be used sparingly. All in all I would be mildly against adding this impl. + use std::ops::DerefMut; trait RngCore {} + impl<R: RngCore + ?Sized, T: DerefMut<Target = R>> RngCore for T {} fn assert_rng_core<T: RngCore>() {} fn main() { assert_rng_core::<String>(); } Error message without impl error[E0277]: the trait bound `std::string::String: RngCore` is not satisfied --> src/main.rs:10:5 | 10 | assert_rng_core::<String>(); | ^^^^^^^^^^^^^^^^^^^^^^^^^ the trait `RngCore` is not implemented for `std::string::String` | = note: required by `assert_rng_core` Error message after adding impl error[E0277]: the trait bound `str: std::marker::Sized` is not satisfied --> src/main.rs:10:5 | 10 | assert_rng_core::<String>(); | ^^^^^^^^^^^^^^^^^^^^^^^^^ `str` does not have a constant size known at compile-time | = help: the trait `std::marker::Sized` is not implemented for `str` = note: required because of the requirements on the impl of `RngCore` for `str` = note: required because of the requirements on the impl of `RngCore` for `std::string::String` = note: required by `assert_rng_core` error[E0277]: the trait bound `str: std::ops::DerefMut` is not satisfied --> src/main.rs:10:5 | 10 | assert_rng_core::<String>(); | ^^^^^^^^^^^^^^^^^^^^^^^^^ the trait `std::ops::DerefMut` is not implemented for `str` | = note: required because of the requirements on the impl of `RngCore` for `str` = note: required because of the requirements on the impl of `RngCore` for `std::string::String` = note: required by `assert_rng_core` https://github.com/rust-lang-nursery/rand/pull/291
gharchive/issue
2018-03-13T11:43:03
2025-04-01T04:35:45.795840
{ "authors": [ "dtolnay", "pitdicker" ], "repo": "rust-lang-nursery/api-guidelines", "url": "https://github.com/rust-lang-nursery/api-guidelines/issues/158", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
480860896
Expose features from rand to avoid compilation breakage on WASM Current error: error: target is not supported, for more information see: https://docs.rs/getrandom/#unsupported-targets --> /home/user/.cargo/registry/src/github.com-1ecc6299db9ec823/getrandom-0.1.9/src/lib.rs:249:9 | 249 | / compile_error!("\ 250 | | target is not supported, for more information see: \ 251 | | https://docs.rs/getrandom/#unsupported-targets\ 252 | | "); | |___________^ In order to use join! and select! macros, we require the user to enable the async-await and nightly features. However those features will pull in rand. In order to work on WASM rand needs to have either the wasm-bindgen or stdweb feature enabled. Otherwise compile time breakage happens due to the getrandom crate. Explanation can be found here: https://docs.rs/getrandom/0.1.9/getrandom/#unsupported-targets This commit exposes those features through futures-util, allowing the end user to enable them when compiling on WASM. This means futures does not need to choose which feature to enable which would take away that decision from the end user. An alternative would be to detect the wasm32 target in Cargo.toml and to choose one of the features automatically. Current approach requires action on part of the user to enable one of those features, which means it should probably be documented somewhere. Currently the documentation of join! and select! does not mention the need for the async-await and nightly features. I propose we add some documentation changes to this PR before merging. Following a similar approach to getrandom by generating a compile time error to explain the situation to the user might be the lowest friction for the end user. eg. when feature async-await is enabled on wasm32, verify that one of the two required features is enabled as well, if not throw error. I can add that here if it seems a desirable solution. I have not tested the emscripten and WASI targets. On some quick testing this alleviates the compile error on my system, but it would be good if someone else had a look, or if we had CI testing for WASM. Ps: another solution would be to see with getrandom why they prefer breaking rather than selecting wasm-bindgen as a default. That would simplify things. @newpavlov ? wasm32-unknown-unknown can not make any assumptions about a target on which it will be executed. It can be a browser, Node.JS, smart-contract, game module system, etc. So we simply don't have any "system" entropy source, thus we can not make wasm-bindgen feature enabled by default. I am not sure if you need wasm-bindgen and stdweb features in this crate. The idea was that users will enable one of those feature only in the top level application crate. @newpavlov The problem is that futures-util pulls in rand which pulls in getrandom. So the end user doesn't necessarily have getrandom in their Cargo.toml. They will depend on futures however. I just verified and adding getrandom with the wasm-bindgen feature in my crate fixes the issue without modification to futures. If that is the desired approach, we need to document this! If that is the desired approach, we need to document this! Yes, I think it's the desired approach. If you have ideas of how we can improve getrandom docs, feel free to open a PR! Ok, I filed rust-random/getrandom#89 I still think we should add section about "target platforms" on the readme of futures, and one explaining the "feature flags" too... I still think we should add section about "target platforms" on the readme of futures I think having to re-document this in every crate that might transitively depend on getrandom is not worth it. A major reason to pull in semi-standard crates like this instead of rerolling randomness in every crate that needs it is to centralise platform support, it should be well documented by getrandom how to get it working on any odd platforms if it has partial support for them. (I would not consider wasm32-unknown-unknown a platform that is always expected to work out of the box with std depending crates since it is lacking fundamental OS features like randomness. If there were no_std support for select then I would expect it to be possible to use that on wasm32-unknown-unknown, which reminds me I should open an issue about getting that working.) @Nemo157 Have you seen the discussion on the getrandom repo? I am proposing some improvements to the docs there. It turns out that this is a features application devs have to put, even if getrandom get's pulled in by a dependency. I would propose afterwards that we do put a section about WASM in the readme where we explain things like this. I shall close this PR to avoid confusion.
gharchive/pull-request
2019-08-14T19:57:28
2025-04-01T04:35:45.806762
{ "authors": [ "Nemo157", "najamelan", "newpavlov" ], "repo": "rust-lang-nursery/futures-rs", "url": "https://github.com/rust-lang-nursery/futures-rs/pull/1804", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1727890561
Expose the Path of Cargo.toml in cargo-metadata output Problem There are a few cases in tooling (I am thinking of cases primarily inside of rust-analyzer, but I'm sure there are others) where it'd be very helpful to deterministically determine the crate's root folder (e.g., the folder containing Cargo.toml) in order to present the crate in a meaningful and idiomatic fashion and without resorting to approximations like "search parent folders of the crate root for a Cargo.toml". For context, both Buck2 and Bazel—which don't enforce structures on the crates that they build—both allow the user to query the build system to determine the file that defines the build target. Proposed Solution When emitting the crate root (the Rust file, not the Cargo.toml!), cargo-metadata should also emit the absolute path of the Cargo.toml that defines the crate. In the case of a Cargo workspace, I'd guess that the output should also return the Cargo.toml that defines Cargo workspace if the crate is not downloaded from crates.io. (I assume this information already exists within Cargo, but it's not exposed and hasn't been asked for before, which is why I didn't see an issue for this request.) Notes I'd be happy to add this feature to Cargo's metadata output. Could you clarify what you are looking for. For example, cargo metadata includes a Package::manifest_path field but I assume you saw that, so I'm trying to understand what is still missing. Could you clarify what you are looking for. For example, cargo metadata includes a Package::manifest_path field but I assume you saw that, so I'm trying to understand what is still missing. That's very generous of you to assume that I saw Package::manifest_path, but I unfortunately did not! I completely missed its existence—that's exactly what I'm looking for. My apologies for the noise; I'll close this issue.
gharchive/issue
2023-05-26T16:08:22
2025-04-01T04:35:45.828203
{ "authors": [ "davidbarsky", "epage" ], "repo": "rust-lang/cargo", "url": "https://github.com/rust-lang/cargo/issues/12187", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2015144966
gc: Determine CLI design for manual cleaning The current implementation from https://github.com/rust-lang/cargo/pull/12634 exposes a cargo clean gc subcommand to handle manually cleaning cache data. It is not clear what the final CLI design should be (and it is not clear exactly what the user scenarios are for when they would want to take manual control). This issue is tracking for determining what the CLI should look like. There are few different considerations: What should the actual subcommand be called? There were a few different considerations: cargo cache is already in use by a third-party command cargo gc (reserved by a third-party, but unused), was part of early proposals (cargo doesn't handle unstable top-level subcommands very well) cargo clean gc — the current implementation cargo clean — just fold the functionality into a single command which handles cleaning caches. @epage has concerns that this is overloading a subcommand used for different types of caches (local vs global, etc.). cargo maintenance — just an idea stolen from git How does this evolve with cleaning target directories, global build caches, etc.? Which flags should be exposed? Currently it exposes a large number of low-level flags. Ideally I would like to only stabilize a small number of higher-level flags. But I don't know what the high-level user scenarios might be, so I'm uncertain exactly what to expose. For example, one idea is to have a flag for "things that can be recreated" and another for "things that require downloading". Also bikeshed the names of the options. Potential sources of prior of other tools with cache management poetry docker git nix, zig dune, etc) Apparently buck2 and Turborepo grow unbounded Quick scan of brew autoremove cleanup "Removes all downloads more than 120 days old. This can be adjusted with HOMEBREW_CLEANUP_MAX_AGE_DAYS." HOMEBREW_CLEANUP_PERIODIC_FULL_DAYS (default 30 days) One complaint that came up was "brew cleanup has not been run in 30 days, running now" ... and then proceeds to run an interminable process in the middle of you attempting to do something else." (mastadon) We should probably step back and enumerate what the required use cases are and the "if it works" use cases. I like the way pnpm approaches this, https://pnpm.io/cli/store Running pnpm store prune is not harmful and has no side effects on your projects. If future installations require removed packages, pnpm will download them again. It is best practice to run pnpm store prune occasionally to clean up the store, but not too frequently. Sometimes, unreferenced packages become required again. This could occur when switching branches and installing older dependencies, in which case pnpm would need to re-download all removed packages, briefly slowing down the installation process. Regarding the concern of deleting crates that might still be in use, I like how rushjs asks you to pass an --unsafe flag to it's purge command, https://rushjs.io/pages/commands/rush_purge/ --unsafe (UNSAFE!) Also delete shared files such as the package manager instances stored in the ".rush" folder in the user's home directory. This is a more aggressive fix that is NOT SAFE to run in a live environment because it will cause other concurrent Rush processes to fail. @juliusl https://pnpm.io/cli/store it sounds like pnpm store prune removes everything that isn't referenced, regardless of age or size. In #13137 I bring up the idea to build on top of the work to track workspaces in #13136 so we pin entries not in current lockfiles. To extend this to clean up everything has the risk is if a project is transient (e.g. removable media) or moved but a new command wasn't run to register the new location. If its manually done with a command, rather than part of the auto-gc, then that might be reasonable, especially if we swap the logic and have a --execute flag rather than a --dry-run flag. Regarding the concern of deleting crates that might still be in use, I like how rushjs asks you to pass an --unsafe flag to it's purge command, I don't think an --unsafe flag is as relevant. For any shared directories, we use filesystem locks to ensure consistent reads/writes. Any "in use" concerns we have are more about "relevant to the user" and not "file descriptors are open" and is mostly relevant for slow networks/systems and offline usage (I don't want a rarely used dependency being removed just before I go on an airplane to do development offline).
gharchive/issue
2023-11-28T19:23:28
2025-04-01T04:35:45.842025
{ "authors": [ "ehuss", "epage", "juliusl" ], "repo": "rust-lang/cargo", "url": "https://github.com/rust-lang/cargo/issues/13060", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
929334907
Failed to update git submodule of dependency Problem Cargo can't clone/update a git submodule of dependency, for example redbpf crate, which contains submodule libbpf. Steps cargo new cargo_issue_git_submodule cd cargo_issue_git_submodule echo 'redbpf = { git = "https://github.com/foniod/redbpf" }' >> Cargo.toml cargo check And we will get an error: error: failed to get `redbpf` as a dependency of package `cargo_issue_git_submodule v0.1.0 (/home/user/cargo_issue_git_submodule)` Caused by: failed to load source for dependency `redbpf` Caused by: Unable to update https://github.com/foniod/redbpf Caused by: failed to update submodule `bpf-sys/libbpf` Caused by: object not found - no match for id (5a10cd2060c19959e2c55ee33421baa358c09c0b); class=Odb (9); code=NotFound (-3) Possible Solution(s) Update submodule manually in ~/.cargo/git/checkouts/redbpf-HASH/b81f069/bpf-sys/libbpf. Notes cargo version: cargo 1.55.0 commit of cargo: 4c27c96645e235d81f6c8dfff03ff9ebaf0ef71d It looks like the commit 5a10cd2060c19959e2c55ee33421baa358c09c0b does not exist on any branch or tag in the repository https://github.com/redsift/libbpf.git. The way cargo fetches submodules needs to assume that it exists on some branch or tag. Unfortunately I can't find any APIs in libgit2 that provide a way to fetch a specific commit, so I don't see a direct way this can be supported. One option is to restructure the net.git-fetch-with-cli option so that it uses git submodule update --init instead of git fetch (or pass the direct revision to git fetch which is supported as of git 2.5). This is something i'm currently running into that i'd really like to see resolved. At the current moment this is making it so that a submodule of one of my dependencies must be on the main/master branch, and that's not exactly doable
gharchive/issue
2021-06-24T15:14:10
2025-04-01T04:35:45.848134
{ "authors": [ "daniel-brenot-apcapital", "ehuss", "saruman9" ], "repo": "rust-lang/cargo", "url": "https://github.com/rust-lang/cargo/issues/9622", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1161797667
Update --timings=json output include data similar to the HTML output Content This PR attempts to bring the JSON output when using the --timings=json flag more inline with the data that is output when requesting HTML. Added custom Serialization impls to the Timings, and Unit structs. Added a report_json function to output the data in a manner similar to the HTML output. Removed the output of the timing reports on stdout during unit completion when the --timings=json flag was used. That information in now collected in the final report. Testing This was my method. There is probably a better way. Built a release version of Cargo. cargo build --release In a test project ran cargo clean && <path-to-cargo>/target/release/cargo build --timings=json -Z unstable-options Verified that JSON output was generated in the <basedir>/target/cargo-timings/ folder. r? @alexcrichton (rust-highfive has picked a reviewer for you, use r? to override) Thanks for the PR, but I'm going to be stepping down from the Cargo team so I'm going to un-assign myself from this. The Cargo team will help review this when they get a chance. r? @ehuss
gharchive/pull-request
2022-03-07T18:54:05
2025-04-01T04:35:45.853273
{ "authors": [ "CraZySacX", "alexcrichton", "rust-highfive" ], "repo": "rust-lang/cargo", "url": "https://github.com/rust-lang/cargo/pull/10463", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
463718854
add env variable to retrieve workspace root path Closes https://github.com/rust-lang/cargo/issues/7036 r? @matklad @alexcrichton The current work-around for this is to launch cargo metadata, which is not to pretty: https://github.com/mitsuhiko/insta/commit/943e7bca885140d6253a0ee97127e0f963403eac The potential gotcha here is that workspace situation is different between locally developed crate, and crate as it exists on crates.io: in the latter case, we remove all workspace structure and just store a flat list of crates. @matklad tried to build a crate with this patch, output this snippet: CARGO_MANIFEST_DIR=/Users/pengtangjin/.cargo/registry/src/github.com-1ecc6299db9ec823/lazy_static-1.3.0 CARGO_WORKSPACE_MANIFEST_DIR=/Users/pengtangjin/.cargo/registry/src/github.com-1ecc6299db9ec823/lazy_static-1.3.0 I thought it was ok, any case dangerous cases I missed? @matklad tried to build a crate with this patch, output this snippet: CARGO_MANIFEST_DIR=xxx/.cargo/registry/src/github.com-1ecc6299db9ec823/lazy_static-1.3.0 CARGO_WORKSPACE_MANIFEST_DIR=xxx/.cargo/registry/src/github.com-1ecc6299db9ec823/lazy_static-1.3.0 I thought it was ok, any case dangerous cases I missed? Thanks for the PR @csmoe and sorry for the delay in review! This sort of feature has a bit of history at this point, so I think it's worth paging that in as well. In addition to https://github.com/rust-lang/cargo/issues/7036 there's a lot more discussion on https://github.com/rust-lang/cargo/issues/3946 along with a closed PR and rationale for why the PR was closed. Could this perhaps address some of the feedback in those threads to indicate why we would merge this instead? Thanks for reviewing :), the problems wasn't addressed. So, closed.
gharchive/pull-request
2019-07-03T12:16:31
2025-04-01T04:35:45.859338
{ "authors": [ "alexcrichton", "csmoe", "matklad" ], "repo": "rust-lang/cargo", "url": "https://github.com/rust-lang/cargo/pull/7089", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
616717623
Expand error message to explain that a string was found With opt-level = "3" this previously said: must be an integer, `z`, or `s`, but found: 3 for ... The error message doesn't make that super clear. This should now be a bit more clear. Fixes #8234 We could even include a bit more saying that "3" should become 3 (either unconditionally or after trying to parse "3" into an integer? cc @steveklabnik r? @alexcrichton (rust_highfive has picked a reviewer for you, use r? to override) may be we should write quotes also \"3\" I would agree yeah it probably makes the most sense to also include the surrounding quotes, but otherwise looks great to me! Adjusted and also fixed the tests. Now they fail in apt install gcc-multilib @bors: r+ :pushpin: Commit 09084a365f612192fa026754b448f5b0be231890 has been approved by alexcrichton :hourglass: Testing commit 09084a365f612192fa026754b448f5b0be231890 with merge db702d3d09de4c1c1284e159090f49b793a7fe4d... :broken_heart: Test failed - checks-azure @bors retry #7858 😦 :hourglass: Testing commit 09084a365f612192fa026754b448f5b0be231890 with merge 55869de80ab8b7bd20cbb9a909e303fc03cf0125... :sunny: Test successful - checks-azure Approved by: alexcrichton Pushing 55869de80ab8b7bd20cbb9a909e303fc03cf0125 to master...
gharchive/pull-request
2020-05-12T14:48:07
2025-04-01T04:35:45.865968
{ "authors": [ "alexcrichton", "badboy", "bors", "ehuss", "rust-highfive", "shirshak55" ], "repo": "rust-lang/cargo", "url": "https://github.com/rust-lang/cargo/pull/8235", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
800744714
Backport #9102 to the 1.49 release branch #9129 but to the branch that matters, this time with the version bump. Does this look correct? r? @alexcrichton (rust-highfive has picked a reviewer for you, use r? to override) :warning: Warning :warning: Pull requests are usually filed against the master branch for this repo, but this one is against rust-1.49.0. Please double check that you specified the right target! Thanks! I don't think it's that worth it to get CI working at this point, so I'm going to hit the merge button on this one. I'll do the publish afterwards as well.
gharchive/pull-request
2021-02-03T22:15:23
2025-04-01T04:35:45.868762
{ "authors": [ "Eh2406", "alexcrichton", "rust-highfive" ], "repo": "rust-lang/cargo", "url": "https://github.com/rust-lang/cargo/pull/9130", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1208540335
Add DragonFly umtx_{sleep, wakeup}. This adds DragonFly's futex functions: https://man.dragonflybsd.org/?command=umtx&section=2 @bors r+ :pushpin: Commit 9579c1dc5878abe52e7aeef229d69284cdc1543c has been approved by Amanieu :hourglass: Testing commit 9579c1dc5878abe52e7aeef229d69284cdc1543c with merge 01b49024c26378a78e48358645a119595518e707... :sunny: Test successful - checks-actions, checks-cirrus-freebsd-11, checks-cirrus-freebsd-12, checks-cirrus-freebsd-13 Approved by: Amanieu Pushing 01b49024c26378a78e48358645a119595518e707 to master...
gharchive/pull-request
2022-04-19T15:41:02
2025-04-01T04:35:45.880092
{ "authors": [ "Amanieu", "bors", "m-ou-se" ], "repo": "rust-lang/libc", "url": "https://github.com/rust-lang/libc/pull/2763", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1145453582
Stacked borrows fails on {ChunksMut,ChunksExactMut}::__iterator_get_unchecked() Description fn main() { let mut arr1 = [0u8; 64]; let arr2 = [0u8; 64]; let mut iter = arr1.chunks_mut(8).zip(arr2.chunks(8)); while let Some((chunk1, chunk2)) = iter.next() { dbg!(chunk2[0]); dbg!(chunk1[0]); iter.next(); dbg!(chunk2[0]); dbg!(chunk1[0]); // error here } } Running this code in Miri will produce the following output: [src\main.rs:6] chunk2[0] = 0 [src\main.rs:7] chunk1[0] = 0 [src\main.rs:9] chunk2[0] = 0 error: Undefined Behavior: no item granting read access to tag <1832> at alloc906 found in borrow stack. --> src\main.rs:10:9 | 10 | dbg!(chunk1[0]); // error here | ^^^^^^^^^^^^^^^ no item granting read access to tag <1832> at alloc906 found in borrow stack. | = help: this indicates a potential bug in the program: it performed an invalid operation, but the rules it violated are still experimental = help: see https://github.com/rust-lang/unsafe-code-guidelines/blob/master/wip/stacked-borrows.md for further information With -Zmiri-track-pointer-tag=1832,1585,12319 note: tracking was triggered --> \library\core\src\slice\iter.rs:1564:19 | 1564 | Self { v: slice, chunk_size: size } | ^^^^^ created tag 1585 | = note: inside `std::slice::ChunksMut::<u8>::new` at \library\core\src\slice\iter.rs:1564:19 = note: inside `core::slice::<impl [u8]>::chunks_mut` at \library\core\src\slice\mod.rs:828:9 note: inside `main` at src\main.rs:4:20 --> src\main.rs:4:20 | 4 | let mut iter = arr1.chunks_mut(8).zip(arr2.chunks(8)); | ^^^^^^^^^^^^^^^^^^ note: tracking was triggered --> src\main.rs:5:21 | 5 | while let Some((chunk1, chunk2)) = iter.next() { | ^^^^^^ created tag 1832 | = note: inside `main` at src\main.rs:5:21 [src\main.rs:6] chunk2[0] = 0 [src\main.rs:7] chunk1[0] = 0 note: tracking was triggered --> \library\core\src\slice\iter.rs:1641:32 | 1641 | let len = cmp::min(self.v.len().unchecked_sub(start), self.chunk_size); | ^^^^^^^^^^^^ popped tracked tag for item [Unique for <1832>] due to Read access for <1585> | = note: inside `<std::slice::ChunksMut<u8> as std::iter::Iterator>::__iterator_get_unchecked` at \library\core\src\slice\iter.rs:1641:32 = note: inside `<std::iter::Zip<std::slice::ChunksMut<u8>, std::slice::Chunks<u8>> as std::iter::adapters::zip::ZipImpl<std::slice::ChunksMut<u8>, std::slice::Chunks<u8>>>::next` at \library\core\src\iter\adapters\zip.rs:278:23 = note: inside `<std::iter::Zip<std::slice::ChunksMut<u8>, std::slice::Chunks<u8>> as std::iter::Iterator>::next` at \library\core\src\iter\adapters\zip.rs:84:9 --> \library\core\src\slice\iter.rs:1642:32 | 1642 | from_raw_parts_mut(self.v.as_mut_ptr().add(start), len) | ^^^^^^^^^^^^^^^^^^^ created tag 12319 | = note: inside `<std::slice::ChunksMut<u8> as std::iter::Iterator>::__iterator_get_unchecked` at \library\core\src\slice\iter.rs:1642:32 --> \library\core\src\slice\mod.rs:483:5 | 483 | / pub const fn as_mut_ptr(&mut self) -> *mut T { 484 | | self as *mut [T] as *mut T 485 | | } | |_____^ popped tracked tag for item [Disabled for <1832>] due to Write access for <12319> | = note: inside `core::slice::<impl [u8]>::as_mut_ptr` at \library\core\src\slice\mod.rs:483:5 = note: inside `<std::slice::ChunksMut<u8> as std::iter::Iterator>::__iterator_get_unchecked` at \library\core\src\slice\iter.rs:1642:32 note: inside `main` at src\main.rs:8:9 --> src\main.rs:8:9 | 8 | iter.next(); | ^^^^^^^^^^^ = note [src\main.rs:9] chunk2[0] = 0 error: Undefined Behavior: no item granting read access to tag <1832> at alloc906 found in borrow stack. --> src\main.rs:10:9 | 10 | dbg!(chunk1[0]); // error here | ^^^^^^^^^^^^^^^ no item granting read access to tag <1832> at alloc906 found in borrow stack. | = help: this indicates a potential bug in the program: it performed an invalid operation, but the rules it violated are still experimental = help: see https://github.com/rust-lang/unsafe-code-guidelines/blob/master/wip/stacked-borrows.md for further information A few things to note here: Access to the mutable chunk will fail. For example arr1.chunks(8).zip(arr2.chunks_mut(8)) will fail on chunk2[0] instead. zip() is required. This is probably due to Zip uses __iterator_get_unchecked() internally rather than next(). chunks_mut() and chunks_exact_mut() can both reproduce. array_chunks_mut() can't reproduce. Environment $ rustc --version --verbose rustc 1.61.0-nightly (45e2c2881 2022-02-20) binary: rustc commit-hash: 45e2c2881d11324d610815bfff097e25c412199e commit-date: 2022-02-20 host: x86_64-pc-windows-msvc release: 1.61.0-nightly LLVM version: 14.0.0 $ cargo miri --version miri 0.1.0 (0db4090 2022-02-12) There's probably more of these kinds of problems that @saethlin has been looking into recently. Huge thanks for reporting this @andylizi! This is definitely an issue with the implementation in core. Perhaps this issue should be moved to rust-lang/rust? The existing definition of ChunksMut: https://github.com/rust-lang/rust/blob/1103d2e914b67c18b0deb86073c26c6aefda761d/library/core/src/slice/iter.rs#L1556-L1559 pub struct ChunksMut<'a, T: 'a> { v: &'a mut [T], chunk_size: usize, } Cannot support the current implementation of __iterator_get_unchecked: https://github.com/rust-lang/rust/blob/1103d2e914b67c18b0deb86073c26c6aefda761d/library/core/src/slice/iter.rs#L1632-L1644 unsafe fn __iterator_get_unchecked(&mut self, idx: usize) -> Self::Item { let start = idx * self.chunk_size; // SAFETY: see comments for `Chunks::__iterator_get_unchecked`. // // Also note that the caller also guarantees that we're never called // with the same index again, and that no other methods that will // access this subslice are called, so it is valid for the returned // slice to be mutable. unsafe { let len = cmp::min(self.v.len().unchecked_sub(start), self.chunk_size); from_raw_parts_mut(self.v.as_mut_ptr().add(start), len) } } The problem is that this method is trying to hand out subslices of a &mut [T] while also holding on to the original. Because it wraps a &mut [T], ChunksMut guarantees that its inner slice always shrinks when it returns a slice. It looks like this guarantee was accidentally introduced and accidentally upheld in implementation by avoiding unsafe... until __iterator_get_unchecked was added. I'm working on a patch that fixes the aliasing problem. It's perhaps worth noting that the specialization on zip that this exists to support has already earned 4 I-unsound issues, and this aliasing issue isn't due to new code. We/I should have caught this already, but didn't. That's quite annoying.
gharchive/issue
2022-02-21T08:09:10
2025-04-01T04:35:45.898482
{ "authors": [ "RalfJung", "andylizi", "saethlin" ], "repo": "rust-lang/miri", "url": "https://github.com/rust-lang/miri/issues/1980", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1251870966
rustup Cc https://github.com/rust-lang/rust/issues/97486 @bors r+ :pushpin: Commit 7cd5fc3de327b9db96918cc895676f2d94c0a44d has been approved by RalfJung :hourglass: Testing commit 7cd5fc3de327b9db96918cc895676f2d94c0a44d with merge 5c3e4b6556431c816d05b262a890905a2a50841a... :sunny: Test successful - checks-actions Approved by: RalfJung Pushing 5c3e4b6556431c816d05b262a890905a2a50841a to master...
gharchive/pull-request
2022-05-29T12:07:01
2025-04-01T04:35:45.901608
{ "authors": [ "RalfJung", "bors" ], "repo": "rust-lang/miri", "url": "https://github.com/rust-lang/miri/pull/2162", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1428625078
Implement condvars for Windows Adds 3 shims for Windows: SleepConditionVariableSRW, WakeConditionVariable, WakeAllConditionVariable to add support for condvars (which fixes #2628). Salvaged from what was removed from #2231 I think this panic is caused by a mistake in #2601. CI failure should be fixed by #2641. I think this panic is caused by a mistake in #2601. That PR landed, so please rebase this one. :) The rebase is done. CI's finally passing :) Great. :) Please squash the commits a little, then we can land this. Squashed. @bors r+ :pushpin: Commit 958ca31d26951c1c1d8be875d704c72efa8aa8b5 has been approved by RalfJung It is now in the queue for this repository. :hourglass: Testing commit 958ca31d26951c1c1d8be875d704c72efa8aa8b5 with merge 32c90ffab5663ec5495220bcb85b199327227878... :sunny: Test successful - checks-actions Approved by: RalfJung Pushing 32c90ffab5663ec5495220bcb85b199327227878 to master...
gharchive/pull-request
2022-10-30T06:05:13
2025-04-01T04:35:45.906399
{ "authors": [ "DrMeepster", "RalfJung", "bors" ], "repo": "rust-lang/miri", "url": "https://github.com/rust-lang/miri/pull/2638", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2302862726
automata: Fix broken universal start states with sparse DFA The state IDs were not remapped, which will usually result in an index out of range error. Is it possible to add a regression test? Sure, let me give it a try tomorrow (it's late here today). Hi, I added a test based on examples in Automaton::is_special_state. I'm asserting that the initial state is .is_start_state(), but this seems to be broken for some reason. Could you help figuring out why? I'm afraid I'm not familiar enough with how the start state ranges are being assigned. Alternatively, I could remove the assert for now and file an issue for it to be fixed later. It looks like start states were simply being removed with set_no_special_start_states, so this was the intended behavior. I have removed the assert now. Let me know if the tests look good to you. Hi, mind giving this a review? Friendly ping, let me know if there's something else I need to address
gharchive/pull-request
2024-05-17T14:11:29
2025-04-01T04:35:45.911445
{ "authors": [ "BurntSushi", "ishitatsuyuki" ], "repo": "rust-lang/regex", "url": "https://github.com/rust-lang/regex/pull/1195", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
367545890
String::from(“…”) => d”…” Dynamic Strings are common but using String::From("...") or "...".to_string() can get tedious, so it'll be more productive to write a string litteral d"..." instead. It's more productive, removes the apparent distance between both types of strings and it's syntactically sweeter. Note that this does not constitute a formal RFC proposal; those are pull requests that follow the template. Please read the README.md file to see how the process works. I'm closing this issue in favor of https://internals.rust-lang.org/t/string-from-d/8528.
gharchive/issue
2018-10-07T11:35:36
2025-04-01T04:35:45.913940
{ "authors": [ "AmjadHD", "Centril" ], "repo": "rust-lang/rfcs", "url": "https://github.com/rust-lang/rfcs/issues/2559", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
139182339
How long should rust-mode keep supporting Emacs 23? Is anyone using rust-mode regularly with emacs 23? (The last version of emacs 23 was released in January 2012...three and a half years before Rust 1.0 was released, for what it's worth.) I'm starting to see signs of slight hiccups for people trying to enhance rust-mode. Nothing hard to overcome yet, but it could get more annoying later. When is it time to cut it off? When emacs 25 comes out? When someone writes a feature that really needs something only available in emacs 24 and above? Sooner than that? This affects issue #139 as well. Personally I think we should just drop emacs23 support now.
gharchive/issue
2016-03-08T05:00:46
2025-04-01T04:35:45.944573
{ "authors": [ "MicahChalmer", "nikomatsakis", "tromey" ], "repo": "rust-lang/rust-mode", "url": "https://github.com/rust-lang/rust-mode/issues/137", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
53918525
Fails on non-privileged user account. The previous installer which only installed rust would ask for you to enter a admin password. The new installer which installs both rust and cargo fail instead requiring you to run it as an admin. The new installer is derived from the old one and I don't believe it ever requested sudo. That said, it might be a good idea for the installer to ask for root if it determines that it can't install. I filed this against rust-installer https://github.com/rust-lang/rust-installer/issues/7 @brson I should I have also said that I run windows 7.
gharchive/issue
2015-01-09T21:43:30
2025-04-01T04:35:45.946385
{ "authors": [ "Jexell", "brson" ], "repo": "rust-lang/rust-packaging", "url": "https://github.com/rust-lang/rust-packaging/issues/5", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
51929514
rust-lang.org is not rendering correctly (blank) Most of rust-lang.org is not rendering correctly for me. Neither with Firefox 34.0.5 nor with Chromium 39.0.2171.95 on Archlinux. Firefox's console does not show any error but Chrome's one tell me : Resource interpreted as Font but transferred with MIME type binary/octet-stream: "http://doc.rust-lang.org/Heuristica-Italic.woff" The MIME type in Firefox is "application/x-font-woff" so I'm not sure it is the issue here. Just wanted to put here: if you experience the problem, the current official Arch packages for Fira Sans should work. https://www.archlinux.org/packages/community/any/ttf-fira-sans/
gharchive/issue
2014-12-14T19:53:25
2025-04-01T04:35:45.949824
{ "authors": [ "ldng", "skade" ], "repo": "rust-lang/rust-www", "url": "https://github.com/rust-lang/rust-www/issues/75", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
467970524
Modules with path still can't be formatted on Windows #1754 issue is still occuring. I tried cargo fmt on my tokio checkout and I get several errors like following: error: couldn't read \\?\D:\tokio\tokio-sync\tests\..\src\mpsc\list.rs: The filename, directory name, or volume label syntax is incorrect. (os error 123) --> \\?\D:\tokio\tokio-sync\tests\fuzz_list.rs:14:5 | 14 | mod list; | ^^^^ code for above is: #[path = "../src/mpsc/list.rs"] #[allow(warnings)] mod list; Can you try the changes in https://github.com/rust-lang/rustfmt/pull/3590 and see if that fixes your problem ? Looks like it doesn't for some users... It is not building: --> /home/95th/.cargo/registry/src/github.com-1ecc6299db9ec823/rustc-ap-rustc_target-407.0.0/abi/mod.rs:112:62 | 112 | ["a", ref a..] => dl.aggregate_align = align(a, "a")?, | ^ expected slice, found &str | = note: expected type `&[&str]` found type `&&str` error[E0308]: mismatched types --> /home/95th/.cargo/registry/src/github.com-1ecc6299db9ec823/rustc-ap-rustc_target-407.0.0/abi/mod.rs:113:58 | 113 | ["f32", ref a..] => dl.f32_align = align(a, "f32")?, | ^ expected slice, found &str | = note: expected type `&[&str]` found type `&&str` error[E0308]: mismatched types --> /home/95th/.cargo/registry/src/github.com-1ecc6299db9ec823/rustc-ap-rustc_target-407.0.0/abi/mod.rs:114:58 | 114 | ["f64", ref a..] => dl.f64_align = align(a, "f64")?, | ^ expected slice, found &str | = note: expected type `&[&str]` found type `&&str` error[E0308]: mismatched types --> /home/95th/.cargo/registry/src/github.com-1ecc6299db9ec823/rustc-ap-rustc_target-407.0.0/abi/mod.rs:117:46 | 117 | dl.pointer_align = align(a, p)?; | ^ expected slice, found &str | = note: expected type `&[&str]` found type `&&str` error[E0308]: mismatched types --> /home/95th/.cargo/registry/src/github.com-1ecc6299db9ec823/rustc-ap-rustc_target-407.0.0/abi/mod.rs:127:35 | 127 | let a = align(a, s)?; | ^ expected slice, found &str | = note: expected type `&[&str]` found type `&&str` error[E0308]: mismatched types --> /home/95th/.cargo/registry/src/github.com-1ecc6299db9ec823/rustc-ap-rustc_target-407.0.0/abi/mod.rs:145:35 | 145 | let a = align(a, s)?; | ^ expected slice, found &str | = note: expected type `&[&str]` found type `&&str` error: aborting due to 13 previous errors Hi @scampi, is there any update on this issue please? I just hit this issue today in my repository. The repository and commit that is failing to rustfmt is: https://github.com/rojo-rbx/rbx-dom/tree/3448e528766b1558ac76addf5e636df797f180ed @95th Sorry for the delay, I had little time available for this. The branch is now compiling. Thanks for giving it a try! This will be fixed in the 2.0 release.
gharchive/issue
2019-07-15T07:04:07
2025-04-01T04:35:46.595006
{ "authors": [ "95th", "LPGhatguy", "scampi", "topecongiro" ], "repo": "rust-lang/rustfmt", "url": "https://github.com/rust-lang/rustfmt/issues/3688", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
876139041
Mask types are missing select implementation I want to be able to: fn f(cond: mask32x4, true_v: mask32x4, false_v: mask32x4) -> mask32x4 { cond.select(true_v, false_v) } I was just thinking about this. It's slightly non-trivial because the masks aren't necessarily SIMD types. But no reason it can't be done.
gharchive/issue
2021-05-05T07:15:37
2025-04-01T04:35:46.605386
{ "authors": [ "calebzulawski", "programmerjake" ], "repo": "rust-lang/stdsimd", "url": "https://github.com/rust-lang/stdsimd/issues/114", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
599389604
Add an environment variable to keep the temp dir Fixes #66 Published as v0.5.29.
gharchive/pull-request
2020-04-14T08:27:05
2025-04-01T04:35:46.653562
{ "authors": [ "phil-opp" ], "repo": "rust-osdev/cargo-xbuild", "url": "https://github.com/rust-osdev/cargo-xbuild/pull/67", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2177494796
Skip URL check for https://blog.rust.careers Certificate has expired. I checked the site to see if I could find an email address to notify that their cert is expired, but didn't find anything from a quick look. I checked the site to see if I could find an email address to notify that their cert is expired, but didn't find anything from a quick look. They mention an email on their main site at https://rust.careers/about. I sent them an email a few days ago, let's hope they fix it soon. Note that this is already the second time that the certificate has expired for that site in the past few months. So I'm fine with keeping the exception even after the certificate is renewed to avoid additional work for us in the future.
gharchive/pull-request
2024-03-10T02:49:45
2025-04-01T04:35:46.656002
{ "authors": [ "nicholasbishop", "phil-opp" ], "repo": "rust-osdev/homepage", "url": "https://github.com/rust-osdev/homepage/pull/206", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1393936735
Question over stdout mutability Hello, I'm attempting to use this crate to implement a basic UEFI bootloader but I've encountered an issue that I'm unclear on how to navigate around. The following code let vendor = system_table.firmware_vendor(); system_table.stdout().output_string(vendor)?; attempts to retrieve a reference to the firmware vendor from the system table, and then print it to the standard output. However, this code fails to compile due to system_table being borrowed as a mutable reference when calling stdout(). error[E0502]: cannot borrow `system_table` as mutable because it is also borrowed as immutable --> src\main.rs:26:5 | 25 | let vendor = system_table.firmware_vendor(); | ------------------------------ immutable borrow occurs here 26 | system_table.stdout().output_string(vendor)?; | ^^^^^^^^^^^^^^^^^^^^^ ------ immutable borrow later used here | | | mutable borrow occurs here Although I've found I can use other means to print the firmware vendor to the standard output, I'd like to avoid having to copy the string data if possible, as it's already present in memory and I'd like to be able to just pass it directly to output_string without overhead. Is there an intended code pattern I should be using here in order to achieve this, or is there no way with the package API to avoid doing a copy? From what I could tell from peeking at the code, stdout is borrowing the system table reference as mutable, and although it's only touching self.table.stdout, this conceptually invalidates the result of firmware_vendor despite only touching self.table.fw_vendor, but I'm unsure if there's a way for the API to represent this or not. It's possible to avoid making a copy of the string, but it requires using unsafe. One way to do it is to use the uefi-services crate, which provides print/println macros (as well as a log implementation). It keeps an internal pointer to the system table, so you can do something like println!("vendor: {}", st.firmware_vendor()) without any borrowing issues. That hides the unsafe from your code, but it is still present. Another alternative, closer to the example you gave, would be to keep a pointer to the vendor string: let vendor: *const _ = st.firmware_vendor(); // Safety: we trust the firmware to not invalidate the vendor pointer while calling `output_string`. unsafe { st.stdout().output_string(&*vendor).unwrap(); }
gharchive/issue
2022-10-02T23:39:48
2025-04-01T04:35:46.660161
{ "authors": [ "nicholasbishop", "ori-sky" ], "repo": "rust-osdev/uefi-rs", "url": "https://github.com/rust-osdev/uefi-rs/issues/522", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2201110138
Fix some new lints/warnings Checklist [ ] Sensible git history (for example, squash "typo" or "fix" commits). See the Rewriting History guide for help. [ ] Update the changelog (if necessary) The merge train pipeline fails: https://github.com/rust-osdev/uefi-rs/actions/runs/8417467100/job/23045999267 @nicholasbishop - unfortunately, I do not have time to look into this. Looks like the nightly compiler now warns if you import something by name that was already imported by glob. So for example: use uefi::prelude::*; use uefi::table::boot::BootServices; // warning: the item `BootServices` is already imported In a couple places this was also happening due to the prelude glob, e.g. an import of Iterator now warns since it's already imported in the prelude. Pushed another commit to fix all those cases.
gharchive/pull-request
2024-03-21T20:23:43
2025-04-01T04:35:46.663415
{ "authors": [ "nicholasbishop", "phip1611" ], "repo": "rust-osdev/uefi-rs", "url": "https://github.com/rust-osdev/uefi-rs/pull/1103", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
660712875
Disable serial tests in CI CI on master recently failed by crashing QEMU while running the serial protocol tests. It didn't fail on the corresponding PR however, making me think this is bug only occurs sometimes. For now, I want to check if disabling these tests keeps CI happy. Rerunning the tests seems to have fixed the issue. I'll investigate if there's any way to prevent them from failing (I'm guessing they timeout during periods of intense activity on GH Actions)
gharchive/pull-request
2020-07-19T09:12:17
2025-04-01T04:35:46.664937
{ "authors": [ "GabrielMajeri" ], "repo": "rust-osdev/uefi-rs", "url": "https://github.com/rust-osdev/uefi-rs/pull/146", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
833141657
Unmaintained dependency - dkjson Details The dkjson package required to load the tags to JSONB (instead of HSTORE) is unmaintained (https://github.com/LuaDist/dkjson). Reported on #81, tracking here to avoid losing sight of this. The dkjson package was suggested by the osm2pgsql project examples (see https://github.com/openstreetmap/osm2pgsql/blob/master/flex-config/places.lua#L5-L7), not sure what other alternatives/options may exist at this time. Planning to use https://github.com/rxi/json.lua The rxi project has >1k stars and 245 forks and had relatively recent activity. Their docs mentioned being faster than other options (including dkjson) so did a quick test. I ran tags.lua to load Colorado 3x w/ dkjson and and 3x w/ rki-json. 4CPU, 8 GB RAM, SSD. Basic command: osm2pgsql --slim --drop \ --output=flex --style=./style/tags.lua \ --cache 3000 -d $PGOSM_CONN ~/pgosm-data/colorado-latest.osm.pbf rki-json.lua loaded in an average of 119 seconds vs 136s for dkjson, so ~13% faster with this option. I looked at a few other options but none were as popular, recently updated, or as simple/straightfoward. I plan to include the json.lua script in the flex-config/ directory with a link back to the project for the ultimate source. Will have to occasionally check on the parent project for updates. The json script can then be included directly (like helpers.lua) without additional prereq to install.
gharchive/issue
2021-03-16T19:46:47
2025-04-01T04:35:46.751663
{ "authors": [ "rustprooflabs" ], "repo": "rustprooflabs/pgosm-flex", "url": "https://github.com/rustprooflabs/pgosm-flex/issues/88", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1083814307
No dense data is shown on dashboard If I select 3 hours on dashboard, I see data points only every 15min (should be 1min): It'll now fetch mixed data if history range <= 24h, else sparse
gharchive/issue
2021-12-18T11:41:06
2025-04-01T04:35:46.764342
{ "authors": [ "io53", "laurijamsa" ], "repo": "ruuvi/com.ruuvi.station.webui", "url": "https://github.com/ruuvi/com.ruuvi.station.webui/issues/97", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }