id
stringlengths
9
10
submitter
stringlengths
2
52
authors
stringlengths
4
6.51k
title
stringlengths
4
246
comments
stringlengths
1
523
journal-ref
stringlengths
4
345
doi
stringlengths
11
120
report-no
stringlengths
2
243
categories
stringlengths
5
98
license
stringclasses
9 values
abstract
stringlengths
33
3.33k
versions
list
update_date
timestamp[s]
authors_parsed
list
prediction
stringclasses
1 value
probability
float64
0.95
1
2212.03282
Darryl Hannan
Darryl Hannan, Steven C. Nesbit, Ximing Wen, Glen Smith, Qiao Zhang, Alberto Goffi, Vincent Chan, Michael J. Morris, John C. Hunninghake, Nicholas E. Villalobos, Edward Kim, Rosina O. Weber and Christopher J. MacLellan
MobilePTX: Sparse Coding for Pneumothorax Detection Given Limited Training Examples
IAAI 2023 (7 pages)
null
null
null
cs.CV
http://creativecommons.org/licenses/by-nc-sa/4.0/
Point-of-Care Ultrasound (POCUS) refers to clinician-performed and interpreted ultrasonography at the patient's bedside. Interpreting these images requires a high level of expertise, which may not be available during emergencies. In this paper, we support POCUS by developing classifiers that can aid medical professionals by diagnosing whether or not a patient has pneumothorax. We decomposed the task into multiple steps, using YOLOv4 to extract relevant regions of the video and a 3D sparse coding model to represent video features. Given the difficulty in acquiring positive training videos, we trained a small-data classifier with a maximum of 15 positive and 32 negative examples. To counteract this limitation, we leveraged subject matter expert (SME) knowledge to limit the hypothesis space, thus reducing the cost of data collection. We present results using two lung ultrasound datasets and demonstrate that our model is capable of achieving performance on par with SMEs in pneumothorax identification. We then developed an iOS application that runs our full system in less than 4 seconds on an iPad Pro, and less than 8 seconds on an iPhone 13 Pro, labeling key regions in the lung sonogram to provide interpretable diagnoses.
[ { "version": "v1", "created": "Tue, 6 Dec 2022 19:33:05 GMT" }, { "version": "v2", "created": "Thu, 8 Dec 2022 03:46:45 GMT" } ]
2022-12-09T00:00:00
[ [ "Hannan", "Darryl", "" ], [ "Nesbit", "Steven C.", "" ], [ "Wen", "Ximing", "" ], [ "Smith", "Glen", "" ], [ "Zhang", "Qiao", "" ], [ "Goffi", "Alberto", "" ], [ "Chan", "Vincent", "" ], [ "Morris", "Michael J.", "" ], [ "Hunninghake", "John C.", "" ], [ "Villalobos", "Nicholas E.", "" ], [ "Kim", "Edward", "" ], [ "Weber", "Rosina O.", "" ], [ "MacLellan", "Christopher J.", "" ] ]
new_dataset
0.961141
2212.03858
Ruohan Gao
Hao Li, Yizhi Zhang, Junzhe Zhu, Shaoxiong Wang, Michelle A Lee, Huazhe Xu, Edward Adelson, Li Fei-Fei, Ruohan Gao, Jiajun Wu
See, Hear, and Feel: Smart Sensory Fusion for Robotic Manipulation
In CoRL 2022. Li and Zhang equal contribution; Gao and Wu equal advising. Project page: https://ai.stanford.edu/~rhgao/see_hear_feel/
null
null
null
cs.RO cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Humans use all of their senses to accomplish different tasks in everyday activities. In contrast, existing work on robotic manipulation mostly relies on one, or occasionally two modalities, such as vision and touch. In this work, we systematically study how visual, auditory, and tactile perception can jointly help robots to solve complex manipulation tasks. We build a robot system that can see with a camera, hear with a contact microphone, and feel with a vision-based tactile sensor, with all three sensory modalities fused with a self-attention model. Results on two challenging tasks, dense packing and pouring, demonstrate the necessity and power of multisensory perception for robotic manipulation: vision displays the global status of the robot but can often suffer from occlusion, audio provides immediate feedback of key moments that are even invisible, and touch offers precise local geometry for decision making. Leveraging all three modalities, our robotic system significantly outperforms prior methods.
[ { "version": "v1", "created": "Wed, 7 Dec 2022 18:55:53 GMT" }, { "version": "v2", "created": "Thu, 8 Dec 2022 05:52:16 GMT" } ]
2022-12-09T00:00:00
[ [ "Li", "Hao", "" ], [ "Zhang", "Yizhi", "" ], [ "Zhu", "Junzhe", "" ], [ "Wang", "Shaoxiong", "" ], [ "Lee", "Michelle A", "" ], [ "Xu", "Huazhe", "" ], [ "Adelson", "Edward", "" ], [ "Fei-Fei", "Li", "" ], [ "Gao", "Ruohan", "" ], [ "Wu", "Jiajun", "" ] ]
new_dataset
0.974294
2212.03957
Shahrzad Haddadan
Suman K.Bera and Jayesh Choudhari and Shahrzad Haddadan and Sara Ahmadian
DeMEtRIS: Counting (near)-Cliques by Crawling
null
null
10.1145/3539597.3570438
null
cs.DS
http://creativecommons.org/licenses/by-nc-sa/4.0/
We study the problem of approximately counting cliques and near cliques in a graph, where the access to the graph is only available through crawling its vertices; thus typically seeing only a small portion of it. This model, known as the random walk model or the neighborhood query model has been introduced recently and captures real-life scenarios in which the entire graph is too massive to be stored as a whole or be scanned entirely and sampling vertices independently is non-trivial in it. We introduce DeMEtRIS: Dense Motif Estimation through Random Incident Sampling. This method provides a scalable algorithm for clique and near clique counting in the random walk model. We prove the correctness of our algorithm through rigorous mathematical analysis and extensive experiments. Both our theoretical results and our experiments show that DeMEtRIS obtains a high precision estimation by only crawling a sub-linear portion on vertices, thus we demonstrate a significant improvement over previously known results.
[ { "version": "v1", "created": "Wed, 7 Dec 2022 21:10:18 GMT" } ]
2022-12-09T00:00:00
[ [ "Bera", "Suman K.", "" ], [ "Choudhari", "Jayesh", "" ], [ "Haddadan", "Shahrzad", "" ], [ "Ahmadian", "Sara", "" ] ]
new_dataset
0.986778
2212.03961
Gyeongmin Choe
Gyeongmin Choe, Beibei Du, Seonghyeon Nam, Xiaoyu Xiang, Bo Zhu, Rakesh Ranjan
FSID: Fully Synthetic Image Denoising via Procedural Scene Generation
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
For low-level computer vision and image processing ML tasks, training on large datasets is critical for generalization. However, the standard practice of relying on real-world images primarily from the Internet comes with image quality, scalability, and privacy issues, especially in commercial contexts. To address this, we have developed a procedural synthetic data generation pipeline and dataset tailored to low-level vision tasks. Our Unreal engine-based synthetic data pipeline populates large scenes algorithmically with a combination of random 3D objects, materials, and geometric transformations. Then, we calibrate the camera noise profiles to synthesize the noisy images. From this pipeline, we generated a fully synthetic image denoising dataset (FSID) which consists of 175,000 noisy/clean image pairs. We then trained and validated a CNN-based denoising model, and demonstrated that the model trained on this synthetic data alone can achieve competitive denoising results when evaluated on real-world noisy images captured with smartphone cameras.
[ { "version": "v1", "created": "Wed, 7 Dec 2022 21:21:55 GMT" } ]
2022-12-09T00:00:00
[ [ "Choe", "Gyeongmin", "" ], [ "Du", "Beibei", "" ], [ "Nam", "Seonghyeon", "" ], [ "Xiang", "Xiaoyu", "" ], [ "Zhu", "Bo", "" ], [ "Ranjan", "Rakesh", "" ] ]
new_dataset
0.998893
2212.03965
Shikhar Tuli
Shikhar Tuli, Chia-Hao Li, Ritvik Sharma, Niraj K. Jha
CODEBench: A Neural Architecture and Hardware Accelerator Co-Design Framework
Published at ACM Transactions on Embedded Computing Systems. Code available at https://github.com/jha-lab/codebench
null
10.1145/3575798
null
cs.AR cs.LG eess.IV
http://creativecommons.org/licenses/by/4.0/
Recently, automated co-design of machine learning (ML) models and accelerator architectures has attracted significant attention from both the industry and academia. However, most co-design frameworks either explore a limited search space or employ suboptimal exploration techniques for simultaneous design decision investigations of the ML model and the accelerator. Furthermore, training the ML model and simulating the accelerator performance is computationally expensive. To address these limitations, this work proposes a novel neural architecture and hardware accelerator co-design framework, called CODEBench. It is composed of two new benchmarking sub-frameworks, CNNBench and AccelBench, which explore expanded design spaces of convolutional neural networks (CNNs) and CNN accelerators. CNNBench leverages an advanced search technique, BOSHNAS, to efficiently train a neural heteroscedastic surrogate model to converge to an optimal CNN architecture by employing second-order gradients. AccelBench performs cycle-accurate simulations for a diverse set of accelerator architectures in a vast design space. With the proposed co-design method, called BOSHCODE, our best CNN-accelerator pair achieves 1.4% higher accuracy on the CIFAR-10 dataset compared to the state-of-the-art pair, while enabling 59.1% lower latency and 60.8% lower energy consumption. On the ImageNet dataset, it achieves 3.7% higher Top1 accuracy at 43.8% lower latency and 11.2% lower energy consumption. CODEBench outperforms the state-of-the-art framework, i.e., Auto-NBA, by achieving 1.5% higher accuracy and 34.7x higher throughput, while enabling 11.0x lower energy-delay product (EDP) and 4.0x lower chip area on CIFAR-10.
[ { "version": "v1", "created": "Wed, 7 Dec 2022 21:38:03 GMT" } ]
2022-12-09T00:00:00
[ [ "Tuli", "Shikhar", "" ], [ "Li", "Chia-Hao", "" ], [ "Sharma", "Ritvik", "" ], [ "Jha", "Niraj K.", "" ] ]
new_dataset
0.994493
2212.03968
Michal Balazia
Tanay Agrawal, Michal Balazia, Philipp M\"uller, Fran\c{c}ois Br\'emond
Multimodal Vision Transformers with Forced Attention for Behavior Analysis
Preprint. Full paper accepted at the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, USA, Jan 2023. 11 pages
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Human behavior understanding requires looking at minute details in the large context of a scene containing multiple input modalities. It is necessary as it allows the design of more human-like machines. While transformer approaches have shown great improvements, they face multiple challenges such as lack of data or background noise. To tackle these, we introduce the Forced Attention (FAt) Transformer which utilize forced attention with a modified backbone for input encoding and a use of additional inputs. In addition to improving the performance on different tasks and inputs, the modification requires less time and memory resources. We provide a model for a generalised feature extraction for tasks concerning social signals and behavior analysis. Our focus is on understanding behavior in videos where people are interacting with each other or talking into the camera which simulates the first person point of view in social interaction. FAt Transformers are applied to two downstream tasks: personality recognition and body language recognition. We achieve state-of-the-art results for Udiva v0.5, First Impressions v2 and MPII Group Interaction datasets. We further provide an extensive ablation study of the proposed architecture.
[ { "version": "v1", "created": "Wed, 7 Dec 2022 21:56:50 GMT" } ]
2022-12-09T00:00:00
[ [ "Agrawal", "Tanay", "" ], [ "Balazia", "Michal", "" ], [ "Müller", "Philipp", "" ], [ "Brémond", "François", "" ] ]
new_dataset
0.9989
2212.03992
Ian McQuillan
Oscar H. Ibarra and Ian McQuillan
State Grammars with Stores
21 pages
Theoretical Computer Science 798, 23-39 (2019)
10.1016/j.tcs.2019.06.024
null
cs.FL
http://creativecommons.org/licenses/by-nc-nd/4.0/
State grammars are context-free grammars where the productions have states associated with them, and a production can only be applied to a nonterminal if the current state matches the state in the production. Once states are added to grammars, it is natural to add various stores, similar to machine models. With such extensions, productions can only be applied if both the state and the value read from each store matches between the current sentential form and the production. Here, generative capacity results are presented for different derivation modes, with and without additional stores. In particular, with the standard derivation relation, it is shown that adding reversal-bounded counters does not increase the capacity, and states are enough. Also, state grammars with reversal-bounded counters that operate using leftmost derivations are shown to coincide with languages accepted by one-way machines with a pushdown and reversal-bounded counters, and these are surprisingly shown to be strictly weaker than state grammars with the standard derivation relation (and no counters). The complexity of the emptiness problem involving state grammars with reversal-bounded counters is also studied.
[ { "version": "v1", "created": "Wed, 7 Dec 2022 22:54:07 GMT" } ]
2022-12-09T00:00:00
[ [ "Ibarra", "Oscar H.", "" ], [ "McQuillan", "Ian", "" ] ]
new_dataset
0.982774
2212.04005
Jinyoung Park
Jinyoung Park, Minseok Son, Seungju Cho, Inyoung Lee, Changick Kim
RainUNet for Super-Resolution Rain Movie Prediction under Spatio-temporal Shifts
NeurIPS 2022, Weather4Cast core challenge
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper presents a solution to the Weather4cast 2022 Challenge Stage 2. The goal of the challenge is to forecast future high-resolution rainfall events obtained from ground radar using low-resolution multiband satellite images. We suggest a solution that performs data preprocessing appropriate to the challenge and then predicts rainfall movies using a novel RainUNet. RainUNet is a hierarchical U-shaped network with temporal-wise separable block (TS block) using a decoupled large kernel 3D convolution to improve the prediction performance. Various evaluation metrics show that our solution is effective compared to the baseline method. The source codes are available at https://github.com/jinyxp/Weather4cast-2022
[ { "version": "v1", "created": "Wed, 7 Dec 2022 23:42:39 GMT" } ]
2022-12-09T00:00:00
[ [ "Park", "Jinyoung", "" ], [ "Son", "Minseok", "" ], [ "Cho", "Seungju", "" ], [ "Lee", "Inyoung", "" ], [ "Kim", "Changick", "" ] ]
new_dataset
0.998956
2212.04018
Kartik Pant
Kartik Anand Pant, Zhanpeng Yang, James M Goppert, and Inseok Hwang
An Open-Source Gazebo Plugin for GNSS Multipath Signal Emulation in Virtual Urban Canyons
13 pages, 8 figures
null
null
null
cs.RO
http://creativecommons.org/licenses/by/4.0/
One of the major errors affecting GNSS signals in urban canyons is GNSS multipath error. In this work, we develop a Gazebo plugin which utilizes a ray tracing technique to account for multipath effects in a virtual urban canyon environment using virtual satellites. This software plugin balances accuracy and computational complexity to run the simulation in real-time for both software-in-the-loop (SITL) and hardware-in-the-loop (HITL) testing. We also construct a 3D virtual environment of Hong Kong and compare the results from our plugin with the GNSS data in the publicly available Urban-Nav dataset, to validate the efficacy of the proposed Gazebo Plugin. The plugin is openly available to all the researchers in the robotics community. https://github.com/kpant14/multipath_sim
[ { "version": "v1", "created": "Thu, 8 Dec 2022 00:44:49 GMT" } ]
2022-12-09T00:00:00
[ [ "Pant", "Kartik Anand", "" ], [ "Yang", "Zhanpeng", "" ], [ "Goppert", "James M", "" ], [ "Hwang", "Inseok", "" ] ]
new_dataset
0.963218
2212.04058
Dalin Zhang
Xinle Wu, Dalin Zhang, Miao Zhang, Chenjuan Guo, Shuai Zhao, Yi Zhang, Huai Wang, Bin Yang
AutoPINN: When AutoML Meets Physics-Informed Neural Networks
null
null
null
null
cs.LG cs.AI
http://creativecommons.org/licenses/by/4.0/
Physics-Informed Neural Networks (PINNs) have recently been proposed to solve scientific and engineering problems, where physical laws are introduced into neural networks as prior knowledge. With the embedded physical laws, PINNs enable the estimation of critical parameters, which are unobservable via physical tools, through observable variables. For example, Power Electronic Converters (PECs) are essential building blocks for the green energy transition. PINNs have been applied to estimate the capacitance, which is unobservable during PEC operations, using current and voltage, which can be observed easily during operations. The estimated capacitance facilitates self-diagnostics of PECs. Existing PINNs are often manually designed, which is time-consuming and may lead to suboptimal performance due to a large number of design choices for neural network architectures and hyperparameters. In addition, PINNs are often deployed on different physical devices, e.g., PECs, with limited and varying resources. Therefore, it requires designing different PINN models under different resource constraints, making it an even more challenging task for manual design. To contend with the challenges, we propose Automated Physics-Informed Neural Networks (AutoPINN), a framework that enables the automated design of PINNs by combining AutoML and PINNs. Specifically, we first tailor a search space that allows finding high-accuracy PINNs for PEC internal parameter estimation. We then propose a resource-aware search strategy to explore the search space to find the best PINN model under different resource constraints. We experimentally demonstrate that AutoPINN is able to find more accurate PINN models than human-designed, state-of-the-art PINN models using fewer resources.
[ { "version": "v1", "created": "Thu, 8 Dec 2022 03:44:08 GMT" } ]
2022-12-09T00:00:00
[ [ "Wu", "Xinle", "" ], [ "Zhang", "Dalin", "" ], [ "Zhang", "Miao", "" ], [ "Guo", "Chenjuan", "" ], [ "Zhao", "Shuai", "" ], [ "Zhang", "Yi", "" ], [ "Wang", "Huai", "" ], [ "Yang", "Bin", "" ] ]
new_dataset
0.998084
2212.04061
Sibendu Paul
Sibendu Paul, Kunal Rao, Giuseppe Coviello, Murugan Sankaradas, Oliver Po, Y. Charlie Hu, Srimat T. Chakradhar
Elixir: A system to enhance data quality for multiple analytics on a video stream
null
null
null
null
cs.CV cs.MA
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
IoT sensors, especially video cameras, are ubiquitously deployed around the world to perform a variety of computer vision tasks in several verticals including retail, healthcare, safety and security, transportation, manufacturing, etc. To amortize their high deployment effort and cost, it is desirable to perform multiple video analytics tasks, which we refer to as Analytical Units (AUs), off the video feed coming out of every camera. In this paper, we first show that in a multi-AU setting, changing the camera setting has disproportionate impact on different AUs performance. In particular, the optimal setting for one AU may severely degrade the performance for another AU, and further the impact on different AUs varies as the environmental condition changes. We then present Elixir, a system to enhance the video stream quality for multiple analytics on a video stream. Elixir leverages Multi-Objective Reinforcement Learning (MORL), where the RL agent caters to the objectives from different AUs and adjusts the camera setting to simultaneously enhance the performance of all AUs. To define the multiple objectives in MORL, we develop new AU-specific quality estimator values for each individual AU. We evaluate Elixir through real-world experiments on a testbed with three cameras deployed next to each other (overlooking a large enterprise parking lot) running Elixir and two baseline approaches, respectively. Elixir correctly detects 7.1% (22,068) and 5.0% (15,731) more cars, 94% (551) and 72% (478) more faces, and 670.4% (4975) and 158.6% (3507) more persons than the default-setting and time-sharing approaches, respectively. It also detects 115 license plates, far more than the time-sharing approach (7) and the default setting (0).
[ { "version": "v1", "created": "Thu, 8 Dec 2022 04:04:58 GMT" } ]
2022-12-09T00:00:00
[ [ "Paul", "Sibendu", "" ], [ "Rao", "Kunal", "" ], [ "Coviello", "Giuseppe", "" ], [ "Sankaradas", "Murugan", "" ], [ "Po", "Oliver", "" ], [ "Hu", "Y. Charlie", "" ], [ "Chakradhar", "Srimat T.", "" ] ]
new_dataset
0.978789
2212.04119
ByungSoo Ko
Young-Jun Lee, Byungsoo Ko, Han-Gyu Kim, Ho-Jin Choi
DialogCC: Large-Scale Multi-Modal Dialogue Dataset
null
null
null
null
cs.CV cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
As sharing images in an instant message is a crucial factor, there has been active research on learning a image-text multi-modal dialogue model. However, training a well-generalized multi-modal dialogue model is challenging because existing multi-modal dialogue datasets contain a small number of data, limited topics, and a restricted variety of images per dialogue. In this paper, we present a multi-modal dialogue dataset creation pipeline that involves matching large-scale images to dialogues based on CLIP similarity. Using this automatic pipeline, we propose a large-scale multi-modal dialogue dataset, DialogCC, which covers diverse real-world topics and various images per dialogue. With extensive experiments, we demonstrate that training a multi-modal dialogue model with our dataset can improve generalization performance. Additionally, existing models trained with our dataset achieve state-of-the-art performance on image and text retrieval tasks. The source code and the dataset will be released after publication.
[ { "version": "v1", "created": "Thu, 8 Dec 2022 07:29:07 GMT" } ]
2022-12-09T00:00:00
[ [ "Lee", "Young-Jun", "" ], [ "Ko", "Byungsoo", "" ], [ "Kim", "Han-Gyu", "" ], [ "Choi", "Ho-Jin", "" ] ]
new_dataset
0.999753
2212.04138
Yiannis Kantaros
Kaiyuan Tan, Jun Wang, Yiannis Kantaros
Targeted Adversarial Attacks against Neural Network Trajectory Predictors
null
null
null
null
cs.LG cs.CR
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Trajectory prediction is an integral component of modern autonomous systems as it allows for envisioning future intentions of nearby moving agents. Due to the lack of other agents' dynamics and control policies, deep neural network (DNN) models are often employed for trajectory forecasting tasks. Although there exists an extensive literature on improving the accuracy of these models, there is a very limited number of works studying their robustness against adversarially crafted input trajectories. To bridge this gap, in this paper, we propose a targeted adversarial attack against DNN models for trajectory forecasting tasks. We call the proposed attack TA4TP for Targeted adversarial Attack for Trajectory Prediction. Our approach generates adversarial input trajectories that are capable of fooling DNN models into predicting user-specified target/desired trajectories. Our attack relies on solving a nonlinear constrained optimization problem where the objective function captures the deviation of the predicted trajectory from a target one while the constraints model physical requirements that the adversarial input should satisfy. The latter ensures that the inputs look natural and they are safe to execute (e.g., they are close to nominal inputs and away from obstacles). We demonstrate the effectiveness of TA4TP on two state-of-the-art DNN models and two datasets. To the best of our knowledge, we propose the first targeted adversarial attack against DNN models used for trajectory forecasting.
[ { "version": "v1", "created": "Thu, 8 Dec 2022 08:34:28 GMT" } ]
2022-12-09T00:00:00
[ [ "Tan", "Kaiyuan", "" ], [ "Wang", "Jun", "" ], [ "Kantaros", "Yiannis", "" ] ]
new_dataset
0.978611
2212.04163
Yijun Wang
Yijun Wang, Rui Lang, Rui Li and Junsong Zhang
NRTR: Neuron Reconstruction with Transformer from 3D Optical Microscopy Images
null
null
null
null
cs.CV cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The neuron reconstruction from raw Optical Microscopy (OM) image stacks is the basis of neuroscience. Manual annotation and semi-automatic neuron tracing algorithms are time-consuming and inefficient. Existing deep learning neuron reconstruction methods, although demonstrating exemplary performance, greatly demand complex rule-based components. Therefore, a crucial challenge is designing an end-to-end neuron reconstruction method that makes the overall framework simpler and model training easier. We propose a Neuron Reconstruction Transformer (NRTR) that, discarding the complex rule-based components, views neuron reconstruction as a direct set-prediction problem. To the best of our knowledge, NRTR is the first image-to-set deep learning model for end-to-end neuron reconstruction. In experiments using the BigNeuron and VISoR-40 datasets, NRTR achieves excellent neuron reconstruction results for comprehensive benchmarks and outperforms competitive baselines. Results of extensive experiments indicate that NRTR is effective at showing that neuron reconstruction is viewed as a set-prediction problem, which makes end-to-end model training available.
[ { "version": "v1", "created": "Thu, 8 Dec 2022 09:35:22 GMT" } ]
2022-12-09T00:00:00
[ [ "Wang", "Yijun", "" ], [ "Lang", "Rui", "" ], [ "Li", "Rui", "" ], [ "Zhang", "Junsong", "" ] ]
new_dataset
0.974197
2212.04166
Yannick Schmitz
Marcel Wagner, Yannick Schmitz and Egon Wanke
On the strong metric dimension of composed graphs
null
null
null
null
cs.CC
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Two vertices $u$ and $v$ of an undirected graph $G$ are strongly resolved by a vertex $w$ if there is a shortest path between $w$ and $u$ containing $v$ or a shortest path between $w$ and $v$ containing $u$. A vertex set $R$ is a strong resolving set for $G$ if for each pair of vertices there is a vertex in $R$ that strongly resolves them. The strong metric dimension of $G$ is the size of a minimum strong resolving set for $G$. We show that a minimum strong resolving set for an undirected graph $G$ can be computed efficiently if and only if a minimum strong resolving set for each biconnected component of $G$ can be computed efficiently.
[ { "version": "v1", "created": "Thu, 8 Dec 2022 09:41:58 GMT" } ]
2022-12-09T00:00:00
[ [ "Wagner", "Marcel", "" ], [ "Schmitz", "Yannick", "" ], [ "Wanke", "Egon", "" ] ]
new_dataset
0.994811
2212.04175
Kan Huang
Kan Huang, Kai Zhang, Ming Liu
GreenEyes: An Air Quality Evaluating Model based on WaveNet
null
null
null
null
cs.LG eess.SP
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Accompanying rapid industrialization, humans are suffering from serious air pollution problems. The demand for air quality prediction is becoming more and more important to the government's policy-making and people's daily life. In this paper, We propose GreenEyes -- a deep neural network model, which consists of a WaveNet-based backbone block for learning representations of sequences and an LSTM with a Temporal Attention module for capturing the hidden interactions between features of multi-channel inputs. To evaluate the effectiveness of our proposed method, we carry out several experiments including an ablation study on our collected and preprocessed air quality data near HKUST. The experimental results show our model can effectively predict the air quality level of the next timestamp given any segment of the air quality data from the data set. We have also released our standalone dataset at https://github.com/AI-Huang/IAQI_Dataset The model and code for this paper are publicly available at https://github.com/AI-Huang/AirEvaluation
[ { "version": "v1", "created": "Thu, 8 Dec 2022 10:28:57 GMT" } ]
2022-12-09T00:00:00
[ [ "Huang", "Kan", "" ], [ "Zhang", "Kai", "" ], [ "Liu", "Ming", "" ] ]
new_dataset
0.979828
2212.04197
Yuekai Jia
Yuekai Jia, Shuang Liu, Wenhao Wang, Yu Chen, Zhengde Zhai, Shoumeng Yan, Zhengyu He
HyperEnclave: An Open and Cross-platform Trusted Execution Environment
null
In 2022 USENIX Annual Technical Conference (USENIX ATC 22), pages 437-454, Carlsbad, CA, July 2022. USENIX Association
null
null
cs.CR cs.AR cs.OS
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
A number of trusted execution environments (TEEs) have been proposed by both academia and industry. However, most of them require specific hardware or firmware changes and are bound to specific hardware vendors (such as Intel, AMD, ARM, and IBM). In this paper, we propose HyperEnclave, an open and cross-platform process-based TEE that relies on the widely-available virtualization extension to create the isolated execution environment. In particular, HyperEnclave is designed to support the flexible enclave operation modes to fulfill the security and performance demands under various enclave workloads. We provide the enclave SDK to run existing SGX programs on HyperEnclave with little or no source code changes. We have implemented HyperEnclave on commodity AMD servers and deployed the system in a world-leading FinTech company to support real-world privacy-preserving computations. The evaluation on both micro-benchmarks and application benchmarks shows the design of HyperEnclave introduces only a small overhead.
[ { "version": "v1", "created": "Thu, 8 Dec 2022 11:23:48 GMT" } ]
2022-12-09T00:00:00
[ [ "Jia", "Yuekai", "" ], [ "Liu", "Shuang", "" ], [ "Wang", "Wenhao", "" ], [ "Chen", "Yu", "" ], [ "Zhai", "Zhengde", "" ], [ "Yan", "Shoumeng", "" ], [ "He", "Zhengyu", "" ] ]
new_dataset
0.998784
2212.04229
Prashant Hari Narayan Rajput
Prashant Hari Narayan Rajput (1), Constantine Doumanidis (2), Michail Maniatakos (2) ((1) NYU Tandon School of Engineering, (2) New York University Abu Dhabi)
ICSPatch: Automated Vulnerability Localization and Non-Intrusive Hotpatching in Industrial Control Systems using Data Dependence Graphs
To appear in the 32nd USENIX Security Symposium, August 2023, Anaheim, CA, USA [16 pages, 12 figures, 5 tables, code available at https://github.com/momalab/ICSPatch]
null
null
null
cs.CR cs.SY eess.SY
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The paradigm shift of enabling extensive intercommunication between the Operational Technology (OT) and Information Technology (IT) devices allows vulnerabilities typical to the IT world to propagate to the OT side. Therefore, the security layer offered in the past by air gapping is removed, making security patching for OT devices a hard requirement. Conventional patching involves a device reboot to load the patched code in the main memory, which does not apply to OT devices controlling critical processes due to downtime, necessitating in-memory vulnerability patching. Furthermore, these control binaries are often compiled by in-house proprietary compilers, further hindering the patching process and placing reliance on OT vendors for rapid vulnerability discovery and patch development. The current state-of-the-art hotpatching approaches only focus on firmware and/or RTOS. Therefore, in this work, we develop ICSPatch, a framework to automate control logic vulnerability localization using Data Dependence Graphs (DDGs). With the help of DDGs, ICSPatch pinpoints the vulnerability in the control application. As an independent second step, ICSPatch can non-intrusively hotpatch vulnerabilities in the control application directly in the main memory of Programmable Logic Controllers while maintaining reliable continuous operation. To evaluate our framework, we test ICSPatch on a synthetic dataset of 24 vulnerable control application binaries from diverse critical infrastructure sectors. Results show that ICSPatch could successfully localize all vulnerabilities and generate patches accordingly. Furthermore, the patch added negligible latency increase in the execution cycle while maintaining correctness and protection against the vulnerability.
[ { "version": "v1", "created": "Thu, 8 Dec 2022 12:26:15 GMT" } ]
2022-12-09T00:00:00
[ [ "Rajput", "Prashant Hari Narayan", "" ], [ "Doumanidis", "Constantine", "" ], [ "Maniatakos", "Michail", "" ] ]
new_dataset
0.981054
2212.04234
Xiaoyang Shan
Lihai Nie, Xiaoyang Shan, Laiping Zhao, Keqiu Li
PKDGA: A Partial Knowledge-based Domain Generation Algorithm for Botnets
12 pages,11 figures
null
null
null
cs.CR
http://creativecommons.org/licenses/by/4.0/
Domain generation algorithms (DGAs) can be categorized into three types: zero-knowledge, partial-knowledge, and full-knowledge. While prior research merely focused on zero-knowledge and full-knowledge types, we characterize their anti-detection ability and practicality and find that zero-knowledge DGAs present low anti-detection ability against detectors, and full-knowledge DGAs suffer from low practicality due to the strong assumption that they are fully detector-aware. Given these observations, we propose PKDGA, a partial knowledge-based domain generation algorithm with high anti-detection ability and high practicality. PKDGA employs the reinforcement learning architecture, which makes it evolve automatically based only on the easily-observable feedback from detectors. We evaluate PKDGA using a comprehensive set of real-world datasets, and the results demonstrate that it reduces the detection performance of existing detectors from 91.7% to 52.5%. We further apply PKDGA to the Mirai malware, and the evaluations show that the proposed method is quite lightweight and time-efficient.
[ { "version": "v1", "created": "Thu, 8 Dec 2022 12:31:57 GMT" } ]
2022-12-09T00:00:00
[ [ "Nie", "Lihai", "" ], [ "Shan", "Xiaoyang", "" ], [ "Zhao", "Laiping", "" ], [ "Li", "Keqiu", "" ] ]
new_dataset
0.996743
2212.04264
Kaan Ak\c{s}it
Ahmet G\"uzel, Jeanne Beyazian, Praneeth Chakravarthula and Kaan Ak\c{s}it
ChromaCorrect: Prescription Correction in Virtual Reality Headsets through Perceptual Guidance
12 pages, 9 figures, 1 table, 1 listing
null
null
null
cs.HC cs.GR cs.LG
http://creativecommons.org/licenses/by-nc-nd/4.0/
A large portion of today's world population suffer from vision impairments and wear prescription eyeglasses. However, eyeglasses causes additional bulk and discomfort when used with augmented and virtual reality headsets, thereby negatively impacting the viewer's visual experience. In this work, we remedy the usage of prescription eyeglasses in Virtual Reality (VR) headsets by shifting the optical complexity completely into software and propose a prescription-aware rendering approach for providing sharper and immersive VR imagery. To this end, we develop a differentiable display and visual perception model encapsulating display-specific parameters, color and visual acuity of human visual system and the user-specific refractive errors. Using this differentiable visual perception model, we optimize the rendered imagery in the display using stochastic gradient-descent solvers. This way, we provide prescription glasses-free sharper images for a person with vision impairments. We evaluate our approach on various displays, including desktops and VR headsets, and show significant quality and contrast improvements for users with vision impairments.
[ { "version": "v1", "created": "Thu, 8 Dec 2022 13:30:17 GMT" } ]
2022-12-09T00:00:00
[ [ "Güzel", "Ahmet", "" ], [ "Beyazian", "Jeanne", "" ], [ "Chakravarthula", "Praneeth", "" ], [ "Akşit", "Kaan", "" ] ]
new_dataset
0.996192
2212.04320
Guodong Yin
Guodong Yin, Mufeng Zhou, Yiming Chen, Wenjun Tang, Zekun Yang, Mingyen Lee, Xirui Du, Jinshan Yue, Jiaxin Liu, Huazhong Yang, Yongpan Liu, Xueqing Li
A 65nm 8b-Activation 8b-Weight SRAM-Based Charge-Domain Computing-in-Memory Macro Using A Fully-Parallel Analog Adder Network and A Single-ADC Interface
Accepted by IEEE 48th European Solid-State Circuits Conference (ESSCIRC 2022)
null
null
null
cs.AR cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Performing data-intensive tasks in the von Neumann architecture is challenging to achieve both high performance and power efficiency due to the memory wall bottleneck. Computing-in-memory (CiM) is a promising mitigation approach by enabling parallel in-situ multiply-accumulate (MAC) operations within the memory with support from the peripheral interface and datapath. SRAM-based charge-domain CiM (CD-CiM) has shown its potential of enhanced power efficiency and computing accuracy. However, existing SRAM-based CD-CiM faces scaling challenges to meet the throughput requirement of high-performance multi-bit-quantization applications. This paper presents an SRAM-based high-throughput ReLU-optimized CD-CiM macro. It is capable of completing MAC and ReLU of two signed 8b vectors in one CiM cycle with only one A/D conversion. Along with non-linearity compensation for the analog computing and A/D conversion interfaces, this work achieves 51.2GOPS throughput and 10.3TOPS/W energy efficiency, while showing 88.6% accuracy in the CIFAR-10 dataset.
[ { "version": "v1", "created": "Wed, 23 Nov 2022 07:52:10 GMT" } ]
2022-12-09T00:00:00
[ [ "Yin", "Guodong", "" ], [ "Zhou", "Mufeng", "" ], [ "Chen", "Yiming", "" ], [ "Tang", "Wenjun", "" ], [ "Yang", "Zekun", "" ], [ "Lee", "Mingyen", "" ], [ "Du", "Xirui", "" ], [ "Yue", "Jinshan", "" ], [ "Liu", "Jiaxin", "" ], [ "Yang", "Huazhong", "" ], [ "Liu", "Yongpan", "" ], [ "Li", "Xueqing", "" ] ]
new_dataset
0.998136
2212.04357
Kaifa Zhao
Kaifa Zhao, Le Yu, Shiyao Zhou, Jing Li, Xiapu Luo, Yat Fei Aemon Chiu, Yutong Liu
A Fine-grained Chinese Software Privacy Policy Dataset for Sequence Labeling and Regulation Compliant Identification
Accepted by EMNLP 2022 (The 2022 Conference on Empirical Methods in Natural Language Processing)
null
null
null
cs.CR cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Privacy protection raises great attention on both legal levels and user awareness. To protect user privacy, countries enact laws and regulations requiring software privacy policies to regulate their behavior. However, privacy policies are written in natural languages with many legal terms and software jargon that prevent users from understanding and even reading them. It is desirable to use NLP techniques to analyze privacy policies for helping users understand them. Furthermore, existing datasets ignore law requirements and are limited to English. In this paper, we construct the first Chinese privacy policy dataset, namely CA4P-483, to facilitate the sequence labeling tasks and regulation compliance identification between privacy policies and software. Our dataset includes 483 Chinese Android application privacy policies, over 11K sentences, and 52K fine-grained annotations. We evaluate families of robust and representative baseline models on our dataset. Based on baseline performance, we provide findings and potential research directions on our dataset. Finally, we investigate the potential applications of CA4P-483 combing regulation requirements and program analysis.
[ { "version": "v1", "created": "Sun, 4 Dec 2022 05:59:59 GMT" } ]
2022-12-09T00:00:00
[ [ "Zhao", "Kaifa", "" ], [ "Yu", "Le", "" ], [ "Zhou", "Shiyao", "" ], [ "Li", "Jing", "" ], [ "Luo", "Xiapu", "" ], [ "Chiu", "Yat Fei Aemon", "" ], [ "Liu", "Yutong", "" ] ]
new_dataset
0.999811
2212.04360
Hongwei Yi
Hongwei Yi, Chun-Hao P. Huang, Shashank Tripathi, Lea Hering, Justus Thies, Michael J. Black
MIME: Human-Aware 3D Scene Generation
Project Page: https://mime.is.tue.mpg.de
null
null
null
cs.CV cs.GR
http://creativecommons.org/licenses/by/4.0/
Generating realistic 3D worlds occupied by moving humans has many applications in games, architecture, and synthetic data creation. But generating such scenes is expensive and labor intensive. Recent work generates human poses and motions given a 3D scene. Here, we take the opposite approach and generate 3D indoor scenes given 3D human motion. Such motions can come from archival motion capture or from IMU sensors worn on the body, effectively turning human movement in a "scanner" of the 3D world. Intuitively, human movement indicates the free-space in a room and human contact indicates surfaces or objects that support activities such as sitting, lying or touching. We propose MIME (Mining Interaction and Movement to infer 3D Environments), which is a generative model of indoor scenes that produces furniture layouts that are consistent with the human movement. MIME uses an auto-regressive transformer architecture that takes the already generated objects in the scene as well as the human motion as input, and outputs the next plausible object. To train MIME, we build a dataset by populating the 3D FRONT scene dataset with 3D humans. Our experiments show that MIME produces more diverse and plausible 3D scenes than a recent generative scene method that does not know about human movement. Code and data will be available for research at https://mime.is.tue.mpg.de.
[ { "version": "v1", "created": "Thu, 8 Dec 2022 15:56:17 GMT" } ]
2022-12-09T00:00:00
[ [ "Yi", "Hongwei", "" ], [ "Huang", "Chun-Hao P.", "" ], [ "Tripathi", "Shashank", "" ], [ "Hering", "Lea", "" ], [ "Thies", "Justus", "" ], [ "Black", "Michael J.", "" ] ]
new_dataset
0.999553
2212.04408
Xuancheng Ren
Jinze Bai, Rui Men, Hao Yang, Xuancheng Ren, Kai Dang, Yichang Zhang, Xiaohuan Zhou, Peng Wang, Sinan Tan, An Yang, Zeyu Cui, Yu Han, Shuai Bai, Wenbin Ge, Jianxin Ma, Junyang Lin, Jingren Zhou, Chang Zhou
OFASys: A Multi-Modal Multi-Task Learning System for Building Generalist Models
null
null
null
null
cs.CV cs.AI cs.CL cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Generalist models, which are capable of performing diverse multi-modal tasks in a task-agnostic way within a single model, have been explored recently. Being, hopefully, an alternative to approaching general-purpose AI, existing generalist models are still at an early stage, where modality and task coverage is limited. To empower multi-modal task-scaling and speed up this line of research, we release a generalist model learning system, OFASys, built on top of a declarative task interface named multi-modal instruction. At the core of OFASys is the idea of decoupling multi-modal task representations from the underlying model implementations. In OFASys, a task involving multiple modalities can be defined declaratively even with just a single line of code. The system automatically generates task plans from such instructions for training and inference. It also facilitates multi-task training for diverse multi-modal workloads. As a starting point, we provide presets of 7 different modalities and 23 highly-diverse example tasks in OFASys, with which we also develop a first-in-kind, single model, OFA+, that can handle text, image, speech, video, and motion data. The single OFA+ model achieves 95% performance in average with only 16% parameters of 15 task-finetuned models, showcasing the performance reliability of multi-modal task-scaling provided by OFASys. Available at https://github.com/OFA-Sys/OFASys
[ { "version": "v1", "created": "Thu, 8 Dec 2022 17:07:09 GMT" } ]
2022-12-09T00:00:00
[ [ "Bai", "Jinze", "" ], [ "Men", "Rui", "" ], [ "Yang", "Hao", "" ], [ "Ren", "Xuancheng", "" ], [ "Dang", "Kai", "" ], [ "Zhang", "Yichang", "" ], [ "Zhou", "Xiaohuan", "" ], [ "Wang", "Peng", "" ], [ "Tan", "Sinan", "" ], [ "Yang", "An", "" ], [ "Cui", "Zeyu", "" ], [ "Han", "Yu", "" ], [ "Bai", "Shuai", "" ], [ "Ge", "Wenbin", "" ], [ "Ma", "Jianxin", "" ], [ "Lin", "Junyang", "" ], [ "Zhou", "Jingren", "" ], [ "Zhou", "Chang", "" ] ]
new_dataset
0.968966
2212.04437
Benjamin Fele
Benjamin Fele and Ajda Lampe and Peter Peer and Vitomir \v{S}truc
C-VTON: Context-Driven Image-Based Virtual Try-On Network
Accepted to WACV 2022
null
10.1109/WACV51458.2022.00226
null
cs.CV
http://creativecommons.org/licenses/by/4.0/
Image-based virtual try-on techniques have shown great promise for enhancing the user-experience and improving customer satisfaction on fashion-oriented e-commerce platforms. However, existing techniques are currently still limited in the quality of the try-on results they are able to produce from input images of diverse characteristics. In this work, we propose a Context-Driven Virtual Try-On Network (C-VTON) that addresses these limitations and convincingly transfers selected clothing items to the target subjects even under challenging pose configurations and in the presence of self-occlusions. At the core of the C-VTON pipeline are: (i) a geometric matching procedure that efficiently aligns the target clothing with the pose of the person in the input images, and (ii) a powerful image generator that utilizes various types of contextual information when synthesizing the final try-on result. C-VTON is evaluated in rigorous experiments on the VITON and MPV datasets and in comparison to state-of-the-art techniques from the literature. Experimental results show that the proposed approach is able to produce photo-realistic and visually convincing results and significantly improves on the existing state-of-the-art.
[ { "version": "v1", "created": "Thu, 8 Dec 2022 17:56:34 GMT" } ]
2022-12-09T00:00:00
[ [ "Fele", "Benjamin", "" ], [ "Lampe", "Ajda", "" ], [ "Peer", "Peter", "" ], [ "Štruc", "Vitomir", "" ] ]
new_dataset
0.999727
2212.04498
Deepak Pathak
Kenneth Shaw, Shikhar Bahl, Deepak Pathak
VideoDex: Learning Dexterity from Internet Videos
Accepted at CoRL 2022. Website at https://video-dex.github.io
null
null
null
cs.RO cs.AI cs.CV cs.LG cs.SY eess.SY
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
To build general robotic agents that can operate in many environments, it is often imperative for the robot to collect experience in the real world. However, this is often not feasible due to safety, time, and hardware restrictions. We thus propose leveraging the next best thing as real-world experience: internet videos of humans using their hands. Visual priors, such as visual features, are often learned from videos, but we believe that more information from videos can be utilized as a stronger prior. We build a learning algorithm, VideoDex, that leverages visual, action, and physical priors from human video datasets to guide robot behavior. These actions and physical priors in the neural network dictate the typical human behavior for a particular robot task. We test our approach on a robot arm and dexterous hand-based system and show strong results on various manipulation tasks, outperforming various state-of-the-art methods. Videos at https://video-dex.github.io
[ { "version": "v1", "created": "Thu, 8 Dec 2022 18:59:59 GMT" } ]
2022-12-09T00:00:00
[ [ "Shaw", "Kenneth", "" ], [ "Bahl", "Shikhar", "" ], [ "Pathak", "Deepak", "" ] ]
new_dataset
0.955025
2104.12290
Gokhan Egri
Gokhan Egri, Todd Zickler
StegaPos: Preventing Unwanted Crops and Replacements with Imperceptible Positional Embeddings
For CVPR 2022 submission, 8 pages (main)
null
null
null
cs.CV cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We present a learned, spatially-varying steganography system that allows detecting when and how images have been altered by cropping, splicing or inpainting after publication. The system comprises a learned encoder that imperceptibly hides distinct positional signatures in every local image region before publication, and an accompanying learned decoder that extracts the steganographic signatures to determine, for each local image region, its 2D positional coordinates within the originally-published image. Crop and replacement edits become detectable by the inconsistencies they cause in the hidden positional signatures. Using a prototype system for small $(400 \times 400)$ images, we show experimentally that simple CNN encoder and decoder architectures can be trained jointly to achieve detection that is reliable and robust, without introducing perceptible distortion. This approach could help individuals and image-sharing platforms certify that an image was published by a trusted source, and also know which parts of such an image, if any, have been substantially altered since publication.
[ { "version": "v1", "created": "Sun, 25 Apr 2021 23:42:29 GMT" }, { "version": "v2", "created": "Wed, 7 Dec 2022 07:11:09 GMT" } ]
2022-12-08T00:00:00
[ [ "Egri", "Gokhan", "" ], [ "Zickler", "Todd", "" ] ]
new_dataset
0.997787
2105.08209
Wojciech Kry\'sci\'nski
Wojciech Kry\'sci\'nski, Nazneen Rajani, Divyansh Agarwal, Caiming Xiong, Dragomir Radev
BookSum: A Collection of Datasets for Long-form Narrative Summarization
19 pages, 12 tables, 3 figures
null
null
null
cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The majority of available text summarization datasets include short-form source documents that lack long-range causal and temporal dependencies, and often contain strong layout and stylistic biases. While relevant, such datasets will offer limited challenges for future generations of text summarization systems. We address these issues by introducing BookSum, a collection of datasets for long-form narrative summarization. Our dataset covers source documents from the literature domain, such as novels, plays and stories, and includes highly abstractive, human written summaries on three levels of granularity of increasing difficulty: paragraph-, chapter-, and book-level. The domain and structure of our dataset poses a unique set of challenges for summarization systems, which include: processing very long documents, non-trivial causal and temporal dependencies, and rich discourse structures. To facilitate future work, we trained and evaluated multiple extractive and abstractive summarization models as baselines for our dataset.
[ { "version": "v1", "created": "Tue, 18 May 2021 00:22:46 GMT" }, { "version": "v2", "created": "Tue, 6 Dec 2022 19:19:35 GMT" } ]
2022-12-08T00:00:00
[ [ "Kryściński", "Wojciech", "" ], [ "Rajani", "Nazneen", "" ], [ "Agarwal", "Divyansh", "" ], [ "Xiong", "Caiming", "" ], [ "Radev", "Dragomir", "" ] ]
new_dataset
0.999591
2112.01238
Kyle McDonald
Kyle McDonald
Ethereum Emissions: A Bottom-up Estimate
Code at https://github.com/kylemcdonald/ethereum-emissions
null
null
null
cs.CY cs.CR math.OC
http://creativecommons.org/licenses/by/4.0/
The Ethereum ecosystem was maintained by a distributed global network of computers that required massive amounts of computational power. Previous work on estimating the energy use and emissions of the Ethereum network has relied on top-down economic analysis and rough estimates of hardware efficiency and emissions factors. In this work we provide a bottom-up analysis that works from hashrate to an energy usage estimate, and from mining locations to an emissions factor estimate, and combines these for an overall emissions estimate. We analyze the entire history of PoW Ethereum, from creation to the merge.
[ { "version": "v1", "created": "Fri, 19 Nov 2021 11:05:48 GMT" }, { "version": "v2", "created": "Fri, 3 Dec 2021 02:18:26 GMT" }, { "version": "v3", "created": "Wed, 7 Dec 2022 08:24:55 GMT" } ]
2022-12-08T00:00:00
[ [ "McDonald", "Kyle", "" ] ]
new_dataset
0.997269
2201.09750
Bilge Celik
Bilge Celik and Prabhant Singh and Joaquin Vanschoren
Online AutoML: An adaptive AutoML framework for online learning
25 pages, 8 figures. Machine Learning S.I.: Automating Data Science (2022)
null
10.1007/s10994-022-06262-0
null
cs.LG cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Automated Machine Learning (AutoML) has been used successfully in settings where the learning task is assumed to be static. In many real-world scenarios, however, the data distribution will evolve over time, and it is yet to be shown whether AutoML techniques can effectively design online pipelines in dynamic environments. This study aims to automate pipeline design for online learning while continuously adapting to data drift. For this purpose, we design an adaptive Online Automated Machine Learning (OAML) system, searching the complete pipeline configuration space of online learners, including preprocessing algorithms and ensembling techniques. This system combines the inherent adaptation capabilities of online learners with the fast automated pipeline (re)optimization capabilities of AutoML. Focusing on optimization techniques that can adapt to evolving objectives, we evaluate asynchronous genetic programming and asynchronous successive halving to optimize these pipelines continually. We experiment on real and artificial data streams with varying types of concept drift to test the performance and adaptation capabilities of the proposed system. The results confirm the utility of OAML over popular online learning algorithms and underscore the benefits of continuous pipeline redesign in the presence of data drift.
[ { "version": "v1", "created": "Mon, 24 Jan 2022 15:37:20 GMT" }, { "version": "v2", "created": "Tue, 10 May 2022 08:57:15 GMT" }, { "version": "v3", "created": "Wed, 7 Dec 2022 10:21:57 GMT" } ]
2022-12-08T00:00:00
[ [ "Celik", "Bilge", "" ], [ "Singh", "Prabhant", "" ], [ "Vanschoren", "Joaquin", "" ] ]
new_dataset
0.974912
2202.03901
George Eskandar
George Eskandar, Sanjeev Sudarsan, Karim Guirguis, Janaranjani Palaniswamy, Bharath Somashekar, Bin Yang
HALS: A Height-Aware Lidar Super-Resolution Framework for Autonomous Driving
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Lidar sensors are costly yet critical for understanding the 3D environment in autonomous driving. High-resolution sensors provide more details about the surroundings because they contain more vertical beams, but they come at a much higher cost, limiting their inclusion in autonomous vehicles. Upsampling lidar pointclouds is a promising approach to gain the benefits of high resolution while maintaining an affordable cost. Although there exist many pointcloud upsampling frameworks, a consistent comparison of these works against each other on the same dataset using unified metrics is still missing. In the first part of this paper, we propose to benchmark existing methods on the Kitti dataset. In the second part, we introduce a novel lidar upsampling model, HALS: Height-Aware Lidar Super-resolution. HALS exploits the observation that lidar scans exhibit a height-aware range distribution and adopts a generator architecture with multiple upsampling branches of different receptive fields. HALS regresses polar coordinates instead of spherical coordinates and uses a surface-normal loss. Extensive experiments show that HALS achieves state-of-the-art performance on 3 real-world lidar datasets.
[ { "version": "v1", "created": "Tue, 8 Feb 2022 14:43:47 GMT" }, { "version": "v2", "created": "Wed, 7 Dec 2022 17:07:54 GMT" } ]
2022-12-08T00:00:00
[ [ "Eskandar", "George", "" ], [ "Sudarsan", "Sanjeev", "" ], [ "Guirguis", "Karim", "" ], [ "Palaniswamy", "Janaranjani", "" ], [ "Somashekar", "Bharath", "" ], [ "Yang", "Bin", "" ] ]
new_dataset
0.999345
2206.12864
Xuefei Yin
Xuefei Yin, Song Wang, Yanming Zhu, Jiankun Hu
A Novel Length-Flexible Lightweight Cancelable Fingerprint Template for Privacy-Preserving Authentication Systems in Resource-Constrained IoT Applications
null
IEEE Internet of Things Journal, 2022
10.1109/JIOT.2022.3204246
null
cs.CR
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Fingerprint authentication techniques have been employed in various Internet of Things (IoT) applications for access control to protect private data, but raw fingerprint template leakage in unprotected IoT applications may render the authentication system insecure. Cancelable fingerprint templates can effectively prevent privacy breaches and provide strong protection to the original templates. However, to suit resource-constrained IoT devices, oversimplified templates would compromise authentication performance significantly. In addition, the length of existing cancelable fingerprint templates is usually fixed, making them difficult to be deployed in various memory-limited IoT devices. To address these issues, we propose a novel length-flexible lightweight cancelable fingerprint template for privacy-preserving authentication systems in various resource-constrained IoT applications. The proposed cancelable template design primarily consists of two components: 1) length-flexible partial-cancelable feature generation based on the designed re-indexing scheme; and 2) lightweight cancelable feature generation based on the designed encoding-nested-difference-XOR scheme. Comprehensive experimental results on public databases~FVC2002 DB1-DB4 and FVC2004 DB1-DB4 demonstrate that the proposed cancelable fingerprint template achieves equivalent authentication performance to state-of-the-art methods in IoT environments, but our design substantially reduces template storage space and computational cost. More importantly, the proposed length-flexible lightweight cancelable template is suitable for a variety of commercial smart cards (e.g., C5-M.O.S.T. Card Contact Microprocessor Smart Cards CLXSU064KC5). To the best of our knowledge, the proposed method is the first length-flexible lightweight, high-performing cancelable fingerprint template design for resource-constrained IoT applications.
[ { "version": "v1", "created": "Sun, 26 Jun 2022 12:47:28 GMT" } ]
2022-12-08T00:00:00
[ [ "Yin", "Xuefei", "" ], [ "Wang", "Song", "" ], [ "Zhu", "Yanming", "" ], [ "Hu", "Jiankun", "" ] ]
new_dataset
0.987155
2207.10805
Xuefei Yin
Xuefei Yin, Yanming Zhu, Yi Xie, Jiankun Hu
PowerFDNet: Deep Learning-Based Stealthy False Data Injection Attack Detection for AC-model Transmission Systems
null
IEEE Open Journal of the Computer Society, 2022
10.1109/OJCS.2022.3199755
null
cs.CR cs.AI cs.LG cs.SY eess.SY
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Recent studies have demonstrated that smart grids are vulnerable to stealthy false data injection attacks (SFDIAs), as SFDIAs can bypass residual-based bad data detection mechanisms. The SFDIA detection has become one of the focuses of smart grid research. Methods based on deep learning technology have shown promising accuracy in the detection of SFDIAs. However, most existing methods rely on the temporal structure of a sequence of measurements but do not take account of the spatial structure between buses and transmission lines. To address this issue, we propose a spatiotemporal deep network, PowerFDNet, for the SFDIA detection in AC-model power grids. The PowerFDNet consists of two sub-architectures: spatial architecture (SA) and temporal architecture (TA). The SA is aimed at extracting representations of bus/line measurements and modeling the spatial structure based on their representations. The TA is aimed at modeling the temporal structure of a sequence of measurements. Therefore, the proposed PowerFDNet can effectively model the spatiotemporal structure of measurements. Case studies on the detection of SFDIAs on the benchmark smart grids show that the PowerFDNet achieved significant improvement compared with the state-of-the-art SFDIA detection methods. In addition, an IoT-oriented lightweight prototype of size 52 MB is implemented and tested for mobile devices, which demonstrates the potential applications on mobile devices. The trained model will be available at \textit{https://github.com/HubYZ/PowerFDNet}.
[ { "version": "v1", "created": "Fri, 15 Jul 2022 00:46:43 GMT" }, { "version": "v2", "created": "Tue, 6 Dec 2022 23:35:01 GMT" } ]
2022-12-08T00:00:00
[ [ "Yin", "Xuefei", "" ], [ "Zhu", "Yanming", "" ], [ "Xie", "Yi", "" ], [ "Hu", "Jiankun", "" ] ]
new_dataset
0.953908
2209.08565
Pranav Page
Pranav S. Page, Kaustubh S. Bhargao, Hrishikesh V. Baviskar, Gaurav S. Kasbekar
Distributed Probabilistic Congestion Control in LEO Satellite Networks
5 pages, 5 figures, conference, poster
null
null
null
cs.NI
http://creativecommons.org/licenses/by/4.0/
In a dense Low Earth Orbit (LEO) satellite constellation, using a centralized algorithm for minimum-delay routing would incur significant signaling and computational overhead. In this work, we exploit the deterministic topology of the constellation to calculate the minimum-delay path between any two nodes in a satellite network. We propose a distributed probabilistic congestion control scheme to minimize end-to-end delay, which is built on top of the existing Datagram Routing Algorithm (DRA). The decision to route packets is taken based on the latest traffic information received from neighbours. We provide an analysis of the congestion caused by a simplified DRA on a uniform infinite mesh of nodes. We compare the proposed congestion control mechanism with the existing congestion control used by the DRA via simulations, and show improvements over the latter.
[ { "version": "v1", "created": "Sun, 18 Sep 2022 13:13:58 GMT" }, { "version": "v2", "created": "Wed, 7 Dec 2022 00:56:00 GMT" } ]
2022-12-08T00:00:00
[ [ "Page", "Pranav S.", "" ], [ "Bhargao", "Kaustubh S.", "" ], [ "Baviskar", "Hrishikesh V.", "" ], [ "Kasbekar", "Gaurav S.", "" ] ]
new_dataset
0.99851
2210.04936
Zhitong Xiong
Zhitong Xiong, Fahong Zhang, Yi Wang, Yilei Shi, Xiao Xiang Zhu
EarthNets: Empowering AI in Earth Observation
28 pages
null
null
null
cs.CV
http://creativecommons.org/licenses/by/4.0/
Earth observation, aiming at monitoring the state of planet Earth using remote sensing data, is critical for improving our daily lives and living environment. With a growing number of satellites in orbit, an increasing number of datasets with diverse sensors and research domains are being published to facilitate the research of the remote sensing community. In this paper, we present a comprehensive review of more than 400 publicly published datasets, including applications like land use/cover, change/disaster monitoring, scene understanding, agriculture, climate change, and weather forecasting. We systematically analyze these Earth observation datasets with respect to five aspects volume, bibliometric analysis, resolution distributions, research domains, and the correlation between datasets. Based on the dataset attributes, we propose to measure, rank, and select datasets to build a new benchmark for model evaluation. Furthermore, a new platform for Earth observation, termed EarthNets, is released as a means of achieving a fair and consistent evaluation of deep learning methods on remote sensing data. EarthNets supports standard dataset libraries and cutting-edge deep learning models to bridge the gap between the remote sensing and machine learning communities. Based on this platform, extensive deep learning methods are evaluated on the new benchmark. The insightful results are beneficial to future research. The platform and dataset collections are publicly available at https://earthnets.github.io/.
[ { "version": "v1", "created": "Mon, 10 Oct 2022 18:09:35 GMT" }, { "version": "v2", "created": "Wed, 7 Dec 2022 15:35:11 GMT" } ]
2022-12-08T00:00:00
[ [ "Xiong", "Zhitong", "" ], [ "Zhang", "Fahong", "" ], [ "Wang", "Yi", "" ], [ "Shi", "Yilei", "" ], [ "Zhu", "Xiao Xiang", "" ] ]
new_dataset
0.998908
2210.11744
Ife Adebara
Ife Adebara, AbdelRahim Elmadany, Muhammad Abdul-Mageed and Alcides Alcoba Inciarte
AfroLID: A Neural Language Identification Tool for African Languages
To appear at EMNLP 2022 Main conference
null
null
null
cs.CL cs.LG
http://creativecommons.org/licenses/by/4.0/
Language identification (LID) is a crucial precursor for NLP, especially for mining web data. Problematically, most of the world's 7000+ languages today are not covered by LID technologies. We address this pressing issue for Africa by introducing AfroLID, a neural LID toolkit for $517$ African languages and varieties. AfroLID exploits a multi-domain web dataset manually curated from across 14 language families utilizing five orthographic systems. When evaluated on our blind Test set, AfroLID achieves 95.89 F_1-score. We also compare AfroLID to five existing LID tools that each cover a small number of African languages, finding it to outperform them on most languages. We further show the utility of AfroLID in the wild by testing it on the acutely under-served Twitter domain. Finally, we offer a number of controlled case studies and perform a linguistically-motivated error analysis that allow us to both showcase AfroLID's powerful capabilities and limitations.
[ { "version": "v1", "created": "Fri, 21 Oct 2022 05:45:50 GMT" }, { "version": "v2", "created": "Mon, 24 Oct 2022 18:25:36 GMT" }, { "version": "v3", "created": "Wed, 7 Dec 2022 04:22:20 GMT" } ]
2022-12-08T00:00:00
[ [ "Adebara", "Ife", "" ], [ "Elmadany", "AbdelRahim", "" ], [ "Abdul-Mageed", "Muhammad", "" ], [ "Inciarte", "Alcides Alcoba", "" ] ]
new_dataset
0.999804
2211.16882
Ashwin Rao
Pranjali Pathre, Anurag Sahu, Ashwin Rao, Avinash Prabhu, Meher Shashwat Nigam, Tanvi Karandikar, Harit Pandya, and K. Madhava Krishna
MVRackLay: Monocular Multi-View Layout Estimation for Warehouse Racks and Shelves
null
IEEE International Conference on Robotics and Biomimetics (ROBIO) 2022
null
null
cs.CV cs.RO
http://creativecommons.org/licenses/by/4.0/
In this paper, we propose and showcase, for the first time, monocular multi-view layout estimation for warehouse racks and shelves. Unlike typical layout estimation methods, MVRackLay estimates multi-layered layouts, wherein each layer corresponds to the layout of a shelf within a rack. Given a sequence of images of a warehouse scene, a dual-headed Convolutional-LSTM architecture outputs segmented racks, the front and the top view layout of each shelf within a rack. With minimal effort, such an output is transformed into a 3D rendering of all racks, shelves and objects on the shelves, giving an accurate 3D depiction of the entire warehouse scene in terms of racks, shelves and the number of objects on each shelf. MVRackLay generalizes to a diverse set of warehouse scenes with varying number of objects on each shelf, number of shelves and in the presence of other such racks in the background. Further, MVRackLay shows superior performance vis-a-vis its single view counterpart, RackLay, in layout accuracy, quantized in terms of the mean IoU and mAP metrics. We also showcase a multi-view stitching of the 3D layouts resulting in a representation of the warehouse scene with respect to a global reference frame akin to a rendering of the scene from a SLAM pipeline. To the best of our knowledge, this is the first such work to portray a 3D rendering of a warehouse scene in terms of its semantic components - Racks, Shelves and Objects - all from a single monocular camera.
[ { "version": "v1", "created": "Wed, 30 Nov 2022 10:32:04 GMT" } ]
2022-12-08T00:00:00
[ [ "Pathre", "Pranjali", "" ], [ "Sahu", "Anurag", "" ], [ "Rao", "Ashwin", "" ], [ "Prabhu", "Avinash", "" ], [ "Nigam", "Meher Shashwat", "" ], [ "Karandikar", "Tanvi", "" ], [ "Pandya", "Harit", "" ], [ "Krishna", "K. Madhava", "" ] ]
new_dataset
0.999592
2212.02936
Constantin Eichenberg
Samuel Weinbach, Marco Bellagente, Constantin Eichenberg, Andrew Dai, Robert Baldock, Souradeep Nanda, Bj\"orn Deiseroth, Koen Oostermeijer, Hannah Teufel, Andres Felipe Cruz-Salinas
M-VADER: A Model for Diffusion with Multimodal Context
22 pages, 14 figures, 2 tables, fixed figure 3
null
null
null
cs.CV
http://creativecommons.org/licenses/by/4.0/
We introduce M-VADER: a diffusion model (DM) for image generation where the output can be specified using arbitrary combinations of images and text. We show how M-VADER enables the generation of images specified using combinations of image and text, and combinations of multiple images. Previously, a number of successful DM image generation algorithms have been introduced that make it possible to specify the output image using a text prompt. Inspired by the success of those models, and led by the notion that language was already developed to describe the elements of visual contexts that humans find most important, we introduce an embedding model closely related to a vision-language model. Specifically, we introduce the embedding model S-MAGMA: a 13 billion parameter multimodal decoder combining components from an autoregressive vision-language model MAGMA and biases finetuned for semantic search.
[ { "version": "v1", "created": "Tue, 6 Dec 2022 12:45:21 GMT" }, { "version": "v2", "created": "Wed, 7 Dec 2022 09:11:18 GMT" } ]
2022-12-08T00:00:00
[ [ "Weinbach", "Samuel", "" ], [ "Bellagente", "Marco", "" ], [ "Eichenberg", "Constantin", "" ], [ "Dai", "Andrew", "" ], [ "Baldock", "Robert", "" ], [ "Nanda", "Souradeep", "" ], [ "Deiseroth", "Björn", "" ], [ "Oostermeijer", "Koen", "" ], [ "Teufel", "Hannah", "" ], [ "Cruz-Salinas", "Andres Felipe", "" ] ]
new_dataset
0.995571
2212.03069
Ngoc Tran
Ngoc N. Tran, Anh Tuan Bui, Dinh Phung, Trung Le
Multiple Perturbation Attack: Attack Pixelwise Under Different $\ell_p$-norms For Better Adversarial Performance
18 pages, 8 figures, 7 tables
null
null
null
cs.CV cs.CR cs.LG
http://creativecommons.org/licenses/by/4.0/
Adversarial machine learning has been both a major concern and a hot topic recently, especially with the ubiquitous use of deep neural networks in the current landscape. Adversarial attacks and defenses are usually likened to a cat-and-mouse game in which defenders and attackers evolve over the time. On one hand, the goal is to develop strong and robust deep networks that are resistant to malicious actors. On the other hand, in order to achieve that, we need to devise even stronger adversarial attacks to challenge these defense models. Most of existing attacks employs a single $\ell_p$ distance (commonly, $p\in\{1,2,\infty\}$) to define the concept of closeness and performs steepest gradient ascent w.r.t. this $p$-norm to update all pixels in an adversarial example in the same way. These $\ell_p$ attacks each has its own pros and cons; and there is no single attack that can successfully break through defense models that are robust against multiple $\ell_p$ norms simultaneously. Motivated by these observations, we come up with a natural approach: combining various $\ell_p$ gradient projections on a pixel level to achieve a joint adversarial perturbation. Specifically, we learn how to perturb each pixel to maximize the attack performance, while maintaining the overall visual imperceptibility of adversarial examples. Finally, through various experiments with standardized benchmarks, we show that our method outperforms most current strong attacks across state-of-the-art defense mechanisms, while retaining its ability to remain clean visually.
[ { "version": "v1", "created": "Mon, 5 Dec 2022 15:38:37 GMT" }, { "version": "v2", "created": "Wed, 7 Dec 2022 18:30:33 GMT" } ]
2022-12-08T00:00:00
[ [ "Tran", "Ngoc N.", "" ], [ "Bui", "Anh Tuan", "" ], [ "Phung", "Dinh", "" ], [ "Le", "Trung", "" ] ]
new_dataset
0.993649
2212.03267
Congyue Deng
Congyue Deng, Chiyu "Max'' Jiang, Charles R. Qi, Xinchen Yan, Yin Zhou, Leonidas Guibas, Dragomir Anguelov
NeRDi: Single-View NeRF Synthesis with Language-Guided Diffusion as General Image Priors
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
2D-to-3D reconstruction is an ill-posed problem, yet humans are good at solving this problem due to their prior knowledge of the 3D world developed over years. Driven by this observation, we propose NeRDi, a single-view NeRF synthesis framework with general image priors from 2D diffusion models. Formulating single-view reconstruction as an image-conditioned 3D generation problem, we optimize the NeRF representations by minimizing a diffusion loss on its arbitrary view renderings with a pretrained image diffusion model under the input-view constraint. We leverage off-the-shelf vision-language models and introduce a two-section language guidance as conditioning inputs to the diffusion model. This is essentially helpful for improving multiview content coherence as it narrows down the general image prior conditioned on the semantic and visual features of the single-view input image. Additionally, we introduce a geometric loss based on estimated depth maps to regularize the underlying 3D geometry of the NeRF. Experimental results on the DTU MVS dataset show that our method can synthesize novel views with higher quality even compared to existing methods trained on this dataset. We also demonstrate our generalizability in zero-shot NeRF synthesis for in-the-wild images.
[ { "version": "v1", "created": "Tue, 6 Dec 2022 19:00:07 GMT" } ]
2022-12-08T00:00:00
[ [ "Deng", "Congyue", "" ], [ "Jiang", "Chiyu \"Max''", "" ], [ "Qi", "Charles R.", "" ], [ "Yan", "Xinchen", "" ], [ "Zhou", "Yin", "" ], [ "Guibas", "Leonidas", "" ], [ "Anguelov", "Dragomir", "" ] ]
new_dataset
0.998365
2212.03273
Tristan Lazard
Tristan Lazard, Marvin Lerousseau, Etienne Decenci\`ere, Thomas Walter
Giga-SSL: Self-Supervised Learning for Gigapixel Images
null
null
null
null
cs.CV cs.LG q-bio.QM
http://creativecommons.org/licenses/by-nc-sa/4.0/
Whole slide images (WSI) are microscopy images of stained tissue slides routinely prepared for diagnosis and treatment selection in medical practice. WSI are very large (gigapixel size) and complex (made of up to millions of cells). The current state-of-the-art (SoTA) approach to classify WSI subdivides them into tiles, encodes them by pre-trained networks and applies Multiple Instance Learning (MIL) to train for specific downstream tasks. However, annotated datasets are often small, typically a few hundred to a few thousand WSI, which may cause overfitting and underperforming models. Conversely, the number of unannotated WSI is ever increasing, with datasets of tens of thousands (soon to be millions) of images available. While it has been previously proposed to use these unannotated data to identify suitable tile representations by self-supervised learning (SSL), downstream classification tasks still require full supervision because parts of the MIL architecture is not trained during tile level SSL pre-training. Here, we propose a strategy of slide level SSL to leverage the large number of WSI without annotations to infer powerful slide representations. Applying our method to The Cancer-Genome Atlas, one of the most widely used data resources in cancer research (16 TB image data), we are able to downsize the dataset to 23 MB without any loss in predictive power: we show that a linear classifier trained on top of these embeddings maintains or improves previous SoTA performances on various benchmark WSI classification tasks. Finally, we observe that training a classifier on these representations with tiny datasets (e.g. 50 slides) improved performances over SoTA by an average of +6.3 AUC points over all downstream tasks.
[ { "version": "v1", "created": "Tue, 6 Dec 2022 19:09:19 GMT" } ]
2022-12-08T00:00:00
[ [ "Lazard", "Tristan", "" ], [ "Lerousseau", "Marvin", "" ], [ "Decencière", "Etienne", "" ], [ "Walter", "Thomas", "" ] ]
new_dataset
0.999427
2212.03287
Guy Amir
Guy Amir, Ziv Freund, Guy Katz, Elad Mandelbaum, Idan Refaeli
veriFIRE: Verifying an Industrial, Learning-Based Wildfire Detection System
To appear in Proceedings of the 25th International Symposium on Formal Methods (FM)
null
null
null
cs.LO cs.LG cs.SE math.OC
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this short paper, we present our ongoing work on the veriFIRE project -- a collaboration between industry and academia, aimed at using verification for increasing the reliability of a real-world, safety-critical system. The system we target is an airborne platform for wildfire detection, which incorporates two deep neural networks. We describe the system and its properties of interest, and discuss our attempts to verify the system's consistency, i.e., its ability to continue and correctly classify a given input, even if the wildfire it describes increases in intensity. We regard this work as a step towards the incorporation of academic-oriented verification tools into real-world systems of interest.
[ { "version": "v1", "created": "Tue, 6 Dec 2022 19:41:08 GMT" } ]
2022-12-08T00:00:00
[ [ "Amir", "Guy", "" ], [ "Freund", "Ziv", "" ], [ "Katz", "Guy", "" ], [ "Mandelbaum", "Elad", "" ], [ "Refaeli", "Idan", "" ] ]
new_dataset
0.999539
2212.03297
Justin Xie
Justin Xie
Fine-Grained Emotional Paraphrasing along Emotion Gradients
null
null
null
null
cs.CL cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Paraphrase generation, a.k.a. paraphrasing, is a common and important task in natural language processing. Emotional paraphrasing, which changes the emotion embodied in a piece of text while preserving its meaning, has many potential applications, e.g., moderating online dialogues and preventing cyberbullying. We introduce a new task of fine-grained emotional paraphrasing along emotion gradients, that is, altering the emotional intensities of the paraphrases in fine grain following smooth variations in affective dimensions while preserving the meanings of the originals. We propose a framework for addressing this task by fine-tuning text-to-text Transformers through multi-task training. We enhance several widely used paraphrasing corpus by annotating the input and target texts with their fine-grained emotion labels. With these labels, fine-tuning text-to-text Transformers on these corpus entails multi-task training. Evaluations of the fine-tuned Transformers on separate test sets show that including fine-grained emotion labels in the paraphrase task significantly improve the chance of obtaining high-quality paraphrases of the desired emotions, i.e., more than doubling the number of exact matches of desired emotions while achieving consistently better scores in paraphrase metrics such as BLEU, ROGUE, and METEOR.
[ { "version": "v1", "created": "Sun, 30 Oct 2022 05:38:22 GMT" } ]
2022-12-08T00:00:00
[ [ "Xie", "Justin", "" ] ]
new_dataset
0.996144
2212.03308
Yashar Salami
Yashar Salami, Vahid Khajehvand, Esmaeil Zeinali
E3C: A Tool for Evaluating Communication and Computation Costs in Authentication and Key Exchange Protocol
20 pages ,10 figures, 4 Table
null
null
null
cs.CR
http://creativecommons.org/licenses/by/4.0/
Today, with the development of blockchain and Internet of Things technologies, we need authentication protocols and key exchanges to communicate with these different technologies. Symmetric and asymmetric encryption methods are used to design authentication and key exchange protocols, each of which has different computation costs. In the Internet of Things systems, due to the limited memory and computation power, researchers are looking the lightweight design protocols so that the pressure caused by the computation of protocols can be minimized. Calculating protocols' computational and communication costs was done manually until now, which was associated with human error. In this paper, we proposed an E3C tool that can calculate the computation and communication costs of the authentication and key exchange protocols. E3C provides the ability to compare several protocols in terms of communication and processing costs and present them in separate charts. Comparing the processing and communication costs of classical and modern protocols manually and with the E3C indicate that the E3C can calculate the processing and communication costs of authentication and key exchange protocols with 99.99% accuracy.
[ { "version": "v1", "created": "Tue, 6 Dec 2022 20:14:26 GMT" } ]
2022-12-08T00:00:00
[ [ "Salami", "Yashar", "" ], [ "Khajehvand", "Vahid", "" ], [ "Zeinali", "Esmaeil", "" ] ]
new_dataset
0.99065
2212.03357
Yuan Yuan
Hao He, Yuan Yuan, Ying-Cong Chen, Peng Cao and Dina Katabi
Contactless Oxygen Monitoring with Gated Transformer
19 pages, Workshop on Learning from Time Series for Health, NeurIPS 2022
null
null
null
cs.LG eess.SP
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
With the increasing popularity of telehealth, it becomes critical to ensure that basic physiological signals can be monitored accurately at home, with minimal patient overhead. In this paper, we propose a contactless approach for monitoring patients' blood oxygen at home, simply by analyzing the radio signals in the room, without any wearable devices. We extract the patients' respiration from the radio signals that bounce off their bodies and devise a novel neural network that infers a patient's oxygen estimates from their breathing signal. Our model, called \emph{Gated BERT-UNet}, is designed to adapt to the patient's medical indices (e.g., gender, sleep stages). It has multiple predictive heads and selects the most suitable head via a gate controlled by the person's physiological indices. Extensive empirical results show that our model achieves high accuracy on both medical and radio datasets.
[ { "version": "v1", "created": "Tue, 6 Dec 2022 22:43:59 GMT" } ]
2022-12-08T00:00:00
[ [ "He", "Hao", "" ], [ "Yuan", "Yuan", "" ], [ "Chen", "Ying-Cong", "" ], [ "Cao", "Peng", "" ], [ "Katabi", "Dina", "" ] ]
new_dataset
0.950186
2212.03383
Zhongtang Luo
Zhongtang Luo, Rohan Murukutla, Aniket Kate
Last Mile of Blockchains: RPC and Node-as-a-service
null
null
null
null
cs.CR
http://creativecommons.org/licenses/by-nc-sa/4.0/
While much research focuses on different methods to secure blockchain, information on the chain needs to be accessed by end-users to be useful. This position paper surveys different ways that end-users may access blockchains. We observe that between the two extremes of running a full node and fully utilizing a trusted third-party service, many solutions regarding light nodes are emerging. We analyze these solutions based on three basic properties of web communication: integrity, availability and privacy. We conclude that currently, the best way to access a blockchain while maintaining these three properties is still to run a full node. We consider it essential that future blockchain accessibility services should be built while considering these three expectations.
[ { "version": "v1", "created": "Wed, 7 Dec 2022 00:31:46 GMT" } ]
2022-12-08T00:00:00
[ [ "Luo", "Zhongtang", "" ], [ "Murukutla", "Rohan", "" ], [ "Kate", "Aniket", "" ] ]
new_dataset
0.9546
2212.03419
Ruth-Ann Armstrong
Ruth-Ann Armstrong, John Hewitt and Christopher Manning
JamPatoisNLI: A Jamaican Patois Natural Language Inference Dataset
14 pages, 3 figures, Findings of EMNLP 2022
null
null
null
cs.CL cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
JamPatoisNLI provides the first dataset for natural language inference in a creole language, Jamaican Patois. Many of the most-spoken low-resource languages are creoles. These languages commonly have a lexicon derived from a major world language and a distinctive grammar reflecting the languages of the original speakers and the process of language birth by creolization. This gives them a distinctive place in exploring the effectiveness of transfer from large monolingual or multilingual pretrained models. While our work, along with previous work, shows that transfer from these models to low-resource languages that are unrelated to languages in their training set is not very effective, we would expect stronger results from transfer to creoles. Indeed, our experiments show considerably better results from few-shot learning of JamPatoisNLI than for such unrelated languages, and help us begin to understand how the unique relationship between creoles and their high-resource base languages affect cross-lingual transfer. JamPatoisNLI, which consists of naturally-occurring premises and expert-written hypotheses, is a step towards steering research into a traditionally underserved language and a useful benchmark for understanding cross-lingual NLP.
[ { "version": "v1", "created": "Wed, 7 Dec 2022 03:07:02 GMT" } ]
2022-12-08T00:00:00
[ [ "Armstrong", "Ruth-Ann", "" ], [ "Hewitt", "John", "" ], [ "Manning", "Christopher", "" ] ]
new_dataset
0.999837
2212.03420
S. Farokh Atashzar
Jacqueline Libby, Aniket A. Somwanshi, Federico Stancati, Gayatri Tyagi, Aadit Patel, Naigam Bhatt, JohnRoss Rizzo, S. Farokh Atashzar
What Happens When Pneu-Net Soft Robotic Actuators Get Fatigued?
null
null
null
null
cs.RO cs.SY eess.SY
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Soft actuators have attracted a great deal of interest in the context of rehabilitative and assistive robots for increasing safety and lowering costs as compared to rigid-body robotic systems. During actuation, soft actuators experience high levels of deformation, which can lead to microscale fractures in their elastomeric structure, which fatigues the system over time and eventually leads to macroscale damages and eventually failure. This paper reports finite element modeling (FEM) of pneu-nets at high angles, along with repetitive experimentation at high deformation rates, in order to study the effect and behavior of fatigue in soft robotic actuators, which would result in deviation from the ideal behavior. Comparing the FEM model and experimental data, we show that FEM can model the performance of the actuator before fatigue to a bending angle of 167 degrees with ~96% accuracy. We also show that the FEM model performance will drop to 80% due to fatigue after repetitive high-angle bending. The results of this paper objectively highlight the emergence of fatigue over cyclic activation of the system and the resulting deviation from the computational FEM model. Such behavior can be considered in future controllers to adapt the system with time-variable and non-autonomous response dynamics of soft robots.
[ { "version": "v1", "created": "Wed, 7 Dec 2022 03:07:33 GMT" } ]
2022-12-08T00:00:00
[ [ "Libby", "Jacqueline", "" ], [ "Somwanshi", "Aniket A.", "" ], [ "Stancati", "Federico", "" ], [ "Tyagi", "Gayatri", "" ], [ "Patel", "Aadit", "" ], [ "Bhatt", "Naigam", "" ], [ "Rizzo", "JohnRoss", "" ], [ "Atashzar", "S. Farokh", "" ] ]
new_dataset
0.99293
2212.03435
Fengyu Yang
Fengyu Yang, Jian Luan, Yujun Wang
Improve Bilingual TTS Using Dynamic Language and Phonology Embedding
Submitted to ICASSP2023
null
null
null
cs.SD cs.CL eess.AS
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In most cases, bilingual TTS needs to handle three types of input scripts: first language only, second language only, and second language embedded in the first language. In the latter two situations, the pronunciation and intonation of the second language are usually quite different due to the influence of the first language. Therefore, it is a big challenge to accurately model the pronunciation and intonation of the second language in different contexts without mutual interference. This paper builds a Mandarin-English TTS system to acquire more standard spoken English speech from a monolingual Chinese speaker. We introduce phonology embedding to capture the English differences between different phonology. Embedding mask is applied to language embedding for distinguishing information between different languages and to phonology embedding for focusing on English expression. We specially design an embedding strength modulator to capture the dynamic strength of language and phonology. Experiments show that our approach can produce significantly more natural and standard spoken English speech of the monolingual Chinese speaker. From analysis, we find that suitable phonology control contributes to better performance in different scenarios.
[ { "version": "v1", "created": "Wed, 7 Dec 2022 03:46:18 GMT" } ]
2022-12-08T00:00:00
[ [ "Yang", "Fengyu", "" ], [ "Luan", "Jian", "" ], [ "Wang", "Yujun", "" ] ]
new_dataset
0.951008
2212.03490
Yue Ma
Yue Ma, Tianyu Yang, Yin Shan, Xiu Li
SimVTP: Simple Video Text Pre-training with Masked Autoencoders
Github: https://github.com/mayuelala/SimVTP
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper presents SimVTP: a Simple Video-Text Pretraining framework via masked autoencoders. We randomly mask out the spatial-temporal tubes of input video and the word tokens of input text and then feed them into a unified autencoder to reconstruct the missing pixels and words. Our SimVTP has several properties: 1) Thanks to the unified autoencoder, SimVTP reconstructs the masked signal of one modality with the help from another modality, which implicitly learns the cross-modal alignment between video tubes and text tokens. 2) SimVTP not only benefits from a high video masking ratio (e.g. 90%) due to the temporal redundancy of video, but also needs a high text masking ratio (e.g. 75%), which is much higher than BERT (e.g. 15%), to achieve optimal performance. This is because the aid of video modality makes text reconstruction less challenging, which thus needs a higher mask ratio to make the pretext harder for useful feature learning. 3) Equipping SimVTP with video-text contrastive learning (VTC) and video-text matching (VTM), which are two commonly used cross-modal training strategies, could further improve the transferable performance significantly. 4) SimVTP is dataefficent, e.g., pre-training only on 10% data of WebVid-2M, SimVTP achieves surprisingly good results (43.8 R@1) on MSRVTT, which is far above recent state-of-the-art methods pre-trained on both CC3M and WebVid-2M. We transfer our pre-trained model to various downstream tasks and achieve superior performance. The codes and models will be released at https://github.com/mayuelala/SimVTP.
[ { "version": "v1", "created": "Wed, 7 Dec 2022 07:14:22 GMT" } ]
2022-12-08T00:00:00
[ [ "Ma", "Yue", "" ], [ "Yang", "Tianyu", "" ], [ "Shan", "Yin", "" ], [ "Li", "Xiu", "" ] ]
new_dataset
0.99845
2212.03517
Siwei Yang
Siwei Yang, Longlong Jing, Junfei Xiao, Hang Zhao, Alan Yuille, Yingwei Li
AsyInst: Asymmetric Affinity with DepthGrad and Color for Box-Supervised Instance Segmentation
null
null
null
null
cs.CV
http://creativecommons.org/licenses/by-nc-nd/4.0/
The weakly supervised instance segmentation is a challenging task. The existing methods typically use bounding boxes as supervision and optimize the network with a regularization loss term such as pairwise color affinity loss for instance segmentation. Through systematic analysis, we found that the commonly used pairwise affinity loss has two limitations: (1) it works with color affinity but leads to inferior performance with other modalities such as depth gradient, (2)the original affinity loss does not prevent trivial predictions as intended but actually accelerates this process due to the affinity loss term being symmetric. To overcome these two limitations, in this paper, we propose a novel asymmetric affinity loss which provides the penalty against the trivial prediction and generalizes well with affinity loss from different modalities. With the proposed asymmetric affinity loss, our method outperforms the state-of-the-art methods on the Cityscapes dataset and outperforms our baseline method by 3.5% in mask AP.
[ { "version": "v1", "created": "Wed, 7 Dec 2022 08:47:10 GMT" } ]
2022-12-08T00:00:00
[ [ "Yang", "Siwei", "" ], [ "Jing", "Longlong", "" ], [ "Xiao", "Junfei", "" ], [ "Zhao", "Hang", "" ], [ "Yuille", "Alan", "" ], [ "Li", "Yingwei", "" ] ]
new_dataset
0.95765
2212.03520
Mordechai Guri
Mordechai Guri
COVID-bit: Keep a Distance of (at least) 2m From My Air-Gap Computer!
This is an significantly extended version of a shorter paper accepted to IEEE TrustCom 2022
null
null
null
cs.CR
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Air-gapped systems are isolated from the Internet due to the sensitive information they handle. This paper presents COVID-bit, a new COVert channel attack that leaks sensitive information over the air from highly isolated systems. The information emanates from the air-gapped computer over the air to a distance of 2m and more and can be picked up by a nearby insider or spy with a mobile phone or laptop. Malware on an air-gapped computer can generate radio waves by executing crafted code on the target system. The malicious code exploits the dynamic power consumption of modern computers and manipulates the momentary loads on CPU cores. This technique allows the malware to control the computer's internal utilization and generate low-frequency electromagnetic radiation in the 0 - 60 kHz band. Sensitive information (e.g., files, encryption keys, biometric data, and keylogging) can be modulated over the emanated signals and received by a nearby mobile phone at a max speed of 1000 bits/sec. We show that a smartphone or laptop with a small \$1 antenna carried by a malicious insider or visitor can be used as a covert receiver. Notably, the attack is highly evasive since it executes from an ordinary user-level process, does not require root privileges, and is effective even within a Virtual Machine (VM). We discuss the attack model and provide technical details. We implement air-gap transmission of texts and files, and present signal generation and data modulation. We test the covert channel and show evaluation results. Finally, we present a set of countermeasures to this air-gap attack.
[ { "version": "v1", "created": "Wed, 7 Dec 2022 08:57:40 GMT" } ]
2022-12-08T00:00:00
[ [ "Guri", "Mordechai", "" ] ]
new_dataset
0.999737
2212.03641
Michele Campobasso
Michele Campobasso, Luca Allodi (Eindhoven University of Technology)
THREAT/crawl: a Trainable, Highly-Reusable, and Extensible Automated Method and Tool to Crawl Criminal Underground Forums
To be published in the Proceedings of the 17th Symposium on Electronic Crime Research (APWG eCrime 2022). Source code of the implemented solution available at https://gitlab.tue.nl/threat-crawl/THREATcrawl/
null
null
null
cs.IR cs.CR
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Collecting data on underground criminal communities is highly valuable both for security research and security operations. Unfortunately these communities live within a constellation of diverse online forums that are difficult to infiltrate, may adopt crawling monitoring countermeasures, and require the development of ad-hoc scrapers for each different community, making the endeavour increasingly technically challenging, and potentially expensive. To address this problem we propose THREAT/crawl, a method and prototype tool for a highly reusable crawler that can learn a wide range of (arbitrary) forum structures, can remain under-the-radar during the crawling activity and can be extended and configured at the user will. We showcase THREAT/crawl capabilities and provide prime evaluation of our prototype against a range of active, live, underground communities.
[ { "version": "v1", "created": "Wed, 7 Dec 2022 13:54:51 GMT" } ]
2022-12-08T00:00:00
[ [ "Campobasso", "Michele", "", "Eindhoven University of Technology" ], [ "Allodi", "Luca", "", "Eindhoven University of Technology" ] ]
new_dataset
0.988116
2212.03810
Kristina Lerman
Kristina Lerman
The Social Emotional Web
The 8th IEEE International Conference on Collaboration and Internet Computing (IEEE CIC 2022)
null
null
null
cs.CY
http://creativecommons.org/licenses/by/4.0/
The social web has linked people on a global scale, transforming how we communicate and interact. The massive interconnectedness has created new vulnerabilities in the form of social manipulation and misinformation. As the social web matures, we are entering a new phase, where people share their private feelings and emotions. This so-called social emotional web creates new opportunities for human flourishing, but also exposes new vulnerabilities. To reap the benefits of the social emotional web, and reduce potential harms, we must anticipate how it will evolve and create policies that minimize risks.
[ { "version": "v1", "created": "Wed, 7 Dec 2022 17:46:22 GMT" } ]
2022-12-08T00:00:00
[ [ "Lerman", "Kristina", "" ] ]
new_dataset
0.997627
1910.06078
Fangli Xu
Fangli Xu, Lingfei Wu, KP Thai, Carol Hsu, Wei Wang, Richard Tong
MUTLA: A Large-Scale Dataset for Multimodal Teaching and Learning Analytics
3 pages, 1 figure, 2 tables workshop paper
null
null
null
cs.CY stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Automatic analysis of teacher and student interactions could be very important to improve the quality of teaching and student engagement. However, despite some recent progress in utilizing multimodal data for teaching and learning analytics, a thorough analysis of a rich multimodal dataset coming for a complex real learning environment has yet to be done. To bridge this gap, we present a large-scale MUlti-modal Teaching and Learning Analytics (MUTLA) dataset. This dataset includes time-synchronized multimodal data records of students (learning logs, videos, EEG brainwaves) as they work in various subjects from Squirrel AI Learning System (SAIL) to solve problems of varying difficulty levels. The dataset resources include user records from the learner records store of SAIL, brainwave data collected by EEG headset devices, and video data captured by web cameras while students worked in the SAIL products. Our hope is that by analyzing real-world student learning activities, facial expressions, and brainwave patterns, researchers can better predict engagement, which can then be used to improve adaptive learning selection and student learning outcomes. An additional goal is to provide a dataset gathered from real-world educational activities versus those from controlled lab environments to benefit the educational learning community.
[ { "version": "v1", "created": "Sat, 5 Oct 2019 03:53:49 GMT" }, { "version": "v2", "created": "Tue, 6 Dec 2022 18:21:33 GMT" } ]
2022-12-07T00:00:00
[ [ "Xu", "Fangli", "" ], [ "Wu", "Lingfei", "" ], [ "Thai", "KP", "" ], [ "Hsu", "Carol", "" ], [ "Wang", "Wei", "" ], [ "Tong", "Richard", "" ] ]
new_dataset
0.999799
2109.02580
Wenxi Liu
Wenxi Liu, Qi Li, Xindai Lin, Weixiang Yang, Shengfeng He, Yuanlong Yu
Ultra-high Resolution Image Segmentation via Locality-aware Context Fusion and Alternating Local Enhancement
Extension of ICCV 2021 "From Contexts to Locality: Ultra-high Resolution Image Segmentation via Locality-aware Contextual Correlation"
null
null
null
cs.CV
http://creativecommons.org/licenses/by/4.0/
Ultra-high resolution image segmentation has raised increasing interests in recent years due to its realistic applications. In this paper, we innovate the widely used high-resolution image segmentation pipeline, in which an ultra-high resolution image is partitioned into regular patches for local segmentation and then the local results are merged into a high-resolution semantic mask. In particular, we introduce a novel locality-aware context fusion based segmentation model to process local patches, where the relevance between local patch and its various contexts are jointly and complementarily utilized to handle the semantic regions with large variations. Additionally, we present the alternating local enhancement module that restricts the negative impact of redundant information introduced from the contexts, and thus is endowed with the ability of fixing the locality-aware features to produce refined results. Furthermore, in comprehensive experiments, we demonstrate that our model outperforms other state-of-the-art methods in public benchmarks. Our released codes are available at: https://github.com/liqiokkk/FCtL.
[ { "version": "v1", "created": "Mon, 6 Sep 2021 16:26:05 GMT" }, { "version": "v2", "created": "Tue, 15 Nov 2022 14:13:49 GMT" }, { "version": "v3", "created": "Tue, 6 Dec 2022 09:44:06 GMT" } ]
2022-12-07T00:00:00
[ [ "Liu", "Wenxi", "" ], [ "Li", "Qi", "" ], [ "Lin", "Xindai", "" ], [ "Yang", "Weixiang", "" ], [ "He", "Shengfeng", "" ], [ "Yu", "Yuanlong", "" ] ]
new_dataset
0.997818
2203.11015
Xianghao Zhan
Xianghao Zhan, Fanjin Wang, Olivier Gevaert
Filter Drug-induced Liver Injury Literature with Natural Language Processing and Ensemble Learning
8 pages, 4 figures
null
10.1109/JBHI.2022.3193365
null
cs.IR cs.LG stat.AP
http://creativecommons.org/licenses/by/4.0/
Drug-induced liver injury (DILI) describes the adverse effects of drugs that damage liver. Life-threatening results including liver failure or death were also reported in severe DILI cases. Therefore, DILI-related events are strictly monitored for all approved drugs and the liver toxicity became important assessments for new drug candidates. These DILI-related reports are documented in hospital records, in clinical trial results, and also in research papers that contain preliminary in vitro and in vivo experiments. Conventionally, data extraction from previous publications relies heavily on resource-demanding manual labelling, which considerably decreased the efficiency of the information extraction process. The recent development of artificial intelligence, particularly, the rise of natural language processing (NLP) techniques, enabled the automatic processing of biomedical texts. In this study, based on around 28,000 papers (titles and abstracts) provided by the Critical Assessment of Massive Data Analysis (CAMDA) challenge, we benchmarked model performances on filtering out DILI literature. Among four word vectorization techniques, the model using term frequency-inverse document frequency (TF-IDF) and logistic regression outperformed others with an accuracy of 0.957 with our in-house test set. Furthermore, an ensemble model with similar overall performances was implemented and was fine-tuned to lower the false-negative cases to avoid neglecting potential DILI reports. The ensemble model achieved a high accuracy of 0.954 and an F1 score of 0.955 in the hold-out validation data provided by the CAMDA committee. Moreover, important words in positive/negative predictions were identified via model interpretation. Overall, the ensemble model reached satisfactory classification results, which can be further used by researchers to rapidly filter DILI-related literature.
[ { "version": "v1", "created": "Wed, 9 Mar 2022 23:53:07 GMT" } ]
2022-12-07T00:00:00
[ [ "Zhan", "Xianghao", "" ], [ "Wang", "Fanjin", "" ], [ "Gevaert", "Olivier", "" ] ]
new_dataset
0.966045
2204.07874
Markus Borg
Markus Borg, Jens Henriksson, Kasper Socha, Olof Lennartsson, Elias Sonnsj\"o L\"onegren, Thanh Bui, Piotr Tomaszewski, Sankar Raman Sathyamoorthy, Sebastian Brink, Mahshid Helali Moghadam
Ergo, SMIRK is Safe: A Safety Case for a Machine Learning Component in a Pedestrian Automatic Emergency Brake System
Accepted for publication in Software Quality Journal
null
null
null
cs.SE cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Integration of Machine Learning (ML) components in critical applications introduces novel challenges for software certification and verification. New safety standards and technical guidelines are under development to support the safety of ML-based systems, e.g., ISO 21448 SOTIF for the automotive domain and the Assurance of Machine Learning for use in Autonomous Systems (AMLAS) framework. SOTIF and AMLAS provide high-level guidance but the details must be chiseled out for each specific case. We initiated a research project with the goal to demonstrate a complete safety case for an ML component in an open automotive system. This paper reports results from an industry-academia collaboration on safety assurance of SMIRK, an ML-based pedestrian automatic emergency braking demonstrator running in an industry-grade simulator. We demonstrate an application of AMLAS on SMIRK for a minimalistic operational design domain, i.e., we share a complete safety case for its integrated ML-based component. Finally, we report lessons learned and provide both SMIRK and the safety case under an open-source licence for the research community to reuse.
[ { "version": "v1", "created": "Sat, 16 Apr 2022 21:28:50 GMT" }, { "version": "v2", "created": "Thu, 15 Sep 2022 12:43:05 GMT" }, { "version": "v3", "created": "Tue, 6 Dec 2022 10:49:12 GMT" } ]
2022-12-07T00:00:00
[ [ "Borg", "Markus", "" ], [ "Henriksson", "Jens", "" ], [ "Socha", "Kasper", "" ], [ "Lennartsson", "Olof", "" ], [ "Lönegren", "Elias Sonnsjö", "" ], [ "Bui", "Thanh", "" ], [ "Tomaszewski", "Piotr", "" ], [ "Sathyamoorthy", "Sankar Raman", "" ], [ "Brink", "Sebastian", "" ], [ "Moghadam", "Mahshid Helali", "" ] ]
new_dataset
0.99461
2205.00180
Mifta Sintaha
Mifta Sintaha, Noor Nashid, Ali Mesbah
Katana: Dual Slicing-Based Context for Learning Bug Fixes
null
null
null
null
cs.SE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Contextual information plays a vital role for software developers when understanding and fixing a bug. Consequently, deep learning-based program repair techniques leverage context for bug fixes. However, existing techniques treat context in an arbitrary manner, by extracting code in close proximity of the buggy statement within the enclosing file, class, or method, without any analysis to find actual relations with the bug. To reduce noise, they use a predefined maximum limit on the number of tokens to be used as context. We present a program slicing-based approach, in which instead of arbitrarily including code as context, we analyze statements that have a control or data dependency on the buggy statement. We propose a novel concept called dual slicing, which leverages the context of both buggy and fixed versions of the code to capture relevant repair ingredients. We present our technique and tool called Katana, the first to apply slicing-based context for a program repair task. The results show Katana effectively preserves sufficient information for a model to choose contextual information while reducing noise. We compare against four recent state-of-the-art context-aware program repair techniques. Our results show Katana fixes between 1.5 to 3.7 times more bugs than existing techniques.
[ { "version": "v1", "created": "Sat, 30 Apr 2022 07:04:41 GMT" }, { "version": "v2", "created": "Thu, 2 Jun 2022 20:43:53 GMT" }, { "version": "v3", "created": "Mon, 5 Dec 2022 22:09:36 GMT" } ]
2022-12-07T00:00:00
[ [ "Sintaha", "Mifta", "" ], [ "Nashid", "Noor", "" ], [ "Mesbah", "Ali", "" ] ]
new_dataset
0.997266
2205.00222
Randy Harsuko
Randy Harsuko and Tariq Alkhalifah
StorSeismic: A new paradigm in deep learning for seismic processing
18 pages, 18 figures
null
10.1109/TGRS.2022.3216660
null
cs.LG eess.SP physics.comp-ph physics.geo-ph
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Machine learned tasks on seismic data are often trained sequentially and separately, even though they utilize the same features (i.e. geometrical) of the data. We present StorSeismic, as a framework for seismic data processing, which consists of neural network pre-training and fine-tuning procedures. We, specifically, utilize a neural network as a preprocessing model to store seismic data features of a particular dataset for any downstream tasks. After pre-training, the resulting model can be utilized later, through a fine-tuning procedure, to perform tasks using limited additional training. Used often in Natural Language Processing (NLP) and lately in vision tasks, BERT (Bidirectional Encoder Representations from Transformer), a form of a Transformer model, provides an optimal platform for this framework. The attention mechanism of BERT, applied here on a sequence of traces within the shot gather, is able to capture and store key geometrical features of the seismic data. We pre-train StorSeismic on field data, along with synthetically generated ones, in the self-supervised step. Then, we use the labeled synthetic data to fine-tune the pre-trained network in a supervised fashion to perform various seismic processing tasks, like denoising, velocity estimation, first arrival picking, and NMO. Finally, the fine-tuned model is used to obtain satisfactory inference results on the field data.
[ { "version": "v1", "created": "Sat, 30 Apr 2022 09:55:00 GMT" } ]
2022-12-07T00:00:00
[ [ "Harsuko", "Randy", "" ], [ "Alkhalifah", "Tariq", "" ] ]
new_dataset
0.999001
2205.07979
Boro Sitnikovski
Boro Sitnikovski
Budge: a programming language and a theorem prover
null
null
null
null
cs.PL cs.CL cs.LO
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We present a simple programming language based on G\"odel numbering and prime factorization, enhanced with explicit, scoped loops, allowing for easy program composition. Further, we will present a theorem prover that allows expressing and working with formal systems. The theorem prover is simple as it relies merely on a substitution rule and set equality to derive theorems. Finally, we will represent the programming language in the theorem prover. We will show the syntax and semantics of both, and then provide a few example programs and their evaluation.
[ { "version": "v1", "created": "Mon, 16 May 2022 20:35:25 GMT" }, { "version": "v2", "created": "Thu, 19 May 2022 12:25:22 GMT" }, { "version": "v3", "created": "Tue, 24 May 2022 13:15:24 GMT" }, { "version": "v4", "created": "Thu, 4 Aug 2022 11:47:22 GMT" }, { "version": "v5", "created": "Tue, 23 Aug 2022 23:23:12 GMT" }, { "version": "v6", "created": "Tue, 6 Dec 2022 11:48:21 GMT" } ]
2022-12-07T00:00:00
[ [ "Sitnikovski", "Boro", "" ] ]
new_dataset
0.999798
2205.08491
Veli Safak
Veli Safak and Aniish Sridhar
Elon Musk's Twitter Takeover: Politician Accounts
null
null
null
null
cs.SI
http://creativecommons.org/licenses/by/4.0/
We provided quantitative data supporting significant changes between the time Twitter acceptance the offer on April 25 and the time the agreement was finalized on October 27. Republican politicians saw significant increases in their follower counts, while Democrat politicians saw significant decreases.
[ { "version": "v1", "created": "Mon, 16 May 2022 14:11:49 GMT" }, { "version": "v2", "created": "Sun, 4 Dec 2022 13:36:05 GMT" }, { "version": "v3", "created": "Tue, 6 Dec 2022 02:57:59 GMT" } ]
2022-12-07T00:00:00
[ [ "Safak", "Veli", "" ], [ "Sridhar", "Aniish", "" ] ]
new_dataset
0.966107
2205.13707
Shucheng Yang
Shucheng Yang, Xiaoping Gao, Ruoting Yang, Jie Ren, and Zhen Wang
A Hybrid Josephson Transmission Line and Passive Transmission Line Routing Framework for Single Flux Quantum Logic
null
null
10.1109/TASC.2022.3206280
null
cs.ET cs.AR
http://creativecommons.org/licenses/by/4.0/
The Single Flux Quantum (SFQ) logic family is a novel digital logic as it provides ultra-fast and energy-efficient circuits. For large-scale SFQ circuit design, specialized electronic design automation (EDA) tools are required due to the differences in logic type, timing constraints and circuit architecture, in contrast to the CMOS logic. In order to improve the overall performance of an SFQ circuit, an efficient routing algorithm should be applied during the layout design to perform accurate timing adjustment for fixing hold violations and optimizing critical paths. Thus, a hybrid Josephson transmission line and passive transmission line routing framework is proposed. It consists of four main modules and an exploration of the potential timing performance based on the given layout placement. The proposed routing tool is demonstrated on seven testbench circuits. The obtained results demonstrate that the operating frequency is greatly improved, and all the hold violations are eliminated for each circuit.
[ { "version": "v1", "created": "Fri, 27 May 2022 01:51:51 GMT" } ]
2022-12-07T00:00:00
[ [ "Yang", "Shucheng", "" ], [ "Gao", "Xiaoping", "" ], [ "Yang", "Ruoting", "" ], [ "Ren", "Jie", "" ], [ "Wang", "Zhen", "" ] ]
new_dataset
0.990859
2208.02313
Monika Kwiatkowski
Dominik Kuhnke, Monika Kwiatkowski, Olaf Hellwich
Image-based Detection of Surface Defects in Concrete during Construction
null
null
null
null
cs.CV cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Defects increase the cost and duration of construction projects as they require significant inspection and documentation efforts. Automating defect detection could significantly reduce these efforts. This work focuses on detecting honeycombs, a substantial defect in concrete structures that may affect structural integrity. We compared honeycomb images scraped from the web with images obtained from real construction inspections. We found that web images do not capture the complete variance found in real-case scenarios and that there is still a lack of data in this domain. Our dataset is therefore freely available for further research. A Mask R-CNN and EfficientNet-B0 were trained for honeycomb detection. The Mask R-CNN model allows detecting honeycombs based on instance segmentation, whereas the EfficientNet-B0 model allows a patch-based classification. Our experiments demonstrate that both approaches are suitable for solving and automating honeycomb detection. In the future, this solution can be incorporated into defect documentation systems.
[ { "version": "v1", "created": "Wed, 3 Aug 2022 19:05:12 GMT" }, { "version": "v2", "created": "Tue, 6 Dec 2022 15:19:33 GMT" } ]
2022-12-07T00:00:00
[ [ "Kuhnke", "Dominik", "" ], [ "Kwiatkowski", "Monika", "" ], [ "Hellwich", "Olaf", "" ] ]
new_dataset
0.999687
2209.11518
Tuan-Anh Vu
Quang-Trung Truong and Tuan-Anh Vu and Tan-Sang Ha and Lokoc Jakub and Yue Him Wong Tim and Ajay Joneja and Sai-Kit Yeung
Marine Video Kit: A New Marine Video Dataset for Content-based Analysis and Retrieval
Camera Ready for MMM 2023, Bergen, Norway
null
null
null
cs.CV cs.IR cs.MM
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Effective analysis of unusual domain specific video collections represents an important practical problem, where state-of-the-art general purpose models still face limitations. Hence, it is desirable to design benchmark datasets that challenge novel powerful models for specific domains with additional constraints. It is important to remember that domain specific data may be noisier (e.g., endoscopic or underwater videos) and often require more experienced users for effective search. In this paper, we focus on single-shot videos taken from moving cameras in underwater environments, which constitute a nontrivial challenge for research purposes. The first shard of a new Marine Video Kit dataset is presented to serve for video retrieval and other computer vision challenges. Our dataset is used in a special session during Video Browser Showdown 2023. In addition to basic meta-data statistics, we present several insights based on low-level features as well as semantic annotations of selected keyframes. The analysis also contains experiments showing limitations of respected general purpose models for retrieval. Our dataset and code are publicly available at https://hkust-vgd.github.io/marinevideokit.
[ { "version": "v1", "created": "Fri, 23 Sep 2022 10:57:50 GMT" }, { "version": "v2", "created": "Fri, 21 Oct 2022 06:53:46 GMT" }, { "version": "v3", "created": "Sat, 3 Dec 2022 15:03:32 GMT" }, { "version": "v4", "created": "Tue, 6 Dec 2022 05:29:30 GMT" } ]
2022-12-07T00:00:00
[ [ "Truong", "Quang-Trung", "" ], [ "Vu", "Tuan-Anh", "" ], [ "Ha", "Tan-Sang", "" ], [ "Jakub", "Lokoc", "" ], [ "Tim", "Yue Him Wong", "" ], [ "Joneja", "Ajay", "" ], [ "Yeung", "Sai-Kit", "" ] ]
new_dataset
0.999697
2210.07128
Aman Madaan
Aman Madaan, Shuyan Zhou, Uri Alon, Yiming Yang, Graham Neubig
Language Models of Code are Few-Shot Commonsense Learners
EMNLP 2022
null
null
null
cs.CL cs.LG
http://creativecommons.org/licenses/by/4.0/
We address the general task of structured commonsense reasoning: given a natural language input, the goal is to generate a graph such as an event -- or a reasoning-graph. To employ large language models (LMs) for this task, existing approaches ``serialize'' the output graph as a flat list of nodes and edges. Although feasible, these serialized graphs strongly deviate from the natural language corpora that LMs were pre-trained on, hindering LMs from generating them correctly. In this paper, we show that when we instead frame structured commonsense reasoning tasks as code generation tasks, pre-trained LMs of code are better structured commonsense reasoners than LMs of natural language, even when the downstream task does not involve source code at all. We demonstrate our approach across three diverse structured commonsense reasoning tasks. In all these natural language tasks, we show that using our approach, a code generation LM (CODEX) outperforms natural-LMs that are fine-tuned on the target task (e.g., T5) and other strong LMs such as GPT-3 in the few-shot setting.
[ { "version": "v1", "created": "Thu, 13 Oct 2022 16:09:36 GMT" }, { "version": "v2", "created": "Mon, 24 Oct 2022 05:29:59 GMT" }, { "version": "v3", "created": "Tue, 6 Dec 2022 15:58:30 GMT" } ]
2022-12-07T00:00:00
[ [ "Madaan", "Aman", "" ], [ "Zhou", "Shuyan", "" ], [ "Alon", "Uri", "" ], [ "Yang", "Yiming", "" ], [ "Neubig", "Graham", "" ] ]
new_dataset
0.953699
2211.00111
Sangdon Park
Sangdon Park and Xiang Cheng and Taesoo Kim
Unsafe's Betrayal: Abusing Unsafe Rust in Binary Reverse Engineering via Machine Learning
null
null
null
null
cs.CR cs.LG cs.SE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Memory-safety bugs introduce critical software-security issues. Rust provides memory-safe mechanisms to avoid memory-safety bugs in programming, while still allowing unsafe escape hatches via unsafe code. However, the unsafe code that enhances the usability of Rust provides clear spots for finding memory-safety bugs in Rust source code. In this paper, we claim that these unsafe spots can still be identifiable in Rust binary code via machine learning and be leveraged for finding memory-safety bugs. To support our claim, we propose the tool textttrustspot, that enables reverse engineering to learn an unsafe classifier that proposes a list of functions in Rust binaries for downstream analysis. We empirically show that the function proposals by textttrustspot can recall $92.92\%$ of memory-safety bugs, while it covers only $16.79\%$ of the entire binary code. As an application, we demonstrate that the function proposals are used in targeted fuzzing on Rust packages, which contribute to reducing the fuzzing time compared to non-targeted fuzzing.
[ { "version": "v1", "created": "Mon, 31 Oct 2022 19:32:18 GMT" }, { "version": "v2", "created": "Tue, 6 Dec 2022 05:50:30 GMT" } ]
2022-12-07T00:00:00
[ [ "Park", "Sangdon", "" ], [ "Cheng", "Xiang", "" ], [ "Kim", "Taesoo", "" ] ]
new_dataset
0.997736
2211.15350
Yanhui Zhang
Yanhui Zhang
Three classes of BCH codes and their duals
25 pages
null
null
null
cs.IT math.IT
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
BCH codes are an important class of cyclic codes, and have wide applicantions in communication and storage systems. However, it is difficult to determine the parameters of BCH codes and only a few cases are known. In this paper, we mainly study three classes of BCH codes with $n=q^{m}-1,\frac{q^{2s}-1}{q+1},\frac{q^{m}-1}{q-1}$. On one hand, we accurately give the parameters of $\mathcal C_{(q,n,\delta,1)}$ and its dual codes. On the other hand, we give the sufficient and necessary conditions for $\mathcal C_{(q,n,\delta,2)}$ being dually-BCH codes.
[ { "version": "v1", "created": "Mon, 28 Nov 2022 14:31:51 GMT" }, { "version": "v2", "created": "Tue, 6 Dec 2022 03:37:40 GMT" } ]
2022-12-07T00:00:00
[ [ "Zhang", "Yanhui", "" ] ]
new_dataset
0.955124
2212.01546
Yi Lei
Yi Lei, Shan Yang, Xinsheng Wang, Qicong Xie, Jixun Yao, Lei Xie, Dan Su
UniSyn: An End-to-End Unified Model for Text-to-Speech and Singing Voice Synthesis
null
null
null
null
cs.SD eess.AS
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Text-to-speech (TTS) and singing voice synthesis (SVS) aim at generating high-quality speaking and singing voice according to textual input and music scores, respectively. Unifying TTS and SVS into a single system is crucial to the applications requiring both of them. Existing methods usually suffer from some limitations, which rely on either both singing and speaking data from the same person or cascaded models of multiple tasks. To address these problems, a simplified elegant framework for TTS and SVS, named UniSyn, is proposed in this paper. It is an end-to-end unified model that can make a voice speak and sing with only singing or speaking data from this person. To be specific, a multi-conditional variational autoencoder (MC-VAE), which constructs two independent latent sub-spaces with the speaker- and style-related (i.e. speak or sing) conditions for flexible control, is proposed in UniSyn. Moreover, supervised guided-VAE and timbre perturbation with the Wasserstein distance constraint are leveraged to further disentangle the speaker timbre and style. Experiments conducted on two speakers and two singers demonstrate that UniSyn can generate natural speaking and singing voice without corresponding training data. The proposed approach outperforms the state-of-the-art end-to-end voice generation work, which proves the effectiveness and advantages of UniSyn.
[ { "version": "v1", "created": "Sat, 3 Dec 2022 05:58:10 GMT" }, { "version": "v2", "created": "Tue, 6 Dec 2022 11:28:30 GMT" } ]
2022-12-07T00:00:00
[ [ "Lei", "Yi", "" ], [ "Yang", "Shan", "" ], [ "Wang", "Xinsheng", "" ], [ "Xie", "Qicong", "" ], [ "Yao", "Jixun", "" ], [ "Xie", "Lei", "" ], [ "Su", "Dan", "" ] ]
new_dataset
0.994518
2212.02375
Hankyu Jang
Hankyu Jang, Daeyoung Kim
D-TensoRF: Tensorial Radiance Fields for Dynamic Scenes
21 pages, 11 figures
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Neural radiance field (NeRF) attracts attention as a promising approach to reconstructing the 3D scene. As NeRF emerges, subsequent studies have been conducted to model dynamic scenes, which include motions or topological changes. However, most of them use an additional deformation network, slowing down the training and rendering speed. Tensorial radiance field (TensoRF) recently shows its potential for fast, high-quality reconstruction of static scenes with compact model size. In this paper, we present D-TensoRF, a tensorial radiance field for dynamic scenes, enabling novel view synthesis at a specific time. We consider the radiance field of a dynamic scene as a 5D tensor. The 5D tensor represents a 4D grid in which each axis corresponds to X, Y, Z, and time and has 1D multi-channel features per element. Similar to TensoRF, we decompose the grid either into rank-one vector components (CP decomposition) or low-rank matrix components (newly proposed MM decomposition). We also use smoothing regularization to reflect the relationship between features at different times (temporal dependency). We conduct extensive evaluations to analyze our models. We show that D-TensoRF with CP decomposition and MM decomposition both have short training times and significantly low memory footprints with quantitatively and qualitatively competitive rendering results in comparison to the state-of-the-art methods in 3D dynamic scene modeling.
[ { "version": "v1", "created": "Mon, 5 Dec 2022 15:57:55 GMT" }, { "version": "v2", "created": "Tue, 6 Dec 2022 04:15:10 GMT" } ]
2022-12-07T00:00:00
[ [ "Jang", "Hankyu", "" ], [ "Kim", "Daeyoung", "" ] ]
new_dataset
0.992037
2212.02564
David Pomerenke
David Pomerenke
INCLUSIFY: A benchmark and a model for gender-inclusive German
null
null
null
null
cs.CL
http://creativecommons.org/licenses/by/4.0/
Gender-inclusive language is important for achieving gender equality in languages with gender inflections, such as German. While stirring some controversy, it is increasingly adopted by companies and political institutions. A handful of tools have been developed to help people use gender-inclusive language by identifying instances of the generic masculine and providing suggestions for more inclusive reformulations. In this report, we define the underlying tasks in terms of natural language processing, and present a dataset and measures for benchmarking them. We also present a model that implements these tasks, by combining an inclusive language database with an elaborate sequence of processing steps via standard pre-trained models. Our model achieves a recall of 0.89 and a precision of 0.82 in our benchmark for identifying exclusive language; and one of its top five suggestions is chosen in real-world texts in 44% of cases. We sketch how the area could be further advanced by training end-to-end models and using large language models; and we urge the community to include more gender-inclusive texts in their training data in order to not present an obstacle to the adoption of gender-inclusive language. Through these efforts, we hope to contribute to restoring justice in language and, to a small extent, in reality.
[ { "version": "v1", "created": "Mon, 5 Dec 2022 19:37:48 GMT" } ]
2022-12-07T00:00:00
[ [ "Pomerenke", "David", "" ] ]
new_dataset
0.999853
2212.02738
Weigang Lv
Weigang Lv, Jiale Bai, Qingli Yan, Hui-Ming Wang
RIS-Assisted Green Secure Communications: Active RIS or Passive RIS?
null
null
null
null
cs.IT math.IT
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Reconfigurable Intelligent Surface (RIS) is one of the promising techniques for 6G wireless communications, and recently has also been shown to be able to improve secure communications. However, there is a "double fading" effect in the reflection link between base station and user, thus passive RIS only achieves a negligible secrecy gain in typical communications scenarios.In this letter, we propose an active RIS-aided multi-antenna physical layer secrecy transmission scheme, where the active RIS can amplify the signal actively. Our aim is to minimize the transmit power subject to the constraint of secrecy rate. To solve the non-convex optimization problem, a penalty-based alternating minimization (AltMin) algorithm is proposed to optimize both the beamformer at the transmitter and the reflection matrix at RIS. Simulation results show that active RIS can resist the impact of "double fading" effect effectively, and is more energy efficient than passive RIS.
[ { "version": "v1", "created": "Tue, 6 Dec 2022 04:09:00 GMT" } ]
2022-12-07T00:00:00
[ [ "Lv", "Weigang", "" ], [ "Bai", "Jiale", "" ], [ "Yan", "Qingli", "" ], [ "Wang", "Hui-Ming", "" ] ]
new_dataset
0.95864
2212.02746
Jiaqi Chen
Jiaqi Chen, Tong Li, Jinghui Qin, Pan Lu, Liang Lin, Chongyu Chen, Xiaodan Liang
UniGeo: Unifying Geometry Logical Reasoning via Reformulating Mathematical Expression
null
null
null
null
cs.AI cs.LG
http://creativecommons.org/licenses/by-nc-sa/4.0/
Geometry problem solving is a well-recognized testbed for evaluating the high-level multi-modal reasoning capability of deep models. In most existing works, two main geometry problems: calculation and proving, are usually treated as two specific tasks, hindering a deep model to unify its reasoning capability on multiple math tasks. However, in essence, these two tasks have similar problem representations and overlapped math knowledge which can improve the understanding and reasoning ability of a deep model on both two tasks. Therefore, we construct a large-scale Unified Geometry problem benchmark, UniGeo, which contains 4,998 calculation problems and 9,543 proving problems. Each proving problem is annotated with a multi-step proof with reasons and mathematical expressions. The proof can be easily reformulated as a proving sequence that shares the same formats with the annotated program sequence for calculation problems. Naturally, we also present a unified multi-task Geometric Transformer framework, Geoformer, to tackle calculation and proving problems simultaneously in the form of sequence generation, which finally shows the reasoning ability can be improved on both two tasks by unifying formulation. Furthermore, we propose a Mathematical Expression Pretraining (MEP) method that aims to predict the mathematical expressions in the problem solution, thus improving the Geoformer model. Experiments on the UniGeo demonstrate that our proposed Geoformer obtains state-of-the-art performance by outperforming task-specific model NGS with over 5.6% and 3.2% accuracies on calculation and proving problems, respectively.
[ { "version": "v1", "created": "Tue, 6 Dec 2022 04:37:51 GMT" } ]
2022-12-07T00:00:00
[ [ "Chen", "Jiaqi", "" ], [ "Li", "Tong", "" ], [ "Qin", "Jinghui", "" ], [ "Lu", "Pan", "" ], [ "Lin", "Liang", "" ], [ "Chen", "Chongyu", "" ], [ "Liang", "Xiaodan", "" ] ]
new_dataset
0.999627
2212.02749
Nariman Habili
Nariman Habili, Ernest Kwan, Weihao Li, Christfried Webers, Jeremy Oorloff, Mohammad Ali Armin, Lars Petersson
A Hyperspectral and RGB Dataset for Building Facade Segmentation
null
null
null
null
cs.CV
http://creativecommons.org/licenses/by-nc-sa/4.0/
Hyperspectral Imaging (HSI) provides detailed spectral information and has been utilised in many real-world applications. This work introduces an HSI dataset of building facades in a light industry environment with the aim of classifying different building materials in a scene. The dataset is called the Light Industrial Building HSI (LIB-HSI) dataset. This dataset consists of nine categories and 44 classes. In this study, we investigated deep learning based semantic segmentation algorithms on RGB and hyperspectral images to classify various building materials, such as timber, brick and concrete.
[ { "version": "v1", "created": "Tue, 6 Dec 2022 04:38:44 GMT" } ]
2022-12-07T00:00:00
[ [ "Habili", "Nariman", "" ], [ "Kwan", "Ernest", "" ], [ "Li", "Weihao", "" ], [ "Webers", "Christfried", "" ], [ "Oorloff", "Jeremy", "" ], [ "Armin", "Mohammad Ali", "" ], [ "Petersson", "Lars", "" ] ]
new_dataset
0.999755
2212.02821
Madhu Raka
Swati Bhardwaj, Mokshi Goyal and Madhu Raka
New Quantum codes from constacyclic codes over a general non-chain ring
null
null
null
null
cs.IT math.IT
http://creativecommons.org/licenses/by-nc-sa/4.0/
Let $q$ be a prime power and let $\mathcal{R}=\mathbb{F}_{q}[u_1,u_2, \cdots, u_k]/\langle f_i(u_i),u_iu_j-u_ju_i\rangle$ be a finite non-chain ring, where $f_i(u_i), 1\leq i \leq k$ are polynomials, not all linear, which split into distinct linear factors over $\mathbb{F}_{q}$. We characterize constacyclic codes over the ring $\mathcal{R}$ and study quantum codes from these. As an application, some new and better quantum codes, as compared to the best known codes, are obtained. We also prove that the choice of the polynomials $f_i(u_i),$ $1 \leq i \leq k$ is irrelevant while constructing quantum codes from constacyclic codes over $\mathcal{R}$, it depends only on their degrees. It is shown that there always exists Quantum MDS code $[[n,n-2,2]]_q$ for any $n$ with $\gcd (n,q)\neq 1.$
[ { "version": "v1", "created": "Tue, 6 Dec 2022 08:32:49 GMT" } ]
2022-12-07T00:00:00
[ [ "Bhardwaj", "Swati", "" ], [ "Goyal", "Mokshi", "" ], [ "Raka", "Madhu", "" ] ]
new_dataset
0.999801
2212.02845
Yan Wang
Yan Wang, Junbo Yin, Wei Li, Pascal Frossard, Ruigang Yang, Jianbing Shen
SSDA3D: Semi-supervised Domain Adaptation for 3D Object Detection from Point Cloud
Accepted by AAAI 2023
null
null
null
cs.CV cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
LiDAR-based 3D object detection is an indispensable task in advanced autonomous driving systems. Though impressive detection results have been achieved by superior 3D detectors, they suffer from significant performance degeneration when facing unseen domains, such as different LiDAR configurations, different cities, and weather conditions. The mainstream approaches tend to solve these challenges by leveraging unsupervised domain adaptation (UDA) techniques. However, these UDA solutions just yield unsatisfactory 3D detection results when there is a severe domain shift, e.g., from Waymo (64-beam) to nuScenes (32-beam). To address this, we present a novel Semi-Supervised Domain Adaptation method for 3D object detection (SSDA3D), where only a few labeled target data is available, yet can significantly improve the adaptation performance. In particular, our SSDA3D includes an Inter-domain Adaptation stage and an Intra-domain Generalization stage. In the first stage, an Inter-domain Point-CutMix module is presented to efficiently align the point cloud distribution across domains. The Point-CutMix generates mixed samples of an intermediate domain, thus encouraging to learn domain-invariant knowledge. Then, in the second stage, we further enhance the model for better generalization on the unlabeled target set. This is achieved by exploring Intra-domain Point-MixUp in semi-supervised learning, which essentially regularizes the pseudo label distribution. Experiments from Waymo to nuScenes show that, with only 10% labeled target data, our SSDA3D can surpass the fully-supervised oracle model with 100% target label. Our code is available at https://github.com/yinjunbo/SSDA3D.
[ { "version": "v1", "created": "Tue, 6 Dec 2022 09:32:44 GMT" } ]
2022-12-07T00:00:00
[ [ "Wang", "Yan", "" ], [ "Yin", "Junbo", "" ], [ "Li", "Wei", "" ], [ "Frossard", "Pascal", "" ], [ "Yang", "Ruigang", "" ], [ "Shen", "Jianbing", "" ] ]
new_dataset
0.99202
2212.02871
Siyuan Zhou
Siyuan Zhou and Chunru Zhan and Biao Wang and Tiezheng Ge and Yuning Jiang and Li Niu
Video Object of Interest Segmentation
13 pages, 8 figures
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this work, we present a new computer vision task named video object of interest segmentation (VOIS). Given a video and a target image of interest, our objective is to simultaneously segment and track all objects in the video that are relevant to the target image. This problem combines the traditional video object segmentation task with an additional image indicating the content that users are concerned with. Since no existing dataset is perfectly suitable for this new task, we specifically construct a large-scale dataset called LiveVideos, which contains 2418 pairs of target images and live videos with instance-level annotations. In addition, we propose a transformer-based method for this task. We revisit Swin Transformer and design a dual-path structure to fuse video and image features. Then, a transformer decoder is employed to generate object proposals for segmentation and tracking from the fused features. Extensive experiments on LiveVideos dataset show the superiority of our proposed method.
[ { "version": "v1", "created": "Tue, 6 Dec 2022 10:21:10 GMT" } ]
2022-12-07T00:00:00
[ [ "Zhou", "Siyuan", "" ], [ "Zhan", "Chunru", "" ], [ "Wang", "Biao", "" ], [ "Ge", "Tiezheng", "" ], [ "Jiang", "Yuning", "" ], [ "Niu", "Li", "" ] ]
new_dataset
0.999872
2212.02896
Pengfei Hu
Pengfei Hu, Zhenrong Zhang, Jianshu Zhang, Jun Du, Jiajia Wu
Multimodal Tree Decoder for Table of Contents Extraction in Document Images
Accepted by ICPR2022
null
null
null
cs.CV cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Table of contents (ToC) extraction aims to extract headings of different levels in documents to better understand the outline of the contents, which can be widely used for document understanding and information retrieval. Existing works often use hand-crafted features and predefined rule-based functions to detect headings and resolve the hierarchical relationship between headings. Both the benchmark and research based on deep learning are still limited. Accordingly, in this paper, we first introduce a standard dataset, HierDoc, including image samples from 650 documents of scientific papers with their content labels. Then we propose a novel end-to-end model by using the multimodal tree decoder (MTD) for ToC as a benchmark for HierDoc. The MTD model is mainly composed of three parts, namely encoder, classifier, and decoder. The encoder fuses the multimodality features of vision, text, and layout information for each entity of the document. Then the classifier recognizes and selects the heading entities. Next, to parse the hierarchical relationship between the heading entities, a tree-structured decoder is designed. To evaluate the performance, both the metric of tree-edit-distance similarity (TEDS) and F1-Measure are adopted. Finally, our MTD approach achieves an average TEDS of 87.2% and an average F1-Measure of 88.1% on the test set of HierDoc. The code and dataset will be released at: https://github.com/Pengfei-Hu/MTD.
[ { "version": "v1", "created": "Tue, 6 Dec 2022 11:38:31 GMT" } ]
2022-12-07T00:00:00
[ [ "Hu", "Pengfei", "" ], [ "Zhang", "Zhenrong", "" ], [ "Zhang", "Jianshu", "" ], [ "Du", "Jun", "" ], [ "Wu", "Jiajia", "" ] ]
new_dataset
0.999704
2212.02935
Richard Preen
Richard J. Preen and Jim Smith
ACRO: A multi-language toolkit for supporting Automated Checking of Research Outputs
null
null
null
null
cs.CR cs.IR cs.SE stat.AP stat.ME
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper discusses the development of an open source tool ACRO, (Automatic Checking of Research Outputs) to assist researchers and data governance teams by distinguishing between: research output that is safe to publish; output that requires further analysis; and output that cannot be published because it creates substantial risk of disclosing private data. ACRO extends the functionality and accessibility of a previous prototype by providing a light-weight 'skin' that sits over well-known analysis tools, and enables access in a variety of programming languages researchers might use. This adds functionality to (i) identify potentially disclosive outputs against a range of commonly used disclosure tests; (ii) suppress outputs where required; (iii) report reasons for suppression; and (iv) produce simple summary documents Trusted Research Environment (TRE) staff can use to streamline their workflow. The ACRO code and documentation are available under an MIT license at https://github.com/AI-SDC/ACRO
[ { "version": "v1", "created": "Tue, 6 Dec 2022 12:45:15 GMT" } ]
2022-12-07T00:00:00
[ [ "Preen", "Richard J.", "" ], [ "Smith", "Jim", "" ] ]
new_dataset
0.996136
2212.03091
Pei Chen
Pei Chen, Wenlin Yao, Hongming Zhang, Xiaoman Pan, Dian Yu, Dong Yu, and Jianshu Chen
ZeroKBC: A Comprehensive Benchmark for Zero-Shot Knowledge Base Completion
ICDMW 2022
null
null
null
cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Knowledge base completion (KBC) aims to predict the missing links in knowledge graphs. Previous KBC tasks and approaches mainly focus on the setting where all test entities and relations have appeared in the training set. However, there has been limited research on the zero-shot KBC settings, where we need to deal with unseen entities and relations that emerge in a constantly growing knowledge base. In this work, we systematically examine different possible scenarios of zero-shot KBC and develop a comprehensive benchmark, ZeroKBC, that covers these scenarios with diverse types of knowledge sources. Our systematic analysis reveals several missing yet important zero-shot KBC settings. Experimental results show that canonical and state-of-the-art KBC systems cannot achieve satisfactory performance on this challenging benchmark. By analyzing the strength and weaknesses of these systems on solving ZeroKBC, we further present several important observations and promising future directions.
[ { "version": "v1", "created": "Tue, 6 Dec 2022 16:02:09 GMT" } ]
2022-12-07T00:00:00
[ [ "Chen", "Pei", "" ], [ "Yao", "Wenlin", "" ], [ "Zhang", "Hongming", "" ], [ "Pan", "Xiaoman", "" ], [ "Yu", "Dian", "" ], [ "Yu", "Dong", "" ], [ "Chen", "Jianshu", "" ] ]
new_dataset
0.98969
2212.03222
William Bruno
William Bruno, Dan Roth
LawngNLI: A Long-Premise Benchmark for In-Domain Generalization from Short to Long Contexts and for Implication-Based Retrieval
Findings of EMNLP 2022
null
null
null
cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Natural language inference has trended toward studying contexts beyond the sentence level. An important application area is law: past cases often do not foretell how they apply to new situations and implications must be inferred. This paper introduces LawngNLI, constructed from U.S. legal opinions with automatic labels with high human-validated accuracy. Premises are long and multigranular. Experiments show two use cases. First, LawngNLI can benchmark for in-domain generalization from short to long contexts. It has remained unclear if large-scale long-premise NLI datasets actually need to be constructed: near-top performance on long premises could be achievable by fine-tuning using short premises. Without multigranularity, benchmarks cannot distinguish lack of fine-tuning on long premises versus domain shift between short and long datasets. In contrast, our long and short premises share the same examples and domain. Models fine-tuned using several past NLI datasets and/or our short premises fall short of top performance on our long premises. So for at least certain domains (such as ours), large-scale long-premise datasets are needed. Second, LawngNLI can benchmark for implication-based retrieval. Queries are entailed or contradicted by target documents, allowing users to move between arguments and evidence. Leading retrieval models perform reasonably zero shot on a LawngNLI-derived retrieval task. We compare different systems for re-ranking, including lexical overlap and cross-encoders fine-tuned using a modified LawngNLI or past NLI datasets. LawngNLI can train and test systems for implication-based case retrieval and argumentation.
[ { "version": "v1", "created": "Tue, 6 Dec 2022 18:42:39 GMT" } ]
2022-12-07T00:00:00
[ [ "Bruno", "William", "" ], [ "Roth", "Dan", "" ] ]
new_dataset
0.999713
2212.03237
Umar Iqbal
Umar Iqbal, Akin Caliskan, Koki Nagano, Sameh Khamis, Pavlo Molchanov, Jan Kautz
RANA: Relightable Articulated Neural Avatars
project page: https://nvlabs.github.io/RANA/
null
null
null
cs.CV
http://creativecommons.org/licenses/by/4.0/
We propose RANA, a relightable and articulated neural avatar for the photorealistic synthesis of humans under arbitrary viewpoints, body poses, and lighting. We only require a short video clip of the person to create the avatar and assume no knowledge about the lighting environment. We present a novel framework to model humans while disentangling their geometry, texture, and also lighting environment from monocular RGB videos. To simplify this otherwise ill-posed task we first estimate the coarse geometry and texture of the person via SMPL+D model fitting and then learn an articulated neural representation for photorealistic image generation. RANA first generates the normal and albedo maps of the person in any given target body pose and then uses spherical harmonics lighting to generate the shaded image in the target lighting environment. We also propose to pretrain RANA using synthetic images and demonstrate that it leads to better disentanglement between geometry and texture while also improving robustness to novel body poses. Finally, we also present a new photorealistic synthetic dataset, Relighting Humans, to quantitatively evaluate the performance of the proposed approach.
[ { "version": "v1", "created": "Tue, 6 Dec 2022 18:59:31 GMT" } ]
2022-12-07T00:00:00
[ [ "Iqbal", "Umar", "" ], [ "Caliskan", "Akin", "" ], [ "Nagano", "Koki", "" ], [ "Khamis", "Sameh", "" ], [ "Molchanov", "Pavlo", "" ], [ "Kautz", "Jan", "" ] ]
new_dataset
0.974585
2009.14115
Adam Kortylewski
Yutong Bai, Angtian Wang, Adam Kortylewski, Alan Yuille
CoKe: Localized Contrastive Learning for Robust Keypoint Detection
Accepted to WACV 2023
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we introduce a contrastive learning framework for keypoint detection (CoKe). Keypoint detection differs from other visual tasks where contrastive learning has been applied because the input is a set of images in which multiple keypoints are annotated. This requires the contrastive learning to be extended such that the keypoints are represented and detected independently, which enables the contrastive loss to make the keypoint features different from each other and from the background. Our approach has two benefits: It enables us to exploit contrastive learning for keypoint detection, and by detecting each keypoint independently the detection becomes more robust to occlusion compared to holistic methods, such as stacked hourglass networks, which attempt to detect all keypoints jointly. Our CoKe framework introduces several technical innovations. In particular, we introduce: (i) A clutter bank to represent non-keypoint features; (ii) a keypoint bank that stores prototypical representations of keypoints to approximate the contrastive loss between keypoints; and (iii) a cumulative moving average update to learn the keypoint prototypes while training the feature extractor. Our experiments on a range of diverse datasets (PASCAL3D+, MPII, ObjectNet3D) show that our approach works as well, or better than, alternative methods for keypoint detection, even for human keypoints, for which the literature is vast. Moreover, we observe that CoKe is exceptionally robust to partial occlusion and previously unseen object poses.
[ { "version": "v1", "created": "Tue, 29 Sep 2020 16:00:43 GMT" }, { "version": "v2", "created": "Wed, 30 Sep 2020 01:32:46 GMT" }, { "version": "v3", "created": "Mon, 23 Nov 2020 16:22:35 GMT" }, { "version": "v4", "created": "Mon, 5 Dec 2022 08:56:16 GMT" } ]
2022-12-06T00:00:00
[ [ "Bai", "Yutong", "" ], [ "Wang", "Angtian", "" ], [ "Kortylewski", "Adam", "" ], [ "Yuille", "Alan", "" ] ]
new_dataset
0.955058
2104.09375
Burak Ekim
Burak Ekim, Elif Sertel
A Multi-Task Deep Learning Framework for Building Footprint Segmentation
International Geoscience and Remote Sensing Symposium (IGARSS), Jul 2021, Brussels, Belgium
null
10.1109/IGARSS47720.2021.9554766
null
cs.CV
http://creativecommons.org/licenses/by/4.0/
The task of building footprint segmentation has been well-studied in the context of remote sensing (RS) as it provides valuable information in many aspects, however, difficulties brought by the nature of RS images such as variations in the spatial arrangements and in-consistent constructional patterns require studying further, since it often causes poorly classified segmentation maps. We address this need by designing a joint optimization scheme for the task of building footprint delineation and introducing two auxiliary tasks; image reconstruction and building footprint boundary segmentation with the intent to reveal the common underlying structure to advance the classification accuracy of a single task model under the favor of auxiliary tasks. In particular, we propose a deep multi-task learning (MTL) based unified fully convolutional framework which operates in an end-to-end manner by making use of joint loss function with learnable loss weights considering the homoscedastic uncertainty of each task loss. Experimental results conducted on the SpaceNet6 dataset demonstrate the potential of the proposed MTL framework as it improves the classification accuracy considerably compared to single-task and lesser compounded tasks.
[ { "version": "v1", "created": "Mon, 19 Apr 2021 15:07:27 GMT" } ]
2022-12-06T00:00:00
[ [ "Ekim", "Burak", "" ], [ "Sertel", "Elif", "" ] ]
new_dataset
0.997504
2108.07366
Anurag Murty Naredla
Anna Lubiw, Anurag Murty Naredla
The Visibility Center of a Simple Polygon
Full-length version of a paper that appeared at the European Symposium of Algorithms 2021
null
null
null
cs.CG
http://creativecommons.org/licenses/by/4.0/
We introduce the \emph{visibility center} of a set of points inside a polygon -- a point $c_V$ such that the maximum geodesic distance from $c_V$ to see any point in the set is minimized. For a simple polygon of $n$ vertices and a set of $m$ points inside it, we give an $O((n+m) \log {(n+m)})$ time algorithm to find the visibility center. We find the visibility center of \emph{all} points in a simple polygon in $O(n \log n)$ time. Our algorithm reduces the visibility center problem to the problem of finding the geodesic center of a set of half-polygons inside a polygon, which is of independent interest. We give an $O((n+k) \log (n+k))$ time algorithm for this problem, where $k$ is the number of half-polygons.
[ { "version": "v1", "created": "Mon, 16 Aug 2021 22:44:32 GMT" }, { "version": "v2", "created": "Wed, 18 Aug 2021 16:03:30 GMT" }, { "version": "v3", "created": "Sun, 4 Dec 2022 23:06:46 GMT" } ]
2022-12-06T00:00:00
[ [ "Lubiw", "Anna", "" ], [ "Naredla", "Anurag Murty", "" ] ]
new_dataset
0.975771
2109.00405
Celyn Walters
Celyn Walters and Simon Hadfield
EVReflex: Dense Time-to-Impact Prediction for Event-based Obstacle Avoidance
To be published in IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2021
null
10.1109/IROS51168.2021.9636327
null
cs.CV cs.RO
http://creativecommons.org/licenses/by-nc-nd/4.0/
The broad scope of obstacle avoidance has led to many kinds of computer vision-based approaches. Despite its popularity, it is not a solved problem. Traditional computer vision techniques using cameras and depth sensors often focus on static scenes, or rely on priors about the obstacles. Recent developments in bio-inspired sensors present event cameras as a compelling choice for dynamic scenes. Although these sensors have many advantages over their frame-based counterparts, such as high dynamic range and temporal resolution, event-based perception has largely remained in 2D. This often leads to solutions reliant on heuristics and specific to a particular task. We show that the fusion of events and depth overcomes the failure cases of each individual modality when performing obstacle avoidance. Our proposed approach unifies event camera and lidar streams to estimate metric time-to-impact without prior knowledge of the scene geometry or obstacles. In addition, we release an extensive event-based dataset with six visual streams spanning over 700 scanned scenes.
[ { "version": "v1", "created": "Wed, 1 Sep 2021 14:34:20 GMT" } ]
2022-12-06T00:00:00
[ [ "Walters", "Celyn", "" ], [ "Hadfield", "Simon", "" ] ]
new_dataset
0.998225
2109.00945
Lynnette Hui Xian Ng
Lynnette Hui Xian Ng, Iain Cruickshank, Kathleen M. Carley
Coordinating Narratives and the Capitol Riots on Parler
null
Computational Mathematics Organizational Theory (2022)
10.1007/s10588-022-09371-2
null
cs.SI cs.CL
http://creativecommons.org/licenses/by/4.0/
Coordinated disinformation campaigns are used to influence social media users, potentially leading to offline violence. In this study, we introduce a general methodology to uncover coordinated messaging through analysis of user parleys on Parler. The proposed method constructs a user-to-user coordination network graph induced by a user-to-text graph and a text-to-text similarity graph. The text-to-text graph is constructed based on the textual similarity of Parler posts. We study three influential groups of users in the 6 January 2020 Capitol riots and detect networks of coordinated user clusters that are all posting similar textual content in support of different disinformation narratives related to the U.S. 2020 elections.
[ { "version": "v1", "created": "Thu, 2 Sep 2021 13:44:59 GMT" } ]
2022-12-06T00:00:00
[ [ "Ng", "Lynnette Hui Xian", "" ], [ "Cruickshank", "Iain", "" ], [ "Carley", "Kathleen M.", "" ] ]
new_dataset
0.97697
2110.02929
Martino Sorbaro
Julian B\"uchel, Gregor Lenz, Yalun Hu, Sadique Sheik, Martino Sorbaro
Adversarial Attacks on Spiking Convolutional Neural Networks for Event-based Vision
9 pages plus Supplementary Material. Accepted in Frontiers in Neuroscience -- Neuromorphic Engineering
null
null
null
cs.CV
http://creativecommons.org/licenses/by/4.0/
Event-based dynamic vision sensors provide very sparse output in the form of spikes, which makes them suitable for low-power applications. Convolutional spiking neural networks model such event-based data and develop their full energy-saving potential when deployed on asynchronous neuromorphic hardware. Event-based vision being a nascent field, the sensitivity of spiking neural networks to potentially malicious adversarial attacks has received little attention so far. We show how white-box adversarial attack algorithms can be adapted to the discrete and sparse nature of event-based visual data, and demonstrate smaller perturbation magnitudes at higher success rates than the current state-of-the-art algorithms. For the first time, we also verify the effectiveness of these perturbations directly on neuromorphic hardware. Finally, we discuss the properties of the resulting perturbations, the effect of adversarial training as a defense strategy, and future directions.
[ { "version": "v1", "created": "Wed, 6 Oct 2021 17:20:05 GMT" }, { "version": "v2", "created": "Mon, 20 Dec 2021 15:34:41 GMT" }, { "version": "v3", "created": "Mon, 5 Dec 2022 12:49:10 GMT" } ]
2022-12-06T00:00:00
[ [ "Büchel", "Julian", "" ], [ "Lenz", "Gregor", "" ], [ "Hu", "Yalun", "" ], [ "Sheik", "Sadique", "" ], [ "Sorbaro", "Martino", "" ] ]
new_dataset
0.983443
2111.03788
Takuma Seno
Takuma Seno, Michita Imai
d3rlpy: An Offline Deep Reinforcement Learning Library
Journal of Machine Learning Research
Journal of Machine Learning Research 23(315) (2022) 1-20;
null
null
cs.LG cs.AI
http://creativecommons.org/licenses/by/4.0/
In this paper, we introduce d3rlpy, an open-sourced offline deep reinforcement learning (RL) library for Python. d3rlpy supports a set of offline deep RL algorithms as well as off-policy online algorithms via a fully documented plug-and-play API. To address a reproducibility issue, we conduct a large-scale benchmark with D4RL and Atari 2600 dataset to ensure implementation quality and provide experimental scripts and full tables of results. The d3rlpy source code can be found on GitHub: \url{https://github.com/takuseno/d3rlpy}.
[ { "version": "v1", "created": "Sat, 6 Nov 2021 03:09:39 GMT" }, { "version": "v2", "created": "Sat, 3 Dec 2022 12:03:07 GMT" } ]
2022-12-06T00:00:00
[ [ "Seno", "Takuma", "" ], [ "Imai", "Michita", "" ] ]
new_dataset
0.999507
2112.13230
Yang Li
Yang Li, Yiting Dong, Dongcheng Zhao, Yi Zeng
N-Omniglot, a large-scale neuromorphic dataset for spatio-temporal sparse few-shot learning
null
null
10.1038/s41597-022-01851-z
null
cs.NE cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Few-shot learning (learning with a few samples) is one of the most important cognitive abilities of the human brain. However, the current artificial intelligence systems meet difficulties in achieving this ability. Similar challenges also exist for biologically plausible spiking neural networks (SNNs). Datasets for traditional few-shot learning domains provide few amounts of temporal information. and the absence of neuromorphic datasets has hindered the development of few-shot learning for SNNs. Here, to the best of our knowledge, we provide the first neuromorphic dataset for few-shot learning using SNNs: N-Omniglot, based on the Dynamic Vision Sensor. It contains 1,623 categories of handwritten characters, with only 20 samples per class. N-Omniglot eliminates the need for a neuromorphic dataset for SNNs with high spareness and tremendous temporal coherence. Additionally, the dataset provides a powerful challenge and a suitable benchmark for developing SNNs algorithms in the few-shot learning domain due to the chronological information of strokes. We also provide the improved nearest neighbor, convolutional network, SiameseNet, and meta-learning algorithm in the spiking version for verification.
[ { "version": "v1", "created": "Sat, 25 Dec 2021 12:41:34 GMT" }, { "version": "v2", "created": "Tue, 28 Dec 2021 11:07:25 GMT" }, { "version": "v3", "created": "Sat, 3 Dec 2022 15:25:33 GMT" } ]
2022-12-06T00:00:00
[ [ "Li", "Yang", "" ], [ "Dong", "Yiting", "" ], [ "Zhao", "Dongcheng", "" ], [ "Zeng", "Yi", "" ] ]
new_dataset
0.995697
2202.04215
Omar Hamido
Omar Costa Hamido
QAC: Quantum-computing Aided Composition
Pre-publication draft, to appear in book 'Quantum Computer Music', E. R. Miranda (Ed.)
null
10.1007/978-3-031-13909-3_8
null
cs.ET cs.HC cs.SD eess.AS quant-ph
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this chapter I will discuss the role of quantum computing in computer music and how it can be integrated to better serve the creative artists. I will start by considering different approaches in current computer music and quantum computing tools, as well as reviewing some previous attempts to integrate them. Then, I will reflect on the meaning of this integration and present what I coined as QAC (Quantum-computing Aided Composition) as well as an early attempt at realizing it. This chapter will also introduce The QAC Toolkit Max package, analyze its performance, and explore some examples of what it can offer to realtime creative practice. Lastly, I will present a real case scenario of QAC in the creative work Disklavier Prelude #3.
[ { "version": "v1", "created": "Wed, 9 Feb 2022 01:17:21 GMT" } ]
2022-12-06T00:00:00
[ [ "Hamido", "Omar Costa", "" ] ]
new_dataset
0.99794
2204.00862
Pei Ke
Pei Ke, Hao Zhou, Yankai Lin, Peng Li, Jie Zhou, Xiaoyan Zhu, Minlie Huang
CTRLEval: An Unsupervised Reference-Free Metric for Evaluating Controlled Text Generation
Accepted by ACL 2022 (Main Conference)
null
null
null
cs.CL cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Existing reference-free metrics have obvious limitations for evaluating controlled text generation models. Unsupervised metrics can only provide a task-agnostic evaluation result which correlates weakly with human judgments, whereas supervised ones may overfit task-specific data with poor generalization ability to other datasets. In this paper, we propose an unsupervised reference-free metric called CTRLEval, which evaluates controlled text generation from different aspects by formulating each aspect into multiple text infilling tasks. On top of these tasks, the metric assembles the generation probabilities from a pre-trained language model without any model training. Experimental results show that our metric has higher correlations with human judgments than other baselines, while obtaining better generalization of evaluating generated texts from different models and with different qualities.
[ { "version": "v1", "created": "Sat, 2 Apr 2022 13:42:49 GMT" }, { "version": "v2", "created": "Mon, 5 Dec 2022 10:11:11 GMT" } ]
2022-12-06T00:00:00
[ [ "Ke", "Pei", "" ], [ "Zhou", "Hao", "" ], [ "Lin", "Yankai", "" ], [ "Li", "Peng", "" ], [ "Zhou", "Jie", "" ], [ "Zhu", "Xiaoyan", "" ], [ "Huang", "Minlie", "" ] ]
new_dataset
0.995242
2204.07003
Paolo Perrone
Sean Moss and Paolo Perrone
Probability monads with submonads of deterministic states - Extended version
16 pages. Extended version of paper accepted for LICS 2022 conference
null
10.1145/3531130.3533355
null
cs.LO math.CT math.PR
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Probability theory can be studied synthetically as the computational effect embodied by a commutative monad. In the recently proposed Markov categories, one works with an abstraction of the Kleisli category and then defines deterministic morphisms equationally in terms of copying and discarding. The resulting difference between 'pure' and 'deterministic' leads us to investigate the 'sober' objects for a probability monad, for which the two concepts coincide. We propose natural conditions on a probability monad which allow us to identify the sober objects and define an idempotent sobrification functor. Our framework applies to many examples of interest, including the Giry monad on measurable spaces, and allows us to sharpen a previously given version of de Finetti's theorem for Markov categories. This is an extended version of the paper accepted for the Logic In Computer Science (LICS) conference 2022. In this document we include more mathematical details, including all the proofs, of the statements and constructions given in the published version.
[ { "version": "v1", "created": "Thu, 14 Apr 2022 14:54:45 GMT" } ]
2022-12-06T00:00:00
[ [ "Moss", "Sean", "" ], [ "Perrone", "Paolo", "" ] ]
new_dataset
0.977297
2205.10956
Wei Yuan
Wei Yuan, Quanjun Zhang, Tieke He, Chunrong Fang, Nguyen Quoc Viet Hung, Xiaodong Hao, Hongzhi Yin
CIRCLE: Continual Repair across Programming Languages
This paper was accepted by ISSTA2022
null
null
null
cs.SE
http://creativecommons.org/licenses/by-nc-sa/4.0/
Automatic Program Repair (APR) aims at fixing buggy source code with less manual debugging efforts, which plays a vital role in improving software reliability and development productivity. Recent APR works have achieved remarkable progress via applying deep learning (DL), particularly neural machine translation (NMT) techniques. However, we observe that existing DL-based APR models suffer from at least two severe drawbacks: (1) Most of them can only generate patches for a single programming language, as a result, to repair multiple languages, we have to build and train many repairing models. (2) Most of them are developed in an offline manner. Therefore, they won't function when there are new-coming requirements. To address the above problems, a T5-based APR framework equipped with continual learning ability across multiple programming languages is proposed, namely \emph{C}ont\emph{I}nual \emph{R}epair a\emph{C}ross Programming \emph{L}anguag\emph{E}s (\emph{CIRCLE}). Specifically, (1) CIRCLE utilizes a prompting function to narrow the gap between natural language processing (NLP) pre-trained tasks and APR. (2) CIRCLE adopts a difficulty-based rehearsal strategy to achieve lifelong learning for APR without access to the full historical data. (3) An elastic regularization method is employed to strengthen CIRCLE's continual learning ability further, preventing it from catastrophic forgetting. (4) CIRCLE applies a simple but effective re-repairing method to revise generated errors caused by crossing multiple programming languages. We train CIRCLE for four languages (i.e., C, JAVA, JavaScript, and Python) and evaluate it on five commonly used benchmarks. The experimental results demonstrate that CIRCLE not only effectively and efficiently repairs multiple programming languages in continual learning settings, but also achieves state-of-the-art performance with a single repair model.
[ { "version": "v1", "created": "Sun, 22 May 2022 23:34:37 GMT" }, { "version": "v2", "created": "Thu, 26 May 2022 09:14:30 GMT" }, { "version": "v3", "created": "Thu, 2 Jun 2022 01:16:24 GMT" }, { "version": "v4", "created": "Sat, 3 Dec 2022 14:24:03 GMT" } ]
2022-12-06T00:00:00
[ [ "Yuan", "Wei", "" ], [ "Zhang", "Quanjun", "" ], [ "He", "Tieke", "" ], [ "Fang", "Chunrong", "" ], [ "Hung", "Nguyen Quoc Viet", "" ], [ "Hao", "Xiaodong", "" ], [ "Yin", "Hongzhi", "" ] ]
new_dataset
0.989381
2207.07195
Duowei Li
Duowei Li (1 and 2), Jianping Wu (1), Feng Zhu (2), Tianyi Chen (2), Yiik Diew Wong (2) ((1) Department of Civil Engineering, Tsinghua University, China, (2) School of Civil and Environmental Engineering, Nanyang Technological University, Singapore)
COOR-PLT: A hierarchical control model for coordinating adaptive platoons of connected and autonomous vehicles at signal-free intersections based on deep reinforcement learning
This paper has been submitted to Transportation Research Part C: Emerging Technologies and is currently under review
Transportation Research Part C: Emerging Technologies 146 (2023): 103933
10.1016/j.trc.2022.103933
null
cs.LG cs.MA cs.SY eess.SY
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Platooning and coordination are two implementation strategies that are frequently proposed for traffic control of connected and autonomous vehicles (CAVs) at signal-free intersections instead of using conventional traffic signals. However, few studies have attempted to integrate both strategies to better facilitate the CAV control at signal-free intersections. To this end, this study proposes a hierarchical control model, named COOR-PLT, to coordinate adaptive CAV platoons at a signal-free intersection based on deep reinforcement learning (DRL). COOR-PLT has a two-layer framework. The first layer uses a centralized control strategy to form adaptive platoons. The optimal size of each platoon is determined by considering multiple objectives (i.e., efficiency, fairness and energy saving). The second layer employs a decentralized control strategy to coordinate multiple platoons passing through the intersection. Each platoon is labeled with coordinated status or independent status, upon which its passing priority is determined. As an efficient DRL algorithm, Deep Q-network (DQN) is adopted to determine platoon sizes and passing priorities respectively in the two layers. The model is validated and examined on the simulator Simulation of Urban Mobility (SUMO). The simulation results demonstrate that the model is able to: (1) achieve satisfactory convergence performances; (2) adaptively determine platoon size in response to varying traffic conditions; and (3) completely avoid deadlocks at the intersection. By comparison with other control methods, the model manifests its superiority of adopting adaptive platooning and DRL-based coordination strategies. Also, the model outperforms several state-of-the-art methods on reducing travel time and fuel consumption in different traffic conditions.
[ { "version": "v1", "created": "Fri, 1 Jul 2022 02:22:31 GMT" } ]
2022-12-06T00:00:00
[ [ "Li", "Duowei", "", "1 and 2" ], [ "Wu", "Jianping", "" ], [ "Zhu", "Feng", "" ], [ "Chen", "Tianyi", "" ], [ "Wong", "Yiik Diew", "" ] ]
new_dataset
0.986756
2208.04563
Tarun Rambha
Saumya Bhatnagar, Tarun Rambha, Gitakrishnan Ramadurai
An Agent-Based Fleet Management Model for First- and Last-Mile Services
null
null
10.1007/s11116-022-10363-z
null
cs.MA
http://creativecommons.org/licenses/by/4.0/
With the growth of cars and car-sharing applications, commuters in many cities, particularly developing countries, are shifting away from public transport. These shifts have affected two key stakeholders: transit operators and first- and last-mile (FLM) services. Although most cities continue to invest heavily in bus and metro projects to make public transit attractive, ridership in these systems has often failed to reach targeted levels. FLM service providers also experience lower demand and revenues in the wake of shifts to other means of transport. Effective FLM options are required to prevent this phenomenon and make public transport attractive for commuters. One possible solution is to forge partnerships between public transport and FLM providers that offer competitive joint mobility options. Such solutions require prudent allocation of supply and optimised strategies for FLM operations and ride-sharing. To this end, we build an agent- and event-based simulation model which captures interactions between passengers and FLM services using statecharts, vehicle routing models, and other trip matching rules. An optimisation model for allocating FLM vehicles at different transit stations is proposed to reduce unserved requests. Using real-world metro transit demand data from Bengaluru, India, the effectiveness of our approach in improving FLM connectivity and quantifying the benefits of sharing trips is demonstrated.
[ { "version": "v1", "created": "Tue, 9 Aug 2022 06:52:28 GMT" }, { "version": "v2", "created": "Sun, 4 Dec 2022 12:56:27 GMT" } ]
2022-12-06T00:00:00
[ [ "Bhatnagar", "Saumya", "" ], [ "Rambha", "Tarun", "" ], [ "Ramadurai", "Gitakrishnan", "" ] ]
new_dataset
0.997786
2209.11388
Haoyu Lu
Haoyu Lu and Mingyu Ding and Nanyi Fei and Yuqi Huo and Zhiwu Lu
LGDN: Language-Guided Denoising Network for Video-Language Modeling
Accepted by NeurIPS2022
null
null
null
cs.CV cs.AI cs.MM
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Video-language modeling has attracted much attention with the rapid growth of web videos. Most existing methods assume that the video frames and text description are semantically correlated, and focus on video-language modeling at video level. However, this hypothesis often fails for two reasons: (1) With the rich semantics of video contents, it is difficult to cover all frames with a single video-level description; (2) A raw video typically has noisy/meaningless information (e.g., scenery shot, transition or teaser). Although a number of recent works deploy attention mechanism to alleviate this problem, the irrelevant/noisy information still makes it very difficult to address. To overcome such challenge, we thus propose an efficient and effective model, termed Language-Guided Denoising Network (LGDN), for video-language modeling. Different from most existing methods that utilize all extracted video frames, LGDN dynamically filters out the misaligned or redundant frames under the language supervision and obtains only 2--4 salient frames per video for cross-modal token-level alignment. Extensive experiments on five public datasets show that our LGDN outperforms the state-of-the-arts by large margins. We also provide detailed ablation study to reveal the critical importance of solving the noise issue, in hope of inspiring future video-language work.
[ { "version": "v1", "created": "Fri, 23 Sep 2022 03:35:59 GMT" }, { "version": "v2", "created": "Mon, 3 Oct 2022 04:14:08 GMT" }, { "version": "v3", "created": "Mon, 5 Dec 2022 07:20:42 GMT" } ]
2022-12-06T00:00:00
[ [ "Lu", "Haoyu", "" ], [ "Ding", "Mingyu", "" ], [ "Fei", "Nanyi", "" ], [ "Huo", "Yuqi", "" ], [ "Lu", "Zhiwu", "" ] ]
new_dataset
0.990015
2211.07971
Charalambos Themistocleous
Charalambos Themistocleous
Discourse and conversation impairments in patients with dementia
Book chapter
null
null
null
cs.CL
http://creativecommons.org/licenses/by/4.0/
Neurodegeneration characterizes individuals with different dementia subtypes (e.g., individuals with Alzheimer's Disease, Primary Progressive Aphasia, and Parkinson's Disease), leading to progressive decline in cognitive, linguistic, and social functioning. Speech and language impairments are early symptoms in individuals with focal forms of neurodegenerative conditions, coupled with deficits in cognitive, social, and behavioral domains. This paper reviews the findings on language and communication deficits and identifies the effects of dementia on the production and perception of discourse. It discusses findings concerning (i) language function, cognitive representation, and impairment, (ii) communicative competence, emotions, empathy, and theory-of-mind, and (iii) speech-in-interaction. It argues that clinical discourse analysis can provide a comprehensive assessment of language and communication skills in individuals, which complements the existing neurolinguistic evaluation for (differential) diagnosis, prognosis, and treatment efficacy evaluation.
[ { "version": "v1", "created": "Tue, 15 Nov 2022 08:18:30 GMT" }, { "version": "v2", "created": "Sat, 3 Dec 2022 16:20:41 GMT" } ]
2022-12-06T00:00:00
[ [ "Themistocleous", "Charalambos", "" ] ]
new_dataset
0.994021
2211.09210
Roberto Daza
Roberto Daza, Aythami Morales, Ruben Tolosana, Luis F. Gomez, Julian Fierrez, Javier Ortega-Garcia
edBB-Demo: Biometrics and Behavior Analysis for Online Educational Platforms
Accepted in "AAAI-23 Conference on Artificial Intelligence (Demonstration Program)"
null
null
null
cs.HC cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We present edBB-Demo, a demonstrator of an AI-powered research platform for student monitoring in remote education. The edBB platform aims to study the challenges associated to user recognition and behavior understanding in digital platforms. This platform has been developed for data collection, acquiring signals from a variety of sensors including keyboard, mouse, webcam, microphone, smartwatch, and an Electroencephalography band. The information captured from the sensors during the student sessions is modelled in a multimodal learning framework. The demonstrator includes: i) Biometric user authentication in an unsupervised environment; ii) Human action recognition based on remote video analysis; iii) Heart rate estimation from webcam video; and iv) Attention level estimation from facial expression analysis.
[ { "version": "v1", "created": "Wed, 16 Nov 2022 20:53:56 GMT" }, { "version": "v2", "created": "Mon, 5 Dec 2022 11:21:31 GMT" } ]
2022-12-06T00:00:00
[ [ "Daza", "Roberto", "" ], [ "Morales", "Aythami", "" ], [ "Tolosana", "Ruben", "" ], [ "Gomez", "Luis F.", "" ], [ "Fierrez", "Julian", "" ], [ "Ortega-Garcia", "Javier", "" ] ]
new_dataset
0.989073
2211.11720
Sheng Shen
Sheng Shen, Shijia Yang, Tianjun Zhang, Bohan Zhai, Joseph E. Gonzalez, Kurt Keutzer, Trevor Darrell
Multitask Vision-Language Prompt Tuning
Preprint
null
null
null
cs.CV cs.CL
http://creativecommons.org/licenses/by/4.0/
Prompt Tuning, conditioning on task-specific learned prompt vectors, has emerged as a data-efficient and parameter-efficient method for adapting large pretrained vision-language models to multiple downstream tasks. However, existing approaches usually consider learning prompt vectors for each task independently from scratch, thereby failing to exploit the rich shareable knowledge across different vision-language tasks. In this paper, we propose multitask vision-language prompt tuning (MVLPT), which incorporates cross-task knowledge into prompt tuning for vision-language models. Specifically, (i) we demonstrate the effectiveness of learning a single transferable prompt from multiple source tasks to initialize the prompt for each target task; (ii) we show many target tasks can benefit each other from sharing prompt vectors and thus can be jointly learned via multitask prompt tuning. We benchmark the proposed MVLPT using three representative prompt tuning methods, namely text prompt tuning, visual prompt tuning, and the unified vision-language prompt tuning. Results in 20 vision tasks demonstrate that the proposed approach outperforms all single-task baseline prompt tuning methods, setting the new state-of-the-art on the few-shot ELEVATER benchmarks and cross-task generalization benchmarks. To understand where the cross-task knowledge is most effective, we also conduct a large-scale study on task transferability with 20 vision tasks in 400 combinations for each prompt tuning method. It shows that the most performant MVLPT for each prompt tuning method prefers different task combinations and many tasks can benefit each other, depending on their visual similarity and label similarity. Code is available at https://github.com/sIncerass/MVLPT.
[ { "version": "v1", "created": "Mon, 21 Nov 2022 18:41:44 GMT" }, { "version": "v2", "created": "Tue, 22 Nov 2022 07:24:16 GMT" }, { "version": "v3", "created": "Mon, 5 Dec 2022 16:31:49 GMT" } ]
2022-12-06T00:00:00
[ [ "Shen", "Sheng", "" ], [ "Yang", "Shijia", "" ], [ "Zhang", "Tianjun", "" ], [ "Zhai", "Bohan", "" ], [ "Gonzalez", "Joseph E.", "" ], [ "Keutzer", "Kurt", "" ], [ "Darrell", "Trevor", "" ] ]
new_dataset
0.988922