arxiv_id
stringlengths 10
10
| published
stringlengths 20
20
| titles
stringlengths 9
243
| authors
sequencelengths 1
389
| abstract
stringlengths 96
3.09k
| categories
sequencelengths 1
10
| selected
bool 2
classes |
---|---|---|---|---|---|---|
2402.12914 | 2024-02-20T11:03:36Z | Large Language Model-based Human-Agent Collaboration for Complex Task
Solving | [
"Xueyang Feng",
"Zhi-Yuan Chen",
"Yujia Qin",
"Yankai Lin",
"Xu Chen",
"Zhiyuan Liu",
"Ji-Rong Wen"
] | In recent developments within the research community, the integration of
Large Language Models (LLMs) in creating fully autonomous agents has garnered
significant interest. Despite this, LLM-based agents frequently demonstrate
notable shortcomings in adjusting to dynamic environments and fully grasping
human needs. In this work, we introduce the problem of LLM-based human-agent
collaboration for complex task-solving, exploring their synergistic potential.
In addition, we propose a Reinforcement Learning-based Human-Agent
Collaboration method, ReHAC. This approach includes a policy model designed to
determine the most opportune stages for human intervention within the
task-solving process. We construct a human-agent collaboration dataset to train
this policy model in an offline reinforcement learning environment. Our
validation tests confirm the model's effectiveness. The results demonstrate
that the synergistic efforts of humans and LLM-based agents significantly
improve performance in complex tasks, primarily through well-planned, limited
human intervention. Datasets and code are available at:
https://github.com/XueyangFeng/ReHAC. | [
"cs.CL",
"cs.HC"
] | false |
2402.12959 | 2024-02-20T12:25:26Z | Prompt Stealing Attacks Against Large Language Models | [
"Zeyang Sha",
"Yang Zhang"
] | The increasing reliance on large language models (LLMs) such as ChatGPT in
various fields emphasizes the importance of ``prompt engineering,'' a
technology to improve the quality of model outputs. With companies investing
significantly in expert prompt engineers and educational resources rising to
meet market demand, designing high-quality prompts has become an intriguing
challenge. In this paper, we propose a novel attack against LLMs, named prompt
stealing attacks. Our proposed prompt stealing attack aims to steal these
well-designed prompts based on the generated answers. The prompt stealing
attack contains two primary modules: the parameter extractor and the prompt
reconstruction. The goal of the parameter extractor is to figure out the
properties of the original prompts. We first observe that most prompts fall
into one of three categories: direct prompt, role-based prompt, and in-context
prompt. Our parameter extractor first tries to distinguish the type of prompts
based on the generated answers. Then, it can further predict which role or how
many contexts are used based on the types of prompts. Following the parameter
extractor, the prompt reconstructor can be used to reconstruct the original
prompts based on the generated answers and the extracted features. The final
goal of the prompt reconstructor is to generate the reversed prompts, which are
similar to the original prompts. Our experimental results show the remarkable
performance of our proposed attacks. Our proposed attacks add a new dimension
to the study of prompt engineering and call for more attention to the security
issues on LLMs. | [
"cs.CR",
"cs.CL"
] | false |
2402.12969 | 2024-02-20T12:36:40Z | GlórIA -- A Generative and Open Large Language Model for Portuguese | [
"Ricardo Lopes",
"João Magalhães",
"David Semedo"
] | Significant strides have been made in natural language tasks, largely
attributed to the emergence of powerful large language models (LLMs). These
models, pre-trained on extensive and diverse corpora, have become increasingly
capable of comprehending the intricacies of language. Despite the abundance of
LLMs for many high-resource languages, the availability of such models remains
limited for European Portuguese. We introduce Gl\'orIA, a robust European
Portuguese decoder LLM. To pre-train Gl\'orIA, we assembled a comprehensive
PT-PT text corpus comprising 35 billion tokens from various sources. We present
our pre-training methodology, followed by an assessment of the model's
effectiveness on multiple downstream tasks. Additionally, to evaluate our
models' language modeling capabilities, we introduce CALAME-PT (Context-Aware
LAnguage Modeling Evaluation for Portuguese), the first Portuguese zero-shot
language-modeling benchmark. Evaluation shows that Gl\'orIA significantly
outperforms existing open PT decoder models in language modeling and that it
can generate sound, knowledge-rich, and coherent PT-PT text. The model also
exhibits strong potential for various downstream tasks. | [
"cs.CL",
"cs.AI"
] | false |
2402.12976 | 2024-02-20T12:53:31Z | The Impact of Demonstrations on Multilingual In-Context Learning: A
Multidimensional Analysis | [
"Miaoran Zhang",
"Vagrant Gautam",
"Mingyang Wang",
"Jesujoba O. Alabi",
"Xiaoyu Shen",
"Dietrich Klakow",
"Marius Mosbach"
] | In-context learning is a popular inference strategy where large language
models solve a task using only a few labelled demonstrations without needing
any parameter updates. Compared to work on monolingual (English) in-context
learning, multilingual in-context learning is under-explored, and we lack an
in-depth understanding of the role of demonstrations in this context. To
address this gap, we conduct a multidimensional analysis of multilingual
in-context learning, experimenting with 5 models from different model families,
9 datasets covering classification and generation tasks, and 56 typologically
diverse languages. Our results reveal that the effectiveness of demonstrations
varies significantly across models, tasks, and languages. We also find that
Llama 2-Chat, GPT-3.5, and GPT-4 are largely insensitive to the quality of
demonstrations. Instead, a carefully crafted template often eliminates the
benefits of demonstrations for some tasks and languages altogether. These
findings show that the importance of demonstrations might be overestimated. Our
work highlights the need for granular evaluation across multiple axes towards a
better understanding of in-context learning. | [
"cs.CL",
"cs.AI"
] | false |
2402.12984 | 2024-02-20T13:13:13Z | Can GNN be Good Adapter for LLMs? | [
"Xuanwen Huang",
"Kaiqiao Han",
"Yang Yang",
"Dezheng Bao",
"Quanjin Tao",
"Ziwei Chai",
"Qi Zhu"
] | Recently, large language models (LLMs) have demonstrated superior
capabilities in understanding and zero-shot learning on textual data, promising
significant advances for many text-related domains. In the graph domain,
various real-world scenarios also involve textual data, where tasks and node
features can be described by text. These text-attributed graphs (TAGs) have
broad applications in social media, recommendation systems, etc. Thus, this
paper explores how to utilize LLMs to model TAGs. Previous methods for TAG
modeling are based on million-scale LMs. When scaled up to billion-scale LLMs,
they face huge challenges in computational costs. Additionally, they also
ignore the zero-shot inference capabilities of LLMs. Therefore, we propose
GraphAdapter, which uses a graph neural network (GNN) as an efficient adapter
in collaboration with LLMs to tackle TAGs. In terms of efficiency, the GNN
adapter introduces only a few trainable parameters and can be trained with low
computation costs. The entire framework is trained using auto-regression on
node text (next token prediction). Once trained, GraphAdapter can be seamlessly
fine-tuned with task-specific prompts for various downstream tasks. Through
extensive experiments across multiple real-world TAGs, GraphAdapter based on
Llama 2 gains an average improvement of approximately 5\% in terms of node
classification. Furthermore, GraphAdapter can also adapt to other language
models, including RoBERTa, GPT-2. The promising results demonstrate that GNNs
can serve as effective adapters for LLMs in TAG modeling. | [
"cs.CL",
"cs.AI"
] | false |
2402.13006 | 2024-02-20T13:41:21Z | Investigating the Impact of Model Instability on Explanations and
Uncertainty | [
"Sara Vera Marjanović",
"Isabelle Augenstein",
"Christina Lioma"
] | Explainable AI methods facilitate the understanding of model behaviour, yet,
small, imperceptible perturbations to inputs can vastly distort explanations.
As these explanations are typically evaluated holistically, before model
deployment, it is difficult to assess when a particular explanation is
trustworthy. Some studies have tried to create confidence estimators for
explanations, but none have investigated an existing link between uncertainty
and explanation quality. We artificially simulate epistemic uncertainty in text
input by introducing noise at inference time. In this large-scale empirical
study, we insert different levels of noise perturbations and measure the effect
on the output of pre-trained language models and different uncertainty metrics.
Realistic perturbations have minimal effect on performance and explanations,
yet masking has a drastic effect. We find that high uncertainty doesn't
necessarily imply low explanation plausibility; the correlation between the two
metrics can be moderately positive when noise is exposed during the training
process. This suggests that noise-augmented models may be better at identifying
salient tokens when uncertain. Furthermore, when predictive and epistemic
uncertainty measures are over-confident, the robustness of a saliency map to
perturbation can indicate model stability issues. Integrated Gradients shows
the overall greatest robustness to perturbation, while still showing
model-specific patterns in performance; however, this phenomenon is limited to
smaller Transformer-based language models. | [
"cs.LG",
"cs.CL"
] | false |
2402.13022 | 2024-02-20T14:02:45Z | SoMeLVLM: A Large Vision Language Model for Social Media Processing | [
"Xinnong Zhang",
"Haoyu Kuang",
"Xinyi Mou",
"Hanjia Lyu",
"Kun Wu",
"Siming Chen",
"Jiebo Luo",
"Xuanjing Huang",
"Zhongyu Wei"
] | The growth of social media, characterized by its multimodal nature, has led
to the emergence of diverse phenomena and challenges, which calls for an
effective approach to uniformly solve automated tasks. The powerful Large
Vision Language Models make it possible to handle a variety of tasks
simultaneously, but even with carefully designed prompting methods, the general
domain models often fall short in aligning with the unique speaking style and
context of social media tasks. In this paper, we introduce a Large Vision
Language Model for Social Media Processing (SoMeLVLM), which is a cognitive
framework equipped with five key capabilities including knowledge &
comprehension, application, analysis, evaluation, and creation. SoMeLVLM is
designed to understand and generate realistic social media behavior. We have
developed a 654k multimodal social media instruction-tuning dataset to support
our cognitive framework and fine-tune our model. Our experiments demonstrate
that SoMeLVLM achieves state-of-the-art performance in multiple social media
tasks. Further analysis shows its significant advantages over baselines in
terms of cognitive abilities. | [
"cs.CL",
"cs.MM"
] | false |
2402.13025 | 2024-02-20T14:08:24Z | CFEVER: A Chinese Fact Extraction and VERification Dataset | [
"Ying-Jia Lin",
"Chun-Yi Lin",
"Chia-Jen Yeh",
"Yi-Ting Li",
"Yun-Yu Hu",
"Chih-Hao Hsu",
"Mei-Feng Lee",
"Hung-Yu Kao"
] | We present CFEVER, a Chinese dataset designed for Fact Extraction and
VERification. CFEVER comprises 30,012 manually created claims based on content
in Chinese Wikipedia. Each claim in CFEVER is labeled as "Supports", "Refutes",
or "Not Enough Info" to depict its degree of factualness. Similar to the FEVER
dataset, claims in the "Supports" and "Refutes" categories are also annotated
with corresponding evidence sentences sourced from single or multiple pages in
Chinese Wikipedia. Our labeled dataset holds a Fleiss' kappa value of 0.7934
for five-way inter-annotator agreement. In addition, through the experiments
with the state-of-the-art approaches developed on the FEVER dataset and a
simple baseline for CFEVER, we demonstrate that our dataset is a new rigorous
benchmark for factual extraction and verification, which can be further used
for developing automated systems to alleviate human fact-checking efforts.
CFEVER is available at https://ikmlab.github.io/CFEVER. | [
"cs.CL",
"cs.AI"
] | false |
2402.13028 | 2024-02-20T14:10:40Z | Heterogeneous Graph Reasoning for Fact Checking over Texts and Tables | [
"Haisong Gong",
"Weizhi Xu",
"Shu wu",
"Qiang Liu",
"Liang Wang"
] | Fact checking aims to predict claim veracity by reasoning over multiple
evidence pieces. It usually involves evidence retrieval and veracity reasoning.
In this paper, we focus on the latter, reasoning over unstructured text and
structured table information. Previous works have primarily relied on
fine-tuning pretrained language models or training homogeneous-graph-based
models. Despite their effectiveness, we argue that they fail to explore the
rich semantic information underlying the evidence with different structures. To
address this, we propose a novel word-level Heterogeneous-graph-based model for
Fact Checking over unstructured and structured information, namely HeterFC. Our
approach leverages a heterogeneous evidence graph, with words as nodes and
thoughtfully designed edges representing different evidence properties. We
perform information propagation via a relational graph neural network,
facilitating interactions between claims and evidence. An attention-based
method is utilized to integrate information, combined with a language model for
generating predictions. We introduce a multitask loss function to account for
potential inaccuracies in evidence retrieval. Comprehensive experiments on the
large fact checking dataset FEVEROUS demonstrate the effectiveness of HeterFC.
Code will be released at: https://github.com/Deno-V/HeterFC. | [
"cs.CL",
"cs.AI"
] | false |
2402.13055 | 2024-02-20T14:43:39Z | Identifying Semantic Induction Heads to Understand In-Context Learning | [
"Jie Ren",
"Qipeng Guo",
"Hang Yan",
"Dongrui Liu",
"Xipeng Qiu",
"Dahua Lin"
] | Although large language models (LLMs) have demonstrated remarkable
performance, the lack of transparency in their inference logic raises concerns
about their trustworthiness. To gain a better understanding of LLMs, we conduct
a detailed analysis of the operations of attention heads and aim to better
understand the in-context learning of LLMs. Specifically, we investigate
whether attention heads encode two types of relationships between tokens
present in natural languages: the syntactic dependency parsed from sentences
and the relation within knowledge graphs. We find that certain attention heads
exhibit a pattern where, when attending to head tokens, they recall tail tokens
and increase the output logits of those tail tokens. More crucially, the
formulation of such semantic induction heads has a close correlation with the
emergence of the in-context learning ability of language models. The study of
semantic attention heads advances our understanding of the intricate operations
of attention heads in transformers, and further provides new insights into the
in-context learning of LLMs. | [
"cs.CL",
"cs.AI"
] | false |
2402.13093 | 2024-02-20T15:36:41Z | Event-level Knowledge Editing | [
"Hao Peng",
"Xiaozhi Wang",
"Chunyang Li",
"Kaisheng Zeng",
"Jiangshan Duo",
"Yixin Cao",
"Lei Hou",
"Juanzi Li"
] | Knowledge editing aims at updating knowledge of large language models (LLMs)
to prevent them from becoming outdated. Existing work edits LLMs at the level
of factual knowledge triplets. However, natural knowledge updates in the real
world come from the occurrences of new events rather than direct changes in
factual triplets. In this paper, we propose a new task setting: event-level
knowledge editing, which directly edits new events into LLMs and improves over
conventional triplet-level editing on (1) Efficiency. A single event edit leads
to updates in multiple entailed knowledge triplets. (2) Completeness. Beyond
updating factual knowledge, event-level editing also requires considering the
event influences and updating LLMs' knowledge about future trends. We construct
a high-quality event-level editing benchmark ELKEN, consisting of 1,515 event
edits, 6,449 questions about factual knowledge, and 10,150 questions about
future tendencies. We systematically evaluate the performance of various
knowledge editing methods and LLMs on this benchmark. We find that ELKEN poses
significant challenges to existing knowledge editing approaches. Our codes and
dataset are publicly released to facilitate further research. | [
"cs.CL",
"cs.AI"
] | false |
2402.13094 | 2024-02-20T15:37:08Z | Digital Comprehensibility Assessment of Simplified Texts among Persons
with Intellectual Disabilities | [
"Andreas Säuberli",
"Franz Holzknecht",
"Patrick Haller",
"Silvana Deilen",
"Laura Schiffl",
"Silvia Hansen-Schirra",
"Sarah Ebling"
] | Text simplification refers to the process of increasing the comprehensibility
of texts. Automatic text simplification models are most commonly evaluated by
experts or crowdworkers instead of the primary target groups of simplified
texts, such as persons with intellectual disabilities. We conducted an
evaluation study of text comprehensibility including participants with and
without intellectual disabilities reading unsimplified, automatically and
manually simplified German texts on a tablet computer. We explored four
different approaches to measuring comprehensibility: multiple-choice
comprehension questions, perceived difficulty ratings, response time, and
reading speed. The results revealed significant variations in these
measurements, depending on the reader group and whether the text had undergone
automatic or manual simplification. For the target group of persons with
intellectual disabilities, comprehension questions emerged as the most reliable
measure, while analyzing reading speed provided valuable insights into
participants' reading behavior. | [
"cs.CL",
"cs.HC"
] | false |
2402.13098 | 2024-02-20T15:47:59Z | ELAD: Explanation-Guided Large Language Models Active Distillation | [
"Yifei Zhang",
"Bo Pan",
"Chen Ling",
"Yuntong Hu",
"Liang Zhao"
] | The deployment and application of Large Language Models (LLMs) is hindered by
their memory inefficiency, computational demands, and the high costs of API
inferences. Traditional distillation methods, which transfer the capabilities
of LLMs to smaller models, often fail to determine whether the knowledge has
been sufficiently transferred, potentially resulting in high costs or
incomplete distillation. In this paper, we propose an Explanation-Guided LLMs
Active Distillation (ELAD) framework that employs an active learning strategy
to optimize the balance between annotation costs and model performance. To
improve efficient sample selection, we introduce an explanation-guided sample
selection method that identifies samples challenging its reasoning by
exploiting uncertainties in explanation steps. Additionally, we present a
customized LLM-annotated explanation revision technique where the teacher model
detects and corrects flaws in the student model's reasoning. Our experiments
across various reasoning datasets demonstrate that our framework significantly
enhances the efficiency of LLM knowledge distillation. | [
"cs.CL",
"cs.AI"
] | false |
2402.13109 | 2024-02-20T16:02:12Z | CIF-Bench: A Chinese Instruction-Following Benchmark for Evaluating the
Generalizability of Large Language Models | [
"Yizhi LI",
"Ge Zhang",
"Xingwei Qu",
"Jiali Li",
"Zhaoqun Li",
"Zekun Wang",
"Hao Li",
"Ruibin Yuan",
"Yinghao Ma",
"Kai Zhang",
"Wangchunshu Zhou",
"Yiming Liang",
"Lei Zhang",
"Lei Ma",
"Jiajun Zhang",
"Zuowen Li",
"Stephen W. Huang",
"Chenghua Lin",
"Wenhu Chen",
"Jie Fu"
] | The advancement of large language models (LLMs) has enhanced the ability to
generalize across a wide range of unseen natural language processing (NLP)
tasks through instruction-following. Yet, their effectiveness often diminishes
in low-resource languages like Chinese, exacerbated by biased evaluations from
data leakage, casting doubt on their true generalizability to new linguistic
territories. In response, we introduce the Chinese Instruction-Following
Benchmark (CIF-Bench), designed to evaluate the zero-shot generalizability of
LLMs to the Chinese language. CIF-Bench comprises 150 tasks and 15,000
input-output pairs, developed by native speakers to test complex reasoning and
Chinese cultural nuances across 20 categories. To mitigate evaluation bias, we
release only half of the dataset publicly, with the remainder kept private, and
introduce diversified instructions to minimize score variance, totaling 45,000
data instances. Our evaluation of 28 selected LLMs reveals a noticeable
performance gap, with the best model scoring only 52.9%, highlighting the
limitations of LLMs in less familiar language and task contexts. This work aims
to uncover the current limitations of LLMs in handling Chinese tasks, pushing
towards the development of more culturally informed and linguistically diverse
models with the released data and benchmark
(https://yizhilll.github.io/CIF-Bench/). | [
"cs.CL",
"cs.AI"
] | false |
2402.13125 | 2024-02-20T16:38:33Z | TreeEval: Benchmark-Free Evaluation of Large Language Models through
Tree Planning | [
"Xiang Li",
"Yunshi Lan",
"Chao Yang"
] | Recently, numerous new benchmarks have been established to evaluate the
performance of large language models (LLMs) via either computing a holistic
score or employing another LLM as a judge. However, these approaches suffer
from data leakage due to the open access of the benchmark and inflexible
evaluation process. To address this issue, we introduce $\textbf{TreeEval}$, a
benchmark-free evaluation method for LLMs that let a high-performance LLM host
an irreproducible evaluation session and essentially avoids the data leakage.
Moreover, this LLM performs as an examiner to raise up a series of questions
under a topic with a tree planing strategy, which considers the current
evaluation status to decide the next question generation and ensures the
completeness and efficiency of the evaluation process. We evaluate $6$ models
of different parameter sizes, including $7$B, $13$B, and $33$B, and ultimately
achieved the highest correlation coefficient with AlpacaEval2.0 using only
around $45$ questions. We also conduct more analysis to show the robustness and
reliability of TreeEval. Our code can be accessed via the provided
https://github.com/Ashura5/TreeEval. | [
"cs.CL",
"cs.AI"
] | false |
2402.13208 | 2024-02-20T18:19:08Z | How do Hyenas deal with Human Speech? Speech Recognition and Translation
with ConfHyena | [
"Marco Gaido",
"Sara Papi",
"Matteo Negri",
"Luisa Bentivogli"
] | The attention mechanism, a cornerstone of state-of-the-art neural models,
faces computational hurdles in processing long sequences due to its quadratic
complexity. Consequently, research efforts in the last few years focused on
finding more efficient alternatives. Among them, Hyena (Poli et al., 2023)
stands out for achieving competitive results in both language modeling and
image classification, while offering sub-quadratic memory and computational
complexity. Building on these promising results, we propose ConfHyena, a
Conformer whose encoder self-attentions are replaced with an adaptation of
Hyena for speech processing, where the long input sequences cause high
computational costs. Through experiments in automatic speech recognition (for
English) and translation (from English into 8 target languages), we show that
our best ConfHyena model significantly reduces the training time by 27%, at the
cost of minimal quality degradation (~1%), which, in most cases, is not
statistically significant. | [
"cs.CL",
"cs.AI"
] | false |
2402.13225 | 2024-02-20T18:37:19Z | AgentMD: Empowering Language Agents for Risk Prediction with Large-Scale
Clinical Tool Learning | [
"Qiao Jin",
"Zhizheng Wang",
"Yifan Yang",
"Qingqing Zhu",
"Donald Wright",
"Thomas Huang",
"W John Wilbur",
"Zhe He",
"Andrew Taylor",
"Qingyu Chen",
"Zhiyong Lu"
] | Clinical calculators play a vital role in healthcare by offering accurate
evidence-based predictions for various purposes such as prognosis.
Nevertheless, their widespread utilization is frequently hindered by usability
challenges, poor dissemination, and restricted functionality. Augmenting large
language models with extensive collections of clinical calculators presents an
opportunity to overcome these obstacles and improve workflow efficiency, but
the scalability of the manual curation process poses a significant challenge.
In response, we introduce AgentMD, a novel language agent capable of curating
and applying clinical calculators across various clinical contexts. Using the
published literature, AgentMD has automatically curated a collection of 2,164
diverse clinical calculators with executable functions and structured
documentation, collectively named RiskCalcs. Manual evaluations show that
RiskCalcs tools achieve an accuracy of over 80% on three quality metrics. At
inference time, AgentMD can automatically select and apply the relevant
RiskCalcs tools given any patient description. On the newly established RiskQA
benchmark, AgentMD significantly outperforms chain-of-thought prompting with
GPT-4 (87.7% vs. 40.9% in accuracy). Additionally, we also applied AgentMD to
real-world clinical notes for analyzing both population-level and risk-level
patient characteristics. In summary, our study illustrates the utility of
language agents augmented with clinical calculators for healthcare analytics
and patient care. | [
"cs.CL",
"cs.AI"
] | false |
2402.13231 | 2024-02-20T18:47:28Z | Investigating Cultural Alignment of Large Language Models | [
"Badr AlKhamissi",
"Muhammad ElNokrashy",
"Mai AlKhamissi",
"Mona Diab"
] | The intricate relationship between language and culture has long been a
subject of exploration within the realm of linguistic anthropology. Large
Language Models (LLMs), promoted as repositories of collective human knowledge,
raise a pivotal question: do these models genuinely encapsulate the diverse
knowledge adopted by different cultures? Our study reveals that these models
demonstrate greater cultural alignment along two dimensions -- firstly, when
prompted with the dominant language of a specific culture, and secondly, when
pretrained with a refined mixture of languages employed by that culture. We
quantify cultural alignment by simulating sociological surveys, comparing model
responses to those of actual survey participants as references. Specifically,
we replicate a survey conducted in various regions of Egypt and the United
States through prompting LLMs with different pretraining data mixtures in both
Arabic and English with the personas of the real respondents and the survey
questions. Further analysis reveals that misalignment becomes more pronounced
for underrepresented personas and for culturally sensitive topics, such as
those probing social values. Finally, we introduce Anthropological Prompting, a
novel method leveraging anthropological reasoning to enhance cultural
alignment. Our study emphasizes the necessity for a more balanced multilingual
pretraining dataset to better represent the diversity of human experience and
the plurality of different cultures with many implications on the topic of
cross-lingual transfer. | [
"cs.CL",
"cs.CY"
] | false |
2402.13234 | 2024-02-20T18:49:41Z | Unlocking Insights: Semantic Search in Jupyter Notebooks | [
"Lan Li",
"Jinpeng Lv"
] | Semantic search, a process aimed at delivering highly relevant search results
by comprehending the searcher's intent and the contextual meaning of terms
within a searchable dataspace, plays a pivotal role in information retrieval.
In this paper, we investigate the application of large language models to
enhance semantic search capabilities, specifically tailored for the domain of
Jupyter Notebooks. Our objective is to retrieve generated outputs, such as
figures or tables, associated functions and methods, and other pertinent
information.
We demonstrate a semantic search framework that achieves a comprehensive
semantic understanding of the entire notebook's contents, enabling it to
effectively handle various types of user queries. Key components of this
framework include:
1). A data preprocessor is designed to handle diverse types of cells within
Jupyter Notebooks, encompassing both markdown and code cells. 2). An innovative
methodology is devised to address token size limitations that arise with
code-type cells. We implement a finer-grained approach to data input,
transitioning from the cell level to the function level, effectively resolving
these issues. | [
"cs.IR",
"cs.CL"
] | false |
2402.13364 | 2024-02-20T20:42:02Z | A Simple but Effective Approach to Improve Structured Language Model
Output for Information Extraction | [
"Yinghao Li",
"Rampi Ramprasad",
"Chao Zhang"
] | Large language models (LLMs) have demonstrated impressive abilities in
generating unstructured natural language according to instructions. However,
their performance can be inconsistent when tasked with producing text that
adheres to specific structured formats, which is crucial in applications like
named entity recognition (NER) or relation extraction (RE). To address this
issue, this paper introduces an efficient method, G&O, to enhance their
structured text generation capabilities. It breaks the generation into a
two-step pipeline: initially, LLMs generate answers in natural language as
intermediate responses. Subsequently, LLMs are asked to organize the output
into the desired structure, using the intermediate responses as context. G&O
effectively separates the generation of content from the structuring process,
reducing the pressure of completing two orthogonal tasks simultaneously. Tested
on zero-shot NER and RE, the results indicate a significant improvement in LLM
performance with minimal additional efforts. This straightforward and adaptable
prompting technique can also be combined with other strategies, like
self-consistency, to further elevate LLM capabilities in various structured
text generation tasks. | [
"cs.CL",
"cs.IR"
] | false |
2402.13414 | 2024-02-20T22:50:41Z | Harnessing Large Language Models as Post-hoc Correctors | [
"Zhiqiang Zhong",
"Kuangyu Zhou",
"Davide Mottin"
] | As Machine Learning (ML) models grow in size and demand higher-quality
training data, the expenses associated with re-training and fine-tuning these
models are escalating rapidly. Inspired by recent impressive achievements of
Large Language Models (LLMs) in different fields, this paper delves into the
question: can LLMs efficiently improve an ML's performance at a minimal cost?
We show that, through our proposed training-free framework LlmCorr, an LLM can
work as a post-hoc corrector to propose corrections for the predictions of an
arbitrary ML model. In particular, we form a contextual knowledge database by
incorporating the dataset's label information and the ML model's predictions on
the validation dataset. Leveraging the in-context learning capability of LLMs,
we ask the LLM to summarise the instances in which the ML model makes mistakes
and the correlation between primary predictions and true labels. Following
this, the LLM can transfer its acquired knowledge to suggest corrections for
the ML model's predictions. Our experimental results on the challenging
molecular predictions show that LlmCorr improves the performance of a number of
models by up to 39%. | [
"cs.LG",
"cs.CL"
] | false |
2402.14850 | 2024-02-20T01:59:11Z | CHATATC: Large Language Model-Driven Conversational Agents for
Supporting Strategic Air Traffic Flow Management | [
"Sinan Abdulhak",
"Wayne Hubbard",
"Karthik Gopalakrishnan",
"Max Z. Li"
] | Generative artificial intelligence (AI) and large language models (LLMs) have
gained rapid popularity through publicly available tools such as ChatGPT. The
adoption of LLMs for personal and professional use is fueled by the natural
interactions between human users and computer applications such as ChatGPT,
along with powerful summarization and text generation capabilities. Given the
widespread use of such generative AI tools, in this work we investigate how
these tools can be deployed in a non-safety critical, strategic traffic flow
management setting. Specifically, we train an LLM, CHATATC, based on a large
historical data set of Ground Delay Program (GDP) issuances, spanning 2000-2023
and consisting of over 80,000 GDP implementations, revisions, and
cancellations. We test the query and response capabilities of CHATATC,
documenting successes (e.g., providing correct GDP rates, durations, and
reason) and shortcomings (e.g,. superlative questions). We also detail the
design of a graphical user interface for future users to interact and
collaborate with the CHATATC conversational agent. | [
"cs.CL",
"cs.AI"
] | false |
2402.14853 | 2024-02-20T05:58:05Z | NL2Formula: Generating Spreadsheet Formulas from Natural Language
Queries | [
"Wei Zhao",
"Zhitao Hou",
"Siyuan Wu",
"Yan Gao",
"Haoyu Dong",
"Yao Wan",
"Hongyu Zhang",
"Yulei Sui",
"Haidong Zhang"
] | Writing formulas on spreadsheets, such as Microsoft Excel and Google Sheets,
is a widespread practice among users performing data analysis. However,
crafting formulas on spreadsheets remains a tedious and error-prone task for
many end-users, particularly when dealing with complex operations. To alleviate
the burden associated with writing spreadsheet formulas, this paper introduces
a novel benchmark task called NL2Formula, with the aim to generate executable
formulas that are grounded on a spreadsheet table, given a Natural Language
(NL) query as input. To accomplish this, we construct a comprehensive dataset
consisting of 70,799 paired NL queries and corresponding spreadsheet formulas,
covering 21,670 tables and 37 types of formula functions. We realize the
NL2Formula task by providing a sequence-to-sequence baseline implementation
called fCoder. Experimental results validate the effectiveness of fCoder,
demonstrating its superior performance compared to the baseline models.
Furthermore, we also compare fCoder with an initial GPT-3.5 model (i.e.,
text-davinci-003). Lastly, through in-depth error analysis, we identify
potential challenges in the NL2Formula task and advocate for further
investigation. | [
"cs.CL",
"cs.AI"
] | false |
2402.14854 | 2024-02-20T06:18:02Z | A Dual-Prompting for Interpretable Mental Health Language Models | [
"Hyolim Jeon",
"Dongje Yoo",
"Daeun Lee",
"Sejung Son",
"Seungbae Kim",
"Jinyoung Han"
] | Despite the increasing demand for AI-based mental health monitoring tools,
their practical utility for clinicians is limited by the lack of
interpretability.The CLPsych 2024 Shared Task (Chim et al., 2024) aims to
enhance the interpretability of Large Language Models (LLMs), particularly in
mental health analysis, by providing evidence of suicidality through linguistic
content. We propose a dual-prompting approach: (i) Knowledge-aware evidence
extraction by leveraging the expert identity and a suicide dictionary with a
mental health-specific LLM; and (ii) Evidence summarization by employing an
LLM-based consistency evaluator. Comprehensive experiments demonstrate the
effectiveness of combining domain-specific information, revealing performance
improvements and the approach's potential to aid clinicians in assessing mental
state progression. | [
"cs.CL",
"cs.AI"
] | false |
2402.14855 | 2024-02-20T06:20:09Z | An LLM Maturity Model for Reliable and Transparent Text-to-Query | [
"Lei Yu",
"Abir Ray"
] | Recognizing the imperative to address the reliability and transparency issues
of Large Language Models (LLM), this work proposes an LLM maturity model
tailored for text-to-query applications. This maturity model seeks to fill the
existing void in evaluating LLMs in such applications by incorporating
dimensions beyond mere correctness or accuracy. Moreover, this work introduces
a real-world use case from the law enforcement domain and showcases QueryIQ, an
LLM-powered, domain-specific text-to-query assistant to expedite user workflows
and reveal hidden relationship in data. | [
"cs.CL",
"cs.AI"
] | false |
2402.14856 | 2024-02-20T12:58:14Z | Comparing Inferential Strategies of Humans and Large Language Models in
Deductive Reasoning | [
"Philipp Mondorf",
"Barbara Plank"
] | Deductive reasoning plays a pivotal role in the formulation of sound and
cohesive arguments. It allows individuals to draw conclusions that logically
follow, given the truth value of the information provided. Recent progress in
the domain of large language models (LLMs) has showcased their capability in
executing deductive reasoning tasks. Nonetheless, a significant portion of
research primarily assesses the accuracy of LLMs in solving such tasks, often
overlooking a deeper analysis of their reasoning behavior. In this study, we
draw upon principles from cognitive psychology to examine inferential
strategies employed by LLMs, through a detailed evaluation of their responses
to propositional logic problems. Our findings indicate that LLMs display
reasoning patterns akin to those observed in humans, including strategies like
$\textit{supposition following}$ or $\textit{chain construction}$. Moreover,
our research demonstrates that the architecture and scale of the model
significantly affect its preferred method of reasoning, with more advanced
models tending to adopt strategies more frequently than less sophisticated
ones. Importantly, we assert that a model's accuracy, that is the correctness
of its final conclusion, does not necessarily reflect the validity of its
reasoning process. This distinction underscores the necessity for more nuanced
evaluation procedures in the field. | [
"cs.CL",
"cs.AI"
] | false |
2402.14858 | 2024-02-20T20:52:57Z | ChatEL: Entity Linking with Chatbots | [
"Yifan Ding",
"Qingkai Zeng",
"Tim Weninger"
] | Entity Linking (EL) is an essential and challenging task in natural language
processing that seeks to link some text representing an entity within a
document or sentence with its corresponding entry in a dictionary or knowledge
base. Most existing approaches focus on creating elaborate contextual models
that look for clues the words surrounding the entity-text to help solve the
linking problem. Although these fine-tuned language models tend to work, they
can be unwieldy, difficult to train, and do not transfer well to other domains.
Fortunately, Large Language Models (LLMs) like GPT provide a highly-advanced
solution to the problems inherent in EL models, but simply naive prompts to
LLMs do not work well. In the present work, we define ChatEL, which is a
three-step framework to prompt LLMs to return accurate results. Overall the
ChatEL framework improves the average F1 performance across 10 datasets by more
than 2%. Finally, a thorough error analysis shows many instances with the
ground truth labels were actually incorrect, and the labels predicted by ChatEL
were actually correct. This indicates that the quantitative results presented
in this paper may be a conservative estimate of the actual performance. All
data and code are available as an open-source package on GitHub at
https://github.com/yifding/In_Context_EL. | [
"cs.CL",
"cs.AI"
] | false |
2402.12654 | 2024-02-20T02:04:38Z | OWSM-CTC: An Open Encoder-Only Speech Foundation Model for Speech
Recognition, Translation, and Language Identification | [
"Yifan Peng",
"Yui Sudo",
"Muhammad Shakeel",
"Shinji Watanabe"
] | There has been an increasing interest in large speech models that can perform
multiple speech processing tasks in a single model. Such models usually adopt
the encoder-decoder or decoder-only architecture due to their popularity and
good performance in many domains. However, autoregressive models can be slower
during inference compared to non-autoregressive models and also have potential
risks of hallucination. Though prior studies observed promising results of
non-autoregressive models for certain tasks at small scales, it remains unclear
if they can be scaled to speech-to-text generation in diverse languages and
tasks. Inspired by the Open Whisper-style Speech Model (OWSM) project, we
propose OWSM-CTC, a novel encoder-only speech foundation model based on
Connectionist Temporal Classification (CTC). It is trained on 180k hours of
public audio data for multilingual automatic speech recognition (ASR), speech
translation (ST), and language identification (LID). Compared to
encoder-decoder OWSM, our OWSM-CTC achieves competitive results on ASR and up
to 25% relative improvement on ST, while it is more robust and 3 to 4 times
faster for inference. OWSM-CTC also improves the long-form ASR result with 20x
speed-up. We will publicly release our codebase, pre-trained model, and
training logs to promote open science in speech foundation models. | [
"cs.CL",
"cs.SD",
"eess.AS"
] | false |
2402.12659 | 2024-02-20T02:16:16Z | The FinBen: An Holistic Financial Benchmark for Large Language Models | [
"Qianqian Xie",
"Weiguang Han",
"Zhengyu Chen",
"Ruoyu Xiang",
"Xiao Zhang",
"Yueru He",
"Mengxi Xiao",
"Dong Li",
"Yongfu Dai",
"Duanyu Feng",
"Yijing Xu",
"Haoqiang Kang",
"Ziyan Kuang",
"Chenhan Yuan",
"Kailai Yang",
"Zheheng Luo",
"Tianlin Zhang",
"Zhiwei Liu",
"Guojun Xiong",
"Zhiyang Deng",
"Yuechen Jiang",
"Zhiyuan Yao",
"Haohang Li",
"Yangyang Yu",
"Gang Hu",
"Jiajia Huang",
"Xiao-Yang Liu",
"Alejandro Lopez-Lira",
"Benyou Wang",
"Yanzhao Lai",
"Hao Wang",
"Min Peng",
"Sophia Ananiadou",
"Jimin Huang"
] | LLMs have transformed NLP and shown promise in various fields, yet their
potential in finance is underexplored due to a lack of thorough evaluations and
the complexity of financial tasks. This along with the rapid development of
LLMs, highlights the urgent need for a systematic financial evaluation
benchmark for LLMs. In this paper, we introduce FinBen, the first comprehensive
open-sourced evaluation benchmark, specifically designed to thoroughly assess
the capabilities of LLMs in the financial domain. FinBen encompasses 35
datasets across 23 financial tasks, organized into three spectrums of
difficulty inspired by the Cattell-Horn-Carroll theory, to evaluate LLMs'
cognitive abilities in inductive reasoning, associative memory, quantitative
reasoning, crystallized intelligence, and more. Our evaluation of 15
representative LLMs, including GPT-4, ChatGPT, and the latest Gemini, reveals
insights into their strengths and limitations within the financial domain. The
findings indicate that GPT-4 leads in quantification, extraction, numerical
reasoning, and stock trading, while Gemini shines in generation and
forecasting; however, both struggle with complex extraction and forecasting,
showing a clear need for targeted enhancements. Instruction tuning boosts
simple task performance but falls short in improving complex reasoning and
forecasting abilities. FinBen seeks to continuously evaluate LLMs in finance,
fostering AI development with regular updates of tasks and models. | [
"cs.CL",
"cs.AI",
"cs.CE"
] | true |
2402.12663 | 2024-02-20T02:23:15Z | SoftQE: Learned Representations of Queries Expanded by LLMs | [
"Varad Pimpalkhute",
"John Heyer",
"Xusen Yin",
"Sameer Gupta"
] | We investigate the integration of Large Language Models (LLMs) into query
encoders to improve dense retrieval without increasing latency and cost, by
circumventing the dependency on LLMs at inference time. SoftQE incorporates
knowledge from LLMs by mapping embeddings of input queries to those of the
LLM-expanded queries. While improvements over various strong baselines on
in-domain MS-MARCO metrics are marginal, SoftQE improves performance by 2.83
absolute percentage points on average on five out-of-domain BEIR tasks. | [
"cs.CL",
"cs.IR",
"cs.LG"
] | false |
2402.12738 | 2024-02-20T06:05:36Z | Can Large Language Models be Used to Provide Psychological Counselling?
An Analysis of GPT-4-Generated Responses Using Role-play Dialogues | [
"Michimasa Inaba",
"Mariko Ukiyo",
"Keiko Takamizo"
] | Mental health care poses an increasingly serious challenge to modern
societies. In this context, there has been a surge in research that utilizes
information technologies to address mental health problems, including those
aiming to develop counseling dialogue systems. However, there is a need for
more evaluations of the performance of counseling dialogue systems that use
large language models. For this study, we collected counseling dialogue data
via role-playing scenarios involving expert counselors, and the utterances were
annotated with the intentions of the counselors. To determine the feasibility
of a dialogue system in real-world counseling scenarios, third-party counselors
evaluated the appropriateness of responses from human counselors and those
generated by GPT-4 in identical contexts in role-play dialogue data. Analysis
of the evaluation results showed that the responses generated by GPT-4 were
competitive with those of human counselors. | [
"cs.CL",
"cs.AI",
"cs.HC"
] | false |
2402.12817 | 2024-02-20T08:38:19Z | On Sensitivity of Learning with Limited Labelled Data to the Effects of
Randomness: Impact of Interactions and Systematic Choices | [
"Branislav Pecher",
"Ivan Srba",
"Maria Bielikova"
] | While learning with limited labelled data can improve performance when the
labels are lacking, it is also sensitive to the effects of uncontrolled
randomness introduced by so-called randomness factors (e.g., varying order of
data). We propose a method to systematically investigate the effects of
randomness factors while taking the interactions between them into
consideration. To measure the true effects of an individual randomness factor,
our method mitigates the effects of other factors and observes how the
performance varies across multiple runs. Applying our method to multiple
randomness factors across in-context learning and fine-tuning approaches on 7
representative text classification tasks and meta-learning on 3 tasks, we show
that: 1) disregarding interactions between randomness factors in existing works
caused inconsistent findings due to incorrect attribution of the effects of
randomness factors, such as disproving the consistent sensitivity of in-context
learning to sample order even with random sample selection; and 2) besides
mutual interactions, the effects of randomness factors, especially sample
order, are also dependent on more systematic choices unexplored in existing
works, such as number of classes, samples per class or choice of prompt format. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2402.12819 | 2024-02-20T08:38:24Z | Fine-Tuning, Prompting, In-Context Learning and Instruction-Tuning: How
Many Labelled Samples Do We Need? | [
"Branislav Pecher",
"Ivan Srba",
"Maria Bielikova"
] | When solving a task with limited labelled data, researchers can either use a
general large language model without further update, or use the few examples to
tune a specialised smaller model. When enough labels are available, the
specialised models outperform the general ones on many NLP tasks. In this work,
we aim to investigate how many labelled samples are required for the
specialised models to achieve this superior performance, while taking the
results variance into consideration. Observing the behaviour of prompting,
in-context learning, fine-tuning and instruction-tuning, identifying their
break-even points when increasing number of labelled training samples across
three tasks of varying complexity, we find that the specialised models often
need only few samples ($100-1000$) to be on par or better than the general
ones. At the same time, the amount of required labelled data strongly depends
on the task complexity and results variance. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2402.12842 | 2024-02-20T09:10:08Z | PromptKD: Distilling Student-Friendly Knowledge for Generative Language
Models via Prompt Tuning | [
"Gyeongman Kim",
"Doohyuk Jang",
"Eunho Yang"
] | Recent advancements in large language models (LLMs) have raised concerns
about inference costs, increasing the need for research into model compression.
While knowledge distillation (KD) is a prominent method for this, research on
KD for generative language models like LLMs is relatively sparse, and the
approach of distilling student-friendly knowledge, which has shown promising
performance in KD for classification models, remains unexplored in generative
language models. To explore this approach, we propose PromptKD, a simple yet
effective method that utilizes prompt tuning - for the first time in KD - to
enable generative language models to transfer student-friendly knowledge.
Unlike previous works in classification that require fine-tuning the entire
teacher model for extracting student-friendly knowledge, PromptKD achieves
similar effects by adding a small number of prompt tokens and tuning only the
prompt with student guidance. Extensive experiments on instruction-following
datasets using the GPT-2 model family show that PromptKD achieves
state-of-the-art performance while adding only 0.0007% of the teacher's
parameters as prompts. Further analysis suggests that distilling
student-friendly knowledge alleviates exposure bias effectively throughout the
entire training process, leading to performance enhancements. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2402.12847 | 2024-02-20T09:20:32Z | Instruction-tuned Language Models are Better Knowledge Learners | [
"Zhengbao Jiang",
"Zhiqing Sun",
"Weijia Shi",
"Pedro Rodriguez",
"Chunting Zhou",
"Graham Neubig",
"Xi Victoria Lin",
"Wen-tau Yih",
"Srinivasan Iyer"
] | In order for large language model (LLM)-based assistants to effectively adapt
to evolving information needs, it must be possible to update their factual
knowledge through continued training on new data. The standard recipe for doing
so involves continued pre-training on new documents followed by
instruction-tuning on question-answer (QA) pairs. However, we find that LLMs
trained with this recipe struggle to answer questions, even though the
perplexity of documents is minimized. We found that QA pairs are generally
straightforward, while documents are more complex, weaving many factual
statements together in an intricate manner. Therefore, we hypothesize that it
is beneficial to expose LLMs to QA pairs before continued pre-training on
documents so that the process of encoding knowledge from complex documents
takes into account how this knowledge is accessed through questions. Based on
this, we propose pre-instruction-tuning (PIT), a method that instruction-tunes
on questions prior to training on documents. This contrasts with standard
instruction-tuning, which learns how to extract knowledge after training on
documents. Extensive experiments and ablation studies demonstrate that PIT
significantly enhances the ability of LLMs to absorb knowledge from new
documents, outperforming standard instruction-tuning by 17.8%. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | true |
2402.12865 | 2024-02-20T09:57:08Z | Backward Lens: Projecting Language Model Gradients into the Vocabulary
Space | [
"Shahar Katz",
"Yonatan Belinkov",
"Mor Geva",
"Lior Wolf"
] | Understanding how Transformer-based Language Models (LMs) learn and recall
information is a key goal of the deep learning community. Recent
interpretability methods project weights and hidden states obtained from the
forward pass to the models' vocabularies, helping to uncover how information
flows within LMs. In this work, we extend this methodology to LMs' backward
pass and gradients. We first prove that a gradient matrix can be cast as a
low-rank linear combination of its forward and backward passes' inputs. We then
develop methods to project these gradients into vocabulary items and explore
the mechanics of how new information is stored in the LMs' neurons. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2402.12991 | 2024-02-20T13:20:39Z | TRAP: Targeted Random Adversarial Prompt Honeypot for Black-Box
Identification | [
"Martin Gubri",
"Dennis Ulmer",
"Hwaran Lee",
"Sangdoo Yun",
"Seong Joon Oh"
] | Large Language Model (LLM) services and models often come with legal rules on
who can use them and how they must use them. Assessing the compliance of the
released LLMs is crucial, as these rules protect the interests of the LLM
contributor and prevent misuse. In this context, we describe the novel problem
of Black-box Identity Verification (BBIV). The goal is to determine whether a
third-party application uses a certain LLM through its chat function. We
propose a method called Targeted Random Adversarial Prompt (TRAP) that
identifies the specific LLM in use. We repurpose adversarial suffixes,
originally proposed for jailbreaking, to get a pre-defined answer from the
target LLM, while other models give random answers. TRAP detects the target
LLMs with over 95% true positive rate at under 0.2% false positive rate even
after a single interaction. TRAP remains effective even if the LLM has minor
changes that do not significantly alter the original function. | [
"cs.LG",
"cs.AI",
"cs.CL",
"cs.CR"
] | false |
2402.13089 | 2024-02-20T15:31:44Z | Towards an empirical understanding of MoE design choices | [
"Dongyang Fan",
"Bettina Messmer",
"Martin Jaggi"
] | In this study, we systematically evaluate the impact of common design choices
in Mixture of Experts (MoEs) on validation performance, uncovering distinct
influences at token and sequence levels. We also present empirical evidence
showing comparable performance between a learned router and a frozen, randomly
initialized router, suggesting that learned routing may not be essential. Our
study further reveals that Sequence-level routing can result in topic-specific
weak expert specialization, in contrast to syntax specialization observed with
Token-level routing. | [
"cs.LG",
"cs.AI",
"cs.CL"
] | false |
2402.13212 | 2024-02-20T18:22:38Z | Soft Self-Consistency Improves Language Model Agents | [
"Han Wang",
"Archiki Prasad",
"Elias Stengel-Eskin",
"Mohit Bansal"
] | Generations from large language models (LLMs) can be improved by sampling and
scoring multiple solutions to select a final answer. Current "sample and
select" methods such as self-consistency (SC) rely on majority voting to score
answers. However, when tasks have many distinct and valid answers, selection by
voting requires a large number of samples. This makes SC prohibitively
expensive for interactive tasks that involve generating multiple actions
(answers) sequentially. After establishing that majority voting fails to
provide consistent gains on such tasks, we demonstrate how to increase success
rates by softening the scoring criterion. We introduce Soft Self-Consistency
(Soft-SC), which replaces SC's discontinuous scoring with a continuous score
computed from model likelihoods, allowing for selection even when actions are
sparsely distributed. Soft-SC improves both performance and efficiency on
long-horizon interactive tasks, requiring half as many samples as SC for
comparable or better performance. For a fixed number of samples, Soft-SC leads
to a 1.3% increase over SC in absolute success rate on writing bash programs, a
6.6% increase on online shopping (WebShop), and a 4.7% increase for an
interactive household game (ALFWorld). Finally, we show that Soft-SC can be
applied to both open-source and black-box models. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2402.13213 | 2024-02-20T18:24:47Z | Softmax Probabilities (Mostly) Predict Large Language Model Correctness
on Multiple-Choice Q&A | [
"Benjamin Plaut",
"Khanh Nguyen",
"Tu Trinh"
] | Although large language models (LLMs) perform impressively on many tasks,
overconfidence remains a problem. We hypothesized that on multiple-choice Q&A
tasks, wrong answers would be associated with smaller maximum softmax
probabilities (MSPs) compared to correct answers. We comprehensively evaluate
this hypothesis on ten open-source LLMs and five datasets, and find strong
evidence for our hypothesis among models which perform well on the original Q&A
task. For the six LLMs with the best Q&A performance, the AUROC derived from
the MSP was better than random chance with p < 10^{-4} in 59/60 instances.
Among those six LLMs, the average AUROC ranged from 60% to 69%. Leveraging
these findings, we propose a multiple-choice Q&A task with an option to abstain
and show that performance can be improved by selectively abstaining based on
the MSP of the initial model response. We also run the same experiments with
pre-softmax logits instead of softmax probabilities and find similar (but not
identical) results. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2402.13228 | 2024-02-20T18:42:34Z | Smaug: Fixing Failure Modes of Preference Optimisation with DPO-Positive | [
"Arka Pal",
"Deep Karkhanis",
"Samuel Dooley",
"Manley Roberts",
"Siddartha Naidu",
"Colin White"
] | Direct Preference Optimisation (DPO) is effective at significantly improving
the performance of large language models (LLMs) on downstream tasks such as
reasoning, summarisation, and alignment. Using pairs of preferred and
dispreferred data, DPO models the \textit{relative} probability of picking one
response over another. In this work, first we show theoretically that the
standard DPO loss can lead to a \textit{reduction} of the model's likelihood of
the preferred examples, as long as the relative probability between the
preferred and dispreferred classes increases. We then show empirically that
this phenomenon occurs when fine-tuning LLMs on common datasets, especially
datasets in which the edit distance between pairs of completions is low. Using
these insights, we design DPO-Positive (DPOP), a new loss function and training
procedure which avoids this failure mode. Surprisingly, we also find that DPOP
significantly outperforms DPO across a wide variety of datasets and downstream
tasks, including datasets with high edit distances between completions. By
fine-tuning with DPOP, we create and release Smaug-34B and Smaug-72B, which
achieve state-of-the-art open-source performance. Notably, Smaug-72B is nearly
2\% better than any other open-source model on the HuggingFace Open LLM
Leaderboard and becomes the first open-source LLM to surpass an average
accuracy of 80\%. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2402.13432 | 2024-02-20T23:54:02Z | DrBenchmark: A Large Language Understanding Evaluation Benchmark for
French Biomedical Domain | [
"Yanis Labrak",
"Adrien Bazoge",
"Oumaima El Khettari",
"Mickael Rouvier",
"Pacome Constant dit Beaufils",
"Natalia Grabar",
"Beatrice Daille",
"Solen Quiniou",
"Emmanuel Morin",
"Pierre-Antoine Gourraud",
"Richard Dufour"
] | The biomedical domain has sparked a significant interest in the field of
Natural Language Processing (NLP), which has seen substantial advancements with
pre-trained language models (PLMs). However, comparing these models has proven
challenging due to variations in evaluation protocols across different models.
A fair solution is to aggregate diverse downstream tasks into a benchmark,
allowing for the assessment of intrinsic PLMs qualities from various
perspectives. Although still limited to few languages, this initiative has been
undertaken in the biomedical field, notably English and Chinese. This
limitation hampers the evaluation of the latest French biomedical models, as
they are either assessed on a minimal number of tasks with non-standardized
protocols or evaluated using general downstream tasks. To bridge this research
gap and account for the unique sensitivities of French, we present the
first-ever publicly available French biomedical language understanding
benchmark called DrBenchmark. It encompasses 20 diversified tasks, including
named-entity recognition, part-of-speech tagging, question-answering, semantic
textual similarity, and classification. We evaluate 8 state-of-the-art
pre-trained masked language models (MLMs) on general and biomedical-specific
data, as well as English specific MLMs to assess their cross-lingual
capabilities. Our experiments reveal that no single model excels across all
tasks, while generalist models are sometimes still competitive. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2402.14851 | 2024-02-20T03:57:55Z | SQL-CRAFT: Text-to-SQL through Interactive Refinement and Enhanced
Reasoning | [
"Hanchen Xia",
"Feng Jiang",
"Naihao Deng",
"Cunxiang Wang",
"Guojiang Zhao",
"Rada Mihalcea",
"Yue Zhang"
] | Modern LLMs have become increasingly powerful, but they are still facing
challenges in specialized tasks such as Text-to-SQL. We propose SQL-CRAFT, a
framework to advance LLMs' SQL generation Capabilities through inteRActive
reFinemenT and enhanced reasoning. We leverage an Interactive Correction Loop
(IC-Loop) for LLMs to interact with databases automatically, as well as
Python-enhanced reasoning. We conduct experiments on two Text-to-SQL datasets,
Spider and Bird, with performance improvements of up to 5.7% compared to the
naive prompting method. Moreover, our method surpasses the current
state-of-the-art on the Spider Leaderboard, demonstrating the effectiveness of
our framework. | [
"cs.CL",
"cs.AI",
"cs.DB"
] | false |
2402.14852 | 2024-02-20T04:17:21Z | HumanEval on Latest GPT Models -- 2024 | [
"Daniel Li",
"Lincoln Murr"
] | In 2023, we are using the latest models of GPT-4 to advance program
synthesis. The large language models have significantly improved the
state-of-the-art for this purpose. To make these advancements more accessible,
we have created a repository that connects these models to Huamn Eval. This
dataset was initally developed to be used with a language model called CODEGEN
on natural and programming language data. The utility of these trained models
is showcased by demonstrating their competitive performance in zero-shot Python
code generation on HumanEval tasks compared to previous state-of-the-art
solutions. Additionally, this gives way to developing more multi-step paradigm
synthesis. This benchmark features 160 diverse problem sets factorized into
multistep prompts that our analysis shows significantly improves program
synthesis over single-turn inputs. All code is open source at
https://github.com/daniel442li/gpt-human-eval . | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2402.14857 | 2024-02-20T17:39:40Z | Is the System Message Really Important to Jailbreaks in Large Language
Models? | [
"Xiaotian Zou",
"Yongkang Chen",
"Ke Li"
] | The rapid evolution of Large Language Models (LLMs) has rendered them
indispensable in modern society. While security measures are typically in place
to align LLMs with human values prior to release, recent studies have unveiled
a concerning phenomenon named "jailbreak." This term refers to the unexpected
and potentially harmful responses generated by LLMs when prompted with
malicious questions. Existing research focuses on generating jailbreak prompts
but our study aim to answer a different question: Is the system message really
important to jailbreak in LLMs? To address this question, we conducted
experiments in a stable GPT version gpt-3.5-turbo-0613 to generated jailbreak
prompts with varying system messages: short, long, and none. We discover that
different system messages have distinct resistances to jailbreak by
experiments. Additionally, we explore the transferability of jailbreak across
LLMs. This finding underscores the significant impact system messages can have
on mitigating LLMs jailbreak. To generate system messages that are more
resistant to jailbreak prompts, we propose System Messages Evolutionary
Algorithms (SMEA). Through SMEA, we can get robust system messages population
that demonstrate up to 98.9% resistance against jailbreak prompts. Our research
not only bolsters LLMs security but also raises the bar for jailbreak,
fostering advancements in this field of study. | [
"cs.CL",
"cs.AI",
"cs.CR"
] | false |
2403.00788 | 2024-02-20T04:26:31Z | PRECISE Framework: GPT-based Text For Improved Readability, Reliability,
and Understandability of Radiology Reports For Patient-Centered Care | [
"Satvik Tripathi",
"Liam Mutter",
"Meghana Muppuri",
"Suhani Dheer",
"Emiliano Garza-Frias",
"Komal Awan",
"Aakash Jha",
"Michael Dezube",
"Azadeh Tabari",
"Christopher P. Bridge",
"Dania Daye"
] | This study introduces and evaluates the PRECISE framework, utilizing OpenAI's
GPT-4 to enhance patient engagement by providing clearer and more accessible
chest X-ray reports at a sixth-grade reading level. The framework was tested on
500 reports, demonstrating significant improvements in readability,
reliability, and understandability. Statistical analyses confirmed the
effectiveness of the PRECISE approach, highlighting its potential to foster
patient-centric care delivery in healthcare decision-making. | [
"cs.CL",
"cs.AI",
"cs.HC",
"cs.LG"
] | false |
2403.08819 | 2024-02-20T04:13:48Z | Thermometer: Towards Universal Calibration for Large Language Models | [
"Maohao Shen",
"Subhro Das",
"Kristjan Greenewald",
"Prasanna Sattigeri",
"Gregory Wornell",
"Soumya Ghosh"
] | We consider the issue of calibration in large language models (LLM). Recent
studies have found that common interventions such as instruction tuning often
result in poorly calibrated LLMs. Although calibration is well-explored in
traditional applications, calibrating LLMs is uniquely challenging. These
challenges stem as much from the severe computational requirements of LLMs as
from their versatility, which allows them to be applied to diverse tasks.
Addressing these challenges, we propose THERMOMETER, a calibration approach
tailored to LLMs. THERMOMETER learns an auxiliary model, given data from
multiple tasks, for calibrating a LLM. It is computationally efficient,
preserves the accuracy of the LLM, and produces better-calibrated responses for
new tasks. Extensive empirical evaluations across various benchmarks
demonstrate the effectiveness of the proposed method. | [
"cs.LG",
"cs.CL",
"stat.ML"
] | false |
2402.12617 | 2024-02-20T00:51:05Z | Generative AI Security: Challenges and Countermeasures | [
"Banghua Zhu",
"Norman Mu",
"Jiantao Jiao",
"David Wagner"
] | Generative AI's expanding footprint across numerous industries has led to
both excitement and increased scrutiny. This paper delves into the unique
security challenges posed by Generative AI, and outlines potential research
directions for managing these risks. | [
"cs.CR",
"cs.AI",
"cs.CL",
"cs.CY",
"cs.LG"
] | false |
2402.13040 | 2024-02-20T14:29:02Z | Text-Guided Molecule Generation with Diffusion Language Model | [
"Haisong Gong",
"Qiang Liu",
"Shu Wu",
"Liang Wang"
] | Text-guided molecule generation is a task where molecules are generated to
match specific textual descriptions. Recently, most existing SMILES-based
molecule generation methods rely on an autoregressive architecture. In this
work, we propose the Text-Guided Molecule Generation with Diffusion Language
Model (TGM-DLM), a novel approach that leverages diffusion models to address
the limitations of autoregressive methods. TGM-DLM updates token embeddings
within the SMILES string collectively and iteratively, using a two-phase
diffusion generation process. The first phase optimizes embeddings from random
noise, guided by the text description, while the second phase corrects invalid
SMILES strings to form valid molecular representations. We demonstrate that
TGM-DLM outperforms MolT5-Base, an autoregressive model, without the need for
additional data resources. Our findings underscore the remarkable effectiveness
of TGM-DLM in generating coherent and precise molecules with specific
properties, opening new avenues in drug discovery and related scientific
domains. Code will be released at: https://github.com/Deno-V/tgm-dlm. | [
"cs.LG",
"cs.AI",
"cs.CE",
"cs.CL",
"q-bio.BM"
] | false |
2402.12613 | 2024-02-20T00:34:58Z | Analysis of Using Sigmoid Loss for Contrastive Learning | [
"Chungpa Lee",
"Joonhwan Chang",
"Jy-yong Sohn"
] | Contrastive learning has emerged as a prominent branch of self-supervised
learning for several years. Especially, CLIP, which applies contrastive
learning to large sets of captioned images, has garnered significant attention.
Recently, SigLIP, a variant of CLIP, has been proposed, which uses the sigmoid
loss instead of the standard InfoNCE loss. SigLIP achieves the performance
comparable to CLIP in a more efficient manner by eliminating the need for a
global view. However, theoretical understanding of using the sigmoid loss in
contrastive learning is underexplored. In this paper, we provide a theoretical
analysis of using the sigmoid loss in contrastive learning, in the perspective
of the geometric structure of learned embeddings. First, we propose the
double-Constant Embedding Model (CCEM), a framework for parameterizing various
well-known embedding structures by a single variable. Interestingly, the
proposed CCEM is proven to contain the optimal embedding with respect to the
sigmoid loss. Second, we mathematically analyze the optimal embedding
minimizing the sigmoid loss for contrastive learning. The optimal embedding
ranges from simplex equiangular-tight-frame to antipodal structure, depending
on the temperature parameter used in the sigmoid loss. Third, our experimental
results on synthetic datasets coincide with the theoretical results on the
optimal embedding structures. | [
"cs.LG"
] | false |
2402.12664 | 2024-02-20T02:26:48Z | Discriminant Distance-Aware Representation on Deterministic Uncertainty
Quantification Methods | [
"Jiaxin Zhang",
"Kamalika Das",
"Sricharan Kumar"
] | Uncertainty estimation is a crucial aspect of deploying dependable deep
learning models in safety-critical systems. In this study, we introduce a novel
and efficient method for deterministic uncertainty estimation called
Discriminant Distance-Awareness Representation (DDAR). Our approach involves
constructing a DNN model that incorporates a set of prototypes in its latent
representations, enabling us to analyze valuable feature information from the
input data. By leveraging a distinction maximization layer over optimal
trainable prototypes, DDAR can learn a discriminant distance-awareness
representation. We demonstrate that DDAR overcomes feature collapse by relaxing
the Lipschitz constraint that hinders the practicality of deterministic
uncertainty methods (DUMs) architectures. Our experiments show that DDAR is a
flexible and architecture-agnostic method that can be easily integrated as a
pluggable layer with distance-sensitive metrics, outperforming state-of-the-art
uncertainty estimation methods on multiple benchmark problems. | [
"cs.LG"
] | false |
2402.12673 | 2024-02-20T02:45:20Z | Beyond Worst-case Attacks: Robust RL with Adaptive Defense via
Non-dominated Policies | [
"Xiangyu Liu",
"Chenghao Deng",
"Yanchao Sun",
"Yongyuan Liang",
"Furong Huang"
] | In light of the burgeoning success of reinforcement learning (RL) in diverse
real-world applications, considerable focus has been directed towards ensuring
RL policies are robust to adversarial attacks during test time. Current
approaches largely revolve around solving a minimax problem to prepare for
potential worst-case scenarios. While effective against strong attacks, these
methods often compromise performance in the absence of attacks or the presence
of only weak attacks. To address this, we study policy robustness under the
well-accepted state-adversarial attack model, extending our focus beyond only
worst-case attacks. We first formalize this task at test time as a regret
minimization problem and establish its intrinsic hardness in achieving
sublinear regret when the baseline policy is from a general continuous policy
class, $\Pi$. This finding prompts us to \textit{refine} the baseline policy
class $\Pi$ prior to test time, aiming for efficient adaptation within a finite
policy class $\Tilde{\Pi}$, which can resort to an adversarial bandit
subroutine. In light of the importance of a small, finite $\Tilde{\Pi}$, we
propose a novel training-time algorithm to iteratively discover
\textit{non-dominated policies}, forming a near-optimal and minimal
$\Tilde{\Pi}$, thereby ensuring both robustness and test-time efficiency.
Empirical validation on the Mujoco corroborates the superiority of our approach
in terms of natural and robust performance, as well as adaptability to various
attack scenarios. | [
"cs.LG"
] | false |
2402.12715 | 2024-02-20T04:49:34Z | Spurious Correlations in Machine Learning: A Survey | [
"Wenqian Ye",
"Guangtao Zheng",
"Xu Cao",
"Yunsheng Ma",
"Xia Hu",
"Aidong Zhang"
] | Machine learning systems are known to be sensitive to spurious correlations
between biased features of the inputs (e.g., background, texture, and secondary
objects) and the corresponding labels. These features and their correlations
with the labels are known as "spurious" because they tend to change with shifts
in real-world data distributions, which can negatively impact the model's
generalization and robustness. In this survey, we provide a comprehensive
review of this issue, along with a taxonomy of current state-of-the-art methods
for addressing spurious correlations in machine learning models. Additionally,
we summarize existing datasets, benchmarks, and metrics to aid future research.
The paper concludes with a discussion of the recent advancements and future
research challenges in this field, aiming to provide valuable insights for
researchers in the related domains. | [
"cs.LG"
] | false |
2402.12722 | 2024-02-20T05:11:20Z | Structural Knowledge Informed Continual Multivariate Time Series
Forecasting | [
"Zijie Pan",
"Yushan Jiang",
"Dongjin Song",
"Sahil Garg",
"Kashif Rasul",
"Anderson Schneider",
"Yuriy Nevmyvaka"
] | Recent studies in multivariate time series (MTS) forecasting reveal that
explicitly modeling the hidden dependencies among different time series can
yield promising forecasting performance and reliable explanations. However,
modeling variable dependencies remains underexplored when MTS is continuously
accumulated under different regimes (stages). Due to the potential distribution
and dependency disparities, the underlying model may encounter the catastrophic
forgetting problem, i.e., it is challenging to memorize and infer different
types of variable dependencies across different regimes while maintaining
forecasting performance. To address this issue, we propose a novel Structural
Knowledge Informed Continual Learning (SKI-CL) framework to perform MTS
forecasting within a continual learning paradigm, which leverages structural
knowledge to steer the forecasting model toward identifying and adapting to
different regimes, and selects representative MTS samples from each regime for
memory replay. Specifically, we develop a forecasting model based on graph
structure learning, where a consistency regularization scheme is imposed
between the learned variable dependencies and the structural knowledge while
optimizing the forecasting objective over the MTS data. As such, MTS
representations learned in each regime are associated with distinct structural
knowledge, which helps the model memorize a variety of conceivable scenarios
and results in accurate forecasts in the continual learning context. Meanwhile,
we develop a representation-matching memory replay scheme that maximizes the
temporal coverage of MTS data to efficiently preserve the underlying temporal
dynamics and dependency structures of each regime. Thorough empirical studies
on synthetic and real-world benchmarks validate SKI-CL's efficacy and
advantages over the state-of-the-art for continual MTS forecasting tasks. | [
"cs.LG"
] | false |
2402.12737 | 2024-02-20T06:04:44Z | Guarantee Regions for Local Explanations | [
"Marton Havasi",
"Sonali Parbhoo",
"Finale Doshi-Velez"
] | Interpretability methods that utilise local surrogate models (e.g. LIME) are
very good at describing the behaviour of the predictive model at a point of
interest, but they are not guaranteed to extrapolate to the local region
surrounding the point. However, overfitting to the local curvature of the
predictive model and malicious tampering can significantly limit extrapolation.
We propose an anchor-based algorithm for identifying regions in which local
explanations are guaranteed to be correct by explicitly describing those
intervals along which the input features can be trusted. Our method produces an
interpretable feature-aligned box where the prediction of the local surrogate
model is guaranteed to match the predictive model. We demonstrate that our
algorithm can be used to find explanations with larger guarantee regions that
better cover the data manifold compared to existing baselines. We also show how
our method can identify misleading local explanations with significantly poorer
guarantee regions. | [
"cs.LG"
] | false |
2402.12780 | 2024-02-20T07:40:11Z | Tackling Byzantine Clients in Federated Learning | [
"Youssef Allouah",
"Sadegh Farhadkhani",
"Rachid GuerraouI",
"Nirupam Gupta",
"Rafael Pinot",
"Geovani Rizk",
"Sasha Voitovych"
] | The possibility of adversarial (a.k.a., {\em Byzantine}) clients makes
federated learning (FL) prone to arbitrary manipulation. The natural approach
to robustify FL against adversarial clients is to replace the simple averaging
operation at the server in the standard $\mathsf{FedAvg}$ algorithm by a
\emph{robust averaging rule}. While a significant amount of work has been
devoted to studying the convergence of federated {\em robust averaging} (which
we denote by $\mathsf{FedRo}$), prior work has largely ignored the impact of
{\em client subsampling} and {\em local steps}, two fundamental FL
characteristics. While client subsampling increases the effective fraction of
Byzantine clients, local steps increase the drift between the local updates
computed by honest (i.e., non-Byzantine) clients. Consequently, a careless
deployment of $\mathsf{FedRo}$ could yield poor performance. We validate this
observation by presenting an in-depth analysis of $\mathsf{FedRo}$ tightly
analyzing the impact of client subsampling and local steps. Specifically, we
present a sufficient condition on client subsampling for nearly-optimal
convergence of $\mathsf{FedRo}$ (for smooth non-convex loss). Also, we show
that the rate of improvement in learning accuracy {\em diminishes} with respect
to the number of clients subsampled, as soon as the sample size exceeds a
threshold value. Interestingly, we also observe that under a careful choice of
step-sizes, the learning error due to Byzantine clients decreases with the
number of local steps. We validate our theory by experiments on the FEMNIST and
CIFAR-$10$ image classification tasks. | [
"cs.LG"
] | false |
2402.12790 | 2024-02-20T07:58:04Z | From Movements to Metrics: Evaluating Explainable AI Methods in
Skeleton-Based Human Activity Recognition | [
"Kimji N. Pellano",
"Inga Strümke",
"Espen Alexander F. Ihlen"
] | The advancement of deep learning in human activity recognition (HAR) using 3D
skeleton data is critical for applications in healthcare, security, sports, and
human-computer interaction. This paper tackles a well-known gap in the field,
which is the lack of testing in the applicability and reliability of XAI
evaluation metrics in the skeleton-based HAR domain. We have tested established
XAI metrics namely faithfulness and stability on Class Activation Mapping (CAM)
and Gradient-weighted Class Activation Mapping (Grad-CAM) to address this
problem. The study also introduces a perturbation method that respects human
biomechanical constraints to ensure realistic variations in human movement. Our
findings indicate that \textit{faithfulness} may not be a reliable metric in
certain contexts, such as with the EfficientGCN model. Conversely, stability
emerges as a more dependable metric when there is slight input data
perturbations. CAM and Grad-CAM are also found to produce almost identical
explanations, leading to very similar XAI metric performance. This calls for
the need for more diversified metrics and new XAI methods applied in
skeleton-based HAR. | [
"cs.LG"
] | false |
2402.12852 | 2024-02-20T09:31:03Z | CCFC++: Enhancing Federated Clustering through Feature Decorrelation | [
"Jie Yan",
"Jing Liu",
"Yi-Zi Ning",
"Zhong-Yuan Zhang"
] | In federated clustering, multiple data-holding clients collaboratively group
data without exchanging raw data. This field has seen notable advancements
through its marriage with contrastive learning, exemplified by
Cluster-Contrastive Federated Clustering (CCFC). However, CCFC suffers from
heterogeneous data across clients, leading to poor and unrobust performance.
Our study conducts both empirical and theoretical analyses to understand the
impact of heterogeneous data on CCFC. Findings indicate that increased data
heterogeneity exacerbates dimensional collapse in CCFC, evidenced by increased
correlations across multiple dimensions of the learned representations. To
address this, we introduce a decorrelation regularizer to CCFC. Benefiting from
the regularizer, the improved method effectively mitigates the detrimental
effects of data heterogeneity, and achieves superior performance, as evidenced
by a marked increase in NMI scores, with the gain reaching as high as 0.32 in
the most pronounced case. | [
"cs.LG"
] | false |
2402.12874 | 2024-02-20T10:09:00Z | Skill or Luck? Return Decomposition via Advantage Functions | [
"Hsiao-Ru Pan",
"Bernhard Schölkopf"
] | Learning from off-policy data is essential for sample-efficient reinforcement
learning. In the present work, we build on the insight that the advantage
function can be understood as the causal effect of an action on the return, and
show that this allows us to decompose the return of a trajectory into parts
caused by the agent's actions (skill) and parts outside of the agent's control
(luck). Furthermore, this decomposition enables us to naturally extend Direct
Advantage Estimation (DAE) to off-policy settings (Off-policy DAE). The
resulting method can learn from off-policy trajectories without relying on
importance sampling techniques or truncating off-policy actions. We draw
connections between Off-policy DAE and previous methods to demonstrate how it
can speed up learning and when the proposed off-policy corrections are
important. Finally, we use the MinAtar environments to illustrate how ignoring
off-policy corrections can lead to suboptimal policy optimization performance. | [
"cs.LG"
] | false |
2402.12930 | 2024-02-20T11:29:57Z | Learning Exceptional Subgroups by End-to-End Maximizing KL-divergence | [
"Sascha Xu",
"Nils Philipp Walter",
"Janis Kalofolias",
"Jilles Vreeken"
] | Finding and describing sub-populations that are exceptional regarding a
target property has important applications in many scientific disciplines, from
identifying disadvantaged demographic groups in census data to finding
conductive molecules within gold nanoparticles. Current approaches to finding
such subgroups require pre-discretized predictive variables, do not permit
non-trivial target distributions, do not scale to large datasets, and struggle
to find diverse results.
To address these limitations, we propose Syflow, an end-to-end optimizable
approach in which we leverage normalizing flows to model arbitrary target
distributions, and introduce a novel neural layer that results in easily
interpretable subgroup descriptions. We demonstrate on synthetic and real-world
data, including a case study, that Syflow reliably finds highly exceptional
subgroups accompanied by insightful descriptions. | [
"cs.LG"
] | false |
2402.12945 | 2024-02-20T12:00:25Z | Stochastic Approximation Approach to Federated Machine Learning | [
"Srihari P V",
"Bharath Bhikkaji"
] | This paper examines Federated learning (FL) in a Stochastic Approximation
(SA) framework. FL is a collaborative way to train neural network models across
various participants or clients without centralizing their data. Each client
will train a model on their respective data and send the weights across to a
the server periodically for aggregation. The server aggregates these weights
which are then used by the clients to re-initialize their neural network and
continue the training. SA is an iterative algorithm that uses approximate
sample gradients and tapering step size to locate a minimizer of a cost
function. In this paper the clients use a stochastic approximation iterate to
update the weights of its neural network. It is shown that the aggregated
weights track an autonomous ODE. Numerical simulations are performed and the
results are compared with standard algorithms like FedAvg and FedProx. It is
observed that the proposed algorithm is robust and gives more reliable
estimates of the weights, in particular when the clients data are not
identically distributed. | [
"cs.LG"
] | false |
2402.12987 | 2024-02-20T13:17:37Z | Towards Robust Graph Incremental Learning on Evolving Graphs | [
"Junwei Su",
"Difan Zou",
"Zijun Zhang",
"Chuan Wu"
] | Incremental learning is a machine learning approach that involves training a
model on a sequence of tasks, rather than all tasks at once. This ability to
learn incrementally from a stream of tasks is crucial for many real-world
applications. However, incremental learning is a challenging problem on
graph-structured data, as many graph-related problems involve prediction tasks
for each individual node, known as Node-wise Graph Incremental Learning (NGIL).
This introduces non-independent and non-identically distributed characteristics
in the sample data generation process, making it difficult to maintain the
performance of the model as new tasks are added. In this paper, we focus on the
inductive NGIL problem, which accounts for the evolution of graph structure
(structural shift) induced by emerging tasks. We provide a formal formulation
and analysis of the problem, and propose a novel regularization-based technique
called Structural-Shift-Risk-Mitigation (SSRM) to mitigate the impact of the
structural shift on catastrophic forgetting of the inductive NGIL problem. We
show that the structural shift can lead to a shift in the input distribution
for the existing tasks, and further lead to an increased risk of catastrophic
forgetting. Through comprehensive empirical studies with several benchmark
datasets, we demonstrate that our proposed method,
Structural-Shift-Risk-Mitigation (SSRM), is flexible and easy to adapt to
improve the performance of state-of-the-art GNN incremental learning frameworks
in the inductive setting. | [
"cs.LG"
] | false |
2402.13108 | 2024-02-20T16:01:42Z | On the Stability of Gradient Descent for Large Learning Rate | [
"Alexandru Crăciun",
"Debarghya Ghoshdastidar"
] | There currently is a significant interest in understanding the Edge of
Stability (EoS) phenomenon, which has been observed in neural networks
training, characterized by a non-monotonic decrease of the loss function over
epochs, while the sharpness of the loss (spectral norm of the Hessian)
progressively approaches and stabilizes around 2/(learning rate). Reasons for
the existence of EoS when training using gradient descent have recently been
proposed -- a lack of flat minima near the gradient descent trajectory together
with the presence of compact forward-invariant sets. In this paper, we show
that linear neural networks optimized under a quadratic loss function satisfy
the first assumption and also a necessary condition for the second assumption.
More precisely, we prove that the gradient descent map is non-singular, the set
of global minimizers of the loss function forms a smooth manifold, and the
stable minima form a bounded subset in parameter space. Additionally, we prove
that if the step-size is too big, then the set of initializations from which
gradient descent converges to a critical point has measure zero. | [
"cs.LG"
] | false |
2402.13196 | 2024-02-20T18:07:59Z | Practical Kernel Tests of Conditional Independence | [
"Roman Pogodin",
"Antonin Schrab",
"Yazhe Li",
"Danica J. Sutherland",
"Arthur Gretton"
] | We describe a data-efficient, kernel-based approach to statistical testing of
conditional independence. A major challenge of conditional independence
testing, absent in tests of unconditional independence, is to obtain the
correct test level (the specified upper bound on the rate of false positives),
while still attaining competitive test power. Excess false positives arise due
to bias in the test statistic, which is obtained using nonparametric kernel
ridge regression. We propose three methods for bias control to correct the test
level, based on data splitting, auxiliary data, and (where possible) simpler
function classes. We show these combined strategies are effective both for
synthetic and real-world data. | [
"cs.LG"
] | false |
2402.13210 | 2024-02-20T18:20:59Z | Bayesian Reward Models for LLM Alignment | [
"Adam X. Yang",
"Maxime Robeyns",
"Thomas Coste",
"Jun Wang",
"Haitham Bou-Ammar",
"Laurence Aitchison"
] | To ensure that large language model (LLM) responses are helpful and
non-toxic, we usually fine-tune a reward model on human preference data. We
then select policy responses with high rewards (best-of-n sampling) or further
optimize the policy to produce responses with high rewards (reinforcement
learning from human feedback). However, this process is vulnerable to reward
overoptimization or hacking, in which the responses selected have high rewards
due to errors in the reward model rather than a genuine preference. This is
especially problematic as the prompt or response diverges from the training
data. It should be possible to mitigate these issues by training a Bayesian
reward model, which signals higher uncertainty further from the training data
distribution. Therefore, we trained Bayesian reward models using Laplace-LoRA
(Yang et al., 2024) and found that the resulting uncertainty estimates can
successfully mitigate reward overoptimization in best-of-n sampling. | [
"cs.LG"
] | false |
2402.13371 | 2024-02-20T20:53:04Z | FIDLAR: Forecast-Informed Deep Learning Architecture for Flood
Mitigation | [
"Jimeng Shi",
"Zeda Yin",
"Arturo Leon",
"Jayantha Obeysekera",
"Giri Narasimhan"
] | In coastal river systems, frequent floods, often occurring during major
storms or king tides, pose a severe threat to lives and property. However,
these floods can be mitigated or even prevented by strategically releasing
water before extreme weather events with hydraulic structures such as dams,
gates, pumps, and reservoirs. A standard approach used by local water
management agencies is the "rule-based" method, which specifies predetermined
pre-releases of water based on historical and time-tested human experience, but
which tends to result in excess or inadequate water release. The model
predictive control (MPC), a physics-based model for prediction, is an
alternative approach, albeit involving computationally intensive calculations.
In this paper, we propose a Forecast Informed Deep Learning Architecture,
FIDLAR, to achieve rapid and optimal flood management with precise water
pre-releases. FIDLAR seamlessly integrates two neural network modules: one
called the Flood Manager, which is responsible for generating water pre-release
schedules, and another called the Flood Evaluator, which assesses these
generated schedules. The Evaluator module is pre-trained separately, and its
gradient-based feedback is used to train the Manager model, ensuring optimal
water pre-releases. We have conducted experiments using FIDLAR with data from a
flood-prone coastal area in South Florida, particularly susceptible to frequent
storms. Results show that FIDLAR is several orders of magnitude faster than
currently used physics-based approaches while outperforming baseline methods
with improved water pre-release schedules. Our code is at
https://github.com/JimengShi/FIDLAR/. | [
"cs.LG"
] | false |
2402.13402 | 2024-02-20T22:12:33Z | Towards accelerating physical discovery via non-interactive and
interactive multi-fidelity Bayesian Optimization: Current challenges and
future opportunities | [
"Arpan Biswas",
"Sai Mani Prudhvi Valleti",
"Rama Vasudevan",
"Maxim Ziatdinov",
"Sergei V. Kalinin"
] | Both computational and experimental material discovery bring forth the
challenge of exploring multidimensional and often non-differentiable parameter
spaces, such as phase diagrams of Hamiltonians with multiple interactions,
composition spaces of combinatorial libraries, processing spaces, and molecular
embedding spaces. Often these systems are expensive or time-consuming to
evaluate a single instance, and hence classical approaches based on exhaustive
grid or random search are too data intensive. This resulted in strong interest
towards active learning methods such as Bayesian optimization (BO) where the
adaptive exploration occurs based on human learning (discovery) objective.
However, classical BO is based on a predefined optimization target, and
policies balancing exploration and exploitation are purely data driven. In
practical settings, the domain expert can pose prior knowledge on the system in
form of partially known physics laws and often varies exploration policies
during the experiment. Here, we explore interactive workflows building on
multi-fidelity BO (MFBO), starting with classical (data-driven) MFBO, then
structured (physics-driven) sMFBO, and extending it to allow human in the loop
interactive iMFBO workflows for adaptive and domain expert aligned exploration.
These approaches are demonstrated over highly non-smooth multi-fidelity
simulation data generated from an Ising model, considering spin-spin
interaction as parameter space, lattice sizes as fidelity spaces, and the
objective as maximizing heat capacity. Detailed analysis and comparison show
the impact of physics knowledge injection and on-the-fly human decisions for
improved exploration, current challenges, and potential opportunities for
algorithm development with combining data, physics and real time human
decisions. | [
"cs.LG"
] | false |
2402.12616 | 2024-02-20T00:50:26Z | Multi-objective Binary Coordinate Search for Feature Selection | [
"Sevil Zanjani Miyandoab",
"Shahryar Rahnamayan",
"Azam Asilian Bidgoli"
] | A supervised feature selection method selects an appropriate but concise set
of features to differentiate classes, which is highly expensive for large-scale
datasets. Therefore, feature selection should aim at both minimizing the number
of selected features and maximizing the accuracy of classification, or any
other task. However, this crucial task is computationally highly demanding on
many real-world datasets and requires a very efficient algorithm to reach a set
of optimal features with a limited number of fitness evaluations. For this
purpose, we have proposed the binary multi-objective coordinate search (MOCS)
algorithm to solve large-scale feature selection problems. To the best of our
knowledge, the proposed algorithm in this paper is the first multi-objective
coordinate search algorithm. In this method, we generate new individuals by
flipping a variable of the candidate solutions on the Pareto front. This
enables us to investigate the effectiveness of each feature in the
corresponding subset. In fact, this strategy can play the role of crossover and
mutation operators to generate distinct subsets of features. The reported
results indicate the significant superiority of our method over NSGA-II, on
five real-world large-scale datasets, particularly when the computing budget is
limited. Moreover, this simple hyper-parameter-free algorithm can solve feature
selection much faster and more efficiently than NSGA-II. | [
"cs.LG",
"cs.NE",
"I.2.6"
] | false |
2402.12625 | 2024-02-20T01:10:12Z | Compact NSGA-II for Multi-objective Feature Selection | [
"Sevil Zanjani Miyandoab",
"Shahryar Rahnamayan",
"Azam Asilian Bidgoli"
] | Feature selection is an expensive challenging task in machine learning and
data mining aimed at removing irrelevant and redundant features. This
contributes to an improvement in classification accuracy, as well as the budget
and memory requirements for classification, or any other post-processing task
conducted after feature selection. In this regard, we define feature selection
as a multi-objective binary optimization task with the objectives of maximizing
classification accuracy and minimizing the number of selected features. In
order to select optimal features, we have proposed a binary Compact NSGA-II
(CNSGA-II) algorithm. Compactness represents the population as a probability
distribution to enhance evolutionary algorithms not only to be more
memory-efficient but also to reduce the number of fitness evaluations. Instead
of holding two populations during the optimization process, our proposed method
uses several Probability Vectors (PVs) to generate new individuals. Each PV
efficiently explores a region of the search space to find non-dominated
solutions instead of generating candidate solutions from a small population as
is the common approach in most evolutionary algorithms. To the best of our
knowledge, this is the first compact multi-objective algorithm proposed for
feature selection. The reported results for expensive optimization cases with a
limited budget on five datasets show that the CNSGA-II performs more
efficiently than the well-known NSGA-II method in terms of the hypervolume (HV)
performance metric requiring less memory. The proposed method and experimental
results are explained and analyzed in detail. | [
"cs.LG",
"cs.NE",
"I.2.6"
] | false |
2402.12626 | 2024-02-20T01:12:59Z | Indiscriminate Data Poisoning Attacks on Pre-trained Feature Extractors | [
"Yiwei Lu",
"Matthew Y. R. Yang",
"Gautam Kamath",
"Yaoliang Yu"
] | Machine learning models have achieved great success in supervised learning
tasks for end-to-end training, which requires a large amount of labeled data
that is not always feasible. Recently, many practitioners have shifted to
self-supervised learning methods that utilize cheap unlabeled data to learn a
general feature extractor via pre-training, which can be further applied to
personalized downstream tasks by simply training an additional linear layer
with limited labeled data. However, such a process may also raise concerns
regarding data poisoning attacks. For instance, indiscriminate data poisoning
attacks, which aim to decrease model utility by injecting a small number of
poisoned data into the training set, pose a security risk to machine learning
models, but have only been studied for end-to-end supervised learning. In this
paper, we extend the exploration of the threat of indiscriminate attacks on
downstream tasks that apply pre-trained feature extractors. Specifically, we
propose two types of attacks: (1) the input space attacks, where we modify
existing attacks to directly craft poisoned data in the input space. However,
due to the difficulty of optimization under constraints, we further propose (2)
the feature targeted attacks, where we mitigate the challenge with three
stages, firstly acquiring target parameters for the linear head; secondly
finding poisoned features by treating the learned feature representations as a
dataset; and thirdly inverting the poisoned features back to the input space.
Our experiments examine such attacks in popular downstream tasks of fine-tuning
on the same dataset and transfer learning that considers domain adaptation.
Empirical results reveal that transfer learning is more vulnerable to our
attacks. Additionally, input space attacks are a strong threat if no
countermeasures are posed, but are otherwise weaker than feature targeted
attacks. | [
"cs.LG",
"cs.CR"
] | false |
2402.12630 | 2024-02-20T01:22:04Z | FAST: An Optimization Framework for Fast Additive Segmentation in
Transparent ML | [
"Brian Liu",
"Rahul Mazumder"
] | We present FAST, an optimization framework for fast additive segmentation.
FAST segments piecewise constant shape functions for each feature in a dataset
to produce transparent additive models. The framework leverages a novel
optimization procedure to fit these models $\sim$2 orders of magnitude faster
than existing state-of-the-art methods, such as explainable boosting machines
\citep{nori2019interpretml}. We also develop new feature selection algorithms
in the FAST framework to fit parsimonious models that perform well. Through
experiments and case studies, we show that FAST improves the computational
efficiency and interpretability of additive models. | [
"stat.ML",
"cs.LG"
] | false |
2402.12646 | 2024-02-20T01:47:25Z | Training Artificial Neural Networks by Coordinate Search Algorithm | [
"Ehsan Rokhsatyazdi",
"Shahryar Rahnamayan",
"Sevil Zanjani Miyandoab",
"Azam Asilian Bidgoli",
"H. R. Tizhoosh"
] | Training Artificial Neural Networks poses a challenging and critical problem
in machine learning. Despite the effectiveness of gradient-based learning
methods, such as Stochastic Gradient Descent (SGD), in training neural
networks, they do have several limitations. For instance, they require
differentiable activation functions, and cannot optimize a model based on
several independent non-differentiable loss functions simultaneously; for
example, the F1-score, which is used during testing, can be used during
training when a gradient-free optimization algorithm is utilized. Furthermore,
the training in any DNN can be possible with a small size of the training
dataset. To address these concerns, we propose an efficient version of the
gradient-free Coordinate Search (CS) algorithm, an instance of General Pattern
Search methods, for training neural networks. The proposed algorithm can be
used with non-differentiable activation functions and tailored to
multi-objective/multi-loss problems. Finding the optimal values for weights of
ANNs is a large-scale optimization problem. Therefore instead of finding the
optimal value for each variable, which is the common technique in classical CS,
we accelerate optimization and convergence by bundling the weights. In fact,
this strategy is a form of dimension reduction for optimization problems. Based
on the experimental results, the proposed method, in some cases, outperforms
the gradient-based approach, particularly, in situations with insufficient
labeled training data. The performance plots demonstrate a high convergence
rate, highlighting the capability of our suggested method to find a reasonable
solution with fewer function calls. As of now, the only practical and efficient
way of training ANNs with hundreds of thousands of weights is gradient-based
algorithms such as SGD or Adam. In this paper we introduce an alternative
method for training ANN. | [
"cs.LG",
"cs.AI",
"I.2.6"
] | false |
2402.12668 | 2024-02-20T02:36:26Z | Randomization Can Reduce Both Bias and Variance: A Case Study in Random
Forests | [
"Brian Liu",
"Rahul Mazumder"
] | We study the often overlooked phenomenon, first noted in
\cite{breiman2001random}, that random forests appear to reduce bias compared to
bagging. Motivated by an interesting paper by \cite{mentch2020randomization},
where the authors argue that random forests reduce effective degrees of freedom
and only outperform bagging ensembles in low signal-to-noise ratio (SNR)
settings, we explore how random forests can uncover patterns in the data missed
by bagging. We empirically demonstrate that in the presence of such patterns,
random forests reduce bias along with variance and increasingly outperform
bagging ensembles when SNR is high. Our observations offer insights into the
real-world success of random forests across a range of SNRs and enhance our
understanding of the difference between random forests and bagging ensembles
with respect to the randomization injected into each split. Our investigations
also yield practical insights into the importance of tuning $mtry$ in random
forests. | [
"stat.ML",
"cs.LG"
] | false |
2402.12687 | 2024-02-20T03:27:53Z | Learning on manifolds without manifold learning | [
"H. N. Mhaskar",
"Ryan O'Dowd"
] | Function approximation based on data drawn randomly from an unknown
distribution is an important problem in machine learning. In contrast to the
prevalent paradigm of solving this problem by minimizing a loss functional, we
have given a direct one-shot construction together with optimal error bounds
under the manifold assumption; i.e., one assumes that the data is sampled from
an unknown sub-manifold of a high dimensional Euclidean space. A great deal of
research deals with obtaining information about this manifold, such as the
eigendecomposition of the Laplace-Beltrami operator or coordinate charts, and
using this information for function approximation. This two step approach
implies some extra errors in the approximation stemming from basic quantities
of the data in addition to the errors inherent in function approximation. In
Neural Networks, 132:253268, 2020, we have proposed a one-shot direct method to
achieve function approximation without requiring the extraction of any
information about the manifold other than its dimension. However, one cannot
pin down the class of approximants used in that paper.
In this paper, we view the unknown manifold as a sub-manifold of an ambient
hypersphere and study the question of constructing a one-shot approximation
using the spherical polynomials based on the hypersphere. Our approach does not
require preprocessing of the data to obtain information about the manifold
other than its dimension. We give optimal rates of approximation for relatively
"rough" functions. | [
"cs.LG",
"stat.ML"
] | false |
2402.12704 | 2024-02-20T04:06:28Z | Quantum Embedding with Transformer for High-dimensional Data | [
"Hao-Yuan Chen",
"Yen-Jui Chang",
"Shih-Wei Liao",
"Ching-Ray Chang"
] | Quantum embedding with transformers is a novel and promising architecture for
quantum machine learning to deliver exceptional capability on near-term devices
or simulators. The research incorporated a vision transformer (ViT) to advance
quantum significantly embedding ability and results for a single qubit
classifier with around 3 percent in the median F1 score on the BirdCLEF-2021, a
challenging high-dimensional dataset. The study showcases and analyzes
empirical evidence that our transformer-based architecture is a highly
versatile and practical approach to modern quantum machine learning problems. | [
"quant-ph",
"cs.LG"
] | false |
2402.12711 | 2024-02-20T04:21:13Z | Achieving Near-Optimal Regret for Bandit Algorithms with Uniform
Last-Iterate Guarantee | [
"Junyan Liu",
"Yunfan Li",
"Lin Yang"
] | Existing performance measures for bandit algorithms such as regret, PAC
bounds, or uniform-PAC (Dann et al., 2017), typically evaluate the cumulative
performance, while allowing the play of an arbitrarily bad arm at any finite
time t. Such a behavior can be highly detrimental in high-stakes applications.
This paper introduces a stronger performance measure, the uniform last-iterate
(ULI) guarantee, capturing both cumulative and instantaneous performance of
bandit algorithms. Specifically, ULI characterizes the instantaneous
performance since it ensures that the per-round regret of the played arm is
bounded by a function, monotonically decreasing w.r.t. (large) round t,
preventing revisits to bad arms when sufficient samples are available. We
demonstrate that a near-optimal ULI guarantee directly implies near-optimal
cumulative performance across aforementioned performance measures. To examine
the achievability of ULI in the finite arm setting, we first provide two
positive results that some elimination-based algorithms and high-probability
adversarial algorithms with stronger analysis or additional designs, can attain
near-optimal ULI guarantees. Then, we also provide a negative result,
indicating that optimistic algorithms cannot achieve a near-optimal ULI
guarantee. Finally, we propose an efficient algorithm for linear bandits with
infinitely many arms, which achieves the ULI guarantee, given access to an
optimization oracle. | [
"cs.LG",
"stat.ML"
] | false |
2402.12714 | 2024-02-20T04:40:00Z | Equivariant Pretrained Transformer for Unified Geometric Learning on
Multi-Domain 3D Molecules | [
"Rui Jiao",
"Xiangzhe Kong",
"Ziyang Yu",
"Wenbing Huang",
"Yang Liu"
] | Pretraining on a large number of unlabeled 3D molecules has showcased
superiority in various scientific applications. However, prior efforts
typically focus on pretraining models on a specific domain, either proteins or
small molecules, missing the opportunity to leverage the cross-domain
knowledge. To mitigate this gap, we introduce Equivariant Pretrained
Transformer (EPT), a novel pretraining framework designed to harmonize the
geometric learning of small molecules and proteins. To be specific, EPT unifies
the geometric modeling of multi-domain molecules via the block-enhanced
representation that can attend a broader context of each atom. Upon transformer
framework, EPT is further enhanced with E(3) equivariance to facilitate the
accurate representation of 3D structures. Another key innovation of EPT is its
block-level pretraining task, which allows for joint pretraining on datasets
comprising both small molecules and proteins. Experimental evaluations on a
diverse group of benchmarks, including ligand binding affinity prediction,
molecular property prediction, and protein property prediction, show that EPT
significantly outperforms previous SOTA methods for affinity prediction, and
achieves the best or comparable performance with existing domain-specific
pretraining models for other tasks. | [
"cs.LG",
"physics.chem-ph"
] | false |
2402.12729 | 2024-02-20T05:39:32Z | Scalable and reliable deep transfer learning for intelligent fault
detection via multi-scale neural processes embedded with knowledge | [
"Zhongzhi Li",
"Jingqi Tu",
"Jiacheng Zhu",
"Jianliang Ai",
"Yiqun Dong"
] | Deep transfer learning (DTL) is a fundamental method in the field of
Intelligent Fault Detection (IFD). It aims to mitigate the degradation of
method performance that arises from the discrepancies in data distribution
between training set (source domain) and testing set (target domain).
Considering the fact that fault data collection is challenging and certain
faults are scarce, DTL-based methods face the limitation of available
observable data, which reduces the detection performance of the methods in the
target domain. Furthermore, DTL-based methods lack comprehensive uncertainty
analysis that is essential for building reliable IFD systems. To address the
aforementioned problems, this paper proposes a novel DTL-based method known as
Neural Processes-based deep transfer learning with graph convolution network
(GTNP). Feature-based transfer strategy of GTNP bridges the data distribution
discrepancies of source domain and target domain in high-dimensional space.
Both the joint modeling based on global and local latent variables and sparse
sampling strategy reduce the demand of observable data in the target domain.
The multi-scale uncertainty analysis is obtained by using the distribution
characteristics of global and local latent variables. Global analysis of
uncertainty enables GTNP to provide quantitative values that reflect the
complexity of methods and the difficulty of tasks. Local analysis of
uncertainty allows GTNP to model uncertainty (confidence of the fault detection
result) at each sample affected by noise and bias. The validation of the
proposed method is conducted across 3 IFD tasks, consistently showing the
superior detection performance of GTNP compared to the other DTL-based methods. | [
"cs.LG",
"cs.AI"
] | false |
2402.12743 | 2024-02-20T06:19:55Z | APT-MMF: An advanced persistent threat actor attribution method based on
multimodal and multilevel feature fusion | [
"Nan Xiao",
"Bo Lang",
"Ting Wang",
"Yikai Chen"
] | Threat actor attribution is a crucial defense strategy for combating advanced
persistent threats (APTs). Cyber threat intelligence (CTI), which involves
analyzing multisource heterogeneous data from APTs, plays an important role in
APT actor attribution. The current attribution methods extract features from
different CTI perspectives and employ machine learning models to classify CTI
reports according to their threat actors. However, these methods usually
extract only one kind of feature and ignore heterogeneous information,
especially the attributes and relations of indicators of compromise (IOCs),
which form the core of CTI. To address these problems, we propose an APT actor
attribution method based on multimodal and multilevel feature fusion (APT-MMF).
First, we leverage a heterogeneous attributed graph to characterize APT reports
and their IOC information. Then, we extract and fuse multimodal features,
including attribute type features, natural language text features and
topological relationship features, to construct comprehensive node
representations. Furthermore, we design multilevel heterogeneous graph
attention networks to learn the deep hidden features of APT report nodes; these
networks integrate IOC type-level, metapath-based neighbor node-level, and
metapath semantic-level attention. Utilizing multisource threat intelligence,
we construct a heterogeneous attributed graph dataset for verification
purposes. The experimental results show that our method not only outperforms
the existing methods but also demonstrates its good interpretability for
attribution analysis tasks. | [
"cs.CR",
"cs.LG"
] | false |
2402.12756 | 2024-02-20T06:49:43Z | Static vs. Dynamic Databases for Indoor Localization based on Wi-Fi
Fingerprinting: A Discussion from a Data Perspective | [
"Zhe Tang",
"Ruocheng Gu",
"Sihao Li",
"Kyeong Soo Kim",
"Jeremy S. Smith"
] | Wi-Fi fingerprinting has emerged as the most popular approach to indoor
localization. The use of ML algorithms has greatly improved the localization
performance of Wi-Fi fingerprinting, but its success depends on the
availability of fingerprint databases composed of a large number of RSSIs, the
MAC addresses of access points, and the other measurement information. However,
most fingerprint databases do not reflect well the time varying nature of
electromagnetic interferences in complicated modern indoor environment. This
could result in significant changes in statistical characteristics of
training/validation and testing datasets, which are often constructed at
different times, and even the characteristics of the testing datasets could be
different from those of the data submitted by users during the operation of
localization systems after their deployment. In this paper, we consider the
implications of time-varying Wi-Fi fingerprints on indoor localization from a
data-centric point of view and discuss the differences between static and
dynamic databases. As a case study, we have constructed a dynamic database
covering three floors of the IR building of XJTLU based on RSSI measurements,
over 44 days, and investigated the differences between static and dynamic
databases in terms of statistical characteristics and localization performance.
The analyses based on variance calculations and Isolation Forest show the
temporal shifts in RSSIs, which result in a noticeable trend of the increase in
the localization error of a Gaussian process regression model with the maximum
error of 6.65 m after 14 days of training without model adjustments. The
results of the case study with the XJTLU dynamic database clearly demonstrate
the limitations of static databases and the importance of the creation and
adoption of dynamic databases for future indoor localization research and
real-world deployment. | [
"cs.LG",
"cs.NI"
] | false |
2402.12761 | 2024-02-20T07:03:59Z | FGAD: Self-boosted Knowledge Distillation for An Effective Federated
Graph Anomaly Detection Framework | [
"Jinyu Cai",
"Yunhe Zhang",
"Zhoumin Lu",
"Wenzhong Guo",
"See-kiong Ng"
] | Graph anomaly detection (GAD) aims to identify anomalous graphs that
significantly deviate from other ones, which has raised growing attention due
to the broad existence and complexity of graph-structured data in many
real-world scenarios. However, existing GAD methods usually execute with
centralized training, which may lead to privacy leakage risk in some sensitive
cases, thereby impeding collaboration among organizations seeking to
collectively develop robust GAD models. Although federated learning offers a
promising solution, the prevalent non-IID problems and high communication costs
present significant challenges, particularly pronounced in collaborations with
graph data distributed among different participants. To tackle these
challenges, we propose an effective federated graph anomaly detection framework
(FGAD). We first introduce an anomaly generator to perturb the normal graphs to
be anomalous, and train a powerful anomaly detector by distinguishing generated
anomalous graphs from normal ones. Then, we leverage a student model to distill
knowledge from the trained anomaly detector (teacher model), which aims to
maintain the personality of local models and alleviate the adverse impact of
non-IID problems. Moreover, we design an effective collaborative learning
mechanism that facilitates the personalization preservation of local models and
significantly reduces communication costs among clients. Empirical results of
the GAD tasks on non-IID graphs compared with state-of-the-art baselines
demonstrate the superiority and efficiency of the proposed FGAD method. | [
"cs.LG",
"cs.CR"
] | false |
2402.12789 | 2024-02-20T07:57:38Z | Fair Classifiers Without Fair Training: An Influence-Guided Data
Sampling Approach | [
"Jinlong Pang",
"Jialu Wang",
"Zhaowei Zhu",
"Yuanshun Yao",
"Chen Qian",
"Yang Liu"
] | A fair classifier should ensure the benefit of people from different groups,
while the group information is often sensitive and unsuitable for model
training. Therefore, learning a fair classifier but excluding sensitive
attributes in the training dataset is important. In this paper, we study
learning fair classifiers without implementing fair training algorithms to
avoid possible leakage of sensitive information. Our theoretical analyses
validate the possibility of this approach, that traditional training on a
dataset with an appropriate distribution shift can reduce both the upper bound
for fairness disparity and model generalization error, indicating that fairness
and accuracy can be improved simultaneously with simply traditional training.
We then propose a tractable solution to progressively shift the original
training data during training by sampling influential data, where the sensitive
attribute of new data is not accessed in sampling or used in training.
Extensive experiments on real-world data demonstrate the effectiveness of our
proposed algorithm. | [
"cs.LG",
"cs.AI"
] | false |
2402.12808 | 2024-02-20T08:27:50Z | Learning Generalization and Regularization of Nonhomogeneous Temporal
Poisson Processes | [
"Son Nguyen Van",
"Hoai Nguyen Xuan"
] | The Poisson process, especially the nonhomogeneous Poisson process (NHPP), is
an essentially important counting process with numerous real-world
applications. Up to date, almost all works in the literature have been on the
estimation of NHPPs with infinite data using non-data driven binning methods.
In this paper, we formulate the problem of estimation of NHPPs from finite and
limited data as a learning generalization problem. We mathematically show that
while binning methods are essential for the estimation of NHPPs, they pose a
threat of overfitting when the amount of data is limited. We propose a
framework for regularized learning of NHPPs with two new adaptive and
data-driven binning methods that help to remove the ad-hoc tuning of binning
parameters. Our methods are experimentally tested on synthetic and real-world
datasets and the results show their effectiveness. | [
"cs.LG",
"stat.ML"
] | false |
2402.12812 | 2024-02-20T08:30:46Z | Scalable Decentralized Algorithms for Online Personalized Mean
Estimation | [
"Franco Galante",
"Giovanni Neglia",
"Emilio Leonardi"
] | In numerous settings, agents lack sufficient data to directly learn a model.
Collaborating with other agents may help, but it introduces a bias-variance
trade-off, when local data distributions differ. A key challenge is for each
agent to identify clients with similar distributions while learning the model,
a problem that remains largely unresolved. This study focuses on a simplified
version of the overarching problem, where each agent collects samples from a
real-valued distribution over time to estimate its mean. Existing algorithms
face impractical space and time complexities (quadratic in the number of agents
A). To address scalability challenges, we propose a framework where agents
self-organize into a graph, allowing each agent to communicate with only a
selected number of peers r. We introduce two collaborative mean estimation
algorithms: one draws inspiration from belief propagation, while the other
employs a consensus-based approach, with complexity of O( r |A| log |A|) and
O(r |A|), respectively. We establish conditions under which both algorithms
yield asymptotically optimal estimates and offer a theoretical characterization
of their performance. | [
"cs.LG",
"cs.DC"
] | false |
2402.12861 | 2024-02-20T09:52:30Z | Bounding Reconstruction Attack Success of Adversaries Without Data
Priors | [
"Alexander Ziller",
"Anneliese Riess",
"Kristian Schwethelm",
"Tamara T. Mueller",
"Daniel Rueckert",
"Georgios Kaissis"
] | Reconstruction attacks on machine learning (ML) models pose a strong risk of
leakage of sensitive data. In specific contexts, an adversary can (almost)
perfectly reconstruct training data samples from a trained model using the
model's gradients. When training ML models with differential privacy (DP),
formal upper bounds on the success of such reconstruction attacks can be
provided. So far, these bounds have been formulated under worst-case
assumptions that might not hold high realistic practicality. In this work, we
provide formal upper bounds on reconstruction success under realistic
adversarial settings against ML models trained with DP and support these bounds
with empirical results. With this, we show that in realistic scenarios, (a) the
expected reconstruction success can be bounded appropriately in different
contexts and by different metrics, which (b) allows for a more educated choice
of a privacy parameter. | [
"cs.LG",
"cs.CR"
] | false |
2402.12867 | 2024-02-20T09:57:49Z | Towards MLOps: A DevOps Tools Recommender System for Machine Learning
System | [
"Pir Sami Ullah Shah",
"Naveed Ahmad",
"Mirza Omer Beg"
] | Applying DevOps practices to machine learning system is termed as MLOps and
machine learning systems evolve on new data unlike traditional systems on
requirements. The objective of MLOps is to establish a connection between
different open-source tools to construct a pipeline that can automatically
perform steps to construct a dataset, train the machine learning model and
deploy the model to the production as well as store different versions of model
and dataset. Benefits of MLOps is to make sure the fast delivery of the new
trained models to the production to have accurate results. Furthermore, MLOps
practice impacts the overall quality of the software products and is completely
dependent on open-source tools and selection of relevant open-source tools is
considered as challenged while a generalized method to select an appropriate
open-source tools is desirable. In this paper, we present a framework for
recommendation system that processes the contextual information (e.g., nature
of data, type of the data) of the machine learning project and recommends a
relevant toolchain (tech-stack) for the operationalization of machine learning
systems. To check the applicability of the proposed framework, four different
approaches i.e., rule-based, random forest, decision trees and k-nearest
neighbors were investigated where precision, recall and f-score is measured,
the random forest out classed other approaches with highest f-score value of
0.66. | [
"cs.SE",
"cs.LG"
] | false |
2402.12868 | 2024-02-20T09:59:33Z | Fast Rates in Online Convex Optimization by Exploiting the Curvature of
Feasible Sets | [
"Taira Tsuchiya",
"Shinji Ito"
] | In this paper, we explore online convex optimization (OCO) and introduce a
new analysis that provides fast rates by exploiting the curvature of feasible
sets. In online linear optimization, it is known that if the average gradient
of loss functions is larger than a certain value, the curvature of feasible
sets can be exploited by the follow-the-leader (FTL) algorithm to achieve a
logarithmic regret. This paper reveals that algorithms adaptive to the
curvature of loss functions can also leverage the curvature of feasible sets.
We first prove that if an optimal decision is on the boundary of a feasible set
and the gradient of an underlying loss function is non-zero, then the algorithm
achieves a regret upper bound of $O(\rho \log T)$ in stochastic environments.
Here, $\rho > 0$ is the radius of the smallest sphere that includes the optimal
decision and encloses the feasible set. Our approach, unlike existing ones, can
work directly with convex loss functions, exploiting the curvature of loss
functions simultaneously, and can achieve the logarithmic regret only with a
local property of feasible sets. Additionally, it achieves an $O(\sqrt{T})$
regret even in adversarial environments where FTL suffers an $\Omega(T)$
regret, and attains an $O(\rho \log T + \sqrt{C \rho \log T})$ regret bound in
corrupted stochastic environments with corruption level $C$. Furthermore, by
extending our analysis, we establish a regret upper bound of
$O\Big(T^{\frac{q-2}{2(q-1)}} (\log T)^{\frac{q}{2(q-1)}}\Big)$ for
$q$-uniformly convex feasible sets, where uniformly convex sets include
strongly convex sets and $\ell_p$-balls for $p \in [1,\infty)$. This bound
bridges the gap between the $O(\log T)$ regret bound for strongly convex sets
($q=2$) and the $O(\sqrt{T})$ regret bound for non-curved sets ($q\to\infty$). | [
"cs.LG",
"stat.ML"
] | false |
2402.12885 | 2024-02-20T10:25:44Z | A Bound on the Maximal Marginal Degrees of Freedom | [
"Paul Dommel"
] | Common kernel ridge regression is expensive in memory allocation and
computation time. This paper addresses low rank approximations and surrogates
for kernel ridge regression, which bridge these difficulties. The fundamental
contribution of the paper is a lower bound on the rank of the low dimensional
approximation, which is required such that the prediction power remains
reliable. The bound relates the effective dimension with the largest
statistical leverage score. We characterize the effective dimension and its
growth behavior with respect to the regularization parameter by involving the
regularity of the kernel. This growth is demonstrated to be asymptotically
logarithmic for suitably chosen kernels, justifying low-rank approximations as
the Nystr\"om method. | [
"stat.ML",
"cs.LG"
] | false |
2402.12916 | 2024-02-20T11:06:42Z | Data Pipeline Training: Integrating AutoML to Optimize the Data Flow of
Machine Learning Models | [
"Jiang Wu",
"Hongbo Wang",
"Chunhe Ni",
"Chenwei Zhang",
"Wenran Lu"
] | Data Pipeline plays an indispensable role in tasks such as modeling machine
learning and developing data products. With the increasing diversification and
complexity of Data sources, as well as the rapid growth of data volumes,
building an efficient Data Pipeline has become crucial for improving work
efficiency and solving complex problems. This paper focuses on exploring how to
optimize data flow through automated machine learning methods by integrating
AutoML with Data Pipeline. We will discuss how to leverage AutoML technology to
enhance the intelligence of Data Pipeline, thereby achieving better results in
machine learning tasks. By delving into the automation and optimization of Data
flows, we uncover key strategies for constructing efficient data pipelines that
can adapt to the ever-changing data landscape. This not only accelerates the
modeling process but also provides innovative solutions to complex problems,
enabling more significant outcomes in increasingly intricate data domains.
Keywords- Data Pipeline Training;AutoML; Data environment; Machine learning | [
"cs.LG",
"cs.AI"
] | false |
2402.12937 | 2024-02-20T11:38:52Z | GRAPHGINI: Fostering Individual and Group Fairness in Graph Neural
Networks | [
"Anuj Kumar Sirohi",
"Anjali Gupta",
"Sayan Ranu",
"Sandeep Kumar",
"Amitabha Bagchi"
] | We address the growing apprehension that GNNs, in the absence of fairness
constraints, might produce biased decisions that disproportionately affect
underprivileged groups or individuals. Departing from previous work, we
introduce for the first time a method for incorporating the Gini coefficient as
a measure of fairness to be used within the GNN framework. Our proposal,
GRAPHGINI, works with the two different goals of individual and group fairness
in a single system, while maintaining high prediction accuracy. GRAPHGINI
enforces individual fairness through learnable attention scores that help in
aggregating more information through similar nodes. A heuristic-based maximum
Nash social welfare constraint ensures the maximum possible group fairness.
Both the individual fairness constraint and the group fairness constraint are
stated in terms of a differentiable approximation of the Gini coefficient. This
approximation is a contribution that is likely to be of interest even beyond
the scope of the problem studied in this paper. Unlike other state-of-the-art,
GRAPHGINI automatically balances all three optimization objectives (utility,
individual, and group fairness) of the GNN and is free from any manual tuning
of weight parameters. Extensive experimentation on real-world datasets
showcases the efficacy of GRAPHGINI in making significant improvements in
individual fairness compared to all currently available state-of-the-art
methods while maintaining utility and group equality. | [
"cs.LG",
"cs.SI"
] | false |
2402.12939 | 2024-02-20T11:50:50Z | Discovering Behavioral Modes in Deep Reinforcement Learning Policies
Using Trajectory Clustering in Latent Space | [
"Sindre Benjamin Remman",
"Anastasios M. Lekkas"
] | Understanding the behavior of deep reinforcement learning (DRL) agents is
crucial for improving their performance and reliability. However, the
complexity of their policies often makes them challenging to understand. In
this paper, we introduce a new approach for investigating the behavior modes of
DRL policies, which involves utilizing dimensionality reduction and trajectory
clustering in the latent space of neural networks. Specifically, we use
Pairwise Controlled Manifold Approximation Projection (PaCMAP) for
dimensionality reduction and TRACLUS for trajectory clustering to analyze the
latent space of a DRL policy trained on the Mountain Car control task. Our
methodology helps identify diverse behavior patterns and suboptimal choices by
the policy, thus allowing for targeted improvements. We demonstrate how our
approach, combined with domain knowledge, can enhance a policy's performance in
specific regions of the state space. | [
"cs.LG",
"cs.AI"
] | false |
2402.12971 | 2024-02-20T12:40:31Z | How Temporal Unrolling Supports Neural Physics Simulators | [
"Bjoern List",
"Li-Wei Chen",
"Kartik Bali",
"Nils Thuerey"
] | Unrolling training trajectories over time strongly influences the inference
accuracy of neural network-augmented physics simulators. We analyze these
effects by studying three variants of training neural networks on discrete
ground truth trajectories. In addition to commonly used one-step setups and
fully differentiable unrolling, we include a third, less widely used variant:
unrolling without temporal gradients. Comparing networks trained with these
three modalities makes it possible to disentangle the two dominant effects of
unrolling, training distribution shift and long-term gradients. We present a
detailed study across physical systems, network sizes, network architectures,
training setups, and test scenarios. It provides an empirical basis for our
main findings: A non-differentiable but unrolled training setup supported by a
numerical solver can yield 4.5-fold improvements over a fully differentiable
prediction setup that does not utilize this solver. We also quantify a
difference in the accuracy of models trained in a fully differentiable setup
compared to their non-differentiable counterparts. While differentiable setups
perform best, the accuracy of unrolling without temporal gradients comes
comparatively close. Furthermore, we empirically show that these behaviors are
invariant to changes in the underlying physical system, the network
architecture and size, and the numerical scheme. These results motivate
integrating non-differentiable numerical simulators into training setups even
if full differentiability is unavailable. We also observe that the convergence
rate of common neural architectures is low compared to numerical algorithms.
This encourages the use of hybrid approaches combining neural and numerical
algorithms to utilize the benefits of both. | [
"physics.comp-ph",
"cs.LG"
] | false |
2402.13037 | 2024-02-20T14:24:00Z | Align Your Intents: Offline Imitation Learning via Optimal Transport | [
"Maksim Bobrin",
"Nazar Buzun",
"Dmitrii Krylov",
"Dmitry V. Dylov"
] | Offline reinforcement learning (RL) addresses the problem of sequential
decision-making by learning optimal policy through pre-collected data, without
interacting with the environment. As yet, it has remained somewhat impractical,
because one rarely knows the reward explicitly and it is hard to distill it
retrospectively. Here, we show that an imitating agent can still learn the
desired behavior merely from observing the expert, despite the absence of
explicit rewards or action labels. In our method, AILOT (Aligned Imitation
Learning via Optimal Transport), we involve special representation of states in
a form of intents that incorporate pairwise spatial distances within the data.
Given such representations, we define intrinsic reward function via optimal
transport distance between the expert's and the agent's trajectories. We report
that AILOT outperforms state-of-the art offline imitation learning algorithms
on D4RL benchmarks and improves the performance of other offline RL algorithms
in the sparse-reward tasks. | [
"cs.LG",
"cs.AI"
] | false |
2402.13081 | 2024-02-20T15:25:56Z | IT Intrusion Detection Using Statistical Learning and Testbed
Measurements | [
"Xiaoxuan Wang",
"Rolf Stadler"
] | We study automated intrusion detection in an IT infrastructure, specifically
the problem of identifying the start of an attack, the type of attack, and the
sequence of actions an attacker takes, based on continuous measurements from
the infrastructure. We apply statistical learning methods, including Hidden
Markov Model (HMM), Long Short-Term Memory (LSTM), and Random Forest Classifier
(RFC) to map sequences of observations to sequences of predicted attack
actions. In contrast to most related research, we have abundant data to train
the models and evaluate their predictive power. The data comes from traces we
generate on an in-house testbed where we run attacks against an emulated IT
infrastructure. Central to our work is a machine-learning pipeline that maps
measurements from a high-dimensional observation space to a space of low
dimensionality or to a small set of observation symbols. Investigating
intrusions in offline as well as online scenarios, we find that both HMM and
LSTM can be effective in predicting attack start time, attack type, and attack
actions. If sufficient training data is available, LSTM achieves higher
prediction accuracy than HMM. HMM, on the other hand, requires less
computational resources and less training data for effective prediction. Also,
we find that the methods we study benefit from data produced by traditional
intrusion detection systems like SNORT. | [
"cs.LG",
"cs.CR"
] | false |
2402.13087 | 2024-02-20T15:29:49Z | How Does Selection Leak Privacy: Revisiting Private Selection and
Improved Results for Hyper-parameter Tuning | [
"Zihang Xiang",
"Chenglong Wang",
"Di Wang"
] | We study the problem of guaranteeing Differential Privacy (DP) in
hyper-parameter tuning, a crucial process in machine learning involving the
selection of the best run from several. Unlike many private algorithms,
including the prevalent DP-SGD, the privacy implications of tuning remain
insufficiently understood. Recent works propose a generic private solution for
the tuning process, yet a fundamental question still persists: is the current
privacy bound for this solution tight?
This paper contributes both positive and negative answers to this question.
Initially, we provide studies affirming the current privacy analysis is indeed
tight in a general sense. However, when we specifically study the
hyper-parameter tuning problem, such tightness no longer holds. This is first
demonstrated by applying privacy audit on the tuning process. Our findings
underscore a substantial gap between the current theoretical privacy bound and
the empirical bound derived even under the strongest audit setup.
The gap found is not a fluke. Our subsequent study provides an improved
privacy result for private hyper-parameter tuning due to its distinct
properties. Our privacy results are also more generalizable compared to prior
analyses that are only easily applicable in specific setups. | [
"cs.LG",
"cs.CR"
] | false |
2402.13114 | 2024-02-20T16:11:59Z | BuffGraph: Enhancing Class-Imbalanced Node Classification via Buffer
Nodes | [
"Qian Wang",
"Zemin Liu",
"Zhen Zhang",
"Bingsheng He"
] | Class imbalance in graph-structured data, where minor classes are
significantly underrepresented, poses a critical challenge for Graph Neural
Networks (GNNs). To address this challenge, existing studies generally generate
new minority nodes and edges connecting new nodes to the original graph to make
classes balanced. However, they do not solve the problem that majority classes
still propagate information to minority nodes by edges in the original graph
which introduces bias towards majority classes. To address this, we introduce
BuffGraph, which inserts buffer nodes into the graph, modulating the impact of
majority classes to improve minor class representation. Our extensive
experiments across diverse real-world datasets empirically demonstrate that
BuffGraph outperforms existing baseline methods in class-imbalanced node
classification in both natural settings and imbalanced settings. Code is
available at https://anonymous.4open.science/r/BuffGraph-730A. | [
"cs.LG",
"cs.AI"
] | false |
2402.13147 | 2024-02-20T17:02:48Z | SubIQ: Inverse Soft-Q Learning for Offline Imitation with Suboptimal
Demonstrations | [
"Huy Hoang",
"Tien Mai",
"Pradeep Varakantham"
] | We consider offline imitation learning (IL), which aims to mimic the expert's
behavior from its demonstration without further interaction with the
environment. One of the main challenges in offline IL is dealing with the
limited support of expert demonstrations that cover only a small fraction of
the state-action spaces. In this work, we consider offline IL, where expert
demonstrations are limited but complemented by a larger set of sub-optimal
demonstrations of lower expertise levels. Most of the existing offline IL
methods developed for this setting are based on behavior cloning or
distribution matching, where the aim is to match the occupancy distribution of
the imitation policy with that of the expert policy. Such an approach often
suffers from over-fitting, as expert demonstrations are limited to accurately
represent any occupancy distribution. On the other hand, since sub-optimal sets
are much larger, there is a high chance that the imitation policy is trained
towards sub-optimal policies. In this paper, to address these issues, we
propose a new approach based on inverse soft-Q learning, where a regularization
term is added to the training objective, with the aim of aligning the learned
rewards with a pre-assigned reward function that allocates higher weights to
state-action pairs from expert demonstrations, and lower weights to those from
lower expertise levels. On standard benchmarks, our inverse soft-Q learning
significantly outperforms other offline IL baselines by a large margin. | [
"cs.LG",
"cs.AI"
] | false |
2402.13148 | 2024-02-20T17:04:06Z | Defending Jailbreak Prompts via In-Context Adversarial Game | [
"Yujun Zhou",
"Yufei Han",
"Haomin Zhuang",
"Taicheng Guo",
"Kehan Guo",
"Zhenwen Liang",
"Hongyan Bao",
"Xiangliang Zhang"
] | Large Language Models (LLMs) demonstrate remarkable capabilities across
diverse applications. However, concerns regarding their security, particularly
the vulnerability to jailbreak attacks, persist. Drawing inspiration from
adversarial training in deep learning and LLM agent learning processes, we
introduce the In-Context Adversarial Game (ICAG) for defending against
jailbreaks without the need for fine-tuning. ICAG leverages agent learning to
conduct an adversarial game, aiming to dynamically extend knowledge to defend
against jailbreaks. Unlike traditional methods that rely on static datasets,
ICAG employs an iterative process to enhance both the defense and attack
agents. This continuous improvement process strengthens defenses against newly
generated jailbreak prompts. Our empirical studies affirm ICAG's efficacy,
where LLMs safeguarded by ICAG exhibit significantly reduced jailbreak success
rates across various attack scenarios. Moreover, ICAG demonstrates remarkable
transferability to other LLMs, indicating its potential as a versatile defense
mechanism. | [
"cs.LG",
"cs.CR"
] | false |
2402.13181 | 2024-02-20T17:48:11Z | DINOBot: Robot Manipulation via Retrieval and Alignment with Vision
Foundation Models | [
"Norman Di Palo",
"Edward Johns"
] | We propose DINOBot, a novel imitation learning framework for robot
manipulation, which leverages the image-level and pixel-level capabilities of
features extracted from Vision Transformers trained with DINO. When interacting
with a novel object, DINOBot first uses these features to retrieve the most
visually similar object experienced during human demonstrations, and then uses
this object to align its end-effector with the novel object to enable effective
interaction. Through a series of real-world experiments on everyday tasks, we
show that exploiting both the image-level and pixel-level properties of vision
foundation models enables unprecedented learning efficiency and generalisation.
Videos and code are available at https://www.robot-learning.uk/dinobot. | [
"cs.RO",
"cs.LG"
] | false |
2402.13204 | 2024-02-20T18:15:11Z | SONATA: Self-adaptive Evolutionary Framework for Hardware-aware Neural
Architecture Search | [
"Halima Bouzidi",
"Smail Niar",
"Hamza Ouarnoughi",
"El-Ghazali Talbi"
] | Recent advancements in Artificial Intelligence (AI), driven by Neural
Networks (NN), demand innovative neural architecture designs, particularly
within the constrained environments of Internet of Things (IoT) systems, to
balance performance and efficiency. HW-aware Neural Architecture Search
(HW-aware NAS) emerges as an attractive strategy to automate the design of NN
using multi-objective optimization approaches, such as evolutionary algorithms.
However, the intricate relationship between NN design parameters and HW-aware
NAS optimization objectives remains an underexplored research area, overlooking
opportunities to effectively leverage this knowledge to guide the search
process accordingly. Furthermore, the large amount of evaluation data produced
during the search holds untapped potential for refining the optimization
strategy and improving the approximation of the Pareto front. Addressing these
issues, we propose SONATA, a self-adaptive evolutionary algorithm for HW-aware
NAS. Our method leverages adaptive evolutionary operators guided by the learned
importance of NN design parameters. Specifically, through tree-based surrogate
models and a Reinforcement Learning agent, we aspire to gather knowledge on
'How' and 'When' to evolve NN architectures. Comprehensive evaluations across
various NAS search spaces and hardware devices on the ImageNet-1k dataset have
shown the merit of SONATA with up to 0.25% improvement in accuracy and up to
2.42x gains in latency and energy. Our SONATA has seen up to sim$93.6% Pareto
dominance over the native NSGA-II, further stipulating the importance of
self-adaptive evolution operators in HW-aware NAS. | [
"cs.NE",
"cs.LG"
] | false |