instruction
stringclasses 1
value | input
stringlengths 260
2.07k
| output
stringclasses 10
values |
---|---|---|
What field is the article from? | Title: Multimodal Stress Detection Using Facial Landmarks and Biometric Signals
Abstract: The development of various sensing technologies is improving measurements of
stress and the well-being of individuals. Although progress has been made with
single signal modalities like wearables and facial emotion recognition,
integrating multiple modalities provides a more comprehensive understanding of
stress, given that stress manifests differently across different people.
Multi-modal learning aims to capitalize on the strength of each modality rather
than relying on a single signal. Given the complexity of processing and
integrating high-dimensional data from limited subjects, more research is
needed. Numerous research efforts have been focused on fusing stress and
emotion signals at an early stage, e.g., feature-level fusion using basic
machine learning methods and 1D-CNN Methods. This paper proposes a multi-modal
learning approach for stress detection that integrates facial landmarks and
biometric signals. We test this multi-modal integration with various
early-fusion and late-fusion techniques to integrate the 1D-CNN model from
biometric signals and 2-D CNN using facial landmarks. We evaluate these
architectures using a rigorous test of models' generalizability using the
leave-one-subject-out mechanism, i.e., all samples related to a single subject
are left out to train the model. Our findings show that late-fusion achieved
94.39\% accuracy, and early-fusion surpassed it with a 98.38\% accuracy rate.
This research contributes valuable insights into enhancing stress detection
through a multi-modal approach. The proposed research offers important
knowledge in improving stress detection using a multi-modal approach. | Computer Vision |
What field is the article from? | Title: 3D Hand Pose Estimation in Egocentric Images in the Wild
Abstract: We present WildHands, a method for 3D hand pose estimation in egocentric
images in the wild. This is challenging due to (a) lack of 3D hand pose
annotations for images in the wild, and (b) a form of perspective
distortion-induced shape ambiguity that arises in the analysis of crops around
hands. For the former, we use auxiliary supervision on in-the-wild data in the
form of segmentation masks & grasp labels in addition to 3D supervision
available in lab datasets. For the latter, we provide spatial cues about the
location of the hand crop in the camera's field of view. Our approach achieves
the best 3D hand pose on the ARCTIC leaderboard and outperforms FrankMocap, a
popular and robust approach for estimating hand pose in the wild, by 45.3% when
evaluated on 2D hand pose on our EPIC-HandKps dataset. | Computer Vision |
What field is the article from? | Title: Domain-Specific Deep Learning Feature Extractor for Diabetic Foot Ulcer Detection
Abstract: Diabetic Foot Ulcer (DFU) is a condition requiring constant monitoring and
evaluations for treatment. DFU patient population is on the rise and will soon
outpace the available health resources. Autonomous monitoring and evaluation of
DFU wounds is a much-needed area in health care. In this paper, we evaluate and
identify the most accurate feature extractor that is the core basis for
developing a deep-learning wound detection network. For the evaluation, we used
mAP and F1-score on the publicly available DFU2020 dataset. A combination of
UNet and EfficientNetb3 feature extractor resulted in the best evaluation among
the 14 networks compared. UNet and Efficientnetb3 can be used as the classifier
in the development of a comprehensive DFU domain-specific autonomous wound
detection pipeline. | Computer Vision |
What field is the article from? | Title: Improving Source-Free Target Adaptation with Vision Transformers Leveraging Domain Representation Images
Abstract: Unsupervised Domain Adaptation (UDA) methods facilitate knowledge transfer
from a labeled source domain to an unlabeled target domain, navigating the
obstacle of domain shift. While Convolutional Neural Networks (CNNs) are a
staple in UDA, the rise of Vision Transformers (ViTs) provides new avenues for
domain generalization. This paper presents an innovative method to bolster ViT
performance in source-free target adaptation, beginning with an evaluation of
how key, query, and value elements affect ViT outcomes. Experiments indicate
that altering the key component has negligible effects on Transformer
performance. Leveraging this discovery, we introduce Domain Representation
Images (DRIs), feeding embeddings through the key element. DRIs act as
domain-specific markers, effortlessly merging with the training regimen. To
assess our method, we perform target adaptation tests on the Cross Instance DRI
source-only (SO) control. We measure the efficacy of target adaptation with and
without DRIs, against existing benchmarks like SHOT-B* and adaptations via
CDTrans. Findings demonstrate that excluding DRIs offers limited gains over
SHOT-B*, while their inclusion in the key segment boosts average precision
promoting superior domain generalization. This research underscores the vital
role of DRIs in enhancing ViT efficiency in UDA scenarios, setting a precedent
for further domain adaptation explorations. | Computer Vision |
What field is the article from? | Title: Exploring Post-Training Quantization of Protein Language Models
Abstract: Recent advancements in unsupervised protein language models (ProteinLMs),
like ESM-1b and ESM-2, have shown promise in different protein prediction
tasks. However, these models face challenges due to their high computational
demands, significant memory needs, and latency, restricting their usage on
devices with limited resources. To tackle this, we explore post-training
quantization (PTQ) for ProteinLMs, focusing on ESMFold, a simplified version of
AlphaFold based on ESM-2 ProteinLM. Our study is the first attempt to quantize
all weights and activations of ProteinLMs. We observed that the typical uniform
quantization method performs poorly on ESMFold, causing a significant drop in
TM-Score when using 8-bit quantization. We conducted extensive quantization
experiments, uncovering unique challenges associated with ESMFold, particularly
highly asymmetric activation ranges before Layer Normalization, making
representation difficult using low-bit fixed-point formats. To address these
challenges, we propose a new PTQ method for ProteinLMs, utilizing piecewise
linear quantization for asymmetric activation values to ensure accurate
approximation. We demonstrated the effectiveness of our method in protein
structure prediction tasks, demonstrating that ESMFold can be accurately
quantized to low-bit widths without compromising accuracy. Additionally, we
applied our method to the contact prediction task, showcasing its versatility.
In summary, our study introduces an innovative PTQ method for ProteinLMs,
addressing specific quantization challenges and potentially leading to the
development of more efficient ProteinLMs with significant implications for
various protein-related applications. | Machine Learning |
What field is the article from? | Title: Toward Reinforcement Learning-based Rectilinear Macro Placement Under Human Constraints
Abstract: Macro placement is a critical phase in chip design, which becomes more
intricate when involving general rectilinear macros and layout areas.
Furthermore, macro placement that incorporates human-like constraints, such as
design hierarchy and peripheral bias, has the potential to significantly reduce
the amount of additional manual labor required from designers. This study
proposes a methodology that leverages an approach suggested by Google's Circuit
Training (G-CT) to provide a learning-based macro placer that not only supports
placing rectilinear cases, but also adheres to crucial human-like design
principles. Our experimental results demonstrate the effectiveness of our
framework in achieving power-performance-area (PPA) metrics and in obtaining
placements of high quality, comparable to those produced with human
intervention. Additionally, our methodology shows potential as a generalized
model to address diverse macro shapes and layout areas. | Machine Learning |
What field is the article from? | Title: A Multi-Center Study on the Adaptability of a Shared Foundation Model for Electronic Health Records
Abstract: Foundation models hold promise for transforming AI in healthcare by providing
modular components that are easily adaptable to downstream healthcare tasks,
making AI development more scalable and cost-effective. Structured EHR
foundation models, trained on coded medical records from millions of patients,
demonstrated benefits including increased performance with fewer training
labels, and improved robustness to distribution shifts. However, questions
remain on the feasibility of sharing these models across different hospitals
and their performance for local task adaptation. This multi-center study
examined the adaptability of a recently released structured EHR foundation
model ($FM_{SM}$), trained on longitudinal medical record data from 2.57M
Stanford Medicine patients. Experiments were conducted using EHR data at The
Hospital for Sick Children and MIMIC-IV. We assessed both adaptability via
continued pretraining on local data, and task adaptability compared to
baselines of training models from scratch at each site, including a local
foundation model. We evaluated the performance of these models on 8 clinical
prediction tasks. In both datasets, adapting the off-the-shelf $FM_{SM}$
matched the performance of GBM models locally trained on all data while
providing a 13% improvement in settings with few task-specific training labels.
With continued pretraining on local data, label efficiency substantially
improved, such that $FM_{SM}$ required fewer than 1% of training examples to
match the fully trained GBM's performance. Continued pretraining was also 60 to
90% more sample-efficient than training local foundation models from scratch.
Our findings show that adapting shared EHR foundation models across hospitals
provides improved prediction performance at less cost, underscoring the utility
of base foundation models as modular components to streamline the development
of healthcare AI. | Machine Learning |
What field is the article from? | Title: Enabling Decision-Support Systems through Automated Cell Tower Detection
Abstract: Cell phone coverage and high-speed service gaps persist in rural areas in
sub-Saharan Africa, impacting public access to mobile-based financial,
educational, and humanitarian services. Improving maps of telecommunications
infrastructure can help inform strategies to eliminate gaps in mobile coverage.
Deep neural networks, paired with remote sensing images, can be used for object
detection of cell towers and eliminate the need for inefficient and burdensome
manual mapping to find objects over large geographic regions. In this study, we
demonstrate a partially automated workflow to train an object detection model
to locate cell towers using OpenStreetMap (OSM) features and high-resolution
Maxar imagery. For model fine-tuning and evaluation, we curated a diverse
dataset of over 6,000 unique images of cell towers in 26 countries in eastern,
southern, and central Africa using automatically generated annotations from OSM
points. Our model achieves an average precision at 50% Intersection over Union
(IoU) (AP@50) of 81.2 with good performance across different geographies and
out-of-sample testing. Accurate localization of cell towers can yield more
accurate cell coverage maps, in turn enabling improved delivery of digital
services for decision-support applications. | Computer Vision |
What field is the article from? | Title: REST: Retrieval-Based Speculative Decoding
Abstract: We introduce Retrieval-Based Speculative Decoding (REST), a novel algorithm
designed to speed up language model generation. The key insight driving the
development of REST is the observation that the process of text generation
often includes certain common phases and patterns. Unlike previous methods that
rely on a draft language model for speculative decoding, REST harnesses the
power of retrieval to generate draft tokens. This method draws from the
reservoir of existing knowledge, retrieving and employing relevant tokens based
on the current context. Its plug-and-play nature allows for seamless
integration and acceleration of any language models, all without necessitating
additional training. When benchmarked on 7B and 13B language models in a
single-batch setting, REST achieves a significant speedup of 1.62X to 2.36X on
code or text generation. The code of REST is available at
https://github.com/FasterDecoding/REST. | Computational Linguistics |
What field is the article from? | Title: GTP-ViT: Efficient Vision Transformers via Graph-based Token Propagation
Abstract: Vision Transformers (ViTs) have revolutionized the field of computer vision,
yet their deployments on resource-constrained devices remain challenging due to
high computational demands. To expedite pre-trained ViTs, token pruning and
token merging approaches have been developed, which aim at reducing the number
of tokens involved in the computation. However, these methods still have some
limitations, such as image information loss from pruned tokens and inefficiency
in the token-matching process. In this paper, we introduce a novel Graph-based
Token Propagation (GTP) method to resolve the challenge of balancing model
efficiency and information preservation for efficient ViTs. Inspired by graph
summarization algorithms, GTP meticulously propagates less significant tokens'
information to spatially and semantically connected tokens that are of greater
importance. Consequently, the remaining few tokens serve as a summarization of
the entire token graph, allowing the method to reduce computational complexity
while preserving essential information of eliminated tokens. Combined with an
innovative token selection strategy, GTP can efficiently identify image tokens
to be propagated. Extensive experiments have validated GTP's effectiveness,
demonstrating both efficiency and performance improvements. Specifically, GTP
decreases the computational complexity of both DeiT-S and DeiT-B by up to 26%
with only a minimal 0.3% accuracy drop on ImageNet-1K without finetuning, and
remarkably surpasses the state-of-the-art token merging method on various
backbones at an even faster inference speed. The source code is available at
https://github.com/Ackesnal/GTP-ViT. | Computer Vision |
What field is the article from? | Title: Differentially Private Pre-Trained Model Fusion using Decentralized Federated Graph Matching
Abstract: Model fusion is becoming a crucial component in the context of
model-as-a-service scenarios, enabling the delivery of high-quality model
services to local users. However, this approach introduces privacy risks and
imposes certain limitations on its applications. Ensuring secure model exchange
and knowledge fusion among users becomes a significant challenge in this
setting. To tackle this issue, we propose PrivFusion, a novel architecture that
preserves privacy while facilitating model fusion under the constraints of
local differential privacy. PrivFusion leverages a graph-based structure,
enabling the fusion of models from multiple parties without necessitating
retraining. By employing randomized mechanisms, PrivFusion ensures privacy
guarantees throughout the fusion process. To enhance model privacy, our
approach incorporates a hybrid local differentially private mechanism and
decentralized federated graph matching, effectively protecting both activation
values and weights. Additionally, we introduce a perturbation filter adapter to
alleviate the impact of randomized noise, thereby preserving the utility of the
fused model. Through extensive experiments conducted on diverse image datasets
and real-world healthcare applications, we provide empirical evidence
showcasing the effectiveness of PrivFusion in maintaining model performance
while preserving privacy. Our contributions offer valuable insights and
practical solutions for secure and collaborative data analysis within the
domain of privacy-preserving model fusion. | Machine Learning |
What field is the article from? | Title: Large-scale Training of Foundation Models for Wearable Biosignals
Abstract: Tracking biosignals is crucial for monitoring wellness and preempting the
development of severe medical conditions. Today, wearable devices can
conveniently record various biosignals, creating the opportunity to monitor
health status without disruption to one's daily routine. Despite widespread use
of wearable devices and existing digital biomarkers, the absence of curated
data with annotated medical labels hinders the development of new biomarkers to
measure common health conditions. In fact, medical datasets are usually small
in comparison to other domains, which is an obstacle for developing neural
network models for biosignals. To address this challenge, we have employed
self-supervised learning using the unlabeled sensor data collected under
informed consent from the large longitudinal Apple Heart and Movement Study
(AHMS) to train foundation models for two common biosignals:
photoplethysmography (PPG) and electrocardiogram (ECG) recorded on Apple Watch.
We curated PPG and ECG datasets from AHMS that include data from ~141K
participants spanning ~3 years. Our self-supervised learning framework includes
participant level positive pair selection, stochastic augmentation module and a
regularized contrastive loss optimized with momentum training, and generalizes
well to both PPG and ECG modalities. We show that the pre-trained foundation
models readily encode information regarding participants' demographics and
health conditions. To the best of our knowledge, this is the first study that
builds foundation models using large-scale PPG and ECG data collected via
wearable consumer devices $\unicode{x2013}$ prior works have commonly used
smaller-size datasets collected in clinical and experimental settings. We
believe PPG and ECG foundation models can enhance future wearable devices by
reducing the reliance on labeled data and hold the potential to help the users
improve their health. | Machine Learning |
What field is the article from? | Title: Universal Knowledge Graph Embeddings
Abstract: A variety of knowledge graph embedding approaches have been developed. Most
of them obtain embeddings by learning the structure of the knowledge graph
within a link prediction setting. As a result, the embeddings reflect only the
semantics of a single knowledge graph, and embeddings for different knowledge
graphs are not aligned, e.g., they cannot be used to find similar entities
across knowledge graphs via nearest neighbor search. However, knowledge graph
embedding applications such as entity disambiguation require a more global
representation, i.e., a representation that is valid across multiple sources.
We propose to learn universal knowledge graph embeddings from large-scale
interlinked knowledge sources. To this end, we fuse large knowledge graphs
based on the owl:sameAs relation such that every entity is represented by a
unique identity. We instantiate our idea by computing universal embeddings
based on DBpedia and Wikidata yielding embeddings for about 180 million
entities, 15 thousand relations, and 1.2 billion triples. Moreover, we develop
a convenient API to provide embeddings as a service. Experiments on link
prediction show that universal knowledge graph embeddings encode better
semantics compared to embeddings computed on a single knowledge graph. For
reproducibility purposes, we provide our source code and datasets open access
at https://github.com/dice-group/Universal_Embeddings | Artificial Intelligence |
What field is the article from? | Title: Empowering Distributed Solutions in Renewable Energy Systems and Grid Optimization
Abstract: This study delves into the shift from centralized to decentralized approaches
in the electricity industry, with a particular focus on how machine learning
(ML) advancements play a crucial role in empowering renewable energy sources
and improving grid management. ML models have become increasingly important in
predicting renewable energy generation and consumption, utilizing various
techniques like artificial neural networks, support vector machines, and
decision trees. Furthermore, data preprocessing methods, such as data
splitting, normalization, decomposition, and discretization, are employed to
enhance prediction accuracy.
The incorporation of big data and ML into smart grids offers several
advantages, including heightened energy efficiency, more effective responses to
demand, and better integration of renewable energy sources. Nevertheless,
challenges like handling large data volumes, ensuring cybersecurity, and
obtaining specialized expertise must be addressed. The research investigates
various ML applications within the realms of solar energy, wind energy, and
electric distribution and storage, illustrating their potential to optimize
energy systems. To sum up, this research demonstrates the evolving landscape of
the electricity sector as it shifts from centralized to decentralized solutions
through the application of ML innovations and distributed decision-making,
ultimately shaping a more efficient and sustainable energy future. | Machine Learning |
What field is the article from? | Title: SortNet: Learning To Rank By a Neural-Based Sorting Algorithm
Abstract: The problem of relevance ranking consists of sorting a set of objects with
respect to a given criterion. Since users may prefer different relevance
criteria, the ranking algorithms should be adaptable to the user needs. Two
main approaches exist in literature for the task of learning to rank: 1) a
score function, learned by examples, which evaluates the properties of each
object yielding an absolute relevance value that can be used to order the
objects or 2) a pairwise approach, where a "preference function" is learned
using pairs of objects to define which one has to be ranked first. In this
paper, we present SortNet, an adaptive ranking algorithm which orders objects
using a neural network as a comparator. The neural network training set
provides examples of the desired ordering between pairs of items and it is
constructed by an iterative procedure which, at each iteration, adds the most
informative training examples. Moreover, the comparator adopts a connectionist
architecture that is particularly suited for implementing a preference
function. We also prove that such an architecture has the universal
approximation property and can implement a wide class of functions. Finally,
the proposed algorithm is evaluated on the LETOR dataset showing promising
performances in comparison with other state of the art algorithms. | Machine Learning |
What field is the article from? | Title: Establishing Performance Baselines in Fine-Tuning, Retrieval-Augmented Generation and Soft-Prompting for Non-Specialist LLM Users
Abstract: Research into methods for improving the performance of large language models
(LLMs) through fine-tuning, retrieval-augmented generation (RAG) and
soft-prompting has tended to focus on the use of highly technical or high-cost
techniques, making many of the newly discovered approaches comparatively
inaccessible to non-technical users. In this paper we tested an unmodified
version of GPT 3.5, a fine-tuned version, and the same unmodified model when
given access to a vectorised RAG database, both in isolation and in combination
with a basic, non-algorithmic soft prompt. In each case we tested the model's
ability to answer a set of 100 questions relating primarily to events that
occurred after September 2021 (the point at which GPT 3.5's training data set
ends). We found that if commercial platforms are used and default settings are
applied with no iteration in order to establish a baseline set of outputs, a
fine-tuned model outperforms GPT 3.5 Turbo, while the RAG approach
out-performed both. The application of a soft prompt significantly improved the
performance of each approach. | Information Retrieval |
What field is the article from? | Title: Detrimental Contexts in Open-Domain Question Answering
Abstract: For knowledge intensive NLP tasks, it has been widely accepted that accessing
more information is a contributing factor to improvements in the model's
end-to-end performance. However, counter-intuitively, too much context can have
a negative impact on the model when evaluated on common question answering (QA)
datasets. In this paper, we analyze how passages can have a detrimental effect
on retrieve-then-read architectures used in question answering. Our empirical
evidence indicates that the current read architecture does not fully leverage
the retrieved passages and significantly degrades its performance when using
the whole passages compared to utilizing subsets of them. Our findings
demonstrate that model accuracy can be improved by 10% on two popular QA
datasets by filtering out detrimental passages. Additionally, these outcomes
are attained by utilizing existing retrieval methods without further training
or data. We further highlight the challenges associated with identifying the
detrimental passages. First, even with the correct context, the model can make
an incorrect prediction, posing a challenge in determining which passages are
most influential. Second, evaluation typically considers lexical matching,
which is not robust to variations of correct answers. Despite these
limitations, our experimental results underscore the pivotal role of
identifying and removing these detrimental passages for the context-efficient
retrieve-then-read pipeline. Code and data are available at
https://github.com/xfactlab/emnlp2023-damaging-retrieval | Computational Linguistics |
What field is the article from? | Title: MAIRA-1: A specialised large multimodal model for radiology report generation
Abstract: We present a radiology-specific multimodal model for the task for generating
radiological reports from chest X-rays (CXRs). Our work builds on the idea that
large language model(s) can be equipped with multimodal capabilities through
alignment with pre-trained vision encoders. On natural images, this has been
shown to allow multimodal models to gain image understanding and description
capabilities. Our proposed model (MAIRA-1) leverages a CXR-specific image
encoder in conjunction with a fine-tuned large language model based on
Vicuna-7B, and text-based data augmentation, to produce reports with
state-of-the-art quality. In particular, MAIRA-1 significantly improves on the
radiologist-aligned RadCliQ metric and across all lexical metrics considered.
Manual review of model outputs demonstrates promising fluency and accuracy of
generated reports while uncovering failure modes not captured by existing
evaluation practices. More information and resources can be found on the
project website: https://aka.ms/maira. | Computational Linguistics |
What field is the article from? | Title: Gen2Sim: Scaling up Robot Learning in Simulation with Generative Models
Abstract: Generalist robot manipulators need to learn a wide variety of manipulation
skills across diverse environments. Current robot training pipelines rely on
humans to provide kinesthetic demonstrations or to program simulation
environments and to code up reward functions for reinforcement learning. Such
human involvement is an important bottleneck towards scaling up robot learning
across diverse tasks and environments. We propose Generation to Simulation
(Gen2Sim), a method for scaling up robot skill learning in simulation by
automating generation of 3D assets, task descriptions, task decompositions and
reward functions using large pre-trained generative models of language and
vision. We generate 3D assets for simulation by lifting open-world 2D
object-centric images to 3D using image diffusion models and querying LLMs to
determine plausible physics parameters. Given URDF files of generated and
human-developed assets, we chain-of-thought prompt LLMs to map these to
relevant task descriptions, temporal decompositions, and corresponding python
reward functions for reinforcement learning. We show Gen2Sim succeeds in
learning policies for diverse long horizon tasks, where reinforcement learning
with non temporally decomposed reward functions fails. Gen2Sim provides a
viable path for scaling up reinforcement learning for robot manipulators in
simulation, both by diversifying and expanding task and environment
development, and by facilitating the discovery of reinforcement-learned
behaviors through temporal task decomposition in RL. Our work contributes
hundreds of simulated assets, tasks and demonstrations, taking a step towards
fully autonomous robotic manipulation skill acquisition in simulation. | Robotics |
What field is the article from? | Title: Efficient Temporally-Aware DeepFake Detection using H.264 Motion Vectors
Abstract: Video DeepFakes are fake media created with Deep Learning (DL) that
manipulate a person's expression or identity. Most current DeepFake detection
methods analyze each frame independently, ignoring inconsistencies and
unnatural movements between frames. Some newer methods employ optical flow
models to capture this temporal aspect, but they are computationally expensive.
In contrast, we propose using the related but often ignored Motion Vectors
(MVs) and Information Masks (IMs) from the H.264 video codec, to detect
temporal inconsistencies in DeepFakes. Our experiments show that this approach
is effective and has minimal computational costs, compared with per-frame
RGB-only methods. This could lead to new, real-time temporally-aware DeepFake
detection methods for video calls and streaming. | Computer Vision |
What field is the article from? | Title: Taking it further: leveraging pseudo labels for field delineation across label-scarce smallholder regions
Abstract: Transfer learning allows for resource-efficient geographic transfer of
pre-trained field delineation models. However, the scarcity of labeled data for
complex and dynamic smallholder landscapes, particularly in Sub-Saharan Africa,
remains a major bottleneck for large-area field delineation. This study
explores opportunities of using sparse field delineation pseudo labels for
fine-tuning models across geographies and sensor characteristics. We build on a
FracTAL ResUNet trained for crop field delineation in India (median field size
of 0.24 ha) and use this pre-trained model to generate pseudo labels in
Mozambique (median field size of 0.06 ha). We designed multiple pseudo label
selection strategies and compared the quantities, area properties, seasonal
distribution, and spatial agreement of the pseudo labels against
human-annotated training labels (n = 1,512). We then used the human-annotated
labels and the pseudo labels for model fine-tuning and compared predictions
against human field annotations (n = 2,199). Our results indicate i) a good
baseline performance of the pre-trained model in both field delineation and
field size estimation, and ii) the added value of regional fine-tuning with
performance improvements in nearly all experiments. Moreover, we found iii)
substantial performance increases when using only pseudo labels (up to 77% of
the IoU increases and 68% of the RMSE decreases obtained by human labels), and
iv) additional performance increases when complementing human annotations with
pseudo labels. Pseudo labels can be efficiently generated at scale and thus
facilitate domain adaptation in label-scarce settings. The workflow presented
here is a stepping stone for overcoming the persisting data gaps in
heterogeneous smallholder agriculture of Sub-Saharan Africa, where labels are
commonly scarce. | Computer Vision |
What field is the article from? | Title: KwaiAgents: Generalized Information-seeking Agent System with Large Language Models
Abstract: Driven by curiosity, humans have continually sought to explore and understand
the world around them, leading to the invention of various tools to satiate
this inquisitiveness. Despite not having the capacity to process and memorize
vast amounts of information in their brains, humans excel in critical thinking,
planning, reflection, and harnessing available tools to interact with and
interpret the world, enabling them to find answers efficiently. The recent
advancements in large language models (LLMs) suggest that machines might also
possess the aforementioned human-like capabilities, allowing them to exhibit
powerful abilities even with a constrained parameter count. In this paper, we
introduce KwaiAgents, a generalized information-seeking agent system based on
LLMs. Within KwaiAgents, we propose an agent system that employs LLMs as its
cognitive core, which is capable of understanding a user's query, behavior
guidelines, and referencing external documents. The agent can also update and
retrieve information from its internal memory, plan and execute actions using a
time-aware search-browse toolkit, and ultimately provide a comprehensive
response. We further investigate the system's performance when powered by LLMs
less advanced than GPT-4, and introduce the Meta-Agent Tuning (MAT) framework,
designed to ensure even an open-sourced 7B or 13B model performs well among
many agent systems. We exploit both benchmark and human evaluations to
systematically validate these capabilities. Extensive experiments show the
superiority of our agent system compared to other autonomous agents and
highlight the enhanced generalized agent-abilities of our fine-tuned LLMs. | Artificial Intelligence |
What field is the article from? | Title: Efficient Toxic Content Detection by Bootstrapping and Distilling Large Language Models
Abstract: Toxic content detection is crucial for online services to remove
inappropriate content that violates community standards. To automate the
detection process, prior works have proposed varieties of machine learning (ML)
approaches to train Language Models (LMs) for toxic content detection. However,
both their accuracy and transferability across datasets are limited. Recently,
Large Language Models (LLMs) have shown promise in toxic content detection due
to their superior zero-shot and few-shot in-context learning ability as well as
broad transferability on ML tasks. However, efficiently designing prompts for
LLMs remains challenging. Moreover, the high run-time cost of LLMs may hinder
their deployments in production. To address these challenges, in this work, we
propose BD-LLM, a novel and efficient approach to Bootstrapping and Distilling
LLMs for toxic content detection. Specifically, we design a novel prompting
method named Decision-Tree-of-Thought (DToT) to bootstrap LLMs' detection
performance and extract high-quality rationales. DToT can automatically select
more fine-grained context to re-prompt LLMs when their responses lack
confidence. Additionally, we use the rationales extracted via DToT to fine-tune
student LMs. Our experimental results on various datasets demonstrate that DToT
can improve the accuracy of LLMs by up to 4.6%. Furthermore, student LMs
fine-tuned with rationales extracted via DToT outperform baselines on all
datasets with up to 16.9\% accuracy improvement, while being more than 60x
smaller than conventional LLMs. Finally, we observe that student LMs fine-tuned
with rationales exhibit better cross-dataset transferability. | Computational Linguistics |
What field is the article from? | Title: Representation Learning with Large Language Models for Recommendation
Abstract: Recommender systems have seen significant advancements with the influence of
deep learning and graph neural networks, particularly in capturing complex
user-item relationships. However, these graph-based recommenders heavily depend
on ID-based data, potentially disregarding valuable textual information
associated with users and items, resulting in less informative learned
representations. Moreover, the utilization of implicit feedback data introduces
potential noise and bias, posing challenges for the effectiveness of user
preference learning. While the integration of large language models (LLMs) into
traditional ID-based recommenders has gained attention, challenges such as
scalability issues, limitations in text-only reliance, and prompt input
constraints need to be addressed for effective implementation in practical
recommender systems. To address these challenges, we propose a model-agnostic
framework RLMRec that aims to enhance existing recommenders with LLM-empowered
representation learning. It proposes a recommendation paradigm that integrates
representation learning with LLMs to capture intricate semantic aspects of user
behaviors and preferences. RLMRec incorporates auxiliary textual signals,
develops a user/item profiling paradigm empowered by LLMs, and aligns the
semantic space of LLMs with the representation space of collaborative
relational signals through a cross-view alignment framework. This work further
establish a theoretical foundation demonstrating that incorporating textual
signals through mutual information maximization enhances the quality of
representations. In our evaluation, we integrate RLMRec with state-of-the-art
recommender models, while also analyzing its efficiency and robustness to noise
data. Our implementation codes are available at
https://github.com/HKUDS/RLMRec. | Information Retrieval |
What field is the article from? | Title: Generating Continuations in Multilingual Idiomatic Contexts
Abstract: The ability to process idiomatic or literal multiword expressions is a
crucial aspect of understanding and generating any language. The task of
generating contextually relevant continuations for narratives containing
idiomatic (or literal) expressions can allow us to test the ability of
generative language models (LMs) in understanding nuanced language containing
non-compositional figurative text. We conduct a series of experiments using
datasets in two distinct languages (English and Portuguese) under three
different training settings (zero-shot, few-shot, and fine-tuned). Our results
suggest that the models are only slightly better at generating continuations
for literal contexts than idiomatic contexts, with exceedingly small margins.
Furthermore, the models studied in this work perform equally well across both
languages, indicating the robustness of generative models in performing this
task. | Computational Linguistics |
What field is the article from? | Title: On Measuring Faithfulness of Natural Language Explanations
Abstract: Large language models (LLMs) can explain their own predictions, through
post-hoc or Chain-of-Thought (CoT) explanations. However the LLM could make up
reasonably sounding explanations that are unfaithful to its underlying
reasoning. Recent work has designed tests that aim to judge the faithfulness of
either post-hoc or CoT explanations. In this paper we argue that existing
faithfulness tests are not actually measuring faithfulness in terms of the
models' inner workings, but only evaluate their self-consistency on the output
level. The aims of our work are two-fold. i) We aim to clarify the status of
existing faithfulness tests in terms of model explainability, characterising
them as self-consistency tests instead. This assessment we underline by
constructing a Comparative Consistency Bank for self-consistency tests that for
the first time compares existing tests on a common suite of 11 open-source LLMs
and 5 datasets -- including ii) our own proposed self-consistency measure
CC-SHAP. CC-SHAP is a new fine-grained measure (not test) of LLM
self-consistency that compares a model's input contributions to answer
prediction and generated explanation. With CC-SHAP, we aim to take a step
further towards measuring faithfulness with a more interpretable and
fine-grained method. Code available at
\url{https://github.com/Heidelberg-NLP/CC-SHAP} | Computational Linguistics |
What field is the article from? | Title: Contrastive Deep Nonnegative Matrix Factorization for Community Detection
Abstract: Recently, nonnegative matrix factorization (NMF) has been widely adopted for
community detection, because of its better interpretability. However, the
existing NMF-based methods have the following three problems: 1) they directly
transform the original network into community membership space, so it is
difficult for them to capture the hierarchical information; 2) they often only
pay attention to the topology of the network and ignore its node attributes; 3)
it is hard for them to learn the global structure information necessary for
community detection. Therefore, we propose a new community detection algorithm,
named Contrastive Deep Nonnegative Matrix Factorization (CDNMF). Firstly, we
deepen NMF to strengthen its capacity for information extraction. Subsequently,
inspired by contrastive learning, our algorithm creatively constructs network
topology and node attributes as two contrasting views. Furthermore, we utilize
a debiased negative sampling layer and learn node similarity at the community
level, thereby enhancing the suitability of our model for community detection.
We conduct experiments on three public real graph datasets and the proposed
model has achieved better results than state-of-the-art methods. Code available
at https://github.com/6lyc/CDNMF.git. | Machine Learning |
What field is the article from? | Title: Cognitive bias in large language models: Cautious optimism meets anti-Panglossian meliorism
Abstract: Traditional discussions of bias in large language models focus on a
conception of bias closely tied to unfairness, especially as affecting
marginalized groups. Recent work raises the novel possibility of assessing the
outputs of large language models for a range of cognitive biases familiar from
research in judgment and decisionmaking. My aim in this paper is to draw two
lessons from recent discussions of cognitive bias in large language models:
cautious optimism about the prevalence of bias in current models coupled with
an anti-Panglossian willingness to concede the existence of some genuine biases
and work to reduce them. I draw out philosophical implications of this
discussion for the rationality of human cognitive biases as well as the role of
unrepresentative data in driving model biases. | Artificial Intelligence |
What field is the article from? | Title: Improving age prediction: Utilizing LSTM-based dynamic forecasting for data augmentation in multivariate time series analysis
Abstract: The high dimensionality and complexity of neuroimaging data necessitate large
datasets to develop robust and high-performing deep learning models. However,
the neuroimaging field is notably hampered by the scarcity of such datasets. In
this work, we proposed a data augmentation and validation framework that
utilizes dynamic forecasting with Long Short-Term Memory (LSTM) networks to
enrich datasets. We extended multivariate time series data by predicting the
time courses of independent component networks (ICNs) in both one-step and
recursive configurations. The effectiveness of these augmented datasets was
then compared with the original data using various deep learning models
designed for chronological age prediction tasks. The results suggest that our
approach improves model performance, providing a robust solution to overcome
the challenges presented by the limited size of neuroimaging datasets. | Machine Learning |
What field is the article from? | Title: RDBench: ML Benchmark for Relational Databases
Abstract: Benefiting from high-quality datasets and standardized evaluation metrics,
machine learning (ML) has achieved sustained progress and widespread
applications. However, while applying machine learning to relational databases
(RDBs), the absence of a well-established benchmark remains a significant
obstacle to the development of ML. To address this issue, we introduce ML
Benchmark For Relational Databases (RDBench), a standardized benchmark that
aims to promote reproducible ML research on RDBs that include multiple tables.
RDBench offers diverse RDB datasets of varying scales, domains, and relational
structures, organized into 4 levels. Notably, to simplify the adoption of
RDBench for diverse ML domains, for any given database, RDBench exposes three
types of interfaces including tabular data, homogeneous graphs, and
heterogeneous graphs, sharing the same underlying task definition. For the
first time, RDBench enables meaningful comparisons between ML methods from
diverse domains, ranging from XGBoost to Graph Neural Networks, under RDB
prediction tasks. We design multiple classification and regression tasks for
each RDB dataset and report averaged results over the same dataset, further
enhancing the robustness of the experimental findings. RDBench is implemented
with DBGym, a user-friendly platform for ML research and application on
databases, enabling benchmarking new ML methods with RDBench at ease. | Machine Learning |
What field is the article from? | Title: On the Exploitability of Reinforcement Learning with Human Feedback for Large Language Models
Abstract: Reinforcement Learning with Human Feedback (RLHF) is a methodology designed
to align Large Language Models (LLMs) with human preferences, playing an
important role in LLMs alignment. Despite its advantages, RLHF relies on human
annotators to rank the text, which can introduce potential security
vulnerabilities if any adversarial annotator (i.e., attackers) manipulates the
ranking score by up-ranking any malicious text to steer the LLM adversarially.
To assess the red-teaming of RLHF against human preference data poisoning, we
propose RankPoison, a poisoning attack method on candidates' selection of
preference rank flipping to reach certain malicious behaviors (e.g., generating
longer sequences, which can increase the computational cost). With poisoned
dataset generated by RankPoison, we can perform poisoning attacks on LLMs to
generate longer tokens without hurting the original safety alignment
performance. Moreover, applying RankPoison, we also successfully implement a
backdoor attack where LLMs can generate longer answers under questions with the
trigger word. Our findings highlight critical security challenges in RLHF,
underscoring the necessity for more robust alignment methods for LLMs. | Artificial Intelligence |
What field is the article from? | Title: Addressing Long-Horizon Tasks by Integrating Program Synthesis and State Machines
Abstract: Deep reinforcement learning excels in various domains but lacks
generalizability and interoperability. Programmatic RL methods (Trivedi et al.,
2021; Liu et al., 2023) reformulate solving RL tasks as synthesizing
interpretable programs that can be executed in the environments. Despite
encouraging results, these methods are limited to short-horizon tasks. On the
other hand, representing RL policies using state machines (Inala et al., 2020)
can inductively generalize to long-horizon tasks; however, it struggles to
scale up to acquire diverse and complex behaviors. This work proposes Program
Machine Policies (POMPs), which bridge the advantages of programmatic RL and
state machine policies, allowing for the representation of complex behaviors
and the address of long-term tasks. Specifically, we introduce a method that
can retrieve a set of effective, diverse, compatible programs. Then, we use
these programs as modes of a state machine and learn a transition function to
transition among mode programs, allowing for capturing long-horizon repetitive
behaviors. Our proposed framework outperforms programmatic RL and deep RL
baselines on various tasks and demonstrates the ability to generalize to even
longer horizons without any fine-tuning inductively. Ablation studies justify
the effectiveness of our proposed search algorithm for retrieving a set of
programs as modes. | Machine Learning |
What field is the article from? | Title: Non-Cross Diffusion for Semantic Consistency
Abstract: In diffusion models, deviations from a straight generative flow are a common
issue, resulting in semantic inconsistencies and suboptimal generations. To
address this challenge, we introduce `Non-Cross Diffusion', an innovative
approach in generative modeling for learning ordinary differential equation
(ODE) models. Our methodology strategically incorporates an ascending dimension
of input to effectively connect points sampled from two distributions with
uncrossed paths. This design is pivotal in ensuring enhanced semantic
consistency throughout the inference process, which is especially critical for
applications reliant on consistent generative flows, including various
distillation methods and deterministic sampling, which are fundamental in image
editing and interpolation tasks. Our empirical results demonstrate the
effectiveness of Non-Cross Diffusion, showing a substantial reduction in
semantic inconsistencies at different inference steps and a notable enhancement
in the overall performance of diffusion models. | Machine Learning |
What field is the article from? | Title: Open Visual Knowledge Extraction via Relation-Oriented Multimodality Model Prompting
Abstract: Images contain rich relational knowledge that can help machines understand
the world. Existing methods on visual knowledge extraction often rely on the
pre-defined format (e.g., sub-verb-obj tuples) or vocabulary (e.g., relation
types), restricting the expressiveness of the extracted knowledge. In this
work, we take a first exploration to a new paradigm of open visual knowledge
extraction. To achieve this, we present OpenVik which consists of an open
relational region detector to detect regions potentially containing relational
knowledge and a visual knowledge generator that generates format-free knowledge
by prompting the large multimodality model with the detected region of
interest. We also explore two data enhancement techniques for diversifying the
generated format-free visual knowledge. Extensive knowledge quality evaluations
highlight the correctness and uniqueness of the extracted open visual knowledge
by OpenVik. Moreover, integrating our extracted knowledge across various visual
reasoning applications shows consistent improvements, indicating the real-world
applicability of OpenVik. | Computational Linguistics |
What field is the article from? | Title: A Survey of Blockchain, Artificial Intelligence, and Edge Computing for Web 3.0
Abstract: Web 3.0, as the third generation of the World Wide Web, aims to solve
contemporary problems of trust, centralization, and data ownership. Driven by
the latest advances in cutting-edge technologies, Web 3.0 is moving towards a
more open, decentralized, intelligent, and interconnected network. However,
increasingly widespread data breaches have raised awareness of online privacy
and security of personal data. Additionally, since Web 3.0 is a sophisticated
and complex convergence, the technical details behind it are not as clear as
the characteristics it presents. In this survey, we conduct an in-depth
exploration of Web 3.0 from the perspectives of blockchain, artificial
intelligence, and edge computing. Specifically, we begin with summarizing the
evolution of the Internet and providing an overview of these three key
technological factors. Afterward, we provide a thorough analysis of each
technology separately, including its relevance to Web 3.0, key technology
components, and practical applications. We also propose decentralized storage
and computing solutions by exploring the integration of technologies. Finally,
we highlight the key challenges alongside potential research directions.
Through the combination and mutual complementation of multiple technologies,
Web 3.0 is expected to return more control and ownership of data and digital
assets back to users. | Cryptography and Security |
What field is the article from? | Title: RAEDiff: Denoising Diffusion Probabilistic Models Based Reversible Adversarial Examples Self-Generation and Self-Recovery
Abstract: Collected and annotated datasets, which are obtained through extensive
efforts, are effective for training Deep Neural Network (DNN) models. However,
these datasets are susceptible to be misused by unauthorized users, resulting
in infringement of Intellectual Property (IP) rights owned by the dataset
creators. Reversible Adversarial Exsamples (RAE) can help to solve the issues
of IP protection for datasets. RAEs are adversarial perturbed images that can
be restored to the original. As a cutting-edge approach, RAE scheme can serve
the purposes of preventing unauthorized users from engaging in malicious model
training, as well as ensuring the legitimate usage of authorized users.
Nevertheless, in the existing work, RAEs still rely on the embedded auxiliary
information for restoration, which may compromise their adversarial abilities.
In this paper, a novel self-generation and self-recovery method, named as
RAEDiff, is introduced for generating RAEs based on a Denoising Diffusion
Probabilistic Models (DDPM). It diffuses datasets into a Biased Gaussian
Distribution (BGD) and utilizes the prior knowledge of the DDPM for generating
and recovering RAEs. The experimental results demonstrate that RAEDiff
effectively self-generates adversarial perturbations for DNN models, including
Artificial Intelligence Generated Content (AIGC) models, while also exhibiting
significant self-recovery capabilities. | Cryptography and Security |
What field is the article from? | Title: A Bi-level Framework for Traffic Accident Duration Prediction: Leveraging Weather and Road Condition Data within a Practical Optimum Pipeline
Abstract: Due to the stochastic nature of events, predicting the duration of a traffic
incident presents a formidable challenge. Accurate duration estimation can
result in substantial advantages for commuters in selecting optimal routes and
for traffic management personnel in addressing non-recurring congestion issues.
In this study, we gathered accident duration, road conditions, and
meteorological data from a database of traffic accidents to check the
feasibility of a traffic accident duration pipeline without accident contextual
information data like accident severity and textual description. Multiple
machine learning models were employed to predict whether an accident's impact
on road traffic would be of a short-term or long-term nature, and then
utilizing a bimodal approach the precise duration of the incident's effect was
determined. Our binary classification random forest model distinguished between
short-term and long-term effects with an 83% accuracy rate, while the LightGBM
regression model outperformed other machine learning regression models with
Mean Average Error (MAE) values of 26.15 and 13.3 and RMSE values of 32.91 and
28.91 for short and long-term accident duration prediction, respectively. Using
the optimal classification and regression model identified in the preceding
section, we then construct an end-to-end pipeline to incorporate the entire
process. The results of both separate and combined approaches were comparable
with previous works, which shows the applicability of only using static
features for predicting traffic accident duration. The SHAP value analysis
identified weather conditions, wind chill and wind speed as the most
influential factors in determining the duration of an accident. | Artificial Intelligence |
What field is the article from? | Title: Not All Large Language Models (LLMs) Succumb to the "Reversal Curse": A Comparative Study of Deductive Logical Reasoning in BERT and GPT Models
Abstract: The "Reversal Curse" refers to the scenario where auto-regressive decoder
large language models (LLMs), such as ChatGPT, trained on "A is B" fail to
learn "B is A", demonstrating a basic failure of logical deduction. This raises
a red flag in the use of GPT models for certain general tasks such as
constructing knowledge graphs, considering their adherence to this symmetric
principle. In our study, we examined a bidirectional LLM, BERT, and found that
it is immune to the reversal curse. Driven by ongoing efforts to construct
biomedical knowledge graphs with LLMs, we also embarked on evaluating more
complex but essential deductive reasoning capabilities. This process included
first training encoder and decoder language models to master the intersection
($\cap$) and union ($\cup$) operations on two sets and then moving on to assess
their capability to infer different combinations of union ($\cup$) and
intersection ($\cap$) operations on three newly created sets. The findings
showed that while both encoder and decoder language models, trained for tasks
involving two sets (union/intersection), were proficient in such scenarios,
they encountered difficulties when dealing with operations that included three
sets (various combinations of union and intersection). Our research highlights
the distinct characteristics of encoder and decoder models in simple and
complex logical reasoning. In practice, the choice between BERT and GPT should
be guided by the specific requirements and nature of the task at hand,
leveraging their respective strengths in bidirectional context comprehension
and sequence prediction. | Computational Linguistics |
What field is the article from? | Title: A comparative analysis between Conformer-Transducer, Whisper, and wav2vec2 for improving the child speech recognition
Abstract: Automatic Speech Recognition (ASR) systems have progressed significantly in
their performance on adult speech data; however, transcribing child speech
remains challenging due to the acoustic differences in the characteristics of
child and adult voices. This work aims to explore the potential of adapting
state-of-the-art Conformer-transducer models to child speech to improve child
speech recognition performance. Furthermore, the results are compared with
those of self-supervised wav2vec2 models and semi-supervised multi-domain
Whisper models that were previously finetuned on the same data. We demonstrate
that finetuning Conformer-transducer models on child speech yields significant
improvements in ASR performance on child speech, compared to the non-finetuned
models. We also show Whisper and wav2vec2 adaptation on different child speech
datasets. Our detailed comparative analysis shows that wav2vec2 provides the
most consistent performance improvements among the three methods studied. | Computational Linguistics |
What field is the article from? | Title: Everything of Thoughts: Defying the Law of Penrose Triangle for Thought Generation
Abstract: Recent advancements in Large Language Models (LLMs) have revolutionized
decision-making by breaking down complex problems into more manageable language
sequences referred to as ``thoughts''. An effective thought design should
consider three key perspectives: performance, efficiency, and flexibility.
However, existing thought can at most exhibit two of these attributes. To
address these limitations, we introduce a novel thought prompting approach
called ``Everything of Thoughts'' (XoT) to defy the law of ``Penrose triangle
of existing thought paradigms. XoT leverages pretrained reinforcement learning
and Monte Carlo Tree Search (MCTS) to incorporate external domain knowledge
into thoughts, thereby enhancing LLMs' capabilities and enabling them to
generalize to unseen problems efficiently. Through the utilization of the
MCTS-LLM collaborative thought revision framework, this approach autonomously
produces high-quality comprehensive cognitive mappings with minimal LLM
interactions. Additionally, XoT empowers LLMs to engage in unconstrained
thinking, allowing for flexible cognitive mappings for problems with multiple
solutions. We evaluate XoT on several challenging multi-solution
problem-solving tasks, including Game of 24, 8-Puzzle, and Pocket Cube. Our
results demonstrate that XoT significantly outperforms existing approaches.
Notably, XoT can yield multiple solutions with just one LLM call, showcasing
its remarkable proficiency in addressing complex problems across diverse
domains. | Artificial Intelligence |
What field is the article from? | Title: Lightweight Face Recognition: An Improved MobileFaceNet Model
Abstract: This paper presents an extensive exploration and comparative analysis of
lightweight face recognition (FR) models, specifically focusing on
MobileFaceNet and its modified variant, MMobileFaceNet. The need for efficient
FR models on devices with limited computational resources has led to the
development of models with reduced memory footprints and computational demands
without sacrificing accuracy. Our research delves into the impact of dataset
selection, model architecture, and optimization algorithms on the performance
of FR models. We highlight our participation in the EFaR-2023 competition,
where our models showcased exceptional performance, particularly in categories
restricted by the number of parameters. By employing a subset of the Webface42M
dataset and integrating sharpness-aware minimization (SAM) optimization, we
achieved significant improvements in accuracy across various benchmarks,
including those that test for cross-pose, cross-age, and cross-ethnicity
performance. The results underscore the efficacy of our approach in crafting
models that are not only computationally efficient but also maintain high
accuracy in diverse conditions. | Computer Vision |
What field is the article from? | Title: One-shot Localization and Segmentation of Medical Images with Foundation Models
Abstract: Recent advances in Vision Transformers (ViT) and Stable Diffusion (SD) models
with their ability to capture rich semantic features of the image have been
used for image correspondence tasks on natural images. In this paper, we
examine the ability of a variety of pre-trained ViT (DINO, DINOv2, SAM, CLIP)
and SD models, trained exclusively on natural images, for solving the
correspondence problems on medical images. While many works have made a case
for in-domain training, we show that the models trained on natural images can
offer good performance on medical images across different modalities
(CT,MR,Ultrasound) sourced from various manufacturers, over multiple anatomical
regions (brain, thorax, abdomen, extremities), and on wide variety of tasks.
Further, we leverage the correspondence with respect to a template image to
prompt a Segment Anything (SAM) model to arrive at single shot segmentation,
achieving dice range of 62%-90% across tasks, using just one image as
reference. We also show that our single-shot method outperforms the recently
proposed few-shot segmentation method - UniverSeg (Dice range 47%-80%) on most
of the semantic segmentation tasks(six out of seven) across medical imaging
modalities. | Computer Vision |
What field is the article from? | Title: IA-LSTM: Interaction-Aware LSTM for Pedestrian Trajectory Prediction
Abstract: Predicting the trajectory of pedestrians in crowd scenarios is indispensable
in self-driving or autonomous mobile robot field because estimating the future
locations of pedestrians around is beneficial for policy decision to avoid
collision. It is a challenging issue because humans have different walking
motions and the interactions between humans and objects in the current
environment, especially between human themselves, are complex. Previous
researches have focused on how to model the human-human interactions, however,
neglecting the relative importance of interactions. In order to address this
issue, we introduce a novel mechanism based on the correntropy, which not only
can measure the relative importance of human-human interactions, but also can
build personal space for each pedestrian. We further propose an Interaction
Module including this data-driven mechanism that can effectively extract
feature representations of dynamic human-human interactions in the scene and
calculate corresponding weights to represent the importance of different
interactions. To share such social messages among pedestrians, we design an
interaction-aware architecture based on the Long Short-Term Memory (LSTM)
network for trajectory prediction. We demonstrate the performance of our model
on two public datasets and the experimental results demonstrate that our model
can achieve better performance than several latest methods with good
performance. | Computer Vision |
What field is the article from? | Title: A Comprehensive and Reliable Feature Attribution Method: Double-sided Remove and Reconstruct (DoRaR)
Abstract: The limited transparency of the inner decision-making mechanism in deep
neural networks (DNN) and other machine learning (ML) models has hindered their
application in several domains. In order to tackle this issue, feature
attribution methods have been developed to identify the crucial features that
heavily influence decisions made by these black box models. However, many
feature attribution methods have inherent downsides. For example, one category
of feature attribution methods suffers from the artifacts problem, which feeds
out-of-distribution masked inputs directly through the classifier that was
originally trained on natural data points. Another category of feature
attribution method finds explanations by using jointly trained feature
selectors and predictors. While avoiding the artifacts problem, this new
category suffers from the Encoding Prediction in the Explanation (EPITE)
problem, in which the predictor's decisions rely not on the features, but on
the masks that selects those features. As a result, the credibility of
attribution results is undermined by these downsides. In this research, we
introduce the Double-sided Remove and Reconstruct (DoRaR) feature attribution
method based on several improvement methods that addresses these issues. By
conducting thorough testing on MNIST, CIFAR10 and our own synthetic dataset, we
demonstrate that the DoRaR feature attribution method can effectively bypass
the above issues and can aid in training a feature selector that outperforms
other state-of-the-art feature attribution methods. Our code is available at
https://github.com/dxq21/DoRaR. | Machine Learning |
What field is the article from? | Title: Knowledgeable Preference Alignment for LLMs in Domain-specific Question Answering
Abstract: Recently, the development of large language models (LLMs) has attracted wide
attention in academia and industry. Deploying LLMs to real scenarios is one of
the key directions in the current Internet industry. In this paper, we present
a novel pipeline to apply LLMs for domain-specific question answering (QA) that
incorporates domain knowledge graphs (KGs), addressing an important direction
of LLM application. As a real-world application, the content generated by LLMs
should be user-friendly to serve the customers. Additionally, the model needs
to utilize domain knowledge properly to generate reliable answers. These two
issues are the two major difficulties in the LLM application as vanilla
fine-tuning can not adequately address them. We think both requirements can be
unified as the model preference problem that needs to align with humans to
achieve practical application. Thus, we introduce Knowledgeable Preference
AlignmenT (KnowPAT), which constructs two kinds of preference set called style
preference set and knowledge preference set respectively to tackle the two
issues. Besides, we design a new alignment objective to align the LLM
preference with human preference, aiming to train a better LLM for
real-scenario domain-specific QA to generate reliable and user-friendly
answers. Adequate experiments and comprehensive with 15 baseline methods
demonstrate that our KnowPAT is an outperforming pipeline for real-scenario
domain-specific QA with LLMs. Our code is open-source at
https://github.com/zjukg/KnowPAT. | Computational Linguistics |
What field is the article from? | Title: Mixture-of-Linear-Experts for Long-term Time Series Forecasting
Abstract: Long-term time series forecasting (LTSF) aims to predict future values of a
time series given the past values. The current state-of-the-art (SOTA) on this
problem is attained in some cases by linear-centric models, which primarily
feature a linear mapping layer. However, due to their inherent simplicity, they
are not able to adapt their prediction rules to periodic changes in time series
patterns. To address this challenge, we propose a Mixture-of-Experts-style
augmentation for linear-centric models and propose Mixture-of-Linear-Experts
(MoLE). Instead of training a single model, MoLE trains multiple linear-centric
models (i.e., experts) and a router model that weighs and mixes their outputs.
While the entire framework is trained end-to-end, each expert learns to
specialize in a specific temporal pattern, and the router model learns to
compose the experts adaptively. Experiments show that MoLE reduces forecasting
error of linear-centric models, including DLinear, RLinear, and RMLP, in over
78% of the datasets and settings we evaluated. By using MoLE existing
linear-centric models can achieve SOTA LTSF results in 68% of the experiments
that PatchTST reports and we compare to, whereas existing single-head
linear-centric models achieve SOTA results in only 25% of cases. Additionally,
MoLE models achieve SOTA in all settings for the newly released Weather2K
datasets. | Machine Learning |
What field is the article from? | Title: Cross-Modal Information-Guided Network using Contrastive Learning for Point Cloud Registration
Abstract: The majority of point cloud registration methods currently rely on extracting
features from points. However, these methods are limited by their dependence on
information obtained from a single modality of points, which can result in
deficiencies such as inadequate perception of global features and a lack of
texture information. Actually, humans can employ visual information learned
from 2D images to comprehend the 3D world. Based on this fact, we present a
novel Cross-Modal Information-Guided Network (CMIGNet), which obtains global
shape perception through cross-modal information to achieve precise and robust
point cloud registration. Specifically, we first incorporate the projected
images from the point clouds and fuse the cross-modal features using the
attention mechanism. Furthermore, we employ two contrastive learning
strategies, namely overlapping contrastive learning and cross-modal contrastive
learning. The former focuses on features in overlapping regions, while the
latter emphasizes the correspondences between 2D and 3D features. Finally, we
propose a mask prediction module to identify keypoints in the point clouds.
Extensive experiments on several benchmark datasets demonstrate that our
network achieves superior registration performance. | Computer Vision |
What field is the article from? | Title: Retrieval-Augmented Code Generation for Universal Information Extraction
Abstract: Information Extraction (IE) aims to extract structural knowledge (e.g.,
entities, relations, events) from natural language texts, which brings
challenges to existing methods due to task-specific schemas and complex text
expressions. Code, as a typical kind of formalized language, is capable of
describing structural knowledge under various schemas in a universal way. On
the other hand, Large Language Models (LLMs) trained on both codes and texts
have demonstrated powerful capabilities of transforming texts into codes, which
provides a feasible solution to IE tasks. Therefore, in this paper, we propose
a universal retrieval-augmented code generation framework based on LLMs, called
Code4UIE, for IE tasks. Specifically, Code4UIE adopts Python classes to define
task-specific schemas of various structural knowledge in a universal way. By so
doing, extracting knowledge under these schemas can be transformed into
generating codes that instantiate the predefined Python classes with the
information in texts. To generate these codes more precisely, Code4UIE adopts
the in-context learning mechanism to instruct LLMs with examples. In order to
obtain appropriate examples for different tasks, Code4UIE explores several
example retrieval strategies, which can retrieve examples semantically similar
to the given texts. Extensive experiments on five representative IE tasks
across nine datasets demonstrate the effectiveness of the Code4UIE framework. | Artificial Intelligence |
What field is the article from? | Title: Probabilistic Tree-of-thought Reasoning for Answering Knowledge-intensive Complex Questions
Abstract: Large language models (LLMs) are capable of answering knowledge-intensive
complex questions with chain-of-thought (CoT) reasoning. However, they tend to
generate factually incorrect reasoning steps when the required knowledge is not
available or up-to-date in models' parameters. Recent works turn to retrieving
external knowledge to augment CoT reasoning. Despite being promising, these
chain-based methods suffer from: 1) Negative retrieval. Unnecessary or
incorrect retrieval may mislead the reasoning; 2) Limited sight. Lacking the
ability to look backward or forward, a local error in one step will propagate
along the chain.
In this paper, we propose a novel approach: Probabilistic Tree-of-thought
Reasoning (ProbTree). First, LLMs translate a complex question into a query
tree, in which each non-root node denotes a sub-question of its parent node.
Then, probabilistic reasoning is conducted over the tree, by solving questions
from leaf to root considering the confidence of both question decomposing and
answering. During reasoning, for leaf nodes, LLMs choose a more confident
answer from Closed-book QA that employs parametric knowledge and Open-book QA
that employs retrieved external knowledge, thus eliminating the negative
retrieval problem. For non-leaf nodes, with the hierarchical structure, LLMs
have broader sights and are able to globally reason with the information from
child nodes, thus recovering from local errors. The experiments on three
Complex QA datasets under the open-domain setting show that our approach
outperforms SOTA methods significantly, demonstrating the effect of
probabilistic tree-of-thought reasoning. | Computational Linguistics |
What field is the article from? | Title: AI-assisted Learning for Electronic Engineering Courses in High Education
Abstract: This study evaluates the efficacy of ChatGPT as an AI teaching and learning
support tool in an integrated circuit systems course at a higher education
institution in an Asian country. Various question types were completed, and
ChatGPT responses were assessed to gain valuable insights for further
investigation. The objective is to assess ChatGPT's ability to provide
insights, personalized support, and interactive learning experiences in
engineering education. The study includes the evaluation and reflection of
different stakeholders: students, lecturers, and engineers. The findings of
this study shed light on the benefits and limitations of ChatGPT as an AI tool,
paving the way for innovative learning approaches in technical disciplines.
Furthermore, the study contributes to our understanding of how digital
transformation is likely to unfold in the education sector. | Computers and Society |
What field is the article from? | Title: Language Agents with Reinforcement Learning for Strategic Play in the Werewolf Game
Abstract: Agents built with large language models (LLMs) have recently achieved great
advancements. However, most of the efforts focus on single-agent or cooperative
settings, leaving more general multi-agent environments underexplored. We
propose a new framework powered by reinforcement learning (RL) to develop
strategic language agents, i.e., LLM-based agents with strategic thinking
ability, for a popular language game, Werewolf. Werewolf is a social deduction
game with hidden roles that involves both cooperation and competition and
emphasizes deceptive communication and diverse gameplay. Our agent tackles this
game by first using LLMs to reason about potential deceptions and generate a
set of strategically diverse actions. Then an RL policy, which selects an
action from the candidates, is learned by population-based training to enhance
the agents' decision-making ability. By combining LLMs with the RL policy, our
agent produces a variety of emergent strategies, achieves the highest win rate
against other LLM-based agents, and stays robust against adversarial human
players in the Werewolf game. | Artificial Intelligence |
What field is the article from? | Title: Efficient Pre-training for Localized Instruction Generation of Videos
Abstract: Procedural videos show step-by-step demonstrations of tasks like recipe
preparation. Understanding such videos is challenging, involving the precise
localization of steps and the generation of textual instructions. Manually
annotating steps and writing instructions is costly, which limits the size of
current datasets and hinders effective learning. Leveraging large but noisy
video-transcript datasets for pre-training can boost performance, but demands
significant computational resources. Furthermore, transcripts contain
irrelevant content and exhibit style variation compared to instructions written
by human annotators. To mitigate both issues, we propose a technique,
Sieve-&-Swap, to automatically curate a smaller dataset: (i) Sieve filters
irrelevant transcripts and (ii) Swap enhances the quality of the text
instruction by automatically replacing the transcripts with human-written
instructions from a text-only recipe dataset. The curated dataset, three orders
of magnitude smaller than current web-scale datasets, enables efficient
training of large-scale models with competitive performance. We complement our
Sieve-\&-Swap approach with a Procedure Transformer (ProcX) for end-to-end step
localization and instruction generation for procedural videos. When this model
is pre-trained on our curated dataset, it achieves state-of-the-art performance
in zero-shot and finetuning settings on YouCook2 and Tasty, while using a
fraction of the computational resources. | Computer Vision |
What field is the article from? | Title: Decision-Making for Autonomous Vehicles with Interaction-Aware Behavioral Prediction and Social-Attention Neural Network
Abstract: Autonomous vehicles need to accomplish their tasks while interacting with
human drivers in traffic. It is thus crucial to equip autonomous vehicles with
artificial reasoning to better comprehend the intentions of the surrounding
traffic, thereby facilitating the accomplishments of the tasks. In this work,
we propose a behavioral model that encodes drivers' interacting intentions into
latent social-psychological parameters. Leveraging a Bayesian filter, we
develop a receding-horizon optimization-based controller for autonomous vehicle
decision-making which accounts for the uncertainties in the interacting
drivers' intentions. For online deployment, we design a neural network
architecture based on the attention mechanism which imitates the behavioral
model with online estimated parameter priors. We also propose a decision tree
search algorithm to solve the decision-making problem online. The proposed
behavioral model is then evaluated in terms of its capabilities for real-world
trajectory prediction. We further conduct extensive evaluations of the proposed
decision-making module, in forced highway merging scenarios, using both
simulated environments and real-world traffic datasets. The results demonstrate
that our algorithms can complete the forced merging tasks in various traffic
conditions while ensuring driving safety. | Artificial Intelligence |
What field is the article from? | Title: Debiasing Multimodal Models via Causal Information Minimization
Abstract: Most existing debiasing methods for multimodal models, including causal
intervention and inference methods, utilize approximate heuristics to represent
the biases, such as shallow features from early stages of training or unimodal
features for multimodal tasks like VQA, etc., which may not be accurate. In
this paper, we study bias arising from confounders in a causal graph for
multimodal data and examine a novel approach that leverages causally-motivated
information minimization to learn the confounder representations. Robust
predictive features contain diverse information that helps a model generalize
to out-of-distribution data. Hence, minimizing the information content of
features obtained from a pretrained biased model helps learn the simplest
predictive features that capture the underlying data distribution. We treat
these features as confounder representations and use them via methods motivated
by causal theory to remove bias from models. We find that the learned
confounder representations indeed capture dataset biases, and the proposed
debiasing methods improve out-of-distribution (OOD) performance on multiple
multimodal datasets without sacrificing in-distribution performance.
Additionally, we introduce a novel metric to quantify the sufficiency of
spurious features in models' predictions that further demonstrates the
effectiveness of our proposed methods. Our code is available at:
https://github.com/Vaidehi99/CausalInfoMin | Machine Learning |
What field is the article from? | Title: Federated Learning for Generalization, Robustness, Fairness: A Survey and Benchmark
Abstract: Federated learning has emerged as a promising paradigm for privacy-preserving
collaboration among different parties. Recently, with the popularity of
federated learning, an influx of approaches have delivered towards different
realistic challenges. In this survey, we provide a systematic overview of the
important and recent developments of research on federated learning. Firstly,
we introduce the study history and terminology definition of this area. Then,
we comprehensively review three basic lines of research: generalization,
robustness, and fairness, by introducing their respective background concepts,
task settings, and main challenges. We also offer a detailed overview of
representative literature on both methods and datasets. We further benchmark
the reviewed methods on several well-known datasets. Finally, we point out
several open issues in this field and suggest opportunities for further
research. We also provide a public website to continuously track developments
in this fast advancing field: https://github.com/WenkeHuang/MarsFL. | Machine Learning |
What field is the article from? | Title: Towards Concept-Aware Large Language Models
Abstract: Concepts play a pivotal role in various human cognitive functions, including
learning, reasoning and communication. However, there is very little work on
endowing machines with the ability to form and reason with concepts. In
particular, state-of-the-art large language models (LLMs) work at the level of
tokens, not concepts.
In this work, we analyze how well contemporary LLMs capture human concepts
and their structure. We then discuss ways to develop concept-aware LLMs, taking
place at different stages of the pipeline. We sketch a method for pretraining
LLMs using concepts, and also explore the simpler approach that uses the output
of existing LLMs. Despite its simplicity, our proof-of-concept is shown to
better match human intuition, as well as improve the robustness of predictions.
These preliminary results underscore the promise of concept-aware LLMs. | Computational Linguistics |
What field is the article from? | Title: nerblackbox: A High-level Library for Named Entity Recognition in Python
Abstract: We present nerblackbox, a python library to facilitate the use of
state-of-the-art transformer-based models for named entity recognition. It
provides simple-to-use yet powerful methods to access data and models from a
wide range of sources, for fully automated model training and evaluation as
well as versatile model inference. While many technical challenges are solved
and hidden from the user by default, nerblackbox also offers fine-grained
control and a rich set of customizable features. It is thus targeted both at
application-oriented developers as well as machine learning experts and
researchers. | Computational Linguistics |
What field is the article from? | Title: More Robots are Coming: Large Multimodal Models (ChatGPT) can Solve Visually Diverse Images of Parsons Problems
Abstract: The advent of large language models is reshaping computing education. Recent
research has demonstrated that these models can produce better explanations
than students, answer multiple-choice questions at or above the class average,
and generate code that can pass automated tests in introductory courses. These
capabilities have prompted instructors to rapidly adapt their courses and
assessment methods to accommodate changes in learning objectives and the
potential for academic integrity violations. While some scholars have advocated
for the integration of visual problems as a safeguard against the capabilities
of language models, new multimodal language models now have vision and language
capabilities that may allow them to analyze and solve visual problems. In this
paper, we evaluate the performance of two large multimodal models on visual
assignments, with a specific focus on Parsons problems presented across diverse
visual representations. Our results show that GPT-4V solved 96.7\% of these
visual problems, struggling minimally with a single Parsons problem.
Conversely, Bard performed poorly by only solving 69.2\% of problems,
struggling with common issues like hallucinations and refusals. These findings
suggest that merely transitioning to visual programming problems might not be a
panacea to issues of academic integrity in the generative AI era. | Computational Linguistics |
What field is the article from? | Title: Methods to Estimate Large Language Model Confidence
Abstract: Large Language Models have difficulty communicating uncertainty, which is a
significant obstacle to applying LLMs to complex medical tasks. This study
evaluates methods to measure LLM confidence when suggesting a diagnosis for
challenging clinical vignettes. GPT4 was asked a series of challenging case
questions using Chain of Thought and Self Consistency prompting. Multiple
methods were investigated to assess model confidence and evaluated on their
ability to predict the models observed accuracy. The methods evaluated were
Intrinsic Confidence, SC Agreement Frequency and CoT Response Length. SC
Agreement Frequency correlated with observed accuracy, yielding a higher Area
under the Receiver Operating Characteristic Curve compared to Intrinsic
Confidence and CoT Length analysis. SC agreement is the most useful proxy for
model confidence, especially for medical diagnosis. Model Intrinsic Confidence
and CoT Response Length exhibit a weaker ability to differentiate between
correct and incorrect answers, preventing them from being reliable and
interpretable markers for model confidence. We conclude GPT4 has a limited
ability to assess its own diagnostic accuracy. SC Agreement Frequency is the
most useful method to measure GPT4 confidence. | Computational Linguistics |
What field is the article from? | Title: Green Resilience of Cyber-Physical Systems
Abstract: Cyber-Physical System (CPS) represents systems that join both hardware and
software components to perform real-time services. Maintaining the system's
reliability is critical to the continuous delivery of these services. However,
the CPS running environment is full of uncertainties and can easily lead to
performance degradation. As a result, the need for a recovery technique is
highly needed to achieve resilience in the system, with keeping in mind that
this technique should be as green as possible. This early doctorate proposal,
suggests a game theory solution to achieve resilience and green in CPS. Game
theory has been known for its fast performance in decision-making, helping the
system to choose what maximizes its payoffs. The proposed game model is
described over a real-life collaborative artificial intelligence system (CAIS),
that involves robots with humans to achieve a common goal. It shows how the
expected results of the system will achieve the resilience of CAIS with
minimized CO2 footprint. | Software Engineering |
What field is the article from? | Title: Quilt: Robust Data Segment Selection against Concept Drifts
Abstract: Continuous machine learning pipelines are common in industrial settings where
models are periodically trained on data streams. Unfortunately, concept drifts
may occur in data streams where the joint distribution of the data X and label
y, P(X, y), changes over time and possibly degrade model accuracy. Existing
concept drift adaptation approaches mostly focus on updating the model to the
new data possibly using ensemble techniques of previous models and tend to
discard the drifted historical data. However, we contend that explicitly
utilizing the drifted data together leads to much better model accuracy and
propose Quilt, a data-centric framework for identifying and selecting data
segments that maximize model accuracy. To address the potential downside of
efficiency, Quilt extends existing data subset selection techniques, which can
be used to reduce the training data without compromising model accuracy. These
techniques cannot be used as is because they only assume virtual drifts where
the posterior probabilities P(y|X) are assumed not to change. In contrast, a
key challenge in our setup is to also discard undesirable data segments with
concept drifts. Quilt thus discards drifted data segments and selects data
segment subsets holistically for accurate and efficient model training. The two
operations use gradient-based scores, which have little computation overhead.
In our experiments, we show that Quilt outperforms state-of-the-art drift
adaptation and data selection baselines on synthetic and real datasets. | Machine Learning |
What field is the article from? | Title: Coordination-free Decentralised Federated Learning on Complex Networks: Overcoming Heterogeneity
Abstract: Federated Learning (FL) is a well-known framework for successfully performing
a learning task in an edge computing scenario where the devices involved have
limited resources and incomplete data representation. The basic assumption of
FL is that the devices communicate directly or indirectly with a parameter
server that centrally coordinates the whole process, overcoming several
challenges associated with it. However, in highly pervasive edge scenarios, the
presence of a central controller that oversees the process cannot always be
guaranteed, and the interactions (i.e., the connectivity graph) between devices
might not be predetermined, resulting in a complex network structure. Moreover,
the heterogeneity of data and devices further complicates the learning process.
This poses new challenges from a learning standpoint that we address by
proposing a communication-efficient Decentralised Federated Learning (DFL)
algorithm able to cope with them. Our solution allows devices communicating
only with their direct neighbours to train an accurate model, overcoming the
heterogeneity induced by data and different training histories. Our results
show that the resulting local models generalise better than those trained with
competing approaches, and do so in a more communication-efficient way. | Machine Learning |
What field is the article from? | Title: Kuro Siwo: 12.1 billion $m^2$ under the water. A global multi-temporal satellite dataset for rapid flood mapping
Abstract: Global floods, exacerbated by climate change, pose severe threats to human
life, infrastructure, and the environment. This urgency is highlighted by
recent catastrophic events in Pakistan and New Zealand, underlining the
critical need for precise flood mapping for guiding restoration efforts,
understanding vulnerabilities, and preparing for future events. While Synthetic
Aperture Radar (SAR) offers day-and-night, all-weather imaging capabilities,
harnessing it for deep learning is hindered by the absence of a large annotated
dataset. To bridge this gap, we introduce Kuro Siwo, a meticulously curated
multi-temporal dataset, spanning 32 flood events globally. Our dataset maps
more than 63 billion m2 of land, with 12.1 billion of them being either a
flooded area or a permanent water body. Kuro Siwo stands out for its
unparalleled annotation quality to facilitate rapid flood mapping in a
supervised setting. We also augment learning by including a large unlabeled set
of SAR samples, aimed at self-supervised pretraining. We provide an extensive
benchmark and strong baselines for a diverse set of flood events from Europe,
America, Africa and Australia. Our benchmark demonstrates the quality of Kuro
Siwo annotations, training models that can achieve $\approx$ 85% and $\approx$
87% in F1-score for flooded areas and general water detection respectively.
This work calls on the deep learning community to develop solution-driven
algorithms for rapid flood mapping, with the potential to aid civil protection
and humanitarian agencies amid climate change challenges. Our code and data
will be made available at https://github.com/Orion-AI-Lab/KuroSiwo | Computer Vision |
What field is the article from? | Title: Masking Hyperspectral Imaging Data with Pretrained Models
Abstract: The presence of undesired background areas associated with potential noise
and unknown spectral characteristics degrades the performance of hyperspectral
data processing. Masking out unwanted regions is key to addressing this issue.
Processing only regions of interest yields notable improvements in terms of
computational costs, required memory, and overall performance. The proposed
processing pipeline encompasses two fundamental parts: regions of interest mask
generation, followed by the application of hyperspectral data processing
techniques solely on the newly masked hyperspectral cube. The novelty of our
work lies in the methodology adopted for the preliminary image segmentation. We
employ the Segment Anything Model (SAM) to extract all objects within the
dataset, and subsequently refine the segments with a zero-shot Grounding Dino
object detector, followed by intersection and exclusion filtering steps,
without the need for fine-tuning or retraining. To illustrate the efficacy of
the masking procedure, the proposed method is deployed on three challenging
applications scenarios that demand accurate masking; shredded plastics
characterization, drill core scanning, and litter monitoring. The numerical
evaluation of the proposed masking method on the three applications is provided
along with the used hyperparameters. The scripts for the method will be
available at https://github.com/hifexplo/Masking. | Computer Vision |
What field is the article from? | Title: How to Bridge the Gap between Modalities: A Comprehensive Survey on Multimodal Large Language Model
Abstract: This review paper explores Multimodal Large Language Models (MLLMs), which
integrate Large Language Models (LLMs) like GPT-4 to handle multimodal data
such as text and vision. MLLMs demonstrate capabilities like generating image
narratives and answering image-based questions, bridging the gap towards
real-world human-computer interactions and hinting at a potential pathway to
artificial general intelligence. However, MLLMs still face challenges in
processing the semantic gap in multimodality, which may lead to erroneous
generation, posing potential risks to society. Choosing the appropriate
modality alignment method is crucial, as improper methods might require more
parameters with limited performance improvement. This paper aims to explore
modality alignment methods for LLMs and their existing capabilities.
Implementing modality alignment allows LLMs to address environmental issues and
enhance accessibility. The study surveys existing modal alignment methods in
MLLMs into four groups: (1) Multimodal Converters that change data into
something LLMs can understand; (2) Multimodal Perceivers to improve how LLMs
perceive different types of data; (3) Tools Assistance for changing data into
one common format, usually text; and (4) Data-Driven methods that teach LLMs to
understand specific types of data in a dataset. This field is still in a phase
of exploration and experimentation, and we will organize and update various
existing research methods for multimodal information alignment. | Computational Linguistics |
What field is the article from? | Title: Adaptive Interventions with User-Defined Goals for Health Behavior Change
Abstract: Physical inactivity remains a major public health concern, having
associations with adverse health outcomes such as cardiovascular disease and
type-2 diabetes. Mobile health applications present a promising avenue for
low-cost, scalable physical activity promotion, yet often suffer from small
effect sizes and low adherence rates, particularly in comparison to human
coaching. Goal-setting is a critical component of health coaching that has been
underutilized in adaptive algorithms for mobile health interventions. This
paper introduces a modification to the Thompson sampling algorithm that places
emphasis on individualized goal-setting by optimizing personalized reward
functions. As a step towards supporting goal-setting, this paper offers a
balanced approach that can leverage shared structure while optimizing
individual preferences and goals. We prove that our modification incurs only a
constant penalty on the cumulative regret while preserving the sample
complexity benefits of data sharing. In a physical activity simulator, we
demonstrate that our algorithm achieves substantial improvements in cumulative
regret compared to baselines that do not share data or do not optimize for
individualized rewards. | Machine Learning |
What field is the article from? | Title: Roles of Scaling and Instruction Tuning in Language Perception: Model vs. Human Attention
Abstract: Recent large language models (LLMs) have revealed strong abilities to
understand natural language. Since most of them share the same basic structure,
i.e. the transformer block, possible contributors to their success in the
training process are scaling and instruction tuning. However, how these factors
affect the models' language perception is unclear. This work compares the
self-attention of several existing LLMs (LLaMA, Alpaca and Vicuna) in different
sizes (7B, 13B, 30B, 65B), together with eye saccade, an aspect of human
reading attention, to assess the effect of scaling and instruction tuning on
language perception. Results show that scaling enhances the human resemblance
and improves the effective attention by reducing the trivial pattern reliance,
while instruction tuning does not. However, instruction tuning significantly
enhances the models' sensitivity to instructions. We also find that current
LLMs are consistently closer to non-native than native speakers in attention,
suggesting a sub-optimal language perception of all models. Our code and data
used in the analysis is available on GitHub. | Computational Linguistics |
What field is the article from? | Title: Culturally Responsive Artificial Intelligence -- Problems, Challenges and Solutions
Abstract: In the contemporary interconnected world, the concept of cultural
responsibility occupies paramount importance. As the lines between nations
become less distinct, it is incumbent upon individuals, communities, and
institutions to assume the responsibility of safeguarding and valuing the
landscape of diverse cultures that constitute our global society. This paper
explores the socio-cultural and ethical challenges stemming from the
implementation of AI algorithms and highlights the necessity for their
culturally responsive development. It also offers recommendations on essential
elements required to enhance AI systems' adaptability to meet the demands of
contemporary multicultural societies. The paper highlights the need for further
multidisciplinary research to create AI models that effectively address these
challenges. It also advocates the significance of AI enculturation and
underlines the importance of regulatory measures to promote cultural
responsibility in AI systems. | Computers and Society |
What field is the article from? | Title: Deep Reinforcement Learning for Community Battery Scheduling under Uncertainties of Load, PV Generation, and Energy Prices
Abstract: In response to the growing uptake of distributed energy resources (DERs),
community batteries have emerged as a promising solution to support renewable
energy integration, reduce peak load, and enhance grid reliability. This paper
presents a deep reinforcement learning (RL) strategy, centered around the soft
actor-critic (SAC) algorithm, to schedule a community battery system in the
presence of uncertainties, such as solar photovoltaic (PV) generation, local
demand, and real-time energy prices. We position the community battery to play
a versatile role, in integrating local PV energy, reducing peak load, and
exploiting energy price fluctuations for arbitrage, thereby minimizing the
system cost. To improve exploration and convergence during RL training, we
utilize the noisy network technique. This paper conducts a comparative study of
different RL algorithms, including proximal policy optimization (PPO) and deep
deterministic policy gradient (DDPG) algorithms, to evaluate their
effectiveness in the community battery scheduling problem. The results
demonstrate the potential of RL in addressing community battery scheduling
challenges and show that the SAC algorithm achieves the best performance
compared to RL and optimization benchmarks. | Machine Learning |
What field is the article from? | Title: Artificial Intelligence Ethics Education in Cybersecurity: Challenges and Opportunities: a focus group report
Abstract: The emergence of AI tools in cybersecurity creates many opportunities and
uncertainties. A focus group with advanced graduate students in cybersecurity
revealed the potential depth and breadth of the challenges and opportunities.
The salient issues are access to open source or free tools, documentation,
curricular diversity, and clear articulation of ethical principles for AI
cybersecurity education. Confronting the "black box" mentality in AI
cybersecurity work is also of the greatest importance, doubled by deeper and
prior education in foundational AI work. Systems thinking and effective
communication were considered relevant areas of educational improvement. Future
AI educators and practitioners need to address these issues by implementing
rigorous technical training curricula, clear documentation, and frameworks for
ethically monitoring AI combined with critical and system's thinking and
communication skills. | Cryptography and Security |
What field is the article from? | Title: Customizable Combination of Parameter-Efficient Modules for Multi-Task Learning
Abstract: Modular and composable transfer learning is an emerging direction in the
field of Parameter Efficient Fine-Tuning, as it enables neural networks to
better organize various aspects of knowledge, leading to improved cross-task
generalization. In this paper, we introduce a novel approach Customized
Polytropon C-Poly that combines task-common skills and task-specific skills,
while the skill parameters being highly parameterized using low-rank
techniques. Each task is associated with a customizable number of exclusive
specialized skills and also benefits from skills shared with peer tasks. A
skill assignment matrix is jointly learned. To evaluate our approach, we
conducted extensive experiments on the Super-NaturalInstructions and the
SuperGLUE benchmarks. Our findings demonstrate that C-Poly outperforms
fully-shared, task-specific, and skill-indistinguishable baselines,
significantly enhancing the sample efficiency in multi-task learning scenarios. | Machine Learning |
What field is the article from? | Title: Dual Conditioned Diffusion Models for Out-Of-Distribution Detection: Application to Fetal Ultrasound Videos
Abstract: Out-of-distribution (OOD) detection is essential to improve the reliability
of machine learning models by detecting samples that do not belong to the
training distribution. Detecting OOD samples effectively in certain tasks can
pose a challenge because of the substantial heterogeneity within the
in-distribution (ID), and the high structural similarity between ID and OOD
classes. For instance, when detecting heart views in fetal ultrasound videos
there is a high structural similarity between the heart and other anatomies
such as the abdomen, and large in-distribution variance as a heart has 5
distinct views and structural variations within each view. To detect OOD
samples in this context, the resulting model should generalise to the
intra-anatomy variations while rejecting similar OOD samples. In this paper, we
introduce dual-conditioned diffusion models (DCDM) where we condition the model
on in-distribution class information and latent features of the input image for
reconstruction-based OOD detection. This constrains the generative manifold of
the model to generate images structurally and semantically similar to those
within the in-distribution. The proposed model outperforms reference methods
with a 12% improvement in accuracy, 22% higher precision, and an 8% better F1
score. | Computer Vision |
What field is the article from? | Title: Unsupervised Lexical Simplification with Context Augmentation
Abstract: We propose a new unsupervised lexical simplification method that uses only
monolingual data and pre-trained language models. Given a target word and its
context, our method generates substitutes based on the target context and also
additional contexts sampled from monolingual data. We conduct experiments in
English, Portuguese, and Spanish on the TSAR-2022 shared task, and show that
our model substantially outperforms other unsupervised systems across all
languages. We also establish a new state-of-the-art by ensembling our model
with GPT-3.5. Lastly, we evaluate our model on the SWORDS lexical substitution
data set, achieving a state-of-the-art result. | Computational Linguistics |
What field is the article from? | Title: RankAug: Augmented data ranking for text classification
Abstract: Research on data generation and augmentation has been focused majorly on
enhancing generation models, leaving a notable gap in the exploration and
refinement of methods for evaluating synthetic data. There are several text
similarity metrics within the context of generated data filtering which can
impact the performance of specific Natural Language Understanding (NLU) tasks,
specifically focusing on intent and sentiment classification. In this study, we
propose RankAug, a text-ranking approach that detects and filters out the top
augmented texts in terms of being most similar in meaning with lexical and
syntactical diversity. Through experiments conducted on multiple datasets, we
demonstrate that the judicious selection of filtering techniques can yield a
substantial improvement of up to 35% in classification accuracy for
under-represented classes. | Computational Linguistics |
What field is the article from? | Title: Unleashing the Creative Mind: Language Model As Hierarchical Policy For Improved Exploration on Challenging Problem Solving
Abstract: Large Language Models (LLMs) have achieved tremendous progress, yet they
still often struggle with challenging reasoning problems. Current approaches
address this challenge by sampling or searching detailed and low-level
reasoning chains. However, these methods are still limited in their exploration
capabilities, making it challenging for correct solutions to stand out in the
huge solution space. In this work, we unleash LLMs' creative potential for
exploring multiple diverse problem solving strategies by framing an LLM as a
hierarchical policy via in-context learning. This policy comprises of a
visionary leader that proposes multiple diverse high-level problem-solving
tactics as hints, accompanied by a follower that executes detailed
problem-solving processes following each of the high-level instruction. The
follower uses each of the leader's directives as a guide and samples multiple
reasoning chains to tackle the problem, generating a solution group for each
leader proposal. Additionally, we propose an effective and efficient
tournament-based approach to select among these explored solution groups to
reach the final answer. Our approach produces meaningful and inspiring hints,
enhances problem-solving strategy exploration, and improves the final answer
accuracy on challenging problems in the MATH dataset. Code will be released at
https://github.com/lz1oceani/LLM-As-Hierarchical-Policy. | Artificial Intelligence |
What field is the article from? | Title: Empowering remittance management in the digitised landscape: A real-time Data-Driven Decision Support with predictive abilities for financial transactions
Abstract: The advent of Blockchain technology (BT) revolutionised the way remittance
transactions are recorded. Banks and remittance organisations have shown a
growing interest in exploring blockchain's potential advantages over
traditional practices. This paper presents a data-driven predictive decision
support approach as an innovative artefact designed for the blockchain-oriented
remittance industry. Employing a theory-generating Design Science Research
(DSR) approach, we have uncovered the emergence of predictive capabilities
driven by transactional big data. The artefact integrates predictive analytics
and Machine Learning (ML) to enable real-time remittance monitoring, empowering
management decision-makers to address challenges in the uncertain digitised
landscape of blockchain-oriented remittance companies. Bridging the gap between
theory and practice, this research not only enhances the security of the
remittance ecosystem but also lays the foundation for future predictive
decision support solutions, extending the potential of predictive analytics to
other domains. Additionally, the generated theory from the artifact's
implementation enriches the DSR approach and fosters grounded and stakeholder
theory development in the information systems domain. | Artificial Intelligence |
What field is the article from? | Title: Dexterous Functional Grasping
Abstract: While there have been significant strides in dexterous manipulation, most of
it is limited to benchmark tasks like in-hand reorientation which are of
limited utility in the real world. The main benefit of dexterous hands over
two-fingered ones is their ability to pickup tools and other objects (including
thin ones) and grasp them firmly to apply force. However, this task requires
both a complex understanding of functional affordances as well as precise
low-level control. While prior work obtains affordances from human data this
approach doesn't scale to low-level control. Similarly, simulation training
cannot give the robot an understanding of real-world semantics. In this paper,
we aim to combine the best of both worlds to accomplish functional grasping for
in-the-wild objects. We use a modular approach. First, affordances are obtained
by matching corresponding regions of different objects and then a low-level
policy trained in sim is run to grasp it. We propose a novel application of
eigengrasps to reduce the search space of RL using a small amount of human data
and find that it leads to more stable and physically realistic motion. We find
that eigengrasp action space beats baselines in simulation and outperforms
hardcoded grasping in real and matches or outperforms a trained human
teleoperator. Results visualizations and videos at https://dexfunc.github.io/ | Robotics |
What field is the article from? | Title: Histopathologic Cancer Detection
Abstract: Early diagnosis of the cancer cells is necessary for making an effective
treatment plan and for the health and safety of a patient. Nowadays, doctors
usually use a histological grade that pathologists determine by performing a
semi-quantitative analysis of the histopathological and cytological features of
hematoxylin-eosin (HE) stained histopathological images. This research
contributes a potential classification model for cancer prognosis to
efficiently utilize the valuable information underlying the HE-stained
histopathological images. This work uses the PatchCamelyon benchmark datasets
and trains them in a multi-layer perceptron and convolution model to observe
the model's performance in terms of precision, Recall, F1 Score, Accuracy, and
AUC Score. The evaluation result shows that the baseline convolution model
outperforms the baseline MLP model. Also, this paper introduced ResNet50 and
InceptionNet models with data augmentation, where ResNet50 is able to beat the
state-of-the-art model. Furthermore, the majority vote and concatenation
ensemble were evaluated and provided the future direction of using transfer
learning and segmentation to understand the specific features. | Computer Vision |
What field is the article from? | Title: Ever Evolving Evaluator (EV3): Towards Flexible and Reliable Meta-Optimization for Knowledge Distillation
Abstract: We introduce EV3, a novel meta-optimization framework designed to efficiently
train scalable machine learning models through an intuitive
explore-assess-adapt protocol. In each iteration of EV3, we explore various
model parameter updates, assess them using pertinent evaluation methods, and
then adapt the model based on the optimal updates and previous progress
history. EV3 offers substantial flexibility without imposing stringent
constraints like differentiability on the key objectives relevant to the tasks
of interest, allowing for exploratory updates with intentionally-biased
gradients and through a diversity of losses and optimizers. Additionally, the
assessment phase provides reliable safety controls to ensure robust
generalization, and can dynamically prioritize tasks in scenarios with multiple
objectives. With inspiration drawn from evolutionary algorithms, meta-learning,
and neural architecture search, we investigate an application of EV3 to
knowledge distillation. Our experimental results illustrate EV3's capability to
safely explore the modeling landscape, while hinting at its potential
applicability across numerous domains due to its inherent flexibility and
adaptability. Finally, we provide a JAX implementation of EV3, along with
source code for experiments, available at:
https://github.com/google-research/google-research/tree/master/ev3. | Machine Learning |
What field is the article from? | Title: VIGraph: Self-supervised Learning for Class-Imbalanced Node Classification
Abstract: Class imbalance in graph data poses significant challenges for node
classification. Existing methods, represented by SMOTE-based approaches,
partially alleviate this issue but still exhibit limitations during imbalanced
scenario construction. Self-supervised learning (SSL) offers a promising
solution by synthesizing minority nodes from the data itself, yet its potential
remains unexplored. In this paper, we analyze the limitations of SMOTE-based
approaches and introduce VIGraph, a novel SSL model based on the
self-supervised Variational Graph Auto-Encoder (VGAE) that leverages
Variational Inference (VI) to generate minority nodes. Specifically, VIGraph
strictly adheres to the concept of imbalance when constructing imbalanced
graphs and utilizes the generative VGAE to generate minority nodes. Moreover,
VIGraph introduces a novel Siamese contrastive strategy at the decoding phase
to improve the overall quality of generated nodes. VIGraph can generate
high-quality nodes without reintegrating them into the original graph,
eliminating the "Generating, Reintegrating, and Retraining" process found in
SMOTE-based methods. Experiments on multiple real-world datasets demonstrate
that VIGraph achieves promising results for class-imbalanced node
classification tasks. | Machine Learning |
What field is the article from? | Title: NVFi: Neural Velocity Fields for 3D Physics Learning from Dynamic Videos
Abstract: In this paper, we aim to model 3D scene dynamics from multi-view videos.
Unlike the majority of existing works which usually focus on the common task of
novel view synthesis within the training time period, we propose to
simultaneously learn the geometry, appearance, and physical velocity of 3D
scenes only from video frames, such that multiple desirable applications can be
supported, including future frame extrapolation, unsupervised 3D semantic scene
decomposition, and dynamic motion transfer. Our method consists of three major
components, 1) the keyframe dynamic radiance field, 2) the interframe velocity
field, and 3) a joint keyframe and interframe optimization module which is the
core of our framework to effectively train both networks. To validate our
method, we further introduce two dynamic 3D datasets: 1) Dynamic Object
dataset, and 2) Dynamic Indoor Scene dataset. We conduct extensive experiments
on multiple datasets, demonstrating the superior performance of our method over
all baselines, particularly in the critical tasks of future frame extrapolation
and unsupervised 3D semantic scene decomposition. | Computer Vision |
What field is the article from? | Title: Large-Scale Multi-Robot Assembly Planning for Autonomous Manufacturing
Abstract: Mobile autonomous robots have the potential to revolutionize manufacturing
processes. However, employing large robot fleets in manufacturing requires
addressing challenges including collision-free movement in a shared workspace,
effective multi-robot collaboration to manipulate and transport large payloads,
complex task allocation due to coupled manufacturing processes, and spatial
planning for parallel assembly and transportation of nested subassemblies. We
propose a full algorithmic stack for large-scale multi-robot assembly planning
that addresses these challenges and can synthesize construction plans for
complex assemblies with thousands of parts in a matter of minutes. Our approach
takes in a CAD-like product specification and automatically plans a full-stack
assembly procedure for a group of robots to manufacture the product. We propose
an algorithmic stack that comprises: (i) an iterative radial layout
optimization procedure to define a global staging layout for the manufacturing
facility, (ii) a graph-repair mixed-integer program formulation and a modified
greedy task allocation algorithm to optimally allocate robots and robot
sub-teams to assembly and transport tasks, (iii) a geometric heuristic and a
hill-climbing algorithm to plan collaborative carrying configurations of robot
sub-teams, and (iv) a distributed control policy that enables robots to execute
the assembly motion plan collision-free. We also present an open-source
multi-robot manufacturing simulator implemented in Julia as a resource to the
research community, to test our algorithms and to facilitate multi-robot
manufacturing research more broadly. Our empirical results demonstrate the
scalability and effectiveness of our approach by generating plans to
manufacture a LEGO model of a Saturn V launch vehicle with 1845 parts, 306
subassemblies, and 250 robots in under three minutes on a standard laptop
computer. | Robotics |
What field is the article from? | Title: Surprisal Driven $k$-NN for Robust and Interpretable Nonparametric Learning
Abstract: Nonparametric learning is a fundamental concept in machine learning that aims
to capture complex patterns and relationships in data without making strong
assumptions about the underlying data distribution. Owing to simplicity and
familiarity, one of the most well-known algorithms under this paradigm is the
$k$-nearest neighbors ($k$-NN) algorithm. Driven by the usage of machine
learning in safety-critical applications, in this work, we shed new light on
the traditional nearest neighbors algorithm from the perspective of information
theory and propose a robust and interpretable framework for tasks such as
classification, regression, and anomaly detection using a single model. Instead
of using a traditional distance measure which needs to be scaled and
contextualized, we use a novel formulation of \textit{surprisal} (amount of
information required to explain the difference between the observed and
expected result). Finally, we demonstrate this architecture's capability to
perform at-par or above the state-of-the-art on classification, regression, and
anomaly detection tasks using a single model with enhanced interpretability by
providing novel concepts for characterizing data and predictions. | Machine Learning |
What field is the article from? | Title: Rule Learning as Machine Translation using the Atomic Knowledge Bank
Abstract: Machine learning models, and in particular language models, are being applied
to various tasks that require reasoning. While such models are good at
capturing patterns their ability to reason in a trustable and controlled manner
is frequently questioned. On the other hand, logic-based rule systems allow for
controlled inspection and already established verification methods. However it
is well-known that creating such systems manually is time-consuming and prone
to errors. We explore the capability of transformers to translate sentences
expressing rules in natural language into logical rules. We see reasoners as
the most reliable tools for performing logical reasoning and focus on
translating language into the format expected by such tools. We perform
experiments using the DKET dataset from the literature and create a dataset for
language to logic translation based on the Atomic knowledge bank. | Computational Linguistics |
What field is the article from? | Title: Towards Possibilities & Impossibilities of AI-generated Text Detection: A Survey
Abstract: Large Language Models (LLMs) have revolutionized the domain of natural
language processing (NLP) with remarkable capabilities of generating human-like
text responses. However, despite these advancements, several works in the
existing literature have raised serious concerns about the potential misuse of
LLMs such as spreading misinformation, generating fake news, plagiarism in
academia, and contaminating the web. To address these concerns, a consensus
among the research community is to develop algorithmic solutions to detect
AI-generated text. The basic idea is that whenever we can tell if the given
text is either written by a human or an AI, we can utilize this information to
address the above-mentioned concerns. To that end, a plethora of detection
frameworks have been proposed, highlighting the possibilities of AI-generated
text detection. But in parallel to the development of detection frameworks,
researchers have also concentrated on designing strategies to elude detection,
i.e., focusing on the impossibilities of AI-generated text detection. This is a
crucial step in order to make sure the detection frameworks are robust enough
and it is not too easy to fool a detector. Despite the huge interest and the
flurry of research in this domain, the community currently lacks a
comprehensive analysis of recent developments. In this survey, we aim to
provide a concise categorization and overview of current work encompassing both
the prospects and the limitations of AI-generated text detection. To enrich the
collective knowledge, we engage in an exhaustive discussion on critical and
challenging open questions related to ongoing research on AI-generated text
detection. | Computational Linguistics |
What field is the article from? | Title: A Central Motor System Inspired Pre-training Reinforcement Learning for Robotic Control
Abstract: Designing controllers to achieve natural motor capabilities for multi-joint
robots is a significant challenge. However, animals in nature are naturally
with basic motor abilities and can master various complex motor skills through
acquired learning. On the basis of analyzing the mechanism of the central motor
system in mammals, we propose a novel pre-training reinforcement learning
algorithm that enables robots to learn rich motor skills and apply them to
complex task environments without relying on external data. We first design a
skill based network similar to the cerebellum by utilizing the selection
mechanism of voluntary movements in the basal ganglia and the basic motor
regulation ability of the cerebellum. Subsequently, by imitating the structure
of advanced centers in the central motor system, we propose a high-level policy
to generate different skill combinations, thereby enabling the robot to acquire
natural motor abilities. We conduct experiments on 4 types of robots and 22
task environments, and the results show that the proposed method can enable
different types of robots to achieve flexible motor skills. Overall, our
research provides a promising framework for the design of neural network motor
controllers. | Robotics |
What field is the article from? | Title: Are Vision Transformers More Data Hungry Than Newborn Visual Systems?
Abstract: Vision transformers (ViTs) are top performing models on many computer vision
benchmarks and can accurately predict human behavior on object recognition
tasks. However, researchers question the value of using ViTs as models of
biological learning because ViTs are thought to be more data hungry than
brains, with ViTs requiring more training data to reach similar levels of
performance. To test this assumption, we directly compared the learning
abilities of ViTs and animals, by performing parallel controlled rearing
experiments on ViTs and newborn chicks. We first raised chicks in impoverished
visual environments containing a single object, then simulated the training
data available in those environments by building virtual animal chambers in a
video game engine. We recorded the first-person images acquired by agents
moving through the virtual chambers and used those images to train self
supervised ViTs that leverage time as a teaching signal, akin to biological
visual systems. When ViTs were trained through the eyes of newborn chicks, the
ViTs solved the same view invariant object recognition tasks as the chicks.
Thus, ViTs were not more data hungry than newborn visual systems: both learned
view invariant object representations in impoverished visual environments. The
flexible and generic attention based learning mechanism in ViTs combined with
the embodied data streams available to newborn animals appears sufficient to
drive the development of animal-like object recognition. | Computer Vision |
What field is the article from? | Title: Physics-Enhanced Multi-fidelity Learning for Optical Surface Imprint
Abstract: Human fingerprints serve as one unique and powerful characteristic for each
person, from which policemen can recognize the identity. Similar to humans,
many natural bodies and intrinsic mechanical qualities can also be uniquely
identified from surface characteristics. To measure the elasto-plastic
properties of one material, one formally sharp indenter is pushed into the
measured body under constant force and retracted, leaving a unique residual
imprint of the minute size from several micrometers to nanometers. However, one
great challenge is how to map the optical image of this residual imprint into
the real wanted mechanical properties, i.e., the tensile force curve. In this
paper, we propose a novel method to use multi-fidelity neural networks (MFNN)
to solve this inverse problem. We first actively train the NN model via pure
simulation data, and then bridge the sim-to-real gap via transfer learning. The
most innovative part is that we use NN to dig out the unknown physics and also
implant the known physics into the transfer learning framework, thus highly
improving the model stability and decreasing the data requirement. This work
serves as one great example of applying machine learning into the real
experimental research, especially under the constraints of data limitation and
fidelity variance. | Machine Learning |
What field is the article from? | Title: Assessing Knowledge Editing in Language Models via Relation Perspective
Abstract: Knowledge Editing (KE) for modifying factual knowledge in Large Language
Models (LLMs) has been receiving increasing attention. However, existing
knowledge editing methods are entity-centric, and it is unclear whether this
approach is suitable for a relation-centric perspective. To address this gap,
this paper constructs a new benchmark named RaKE, which focuses on Relation
based Knowledge Editing. In this paper, we establish a suite of innovative
metrics for evaluation and conduct comprehensive experiments involving various
knowledge editing baselines. We notice that existing knowledge editing methods
exhibit the potential difficulty in their ability to edit relations. Therefore,
we further explore the role of relations in factual triplets within the
transformer. Our research results confirm that knowledge related to relations
is not only stored in the FFN network but also in the attention layers. This
provides experimental support for future relation-based knowledge editing
methods. | Computational Linguistics |
What field is the article from? | Title: Nexus at ArAIEval Shared Task: Fine-Tuning Arabic Language Models for Propaganda and Disinformation Detection
Abstract: The spread of disinformation and propagandistic content poses a threat to
societal harmony, undermining informed decision-making and trust in reliable
sources. Online platforms often serve as breeding grounds for such content, and
malicious actors exploit the vulnerabilities of audiences to shape public
opinion. Although there have been research efforts aimed at the automatic
identification of disinformation and propaganda in social media content, there
remain challenges in terms of performance. The ArAIEval shared task aims to
further research on these particular issues within the context of the Arabic
language. In this paper, we discuss our participation in these shared tasks. We
competed in subtasks 1A and 2A, where our submitted system secured positions
9th and 10th, respectively. Our experiments consist of fine-tuning transformer
models and using zero- and few-shot learning with GPT-4. | Computational Linguistics |
What field is the article from? | Title: Fine-Tuning InstructPix2Pix for Advanced Image Colorization
Abstract: This paper presents a novel approach to human image colorization by
fine-tuning the InstructPix2Pix model, which integrates a language model
(GPT-3) with a text-to-image model (Stable Diffusion). Despite the original
InstructPix2Pix model's proficiency in editing images based on textual
instructions, it exhibits limitations in the focused domain of colorization. To
address this, we fine-tuned the model using the IMDB-WIKI dataset, pairing
black-and-white images with a diverse set of colorization prompts generated by
ChatGPT. This paper contributes by (1) applying fine-tuning techniques to
stable diffusion models specifically for colorization tasks, and (2) employing
generative models to create varied conditioning prompts. After finetuning, our
model outperforms the original InstructPix2Pix model on multiple metrics
quantitatively, and we produce more realistically colored images qualitatively.
The code for this project is provided on the GitHub Repository
https://github.com/AllenAnZifeng/DeepLearning282. | Computer Vision |
What field is the article from? | Title: Investigating YOLO Models Towards Outdoor Obstacle Detection For Visually Impaired People
Abstract: The utilization of deep learning-based object detection is an effective
approach to assist visually impaired individuals in avoiding obstacles. In this
paper, we implemented seven different YOLO object detection models
\textit{viz}., YOLO-NAS (small, medium, large), YOLOv8, YOLOv7, YOLOv6, and
YOLOv5 and performed comprehensive evaluation with carefully tuned
hyperparameters, to analyze how these models performed on images containing
common daily-life objects presented on roads and sidewalks. After a systematic
investigation, YOLOv8 was found to be the best model, which reached a precision
of $80\%$ and a recall of $68.2\%$ on a well-known Obstacle Dataset which
includes images from VOC dataset, COCO dataset, and TT100K dataset along with
images collected by the researchers in the field. Despite being the latest
model and demonstrating better performance in many other applications, YOLO-NAS
was found to be suboptimal for the obstacle detection task. | Computer Vision |
What field is the article from? | Title: Modeling subjectivity (by Mimicking Annotator Annotation) in toxic comment identification across diverse communities
Abstract: The prevalence and impact of toxic discussions online have made content
moderation crucial.Automated systems can play a vital role in identifying
toxicity, and reducing the reliance on human moderation.Nevertheless,
identifying toxic comments for diverse communities continues to present
challenges that are addressed in this paper.The two-part goal of this study is
to(1)identify intuitive variances from annotator disagreement using
quantitative analysis and (2)model the subjectivity of these viewpoints.To
achieve our goal, we published a new
dataset\footnote{\url{https://github.com/XXX}} with expert annotators'
annotations and used two other public datasets to identify the subjectivity of
toxicity.Then leveraging the Large Language Model(LLM),we evaluate the model's
ability to mimic diverse viewpoints on toxicity by varying size of the training
data and utilizing same set of annotators as the test set used during model
training and a separate set of annotators as the test set.We conclude that
subjectivity is evident across all annotator groups, demonstrating the
shortcomings of majority-rule voting. Moving forward, subjective annotations
should serve as ground truth labels for training models for domains like
toxicity in diverse communities. | Artificial Intelligence |
What field is the article from? | Title: Content Augmented Graph Neural Networks
Abstract: In recent years, graph neural networks (GNNs) have become a popular tool for
solving various problems over graphs. In these models, the link structure of
the graph is typically exploited and nodes' embeddings are iteratively updated
based on adjacent nodes. Nodes' contents are used solely in the form of feature
vectors, served as nodes' first-layer embeddings. However, the filters or
convolutions, applied during iterations/layers to these initial embeddings lead
to their impact diminish and contribute insignificantly to the final
embeddings. In order to address this issue, in this paper we propose augmenting
nodes' embeddings by embeddings generating from their content, at higher GNN
layers. More precisely, we propose models wherein a structural embedding using
a GNN and a content embedding are computed for each node. These two are
combined using a combination layer to form the embedding of a node at a given
layer. We suggest methods such as using an auto-encoder or building a content
graph, to generate content embeddings. In the end, by conducting experiments
over several real-world datasets, we demonstrate the high accuracy and
performance of our models. | Machine Learning |
What field is the article from? | Title: Best uses of ChatGPT and Generative AI for computer science research
Abstract: Generative Artificial Intelligence (AI), particularly tools like OpenAI's
popular ChatGPT, is reshaping the landscape of computer science research. Used
wisely, these tools can boost the productivity of a computer research
scientist. This paper provides an exploration of the diverse applications of
ChatGPT and other generative AI technologies in computer science academic
research, making recommendations about the use of Generative AI to make more
productive the role of the computer research scientist, with the focus of
writing new research papers. We highlight innovative uses such as brainstorming
research ideas, aiding in the drafting and styling of academic papers and
assisting in the synthesis of state-of-the-art section. Further, we delve into
using these technologies in understanding interdisciplinary approaches, making
complex texts simpler, and recommending suitable academic journals for
publication. Significant focus is placed on generative AI's contributions to
synthetic data creation, research methodology, and mentorship, as well as in
task organization and article quality assessment. The paper also addresses the
utility of AI in article review, adapting texts to length constraints,
constructing counterarguments, and survey development. Moreover, we explore the
capabilities of these tools in disseminating ideas, generating images and
audio, text transcription, and engaging with editors. We also describe some
non-recommended uses of generative AI for computer science research, mainly
because of the limitations of this technology. | Artificial Intelligence |
What field is the article from? | Title: Towards the Law of Capacity Gap in Distilling Language Models
Abstract: Language model (LM) distillation is a trending area that aims to distil the
knowledge resided in a large teacher LM to a small student one. While various
methods have been proposed to push the distillation to its limits, it is still
a pain distilling LMs when a large capacity gap is exhibited between the
teacher and the student LMs. The pain is mainly resulted by the curse of
capacity gap, which describes that a larger teacher LM cannot always lead to a
better student LM than one distilled from a smaller teacher LM due to the
affect of capacity gap increment. That is, there is likely an optimal point
yielding the best student LM along the scaling course of the teacher LM. Even
worse, the curse of capacity gap can be only partly yet not fully lifted as
indicated in previous studies.
However, the tale is not ever one-sided. Although a larger teacher LM has
better performance than a smaller teacher LM, it is much more
resource-demanding especially in the context of recent large LMs (LLMs).
Consequently, instead of sticking to lifting the curse, leaving the curse as is
should be arguably fine. Even better, in this paper, we reveal that the optimal
capacity gap is almost consistent across different student scales and
architectures, fortunately turning the curse into the law of capacity gap. The
law later guides us to distil a 3B student LM (termed MiniMA) from a 7B teacher
LM (adapted LLaMA2-7B). MiniMA is demonstrated to yield a new
compute-performance pareto frontier among existing 3B LMs on commonly used
benchmarks, and its instruction-tuned version (termed MiniChat) outperforms a
wide range of 3B competitors in GPT4 evaluation and could even compete with
several 7B chat models. | Computational Linguistics |
What field is the article from? | Title: Learning interactions to boost human creativity with bandits and GPT-4
Abstract: This paper considers how interactions with AI algorithms can boost human
creative thought. We employ a psychological task that demonstrates limits on
human creativity, namely semantic feature generation: given a concept name,
respondents must list as many of its features as possible. Human participants
typically produce only a fraction of the features they know before getting
"stuck." In experiments with humans and with a language AI (GPT-4) we contrast
behavior in the standard task versus a variant in which participants can ask
for algorithmically-generated hints. Algorithm choice is administered by a
multi-armed bandit whose reward indicates whether the hint helped generating
more features. Humans and the AI show similar benefits from hints, and
remarkably, bandits learning from AI responses prefer the same prompting
strategy as those learning from human behavior. The results suggest that
strategies for boosting human creativity via computer interactions can be
learned by bandits run on groups of simulated participants. | Artificial Intelligence |
What field is the article from? | Title: Classification for everyone : Building geography agnostic models for fairer recognition
Abstract: In this paper, we analyze different methods to mitigate inherent geographical
biases present in state of the art image classification models. We first
quantitatively present this bias in two datasets - The Dollar Street Dataset
and ImageNet, using images with location information. We then present different
methods which can be employed to reduce this bias. Finally, we analyze the
effectiveness of the different techniques on making these models more robust to
geographical locations of the images. | Computer Vision |
What field is the article from? | Title: Learning Reusable Manipulation Strategies
Abstract: Humans demonstrate an impressive ability to acquire and generalize
manipulation "tricks." Even from a single demonstration, such as using soup
ladles to reach for distant objects, we can apply this skill to new scenarios
involving different object positions, sizes, and categories (e.g., forks and
hammers). Additionally, we can flexibly combine various skills to devise
long-term plans. In this paper, we present a framework that enables machines to
acquire such manipulation skills, referred to as "mechanisms," through a single
demonstration and self-play. Our key insight lies in interpreting each
demonstration as a sequence of changes in robot-object and object-object
contact modes, which provides a scaffold for learning detailed samplers for
continuous parameters. These learned mechanisms and samplers can be seamlessly
integrated into standard task and motion planners, enabling their compositional
use. | Robotics |
What field is the article from? | Title: Modeling the Telemarketing Process using Genetic Algorithms and Extreme Boosting: Feature Selection and Cost-Sensitive Analytical Approach
Abstract: Currently, almost all direct marketing activities take place virtually rather
than in person, weakening interpersonal skills at an alarming pace.
Furthermore, businesses have been striving to sense and foster the tendency of
their clients to accept a marketing offer. The digital transformation and the
increased virtual presence forced firms to seek novel marketing research
approaches. This research aims at leveraging the power of telemarketing data in
modeling the willingness of clients to make a term deposit and finding the most
significant characteristics of the clients. Real-world data from a Portuguese
bank and national socio-economic metrics are used to model the telemarketing
decision-making process. This research makes two key contributions. First,
propose a novel genetic algorithm-based classifier to select the best
discriminating features and tune classifier parameters simultaneously. Second,
build an explainable prediction model. The best-generated classification models
were intensively validated using 50 times repeated 10-fold stratified
cross-validation and the selected features have been analyzed. The models
significantly outperform the related works in terms of class of interest
accuracy, they attained an average of 89.07\% and 0.059 in terms of geometric
mean and type I error respectively. The model is expected to maximize the
potential profit margin at the least possible cost and provide more insights to
support marketing decision-making. | Machine Learning |
Subsets and Splits