text
stringlengths 0
1.22M
|
---|
Align, Reason and Learn: Enhancing Medical Vision-and-Language Pre-training with Knowledge
Zhihong Chen
Shenzhen Research Institute of Big Data,
The Chinese University of Hong Kong,ShenzhenChina
[email protected]
,
Guanbin Li
Sun Yat-sen University,GuangzhouChina
[email protected]
and
Xiang Wan
Shenzhen Research Institute of Big Data,
The Chinese University of Hong Kong,ShenzhenChina
Pazhou Lab,GuangzhouChina
[email protected]
(2022)
Abstract.
Medical vision-and-language pre-training (Med-VLP) has received considerable attention owing to its applicability to extracting generic vision-and-language representations from medical images and texts.
Most existing methods mainly contain three elements: uni-modal encoders (i.e., a vision encoder and a language encoder), a multi-modal fusion module, and pretext tasks, with few studies considering the importance of medical domain expert knowledge and explicitly exploiting such knowledge to facilitate Med-VLP.
Although there exist knowledge-enhanced vision-and-language pre-training (VLP) methods in the general domain, most require off-the-shelf toolkits (e.g., object detectors and scene graph parsers), which are unavailable in the medical domain.
In this paper, we propose a systematic and effective approach to enhance Med-VLP by structured medical knowledge from three perspectives.
First, considering knowledge can be regarded as the intermediate medium between vision and language, we align the representations of the vision encoder and the language encoder through knowledge.
Second, we inject knowledge into the multi-modal fusion model to enable the model to perform reasoning using knowledge as the supplementation of the input image and text.
Third, we guide the model to put emphasis on the most critical information in images and texts by designing knowledge-induced pretext tasks.
To perform a comprehensive evaluation and facilitate further research, we construct a medical vision-and-language benchmark including three tasks.
Experimental results illustrate the effectiveness of our approach, where state-of-the-art performance is achieved on all downstream tasks.
Further analyses explore the effects of different components of our approach and various settings of pre-training.111The source code is available at https://github.com/zhjohnchan/ARL.
vision-and-language; knowledge-enhanced learning; multi-modal pre-training; medical analysis
††journalyear: 2022††copyright: acmcopyright††conference: Proceedings of the 30th ACM International Conference on Multimedia ; October 10–14, 2022; Lisboa, Portugal.††booktitle: Proceedings of the 30th ACM International Conference on Multimedia (MM ’22), October 10–14, 2022, Lisboa, Portugal††price: 15.00††isbn: 978-1-4503-9203-7/22/10††doi: 10.1145/3503161.3547948††submissionid: 814††ccs: Computing methodologies Artificial intelligence††ccs: Computing methodologies Machine learning††ccs: Computing methodologies Multi-task learning
1. Introduction
Medical data streams from various sources, among which vision and language are two critical ones.
It includes image data (e.g., radiography, magnetic resonance imaging, and computed tomography) and text data (e.g., radiology reports and medical texts).
Medical vision-and-language pre-training (Med-VLP) aims to jointly process data from these two modalities to learn generalizable multi-modal representations from large-scale medical image-text data.
It enables a vision-and-language model to address a wide range of medical vision-and-language tasks (e.g., medical visual question answering (Med-VQA), medical image-text classification (Med-ITC), and medical image-text retrieval (Med-ITR)), which can be crucial for alleviating the data scarcity problem in the medical field.
In the past few years, vision-and-language pre-training (VLP) has drawn sustaining attention (Chen et al., 2020; Kim et al., 2021; Su et al., 2019; Tan and Bansal, 2019; Lu et al., 2019; Li et al., 2020b; Cui et al., 2021) and achieved state-of-the-art performance on many vision-and-language tasks in the general domain.
In general, a VLP system consists of three elements: (i) uni-modal encoders (i.e., a vision encoder and a language encoder) that encode images and texts into image and text features, respectively; (ii) a multi-modal fusion module that performs the fusion of the encoded image and text features; (iii) pretext tasks (e.g., masked image modeling (MIM), masked language modeling (MLM), and image-text matching (ITM)) that assist the learning of VLP models.
More recently, some studies (Li et al., 2020a; Khare et al., 2021; Gong et al., 2021; Moon et al., 2021) applied VLP to the medical domain and significantly improved the performance for medical vision-and-language tasks (especially for Med-VQA).
These methods are superior in capturing the mappings between images and texts and thus enable the pre-trained models to understand the complicated cross-modal information.
For example, (Khare et al., 2021; Gong et al., 2021) proposed to perform the pre-training on medical image-text pairs to capture medical knowledge, and the evaluation on Med-VQA has demonstrated the validity of their proposed methods.
Although these methods have motivated the learning of image-text correspondences through well-designed model architectures and pretext tasks, most of them disregard the complementary information (i.e., knowledge) shared by different modalities and still lack the explicit knowledge modeling for Med-VLP.
Even in the general domain, there are only a few VLP studies (Li et al., 2020b; Yu et al., 2021; Chen et al., 2021; Cui et al., 2021) on incorporating external knowledge into the pre-training process.
For instance, ERNIE-ViL (Yu et al., 2021) constructed a scene graph from the input text to build the semantic connections between vision and language and emphasized the importance of keywords (e.g., objects, attributes, and relationships between objects) through the designs of pretext tasks.
ROSITA (Cui et al., 2021) used a unified scene graph shared by the input image and text to enhance the
semantic alignments between vision and language.
Similarly, KB-VLP (Chen et al., 2021) used object tags detected from images and knowledge graph embeddings extracted from texts to enhance the learning of knowledge-aware representations.
However, the aforementioned studies require off-the-shelf toolkits (e.g., object detectors and scene graph parsers), which are generally unavailable in the medical domain.
Furthermore, they might be limited in scalability as their performance depends heavily on the reliability of the object detectors or scene graph parsers.
Therefore, it is expected to have a better solution to exploit external knowledge more appropriately and systematically and further improve the generalization ability of Med-VLP methods.
In this paper, we propose a systematic approach to Med-VLP enhanced by structured expert domain knowledge from the Unified Medical Language System (Bodenreider, 2004) (UMLS), a large medical knowledge base containing many biomedical terminologies with the associated information, such as synonyms and categorical groupings.
To ensure the effectiveness and efficiency of our approach, structured knowledge is injected into the Med-VLP system from three perspectives:
(i) Aligning Through Knowledge: It uses knowledge as the intermediate medium between vision and language to align the image and text features encoded by the uni-modal encoders;
(ii) Reasoning Using Knowledge: It develops a knowledge-enhanced multi-modal fusion module to integrate knowledge into the interaction process of the image and text features;
(iii) Learning From Knowledge: It constructs knowledge-induced pretext tasks to assist the model in capturing underlying critical medical information of the images and texts to promote the medical vision-and-language understanding.
As a result, the proposed method is able to learn cross-modal domain-specific knowledge from large-scale medical image-text datasets and medical knowledge bases to promote the learning of semantically aligned and knowledge-aware image and text representations.
We perform the pre-training on three large-scale medical image-text datasets, i.e., ROCO (Pelka et al., 2018), MedICaT (Subramanian et al., 2020), and MIMIC-CXR (Johnson et al., 2019).
To verify the effectiveness of our approach and facilitate further research, we construct a medical vision-and-language understanding benchmark including three tasks (i.e., Med-VQA, Med-ITC, and Med-ITR).
Experimental results demonstrate the effectiveness of our approach, where state-of-the-art performance is achieved on all datasets.
2. Related Work
Vision-and-Language Pre-training (VLP)
Motivated by the success of the self-supervised pre-training recipe of BERT in NLP, there has been an increasing interest in developing VLP methods to address a wide range of vision-and-language tasks.
In general, VLP methods can be categorized with respect to three perspectives.
For the designs of the uni-modal encoders, different methods adopt different image features (e.g., region features (Li et al., 2019; Lu et al., 2019), patch embeddings (Kim et al., 2021; Li et al., 2021; Wang et al., 2021b), and grid features (Huang et al., 2020)) and distinct text features (e.g., statistic embeddings (Kim et al., 2021) and dynamic embeddings (Dou et al., 2021)).
For multi-modal fusion modules, existing methods can be classified into two categories (i.e., single-stream and dual-stream).
In specific, for the single-stream fusion, the models (Li et al., 2019; Chen et al., 2020; Su et al., 2019; Li et al., 2020b) use a single Transformer for early and unconstrained fusion between modalities; for the dual-stream fusion, the models (Tan and Bansal, 2019; Lu et al., 2019; Yu et al., 2021) adopt the co-attention mechanism to interact different modalities.
For pretext tasks, inspired by uni-modal pre-training schemes such as MLM (Devlin et al., 2019; Liu et al., 2019) and causal language modeling (Brown et al., 2020), existing studies explore a variety of pre-training tasks, including MLM (Li et al., 2019; Lu et al., 2019; Tan and Bansal, 2019), MIM (Lu et al., 2019; Chen et al., 2020), ITM (Li et al., 2019; Zhang et al., 2021), image-text contrastive (Li et al., 2021) and prefix language modeling (Wang et al., 2021b).
This paper adopts a purely Transformer-based backbone architecture using the dual-stream fusion with ViT-based grid features and BERT-based dynamic text features and three common pretext tasks (i.e., MLM, MIM, and ITM).
Medical Vision-and-Language Pre-training (Med-VLP)
Being one of the applications and extensions of VLP to the medical domain, Med-VLP aims to understand the content and relations between medical images and their corresponding texts.
It can be traced back to (Li et al., 2020a), which explored the performance of four vision-and-language models pre-trained in the general domain on a disease classification task.
Then MMBERT (Khare et al., 2021), PubMedCLIP (Eslami et al., 2021), and MedViLL (Moon et al., 2021) performed pre-training on medical image-text data before fine-tuning on the downstream tasks.
Compared with these studies, we design a more appropriate and systematic scheme for Med-VLP from four aspects (i.e., pre-training datasets, model designs, pre-training tasks, and evaluation benchmarks).
Knowledge-Enhanced Pre-training
For uni-modal pre-training in CV and NLP, many works have investigated how to incorporate knowledge into the pre-trained models.
According to the knowledge injection schemes, existing studies can be classified into four categories: embeddings combination (Zhang et al., 2019; Peters et al., 2019), data structure compatibility (Liu et al., 2020; Sun et al., 2020; He et al., 2020), knowledge supervision (Sun et al., 2019; Wang et al., 2021a), and neural-symbolic methods (Amizadeh et al., 2020).
For VLP, knowledge can be acquired from both the image and text modalities, and there are several works (Li et al., 2020b; Yu et al., 2021; Cui et al., 2021) studying to integrate knowledge into their methods.
ERNIE-ViL (Yu et al., 2021) built detailed semantic alignments between vision and language based on the scene graph parsed from the text.
ROSITA (Cui et al., 2021) proposed integrating extra cross-modal knowledge mappings to enhance the learning of semantic alignments between vision and language.
Different from them, we revisit existing knowledge-enhanced methods and propose to inject knowledge from three VLP-specific perspectives without requiring object detectors or scene graph parsers, which are unavailable in the medical domain.
3. The Proposed Approach
We follow the standard pre-train-and-fine-tune paradigm for medical vision-and-language understanding.
In the pre-training stage, the framework develops a variety of pretext tasks to train the Med-VLP model using medical image-text pairs.
In the fine-tuning stage, the pre-trained Med-VLP model is transferred to various medical vision-and-language downstream tasks.
An overview of the proposed approach is demonstrated in Figure 2, and the details of the general Med-VLP framework, the knowledge extraction process (shown in Figure 1), and the injection of knowledge into the general Med-VLP framework are introduced in the following subsections.
3.1. The General Med-VLP Framework
The general Med-VLP framework can be partitioned into three major components, i.e., the uni-modal encoders, the multi-modal fusion module, and the pretext tasks.
The overall description of the three components is detailed below.
Uni-modal Encoders
In the Med-VLP framework, there is a vision encoder and a language encoder, which encode the input image and text into image and text features, respectively.
For the vision encoder, we study the use of vision Transformer (Dosovitskiy et al., 2020) (ViT).
In ViT, an input image $\boldsymbol{I}\in\mathbb{R}^{H\times W\times C}$ is first segmented into patches $\{\boldsymbol{x_{1}^{v}},\boldsymbol{x_{2}^{v}},\ldots,\boldsymbol{x_{N_{v}}^{v}}\}$, where $H\times W$ is the image resolution, $C$ is the number of channels, $N_{v}$ is the number of patches, $\boldsymbol{x_{i}^{v}}\in\mathbb{R}^{P^{2}\times C}$ and $P\times P$ is the patch resolution.
Then the patches $\{\boldsymbol{x_{1}^{v}},\boldsymbol{x_{2}^{v}},\ldots,\boldsymbol{x_{N_{v}}^{v}}\}$ are flattened and linearly projected into patch embeddings through a linear transformation $\boldsymbol{E^{v}}\in\mathbb{R}^{P^{2}C\times D}$ and a special learnable token embedding $\boldsymbol{x_{I}^{v}}\in\mathbb{R}^{D}$ is prepended for the aggregation of visual information.
Therefore, the input image representations are obtained via summing up the patch embeddings and learnable 1D position embeddings $\boldsymbol{E^{v}_{pos}}\in\mathbb{R}^{(N_{v}+1)\times D}$:
(1)
$$\boldsymbol{X^{v}}=[\boldsymbol{x_{I}^{v}};\boldsymbol{x_{1}^{v}}\boldsymbol{E^{v}};\boldsymbol{x_{2}^{v}}\boldsymbol{E^{v}};\ldots;\boldsymbol{x_{N_{v}}^{v}}\boldsymbol{E^{v}}]+\boldsymbol{E^{v}_{pos}}.$$
Then $\boldsymbol{X^{v}}$ is fed into a Transformer model with $L_{v}$ Transformer layers.
Finally, we obtain the contextualized image representations $\boldsymbol{H^{v}}=[\boldsymbol{h^{v}_{I}};\boldsymbol{h^{v}_{1}};\boldsymbol{h^{v}_{2}};...;\boldsymbol{h^{v}_{N_{v}}}]$.
For the language encoder, we follow BERT (Devlin et al., 2019) to tokenize the input text to subword tokens $\{x_{1}^{l},x_{2}^{l},\ldots,x_{N_{l}}^{l}\}$ by WordPiece (Wu et al., 2016) and then represent subword tokens as $\{\boldsymbol{x_{1}^{l}},\boldsymbol{x_{2}^{l}},\ldots,\boldsymbol{x_{N_{l}}^{l}}\}$, where $\boldsymbol{x_{i}^{l}}\in\mathbb{R}^{V}$ are the one-hot form of $x_{i}^{l}$, $V$ is the vocabulary size, and $N_{l}$ is the number of tokens.
Subsequently, the tokens are linearly projected into embeddings through a linear transformation $\boldsymbol{E^{l}}\in\mathbb{R}^{V\times D}$.
Afterwards, a start-of-sequence token embedding $\boldsymbol{x_{T}^{l}}\in\mathbb{R}^{D}$ and a special boundary token embedding $\boldsymbol{x_{SEP}^{l}}\in\mathbb{R}^{D}$ are added to the text sequence.
Therefore, the input text representations are computed via summing up the token embeddings and text position embeddings $\boldsymbol{E^{l}_{pos}}\in\mathbb{R}^{(N_{l}+2)\times D}$:
(2)
$$\boldsymbol{X^{l}}=[\boldsymbol{x_{T}^{l}};\boldsymbol{x_{1}^{l}E^{l}};\ldots;\boldsymbol{x_{N_{l}}^{l}E^{l}};\boldsymbol{{x}_{SEP}^{l}}]+\boldsymbol{E^{l}_{pos}}.$$
Then $\boldsymbol{X^{l}}$ is fed into a Transformer model with $L_{l}$ Transformer layers.
Finally, we obtain the contextualized text representations $\boldsymbol{H^{l}}=[\boldsymbol{h^{l}_{T}};\boldsymbol{h^{l}_{1}};\boldsymbol{h^{l}_{2}};\ldots;\boldsymbol{h^{l}_{N_{l}}};\boldsymbol{h^{l}_{SEP}}]$.
Multi-model Fusion Module
We adopt the co-attention mechanism in the multi-modal fusion module to fuse the contextualized representations from images and texts.
In detail, the multi-modal fusion module consists of two Transformer models for vision and language, respectively, each of which is a stack of $L_{m}$ Transformer layers.
In each Transformer layer, there are three sub-layers, i.e., a self-attention sub-layer, a cross-attention sub-layer, and a feed-forward sub-layer.
The attention mechanism is applied in the self-attention and cross-attention sub-layers and is defined as
(3)
$$\text{ATTN}(\boldsymbol{Q},\boldsymbol{K},\boldsymbol{V})=\operatorname{softmax}\left(\boldsymbol{Q}\boldsymbol{K}^{\top}/\sqrt{D_{k}}\right)\cdot\boldsymbol{V},$$
where $\boldsymbol{Q}$, $\boldsymbol{K}$, and $\boldsymbol{V}$ are the query, key, value matrices linearly transformed from the corresponding input sequences, respectively, and $D_{k}$ is the dimension of $\boldsymbol{K}$.
In the self-attention sub-layer, the representations interact within modalities:
(4)
$$\displaystyle\boldsymbol{H^{vs}}$$
$$\displaystyle=\text{ATTN}(\boldsymbol{H^{v}},\boldsymbol{H^{v}},\boldsymbol{H^{v}}),$$
(5)
$$\displaystyle\boldsymbol{H^{ls}}$$
$$\displaystyle=\text{ATTN}(\boldsymbol{H^{l}},\boldsymbol{H^{l}},\boldsymbol{H^{l}}),$$
where $\boldsymbol{H^{vs}}$ and $\boldsymbol{H^{ls}}$ are the self-attention outputs for vision and language, respectively.
Then residual connections followed by layer normalization are employed to $\boldsymbol{H^{vs}}$ and $\boldsymbol{H^{ls}}$ and we denote the results as $\boldsymbol{H^{vs}}$ and $\boldsymbol{H^{ls}}$, respectively, for simplicity.
In the cross-attention sub-layer, the representations interact across modalities to integrate cross-modal information into their representations:
(6)
$$\displaystyle\boldsymbol{H^{vc}}=\text{ATTN}(\boldsymbol{H^{vs}},\boldsymbol{H^{ls}},\boldsymbol{H^{ls}}),$$
(7)
$$\displaystyle\boldsymbol{H^{lc}}=\text{ATTN}(\boldsymbol{H^{ls}},\boldsymbol{H^{vs}},\boldsymbol{H^{vs}}),$$
where $\boldsymbol{H^{vc}}$ and $\boldsymbol{H^{lc}}$ are the cross-attention outputs for vision and language, respectively.
Similarly, residual connections followed by layer normalization are employed to $\boldsymbol{H^{vc}}$ and $\boldsymbol{H^{lc}}$, and we denote the results as $\boldsymbol{H^{vc}}$ and $\boldsymbol{H^{lc}}$, respectively, for simplicity.
Finally, $\boldsymbol{H^{vc}}$ and $\boldsymbol{H^{lc}}$ are input to the feed-forward sub-layer (i.e., a multi-layer perceptron (MLP)) to obtain the multi-modal representations $\boldsymbol{Z^{v}}=[\boldsymbol{z^{v}_{I}};\boldsymbol{z^{v}_{1}};\boldsymbol{z^{v}_{2}};\ldots;\boldsymbol{z^{v}_{N_{v}}}]$ for vision and $\boldsymbol{Z^{l}}=[\boldsymbol{z^{l}_{T}};\boldsymbol{z^{l}_{1}};\boldsymbol{z^{l}_{2}};\ldots;\boldsymbol{z^{l}_{N_{l}}};\boldsymbol{z^{l}_{SEP}}]$ for language.
Pretext Tasks
Given the aforementioned structure (denoted as $\mathcal{M}_{\theta}$) with its parameters $\theta$, the Med-VLP framework develops various pretext tasks (e.g., masked language modeling (MLM), masked image modeling (MIM), and image-text matching (ITM)) to guide the learning of $\theta$.
Assuming there are $S$ pretext tasks, the learning process can be formalized as
(8)
$$\theta^{*},\theta_{1}^{*},...,\theta_{S}^{*}=\mathop{\arg\min}_{\theta,\theta_{1},...,\theta_{S}}\sum_{i=1}^{S}L_{i}(Y_{i},\mathcal{D}_{\theta_{i}}(\mathcal{M}_{\theta}(I,T)),$$
where $L_{i}$ are the loss functions of pretext tasks, $Y_{i}$ are the corresponding ground-truth labels, and $\mathcal{D}_{\theta_{i}}$ are the prediction heads with their parameters $\theta_{i}$.
3.2. Knowledge Extraction
Although knowledge graphs (KGs) have shown their effectiveness in many natural language processing (NLP) tasks (Zhang et al., 2019; Peters et al., 2019; Liu et al., 2020) and computer vision (CV) tasks (Marino et al., 2017; Wang et al., 2018), the existing Med-VLP methods rarely consider incorporating KGs to provide rich structured knowledge for better vision-and-language understanding.
Therefore, we propose to enhance Med-VLP by leveraging external domain expert knowledge from UMLS.
In doing so, we extract knowledge through two stages, as illustrated in Figure 1.
The first stage is to apply a named entity recognition, and linking tool ScispaCy (Neumann et al., 2019) to pre-process the texts in the pre-training corpus to link entities in the texts to the UMLS knowledge base for entity disambiguation.
Therefore, for each image-text pair, there is an entity sequence $ES=\{x_{1}^{e},x_{2}^{e},\ldots,x_{N_{es}}^{e}\}$ aligning to the token sequence $T=\{x_{1}^{l},x_{2}^{l},\ldots,x_{N_{l}}^{l}\}$, where $x_{i}^{e}$ are the extracted entities and $N_{es}$ is the length of the entity sequence.
Following (Diao et al., 2020), to record the position of the extracted entities, we adopt an entity matching matrix $\boldsymbol{P}\in\mathbb{R}^{N_{l}\times N_{es}}$, where each element is represented by
(9)
$$\boldsymbol{P}_{ij}=\begin{cases}1&x_{i}^{l}\in x_{j}^{e}\\
0&x_{i}^{l}\notin x_{j}^{e}\end{cases},$$
where $P$ is employed to assist the interaction between the text and the entity sequence in the knowledge injection process (as described in the next subsection).
After pre-processing all the texts, we can obtain an entity set $\mathcal{E}=\{e_{i}\}_{i=1}^{N_{e}}$ containing all the $N_{e}$ entities related to the pre-training corpus.
The second stage is to extract relevant knowledge graph triples from the UMLS knowledge base once both the head and tail entities of the triple are in the entity set $\mathcal{E}$.
We denote the extracted knowledge graph (i.e., a sub-graph of the UMLS knowledge base) as the knowledge set $\mathcal{G}=\{k_{i}=(h_{i},r_{i},t_{i})\}_{i=1}^{N_{g}}$, where $N_{g}$ is the number of knowledge graph triples, $k_{i}$ are the knowledge graph triples, and $h_{i}$, $r_{i}$ and $t_{i}$ represent the head entity, relation, and tail entity, respectively.
3.3. Knowledge Injection
To integrate knowledge into the general Med-VLP framework, first, we perform knowledge representation following two steps:
(i) We apply knowledge representation learning algorithm (e.g., TransE (Bordes et al., 2013)) to the knowledge graph $\mathcal{G}=\{k_{i}=(h_{i},r_{i},t_{i})\}_{i=1}^{N_{g}}$ to obtain the entity embeddings $\{\boldsymbol{e_{i}}\}_{i=1}^{N_{e}}$, where $\boldsymbol{e_{i}}\in\mathbb{R}^{D_{e}}$ and $D_{e}$ is the dimension of the entity embeddings;
(ii) We adopt Graph Neural Networks (e.g., Graph Attention Networks (Veličković et al., 2018)) to take account of the whole structure of the graph to aggregate local information in the graph neighborhood for each node, and obtain the entity representations (denoted as $\{\boldsymbol{e_{i}}\}_{i=1}^{N_{e}}$ for simplicity, where $\boldsymbol{e_{i}}\in\mathbb{R}^{D_{e}}$).
Afterwards, given the input image $I$ and text $T=\{x_{1}^{l},x_{2}^{l},\ldots,x_{N_{l}}^{l}\}$ with its corresponding entity sequence $ES=\{x_{1}^{e},x_{2}^{e},\ldots,x_{N_{es}}^{e}\}$, we develop three essential and systematic designs to inject knowledge from the following perspectives:
(i) Aligning Through Knowledge
Knowledge can be regarded as the intermediate medium between vision and language, where knowledge can be used as an explanation of the meaning behind both images and texts.
In most cases, entities in knowledge graph triples can be treated as anchor points that appear in the image and are mentioned in the accompanying text.
Motivated by this fact, we propose to align the image representations and the text representations from uni-modal encoders through knowledge.
Similar to (Li et al., 2021), it serves two purposes:
It improves the unimodal encoders to better understand the semantic meaning of images and texts;
It eases the learning of semantic alignments between images and texts.
Formally, given the aggregated image representation $\boldsymbol{h^{v}_{I}}$ and text representation $\boldsymbol{h^{l}_{T}}$, we calculate the image-knowledge and text-knowledge similarity followed by a sigmoid function:
(10)
$$\displaystyle p_{i}^{v}$$
$$\displaystyle=\text{sigmoid}(\boldsymbol{e_{i}}^{\top}\boldsymbol{W_{vk}}\boldsymbol{h^{v}_{I}}),i=1,...,N_{e},$$
(11)
$$\displaystyle p_{i}^{l}$$
$$\displaystyle=\text{sigmoid}(\boldsymbol{e_{i}}^{\top}\boldsymbol{W_{lk}}\boldsymbol{h^{l}_{T}}),i=1,...,N_{e},$$
where $\boldsymbol{W_{vk}}\in\mathbb{R}^{D_{e}\times D}$ and $\boldsymbol{W_{lk}}\in\mathbb{R}^{D_{e}\times D}$ are trainable weights for the linear transformation.
Therefore, the alignments for image-knowledge and text-knowledge are learned explicitly through minimizing the following functions:
(12)
$$\displaystyle L_{vk}$$
$$\displaystyle=-\sum_{i=1}^{N_{e}}(y_{i}\log p_{i}^{v}+\left(1-y_{i}\right)\log\left(1-p_{i}^{v}\right)),$$
(13)
$$\displaystyle L_{lk}$$
$$\displaystyle=-\sum_{i=1}^{N_{e}}(y_{i}\log p_{i}^{l}+\left(1-y_{i}\right)\log\left(1-p_{i}^{l}\right)),$$
where $y_{i}$ can be defined as
(14)
$$y_{i}=\begin{cases}1&e_{i}\in ES\\
0&e_{i}\notin ES\end{cases}.$$
Therefore, knowledge is employed as an intermediate medium to enhance and smooth image-text mappings by doing so.
(ii) Reasoning Using Knowledge
As the supplementation of the input image and text, knowledge can also be utilized to assist the reasoning of the Med-VLP model.
In doing so, we enhance the multi-modal fusion module by knowledge.
Formally, given the entity sequence $ES=\{x_{1}^{e},x_{2}^{e},...,x_{N_{es}}^{e}\}$, first, we extract its entity representations $\boldsymbol{H^{e}}=[\boldsymbol{h^{e}_{1}};\boldsymbol{h^{e}_{2}};\ldots;\boldsymbol{h^{e}_{N_{es}}}]$.
Second, we apply self-attention to the entity representations to encode the contextualized information:
(15)
$$\boldsymbol{H^{es}}=\text{ATTN}(\boldsymbol{H^{e}},\boldsymbol{H^{e}},\boldsymbol{H^{e}}),$$
where $\boldsymbol{H^{es}}$ is the self-attention outputs of entity representations.
Residual connections followed by layer normalization are employed to $\boldsymbol{H^{es}}$, and we denote the outputs as $\boldsymbol{H^{es}}$ for simplicity.
Third, to interact the entities with the image, since there is no available toolkits to structuralize the input image to construct mappings between image patches and entities, we directly perform cross-attention on $\boldsymbol{H^{es}}$ and $\boldsymbol{H^{vs}}$:
(16)
$$\boldsymbol{H^{ec}}=\text{ATTN}(\boldsymbol{H^{es}},\boldsymbol{H^{vs}},\boldsymbol{H^{vs}}),$$
where $\boldsymbol{H^{ec}}$ is the cross-attention outputs of entity representations.
Residual connections followed by layer normalization are employed to $\boldsymbol{H^{ec}}$, and we denote the outputs as $\boldsymbol{H^{ec}}$ for simplicity.
Fourth, since the mappings between the entities and the text are recorded by $\boldsymbol{P}$, we can use it to fuse $\boldsymbol{H^{ec}}$ and $\boldsymbol{H^{ls}}$ through:
(17)
$$\boldsymbol{\tilde{H}^{lc}}=\boldsymbol{P}\boldsymbol{H^{ec}}+\boldsymbol{H^{lc}},$$
where $\boldsymbol{\tilde{H}^{lc}}$ is the text representations encoded with the image and entity information.
Finally, we apply residual connections followed by layer normalization to $\boldsymbol{\tilde{H^{lc}}}$ and input it to the feed-forward sub-layer to complete the knowledge-enhanced multi-modal fusion.
In the meantime, $\boldsymbol{H^{ec}}$ is input to another feed-forward sub-layer to produce the representations $\boldsymbol{Z^{e}}$ for the next layer.
(iii) Learning From Knowledge
Knowledge can help us to induce more sophisticated pretext tasks to guide the model to learn more informative representations.
In our paper, we follow (Sun et al., 2019) to design a knowledge-induced mask generation strategy.
Specifically, when performing the MLM task given the input text $T=\{x_{1}^{l},x_{2}^{l},\ldots,x_{N_{l}}^{l}\}$ with its entity sequence $ES=\{x_{1}^{e},x_{2}^{e},\ldots,x_{N_{es}}^{e}\}$, we do not mask subword tokens in $T$ randomly.
Instead, we randomly sample entities from $ES$ and then mask consecutive spans of subword tokens belonging to the sampled entities.
Since entities can be abstract or have a physical existence, it can force the model to focus on critical medical information in both images and texts.
Therefore, knowledge can be injected into the Med-VLP framework in a systematic way through the above three designs.
4. Experimental Settings
4.1. Pre-training Setup
Datasets
In our experiments, we perform the pre-training on three datasets, which are described as follows:
•
ROCO (Pelka et al., 2018): a dataset of radiology figure-caption pairs from PubMed Central, an open-access biomedical literature database.
It has over 81,000 radiology images (from various imaging modalities) and their corresponding captions.
•
MedICaT (Subramanian et al., 2020): a dataset of medical figure-caption pairs also extracted from PubMed Central.
Different from ROCO, 75% of its figures are compound figures, including several sub-figures.
It contains over 217,000 images with their captions and inline textual references.
•
MIMIC-CXR (Johnson et al., 2019): the largest radiology dataset to date that consists of 473,057 chest X-ray images (in frontal or lateral views) and 206,563 reports from 63,478 patients from the Beth Israel Deaconess Medical Center.
For all the datasets, we exclude those samples with the length of their texts less than 3.
For ROCO and MedICaT, we filter non-radiology samples, and for MIMIC-CXR, we only keep images in the frontal view.
As for the dataset split, we adopt the official splits of ROCO and MIMIC-CXR.
For MedICaT, we randomly sample 1,000 image-text pairs for validation and 1,000 for test, and the remaining image-text pairs are used for training.
Implementation Details
For the uni-modal encoders, we use the vision encoder with CLIP-ViT-B (Radford et al., 2021) ($L_{v}=12$) and the language encoder with RoBERTa-base (Liu et al., 2019) ($L_{l}=12$).
For the multi-modal fusion module, we set the number of Transformer layers $L_{m}=6$, and the dimension of the hidden states $D=768$ with the number of heads set to 12.
For knowledge representation and injection, we set the dimension of the hidden states $D_{e}=256$.
For the pretext tasks, we adopt (knowledge-enhanced) MLM, MIM (He et al., 2021), and ITM, where the masking ratios of MLM and MIM are set to 15% and 75%, respectively.
For the optimization, the models are trained with AdamW optimizer (Loshchilov and Hutter, 2018) for 100,000 steps with the learning rates for the uni-modal encoders and the remaining parameters set to 1e-5 and 5e-5, respectively.
The warm-up ratio is set to 10%, and the learning rate is linearly decayed to 0 after warm-up.
Besides, we use center-crop to resize each image to the size of 288$\times$288.
4.2. Vision-and-Language Transfer Tasks
To evaluate the performance, we construct a medical vision-and-language understanding benchmark including three tasks.
The details of the tasks and fine-tuning strategies are described below.
Medical Visual Question Answering (Med-VQA)
This task requires the model to answer natural language questions about a medical image.
We adopt three publicly available Med-VQA datasets (i.e., VQA-RAD (Lau et al., 2018), SLACK (Liu et al., 2021b) and VQA-2019 (Abacha et al., 2019)), where VQA-RAD consists of 315 images and 3515 questions, SLACK contains 642 images and 14,028 questions, and VQA-2019 contains 4,200 images and 15,292 questions.
To fine-tune on this task, we regard it as a multi-label classification task and feed the concatenation of the image and text representations to a two-layer MLP to predict the corresponding answer.
During training, the models are trained with a binary cross-entropy loss with a batch size of 64.
Medical Image-Text Classification (Med-ITC)
This task aims to produce the classification label given an image-text pair.
We evaluate the performance on MELINDA (Wu et al., 2021), a Biomedical Experiment Method Classification dataset that contains 5,371 image-text pairs.
To fine-tune on this task, we learn a two-layer MLP on top of the concatenation of the image and text representations.
We train the models with a cross-entropy loss with a batch size of 16 over a maximum of 20 epochs.
Medical Image-Text Retrieval (Med-ITR)
The target of this task is to calculate a similarity score between an image and a text and then perform cross-modal retrieval.
There are two subtasks for this task, where image-to-text (I2T) retrieval requires retrieving the most relevant texts from a large pool of texts given an image, and vice versa for text-to-image (T2I) retrieval.
We conduct experiments on the ROCO dataset and measure both zero-shot and fine-tuned performance.
To fine-tune on this task, we initialize the similarity score head from the pre-trained ITM head.
The model is tuned with cross-entropy loss to maximize the scores on positive pairs with 15 random texts sampled as negative samples with a batch size of 256 over a maximum of 10 epochs.
During the evaluation, we sample 2,000 image-text pairs from the ROCO test set and report the results on the sampled 2,000 image-text pairs due to the large time complexity of the ranking process.222The time complexity of the ranking process is $O(N^{2})$, where $N$ is the sample number.
For all tasks, we use the AdamW optimizer with the learning rate set to 5e-6 and 2.5e-5 for the model backbone and prediction heads, respectively, and the warm-up ratio set to 10%.
For the evaluation metrics, we adopt accuracy for Med-VQA and Med-ITC, and Recall@K333Recall@K corresponds to whether the ground truth is included among top K results. (K=1, 5, 10) for Med-ITR.
5. Experimental Results
5.1. Main Results
We compare the proposed approach with existing methods on the same datasets, with their results reported in Table 1, 2, and 3.
There are several observations drawn from different aspects.
First, our approach achieves the best performance on all tasks, which confirms the validity of the proposed pre-training approach.
Second, for Med-VQA, our approach achieves significant improvements even compared with those pre-training methods (e.g., MTPT and MMBERT), which indicates the usefulness of incorporating knowledge into the pre-training process.
Third, for Med-ITC, the results of those strong baselines (i.e., RoBERTa, SciBERT, and ViL-BERT) are achieved by continued pre-training on the MELINDA dataset.
Our approach achieves better performance without such requirements, which indicates that an appropriate design can alleviate the need for continued pre-training on downstream datasets.
Fourth, for Med-ITR, the proposed approach achieves a substantial improvement compared with the state-of-the-art methods, where ViLT and METER are two strong baselines in the general domain.
This shows that it is necessary to design an appropriate approach (including pre-training data and methods) for the medical domain.
5.2. Ablation Studies
To illustrate the effectiveness of our proposed approach, we perform an ablation study on the three proposed knowledge injection designs.
The experiments are conducted on the VQA-RAD dataset, and the results are reported in Table 4.
We have the following observations.
First, for the model parameters, only the RK design brings extra 12M parameters ($\sim$3.4% of the whole model) while other designs do not add additional parameters, which justifies introducing knowledge to Med-VLP through our approach can be done with a small price.
Second, the results of pre-training with two designs (ID 5, 6, and 7) and one design (ID 2, 3, and 4) are consistently better than those of pre-training with one design (ID 2, 3, and 4) and without any knowledge injection design (ID 1), respectively.
This demonstrates the excellent compatibility and complementarity of our design perspectives, which is critical in a multi-component approach and allows us to develop more designs under such a framework.
Third, injecting the knowledge into the multi-modal fusion module (ID 3) achieves a significant improvement.
The reason behind this might be that knowledge (i.e., entities here) serves three functions:
(i) It smooths the interaction process of the image and text representations;
(ii) It provides information at a greater granularity than words;
(iii) It removes ambiguity between diverse words by linking to the knowledge base.
Fourth, performing the aligning process can further improve the performance of the RK design (ID 5), which can be explained by the fact that the aligning processing can produce better knowledge representations for the RK process.
Fifth, our full approach (ID 8) achieves the best performance, which confirms the effectiveness of the proposed framework for medical knowledge injection.
5.3. Qualitative Analysis
To further investigate the effectiveness of our approach, we perform qualitative analysis on some cases with their results shown in Figure 3.
For “aligning through knowledge”, in the input text of this case, there is an entity “Brain magnetic resonance
imaging” which links to the entity “C4028269: Nuclear magnetic resonance imaging brain” in the UMLS knowledge base.
The sub-figures (3(C) and 3(D)) show that the image and text representations produced by the uni-modal encoders have high similarity with the entity representation of “C4028269”, which implicitly pulls the image and text representations close (as shown in the sub-figure 3(E)).
For “reasoning using knowledge”, the sub-figures 3(H), 3(I), and 3(J) illustrate that using entities is beneficial for aligning the text with the image, where the learned entity-image attention mappings are better than the subword-image mappings.
The reason behind this is that entities are more complete semantic units.
In contrast, words (or subwords) have a smaller granularity than entities, making the correspondences between images and texts more challenging to learn.
For “learning from knowledge”, the masked medical entities are correctly recovered by the pre-trained model (as shown in the sub-figure 3(O)) since the knowledge-induced pretext task guides the model to put more emphasis on the medical knowledge-related information.
In addition, the masked and recovered images are also shown in the sub-figures 3(K) and 3(M), respectively, which shows the high quality of the image reconstruction.
In summary, these cases reveal that injecting knowledge through the three proposed designs is essential in modeling the hidden structures among the images and texts better to promote Med-VLP.
6. Conclusion
In this paper, we propose to pre-train the medical vision-and-language model with medical domain knowledge, where the knowledge is injected into the Med-VLP framework from three aspects: (i) aligning the image and text representations through knowledge before their interaction; (ii) treating knowledge as the supplementation of the input image and text to assist the reasoning during the multi-modal fusion process; (iii) utilizing knowledge to induce more sophisticated pretext tasks to guide the model put more emphasis on the critical medical information.
To perform a comprehensive evaluation and facilitate further research, we construct a medical vision-and-language understanding benchmark, including three tasks (i.e., Med-VQA, Med-ITC, and Med-ITR).
Experimental results on the downstream datasets demonstrate the effectiveness of our approach, which achieves state-of-the-art performance.
Further analyses investigate the effects of different components in our approach and show that our approach is able to better learn the correspondences between vision and language so as to produce more generic and effective vision-and-language representations.
Acknowledgement
This work is supported in part by the Chinese Key-Area Research and Development Program of Guangdong Province (2020B0101350001), in part by the Guangdong Basic and Applied Basic Research Foundation (2020B1515020048), in part by the National Natural Science Foundation of China (61976250), in part by the Guangzhou Science and technology project (No.202102020633), and is also supported by the Guangdong Provincial Key Laboratory of Big Data Computing, The Chinese University of Hong Kong, Shenzhen.
References
(1)
Abacha et al. (2019)
Asma Ben Abacha, Sadid A.
Hasan, Vivek Datla, Joey Liu,
Dina Demner-Fushman, and Henning
Müller. 2019.
VQA-Med: Overview of the Medical Visual Question
Answering Task at ImageCLEF 2019. In CLEF.
Amizadeh et al. (2020)
Saeed Amizadeh, Hamid
Palangi, Alex Polozov, Yichen Huang,
and Kazuhito Koishida. 2020.
Neuro-Symbolic Visual Reasoning: Disentangling. In
International Conference on Machine Learning.
PMLR, 279–290.
Beltagy et al. (2019)
Iz Beltagy, Kyle Lo,
and Arman Cohan. 2019.
SciBERT: A Pretrained Language Model for Scientific
Text. In Proceedings of the 2019 Conference on
Empirical Methods in Natural Language Processing and the 9th International
Joint Conference on Natural Language Processing (EMNLP-IJCNLP).
3615–3620.
Bodenreider (2004)
Olivier Bodenreider.
2004.
The Unified Medical Language System (UMLS):
integrating biomedical terminology.
Nucleic acids research
32, Database issue
(2004), D267–D270.
Bordes et al. (2013)
Antoine Bordes, Nicolas
Usunier, Alberto Garcia-Duran, Jason
Weston, and Oksana Yakhnenko.
2013.
Translating embeddings for modeling
multi-relational data.
Advances in neural information processing
systems 26 (2013).
Brown et al. (2020)
Tom Brown, Benjamin Mann,
Nick Ryder, Melanie Subbiah,
Jared D Kaplan, Prafulla Dhariwal,
Arvind Neelakantan, Pranav Shyam,
Girish Sastry, Amanda Askell,
et al. 2020.
Language models are few-shot learners.
Advances in neural information processing
systems 33 (2020),
1877–1901.
Chen et al. (2021)
Kezhen Chen, Qiuyuan
Huang, Yonatan Bisk, Daniel McDuff,
and Jianfeng Gao. 2021.
KB-VLP: Knowledge Based Vision and Language
Pretraining. In Proceedings of the 38 th
International Conference on Machine Learning, PMLR 139, 2021. ICML, workshop,
2021.
Chen et al. (2020)
Yen-Chun Chen, Linjie Li,
Licheng Yu, Ahmed El Kholy,
Faisal Ahmed, Zhe Gan,
Yu Cheng, and Jingjing Liu.
2020.
Uniter: Universal image-text representation
learning. In European conference on computer
vision. Springer, 104–120.
Cui et al. (2021)
Yuhao Cui, Zhou Yu,
Chunqi Wang, Zhongzhou Zhao,
Ji Zhang, Meng Wang, and
Jun Yu. 2021.
ROSITA: Enhancing Vision-and-Language Semantic
Alignments via Cross-and Intra-modal Knowledge Integration. In
Proceedings of the 29th ACM International
Conference on Multimedia. 797–806.
Devlin et al. (2019)
Jacob Devlin, Ming-Wei
Chang, Kenton Lee, and Kristina
Toutanova. 2019.
BERT: Pre-training of Deep Bidirectional
Transformers for Language Understanding. In
Proceedings of the 2019 Conference of the North
American Chapter of the Association for Computational Linguistics: Human
Language Technologies, Volume 1 (Long and Short Papers).
4171–4186.
Diao et al. (2020)
Shizhe Diao, Jiaxin Bai,
Yan Song, Tong Zhang, and
Yonggang Wang. 2020.
ZEN: Pre-training Chinese Text Encoder Enhanced by
N-gram Representations. In Findings of the
Association for Computational Linguistics: EMNLP 2020.
4729–4740.
Dosovitskiy et al. (2020)
Alexey Dosovitskiy, Lucas
Beyer, Alexander Kolesnikov, Dirk
Weissenborn, Xiaohua Zhai, Thomas
Unterthiner, Mostafa Dehghani, Matthias
Minderer, Georg Heigold, Sylvain Gelly,
et al. 2020.
An Image is Worth 16x16 Words: Transformers for
Image Recognition at Scale. In International
Conference on Learning Representations.
Dou et al. (2021)
Zi-Yi Dou, Yichong Xu,
Zhe Gan, Jianfeng Wang,
Shuohang Wang, Lijuan Wang,
Chenguang Zhu, Zicheng Liu,
Michael Zeng, et al.
2021.
An Empirical Study of Training End-to-End
Vision-and-Language Transformers.
arXiv preprint arXiv:2111.02387
(2021).
Eslami et al. (2021)
Sedigheh Eslami, Gerard de
Melo, and Christoph Meinel.
2021.
Does CLIP Benefit Visual Question Answering in the
Medical Domain as Much as it Does in the General Domain?
arXiv preprint arXiv:2112.13906
(2021).
Gong et al. (2021)
Haifan Gong, Guanqi Chen,
Sishuo Liu, Yizhou Yu, and
Guanbin Li. 2021.
Cross-Modal Self-Attention with Multi-Task
Pre-Training for Medical Visual Question Answering. In
Proceedings of the 2021 International Conference on
Multimedia Retrieval. 456–460.
He et al. (2020)
Bin He, Di Zhou,
Jinghui Xiao, Xin Jiang,
Qun Liu, Nicholas Jing Yuan, and
Tong Xu. 2020.
BERT-MK: Integrating Graph Contextualized Knowledge
into Pre-trained Language Models. In Findings of
the Association for Computational Linguistics: EMNLP 2020.
2281–2290.
He et al. (2021)
Kaiming He, Xinlei Chen,
Saining Xie, Yanghao Li,
Piotr Dollár, and Ross Girshick.
2021.
Masked autoencoders are scalable vision learners.
arXiv preprint arXiv:2111.06377
(2021).
He et al. (2016)
Kaiming He, Xiangyu
Zhang, Shaoqing Ren, and Jian Sun.
2016.
Deep residual learning for image recognition. In
Proceedings of the IEEE conference on computer
vision and pattern recognition. 770–778.
Hochreiter and Schmidhuber (1997)
Sepp Hochreiter and
Jürgen Schmidhuber. 1997.
Long short-term memory.
Neural computation 9,
8 (1997), 1735–1780.
Huang et al. (2020)
Zhicheng Huang, Zhaoyang
Zeng, Bei Liu, Dongmei Fu, and
Jianlong Fu. 2020.
Pixel-bert: Aligning image pixels with text by deep
multi-modal transformers.
arXiv preprint arXiv:2004.00849
(2020).
Johnson et al. (2019)
Alistair EW Johnson, Tom J
Pollard, Nathaniel R Greenbaum, Matthew P
Lungren, Chih-ying Deng, Yifan Peng,
Zhiyong Lu, Roger G Mark,
Seth J Berkowitz, and Steven Horng.
2019.
MIMIC-CXR-JPG, a large publicly available database
of labeled chest radiographs.
arXiv preprint arXiv:1901.07042
(2019).
Khare et al. (2021)
Yash Khare, Viraj Bagal,
Minesh Mathew, Adithi Devi,
U Deva Priyakumar, and CV Jawahar.
2021.
Mmbert: Multimodal bert pretraining for improved
medical vqa. In 2021 IEEE 18th International
Symposium on Biomedical Imaging (ISBI). IEEE, 1033–1036.
Kim et al. (2018)
Jin-Hwa Kim, Jaehyun Jun,
and Byoung-Tak Zhang. 2018.
Bilinear attention networks.
Advances in Neural Information Processing
Systems 31 (2018).
Kim et al. (2021)
Wonjae Kim, Bokyung Son,
and Ildoo Kim. 2021.
Vilt: Vision-and-language transformer without
convolution or region supervision. In
International Conference on Machine Learning.
PMLR, 5583–5594.
Lau et al. (2018)
Jason J Lau, Soumya
Gayen, Asma Ben Abacha, and Dina
Demner-Fushman. 2018.
A dataset of clinically generated visual questions
and answers about radiology images.
Scientific data 5,
1 (2018), 1–10.
Li et al. (2021)
Junnan Li, Ramprasaath
Selvaraju, Akhilesh Gotmare, Shafiq
Joty, Caiming Xiong, and Steven
Chu Hong Hoi. 2021.
Align before fuse: Vision and language
representation learning with momentum distillation.
Advances in Neural Information Processing
Systems 34 (2021).
Li et al. (2019)
Liunian Harold Li, Mark
Yatskar, Da Yin, Cho-Jui Hsieh, and
Kai-Wei Chang. 2019.
Visualbert: A simple and performant baseline for
vision and language.
arXiv preprint arXiv:1908.03557
(2019).
Li et al. (2020b)
Xiujun Li, Xi Yin,
Chunyuan Li, Pengchuan Zhang,
Xiaowei Hu, Lei Zhang,
Lijuan Wang, Houdong Hu,
Li Dong, Furu Wei, et al.
2020b.
Oscar: Object-semantics aligned pre-training for
vision-language tasks. In European Conference on
Computer Vision. Springer, 121–137.
Li et al. (2020a)
Yikuan Li, Hanyin Wang,
and Yuan Luo. 2020a.
A comparison of pre-trained vision-and-language
models for multimodal representation learning across medical images and
reports. In 2020 IEEE International Conference on
Bioinformatics and Biomedicine (BIBM). IEEE, 1999–2004.
Liu et al. (2021a)
Bo Liu, Li-Ming Zhan,
and Xiao-Ming Wu. 2021a.
Contrastive Pre-training and Representation
Distillation for Medical Visual Question Answering Based on Radiology
Images. In International Conference on Medical
Image Computing and Computer-Assisted Intervention. Springer,
210–220.
Liu et al. (2021b)
Bo Liu, Li-Ming Zhan,
Li Xu, Lin Ma, Yan
Yang, and Xiao-Ming Wu.
2021b.
SLAKE: A semantically-labeled knowledge-enhanced
dataset for medical visual question answering. In
2021 IEEE 18th International Symposium on
Biomedical Imaging (ISBI). IEEE, 1650–1654.
Liu et al. (2020)
Weijie Liu, Peng Zhou,
Zhe Zhao, Zhiruo Wang,
Qi Ju, Haotang Deng, and
Ping Wang. 2020.
K-bert: Enabling language representation with
knowledge graph. In Proceedings of the AAAI
Conference on Artificial Intelligence, Vol. 34.
2901–2908.
Liu et al. (2019)
Yinhan Liu, Myle Ott,
Naman Goyal, Jingfei Du,
Mandar Joshi, Danqi Chen,
Omer Levy, Mike Lewis,
Luke Zettlemoyer, and Veselin
Stoyanov. 2019.
Roberta: A robustly optimized bert pretraining
approach.
arXiv preprint arXiv:1907.11692
(2019).
Loshchilov and Hutter (2018)
Ilya Loshchilov and
Frank Hutter. 2018.
Decoupled Weight Decay Regularization. In
International Conference on Learning
Representations.
Lu et al. (2019)
Jiasen Lu, Dhruv Batra,
Devi Parikh, and Stefan Lee.
2019.
Vilbert: Pretraining task-agnostic visiolinguistic
representations for vision-and-language tasks.
Advances in neural information processing
systems 32 (2019).
Marino et al. (2017)
Kenneth Marino, Ruslan
Salakhutdinov, and Abhinav Gupta.
2017.
The More You Know: Using Knowledge Graphs for Image
Classification. In 2017 IEEE Conference on
Computer Vision and Pattern Recognition (CVPR). IEEE,
20–28.
Moon et al. (2021)
Jong Hak Moon, Hyungyung
Lee, Woncheol Shin, and Edward Choi.
2021.
Multi-modal Understanding and Generation for
Medical Images and Text via Vision-Language Pre-Training.
arXiv preprint arXiv:2105.11333
(2021).
Neumann et al. (2019)
Mark Neumann, Daniel
King, Iz Beltagy, and Waleed Ammar.
2019.
ScispaCy: Fast and Robust Models for Biomedical
Natural Language Processing. In Proceedings of the
18th BioNLP Workshop and Shared Task. 319–327.
Nguyen et al. (2019)
Binh D Nguyen, Thanh-Toan
Do, Binh X Nguyen, Tuong Do,
Erman Tjiputra, and Quang D Tran.
2019.
Overcoming data limitation in medical visual
question answering. In International Conference on
Medical Image Computing and Computer-Assisted Intervention. Springer,
522–530.
Pelka et al. (2018)
Obioma Pelka, Sven
Koitka, Johannes Rückert, Felix
Nensa, and Christoph M Friedrich.
2018.
Radiology Objects in COntext (ROCO): a multimodal
image dataset.
In Intravascular Imaging and Computer
Assisted Stenting and Large-Scale Annotation of Biomedical Data and Expert
Label Synthesis. Springer, 180–189.
Peters et al. (2019)
Matthew E Peters, Mark
Neumann, Robert Logan, Roy Schwartz,
Vidur Joshi, Sameer Singh, and
Noah A Smith. 2019.
Knowledge Enhanced Contextual Word
Representations. In Proceedings of the 2019
Conference on Empirical Methods in Natural Language Processing and the 9th
International Joint Conference on Natural Language Processing
(EMNLP-IJCNLP). 43–54.
Radford et al. (2021)
Alec Radford, Jong Wook
Kim, Chris Hallacy, Aditya Ramesh,
Gabriel Goh, Sandhini Agarwal,
Girish Sastry, Amanda Askell,
Pamela Mishkin, Jack Clark,
et al. 2021.
Learning transferable visual models from natural
language supervision. In International Conference
on Machine Learning. PMLR, 8748–8763.
Su et al. (2019)
Weijie Su, Xizhou Zhu,
Yue Cao, Bin Li, Lewei
Lu, Furu Wei, and Jifeng Dai.
2019.
VL-BERT: Pre-training of Generic Visual-Linguistic
Representations. In International Conference on
Learning Representations.
Subramanian et al. (2020)
Sanjay Subramanian,
Lucy Lu Wang, Ben Bogin,
Sachin Mehta, Madeleine van Zuylen,
Sravanthi Parasa, Sameer Singh,
Matt Gardner, and Hannaneh Hajishirzi.
2020.
MedICaT: A Dataset of Medical Images, Captions, and
Textual References. In Findings of the Association
for Computational Linguistics: EMNLP 2020. 2112–2120.
Sun et al. (2020)
Tianxiang Sun, Yunfan
Shao, Xipeng Qiu, Qipeng Guo,
Yaru Hu, Xuan-Jing Huang, and
Zheng Zhang. 2020.
CoLAKE: Contextualized Language and Knowledge
Embedding. In Proceedings of the 28th
International Conference on Computational Linguistics.
3660–3670.
Sun et al. (2019)
Yu Sun, Shuohuan Wang,
Yukun Li, Shikun Feng,
Xuyi Chen, Han Zhang,
Xin Tian, Danxiang Zhu,
Hao Tian, and Hua Wu.
2019.
Ernie: Enhanced representation through knowledge
integration.
arXiv preprint arXiv:1904.09223
(2019).
Tan and Bansal (2019)
Hao Tan and Mohit
Bansal. 2019.
LXMERT: Learning Cross-Modality Encoder
Representations from Transformers. In Proceedings
of the 2019 Conference on Empirical Methods in Natural Language Processing
and the 9th International Joint Conference on Natural Language Processing
(EMNLP-IJCNLP). 5100–5111.
Veličković et al. (2018)
Petar Veličković,
Guillem Cucurull, Arantxa Casanova,
Adriana Romero, Pietro Liò, and
Yoshua Bengio. 2018.
Graph Attention Networks. In
International Conference on Learning
Representations.
Wang et al. (2021a)
Xiaozhi Wang, Tianyu Gao,
Zhaocheng Zhu, Zhengyan Zhang,
Zhiyuan Liu, Juanzi Li, and
Jian Tang. 2021a.
KEPLER: A unified model for knowledge embedding and
pre-trained language representation.
Transactions of the Association for
Computational Linguistics 9 (2021),
176–194.
Wang et al. (2018)
Xiaolong Wang, Yufei Ye,
and Abhinav Gupta. 2018.
Zero-shot recognition via semantic embeddings and
knowledge graphs. In Proceedings of the IEEE
conference on computer vision and pattern recognition.
6857–6866.
Wang et al. (2021b)
Zirui Wang, Jiahui Yu,
Adams Wei Yu, Zihang Dai,
Yulia Tsvetkov, and Yuan Cao.
2021b.
Simvlm: Simple visual language model pretraining
with weak supervision.
arXiv preprint arXiv:2108.10904
(2021).
Wu et al. (2021)
Te-Lin Wu, Shikhar Singh,
Sayan Paul, Gully Burns, and
Nanyun Peng. 2021.
MELINDA: A Multimodal Dataset for Biomedical
Experiment Method Classification. In Proceedings
of the AAAI Conference on Artificial Intelligence,
Vol. 35. 14076–14084.
Wu et al. (2016)
Yonghui Wu, Mike
Schuster, Zhifeng Chen, Quoc V Le,
Mohammad Norouzi, Wolfgang Macherey,
Maxim Krikun, Yuan Cao,
Qin Gao, Klaus Macherey, et al.
2016.
Google’s neural machine translation system:
Bridging the gap between human and machine translation.
arXiv preprint arXiv:1609.08144
(2016).
Yang et al. (2016)
Zichao Yang, Xiaodong He,
Jianfeng Gao, Li Deng, and
Alex Smola. 2016.
Stacked attention networks for image question
answering. In Proceedings of the IEEE conference
on computer vision and pattern recognition. 21–29.
Yu et al. (2021)
Fei Yu, Jiji Tang,
Weichong Yin, Yu Sun,
Hao Tian, Hua Wu, and
Haifeng Wang. 2021.
ERNIE-ViL: Knowledge Enhanced Vision-Language
Representations through Scene Graphs. In
Proceedings of the AAAI Conference on Artificial
Intelligence, Vol. 35. 3208–3216.
Yu et al. (2017)
Zhou Yu, Jun Yu,
Jianping Fan, and Dacheng Tao.
2017.
Multi-modal factorized bilinear pooling with
co-attention learning for visual question answering. In
Proceedings of the IEEE international conference on
computer vision. 1821–1830.
Zhan et al. (2020)
Li-Ming Zhan, Bo Liu,
Lu Fan, Jiaxin Chen, and
Xiao-Ming Wu. 2020.
Medical visual question answering via conditional
reasoning. In Proceedings of the 28th ACM
International Conference on Multimedia. 2345–2354.
Zhang et al. (2021)
Pengchuan Zhang, Xiujun
Li, Xiaowei Hu, Jianwei Yang,
Lei Zhang, Lijuan Wang,
Yejin Choi, and Jianfeng Gao.
2021.
Vinvl: Revisiting visual representations in
vision-language models. In Proceedings of the
IEEE/CVF Conference on Computer Vision and Pattern Recognition.
5579–5588.
Zhang et al. (2019)
Zhengyan Zhang, Xu Han,
Zhiyuan Liu, Xin Jiang,
Maosong Sun, and Qun Liu.
2019.
ERNIE: Enhanced Language Representation with
Informative Entities. In Proceedings of the 57th
Annual Meeting of the Association for Computational Linguistics.
1441–1451. |
Abstract
Low $x$ deep-inelastic $ep$ scattering data, taken in 1994 at the H1
detector at HERA, are analysed in the Breit frame of reference.
The evolution of the peak and width of the current hemisphere fragmentation
function is presented as a function of $Q$ and compared
with $e^{+}e^{-}$ results at equivalent centre of mass energies.
Differences between the average charged multiplicity and the multiplicity
of $e^{+}e^{-}$ annihilations at low energies are analysed.
Invariant energy spectra are compared with MLLA predictions.
Distributions of multiplicity are presented as functions of Bjorken-$x$
and $Q^{2}$, and KNO scaling is discussed.
DESY 97-108 ISSN 0418-9833
June 1997
Evolution of $ep$ Fragmentation and Multiplicity
Distributions in the Breit Frame
H1 Collaboration
C. Adloff${}^{35}$, S. Aid${}^{13}$, M. Anderson${}^{23}$, V. Andreev${}^{26}$, B. Andrieu${}^{29}$, V. Arkadov${}^{36}$, C. Arndt${}^{11}$, I. Ayyaz${}^{30}$, A. Babaev${}^{25}$, J. Bähr${}^{36}$, J. Bán${}^{18}$, Y. Ban${}^{28}$, P. Baranov${}^{26}$, E. Barrelet${}^{30}$, R. Barschke${}^{11}$, W. Bartel${}^{11}$, U. Bassler${}^{30}$, H.P. Beck${}^{38}$, M. Beck${}^{14}$, H.-J. Behrend${}^{11}$, A. Belousov${}^{26}$, Ch. Berger${}^{1}$, G. Bernardi${}^{30}$, G. Bertrand-Coremans${}^{4}$, R. Beyer${}^{11}$, P. Biddulph${}^{23}$, P. Bispham${}^{23}$, J.C. Bizot${}^{28}$, K. Borras${}^{8}$, F. Botterweck${}^{27}$, V. Boudry${}^{29}$, S. Bourov${}^{25}$, A. Braemer${}^{15}$, W. Braunschweig${}^{1}$, V. Brisson${}^{28}$, W. Brückner${}^{14}$, P. Bruel${}^{29}$, D. Bruncko${}^{18}$, C. Brune${}^{16}$, R. Buchholz${}^{11}$, L. Büngener${}^{13}$, J. Bürger${}^{11}$, F.W. Büsser${}^{13}$, A. Buniatian${}^{4}$, S. Burke${}^{19}$, M.J. Burton${}^{23}$, G. Buschhorn${}^{27}$, D. Calvet${}^{24}$, A.J. Campbell${}^{11}$, T. Carli${}^{27}$, M. Charlet${}^{11}$, D. Clarke${}^{5}$, B. Clerbaux${}^{4}$, S. Cocks${}^{20}$, J.G. Contreras${}^{8}$, C. Cormack${}^{20}$, J.A. Coughlan${}^{5}$, A. Courau${}^{28}$, M.-C. Cousinou${}^{24}$, B.E. Cox${}^{23}$, G. Cozzika${}^{9}$, D.G. Cussans${}^{5}$, J. Cvach${}^{31}$, S. Dagoret${}^{30}$, J.B. Dainton${}^{20}$, W.D. Dau${}^{17}$, K. Daum${}^{40}$, M. David${}^{9}$, C.L. Davis${}^{19,41}$, A. De Roeck${}^{11}$, E.A. De Wolf${}^{4}$, B. Delcourt${}^{28}$, M. Dirkmann${}^{8}$, P. Dixon${}^{19}$, W. Dlugosz${}^{7}$, C. Dollfus${}^{38}$, K.T. Donovan${}^{21}$, J.D. Dowell${}^{3}$, H.B. Dreis${}^{2}$, A. Droutskoi${}^{25}$, J. Ebert${}^{35}$, T.R. Ebert${}^{20}$, G. Eckerlin${}^{11}$, V. Efremenko${}^{25}$, S. Egli${}^{38}$, R. Eichler${}^{37}$, F. Eisele${}^{15}$, E. Eisenhandler${}^{21}$, E. Elsen${}^{11}$, M. Erdmann${}^{15}$, A.B. Fahr${}^{13}$, L. Favart${}^{28}$, A. Fedotov${}^{25}$, R. Felst${}^{11}$, J. Feltesse${}^{9}$, J. Ferencei${}^{18}$, F. Ferrarotto${}^{33}$, K. Flamm${}^{11}$, M. Fleischer${}^{8}$, M. Flieser${}^{27}$, G. Flügge${}^{2}$, A. Fomenko${}^{26}$, J. Formánek${}^{32}$, J.M. Foster${}^{23}$, G. Franke${}^{11}$, E. Gabathuler${}^{20}$, K. Gabathuler${}^{34}$, F. Gaede${}^{27}$, J. Garvey${}^{3}$, J. Gayler${}^{11}$, M. Gebauer${}^{36}$, R. Gerhards${}^{11}$, A. Glazov${}^{36}$, L. Goerlich${}^{6}$, N. Gogitidze${}^{26}$, M. Goldberg${}^{30}$, D. Goldner${}^{8}$, K. Golec-Biernat${}^{6}$, B. Gonzalez-Pineiro${}^{30}$, I. Gorelov${}^{25}$, C. Grab${}^{37}$, H. Grässler${}^{2}$, T. Greenshaw${}^{20}$, R.K. Griffiths${}^{21}$, G. Grindhammer${}^{27}$, A. Gruber${}^{27}$, C. Gruber${}^{17}$, T. Hadig${}^{1}$, D. Haidt${}^{11}$, L. Hajduk${}^{6}$, T. Haller${}^{14}$, M. Hampel${}^{1}$, W.J. Haynes${}^{5}$, B. Heinemann${}^{11}$, G. Heinzelmann${}^{13}$, R.C.W. Henderson${}^{19}$, H. Henschel${}^{36}$, I. Herynek${}^{31}$, M.F. Hess${}^{27}$, K. Hewitt${}^{3}$, K.H. Hiller${}^{36}$, C.D. Hilton${}^{23}$, J. Hladký${}^{31}$, M. Höppner${}^{8}$, D. Hoffmann${}^{11}$, T. Holtom${}^{20}$, R. Horisberger${}^{34}$, V.L. Hudgson${}^{3}$, M. Hütte${}^{8}$, M. Ibbotson${}^{23}$, Ç. İşsever${}^{8}$, H. Itterbeck${}^{1}$, A. Jacholkowska${}^{28}$, C. Jacobsson${}^{22}$, M. Jacquet${}^{28}$, M. Jaffre${}^{28}$, J. Janoth${}^{16}$, D.M. Jansen${}^{14}$, L. Jönsson${}^{22}$, D.P. Johnson${}^{4}$, H. Jung${}^{22}$, P.I.P. Kalmus${}^{21}$, M. Kander${}^{11}$, D. Kant${}^{21}$, U. Kathage${}^{17}$, J. Katzy${}^{15}$, H.H. Kaufmann${}^{36}$, O. Kaufmann${}^{15}$, M. Kausch${}^{11}$, S. Kazarian${}^{11}$, I.R. Kenyon${}^{3}$, S. Kermiche${}^{24}$, C. Keuker${}^{1}$, C. Kiesling${}^{27}$, M. Klein${}^{36}$, C. Kleinwort${}^{11}$, G. Knies${}^{11}$, T. Köhler${}^{1}$, J.H. Köhne${}^{27}$, H. Kolanoski${}^{39}$, S.D. Kolya${}^{23}$, V. Korbel${}^{11}$, P. Kostka${}^{36}$, S.K. Kotelnikov${}^{26}$, T. Krämerkämper${}^{8}$, M.W. Krasny${}^{6,30}$, H. Krehbiel${}^{11}$, D. Krücker${}^{27}$, A. Küpper${}^{35}$, H. Küster${}^{22}$, M. Kuhlen${}^{27}$, T. Kurča${}^{36}$, J. Kurzhöfer${}^{8}$, B. Laforge${}^{9}$, M.P.J. Landon${}^{21}$, W. Lange${}^{36}$, U. Langenegger${}^{37}$, A. Lebedev${}^{26}$, F. Lehner${}^{11}$, V. Lemaitre${}^{11}$, S. Levonian${}^{29}$, M. Lindstroem${}^{22}$, F. Linsel${}^{11}$, J. Lipinski${}^{11}$, B. List${}^{11}$, G. Lobo${}^{28}$, J.W. Lomas${}^{23}$, G.C. Lopez${}^{12}$, V. Lubimov${}^{25}$, D. Lüke${}^{8,11}$, L. Lytkin${}^{14}$, N. Magnussen${}^{35}$, H. Mahlke-Krüger${}^{11}$, E. Malinovski${}^{26}$, R. Maraček${}^{18}$, P. Marage${}^{4}$, J. Marks${}^{15}$, R. Marshall${}^{23}$, J. Martens${}^{35}$, G. Martin${}^{13}$, R. Martin${}^{20}$, H.-U. Martyn${}^{1}$, J. Martyniak${}^{6}$, T. Mavroidis${}^{21}$, S.J. Maxfield${}^{20}$, S.J. McMahon${}^{20}$, A. Mehta${}^{5}$, K. Meier${}^{16}$, P. Merkel${}^{11}$, F. Metlica${}^{14}$, A. Meyer${}^{11}$, A. Meyer${}^{13}$, H. Meyer${}^{35}$, J. Meyer${}^{11}$, P.-O. Meyer${}^{2}$, A. Migliori${}^{29}$, S. Mikocki${}^{6}$, D. Milstead${}^{20}$, J. Moeck${}^{27}$, F. Moreau${}^{29}$, J.V. Morris${}^{5}$, E. Mroczko${}^{6}$, D. Müller${}^{38}$, T. Walter${}^{38}$, K. Müller${}^{11}$, P. Murín${}^{18}$, V. Nagovizin${}^{25}$, R. Nahnhauer${}^{36}$, B. Naroska${}^{13}$, Th. Naumann${}^{36}$, I. Négri${}^{24}$, P.R. Newman${}^{3}$, D. Newton${}^{19}$, H.K. Nguyen${}^{30}$, T.C. Nicholls${}^{3}$, F. Niebergall${}^{13}$, C. Niebuhr${}^{11}$, Ch. Niedzballa${}^{1}$, H. Niggli${}^{37}$, G. Nowak${}^{6}$, T. Nunnemann${}^{14}$, M. Nyberg-Werther${}^{22}$, H. Oberlack${}^{27}$, J.E. Olsson${}^{11}$, D. Ozerov${}^{25}$, P. Palmen${}^{2}$, E. Panaro${}^{11}$, A. Panitch${}^{4}$, C. Pascaud${}^{28}$, S. Passaggio${}^{37}$, G.D. Patel${}^{20}$, H. Pawletta${}^{2}$, E. Peppel${}^{36}$, E. Perez${}^{9}$, J.P. Phillips${}^{20}$, A. Pieuchot${}^{24}$, D. Pitzl${}^{37}$, R. Pöschl${}^{8}$, G. Pope${}^{7}$, B. Povh${}^{14}$, S. Prell${}^{11}$, K. Rabbertz${}^{1}$, P. Reimer${}^{31}$, H. Rick${}^{8}$, S. Riess${}^{13}$, E. Rizvi${}^{21}$, P. Robmann${}^{38}$, R. Roosen${}^{4}$, K. Rosenbauer${}^{1}$, A. Rostovtsev${}^{30}$, F. Rouse${}^{7}$, C. Royon${}^{9}$, K. Rüter${}^{27}$, S. Rusakov${}^{26}$, K. Rybicki${}^{6}$, D.P.C. Sankey${}^{5}$, P. Schacht${}^{27}$, S. Schiek${}^{13}$, S. Schleif${}^{16}$, P. Schleper${}^{15}$, W. von Schlippe${}^{21}$, D. Schmidt${}^{35}$, G. Schmidt${}^{13}$, L. Schoeffel${}^{9}$, A. Schöning${}^{11}$, V. Schröder${}^{11}$, E. Schuhmann${}^{27}$, B. Schwab${}^{15}$, F. Sefkow${}^{38}$, A. Semenov${}^{25}$, V. Shekelyan${}^{11}$, I. Sheviakov${}^{26}$, L.N. Shtarkov${}^{26}$, G. Siegmon${}^{17}$, U. Siewert${}^{17}$, Y. Sirois${}^{29}$, I.O. Skillicorn${}^{10}$, T. Sloan${}^{19}$, P. Smirnov${}^{26}$, M. Smith${}^{20}$, V. Solochenko${}^{25}$, Y. Soloviev${}^{26}$, A. Specka${}^{29}$, J. Spiekermann${}^{8}$, S. Spielman${}^{29}$, H. Spitzer${}^{13}$, F. Squinabol${}^{28}$, P. Steffen${}^{11}$, R. Steinberg${}^{2}$, J. Steinhart${}^{13}$, B. Stella${}^{33}$, A. Stellberger${}^{16}$, J. Stier${}^{11}$, J. Stiewe${}^{16}$, U. Stößlein${}^{36}$, K. Stolze${}^{36}$, U. Straumann${}^{15}$, W. Struczinski${}^{2}$, J.P. Sutton${}^{3}$, S. Tapprogge${}^{16}$, M. Taševský${}^{32}$, V. Tchernyshov${}^{25}$, S. Tchetchelnitski${}^{25}$, J. Theissen${}^{2}$, G. Thompson${}^{21}$, P.D. Thompson${}^{3}$, N. Tobien${}^{11}$, R. Todenhagen${}^{14}$, P. Truöl${}^{38}$, G. Tsipolitis${}^{37}$, J. Turnau${}^{6}$, E. Tzamariudaki${}^{11}$, P. Uelkes${}^{2}$, A. Usik${}^{26}$, S. Valkár${}^{32}$, A. Valkárová${}^{32}$, C. Vallée${}^{24}$, P. Van Esch${}^{4}$, P. Van Mechelen${}^{4}$, D. Vandenplas${}^{29}$, Y. Vazdik${}^{26}$, P. Verrecchia${}^{9}$, G. Villet${}^{9}$, K. Wacker${}^{8}$, A. Wagener${}^{2}$, M. Wagener${}^{34}$, R. Wallny${}^{15}$, B. Waugh${}^{23}$, G. Weber${}^{13}$, M. Weber${}^{16}$, D. Wegener${}^{8}$, A. Wegner${}^{27}$, T. Wengler${}^{15}$, M. Werner${}^{15}$, L.R. West${}^{3}$, S. Wiesand${}^{35}$, T. Wilksen${}^{11}$, S. Willard${}^{7}$, M. Winde${}^{36}$, G.-G. Winter${}^{11}$, C. Wittek${}^{13}$, M. Wobisch${}^{2}$, H. Wollatz${}^{11}$, E. Wünsch${}^{11}$, J. Žáček${}^{32}$, D. Zarbock${}^{12}$, Z. Zhang${}^{28}$, A. Zhokin${}^{25}$, P. Zini${}^{30}$, F. Zomer${}^{28}$, J. Zsembery${}^{9}$, and
M. zurNedden${}^{38}$, ${}^{1}$ I. Physikalisches Institut der RWTH, Aachen, Germany${}^{a}$
${}^{2}$ III. Physikalisches Institut der RWTH, Aachen, Germany${}^{a}$
${}^{3}$ School of Physics and Space Research, University of Birmingham,
Birmingham, UK${}^{b}$
${}^{4}$ Inter-University Institute for High Energies ULB-VUB, Brussels;
Universitaire Instelling Antwerpen, Wilrijk; Belgium${}^{c}$
${}^{5}$ Rutherford Appleton Laboratory, Chilton, Didcot, UK${}^{b}$
${}^{6}$ Institute for Nuclear Physics, Cracow, Poland${}^{d}$
${}^{7}$ Physics Department and IIRPA,
University of California, Davis, California, USA${}^{e}$
${}^{8}$ Institut für Physik, Universität Dortmund, Dortmund,
Germany${}^{a}$
${}^{9}$ CEA, DSM/DAPNIA, CE-Saclay, Gif-sur-Yvette, France
${}^{10}$ Department of Physics and Astronomy, University of Glasgow,
Glasgow, UK${}^{b}$
${}^{11}$ DESY, Hamburg, Germany${}^{a}$
${}^{12}$ I. Institut für Experimentalphysik, Universität Hamburg,
Hamburg, Germany${}^{a}$
${}^{13}$ II. Institut für Experimentalphysik, Universität Hamburg,
Hamburg, Germany${}^{a}$
${}^{14}$ Max-Planck-Institut für Kernphysik,
Heidelberg, Germany${}^{a}$
${}^{15}$ Physikalisches Institut, Universität Heidelberg,
Heidelberg, Germany${}^{a}$
${}^{16}$ Institut für Hochenergiephysik, Universität Heidelberg,
Heidelberg, Germany${}^{a}$
${}^{17}$ Institut für Reine und Angewandte Kernphysik, Universität
Kiel, Kiel, Germany${}^{a}$
${}^{18}$ Institute of Experimental Physics, Slovak Academy of
Sciences, Košice, Slovak Republic${}^{f,j}$
${}^{19}$ School of Physics and Chemistry, University of Lancaster,
Lancaster, UK${}^{b}$
${}^{20}$ Department of Physics, University of Liverpool,
Liverpool, UK${}^{b}$
${}^{21}$ Queen Mary and Westfield College, London, UK${}^{b}$
${}^{22}$ Physics Department, University of Lund,
Lund, Sweden${}^{g}$
${}^{23}$ Physics Department, University of Manchester,
Manchester, UK${}^{b}$
${}^{24}$ CPPM, Université d’Aix-Marseille II,
IN2P3-CNRS, Marseille, France
${}^{25}$ Institute for Theoretical and Experimental Physics,
Moscow, Russia
${}^{26}$ Lebedev Physical Institute, Moscow, Russia${}^{f}$
${}^{27}$ Max-Planck-Institut für Physik,
München, Germany${}^{a}$
${}^{28}$ LAL, Université de Paris-Sud, IN2P3-CNRS,
Orsay, France
${}^{29}$ LPNHE, Ecole Polytechnique, IN2P3-CNRS,
Palaiseau, France
${}^{30}$ LPNHE, Universités Paris VI and VII, IN2P3-CNRS,
Paris, France
${}^{31}$ Institute of Physics, Czech Academy of
Sciences, Praha, Czech Republic${}^{f,h}$
${}^{32}$ Nuclear Center, Charles University,
Praha, Czech Republic${}^{f,h}$
${}^{33}$ INFN Roma 1 and Dipartimento di Fisica,
Università Roma 3, Roma, Italy
${}^{34}$ Paul Scherrer Institut, Villigen, Switzerland
${}^{35}$ Fachbereich Physik, Bergische Universität Gesamthochschule
Wuppertal, Wuppertal, Germany${}^{a}$
${}^{36}$ DESY, Institut für Hochenergiephysik,
Zeuthen, Germany${}^{a}$
${}^{37}$ Institut für Teilchenphysik,
ETH, Zürich, Switzerland${}^{i}$
${}^{38}$ Physik-Institut der Universität Zürich,
Zürich, Switzerland${}^{i}$
${}^{39}$ Institut für Physik, Humboldt-Universität,
Berlin, Germany${}^{a}$
${}^{40}$ Rechenzentrum, Bergische Universität Gesamthochschule
Wuppertal, Wuppertal, Germany${}^{a}$
${}^{41}$ Visitor from Physics Dept. University Louisville, USA
${}^{a}$ Supported by the Bundesministerium für Bildung, Wissenschaft,
Forschung und Technologie, FRG,
under contract numbers 6AC17P, 6AC47P, 6DO57I, 6HH17P, 6HH27I,
6HD17I, 6HD27I, 6KI17P, 6MP17I, and 6WT87P
${}^{b}$ Supported by the UK Particle Physics and Astronomy Research
Council, and formerly by the UK Science and Engineering Research
Council
${}^{c}$ Supported by FNRS-NFWO, IISN-IIKW
${}^{d}$ Partially supported by the Polish State Committee for Scientific
Research, grant no. 115/E-343/SPUB/P03/120/96
${}^{e}$ Supported in part by USDOE grant DE F603 91ER40674
${}^{f}$ Supported by the Deutsche Forschungsgemeinschaft
${}^{g}$ Supported by the Swedish Natural Science Research Council
${}^{h}$ Supported by GA ČR grant no. 202/96/0214,
GA AV ČR grant no. A1010619 and GA UK grant no. 177
${}^{i}$ Supported by the Swiss National Science Foundation
${}^{j}$ Supported by VEGA SR grant no. 2/1325/96
1 Introduction
In a previous publication [1] the H1 experiment analysed hadron
production in $ep$ deep–inelastic scattering (DIS) in the
Breit frame [2, 3] of reference. We showed similarities between
hadronic distributions in the ‘current’ hemisphere of this frame of reference
and those in a given single hemisphere of an $e^{+}e^{-}$ annihilation interaction.
A similar analysis [4] was performed by the ZEUS experiment. The
first H1 study analysed 1993 data resulting from $e^{-}p$ collisions. The 1994
run used incident positrons and this analysis has more than six times as many
events despite tighter cuts on event quality. Since the $Q^{2}$ values
investigated are still in general below the kinematic region where anything
other than virtual photon exchange is relevant, the results of the two
analyses are directly comparable.
It is the object of this analysis to study the evolution of the
fragmentation function and charged hadronic multiplicity distributions.
One aim is to continue the comparison with equivalent $e^{+}e^{-}$ results as a test
of the universality of properties of the partons found within a proton and
those of the quarks produced in pairs from the vacuum. Another aim is to show
sensitivity to the running of the QCD coupling constant and to leading order
(LO) QCD processes from consideration of such hadronic final states.
After a description of the relevant parts of the apparatus (section 2)
and a discussion of the treatment of the data (section 3), this paper describes
the Breit frame of reference (section 4) and extends (section 5) the earlier
analysis of the evolution of the fragmentation function with $Q^{2}$, where
$q^{2}=-Q^{2}$ is the square of the four-momentum transferred from the incident
lepton. The average charged multiplicity dependence on $Q$ (section 6) is
used to study the effects of LO QCD processes. For the first time we present
invariant energy spectra (section 7) from HERA and H1 multiplicity
distributions (section 8) in the Breit frame.
2 The H1 Detector
The H1 detector is described in detail elsewhere [5]. Here,
we give only a short description of the components more relevant to this
analysis. We employ a right-handed coordinate system with the
positive (forward) $z$ axis being in the direction of the incoming proton
beam. Polar angles are measured from this direction.
Momentum measurements of charged particles are provided, in the central
region of the apparatus, by two cylindrical and co-axial drift
chambers [6] for ($r,\phi$) measurement which have significantly fewer
dead cells compared with the situation in 1993. These detectors are
supplemented by two $z$-chambers. In the forward (proton) direction the
equivalent measurements are provided by three Radial and three Planar drift
chamber modules, the Forward Track Detector [7]. All these track
detectors are inside a uniform 1.15 T magnetic field. Track segments from all
devices are combined to give efficient detection and momentum measurement with
$\delta p/p^{2}$ $\stackrel{{\scriptstyle\textstyle<}}{{\sim}}$1% /GeV for
most of the angular range used in this analysis, $10^{\circ}<\theta<160^{\circ}$.
In the polar angle range $4^{\circ}<\theta<153^{\circ}$ the trackers
are surrounded by a fine-grained liquid argon (LAr) sampling
calorimeter [8] with lead and steel absorber in the electromagnetic
and hadronic sections respectively. The calorimeter cells measure hadronic
energy flow and the energy of the scattered electron for high $Q^{2}$ events.
The LAr calorimeter is complemented by a backward electromagnetic,
lead-scintillator, calorimeter (BEMC [9]) covering the angular range
$151^{\circ}<\theta<176^{\circ}$. Behind this there is a double–layer scintillator
hodoscope which gives efficient background rejection by time of flight
measurement. The data are derived from two separate samples corresponding to
the scattered positron being triggered by, and detected in, either the BEMC or
the LAr calorimeter. The transition is at a $Q^{2}$ of about 100 GeV${}^{2}$ and these
samples are referred to in this paper as, respectively, low or high $Q^{2}$ data.
The trigger is already fully efficient at positron energies of 10 GeV.
3 Data Selection and Corrections
The full data sample, triggered as above, consists of some $150\,000$
events taken at an $ep$ centre of mass energy, $\sqrt{s}$, of 300 GeV. Event
kinematic variables used in this analysis are calculated using only the
scattered lepton, which gives both the best resolution in the chosen region
and clear freedom from bias on the hadronic system studied. Throughout, the
events are selected to have an identified scattered positron [10] with
an energy $E_{e}^{\prime}>14$ GeV.
To reject beam associated background, we demand that there be no veto
from the time of flight system and that a vertex is found for the event
within 30 cm of the nominal vertex position. The low (high) $Q^{2}$ data
referred to in the above section are selected to be within the limits
$12<Q^{2}<100$ GeV${}^{2}$ ($100<Q^{2}<8000$ GeV${}^{2}$). The dimensionless inelasticity
variable, $y=Q^{2}/xs$, is required to be in the range $0.05<y<0.6$. These
conditions ensure that contamination from mis-identified photoproduction events
is below the 1% level. Together these cuts remove $\sim$ 45% of the original
sample. In addition, a cut is placed to safely exclude a further $\sim$ 5%
of diffractive events which are not well-modelled with the DIS Monte Carlo
programs used in this analysis and which have no analogue in $e^{+}e^{-}$
interactions. We require that a total cluster energy of at least 0.5 GeV
should be observed in the region $4.4^{\circ}<\theta<15^{\circ}$.
Events are then selected in which a massless quark would be scattered
through $10^{\circ}<\theta<150^{\circ}$ using four momentum conservation with the
scattered positron. This is inside the acceptance of the H1 track detectors
and thus minimises corrections, at the cost of removing a further $\sim$ 24%
of events.
There is a source of error arising from QED radiation which comes
about because of an incorrect, typically overestimated, value of $Q^{2}$. The
error in the boost leads to a miscalculation of the direction of the Breit
frame axis. In most cases, this leads to an apparently empty, or at least
depleted, region of phase space where the scattered quark fragments are
expected. With improved statistics this has become a significant
effect. To reduce the size of the necessary QED corrections, the new analysis
utilises more severe cuts on the mass of the total hadronic system,
$W$, of $W^{2}>4400\,$ GeV${}^{2}$ both using the scattered lepton variables,
and separately from the hadronic system variables with the Jaquet Blondel
method [11].
Following all of the selection procedures there are $20\,810$ events in
the low $Q^{2}$ sample and $1\,250$ events at high $Q^{2}$.
In addition to these event selections there are also cuts
made to reject badly measured tracks. Any tracks in the central chambers
with transverse momentum below 150 MeV and those in the forward track detectors
with momentum below 500 MeV are removed. We also remove tracks failing minimum
requirements on the number of hits in a given chamber and the quality of the
track fit. In order to have only the
primary multiplicity, cuts are also made to exclude tracks not originating from
the interaction vertex. There remains a small excess contribution due to the
decays of short-lived strange particles.
To correct for this excess and for acceptance losses, we utilise the
DJANGO6 [12] Monte Carlo event generator. This combines a LO perturbative
QCD matrix element calculation and the colour dipole model of hadronisation
with a calculation of QED radiative effects. Radiative effects remaining after
our event selections have been corrected by comparing the results of Monte
Carlo calculations with and without the inclusion of QED radiative effects.
These corrections are at the $\sim$10%
level. The detector response is simulated using a program based
on the GEANT [13] package and the simulated events are reconstructed
and selected using exactly the same analysis chain as with real data.
The total bin-by-bin corrections made throughout this analysis are generally
well within $\pm$20% and vary smoothly in any given distribution.
The largest individual sources of systematic error for this analysis
originate in possible calibration errors of $\pm$1% for the BEMC and $\pm$3%
for the LAr electromagnetic calorimeter energy scales. These directly affect
the accuracy of Lorentz boosts and give rise to an uncertainty in the number
of tracks of $\sim$5% at low $Q^{2}$ and $\sim$8% at high $Q^{2}$, irrespective
of other kinematic selection. The corresponding systematic error in the Monte
Carlo derived
acceptance correction functions has been estimated to be $\sim$ 2% using
several different generators [14, 15]. Visual scans of real and simulated
data have ascertained that the efficiency of the track detectors is modelled
to an accuracy of better than 2% by our Monte Carlo simulations.
The full experimental details of how the corrected distributions of
this analysis are obtained from the data may be found in reference [16].
4 The Breit Frame of Reference
The $ep$ Breit frame is aligned with the hadronic centre of mass (HCM) but
boosted along a common $z$ direction such that the incident virtual photon has
zero energy, zero transverse momentum and a $z$ component of momentum $-Q$.
As with the laboratory frame of reference, we choose the positive $z$ axis to
be in the direction of the incoming proton. The negative
$z$ direction is referred to as the ‘current’ hemisphere of the
interaction. Our earlier Breit frame analysis [1] showed that
multiplicities in the current region of the Breit frame depend on $Q$ and
not on Bjorken $x$ ($\sim Q^{2}/W^{2}$), as opposed to the HCM where multiplicities
depend on the natural scale, $W$ [17]. In the naïve quark parton
model (QPM) the massless incoming quark has energy $Q/2$ and $z$ component
of momentum $+Q/2$, carrying an approximate fraction $x$ of the proton’s
momentum. After scattering it still has energy $Q/2$, with momentum $-Q/2$.
By comparison, we thus take the equivalent to the $e^{+}e^{-}$ centre of mass
energy, $E^{*}$, to be $Q$.
As compared with the HCM the Breit frame current hemisphere is
dominated by the fragments of the struck quark alone; the ‘spectator’ proton
remnants go entirely into the ‘target’ hemisphere, with much higher momentum.
There is excellent acceptance for the current hemisphere in the central H1
detector. For example, even in the highest $Q^{2}$ interval of this analysis
there is only an $\sim$8% contribution of tracks from the Forward Track
Detector. The HCM current hemisphere, in contrast, has an energy scale of
$W/2$ rather than $Q/2$ and generally
can not be seen in its entirety. This makes comparisons with a
complete $e^{+}e^{-}$ interaction hemisphere somewhat easier in the Breit frame.
When QCD corrections to the parton model are considered, the incident
parton carries, in general, a proportion of the incident proton’s momentum
that is larger than $x$ and has an energy larger than $Q/2$.
Furthermore, since the four-momentum of the photon is fixed the energy seen in
the current hemisphere may be either greater than or less than $Q/2$. In an
$e^{+}e^{-}$ interaction it is also the case that the energy in any given hemisphere
may be greater than or less than $E^{*}/2$. The effects of final state radiation
in $ep$ DIS and $e^{+}e^{-}$ interactions are similar, but other LO QCD processes which
affect DIS, initial state QCD radiation and boson gluon fusion (BGF), have no
equivalent in $e^{+}e^{-}$ annihilation. Note that, to order $\alpha_{s}$, i.e.
two exiting partons as well as the spectator system, DIS events may have
no energy in the current hemisphere when the incident parton has energy
above $Q$, other than that due to hadronisation [3]. Momentum
conservation ensures that this situation never exists in any given hemisphere
of an $e^{+}e^{-}$ interaction.
5 Evolution of the Fragmentation Function
The $ep$ Breit frame equivalent of the $e^{+}e^{-}$ scaled hadron momentum
$x_{p}=2p_{hadron}/E^{*}$ is $x_{p}=2p_{hadron}/Q$, where only hadrons in the
current hemisphere are considered. The event normalised distribution
$D(x_{p},Q^{2})=(1/N_{evts})\times dn_{tracks}^{\pm}/dx_{p}$, the fragmentation
function, characterises the processes by which partons shower and then
hadronise. In this paper, the intention is to present the spectra of
charged particles originating from the primary vertex following the
fragmentation of light quarks.
At high enough energies and for light enough final state particles,
fragmentation functions approximately scale and are “soft”, rising rapidly
at small $x_{p}$ and peaking near $x_{p}=0$. As the energy of the
initial parton increases, $D(x_{p},Q^{2})$ evolves into an even softer function
with increased probability for low $x_{p}$ hadrons at the expense of high $x_{p}$.
This scaling violation can be seen with the low and high $Q^{2}$ data
as a function of $x_{p}$ in Fig. 1(a) or in Fig. 2 where the
fragmentation function is plotted as a function of $Q$ for
different intervals of $x_{p}$. Typically, in this analysis, the intervals
of $x_{p}$ are much greater than the resolution ($\sim$6% rms). In these figures
the H1 data are compared with equivalent data from $e^{+}e^{-}$
experiments [18] plotted as a function of $E^{*}$. Most published
$e^{+}e^{-}$ results refer to full event multiplicities. Here, as elsewhere in this
analysis, track multiplicity data from $e^{+}e^{-}$ experiments have been halved to
correspond to the fragmentation of one timelike quark. Note that H1 as a
single experiment is able to measure the violation for a wide range of $Q$.
Fig. 2 also displays a prediction from a leading order QCD plus
hadronisation Monte Carlo (DJANGO6 [12]) calculation. The scaling
violation effect has a similar origin to the scaling violations in structure
functions, and with next to leading order calculations may be used as a test
of perturbative QCD [19].
When the fragmentation function is plotted as a function of the variable
$\xi=\ln(1/x_{p})$ the turn-over region is expanded. According to the
expectations of the Modified Leading Logarithmic Approximation (MLLA) and
Local Parton Hadron Duality (LPHD), $D(\xi)$ is Gaussian in the
neighbourhood of the peak [20]. Even the high statistics of this,
compared with our earlier, analysis see no significant deviation from this
Gaussian behaviour, as is clear in Fig. 1(b). Most track measurement
problems occur with particles of very high or very low momenta and these are
correlated with low and high $\xi$ respectively. To avoid any dependence on
these areas the fits of this analysis are restricted to a region within
one unit of the dimensionless $\xi$ on either side of the mean, but tests in
which this range is varied by $\pm$20% show no systematic difference.
The evolution of the fragmentation function may be summarised
by the $Q$, or $E^{*}$, dependence of the peak and width (dispersion)
values of the fitted Gaussian as is shown in Fig. 3 and in
Table 1. It is notable that the systematic effects discussed in
section 3 contribute only a total 2% uncertainty to the peak value and have
no significant effect on the width measurement. The results are
compatible with those published earlier, which had considerably lower precision,
as well as with those of an analysis [4] by the ZEUS collaboration.
The results are also compatible with various $e^{+}e^{-}$
experiments [18, 21, 22], where the relevant evolution variable is
the centre of mass energy, $E^{*}$. For this comparison and for the consideration
of multiplicities in the next section, these published $e^{+}e^{-}$ data have
been re-fitted by us to be in a directly comparable form. As well as the
previously mentioned factor of two to convert $e^{+}e^{-}$ event multiplicities to
quark multiplicities, we also reduce published total multiplicity results by
8% to account for $K^{0}$ and $\Lambda$ decay tracks [23, 24], and by a
variable factor below 3% to correct for the
increased multiplicity from $b$ quark fragmentation [25] in $e^{+}e^{-}$ .
Assuming gluon coherence MLLA/LPHD predicts a dependence of the
evolution of the peak and width of the distribution on the dimensionless
variable $Y=\ln(Q/\Lambda_{\hbox{\it eff}})$, where $\Lambda_{\hbox{\it eff}}$ is a scale parameter determining
a presumed cut–off of a parton shower. The prediction [20] gives the
peak,
$\xi_{peak}$ and the width, $\xi_{width}$, of the Gaussian approximation to be
$$\xi_{peak}=0.5Y+c_{2}\sqrt{Y}+{\cal K}$$
$$\xi_{width}=\sqrt{Y^{3\over 2}/2c_{1}}{\rm,}$$
where $c_{1}=\sqrt{36N_{c}/b}$ and $c_{2}=B\sqrt{{{1}\over{16}}b/N_{c}}$ with
$b={{11}\over{3}}N_{c}-{{2}\over{3}}N_{f}$ and
$B=({{11}\over{3}}N_{c}+{{2}\over{3}}N_{f}/N_{c}^{2})/b$ are constants dependent only
on the number of contributing colours, $N_{c}$, and flavours, $N_{f}$ within the
parton shower. For ease of comparison, we follow [21, 22] in the
assumption that this will be dominated by the light quarks and set $N_{f}=3$.
The term ${\cal K}$ contains higher order corrections and is expected to be
roughly constant and of order 1. The result of a simultaneous fit
($\chi^{2}$/NDF of 10/20) to the peak and width values obtained from the present
H1 data alone is shown as a solid line in Fig. 3 and yields the
values $\Lambda_{\hbox{\it eff}}=0.21\pm 0.02$ GeV and ${\cal K}=-0.43\pm 0.06$, in agreement with
the values $\Lambda_{\hbox{\it eff}}=0.21\pm 0.02$ GeV and ${\cal K}=-0.32\pm 0.06$ obtained in an
analysis [21] of the $\xi_{peak}$ evolution of combined $e^{+}e^{-}$ data.
6 The Average Charged Particle Multiplicity
The area underneath the fragmentation function is the average
charged multiplicity and this is given, after all corrections, in
Table 1. It is shown as a function of $Q$ in Fig. 4, with ZEUS
results [4] and with a curve which accurately
parameterises [24] the
average charged multiplicity from many $e^{+}e^{-}$ experiments as a function of $E^{*}$.
The error associated with this fit at any given energy is at the percent level.
There are predictions [26] that both the absolute
average charged DIS multiplicity and the shape of the evolution with $Q^{2}$
should agree with that of $e^{+}e^{-}$ experiments. As previous results indicated,
although there
was very good agreement between H1 and ZEUS [4] analyses and those of
$e^{+}e^{-}$ at high $Q(E^{*})$, the DIS analyses give smaller average charged
multiplicities at low values of $Q(E^{*})$. The new data confirm that this is a
large effect.
The discussion of section 4 indicated that the LO processes
which are present in $ep$ but not in $e^{+}e^{-}$ interactions can produce a
depletion of the current region of the Breit frame. Pure
QPM interactions or events with only final state radiation (i.e. $e^{+}e^{-}$
like events) might both be expected to have a current region hadronic state
characteristic of a quark recoiling with energy $\sim Q/2$ and with only
a $z$ component of momentum. BGF events with a depleted or empty current
hemisphere
or events with significant initial state QCD radiation would, on the contrary,
have a current region hadronic final state with significantly less energy
at an angle to the $z$ axis. However, hadronisation effects might well
mask these naïve expectations.
In order to investigate this experimentally, we add the four–momenta of
all calorimeter energy clusters in the current region of the Breit frame. The
energy component of the resultant four–momentum vector is defined to be
$E_{z<0}$ and the angle it makes with the $z$ axis to be $\Theta_{BF}$. We
then plot
$E_{z<0}/Q$ against $\cos\Theta_{BF}$, as shown in Fig. 5(a,b)
for the low and high $Q^{2}$ samples. From the above discussion we would expect
that $e^{+}e^{-}$ like events might cluster near ($\cos\Theta_{BF}=-1,E_{z<0}/Q=1/2$) in this plot, whereas events from the other LO processes would
either not appear in this plot or be spread away from this point, typically
towards lower energies and larger angles in the current hemisphere, that is to
the lower right-hand corner of this plot.
The high $Q^{2}$ sample, where $\alpha_{s}$ is small, $x$ is large, and the proton
has a
much smaller gluon content, shows obvious clustering near the (-1,$0.5$) point,
while this is much less evident at low $Q^{2}$.
Monte Carlo simulations qualitatively confirm the expectations but it
is hard to be quantitative, both because of the difficulty in strict
definitions of LO QCD event classes and because experimental resolution has a
large contribution to event migration on this plot. Nevertheless,
it is possible [27] to investigate the effect of choosing ever
more $e^{+}e^{-}$ like events by selecting according to the variable $R$, where
$R^{2}=(E_{z<0}/Q-0.5)^{2}+(\cos\Theta_{BF}+1)^{2}$. In order to compare with
$e^{+}e^{-}$ results, we divide $\left<n(R)\right>$, the average multiplicity of
events in an annulus of width $\Delta R=0.1$ at radius $R$, by the equivalent
single–hemisphere $e^{+}e^{-}$ multiplicity. There is a very small variation of $\left<Q\right>$ with $R$, so this $e^{+}e^{-}$ multiplicity is evaluated at an energy
corresponding to the average value of $Q$ of the H1 data in that annulus. The
behaviour of this ratio as a function of $R$ can be seen, with statistical
error bars only, in Fig. 6(a,b). The systematic errors corresponding
to the shaded region are as discussed in section 3. Events with a totally
empty current hemisphere have zero multiplicity but no defined value of $R$,
and are thus not included in these plots. As $R\to 0$ the average charged
hadronic multiplicity increases, possibly even beyond the corresponding
$e^{+}e^{-}$ values, although it is not inconsistent with $e^{+}e^{-}$ values given the
systematic errors involved. Our Monte Carlo simulations also show this general
behaviour.
The above result and discussions motivate a cut to remove extreme low
energy and off–axis events. This is shown as a solid line in
Fig. 5(a,b) joining the points (-1,0) and (0,0.5) in this plane and
is termed the ‘Breit frame energy flow selection’ in the rest of this paper. It
produces samples of $11\,735$ events and $1\,005$ events at low and high $Q^{2}$
respectively, which are less affected by the low–order QCD processes peculiar
to DIS.
This selection is by no means as exact as the technique of letting
$R\to 0$ but in circumstances where the data are already binned in other
variables, e.g. energy spectra or multiplicity distributions, gives an
indication of whether or not distributions move in the direction of closer
similarity between this data and $e^{+}e^{-}$ data.
As is clear from Table 1, the peak and width values of the
fragmentation function are insensitive to this energy flow selection procedure.
The average multiplicity, however, is sensitive and as expected increases
with this selection, especially at low $Q^{2}$.
7 Invariant Energy Spectra
The evolution of the fragmentation function at low energies, or high values of
$\xi$, may be better studied using the event–normalised “invariant”
spectrum, $(1/N_{evts})\times Edn_{tracks}^{\pm}/d^{3}p$. According to the
predictions of perturbative QCD based on the MLLA and LPHD, the
hadronic spectrum at low momentum should be nearly independent of the energy of
the parent parton [28]. Here we follow [28] in taking
$d^{3}p=4\pi p^{2}_{hadron}dp_{hadron}$
and by computing the energy, $E$, of each track as if it had a mass
$Q_{0}(=$270 MeV$\approx\Lambda_{QCD})$ given by a presumed cut-off mass for the
parton shower, i.e. $E=\sqrt{Q_{0}^{2}+p_{hadron}^{2}}$. The data are corrected
for acceptance with the same bin-by-bin method described in section 3. The
correction function is both smooth and close to unity, having an average value
of around 1.15. Although there is always a loss of track acceptance at low
laboratory momentum the boost reduces this loss in the Breit frame. This means
that $ep$ experiments can get closer to the critical low energy limit without
serious loss of acceptance, than can $e^{+}e^{-}$ experiments.
The results are shown in Fig. 7(a,b) where the indicated errors contain
contributions from statistical and (negligible) momentum measurement effects
as well as systematic effects that have already been discussed and which grow
slightly with track energy. The solid lines show MLLA/LPHD predictions at
the relevant $Q$ values for a simple quark behaviour using the calculations
of [28] which the authors show are in agreement with $e^{+}e^{-}$ data. It is
clear that at low $Q^{2}$ these predictions for the behaviour of quark fragments
are in disagreement with the data but are significantly closer to data
subject to the Breit frame energy flow selection described in the previous
section. This selection makes little difference at high $Q^{2}$, where agreement
is better anyway. The QCD predictions are very sensitive to the ‘running’ of the
strong coupling constant, $\alpha_{s}$, as the energy scale alters in the parton
shower. A prediction is shown for the same average energy, normalised to
interpolated $\left<n^{\pm}\right>$ multiplicities in $e^{+}e^{-}$ experiments, but
utilising a fixed value of $\alpha_{s}$ (set to 0.215, a value which roughly
describes the energy dependence of $\left<n^{\pm}\right>$ and the slope of the
first moment of the energy spectrum in $e^{+}e^{-}$ experiments [29]). The
data show that such an assumption is clearly untenable. This calculation is
only in leading order so this cannot yet be turned
around to compute a meaningful value of $\Lambda_{QCD}$.
Data subject to the energy flow selection are re-plotted in six intervals of
$Q^{2}$ in Fig. 7(c) in order to display the clear evidence for a common
limit at low particle energy independent of event momentum transfer.
8 Charged Particle Multiplicity Distributions
Charged hadron multiplicity distributions are considerably more
difficult to measure than average values. In this analysis we have tested a
number of ways of extracting true distributions from observed distributions. The best method is
found to be a simple bin-by-bin correction procedure on the multiplicity
distribution. This has, however, the problem of not modelling correlations
which then gives the method a dependence on the Monte Carlo generator.
We tried two further methods to solve this difficulty.
It is straightforward to use a Monte Carlo simulation to form a generated,
or ‘true’, to reconstructed distribution transfer matrix, which is
physics-generator independent. Inversion of this matrix in order to extract a
true
distribution from a raw input distribution has many problems due to the small
number of high multiplicity events, resulting in the matrix becoming singular
and unstable. We have also tried a method [17, 30] of
approximating the inverted matrix with its inverse and then re-weighting the
input Monte Carlo distribution in successive approximation iterations to
minimise dependence on the initial generated distribution.
The systematic errors of all three techniques have been estimated by
degrading the output distributions back to the ‘raw’ (reconstructed) data level
using the reliable ‘generated to reconstructed’ matrix. This can then be
compared with the original, real input distribution. We find that the larger
dependence on the generated distribution of the bin-by-bin correction method
is still offset by overall smaller systematic errors and the ability to get
meaningful corrections at high multiplicity when the statistics are low. We
thus choose this method to present our results. Quantitatively, the average
correction made per bin is 16%(32%) at low (high) $Q^{2}$ with an error of
method assessed as 10%(17%).
The resulting multiplicity distributions are
given in Table 2 and shown in Fig. 8 in the form of
$P(n)$, the probability of observing a charged multiplicity $n$.
These results are for the multiplicity distribution in the current hemisphere
of all events and do not include any energy flow selection. The bins of
$x$ and $Q^{2}$ are exclusive but, because of the event selections, are not
totally occupied at the edges.
This leads to only minor discrepancies in the average values of $x$ and $Q^{2}$
between neighbouring bins. In general the observation is that the results are
compatible with a gradual shift to higher multiplicities with increasing $Q^{2}$
but with only a small change as a function of $x$ which might be due to minor
bin edge effects. It is interesting to note the many zero-charged multiplicity
events, especially at low $Q^{2}$ where we might expect higher-order QCD
processes to dominate.
Integrating over $x$ gives smaller statistical errors and a
greater reach in multiplicity. The evolution with $Q$ of the shape of the
multiplicity distribution may be studied by normalising with the
average charged hadron multiplicity to give the KNO [31] form
$\Psi(z)=\left<n\right>\times P(n)$, where $z=n/\left<n\right>$. Equivalent
single-hemisphere $e^{+}e^{-}$ distributions [24, 32, 33] exhibit clear
scaling in this variable but only for high ($E^{*}>20$ GeV) energies.
In an analysis [17] of multiplicity distributions in a limited region
of the hadronic centre of mass frame, the H1 collaboration has also observed
good agreement with a common distribution as $W$ increases between 80 and
220 GeV. The results of this
analysis in the Breit frame, shown in Fig. 9(a), are for $Q$ values
much lower than the $W$ values and show a clear violation of KNO scaling,
as also reported by the ZEUS collaboration [4]. For reference, we
display two sets of lines joining $e^{+}e^{-}$ data at an energy close to our largest
$Q$ range [32] and also at LEP energies [24]. As with $e^{+}e^{-}$
data, the H1 low $Q^{2}$ distribution starts with a flatter curve and a large
contribution from low multiplicities. Then, as $Q^{2}$ increases, the data get
closer to the high energy scaling $e^{+}e^{-}$ distributions.
In Fig. 9(b) the distribution for low $Q^{2}$ events is compared
with results from the $e^{+}e^{-}$ JETSET [34] parton shower model, using
parameter settings determined by the DELPHI collaboration [35] at an
energy set to be close to the mean value of the low $Q^{2}$ data. The
data
distribution is again much flatter but a better agreement is achieved when
using the Breit frame energy flow selection discussed in section 6.
This KNO scaling violation is also seen in $ep$ DIS Monte Carlo generated
events, where it is to a large extent caused by leading
order perturbative QCD events, especially BGF events, which, as has already
been noted, have significantly lower multiplicities in the current hemisphere
of the Breit frame. It thus appears
that $ep$ and $e^{+}e^{-}$ data share a flattening of the KNO distribution at low
energies but that there is an additional
sensitivity to the effects of higher order QCD processes in the Breit frame
which do not contribute to $e^{+}e^{-}$ events.
9 Summary
The Breit frame fragmentation function analysis
has been repeated with the higher statistics of 1994 data and is
now extended to give multiplicity distributions as functions of both
$x$ and $Q^{2}$. Invariant energy spectra are shown to be incompatible with a
calculation utilising a constant value for $\alpha_{s}$. We discuss a
method of discriminating against events in kinematic regions dominated by
QCD processes which do not contribute to $e^{+}e^{-}$ interactions.
We show that some discrepancies between $ep$ Breit frame current hemisphere
and $e^{+}e^{-}$ average charged multiplicities, including an observed violation of
KNO scaling at low $Q^{2}$, decrease as conditions are progressively made more
comparable. With a selection which gives similar kinematic conditions to that
of a single hemisphere $e^{+}e^{-}$ annihilation, the $ep$ energy spectra can be shown
to be compatible with a common, $Q^{2}$-independent, low momentum limit.
In summary, the data
give strong support to the concept of quark fragmentation universality.
Acknowledgments
We are grateful to the HERA machine group whose outstanding
efforts have made and continue to make this experiment possible. We thank
the engineers and technicians for their work in constructing and now
maintaining the H1 detector, our funding agencies for financial support, the
DESY technical staff for continual assistance, and the DESY directorate for the
hospitality which they extend to the non–DESY members of the collaboration.
We are grateful to S. Lupia for kind assistance with obtaining MLLA/LPHD
predictions.
References
[1]
H1 Collaboration, S. Aid et al., Nucl. Phys. 445B (1995) 3.
[2]
H.F. Jones, Nuovo Cimento 40A (1965) 1018.
[3]
K.H. Streng et al., Z. Phys. C2 (1979) 237.
[4]
ZEUS Collaboration, M. Derrick et al., Z. Phys. C67 (1995) 93.
[5]
H1 Collaboration, I. Abt et al., DESY preprint 93-103 (1993),
Nucl. Instr. Meth. A386 (1997) 310 (Vol 1) and ibid. p.348 (Vol 2).
[6]
J. Bürger et al., Nucl. Instr. Meth. A279 (1989) 217.
[7]
H1 FTD group, S. Burke et al., Nucl. Instr. Meth. A373 (1996) 227.
[8]
H1 calorimeter group, B. Andrieu et al., Nucl. Instr. Meth. A336 (1993) 460.
[9]
H1 BEMC group, J. Ban et al., Nucl. Instr. Meth. A372 (1996) 399.
[10]
H1 Collaboration, S. Aid et al., Nucl. Phys. 470B (1996) 3.
[11]
Proceedings of the study of an ep facility for Europe, Ed. U. Amaldi,
DESY 79/48 (1979) p391-4.
[12]
G.A. Schuler and H. Spiesberger, Proceedings of the Workshop: Physics at HERA,
vol. 3 eds. W. Buchmüller, G. Ingelman, DESY (1992) 1419.
We use version 6.1 with QED radiation simulated with HERACLES 4.4 when
correcting data. QCD radiation is simulated by the colour dipole model
(ARIADNE 4.03/4.05) and hadronisation with the LUND string model in JETSET 7.4.
We use MRSH parton density parametrizations at low $Q^{2}$ and GRV at high $Q^{2}$.
[13]
R. Brun et al., GEANT3 User’s guide, CERN-DD/EE 84-1, Geneva (1987).
[14]
LEPTO 6.5 Generator with GRV parton density functions ;
G. Ingelman, Proceedings of the Workshop: Physics at
HERA, vol. 3 eds. W. Buchmüller, G. Ingelman, DESY (1991) 1366.
[15]
HERWIG 5.8 Generator with MRSH parton density function, and no soft underlying
event ; G. Marchesini et al., Computer Phys. Comm. 67 (1992) 32.
[16]
P. Dixon, University of Lancaster thesis, in preparation.
[17]
H1 Collaboration, S. Aid et al., Z. Phys. C72 (1996) 573.
[18]
TASSO Collaboration, W. Braunschweig et al., Z. Phys. C47 (1990) 187,
Phys. Lett. 311B (1993) 408;
DELPHI Collaboration, P. Abreu et al., Phys. Lett. 311B (1993) 408;
OPAL Collaboration, G. Alexander et al., Z. Phys. C72 (1996) 191.
[19]
G.D. Cowan Glasgow XXVII HEP Conference 1994 proceedings p.883, IoP Publishing
Ltd. 1995; J. Binnewies,
B.A. Kniehl and G. Kramer Phys. Rev. 52D (1995) 4947 and Phys. Rev. 53D (1996) 3573.
[20]
Yu.L. Dokshitzer, V.A. Khoze, A.H. Mueller and S.I. Troyan, “Basics of
Perturbative QCD” Editions Frontières (1991); Ya.I. Azimov,
Yu.L. Dokshitzer, V.A. Khoze, and S.I. Troyan, Z. Phys. C31 (1986) 213,
Z. Phys. C27 (1985) 65; L.V. Gribov, Yu.L. Dokshitzer, V.A. Khoze, and
S.I. Troyan, Phys. Lett. 202B (1988) 276.
[21]
OPAL Collaboration, M.Z. Akrawy et al., Phys. Lett. 247B (1990) 617.
[22]
ALEPH Collaboration, R. Barate et al., CERN PPE 96-186.
[23]
TASSO Collaboration, M. Althoff et al., Z. Phys. C22 (1984) 307.
[24]
OPAL Collaboration, P.D. Acton et al., Z. Phys. C53 (1992) 539.
[25]
OPAL Collaboration, R. Akers et al., Z. Phys. C61 (1994) 209.
[26]
A.Basetto, Nucl. Phys. 202B (1982) 493.
[27]
D. Kant, University of London thesis, also published by the Rutherford Appleton
laboratory as RAL-TH-96-008.
[28]
V.A. Khoze, S. Lupia, and W. Ochs, Phys. Lett. 394B (1997) 179.
[29]
S. Lupia, private communication.
[30]
P. Van Mechelen, University of Antwerp thesis, in preparation.
[31]
Z. Koba, H.B. Nielsen and P. Olesen, Nucl. Phys. 40B (1972) 317.
[32]
TASSO Collaboration, W. Braunschweig et al., Z. Phys. C45 (1989) 193.
[33]
DELPHI Collaboration, P. Abreu et al. Z. Phys. C56 (1992) 63.
[34]
T. Sjöstrand, Computer Phys. Comm. 39 (1986) 32; T. Sjöstrand, M. Bengtsson,
Computer Phys. Comm. 43 (1987) 32.
[35]
DELPHI Collaboration, P. Abreu et al. Z. Phys. C73 (1996) 11. |
Asymptotically exact theory of fiber-reinforced composite beams
K. C. Le${}^{a,b}$111Email: [email protected], T. M. Tran${}^{a,b}$
${}^{a}$Materials Mechanics Research Group, Ton Duc Thang University, Ho Chi Minh City, Vietnam
${}^{b}$Faculty of Civil Engineering, Ton Duc Thang University, Ho Chi Minh City, Vietnam
Abstract
An asymptotic analysis of the energy functional of a fiber-reinforced composite beam with a periodic microstructure in its cross section is performed. From this analysis the asymptotically exact energy as well as the 1-D beam theory of first order is derived. The effective stiffnesses of the beam are calculated in the most general case from the numerical solution of the cell and homogenized cross-sectional problems.
keywords:
fiber-reinforced, composite, beam, periodic microstructure, variational-asymptotic method.
††journal: Composite Structures
1 Introduction
Fibre reinforced composite (FRC) beams are widely used in civil, mechanical, and aerospace engineering due to their low weight, high strength, and good damping properties (Hodges, 1990; Chandra et al., 1999; Librescu and Song, 2005). The exact treatment of FRC beams within the 3D elasticity is only possible in a few exceptional cases due to their complicated microstructure (see e.g. (Rodriguez et al., 2001; Sendeckyj, 2016) and the references therein). For this reason, different approaches have been developed depending on the type of beams. If FRC beams are thick, no exact one-dimensional theory can be constructed, so only the numerical methods or approximate semi-analytical methods applied to three-dimensional elasticity theory make sense. However, if the FRC beams are thin, the reduction from the three-dimensional to the one-dimensional theory is possible. This dimension reduction can be made asymptotically exact in the limit when the thickness-to-length ratio of the beam goes to zero. The rigorous derivation of the asymptotically exact one-dimensional beam theory based on the variational-asymptotic method (VAM) developed by Berdichevsky (1979) was first performed in (Berdichevsky, 1981). His asymptotic analysis shows that the static (or dynamic) three-dimensional problem of the beam can be split into two problems: (i) the two-dimensional cross-sectional problem and (ii) the one-dimensional variational problem whose energy functional should be found from the solution of the cross-sectional problem. The latter has been solved both for anisotropic homogeneous beams and for inhomogeneous beams with the constant Poisson’s ratio (Berdichevsky, 1981). In addition to these findings, Berdichevsky (1981) has shown that the average energy as well as the extension and bending stiffnesses of FRC beams with piecewise constant Poisson’s ratio must be larger than those of FRC beams with constant Poisson’s ratio (see also (Muskhelishvili, 2013)). However, to our knowledge, the question of how the corrections in energy and stiffnesses depend on the difference in Poisson’s ratio of fibers and matrix of FRC beams remains still an issue. It should be noted that VAM has been further developed in connection with the numerical analysis of cross-sectional problems for geometrically nonlinear composite beams by Hodges, Yu and colleagues in a series of papers (Hodges et al., 1992; Yu et al., 2002, 2005; Yu and Blair, 2012). Note also that VAM has been used, among others, to derive the 2D theory of homogeneous piezoelectric shells (Le, 1986b), the 2D theory of purely elastic anisotropic and inhomogeneous shells (Berdichevsky, 2009), the 2D theory of purely elastic sandwich plates and shells (Berdichevsky, 2010a, b), the theory of smart beams (Roy et al., 2007), the theory of low and high frequency vibrations of laminate composite shells (Lee and Hodges, 2009a, b), and more recently, the theory of smart sandwich and functionally graded shells (Le and Yi, 2016; Le, 2017).
For FRC beams that have the periodic microstructure in the cross section, an additional small parameter appears in the cross-sectional problems: The ratio between the length of the periodic cell and the characteristic size of the cross section. This allows an additional asymptotic analysis of the cross-sectional problems to simplify them. By solving the cell problems imposed with the periodic boundary conditions according to the homogenization technique (Braides, 2002; Milton, 2003), one finds the effective coefficients in the homogenized cross-sectional problems, which can then be solved analytically or numerically. The aim of this paper is to derive and solve the cell and homogenized cross-sectional problems for unidirectional FRC beams whose cross section has the periodic microstructure with piecewise constant Poisson’s ratio. The solution of the cell problems found with the finite element method is used to calculate the asymptotically exact energy and the strain and bending stiffnesses in the 1-D theory of FRC beams. Thus, we determine the dependence of the latter quantities on the shape and volume fraction of the fibers and on the difference in Poisson’s ratio of fibers and matrix that solves the above mentioned issue.
The paper is organized as follows. After this short introduction the variational formulation of the problem is given in Section 2. Sections 3 and 4 are devoted to the multi-scaled asymptotic analysis of the energy functional of FRC beams leading to the cross-sectional and cell problems. In Section 5 the cell problems are solved by the finite element method. Section 6 provides the solutions of the homogenized cross-sectional problems. Section 7 present one-dimensional theory of FRC beams. Finally, Section 8 concludes the paper.
2 Variational formulation for FRC beams
Consider a FRC beam that occupies the domain $\mathcal{B}=\mathcal{A}\times(0,L)$ of the 3-D euclidean space in its undeformed state. Let $x_{3}\equiv x$ be the coordinate along the beam axis. The cross section of the beam in the $(x_{1},x_{2})$-plane, $\mathcal{A}$, consist of two separated 2-D sub-domains $\mathcal{A}^{m}$ and $\mathcal{A}^{f}$ such that the matrix occupies the domain $\mathcal{A}^{m}\times(0,L)$, while the uni-directional fibers occupy the domain $\mathcal{A}^{f}\times(0,L)$. We choose the origin of the $(x_{1},x_{2})$-coordinates so that it matches the centroid of $\mathcal{A}$. We assume that the fibers are periodically situated in the matrix and the bond between the fibers and the matrix is perfect (see Fig. 1 representing the cross section of the beam where $\mathcal{A}^{f}$ is the set of gray circles). We first consider the beam in equilibrium, whose deformation is completely determined by the displacement field ${\bf u}(\mathbf{x})$. Let, for simplicity, the edge $x=0$ of the beam be clamped, while at the other edge $x=L$ of the beam the traction $\mathbf{t}(x_{1},x_{2})$ be specified. Gibbs’ variational principle (Berdichevsky, 2009) states that the true displacement $\check{\mathbf{u}}(\mathbf{x})$ minimizes the energy functional
$$I[\mathbf{u}(\mathbf{x})]=\int_{\mathcal{B}}W(\mathbf{x},\boldsymbol{%
\varepsilon})\differential[3]{x}-\int_{\mathcal{A}}\mathbf{t}(x_{1},x_{2})%
\dotproduct\mathbf{u}(x_{1},x_{2},L)\differential[2]{x}$$
(1)
among all admissible displacements $\mathbf{u}(\mathbf{x})$ satisfying the boundary conditions
$$\mathbf{u}(x_{1},x_{2},0)=0.$$
Here, $\differential[3]{x}=\differential x_{1}\differential x_{2}\differential x$ is the volume element, $\differential[2]{x}=\differential x_{1}\differential x_{2}$ the area element, and the dot denotes the scalar product. Function $W(\mathbf{x},\boldsymbol{\varepsilon})$, called stored energy density, reads
$$W(\mathbf{x},\boldsymbol{\varepsilon})=\frac{1}{2}\lambda(\mathbf{x})(\tr%
\boldsymbol{\varepsilon})^{2}+\mu(\mathbf{x})\boldsymbol{\varepsilon}\mathbf{:%
}\boldsymbol{\varepsilon},$$
where $\boldsymbol{\varepsilon}$ is the strain tensor
$$\boldsymbol{\varepsilon}=\frac{1}{2}(\gradient\mathbf{u}+(\gradient\mathbf{u})%
^{T}).$$
The problem is to replace the three-dimensional energy functional by an approximate one-dimensional energy functional for a thin FRC beam, whose functions depend only on the longitudinal co-ordinate $x$. The possibility of reduction of the three- to the one-dimensional functional is related to the smallness of two parameters: (i) the ratio between the thickness $h$ of the cross section and the length $L$ of the beam, and (ii) the ratio between the size $c$ of the periodic cell and $h$. By using the variational-asymptotic method, the one-dimensional energy functional will be constructed below in which terms of the order $h/L$ and $\epsilon=c/h$ are neglected as compared with unity (the first-order or “classical” approximation). Formally, this corresponds to the limits $h\to 0$ and $\epsilon\to 0$.
In order to perform the variational-asymptotic analysis, it is convenient to use the index notation, with Greek indices running from 1 to 2 and denoting the component of vectors and tensors in the $(x_{1},x_{2})$-plane. Summation over repeated Greek indices is understood. Index 3 of the coordinate $x_{3}$, the displacement $u_{3}$, and the traction $t_{3}$ is dropped for short. To fix the domain of the transverse co-ordinates in the passage to the limit $h\to 0$, we introduce the dimensionless co-ordinates
$$y_{\alpha}=\frac{x_{\alpha}}{h},\quad y_{\alpha}\in\bar{\mathcal{A}},$$
and transform the energy functional of the beam to
$$I=\int_{0}^{L}\int_{\bar{\mathcal{A}}}h^{2}W(y_{\alpha},\boldsymbol{%
\varepsilon})\differential[2]{y}\differential{x}-\int_{\bar{\mathcal{A}}}h^{2}%
\mathbf{t}(y_{\alpha})\dotproduct\mathbf{u}(y_{\alpha},L)\differential[2]{y}.$$
(2)
The transverse coordinates $y_{\alpha}$ play the role of the “fast” variables as opposed to the slow variable $x$. Regarding $\mathbf{u}$ as function of $y_{\alpha}$ and $x$, we separate the fast variables from the slow one. Now $h$ enters the action functional explicitly through the components of the strain tensor $\boldsymbol{\varepsilon}$
$$\varepsilon_{\alpha\beta}=\frac{1}{h}u_{(\alpha;\beta)},\quad 2\varepsilon_{%
\alpha 3}=\frac{1}{h}u_{;\alpha}+u_{\alpha,x},\quad\varepsilon_{33}=u_{,x}.$$
(3)
Here and below, the semicolon preceding Greek indices denotes the derivatives with respect to $y_{\alpha}$, while the parentheses surrounding a pair of indices the symmetrization operation.
Besides $y_{\alpha}$ there are also much faster variables
$$z_{\alpha}=\frac{x_{\alpha}}{c}=\frac{y_{\alpha}}{\epsilon}$$
associated with the periodicity of the elastic moduli of this composite materials in the transverse directions leading to the fast oscillation of the stress and strain fields in these directions. In order to characterize these faster changes of the stress and strain fields in the transverse directions we will assume that the displacement field $\mathbf{u}$ is a composite function of $\mathbf{y}$ and $x$ such that
$$\mathbf{u}=\mathbf{u}(\mathbf{y},\mathbf{y}/\epsilon,x)=\mathbf{u}(\mathbf{y},%
\mathbf{z},x),$$
where $\mathbf{y},\mathbf{z}$ are two-dimensional vectors with the components $y_{\alpha}$, $z_{\alpha}$, respectively, and where $\mathbf{u}(\mathbf{y},\mathbf{z},x)$ is a doubly periodic function in $\mathbf{z}$ with the period 1. This is a typical multi-scale Ansatz to composite structures. The asymptotic analysis must therefore be performed twice, first in the limit $h\to 0$ to realize the dimension reduction, and then in the limit $\epsilon\to 0$ to homogenize the cross-sectional problems and solve them.
3 Dimension reduction
Before starting the asymptotic analysis of the energy functional in the limit $h\to 0$ let us transform the stored energy density to another form more convenient for this purpose (Le, 1999). We note that, among terms of $W(\mathbf{z},\boldsymbol{\varepsilon})$, the derivatives $w_{(\alpha;\beta)}/h$ in $\varepsilon_{\alpha\beta}$ and $w_{;\alpha}/h$ in $\varepsilon_{\alpha 3}$ are the main ones in the asymptotic sense. Therefore it is convenient to single out the components $\varepsilon_{\alpha\beta}$ and $\varepsilon_{\alpha 3}$ in the stored energy density. We represent the latter as the sum of three quadratic forms $W_{\parallel}$, $W_{\angle}$, and $W_{\perp}$ according to
$$W_{\parallel}=\min_{\varepsilon_{\alpha\beta},\varepsilon_{\alpha 3}}W,\quad W%
_{\angle}=\min_{\varepsilon_{\alpha\beta}}(W-W_{\parallel}),\quad W_{\perp}=W-%
W_{\parallel}-W_{\angle}.$$
The “longitudinal” energy density $W_{\parallel}$ depends only on $\varepsilon_{33}$ and coincides with $W$ when the stresses $\sigma_{\alpha\beta}$ and $\sigma_{\alpha 3}$ vanish; the “shear” energy density $W_{\angle}$ depends only on $\epsilon_{\alpha 3}$; the remaining part $W_{\perp}$ is called the “transverse” energy density. From the definitions of $W_{\parallel}$, $W_{\angle}$ and
$W_{\perp}$ one easily finds out that
$$\displaystyle W_{\parallel}=\frac{1}{2}E(\varepsilon_{33})^{2},\quad E=\frac{%
\mu(3\lambda+2\mu)}{\lambda+\mu},$$
$$\displaystyle W_{\angle}=2\mu\varepsilon_{\alpha 3}\varepsilon_{\alpha 3},$$
$$\displaystyle W_{\perp}=\frac{1}{2}\lambda(\varepsilon_{\alpha\alpha}+2\nu%
\varepsilon_{33})^{2}+\mu(\varepsilon_{\alpha\beta}+\nu\delta_{\alpha\beta}%
\varepsilon_{33})(\varepsilon_{\alpha\beta}+\nu\delta_{\alpha\beta}\varepsilon%
_{33}),$$
where $E$ is the Young modulus and $\nu=\lambda/2(\lambda+\mu)$ the Poisson ratio. Note that $E$, $\nu$ as well as Lame’s constants $\lambda$, $\mu$ are doubly periodic functions of the fast variables $\mathbf{z}$. Note also the following identities
$$\displaystyle W_{\angle}$$
$$\displaystyle=\sigma_{\alpha 3}\varepsilon_{\alpha 3},$$
$$\displaystyle W_{\perp}$$
$$\displaystyle=\frac{1}{2}(\sigma_{\alpha\beta}\varepsilon_{\alpha\beta}+\sigma%
_{33}\varepsilon_{33}-E(\varepsilon_{33})^{2}).$$
We could start the variational-asymptotic analysis in the limit $h\to 0$ with the determination of the set $\mathcal{N}$ according to its general scheme (Le, 1999). As a result, it would turn out that, at the first step, the function ${\bf u}$ does not depend on the transverse co-ordinates $\mathbf{y}$ (and $\mathbf{z}$): ${\bf u}={\bf v}(x)$; at the second step the function ${\bf u}^{\star}$ becomes a linear function of $\mathbf{y}$ and involves one more degree of freedom $\varphi(x)$ representing the twist angle; and at the next step ${\bf u}^{\star\star}$ is completely determined through ${\bf u}$ and $\varphi$. Thus, the set $\mathcal{N}$ according to the variational-asymptotic scheme consists of functions ${\bf v}(x)$ and $\varphi(x)$. We will pass over these long, but otherwise standard, deliberations and make a change of unknown functions according to
$$\begin{split}\displaystyle u_{\alpha}(\mathbf{y},\mathbf{z},x)=v_{\alpha}(x)-%
he_{\alpha\beta}\varphi(x)y_{\beta}+hw_{\alpha}(\mathbf{y},\mathbf{z},x),\\
\displaystyle u(\mathbf{y},\mathbf{z},x)=v(x)-hv_{\alpha,x}(x)y_{\alpha}+hw(%
\mathbf{y},\mathbf{z},x),\end{split}$$
(4)
where $e_{\alpha\beta}$ are the two-dimensional permutation symbols ($e_{11}=e_{22}=0,e_{12}=-e_{21}=1$). By redefining $\mathbf{v}(x)$ and $\varphi(x)$ if necessary we can impose on functions $w_{\alpha}$ and $w$ the following constraints
$$\begin{split}\displaystyle\langle w_{\alpha}\rangle=0,\quad e_{\alpha\beta}%
\langle w_{\alpha;\beta}\rangle=0,\\
\displaystyle\quad\langle w\rangle=0,\quad\langle.\rangle\equiv\int_{\bar{%
\mathcal{A}}}.\differential[2]{y}.\end{split}$$
(5)
According to these constraints $v_{\alpha}(x)$ and $v(x)$ describe the mean displacements of the beam, while $\varphi(x)$ corresponds to the mean rotation of its cross section. Equations (4) and (5) set up a one-to-one correspondence between $u_{\alpha},u$ and the set of functions $v_{\alpha},v,\varphi,w_{\alpha},w$ and determine the change in the unknown functions $\{u_{\alpha},u\}\to\{v_{\alpha},v,\varphi,w_{\alpha},w\}$.
Based on the Saint-Venant principle for elastic beams, we may assume that the domain occupied by the beam consists of the inner domain and two boundary layers near the edges of the beam with width of the order $h$ where the stress and strain states are essentially three-dimensional. Then functional (2) can be decomposed into the sum of two functionals, an inner one for which an iteration process will be applied, and a boundary layer functional. When searching for $w_{\alpha}$ and $w$ the boundary layer functional can be neglected in the first-order approximation. Therefore, the dimension reduction problem reduces to finding the minimizer $\check{w}_{\alpha}$ and $\check{w}$ of the inner functional that, in the limit $h\to 0$, can be identified with the functional (2) without the last term.
We now fix $v_{\alpha},v,\varphi$ and seek $w_{\alpha},w$. Substituting (4) into the action functional (2) with the last term being removed, we will keep in it the asymptotically principal terms containing $w_{\alpha},w$ and neglect all other terms. The estimations based on Eqs. (3) and (4) lead to the asymptotic formulas
$$\varepsilon_{\alpha\beta}=w_{(\alpha;\beta)},\quad 2\varepsilon_{\alpha 3}=w_{%
;\alpha}-he_{\alpha\beta}\Omega y_{\beta},\quad\varepsilon_{33}=\gamma+h\Omega%
_{\alpha}y_{\alpha},$$
(6)
where $\gamma$, $\Omega_{\alpha}$, and $\Omega$ are the measures of elongation, bending, and twist defined by
$$\gamma=v_{,x},\quad\Omega_{\alpha}=-v_{\alpha,xx},\quad\Omega=\varphi_{,x}.$$
According to equations (6) the partial derivatives of $w_{\alpha},w$ with respect to
$x$ do not enter the energy functional. As $x$ becomes the formal parameter, we may drop the integral over $x$ and reduce the determination of $w_{\alpha},w$ to the uncoupled minimization problems for every fixed $x$ of the functionals
$$\displaystyle I_{\perp}[w_{\alpha}]=h^{2}\langle\frac{1}{2}\lambda(\mathbf{z})%
[w_{\alpha;\alpha}+2\nu(\mathbf{z})(\gamma+h\Omega_{\sigma}y_{\sigma})]^{2}\\
\displaystyle+\mu(\mathbf{z})[w_{(\alpha;\beta)}+\nu(\mathbf{z})\delta_{\alpha%
\beta}(\gamma+h\Omega_{\gamma}y_{\gamma})][w_{(\alpha;\beta)}+\nu(\mathbf{z})%
\delta_{\alpha\beta}(\gamma+h\Omega_{\delta}y_{\delta})]\rangle,$$
(7)
$$I_{\angle}[w]=h^{2}\langle\frac{1}{2}\mu(\mathbf{z})(w_{;\alpha}-h\Omega e_{%
\alpha\beta}y_{\beta})(w_{;\alpha}-h\Omega e_{\alpha\gamma}y_{\gamma})\rangle.$$
(8)
The minima are searched among all admissible functions $w_{\alpha},w$ satisfying the constraints (5). Functionals (7) and (8) represent the transverse and shear strain energies, integrated over the cross section of the beam. They are positive definite and convex, so the existence of their minimizers $\check{w}_{\alpha},\check{w}$ is guaranteed. We shall see in the next Sections that the minimum of (7) is equal to zero if $\nu$ is equal for both matrix and fiber, while that of (8) is equal to $1/2C\Omega^{2}$, with $C$ the torsional rigidity.
4 Homogenization
Consider now the other limit $\epsilon\to 0$. In this limit $\mathbf{y}$ plays the role of the “slow” variable, while $\mathbf{z}=\mathbf{y}/\epsilon$ becomes the fast variable. We start with the cross-sectional problem of minimizing functional (8) among $w(\mathbf{y},\mathbf{z},x)$ satisfying the constraint (5)${}_{3}$, where $\mu(\mathbf{z})$, expressed in terms of the fast variable $\mathbf{z}=\mathbf{y}/\epsilon$, is a doubly periodic function with period 1. Since $x$ is fixed in this cross-sectional problem, we shall drop this formal variable of $w$ in this Section. Following the homogenization technique based also on the variational-asymptotic method (Berdichevsky, 2009), we look for the minimizer in the form
$$w(\mathbf{y},\mathbf{z})=\psi(\mathbf{y})+\epsilon\chi(\mathbf{z}),$$
(9)
where $\chi(\mathbf{z})$ is a doubly periodic function with period 1. Note that $\chi$ may depend on the slow variable $\mathbf{y}$, but this dependency is suppressed for short. In addition to the constraint (5)${}_{3}$ we may impose the following constraint on $\chi(\mathbf{z})$
$$\langle\!\langle\chi\rangle\!\rangle\equiv\int_{\mathcal{C}}\chi(\mathbf{z})%
\differential[2]{z}=0,$$
(10)
where $\mathcal{C}=(0,1)\times(0,1)$ is the unit periodic cell. In this case $\psi(\mathbf{y})$ can be interpreted as the average value of $w(\mathbf{y},\mathbf{z})$ over the cell. Although the second term in (9) is small compared to $\psi(\mathbf{y})$ and goes to zero in the limit $\epsilon\to 0$, it contribution to the shear strains as the gradient of $w(\mathbf{y},\mathbf{z})$ has the same order as the gradient of $\psi(\mathbf{y})$ as seen from the asymptotic formula
$$w_{;\alpha}=\psi_{;\alpha}+\chi_{|\alpha},\quad\alpha=1,2.$$
(11)
Here, the vertical bar followed by an index $\alpha$ denotes the partial derivative with respect to the corresponding fast variable $z_{\alpha}$.
Inserting (11) into the energy functional (8), we get
$$I_{\angle}[\psi(\mathbf{y}),\chi(\mathbf{y}/\epsilon)]=h^{2}\langle\frac{1}{2}%
\mu(\mathbf{y}/\epsilon)(\chi_{|\alpha}+\xi_{\alpha})(\chi_{|\alpha}+\xi_{%
\alpha})\rangle,$$
with $\xi_{\alpha}=\psi_{;\alpha}-h\Omega e_{\alpha\beta}y_{\beta}$ being the function of the “slow” variable $\mathbf{y}$. We replace the double integral $\langle.\rangle$ by the sum of double integrals over the cells. Then
$$I_{\angle}[\psi(\mathbf{y}),\chi(\mathbf{y}/\epsilon)]=h^{2}\sum_{n=1}^{N}\int%
_{\mathcal{C}_{n}}\frac{1}{2}\mu(\mathbf{y}/\epsilon)(\chi_{|\alpha}+\xi_{%
\alpha})(\chi_{|\alpha}+\xi_{\alpha})\differential[2]{y},$$
(12)
where $N$ is the total number of cells and $\mathcal{C}_{n}$ is the cell $((i-1)\epsilon,i\epsilon)\times((j-1)\epsilon,j\epsilon)$. We minimize this functional in two steps: (i) Fix $\psi(\mathbf{y})$ and minimize the functional among doubly periodic $\chi(\mathbf{y}/\epsilon)$, (ii) Minimize the obtained functional among $\psi(\mathbf{y})$. Because $\chi(\mathbf{y}/\epsilon)$ is doubly periodic with respect to $\mathbf{y}$, we can minimize each integral in the sum independently. Besides, as $\xi_{\alpha}(\mathbf{y})$ change slowly in one cell, we may regard them in each cell integral as constants equal to their value in the middle of the cell. It is convenient to change the variable $\mathbf{y}$ in the cell integrals to $\mathbf{z}=\mathbf{y}/\epsilon-(i-1,j-1)$. Then the minimization of cell integrals reduces to the following cell problem: Minimize the functional
$$\langle\!\langle\frac{1}{2}\mu(\mathbf{z})(\chi_{|\alpha}+\xi_{\alpha})(\chi_{%
|\alpha}+\xi_{\alpha})\rangle\!\rangle$$
(13)
among doubly periodic functions $\chi(\mathbf{z})$ satisfying the constraint (10), where, as before, $\langle\!\langle.\rangle\!\rangle$ is the double integral in $\mathbf{z}$ over the unit cell. We denote the obtained minimum by $\bar{W}_{\angle}(\xi_{\alpha})$ and call it the average shear energy density. Then, replacing the sum in (12) by the integral in the limit $\epsilon\to 0$, we arrive at the following homogenized cross-sectional problem: Minimize the average functional
$$\bar{I}_{\angle}[\psi(\mathbf{y})]=h^{2}\langle\bar{W}_{\angle}(\psi_{;\alpha}%
-h\Omega e_{\alpha\beta}y_{\beta})\rangle$$
(14)
among $\psi(\mathbf{y})$ satisfying the condition
$$\langle\psi\rangle=0.$$
The standard calculus of variations shows that the minimizer of (13) satisfies the equation
$$[\mu(\mathbf{z})(\chi_{|\alpha}+\xi_{\alpha})]_{|\alpha}=0$$
(15)
in the unit cell, where $\boldsymbol{\xi}$ can be regarded as the constant vector in the cell (since it does not depend on the fast variable $\mathbf{z}$). This equation is subjected to the periodic boundary condition and the constraint (10). Besides, as the shear modulus suffers a jump at the boundary $\partial_{i}$ between the fiber and the matrix, the continuity of the traction
$$[\![\mu(\mathbf{z})(\chi_{|\alpha}+\xi_{\alpha})]\!]\nu_{\alpha}=0$$
(16)
should be fulfilled at this internal boundary $\partial_{i}$, where $\boldsymbol{\nu}$ is the unit normal vector outward to the fiber and $[\![.]\!]$ denotes the jump.
After finding the minimizer $\check{\chi}(\mathbf{z})$ as a solution to the boundary-value problem (15)-(16), we substitute it back into functional (13) to calculate the average shear energy density
$$\displaystyle\bar{W}_{\angle}(\xi_{\alpha})$$
$$\displaystyle=\langle\!\langle\frac{1}{2}\mu(\mathbf{z})(\check{\chi}_{|\alpha%
}+\xi_{\alpha})(\check{\chi}_{|\alpha}+\xi_{\alpha})\rangle\!\rangle$$
$$\displaystyle=\langle\!\langle\frac{1}{2}\mu(\mathbf{z})(\check{\chi}_{|\alpha%
}+\xi_{\alpha})\check{\chi}_{|\alpha}\rangle\!\rangle+\langle\!\langle\frac{1}%
{2}\mu(\mathbf{z})(\check{\chi}_{|\alpha}+\xi_{\alpha})\xi_{\alpha}\rangle\!\rangle.$$
(17)
As the minimizer $\check{\chi}$ satisfies (15)-(16), the first integral must vanish. Using the constancy of $\xi_{\alpha}$, we reduce the second integral to
$$\bar{W}_{\angle}(\xi_{\alpha})=\frac{1}{2}\xi_{\alpha}\langle\!\langle\mu(%
\mathbf{z})(\check{\chi}_{|\alpha}+\xi_{\alpha})\rangle\!\rangle.$$
The integrand is the shear stress $\sigma_{3\alpha}$, so the integral gives the average shear stress $\bar{\sigma}_{3\alpha}$. On the other side, due to the constraint (10),
$$\langle\!\langle\check{\chi}_{|\alpha}+\xi_{\alpha}\rangle\!\rangle=\xi_{%
\alpha},$$
so $\xi_{\alpha}$ is the average engineering shear strain $\bar{\varepsilon}_{3\alpha}$. We define the effective elastic shear moduli $\mu^{*}_{\alpha\beta}$ by the linear equation
$$\bar{\sigma}_{3\alpha}=\langle\!\langle\mu(\mathbf{z})(\check{\chi}_{|\alpha}+%
\xi_{\alpha})\rangle\!\rangle=\mu^{*}_{\alpha\beta}\xi_{\beta}.$$
(18)
Note that the homogenized material must not necessarily be isotropic even if the components of the composite are. So, in general $\mu^{*}_{\alpha\beta}$ is a tensor of second order. With this we get the average shear energy density in terms of $\xi_{\alpha}=\psi_{;\alpha}-h\Omega e_{\alpha\beta}y_{\beta}$
$$\bar{W}_{\angle}(\xi_{\alpha})=\frac{1}{2}\mu^{*}_{\alpha\beta}\xi_{\alpha}\xi%
_{\beta}.$$
(19)
We want to show that the tensor $\mu^{*}_{\alpha\beta}$ in (18) is symmetric. Indeed, since $\bar{W}_{\angle}(\xi_{\alpha})$ in (19) is a quadratic form, we can replace $\mu^{*}_{\alpha\beta}$ there by the symmetric tensor $\mu^{*}_{(\alpha\beta)}=\frac{1}{2}(\mu^{*}_{\alpha\beta}+\mu^{*}_{\beta\alpha})$. If we substitute (19) into (17) and differentiate it with respect to $\xi_{\alpha}$, we get
$$\mu^{*}_{(\alpha\beta)}\xi_{\beta}=\langle\!\langle\mu(\mathbf{z})(\check{\chi%
}_{|\alpha}+\xi_{\alpha})\rangle\!\rangle=\mu^{*}_{\alpha\beta}\xi_{\beta},$$
which proves the symmetry of $\mu^{*}_{\alpha\beta}$.
We turn now to the cross-sectional problem (7). Before doing the asymptotic analysis for it in the limit $\epsilon\to 0$ let us prove the following remarkable property: The minimum of functional (7) is zero if the Poisson ratio $\nu$ is an equal constant for both fiber and matrix (Berdichevsky, 1981). Indeed, let us choose the minimizer in the form
$$\check{w}_{\alpha}=-\nu\gamma\delta_{\alpha\beta}y_{\beta}-\frac{1}{2}\nu a_{%
\alpha\beta\gamma}(y_{\beta}y_{\gamma}-\langle y_{\beta}y_{\gamma}\rangle/|%
\bar{\mathcal{A}}|),$$
(20)
where $|\bar{\mathcal{A}}|$ is the area of $\bar{\mathcal{A}}$ and
$$a_{\alpha\beta\gamma}=h(\delta_{\alpha\beta}\Omega_{\gamma}+\delta_{\alpha%
\gamma}\Omega_{\beta}-\delta_{\beta\gamma}\Omega_{\alpha}).$$
Note that the tensor $a_{\alpha\beta\gamma}$ is symmetric with the last two indices, and
$$\check{w}_{\alpha;\beta}=-\nu\gamma\delta_{\alpha\beta}-a_{\alpha\beta\gamma}y%
_{\gamma}.$$
Due to our choice of the origin, $\langle y_{\alpha}\rangle=0$. Therefore the field (20) satisfies the constraints (5)${}_{1,2}$. It is easy to check that the transverse energy evaluated at $\check{w}_{\alpha}$ vanishes identically. Since functional (7) is non-negative definite, its minimum is obviously zero in this case.
The asymptotic analysis of problem (7) for changeable $\nu$ is quite similar to that of problem (8). In this case the minimizer is sought in the form
$$w_{\alpha}(\mathbf{y},\mathbf{z})=\psi_{\alpha}(\mathbf{y})+\epsilon\chi_{%
\alpha}(\mathbf{z}),$$
where $\chi_{\alpha}$ satisfy the constraints
$$\langle\!\langle\chi_{\alpha}\rangle\!\rangle=0.$$
(21)
To determine $\chi_{\alpha}$ we need to solve the following cell problem: Minimize the functional
$$\displaystyle\langle\!\langle\frac{1}{2}\lambda(\mathbf{z})(\chi_{\alpha|%
\alpha}+\bar{\varepsilon}_{\alpha\alpha}+2\nu(\mathbf{z})\xi)^{2}\\
\displaystyle+\mu(\mathbf{z})(\chi_{(\alpha|\beta)}+\bar{\varepsilon}_{\alpha%
\beta}+\nu(\mathbf{z})\delta_{\alpha\beta}\xi)(\chi_{(\alpha|\beta)}+\bar{%
\varepsilon}_{\alpha\beta}+\nu(\mathbf{z})\delta_{\alpha\beta}\xi)\rangle\!\rangle$$
(22)
among doubly periodic functions $\chi_{\alpha}(\mathbf{z})$ satisfying the constraint (21), where
$$\bar{\varepsilon}_{\alpha\beta}=\psi_{(\alpha|\beta)},\quad\xi=\gamma+h\Omega_%
{\alpha}y_{\alpha}.$$
Note that this problem is quite similar to the problem of determining the effective thermal expansion of composite material with periodic microstructure, where $\xi$ plays the role of the temperature increase (Sigmund and Torquato, 1997). Let the minimum of functional (22) be $\bar{W}_{\perp}(\bar{\varepsilon}_{\alpha\beta},\xi)$. Then the determination of $\psi_{\alpha}$ reduces to minimizing the following average functional
$$\bar{I}_{\perp}[\psi_{\alpha}(\mathbf{y})]=h^{2}\langle\bar{W}_{\perp}(\psi_{(%
\alpha;\beta)},\gamma+h\Omega_{\alpha}y_{\alpha})\rangle$$
(23)
among $\psi_{\alpha}(\mathbf{y})$ satisfying the constraints
$$\langle\psi_{\alpha}\rangle=0,\quad e_{\alpha\beta}\langle\psi_{\alpha;\beta}%
\rangle=0.$$
(24)
The minimizer of functional (22) satisfy the equilibrium equations
$$[\lambda(\mathbf{z})(\chi_{\beta|\beta}+\bar{\varepsilon}_{\beta\beta}+\xi)]_{%
|\alpha}+[2\mu(\mathbf{z})(\chi_{(\alpha|\beta)}+\bar{\varepsilon}_{\alpha%
\beta})]_{|\beta}=0$$
in the unit cell, where $\bar{\varepsilon}_{\alpha\beta}$ and $\xi$ can be regarded as the constant tensor and scalar in the cell (since they do not depend on the fast variable $\mathbf{z}$). These equations are subjected to the periodic boundary conditions and the constraints (21). Besides, as the elastic moduli suffers jumps at the boundary $\partial_{i}$ between the fiber and the matrix, the continuity of the traction
$$[\![\lambda(\mathbf{z})(\chi_{\beta|\beta}+\bar{\varepsilon}_{\beta\beta}+\xi)%
\nu_{\alpha}+2\mu(\mathbf{z})(\chi_{(\alpha|\beta)}+\bar{\varepsilon}_{\alpha%
\beta})\nu_{\beta}]\!]=0$$
should be fulfilled there. Since the functional (22) is quadratic, its minimum must be a quadratic form of $\bar{\varepsilon}_{\alpha\beta}$ and $\xi$
$$\bar{W}_{\perp}(\bar{\varepsilon}_{\alpha\beta},\xi)=\frac{1}{2}C^{*}_{\alpha%
\beta\gamma\delta}\bar{\varepsilon}_{\alpha\beta}\bar{\varepsilon}_{\gamma%
\delta}+D^{*}_{\alpha\beta}\bar{\varepsilon}_{\alpha\beta}\xi+\frac{1}{2}F^{*}%
\xi^{2}.$$
(25)
Using the same arguments as in the previous case, we can prove the following symmetry properties of the effective moduli
$$\displaystyle C^{*}_{\alpha\beta\gamma\delta}=C^{*}_{\beta\alpha\gamma\delta}=%
C^{*}_{\alpha\beta\delta\gamma}=C^{*}_{\gamma\delta\alpha\beta},$$
$$\displaystyle D^{*}_{\alpha\beta}=D^{*}_{\beta\alpha}.$$
We also introduce the effective tensor of Poisson’s ratios by
$$\nu^{*}_{\alpha\beta}=C^{*(-1)}_{\alpha\beta\gamma\delta}D_{\gamma\nu},$$
where $C^{*(-1)}_{\alpha\beta\gamma\delta}$ is the tensor of elastic compliances defined by
$$C^{*(-1)}_{\alpha\beta\gamma\delta}C_{\gamma\delta\zeta\eta}=\delta_{\alpha(%
\zeta}\delta_{\beta\eta)}.$$
It is easy to see that $\nu_{\alpha\beta}$ is symmetric.
5 Numerical solution of the cell problems
It is convenient to rewrite functional (22) by changing the sign of $\chi_{\alpha}(\mathbf{y})$
$$\frac{1}{2}\langle\!\langle C_{\alpha\beta\gamma\delta}(\mathbf{z})(-\chi_{(%
\alpha|\beta)}+\bar{\varepsilon}_{\alpha\beta}+\alpha_{\alpha\beta}(\mathbf{z}%
)\xi)(-\chi_{(\gamma|\delta)}+\bar{\varepsilon}_{\gamma\delta}+\alpha_{\gamma%
\delta}(\mathbf{z})\xi)\rangle\!\rangle,$$
(26)
where
$$C_{\alpha\beta\gamma\delta}(\mathbf{z})=\lambda(\mathbf{z})\delta_{\alpha\beta%
}\delta_{\gamma\delta}+\mu(\mathbf{z})(\delta_{\alpha\gamma}\delta_{\beta%
\gamma}+\delta_{\alpha\delta}\delta_{\beta\gamma}),\quad\alpha_{\alpha\beta}(%
\mathbf{z})=\nu(\mathbf{z})\delta_{\alpha\beta}.$$
The minimizer of functional (26) satisfies the variational equation
$$\langle\!\langle C_{\alpha\beta\gamma\delta}(\mathbf{z})\chi_{(\gamma|\delta)}%
\delta\chi_{\alpha|\beta}\rangle\!\rangle=\langle\!\langle C_{\alpha\beta%
\gamma\delta}(\mathbf{z})(\bar{\varepsilon}_{\gamma\delta}+\alpha_{\gamma%
\delta}(\mathbf{z})\xi)\delta\chi_{\alpha|\beta}\rangle\!\rangle$$
(27)
for all doubly periodic functions $\delta\chi_{\alpha}$. Eq. (27) will be solved by the finite element method (Hughes, 2012; Andreasen, 2014). For this purpose it is convenient to change from tensor notation to matrix notation (Hughes, 2012), in which Eq. (27) becomes
$$\langle\!\langle\boldsymbol{\varepsilon}^{T}(\boldsymbol{\psi})\mathbf{C}(%
\mathbf{z})\boldsymbol{\varepsilon}(\boldsymbol{\chi})\rangle\!\rangle=\langle%
\!\langle\boldsymbol{\varepsilon}^{T}(\boldsymbol{\psi})\mathbf{C}(\mathbf{z})%
(\bar{\boldsymbol{\varepsilon}}+\boldsymbol{\alpha}(\mathbf{z})\xi)\rangle\!\rangle,$$
(28)
where $\boldsymbol{\psi}=\delta\boldsymbol{\chi}$, $\boldsymbol{\varepsilon}(\boldsymbol{\chi})$ is the strain vector given by
$$\boldsymbol{\varepsilon}(\boldsymbol{\chi})=(\chi_{1|1},\chi_{2|2},\chi_{1|2}+%
\chi_{2|1})^{T},$$
while
$$\mathbf{C}(\mathbf{z})=\lambda(\mathbf{z})\begin{bmatrix}1&1&0\\
1&1&0\\
0&0&0\end{bmatrix}+\mu(\mathbf{z})\begin{bmatrix}2&0&0\\
0&2&0\\
0&0&1\end{bmatrix},\quad\boldsymbol{\alpha}(\mathbf{z})=\nu(\mathbf{z})\begin{%
bmatrix}1\\
1\\
0\end{bmatrix}.$$
(29)
The discretization of this equation based on the bilinear isoparametric elements is standard. The periodic boundary conditions are imposed by identifying the nodes on opposite sides of the unit cell. This is implemented by using the matrix edofMat for a full, regular grid to index into a periodic version of the grid (Andreasen, 2014). The global stiffness matrix is
$$\mathbf{K}=\mathop{\vphantom{\sum}\mathchoice{\vbox{\hbox{\huge$\mathrm{A}$}}}%
{\vbox{\hbox{\Large$\mathrm{A}$}}}{\mathrm{A}}{\mathrm{A}}}\displaylimits_{e=1%
}^{N}(\mathbf{k}_{e}),\quad\mathbf{k}_{e}=\int_{A_{e}}\mathbf{B}^{T}_{e}%
\mathbf{C}_{e}\mathbf{B}_{e}\differential[2]{z},$$
where $\mathop{\vphantom{\sum}\mathchoice{\vbox{\hbox{\huge$\mathrm{A}$}}}{\vbox{%
\hbox{\Large$\mathrm{A}$}}}{\mathrm{A}}{\mathrm{A}}}\displaylimits$ is the assembly operator taken over the total number $N$ of finite elements. The matrix $\mathbf{B}_{e}$ is the element strain-displacement matrix, $A_{e}$ is the domain of element $e$, and $\mathbf{C}_{e}$ is the constitutive matrix for the element. The indicator matrix is introduced that specifies whether the element is in $\mathcal{A}^{m}$ ($n_{e}=1$) or in $\mathcal{A}^{f}$ ($n_{e}=2$). The piecewise constant $\mathbf{C}(\mathbf{z})$ from (29) takes value $\mathbf{C}_{e}$ in the element in accordance with this indicator matrix.
The discretization of the right hand side of (28) yields the loads $\mathbf{f}^{(i)}$
$$\mathbf{f}^{(i)}=\mathop{\vphantom{\sum}\mathchoice{\vbox{\hbox{\huge$\mathrm{%
A}$}}}{\vbox{\hbox{\Large$\mathrm{A}$}}}{\mathrm{A}}{\mathrm{A}}}%
\displaylimits_{e=1}^{N}(\mathbf{f}^{(i)}_{e}),\quad\mathbf{f}^{(i)}_{e}=\int_%
{A_{e}}\mathbf{B}^{T}_{e}\mathbf{C}_{e}\bar{\boldsymbol{\varepsilon}}^{(i)}%
\differential[2]{z},\quad i=1,2,3$$
which correspond to the average macroscopic strains
$$\bar{\boldsymbol{\varepsilon}}^{(1)}=(1,0,0)^{T},\quad\bar{\boldsymbol{%
\varepsilon}}^{(2)}=(0,1,0)^{T},\quad\bar{\boldsymbol{\varepsilon}}^{(3)}=(0,0%
,1)^{T},$$
(30)
and
$$\mathbf{f}^{(4)}=\mathop{\vphantom{\sum}\mathchoice{\vbox{\hbox{\huge$\mathrm{%
A}$}}}{\vbox{\hbox{\Large$\mathrm{A}$}}}{\mathrm{A}}{\mathrm{A}}}%
\displaylimits_{e=1}^{N}(\mathbf{f}^{(4)}_{e}),\quad\mathbf{f}^{(4)}_{e}=\int_%
{A_{e}}\mathbf{B}^{T}_{e}\mathbf{C}_{e}\boldsymbol{\alpha}_{e}\differential[2]%
{z},$$
which correspond to $\xi=1$, with $\boldsymbol{\alpha}_{e}$ being the constitutive vector from (29) that takes values in the element in accordance with the indicator matrix. The displacement fields are computed by solving the corresponding linear equations with four load-cases
$$\mathbf{K}\boldsymbol{\chi}^{(i)}=\mathbf{f}^{(i)},\quad i=1,2,3,4.$$
When the displacements are obtained, the elements of the effective matrix $\bf{C}^{*}$ are found as:
$$C^{*}_{ij}=\sum_{e=1}^{N}\int_{A_{e}}(\boldsymbol{\chi}_{e}^{0(i)}-\boldsymbol%
{\chi}_{e}^{(i)})^{T}\mathbf{k}_{e}(\boldsymbol{\chi}_{e}^{0(j)}-\boldsymbol{%
\chi}_{e}^{(j)})\differential[2]{z},\quad i,j=1,2,3,$$
where $\boldsymbol{\chi}_{e}^{0(i)}$ are the three displacement fields corresponding to the unit strains from (30), and $\boldsymbol{\chi}_{e}^{(i)}$ contains three columns corresponding to the three displacement fields resulting from globally enforcing the unit strains (30). The indices in parentheses refer to the column number. The components of the effective moduli $D^{*}_{i}$ ($D^{*}_{\alpha\beta}$) due to $\xi$ are computed according to
$$D^{*}_{i}=\sum_{e=1}^{N}\int_{A_{e}}(\boldsymbol{\alpha}_{e}-\mathbf{B}_{e}%
\boldsymbol{\chi}_{e}^{(4)})^{T}\mathbf{C}_{e}(\bar{\boldsymbol{\varepsilon}}^%
{i}-\mathbf{B}_{e}\boldsymbol{\chi}_{e}^{(i)})\differential[2]{z},\quad i=1,2,3.$$
Having $D^{*}_{i}$ we can also compute the effective “Poisson” ratios $\nu^{*}_{i}$ ($\nu^{*}_{\alpha\beta}$) as follows
$$\nu^{*}_{i}=(C^{*}_{ij})^{-1}D^{*}_{j}.$$
(31)
Finally, the effective coefficient $F^{*}$ is given by
$$F^{*}=\sum_{e=1}^{N}\int_{A_{e}}(\boldsymbol{\alpha}_{e}-\mathbf{B}_{e}%
\boldsymbol{\chi}_{e}^{(4)})^{T}\mathbf{C}_{e}(\boldsymbol{\alpha}_{e}-\mathbf%
{B}_{e}\boldsymbol{\chi}_{e}^{(4)})\differential[2]{z}.$$
The Matlab-code homogenizecs.m to solve the cell problem and compute the effective moduli which is a modification of the code homogenize.m written by Andreasen (2014) is presented in the Appendix.
For the FRC bar it will be shown in the next Section that the minimum of functional (23) is $\frac{1}{2}(F^{*}-\nu^{*}_{\alpha\beta}C^{*}_{\alpha\beta\gamma\delta}\nu^{*}_%
{\gamma\delta})h^{2}\langle\xi^{2}\rangle$. Therefore it makes sense to investigate the quantity $H^{*}=F^{*}-\nu^{*}_{\alpha\beta}C^{*}_{\alpha\beta\gamma\delta}\nu^{*}_{%
\gamma\delta}$ giving the correction to the stiffnesses on extension and bending. We take the microstructure of the composite in such a way that the cross sections of the fibers are ellipses of half-axes $r$ and $r/a$ placed in the middle of the unit quadratic periodic cells. Then the region occupied by one fiber in the unit cell is given by the equation
$$\frac{(z_{1}-1/2)^{2}}{r^{2}}+\frac{(z_{2}-1/2)^{2}}{(r/a)^{2}}\leq 1.$$
Note that for the circle with $a=1$ the effective Poisson ratio tensor is $\nu^{*}_{\alpha\beta}=\nu^{*}\delta_{\alpha\beta}$, however this property is no longer valid for ellipses with $a>1$. Assigning the index 1 to the matrix and index 2 to the fiber, we choose the Young moduli and Poisson ratios as follows: $E_{1}=1\,$GPa, $E_{2}=2\,$GPa, $\nu_{1}=0.1$, $\nu_{2}=0.4$. The plot of $H^{*}$ as function of the largest half-axis $r$ for four different aspect ratios of the ellipses $a=1,2,4,8$ is presented in Fig. 2. We see that the correction $H^{*}$ varies in the range $(0,0.03)\,$GPa which is not so large compared with the mean value of the Young moduli. Since the volume fraction of the fibers is $s_{2}=\pi r^{2}/a$, we can also plot $H^{*}$ as function of this volume fraction.
Let us now fix the largest half-axis of elliptical cross section of the fibers to be $r=0.4$, choose $E_{1}=1\,$GPa, $E_{2}=2\,$GPa, $\nu_{1}=0.3$, vary the Poisson ratio of the fiber $\nu_{2}$ in its admissible range between $-1$ and $0.5$, and plot $H^{*}$ as function of $\nu_{2}$. Looking at this plot shown in Fig. 3 we see that the larger the difference between the Poisson ratios, the larger is the correction $H^{*}$. For $\nu_{2}=0.3$ the correction $H^{*}$ vanishes as expected. For $\nu_{2}$ near $-1$ which is not unrealistic (Milton, 1992) the correction is more than one third of the mean Young modulus which can no longer be neglected.
The solution of the cell problem (13) by the finite element method is quite similar. Its solution can be obtained from the previous problem if we put $\lambda(\mathbf{z})=0$. Below we present the numerical solution for the fibers with circular cross sections periodically embedded in the matrix. The unit cell is a square of length 1 with the circle of radius $r$ in the middle representing the cross section of the fiber. In this case the effective tensor is $\mu^{*}_{\alpha\beta}=\mu^{*}\delta_{\alpha\beta}$. The plot of $\mu^{*}$ as function of $s_{2}=\pi r^{2}$ is presented in Fig. 4. For the numerical simulation we choose $\mu_{1}=1\,$GPa, $\mu_{2}=2\,$GPa.
6 Solution of the homogenized cross-sectional problems
Let us first consider the homogenized cross-sectional problem (23) with the average transverse energy being given by (25). It is convenient to present the latter formula in the form
$$\bar{W}_{\perp}(\bar{\varepsilon}_{\alpha\beta},\xi)=\frac{1}{2}C^{*}_{\alpha%
\beta\gamma\delta}(\bar{\varepsilon}_{\alpha\beta}+\nu^{*}_{\alpha\beta}\xi)(%
\bar{\varepsilon}_{\gamma\delta}+\nu^{*}_{\gamma\delta}\xi)+\frac{1}{2}H^{*}%
\xi^{2},$$
(32)
where $\xi=\gamma+h\Omega_{\alpha}y_{\alpha}$. Note that the second term in (32) does not depend on $\psi_{\alpha}(\mathbf{y})$, so we need to minimize functional (23) that contains just the first term. We want to show that the minimum of the functional (23) that contains just the first term among $\psi_{\alpha}(\mathbf{y})$ satisfying the constraints (24) is equal to zero. Indeed, following the same line of reasoning as in Section 4 we choose the minimizer in the form
$$\check{\psi}_{\alpha}=-\nu^{*}_{\alpha\beta}\gamma y_{\beta}-\frac{1}{2}a_{%
\alpha\beta\gamma}(y_{\beta}y_{\gamma}-\langle y_{\beta}y_{\gamma}\rangle/|%
\bar{\mathcal{A}}|),$$
where
$$a_{\alpha\beta\gamma}=h(\nu^{*}_{\alpha\beta}\Omega_{\gamma}+\nu^{*}_{\alpha%
\gamma}\Omega_{\beta}-\nu^{*}_{\beta\gamma}\Omega_{\alpha}).$$
It is easy to check that $\check{\psi}_{\alpha}$ satisfy the constraints (24) and that the first term evaluated at these functions vanishes identically, so
$$\min_{\psi_{\alpha}(\mathbf{y})\in\eqref{eq:comp17a}}\bar{I}_{\perp}[\psi_{%
\alpha}(\mathbf{y})]=\frac{1}{2}H^{*}h^{2}\langle\xi^{2}\rangle.$$
Adding the average longitudinal energy density $h^{2}\langle W_{\parallel}\rangle=\frac{1}{2}\bar{E}\langle\xi^{2}\rangle$ to this average transverse energy density and integrating $\xi^{2}=(\gamma+h\Omega_{\alpha}y_{\alpha})^{2}$ over the cross section, we find the final energy density of extension and bending of FRC beam in the form
$$\Phi(\gamma,\Omega_{\alpha})=\frac{1}{2}(\bar{E}+H^{*})(|\mathcal{A}|\gamma^{2%
}+I_{\alpha\beta}\Omega_{\alpha}\Omega_{\beta}),$$
where
$$\bar{E}=\langle\!\langle E\rangle\!\rangle=s_{1}E_{1}+s_{2}E_{2},\quad I_{%
\alpha\beta}=h^{4}\langle y_{\alpha}y_{\beta}\rangle.$$
Thus, $H^{*}$ is the correction to the stiffnesses on extension and bending of FRC beam. To see how this correction changes the stiffnesses on extension and bending of FRC beam we plot both $\bar{E}$ and $\bar{E}+H^{*}$ as function of the volume fraction of fibers with circular cross sections placed in the middle of the quadratic periodic cell (see Fig. 5). The chosen elastic moduli are: $E_{1}=1\,$GPa, $E_{2}=2\,$GPa, $\nu_{1}=0.1$, $\nu_{2}=0.4$. The correction is about 3 percent of $\bar{E}$ in this case. Note that this correction becomes one third of $\bar{E}$ if $\nu_{2}$ is near $-1$.
We now turn to the minimization problem (14). First of all it is easy to see that the constraint $\langle\psi\rangle=0$ does not affect the minimum value of (14), because the latter is invariant with respect to the change of unknown function $\psi\to\psi+c$, with $c$ being a constant. By such the change one can always achieve the fulfillment of the constraint $\langle\psi\rangle=0$. The minimizer $\check{\psi}$ of $\bar{I}_{\angle}$ is proportional to $h\Omega$. Therefore the torsional rigidity $C$ can be calculated by
$$C=h^{4}\inf_{\bar{\psi}}\,\langle\mu^{*}\delta_{\alpha\beta}(\bar{\psi}_{|%
\alpha}-e_{\alpha\sigma}y_{\sigma})(\bar{\psi}_{|\beta}-e_{\beta\kappa}y_{%
\kappa})\rangle,$$
(33)
where $\bar{\psi}=\psi/h\Omega$. The solution to this problem is well-known for the elliptical and rectangular cross sections (Le, 1999). For the FRC beam with the elliptical cross section described by the equation
$$\displaystyle c_{\alpha\beta}y_{\alpha}y_{\beta}\leq 1,$$
where $c_{\alpha\beta}$ are the components of a positive definite symmetric second-rank tensor, the torsional rigidity reads
$$C=\frac{4\mu^{*}}{c^{2}}c_{\mu\lambda}c_{\lambda\tau}I_{\mu\tau},$$
(34)
where $I_{\mu\tau}=h^{4}\langle y_{\mu}y_{\tau}\rangle$ are the moments of inertia of the cross section.
In the co-ordinate system associated with the principal axes of the ellipse
$$\displaystyle c_{11}=\frac{h^{2}}{b_{1}^{2}},\quad c_{22}=\frac{h^{2}}{b_{2}^{%
2}},\quad c_{12}=0,$$
$$\displaystyle I_{11}=\frac{1}{4}\pi b_{1}^{3}b_{2},\quad I_{22}=\frac{1}{4}\pi
b%
_{2}^{3}b_{1},\quad I_{12}=0,$$
where $b_{1},b_{2}$ are the half-lengths of the major and minor axes. Substituting these formulas into (34) we get finally
$$\displaystyle C=\frac{\pi\mu^{*}b_{1}^{3}b_{2}^{3}}{b_{1}^{2}+b_{2}^{2}}=\frac%
{4\mu^{*}}{(I^{-1})_{\alpha\alpha}},$$
where $(I^{-1})_{\alpha\beta}$ is a tensor inverse to $I_{\alpha\beta}$.
It is easy to see that (34) is also the solution to the problem (33) for a hollow elliptical cross section
$$\lambda^{2}\leq c_{\alpha\beta}y_{\alpha}y_{\beta}\leq 1,\quad\lambda<1.$$
Similar calculations give
$$\displaystyle C=\frac{\pi\mu^{*}(1-\lambda^{4})b_{1}^{3}b_{2}^{3}}{b_{1}^{2}+b%
_{2}^{2}}.$$
For the rectangular cross section of width $a$ and height 1, where $a<1$, the solution is found in form of an infinite series (Le, 1999). The torsional rigidity is given by the formula
$$\displaystyle C=\mu^{*}ch^{4}a^{3},$$
where
$$\displaystyle c=\frac{1}{3}-\frac{64a}{\pi^{5}}\sum_{n=0}^{\infty}\frac{\tanh(%
\pi(2n+1)/2a)}{(2n+1)^{5}}.$$
For other cross sections the problem (33) can be solved by the finite element method.
7 1-D beam theory
Summarizing the results obtained in Sections 3-6, we can now reduce the 3-D problem of equilibrium of the FRC beam to the following 1-D variational problem: Minimize the energy functional
$$J[\mathbf{v},\varphi]=\int_{0}^{L}\Phi(\gamma,\Omega_{\alpha},\Omega)%
\differential{x}-\mathbf{f}\dotproduct\mathbf{v}(L)+m_{\alpha}v_{\alpha,x}(L)-%
m\varphi(L)$$
among functions $\mathbf{v}$ and $\varphi$ satisfying the kinematic boundary conditions
$$\mathbf{v}(0)=0,\quad v_{\alpha,x}(0)=0,\quad\varphi(0)=0.$$
In this functional the stored energy density $\Phi(\gamma,\Omega_{\alpha},\Omega)$ is given by
$$\Phi(\gamma,\Omega_{\alpha},\Omega)=\frac{1}{2}(\bar{E}+H^{*})(|\mathcal{A}|%
\gamma^{2}+I_{\alpha\beta}\Omega_{\alpha}\Omega_{\beta})+\frac{1}{2}C\Omega^{2},$$
while the resultant forces and moments acting at $x=L$ are equal to
$$\displaystyle\mathbf{f}$$
$$\displaystyle=\int_{\mathcal{A}}\mathbf{t}(x_{1},x_{2})\differential[2]{x},$$
$$\displaystyle m_{\alpha}$$
$$\displaystyle=\int_{\mathcal{A}}x_{\alpha}t(x_{1},x_{2})\differential[2]{x},%
\quad m=\int_{\mathcal{A}}e_{\alpha\beta}x_{\alpha}t_{\beta}(x_{1},x_{2})%
\differential[2]{x}.$$
The standard calculus of variations shows that $\mathbf{v}$ and $\varphi$ satisfy the equilibrium equations
$$T_{,x}=0,\quad M_{\alpha,xx}=0,\quad M_{,x}=0,$$
(35)
where
$$T=\frac{\partial\Phi}{\partial\gamma}=(\bar{E}+H^{*})|\mathcal{A}|\gamma,\quad
M%
_{\alpha}=\frac{\partial\Phi}{\partial\Omega_{\alpha}}=(\bar{E}+H^{*})I_{%
\alpha\beta}\Omega_{\beta},\quad M=\frac{\partial\Phi}{\partial\Omega}=C\Omega.$$
Besides, the following boundary conditions must be fulfilled at $x=L$:
$$T(L)=f,\quad M_{\alpha}(L)=m_{\alpha},\quad M_{\alpha,x}=f_{\alpha},\quad M=m.$$
(36)
Using the technique of Gamma-convergence (Braides, 2002; Milton, 2003), one can prove that the solution of (35)-(36) converges to the minimizer of (1) in the energetic norm as $h/L\to 0$ and $\epsilon=c/h\to 0$.
It is easy to extend this one-dimensional theory to the dynamics of FRC beam, where functions $\mathbf{v}$ and $\varphi$ depend on $x$ and $t$. One need just to include into the one-dimensional functional the kinetic energy density which, after the dimension reduction and homogenization in accordance with (4), takes the form
$$\Theta=\frac{1}{2}\bar{\rho}(|\mathcal{A}|\dot{\mathbf{v}}\dotproduct\dot{%
\mathbf{v}}+I_{\alpha\alpha}\dot{\varphi}^{2}),$$
(37)
with $\bar{\rho}=\langle\!\langle\rho\rangle\!\rangle=s_{1}\rho_{1}+s_{2}\rho_{2}$ being the mass density averaged over the unit cell. Hamilton’s variational principle for 1-D beam theory states that, among all admissible functions $\mathbf{v}(x,t)$ and $\varphi(x,t)$ satisfying the initial and end conditions as well as the kinematic boundary conditions, the true displacement $\check{\mathbf{v}}$ and rotation $\check{\varphi}$ are the extremal of the action functional
$$J[\mathbf{v},\varphi]=\int_{t_{0}}^{t_{1}}\int_{0}^{L}(\Theta-\Phi)%
\differential{x}\differential{t}+\int_{t_{0}}^{t_{1}}(\mathbf{f}\dotproduct%
\mathbf{v}(L)-m_{\alpha}v_{\alpha,x}(L)+m\varphi(L))\differential{t}.$$
The Euler equations become
$$\bar{\rho}|\mathcal{A}|\ddot{v}=T_{,x},\quad\bar{\rho}|\mathcal{A}|\ddot{v}_{%
\alpha}=M_{\alpha,xx},\quad\bar{\rho}I_{\alpha\alpha}\ddot{\varphi}=M_{,x}.$$
These equations are subjected to the boundary conditions (37) and the initial conditions. Similar to the static case, it can be proved that the solution of this 1-D dynamic theory converges to the solution of the 3-D theory in the limits $h/L\to 0$ and $\epsilon=c/h\to 0$ and for low frequency vibrations (Le, 1999).
8 Conclusion
It is shown in this paper that the rigorous first order approximate 1-D theory of thin FRC beams can be derived from the exact 3-D elasticity theory by the variational-asymptotic method. The developed finite element code can be used to solve cross-sectional problems with arbitrary elastic moduli of the fibers and matrix as well as arbitrary distributions and shapes of fibers. The extension of this multi-scaled asymptotic analysis to curved and naturally twisted FRC beams is straightforward (Le, 1999). As seen from (Le, 1999), the extension, bending, and torsion modes of beams are coupled in that case. The VAM combined with the finite element cross-sectional analysis can also be applied to the nonlinear FRC beams with periodic microstructure in the spirit of (Yu et al., 2002, 2005; Yu and Tang, 2007). Finally, let us mention the FRC beams with randomly distributed fibers under torsion analyzed in (Barretta et al., 2015). The analysis of extension and bending of such beams requires the solution of the plane strain problem which is quite challenging.
Appendix: Matlab-code
⬇
1function [CH,DH,FH] = homogenizecs(lx, ly, lambda, mu, phi, x)
2%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
3% lx = Unit cell length in x-direction.
4% ly = Unit cell length in y-direction.
5% lambda = Lame’s first parameter for both materials. Two entries.
6% mu = Lame’s second parameter for both materials. Two entries.
7% phi = Angle between horizontal and vertical cell wall. Degrees
8% x = Material indicator matrix. Size used to determine nelx/nely
9%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
10%% INITIALIZE
11% Compute contraction ratios
12nu = [lambda(1)/(2*(lambda(1) + mu(1)))...
13 lambda(2)/(2*(lambda(2) + mu(2)))];
14% Deduce discretization
15[nely, nelx] = size(x);
16% Stiffness matrix consists of two parts, one belonging to lambda and
17% one belonging to mu. Same goes for load vector
18dx = lx/nelx; dy = ly/nely;
19nel = nelx*nely;
20[keLambda, keMu, feLambda, feMu, feXiLambda, feXiMu] = elementMatVec(dx/2, dy/2, phi);
21% Node numbers and element degrees of freedom for full (not periodic) mesh
22nodenrs = reshape(1:(1+nelx)*(1+nely),1+nely,1+nelx);
23edofVec = reshape(2*nodenrs(1:end-1,1:end-1)+1,nel,1);
24edofMat = repmat(edofVec,1,8)+repmat([0 1 2*nely+[2 3 0 1] -2 -1],nel,1);
25%% IMPOSE PERIODIC BOUNDARY CONDITIONS
26% Use original edofMat to index into list with the periodic dofs
27nn = (nelx+1)*(nely+1); % Total number of nodes
28nnP = (nelx)*(nely); % Total number of unique nodes
29nnPArray = reshape(1:nnP, nely, nelx);
30% Extend with a mirror of the top border
31nnPArray(end+1,:) = nnPArray(1,:);
32% Extend with a mirror of the left border
33nnPArray(:,end+1) = nnPArray(:,1);
34% Make a vector into which we can index using edofMat:
35dofVector = zeros(2*nn, 1);
36dofVector(1:2:end) = 2*nnPArray(:)-1;
37dofVector(2:2:end) = 2*nnPArray(:);
38edofMat = dofVector(edofMat);
39ndof = 2*nnP; % Number of dofs
40%% ASSEMBLE STIFFNESS MATRIX
41% Indexing vectors
42iK = kron(edofMat,ones(8,1))’;
43jK = kron(edofMat,ones(1,8))’;
44% Material properties in the different elements
45lambda = lambda(1)*(x==1) + lambda(2)*(x==2);
46mu = mu(1)*(x==1) + mu(2)*(x==2);
47nu = nu(1)*(x==1) + nu(2)*(x==2);
48% The corresponding stiffness matrix entries
49sK = keLambda(:)*lambda(:).’ + keMu(:)*mu(:).’;
50K = sparse(iK(:), jK(:), sK(:), ndof, ndof);
51%% LOAD VECTORS AND SOLUTION
52% Assembly three load cases corresponding to the three strain cases
53sF = feLambda(:)*lambda(:).’+feMu(:)*mu(:).’;
54lambda_nu = lambda.*nu;
55mu_nu = mu.*nu;
56sF=[sF; feXiLambda(:)*lambda_nu(:).’ + feXiMu(:)*mu_nu(:).’];
57iF = repmat(edofMat’,4,1);
58jF = [ones(8,nel); 2*ones(8,nel); 3*ones(8,nel); 4*ones(8,nel)];
59F = sparse(iF(:), jF(:), sF(:), ndof, 4);
60% Solve (remember to constrain one node)
61chi(3:ndof,:) = K(3:ndof,3:ndof)\F(3:ndof,:);
62chi_4=chi(:,4);
63%% HOMOGENIZATION
64% The displacement vectors corresponding to the unit strain cases
65chi0 = zeros(nel, 8, 3);
66% The element displacements for the three unit strains
67chi0_e = zeros(8, 4);
68ke = keMu + keLambda; % Here the exact ratio does not matter, because
69fe = feMu + feLambda; % it is reflected in the load vector
70fe = [fe feXiLambda+feXiMu];
71chi0_e([3 5:end],:) = ke([3 5:end],[3 5:end])\fe([3 5:end],:);
72chi0_4_e=nu(:)*chi0_e(:,4)’;
73% epsilon0_11 = (1, 0, 0)
74chi0(:,:,1) = kron(chi0_e(:,1)’, ones(nel,1));
75% epsilon0_22 = (0, 1, 0)
76chi0(:,:,2) = kron(chi0_e(:,2)’, ones(nel,1));
77% epsilon0_12 = (0, 0, 1)
78chi0(:,:,3) = kron(chi0_e(:,3)’, ones(nel,1));
79CH = zeros(3);
80DH = [0; 0; 0];
81cellVolume = lx*ly;
82sumLambda = ((chi0_4_e(:,:)-chi_4(edofMat))*...
83 keLambda).*(chi0_4_e(:,:)-chi_4(edofMat));
84sumMu = ((chi0_4_e(:,:)-chi_4(edofMat))*keMu).*...
85 (chi0_4_e(:,:)-chi_4(edofMat));
86sumLambda = reshape(sum(sumLambda,2), nely, nelx);
87sumMu = reshape(sum(sumMu,2), nely, nelx);
88FH = 1/cellVolume*sum(sum(lambda.*sumLambda+mu.*sumMu));
89for i = 1:3
90 sumLambda = ((chi0_4_e(:,:)-chi_4(edofMat))*...
91 keLambda).*(chi0(:,:,i)-chi(edofMat+(i-1)*ndof));
92 sumMu = ((chi0_4_e(:,:)-chi_4(edofMat))*keMu).*...
93 (chi0(:,:,i)-chi(edofMat+(i-1)*ndof));
94 sumLambda = reshape(sum(sumLambda,2), nely, nelx);
95 sumMu = reshape(sum(sumMu,2), nely, nelx);
96 DH(i) = 1/cellVolume*sum(sum(lambda.*sumLambda+mu.*sumMu));
97 for j = 1:3
98 sumLambda = ((chi0(:,:,i) - chi(edofMat+(i-1)*ndof))*keLambda).*...
99 (chi0(:,:,j) - chi(edofMat+(j-1)*ndof));
100 sumMu = ((chi0(:,:,i) - chi(edofMat+(i-1)*ndof))*keMu).*...
101 (chi0(:,:,j) - chi(edofMat+(j-1)*ndof));
102 sumLambda = reshape(sum(sumLambda,2), nely, nelx);
103 sumMu = reshape(sum(sumMu,2), nely, nelx);
104 % Homogenized elasticity tensor
105 CH(i,j) = 1/cellVolume*sum(sum(lambda.*sumLambda + mu.*sumMu));
106 end
107end
108disp(’--- Homogenized elasticity CH ---’); disp(CH)
109disp(’--- Homogenized elasticity DH ---’); disp(DH)
110disp(’--- Homogenized elasticity FH ---’); disp(FH)
111%% COMPUTE ELEMENT STIFFNESS MATRIX AND FORCE VECTOR (NUMERICALLY)
112function [keLambda, keMu, feLambda, feMu, feXiLambda, feXiMu] = elementMatVec(a, b, phi)
113% Constitutive matrix contributions
114CMu = diag([2 2 1]); CLambda = zeros(3); CLambda(1:2,1:2) = 1;
115% Two Gauss points in both directions
116xx=[-1/sqrt(3), 1/sqrt(3)]; yy = xx;
117ww=[1,1];
118% Initialize
119keLambda = zeros(8,8); keMu = zeros(8,8);
120feLambda = zeros(8,3); feMu = zeros(8,3); feXiLambda = zeros(8,1);feXiMu = zeros(8,1);
121L = zeros(3,4); L(1,1) = 1; L(2,4) = 1; L(3,2:3) = 1;
122for ii=1:length(xx)
123for jj=1:length(yy)
124% Integration point
125x = xx(ii); y = yy(jj);
126% Differentiated shape functions
127dNx = 1/4*[-(1-y) (1-y) (1+y) -(1+y)];
128dNy = 1/4*[-(1-x) -(1+x) (1+x) (1-x)];
129% Jacobian
130J = [dNx; dNy]*[-a a a+2*b/tan(phi*pi/180) 2*b/tan(phi*pi/180)-a; ...
131-b -b b b]’;
132detJ = J(1,1)*J(2,2) - J(1,2)*J(2,1);
133invJ = 1/detJ*[J(2,2) -J(1,2); -J(2,1) J(1,1)];
134% Weight factor at this point
135weight = ww(ii)*ww(jj)*detJ;
136% Strain-displacement matrix
137G = [invJ zeros(2); zeros(2) invJ];
138dN = zeros(4,8);
139dN(1,1:2:8) = dNx;
140dN(2,1:2:8) = dNy;
141dN(3,2:2:8) = dNx;
142dN(4,2:2:8) = dNy;
143B = L*G*dN;
144% Element matrices
145keLambda = keLambda + weight*(B’ * CLambda * B);
146keMu = keMu + weight*(B’ * CMu * B);
147% Element loads
148feLambda = feLambda + weight*(B’ * CLambda * diag([1 1 1]));
149feMu = feMu + weight*(B’ * CMu * diag([1 1 1]));
150feXiLambda = feXiLambda + weight*(B’ * CLambda* [1; 1; 0]);
151feXiMu=feXiMu+ weight*(B’ * CMu* [1; 1; 0]);
152end
153end
References
Andreasen (2014)
Andreassen, E., Andreasen, C. S., 2014. How to determine composite material properties using numerical homogenization. Computational Materials Science 83, 488–495.
Barretta et al. (2015)
Barretta, R., Luciano, R., Willis, J. R., 2015. On torsion of random composite beams. Composite Structures 132, 915–922.
Berdichevsky (1979)
Berdichevsky, V. L., 1979. Variational-asymptotic method of constructing a theory of shells. Journal of Applied Mathematics and Mechanics 43 (4), 711–736.
Berdichevsky (1981)
Berdichevsky, V. L., 1981. On the energy of an elastic rod. Journal of Applied Mathematics and Mechanics 45(4), 518–529.
Berdichevsky (2009)
Berdichevsky, V. L., 2009. Variational Principles of Continuum Mechanics, vols. 1 and 2. Springer Verlag, Berlin.
Berdichevsky (2010a)
Berdichevsky, V. L., 2010a. An asymptotic theory of sandwich plates. International Journal of Engineering Science 48 (3), 383–404.
Berdichevsky (2010b)
Berdichevsky, V. L., 2010b. Nonlinear theory of hard-skin plates and shells. International Journal of Engineering Science 48 (3), 357–369.
Braides (2002)
Braides, A., 2002. Gamma-convergence for Beginners. Vol. 22. Clarendon Press.
Chandra et al. (1999)
Chandra, R., Singh, S. P., Gupta, K., 1999. Damping studies in fiber-reinforced composites–a review. Composite Structures 46(1), 41–51.
Hodges (1990)
Hodges, D. H., 1990. Review of composite rotor blade modeling. AIAA journal 28(3), 561–565.
Hodges et al. (1992)
Hodges, D. H., Atilgan, A. R., Cesnik, C. E., Fulton, M. V., 1992. On a simplified strain energy function for geometrically nonlinear behaviour of anisotropic beams. Composites Engineering 2(5-7), 513–526.
Hughes (2012)
Hughes, T. J., 2012. The finite element method: linear static and dynamic finite element analysis. Courier Corporation.
Le (1986b)
Le, K. C., 1986b. The theory of piezoelectric shells. Journal of
Applied Mathematics and Mechanics 50 (1), 98–105.
Le (1999)
Le, K. C., 1999. Vibrations of shells and rods. Springer Verlag.
Le and Yi (2016)
Le, K. C., Yi, J.-H., 2016. An asymptotically exact theory of smart sandwich
shells. International Journal of Engineering Science 106, 179–198.
Le (2017)
Le, K. C., 2017. An asymptotically exact theory of functionally graded piezoelectric shells. International Journal of Engineering Science 112, 42-62.
Lee and Hodges (2009a)
Lee, C.-Y., Hodges, D. H., 2009a. Dynamic variational-asymptotic
procedure for laminated composite shells - Part I: Low-frequency vibration
analysis. Journal of Applied Mechanics 76 (1), 011002.
Lee and Hodges (2009b)
Lee, C.-Y., Hodges, D. H., 2009b. Dynamic variational-asymptotic
procedure for laminated composite shells - Part II: High-frequency vibration
analysis. Journal of Applied Mechanics 76 (1), 011003.
Librescu and Song (2005)
Librescu, L. and Song, O., 2005. Thin-walled composite beams: theory and application (Vol. 131). Springer Science & Business Media.
Milton (1992)
Milton, G. W., 1992. Composite materials with Poisson’s ratios close to -1. Journal of the Mechanics and Physics of Solids 40(5), 1105–1137.
Milton (2003)
Milton, G.W., 2003. Theory of Composites. Cambridge Monographs on Applied and Computational Mathematics.
Muskhelishvili (2013)
Muskhelishvili, N. I., 2013. Some basic problems of the mathematical theory of elasticity. Springer Science & Business Media.
Rodriguez et al. (2001)
Rodriguez-Ramos, R., Sabina, F.J., Guinovart-Diaz, R., Bravo-Castillero, J., 2001. Closed-form expressions for the effective coefficients of a fiber-reinforced composite with transversely isotropic constituents–I. Elastic and square symmetry. Mechanics of Materials 33(4), 223–235.
Roy et al. (2007)
Roy, S., Yu, W., Han, D., 2007. An asymptotically correct classical model for smart beams. International Journal of Solids and Structures 44 (25), 8424–8439.
Sendeckyj (2016)
Sendeckyj, G. P. (ed.), 2016. Mechanics of Composite Materials: Composite Materials (Vol. 2). Elsevier.
Sigmund and Torquato (1997)
Sigmund, O., Torquato, S., 1997. Design of materials with extreme thermal expansion using a three-phase topology optimization method. Journal of the Mechanics and Physics of Solids 45 (6), 1037–1067.
Yu et al. (2002)
Yu, W., Hodges, D. H., Volovoi, V., Cesnik, C. E., 2002. On Timoshenko-like modeling of initially curved and twisted composite beams. International Journal of Solids and Structures 39(19), 5101–5121.
Yu et al. (2005)
Yu, W., Hodges, D. H., Volovoi, V. V., Fuchs, E. D., 2005. A generalized Vlasov theory for composite beams. Thin-Walled Structures 43(9), 1493–1511.
Yu and Blair (2012)
Yu, W. and Blair, M., 2012. GEBT: A general-purpose nonlinear analysis tool for composite beams. Composite Structures 94(9), 2677–2689.
Yu and Tang (2007)
Yu, W., Tang, T., 2007. Variational asymptotic method for unit cell homogenization of periodically heterogeneous materials. International Journal of Solids and Structures 44(11-12), 3738–3755. |
Thermalization in many-particle quantum walks
Dibwe Pierrot Musumbu
[email protected]
Center for Theoretical Physics of the Polish Academy of Science, Al. Lotnikow 32/46, 02-668 Warsaw, Poland
Maria Przybylska
[email protected]
Institute of Physics, University of Zielona Góra, Licealna 9, 65–417 Zielona Góra, Poland
Andrzej J. Maciejewski
[email protected]
Institute of Astronomy, University of Zielona Góra, Licealna 9, PL-65–417 Zielona Góra, Poland.
Abstract
Many-particles quantum walks of particles obeying Bose statistics moving on graphs of various topologies are introduced. A single coin tossing commands the conditional shift operation over the whole graph. Vertices particle densities, the mean values of the phase space variables, second order spatial correlation and counting statistics are evaluated and simulated. Evidence of an universal dynamics is presented.
Many-particle, bosons, thermalization, quantum walks
pacs: 03.65.Ge,02.30.Ik,42.50.Pq
I Introduction
Classical random walks have proven to be a powerful tool in both physics and mathematics through their numerous applications in
algorithmics: Markov processes, Monte Carlo simulations, etc…Recent developments of quantum computation and quantum information raises
renewed interests directed toward advantages in quantum algorithms Aharonov:01 ; Nayak:01 ; Milburn:01 ; Childs:01 and aiming to adapt random walks to the quantum world.
Problems such as bosons sampling Seshadreesan:01 , development of quantum metrology, simulations of many-body quantum systems Knill:01 , waveguides arrays Peruzzo:01 ; Hamilton:01 ; DeNicola:01 and many others, have driven the newly developed quantum walks toward the use of many quantum walkers.
In recent developments, the studies on quantum walks are also directed toward understanding the dynamics of many-particle systems on lattices
laflamme:01 ; laflamme:02 ; Mayer01 ; Xue02 ; Sheridan:01 . In that regards, studying many-particles on graph implies exploring individual vertices dynamics and how much they affect the entire system. Furthermore, looking at the evolution of both individual vertices and the entire system enables to explore how long the system can be confined in a certain regime. This is similar to measuring how fast the quantum walk spreads or how confined the quantum walk stays in a small neighborhood Aharonov:01 .
In this work we study many-particle quantum walks on closed graphs: the cyclic graph, the double hexagon graph and the Petersen graph. We simulate discrete time quantum walks on these graphs for twelve bosons on ten vertices. We focus on the universality of dynamics appearing in the counting statistics and we analyze how vertices occupation numbers evolve in time. We observe the universal behavior of the counting statistics that is independent of the types of graph and the initial conditions of the systems as long as the number of vertices of the graphs is the same and the number of quantum walkers remains unchanged.
We recall that the quantum walk is a unitary process and therefore, its probability distribution and many other observables do not converge julia:02 . Nevertheless
we must precise that a step in discrete time quantum walks is a two stage operation involving the state space and an additional degree of freedom known as the coin
degrees Childs04 . This puts a restriction on the unitarity in discrete quantum walks but it can be overcomes as suggested in Watrous:01 . Consequently, we do not expect the dynamics into our systems to converge. We control how long the quantum walks is confined on a vertex and how this affects the evolution of the entire quantum walks. We use the vertices particle distribution, the multimode phase space dynamics, the vertices counting statistics as well as the second order position correlation functions to study time evolution of our systems. In the counting statistics we
observed a change of regime.
The title of this article introduce the concept of thermalization in many-particle quantum walks.
In classical systems thermalization is associated with the time evolution of physical observables such as momentum, energy etc…
toward a Boltzmann distribution independently of theirs initial conditions. This behavior is explained by classical physics
in the following way. Almost all particle trajectories independently of theirs initial conditions, quickly begin to look alike
because theirs dynamics is governed by nonlinear equations that drive them to explore the constant-energy manifold ergodically.
However, if the classical system possesses additional first integrals that are functionally independent of the hamiltonian and
each other, then the phase curves of the system of particles are confined to a highly restricted region of
the energy manifold. Hence, the statistical predictions fail and the system does not thermalize Gallavotti:99 . But the time evolution of isolated quantum systems is linear and theirs spectra are discrete Krylov:01 . Moreover in isolated quantum systems, the conditions directing conserved quantities to provide independent constraints on the relaxation dynamics are not well understood Rigol01 ; Rigol02 ; Rigol03 . Roughly speaking,
thermalization in isolated quantum systems can be understood as a time evolution to a state that belongs to a proper subspace of the Hilbert space of the system. The thermalization of a quantum dynamics corresponds with an observable that mean values behave according to the prediction of an appropriate statistical mechanical ensemble (represented by a certain density matrix). In our simulations
of the quantum walks on the three systems, we have observed such a behavior for the particles counting statistics. We need to precise that we are analyzing the quantum thermalization with respect to the eigenstates thermalization hypothesis Deutsch ; Srednicki . This means that the dimension of the fraction of the Hilbert space containing the thermalized state remains constant in time. We want to avoid the generalities by reserving a more detailed analysis in the conclusion.
This article is planned as follows. We start by introducing the mathematical formulation of quantum walks using identical particles in the spatial field representation in order to construct the graph many-particle state (GMP state). Next we construct the conditional shift operator needed for the implementation of the shared coin’s many-particle quantum walks using the topology of the graph. After, we present the results of the simulations for many-particle quantum walks. In each case, we present the evolution of the vertices particles distributions, multimode phase space dynamics, the second order position correlation functions and the vertices counting statistics. The last section contains our conclusions and gives a short outlook.
II The graph many-particle state
Let us consider a family of operators $\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}x\mbox{\small{)}}}$ acting on the vacuum state $|\Omega\rangle$ such that they create position eigenvector $|x\rangle$:
$$\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}x\mbox{\small{)}}}|\Omega\rangle=|x\rangle.$$
(1)
This field operator creates a particle at position $x$. Similarly we have the family of operators $\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}x\mbox{\small{)}}}$ called the field annihilation operators defined as
$$\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}x\mbox{\small{)}}}|x\rangle=|\Omega\rangle.$$
(2)
In this work, we use discrete field operators therefore when considering the integration over space we will use summation over all the position $x$. We recall that the field operators satisfying a set of axioms were introduced by Jordan and Wigner Wigner01 in order to describe identical particles. Those axioms lead to two types of fields describing the two types of particles: bosons and fermions. The creation $\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}x_{\mbox{\tiny{$2$}}}\mbox{\small{)}}}$ and the annihilation $\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}x_{\mbox{\tiny{$1$}}}\mbox{\small{)}}}$
and the field operator satisfy a commutator algebra,
$$\displaystyle[\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}x_{\mbox{\tiny{$1$}}}\mbox{\small{)}}},\,\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}x_{\mbox{\tiny{$2$}}}\mbox{\small{)}}}]$$
$$\displaystyle=$$
$$\displaystyle\delta(x_{\mbox{\tiny{$1$}}}-x_{\mbox{\tiny{$2$}}}),$$
(3)
$$\displaystyle[\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}x_{\mbox{\tiny{$1$}}}\mbox{\small{)}}},\,\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}x_{\mbox{\tiny{$2$}}}\mbox{\small{)}}}]$$
$$\displaystyle=$$
$$\displaystyle 0,$$
(4)
$$\displaystyle[\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}x_{\mbox{\tiny{$1$}}}\mbox{\small{)}}},\,\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}x_{\mbox{\tiny{$2$}}}\mbox{\small{)}}}]$$
$$\displaystyle=$$
$$\displaystyle 0,$$
(5)
for bosons, and the anti commutator algebra
$$\displaystyle\{\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}x_{\mbox{\tiny{$1$}}}\mbox{\small{)}}},\,\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}x_{\mbox{\tiny{$2$}}}\mbox{\small{)}}}\}$$
$$\displaystyle=$$
$$\displaystyle\delta(x_{\mbox{\tiny{$1$}}}-x_{\mbox{\tiny{$2$}}}),$$
(6)
$$\displaystyle\{\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}x_{\mbox{\tiny{$1$}}}\mbox{\small{)}}},\,\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}x_{\mbox{\tiny{$2$}}}\mbox{\small{)}}}\}$$
$$\displaystyle=$$
$$\displaystyle 0,$$
(7)
$$\displaystyle\{\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}x_{\mbox{\tiny{$1$}}}\mbox{\small{)}}},\,\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}x_{\mbox{\tiny{$2$}}}\mbox{\small{)}}}\}$$
$$\displaystyle=$$
$$\displaystyle 0,$$
(8)
for fermions.
Suppose that we want to populate the vertex at the position $x$ with $n$ particles. Mathematically we can write
$${\mbox{\large{(}}\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}x\mbox{\small{)}}}\mbox{\large{)}}}^{n}|\Omega\rangle=\sqrt{(n_{x})!}|n_{x}\rangle,$$
(9)
where $|n_{x}\rangle$ is the eigenstate of the operator $\hat{\bf n}_{x}=\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}x\mbox{\small{)}}}\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}x\mbox{\small{)}}}$. For the moment we are only
interested by the number of particles at vertex $x$, any other property of particles such as modes are irrelevant.
Since we will shift particles between different vertices therefore in a discrete setting, the position of a vertex will be indexed ${x}_{\alpha}$ in order to distinguish different vertices. In that regards, when considering two vertices with different number of particles, we can write $|{n}_{x_{\alpha\mbox{\tiny{$1$}}}}\rangle$ and $|{n}_{x_{\alpha\mbox{\tiny{$2$}}}}\rangle$. For the sake of notations simplification we will use:
$|{n}_{\alpha_{\mbox{\tiny{$1$}}}}\rangle$ instead $|{n}_{x_{\alpha\mbox{\tiny{$1$}}}}\rangle$ and $|{n}_{\alpha_{\mbox{\tiny{$2$}}}}\rangle$ instead $|{n}_{x_{\alpha\mbox{\tiny{$2$}}}}\rangle$.
We consider $N$ indistinguishable bosons distributed on a $M$-vertices graph. The Hilbert space of such a system is spanned by
$$D\mbox{\small{(}}\mbox{\small{$N,M$}}\mbox{\small{)}}=\bigg{(}{\mbox{\small{$M+N-1$}}\atop{\mbox{\small{$N$}}}}\bigg{)}$$
(10)
vectors $|{\bf n}\rangle=|n_{1},\ldots,n_{M}\rangle=|{n}_{\alpha_{\mbox{\tiny{$1$}}}}\rangle\otimes|{n}_{\alpha_{\mbox{\tiny{$2$}}}}\rangle\otimes\ldots\otimes|{n}_{\alpha_{\mbox{\tiny{$M$}}}}\rangle$ of magnitude $N=\sum_{\alpha}^{M}n_{\alpha}$. In quantum optics and quantum matter, laser light have been used to confine cold atoms on lattices. In such a case the word configuration is used to indicate a specific way of arranging cold atoms on lattice Immanuel . We use the same word configuration to indicate the vectors $|{\bf n}\rangle$. The components of the $|{\bf n}\rangle$’s represent the position occupation numbers $n_{\alpha}$ of individual vertices. The sum of the $n_{\alpha}$’s over the entire graph is equal to the total number of quantum walkers $N$. Individual possible configuration will be denoted as $|{\bf n}_{\ell}\rangle$, where $\ell=1,\ldots,D\mbox{\small{(}}\mbox{\small{$N,M$}}\mbox{\small{)}}$.
The set of all configuration is called Hilbert space of configurations. In general, a state of a $M$-vertices graph
containing $N$ particles can be represented as
$$\big{|}\Psi\big{\rangle}=\sum_{\ell}C_{\ell}|{\bf n}_{\ell}\rangle,$$
(11)
where $|{\bf n}_{\ell}\rangle$ generate a basis of the Hilbert space of many-particles on a graph. We use Dirac bra ket notation for simplicity.
In addition, we need an auxiliary Hilbert space (the so called coin’s Hilbert space) spanned by the vectors specifying the directions of the edges connected to any given vertex.
The number of edges ending at a vertex is called the vertex degree of this vertex. The degree of a graph is the largest of its vertices degrees Reinhard:01 . For a graph of degree $d$, the
coin’s Hilbert space is spanned by the basis $\{|v_{{}_{1}}\rangle,\ldots,|v_{d}\rangle\}$. These $d$ vectors are also known as
chiralities Ambainis02 of the coin and we will elaborate more on them in the next section. The tensor product of the coin’s Hilbert space and the configurations Hilbert space is equivalent to attaching a single coin to a configuration $|{\bf n}_{\ell}\rangle$. Therefore, a single coin is shared by all particles under the fixed configuration. The GMP states in Eq. (11) augmented with the coins chiralities
becomes
$$|\Psi_{r}\rangle=\sum_{\ell}\sum_{i}\frac{C_{i\ell}^{r}}{\mathcal{K}_{r}}|v_{i},\,{\bf n}_{\ell}\rangle,$$
(12)
where $|v_{i},\,{\bf n}_{\ell}\rangle=|v_{i}\rangle\otimes|{\bf n}_{\ell}\rangle$.
The state $|\Psi_{\mbox{\tiny{r}}}\rangle$ represents the system of $N$ particles distributed on $M$ vertices of the graph, augmented with the directional amplitudes. We use the abbreviation ’GMP state’ to indicate $|\Psi_{\mbox{\tiny{r}}}\rangle$. The index $r$ indicates the time step of the quantum walks and
$\mathcal{K}_{r}$ is the normalization constant
$$[\mathcal{K}_{r}]^{2}=\sum_{\ell}\sum_{i}|C_{i\ell}^{r}|^{2}.$$
(13)
Since $|n_{\alpha}\rangle$ can be created from the vacuum $|\Omega\rangle$, see Eq. (9), we have constructed the GMP state using field operator scheme Wigner:01 .
Now we want to equip the vertices populations with some additional properties. To this end we recall that every particle
created at a given vertex is characterized by its mode ${\upeta}$. Let us consider a position field operator $\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}{x}_{\alpha}\mbox{\small{)}}}$ and the related Fock creation operator ${\hat{\bf b}}_{\upeta}^{\dagger}$ as well as the conjugated operators $\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}{x}_{\alpha}\mbox{\small{)}}}$ and ${\hat{\bf b}}_{\upeta}$:
$$\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}{x}_{\alpha}\mbox{\small{)}}}=\sum_{\upeta}\frac{e^{-i{\varphi}_{\upeta}{x}_{\alpha}}}{\sqrt{N}}{\hat{\bf b}}_{\upeta}^{\dagger}\,\,\,{\mbox{and}}\,\,\,\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}{x}_{\alpha}\mbox{\small{)}}}=\sum_{\upeta}\frac{e^{i{\varphi}_{\upeta}{x}_{\alpha}}}{\sqrt{N}}{\hat{\bf b}}_{\upeta},$$
(14)
where $\varphi_{\upeta}$ denotes the phase of a mode $\upeta$ and ${x}_{\alpha}$ is the position. The boson Fock operators obey the following commutation rules
$$\big{[}{\hat{\bf b}}_{\upeta_{\mbox{\tiny{1}}}},\,{\hat{\bf b}}^{\dagger}_{\upeta_{\mbox{\tiny{2}}}}\big{]}=\delta_{{\upeta_{\mbox{\tiny{1}}}}{\upeta_{\mbox{\tiny{2}}}}},\,\,\,\big{[}{\hat{\bf b}}_{\upeta_{\mbox{\tiny{1}}}},\,{\hat{\bf b}}_{\upeta_{\mbox{\tiny{2}}}}\big{]}=0,\,\,\,\big{[}{\hat{\bf b}}^{\dagger}_{\upeta_{\mbox{\tiny{1}}}},\,{\hat{\bf b}}^{\dagger}_{\upeta_{\mbox{\tiny{2}}}}\big{]}=0.$$
(15)
The boson Fock representation enables us to rewrite the GMP state in Eq.
(12) in multimode boson Fock representation as:
$$|\Psi_{r}\rangle=\sum_{\ell}\sum_{j}\frac{C_{j\ell}^{r}}{\mathcal{K}_{r}}|v_{j}\rangle\bigotimes_{\alpha=1}^{M}{1\over{\sqrt{(n_{\alpha})!}}}\bigg{[}\sum_{\upeta}\frac{e^{-i{\varphi}_{\upeta}{x}_{\alpha}}}{\sqrt{N}}{\hat{\bf b}}_{\upeta}^{\dagger}\bigg{]}^{n_{\alpha}}|\Omega\rangle,$$
(16)
where $|\Omega\rangle$ is the vacuum state.
III The conditional shift operator
Before entering in the design of our conditional shift operator we need to give some remarks about distinction between continuous time quantum walks and discrete time quantum walks. Continuous time quantum walks are analogous to classical diffusion process where the stochastic matrix is substituted by the adjacency matrix of the graph and the probabilities are replaced by the amplitudes. This makes the continuous quantum walks as the process described by the Schrödinger equation where the Hamiltonian is the adjacency matrix Childs04 . Since the adjacency matrix is real and symmetric, the time evolution is unitary. Moreover it was proved that there is no discrete time process that is translationally invariant on a $d$-dimensional graph Meyer:01 ; Meyer:02 . Concerning our many-particle quantum walks, the conditional shift operator is the product of the field annihilation and creation operators, the adjacency matrix and the coin operator (Hadamard operator) as we will show in this section. During the implementation of the quantum walks, the action of the annihilation and creation operators in the conditional shifting of many-particles doesn’t preserve the length of configurations and therefore destroys the unitarity. For this reason we need a step by step normalization of the GMP state. We also notice that the dimension of the Hilbert space of the discrete time quantum walks is greater than that of a continuous time quantum walks.
As it was already noted by several authors, the structure of the GMP state suggests that the operator implementation of the quantum walks
is a two stages operation. The first stage is performed using the coin’s tossing operation.
It is defined by a coin tossing operator that is unitary operator acting on the coin’s Hilbert space. After the coin tossing the particle shifting is performed with respect to the selected direction. In a
single particle quantum walk on a line, according to Nayak:01 ; julia:01 ; julia:02 , the coin tossing is performed using the Hadamard gate
$${\bf H}={1\over{\sqrt{2}}}\left(\begin{array}[]{cc}1&1\\
1&-1\end{array}\right).$$
(17)
On a line the coin Hilbert space is two dimensional (i.e. $d=2$) and spanned by vectors $|v_{1}\rangle$ and $|v_{2}\rangle$ that correspond to the left-hand side move and right-hand move respectively. Considering the vectors $|v_{1}\rangle=\big{(}{\mbox{\tiny{1}}\atop{\mbox{\tiny{0}}}}\big{)}$ and $|v_{2}\rangle=\big{(}{\mbox{\tiny{0}}\atop{\mbox{\tiny{1}}}}\big{)}$ of the coins Hilbert space, the coin tossing respectively transforms $|v_{1}\rangle$ and $|v_{2}\rangle$ in following ways
$$\displaystyle{\bf H}|v_{1}\rangle$$
$$\displaystyle=$$
$$\displaystyle{1\over{\sqrt{2}}}(|v_{1}\rangle+|v_{2}\rangle),$$
(18)
$$\displaystyle{\bf H}|v_{2}\rangle$$
$$\displaystyle=$$
$$\displaystyle{1\over{\sqrt{2}}}(|v_{1}\rangle-|v_{2}\rangle).$$
(19)
This coin tossing operation is combined with the particle shifting operation where the shift is performed using the following operator
$${\bf S}=|v_{2}\rangle\langle v_{2}|\otimes\sum_{\nu\in\mathbb{Z}}|\nu-1\rangle\langle\nu|+|v_{1}\rangle\langle v_{1}|\otimes\sum_{\nu\in\mathbb{Z}}|\nu+1\rangle\langle\nu|.$$
(20)
This kind of shifting is well constructed for a single-particle quantum walk on the line where the structure of the
graph is such that adjacent vertices are also successive vertices. In such a setup the projector
$|\nu+1\rangle\langle\nu|\equiv|\nu+1\rangle\langle 0|\cdot|0\rangle\langle\nu|$
is equivalent to the product $\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}{x}_{\nu+1}\mbox{\small{)}}}\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}{x}_{\nu}\mbox{\small{)}}}$ of field creation and annihilation operators. This line constructed scheme of coin tossing and particle shifting can be extended to other types of graphs and many-particle shifting. Bearing that in mind, we define the coin’s tossing operator ${\bf H}_{d}$ as a $d$-order Hadamard operator given by the following formula
$${\bf H}_{d}=\sum_{j,k}h_{jk}|v_{j}\rangle\langle v_{k}|\,\,\,\,\,j,\,k\,\in\,1,2,3,\ldots,d.$$
(21)
These $h_{\jmath k}$ are normalized roots of the unity of order $d$.
The operation on the coin’s Hilbert space provides the direction to the quantum walk. In addition, we need the shifting operation which implements the
movement of quantum walkers from theirs previous positions (one configuration) to their next positions (other configuration) at each step. Such an operation over the graph uses the property of
graph connectivity and the field creation and annihilation operations. In fact, particles can only be shifted between adjacent vertices, therefore
the adjacency matrix is involved in the shifting operation. For the $M$-vertices graph, the adjacency matrix
is a $M$-dimension square matrix with entries
$$\displaystyle A_{\mu\nu}$$
$$\displaystyle=$$
$$\displaystyle\begin{cases}1&~{}~{}\text{when $\mu$ and $\nu$ are connected,}\\
0&~{}~{}\text{when $\mu$ and $\nu$ are not connected.}\end{cases}$$
In addition to the adjacency matrix, the field creation and annihilation
operators interfere every time when a particle is shifted from vertex $\mu$ to vertex $\nu$. In other words, the annihilation operator acts on vertex
$\mu$ removing one occupant while the creation operator acts on vertex $\nu$ adding one more occupant.
For example for the cyclic graph in Fig.1a, the adjacency matrix is a $10\times 10$ matrix ${\bf A}=\big{(}A_{\nu\mu}\big{)}$; $\mu,\,\nu\,=\,1,\ldots,10$ of the form
$${\bf A}=\left(\begin{array}[]{ccccccc}0&1&\cdots&0&1\\
1&0&\cdots&0&0\\
\vdots&\vdots&\ddots&\vdots&\vdots\\
0&0&\cdots&0&1\\
1&0&\cdots&1&0\end{array}\right).\vspace*{-0.2cm}$$
(22)
Graphs considered in quantum walks are undirected and quantum particles on them can spread in all directions. For the cyclic graph in Fig.1a. the quantum walks on each edge $(\mu,\,\nu)$ are realized both ways clockwise and counterclockwise and this property enables to write its
adjacency matrix as the sum of its two mutually transposed components
$${\bf A}={\bf A}_{\mbox{\tiny{L}}}+{\bf A}_{\mbox{\tiny{L}}}^{\mbox{\tiny{T}}},$$
(23)
where
$${\bf A}_{\mbox{\tiny{L}}}=\left(\begin{array}[]{ccccccc}0&0&\cdots&0&1\\
1&0&\cdots&0&0\\
\vdots&\vdots&\ddots&\vdots&\vdots\\
0&0&\cdots&0&0\\
0&0&\cdots&1&0\end{array}\right).\vspace*{-0.2cm}$$
(24)
${\bf A}_{\mbox{\tiny{L}}}$ is the adjacency matrix for directed cyclic graph corresponding to counterclockwise walks on this graph.
Applying this property of the adjacency matrix to the vertex particle annihilation and creation operations enables to write the shift operator for the
cyclic graph in Fig.1a in the following form
$${\bf S}=\mathcal{C}^{\dagger}{\bf A}_{\mbox{\tiny{L}}}\mathcal{C}\otimes|v_{1}\rangle\langle v_{1}|+\mathcal{C}^{\dagger}{\bf A}_{\mbox{\tiny{L}}}^{\mbox{\tiny{T}}}\mathcal{C}\otimes|v_{2}\rangle\langle v_{2}|,$$
(25)
where:
$$\begin{cases}\mathcal{C}&=\big{(}\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}{x}_{\mbox{\tiny{$1$}}}\mbox{\small{)}}},\ldots,\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}{x}_{\mbox{\tiny{$M$}}}\mbox{\small{)}}}\big{)}^{\mbox{\tiny{T}}},\\
\mathcal{C}^{\dagger}&=\big{(}\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}{x}_{\mbox{\tiny{$1$}}}\mbox{\small{)}}},\ldots,\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}{x}_{\mbox{\tiny{$M$}}}\mbox{\small{)}}}\big{)}.\end{cases}$$
(26)
The similar decomposition of the adjacency matrix as in Eq. (23) is valid for the line graph. For two degrees undirected graphs, two directions of quantum walks correspond to two components of the adjacency matrix ${\bf A}$ in Eq. (22). In general let us consider a $d$ degrees graph, each direction induces an automorphism of the graph. The automorphism of graph are elements of the permutation group defined on the graph. On any given vertex $\mu$ we define an permutation in direction $k$, $P^{k}\mbox{\small{(}}\mu\mbox{\small{)}}\equiv\nu$. The set of all permutations $P^{k}\mbox{\small{(}}\mu\mbox{\small{)}}\equiv\nu$ over the entire graph defines a class of equivalence in the permutation group. Such a class of equivalence is a graph. The adjacency matrix $(A_{\nu\mu}^{k})$ of such a graph corresponding to direction $k$ is a component of the adjacency matrix of the graph $(A_{\nu\mu})$. For example for the Petersen graph given in Fig.1c one can define four directions i.e. $d=4$. The first direction $\mu\rightarrow\nu$ if $\mu-\nu=-1$ and the second direction $\mu\rightarrow\nu$ if $\mu-\nu=+1$. The third direction $\mu\rightarrow\nu$ if $\mu-\nu=-5$, the fourth direction $\mu\rightarrow\nu$ if $\mu-\nu=+5$. The undirected move on each direction generates two mutually transposed components of the adjacency matrix. Thus for the Petersen graph we can have four components ${\bf A}^{k}=(A_{\nu\mu}^{k})$ such that: ${\bf A}^{1}=({\bf A}^{2})^{\mbox{\tiny{T}}}$ and ${\bf A}^{3}=({\bf A}^{4})^{\mbox{\tiny{T}}}$. In general the action of the permutation group over the graph splits the adjacency matrix into $d$ components.
Therefore we generalize the shift operator given in Eq. (25) to the following form
$${\bf S}=\sum_{\mu,\nu}\sum_{k}\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}{x}_{\nu}\mbox{\small{)}}}A_{\nu\mu}^{k}\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}{x}_{\mu}\mbox{\small{)}}}\otimes|v_{k}\rangle\langle v_{k}|.$$
(27)
Here $A_{\nu\mu}^{k}$ indicates the action of the permutation group of the graph on its adjacency matrix. Such an operation associates an element $A_{\nu\mu}$ with the chirality $k$ of the coins. Consequently action of permutation group operation coordinates the choice of vertices on which the boson
annihilation and boson creation act. On the line or the cyclic graph in Fig.1a, the permutation becomes just the transposition. The product of the Hadamard operator ${\bf H}_{d}$ defined in Eq. (21) and the shift operator ${\bf S}$ in Eq.(27), gives the conditional shift operator $\widehat{\mathcal{S}}$
$$\widehat{\mathcal{S}}=\sum_{\mu,\nu}\sum_{k}\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}{x}_{\nu}\mbox{\small{)}}}A_{\nu\mu}^{k}\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}{x}_{\mu}\mbox{\small{)}}}\otimes\sum_{j}h_{jk}|v_{j}\rangle\langle v_{k}|.$$
(28)
In the case of the cyclic graph shown in Fig.1a, we have $d=2,\,\,M=10,\,\,h_{11}=h_{12}=h_{12}={1\over{\sqrt{2}}}$ and $h_{22}=-{1\over{\sqrt{2}}}$. In such a case the matrix transposition
operation is the only symmetry splitting the adjacency matrix therefore ${\bf A}_{\mbox{\tiny{L}}}=\big{(}A_{\nu\mu}^{\mbox{\tiny{1}}}\big{)}$ and ${\bf A}_{\mbox{\tiny{L}}}^{\mbox{\tiny{T}}}=\big{(}A_{\nu\mu}^{\mbox{\tiny{2}}}\big{)}$ with $\mu,\nu=1,\ldots,10$.
It follows
$$\widehat{\mathcal{S}}=\sum_{\mu,\nu=1}^{10}\sum_{k=1}^{2}\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}{x}_{\nu}\mbox{\small{)}}}A_{\nu\mu}^{k}\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}{x}_{\mu}\mbox{\small{)}}}\otimes\sum_{j=1}^{2}h_{jk}|v_{j}\rangle\langle v_{k}|.$$
(29)
Suppose that at a $r^{\mbox{\tiny{th}}}$ step such a system is represented by a GMP state in Eq.(16). Then one step later it becomes
$$|\Psi_{r+{\mbox{\tiny{1}}}}\rangle=\widehat{\mathcal{S}}|\Psi_{r}\rangle=\sum_{\ell_{\mbox{\tiny{1}}}}\sum_{j}\frac{C_{j\ell_{\mbox{\tiny{1}}}}^{r+1}}{\mathcal{K}_{r+1}}\big{|}v_{j}{\bf n}_{\ell_{\mbox{\tiny{1}}}}\big{\rangle},$$
(30)
where
$$\displaystyle C_{j\ell_{\mbox{\tiny{1}}}}^{r+1}$$
$$\displaystyle=$$
$$\displaystyle\sum_{\mu,\nu}\sum_{k}\frac{h_{jk}C_{k\ell}^{r}}{\mathcal{K}_{r}}\mbox{\large{$f$}}_{\mu\nu}^{k},$$
(31)
$$\displaystyle|{\bf n}_{\ell_{\mbox{\tiny{1}}}}\rangle$$
$$\displaystyle\sim$$
$$\displaystyle{\bf S}\,|{\bf n}_{\ell}\rangle,$$
(32)
$$\displaystyle\mbox{\large{$f$}}_{\mu\nu}^{k}$$
$$\displaystyle=$$
$$\displaystyle A_{\nu\mu}^{k}h_{jk}\sqrt{{n}_{\mu}({n}_{\nu}+1)},$$
(33)
$$\displaystyle[\mathcal{K}_{r+1}]^{2}$$
$$\displaystyle=$$
$$\displaystyle\sum_{\ell_{\mbox{\tiny{1}}}}\sum_{j}|C_{j\ell_{\mbox{\tiny{1}}}}^{r+1}|^{2}.$$
(34)
The term $\mbox{\large{$f$}}_{\mu\nu}^{k}$ contains residuals of field operators actions and depends on graph’s adjacency matrix. It shows how the graph structure is encoded in the GMP state
during the evolution. In addition, at every step the GMP state must be normalized that is why the normalization constant $\mathcal{K}_{r}$ is step $r$ dependent. Considering the GMP state obtained in Eq.(30) we observe that the conditional shifting plays the role in redefining the amplitude of the configuration over the GMP state. The triple sum in the recursion relations in Eq.(31) defines the relation between the amplitudes $C_{j\ell_{\mbox{\tiny{1}}}}^{r+1}$ and $C_{k\ell}^{r}$ during the step implementation and couples all the parent configurations ${\bf n}_{\ell}$ involved in the induction of the new configurations ${\bf n}_{\ell_{\mbox{\tiny{1}}}}$ that constitute the evolved GMP state. In other words, the whole dynamics of many-particle quantum walks becomes walks over the configurations Hilbert space.
In addition, mathematical expression of the evolution of amplitudes can be used as indicators of interfering
parent configurations dragging the system in the unexplored regions of the configurations Hilbert space or trapping the system in a specific subspace of this Hilbert space.
IV Results and conclusions
The probability of reaching a specific configuration ${\bf n}_{\ell}$ at the step $r$ is given by
$$P_{\ell}^{r}=\sum_{j}\bigg{|}\frac{C_{j\ell}^{r}}{\mathcal{K}_{r}}\bigg{|}^{2}.$$
(35)
It is also the joint probability of the $M$-tuples occupation numbers $n_{\alpha}$’s over the graph.
The $q^{\mbox{\tiny{th}}}$ moment of the occupation number on a vertex $\alpha$ is given by:
$$\langle(n_{\alpha}\mbox{\small{(}}r\mbox{\small{)}})^{q}\rangle=\sum_{\ell}(n_{\alpha})^{q}P_{\ell}^{r},$$
(36)
where $\langle\cdot\rangle$ denote the mean value.
These moments are computationally accessible for any value of $q$. Frequently only $q=1,2$ and $3$ are used. The first moment represents the vertex expected particles occupation numbers $\langle n_{\alpha}\mbox{\small{(}}r\mbox{\small{)}}\rangle$. The $\langle n_{\alpha}\mbox{\small{(}}r\mbox{\small{)}}\rangle$ are also directly calculable using the GMP state after a specific number of steps $r$ by evaluating
$$\langle n_{\alpha}\mbox{\small{(}}r\mbox{\small{)}}\rangle=\langle\Psi_{r}|\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}{x}_{\alpha}\mbox{\small{)}}}\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}{x}_{\alpha}\mbox{\small{)}}}|\Psi_{r}\rangle=\sum_{\ell}\sum_{j}\bigg{|}\frac{C_{j\ell}^{r}}{\mathcal{K}_{r}}\bigg{|}^{2}n_{\alpha}.$$
(37)
The Figs. 1(a), 1(b) and 1(c), present the expected vertices particle distribution $\langle n_{\alpha}\mbox{\small{(}}r\mbox{\small{)}}\rangle$ for a selected number of steps. We considered all three systems starting with an initial state of the form
$$|\Psi_{\mbox{\tiny{0}}}\rangle=\frac{C_{11}^{\mbox{\tiny{0}}}}{\mathcal{K}_{\mbox{\tiny{0}}}}|v_{\mbox{\tiny{1}}}{\bf n}_{\mbox{\tiny{1}}}\rangle+\frac{C_{22}^{\mbox{\tiny{0}}}}{\mathcal{K}_{\mbox{\tiny{0}}}}|v_{\mbox{\tiny{2}}}{\bf n}_{\mbox{\tiny{2}}}\rangle,$$
(38)
that is a linear combination of two configurations with the same amplitudes $\frac{C_{11}^{\mbox{\tiny{0}}}}{\mathcal{K}_{\mbox{\tiny{0}}}}={-i\over\sqrt{2}}$ and $\frac{C_{22}^{\mbox{\tiny{0}}}}{\mathcal{K}_{\mbox{\tiny{0}}}}={1\over\sqrt{2}}$. We also considered $M=10$ vertices and $N=12$ particles initially distributed on a single vertex in two initial configurations: $|{\bf n}_{\mbox{\tiny{1}}}\rangle=|0,0,N,0,0,0,0,0,0,0\rangle$ and $|{\bf n}_{\mbox{\tiny{2}}}\rangle=|0,0,0,0,N,0,0,0,0,0\rangle$.
In addition to the vertices particle density, we also evaluate the phase space evolution, the second order spatial correlations and the vertices counting statistics.
This scheme of many-particle quantum walks offers the possibility of accessing all types of correlations. The phase space is described by evaluating the position operator $\langle{x}_{\upeta}\mbox{\small{(}}r\mbox{\small{)}}\rangle=\langle\Psi_{r}|\hat{\bf x}_{\upeta}|\Psi_{r}\rangle$
$$\langle{x}_{\upeta}\mbox{\small{(}}r\mbox{\small{)}}\rangle=\sum_{\ell}\sum_{j}\bigg{|}\frac{C_{j\ell}^{r}(n_{\alpha})!}{\mathcal{K}_{r}n_{\alpha}^{n_{\alpha}}}\bigg{|}^{2}\sum_{\{{n}_{\upeta}\}}\prod_{\upeta}\cos[{\varphi}_{\upeta}{x}_{\alpha}],$$
(39)
the momentum operator $\langle{p}_{\upeta}\mbox{\small{(}}r\mbox{\small{)}}\rangle=\langle\Psi_{r}|\hat{\bf p}_{\upeta}|\Psi_{r}\rangle$
$$\langle{p}_{\upeta}\mbox{\small{(}}r\mbox{\small{)}}\rangle=\sum_{\ell}\sum_{j}\bigg{|}\frac{C_{j\ell}^{r}(n_{\alpha})!}{\mathcal{K}_{r}n_{\alpha}^{n_{\alpha}}}\bigg{|}^{2}\sum_{\{{n}_{\upeta}\}}\prod_{\upeta}\sin[{\varphi}_{\upeta}{x}_{\alpha}],$$
(40)
and the energy operator $\langle E_{\upeta}\mbox{\small{(}}r\mbox{\small{)}}\rangle=\langle\Psi_{r}|\hat{\bf E}_{\upeta}|\Psi_{r}\rangle$
$$\displaystyle\langle{E}_{\upeta}\mbox{\small{(}}r\mbox{\small{)}}\rangle$$
$$\displaystyle=$$
$$\displaystyle\sum_{\ell}\sum_{j}\bigg{|}\frac{C_{j\ell}^{r}(n_{\alpha})!}{\mathcal{K}_{r}n_{\alpha}^{n_{\alpha}}}\bigg{|}^{2}$$
(41)
$$\displaystyle\quad\times\sum_{\{{n}_{\upeta}\}}\prod_{\upeta}\bigg{(}{n}_{\upeta}+{1\over 2}-\cos[2{\varphi}_{\upeta}{x}_{\alpha}]\bigg{)},$$
calculated in each mode and at every step. These three quantities in Eqs. (39, 40 and 41) represent the phase space coordinates in each mode. We recall that these operators are respectively defined by: $\hat{\bf x}_{\upeta}={1\over\sqrt{2}}(\hat{\bf b}_{\upeta}+\hat{\bf b}_{\upeta}^{\dagger})$, $\hat{\bf p}_{\upeta}={i\over\sqrt{2}}(\hat{\bf b}_{\upeta}-\hat{\bf b}_{\upeta}^{\dagger})$, and $\hat{\bf E}_{\upeta}={1\over 2}\hat{\bf p}_{\upeta}^{\dagger}\hat{\bf p}_{\upeta}$. The constants such as the mass, the frequency and $\hbar$ are set to 1. The Figs. 1(d), 1(e) and 1(f) present the results of 400 steps simulations where the color bar is the energy $\langle{E}_{\upeta}\mbox{\small{(}}r\mbox{\small{)}}\rangle$ in Eq (41).
Looking at the position second order correlations at step $r$, we use the spacial second order correlation function given by
$$g_{(\alpha_{\mbox{\tiny{1}}},\alpha_{\mbox{\tiny{2}}},r)}^{\mbox{\tiny{(2)}}}=\frac{\langle\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}{x}_{\alpha_{\mbox{\tiny{1}}}},r\mbox{\small{)}}}\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}{x}_{\alpha_{\mbox{\tiny{2}}}},r\mbox{\small{)}}}\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}{x}_{\alpha_{\mbox{\tiny{2}}}},r\mbox{\small{)}}}\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}{x}_{\alpha_{\mbox{\tiny{1}}}},r\mbox{\small{)}}}\rangle}{\langle\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}{x}_{\alpha_{\mbox{\tiny{2}}}},r\mbox{\small{)}}}\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}{x}_{\alpha_{\mbox{\tiny{2}}}},r\mbox{\small{)}}}\rangle\langle\widehat{\mathbf{\Uppsi}}^{\dagger}{\mbox{\small{(}}{x}_{\alpha_{\mbox{\tiny{1}}}},r\mbox{\small{)}}}\widehat{\mathbf{\Uppsi}}{\mbox{\small{(}}{x}_{\alpha_{\mbox{\tiny{1}}}},r\mbox{\small{)}}}\rangle}.$$
(42)
These quantities are used to explore the influence of other vertices on a fixed vertex occupation number dynamics.
The walks on the configurations Hilbert space are controlled by vertex occupation numbers. In Figs. 2(a) – 2(l), we present the results obtained for the $g_{(\alpha_{\mbox{\tiny{1}}},\alpha_{\mbox{\tiny{2}}},r)}^{\mbox{\tiny{(2)}}}$ for 400 steps of 12 quantum walkers over the three graphs in Fig. 1 where we present
the $\mbox{30}^{\mbox{\tiny{th}}}$ step (Figs. 2(a) – 2(c)), the $\mbox{50}^{\mbox{\tiny{th}}}$ step (Figs. 2(d) – 2(f)), the $\mbox{100}^{\mbox{\tiny{th}}}$ step (Figs. 2(g) – 2(i)) and the $\mbox{400}^{\mbox{\tiny{th}}}$ step (Figs. 2(j) – 2(l)).
There are various methods of exploring the evolution of the configurations Hilbert space. In this case we are focusing on microscopic evolutions by looking at a specific vertex $\alpha$. To such a vertex we join a counting $n_{\alpha}$. This is equivalent to fine tuning the detector to record a count only when $n_{\alpha}$ walkers report on vertex $\alpha$. This counting combines the combinatoric problem of distributing of $N$ indistinguishable bosons into $M$ vertices to the probability of a given configuration as in Eq. (35). The probability of finding $n_{\alpha}$ bosons on vertex $\alpha$ considered as the combinatorial problem of distributing of $N$ indistinguishable bosons into $M$ vertices is given by: ${D\mbox{\tiny{(}}\mbox{\tiny{$N$}}-n_{\alpha},\mbox{\tiny{$M$}}\mbox{\tiny{)}}\over{MD\mbox{\tiny{(}}\mbox{\tiny{$N,M$}}\mbox{\tiny{)}}}}$ where $D\mbox{\tiny{(}}\cdot,\cdot\mbox{\tiny{)}}$ is defined in Eq.(10)
We define the probability $P_{n_{\alpha}}^{r}$ of finding $n_{\alpha_{\mbox{\tiny{1}}}}=n_{\alpha}$ particles on vertex $\alpha$ at step $r$ as:
$$P_{n_{\alpha}}^{r}=\sum_{j}\frac{(C_{j\ell}^{r})^{*}C_{j\ell}^{r}}{[\mathcal{K}_{r}]^{2}M}\frac{D\mbox{\small{(}}N-n_{\alpha},M-1\mbox{\small{)}}}{D\mbox{\small{(}}N,M\mbox{\small{)}}}.$$
(43)
Considering the evolution of vertex by vertex counting statistics we observe a change of regime in the time step evolution counting statistics. This change appears in all the three graphs but at different time steps. For example for all these three systems, when the quantum walks start from the initial conditions in Eq.(38) the change of regime occurs after $r=94$ steps for the cyclic graph in Fig. 1a, after $r=70$ steps for the double hexagon graph in Fig. 1b and after $r=48$ steps for the Petersen graph in Fig. 1c. Before the change of regime the number of the counts grows quickly specially for the small number of particles on a vertex see Figs. 3(a)–3(d), Figs. 3(b)–3(e) and Figs. 3(c)–3(f). After the change of regime the counting statistics remains consistently the same for all the three graphs and for different initial conditions. During our simulations we also monitored what fraction of the configuration Hilbert space was effectively contributing in the GMP state with nonzero amplitudes. Let us simply name it the effective Hilbert space. Moreover, the observation of regime change in the counting statistics time evolution corresponds with the change of behavior of the dimension of the effective Hilbert space. Before the change of the regime this dimension grows quickly e.g for the cyclic graph from $7900$ at step $r=30$ to $68632$ at step $r=50$ (Corresponding numbers for the double hexagon graph are: $14507$, $115052$). At the change of regime step, the dimension of the effective Hilbert space reaches one of the two critical values $146860$ or $147070$ and at the subsequent steps oscillates between these two values. We say that this behavior is universal because it is present in different types of graphs and for different initial conditions as long as the number of vertices over the considered graphs is the same and the number quantum walkers remains the same. In this work we have present only the results
for one initial condition in Eq. (38) shown in Figs. 3(a)-3(l). We observed the same phenomenon for other initial conditions. It means that the vertices counting statistics shows universal behavior after the change of regime step for each system. It must also be noted that the change of regime step depends on the initial conditions for the designated graph in Fig.1.
Phenomenologically it can be observed that during the change of regime, the systems enter into a two degrees of freedom phase. In other words, during successive steps, the GMP states in Eq. (30) related to the three graphs evolve between two subspaces of the configuration Hilbert space of dimensions 146860 and 147070, respectively. Looking at generic many-particle quantum systems, it is conjectured that thermalization is possible at the level of individual eigenstates Deutsch ; Srednicki . Rigol et al in works Rigol01 ; Rigol03 ; Rigol02 on thermalization in generic isolated quantum systems suggest that despite the fact that an isolated quantum system doesn’t thermalize like a classical system, it is possible to identify an observable for which the mean values behave such as it is predicted using an appropriate statistical mechanical ensemble (represented by a certain density matrix). Authors of Rigol02 use the evolution of particles-mode distribution $n_{\upeta}$ to monitor numerically the thermal relaxation. In this work, we have chosen the vertices counting statistics to monitor the time evolution of the system. Under the hypothesis of eigenstate thermalization, the system evolves from an initial state with non-thermalized eigenstates to a final state with thermalized eigenstates. Therefore it is expected that individual many-particle eigenstates thermalize without affecting the dimension of the effective configuration Hilbert space. Consequently, under the hypothesis of eigenstates thermalization the GMP state during the shared coins many-particle quantum walks on closed graphs evolves toward a subspace of the configuration Hilbert space called effective Hilbert space of desired dimension. However, we are observing a cyclic changes of the dimension of the effective configuration Hilbert space for systems on graphs in Figs. 1 where some eigenstates coalesce during the evolution from the effective Hilbert space of dimension $147070$ to the effective Hilbert space of dimension $146860$ and split during the reverse evolution.
Moreover, we notice that the dimension of the effective Hilbert space oscillates between values that are roughly a third of dimension of the configurations Hilbert space. Then under the hypothesis of ergodicity, there must exist additional conserved physical observables including the Hamiltonian that are functionally independent each other. In such a case the evolution of the system will be restricted to a subspace of the configurations Hilbert space and the uniform distribution for the many-particle quantum walks becomes inaccessible. Such cases have been experimentally observed like in Fermi-Pasta-Ulam numerical experiment FerUlam01 . Nevertheless, looking at the evolution of the second order spacial correlations in Figs.2(j)–2(l), one can observe that the quantum walks have a finite confinement time and they are slowly evolving toward a uniform distribution in effective Hilbert space, see Fig 1(c). In conclusion, we conjecture that these systems become integrable Toshiya01 ; JHarnard01 ; Izergin:01 after the regime change steps. To understand the limits of this conjecture, it is useful to explore other observables involved in the dynamics. Here we have in mind higher order correlations that might be of relevance as well as other observables directly connected to the microscopic evolution of the configurations Hilbert space.
This research was supported by grant No. DEC-2011/02/A/ST1/00208 of National Science Centre of Poland.
References
(1)
D. Aharonov, A. Ambainis, J. Kempe, and U. Vazirani.
Quantum walks on graphs.
In Proceedings of the Thirty-third Annual ACM Symposium on
Theory of Computing, STOC ’01, pages 50–59, New York, NY, USA, 2001. ACM.
(2)
Y. Aharonov, L. Davidovich, and N. Zagury.
Quantum random walks.
Phys. Rev. A, 48(2):1687–1690, 1993.
(3)
A. Ambainis, E. Bach, A. Nayak, A.and Vishwanath, and J. Watrous.
One-dimensional quantum walks.
In Proceedings of the Thirty-third Annual ACM Symposium on
Theory of Computing, STOC ’01, pages 37–49. ACM, 2001.
(4)
I. Bloch, J. Dalibard, and W. Zwerger.
Many-body physics with ultracold gases.
Rev. Mod. Phys., 80(3):885–964, 2008.
(5)
A. Childs and J. Goldstone.
Spatial search by quantum walks.
Phys. Rev. A, 70(2):022314, 2004.
(6)
Andrew M. Childs.
On the relationship between continuous- and discrete-time quantum
walk.
Comm. Math. Phys., 294(2):581–603, 2010.
(7)
F. De Nicola, L. Sansoni, A. Crespi, R. Ramponi, R. Osellame, V. Giovannetti,
R. Fazio, P. Mataloni, and F. Sciarrino.
Quantum simulation of bosonic-fermionic noninteracting particles in
disordered systems via a quantum walk.
Phys. Rev. A, 89(3), 2014.
(8)
J.M. Deutsch.
Quantum statistical mechanics in a closed system.
Phys. Rev. A, 43(4):2046–2049, 1991.
(9)
Reinhard Diestel.
Graph Theory. IV, volume 273 of Graduate Texts in
Mathematics.
Springer-Verlag, Heidelberg, 2012.
(10)
L. Euler.
Solutio problematis ad geometriam situs pertinentis.
Commentarii Academiae Scientiarum Imperialis Petropolitanae,
8:128–140, 1736.
(11)
E. Fermi, J. Pasta, and S. Ulam.
Studies of nonlinear problems.
Technical report, Los Alamos, 1955.
(12)
Giovanni Gallavotti.
Statistical mechanics.
Texts and Monographs in Physics. Springer-Verlag, Berlin, 1999.
A short treatise.
(13)
C.S. Hamilton, R. Kruse, L. Sansoni, C. Silberhorn, and I. Jex.
Driven quantum walks.
Phys. Rev. Lett., 113(8), 2014.
(14)
J. Harnard, G. Sadidussi, and P. P Winternitz.
Integrable systems: From classical to quantum, volume 26 of
CRM Proceedings & Lecture Notes.
AMS and Centre de Recherches Mathématiques, 2000.
(15)
A.G. Izergin and V.E. Korepin.
The quantum inverse scattering method approach to correlation
functions.
Comm. Math. Phys., 94(1):67–92, 1984.
(16)
P. Jordan and E. Wigner.
Über das Paulische äquivalenzverbot (On the Pauli
exclusion principle).
Z. Phys., 47:631, 1928.
(17)
P. Jordan and E. Wigner.
Über das Paulische äquivalenzverbot (On the Pauli
exclusion principle).
Z. Phys., 47:631, 1928.
(18)
Mayer K., Tichy M.C., Mintert F., Konrad T., and Buchleitner A.
Counting statistics of many-particle quantum walks.
Phys. Rev. A, 83(6), 2011.
(19)
J. Kempe.
Quantum random walks: an introductory overview.
Contemp. Phys., 44(4):307–327, 2003.
(20)
T. Kinoshita, T. Wenger, and D. S. Weiss.
A quantum Newton’s cradle.
Nature, 440(7086):900–903, 2006.
(21)
E. Knill.
Physics: Quantum computing.
Nature, 463(7280):441–443, 2010.
(22)
Nikolai Sergeevich Krylov.
Works on the foundations of statistical physics.
Princeton University Press, Princeton, N.J., 1979.
(23)
David A. Meyer.
From quantum cellular automata to quantum lattice gases.
J. Statist. Phys., 85(5-6):551–574, 1996.
(24)
David A. Meyer.
On the absence of homogeneous scalar unitary cellular automata.
Phys. Lett. A, 223(5):337–340, 1996.
(25)
A. Nayak and A Vishwanath.
Quantum walk on the line.
quant-ph/0010117, 0010117, 2000.
(26)
Xue P. and Sanders B.C.
Two quantum walkers sharing coins.
Phys. Rev. A, 85(2), 2012.
(27)
A. Peruzzo, M. Lobino, J.C.F. Matthews, N. Matsuda, A. Politi, K. Poulios,
X.-Q. Zhou, Y. Lahini, N. Ismail, K. W rhoff, Y. Bromberg, Y. Silberberg,
M.G. Thompson, and J.L. OBrien.
Quantum walks of correlated photons.
Science, 329(5998):1500–1503, 2010.
(28)
M. Rigol.
Breakdown of thermalization in finite one-dimensional systems.
Phys. Rev. Lett, 103(10):100403,(4pp), 2009.
(29)
M. Rigol.
Quantum quenches and thermalization in one-dimensional fermionic
systems.
Phys. Rev. A, 80(5):053607,(12pp), 2009.
(30)
M. Rigol, V. Dunjko, and M. Olshanii.
Thermalization and its mechanism for generic isolated quantum
systems.
Nature, 452(7189):854–858, 2008.
(31)
C. A. Ryan, M. Laforest, J. C. Boileau, and R. Laflamme.
Experimental implementation of a discrete-time quantum random walk on
an nmr quantum-information processor.
Phys. Rev. A, 72:062317, 2005.
(32)
K.P. Seshadreesan, J.P. Olson, K.R. Motes, P.P. Rohde, and J.P. Dowling.
Boson sampling with displaced single-photon fock states versus
single-photon-added coherent states: The quantum-classical divide and
computational-complexity transitions in linear optics.
Phys. Rev. A, 91(2), 2015.
(33)
L. Sheridan, N. Paunkovi?, Y. Omar, and S. Bose.
Discrete time quantum walk on a line with two particles.
In. J. Quantum Inf., 4(3):573–583, 2006.
(34)
M. Srednicki.
Chaos and quantum thermalization.
Phys. Rev. E, 50(2):888–901, 1994.
(35)
B. C. Travaglione and G. J. Milburn.
Implementing the quantum random walk.
Phys. Rev. A, 65(3):032310, 2002.
(36)
J. Watrous.
Quantum simulations of classical random walks and undirected graph
connectivity.
J. Comput. System Sci., 62(2):376–391, 2001. |
Uncertainty and Analyticity
Vladimir V. Kisil
School of Mathematics,
University of Leeds,
Leeds LS2 9JT,
England,
Web: http://www.maths.leeds.ac.uk/~kisilv/
[email protected]
(Date:: December 1, 2013)
Abstract.
We describe a connection between minimal uncertainty states and
holomorphy-type conditions on the images of the respective wavelet
transform. The most familiar example is the Fock–Segal–Bargmann
transform generated by the Gaussian, however, this also occurs under
more general assumptions.
Key words and phrases:Quantum mechanics, classical mechanics, Heisenberg
commutation relations, observables, Heisenberg group,
Fock–Segal–Bargmann space, $\mathrm{SU}(1,1)$, Hardy space
1991 Mathematics Subject Classification: Primary 81P05; Secondary 22E27
On leave from Odessa University.
\BibSpecAlias
incollectioninproceedings
1. Introduction
There are two and a half main examples of reproducing kernel spaces of
analytic function. One is the Fock–Segal–Bargmann (FSB) space and
others (one and a half)—the Bergman and Hardy spaces. The first
space is generated by the Heisenberg
group \citelist[Kisil11c]*§ 7.3 [Folland89]*§ 1.6,
two others—by the group $\mathrm{SU}(1,1)$ [Kisil11c]*§ 4.2 (this
explains our way of counting).
Those spaces have the following properties, which make their study
particularly pleasant and fruitful:
(i)
There is a group, which acts transitively on functions’
domain.
(ii)
There is a reproducing kernel.
(iii)
The space consists of holomorphic functions.
Furthermore, for FSB space there is the following property:
iv.
The reproducing kernel is generated by a function, which minimises
the uncertainty for coordinate and momentum observables.
It is known, that a transformation group is responsible for the
appearance of the reproducing
kernel [AliAntGaz00]*Thm. 8.1.3. This paper shows that the last
two properties are equivalent and connected to the group as well.
2. The Uncertainty Relation
In quantum mechanics [Folland89]*§ 1.1, an observable
(self-adjoint operator on a Hilbert space $\mathcal{H}{}$) $A$ produces the
expectation value $\bar{A}$ on a state (unit vector)
$\phi\in\mathcal{H}{}$ by $\bar{A}=\left\langle A\phi,\phi\right\rangle$. Then, the dispersion can
be evaluated as follow:
(1)
$$\Delta_{\phi}^{2}(A)=\left\langle(A-\bar{A})^{2}\phi,\phi\right\rangle=\left%
\langle(A-\bar{A})\phi,(A-\bar{A})\phi\right\rangle=\left\|(A-\bar{A})\phi%
\right\|^{2}.$$
The following theorem links obstructions of exact measurements with
non-commutativity of observables.
Theorem 1 (The Uncertainty relation).
If $A$ and $B$ are self-adjoint operators on a Hilbert space $\mathcal{H}{}$, then
(2)
$$\textstyle\left\|(A-a)u\right\|\left\|(B-b)u\right\|\geq\frac{1}{2}\left|\left%
\langle(AB-BA)u,u\right\rangle\right|,$$
for any $u\in\mathcal{H}{}$ from the domains of $AB$ and $BA$ and $a$,
$b\in\mathbb{R}{}$. Equality holds precisely when $u$ is a
solution of $((A-a)+\mathrm{i}r(B-b))u=0$ for some real $r$.
Proof.
The proof is well-known [Folland89]*§ 1.3, but
it is short, instructive and relevant for the following discussion,
thus we include it in full. We start from simple algebraic
transformations:
(3)
$$\displaystyle\left\langle(AB-BA)u,u\right\rangle$$
$$\displaystyle=$$
$$\displaystyle\left\langle(A-a)(B-b)-(B-b)(A-a))u,u\right\rangle$$
$$\displaystyle=$$
$$\displaystyle\left\langle(B-b)u,(A-a)u\right\rangle-\left\langle(A-a))u,(B-b)u\right\rangle$$
$$\displaystyle=$$
$$\displaystyle 2\mathrm{i}\Im\left\langle(B-b)u,(A-a)u\right\rangle$$
Then by the Cauchy–Schwartz inequality:
$$\textstyle\frac{1}{2}\left\langle(AB-BA)u,u\right\rangle\leq\left|\left\langle%
(B-b)u,(A-a)u\right\rangle\right|\leq\left\|(B-b)u\right\|\left\|(A-a)u\right\|.$$
The equality holds if and only if $(B-b)u$ and $(A-a)u$ are
proportional by a purely imaginary scalar.
∎
The famous application of the above theorem is the following
fundamental relation in quantum
mechanics. Recall [Folland89]*§ 1.2, that the one-dimensional
Heisenberg group $\mathbb{H}^{1}{}$ consists of points
$(s,x,y)\in\mathbb{R}^{3}{}$, with the group law:
(4)
$$\textstyle(s,x,y)*(s^{\prime},x^{\prime},y^{\prime})=(s+s^{\prime}+\frac{1}{2}%
(xy^{\prime}-x^{\prime}y),x+x^{\prime},y+y^{\prime}).$$
This is a nilpotent step two Lie group. By the Stone–von Neumann
theorem [Folland89]*§ 1.5, any infinite-dimensional
unitary irreducible representation of $\mathbb{H}^{1}{}$ is unitary
equivalent to the Schrödinger representation ${\rho_{\hslash}}$ in
$\mathcal{L}_{2}{}(\mathbb{R}{})$ parametrised by the Planck constant
$\hslash\in\mathbb{R}{}\setminus\{0\}$. A physically consistent form
of ${\rho_{\hslash}}$ is [Kisil10a]*(3.5):
(5)
$$[{\rho_{\hslash}}(s,x,y){f}\,](q)=e^{-2\pi\mathrm{i}\hslash(s+xy/2)-2\pi%
\mathrm{i}xq}\,{f}(q+{\hslash}y).$$
Elements of the Lie algebra $\mathfrak{h}_{1}$, corresponding to the
infinitesimal generators $X$ and $Y$ of one-parameters subgroups
$(0,t/(2\pi),0)$ and $(0,0,t)$ in $\mathbb{H}^{1}{}$, are
represented in (5) by the
(unbounded) operators $M$ and $D$ on
$\mathcal{L}_{2}{}(\mathbb{R}{})$:
(6)
$$\textstyle M=-\mathrm{i}q,\quad D=\hslash\frac{d}{dq},\quad\text{with the
commutator}\quad[M,D]=\mathrm{i}\hslash I.$$
In quantum mechanics, $f(q)\in\mathcal{L}_{2}{}(\mathbb{R}{})$ is
interpreted as a wave function of a particle, $M$ and $D$ are the
observables of coordinate and momentum.
Corollary 2 (Heisenberg–Kennard uncertainty relation).
For the coordinate $M$ and momentum $D$ observables we have the
Heisenberg–Kennard uncertainty relation:
(7)
$$\Delta_{\phi}(M)\cdot\Delta_{\phi}(D)\geq\frac{h}{2}.$$
The equality holds if and only if
$\phi(x)=e^{-cx^{2}}$, $c\in\mathbb{R}_{+}{}$ is the vacuum state
in the Schrödinger model.
Proof.
The relations follows from the commutator $[M,D]=\mathrm{i}\hslash I$,
which, in turn, is the representation of the Lie algebra
$\mathfrak{h}_{n}$ of the Heisenberg group. The minimal uncertainty
state in the Schrodinger representation is a solution of the
differential equation: $(M-\mathrm{i}rD)\phi=0$, or, explicitly:
(8)
$$(M-\mathrm{i}rD)\phi=-\mathrm{i}\left(q+r{\hslash}\frac{d}{dq}\right)\phi(q)=0.$$
The solution is the Gaussian $\phi(x)=e^{-cq^{2}}$,
$c=\frac{1}{2r\hslash}$. For $c>0$, this function is in
$\mathcal{L}_{2}{}(\mathbb{R}{})$.
∎
It is common to say that the Gaussian $\phi(q)=e^{-cq^{2}}$ represents
the ground state, which minimises the uncertainty of coordinate and
momentum.
3. Wavelet transform and analyticity
3.1. Induced wavelet transform
The following object is common in quantum
mechanics [Kisil02e], signal processing, harmonic
analysis [Kisil12d], operator theory [Kisil12b, Kisil13a]
and many other areas [Kisil11c]. Therefore, it has various
names [AliAntGaz00]: coherent states, wavelets, matrix
coefficients, etc. In the most fundamental
situation [AliAntGaz00]*Ch. 8, we start from an irreducible
unitary representation ${\rho}$ of a Lie group $G$ in a Hilbert
space $\mathcal{H}{}$. For a vector $f\in\mathcal{H}{}$ (called mother wavelet, vacuum
state, etc.), we define the map $\mathcal{W}_{f}$ from $\mathcal{H}{}$ to a space
of functions on $G$ by:
(9)
$$[\mathcal{W}_{f}v](g)=\tilde{v}(g)=\left\langle v,{\rho}(g)f\right\rangle.$$
Under the above assumptions, $\tilde{v}(g)$ is a bounded continuous
function on $G$. The map $\mathcal{W}_{f}$ intertwines ${\rho}(g)$
with the left shifts:
(10)
$$\mathcal{W}_{f}\circ{\rho}(g)=\Lambda(g)\circ\mathcal{W}_{f},\qquad\text{ %
where }\Lambda(g):\tilde{v}(g^{\prime})\mapsto\tilde{v}(g^{-1}g^{\prime}).$$
Thus, the image $\mathcal{W}_{f}\mathcal{H}{}$ is invariant under the left shifts on
$G$. If ${\rho}$ is square integrable and $f$ is
admissible [AliAntGaz00]*§ 8.1, then $\tilde{v}(g)$ is
square-integrable with respect to the Haar measure on $G$. At this
point, none of admissible vectors has an advantage over others.
It is common [Kisil11c]*§ 5.1, that there exists a subgroup
$H\subset G$ and a respective $f\in\mathcal{H}{}$ such that ${\rho}(h)f=\chi(h)f$ for some character $\chi$ of $H$. In this case, it
is enough to know values of $\tilde{v}(\mathsf{s}(x))$, for any
continuous section $\mathsf{s}$ from the homogeneous space $X=G/H$ to
$G$. The map $v\mapsto\tilde{v}(x)=\tilde{v}(\mathsf{s}(x))$
intertwines ${\rho}$ with the representation ${\rho_{\chi}}$ in a
certain function space on $X$ induced by the character $\chi$ of
$H$ [Kirillov76]*§ 13.2. We call the map $\mathcal{W}_{f}:v\mapsto\tilde{v}$ the induced wavelet
transform [Kisil11c]*§ 5.1.
For example, if $G=\mathbb{H}^{1}{}$, $H=\{(s,0,0)\in\mathbb{H}^{1}{}:\ s\in\mathbb{R}{}\}$ and its character
$\chi_{\hslash}(s,0,0)=e^{2\pi\mathrm{i}\hslash s}$, then any vector
$f\in\mathcal{L}_{2}{}(\mathbb{R}{})$ satisfies ${\rho_{\hslash}}(s,0,0)f=\chi_{\hslash}(s)f$ for the
representation (5). Thus, we still
do not have a reason to prefer any admissible vector to
others.
3.2. Right shifts and analyticity
To discover some preferable mother wavelets, we use the following a
general result from [Kisil11c]*§ 5. Let $G$ be a locally
compact group and ${\rho}$ be its representation in a Hilbert
space $\mathcal{H}{}$. Let $[\mathcal{W}_{f}v](g)=\left\langle v,{\rho}(g)f\right\rangle$ be the
wavelet transform defined by a vacuum state $f\in\mathcal{H}{}$. Then, the
right shift $R(g):[\mathcal{W}f](g^{\prime})\mapsto[\mathcal{W}f](g^{\prime}g)$ for
$g\in G$ coincides with the wavelet transform
$[\mathcal{W}_{f_{g}}](g^{\prime})=\left\langle v,{\rho}(g^{\prime})f_{g}\right\rangle$ defined by the vacuum
state $f_{g}={\rho}(g)f$. In other words, the covariant transform
intertwines right shifts on the group $G$ with the
associated action ${\rho}$ on vacuum states,
cf. (10):
(11)
$$R(g)\circ\mathcal{W}_{f}=\mathcal{W}_{{\rho}(g)f}.$$
Although, the above observation is almost trivial, applications of the following
corollary are not.
Corollary 3 (Analyticity of the wavelet transform, [Kisil11c]*§ 5).
Let $G$ be a group and $dg$ be a measure on $G$. Let
${\rho}$ be a unitary representation of $G$, which can be
extended by integration to a vector space $V$ of functions or
distributions on $G$. Let a mother wavelet $f\in\mathcal{H}{}$ satisfy the
equation
$$\int_{G}a(g)\,{\rho}(g)f\,dg=0,$$
for a fixed distribution $a(g)\in V$. Then any wavelet transform
$\tilde{v}(g)=\left\langle v,{\rho}(g)f\right\rangle$ obeys the condition:
(12)
$$D\tilde{v}=0,\qquad\text{where}\quad D=\int_{G}\bar{a}(g)\,R(g)\,dg,$$
with $R$ being the right regular representation of $G$.
Some applications (including discrete one) produced by the $ax+b$ group
can be found in [Kisil12d]*§ 6. We turn to the Heisenberg
group now.
Example 4 (Gaussian and FSB transform).
The Gaussian $\phi(x)=e^{-cq^{2}/2}$ is a null-solution of the
operator $\hslash cM-\mathrm{i}D$. For the centre $Z=\{(s,0,0):\ s\in\mathbb{R}{}\}\subset\mathbb{H}^{1}{}$, we define the section
$\mathsf{s}:\mathbb{H}^{1}{}/Z\rightarrow\mathbb{H}^{1}{}$ by
$\mathsf{s}(x,y)=(0,x,y)$. Then, the corresponding induced wavelet
transform is:
(13)
$$\tilde{v}(x,y)=\left\langle v,{\rho}(\mathsf{s}(x,y))f\right\rangle=\int_{%
\mathbb{R}{}}v(q)\,e^{\pi\mathrm{i}\hslash xy-2\pi\mathrm{i}xq}\,e^{-c(q+{%
\hslash}y)^{2}/2}\,dq.$$
The infinitesimal generators $X$ and $Y$ of one-parameters subgroups
$(0,t/(2\pi),0)$ and $(0,0,t)$ are represented through the right
shift in (4) by
$$\textstyle dR(X)=-\frac{1}{4\pi}y\partial_{s}+\frac{1}{2\pi}\partial_{x},\quad
dR%
(Y)=\frac{1}{2}x\partial_{s}+\partial_{y}.$$
For the representation induced
by the character $\chi_{\hslash}(s,0,0)=e^{2\pi\mathrm{i}\hslash s}$ we
have $\partial_{s}=2\pi\mathrm{i}\hslash I$. Cor. 3 ensures that the operator
(14)
$$\hslash c\cdot dR(X)+\mathrm{i}\cdot dR(Y)=-\frac{\hslash}{2}(2\pi x+{\mathrm{%
i}\hslash c}y)+\frac{\hslash c}{2\pi}\partial_{x}+\mathrm{i}\partial_{y}$$
annihilate any $\tilde{v}(x,y)$ from (13).
The integral (13) is known as
Fock–Segal–Bargmann (FSB) transform and in the most common case
the values $\hslash=1$ and $c=2\pi$ are used. For these,
operator (14) becomes $-\pi(x+\mathrm{i}y)+(\partial_{x}+\mathrm{i}\partial_{y})=-\pi z+2\partial_{%
\bar{z}}$ with
$z=x+\mathrm{i}y$. Then
the function $V(z)=e^{\pi z\bar{z}/2}\,\tilde{v}(z)=e^{\pi(x^{2}+y^{2})/2}\,\tilde{v}(x,y)$
satisfies the Cauchy–Riemann equation $\partial_{\bar{z}}V(z)=0$.
This example
shows, that the Gaussian is a preferred vacuum state (as producing
analytic functions through FSB transform) exactly for the same reason
as being the minimal uncertainty state: the both are derived from the
identity $(\hslash cM+\mathrm{i}D)e^{-cq^{2}/2}=0$.
3.3. Uncertainty and analyticity
The main result of this paper is a generalisation of the previous
observation, which bridge together Cor. 3 and
Thm. 1. Let $G$, $H$, ${\rho}$ and
$\mathcal{H}{}$ be
as before. Assume, that the homogeneous space
$X=G/H$ has a (quasi-)invariant measure
$d\mu(x)$ [Kirillov76]*§ 13.2. Then, for a function (or a
suitable distribution) $k$ on $X$ we can define the integrated
representation:
(15)
$${\rho}(k)=\int_{X}k(x){\rho}(\mathsf{s}(x))\,d\mu(x),$$
which is (possibly, unbounded) operators on (possibly, dense subspace
of) $\mathcal{H}{}$. In particular, $R(k)$ denotes the
integrated right shifts, for $H=\{e\}$.
Theorem 5.
Let $k_{1}$ and $k_{2}$ be two distributions on
$X$ with the respective integrated representations
${\rho}(k_{1})$ and ${\rho}(k_{1})$. The following are
equivalent:
(i)
A vector $f\in\mathcal{H}{}$ satisfies the identity
$$\Delta_{f}({\rho}(k_{1}))\cdot\Delta_{f}({\rho}(k_{2}))=\left|\left\langle[{%
\rho}(k_{1}),{\rho}(k_{1})]f,f\right\rangle\right|.$$
(ii)
The image of the wavelet transform $\mathcal{W}_{f}:v\mapsto\tilde{v}(g)=\left\langle v,{\rho}(g)f\right\rangle$ consists of functions
satisfying the equation $R(k_{1}+\mathrm{i}rk_{2})\tilde{v}=0$ for
some $r\in\mathbb{R}{}$, where $R$ is the integrated
form (15) of the right regular representation
on $G$.
Proof.
This is an immediate consequence of a combination of
Thm. 1 and Cor. 3.
∎
Example 4 is a particular case of this theorem
with $k_{1}(x,y)=\delta^{\prime}_{x}(x,y)$ and $k_{2}(x,y)=\delta^{\prime}_{y}(x,y)$
(partial derivatives of the delta function), which represent vectors
$X$ and $Y$ from the Lie algebra $\mathfrak{h}_{1}$. The next
example will be of this type too.
3.4. Hardy space
Let $\mathrm{SU}(1,1)$ be the group of $2\times 2$ complex matrices of the
form $\begin{pmatrix}\alpha&\beta\\
\bar{\beta}&\bar{\alpha}\end{pmatrix}$ with the unit determinant
$\left|\alpha\right|^{2}-\left|\beta\right|^{2}=1$. A standard basis in the
Lie algebra $\mathfrak{su}_{1,1}$ is
$$A=\frac{1}{2}\begin{pmatrix}0&-\mathrm{i}\\
\mathrm{i}&0\end{pmatrix},\quad B=\frac{1}{2}\begin{pmatrix}0&1\\
1&0\end{pmatrix},\quad Z=\begin{pmatrix}\mathrm{i}&0\\
0&-\mathrm{i}\end{pmatrix}.$$
The respective one-dimensional subgroups consist of matrices:
$$e^{tA}=\begin{pmatrix}\cosh\frac{t}{2}&-\mathrm{i}\sinh\frac{t}{2}\\
\mathrm{i}\sinh\frac{t}{2}&\cosh\frac{t}{2}\end{pmatrix},\ e^{tB}=\begin{%
pmatrix}\cosh\frac{t}{2}&\sinh\frac{t}{2}\\
\sinh\frac{t}{2}&\cosh\frac{t}{2}\end{pmatrix},\ e^{tZ}=\begin{pmatrix}e^{%
\mathrm{i}t}&0\\
0&e^{-\mathrm{i}t}\end{pmatrix}.$$
The last subgroup—the maximal compact subgroup of $\mathrm{SU}(1,1)$—is
usually denoted by $K$. The commutators of the basis elements are
(16)
$$[Z,A]=2B,\qquad[Z,B]=-2A,\qquad[A,B]=-\frac{1}{2}Z.$$
Let $\mathbb{T}{}$ denote the unit circle in $\mathbb{C}{}$ with
the rotation-invariant measure. The
mock discrete representation of $\mathrm{SU}(1,1)$ [Lang85]*§ VI.6 acts
on $\mathcal{L}_{2}{}(\mathbb{T}{})$ by unitary transformations
(17)
$$[{\rho_{1}}(g)f](z)=\frac{1}{(\bar{\beta}z+\bar{\alpha})}\,f\left(\frac{{%
\alpha}z+\beta}{\bar{\beta}z+\bar{\alpha}}\right),\qquad g^{-1}=\begin{pmatrix%
}\alpha&\beta\\
\bar{\beta}&\bar{\alpha}\end{pmatrix}.$$
The respective derived representation of the $\mathfrak{su}_{1,1}$
basis is:
$$\textstyle d{\rho^{A}_{1}}=\frac{\mathrm{i}}{2}(z+(z^{2}+1)\partial_{z}),\quad
d%
{\rho^{B}_{1}}=\frac{1}{2}(z+(z^{2}-1)\partial_{z}),\quad d{\rho^{Z}_{1}}=-%
\mathrm{i}I-2\mathrm{i}z\partial_{z}.$$
Thus, $d{\rho^{B+\mathrm{i}A}_{1}}=-\partial_{z}$ and the function
$f_{+}(z)\equiv 1$ satisfies $d{\rho^{B+\mathrm{i}A}_{1}}f_{+}=0$. Recalling
the commutator $[A,B]=-\frac{1}{2}Z$ we note that
${\rho_{1}}(e^{tZ})f_{+}=e^{\mathrm{i}t}f_{+}$. Therefore, there is the
following identity for dispersions on this state:
$$\textstyle\Delta_{f_{+}}(d{\rho^{A}_{1}})\cdot\Delta_{f_{+}}(d{\rho^{B}_{1}})=%
\frac{1}{2},$$
with the minimal value of uncertainty among all eigenvectors of the operator
${\rho_{1}}(e^{tZ})$.
Furthermore, the vacuum state $f_{+}$ generates the induced wavelet
transform for the subgroup $K=\{e^{tZ}\,\mid\,t\in\mathbb{R}{}\}$. We
identify $\mathrm{SU}(1,1)/K$ with the open unit disk $D=\{w\in\mathbb{C}{}\,\mid\,\left|w\right|<1\}$ \citelist[Kisil11c]*§ 5.5 [Kisil13a]. The map
$\mathsf{s}:\mathrm{SU}(1,1)/K\rightarrow\mathrm{SU}(1,1)$ is defined as
$\mathsf{s}(w)=\frac{1}{\sqrt{1-\left|w\right|^{2}}}\begin{pmatrix}1&w\\
\bar{w}&1\end{pmatrix}$. Then, the induced wavelet transform is:
$$\displaystyle\tilde{v}(w)=\left\langle v,{\rho_{1}}(\mathsf{s}(w))f_{+}\right\rangle$$
$$\displaystyle=\frac{1}{2\pi\sqrt{1-\left|w\right|^{2}}}\int_{\mathbb{T}{}}%
\frac{v(e^{\mathrm{i}\theta})\,d\theta}{1-we^{-\mathrm{i}\theta}}$$
$$\displaystyle=\frac{1}{2\pi\mathrm{i}\sqrt{1-\left|w\right|^{2}}}\int_{\mathbb%
{T}{}}\frac{v(e^{\mathrm{i}\theta})\,de^{\mathrm{i}\theta}}{e^{\mathrm{i}%
\theta}-w}.$$
Clearly, this is the Cauchy integral up to the factor
$\frac{1}{\sqrt{1-\left|w\right|^{2}}}$, which presents the conformal
metric on the unit disk. Similarly, we can consider the operator
$d{\rho^{B-\mathrm{i}A}_{1}}=z+z^{2}\partial_{z}$ and the function
$f_{-}(z)=\frac{1}{z}$ simultaneously solving the equations
$d{\rho^{B-\mathrm{i}A}_{1}}f_{-}=0$ and ${\rho_{1}}(e^{tZ})f_{-}=e^{-\mathrm{i}t}f_{-}$. It produces the integral with the conjugation Cauchy kernel.
Finally, we can calculate the operator (12)
annihilating the image of the wavelet transform. In the coordinates
$(w,t)\in(\mathrm{SU}(1,1)/K)\times K$, the restriction to the induced
subrepresentation is, cf. [Lang85]*§ IX.5:
$$\displaystyle\mathfrak{L}^{B-\mathrm{i}A}{}=\exp(2t\mathrm{i})(-\frac{1}{2}w+(%
1-\left|w\right|^{2})\partial_{\bar{w}}).$$
Furthermore, if $\mathfrak{L}^{B-\mathrm{i}A}{}\tilde{v}(w)=0$, then
$\partial_{\bar{w}}(\sqrt{1-w\bar{w}}\cdot\tilde{v}(w))=0$. That is,
$V(w)=\sqrt{1-w\bar{w}}\cdot\tilde{v}(w)$ is a holomorphic
function on the unit disk.
Similarly, we can treat representations of $\mathrm{SU}(1,1)$ in the space of
square integrable functions on the unit disk. The irreducible
components of this representation are isometrically
isomorphic [Kisil11c]*§ 4–5 to the weighted Bergman spaces of
(purely poly-)analytic functions on the unit, cf. [Vasilevski99].
References |
Photon-phonon-photon transfer in optomechanics
Andrey A. Rakhubovsky
Radim Filip
Department of Optics, Palacký University, 17. Listopadu 12, 771 46 Olomouc, Czech Republic
(January 14, 2021)
Abstract
We consider transfer of a highly nonclassical quantum state through an optomechanical system. That is we investigate a protocol consisting of sequential upload, storage and reading out of the quantum state from a mechanical mode of an optomechanical system. We show that provided the input state is in a test-bed single-photon Fock state, the Wigner function of the recovered state can have negative values at the origin, which is a manifest of nonclassicality of the quantum state of the macroscopic mechanical mode and the overall transfer protocol itself. Moreover, we
prove that the recovered state is quantum non-Gaussian for wide range of setup parameters. We verify that current electromechanical and optomechanical experiments can test this complete transfer of single photon.
1 Introduction
Quantum optomechanics [1, 2] can reversibly interconnect radiation to mechanical oscillators at the level of individual quanta [3, 4, 5, 6]
This connection has both fundamental and also practical consequences.
At fundamental side, it controllably converts a single photon to single phonon and then, back to individual photons.
It proves fundamental aspect that the information carried by the physical state of single quantum can be transferred between totally different physical systems.
Also, it says message to quantum thermodynamics [7], that minimal portion of energy can substantially change its form.
At practical side, such the high-fidelity photon-phonon-photon transducers can be used to sense mechanical motion at levels of single phonon level.
It can stimulate quantum metrology of mechanical motion [8] beyond the current state of the art.
Gaussian quantum entanglement [9] and squeezed states [10] between microwave radiation and mechanical oscillator have been experimentally generated and read out using resolved sideband regime.
However, Gaussian states are not sufficient resources for advanced quantum technology.
It requires non-Gaussian quantum states [11] to open full space of the applications.
Single photon state is basic non-Gaussian resource for optical technology, therefore the single phonon states are as well basic for non-Gaussian quantum optomechanics.
A mutual interconversion between a photon and a phonon is thus an important proof of this fundamental correspondence.
A single photon is simultaneously a very good test-bed to verify quantum nature of the physical interface between ligth and mechanics.
Recently, nonclassical correlations of phonon with photon at optical frequencies have been unconditionally generated [12]. This opens future possibilities to exploit optomechanics operated at single phonon level.
The transducers between optical, mechanical and microwave modes have been considered [13, 14] and realized in continuous-wave regime [15, 16, 17, 18].
However, the continuous wave regime with steady states is not suitable when quantum processing and metrology runs in time slots.
In this case, all operations are pulsed with optimized temporal shapes.
Theory proposals of photon-to-phonon transfer in the pulsed [19, 20, 21, 22] and stroboscopic regime [23] have been recently analyzed.
Also squeezed light improved read-out of mechanical state has been separately investigated [24].
In this paper, we analyze complete photon-phonon-photon coherent transfer for microwave and optical experiments in the pulsed and red side-band resolved regime.
We investigate in detail possibility to transfer the negativity of Wigner function and more general quantum non-Gaussianity through an optomechanical system in contrast to studies of engineering of a nonclassical state of a mechanical mode [25, 26].
Transfer of both negativity and non-Gaussianity is much more demanding than the transfer of Gaussian entanglement.
We verify that the electromechanical platforms [9, 27] with microwave radiation are already feasible for such transduction preserving negativity of Wigner function.
The optomechanical experiments require coupling higher than critical to reach this possibility.
To preserve quantum non-Gaussianity of single photon, optical experiment [12] is however already sufficient.
Our analysis can be extended to other systems [28].
We numerically verify all the predictions using analysis beyond adiabatic elimination of intracavity field.
We use single photon as a basic test bed for photon-phonon-photon channel, however, it verifies basic quality of coherent transfer for other non-Gaussian states.
Our results therefore open path to use photon-phonon-photon transducer for many future experiments in quantum metrology, quantum nonlinear optomechanics and quantum thermodynamics.
2 Summary of the main results
It has been shown previously [30] that in presence of a strong red-detuned coherent pulse, the quantum states of the mechanical mode of an optomechanical cavity and the resonant with the cavity pulse can in principle swapped. That is, the quantum state of an incident pulse is transferred to the mechanical mode and the state of the latter is mapped to the reflected light. The scheme analyzed here uses this state swap twice: first time to upload a highly nonclassical (single-photon) quantum state to the mechanical mode and second time to read the evolved state out. We were inspired by the quality of state read out in experiment [9].
First, the pulse in single-photon state in the optical mode described by annihilation operator $A^{\text{in}}$ is shone upon the cavity at its resonance frequency $\omega_{\mathrm{cav}}$. A strong classical coherent pulse with amplitude $\alpha_{1}$ at frequency $\omega_{\mathrm{cav}}-\omega_{m}$ enables the state swap, so that the mechanical state after the pulse reads [30, 24]
$$a_{m}(\tau)=\sqrt{T_{1}}A^{\text{in}}+\sqrt{1-T_{1}}a_{m}(0),$$
(1)
where $T_{1}=1-\exp(-2g_{0}^{2}N_{1}\tau_{1}/\kappa)$ is the transmittivity associated with the partial state swap. Here $g_{0}$ is the single-photon optomechanical coupling strength enhanced by the mean intracavity photon number $N_{1}=|\alpha_{1}|^{2}$. If the interaction time $\tau_{1}$ or optomechanical coupling $g_{0}\sqrt{N_{1}}$ is sufficient, the optical state is perfectly mapped to the mechanical mode. In practice, high optomechanical coupling may be inaccessible, or cause unwanted heating due to absorption in mirrors. The temporal duration $\tau_{1}$ of the pulse is as well limited by mechanical decoherence.
After the upload, the mechanical state is left to evolve for time $\tau_{\mathrm{s}}$. During this time the mechanical environment admixes thermal noise to the quantum state of the mechanical mode. This effect can be described by an effective transmittance $\delta=e^{-\gamma\tau_{\mathrm{s}}}$, where $\gamma$ is the mechanical damping rate. This admixture happens as well during the upload stage (and consequent readout), but we omit it for the state swap stages as this effect can be reduced by shortening the pulses durations.
Finally, another state swap is performed to read the mechanical state out. In a full analogy with Eq. 1 the state of the output optical mode reads
$$A^{\text{out}}=\sqrt{T_{2}}a_{m}(\tau_{1}+\tau_{\mathrm{s}})-\sqrt{1-T_{2}}A^{%
\text{vac}},$$
(2)
where $A^{\text{vac}}$ describes a vacuum mode incident to the cavity, $T_{2}=1-\exp(-2g_{0}^{2}N_{2}\tau_{2}/\kappa)$, $N_{2}=|\alpha_{2}|^{2}$.
The complete channel from the input optical state with annihilation operator $A^{\text{in}}$ to the output one with $A^{\text{out}}$ will be degraded by optical losses caused by e.g. imperfect coupling to the optomechanical cavity. The optical loss performs admixture of vacuum to the signal mode and is characterized by transmittivity $\eta=\kappa_{e}/\kappa$, where $\kappa$ and $\kappa_{e}$ are respectively total cavity decay rate and the cavity decay rate due to the coupling to the waveguide.
All the steps constituting the channel thus can be described by known transmittivities and therefore the quadratures of the output optical state can be expressed in terms of quadratures $Q^{\text{in}}$ of input state and quadratures of added noise $Q^{\text{N}}$
$$\displaystyle Q^{\text{out}}=\sqrt{\mathcal{T}}Q^{\text{in}}+\sqrt{1-\mathcal{%
T}}Q^{\text{N}},\quad Q=X,Y,$$
(3)
$$\displaystyle X^{k}=A^{k}{}^{\dagger}+A^{k},\quad Y^{k}=i(A^{k}{}^{\dagger}-A^%
{k}).$$
(4)
Here the transmittivity equals $\mathcal{T}=T_{1}T_{2}\eta^{2}\delta$.
The noisy mode $Q^{\text{N}}$ in the equation above comprises optical noises coming from imperfect coupling, thermal mechanical noise and the initial mechanical state. We assume all these modes to be zero-mean Gaussian noises with known variances. From the latter we construct the covariance matrix of $Q^{\text{N}}$ which is a diagonal $2\times 2$ matrix with the diagonal elements equal
$$V_{N}=1+\frac{T_{2}\eta}{1-\mathcal{T}}\Big{[}2n_{0}\delta(1-T_{1})+2n_{%
\mathrm{th}}(1-\delta)\Big{]}\ \stackrel{{\scriptstyle n_{0}=n_{\mathrm{th}}}}%
{{=}}\ 1+2n_{\mathrm{th}}\frac{\eta T_{2}(1-T_{1}\delta)}{1-\mathcal{T}}.$$
(5)
Here we assumed that the mechanical mode and its environment are initially in thermal states with occupation $n_{0}$ and $n_{\mathrm{th}}$ respectively and all the auxiliary optical modes are in vacuum.
In order to transfer negativity of the Wigner function of a single-photon state, the variance Eq. 5 should satisfy $V_{\mathrm{N}}<V_{\mathrm{N}}^{\text{neg}}\equiv\mathcal{T}/(1-\mathcal{T})$. Our goal is thus to increase the transmittivity $\mathcal{T}$ preserving moderate added noise variance $V_{\mathrm{N}}$. From the expression for $\mathcal{T}$ it is evident that it can be increased by increasing either of $T_{1}$ and $T_{2}$. However, increasing $T_{1}$ not only increases $\mathcal{T}$ but decreases $V_{\mathrm{N}}$ by diminishing the ratio of the initial thermal mechanical state in the output. Increase of $T_{2}$ is not so crucially important and it causes increase of both $\mathcal{T}$ and $V_{\mathrm{N}}$. These simple considerations can be proven by calculating partial derivatives
$$\displaystyle\left(\frac{\partial V_{\mathrm{N}}}{\partial\mathcal{T}}\right)_%
{T_{2}}$$
$$\displaystyle=-2n_{0}\frac{1-\delta\eta^{2}T_{2}}{\eta(1-\mathcal{T})^{2}}+2n_%
{\mathrm{th}}\frac{(1-\delta)\eta T_{2}}{(1-\mathcal{T})^{2}}\ \stackrel{{%
\scriptstyle n_{0}=n_{\mathrm{th}}}}{{=}}\ -2n_{\mathrm{th}}\frac{1-\eta^{2}T_%
{2}}{\eta(1-\mathcal{T})^{2}}\leqslant 0,$$
(6)
$$\displaystyle\left(\frac{\partial V_{\mathrm{N}}}{\partial\mathcal{T}}\right)_%
{T_{1}}$$
$$\displaystyle=2n_{0}\frac{1-T_{1}}{\eta(1-\mathcal{T})^{2}T_{1}}+2n_{\mathrm{%
th}}\frac{1-\delta}{\delta\eta(1-\mathcal{T})^{2}T_{1}}\ \stackrel{{%
\scriptstyle n_{0}=n_{\mathrm{th}}}}{{=}}\ 2n_{\mathrm{th}}\frac{1-T_{1}\delta%
}{T_{1}\eta\delta(1-\mathcal{T})^{2}}\geqslant 0.$$
(7)
The principal dependence of $V_{\mathrm{N}}$ on $\mathcal{T}$ is shown for different transmittances $T_{1,2}$ in Fig. 2. The mechanical mode is considered to be at equilibrium with the environment at $n_{0}=n_{\mathrm{th}}=1$. Even this low occupation creates a demand for a very efficient state upload, so that increasing $T_{1}$ with constant $T_{2}$ helps to suppress the added noise dramatically (dashed lines). It is worth noting that for certain values of $T_{1}$ the successful transfer of the negativity is impossible regardless of $T_{2}$. Note, it is evident that is much easier to reach entanglement preserving channel than to transfer negative Wigner function. From the Fig. 2 that shows the analysis following the full solution (see Section 3) we conclude that the current electromechanical setups [9, 27] are capable of the transfer of the negativity of the Wigner function provided there is a single-photon Fock state at input.
A major impediment to the successful photon-phonon-photon transfer preserving the negativity of Wigner function is the initial occupation of the mechanical mode. The influence of the mechanical bath during the state swap stages can in principle be suppressed by taking shorter pulses, however, the nonzero initial occupation has to be compensated for either by precooling of the mechanical mode or by a very efficient upload. In spite of our considered experimental scheme the precooling happens to be an additional resource that can relax the requirements to the upload of the quantum state stage. The considerations without the additional precooling therefore happen to be more conservative as such regime is more challenging experimentally and it requires the upload efficiencies close to unity. These efficiencies, however, are realistic at least in the domain of state-of-the-art electromechanics [9]. There is therefore an open way to experimentally test photon-phonon-photon channel preserving negativity of Wigner function.
We should note that in order to transmit negativity of the Wigner function, one should have $\mathcal{T}>1/2$ which places restrictions on each of individual transmittances, particularly on the efficiency of the coupling, $\eta$. That is, critically coupled systems with $\eta=0.5$ are incapable of transferring the negativity regardless of the effectiveness of state upload and readout. For these systems we consider another measure of nonclassicality, namely the quantum non-Gaussianity. A convenient criterion [31, 32] allows to determine whether a state can not be represented as a mixture of Gaussian states. The criterion establishes an upper boundary for the probability of the single photon state $p_{1}$ for a fixed probability of vacuum $p_{0}$. This boundary can be parametrized as follows:
$$p_{0}=\frac{e^{-e^{r}\sinh r}}{\cosh r},\quad p_{1,G}=\frac{e^{4r}-1}{4}\frac{%
e^{-e^{r}\sinh r}}{\cosh^{3}r},$$
(8)
with $r\in[0,+\infty)$. To evaluate the non-Gaussian properties of the restored state of the field we therefore need to evaluate the probabilities $p_{0}$ and $p_{1}$ corresponding to this state and certify if they belong to the region accessible to Gaussian states. The latter is shaded at Fig. 3. Experimentally both probabilities can be determined from homodyne detection scheme used to verify negativity of Wigner function [31]. Alternatively, single photon counters can be used [33].
In order to evaluate $p_{0}$ and $p_{1}$, we use our estimates for the transmittance $\mathcal{T}$ and added noise $V_{\mathrm{N}}$ to calculate the Wigner function of the recovered state, from which we find the needed probabilities. Results of this calculation are visualized for recent electro- and optomechanical experiments at Fig. 3. The demonstration of quantum non-Gaussianity transferred by a photon-phonon-photon channel is feasible with both platforms. The required efficiencies, particularly ones of the readout, are significantly lower for the electromechanics than that needed for the transfer of negativity of Wigner function. Note that for the weakly coupled optomechanical system, the transfer of negativity is impossible regardless of the efficiencies of the state swap. Due to the initial occupation of the mechanical modes a very efficient state upload is required.
3 Methods
For a rigorous analysis that accounts for the imperfections properly, we start with the linearised Heisenberg-Langevin equations for the optomechanical system with a strong resonantly red-detuned pump. In terms of annihilation operators $a_{c}$ and $a_{m}$ of optical and mechanical modes respectively the equations read [24]:
$$\displaystyle\dot{a}_{c}$$
$$\displaystyle=-\kappa a_{c}-iga_{m}+\sqrt{2\kappa}\left(\sqrt{\eta}a^{\text{in%
}}(t)+\sqrt{1-\eta}a^{\text{vac}}(t)\right),$$
(9)
$$\displaystyle\dot{a}_{m}$$
$$\displaystyle=-\frac{\gamma}{2}a_{m}-iga_{c}+\sqrt{\gamma}a^{\text{th}}(t).$$
(10)
Here $\gamma$ is mechanical viscous damping coefficient, $a^{\text{in}}$ and $a^{\text{vac}}$ are the input and auxiliary vacuum optical modes respectively, and $a^{\text{th}}$ describes the Markovian noises from the mechanical environment. The coupling equals $g=g_{k}\equiv g_{0}\sqrt{N_{k}}$ for $k$-th pulse and $g=0$ during the storage between the pulses. With help of the input-output relation
$$a_{2}^{\text{out}}(t)=-a_{2}^{\text{in}}(t)+\sqrt{2\kappa_{e}}a_{c}(t),$$
(11)
we are able to construct the relation connecting the bosonic operator of the input optical mode $A^{\text{in}}$ with the output $A^{\text{out}}$ in analogy with Eq. 3.
The bosonic operators are defined as
$$A^{\text{k}}=\dfrac{\int_{0}^{\tau}a^{\text{k}}(t)f^{\text{k}}(t)dt}{\sqrt{%
\int_{0}^{\tau}|f^{\text{k}}(t)|^{2}dt}},\quad\text{ k = in,out }.$$
(12)
The pulse shapes $f^{\text{in}}(t)$ and $f^{\text{out}}(t)$ are determined from the Eqs. 9 and 10. Generally expressions for these functions are quite complicated hence we do not write those explicitly here. In the limit of high $\kappa\gg g,\gamma$, where the cavity mode can be approximately eliminated [30],
$$A^{\text{in}}=\sqrt{\frac{2G_{1}}{e^{2G_{1}\tau_{1}}-1}}\int_{\tau_{1}}\mathrm%
{d}t\>a_{1}^{\text{in}}(t)e^{G_{1}t},\quad A^{\text{out}}=\sqrt{\frac{2G_{2}}{%
1-e^{-2G_{2}\tau_{2}}}}\int_{\tau_{2}}\mathrm{d}t\>a_{2}^{\text{out}}(t)e^{-G_%
{2}t},$$
(13)
with $G_{i}=g_{0}^{2}N_{i}/\kappa$.
Knowing the statistics of each individual constituent of the noisy mode $Q^{\text{N}}$, we can evaluate the variance of the added noise: $\text{var}\>{X^{\text{N}}}=\text{var}\>{Y^{\text{N}}}=V_{\mathrm{N}}$ e.g., following the steps outlined in [34], Appendix B. It is visualized as a function of the transmittivity $V_{\mathrm{N}}(\mathcal{T})$ in Fig. 2.
The effective transmittivities $T_{1}$ and $T_{2}$ associated with respectively upload and readout stages present a convenient way to parametrize this dependence. Each of these transmittivities can be increased via increase of either corresponding pump power (intracavity photon number) or pulse duration. The latter way is accompanied by increase of the mechanical environment impact. In the Fig. 2 we assume the coupling close to the highest reported and change $\tau_{1,2}$ to produce change of $T_{1,2}$. We use the following dimensionless parameters to produce the plots: for the blue lines (as reported in [9]): $\tau_{1}=\tau_{2}\leq 100/\kappa$, $g_{1}=g_{2}=0.25\kappa$, $\tau_{s}=3/\kappa$, $\gamma=2\times 10^{-4}\kappa$, $n_{\mathrm{th}}=n_{0}=20$, $\eta=0.83$. For the green lines (following [27]) $\tau_{1}=\tau_{2}\leq 10^{3}/\kappa$, $g_{1}=g_{2}=0.07\kappa$, $\tau_{s}=100/\kappa$, $\gamma=2\times 10^{-5}\kappa$, $n_{\mathrm{th}}=n_{0}=18$, $\eta=0.91$.
Evaluation of the transfer of quantum nongaussianity requires computation of the probability of vacuum and of single photon. If the restored optical quantum state is $\rho_{f}$, these probabilities equal $p_{0}=\langle 0|\rho_{f}|0\rangle$ and $p_{1}=\langle 1|\rho_{f}|1\rangle$ and can be determined by homodyne tomography [31]. The quantum state $\rho_{f}$ is found as a result of a single photon $\left|1\right\rangle$ passing a virtual beamsplitter with transmittance $\mathcal{T}$ that adds noise with variance $V_{\mathrm{N}}$. The curves in Fig. 3 are produced for same numerical parameters as Fig. 2. To produce Fig. 3 we use parameters reported in [12], $\tau_{1}\leq 3\times 10^{3}/\kappa$, $\tau_{2}\leq 2\times 10^{3}/\kappa$, $g_{1}=g_{2}=0.04\kappa$, $\tau_{s}=100/\kappa$, $\gamma=7\times 10^{-6}\kappa$, $n_{th}=n_{0}=1$, $\eta=0.5$ and $\eta=0.25$ respectively for yellow and purple lines.
Acknowledgements
The authors acknowledge finanacial support from Grant No. GB14-36681G of the Czech Science Foundation. A.R acknowledges institutional support of the Faculty of Science, Palacký University.
Author contributions
Both A.R. and R.F. contributed jointly to theory. R.F. psoposed the theoretical concept. A.R. carried out the detailed calculations. Both authors contributed to the interpretation of the work and the writing of the manuscript.
Additional Information
The authors declare no competing financial interests.
References
[1]
Aspelmeyer, M., Kippenberg, T. J. &
Marquardt, F.
Cavity optomechanics.
Reviews of Modern Physics
86, 1391–1452
(2014).
URL http://link.aps.org/doi/10.1103/RevModPhys.86.1391.
DOI 10.1103/RevModPhys.86.1391.
[2]
Khalili, F. Y. & Danilishin, S. L.
Quantum Optomechanics.
In Visser, T. D. (ed.)
Progress in Optics, vol. 61,
113–236 (Elsevier,
2016).
URL http://www.sciencedirect.com/science/article/pii/S0079663815000244.
[3]
Thompson, J. D. et al.
Strong dispersive coupling of a high-finesse cavity
to a micromechanical membrane.
Nature 452,
72–75 (2008).
URL http://www.nature.com/nature/journal/v452/n7183/full/nature06715.html.
DOI 10.1038/nature06715.
[4]
Gröblacher, S., Hammerer, K.,
Vanner, M. R. & Aspelmeyer, M.
Observation of strong coupling between a
micromechanical resonator and an optical cavity field.
Nature 460,
724–727 (2009).
URL http://www.nature.com/nature/journal/v460/n7256/abs/nature08171.html.
DOI 10.1038/nature08171.
[5]
Verhagen, E., Deléglise, S.,
Weis, S., Schliesser, A. &
Kippenberg, T. J.
Quantum-coherent coupling of a mechanical oscillator
to an optical cavity mode.
Nature 482,
63–67 (2012).
URL http://www.nature.com/nature/journal/v482/n7383/full/nature10787.html.
DOI 10.1038/nature10787.
[6]
Cohen, J. D. et al.
Phonon counting and intensity interferometry of a
nanomechanical resonator.
Nature 520,
522–525 (2015).
URL http://www.nature.com/nature/journal/v520/n7548/full/nature14349.html.
DOI 10.1038/nature14349.
ArXiv: 1410.1047.
[7]
Vinjanampathy, S. & Anders, J.
Quantum Thermodynamics.
Contemporary Physics
57, 545–579
(2016).
URL http://arxiv.org/abs/1508.06099.
DOI 10.1080/00107514.2016.1201896.
ArXiv: 1508.06099.
[8]
Schnabel, R.
Squeezed states of light and their applications in
laser interferometers.
arXiv:1611.03986 [quant-ph]
(2016).
URL http://arxiv.org/abs/1611.03986.
ArXiv: 1611.03986.
[9]
Palomaki, T. A., Teufel, J. D.,
Simmonds, R. W. & Lehnert, K. W.
Entangling Mechanical Motion with Microwave
Fields.
Science 342,
710–713 (2013).
URL http://www.sciencemag.org/content/342/6159/710.
DOI 10.1126/science.1244563.
[10]
Pirkkalainen, J.-M., Damskägg, E.,
Brandt, M., Massel, F. &
Sillanpää, M. A.
Squeezing of Quantum Noise of Motion in a
Micromechanical Resonator.
Physical Review Letters
115, 243601
(2015).
URL http://link.aps.org/doi/10.1103/PhysRevLett.115.243601.
DOI 10.1103/PhysRevLett.115.243601.
ArXiv: 1507.04209.
[11]
Mari, A. & Eisert, J.
Positive Wigner Functions Render Classical
Simulation of Quantum Computation Efficient.
Physical Review Letters
109, 230503
(2012).
URL http://link.aps.org/doi/10.1103/PhysRevLett.109.230503.
DOI 10.1103/PhysRevLett.109.230503.
ArXiv: 1208.3660.
[12]
Riedinger, R. et al.
Non-classical correlations between single photons and
phonons from a mechanical oscillator.
Nature 530,
313–316 (2016).
URL http://www.nature.com/nature/journal/v530/n7590/full/nature16536.html#affil-auth.
DOI 10.1038/nature16536.
ArXiv: 1512.05360.
[13]
Safavi-Naeini, A. H. & Painter, O.
Proposal for an optomechanical traveling wave
phonon–photon translator.
New Journal of Physics
13, 013017
(2011).
URL http://stacks.iop.org/1367-2630/13/i=1/a=013017.
DOI 10.1088/1367-2630/13/1/013017.
ArXiv: 1009.3529.
[14]
Tian, L.
Adiabatic State Conversion and Pulse
Transmission in Optomechanical Systems.
Physical Review Letters
108, 153604
(2012).
URL http://link.aps.org/doi/10.1103/PhysRevLett.108.153604.
DOI 10.1103/PhysRevLett.108.153604.
[15]
Hill, J. T., Safavi-Naeini, A. H.,
Chan, J. & Painter, O.
Coherent optical wavelength conversion via cavity
optomechanics.
Nature Communications
3, 1196 (2012).
URL http://www.nature.com/ncomms/journal/v3/n11/abs/ncomms2201.html.
DOI 10.1038/ncomms2201.
ArXiv: 1206.0704.
[16]
Andrews, R. W. et al.
Bidirectional and efficient conversion between
microwave and optical light.
Nature Physics
10, 321–326
(2014).
URL http://www.nature.com/nphys/journal/v10/n4/full/nphys2911.html.
DOI 10.1038/nphys2911.
ArXiv: 1310.5276.
[17]
Andrews, R. W., Reed, A. P.,
Cicak, K., Teufel, J. D. &
Lehnert, K. W.
Quantum-enabled temporal and spectral mode conversion
of microwave signals.
Nature Communications
6, 10021 (2015).
URL http://www.nature.com/ncomms/2015/151130/ncomms10021/full/ncomms10021.html.
DOI 10.1038/ncomms10021.
ArXiv: 1506.02296.
[18]
Lecocq, F., Clark, J.,
Simmonds, R., Aumentado, J. &
Teufel, J.
Mechanically Mediated Microwave Frequency
Conversion in the Quantum Regime.
Physical Review Letters
116, 043601
(2016).
URL http://link.aps.org/doi/10.1103/PhysRevLett.116.043601.
DOI 10.1103/PhysRevLett.116.043601.
ArXiv: 1512.00078.
[19]
Khalili, F. et al.
Preparing a Mechanical Oscillator in
Non-Gaussian Quantum States.
Physical Review Letters
105, 070403
(2010).
URL http://link.aps.org/doi/10.1103/PhysRevLett.105.070403.
DOI 10.1103/PhysRevLett.105.070403.
[20]
McGee, S. A., Meiser, D.,
Regal, C. A., Lehnert, K. W. &
Holland, M. J.
Mechanical resonators for storage and transfer of
electrical and optical quantum states.
Physical Review A
87, 053818
(2013).
URL http://link.aps.org/doi/10.1103/PhysRevA.87.053818.
DOI 10.1103/PhysRevA.87.053818.
ArXiv: 1305.6962.
[21]
Rakhubovsky, A. A., Vostrosablin, N. &
Filip, R.
Squeezer-based pulsed optomechanical interface.
Physical Review A
93, 033813
(2016).
URL http://link.aps.org/doi/10.1103/PhysRevA.93.033813.
DOI 10.1103/PhysRevA.93.033813.
ArXiv: 1511.08611.
[22]
Bennett, J. S. et al.
A quantum optomechanical interface beyond the
resolved sideband limit.
New Journal of Physics
18, 053030
(2016).
URL http://stacks.iop.org/1367-2630/18/i=5/a=053030.
DOI 10.1088/1367-2630/18/5/053030.
ArXiv: 1510.05368.
[23]
Hoff, U. B., Kollath-Bönig, J.,
Neergaard-Nielsen, J. S. & Andersen,
U. L.
Measurement-Induced Macroscopic Superposition
States in Cavity Optomechanics.
Physical Review Letters
117, 143601
(2016).
URL http://link.aps.org/doi/10.1103/PhysRevLett.117.143601.
DOI 10.1103/PhysRevLett.117.143601.
ArXiv: 1601.01663.
[24]
Filip, R. & Rakhubovsky, A. A.
Transfer of non-Gaussian quantum states of
mechanical oscillator to light.
Physical Review A
92, 053804
(2015).
URL http://link.aps.org/doi/10.1103/PhysRevA.92.053804.
DOI 10.1103/PhysRevA.92.053804.
[25]
Rogers, B., Gullo, N. L.,
De Chiara, G., Palma, G. M. &
Paternostro, M.
Hybrid optomechanics for Quantum Technologies.
Quantum Measurements and Quantum Metrology
2, 11–43 (2014).
URL http://arxiv.org/abs/1402.1195.
DOI 10.2478/qmetro-2014-0002.
ArXiv: 1402.1195.
[26]
Teklu, B., Ferraro, A.,
Paternostro, M. & Paris, M. G. A.
Nonlinearity and nonclassicality in a nanomechanical
resonator.
EPJ Quantum Technology
2, 16 (2015).
URL http://epjquantumtechnology.springeropen.com/articles/10.1140/epjqt/s40507-015-0029-x.
DOI 10.1140/epjqt/s40507-015-0029-x.
ArXiv: 1501.03767.
[27]
Ockeloen-Korppi, C. F. et al.
Low-noise amplification and frequency conversion with
a multiport microwave optomechanical device.
arXiv:1602.05779 [cond-mat,
physics:quant-ph] (2016).
URL http://arxiv.org/abs/1602.05779.
ArXiv: 1602.05779.
[28]
Xia, K., Vanner, M. R. &
Twamley, J.
An opto-magneto-mechanical quantum interface between
distant superconducting qubits.
Scientific Reports
4, 5571 (2014).
URL http://www.nature.com/srep/2014/140704/srep05571/full/srep05571.html.
DOI 10.1038/srep05571.
[29]
Chan, J. et al.
Laser cooling of a nanomechanical oscillator into its
quantum ground state.
Nature 478,
89–92 (2011).
URL http://arxiv.org/abs/1106.3614.
DOI 10.1038/nature10461.
ArXiv:1106.3614 [quant-ph].
[30]
Hofer, S. G., Wieczorek, W.,
Aspelmeyer, M. & Hammerer, K.
Quantum entanglement and teleportation in pulsed
cavity optomechanics.
Physical Review A
84, 052327
(2011).
URL http://link.aps.org/doi/10.1103/PhysRevA.84.052327.
DOI 10.1103/PhysRevA.84.052327.
[31]
Filip, R. & Mišta, L.
Detecting Quantum States with a Positive
Wigner Function beyond Mixtures of Gaussian States.
Physical Review Letters
106, 200401
(2011).
URL http://link.aps.org/doi/10.1103/PhysRevLett.106.200401.
DOI 10.1103/PhysRevLett.106.200401.
[32]
Ježek, M. et al.
Experimental test of the strongly nonclassical
character of a noisy squeezed single-photon state.
Physical Review A
86, 043813
(2012).
URL http://link.aps.org/doi/10.1103/PhysRevA.86.043813.
DOI 10.1103/PhysRevA.86.043813.
ArXiv: 1206.7057.
[33]
Straka, I. et al.
Quantum non-Gaussian Depth of Single-Photon
States.
Physical Review Letters
113, 223603
(2014).
URL http://link.aps.org/doi/10.1103/PhysRevLett.113.223603.
DOI 10.1103/PhysRevLett.113.223603.
ArXiv: 1403.4194.
[34]
Vostrosablin, N., Rakhubovsky, A. A. &
Filip, R.
Pulsed quantum interaction between two distant
mechanical oscillators.
Physical Review A
94, 063801
(2016).
URL http://link.aps.org/doi/10.1103/PhysRevA.94.063801.
DOI 10.1103/PhysRevA.94.063801.
ArXiv: 1605.05932. |
Observation of a dynamical QCD string
L. Ya. Glozman
Institute for Physics,
University of Graz, Universitätsplatz 5, A-8010 Graz, Austria
Abstract
Mesons constructed from the quark propagators without the
lowest-lying eigenmodes of the Dirac operator reveal not only restored
$SU(2)_{L}\times SU(2)_{R}$ chiral and $U(1)_{A}$ symmetries, but actually a
higher symmetry. All possible chiral and $U(1)_{A}$ multiplets for the
states of the same spin are degenerate, i.e., the energy of the observed quantum
levels does not depend on the spin orientation of quarks in the system and
their parities. The quark-spin independence of the energy levels implies
absence of the magnetic interactions in the system. The ultrarelativistic
quark-antiquark system with only the color-electric interactions can
be interpreted (or defined) as a dynamical QCD string.
pacs: 11.30.Rd, 12.38.Gc, 11.25.-w
I Introduction
With the static color charges one observes on the lattice
a color-electric flux tube B , which, if the distance
between the static quarks is large, can be approximated as a string.
This string is nondynamical (in the sense that its ends are fixed).
In the light quark systems a crucial aspect is a relativistic
motion of quarks at the ends of a possible string. There is no
consistent theory of the dynamical QCD string with quarks at the ends.
It is even apriori unclear whether such a picture has something to do with
reality. In this case the chiral symmetry as well as its
dynamical braking should be relevant.
Both confinement and chiral symmetry breaking dynamics
are important for the hadronic mass generation. Their
interplay is responsible for a rather complicated structure
of the hadron spectra in the light quark sector. A large
degeneracy is seen in the highly excited mesons G1 ; G2 ,
which is, however, absent in the observed spectrum below 1.8 GeV.
The low-lying hadron spectra should be strongly affected by the
chiral symmetry breaking dynamics.
In order to disentangle the confinement physics from the
chiral symmetry breaking dynamics we remove on the lattice
from the valence quark
propagators the lowest-lying quasi-zero modes of the Dirac operator
keeping at the same time the gluon gauge configurations intact
LS ; GLS ; DGL . Indeed,
the quark condensate of the vacuum is related to
a density of the lowest quasi-zero eigenmodes of the
Dirac operator C :
$$<0|\bar{q}q|0>=-\pi\rho(0).$$
(1)
We subtract from the valence quark propagators their lowest-lying
chiral modes, which are a tiny part of the full amount of modes,
$$S_{RD(k)}=S_{Full}-\sum_{i=1}^{k}\,\frac{1}{\lambda_{i}}\,|\lambda_{i}\rangle%
\langle\lambda_{i}|\\
,$$
(2)
where $\lambda_{i}$ and $|\lambda_{i}\rangle$ are the eigenvalues and
the corresponding eigenvectors of the
Dirac operator. Given these reduced quark propagators we study the existence
of hadrons and
their masses upon reduction of the lowest-lying modes responsible for the
chiral symmetry breaking.
II Lattice technology
In contrast to refs. LS ; GLS , where a chirally not
invariant Wilson lattice Dirac operator was employed, we adopt now
DGL
a manifestly chiral-invariant overlap Dirac operator N .
The quark propagators were generously provided by the JLQCD
collaboration KEK ; Aoki:2012pma ; Noaki:2008iy . For some technical details of the work
we refer the reader to ref. DGL .
III Observations, symmetries and their string interpretation
At this first stage of the study we have investigated
the ground state and the excited states of all possible $\bar{q}q$
isovector
$J=1$ mesons, i.e., $\rho(1^{--})$, $a_{1}(1^{++})$, $b_{1}(1^{+-})$.
An exponential decay of the correlation function is interpreted as
a physical state and its mass is extracted. The quality of the
exponential decay improves with the reduction of the lowest modes.
The evolution of masses of the ground and excited states
upon reduction of the low-lying modes is shown in Fig. 1.
At the truncation energy about 50 MeV an onset of a degeneracy of the
states , $\rho,\rho^{\prime},a_{1},b_{1}$, as well as a degeneracy of their excited
states is seen. This degeneracy indicates a symmetry.
All possible multiplets of the $SU(2)_{L}\times SU(2)_{R}$ group for the $J=1$
mesons are shown in Table 1 G1 ; G2 . In the chirally symmetric world
there
must be two independent $\rho$-mesons that belong to different
chiral representations. The first one is a member of the $(0,1)+(1,0)$
representation. It can be created from the vacuum only by the operators
that have the same chiral structure, e.g. by the vector-isovector current
$\bar{q}\gamma^{i}\vec{\tau}q$.
Its chiral partner is the axial vector meson, $a_{1}$, that is created by
the axial-vector isovector current. When chiral symmetry is restored this $\rho$-meson
must be degenerate with the $a_{1}$ state. Another $\rho$-meson, along with
its chiral partner $h_{1}$, is a member of the
$(1/2,1/2)_{b}$ representation. It can be created only by the
operators that have the same
chiral structure, e.g. by $\bar{q}\sigma^{0i}\vec{\tau}q$.
In the real world (i.e. with broken chiral symmetry)
each $\rho$-state (i.e. $\rho$ and $\rho^{\prime}$) is a mixture of these two
representations and they are well split. Upon subtraction of 10 lowest
modes we observe two independent degenerate $\rho$-mesons. One of them
couples only to the vector current and does not couple to the
$\bar{q}\sigma^{0i}\vec{\tau}q$ operator. The other $\rho$ meson - the
other way around. This means that one of these degenerate $\rho$-states
belongs to the $(0,1)+(1,0)$ multiplet and the other one is a member
of the $(1/2,1/2)_{b}$ representation.
A degeneracy of the $(0,1)\oplus(1,0)$ $\rho$-meson with the $a_{1}$ meson
is a clear signal of the chiral $SU(2)_{L}\times SU(2)_{R}$ restoration.
Consequently, a similar degeneracy should be observed in all other chiral pairs
from Table 1.
The $U(1)_{A}$ symmetry transforms the $b_{1}$ state into the $(1/2,1/2)_{b}$
$\rho$-meson G1 ; G2 . Their degeneracy indicates a restoration of
the $U(1)_{A}$ symmetry. We conclude
that simultaneously both $SU(2)_{L}\times SU(2)_{R}$ and $U(1)_{A}$ symmetries
get restored.
The restored $SU(2)_{L}\times SU(2)_{R}\times U(1)_{A}$ symmetry requires a
degeneracy of four mesons that belong to $(1/2,1/2)_{a}$
and $(1/2,1/2)_{b}$ chiral multiplets G1 ; G2 , see Table 1.
This symmetry does not require, however, a degeneracy of these four states with other
mesons, in particular with $a_{1}$ and its chiral partner $\rho$. We clearly
see the latter degeneracy. This implies that there is some higher symmetry,
that includes
$SU(2)_{L}\times SU(2)_{R}\times U(1)_{A}$ as a subgroup. This higher symmetry requires a
degeneracy of all eight mesons from the Table 1.
All these eight mesons can be combined to a reducible $\bar{q}q$
representation, which is a product of two fundamental chiral
quark representations CJ :
$$\displaystyle[(0,1/2)+(1/2,0)]\times[(0,1/2)+(1/2,0)]=$$
$$\displaystyle(0,0)+(1/2,1/2)_{a}+(1/2,1/2)_{b}+[(0,1)+(1,0)]\;.$$
(3)
They exhaust all possible chiralities of quarks and antiquarks, i.e.,
their spin orientations,
as well as possible spatial and charge parities for non-exotic mesons.
The observed degeneracy of all these eight mesons suggests that the higher
symmetry, mentioned above, should combine
the $U(1)_{A}$ and the $SU(2)_{L}\times SU(2)_{R}$ rotations in the isospin
space with the
$SU(2)_{S}$ spin-symmetry. The latter one is due to independence of the energy
on the orientations
of the quark spins. The higher symmetry group that combines
all eight mesons from the Table 1 into one multiplet
of dimension 16 should be
$SU(2\cdot N_{f})$.
The quark-spin independence of the energy levels implies that there
are no magnetic interactions in the system, i.e., the spin-orbit
force, the color-magnetic (hyperfine) and tensor interactions
are absent G3 .
The energy of the system is entirely due to interactions of the color
charges via the color-electric field and due to a relativistic motion
of the system. We interpret (or define) such a system as a dynamical QCD
string. Note a significant qualitative difference with the case of a motion of
an electrically charged fermion in a static electric field or with a relative motion
of two fermions with the electric charge.
In the latter cases there exist
spin-orbit and spin-spin forces that are a manifestation of the magnetic interaction in the system.
In our case such a magnetic interaction is absent, which implies that both
quarks are at rest with respect to the electric field (that
moves together with
quarks). It is this circumstance that suggests to interpret (define) our system as a
dynamical QCD string.
The observed radial levels at the truncation energy 65 MeV,
at which we see the onset of the symmetry,
are approximately equidistant and can be described through
the simple relation
$$E_{n_{r}}=(n_{r}+1)\hbar\omega,~{}~{}~{}n_{r}=0,1,...$$
(4)
The extracted value of the radial
string excitation quantum
amounts to $\hbar\omega=(900\pm 70)$ MeV.
At the moment we cannot exclude, however,
the quadratic relation $E_{n_{r}}^{2}\sim(n_{r}+1)$,
because the excited level can be shifted up due to rather
small finite lattice volume.
There is an interesting aspect of the dynamical
QCD string that crucially distinguishes it from the
Nambu-Goto open string. The energy of the the Nambu-Goto open bosonic
string is determined by its orbital angular momentum $L$, $M^{2}\sim L$.
For the dynamical QCD string that contains chiral quarks at the ends
the orbital angular momentum $L$ of the relative motion
is not conserved GN .
For instance, two orthogonal $\rho$-mesons at the same energy
level are represented by the mutually orthogonal fixed superpositions of the
$S$- and $D$-waves.
$$\displaystyle\displaystyle|(0,1)+(1,0);1~{}1^{--}\rangle$$
$$\displaystyle=$$
$$\displaystyle\sqrt{\tfrac{2}{3}}\,|1;{}^{3}S_{1}\rangle+\sqrt{\tfrac{1}{3}}\,|%
1;{}^{3}D_{1}\rangle,$$
$$\displaystyle\displaystyle|(1/2,1/2)_{b};1~{}1^{--}\rangle$$
$$\displaystyle=$$
$$\displaystyle\sqrt{\tfrac{1}{3}}\,|1;{}^{3}S_{1}\rangle-\sqrt{\tfrac{2}{3}}\,|%
1;{}^{3}D_{1}\rangle.$$
Acknowledgments
I am very grateful to M. Denissenya and C. B. Lang
for our common efforts in lattice simulations. The JLQCD collaboration
is acknowledged for their suggestion to use their overlap gauge
configurations and
quark propagators. This work is partially supported by
the Austrian Science
Fund (FWF) through the grant P26627-N16.
References
(1)
G. S. Bali,
Phys. Rept. 343, 1 (2001).
(2)
L. Y. Glozman,
Phys. Lett. B 587, 69 (2004).
(3)
L. Y. Glozman,
Phys. Rept. 444, 1 (2007).
(4)
C. B. Lang and M. Schröck,
Phys. Rev. D 84, 087704 (2011).
(5)
L. Y. Glozman, C. B. Lang and M. Schröck,
Phys. Rev. D 86, 014507 (2012).
(6)
M. Denissenya, L. Y. Glozman and C. B. Lang,
Phys. Rev. D 89, 077502 (2014).
(7)
A. Casher,
Phys. Lett. B 83, 395 (1979).
(8)
H. Neuberger,
Phys. Lett. B 417, 141 (1998).
(9)
S. Aoki et al. [JLQCD Collaboration],
Phys. Rev. D 78, 014508 (2008).
(10)
S. Aoki et al.
PTEP 2012, 01A106 (2012).
(11)
J. Noaki et al. [JLQCD and TWQCD Collaborations],
Phys. Rev. Lett. 101, 202004 (2008).
(12)
T. D. Cohen and X. -D. Ji,
Phys. Rev. D 55, 6870 (1997).
(13)
L. Y. Glozman,
Phys. Lett. B 541, 115 (2002).
(14)
L. Y. Glozman and A. V. Nefediev,
Phys. Rev. D 76, 096004 (2007);
Phys. Rev. D 80, 057901 (2009). |
Testing of quantum nonlocal correlations under constrained free will and imperfect detectors
Abhishek Sadhu
[email protected]
Light and Matter Physics Group, Raman Research Institute, Bengaluru, Karnataka 560080, India
Siddhartha Das
[email protected]
Center for Security, Theory and Algorithmic Research, Centre of Quantum Science and Technology, International Institute of Information Technology, Hyderabad, Gachibowli, Telangana 500032, India
Abstract
In this work, we deal with relaxation of two central assumptions in the standard locally realistic hidden variable (LRHV) inequalities: free will in choosing of measurement settings and presence of perfect detectors at the measurement devices. Violation of LRHV inequalities by quantum correlations imply that given correlations are quantum nonlocal. However, in an adversarial situation, there could be a hidden variable introducing bias in the selection of measurement settings but observers with no access to hidden variable could be unaware of the bias. Whereas in practice, detectors don’t have perfect efficiency. A main focus of this paper is introduction of the framework where given a quantum state with nonlocal behavior under constrained free will, we can determine the threshold values of detector parameters (detector inefficiency and dark counts) such that the detectors are robust enough to certify nonlocality. We also introduce a new class of LRHV inequalities with constrained free will and discuss their implications in testing of quantum nonlocal correlations.
I Introduction
The randomness acts as a primitive resource for various desirable information processing and computational tasks, e.g., cryptography shannon1945mathematical ; bennett1984proceedings ; carl2006denial ; das2021universal ; primaatmaja2022security , statistical sampling yates1946review ; dix2002chance , probabilistic algorithms rabin1980probabilistic ; leon1988probabilistic ; sipser1983complexity , genetic algorithms holland1992adaptation ; morris1998automated ; deaven1995molecular , simulated annealing algorithms ram1996parallel ; fleischer1995simulated ; kirkpatrick1983optimization , neural networks bishop1994neural ; harrald1997evolving ; zhang2000neural , etc. Generation and certification of private randomness also form major goal of cryptographic protocols shannon1945mathematical ; acin2012randomness ; colbeck2012free ; RNGColbeck_2011 ; das2021universal ; primaatmaja2022security . The quantum theory allows for the generation and certification of true randomness einstein1935can ; bell1964einstein ; PhysRevLett.67.661 ; colbeck2012free ; das2021universal ; primaatmaja2022security , which otherwise seems to be impossible within the classical theory.
In a seminal paper bell1964einstein , J.S. Bell showed that the statistical predictions of the quantum mechanics cannot be explained by local realistic hidden variable (LRHV) theories. Several LRHV inequalities, also called Bell-type inequalities, that are based on two physical assumptions— existence of local realism and no-signalling criterion einstein1935can , have been derived since then (see e.g., brunner2014bell ; HSD15 and references therein). Violation of these LRHV inequalities brunner2014bell by quantum systems would imply presence of quantum nonlocal correlations that are not explainable by any LRHV theories. Hence, the randomness generated from the behavior of quantum systems with nonlocal correlations can be deemed to exhibit true randomness. Presence of these nonlocal correlations allow for the generation and certification of randomness and secret key in a device-independent way PhysRevLett.67.661 ; QKDAcin2006 ; acin2016certified ; kaur2021upper ; primaatmaja2022security ; zapatero2022advances ; wooltorton2022tight .
It is known that the experimental verification of the Bell’s inequality requires additional assumptions which could lead to incurring loopholes such as locality loophole bell2004speakable , freedom-of-choice (measurement independence or free will) loophole clauser1974experimental ; bell1985exchange , fair-sampling loophole (detection loophole) pearle1970hidden ; clauser1974experimental . In recent major breakthroughs hensen2015loophole ; giustina2015significant ; shalm2015strong ; big2018challenging , incompatibility of quantum mechanics with LRHV theories has been demonstrated by considerably loophole-free experiments showing violation of Bell’s inequality by quantum states with quantum nonlocal correlations. In an experimental setup, locality assumption requires space like separation between the measurement events einstein1935can ; bell1964einstein which leads to locality loophole aspect1982experimental ; weihs1998violation . In a Bell experiment with photonic setup, the detection events of the two parties are identified as belonging to the same pair, by observing whether the difference in the time of detection is small larsson2014bell . It makes possible for an adversary to fake the observed quantum correlations by exploiting the time difference between the detection events of the parties larsson2004bell giving rise to the detection loophole, which can be closed by using predefined coincidence detection windows larsson2004bell .
The free will assumption 111It should be noted that the free will assumption mentioned in this paper is also called measurement independence. This assumption relates to the possible correlations between the choice of measurement settings for the two parties, which can affect the observed experimental statistics. in the Bell-type inequality states that the parties (users) can choose the measurement settings freely or, use uncorrelated random number generators. In an experimental demonstration presented in big2018challenging human random decisions were used to choose the measurement settings in an attempt to close this loophole. In practice, the detectors are imperfect which introduce dark counts eisaman2011invited and detection inefficiencyeisaman2011invited . In this paper, we discuss implications of bias in the freedom-of-choice and non-ideal detectors on the LRHV tests of quantum nonlocal correlations.
To the best of our knowledge, the assumption of free will was relaxed in hall2010local using a distance measure based quantification of the measurement dependence. It was shown that the Bell-CHSH inequality can be violated by sacrificing equal amount of free will for both parties. This result was extended to the scenario of parties having different amounts of free will in friedman2019relaxed and for one of the parties in banik2012optimal . In an alternate approach, measurement dependence was quantified in putz2014arbitrarily ; putz2016measurement by bounding the probability of choosing of the measurement settings to be in a given range. Following this approach, tests for nonlocality has been constructed putz2014arbitrarily ; zhao2022tilted . These inequalities has been applied to randomness amplification protocols kessler2020device ; zhao2022tilted . However, the consideration of imperfect detectors in the implication of these measurement-dependent LRHV inequalities are still lacking as the above mentioned works assumed perfect detection while allowing for relaxation in measurement dependence.
A main focus of this work is to study the implications of imperfect detectors and constrained free will on the test of quantum nonlocal correlations. We adapt the approach discussed in sauer2020quantum to model imperfect detectors for the Bell experiment as a sequential application of a perfectly working inner box followed by a lossy outer box. The inner box contains a quantum source whose behavior is nonlocal under constrained free will, i.e., violates certain measurement dependent LRHV inequality. The outer box separately introduces detector inefficiency and dark counts for each party. Using this model, we determine the threshold values of the detector parameters that make detectors robust for testing of quantum nonlocality under constrained free will (e.g., see Fig. 1 with details in Section IV). Next for the scenario of perfect detectors, we compare the implications two different approaches presented in putz2014arbitrarily and hall2010local to quantify measurement dependence (a) by bounding the probability of choosing the measurement settings $x$ (for Alice’s side) and $y$ (for Bob’s side) conditioned on a hidden variable $\lambda$ to be in the range $[l,1-3l]$ putz2014arbitrarily and (b) by using a distance measure $M$ to quantify measurement settings distinguishability hall2010local . This comparison is made in the 2 (party) - 2 (measurement settings per party) - 2 (outcome per measurement) scenario and their effects on the certification of the nonlocality. We also introduce a new set of measurement dependent LRHV (MDL) inequalities by introducing distance-based measurement dependent quantity in adapted AMP tilted Bell inequalityacin2012randomness and discuss implications and trade-off between measurement dependence parameters and tilted parameters for the certification of quantum nonlocal correlations.
Throughout this paper we limit our discussions to LRHV for bipartite quantum systems. The structure of this paper is as follows. In Section II, we introduce the framework of locality and measurement dependence, which is a constraint that limits the free will of the user. We present a model where the user is tricked by an adversary into thinking that they have freedom-of-choice for the measurement. In Section III, we consider and compare two different approaches to quantify the measurement dependence in terms of parameters $l$ and $M$ mentioned earlier. We determine the critical values of these parameters necessary for the certification of nonlocality in the measurement dependence settings and compare them with amount of violation obtained for the Bell-CHSH inequality, tilted Hardy relations and the tilted Bell inequalities. In Section IV, We determine the threshold values of the detector parameters namely inefficiency and dark count probability such that the detectors are robust enough to certify nonlocality in presence of constrained free will. In Section V, we introduce a new set of LRHV inequalities adapted from the AMP tilted Bell inequality using distance measures based measurement dependence quantities. We use these inequalities to observe the effect of relaxing the free will assumption for either parties on the certification of quantum nonlocal correlations.
II The adverserial role in choice of measurement settings
Formally a quantum behavior of a device with bipartite state $\rho_{AB}$ is given by its quantum representation $\{p(ab|xy)\}$, where $p(ab|xy)=\operatorname{Tr}[\Gamma^{x}_{a}\otimes\Gamma^{y}_{b}\rho]$ with $\{\Gamma^{x}_{a}\}_{a}$ and $\{\Gamma^{y}_{b}\}_{b}$ being Positive Operator Valued Measures (POVMs) for each input $(x,y)$ to the device, for $x\in\mathcal{X}$ and $y\in\mathcal{Y}$. If a state $\psi_{AB}=\outerproduct{\psi}{\psi}_{AB}$ is pure, then we may simply represent it as a ket $\ket{\psi}_{AB}$ instead of density operator $\psi_{AB}$. Henceforth, the abbreviation MDL inequalities would stand for measurement dependent LRHV inequalities.
We consider the Bell scenario where two parties, Alice and Bob, share a bipartite quantum state $\rho_{AB}$. Each party can choose to perform one of two measurements available to them, i.e., $|\mathcal{X}|=2=|\mathcal{Y}|$. We attribute POVM $\{\Lambda^{x}_{a}\}_{x}$ to Alice and $\{\Lambda^{y}_{b}\}_{y}$ to Bob. Each of these measurements can have two outcomes. We denote measurement outcomes for Alice and Bob by $a$ and $b$, respectively. The statistics of the measurement outcomes in the experiment can then be described by the probability distribution $\textbf{P}=\{p(ab|xy)\}$ which is also termed as behavior. In this framework, let there exist a hidden variable $\lambda$ belonging to some hidden-variable space, $\Lambda.$ The probability distribution of the outputs conditioned on the inputs can then be expressed in the form of
$$p(ab|xy)=\sum_{\lambda\in\Lambda}p(ab|xy\lambda)p(\lambda|xy).$$
(1)
The hidden variable $\lambda$ (distribution according to $p(\lambda)$) can provide an explanation of the observed experimental (measurement) statistics. In each run of the experiment, there exists a fixed $\lambda$ that describes the outcome of the experimental trial following the distribution $p(ab|xy\lambda)$. After multiple runs of the experiment, the output statistics are described by sampling from the distribution $p(\lambda|xy)$.
In an adversarial scenario, Alice and Bob can believe that they are choosing all the settings with equal probability, i.e., $p(xy)=\frac{1}{4}~{}\forall~{}(x,y)$ while an adversary biases their choice in the scale $\lambda$. The adversary can distribute the settings chosen by Alice and Bob according to
$$p(xy)=\sum_{\lambda\in\Lambda}p(xy|\lambda)p(\lambda).$$
(2)
In Eq. (2), let $\lambda$ take two values, $\lambda_{1}$ and $\lambda_{2}$, whose probability distributions are given as $p\text{$(\lambda_{1}$)}=\sin^{2}(\theta_{\lambda})$ and $p\text{$(\lambda_{2}$)}=\cos^{2}(\theta_{\lambda})$, respectively. In the simplest scenario $x$ and $y$ can each take values $0$ or $1$ and there are four possible ways in which the measurement settings can be chosen by Alice and Bob, i.e., $(x,y)\in\{(0,0),(0,1),(1,0),(1,1)\}.$ Let the probability of choosing the measurement setting $(x,y)$ conditioned on the hidden variable be distributed according to Table 1.
To observe the effect of the conditional probability distribution for choice of $x$ and $y$ from Table 1 on $p(xy)$, consider the following examples:
(i)
For the case of $\theta_{s_{2}}=0.6847$, $\phi_{s_{2}}=1.2491$, $\phi_{s_{1}}=0.8861$, $\delta_{s_{2}}=1.2491$, $\theta_{\lambda}=0.785398$, $\theta_{s_{1}}=0.50413$, and $\delta_{s_{1}}=0.175353$. It can be seen that this choice of parameters set $p(xy)=0.25~{}\forall~{}(x,y)$.
(ii)
For the case of $\theta_{s_{2}}=0.5796$, $\phi_{s_{2}}=1.2491$, $\phi_{s_{1}}=0.6847$, $\delta_{s_{2}}=0.75$, $\theta_{\lambda}=0.57964$, $\theta_{s_{1}}=0.793732$, and $\delta_{s_{1}}=1.06395$. It can be seen that this choice of parameters set $p(xy)=0.25~{}\forall~{}(x,y)$.
The above examples show that by proper choice of parameters in Table 1, the adversary can trick Alice and Bob into thinking that they have free will in choosing the measurement setting. In the presence of a bias in the choice of measurement settings in the $\lambda$ scale which Alice and Bob are unaware of, the following constraints blasiak2021violations can be imposed on the conditional joint probability distribution:
a.
The signal locality, i.e., no-signalling, assumption imposes the factorisability constraint on the conditional joint probability distribution,
$$p(ab|xy\lambda)=p(a|x\lambda)p(b|y\lambda).$$
(3)
b.
The measurement independence, i.e., freedom-of-choice or free will, assumption requires that $\lambda$ does not contain any information about $x$ and $y$ which is equivalent to stating
$$\displaystyle p(\lambda|xy)=p(\lambda)$$
or equivalently,
$$\displaystyle p(xy|\lambda)=p(xy).$$
(4)
The relaxation of the measurement independence assumption, which we call measurement dependence indicates an adversary can infer the values of $x$ and $y$ by knowing $\lambda$.
III Quantifying measurement dependence
In this section, we first review two different approaches considered in putz2014arbitrarily ; zhao2022tilted and hall2010local ; banik2012optimal ; friedman2019relaxed to quantify the measurement dependence. We then compare these two approaches and observe their effects on the tests of quantum nonlocal behaviors.
III.1 Review of MDL inequalities from prior works
We review approach discussed in putz2014arbitrarily ; putz2016measurement to quantify measurement dependence by bounding the probability of choice of measurement settings conditioned on a hidden variable to be in a specific range (Section III.1.1). Then we review approach discussed in hall2010local ; banik2012optimal ; friedman2019relaxed to quantify measurement dependence using a distance measure (Section III.1.2).
III.1.1 Bound on the probability of choosing the measurement settings
In the works putz2014arbitrarily ; putz2016measurement , we observe that the probability of Alice and Bob to choose measurement settings $x$ and $y$ conditioned on $\lambda$ can be bounded in the form
$$l\leq p(x,y|\lambda)\leq h,$$
(5)
where $0\leq l\leq p(x,y|\lambda)\leq h\leq 1$. If Alice and Bob each chooses from two possible values of measurement settings, then $l=h=0.25$ corresponds to the complete measurement independence and $l,h\neq 0.25$ represents a bias in the choice of measurement settings.
In the 2 (user) - 2 (measurement settings per user) - 2 (outcome per measurement) scenario with $a,b\in\{+1,-1\}$ and $x,y\in\{0,1\}$, it was shown in putz2014arbitrarily ; putz2016measurement that all the measurement dependent local correlations satisfy the PRBLG MDL inequality
$$lp(++00)-h(p(+-01)+p(-+10)+p(++11))\leq 0.$$
(6)
A two-dimensional slice in the non-signalling space is shown in Fig. 2 (figure from putz2014arbitrarily ). In the figure, the quantum set is bounded by the green line and the set of non-signalling correlations lie within the black triangle.
In Fig. 2, the red dotted line corresponds to Eq. (6) with $h=1-3l$. If we set $h=1-3l,$ the PRBLG MDL inequality tilts from the Bell-CHSH inequality ($l=0.25$) to the non-signalling border ($l=0$). For $h=1-3l$, Eq. (6) modifies to the form
$$\displaystyle lp(++00)-(1-3l)(p(+-01)+p(-+10)$$
$$\displaystyle+p(++11))\leq 0.$$
(7)
We note that if Alice and Bob believes they have complete measurement dependence, i.e., $p(xy)=0.25\hskip 3.0pt\forall(x,y),$ then Eq. (7) reduces to the form
$$\displaystyle lp(++|00)-(1-3l)(p(+-|01)$$
$$\displaystyle+p(-+|10)+p(++|11))\leq 0.$$
(8)
It follows from zhao2022tilted that invoking the PRBLG MDL inequality (7) in the tilted Hardy test zhao2022tilted , we obtain the ZRLH MDL inequality. The ZRLH MDL inequality takes the form of
$$\displaystyle l[p(++|00)+wp(--|00)-\max\{0,w\}]$$
$$\displaystyle-(1-3l)[p(+-|01)+p(-+|10)+p(++|11)]\leq 0,$$
(9)
where $w$ is the tilting parameter taking real numbers in the range, $w\in(-0.25,1).$ We call the behaviors that violate Eqs. (8) and (9) as quantum nonlocal in the presence of measurement dependence.
III.1.2 Distance measure to quantify measurement distinguishability
We discussed in Section II that the experimental statistics described by the joint probability distribution $p(ab|xy)$ can be explained by $\lambda\in\Lambda$ in the following form,
$$p(ab|xy)=\int{\rm d}\lambda p(ab|xy\lambda)p(\lambda|xy).$$
(10)
The assumption of the measurement independence constrains the probability distribution of measurement settings via
$$p(\lambda|xy)=p(\lambda).$$
(11)
Eq. (11) implies that no extra information about $\lambda$ can be obtained from the knowledge of $x$ and $y$. This is equivalent to saying
$$p(xy|\lambda)=\frac{p(\lambda|xy)p(xy)}{p(\lambda)}=p(xy).$$
(12)
Eq. (12) implies that Alice and Bob has complete freedom in choosing the measurement settings $x$ and $y$ respectively. If $x\in U\equiv\{x_{1},x_{2}\}$ and $y\in V\equiv\{y_{1},y_{2}\}$, measurement dependence implies $p(\lambda|x_{1},y_{1})\neq p(\lambda|x_{2},y_{2})$. Distinguishability between $p(\lambda|x_{1},y_{1})$ and $p(\lambda|x_{2},y_{2})$ can be quantified using a distance measure defined in hall2010local ,
$$M=\int{\rm d}\lambda|p(\lambda|x_{1},y_{1})-p(\lambda|x_{2},y_{2})|.$$
(13)
We can express the probability to successfully distinguish $p(\lambda|x_{1},y_{1})$ and $p(\lambda|x_{2},y_{2})$ based on the knowledge of $\lambda$ as
$$F=\frac{1}{2}\left(1+\frac{M}{2}\right).$$
(14)
When $M=0$, we have $F=\frac{1}{2}$. This implies that the probability to distinguish the pairs of measurement settings is $\frac{1}{2}$, and the hidden variable does not have any information about the measurement settings. This is consistent with maximum free will that Alice and Bob has while choosing the measurement settings. For $M=2$, we have $F=1$. This implies that the probability to distinguish the pairs of measurement settings is upper bounded by $1$, and the hidden variable has complete information about the measurement settings. The local degrees of measurement dependence for Alice and Bob as introduced in hall2010local is given by
$$M_{1}\equiv\max\left\{\int{\rm d}\lambda\absolutevalue{p(\lambda|x_{1},y_{1})-p(\lambda|x_{2},y_{1})},\int{\rm d}\lambda\absolutevalue{p(\lambda|x_{1},y_{2})-p(\lambda|x_{2},y_{2})}\right\},$$
(15)
$$M_{2}\equiv\max\left\{\int{\rm d}\lambda\absolutevalue{p(\lambda|x_{1},y_{1})-p(\lambda|x_{1},y_{2})},\int{\rm d}\lambda\absolutevalue{p(\lambda|x_{2},y_{1})-p(\lambda|x_{2},y_{2})}\right\},$$
(16)
$M_{1}$ quantifies the measurement dependence for Alice’s settings keeping Bob’s settings fixed, and similarly the other way round for $M_{2}.$
The above parameters will be useful in deriving the bounds on the AMP tilted Bell inequalities acin2012randomness in the Section V.
III.2 Comparison and discussion on implications of different measurement dependent LRHV inequalities
In this section, we check for the allowed values of measurement dependence parameter $l$ to ensure nonlocality of quantum behaviors that are used to obtain randomness.
Consider the behavior in Table 2 that violates the AMP tilted Bell inequality (34). In the limit of $\alpha\rightarrow\infty,$ close to two bits of randomness can be obtained from this behavior by violating the AMP tilted Bell inequality acin2012randomness . For this behavior, the PRBLG MDL inequality (8) reduces to
$$l\leq\frac{3\alpha t+\alpha\cos(2\phi)t-2\sqrt{2}\alpha\sin^{2}(\phi)+\sqrt{2}(\cos(4\phi)-1)}{10\alpha t+4\alpha\cos(2\phi)\left(t+\sqrt{2}\right)-2\sqrt{2}\left(\alpha+3\sin^{2}(2\phi)\right)}$$
(17)
where $t=\sqrt{-\frac{\cos(4\phi)}{\alpha^{2}}+\frac{1}{\alpha^{2}}+2}.$ In the limit of $\alpha\rightarrow\infty,$ from Eq. (17) we have $l\leq 0.25.$
We see that in the limit of $\alpha\rightarrow\infty$ the behavior in Table 2 does not violate any PRBLG MDL inequality. For $\alpha=1$ (which is equivalent to Bell-CHSH inequality) and $\phi=\frac{\pi}{4}$ (the maximum violation) the behavior behavior in Table 2 violates the family of PRBLG MDL inequalities with $l>0.2023.$
We observe in Fig. 3 that for a fixed value of $\alpha$, the allowed values of $\phi$ from Table 2 that violate the PRBLG MDL inequality given by Eq. (8) increases with the increase in $l$. Also as $\alpha$ increases, for a particular value of $l$, there is a decrease in the number of the allowed values of $\phi$ from Table 2 that violate the AMP tilted Bell inequality.
It was shown in zhao2022tilted that the state $\psi_{g}=\outerproduct{\psi_{g}}{\psi_{g}}$ (18) and measurement settings $\{A_{0}^{g},A_{1}^{g},B_{0}^{g},B_{1}^{g}\}$ can be used to obtain $1$ bit of local randomness and close to $1.6806$ bits of global randomness, where
$$\ket{\psi_{g}}=\cos(\frac{\theta}{2})\ket{00}-\sin(\frac{\theta}{2})\ket{11},$$
(18)
$$\displaystyle A_{0}^{g}=B_{0}^{g}=$$
$$\displaystyle\frac{-\sqrt{2}\sin(\theta)\sqrt{\sin(\theta)}}{(2-\sin(\theta))\sqrt{\sin(\theta)+1}}\sigma_{x}$$
(19)
$$\displaystyle+\frac{-(\sin(\theta)+2)\sqrt{1-\sin(\theta)}}{(2-\sin(\theta))\sqrt{\sin(\theta)+1}}\sigma_{z},$$
$$\displaystyle A_{1}^{g}=B_{1}^{g}=\frac{\sqrt{2}\sqrt{\sin(\theta)}}{\sqrt{\sin(\theta)+1}}\sigma_{x}-\frac{\sqrt{1-\sin(\theta)}}{\sqrt{\sin(\theta)+1}}\sigma_{z},$$
(20)
such that $\theta=\sin^{-1}{(3-\sqrt{4w+5})}$ and $A_{x}^{g},B_{y}^{g}$ for $x,y\in\{0,1\}$ denote measurement settings for Alice and Bob, respectively.
Observation 1.
Let us consider the behavior given by the state $\psi_{g}$ (18) and measurement settings $\{A_{0}^{g},A_{1}^{g},B_{0}^{g},B_{1}^{g}\}$. For this behavior, the PRBLG MDL inequality given by Eq. (8) reduces to
$$\displaystyle\frac{2l}{(\xi-1)^{2}}\left(31\xi+w\left(4\xi-34\right)-69\right)$$
$$\displaystyle+\frac{2l(\xi-5)\sqrt{\xi-2}}{\sqrt{4-\xi}(\xi-1)}\sqrt{-4w+6\xi-13}$$
$$\displaystyle-\frac{2(1-3l)}{1-\xi}\bigg{(}8w-10\xi+\frac{\sqrt{\xi-2}\sin\left(2\sin^{-1}\left(3-\xi\right)\right)}{\sqrt{4-\xi}}$$
$$\displaystyle+22\bigg{)}\leq 0,$$
(21)
where $\xi=\sqrt{4w+5}.$ On inspection we see that for all $w\in(-0.25,1),$ Eq. (21) reduces to $l\leq 0.$ The behavior specified by the state $\psi_{g}$ and measurement settings $\{A_{0}^{g},A_{1}^{g},B_{0}^{g},B_{1}^{g}\}$ is nonlocal for $l>0$.
For the given behavior, the ZRLH MDL inequality (9) reduces to
$$\displaystyle\frac{1}{2\left(\xi-1\right)}(2l\left(-4w+7\xi-15\right)+$$
$$\displaystyle\frac{(1-3l)\sqrt{\xi-2}\sin\left(2\sin^{-1}\left(3-\xi\right)\right)}{\sqrt{4-\xi}}+8w-10\xi+22)$$
$$\displaystyle-l\max(0,w)\leq 0.$$
(22)
On inspection we see that for all $w\in(-0.25,1),$ Eq. (22) reduces to $l\leq 0$.
That is, the behavior given by state $\psi_{g}$ and measurement settings $\{A_{0}^{g},A_{1}^{g},B_{0}^{g},B_{1}^{g}\}$ violates both PRBLG and ZRLH MDL inequalities and its quantum nonlocality can be certified for all possible $l>0$.
IV Imperfect detector and constrained free will
We model the detection units of Alice and Bob using a two-box approach following sauer2020quantum . There is an inner box containing a quantum source generating bipartite quantum states whose behavior is nonlocal under constrained free will but assuming that detectors are perfect. Nonlocality of the quantum behavior in the inner box is tested by violation of a given MDL inequality. The output of the inner box is quantum nonlocal behavior that violates a given MDL inequality. An outer box introduces the detector imperfections, namely the detection inefficiency and dark counts. The quantum nonlocal behavior obtained from the inner box gets mapped at the outer box to the output behavior with detector imperfection parameters. The output behavior then undergoes an LRHV test based on which we are able to determine threshold values of detection inefficiency and dark counts such that the given quantum nonlocal behavior can still be certified to be nonlocal with imperfect detectors. A deviation from a two-box approach in sauer2020quantum is the introduction of the measurement dependence assumption to the working of the inner box. Alice and Bob has access only to the input settings and the outputs of the outer box. We assume that either party (Alice and Bob) has access to two identical detectors that can distinguish between the orthogonal outputs.
The measurement outcomes for the inner box are labeled as $a^{id},b^{id}$, respectively. We note that $a^{id}$ and $b^{id}$ can each take values from the set $\{+,-\}.$ Introducing non-unit detection efficiency, $0\leq\eta\leq 1$, and non-zero dark count probability, $0\leq\delta\leq 1$ in the outer box, the ideal two outcome scenario becomes a 4-outcome scenario with the addition of no-detection event $\Phi$, and the dark-count event $\chi.$ These events are defined in the following way:
$\Phi:$
One particle is sent to the party and none of the two detectors of that party click.
$\chi:$
One particle is sent to the party and both the detectors of the party click.
We label the measurement outcomes for Alice and Bob obtained from the outer box as $a^{ob},b^{ob}$, respectively. We note that $a^{ob}$ and $b^{ob}$ can each take values from the set $\{+,-,\Phi,\chi\}.$ Furthermore, we assume that Alice and Bob’s detection units have identical values for $\eta$ and $\delta$. The conditional probability of observing the outcome $t$ from the outer box conditioned on observing $t^{id}$ in the ideal scale is given by $p(t|t^{id})$ with $t\in\{a^{ob},b^{ob}\}$ and $t^{id}\in\{a^{id},b^{id}\}.$ The observed joint probabilities can then be expressed as sauer2020quantum
$$p(a^{ob}b^{ob}|xy)=\sum_{a^{id},b^{id}}p(a^{ob}|a^{id})p(b^{ob}|b^{id})p(a^{id}b^{id}|xy).$$
(23)
We relax the free will assumption in the inner box by introducing the hidden variable $\lambda\in\Lambda$. Considering this assumption, $p(a^{id}b^{id}|xy)$ takes the form of
$$p(a^{id}b^{id}|xy)=\sum_{\lambda\in\Lambda}p(a^{id}b^{id}|xy\lambda)p(\lambda|xy).$$
(24)
The hidden variable, $\lambda$ (distributed according to $p(\lambda)$) provides an explanation of the observed experimental statistics of the inner box. The distribution of settings that are chosen by Alice and Bob depends on $\lambda$ via the following relation,
$$p(xy)=\sum_{\lambda\in\Lambda}p(xy|\lambda)p(\lambda).$$
(25)
If we impose the locality condition from Eq. (3) on the experimental statistics of the inner box, we arrive at the following factorisability constraint,
$$p(a^{id}b^{id}|xy\lambda)=p(a^{id}|x\lambda)p(b^{id}|y\lambda).$$
(26)
Also, if we impose the measurement independence assumption from Eq. (4), we arrive at the following constraint,
$$p(xy|\lambda)=p(xy)\text{(or equivalently, }p(\lambda|xy)=p(\lambda)).$$
(27)
The output statistics of the outer box for imperfect detectors can depend on the output of the inner box in the following four ways sauer2020quantum :
i)
No particle is detected on either of the detectors and no dark count detection event takes place. We can then write the following:
$$p(a|a^{id})=(1-\eta)(1-\delta)^{2}.$$
(28)
ii)
No particle is detected by the detector that should have detected it and a dark count takes place in the other detector. We can then write the following:
$$p(a|a^{id})=(1-\eta)(1-\delta)\delta.$$
(29)
iii)
Either the particle is detected by one of the detectors and a dark count takes place in the other detector or the particle is not detected and dark counts take place in both the detectors. We can then write the following:
$$\displaystyle p(a|a^{id})$$
$$\displaystyle=\eta\delta+(1-\eta)\delta^{2}$$
(30)
$$\displaystyle=\delta[1-(1-\eta)(1-\delta)].$$
iv)
Either the particle is detected and no dark count takes place or the particle is not detected and a dark count takes place in the detector in which the particle should have been registered. We can then write the following:
$$\displaystyle p(a|a^{id})$$
$$\displaystyle=\eta(1-\delta)+(1-\eta)\delta^{2}$$
(31)
$$\displaystyle=(1-\delta)[1-(1-\eta)(1-\delta)].$$
The quantum nonlocal behavior obtained from the inner box after getting mapped to the output behavior with detector imperfection parameters remain nonlocal if it violates the inequality sauer2020quantum
$$\displaystyle p(++|00)+p(++|01)+p(++|10)-p(++|11)$$
$$\displaystyle-p_{A}(+|0)-p_{B}(+|0)\leq 0,$$
(32)
where $p_{A}(o|s)$ and $p_{B}(o|s)$ are the probabilities of Alice and Bob to obtain the outcome $o$ on measuring $s.$
At first, let us assume there is a quantum source in the inner box that generating a bipartite quantum state whose behavior $\{p(a^{id},b^{id}|xy)\}$ violates the PRBLG MDL inequality given by Eq. (8) assuming that detectors are perfect. The behavior $\{p(a^{id},b^{id}|xy)\}$ obtained from the inner box gets mapped at the outer box to the behavior $\{p(a^{ob},b^{ob}|xy)\}$ with the introduction of the detector inefficiency $\eta$ and dark count probability $\delta$. The behavior $\{p(a^{ob},b^{ob}|xy)\}$ is then inserted in Eq. (32) to obtain the critical detector parameters using Algorithm 1. For a fixed value of $\delta$ we obtain the minimum value of $\eta$ that violates Eq. (32) using Algorithm 1. We abbreviate left-hand side as LHS.
We plot the critical values of $\eta$ and $\delta$ obtained using Algorithm 1 in Fig. 1.
Observation 2.
From Fig. 1, we see that the minimum value of $\eta$ for a given $\delta$ is the most demanding for $l=0$ and decreases as $l$ increases. For a fixed value of $l$, the minimum value of $\eta$ increases monotonically with the dark count probability. We note that for $\delta=0,$ we have $\eta=0.667$ for all the values of $l$.
We next assume there is a quantum source in the inner box that generating a bipartite quantum state whose behavior $\{p(a^{id},b^{id}|xy)\}$ violates the ZRLH MDL inequality given by Eq. (9) assuming that detectors are perfect. The behavior $\{p(a^{id},b^{id}|xy)\}$ obtained from the inner box gets mapped at the outer box to the behavior $\{p(a^{ob},b^{ob}|xy)\}$ with the introduction of the detector inefficiency $\eta$ and dark count probability $\delta$. The behavior $\{p(a^{ob},b^{ob}|xy)\}$ is then inserted in Eq. (32) to obtain the critical detector parameters using Algorithm 1. For a fixed value of $\delta$ we obtain the minimum value of $\eta$ that violates Eq. (32) using Algorithm 1. We abbreviate left-hand side as LHS.
We plot the critical values of $\eta$ and $\delta$ from Algorithm 2 for $w=0$ in Fig. (4).
Observation 3.
The ZRLH MDL inequality given by Eq. (9) reduces to the PRBLG MDL inequality Eq. (8) for $w=0$. For $l=0$, Eq. (9) and Eq. (8) reduce to the same form and is independent of $w$. For these two cases, we observe similar dependence for the threshold detector parameters as can be seen from Fig. 4. Also, we observe from Fig. 4 that for a fixed value of $l$, the minimum detection efficiency increases monotonically with the dark count probability.
We comment here that using Algorithm 2, one can calculate the detector requirements for other values of $(w,l)$ not mentioned in this section. Next, we consider the state $\ket{\psi_{g}}$ and measurement settings $\{A_{0}^{g},A_{1}^{g},B_{0}^{g},B_{1}^{g}\}$ with $\theta=1.13557.$ This choice of state and measurement settings was shown to ensure $1.6806$ bits of global randomness in zhao2022tilted .
Observation 4.
Consider the quantum source in the inner box generates bipartite quantum state $\ket{\psi_{g}}.$ Let Alice and Bob choose the measurement settings $\{A_{0}^{g},A_{1}^{g},B_{0}^{g},B_{1}^{g}\},$ with $\theta=1.13557.$ With this choice of state and measurement settings we obtain the behavior $\{p(a^{id},b^{id}|xy)\}$ assuming that the detectors are perfect. The output behavior $\{p(a^{ob},b^{ob}|xy)\}$ obtained from the outer box is evaluated using Eq. (23). The behavior $\{p(a^{ob},b^{ob}|xy)\}$ is nonlocal if we have a violation of the inequality
$$\displaystyle\delta\eta[\delta(9.50424\delta-4\delta^{2}-9.00848)+4.25636]$$
$$\displaystyle+\eta^{2}[\delta(2\delta^{3}-5.50424\delta^{2}+5.82473\delta-3.13675)$$
$$\displaystyle+0.816258]+2\delta(\delta-1)(\delta^{2}-\delta+1)$$
$$\displaystyle-0.752119\eta\leq 0.$$
(33)
The allowed values of the pair $(\eta,\delta)$ for which $p(a^{ob},b^{ob}|xy)$ is quantum nonlocal shown in Fig. 5.
In Fig. 5 we observe that for $\delta=0$ the minimum detection efficiency $\eta_{\text{crit}}\approx 0.92$ is required to ensure $p(a^{ob},b^{ob}|xy)$ is quantum nonlocal. We also observe that minimum values of $\eta$ increases monotonically with increase in $\delta$.
V The Tilted Bell inequality
The AMP tilted Bell inequality is given as acin2012randomness
$$I_{\alpha}^{\beta}:=\beta\langle x_{1}\rangle+\alpha\langle x_{1}y_{1}\rangle+\alpha\langle x_{1}y_{2}\rangle+\langle x_{2}y_{1}\rangle-\langle x_{2}y_{2}\rangle\leq\beta+2\alpha.$$
(34)
It plays an important role in (a) demonstrating the inequivalence between the amount of certified randomness and the amount of nonlocality acin2012randomness , (b) self-testing of all bipartite pure entangled states coladangelo2017all , (c) protocol for device-independent quantum random number generation with sublinear amount of quantum communication bamps2018device , (d) unbounded randomness certification from a single pair of entangled qubits with sequential measurements curchod2017unbounded .
In the following proposition, we obtain the bound on $I_{\alpha}^{\beta}$ when the measurement independence assumption is relaxed (see Appendix A for proof).
Proposition 1.
The AMP tilted Bell expression $I_{\alpha}^{\beta}$ in the presence of locality and the relaxed measurement independence is bounded by
$$I_{\alpha}^{\beta}\leq\beta+2\alpha+\min\{\alpha(M_{1}+\min\{M_{1},M_{2}\})+M_{2},2\},$$
(35)
where $M_{1}$ and $M_{2}$ are the measurement dependence parameters for Alice and Bob.
V.1 Comparison of one-sided and two-sided measurement dependence to ensure quantum representation
Let Alice and Bob have free will in choosing the measurement settings, then the maximum violation of $I_{\alpha}^{\beta}$ obtained by quantum nonlocal behaviors is given by acin2012randomness
$$I_{\alpha}^{\beta}\leq 2\sqrt{(1+\alpha^{2})(1+\frac{\beta^{2}}{4})}.$$
(36)
As direct consequences of Proposition 1, we have following corollaries.
Corollary 1.
When $M_{1}$ = $M_{2}$ = $M$, the quantum nonlocal behaviors that maximally violate Eq. (34) with the amount of violation given by Eq. (36), remains nonlocal for
$$M<\frac{-2\alpha-\beta+\sqrt{(1+\alpha^{2})(4+\beta^{2})}}{1+2\alpha}.$$
(37)
For $\alpha=1\text{ and }\beta=0$ (the Bell-CHSH scenario), Eq. (37) reduces to $M<\frac{2}{3}(\sqrt{2}-1)\approx 0.276$ and is consistent with the observation in friedman2019relaxed .
Corollary 2.
When $M_{1}=0$ and $M_{2}=M$, the quantum nonlocal behaviors that maximally violate of Eq. (34) with the amount of violation given by Eq. (36), remains nonlocal for
$$M<-2\alpha-\beta+\sqrt{(1+\alpha^{2})(4+\beta^{2})}$$
(38)
For $\beta=0\text{ and }\alpha=1$ (the Bell-CHSH scenario.), Eq. (38) reduces to $M<2(\sqrt{2}-1)\approx 0.828$ and is consistent with the observation in friedman2019relaxed .
Corollary 3.
When $M_{1}=M$ and $M_{2}=0$, the quantum nonlocal behaviors that maximally violate of Eq. (34) with the amount of violation given by Eq. (36), remains nonlocal for
$$M<\frac{1}{\alpha}(-2\alpha-\beta+\sqrt{(1+\alpha^{2})(4+\beta^{2})})$$
(39)
For $\beta=0\text{ and }\alpha=1$ (the Bell-CHSH scenario.), Eq. (39) reduces to $M<2(\sqrt{2}-1)\approx 0.828$ and is consistent with the observation in friedman2019relaxed .
We consider $\beta=0$ and $\alpha\geq 1$ and plot in Fig. 6, the upper bound on $M$ as a function of $\alpha$ for MDL behaviors. The values of $M$ and $\alpha$ from Fig. 6 ensures that the quantum nonlocal behaviors that maximally violate Eq. (34) remains nonlocal in the presence of measurement dependence.
We observe in Fig. 6 that for $\beta=0$ and $\alpha\geq 1$, the introduction of one-sided measurement dependence allows for higher values of $M$ (implying lower degree of freedom-of-choice) as compared to introducing both-sided measurement dependence. Also, for the case of one-sided measurement dependence, Alice can have higher values of $M$ as compared to Bob.
V.2 Bounds on the measurement dependence for testing nonlocality
We observe in Proposition 1 that in the presence of relaxed measurement dependence, the behaviors $\{p(ab|xy)\}$, that violates Eq. (40) are nonlocal.
$$I_{\alpha}^{\beta}\leq\beta+2\alpha+\min\{\alpha(M_{1}+\min\{M_{1},M_{2}\})+M_{2},2\}$$
(40)
In the following, we discuss some of the cases where Eq. (40) can be violated.
a.
We consider a situation when both Alice and Bob have the same measurement dependence, we have $M_{1}=M_{2}=M$. For this case, violating Eq. (40) requires
$$I_{\alpha}^{\beta}>\beta+2\alpha+\min\{(2\alpha+1)M,2\}.$$
(41)
If $(2\alpha+1)M\geq 2$, $I_{\alpha}^{\beta}$ reaches the no-signalling boundary. Whereas, if $(2\alpha+1)M<2$, then Eq. (41) reduces to inequality
$$M<\frac{I_{\alpha}^{\beta}-\beta-2\alpha}{2\alpha+1}.$$
(42)
b.
We consider the situation when only Bob has measurement dependence, we have $M_{1}=0,M_{2}=M.$ For this case, violating Eq. (40) requires
$$I_{\alpha}^{\beta}>\beta+2\alpha+\min\{M,2\}.$$
(43)
If $M=2$, $I_{\alpha}^{\beta}$ reaches the no-signalling boundary. Whereas, if $M<2$, then Eq. (43) to the inequality
$$M<I_{\alpha}^{\beta}-\beta-2\alpha.$$
(44)
c.
We consider the situation when only Alice has measurement dependence, we have $M_{1}=M,M_{2}=0.$ For this case, violating Eq. (40) requires
$$I_{\alpha}^{\beta}>\beta+2\alpha+\min\{\alpha M,2\}.$$
(45)
If $\alpha M\geq 2$, $I_{\alpha}^{\beta}$ reaches the no-signalling boundary. Whereas if $\alpha M<2$, then Eq. (45) reduces to the inequality
$$M<\frac{I_{\alpha}^{\beta}-\beta-2\alpha}{\alpha}.$$
(46)
We note that for $\alpha=1$ and $\beta=0,$ the values of $M_{1}$ and $M_{2}$ that violates Eq. (40) are in agreement with that obtained in friedman2019relaxed .
Consider there exists some behavior $\mathbf{P}^{\prime}=\{p^{\prime}(ab|xy)\}$ that violates Eq. (34) with the amount of violation given by $I_{\alpha}^{\beta}=I^{\prime}(\alpha,\beta,\mathbf{P}^{\prime})$. For $\alpha=1,\beta=8$, we show in Fig. 7 the possible values of measurement dependence parameters $M_{1}$ and $M_{2}$ for which the amount of violation of Eq. (34) given by $I^{\prime}$ cannot be described by a deterministic measurement dependent local model. The plots for the other combinations of $(\alpha,\beta)$ is given in Appendix B.
VI Discussion
In this paper, two different approaches in quantifying measurement dependence via parameters $l$ and $(M_{1},M_{2})$ from standard literature is presented and a bound on these parameters for certifying nonlocality of different behaviors is obtained. It is observed that the behavior certifying close to 2 bits of randomness remains measurement dependent nonlocal only in the limit of complete measurement independence, i.e., in the limit of $l\rightarrow 0.25.$ The behavior that provides 1.6806 bits of global randomness from the violation of the tilted Hardy relations is measurement dependent nonlocal for arbitrarily small values of $l.$ This motivates further study on obtaining an inequality that provides close to two bits of global randomness in the limit of arbitrarily low measurement dependence, a problem we leave for future study. Deviating from the conventional approach of assuming perfect detectors, we present a framework to determine the threshold values of the detector parameters that are robust enough to certify nonlocality of given quantum behaviors. This is an important step towards experimentally obtaining nonlocality in the presence of relaxed measurement independence. For an illustration, we presented the critical requirements for generating 1.6806 bits of tilted-Hardy certified global randomness. The detector parameters obtained from this study is expected to have important applications in experimentally implementing various information processing tasks that rely on quantum nonlocality as a resource.
The modified analytical bound on the AMP tilted Bell inequality in terms of $M_{1}$ and $M_{2}$ has been obtained. Using the analytical bound, the bounds on $M_{1}$ and $M_{2}$ to ensure quantumness and nonlocality are observed. It is observe that one-sided measurement dependence is more advantageous from the point of view of the user as compared to two-sided measurement dependence. The analytical bound obtained is expected to have applications in self-testing of quantum states and other device independent information processing protocols like randomness generation and secure communication.
For future work, it would be interesting to see implications of measurement dependence in multipartite Bell-type inequalities brunner2014bell ; HSD15 for the certification of multipartite nonlocality and device-independent conference keys RMW18 ; holz2020genuine ; HWD22 .
Note— We noticed the related work “Quantum nonlocality in presence of strong measurement dependence” by Ivan Šupić, Jean-Daniel Bancal, and Nicolas Brunner recently posted as arXiv:2209.02337.
Acknowledgements.
The authors thank Karol Horodecki for discussions. A.S. acknowledges the PhD fellowship from the Raman Research Institute, Bangalore. A.S. thanks the International Institute of Information Technology, Hyderabad for the hospitality during his visit in the summer of 2022 where part of this work was done.
Appendix A Calculating the modified local bound for the AMP tilted Bell inequality
We consider the AMP tilted-Bell inequality introduced in acin2012randomness and given by,
$$I_{\alpha}^{\beta}=\beta\langle x_{1}\rangle+\alpha\langle x_{1}y_{1}\rangle+\alpha\langle x_{1}y_{2}\rangle+\langle x_{2}y_{1}\rangle-\langle x_{2}y_{2}\rangle$$
(47)
The determinism assumption states that the measurement outcomes are deterministic functions of the choice of settings and the hidden variable $\lambda,$ i.e., $a=A(x,\lambda)$ and $b=B(x,\lambda)$ with $p(a|x\lambda)=\delta_{a,A(x,\lambda)}$ and $p(b|y\lambda)=\delta_{b,B(y,\lambda)}.$ If we assume that the choice of the measurement settings on the side of Alice and Bob can depend on some hidden variable, $\lambda$, the correlations of Alice and Bob can be expressed as
$$\langle xy\rangle=\int{\rm d}\lambda p(\lambda|x,y)A(x,\lambda)B(y,\lambda).$$
(48)
The correlation function of Alice can also be written as
$$\langle x\rangle=\int{\rm d}\lambda p(\lambda|x)A(x,\lambda).$$
(49)
Applying Eq. (48) and Eq. (49) to Eq. (47) we have
$$\displaystyle I_{\alpha}^{\beta}=\beta\int$$
$$\displaystyle{\rm d}\lambda p(\lambda|x_{1})A(x_{1},\lambda)$$
(50)
$$\displaystyle+\alpha\int{\rm d}\lambda p(\lambda|x_{1},y_{1})A(x_{1},\lambda)B(y_{1},\lambda)$$
$$\displaystyle+\alpha\int{\rm d}\lambda p(\lambda|x_{1},y_{2})A(x_{1},\lambda)B(y_{2},\lambda)$$
$$\displaystyle+\int{\rm d}\lambda p(\lambda|x_{2},y_{1})A(x_{2},\lambda)B(y_{1},\lambda)$$
$$\displaystyle-\int{\rm d}\lambda p(\lambda|x_{2},y_{2})A(x_{2},\lambda)B(y_{2},\lambda).$$
Adding and subtracting the terms
$$\displaystyle\alpha\int{\rm d}\lambda p(\lambda|x_{1},y_{2})A(x_{1},\lambda)B(y_{1},\lambda)\text{ and}$$
$$\displaystyle\int{\rm d}\lambda p(\lambda|x_{2},y_{2})A(x_{2},\lambda)B(y_{1},\lambda)$$
to Eq. (50) we obtain
$$\displaystyle I_{\alpha}^{\beta}$$
$$\displaystyle=\beta\int{\rm d}\lambda p(\lambda|x_{1})A(x_{1},\lambda)$$
(51)
$$\displaystyle\qquad+\alpha\int{\rm d}\lambda A(x_{1},\lambda)B(y_{1},\lambda)\bigg{[}p(\lambda|x_{1},y_{1})$$
$$\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad-p(\lambda|x_{1},y_{2})\bigg{]}$$
$$\displaystyle\qquad+\int{\rm d}\lambda A(x_{2},\lambda)B(y_{1},\lambda)\bigg{[}p(\lambda|x_{2},y_{1})$$
$$\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad-p(\lambda|x_{2},y_{2})\bigg{]}$$
$$\displaystyle\qquad+\alpha\int{\rm d}\lambda p(\lambda|x_{1},y_{2})\bigg{[}A(x_{1},\lambda)B(y_{2},\lambda)$$
$$\displaystyle\qquad\qquad\qquad\qquad\qquad+A(x_{1},\lambda)B(y_{1},\lambda)\bigg{]}$$
$$\displaystyle\qquad-\int{\rm d}\lambda p(\lambda|x_{2},y_{2})\bigg{[}A(x_{2},\lambda)B(y_{2},\lambda)$$
$$\displaystyle\qquad\qquad\qquad\qquad-A(x_{2},\lambda)B(y_{1},\lambda)\bigg{]}.$$
We observe that Eq. (51) is bounded by $I_{\alpha}^{\beta}\leq\max[T_{1}]+\max[T_{2}]+\max[T_{3}]$. We also note that as $p(\lambda|x_{1})$ is a normalised probability distribution, $\int{\rm d}\lambda p(\lambda|x_{1})=1$. This implies that the maximum value of the quantity $\beta\int{\rm d}\lambda p(\lambda|x_{1})A(x_{1},\lambda)$ is given by $\beta$ when $A(x_{1},\lambda)$ is set to $1$. With these observations, $T_{1}$ can be simplified as
$$\displaystyle T_{1}=$$
$$\displaystyle\beta\int{\rm d}\lambda p(\lambda|x_{1})A(x_{1},\lambda)$$
(52)
$$\displaystyle+\alpha\int{\rm d}\lambda A(x_{1},\lambda)B(y_{1},\lambda)\bigg{[}p(\lambda|x_{1},y_{1})$$
$$\displaystyle\qquad\qquad-p(\lambda|x_{1},y_{2})\bigg{]}\leq\beta+\alpha M_{2}.$$
where we have set $B(y_{1},\lambda)=1.$ To evaluate the maximum value of $T_{2}$ we set $A(x_{2},\lambda)=1$ and obtain
$$\displaystyle T_{2}=$$
$$\displaystyle\int{\rm d}\lambda A(x_{2},\lambda)B(y_{1},\lambda)\bigg{[}p(\lambda|x_{2},y_{1})-p(\lambda|x_{2},y_{2})\bigg{]}$$
$$\displaystyle\leq M_{2}.$$
(53)
We evaluate $T_{3}$ as follows,
$$\displaystyle T_{3}$$
$$\displaystyle=$$
$$\displaystyle\alpha\int{\rm d}\lambda p(\lambda|x_{1},y_{2})\bigg{[}A(x_{1},\lambda)B(y_{2},\lambda)+A(x_{1},\lambda)B(y_{1},\lambda)\bigg{]}$$
$$\displaystyle-\int{\rm d}\lambda p(\lambda|x_{2},y_{2})\bigg{[}A(x_{2},\lambda)B(y_{2},\lambda)-A(x_{2},\lambda)B(y_{1},\lambda)\bigg{]}$$
$$\displaystyle=$$
$$\displaystyle\int{\rm d}\lambda A(x_{1},\lambda)B(y_{2},\lambda)\bigg{[}\alpha p(\lambda|x_{1},y_{2})$$
(54)
$$\displaystyle\qquad\qquad\qquad\qquad-p(\lambda|x_{2},y_{2})\frac{A(x_{2},\lambda)}{A(x_{1},\lambda)}\bigg{]}$$
$$\displaystyle+\int{\rm d}\lambda A(x_{1},\lambda)B(y_{1},\lambda)\bigg{[}\alpha p(\lambda|x_{1},y_{2})$$
$$\displaystyle\qquad\qquad\qquad\qquad+p(\lambda|x_{2},y_{2})\frac{A(x_{2},\lambda)}{A(x_{1},\lambda)}\bigg{]}.$$
To get the maximum value of $T_{3}$, we set the values of $A(x_{1},\lambda),B(y_{1},\lambda),A(x_{1},\lambda),B(y_{2},\lambda)$ to one. This then implies,
$$\displaystyle T_{3}=$$
$$\displaystyle\int{\rm d}\lambda\bigg{[}\alpha p(\lambda|x_{1},y_{2})-p(\lambda|x_{2},y_{2})\bigg{]}$$
(55)
$$\displaystyle+\int{\rm d}\lambda\bigg{[}\alpha p(\lambda|x_{1},y_{2})+p(\lambda|x_{2},y_{2})\bigg{]}.$$
We evaluate the first term of the Eq. (55) as
$$\displaystyle\int{\rm d}\lambda\bigg{[}\alpha p(\lambda|x_{1},y_{2})-p(\lambda|x_{2},y_{2})\bigg{]}$$
$$\displaystyle=$$
$$\displaystyle\int{\rm d}\lambda\bigg{[}\alpha p(\lambda|x_{1},y_{2})-\alpha p(\lambda|x_{2},y_{2})+\alpha p(\lambda|x_{2},y_{2})$$
(56)
$$\displaystyle\qquad-p(\lambda|x_{2},y_{2})\bigg{]}$$
$$\displaystyle=$$
$$\displaystyle\int{\rm d}\lambda\bigg{[}\alpha\bigg{(}p(\lambda|x_{1},y_{2})-p(\lambda|x_{2},y_{2})\bigg{)}$$
(57)
$$\displaystyle\qquad+(\alpha-1)p(\lambda|x_{2},y_{2})\bigg{]}$$
$$\displaystyle=$$
$$\displaystyle\alpha\int{\rm d}\lambda\bigg{[}p(\lambda|x_{1},y_{2})-p(\lambda|x_{2},y_{2})\bigg{]}$$
(58)
$$\displaystyle\qquad+(\alpha-1)\int{\rm d}\lambda p(\lambda|x_{2},y_{2}).$$
In Eq. (58) we note that
$$\alpha\int{\rm d}\lambda\bigg{(}p(\lambda|x_{1},y_{2})-p(\lambda|x_{2},y_{2})\bigg{)}\leq\alpha M_{1}.$$
We note that as $p(\lambda|x_{2},y_{2})$ is a normalised probability distribution, $\int{\rm d}\lambda p(\lambda|x_{2},y_{2})=1.$ This then implies
$$\int{\rm d}\lambda\bigg{[}\alpha p(\lambda|x_{1},y_{2})-p(\lambda|x_{2},y_{2})\bigg{]}\leq\alpha M_{1}+\alpha-1,$$
(59)
which on simplification reduces to
$$\int{\rm d}\lambda\bigg{[}\alpha p(\lambda|x_{1},y_{2})-p(\lambda|x_{2},y_{2})\bigg{]}\leq\alpha(M_{1}+1)-1.$$
(60)
We proceed in the same way for the second term in Eq. (55) as follows,
$$\displaystyle\int{\rm d}\lambda\bigg{[}\alpha p(\lambda|x_{1},y_{2})+p(\lambda|x_{2},y_{2})\bigg{]}$$
$$\displaystyle=$$
$$\displaystyle\int{\rm d}\lambda\bigg{[}\alpha p(\lambda|x_{1},y_{2})+\alpha p(\lambda|x_{2},y_{2})-\alpha p(\lambda|x_{2},y_{2})$$
(61)
$$\displaystyle\qquad+p(\lambda|x_{2},y_{2})\bigg{]}$$
$$\displaystyle=$$
$$\displaystyle\int{\rm d}\lambda\alpha\bigg{[}p(\lambda|x_{1},y_{2})+p(\lambda|x_{2},y_{2})\bigg{]}$$
(62)
$$\displaystyle\qquad-(\alpha-1)\int{\rm d}\lambda p(\lambda|x_{2},y_{2})$$
$$\displaystyle=$$
$$\displaystyle\alpha\int{\rm d}\lambda\bigg{[}p(\lambda|x_{1},y_{2})+p(\lambda|x_{2},y_{2})\bigg{]}$$
(63)
$$\displaystyle\qquad+(1-\alpha)\int{\rm d}\lambda p(\lambda|x_{2},y_{2}).$$
We note that as $p(\lambda|x_{1},y_{2})$, $p(\lambda|x_{2},y_{2})$ are normalised probability distributions, $\int{\rm d}\lambda p(\lambda|x_{1},y_{2})=1$ and $\int{\rm d}\lambda p(\lambda|x_{2},y_{2})=1.$ This then implies
$$\int{\rm d}\lambda\bigg{[}\alpha p(\lambda|x_{1},y_{2})+p(\lambda|x_{2},y_{2})\bigg{]}=\alpha+1.$$
(64)
If we insert Eq. (60) and Eq. (64) in Eq. (55), we obtain
$$\displaystyle T_{3}\leq\alpha(M_{1}+2).$$
(65)
Now combining Eq. (52), Eq. (53) and Eq. (65), we have the bound on $I_{\alpha}^{\beta}$ as
$$\displaystyle I_{\alpha}^{\beta}$$
$$\displaystyle\leq T_{1}+T_{2}+T_{3}$$
(69)
$$\displaystyle\leq\beta+\alpha M_{2}+M_{2}+\alpha(M_{1}+2)$$
$$\displaystyle\leq\beta+2\alpha+\alpha M_{1}+(\alpha+1)M_{2}$$
$$\displaystyle\leq\beta+2\alpha+\alpha(M_{1}+M_{2})+M_{2}.$$
We again start by considering Eq. (50) as
$$\displaystyle I_{\alpha}^{\beta}=$$
$$\displaystyle\beta\int{\rm d}\lambda p(\lambda|x_{1})A(x_{1},\lambda)$$
(70)
$$\displaystyle\qquad+\alpha\int{\rm d}\lambda p(\lambda|x_{1},y_{1})A(x_{1},\lambda)B(y_{1},\lambda)$$
$$\displaystyle\qquad+\alpha\int{\rm d}\lambda p(\lambda|x_{1},y_{2})A(x_{1},\lambda)B(y_{2},\lambda)$$
$$\displaystyle\qquad+\int{\rm d}\lambda p(\lambda|x_{2},y_{1})A(x_{2},\lambda)B(y_{1},\lambda)$$
$$\displaystyle\qquad-\int{\rm d}\lambda p(\lambda|x_{2},y_{2})A(x_{2},\lambda)B(y_{2},\lambda).$$
Adding and subtracting the terms $\alpha\int{\rm d}\lambda p(\lambda|x_{2},y_{1})A(x_{1},\lambda)B(y_{1},\lambda)$ and $\alpha\int{\rm d}\lambda p(\lambda|x_{2},y_{2})A(x_{1},\lambda)B(y_{2},\lambda)$ to Eq. (50) we obtain
$$\displaystyle I_{\alpha}^{\beta}$$
$$\displaystyle=\beta\int{\rm d}\lambda p(\lambda|x_{1})A(x_{1},\lambda)$$
(71)
$$\displaystyle\qquad+\alpha\int{\rm d}\lambda A(x_{1},\lambda)B(y_{1},\lambda)\bigg{[}p(\lambda|x_{1},y_{1})$$
$$\displaystyle\qquad\qquad\qquad\qquad\qquad-p(\lambda|x_{2},y_{1})\bigg{]}$$
$$\displaystyle\qquad+\alpha\int{\rm d}\lambda A(x_{1},\lambda)B(y_{2},\lambda)\bigg{[}p(\lambda|x_{1},y_{2})$$
$$\displaystyle\qquad\qquad\qquad\qquad\qquad-p(\lambda|x_{2},y_{2})\bigg{]}$$
$$\displaystyle\qquad+\int{\rm d}\lambda A(x_{2},\lambda)B(y_{1},\lambda)\bigg{[}p(\lambda|x_{2},y_{1})$$
$$\displaystyle\qquad\qquad\qquad-\frac{B(y_{2},\lambda)}{B(y_{1},\lambda)}p(\lambda|x_{2},y_{2})\bigg{]}$$
$$\displaystyle\qquad+\alpha\int{\rm d}\lambda A(x_{1},\lambda)B(y_{1},\lambda)\bigg{[}p(\lambda|x_{2},y_{1})$$
$$\displaystyle\qquad\qquad\qquad+\frac{B(y_{2},\lambda)}{B(y_{1},\lambda)}p(\lambda|x_{2},y_{2})\bigg{]}.$$
We observe that Eq. (71) is bounded by $I_{\alpha}^{\beta}\leq\max[t_{1}]+\max[t_{2}]+\max[t_{3}]$. We also note that as $p(\lambda|x_{1})$ is a normalised probability distribution, $\int{\rm d}\lambda p(\lambda|x_{1})=1$. This implies that the maximum value of the quantity $\beta\int{\rm d}\lambda p(\lambda|x_{1})A(x_{1},\lambda)$ then takes the value of $\beta$ when $A(x_{1},\lambda)$ is set to $1$. With these observations, $t_{1}$ can be simplified as,
$$\displaystyle t_{1}$$
$$\displaystyle=\beta\int{\rm d}\lambda p(\lambda|x_{1})A(x_{1},\lambda)$$
(72)
$$\displaystyle\qquad+\alpha\int{\rm d}\lambda A(x_{1},\lambda)B(y_{1},\lambda)\bigg{[}p(\lambda|x_{1},y_{1})$$
$$\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad\qquad-p(\lambda|x_{2},y_{1})\bigg{]}$$
$$\displaystyle\leq\beta+\alpha M_{1}.$$
Similarly, $t_{2}$ can be simplified as
$$\displaystyle t_{2}$$
$$\displaystyle=\alpha\int{\rm d}\lambda A(x_{1},\lambda)B(y_{2},\lambda)\bigg{[}p(\lambda|x_{1},y_{2})$$
(73)
$$\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad-p(\lambda|x_{2},y_{2})\bigg{]}$$
$$\displaystyle\leq\alpha M_{1},$$
and for $t_{3},$ we have the expression
$$\displaystyle t_{3}$$
$$\displaystyle=\int{\rm d}\lambda A(x_{2},\lambda)B(y_{1},\lambda)\bigg{[}p(\lambda|x_{2},y_{1})$$
(74)
$$\displaystyle\qquad\qquad\qquad-\frac{B(y_{2},\lambda)}{B(y_{1},\lambda)}p(\lambda|x_{2},y_{2})\bigg{]}$$
$$\displaystyle\qquad+\alpha\int{\rm d}\lambda A(x_{1},\lambda)B(y_{1},\lambda)\bigg{[}p(\lambda|x_{2},y_{1})$$
$$\displaystyle\qquad\qquad\qquad+\frac{B(y_{2},\lambda)}{B(y_{1},\lambda)}p(\lambda|x_{2},y_{2})\bigg{]}.$$
We set the values of $A(x_{1},\lambda),B(y_{1},\lambda),A(x_{2},\lambda),B(y_{2},\lambda)$ to one and obtain
$$\displaystyle t_{3}$$
$$\displaystyle=\int{\rm d}\lambda\bigg{[}p(\lambda|x_{2},y_{1})-p(\lambda|x_{2},y_{2})\bigg{]}$$
(75)
$$\displaystyle\qquad+\alpha\int{\rm d}\lambda\bigg{[}p(\lambda|x_{2},y_{1})+p(\lambda|x_{2},y_{2})\bigg{]}.$$
The first term in Eq, (75) is bounded by
$$\int{\rm d}\lambda\bigg{[}p(\lambda|x_{2},y_{1})-p(\lambda|x_{2},y_{2})\bigg{]}\leq M_{2}.$$
(76)
We note that as $p(\lambda|x_{2},y_{1})$ and $p(\lambda|x_{2},y_{2})$ are normalised probability distribution, $\int{\rm d}\lambda p(\lambda|x_{2},y_{1})=1$ and $\int{\rm d}\lambda p(\lambda|x_{2},y_{2})=1.$ Eq. (75) then takes the form of
$$t_{3}\leq M_{2}+2\alpha.$$
(77)
Now combining the values of $t_{1},$ $t_{2}$ and $t_{3}$, we have the bound on $I_{\alpha}^{\beta}$ as
$$\displaystyle I_{\alpha}^{\beta}$$
$$\displaystyle\leq t_{1}+t_{2}+t_{3}$$
(80)
$$\displaystyle\leq\beta+\alpha M_{1}+\alpha M_{1}+M_{2}+2\alpha$$
$$\displaystyle\leq\beta+2\alpha+2\alpha M_{1}+M_{2}.$$
We note that the bound on $I_{\alpha}^{\beta}$ must be the minimum of Eqs. (69) and (80) and takes the form of
$$I_{\alpha}^{\beta}\leq\beta+2\alpha+\alpha[M_{1}+\min\{M_{1},M_{2}\}]+M_{2}.$$
(81)
We observe that for the case of $\beta=0$ and $\alpha=1,$
$$I_{1}^{0}\leq 2+M_{1}+M_{2}+\min\{M_{1},M_{2}\},$$
(82)
which is in agreement with that obtained in friedman2019relaxed . We note that the maximum value that $I_{\alpha}^{\beta}$ [Eq. (47)] can take is $\beta+2\alpha+2$ (We get this bound by setting $\langle x_{1}\rangle=1$, $\langle x_{1}y_{1}\rangle=1$, $\langle x_{1}y_{2}\rangle=1$, $\langle x_{2}y_{1}\rangle=1$, $\langle x_{2}y_{2}\rangle=-1$) and arrive at
$$I_{\alpha}^{\beta}\leq\beta+2\alpha+\min\{\alpha(M_{1}+\min\{M_{1},M_{2}\})+M_{2},2\}.$$
(83)
Appendix B Bounds on the measurement dependence to certify nonlocality
We observe in Proposition 1 that in the presence of relaxed measurement dependence, the behaviors $\{p(ab|xy)\}$, that violates Eq. (84) are nonlocal.
$$I_{\alpha}^{\beta}\leq\beta+2\alpha+\min\{\alpha(M_{1}+\min\{M_{1},M_{2}\})+M_{2},2\}$$
(84)
Consider there exists some behavior $\mathbf{P}^{\prime}=\{p^{\prime}(ab|xy)\}$ that violates Eq. (34) with the amount of violation given by $I_{\alpha}^{\beta}=I^{\prime}(\alpha,\beta,\mathbf{P}^{\prime})$. We show in Fig. 8 and Fig. 9 the possible values of measurement dependence parameters $M_{1}$ and $M_{2}$ for which the amount of violation of Eq. (34) given by $I^{\prime}$ cannot be described by a deterministic measurement dependent local model. We observe that Fig. (8) is in agreement with that obtained in friedman2019relaxed .
References
[1]
Claude E Shannon.
A mathematical theory of cryptography.
Mathematical Theory of Cryptography, 1945.
[2]
C. H. Bennett and G. Brassard.
Quantum cryptography: Public key distribution and coin tossing.
Proceedings of the IEEE International Conference on Computers,
Systems and Signal Processing, pages 175–179, 1984.
[3]
Glenn Carl, George Kesidis, Richard R Brooks, and Suresh Rai.
Denial-of-service attack-detection techniques.
IEEE Internet Computing, 10(1):82–89, 2006.
[4]
Siddhartha Das, Stefan Bäuml, Marek Winczewski, and Karol Horodecki.
Universal limitations on quantum key distribution over a network.
Physical Review X, 11:041016, October 2021.
[5]
Ignatius W Primaatmaja, Koon Tong Goh, Ernest Y-Z Tan, John T-F Khoo, Shouvik
Ghorai, and Charles C-W Lim.
Security of device-independent quantum key distribution protocols: a
review.
2022.
arXiv:2206.04960.
[6]
Frank Yates.
A review of recent statistical developments in sampling and sampling
surveys.
Journal of the Royal Statistical Society, 109(1):12–43, 1946.
[7]
Alan Dix and Geoff Ellis.
By chance enhancing interaction with large data sets through
statistical sampling.
In Proceedings of the Working Conference on Advanced Visual
Interfaces, pages 167–176, 2002.
[8]
Michael O Rabin.
Probabilistic algorithm for testing primality.
Journal of Number Theory, 12(1):128–138, 1980.
[9]
Jeffrey S Leon.
A probabilistic algorithm for computing minimum weights of large
error-correcting codes.
IEEE Transactions on Information Theory, 34(5):1354–1359,
1988.
[10]
Michael Sipser.
A complexity theoretic approach to randomness.
In Proceedings of the fifteenth annual ACM symposium on Theory
of Computing, pages 330–335, 1983.
[11]
John H Holland.
Adaptation in natural and artificial systems: an introductory
analysis with applications to biology, control, and artificial intelligence.
MIT Press, 1992.
[12]
Garrett M Morris, David S Goodsell, Robert S Halliday, Ruth Huey, William E
Hart, Richard K Belew, and Arthur J Olson.
Automated docking using a Lamarckian genetic algorithm and an
empirical binding free energy function.
Journal of Computational Chemistry, 19(14):1639–1662, 1998.
[13]
David M Deaven and Kai-Ming Ho.
Molecular geometry optimization with a genetic algorithm.
Physical Review Letters, 75(2):288, 1995.
[14]
D Janaki Ram, TH Sreenivas, and K Ganapathy Subramaniam.
Parallel simulated annealing algorithms.
Journal of Parallel and Distributed Computing, 37(2):207–212,
1996.
[15]
Mark Fleischer.
Simulated annealing: past, present, and future.
In Winter Simulation Conference Proceedings, 1995., pages
155–161. IEEE, 1995.
[16]
Scott Kirkpatrick, C Daniel Gelatt Jr, and Mario P Vecchi.
Optimization by simulated annealing.
Science, 220(4598):671–680, 1983.
[17]
Chris M Bishop.
Neural networks and their applications.
Review of Scientific Instruments, 65(6):1803–1832, 1994.
[18]
Paul G Harrald and Mark Kamstra.
Evolving artificial neural networks to combine financial forecasts.
IEEE Transactions on Evolutionary Computation, 1(1):40–52,
1997.
[19]
Guoqiang Peter Zhang.
Neural networks for classification: a survey.
IEEE Transactions on Systems, Man, and Cybernetics, Part C
(Applications and Reviews), 30(4):451–462, 2000.
[20]
Antonio Acín, Serge Massar, and Stefano Pironio.
Randomness versus nonlocality and entanglement.
Physical Review Letters, 108(10):100402, 2012.
[21]
Roger Colbeck and Renato Renner.
Free randomness can be amplified.
Nature Physics, 8(6):450–453, 2012.
[22]
Roger Colbeck and Adrian Kent.
Private randomness expansion with untrusted devices.
Journal of Physics A: Mathematical and Theoretical,
44(9):095305, February 2011.
[23]
Albert Einstein, Boris Podolsky, and Nathan Rosen.
Can quantum-mechanical description of physical reality be considered
complete?
Physical Review, 47(10):777, 1935.
[24]
John S Bell.
On the Einstein Podolsky Rosen paradox.
Physics Physique Fizika, 1(3):195, 1964.
[25]
Artur K. Ekert.
Quantum cryptography based on Bell’s theorem.
Physical Review Letters, 67:661–663, August 1991.
[26]
Nicolas Brunner, Daniel Cavalcanti, Stefano Pironio, Valerio Scarani, and
Stephanie Wehner.
Bell nonlocality.
Reviews of Modern Physics, 86(2):419, 2014.
[27]
Dipankar Home, Debashis Saha, and Siddhartha Das.
Multipartite bell-type inequality by generalizing Wigner’s
argument.
Physical Review A, 91:012102, January 2015.
[28]
Antonio Acín, Serge Massar, and Stefano Pironio.
Efficient quantum key distribution secure against no-signalling
eavesdroppers.
New Journal of Physics, 8(8):126–126, August 2006.
[29]
Antonio Acín and Lluis Masanes.
Certified randomness in quantum physics.
Nature, 540(7632):213–219, 2016.
[30]
Eneet Kaur, Karol Horodecki, and Siddhartha Das.
Upper bounds on device-independent quantum key distribution rates in
static and dynamic scenarios.
July 2021.
arXiv:2107.06411.
[31]
Víctor Zapatero, Tim van Leent, Rotem Arnon-Friedman, Wen-Zhao Liu, Qiang
Zhang, Harald Weinfurter, and Marcos Curty.
Advances in device-independent quantum key distribution.
2022.
arXiv:2208.12842.
[32]
Lewis Wooltorton, Peter Brown, and Roger Colbeck.
Tight analytic bound on the trade-off between device-independent
randomness and nonlocality.
2022.
arXiv:2205.00124.
[33]
John S Bell.
Speakable and unspeakable in quantum mechanics: Collected
papers on quantum philosophy.
Cambridge University Press, 2004.
[34]
John F Clauser and Michael A Horne.
Experimental consequences of objective local theories.
Physical Review D, 10(2):526, 1974.
[35]
John S Bell, Abner Shimony, Michael A Horne, and John F Clauser.
An exchange on local beables.
Dialectica, 39(2):85–110, 1985.
[36]
Philip M Pearle.
Hidden-variable example based upon data rejection.
Physical Review D, 2(8):1418, 1970.
[37]
Bas Hensen, Hannes Bernien, Anaïs E Dréau, Andreas Reiserer, Norbert
Kalb, Machiel S Blok, Just Ruitenberg, Raymond FL Vermeulen, Raymond N
Schouten, Carlos Abellán, et al.
Loophole-free Bell inequality violation using electron spins
separated by 1.3 kilometres.
Nature, 526(7575):682–686, 2015.
[38]
Marissa Giustina, Marijn AM Versteegh, Sören Wengerowsky, Johannes
Handsteiner, Armin Hochrainer, Kevin Phelan, Fabian Steinlechner, Johannes
Kofler, Jan-Åke Larsson, Carlos Abellán, et al.
Significant-loophole-free test of Bell’s theorem with entangled
photons.
Physical Review Letters, 115(25):250401, 2015.
[39]
Lynden K Shalm, Evan Meyer-Scott, Bradley G Christensen, Peter Bierhorst,
Michael A Wayne, Martin J Stevens, Thomas Gerrits, Scott Glancy, Deny R
Hamel, Michael S Allman, et al.
Strong loophole-free test of local realism.
Physical Review Letters, 115(25):250402, 2015.
[40]
The BIG Bell Test Collaboration.
Challenging local realism with human choices.
Nature, 557(7704):212–216, 2018.
[41]
Alain Aspect, Jean Dalibard, and Gérard Roger.
Experimental test of Bell’s inequalities using time-varying
analyzers.
Physical Review Letters, 49(25):1804, 1982.
[42]
Gregor Weihs, Thomas Jennewein, Christoph Simon, Harald Weinfurter, and Anton
Zeilinger.
Violation of Bell’s inequality under strict Einstein locality
conditions.
Physical Review Letters, 81(23):5039, 1998.
[43]
Jan-Åke Larsson, Marissa Giustina, Johannes Kofler, Bernhard Wittmann,
Rupert Ursin, and Sven Ramelow.
Bell-inequality violation with entangled photons, free of the
coincidence-time loophole.
Physical Review A, 90(3):032107, 2014.
[44]
J-Å Larsson and Richard David Gill.
Bell’s inequality and the coincidence-time loophole.
EPL (Europhysics Letters), 67(5):707, 2004.
[45]
It should be noted that the free will assumption mentioned in this paper is
also called measurement independence. This assumption relates to the possible
correlations between the choice of measurement settings for the two parties,
which can affect the observed experimental statistics.
[46]
Matthew D Eisaman, Jingyun Fan, Alan Migdall, and Sergey V Polyakov.
Invited review article: Single-photon sources and detectors.
Review of Scientific Instruments, 82(7):071101, 2011.
[47]
Michael JW Hall.
Local deterministic model of singlet state correlations based on
relaxing measurement independence.
Physical Review Letters, 105(25):250404, 2010.
[48]
Andrew S Friedman, Alan H Guth, Michael JW Hall, David I Kaiser, and Jason
Gallicchio.
Relaxed Bell inequalities with arbitrary measurement dependence for
each observer.
Physical Review A, 99(1):012121, 2019.
[49]
Manik Banik, MD Rajjak Gazi, Subhadipa Das, Ashutosh Rai, and Samir Kunkri.
Optimal free will on one side in reproducing the singlet correlation.
Journal of Physics A: Mathematical and Theoretical,
45(20):205301, 2012.
[50]
Gilles Pütz, Denis Rosset, Tomer Jack Barnea, Yeong-Cherng Liang, and
Nicolas Gisin.
Arbitrarily small amount of measurement independence is sufficient to
manifest quantum nonlocality.
Physical Review Letters, 113(19):190402, 2014.
[51]
Gilles Pütz and Nicolas Gisin.
Measurement dependent locality.
New Journal of Physics, 18(5):055006, 2016.
[52]
Shuai Zhao, Ravishankar Ramanathan, Yuan Liu, and Paweł Horodecki.
Tilted Hardy paradoxes for device-independent randomness
extraction.
May 2022.
arXiv:2205.02751.
[53]
Max Kessler and Rotem Arnon-Friedman.
Device-independent randomness amplification and privatization.
IEEE Journal on Selected Areas in Information Theory,
1(2):568–584, 2020.
[54]
Alexander Sauer and Gernot Alber.
Quantum bounds on detector efficiencies for violating bell
inequalities using semidefinite programming.
Cryptography, 4(1):2, 2020.
[55]
Pawel Blasiak, Emmanuel M Pothos, James M Yearsley, Christoph Gallus, and Ewa
Borsuk.
Violations of locality and free choice are equivalent resources in
Bell experiments.
Proceedings of the National Academy of Sciences,
118(17):e2020569118, 2021.
[56]
Andrea Coladangelo, Koon Tong Goh, and Valerio Scarani.
All pure bipartite entangled states can be self-tested.
Nature Communications, 8(1):1–5, 2017.
[57]
Cédric Bamps, Serge Massar, and Stefano Pironio.
Device-independent randomness generation with sublinear shared
quantum resources.
Quantum, 2:86, 2018.
[58]
Florian J Curchod, Markus Johansson, Remigiusz Augusiak, Matty J Hoban, Peter
Wittek, and Antonio Acín.
Unbounded randomness certification using sequences of measurements.
Physical Review A, 95(2):020102, 2017.
[59]
Jérémy Ribeiro, Gláucia Murta, and Stephanie Wehner.
Fully device-independent conference key agreement.
2022.
arXiv:1708.00798.
[60]
Timo Holz, Hermann Kampermann, and Dagmar Bruß.
Genuine multipartite bell inequality for device-independent
conference key agreement.
Physical Review Research, 2(2):023251, 2020.
[61]
Karol Horodecki, Marek Winczewski, and Siddhartha Das.
Fundamental limitations on the device-independent quantum conference
key agreement.
Physical Review A, 105:022604, February 2022. |
Topological semigroups and universal spaces related to
extension dimension
A. Chigogidze
Department of Mathematics and Statistics,
University of Saskatchewan, McLean Hall, 106 Wiggins Road,
Saskatoon, SK, S7N 5E6, Canada
[email protected]
,
A. Karasev
Department of Mathematics and Statistics,
University of Saskatchewan, McLean Hall, 106 Wiggins Road,
Saskatoon, SK, S7N 5E6, Canada
[email protected]
and
M. Zarichnyi
Department of Mechanics and Mathematics, Lviv State University,
Universitetska 1, 290602 Lviv, Ukraine
[email protected]
(Date:: )
Abstract.
It is proved that there is no structure of left
(right) cancelative semigroup on $[L]$-dimensional universal space
for the class of separable compact spaces of extensional
dimension $\leq[L]$. Besides, we note that the homeomorphism group of
$[L]$-dimensional space whose nonempty open sets are universal for
the class of separable compact spaces of extensional
dimension $\leq[L]$ is totally disconnected.
Key words and phrases:Extension dimension, universal space, semigroup
1991 Mathematics Subject Classification: 54H15, 54F45, 55M10
The first author was partially supported by NSERC research grant.
1. Preliminaries
Let $L$ be a CW-complex and $X$ a Tychonov space. The Kuratowski notation $X\tau L$ means that, for any continuous map
$f\colon A\to L$ defined on a closed subset $A$ of $X$, there
exists an extension $\bar{f}\colon X\to L$ onto X. This notation
allows us to define the preorder relation $\preceq$ onto the class
of CW-complexes: $L\preceq L^{\prime}$ iff, for every Tychonov space $X$,
$X\tau L$ implies $X\tau L^{\prime}$ [1].
The preorder relation $\preceq$ naturally generates the equivalence
relation $\sim$: $L\sim L^{\prime}$ iff $L\preceq L^{\prime}$ and $L^{\prime}\preceq L$. We
denote by $[L]$ the equivalence class of $L$.
The following notion is introduced by A. Dranishnikov (see, [5] and [4]). The
extension dimension of a Tychonov space $X$ is less than or
equal to $[L]$ (briefly, $\mathop{\mathrm{ext-dim}}(X)\leq[L]$) if $X\tau L$.
We say that a Tychonov space $Y$ is said to be a universal space for
the class of compact
metric spaces $X$ with $\mathop{\mathrm{ext-dim}}(X)\leq[L]$ if $Y$ contains a topological copy
of every compact
metric space $X$ with $\mathop{\mathrm{ext-dim}}(X)\leq[L]$. See [1] and [2] for
existence of universal spaces.
In what follows we will need the following statement which appears in [3] as Lemma 3.2.
Proposition 1.1.
Let $i_{0}=\min\{i:\pi_{i}(L)\neq 0\}$. Then $\mathop{\mathrm{ext-dim}}(S^{i_{0}})\leq[L]$.
2. Main theorem
Recall that a semigroup $S$ (whose operation is denoted as multiplication)
is called a left cancelation
semigroup if $xy=xz$ implies $y=z$ for every $x,y,z\in S$.
Theorem 2.1.
Let $L$ be a connected CW-complex and Let $Y$ be a universal space for the class of compact
metric spaces $X$ with $\mathop{\mathrm{ext-dim}}(X)\leq[L]$. If $\mathop{\mathrm{ext-dim}}(Y)=[L]$, then there
is no structure of left (right) cancelation semigroup on $Y$
compatible with its topology.
Proof.
Suppose the contrary and let $Y$ be a left cancelation semigroup.
Let $\alpha(\coprod_{j=1}^{\infty}S^{i_{0}}_{j})$ be the Alexandrov
compactification of the countable topological sum of copies $S^{i_{0}}_{j}$ of
the sphere
$S^{i_{0}}$, where $i_{0}=\min\{i:\pi_{i}(L)\neq 0\}$. By the countable sum
theorem for extension dimension and Proposition 1.1, $\mathop{\mathrm{ext-dim}}(\alpha(\coprod_{j=1}^{\infty}S^{i_{0}}_{j}))\leq[L]$ and, since $Y$ is universal,
$Y$ contains a copy of
$\alpha(\coprod_{j=1}^{\infty}S^{i_{0}}_{j})$. We will assume that
$\alpha(\coprod_{j=1}^{\infty}S^{i_{0}}_{j})\subset Y$. Besides, since
$\mathop{\mathrm{ext-dim}}(Y)\geq[S^{1}]$, we see that $Y$ contains an arc $J$. Let $a,b$ be
endpoints of $J$. There exists $j_{0}$ such that $aS^{i_{0}}_{j_{0}}\cap bS^{i_{0}}_{j_{0}}=\emptyset$. By Proposition 1.1, there exists
a map $f\colon aS^{i_{0}}_{j_{0}}\cup bS^{i_{0}}_{j_{0}}\to L$ such that
$f|aS^{i_{0}}_{j_{0}}$ is a constant map and $f|bS^{i_{0}}_{j_{0}}$ is not
null-homotopic. Extend map $f$ to a map $\bar{f}\colon Y\to L$. Let
$g\colon[0,1]\to J$ be a homeomorphism, then the map $F\colon S^{i_{0}}_{j_{0}}\times[0,1]\to L$, $\bar{f}(x,t)=g(t)x$, is a homotopy
that contradicts to the fact that $f|bS^{i_{0}}_{j_{0}}$ is not
null-homotopic.
∎
The homeomorphism group $\mathrm{Homeo}(X)$ of a space $X$ is
endowed with the compact-open topology.
Theorem 2.2.
Suppose $\mathop{\mathrm{ext-dim}}(X)=[L]$ and every nonempty open subset of $X$ is
universal for the class of separable metric spaces $X$ with
$\mathop{\mathrm{ext-dim}}(X)\leq[L]$. Then the homeomorphism group $\mathrm{Homeo}(X)$ is
totally disconnected.
Proof.
Suppose the contrary. Let $h\in\mathrm{Homeo}(X)$,
$h\neq\mathrm{id}_{X}$. There exists $x\in X$ such that $h(x)\neq x$
and, therefore, there exists a neighborhood $U$ of $x$ such that
$h(U)\cap U=\emptyset$. Since $U$ is universal for the class of
separable metric spaces $X$ with $\mathop{\mathrm{ext-dim}}(X)\leq[L]$, there exists an
embedding of $S^{i_{0}}$ into $U$, where $i_{0}$ is as in Proposition
1.1. We may suppose that $S^{i_{0}}\subset U$. There exists a
map $f\colon S^{i_{0}}\cup h(S^{i_{0}})\to L$ such that the restriction
$f|S^{i_{0}}$ is not null-homotopic while the restriction
$f|h(S^{i_{0}})$ is null-homotopic. Since $\mathop{\mathrm{ext-dim}}(X)\leq[L]$, there exists an
extension $\bar{f}\colon X\to L$ of the map $f$. The set
$$W=\{g\in\mathrm{Homeo}(X):\bar{f}|g(S^{i_{0}})\text{ is not null-homotopic }\}$$
is an open and
closed subset of $\mathrm{Homeo}(X)$. We see that $W$ is a neighborhood
of unity that does not contain $h$.
∎
3. Open problems
Note that the case $L=S^{n}$ corresponds to the case of covering dimension.
In this case, the topology of homeomorphism groups of some universal spaces has been
investigated by many authors (see the survey [6]).
In particular, it is known (see [8] and [7]) that the homeomorphism
group of the $n$-dimensional Menger compactum $M^{n}$ (note that $M^{n}$ satisfies the conditions of
Theorem 2.2 with $L=S^{n}$) is one-dimensional.
Let $[L]\geq[S^{1}]$ and $X$ be as in Theorem 2.2. Is
$\dim(\mathrm{Homeo}(X))\geq 1$?
Another version: Is there $X$ that satisfies the conditions of
Theorem 2.2 and such that $\dim(\mathrm{Homeo}(X))\geq 1$?
References
[1]
A. Chigogidze, Cohomological dimension of Tychonov spaces. Topology Appl. 79 (1997), no. 3, 197–228.
[2]
A. Chigogidze, V. Valov, Universal metric spaces and extension dimension.
Geometric topology: Dubrovnik
1998. Topology Appl. 113 (2001), no. 1-3, 23–27.
[3]
A. Chigogidze, M. Zarichnyi, On absolute extensors modulo a complex. Topology Appl. 86 (1998), no. 2,
169–178.
[4]
A. Dranishnikov, J. Dydak, Extension dimension and extension types. Tr. Mat. Inst. Steklova 212 (1996),
Otobrazh. i Razmer., 61–94; translation in Proc. Steklov Inst. Math. 1996, no. 1 (212), 55–88
[5]
A. N. Dranishnikov, The Eilenberg-Borsuk theorem for mappings in an arbitrary
complex. (Russian) Mat. Sb. 185
(1994), no. 4, 81–90.
[6]
Chigogidze, Alex; Kawamura, Kazuhiro; Tymchatyn, E. D. Menger manifolds. Continua (Cincinnati, OH, 1994),
37–88, Lecture Notes in Pure and Appl. Math., 170, Dekker, New York, 1995.
[7]
L. G. Oversteegen, E. D. Tymchatyn, On the dimension of certain totally disconnected spaces. Proc. Amer. Math.
Soc. 122 (1994), no. 3, 885–891.
[8]
B. L. Brechner, On the dimensions of certain spaces of homeomorphisms.
Trans. Amer. Math. Soc. 121 (1966)
516–548. |
A Panoramic 3D Vision System from Multiple RGBD Sensors
Hang Liu${}^{1}$, Hengyu Li${}^{1*}$, Xiahua Liu${}^{2}$, Jun Luo${}^{1}$, Shaorong Xie${}^{1}$, Yu Sun${}^{1,3}$
${}^{1}$Hang Liu, Hengyu Li, Jun Luo, Shaorong Xie with the School of Mechatronic Engineering and Automation, Shanghai University, China. Hengyu Li is the corresponding author. [email protected]@shu.edu.cn${}^{2}$Xiahua Liu with the Department of Engineering Mathematics, University of Bristol.${}^{3}$Yu Sun with the Department of Mechanical and Industrial Engineering, University of
Toronto, Canada, and the Shanghai University. [email protected] work was supported by the National Natural Science Foundation of China (grant numbers 61525305 and 61625304), the Shanghai Natural Science Foundation (grant numbers 17ZR1409700 and 18ZR1415300), and the basic research project of Shanghai Municipal Science and Technology Commission (grant number 16JC1400900).
Abstract
This letter presents a panoramic 3D vision system that was designed specifically for mapping the indoor environment. The system consists of multiple RGBD sensors that cover a 360${}^{\circ}$ field of view, simplifying the reconstruction of indoor environments. A new extrinsic calibration approach is proposed to estimate the relative poses between the RGBD sensors. This approach relies on well-matched keypoints from a descriptor-based pattern, in this case, of a minimally overlapping field of view between sensors. A distributed capturing system was also designed to provide panoramic RGBD frames in real time. Experiments validate the accuracy of the proposed calibration approach and the efficiency of the proposed panoramic 3D vision system for indoor mapping.
I INTRODUCTION
In recent years, indoor scene reconstruction and robot navigation have attracted much attention with the advent of low-cost and efficient depth and color (RGBD) devices such as the Microsoft Kinect, Intel RealSense, and Structure Sensor. The depth sensors of these devices can provide a depth map with a VGA resolution (640x480) at video-rate (e.g., 30 Hz) using efficient light-coding technology that avoids the challenging task of dense 3D reconstruction from color images. The 3D models reconstructed using these depth sensors have been used to generate more realistic 3D content for virtual reality (VR) and help align the rendered virtual objects with real scenes for augmented reality (AR). Furthermore, the direct depth sensing capability of these depth sensors are particularly suitable for helping robots manipulate and navigate in an unknown environment.
With an RGBD sensor, the simultaneous localization and mapping (SLAM)-based approach is mainly used for fusing the point cloud frames to reconstruct the indoor scenes [1, 2, 3]. However, hundreds or even thousands of frames must be captured in state-of-the-art SLAM systems to reconstruct a common indoor environment such as a room or an office [4] because of two problems:
1) The field of view (FOV) of depth sensors is limited; thus, only a small part of the scene is represented in a single frame. The Kinect, for example, has a horizontal FOV of 57${}^{\circ}$, which is much smaller than the horizontal 240${}^{\circ}$ FOV of the Hokuyo URG-04LX-UG01, a laser scanner with a similar range and measurement accuracy compared with the Kinect [5].
2) To track the poses of depth sensors to effectively fuse multiple point cloud frames, consecutive captured frames should be captured to have sufficient scene overlap. Typically, more than ninety percent of overlap is required, which further increases the number of frames for reconstruction.
A solution to the problem is to use a multicamera setup in which the cameras face different directions to sample different sections of the environment. A typical multicamera setup consists of multiple color cameras [6, 7] or RGBD cameras [8, 9].
In this letter, we focus on a panoramic 3D vision system that consists of multiple RGBD sensors for indoor mapping. A new extrinsic calibration method that relies on a descriptor-based pattern to provide well-matched keypoints is proposed to estimate the relative poses between the RGBD cameras with minimal overlapping FoV. A distributed capturing system based on local area network (LAN) is also proposed to provide multiple RGBD frames in real time. The experiments using real scenes validate these features of our study.
II RELATED WORK
Panoramic cameras cover a wider field of view and therefore are applied in mapping systems to provide better localization constraints that can improve the accuracy of structure and motion estimation. One type of panoramic camera is the central catadioptric camera [10], which uses a combination of lenses and mirrors placed in a carefully arranged configuration to capture a much wider FOV. This type of camera can capture 360${}^{\circ}$ with a single camera and therefore is widely used in SLAM systems [11, 12]. Unfortunately, this type of camera has a poor and varying spatial resolution, lowering the measurement accuracy of a single camera pixel [13].
Another approach is to use a multicamera panoramic camera system. Although there are some complications from using multiple cameras, such as additional image data and more difficult calibration, the main advantage of a multicamera system is that it enables the construction of a high resolution 360${}^{\circ}$ image. Studies using this type of multicamera systems appear in [14, 15].
SLAM systems that contain only color cameras can only compute sparse or semidense maps. However, applications such as augmented reality or obstacle avoidance for robots require a dense map of the environment. Thus, some range sensors are introduced in panoramic SLAM systems from laser range finders [11, 16] to stereo matching with triangulation [17] and structured light RGBD systems [18, 19]. Laser approaches are expensive and computing the depth from stereo image pairs is time-consuming. The structured light RGBD systems are low cost and have acceptable measuring accuracy in close ranges, which are suitable for direct 3D SLAM systems in indoor environments.
Complications can occur with the use of multiple cameras, such as more difficult calibration and additional image data. Fernandez-Moral et al. proposed to calibrate multiple RGBD sensors by finding and matching planes [8] or lines [20]. However, new algorithms for extracting and matching these planes or lines require redesign. The reason why classical extrinsic calibration strategies such as the chessboard-based method and the keypoints-based method cannot be applied is because the overlapping FoV between RGBD sensors cannot meet the overlap requirements of these methods. Chessboard-based methods require that chessboards be completely visible in the views of all cameras that must be calibrated. If the overlapping FoV is too small (e.g., thirty percent overlap), the overlapping image regions cannot provide correctly matched keypoints. Because the $N$ number of cameras in a panoramic vision system provide $N$ times the image data, it is also important to find effective ways to efficiently capture and transfer all the data.
In the following, we give the details of our calibration approach. Then, the multi-RGBD camera setup and a LAN-based distributed capturing system are presented in section IV. In addition, we analyze the precision of the proposed calibration approach and show the efficiency of indoor scene reconstruction using the proposed panoramic 3D vision system. Finally, the conclusions are summarized.
III EXTRINSIC CALIBRATION
In this section, we address the problem of estimating extrinsic calibration (i.e., relative poses) between RGBD sensors that have little overlapping FoV. Fig. 1 shows a setup of the proposed panoramic 3D vision system that consists of 12 Kinect v1 sensors. All Kinects are vertically positioned for a more compact design. The Kinect v1 sensor has an angular field of view (FoV) of 43${}^{\circ}$ from the vertical. The overlap FoV of two neighboring sensors is only approximately 30 percent of the vertical FoV of each sensor.
We propose to solve the calibration problem of multiple RGBD sensor with little overlapping FoV using a feature descriptor-based calibration pattern [21], which can provide robust and accurate matched feature points in this case of minimal overlapping FoV. Based on these matched 2D feature points and depth maps from the depth sensors, we construct two 3D point sets to estimate poses by bundle adjustment. Then, a pose graph optimization method is used to refine the estimated poses.
III-A Initial estimation of poses
The descriptor-based calibration pattern is composed of several noise images at different scales in accordance with the mechanism of SIFT/SURF. Compared with natural scenes, this pattern contains a high number of detectable features that can be easily detected by a camera at varying distances. Thus, the descriptor-based pattern can provide many and more accurately matched keypoints between two cameras in the case of minimal overlapping FoV.
In Fig. 2, the detected SURFT keypoints of such a pattern is represented by green circles. Fig. 2 shows the feature detection results of a nature scene. There are many more keypoints in Fig. 2. Fig. 2 and Fig. 2 show the matched keypoints of a pair of images captured by two cameras with an approximately thirty percent overlapping FoV. Well-matched keypoints are connected by green lines; poorly matched keypoints are connected by red lines. The matching results are generated by the Flann-based descriptor matcher implemented in OpenCV, where matched keypoints are accepted only if its descriptor distance is less than 2.0 times the minimum descriptor distances. Fig. 2 shows that the descriptor-based pattern provides more well-matched keypoints. In this work, the keypoints detector can only be applied to the overlapping regions of neighboring images to reduce the possibility of mismatch.
Then, based on these matched keypoints, we must find their corresponding depth values from the depth maps generated by the depth camera in the RGBD sensor to construct 3D point sets to estimate the poses. Due to the different spatial positions and intrinsic parameters of the IR camera of the depth sensor and of the color camera in the RGBD sensor, the depth map is not aligned with the color image. Because the depth and color cameras share enough overlapping FoV to detect the entire chessboard, we use the highly accurate chessboard calibration method to obtain the relative poses between the color and depth camera and their intrinsic parameters to align the depth map with the color image.
Let $[u,v]$ represent the coordinates of a keypoint, $Z$ represents its depth value in $[u,v]$ of corresponding aligned depth map, and $[X,Y,Z]^{T}$ represent the 3D point of $[u,v]$ in the color camera coordinate system. According to the pinhole camera model, the values of $X$ and $Y$ can be calculated as
$$\begin{array}[]{l}X{\rm{=}}(u-{u_{0}})Z/{f_{x}},\\
Y=(v-{v_{0}})Z/{f_{y}}.\end{array}$$
(1)
We obtain two 3D point sets $\{p_{i}\}$, $\{p^{\prime}_{i}\}$; $i=1,2,...,N$, based on $(\ref{eq1})$. $p_{i}$ and $p^{\prime}_{i}$ are $3\times 1$ column matrices. The relative poses between these two 3D point sets can be found by minimizing
$$\mathop{\min}\limits_{\xi}=\frac{1}{2}\sum\limits_{i=1}^{n}{\left\|{\left({{{p%
^{\prime}}_{i}}-\exp(\xi\hat{~{}}){p_{i}}}\right)}\right\|},$$
(2)
where $\xi\in se(3)$ is a vector with six dimensions that represents the camera pose, $\hat{~{}}$ maps $\xi$ to a matrix with four mentions ${R^{4\times 4}}$. $\xi\hat{~{}}$ is mapped to $T\in SE(3)$ by the exponential map
$exp()$. We use bundle adjustment to jointly solve all camera poses [22]. The derivative with respect to the camera pose of an error element is given by
$$-(\exp(\xi\hat{~{}})p)\hat{~{}}$$
(3)
where $\hat{~{}}$ denotes the mapping from ${\exp(\xi\hat{~{}})p}$ represented by $[x,y,z]^{T}$ to corresponding antisymmetric matrix
$\left[{\begin{array}[]{*{20}{c}}0&{-z}&y\\
z&0&{-x}\\
{-y}&x&0\end{array}}\right]$, the detailed derivation process can be found in the appendices.
III-B Pose Graph Optimization
Assume that there are $N$ RGBD sensors in the panoramic 3D vision system. Let $T_{i,j}$ denote the estimated relative poses between two adjacent RGBD sensors, $i=1,2,...,N$; $j=i+1$, when $i$ equals $N$, $j$ is 1. When the pose of the first RGBD sensor is set to $[0,0,0,1]^{T}$, we can calculate the pose $x_{i}$ for each RGBD sensor based on $T_{i,j}$. Due to the pose estimation error, the pose that is calculated by ${T_{N,1}}{x_{N}}$ is not equal to $[0,0,0,1]^{T}$. Fortunately, this panoramic setup of RGBD sensors provides a definite loop-closure constraint for optimizing the estimated poses. We can solve this problem by using the popular pose graph optimization method in SLAM [22].
According to pose graph optimization theory, the problem can be solved by finding the minimum of a function of this form:
$${x^{*}}=\arg{\kern 1.0pt}{\kern 1.0pt}\mathop{\min}\limits_{x}\sum\limits_{(i,%
j)\in S}{r_{i,j}^{T}}(x){\Lambda_{i,j}}r{}_{i,j}(x),$$
(4)
where $x={\left({x_{1}^{T},{\kern 1.0pt}...,x_{n}^{T}}\right)^{T}}$ is a vector of poses, $r{}_{i,j}$ is the residual of the predicted and observed relative poses between the $i-$th and $j-$th node, ${\Lambda_{i,j}}$ denotes the measurement information matrix, and $S$ represents the set of edges that connect the nodes.
IV EXPERIMENTS AND RESULTS
In this section, we evaluate the accuracy of the proposed extrinsic calibration method and show the efficiency of constructing indoor environments using the panoramic 3D vision system.
IV-A System Setup
We designed a camera rig that consists of twelve Kinect v1 RGBD sensors mounted in a radial configuration (see Fig. 1). We also designed a distributed system built on top of a local area network (LAN) to capture the RGBD frames from all twelve sensors in real time. The distributed capturing system (see Fig. 3) comprises twelve Raspberry Pi single board computers, a gigabit switch and a high-performance computer. Raspberry Pi is used to obtain the RGBD frames from each Kinect and send the data to the high-performance computer through the LAN using User Datagram Protocol. The high-performance computer is used to receive and process the RGBD frames. Both the depth and color frames are set to a size of $640\times 480$; the values of a single pixel in the depth frame and color frame have a size of 2 bytes and 3 bytes, respectively. Thus, the size of the RGBD frames from all twelve Kinects is 17.58 million bytes, which can be sent to the high-performance computer through the gigabit switch at 7 fps. This frame rate can be increased by decreasing the size of RGBD frames from Kinect.
IV-B Precision of Extrinsic Calibration
To evaluate the precision of the proposed extrinsic calibration method, we use a motion capture system to obtain the ground truth relative poses between the sensors. The motion capture system requires at least three reflective markers to track the pose of a rigid body, such as the Kinect and the chessboard in our experiments. We attached four reflective markers to both the Kinect and the chessboard (see Fig. 4). We placed the four markers on the outer corners of the chessboard, such that the relative poses between the chessboard and the motion capture system and the Kinect color camera are known. The motion capture system can track the poses of the markers that are attached to the Kinect to determine the poses of the Kinect color camera.
For convenience, we only attached markers to one Kinect and placed this Kinect on twelve different positions in the camera rig to capture RGD frames. The precision of the extrinsic calibration is evaluated using twelve pairs of RGBD frames from the captured twelve RGBD frames.
While determining the keypoints correspondences for each pair of RGBD frames, we increase the distance threshold of the keypoints (denoted by $dist\_thresh$) from one and a half times the minimum distance to the maximum distance (ten times the minimum distance in average) to analyze the calibration error in rotation and translation with respect to the keypoints correspondences.
In TABLE I, the average residual error in rotation and translation is presented for the initial estimation (denoted by Ini.) and optimized estimation (denoted by Opti.) using pose graph optimization with loop closure. It can be seen that the average residual error is reduced when raising the distance threshold to consider more keypoint correspondences. The residual error in both the rotation and translation for the optimized estimation is generally less than the residual error for the initial estimation. When the distance threshold is set to be larger than four times the minimum distance, the residual error remains at the same level. The aligned point clouds before and after pose graph optimization with loop closure can also be observed in Fig. 5. It can be seen in the red circle on the bottom of Fig. 5 that the point clouds are misaligned, however, these point clouds align well in Fig. 5 with optimized poses with loop closure constraint.
The operating range of the Kinect v1 depth sensor is between 0.5 m to 5.0 m [23]. Using an output panoramic frame can reconstruct scenes within the circle with a radius of 5 meters. Thus the reconstruction of some indoor scenes becomes very efficient using the proposed panoramic 3D vision system. Fig. 6 presents the reconstruction result of a bedroom (3.3 m $\times$ 3 m) and a living room (9 m $\times$ 3.5 m) using only one output panoramic frame from the proposed system.
V CONCLUSION
In this letter, we present a panoramic 3D vision system from multiple RGBD sensors. A new methodology that relies on well-matched keypoints provided by a feature descriptor-based calibration pattern has been proposed to calibrate the extrinsic parameters of the RGBD sensors in the system. We have also designed a LAN-based distributed system that enables the system to provide panoramic RGBD frames in real time. The reconstruction of indoor scenes is efficiently and conveniently performed using this panoramic 3D vision system considering the time consumption and captured RGBD frames. All the experiments have validated the accuracy of the proposed calibration method and the efficiency in reconstruction.
Directions for future work include research on integrating the proposed panoramic 3D vision system with state-of-the art SLAM systems for localization and mapping at large scales. The proposed system provides a 360${}^{\circ}$ FoV, leading to better constraints for localization and can be potentially used to reduce localization and mapping errors. It is particularly important to investigate direct registration methods such as [24, 25], which do not depend on time-consuming keypoint detectors or descriptors for large scale SLAM, using this panoramic 3D vision system.
APPENDIX
The detailed derivation process of the derivative with respect to the camera pose of a error element is as follows,
$$\begin{array}[]{l}\frac{{\partial(\exp(\xi\hat{~{}})p)}}{{\partial\xi}}=%
\mathop{\lim}\limits_{\delta\xi\to 0}\frac{{\exp(\delta\xi\hat{~{}})\exp(\xi%
\hat{~{}})p-\exp(\xi\hat{~{}})p}}{{\delta\xi}}\\
{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}%
{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}%
{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}\approx%
\mathop{\lim}\limits_{\delta\xi\to 0}\frac{{(I+\delta\xi\hat{~{}})\exp(\xi\hat%
{~{}})p-\exp(\xi\hat{~{}})p}}{{\delta\xi}}\\
{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}%
{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}%
{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}=\mathop{\lim%
}\limits_{\delta\xi\to 0}\frac{{\delta\xi\hat{~{}}\exp(\xi\hat{~{}})p}}{{%
\delta\xi}}=\mathop{\lim}\limits_{\delta\xi\to 0}\frac{{-(\exp(\xi\hat{~{}})p)%
\hat{~{}}\delta\xi}}{{\delta\xi}}\\
{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}%
{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}%
{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}{\kern 1.0pt}=-(\exp(\xi%
\hat{~{}})p)\hat{~{}}\end{array}$$
(5)
References
[1]
R. Mur-Artal and J. D. Tardos, “Orb-slam2: An open-source slam system for
monocular, stereo, and rgb-d cameras,” IEEE Transactions on Robotics,
vol. 33, no. 5, pp. 1255–1262, 2017.
[2]
T. Whelan, R. F. Salas-Moreno, B. Glocker, A. J. Davison, and
S. Leutenegger, “Elasticfusion: real-time dense slam and light source
estimation,” The International Journal of Robotics Research, vol. 35,
no. 14, pp. 1697–1716, 2016.
[3]
C. Kerl, J. Sturm, and D. Cremers, “Dense visual slam for rgb-d
cameras,” in 2013 IEEE/RSJ International Conference on Intelligent
Robots and Systems, 2013, pp. 2100–2106.
[4]
J. Sturm, N. Engelhard, F. Endres, W. Burgard, and D. Cremers, “A benchmark
for the evaluation of rgb-d slam systems,” in Proc. of the
International Conference on Intelligent Robot Systems (IROS), Oct. 2012.
[5]
S. Zug, F. Penzlin, A. Dietrich, T. T. Nguyen, and S. Albert, “Are laser
scanners replaceable by kinect sensors in robotic applications?” in
Robotic and Sensors Environments (ROSE), 2012 IEEE International
Symposium on. IEEE, 2012, pp.
144–149.
[6]
L. Heng, G. H. Lee, and M. Pollefeys, “Self-calibration and visual slam with a
multi-camera system on a micro aerial vehicle,” Autonomous Robots,
vol. 39, no. 3, pp. 1–19, 2015.
[7]
M. J. Tribou, A. Harmat, D. W. L. Wang, I. Sharf, and S. L. Waslander,
“Multi-camera parallel tracking and mapping with non-overlapping fields of
view,” International Journal of Robotics Research, vol. 34, no. 12,
2015.
[8]
E. Fernandez-Moral, J. Gonzalez-Jimenez, P. Rives, and V. Arevalo, “Extrinsic
calibration of a set of range cameras in 5 seconds without pattern,” in
Ieee/rsj International Conference on Intelligent Robots and Systems,
2014, pp. 429–435.
[9]
P. Rives, “Scene structure registration for localization and mapping,”
Robotics and Autonomous Systems, vol. 75, no. PB, pp. 649–660, 2016.
[10]
S. K. Nayar, “Catadioptric omnidirectional camera,” in Proceedings of
IEEE Computer Society Conference on Computer Vision and Pattern Recognition,
1997, pp. 482–488.
[11]
C. Zou, B. He, L. Zhang, J. Zhang, and Z. Deng, “An automatic
calibration between an omni-directional camera and a laser rangefinder for
dynamic scenes reconstruction,” in 2016 IEEE International Conference
on Robotics and Biomimetics (ROBIO), 2016, pp. 1528–1534.
[12]
M. Schönbein and A. Geiger, “Omnidirectional 3d reconstruction in
augmented manhattan worlds,” in 2014 IEEE/RSJ International Conference
on Intelligent Robots and Systems (IROS 2014), 2014, pp. 716–723.
[13]
Z. Zhang, H. Rebecq, C. Forster, and D. Scaramuzza, “Benefit of large
field-of-view cameras for visual odometry,” in 2016 IEEE International
Conference on Robotics and Automation (ICRA), 2016, pp. 801–808.
[14]
M. Kaess and F. Dellaert, “Probabilistic structure matching for visual
slam with a multi-camera rig,” Computer Vision and Image
Understanding, vol. 114, no. 2, pp. 286–296, 2010.
[15]
J.-P. Tardif, Y. Pavlidis, and K. Daniilidis, “Monocular visual odometry
in urban environments using an omnidirectional camera,” in 2008
IEEE/RSJ International Conference on Intelligent Robots and Systems, 2008,
pp. 2531–2538.
[16]
P. D. Gashongore, K. Kawasue, K. Yoshida, and R. Aoki, “Indoor space
3d visual reconstruction using mobile cart with laser scanner and cameras,”
in Eighth International Conference on Graphic and Image Processing
(ICGIP 2016), vol. 10225, 2017.
[17]
M. Meilland, A. I. Comport, and P. Rives, “Dense omnidirectional rgb-d
mapping of large-scale outdoor environments for real-time localization and
autonomous navigation,” Journal of Field Robotics, vol. 32, no. 4,
pp. 474–503, 2015.
[18]
T. Gokhool, R. Martins, P. Rives, and N. Despre, “A compact spherical
rgbd keyframe-based representation,” in 2015 IEEE International
Conference on Robotics and Automation (ICRA), 2015, pp. 4273–4278.
[19]
T. Gokhool, M. Meilland, P. Rives, and E. Fernández-Moral, “A dense
map building approach from spherical rgbd images,” in 2014
International Conference on Computer Vision Theory and Applications
(VISAPP), 2014, pp. 656–663.
[20]
A. Perez-Yus, E. Fernandez-Moral, G. Lopez-Nicolas, J. J. Guerrero, and
P. Rives, “Extrinsic calibration of multiple rgb-d cameras from line
observations,” IEEE Robotics and Automation Letters, vol. PP, no. 99,
pp. 1–1, 2017.
[21]
B. Li, L. Heng, K. Koser, and M. Pollefeys, “A multiple-camera system
calibration toolbox using a feature descriptor-based calibration pattern,”
in Ieee/rsj International Conference on Intelligent Robots and
Systems, 2014, pp. 1301–1307.
[22]
R. Kummerle, G. Grisetti, H. Strasdat, K. Konolige, and W. Burgard, “G 2 o: A
general framework for graph optimization,” pp. 3607–3613, 2011.
[23]
K. Khoshelham and S. O. Elberink, “Accuracy and resolution of kinect depth
data for indoor mapping applications,” Sensors, vol. 12, no. 2, p.
1437, 2012.
[24]
J. Engel, V. Koltun, and D. Cremers, “Direct sparse odometry,” IEEE
transactions on pattern analysis and machine intelligence, vol. 4, 2017.
[25]
C. Forster, M. Pizzoli, and D. Scaramuzza, “Svo: Fast semi-direct monocular
visual odometry,” in Robotics and Automation (ICRA), 2014 IEEE
International Conference on. IEEE,
2014, pp. 15–22. |
Spin-selective optical absorption of singly charged excitons in a quantum dot
Alexander Högele
Martin Kroner
Stefan Seidl
Khaled Karrai
Center for NanoScience, Department für Physik, Ludwig-Maximilians-Universität,
Geschwister-Scholl-Platz 1, 80539 München, Germany
Mete Atatüre
Jan Dreiser
Atac Imamoğlu
Institute of Quantum Electronics, ETH Hönggerberg HPT G12, CH-8093 Zürich, Switzerland
Richard J. Warburton
School of Engineering and Physical Sciences,
Heriot-Watt University, Edinburgh EH14 4AS, UK
Antonio Badolato
Brian D. Gerardot
Pierre M. Petroff
Materials Department, University of California, Santa Barbara, California 93106, USA
(October 14, 2004)
Abstract
We report high resolution laser absorption spectroscopy of a
single InGaAs/GaAs self-assembled quantum dot embedded in a
field-effect structure. We show experimentally that the interband
optical absorption to the lower Zeeman branch of the singly
charged exciton is strongly inhibited due to spin (Pauli) blockade
of the optical transition. At high magnetic fields the optical
absorption to the upper Zeeman branch dominates the absorption
spectrum. We find however that the spin blockade is not complete
and a 10% leakage remains at high magnetic fields. Applying a
gate voltage to empty the dot of its resident electron turns the
spin blockade off. This effect is observed at 1.5 K and up to
9 Tesla.
The coherence time of an excess electron spin strongly confined in
a quantum dot (QD) structure is expected to be orders of magnitude
longer than the typical timescales required for its coherent
manipulation Golovach ; Hanson ; Kroutvar ; Elzerman . Motivated
by this observation, several groups have proposed to use single QD
spins as quantum bits (qubits) Loss , and to manipulate,
couple and measure individual spins using either transport
Elzerman or optical techniques Imamoglu . In the case
of self-assembled InGaAs QDs with strong confinement along the
growth direction, the lowest energy optical transitions are those
arising from the excitation of a $J_{z}$=$+\frac{3}{2}$
$(J_{z}$=$-\frac{3}{2})$ valence electron to a
$S_{z}$=$+\frac{1}{2}$ $(S_{z}$=$-\frac{1}{2})$ conduction state.
If the QD already has an excess conduction electron, only one of
these optical transitions is allowed; the other is spin (Pauli)
blocked Warburton1 . In contrast, a neutral QD with
asymmetric confinement potential always has a pair of exciton
transitions. It has been suggested that Pauli blocking of
absorption or fluorescence can be used to implement high
efficiency all-optical single-spin measurements
Imamoglu ; Calarco and conditional spin dynamics
Calarco .
In this Letter, we report resonant absorption measurements on a
single QD charged with a single excess electron in the regime of
Pauli blocking. We observe that for high magnetic fields
(B$>$5 Tesla) where the electron is (with high probability) in the
lowest energy $S_{z}$=$+\frac{1}{2}$ state, the absorption of a
left-hand-circularly polarized ($\sigma^{-}$) laser is suppressed
by a factor of 10 as compared to that of a right-hand-circularly
polarized ($\sigma^{+}$) laser. In contrast, we observe that both
$\sigma^{+}$ and $\sigma^{-}$ transmission dips have equal
strength for a neutral QD, irrespective of the applied magnetic
field. Using the fact that the strength of $\sigma^{+}$ and
$\sigma^{-}$ absorption in a charged QD is a measure of the spin
polarization of the resident electron, we were able to determine
the electron g-factor by fitting the magnetic field dependence of
the transmission data. Our results represent a first step towards
all-optical coherent spin measurement.
The InGaAs dots investigated in this work were self-assembled
in the Stranski-Krastanow growth mode by molecular beam
epitaxy. The QDs are separated from a highly n-doped GaAs back contact by 25 nm of intrinsic GaAs which acts as a tunnel barrier. The electrons are prevented from tunneling to the gate
electrode, the metalized sample surface, by a 110 nm thick
GaAs/AlAs superlattice blocking barrier. The whole structure forms
a field-effect device. The electron occupation of the dots is controlled by
applying a gate voltage and monitored with the QD photoluminescence charging diagrams
Warburton2 . An individual QD is identified to be spectrally
separated by typically more than 5 meV from neighboring dots assuring that the observed features originate
unambiguously from a single QD. We used high resolution laser
spectroscopy and a low temperature confocal microscope to measure
the differential transmission through the QDs. This technique Alen ; PhysE ; PRL gives optical spectra as shown in
Fig. 1. The spectral resolution is much less than the QD
absorption width. A dip in the transmission is obtained when the resonantly scattered laser light is diverted away from the photodetector placed immediately behind the sample.
In the absence of a magnetic field, a negatively charged exciton
(X${}^{1-}$) strongly confined in a QD shows a single unpolarized
absorption peak in the optical interband spectrum, as demonstrated
in Fig. 1 with high resolution transmission spectroscopy.
In the case of X${}^{1-}$, electrons are forced to fill the lowest
energy $s$-shell such that the total electron spin is 0 in
accordance with the Pauli principle. As a consequence, the
exchange interaction between the confined electrons and the hole
is suppressed so that the optical spectrum is free of fine
structure features irrespective of the asymmetry of the confining
potential PRL ; Bayer1 . This is not the case for the neutral
exciton X${}^{0}$: the electron-hole exchange leads generally to a
splitting into two linearly polarized dipole-allowed optical
transitions PRL ; Gammon ; Bayer2 . For the dot investigated
here the measured neutral exciton splitting was 10 $\mu$eV.
In a high magnetic field and at low enough temperatures, the
ground state of a singly charged QD is that of a spin-polarized
electron in the lowest Zeeman level. As shown in Fig. 2a,
an interband photon can only be absorbed under the condition that
the photo-generated electron is of opposite spin. In principle, at
T = 0 K this translates into a single absorption line in the
optical spectrum irrespective of the applied magnetic field. In
spite of the applied magnetic field, the exciton absorption should
not show a Zeeman splitting, i.e. the optical transition at the
lower exciton Zeeman energy should be spin blocked due to the
Pauli principle independently of the photon polarization.
We have investigated Pauli blocking experimentally.
Fig. 2b shows transmission spectra in the presence of a
magnetic field. At T = 1.5 K and under magnetic field two
well-resolved transmission resonances are observed. Both lines
correspond to the two optical transitions shown in
Fig. 2a. The energy positions of the resonances are
plotted in Fig. 3a showing clearly a behavior quadratic
in field, consistent with the exciton diamagnetic shift
Schulhauser . The energy splitting, plotted in
Fig. 3b, is proportional to the applied magnetic field
consistent with a Zeeman splitting of the charged exciton. We have
conducted the transmission measurements as a function of the light
polarization. At finite magnetic fields the two dipole-allowed
transitions have right and left circular polarizations.
Fig. 2b, c and d show the corresponding spectra using
linearly polarized light. At low temperatures, the strength of the
low energy resonance decreases with increasing magnetic field
while the opposite is true for the higher energy resonance.
Fig. 4a shows the relative weight of the resonances as a
function of magnetic field. For each magnetic field the
polarization was optimized to maximize independently each of the
two resonances. At high magnetic field it can clearly be seen that
the absorption of the higher energy branch dominates. The low
energy branch is dramatically inhibited, saturating at high field
at a level of one tenth of the stronger peak, without ever
completely disappearing (Fig. 4b). The high magnetic
field limit confirms roughly the expected Pauli blocking behavior
in that one resonance clearly prevails.
A further confirmation of the Pauli blockade picture is obtained
by increasing the temperature to T = 12 K. In this case, the
spectra obtained at B = 4 T show two Zeeman-split absorptions with
comparable peak strengths (Fig.s 2c and 2d). The
interpretation is that the resident electron is thermally
activated and occupies both spin states with about equal
probability. Lowering the temperature to 1.5 K clearly favors the
higher Zeeman branch confirming that the spin polarization of the
resident electron is the source of the Pauli blocking.
At T = 1.5 K the degree of spin polarization depends on the
magnetic field. We have used a two-level Boltzmann statistical
distribution for the spin state occupation. Considering that the
relative strength of both resonance peaks is a measure of the spin
polarization of the resident electron, we fitted the evolution of
the data in magnetic field, as shown in Fig. 4a. As a
fit parameter we used a Landé factor for the resident electron
of $\mid$g${}_{e}$$\mid$=1.8$\pm$0.2. This value is about twice as
large as the one reported for InGaAs/GaAs self-assembled QDs in
Ref. Bayer2 . However, several other dots emitting around 1.3 eV close to the center of the inhomogeneous QD energy distribution exhibit an electron g-factor of 0.6 as measured at 4.2 K. The deviation of the g-factor from dot to dot is not surprising since the value strongly depends on the individual In content and strain distribution. To obtain an optimal fit, we needed to
introduce empirically a limit to the maximum achievable electron
spin polarization of about 90% in order to account for the fact
that the lower energy absorption branch does not vanish at high
fields, as shown in Fig. 4b. This result indicates that
the Pauli blockade of the optical transition is not complete as
ideally predicted. It was also observed for the dots with the electron g-factor of 0.6 consistent with the temperature and the Zeeman energy. At the present time, the precise origin for the
apparent limitation of the Pauli blocking is still unknown and
requires more detailed measurements and analysis.
In conclusion, using high-resolution transmission spectroscopy we
have confirmed that a singly negatively charged exciton confined
in a single self-assembled QD has a dominant absorption line at
the higher Zeeman energy branch as a result of the spin
polarization of the resident electron. Unexpectedly, at high
fields the absorption at the lower Zeeman branch, although
drastically inhibited as expected from Pauli blocking, is still
not completely extinguished.
We would like to thank A. O. Govorov for helpful discussions.
Financial support for this work was provided in Germany by the DFG
grant no. SFB 631, in Switzerland by NCCR Quantum Photonics grant
no. 6993.1 and in the UK by the EPSRC.
References
(1)
V. N. Golovach, A. Khaetskii, and D. Loss, Phys. Rev. Lett. 93, 016601 (2004).
(2)
R. Hanson, B. Witkamp, L. M. K. Vandersypen, L. H. Willems van Beveren, J. M. Elzerman, and L. P. Kouwenhoven, Phys. Rev. Lett. 91, 196802 (2003).
(3)
M. Kroutvar, Y. Ducommun, D. Heiss, M. Bichler, D. Schuh, G. Abstreiter, and J. J. Finley, Nature (London) 432, 81 (2004).
(4)
J. M. Elzerman, R. Hanson, L. H. Willems van Beveren, B. Witkamp, L. M. K. Vandersypen, and L. P. Kouwenhoven, Nature (London) 430, 431 (2004).
(5)
D. Loss and D. P. DiVincenzo, Phys. Rev. A 57, 120 (1998).
(6)
A. Imamoğlu, D. D. Awschalom, G. Burkard, D. P. DiVincenzo, D. Loss, M. Sherwin, and A. Small, Phys. Rev. Lett. 83, 4204 (1999).
(7)
R. J. Warburton, C. S. Dürr, K. Karrai, J. P. Kotthaus, G. Medeiros-Ribeiro, and P. M. Petroff, Phys. Rev. Lett. 79, 5282 (1997).
(8)
T. Calarco, A. Datta, P. Fedichev, E. Pazy, and P. Zoller, Phys. Rev. A 68, 012310 (2003).
(9)
R. J. Warburton, C. Schäflein, D. Haft, F. Bickel, A. Lorke, K. Karrai, J. M. Garcia, W. Schoenfeld, and P. M. Petroff, Nature (London) 405, 926 (2000).
(10)
B. Alén, F. Bickel, K. Karrai, R. J. Warburton, and P. M. Petroff, Appl. Phys. Lett. 83, 2235 (2003).
(11)
A. Högele, B. Alén, F. Bickel, R. J. Warburton, P. M. Petroff, and K. Karrai, Physica E 21, 175 (2004).
(12)
A. Högele, S. Seidl, M. Kroner, K. Karrai, R. J. Warburton, B. D. Gerardot, and P. M. Petroff, Phys. Rev. Lett. 93, 217401 (2004).
(13)
M. Bayer, G. Ortner, O. Stern, A. Kuther, A. A. Gorbunov, A. Forchel, P. Hawrylak, S. Fafard, K. Hinzer, T. L. Reinecke, S. N. Walck, J. P. Reithmaier, F. Klopf, and F. Schäfer, Phys. Rev. B 65, 195315 (2002).
(14)
D. Gammon, E. S. Snow, B. V. Shanabrook, D. S. Katzer, and D. Park, Phys. Rev. Lett. 76, 3005 (1996).
(15)
M. Bayer, A. Kuther, A. Forchel, A. Gorbunov, V. B. Timofeev, F. Schäfer, J. P. Reithmaier, T. L. Reinecke, and S. N. Walck, Phys. Rev. Lett. 82, 1748 (1999).
(16)
C. Schulhauser, D. Haft, R. J. Warburton, K. Karrai, A. O. Govorov, A. V. Kalameitsev, A. Chaplik, W. Schoenfeld, J. M. Garcia, and P. M. Petroff, Phys. Rev. B 66, 193303 (2002). |
Approach for modelling quantum-mechanical collapse
A. Yu. Ignatiev
[email protected]
Theoretical Physics Research Institute,
Melbourne 3163,
Australia
Abstract
A long-standing quantum-mechanical puzzle is whether the collapse of the wave function is a real physical process or simply an epiphenomenon. This puzzle lies at the heart of the measurement problem.
One way to choose between the alternatives is to assume that one or the other is correct and attempt to draw physical, observable consequences which then could be empirically verified or ruled out.
As a working hypothesis, we propose simple models of collapse as a real physical process for direct binary symmetric measurements made on one particle.
This allows one to construct irreversible unstable
Schrödinger equations capable of describing
continuously the process of collapse induced by the interaction of the quantum system with the measuring device. Due to unknown initial conditions the collapse outcome remains unpredictable so no contradictions with quantum mechanics arise.
Our theoretical framework predicts a finite time-scale of the collapse and links with experiment.
Sensitive probes of the
collapse dynamics could be done using Bose-Einstein condensates,
ultracold neutrons or ultrafast optics.
If confirmed, the formulation could be relevant to the transition from quantum fluctuations to classical inhomogeneities in early cosmology and to establishing the ultimate limits on the speed of quantum computation and information processing.
pacs: 03.65.Ta, 03.65.Sq, 03.75.Nt, 78.47.J-
I Introduction
Modern technology has turned the in-depth studies of quantum mechanics into a fast-growing research area.
The physical meaning and the evolution of quantum state are among the focal points in these studies l ; vN ; pbr .
The wave function collapse (also known as the reduction of the wave packet) is
an irreversible, unpredictable change of quantum state which occurs during the measurement and which cannot be described by the reversible Schrödinger
equation. The nature and physical mechanisms of collapse are among the most debatable issues of quantum mechanics, and our understanding of this mysterious process remains rather limited.
as the Many-Worlds interpretation of quantum mechanics.
Collapse has been discussed from numerous, often very different perspectives, starting from purely philosophical and ending with purely mathematical. Philosophical discussions focus on various interpretations of quantum mechanics, and point out that in some of them, collapse is not a physical process but rather a (trivial) mental process of updating the available information.
This approach is often illustrated with the example of picking balls from a box with one black and one white ball. Initially, the probability to pick the white ball is 1/2. But after the first picked ball turns out black, then the probability to pick up the white ball becomes 1. In this case, which is completely classical, even the term ‘collapse’ itself seems unnecessary. Indeed, before the advent of quantum mechanics, no-one was talking about collapses in the theory of probability.
Such a viewpoint emerged when the options for quantum experimenting were limited by the level of technology. Since then, completely new kinds of experiments became a reality, and it is now possible to consider the collapse problem afresh.
Besides, the quantum collapse without observers has become a vital issue in cosmology where a physical mechanism is needed for turning quantum fluctuations into classical which would then evolve into galaxies.
As a consequence, the modern approach consists in translating philosophical questions into physical. In this vein, known as “experimental metaphysics”, we do not postulate ‘absence of collapse’ or otherwise, but rather ask how we could find out if there is a collapse or not using ordinary physical methods and principles.
The problem of the collapse under measurement is intimately related (but not identical) to the problem of quantum-to-classical transition or the transition from possible to actual.
The theory of environmental decoherence has made impressive progress in this area over recent decades. Optimistically, a view was advocated by some authors that decoherence has solved the measurement problem and the concept of collapse has become superfluous.
However, this issue is far from being settled (see, e.g., adler and references therein).
The conventional quantum mechanics models the transition from quantum to classical as a black box.
The decoherence theory has shone some light on that box, but do we know what is inside? What happens at the final, most intriguing stage of the measurement process? Only 15 percent of experts believe that the decoherence program has solved the measurement problem SKZ .
For this reason, it seems warranted to consider first the case of ‘pure’ collapse, i.e., neglect the environmental decoherence altogether. This approach can be justified in the end by comparing the two characteristic time-scales: for collapse and for decoherence. If the former is much shorter than the latter, then decoherence indeed can be neglected. This case is also more interesting from a practical point of view as in quantum computing decoherence needs to be reduced as much as possible.
II Non-linear effective equations for collapse
The problem of collapse during the measurement is essentially a many body problem (or even an infinitely many degrees of freedom problem). For this reason, exact equations governing the collapse process are hard to obtain, and even if obtained, would be difficult to use.
In such cases, a common method to proceed is to eliminate the unwanted degrees of freedom by applying the phenomenological, effective theory approach. The essence of this approach is to ‘integrate out’ the unknown quantities. This means introducing new terms built out of the known quantities subject to certain symmetry requirements and multiplying them by unknown coefficients. This has become a standard practice of quantum field theory over the recent decades but is less appreciated in the context of quantum mechanics.
When introducing new terms, our strategy is coservative, and we will only rely on the most general physical criteria: instability, irreversibility, and symmetry breaking. The first two features have been advocated already by Niels Bohr, while the third one is among the most universal phenomena in contemporary physics.
We start with the familiar setting where collapse is supposed to occur: a quantum system with a wave function plus a measuring apparatus.
Following the spirit of the effective-theory, phenomenological type of approach,
we set ourselves the following task: Take the Schrödinger equation and add new terms which are made entirely out of the wave function and unknown coefficients. The new equations should be able to describe both ordinary evolution and the measurement process (i.e., the collapse of the wave function).
This may sound as an extraordinary proposal which is bound to fail. Indeed, it seems that nothing could be so different as the continuous, deterministic evolution and the sudden, unpredictable collapse. Further, it seems unavoidable that some random, stochastic sources, or an additional physical phenomena should necessarily be introduced to account for the collapse as was done, for instance, in Refs.grw ; p89 ; w11 ; pen ; dio .
gravity-induced collapse ideas of Penrose pen and Diósi dio .
Nevertheless, we require from the outset that no new stochastic sources or gravitation should be invoked in explaining the collapse 111To avoid confusion, we emphasize that the Continuous Spontaneous Localization (CSL) theory of Ghirardi-
Rimini-Weber and Pearle grw ; p89 ; w11 and the gravity-induced collapse models, for instance, those of Penrose pen and Diósi dio are in principle different from our proposal because they address an essentially different problem: collapse during a free evolution and not during a measurement. So it would be meaningless to ask which theory is better. More details are in Sec. V..
The main difficulty in this search is the fact that quantum and classical systems essentially speak different languages and it is hard to make them talk to each other.
The quantum language is based on operators while the classical one—on ordinary functions.
To get around this obstacle, we can try to rewrite quantum mechanics in a more classical form, and thus to obtain a zeroth approximation suitable for our effective theory.
The most convenient for us is the approach known as the complex quantum Hamilton-Jacobi (CQHJ) formulation lp ; john ; yang ; gdt ; poi ; chou which has attracted growing attention recently.
It is natural to expect that the new formulation could give us new insights into fundamental long-standing issues such as the mechanism of quantum collapse.
For brevity and simplicity, we focus first
on the simplest yet typical case of the wave function collapse. This case is realised in a one-dimensional Ôsymmetric binary measurementÕ of the momentum where only two outcomes are possible and the
initial wave function is symmetric with respect to these
outcomes.
It can be shown (see Refs.lp ; john ; yang ; gdt ; poi ; chou ; ign and Appendix) that
the Schrödinger equation in this case becomes
$$p_{t}=-\nabla H,$$
(1)
where
$$p(x,t)=\frac{\hbar}{i}\frac{\psi^{\prime}}{\psi},\;\;\;H=V+\frac{p^{2}}{2m}+%
\frac{\hat{p}p}{2m}.$$
(2)
The parity-symmetric wave function before the collapse is the 1:1 superposition of two plane waves with opposite directions:
$$\psi=\frac{1}{\sqrt{2}}(e^{ikx}+e^{-ikx}).$$
(3)
We need to construct the phenomenological equation that would collapse this wave function into either $\psi_{+}=e^{ikx}\;\;\;{\rm or}\;\;\;\psi_{-}=e^{-ikx}$.
Following Bohr, we assume that the collapse is caused by the interaction of the quantum system with a semi-classical measuring device. Usually, the interaction is described by the Hamiltonian (or potential) that does not depend on the quantum state. However, such interactions cannot cause the collapse and should be extended to include state-dependent interactions. The phenomenological state-dependent potentials (or their analogues) are familiar in many areas of quantum theory (for instance, equations of Ginzburg-Landau, Gross-Pitaevsky, quantum chemistry etc.).
A state-dependent interaction is not a fundamental
interaction but an approximate one that is obtained, e.g., as a result
of mean-field approximation which assumes a definite classical value for a mean field. So the proposed approach does not attempt to derive the definite outcomes from a fundamental theory.
The state-dependent potential $V_{S}$ should be added to the ordinary potential $V$ in Eq. (22) to obtain:
$$p_{t}=-\nabla(H+V_{S}).$$
(4)
The gradient of the potential $V_{S}$ is a more convenient quantity than the potential itself, and we introduce a special term for it. It is natural to refer to this gradient as some kind of generalized force; in the context of our problem, we will call it “the collapsing force”:
$$F_{c}=-\nabla V_{S}.$$
(5)
The fact that the potential $V_{S}$ should be state-dependent can be equivalently stated as requirement that the collapsing force $F_{c}$ should be $p$-dependent.
Due to the shortness of collapse, the interaction with the apparatus is expected to be dominant during the measurement process, so our Eq. (4) can be written in a simpler form
$$p_{t}=F_{c}.$$
(6)
To find the form of $F_{c}$, we invoke the general principles as formulated in the beginning:
(1) Symmetry;
(2) Irreversibility;
(3) Instability.
It is natural to start with the simplest kind of functions, such as low-degree polynomials (which can also be viewed as the first terms of a Taylor expansion):
$$F_{c}=c_{0}+c_{1}p+c_{2}p^{2}+c_{3}p^{3}+\dots$$
(7)
According to Eq. (6), the force must be a vector under parity symmetry ($p\rightarrow-p$) as expected.
Therefore, the even-power terms in this expansion, $c_{0}$ and $c_{2}p^{2}$, should be dropped as violating this symmetry.
The requirement that the measurement gives two outcomes, $p=+q$ or $p=-q$, translates into the condition that $F_{c}$ must vanish at $p=\pm q$:
$$F_{c}(+q)=F_{c}(-q)=0.$$
(8)
Therefore, we must have $c_{1}=-c_{3}q^{2}$, and introducing $g=-c_{3}$ we obtain
$$F_{c}=gp(q^{2}-p^{2}).$$
(9)
Consequently, $g$ is interpreted as the effective coupling constant controlling the strength of interaction between the system and apparatus.
In general, $g$ could be a complex number, but in order to satisfy the condition of irreversibility we assume that it is real.
Altogether, our general equation (6) takes on a ‘collapsible’ form
$$p_{t}=gp(q^{2}-p^{2}).$$
(10)
In addition to Eq. (5), the collapsing force can be also written in the ‘potential’ form in ‘momentum space’ with the Higgs-type ‘collapsing potential’:
$F_{c}=-\partial V_{c}/\partial p,\;\;V_{c}=(g/2)(q^{2}-p^{2})^{2}$ (Fig.1).
This illustrates the remarkable link between our model and the phenomena of symmetry breaking, phase transitions, and synergetics. The collapsing potential can be compared to the ‘mean field’ and the measured quantity—to the ‘order parameter’.
Solving Eq. (10) with the initial condition (3),
yields
$$\displaystyle p(x,t)=\frac{qp_{0}(x)B(t)}{\sqrt{p_{0}^{2}(x)[B^{2}(t)-1]+q^{2}%
}},$$
(11)
$$\displaystyle\;\;B(t)=\exp{(gq^{2}t)},\;\;\;p_{0}(x)=iq\tan{kx}.$$
(12)
Finally, using the definition (2) we can determine the continuous behavior of the wave function all the way through the measuring process—in other words, the process of the wave function collapse:
$$\psi(x,t)=\frac{B(t)\cos{kx}+\sqrt{1-B^{2}(t)\sin^{2}{kx}}}{B(t)+1}N(t),$$
(13)
where $N(t)$ is the normalization factor.
In particular, we obtain the following formula for the characteristic timescale $\tau_{c}$ of the collapse:
$\tau_{c}\sim 1/(gq^{2})$.
Asymptotically, when $B(t)\gg 1$, the wave function (13) can take one of the two forms (in each domain of length $\pi/2k$):
$$\psi_{+}=e^{ikx}\;\;\;{\rm or}\;\;\;\psi_{-}=e^{-ikx}.$$
(14)
This two-valuedness arises due to the two-valuedness of the square roots in Eqs. (11,13).
An interesting question now is: How does the wave function ‘know’ which way to collapse: $\psi_{+}$ or $\psi_{-}$?
The answer is that if the initial wave function has the exact form (3), then it does not know, i.e., the two-valued ambiguity in (14) cannot be resolved.
This is because the initial $p_{0}$ is purely imaginary. If, on the other hand, $p_{0}$ has a non-zero real part $Re\;p_{0}$, then the definite outcome arises: In case of positive real part, $Re\;p_{0}>0$, the asymptotic wave function is $\psi_{+}=e^{ikx}$; if the real part is negative, $Re\;p_{0}<0$, the asymptotic wave function is $\psi_{-}=e^{-ikx}$.
Thus the measurement outcome is completely determined by the sign of the real part of $p_{0}$.
It is natural to presume that $Re\;p_{0}$ can take up positive and negative values with equal probability as required by the Born rule.
We emphasize that only infinitesimal non-zero values of $Re\;p_{0}$ are required for the emergence of a definite outcome in the scheme based on the simple equation (10).
A question can be raised whether $Re\;p_{0}$ can be considered a hidden variable, the dependence on which of the outcome is a step function.
Using a step function to represent the outcome dependence on $Re\;p_{0}$ leads to the following problems:
If the dependence of the outcome on $Re\;p_{0}$ is represented by a step function, the domain of this function should be infinitesimally small. If finite values of $Re\;p_{0}$ are allowed to be included in that domain, then these values would be incompatible with the state described by the wave function (3). But if we consider other states compatible with finite values of $Re\;p_{0}$, then such a step function would not be a correct predictor of the outcome.
Therefore, the outcome dependence on $Re\;p_{0}$ has a more singular character than the step function. By the same argument, the outcome dependence on
$Re\;p_{0}$ cannot be represented by any function with finite domain. Also, it cannot be represented by a discrete binary variable $\lambda$ taking on values $\pm 1$ because in that case the dependence of this ‘stand alone’ variable on $Re\;p_{0}$ would be lost.
For these reasons we prefer not to call $Re\;p_{0}$ a hidden variable.
The real issue, though, is not so much in the name but in a possible contradiction with various no-go theorems which have been formulated in relation to theories with hidden variables of various sorts. These issues are discussed in Sec. V.F where it is shown that no contradictions arise.
III Interpretation
A few remarks are in order concerning the interpretation of our ‘collapsible equation’, Eq. (10).
(a) It is not supposed to be a fundamental equation describing exactly the collapse process in every detail. Rather, it is a phenomenological, effective approximation which is meant to capture the main features of the collapse: instability, irreversibility, and symmetry breaking, which could lead to observable consequences such as the finite time-scale of the collapse process.
In any case, the construction of more elaborate models of the wave function collapse seems to require more experimental data than is available at the moment.
(b) In contrast with the Bohm-Bub models of explicit collapse bb , Eq. (10) does not contain hidden variables.
There is a large number of works that replace the Schrödinger equation with a non-linear one (see, e.g., bbm ; wn ). In contrast, our collapsible equations complement the Schrödinger equation. They have a different mathematical structure and distinct observational consequences.
(c) The “superluminality” argument which is often invoked to limit the use of non-linear equations in quantum mechanics does not apply in our case for the reasons discussed in Sec. V.
IV Experimental consequences
As with any phenomenological equation, care is required in comparing its predictions with experiment. Although our derivation allows one to establish some necessary conditions of its validity, these conditions may not be sufficient, and only experiment can decide on this issue.
First of all, a certain balance between the characteristic times involved should be observed. Ideally, one would aim at the interaction-switching time much less than the collapse time, and the collapse time much less than the ‘energy’ time-scale $\hbar/\tau_{E}$ where $E$ is the characteristic energy scale involved in a particular experiment.
One scheme of an experiment could look as follows. The initial state is created, then the measurement interaction is switched on at the instant $t=0$, then the measurement result is read-off. In the conventional picture, the interval between the second and third events can be as small as possible (it is bounded from below by the resolution time of the detector). However, in our picture, the collapse time cannot be less than $\tau_{c}$. Therefore, if the time-scale $\tau_{c}$ exceeds the resolution time-scale, the theory can be tested.
To compare the sensitivities of different kinds of experiments, it is convenient to introduce the dimensionless collapse time $\kappa=\tau_{c}/\tau_{E}$.
Except for the pioneering work of Papaliolios pap back in 1967, no dedicated searches for the finite time of the collapse have been performed (to my knowledge). So we attempt to obtain rough estimates on the allowed values of $\tau_{c}$ and $\kappa$ from various other experimental results assuming that null effects were observed in them on a time scale of $\tau_{m}$ which is the shortest time scale probed in a specific experiment (Table I).
It is natural to start with the experiments gkz ; bhi that tested specific non-linear versions of the Schrödinger equation bbm ; wn ; mdo . However, they turn out not to be the most sensitive for our purposes.
The well-known observation of ‘quantum jumps’ nsd is another natural place to look for constraints, but again they are found to be rather weak.
Next, we turn to the ultrafast measurements using femtosecond pulses of laser light blr .
Here, the limits begin to improve.
Finally, an excellent source of constraints appears to be experiments with Bose-Einstein condensates.
The
observed dynamics
turned out surprising in many respects, and it would be interesting to see if the idea of a finite-time collapse could be helpful there.
This area is virtually unexplored, and there is not one, but many promising directions of attack. Consequently, without attempting to be exhaustive, we restrict ourselves to outlining just two general strategies.
One approach is to look for the longest characteristic times $\tau_{E}$, comparable or better than those achieved in the BEC experiments. For example, the distances between the energy levels of ultra-cold neutrons bouncing from a horizontal mirror in the earth gravity field are of the order of $10^{-12}$ eV (or peV) dur which corresponds to $\tau_{E}\sim 10^{-3}$ s. Because times, such as the time of flight, can be measured with much better accuracy, this type of experiment (suggested for other purposes) could be a useful starting point for a test proposal.
Alternative strategy would be to start from the experiments with the shortest possible times $\tau_{m}$, such as those involving ultrashort pulses. This is a rapidly evolving area progressing from the femtosecond scale to atto-, zepto-, and even yoctoseconds ($10^{-24}$ s).
We emphasize that if the theory is confirmed that would not mean at all that the conventional quantum mechanics is refuted. It would only mean that the approximation of the instantaneous collapse is too crude and should be replaced by a more realistic approximation of a finite time-collapse. As was mentioned in the introduction, the assumption of instantaneous collapse is just an approximation rather than a fundamental principle of quantum theory, and the validity of this approximation does not affect the validity of quantum theory as a whole.
Similarly, if our specific Eq. (10) is experimentally excluded, that would not mean that the whole approach is invalidated, because a variety of different equations could be built on the same conceptual basis which could lead to different experimental predictions.
V Discussion
In constructing the collapse-enabled Schrödinger equations, our method is based on three well-established principles:
1. Symmetry breaking. The binary symmetric measurements on one particle in the absence of decoherence are the fundamental prototype measurements that should be studied first, both theoretically and experimentally.
Furthermore, the collapsible equations for such measurements should be constructed phenomenologically using the minimal set of most general assumptions.
Specifically, we single out the following two conditions that are widely accepted as reliable, non-controversial characteristics of the collapse process:
2. Irreversibility. This is generally acknowledged as the most important, essential feature of the collapse.
3. Instability (or amplification). The key role of these principles in the theory of measurement was succinctly formulated by Bohr as follows:“…every atomic phenomenon is closed in the sense that its observation is based on registrations obtained by means of suitable amplification devices with irreversible functioning such as, for example, permanent marks on the photographic plate caused by the penetration of electrons into the emulsion.” (italics added—A.I.). In the earlier literature, the same principles were often described in the language of “uncontrollable perturbations” arising when the quantum system interacts with the measuring device.
We note that the requirement of finite time is not among our basic principles. Rather, it comes out as a natural consequence of them.
Our method reflects the fact that decades of concerted efforts towards clarifying the fundamental mechanisms of collapse and the measurement process have met with limited success. This means that the collapse problem is an extremely challenging as well as an highly interdisciplinary one. It involves several distinct areas of theoretical physics and modern mathematics. As a consequence, there is little hope of progress unless one uses the phenomenological approach first and ask questions about first principles later.
Wave function collapse can be regarded as a ‘macroscopic quantum phenomenon’. Therefore we can turn for inspiration to the successful theories of other macroscopic quantum phenomena such as superfluidity and superconductivity.
In particular, it would be interesting here to draw a parallel with the puzzle of superconductivity, which took almost 50 years to unravel and it is still at the forefront of physics research.
This ultimate success was based on the consistent use of the phenomenological, effective-theory type of approaches before embarking on the final quest for a microscopic theory. The same strategy could be worth trying in the collapse problem.
The analogy can be continued in another respect. The phenomenological theory of superconductivity of Ginzburg and Landau uses a non-linear equation for the wave function. This does not mean, however, that the superposition principle is violated. Likewise, our use of a non-linear collapsing equation does not necessarily mean the violation of that principle because it is meant to be understood as a description at the phenomenological level which is not necessarily fundamental.
Many alternative models can be constructed by choosing different representations, forms of the collapsing force and/or adding hidden variables to the fundamental general equations belonging to the family exemplified by Eqs. (4) and (10).
However, given the probable complexity of the collapse phenomenon, it would be desirable to focus first (both theoretically and experimentally) on the most fundamental, the “purest” form of collapse—the binary symmetric measurement.
With suitable modifications, our methods could be applied in other subfields of quantum theory where a non-potential generalization of the Schrödinger equation is required by the conditions of the problem. Examples include the theory of dissipative systems, motion under the forces of radiative friction, etc. More generally, one can expect that interesting connections with nanoscience, quantum gravity and quantum cosmology could also be possible.
V.1 Schematic of main finding
Quantum state collapse is represented as a kinetic process of irreversible symmetry breaking driven by the instability of the initial state (Fig. 1-3).
V.2 Physical ideas underlying our framework
Our approach and results may look deceptively simple and questions could
be raised as to what makes this possible.
The explanation is that all the hard work is done by combining six powerful insights
from several fronts
of theoretical physics, which have not been put together before:
-complex quantum Hamilton-Jacobi formulation of quantum mechanics
-principle of symmetry breaking
-the effective theory approach
-the idea of irreversibility
-the measured quantity as the order parameter
-Bohr’s insight that quantum phenomena occur due to irreversible amplifying interaction of a classical and quantum systems.
V.3 Where do we stand in terms of various interpretations?
Although our formulation is designed to be interpretation-neutral, it can also be seen as the logical development of the conventional Copenhagen interpretation.
As discussed in the main text, those who favour the statistical ensemble interpretation of quantum mechanics would probably be skeptical about our approach. In the spirit of “experimental metaphysics”, we believe that experimental tests may help to clarify or update one’s philosophical position.
The same can be said about the Many-Worlds interpretation. In fact, experimental searches for collapse dynamics are needed from the standpoint of any interpretation because the results of such experiments will help to decide which one is closer to Nature.
If these searches fail to produce evidence for collapse, this would be a serious argument in favour of the collapse-denying interpretations, including Many-Worlds and statistical. Such arguments are hard to achieve from just theoretical reasoning.
Also, it is well-known that in some situations the wave function is ‘more real’ than in others (superconductivity, Bose-Einstein condensation etc.). Consequently, collapse may be a more real phenomenon in those situations.
In any case, the wave function satisfying our collapsible equation should be interpreted as an ‘effective’ or ‘interpolating’ wave function rather than a ‘genuine’ wave function which ceases to exist once the system starts to interact with the measuring device, according to the standard (von Neumann’s) model of the measurement process.
V.4 Comparison with the approaches of Bohr and von Neumann
In any treatment of the measurement problem
the question arises: is the measuring apparatus to be treated as a quantum system or not? On this crucial point, the opinions of quantum theorists appear to be split.
One school of thought (Bohr, Heisenberg, Dirac, and Born) accepts Bohr’s idea that the measuring device should be described approximately as classical and not quantum-mechanical. Consequently, the wave function of the apparatus is not introduced and the Schrödinger equation for the combined set “system plus apparatus” is not discussed or solved.
The alternative school (von Neumann, Landau and Peierls, Pauli, and most modern textbooks) assumes that the measuring device should be treated on the same footing as any other quantum system. As a consequence, most modern treatments introduce the wave function of the apparatus as a matter of fact that does not require further justification.
At the moment, one of the fruitful if challenging issues in quantum mechanics is the question whether the superposition principle is universally valid or not. Obviously, this is just another way to ask the same question.
Again, this problem is attacked not by purely philosophical discussion, but in the spirit of ‘experimental metaphysics’. However, despite impressive progress in the area, it would be fair to say that the final solution is not yet on the horizon.
We do not introduce explicitly the wave function of the measuring device. This is in accord with Bohr’s ideas that the the measuring device is best
described approximately and classically rather than quantum-mechanically. In fact, Bohr appears to
have reservations regarding the crucial step of “quantizing the apparatus” (first made by von Neumann) as he did not discussed it in his published works.
From the modern perspective, this may be viewed as a flaw or even a serious defect of such an interpretation.
On the other hand, Bohr’s reservation
might probably be justified because to this day, the efforts spent in quantizing the apparatus have not been completely successful. Moreover, this problem is considered fundamental but intractable in the forseeable future.
For example,
it was pointed out bh that no observer can write down his own wave function just because the human brain capacity is not large enough for that. A similar type of argument goes for an inanimate device as well.
We, of course, do not pretend to solve this part of the problem. Instead, we go around it and concentrate on the behaviour of the quantum system rather than the measuring device. In this way, we act in the spirit of ‘effective’ theory, which is well-known and extremely fruitful in the area of quantum fields (see, e.g., wei ), but is less appreciated in the context of quantum mechanics.
The essence of the effective theory approach is to integrate out all unknown degrees of freedom and focus on what is essential and observable.
V.5 Relation to decoherence
Although the theory of decoherence made an important step towards solving the collapse problem, it has not solved it.
Our paper does not include decoherence explicitly, but this is not because it is irrelevant. The real reason is that the collapse puzzle is a difficult and unsolved task. The common approach of a theoretical physicist in such cases is to tackle first an idealized setting where decoherence is ignored.
Moreover, this simple setting could be much more realistic than it seems at first sight. Indeed, we could imagine physical conditions under which the effects of decoherence are greatly reduced, such as very low noise, low temperature and similar types of environment. In this ‘very clean’ environment we expect that decoherence will be greatly affected, but the collapse will not. Consequently, analysis of collapse in the absence of decoherence is not merely an academic exercise, but a viable physical model for specific environmental conditions.
V.6 The Einstein-Podolsky-Rosen set-up and no-go theorems
In this paper, we do not treat any EPR-related issues leaving them for further work. We feel that the phenomenon of collapse, if real, could be complicated and not necessarily universal for all types of systems and experiments. One may even wonder if the quantum collapse could be as complex as its gravitational cousin. In any case, it seems prudent to approach this difficult problem in a step-by-step manner starting with the simplest possible set-up: a measurement made on one non-relativistic particle.
As Braginsky and Khalili note in Ref. bh (p. 28): “The presence of two or more degrees of freedom can change the character of the measurement substantially”.
Perhaps, to emphasise the differences, it could be helpful to introduce special terms like, for example, “contact” and “non-contact” collapse, the latter describing the EPR-type situations.
We also note that under a different set of assumptions, an EPR-type experimental search for a finite-time collapse was recently proposed in Ref. MP .
Because of the fact that our simple model (10) treats only the case of one particle and two measurement outcomes, it avoids tensions with Bell’s inequalities bell as well as various other no-go theorems due to Gleason glea ; Kochen and Specker ks ; Conway and Kochen (or “the Free Will theorem”) ck ; Colbeck and Renner cr ; Pusey, Barrett and Rudolph pbr .
More specifically, both Gleason’s and Kochen-Specker’s theorems require that the dimensionality of Hilbert space be greater than two glea ; ks —which does not apply in our case (10).
The Free Will theorem, in both original and strengthened formulations, requires a two-particle state due to its TWIN axiom ck . Therefore, it does not clash with our model (10) either.
Colbeck and Renner ck extended Bell’s prohibition of local hidden variables by considering hidden variables that have both local and global components and ruling them out as well. Again, our model (10) is immune from this exclusion theorem because it (the theorem) requires at least a two-particle system, among other things cr .
Pusey, Barrett and Rudolph pbr have recently proved another no-go theorem excluding a certain class of theories with hidden parameters $\lambda$.
The following quote from Ref. pbr summarizes their work nicely: “Our main result is that for distinct quantum states
$|\psi_{0}\rangle$ and $|\psi_{1}\rangle$, if the distributions $\mu_{0}(\lambda)$ and $\mu_{1}(\lambda)$ overlap (more precisely: if $\Delta$, the intersection of their supports, has non-zero measure) then there is a contradiction with the predictions of quantum theory.”
It follows immediately that this result does not apply to our model (10) because for distinct quantum states the distributions of our infinitesimal parameters do not overlap.
V.7 Non-linear quantum mechanics and superluminality
It has been proposed by Gisin gis among others that non-linear modifications of quantum mechanics could lead to faster than light communication. However, several points have to be taken into account here:
1) Our collapsible Schrödinger equations are not of the type considered in Ref. gis . Our non-linearity arises only at the measurement stage, while equations studied in Ref. gis are non-linear always. For this reason, superluminal communication is not made possible by our equations and the objection does not apply.
2) For completeness, we also mention that in subsequent papers fss ; kent ; bgh it was pointed out that the argument of Ref.gis is not free from possible loopholes. However, due to the above point, we do not need to appeal to these loopholes in order to demonstrate that our equations are free from the superluminality objection.
V.8 Is consciousness needed for collapse?
In our approach, an observer or the consciousness is not required for collapse to occur (cf. tha ). Moreover, our experimental upper limits on the collapse time (Table 1) show that this process is too fast for human consciousness which typically operates on much longer time-scales.
V.9 Collapse and relativity
Relativistic generalisations of the collapse concept are notoriously difficult and are not discussed here. We would like to stress, however, that at the moment the proposed collapse mechanism does not appear to be incompatible with relativity. On the contrary, a relativistic generalisation of our collapsible equations seems a worthwhile direction to pursue.
V.10 de Broglie-Bohm formulation
Some readers may get an impression that we are using or advocating some version of the de Broglie-Bohm (dBB) interpretation of quantum mechanics hol .
We emphasise that this is not the case 222This phrase should not be construed as an implicit criticism of such interpretations.. What we do is just borrowing a useful form of rewriting the Schrödinger equation. This can be considered as a purely mathematical transformation that is completely neutral with regard to interpretational issues. The dBB interpretation involves much more than such a rewriting, e.g., particle trajectories. These extra features are not necessary in our approach.
V.11 Other types of measurements
Today we know many more types of measurements than in 1930s.
Examples include weak, indirect, non-demolition, continuous, interrupted and other varieties of measurements. We emphasise that all these are beyond the scope of the present paper.
Our focus here is only on the simplest, “canonical” kind of measurement, which was also historically one of the first examples where the measurement problem was born. Without a thorough understanding of this example, there is little chance of progress in the more intricate cases.
Moreover, it would be perhaps over-optimistic to expect that one and the same theory can describe all kinds of measurements. (It would be similar to expecting, for instance, that one and the same theory can describe both strong and gravitational interactions.)
V.12 Theories of dynamical collapse
On a first view, our framework can be seen as a competitor to the well-known theories that represent collapse as a continuous dynamical process bas . In fact, it is not. The main reason is that the two approaches set different goals from the start even if the term “collapse” happens to figure prominently in both of them.
Briefly, this term is used in two rather different senses so one wonders if two different terms would be more appropriate here.
Our collapse has been induced by the measurement (“M-collapse”).
In the theories of dynamical collapse, it is occurring spontaneously, during the free evolution (“S-collapse”). This is a considerable difference with far-reaching consequences.
The theories of dynamical collapse do not address the mechanism of M-collapse. Likewise, we do not discuss the S-collapse. We ask the question: what happens to the wave function during the measurement? The theories of dynamical collapse ask: Why do the wave functions not spread out so much that the classical picture becomes blurred (problem of macro-objectification)?
It might be possible that in the end, the answers to both questions could come from one and the same ultimate source. However, for all similarities between the two problems there seem to be as many differences, so it would be reasonable to treat them as separate issues for the time being.
Our model and the dynamical collapse theories are made for two different purposes; they have different mathematical structure and different observational consequences.
V.13 Meaning of collapsing force
Introduction of the collapsing force (Fig. 4) is the main phenomenological element of our program.
As usual in such cases, we do not attempt to derive it from first principles, but rather postulate it on the combined strength of a series of independent arguments:
—The irreversible collapse during a measurement can be viewed as sort of relaxation process. Then the appearance of a collapsing force is naturally understood as the emergence of a relaxation-driving dissipative force without which relaxation would be impossible.
—In the same vein, the kinetic mechanism of collapse invites a parallel with the classical kinetic theory. The collapsing force, then, plays a similar role to the collision term in the Boltzmann equation.
—The task of constructing the collapsible Schrödinger equation can be formulated as the problem of finding a smooth interpolation between the uncollapsed and collapsed wave functions. A new term that is needed for such an interpolation would be equivalent to a collapsing force.
—What arguments force us to assume that the collapse occurs instantaneously, given that we are not aware of any other instantaneous processes in Nature?
—If, on the other hand, the collapse is not instantaneous but takes a finite time, then it seems plausible for purely mathematical reasons that one or another kind of extra terms (‘generalised collapsing forces’) should be introduced into the CQHJ-Schrödinger equation to produce this finite time.
—Within the CQHJ approach, the Schrödinger equation acquires a form so similar to Newton’s second law that it almost ‘begs for’ introducing
such a new force and exploring its consequences.
V.14 Is quantum mechanics to be modified?
Do our equations require a modification of standard quantum mechanics? At first sight, the answer to this would be “Yes”, but we believe that such a conclusion would be premature and unwarranted.
As was mentioned in the main text, our goal was not to modify quantum mechanics, but to obtain effective equations for the collapse process.
The fact that our equations are non-linear, and the Schrödinger equation is linear is not enough, in our view, to claim that these equations violate quantum mechanics.
The reason is that the Schrödinger equation is exact, and our equation is approximate. But it is well-known that the exact and linear equation (Schrödinger) can lead to approximate non-linear ones when suitable approximations are made (equations of Hartree-Fock, Gross-Pitaevsky etc.).
Indeed, as emphasized by Fock foc , in quantum theory the distinction between “fundamental” and “effective” concepts becomes a matter of convention because approximations play the principal, not a secondary role. For example, from the perspective of quantum field theory, the Schrödinger equation itself becomes an approximate “effective” equation, rather than a “fundamental” law.
Finally, from purely experimental perspective, the ‘fundamental or effective’ question is hardly the most
pressing one.
VI Conclusions and outlook
To sum up, a new class of quantum-mechanical non-linear equations for the wave function collapse has been proposed
that combines conceptual simplicity and predictive power.
Also, we have reported a set of constraints and outlined future opportunities for the sensitive experimental tests.
It is exciting to realize that only two or three orders of magnitude separate us from deeper, dedicated probes of the wave function collapse. The results will help to clarify both foundational and practical issues in quantum mechanics such as the status of the Many-Worlds interpretation and the ultimate limits on the speed of quantum computation and information processing.
Acknowledgments. The author thanks M. Albrow, R. Dörner, L. Ignatieva, V. A. Kuzmin, J.-P. Magnot, M. E. Shaposhnikov, and F. Thaheld for taking time for reading, valuable comments and discussions.
*
Appendix A Complex quantum Hamilton-Jacobi equation
To make the article self-contained, we provide below a
derivation of the complex quantum Hamilton-Jacobi equation.
Let us start with the ordinary Schrödinger equation for a particle of mass $m$ in the potential $V({\bf r})$:
$$i\hbar\psi_{t}=-\frac{\hbar^{2}}{2m}\Delta\psi+V\psi,$$
(15)
and
introduce a new variable ${\bf p}({\bf r},t)$:
$${\bf p}=\frac{\hbar}{i}\frac{\nabla\psi}{\psi}.$$
(16)
Our goal is to obtain a closed dynamical equation for ${\bf p}$.
Differentiating Eq. (16) with respect to time, we find:
$${\bf p}_{t}=\frac{\hbar}{i}\frac{\nabla\psi_{t}}{\psi}-{\bf p}\frac{\psi_{t}}{%
\psi}.$$
(17)
It is convenient to start $\psi$-elimination from the second term of Eq. (17). For this purpose, we first rewrite $\Delta\psi$ as
$$\Delta\psi=\nabla\left(\frac{i}{\hbar}{\bf p}\psi\right)=\frac{i}{\hbar}\left(%
\psi\nabla\cdot{\bf p}+{\bf p}\nabla\psi\right).$$
(18)
Inserting this into the Schrödinger equation (15) and multiplying both sides by $i{\bf p}/(\hbar\psi)$, we obtain:
$$-{\bf p}\frac{\psi_{t}}{\psi}=\frac{1}{2m}{\bf p}\nabla\cdot{\bf p}+\frac{i}{2%
m\hbar}{\bf p}p^{2}+\frac{i}{\hbar}{\bf p}V.$$
(19)
To eliminate $\psi$ from the first term of Eq. (17), we substitute the right-hand side of Eq. (18) into the Schrödinger equation (15) instead of $\Delta\psi$, then take the gradient of both parts and divide both of them by $\psi$, which yields:
$$\displaystyle-i\hbar\frac{\nabla\psi_{t}}{\psi}=\frac{i\hbar}{2m}\nabla(\nabla%
\cdot{\bf p})-\frac{1}{2m}{\bf p}\nabla\cdot{\bf p}-$$
$$\displaystyle\frac{1}{2m}\nabla p^{2}-\frac{i}{2\hbar m}p^{2}{\bf p}-\nabla V-%
\frac{i}{\hbar}{\bf p}V.$$
(20)
We now add this formula, part-by-part, to Eq. (19) and then use Eq. (17) to finally obtain:
$${\bf p}_{t}=-\nabla V-\frac{1}{2m}\nabla p^{2}+\frac{i\hbar}{2m}\nabla(\nabla%
\cdot{\bf p}).$$
(21)
So the complete elimination of the wave function is indeed possible, and Eq. (21) reproduces the dynamical equation that
plays the role of the Schrödinger equation in the CQHJ approach.
A slight rearrangement gives it an elegant shape:
$$\displaystyle{\bf p}_{t}=-\nabla H,$$
(22)
$$\displaystyle{\hat{\bf p}}=-i\hbar\nabla,\;\;\;H=V+\frac{p^{2}}{2m}+\frac{{%
\hat{\bf p}}{\bf p}}{2m}.$$
(23)
References
(1)
Leggett, A. J. The quantum measurement problem. Science 307, 871-872 (2005)
(2)
von Neumann, J. Mathematische Grundlagen der Quantenmechanik. (Julius Springer-Verlag, Berlin, 1932) [English transl.: Princeton University Press, Princeton, N.J., 1955].
(3)
Pusey, M. F., Barrett, J. & Rudolph, T. On the reality of the quantum state. Nature Physics 8, 476-479 (2012).
(4)
Nielsen, M. A. & Chuang, I.L. Quantum computation and quantum information. (Cambridge University Press, Cambridge, 2010).
(5)
Braginsky, V. B. & Khalili, F. Ya. Quantum Measurement. (Cambridge University Press, Cambridge, 1992).
(6)
Adler, S. Why decoherence has not solved the measurement problem. Stud. Hist. & Phil. Sci. 34, 135-142 (2003).
(7)
Schlosshauer, M., Kofler, J. and Zeilinger, A. A Snapshot of Foundational Attitudes Toward Quantum Mechanics. arXiv:1301.1069 [quant-ph].
(8)
Ghirardi, G.C., Rimini, A., & Weber, T. Unified dynamics for microscopic and macroscopic systems. Phys. Rev. D 34, 470-491 (1986).
(9)
Pearle, P. Combining stochastic dynamical state-vector reduction with spontaneous localization. Phys. Rev. A 39, 2277-2289 (1989).
(10)
Weinberg, S. Collapse of the state vector. Phys. Rev. A 85, 062116 (2012).
(11)
Penrose, R. On gravity’s role in
quantum state reduction. Gen. Rel.
Grav. 28, 581-609 (1996).
(12)
Diósi, L. The frictional Schrödinger-Newton equation in models of wave function collapse. J. Phys.: Conf. Ser. 67, 012024 (2007).
(13)
Leacock, R. A. & Padgett, M. J. Hamilton-Jacobi theory and the quantum action variable. Phys. Rev. Lett. 50, 3-6 (1983).
(14)
John, M. V. Modified de Broglie-Bohm approach to quantum mechanics. Found. Phys. Lett.
15, 329-343 (2002).
(15)
Yang, C.-D. Quantum dynamics of hydrogen atom in complex space. Ann. Phys. 319, 399-443 (2005).
(16)
Goldfarb, Y., Degani, I. & Tannor, D. J. Bohmian mechanics with complex action: A new trajectory-based formulation of quantum mechanics. J. Chem. Phys. 125, 231103
(2006).
(17)
Poirier, B. Flux continuity and probability conservation in complexified Bohmian mechanics. Phys. Rev. A 77, 022114 (2008).
(18)
Chou, C. C., Sanz, Á. S., Miret-Artés, S. & Wyatt, R. E. Hydrodynamic view of wave-packet interference: quantum caves. Phys. Rev. Lett. 102, 250401
(2009).
(19)
Ignatiev, A. Yu. How fast is the wave function collapse? arXiv:1204.3373 [quant-ph]; J. Phys.: Conf. Ser. 410, 012153 (2013).
(20)
Bohm, D. & Bub, J. A
proposed solution of the measurement problem in
quantum mechanics by a hidden variable theory.
Rev. Mod. Phys. 38, 453-459 (1966).
(21)
Pearle, P. M. Reduction of the state
vector by a nonlinear Schrödinger equation.
Phys.
Rev. D 13, 857-868 (1976).
(22)
Bialynicki-Birula, I. & J. Mycielski,
Nonlinear wave mechanics.
Ann. Phys. 100, 62-93 (1976).
(23)
Weinberg, S.
Precision tests of quantum mechanics.
Phys. Rev. Lett. 62,
485-488 (1989).
(24)
Mizrahi, S., Dodonov, V.V. and Otero, D.
Nonlinear Schrdinger-Liouville equation with antihermitian terms. Phys. Scr. 57, 24 (1998).
(25)
Papaliolios, C. Experimental
test of a hidden-variable quantum theory.
Phys. Rev. Lett. 18, 622-625 (1967).
(26)
Gähler, R., Klein, A. G. & Zeilinger, A. Neutron optical tests of nonlinear wave mechanics. Phys. Rev. A 23, 1611-1617
(1981).
(27)
Nagourney, W., Sandberg, J. & Dehmelt, H. Shelved optical electron amplifier: observation of quantum jumps. Phys. Rev. Lett. 56, 2797-2788 (1986).
(28)
Bollinger, J. J. et al. Test of the linearity of quantum mechanics by rf spectroscopy of the ${}^{9}$Be${}^{+}$ ground state. Phys. Rev. Lett. 63, 1031-1034
(1989).
(29)
Bauer, M. et al. Direct observation of surface chemistry using ultrafast soft-X-ray pulses. Phys. Rev. Lett. 87, 025501
(2001).
(30)
Donley, E. A. et al. Dynamics of collapsing and exploding Bose-Einstein condensates. Nature 412, 295-299 (2001).
(31)
Zbinden, H.
Brendel, J., Gisin, N. & Tittel, W. Experimental test of non-local quantum correlation in relativistic conÞgurations. Phys. Rev. A 63, 022111 (2001).
(32)
Jenke, T. et al. Realization of a gravity-resonance-spectroscopy technique. Nature Physics 7, 468-472 (2011).
(33)
Durstberger-Rennhofer, K., Jenke, T. & Abele, H. Probing the neutronÕs electric neutrality with Ramsey spectroscopy of gravitational quantum states of ultracold neutrons. Phys. Rev. D 84, 036004
(2011).
(34)
Plonka-Spehr, C. et al. An optical device for ultra-cold neutrons—Investigation of systematic effects and applications. Nucl. Instrum. & Methods A 618, 239-247 (2010).
(35)
Moreno, M. G. M. and Parisio, F. Investigation of the collapse of quantum states using entangled photons. Phys. Rev. A 88, 012118 (2013).
(36)
Bell, J.S.
On the Einstein-Podolsky-Rosen paradox. Physics 1, 195 (1964);
Brunner, N., Cavalcanti, D., Pironio, S., Scarani, V. and Wehner, S. Bell nonlocality.
arXiv: 1303.2849 [quant-ph].
(37)
Gleason, A. M.
Measures on the
closed subspaces of a Hilbert space. J. Math.
Mech. 6, 885 (1957).
(38)
Kochen, S. and Specker, E. J. The Problem of Hidden Variables in Quantum Mechanics. J. Math.
Mech. 17, 59 (1967).
(39)
Conway J. and Kochen, S. The Free Will Theorem. Found. Phys. 36, 1441 (2006); arXiv: 0807.3286 [quant-ph].
(40)
Colbeck, R. and Renner, R. Hidden Variable Models for Quantum Theory Cannot Have Any Local Part.
Phys. Rev. Lett. 101, 050403 (2008).
(41)
Gisin, N. Stochastic quantum dynamics and relativity. Helv. Phys. Acta 62, 363-371
(1989).
(42)
Ferrero, M., Salgado, D. & Sánchez-Gómez, J.L. Nonlinear quantum evolution does not imply supraluminal communication. Phys. Rev. A 70, 014101 (2004).
(43)
Kent, A. Nonlinearity without superluminality. Phys. Rev. A 72, 012108 (2005).
(44)
Brody, D.C., Gustavsson, A.C.T. & Hughston, L.P. Nonlinearity and constrained quantum motion. J. Phys. A: Math. Theor. 43, 082003 (2010).
(45)
Weinberg, S. The Quantum Theory of Fields, v. 1. (Cambridge University Press, Cambridge, 1995).
(46)
Thaheld, F. H. A modified approach to the measurement problem: Objective reduction in
the retinal molecule prior to conformational change. BioSystems 92, 114 (2008).
(47)
Holland, P.R. The Quantum Theory of Motion. (Cambridge University Press, Cambridge, 1993).
(48)
Bassi, A., Lochan, K., Satin, S., Singh, T.P., and Ulbricht, H. Models of wave-function collapse, underlying theories, and experimental tests. Rev. Mod. Phys. 85, 471 (2013).
(49)
Fock, V. A. The principal role of approximate methods in theoretical physics. Usp. Fiz. Nauk 16, 1070 (1936). |
Pose for Action – Action for Pose
Umar Iqbal
Computer Vision Group, University of Bonn, Germany
{uiqbal, garbade, gall}@iai.uni-bonn.de
Martin Garbade
Computer Vision Group, University of Bonn, Germany
{uiqbal, garbade, gall}@iai.uni-bonn.de
Juergen Gall
Computer Vision Group, University of Bonn, Germany
{uiqbal, garbade, gall}@iai.uni-bonn.de
Abstract
In this work we propose to utilize information about human actions to improve pose estimation in monocular videos. To this end, we present a pictorial structure model that exploits high-level information about activities to incorporate higher-order part dependencies by modeling action specific appearance models and pose priors. However, instead of using an additional expensive action recognition framework, the action priors are efficiently estimated by our pose estimation framework. This is achieved by starting with a uniform action prior and updating the action prior during pose estimation. We also show that learning the right amount of appearance sharing among action classes improves the pose estimation. We demonstrate the effectiveness of the proposed method on two challenging datasets for pose estimation and action recognition with over 80,000 test images.111The models and source code are available at http://pages.iai.uni-bonn.de/iqbal_umar/action4pose/.
I Introduction
Human pose estimation from RGB images or videos is a challenging problem in computer vision, especially for realistic and unconstrained data taken from the Internet. Popular approaches for pose estimation [6, 18, 38, 5, 2, 29] adopt the pictorial structure (PS) model, which resembles the human skeleton and allows for efficient inference in case of tree structures [8, 7]. Even if they are trained discriminatively, PS models struggle to cope with the large variation of human pose and appearance.
This problem can be addressed by conditioning the PS model on additional observations from the image.
For instance, [18] detects poselets, which are examples of consistent appearance and body part configurations, and condition the PS model on these.
Instead of conditioning the PS model on predicted configurations of body parts from an image, we propose to condition the PS model on high-level information like activity.
Intuitively, the information about the activity of a person can provide a strong cue about the pose and vice versa the activity can be estimated from pose. There have been only few works [39, 42, 15] that couple action recognition and pose estimation to improve pose estimation. In [39], action class confidences are used to initialize an optimization scheme for estimating the parameters of a subject-specific 3D human model in indoor multi-view scenarios. In [42], a database of 3D poses is used to learn a cross-modality regression forest that predicts the 3D poses from a sequence of 2D poses, which are estimated by [38]. In addition, the action is detected and the 3D poses corresponding to the predicted action are used to refine the pose. However, both approaches cannot be applied to unconstrained monocular videos. While [39] requires a subject-specific model and several views, [42] requires 3D pose data for training. More recently, [15] proposed an approach to jointly estimate action classes and refine human poses. The approach decomposes the human poses estimated at each video frame into sub-parts and tracks these sub-parts across time according to the parameters learned for each action. The action class and joint locations corresponding to the best part-tracks are selected as estimates for the action class and poses.
The estimation of activities, however, comes at high computational cost since the videos are pre-processed by several approaches, one for pose estimation [16] and two for extracting action related features [32, 34].
In this work, we present a framework for pose estimation that infers and integrates activities with a very small computational overhead compared to an approach that estimates the pose only. This is achieved by an action conditioned pictorial structure (ACPS) model for 2D human pose estimation that incorporates priors over activities. The framework of the approach is illustrated in Figure 1. We first infer the poses for each frame with a uniform distribution over actions.
While the binaries of the ACPS are modeled by Gaussian mixture models, which depend on the prior distribution over the action classes, the unaries of the ACPS model are estimated by action conditioned regression forests. To this end, we modify the approach [5], which consists of two layers of random forests, on two counts. Firstly, we replace the first layer by a convolutional network and use convolutional channel features to train the second layer, which consists of regression forests. Secondly, we condition the regression forests on a distribution over actions and learn the sharing among action classes. In our experiments, we show that these modifications increase the pose estimation accuracy by more than 40% compared to [5]. After the poses are inferred with a uniform distribution over actions, we update the action prior and the ACPS model based on the inferred poses to obtain the final pose estimates. Since the update procedure is very efficient, we avoid the computational expensive overhead of [15].
We evaluate our approach on the challenging J-HMDB [13] and Penn-Action [44] datasets, which consist of videos collected from the Internet and contain large amount of scale and appearance variations. In our experiments, we provide a detailed analysis of the impact of conditioning unaries and binaries on a distribution over actions and the benefit of appearance sharing among action classes. We demonstrate the effectiveness of the proposed approach for pose estimation and action recognition on both datasets. Compared to [15], the pose estimation accuracy is improved by over 30%.
II Related Work
State-of-the-art approaches for pose estimation are mostly based on neural networks [30, 28, 17, 35, 20, 21, 10] or on the pictorial structure framework [38, 18, 5, 43].
Several approaches have been proposed to improve the accuracy of PS models for human pose estimation. For instance, joint dependencies can be modeled not only by the PS model, but also by a mid-level image representation such as poselets [18], exemplars [23] or data dependent probabilities learned by a neural network [1].
Pose estimation in videos can be improved by taking temporal information or motion cues into account [16, 2, 45, 10, 14, 12]. In [16] several pose hypotheses are generated for each video frame and a smooth configuration of poses over time is selected from all hypotheses. Instead of complete articulated pose, [22] and [2] track individual body parts and regularize the trajectories of the body parts through the location of neighboring parts. Similar in spirit, the approach in [43] jointly tracks symmetric body parts in order to better incorporate spatio-temporal constraints, and also to avoid double-counting. Optical flow information has also been used to enhance detected poses at each video frame by analyzing body motion in adjacent frames [9, 45].
Recent approaches for human pose estimation use different CNN architectures to directly obtain the heatmaps of body parts [28, 17, 35, 20, 21, 10]. [28, 35, 20] and [21] use fully convolutional neural network architectures, where [35] proposes a multi-staged architecture that sequentially refines the output in each stage. For pose estimation in videos, [17] combines the heatmaps of body parts from multiple frames with optical flow information to leverage the temporal information in videos. More recently, [10] proposes a convolutional recurrent neural network architecture that takes as input a set of video frames and sequentially estimates the body part locations in each frame, while also using the information of estimated body parts in previous frames.
As done in this work, a few works also combine both PS model and CNNs [1, 29]. In contrast to the approaches that use temporal information in videos for pose refinement, we utilize the detected poses in each video frame to extract high-level information about the activity and use it to refine the poses.
Action recognition based on 3D human poses has been investigated in many works [41]. With the progress in the area of 2D human pose estimation in recent years, 2D poses have also gained an increased attention for action recognition [37, 26, 6, 13, 19, 3]. However, utilizing action recognition to aid human pose estimation is not well studied, in particular not in the context of 2D human pose estimation. There are only a few works [39, 42, 31, 40, 15] that showed the benefit of it. The approaches in [39, 42] rely on strong assumption.
The approach [39] assumes that a person-specific 3D model is given and considers pose estimation in the context of multiple synchronized camera views. The approach [42] focuses on 3D pose estimation from monocular videos and assumes that 3D pose data is available for all actions.
The approach [31] adopts a mixture of PS models, and learns a model for each action class. For a given image, each model is weighted by the confidence scores of an additional action recognition system and the pose with the maximum weight is taken. A similar approach is adopted in [40] to model object-pose relations. These approaches, however, do not scale with the number of action classes since each model needs to be evaluated.
The closest to our work is the recent approach of [15] that jointly estimates the action classes and refines human poses. The approach first estimates human poses at each video frame and decomposes them into sub-parts. These sub-parts are then tracked across video frames based on action specific spatio-temporal constraints. Finally, the action labels and joint locations are inferred from the part tracks that maximize a defined objective function. While the approach shows promising results, it does not re-estimate the parts but only re-combines them over frames i. e., only the temporal constraints are influenced by an activity. Moreover, it relies on two additional activity recognition approaches based on optical flow and appearance features to obtain good action recognition accuracy that results in a very large computational overhead as compared to an approach that estimates activities using only the pose information. In this work, we show that additional action recognition approaches are not required, but predict the activities directly from a sequence of poses. In contrast to [15], we condition the pose model itself on activities and re-estimate the entire pose per frame.
III Overview
Our method exploits the fact that the information about the activity of a subject provides a cue about pose and appearance of the subject, and vice versa.
In this work we utilize the high-level information about a person’s activity to leverage the performance of pose estimation, where the activity information is obtained from previously inferred poses. To this end, we propose an action conditioned pictorial structure (PS) that incorporates action specific appearance and kinematic models. If we have only a uniform prior over the action classes, the model is a standard PS model, which we will briefly discuss in Section IV. Figure 1 depicts an overview of the proposed framework.
IV Pictorial Structure
We adopt the joint representation [5] of the PS model [8], where the vector $\mathbb{x}_{j}\in\mathcal{X}$ represents the 2D location of the $j^{th}$ joint in image $\mathbb{I}$, and $\mathcal{X}$ = $\{\mathbb{x}_{j}\}_{j\in\mathcal{J}}$ is the set of all body joints. The structure of a human body is represented by a kinematic tree with nodes of the tree being the joints ${j}$ and edges $\mathcal{E}$ being the kinematic constraints between a joint $j$ and its unique parent joint $p$ as illustrated in Figure 1. The pose configuration in a single image is then inferred by maximizing the following posterior distribution:
$$\mathrm{p}(\mathcal{X}|\mathbb{I})\propto\prod_{j\in\mathcal{J}}\phi_{j}(\mathbb{x}_{j}|\mathbb{I})\prod_{j,p\in\mathcal{E}}\psi_{jp}(\mathbb{x}_{j},\mathbb{x}_{p})$$
(1)
where the unary potentials $\phi_{j}(\mathbb{x}_{j}|\mathbb{I})$ represent the likelihood of the $j^{th}$ joint at location $\mathbb{x}_{j}$. The binary potentials $\psi_{jp}(\mathbb{x}_{j},\mathbb{x}_{p})$ define the deformation cost for the joint-parent pair $(j,p)$, and are often modeled by Gaussian distributions for an exact and efficient solution using a distance transform [8].
We describe the unary and binary terms in Section IV-A and Section IV-B, respectively. In Section V-A, we then discuss how these can be adapted to build an action conditioned PS model.
IV-A Unary Potentials
Random regression forests have been proven to be robust for the task of human pose estimation [24, 27, 5]. A regression forest $\mathcal{F}$ consists of a set of randomized regression trees, where each tree $T$ is composed of split and leaf nodes. Each split node represents a weak classifier which passes an input image patch $P$ to a subsequent left or right node until a leaf node $L_{T}$ is reached. As in [5], we use a separate regression forest for each body joint. Each tree is trained with a set of randomly sampled images from the training data. The patches around the annotated joint locations are considered as foreground and all others as background. Each patch consists of a joint label $c\in\{0,j\}$, a set of image features $F_{P}$, and its 2D offset $\mathbb{d}_{P}$ from the joint center.
During training, a splitting function is learned for each split node by randomly selecting and maximizing a goodness measure for regression or classification.
At the leaves the class probabilities $\mathrm{p}(c|L_{T})$ and the distribution of offset vectors $\mathrm{p}(\mathbb{d}|L_{T})$ are stored.
During testing, patches are densely extracted from the input image $\mathbb{I}$ and are passed through the trained trees.
Each patch centered at location $\mathbb{y}$ ends in a leaf node $L_{T}(P(\mathbb{y}))$ for each tree $T\in\mathcal{F}$.
The unary potentials $\phi_{j}$ for the joint $j$ at location $\mathbb{x}_{j}$ are then given by
$$\phi_{j}(\mathbb{x}_{j}|\mathbb{I})=\sum_{\mathbb{y}\in\Omega}\dfrac{1}{|\mathcal{F}|}\sum_{T\in\mathcal{F}}\Big{\{}\mathrm{p}(c=j|L_{T}(P(\mathbb{y})))\\
\cdot\mathrm{p}(\mathbb{x}-\mathbb{y}|L_{T}(P(\mathbb{y}))\Big{\}}.$$
(2)
In [5] a two layer approach is proposed. The first layer consists of classification forests that classify image patches according to the body parts using a combination of color features, HOG features, and the output of a skin color detector. The second layer consists of regression forests that predict the joint locations using the features of the first layer and the output of the first layer as features. For both layers, the split nodes compare feature values at different pixel locations within a patch of size $24\times 24$ pixels.
We propose to replace the first layer by a convolutional network and extract convolutional channel features (CCF) [36] from the intermediate layers of the network to train the regression forests of the second layer. In [36] several pre-trained network architectures have been evaluated for pedestrian detection using boosting as classifier. The study shows that the “conv3-3” layer of the VGG-16 net [25] trained on the ImageNet (ILSVRC-2012) dataset performs very well even without fine tuning, but it is indicated that the optimal layer depends on the task. Instead of pre-selecting a layer, we use regression forests to select the features based on the layers “conv2-2”, “conv3-3”, “conv4-3”, and “conv5-3”. An example of the CCF extracted from an image is shown in Figure 2. Since these layers are of lower dimensions than the original image, we upsample them using linear interpolation to make their dimensions equivalent to the input image. This results in a 1408 (128+256+512+512) dimensional feature representation for each pixel. As split nodes in the regression forests, we use axis-aligned split functions. For an efficient feature extraction at multiple image scales, we use patchwork as proposed in [11] to perform the forward pass of the convolutional network only once.
IV-B Binary Potentials
Binary potentials $\psi_{jp}(\mathbb{x}_{j},\mathbb{x}_{p})$ are modeled as a Gaussian mixture model for each joint $j$ with respect to its parent joint $p$ in the kinematic tree. As in [5], we obtain the relative offsets between child and parent joints from the training data and cluster them into $k=1,\dots,K$ clusters using k-means clustering. Each cluster $k$ takes the form of a weighted Gaussian distribution as
$$\psi_{jp}(\mathbb{x}_{j},\mathbb{x}_{p})=w_{jp}^{k}\cdot\\
\exp\left(-\dfrac{1}{2}\left(\mathbb{d}_{jp}-\mu_{jp}^{k}\right)^{T}\left(\Sigma_{jp}^{k}\right)^{-1}\left(\mathbb{d}_{jp}-\mu_{jp}^{k})\right)\right)$$
(3)
with mean $\mu_{jp}^{k}$ and covariance $\Sigma_{jp}^{k}$, where $\mathbb{d}_{jp}=(\mathbb{x}_{j}-\mathbb{x}_{p})$. The weights $w_{jp}^{k}$ are set according to the cluster frequency $\mathrm{p}(k|j,p)^{\alpha}$ with a normalization constant $\alpha=0.1$ [5].
For inference, we select the best cluster $k$ for each joint
by computing the max-marginals for the root node and backtrack the best pose configuration from the maximum of the max-marginals.
V Action Conditioned Pose Estimation
As illustrated in Figure 1, our goal is to estimate the pose $\mathcal{X}$ conditioned by the distribution $\mathrm{p_{{}_{A}}}$ for a set of action classes $a\in\mathcal{A}$. To this end, we introduce in Section V-A a pictorial structure model that is conditioned on $\mathrm{p_{{}_{A}}}$. Since we do not assume any prior knowledge of the action, we estimate the pose first with the uniform distribution $\mathrm{p_{{}_{A}}}(a)=1/|\mathcal{A}|,\forall a\in\mathcal{A}$. The estimated poses for $N$ frames are then used to estimate the probabilities of the action classes $\mathrm{p_{{}_{A}}}(a|\mathcal{X}_{n=1\dots N}),\forall a\in\mathcal{A}$ as described in Section V-B. Finally, the poses $\mathcal{X}_{n}$ are updated based on the distribution $\mathrm{p_{{}_{A}}}$.
V-A Action Conditioned Pictorial Structure
In order to integrate the distribution $\mathrm{p_{{}_{A}}}$ of the action classes obtained from the action classifier into (1), we make the unaries and binaries dependent on $\mathrm{p_{{}_{A}}}$:
$$\mathrm{p}(\mathcal{X}|\mathrm{p_{{}_{A}}},\mathbb{I})\propto\prod_{j\in\mathcal{J}}\phi_{j}(\mathbb{x}_{j}|\mathrm{p_{{}_{A}}},\mathbb{I})\cdot\prod_{j,p\in\mathcal{E}}\psi_{jp}(\mathbb{x}_{j},\mathbb{x}_{p}|\mathrm{p_{{}_{A}}}).$$
(4)
While the unary terms are discussed in Section V-A1, the binaries $\psi_{jp}(\mathbb{x}_{j},\mathbb{x}_{p}|\mathrm{p_{{}_{A}}})$ are represented by Gaussians as in (3). However, instead of computing mean and covariance from all training poses with equal weights, we weight each training pose based on its action class label and $\mathrm{p_{{}_{A}}}(a)$. In our experiments, we will also investigate the case where $\mathrm{p_{{}_{A}}}(a)$ is simplified to
$$\mathrm{p_{{}_{A}}}(a)=\begin{cases}1&\text{if}\quad a=\operatorname*{argmax}_{a^{\prime}}\mathrm{p_{{}_{A}}}(a^{\prime}|\mathcal{X}_{n=1\dots N})\\
0&\text{otherwise.}\end{cases}$$
(5)
V-A1 Conditional Joint Regressors
Figure 3 shows examples of patches of the wrist extracted from images of different action classes. We can see a large amount of appearance variation across different classes regardless of the fact that all patches belong to the same body joint. However, it can also be seen that within individual activities this variation is relatively small. We exploit this observation and propose action specific unary potentials for each joint $j$. To this end we adopt conditional regression forests [4, 27] that have been proven to be robust for facial landmark detection in [4] and 3D human pose estimation in [27].
While [4] trains a separate regression forest for each head pose and selects a specific regression forest conditioned on the output of a face pose detector, [27] proposes partially conditioned regression forests, where a forest is jointly trained for a set of discrete states of a human attribute like human orientation or height and the conditioning only happens at the leaf nodes. Since the continuous attributes are discretized, interpolation between the discrete states is achieved by sharing the votes.
In this work we resort to partially conditional forests due to its significantly reduced training time and memory requirements.
During training we augment each patch $P$ with its action class label $a$. Instead of $\mathrm{p}(c|L_{T})$ and $\mathrm{p}(\mathbb{d}|L_{T})$, the leaf nodes model the conditional probabilities $\mathrm{p}(c|a,L_{T})$ and $\mathrm{p}(\mathbb{d}|a,L_{T})$. Given the distribution over action classes $\mathrm{p_{{}_{A}}}$, we obtain the conditional unary potentials:
$$\displaystyle\phi_{j}(\mathbb{x}_{j}|\mathrm{p_{{}_{A}}},\mathbb{I})$$
$$\displaystyle=\sum_{\mathbb{y}\in\Omega}\dfrac{1}{|\mathcal{F}|}\sum_{T\in\mathcal{F}}\sum_{a\in\mathcal{A}}\Big{\{}\mathrm{p_{{}_{A}}}(a)$$
$$\displaystyle\cdot\mathrm{p}(c=j|a,L_{T}(P(\mathbb{y}))).\mathrm{p}(\mathbb{x}-\mathbb{y}|a,L_{T}(P(\mathbb{y}))\Big{\}}$$
$$\displaystyle=\sum_{a\in\mathcal{A}}\phi_{j}(\mathbb{x}_{j}|a,\mathbb{I})\mathrm{p_{{}_{A}}}(a).$$
(6)
Since the terms $\phi_{j}(\mathbb{x}_{j}|a,\mathbb{I})$ need to be computed only once for an image $\mathbb{I}$, $\phi_{j}(\mathbb{x}_{j}|\mathrm{p_{{}_{A}}},\mathbb{I})$ can be efficiently computed after an update of $\mathrm{p_{{}_{A}}}$.
V-A2 Appearance Sharing Across Actions
Different actions sometimes share body pose configurations and appearance of parts as shown in Figure 3. We therefore propose to learn the sharing among action classes within a conditional regression forest. To this end, we replace the term $\phi_{j}(\mathbb{x}_{j}|a,\mathbb{I})$ in (V-A1) by a weighted combination of the action classes:
$$\phi^{\textit{sharing}}_{j}(\mathbb{x}_{j}|a,\mathbb{I})=\sum_{a^{\prime}\in\mathcal{A}}\gamma_{a}(a^{\prime})\phi_{j}(\mathbb{x}_{j}|a^{\prime},\mathbb{I})$$
(7)
where the weights $\gamma_{a}(a^{\prime})$ represent the amount of sharing between action class $a$ and $a^{\prime}$. To learn the weights $\gamma_{a}$ for each class $a\in\mathcal{A}$, we apply the trained conditional regression forests to a set of validation images scaled to a constant body size and maximize the response of (7) at the true joint locations and minimize it at non-joint locations. Formally, this can be stated as
$$\gamma_{a}=\operatorname*{argmax}_{\gamma}\sum_{n_{a}}\sum_{j}\Bigg{\{}\sum_{a^{\prime}\in\mathcal{A}}\gamma(a^{\prime})\phi^{*}_{j}\left(\mathbb{x}^{\textit{gt}}_{j,n_{a}}|{a^{\prime}},\mathbb{I}_{n_{a}}\right)\\
-\max_{\mathbb{x}\in\mathbb{X}_{j,n_{a}}^{\textit{neg}}}\left(\sum_{a^{\prime}\in\mathcal{A}}\gamma(a^{\prime})\phi^{*}_{j}\left(\mathbb{x}|{a^{\prime}},\mathbb{I}_{n_{a}}\right)\right)\Bigg{\}}-\lambda\|\gamma\|^{2}$$
(8)
subject to $\sum_{a^{\prime}\in\mathcal{A}}\gamma(a^{\prime})=1$ and $\gamma(a^{\prime})\geq 0$. $\mathbb{I}_{n_{a}}$ denotes the $n^{th}$ scaled validation image of action class $a$, $\mathbb{x}^{\textit{gt}}_{j,n_{a}}$ is the annotated joint position for joint $j$ in image $\mathbb{I}_{n_{a}}$, and $\mathbb{X}_{j,n_{a}}^{\textit{neg}}$ is a set of image locations which are more than $5$ pixels away from $\mathbb{x}^{\textit{gt}}_{j,n_{a}}$.
The set of negative samples is obtained by computing $\phi^{*}_{j}(\mathbb{x}|{a^{\prime}},\mathbb{I}_{n_{a}})$ and taking the $10$ strongest modes, which do not correspond to $\mathbb{x}^{\textit{gt}}_{j,n_{a}}$, for each image. For optimization, we use the smoothed unaries
$$\phi_{j}^{*}(\mathbb{x}|a,\mathbb{I})=\sum_{\mathbb{y}\in\Omega}\exp\left(-\dfrac{{\|\mathbb{x}-\mathbb{y}\|}^{2}}{\sigma^{2}}\right)\phi_{j}(\mathbb{y}|a,\mathbb{I})$$
(9)
with $\sigma=3$ and replace $\max$ by the softmax function to make (8) differentiable. The last term in (8) is a regularizer that prefers sharing, i. e., $\|\gamma\|^{2}$ attains its minimum value for uniform weights. In our experiments, we use $\lambda=0.4$ as weight for the regularizer.
We optimize the objective function by constrained local optimization using uniform weights for initialization $\gamma(a^{\prime})=1/|\mathcal{A}|$. In our experiments, we observed that similar weights are obtained when the optimization is initialized by $\gamma(a)=1$ and $\gamma(a^{\prime})=0$ for $a^{\prime}\neq a$, indicating that the results are not sensitive to the initialization.
In (8), we learn the weights $\gamma_{a}$ for each action class but we could also optimize for each joint independently. In our experiments, however, we observed that this resulted in over-fitting.
V-B Action Classification
For pose-based action recognition, we use the bag-of-word approach proposed in [13].
From the estimated joint positions $\mathcal{X}_{n=1\dots N}$, a set of features called NTraj+ is computed that encodes spatial and directional joint information. Additionally, differences between successive frames are computed to encode the dynamics of the joint movements. Since we use a body model with 13 joints, we compute the locations of missing joints (neck and belly) in order to obtain the same $15$ joints as in [13]. We approximate the neck position as the mean of the face and the center of shoulder joints. The belly position is approximated by the mean of the shoulder and hip joints.
For each of the $3,223$ descriptor types, a codebook is generated by running k-means $8$ times on all training samples and choosing the codebook with minimum compactness. These codebooks are used to extract a histogram for each descriptor type and video.
For classification, we use an SVM classifier in a multi-channel setup. To this end, for each descriptor type $t$, we compute a distance matrix $D_{t}$ that contains the $\chi^{2}$-distance between the histograms $(h_{i}^{t},h_{j}^{t})$ of all video pairs $(v_{i},v_{j})$. We then obtain the kernel matrix that we use for the multi-class classification as follows
$$K(v_{i},v_{j})=\exp\left(-\frac{1}{L}\sum^{L}_{t=1}\frac{D_{t}(h_{i}^{t},h_{j}^{t})}{\mu^{t}}\right)$$
(10)
where $L$ is the number of descriptor types and $\mu^{t}$ is the mean of the distance matrix $D_{t}$. For classification we use a one-vs-all approach with $C=100$ for the SVM.
VI Experiments
In order to evaluate the proposed method, we follow the same protocol as proposed in [15]. In particular, we evaluate the proposed method on two challenging datasets, namely sub-J-HMDB [13]
and the Penn-Action dataset [44]. Both datasets provide annotated 2D poses and activity labels for each video. They consist of videos collected from the Internet and contain large amount of scale and appearance variations, low resolution frames, occlusions and foreshortened body poses. This makes them very challenging for human pose estimation.
While sub-J-HMDB [13] comprises videos from 12 action categories with fully visible persons, the Penn-Action dataset comprises videos from 15 action categories with a large amount of body part truncations. As in [15], we discard the activity class “playing guitar” since it does not contain any fully visible person.
For testing on sub-J-HMDB, we follow the 3-fold cross validation protocol proposed by [13]. On average for each split, this includes $229$ videos for training and $87$ videos for testing with $8,124$ and $3,076$ frames, respectively.
The Penn-Action dataset consists of $1,212$ videos for training and $1,020$ for testing with $85,325$ and $74,308$ frames, respectively.
To evaluate the performance of pose estimation, we use the APK (Average Precision of Keypoints) metric [38, 15].
VI-A Implementation Details
For the Penn-Action dataset, we split the training images half and half into a training set and a validation set. Since the dataset sub-J-HMDB is smaller, we create a validation set by mirroring the training images. The training images are scaled such that the mean upper body size is $40$ pixels. Each forest consists of $20$ trees, where $10$ trees are trained on the training and $10$ on the validation set, with a maximum depth of $20$ and a minimum of $20$ patches per leaf. We train each tree with $50,000$ positive and $50,000$ negative patches extracted from $5,000$ randomly selected images and generate $40,000$ split functions at each node.
For the binary potentials, we use $k=24$ mixtures per part.
For learning the appearance sharing among action classes (Section V-A2) and training the action classifier (Section V-B), we use the $10$ trees trained on the training set and apply them to the validation set. The action classifier and the sharing are then learned on the validation set.
For pose estimation, we create an image pyramid and perform inference at each scale independently. We then select the final pose from the scale with the highest posterior (4). In our experiments, we use $4$ scales with scale factor $0.8$.
The evaluation of 260 trees (20 trees for each of the 13 joints) including feature extraction takes roughly 15 seconds on average.222Measured on a 3.4GHz Intel processor using only one core with NVidia GeForce GTX 780 GPU. The image size for all videos in sub-J-HMDB is $320\times 240$ pixels. Inference with the PS model for all 4 scales takes around 1 second. The action recognition with feature computation takes only 0.18 seconds per image and it does not increase the time for pose estimation substantially.
VI-B Pose Estimation
We first evaluate the impact of the convolutional channel features (CCF) for pose estimation on split-1 of sub-J-HMDB. The results in Table I show that the CCF outperform the combination of color features, HOG features, and the output of a skin color detector, which is used in [5].
In Table I(a) we evaluate the proposed ACPS model under different settings on split-1 of sub-J-HMDB when using CCF features for joint regressors. We start with the first step of our framework where neither the unaries nor the binaries depend on the action classes. This is equivalent to the standard PS model described in Section IV, which achieves an average joint estimation accuracy of $51.5\%$. Given the estimated poses, the pose-based action recognition approach described in Section V-B achieves an action recognition accuracy of $66.3\%$ for split-1.
Having estimated the action priors $\mathrm{p_{{}_{A}}}$, we first evaluate action conditioned binary potentials while keeping the unary potentials as in the standard PS model. As described in Section V-A, we can use in our model the probabilities $\mathrm{p_{{}_{A}}}$ or replace them by the distribution (5), which considers only the classified action class. The first setting is denoted by “Cond. ($\mathrm{p_{{}_{A}}}$)” and the second by “Cond. (5)”. It can be seen that the conditional binaries based on (5) already outperform the baseline by improving the accuracy from $51.5\%$ to $53.8\%$. However, taking the priors from all classes slightly decreases the performance. This shows that conditioning the binary potentials on the most probable class is a better choice than using priors from all classes.
Secondly, we analyze how action conditioned unary potentials affect pose estimation. For the unaries, we have the same options “Cond. ($\mathrm{p_{{}_{A}}}$)” and “Cond. (5)” as for the binaries. In addition, we can use appearance sharing as described in Section V-A2, which is denoted by “AS”. For all three binaries, the conditional unaries based on (5) decrease the performance.
Since the conditional unaries based on (5) are specifically designed for each action class, they do not generalize well in case of a misclassified action class..
However, adding appearance sharing to the conditional unaries boost the performance for both conditioned on (5) and $\mathrm{p_{{}_{A}}}$. Adding appearance sharing outperforms all other unaries without appearance sharing, i. e., conditional unaries, independent unaries and the unaries conditioned on $\mathrm{p_{{}_{A}}}$. For all unaries, the binaries conditioned on (5) outperform the other binaries.
This shows that appearance sharing and binaries conditioned on the most probable class performs best, which gives an improvement of the baseline from $51.5\%$ to $55.3\%$.
In Table I(b), we also evaluate the proposed ACPS when using the weaker features from [5]. Although the accuracies as compared to CCF features are lower, the benefit of the proposed method remains the same. For the rest of this paper, we will use CCF for all our experiments.
In Table III we compare the proposed action conditioned PS model with other state-of-the-art approaches on all three splits of sub-J-HMDB. In particular, we provide a comparison with [5, 38, 15, 16, 2, 1]. The accuracies for the approaches [38, 15, 16, 2] are taken from [15] where the APK threshold $0.2$ is used. We also evaluated the convolutional network based approach [1] using the publicly available source code trained on sub-J-HMDB. Our approach outperfroms the other methods by a margin, and notably improves wrist localization by more than 5% as compared to the baseline.
Table IV compares the proposed ACPS with the state-of-the-art on the Penn-Action dataset. The accuracies for the approaches [38, 15, 16] are taken from [15]. We can see that the proposed method improves the baseline from 75.5% to 81.1%, while improving the elbow and wrist localization accuracy by more than 7% and 10%, respectively. The proposed method also significantly outperforms other approaches Only the approach [10] achieves a higher accuracy than our method. [10], however, uses a better multi-staged CNN architecture as baseline compared to our network for computing CCF features. Since the gain of ACPS compared to our baseline even increases when better features are used as shown in Table I(a) & Table I(b), we expect at least a similar performance gain when we use the baseline architecture from [10] for ACPS.
VI-C Action Recognition
In Table VI, we compare the action recognition accuracy obtained by our approach with state-of-the-approaches for action recognition. On sub-J-HMDB, the obtained accuracy using only pose as feature is comparable to the other approaches. Only the recent work [3] which combines pose, CNN, and motion features achieves a better action recognition accuracy. However, if we combine our pose-based action recognition with Fisher vector encoding of improved dense trajectories [33] using late fusion, we outperform other methods that also combine pose and appearance. The results are similar on the Penn-Action dataset.
In Table V, we report the effect of different action recognition approaches on pose estimation. We report the pose estimation accuracy for split-1 of sub-J-HMDB when the action classes are not inferred by our framework, but estimated using improved dense trajectories with Fisher vector encoding (IDT-FV) [33] or the fusion of our pose-based method and IDT-FV. Although the action recognition rate is higher when pose and IDT-FV are combined, the pose estimation accuracy is not improved. If the action classes are not predicted but are provided (GT),
the accuracy improves slightly for sub-J-HMDB and
from 64.8% to 68.1% for the Penn-Action dataset. We also experimented with several iterations in our framework, but the improvements compared to the achieved accuracy of $51.6\%$ were not more than 0.1% on all three splits of sub-J-HMDB.
VII Conclusion
In this paper, we have demonstrated that action recognition can be efficiently utilized to improve human pose estimation on realistic data. To this end, we presented a pictorial structure model that incorporates high-level activity information
by conditioning the unaries and binaries on a prior distribution over action labels. Although the action priors can be estimated by an accurate, but expensive action recognition system, we have shown that the action priors can also be efficiently estimated during pose estimation without substantially increasing the computational time of the pose estimation. In our experiments, we thoroughly analyzed various combinations of unaries and binaries and showed that learning the right amount of appearance sharing among action classes improves the pose estimation accuracy. While we expect further improvements by using a more sophisticated CNN architecture as baseline and by including a temporal model, the proposed method has already shown its effectiveness on two challenging datasets for pose estimation and action recognition.
Acknowledgements: The work was partially supported by the ERC Starting Grant ARCA (677650).
References
[1]
X. Chen and A. L. Yuille.
Articulated pose estimation by a graphical model with image dependent
pairwise relations.
In NIPS, 2014.
[2]
A. Cherian, J. Mairal, K. Alahari, and C. Schmid.
Mixing Body-Part Sequences for Human Pose Estimation.
In CVPR, 2014.
[3]
G. Chéron, I. Laptev, and C. Schmid.
P-cnn: Pose-based cnn features for action recognition.
In CVPR, 2015.
[4]
M. Dantone, J. Gall, G. Fanelli, and L. Van Gool.
Real-time facial feature detection using conditional regression
forests.
In CVPR, 2012.
[5]
M. Dantone, C. Leistner, J. Gall, and L. Van Gool.
Body parts dependent joint regressors for human pose estimation in
still images.
TPAMI, 2014.
[6]
C. Desai and D. Ramanan.
Detecting actions, poses, and objects with relational phraselets.
In ECCV, 2012.
[7]
P. F. Felzenszwalb, R. B. Girshick, D. McAllester, and D. Ramanan.
Object detection with discriminatively trained part-based models.
TPAMI, 2010.
[8]
P. F. Felzenszwalb and D. P. Huttenlocher.
Pictorial structures for object recognition.
IJCV, 2005.
[9]
K. Fragkiadaki, H. Hu, and J. Shi.
Pose from flow and flow from pose.
In CVPR, 2013.
[10]
G. Gkioxari, A. Toshev, and N. Jaitly.
Chained predictions using convolutional neural networks.
In ECCV, 2016.
[11]
F. Iandola, M. Moskewicz, S. Karayev, R. Girshick, T. Darrell, and K. Keutzer.
Densenet: Implementing efficient convnet descriptor pyramids.
arXiv preprint arXiv:1404.1869, 2014.
[12]
E. Insafutdinov, L. Pishchulin, B. Andres, M. Andriluka, and B. Schiele.
Deepercut: A deeper, stronger, and faster multi-person pose
estimation model.
In ECCV, 2016.
[13]
H. Jhuang, J. Gall, S. Zuffi, C. Schmid, and M. Black.
Towards understanding action recognition.
In ICCV, 2013.
[14]
A. Newell, K. Yang, and J. Deng.
Stacked hourglass networks for human pose estimation.
In ECCV, 2016.
[15]
B. X. Nie, C. Xiong, and S.-C. Zhu.
Joint action recognition and pose estimation from video.
In CVPR, 2015.
[16]
D. Park and D. Ramanan.
N-best maximal decoders for part models.
In ICCV, 2011.
[17]
T. Pfister, J. Charles, and A. Zisserman.
Flowing convnets for human pose estimation in videos.
In ICCV, 2015.
[18]
L. Pishchulin, M. Andriluka, P. Gehler, and B. Schiele.
Poselet conditioned pictorial structures.
In CVPR, 2013.
[19]
L. Pishchulin, M. Andriluka, and B. Schiele.
Fine-grained activity recognition with holistic and pose based
features.
In GCPR, 2014.
[20]
L. Pishchulin, E. Insafutdinov, S. Tang, B. Andres, M. Andriluka, P. V. Gehler,
and B. Schiele.
Deepcut: Joint subset partition and labeling for multi person pose
estimation.
In CVPR, 2016.
[21]
U. Rafi, J. Gall, and B. Leibe.
An efficient convolutional network for human pose estimation.
In BMVC, 2016.
[22]
V. Ramakrishna, T. Kanade, and Y. Sheikh.
Tracking human pose by tracking symmetric parts.
In CVPR, 2013.
[23]
B. Sapp, C. Jordan, and B. Taskar.
Adaptive pose priors for pictorial structures.
In CVPR, 2010.
[24]
J. Shotton, A. Fitzgibbon, M. Cook, T. Sharp, M. Finocchio, R. Moore,
A. Kipman, and A. Blake.
Real-time human pose recognition in parts from single depth images.
In CVPR, 2011.
[25]
K. Simonyan and A. Zisserman.
Very deep convolutional networks for large-scale image recognition.
CoRR, abs/1409.1556, 2014.
[26]
V. K. Singh and R. Nevatia.
Action recognition in cluttered dynamic scenes using pose-specific
part models.
In ICCV, 2011.
[27]
M. Sun, P. Kohli, and J. Shotton.
Conditional regression forests for human pose estimation.
In CVPR, 2012.
[28]
J. Tompson, R. Goroshin, A. Jain, Y. LeCun, and C. Bregler.
Efficient object localization using convolutional networks.
In CVPR, 2015.
[29]
J. Tompson, A. Jain, Y. LeCun, and C. Bregler.
Joint training of a convolutional network and a graphical model for
human pose estimation.
In NIPS, 2014.
[30]
A. Toshev and C. Szegedy.
Deeppose: Human pose estimation via deep neural networks.
In CVPR, 2014.
[31]
N. Ukita.
Iterative action and pose recognition using global-and-pose features
and action-specific models.
In ICCV Workshop, 2013.
[32]
H. Wang, A. Kläser, C. Schmid, and C.-L. Liu.
Action Recognition by Dense Trajectories.
In CVPR, 2011.
[33]
H. Wang and C. Schmid.
Action recognition with improved trajectories.
In ICCV, 2013.
[34]
J. Wang, X. Nie, Y. Xia, Y. Wu, and S.-C. Zhu.
Cross-view action modeling, learning, and recognition.
In CVPR, 2014.
[35]
S.-E. Wei, V. Ramakrishna, T. Kanade, and Y. Sheikh.
Convolutional pose machines.
In CVPR, 2016.
[36]
B. Yang, J. Yan, Z. Lei, and S. Z. Li.
Convolutional channel features for pedestrian, face and edge
detection.
In ICCV, 2015.
[37]
W. Yang, Y. Wang, and G. Mori.
Recognizing human actions from still images with latent poses.
In CVPR, 2010.
[38]
Y. Yang and D. Ramanan.
Articulated human detection with flexible mixtures of parts.
TPAMI, 2013.
[39]
A. Yao, J. Gall, and L. Van Gool.
Coupled action recognition and pose estimation from multiple views.
IJCV, 2012.
[40]
B. Yao and L. Fei-Fei.
Recognizing human-object interactions in still images by modeling the
mutual context of objects and human poses.
TPAMI, 2012.
[41]
M. Ye, Q. Zhang, L. Wang, J. Zhu, R. Yang, and J. Gall.
A survey on human motion analysis from depth data.
In Time-of-Flight and Depth Imaging. Sensors, Algorithms, and
Applications, 2013.
[42]
T.-H. Yu, T.-K. Kim, and R. Cipolla.
Unconstrained monocular 3d human pose estimation by action detection
and cross-modality regression forest.
In CVPR, 2013.
[43]
D. Zhang and M. Shah.
Human pose estimation in videos.
In ICCV, 2015.
[44]
W. Zhang, M. Zhu, and K. G. Derpanis.
From actemes to action: A strongly-supervised representation for
detailed action understanding.
In ICCV, 2013.
[45]
S. Zuffi, J. Romero, C. Schmid, and M. J. Black.
Estimating human pose with flowing puppets.
In ICCV, 2013. |
A Clear View of a Cloudy Brown Dwarf Companion from High-Resolution Spectroscopy
Jerry W. Xuan
Department of Astronomy, California Institute of Technology, Pasadena, CA 91125, USA
Jason Wang
51 Pegasi b Fellow
Department of Astronomy, California Institute of Technology, Pasadena, CA 91125, USA
Center for Interdisciplinary Exploration and Research in Astrophysics (CIERA) and Department of Physics and Astronomy, Northwestern University, Evanston, IL 60208, USA
Jean-Baptiste Ruffio
Department of Astronomy, California Institute of Technology, Pasadena, CA 91125, USA
Heather Knutson
Division of Geological & Planetary Sciences, California Institute of Technology, Pasadena, CA 91125, USA
Dimitri Mawet
Department of Astronomy, California Institute of Technology, Pasadena, CA 91125, USA
Jet Propulsion Laboratory, California Institute of Technology, 4800 Oak Grove Dr.,Pasadena, CA 91109, USA
Paul Mollière
Max-Planck-Institut für Astronomie, Königstuhl 17, 69117 Heidelberg, Germany
Jared Kolecki
Department of Astronomy, The Ohio State University, 100 W 18th Ave, Columbus, OH 43210 USA
Arthur Vigan
Aix Marseille Univ, CNRS, CNES, LAM, Marseille, France
Sagnick Mukherjee
Department of Astronomy and Astrophysics, University of California, Santa Cruz, CA 95064, USA
Nicole Wallack
Division of Geological & Planetary Sciences, California Institute of Technology, Pasadena, CA 91125, USA
Ji Wang
Department of Astronomy, The Ohio State University, 100 W 18th Ave, Columbus, OH 43210 USA
Ashley Baker
Department of Astronomy, California Institute of Technology, Pasadena, CA 91125, USA
51 Pegasi b Fellow
Randall Bartos
Jet Propulsion Laboratory, California Institute of Technology, 4800 Oak Grove Dr.,Pasadena, CA 91109, USA
Geoffrey A. Blake
Division of Geological & Planetary Sciences, California Institute of Technology, Pasadena, CA 91125, USA
Charlotte Z. Bond
UK Astronomy Technology Centre, Royal Observatory, Edinburgh EH9 3HJ, United Kingdom
Marta Bryan
Department of Astronomy, University of California at Berkeley, CA 94720, USA
51 Pegasi b Fellow
Benjamin Calvin
Department of Astronomy, California Institute of Technology, Pasadena, CA 91125, USA
Department of Physics & Astronomy, 430 Portola Plaza, University of California, Los Angeles, CA 90095, USA
Sylvain Cetre
W. M. Keck Observatory, 65-1120 Mamalahoa Hwy, Kamuela, HI, USA
Mark Chun
Institute for Astronomy, University of Hawaii at Hilo, 640 N Aohoku Pl, Hilo, HI 96720, USA
Jacques-Robert Delorme
W. M. Keck Observatory, 65-1120 Mamalahoa Hwy, Kamuela, HI, USA
Department of Astronomy, California Institute of Technology, Pasadena, CA 91125, USA
Greg Doppmann
W. M. Keck Observatory, 65-1120 Mamalahoa Hwy, Kamuela, HI, USA
Daniel Echeverri
Department of Astronomy, California Institute of Technology, Pasadena, CA 91125, USA
Luke Finnerty
Department of Physics & Astronomy, 430 Portola Plaza, University of California, Los Angeles, CA 90095, USA
Michael P. Fitzgerald
Department of Physics & Astronomy, 430 Portola Plaza, University of California, Los Angeles, CA 90095, USA
Katelyn Horstman
Department of Astronomy, California Institute of Technology, Pasadena, CA 91125, USA
Julie Inglis
Division of Geological & Planetary Sciences, California Institute of Technology, Pasadena, CA 91125, USA
Nemanja Jovanovic
Department of Astronomy, California Institute of Technology, Pasadena, CA 91125, USA
Ronald López
Department of Physics & Astronomy, 430 Portola Plaza, University of California, Los Angeles, CA 90095, USA
Emily C. Martin
Department of Astronomy & Astrophysics, University of California, Santa Cruz, CA95064, USA
Evan Morris
Department of Astronomy & Astrophysics, University of California, Santa Cruz, CA95064, USA
Jacklyn Pezzato
Department of Astronomy, California Institute of Technology, Pasadena, CA 91125, USA
Sam Ragland
W. M. Keck Observatory, 65-1120 Mamalahoa Hwy, Kamuela, HI, USA
Bin Ren
Department of Astronomy, California Institute of Technology, Pasadena, CA 91125, USA
Garreth Ruane
Department of Astronomy, California Institute of Technology, Pasadena, CA 91125, USA
Jet Propulsion Laboratory, California Institute of Technology, 4800 Oak Grove Dr.,Pasadena, CA 91109, USA
Ben Sappey
Center for Astrophysics and Space Sciences, University of California, San Diego, La Jolla, CA 92093
Tobias Schofield
Department of Astronomy, California Institute of Technology, Pasadena, CA 91125, USA
Andrew Skemer
Department of Astronomy & Astrophysics, University of California, Santa Cruz, CA95064, USA
Taylor Venenciano
Physics and Astronomy Department, Pomona College, 333 N. College Way, Claremont, CA 91711, USA
J. Kent Wallace
Jet Propulsion Laboratory, California Institute of Technology, 4800 Oak Grove Dr.,Pasadena, CA 91109, USA
Peter Wizinowich
W. M. Keck Observatory, 65-1120 Mamalahoa Hwy, Kamuela, HI, USA
Abstract
Direct imaging studies have mainly used low-resolution spectroscopy ($R\sim 20-100$) to study the atmospheres of giant exoplanets and brown dwarf companions, but the presence of clouds has often led to degeneracies in the retrieved atmospheric abundances (e.g. C/O, metallicity). This precludes clear insights into the formation mechanisms of these companions. The Keck Planet Imager and Characterizer (KPIC) uses adaptive optics and single-mode fibers to transport light into NIRSPEC ($R\sim 35,000$ in $K$ band), and aims to address these challenges with high-resolution spectroscopy. Using an atmospheric retrieval framework based on petitRADTRANS, we analyze KPIC high-resolution spectrum ($2.29-2.49~{}\mu$m) and archival low-resolution spectrum ($1-2.2~{}\mu$m) of the benchmark brown dwarf HD 4747 B ($m=67.2\pm 1.8~{}M_{\rm{Jup}}$, $a=10.0\pm 0.2$ au, $T_{\rm eff}\approx 1400$ K). We find that our measured C/O and metallicity for the companion from the KPIC high-resolution spectrum agree with that of its host star within $1-2\sigma$. The retrieved parameters from the $K$ band high-resolution spectrum are also independent of our choice of cloud model. In contrast, the retrieved parameters from the low-resolution spectrum are highly sensitive to our chosen cloud model. Finally, we detect CO, H${}_{2}$O, and CH${}_{4}$ (volume mixing ratio of log(CH${}_{4}$)=$-4.82\pm 0.23$) in this L/T transition companion with the KPIC data. The relative molecular abundances allow us to constrain the degree of chemical disequilibrium in the atmosphere of HD 4747 B, and infer a vertical diffusion coefficient that is at the upper limit predicted from mixing length theory.
††facilities: Keck (KPIC)††software: petitRADTRANS (Mollière et al., 2019), dynesty (Speagle, 2020), PyAstronomy (https://github.com/sczesla/PyAstronomy)
1 Introduction
The Keck Planet Imager and Characterizer (KPIC) is a new suite of instrument upgrades at Keck II, including a single-mode fiber injection unit (Mawet et al., 2017; Delorme et al., 2021) that feeds light into the upgraded NIRSPEC (Martin et al., 2018; López et al., 2020), enabling high-resolution spectroscopy (HRS111We will use HRS to abbreviate both high-resolution spectroscopy (the technique) and high-resolution spectra (the data) in this paper. The same is true for LRS: low-resolution spectra or spectroscopy.) at $R\sim 35,000$ in $K$ band. By using single-mode fibers to inject light from planets and brown dwarfs at high-contrast, KPIC provides suppression of the stellar point-spread function at the fiber input and a stable line spread function that is independent of incoming wavefront aberrations (Mawet et al., 2017; Wang et al., 2021b). By observing at high-resolution, we can further distinguish between star and planet light from their spectral differences (Wang et al., 2017; Mawet et al., 2017). Recently, Wang et al. (2021b) published the KPIC detections of HR 8799 c, d, e, demonstrating the ability of KPIC to detect molecular lines and measure the rotational line broadening of planets at high contrast ($\Delta K\approx 11$) and small separations ($\approx 0.4$”) from their host star.
The atmospheric composition of a substellar object holds a wealth of information about its formation, accretion, and evolutionary history, as well as fundamental physical processes that shape its atmosphere. It is therefore important to assess how well KPIC and other fiber-fed, high-resolution spectrographs (e.g. Subaru/REACH Kotani et al. 2020 and VLT/HiRISE Otten et al. 2021) can measure the atmospheric compositions of directly imaged planets and brown dwarfs. Specifically, previous studies of gas giant planet formation have highlighted the carbon-to-oxygen ratio (C/O) and metallicity (e.g. [C/H]) of the atmosphere as informative probes of formation history (e.g. Öberg et al., 2011; Madhusudhan, 2012; Piso et al., 2015). To first order, a companion with a C/O and metallicity similar to that of its host star is consistent with formation via gravitational instability in a protostellar disk or fragmentation in a molecular cloud, akin to binary star formation (Bate et al., 2002). On the other hand, differences between the companion and stellar C/O are suggestive of core accretion (Pollack et al., 1996) as the likely formation mechanism, and in that scenario, could be used to constrain where the companion formed in the disk relative to ice lines of major C- and O-bearing molecules (e.g. H${}_{2}$O, CO${}_{2}$, and CO). This picture can be complicated by a variety of effects such as the relative amount of solids incorporated into the planet’s atmosphere (e.g. Madhusudhan et al., 2014; Öberg & Bergin, 2016; Mordasini et al., 2016; Gravity Collaboration et al., 2020; Pelletier et al., 2021).
So far, atmospheric characterization of directly imaged companions has mostly relied on low-resolution spectroscopy (LRS) with resolving powers of $R\approx 20-100$. LRS is sensitive to continuum emission originating from the deepest observable layer of the atmosphere and modified by opacity sources further up. Many of these companions have temperatures warm enough for silicate clouds to condense in their atmospheres (Marley & Robinson, 2015), and there is much evidence that cloud opacity plays an important role in the LRS of directly imaged companions and brown dwarfs with L or L/T transition spectral types (e.g. Skemer et al., 2014; Burningham et al., 2017; Gravity Collaboration et al., 2020). However, due to our limited knowledge of cloud physics, a reliable assessment of atmospheric abundances from LRS could be fraught with degeneracies between clouds, the pressure-temperature profile, and chemical abundances (e.g. Burningham et al., 2017). In addition, the retrieval results can also be highly sensitive to systematics in different data sets that are combined to obtain a wider wavelength coverage (Wang et al., 2020a). More encouragingly, Mollière et al. (2020) report atmospheric abundances that are relatively robust to clouds and model choices, though Burningham et al. (2021) show that issues such as an unphysically small radius could persist despite improvements in cloud modeling and extensive wavelength coverage ($1-15\mu$m).
Recently, Wang et al. (2022) presented the first atmospheric free retrievals at high-resolution for a directly imaged companion. They studied the L-type brown dwarf HR 7672 B ($T_{\rm eff}\approx 1800$ K) using KPIC HRS and near-infrared photometry, and measured carbon and oxygen abundances that are consistent within $<1.5\sigma$ to that of its host star. In this paper, we present a detailed atmospheric study of HD 4747 B using both KPIC HRS ($K$ band) and archival low-resolution spectra (LRS) from 1-2.2 $\mu$m that we re-extract in a uniform manner. While the KPIC HRS resolves individual molecular lines and conains direct information about a companion’s atmospheric abundances, LRS provides spectral shape and luminosity measurements, which has the potential to complement the HRS.
Compared to HR 7672 B, HD 4747 B is a colder L/T transition object ($T_{\rm eff}\approx 1400$ K) with strong evidence for clouds and a similar color to directly imaged planets such as HR 8799 c,d,e (Crepp et al., 2018; Peretti et al., 2019). Like HR 7672 B, the wealth of prior knowledge available for HD 4747 B makes it a valuable benchmark object to test whether we can make robust inferences with spectroscopic data. First, we are able to precisely measure the dynamical mass of HD 4747 B (§ 2.2). Mass is a fundamental quantity that is poorly constrained for most directly imaged companions (Bowler, 2016). Furthermore, given its high mass, HD 4747 B is expected to have formed via direct gravitational collapse in the same cloud or disk as its host star, which means that we can assume chemical homogeneity: the brown dwarf and primary star should share the same chemical composition. Finally, with the companion mass, observed luminosity, and stellar age, we can independently estimate the brown dwarf’s radius from evolutionary models.
In this paper, we use the open-source radiative transfer code petitRADTRANS (Mollière et al., 2019, 2020) to fit the HRS and LRS for HD 4747 B in a retrieval framework. The main goals of our study are to measure the atmospheric composition of this brown dwarf companion using both the HRS and LRS, and to present a detailed characterization of its atmosphere, including constraints on clouds, chemical equilibrium or disequilibrium, and the detection of CH${}_{4}$. In this process, we also explore the relative advantages and disadvantages of HRS versus LRS.
This paper is organized as follows: in § 2, we summarize the system properties including our mass measurement for HD 4747 B. Our spectroscopic data and data reduction procedure is described in § 3. We then discuss our spectral analysis framework in § 4. We present individual and joint retrievals of the HRS and LRS in § 5, § 6 and § 7, respectively. We summarize the lessons learned in § 8, and conclude in § 9.
2 System properties
2.1 Host star
In this section, we summarize relevant properties of the host star. HD 4747 is a main-sequence, solar-type star located $\approx 19$ parsec away based on its Gaia eDR3 parallax (Brown et al., 2021). Chromospheric emission in the Ca II H&K lines are visible in the stellar spectrum (log$R_{\rm HK}=-4.72\pm 0.02$), which Peretti et al. (2019) used to derive an age of $2.3\pm 1.4$ Gyr from the age-log$R_{\rm HK}$ calibration of Mamajek & Hillenbrand (2008). This agrees with the gyro-chronological age estimate of $3.3^{+2.3}_{-1.9}$ Gyr from Crepp et al. (2018). These studies also converged on $T_{\rm eff}$ around 5300–5400 K, and a surface gravity log($g$) of 4.5–4.65. Of particular relevance to this study are the C/O ratio and metallicity of the host star, since we expect these to be roughly similar to those of the brown dwarf. HD 4747 is found to have a sub-solar metallicity, with [Fe/H]=$-0.23\pm 0.05$ from Peretti et al. (2019) and [Fe/H]=$-0.22\pm 0.04$ from Crepp et al. (2018). Previous studies including Brewer et al. (2016) and Peretti et al. (2019) also measured the elemental abundances for the star, but either did not take into account non-local thermodynamic equilibrium (LTE) effects on their oxygen abundances (Amarsi et al., 2019) or do not quote error bars. We instead carry out a new analysis using the method described in Kolecki et al. (2021) to derive the abundances for different elements, and correct the results to account for 3D non-LTE effects (Amarsi et al., 2019) on the results. For this analysis, we used an archival spectrum from FEROS (Kaufer et al., 1997) which covers 350-920 nm at $R=48,000$. Using this spectrum, we measure the equivalent widths of absorption lines and compare them to model stellar atmospheres in an iterative approach using the MOOG code (Sneden, 1973). From our derived carbon and oxygen abundances, we find C/O=$0.48\pm 0.08$. The iron abundance is [Fe/H]=$-0.30\pm 0.5$, in agreement with previous studies.
Since Fe condenses out for temperatures below $\approx 1800$ K (Marley & Robinson, 2015), it is not a relevant gaseous absorber in the photosphere of HD 4747 B. Therefore, the more useful metrics for comparison are C and O. From our analysis above, we find [C/H]$=-0.08\pm 0.06$ and [O/H]$=-0.02\pm 0.04$ for the host star. [C/H] is defined as $\log_{10}({N_{\rm C}/N_{\rm H}})_{\rm star}$ - $\log_{10}({N_{\rm C}/N_{\rm H}})_{\rm sun}$, where $N_{\rm C}$ and $N_{\rm H}$ are the number fraction of C and H respectively. [O/H] is defined similarly. We adopt Asplund et al. (2009) as our solar reference in order to be consistent with petitRADTRANS, which we use to model the atmosphere of HD 4747 B.
2.2 Orbit and Dynamical mass
The orbit and mass of HD 4747 B have been measured by several studies using relative astrometry from Keck/NIRC2, host star radial velocities (RV) from Keck/HIRES, and Gaia-Hipparcos absolute astrometry (Brandt et al., 2019; Xuan & Wyatt, 2020). Here, we take advantage of 23 yr of RV observations published in Rosenthal et al. (2021) and the improved precision of the Hipparcos-Gaia Catalog of Accelerations (HGCA) (Brandt, 2021) based on Gaia eDR3 (Brown et al., 2021) to update the orbit and mass of HD 4747 B. HD 4747 B shows significant proper motion anomalies (PMa) in both the Gaia and Hipparcos epochs, with S/N of $77.2$ and $9.1$, and the position angle and amplitude of the PMa is consistent with being induced by the brown dwarf companion. For the relative astrometry, we use data points tabulated in Brandt et al. (2019), except for the two GPI epochs measured by Crepp et al. (2018), which we replaced with our new measurements from § 3.2. We choose not to use the companion RV as measured by KPIC for this fit, because it does not appreciably improve our already well-constrained orbital solution.
To fit the relative astrometry, radial velocity, and absolute astrometry from Gaia and Hipparcos together, we use the orvara package (Brandt et al., 2021b), which is designed to jointly fit these types of data and takes into account the Gaia and Hipparcos astrometry at the epoch astrometry level using htof (Brandt et al., 2021a). We use the priors listed in Table 4 of Brandt et al. (2021b) for the fitted parameters. The posterior is sampled using the parallel-tempering MCMC sampler (Vousden et al., 2016), a fork of emcee (Foreman-Mackey et al., 2013) over 50000 steps with 10 temperatures and 100 walkers. The fits converged as determined by visual inspection of the chains, and we discarded the first ten percent as burn-in. In orvara, the system parallax and other linear parameters are analytically marginalized out to speed up the fits.
The resulting orbit and mass measurements are tabulated in Table 1, while the model fits are shown in Appendix A. We find a companion mass of $m=67.2\pm 1.8~{}M_{\rm{Jup}}$, which is consistent with previous values, but more precise. We checked the orvara results with a second fit where we model the Gaia and Hipparcos astrometry using the methodology in Xuan & Wyatt (2020). This gives $m=67.1\pm 2.0~{}M_{\rm{Jup}}$, consistent with the orvara result. Furthermore, the companion mass and orbital parameters we find are also consistent with results from orbit fits that only use RV and imaging data (no Gaia-Hipparcos astrometry) from Peretti et al. (2019) and Crepp et al. (2018). We adopt the companion mass from our orvara fit for the spectral analysis in this paper.
3 Spectroscopic data
3.1 High-resolution spectroscopy
3.1.1 KPIC Observations
We observed HD 4747 B on UT 2020 September 28 with Keck/NIRSPEC. The data were collected using the first version of the KPIC fiber injection unit (FIU) (Delorme et al., 2021). The FIU is located downstream of the Keck II adaptive optics system and is used to inject light from a selected object into one of the single mode fibers connected to NIRSPEC. We obtain spectrum in K band, which is broken up into nine echelle orders from 1.94-2.49 $\mu$m. The observing strategy is identical to that of Wang et al. (2021b). In short, we placed the companion on the fiber with the highest throughput and acquired six exposures of 600 seconds each, for a total integration time of 1 hour. The relative astrometry of the companion was computed using whereistheplanet.com (Wang et al., 2021a), based on data in Peretti et al. (2019). For calibration purposes, we acquired a pair of 60 second exposures of the host star before observing the companion, and a pair of 60 second exposures of a telluric standard star (HIP 6960) after the companion exposures so as to share nearly the same airmass. Using exposures on the host star, we calculated an end-to-end throughput from the top of the atmosphere to the detector of $1.8-2.0\%$ during the observations.
3.1.2 Data reduction
To extract the spectra from the raw data, we follow the procedure outlined in Wang et al. (2021b), which the KPIC team has implemented in a public Python pipeline.222https://github.com/kpicteam/kpic_pipeline The images for all objects were reduced in the same way. First, we removed the thermal background from the images using combined instrument background frames taken during daytime. As shown in Wang et al. (2021b), the thermal background of our data is dominated by the warm optics rather than the sky background. We also remove persistent bad pixels identified using the background frames. Then, we use data from the telluric standard star to fit the trace of each column in the four fibers and nine spectral orders, which give us the position and standard deviation of the PSF (in spatial direction) at each column. The trace positions and widths were additionally smoothed using a cubic spline in order to mitigate random noise. We adopt the trace locations and widths as the line spread function (LSF) positions and widths in the dispersion dimension.
For every frame, we then extracted the 1D spectra in each column of each order. To remove residual background light, we subtracted the median of pixels that are at least 5 pixels away from every pixel in each column. Finally, we used optimal extraction to sum the flux using weights defined by the 1D Gaussian LSF profiles calculated from spectra of the telluric star.
The extracted spectra have a median signal-to-noise ratio (S/N) of $\approx 8$ per pixel element, which has a typical width of $0.2~{}\textup{\AA}$, and consists of a mixture of light from the brown dwarf companion and stellar speckles. The S/N of KPIC is optimized for wavelengths around $2.3~{}\mu$m, where CO has a series of strong absorption lines. For our analysis, we use three spectral orders from 2.29-2.49 $\mu$m, which contain the strongest absorption lines from the companion and have relatively few telluric absorption lines. Note that the three spectral orders have gaps in between them, so we have data over $\approx 0.13~{}\mu$m (instead of $0.2~{}\mu$m; see Fig. 3).
As a preliminary analysis, we cross-correlate our KPIC spectra with single-molecule templates assuming $T_{\rm eff}=1400$ K and log($g$)=5.5 from the Sonora model grid (Marley et al., 2021). In short, we estimate the maximum likelihood value for both the single-molecule companion flux and speckle flux in the data as a function of RV (radial velocity) shift using the method described in Wang et al. (2021b), which is based on Ruffio (2019). We find that H${}_{2}$O and CO are detected with S/N of 8.5 and 13.5 respectively (Fig. 1). CH${}_{4}$ is not detected with statistical confidence in this crude analysis, but we present evidence for a weak CH${}_{4}$ detection in § 5.4.
3.2 Low-resolution spectroscopy
3.2.1 Gemini Planet Imager IFS
The Gemini Planet Imager (GPI) observed HD 4747 B on UT 2015 December 24 and 25, in the $K_{1}$ (1.90-2.19 $\mu$m, $R=66$) and $H$ (1.50-1.80 $\mu$m, $R=46.5$) bands, respectively, and the data were published in Crepp et al. (2018). After doing some fits to the published spectrum, we found that the average flux levels of the $K_{1}$ and $H$ bands are inconsistent, and the error bars appear to be significantly over-estimated.
We therefore re-extracted the GPI spectrum using the pyKLIP package (Wang et al., 2015), which models a stellar point spread function (PSF) with Karhunen-Loève Image Processing (KLIP, also known as Principal Component Analysis) following the framework in Soummer et al. (2012) and Pueyo (2016). We tested various model choices to minimize the residuals after stellar PSF subtraction while preserving the companion signal. A key parameter we tuned was the number of Karhunen-Loève (KL) modes. KL modes represent an orthogonal basis for patterns in the images that are used to model the stellar PSF. We chose 5 and 12 KL modes to subtract the stellar PSF in the $H$ and $K_{1}$ band data, respectively. After subtracting the stellar PSF, we first extracted the companion’s relative astrometry in terms of separation and position angle, which are reported in Appendix B (Table 5). Then, we extracted the flux at the companion’s determined location as a function of wavelength, which gave us the raw spectrum. Note that rather than using spectral differential imaging (SDI) to subtract the stellar PSF, we only used angular differential imaging (ADI). For a bright companion like HD 4747 B, ADI is more than sufficient to properly remove the PSF of the star given sufficient parallactic angle rotation.
To flux-calibrate the raw spectrum, we used the satellite spot flux ratios333https://www.gemini.edu/instrumentation/gpi/capability to find the companion-to-star flux ratio. To obtain the observed flux density of the companion, we empirically determined the flux scaling factor $R^{2}/d^{2}$ by fitting a PHOENIX model (Husser et al., 2013) of the star ($T_{\rm eff}=5400$, log($g$)=4.5, and [Fe/H] = -0.5) using the star’s 2MASS J, H, K (Cutri et al., 2003) and the Gaia G band magnitudes (Riello et al., 2021). The zeropoint fluxes and filter transmission of the photometric bands are downloaded from the SVO Filter Service444http://svo2.cab.inta-csic.es/theory/fps/ and the Gaia website.555https://www.cosmos.esa.int/web/gaia/edr3-passbands To obtain measurement uncertainties, we injected 20 fake companions at the same separation and equally spaced position angles in the data, and repeated the same spectral extraction process. We avoided using the fake injections that were within $20\degree$ of the real companion to avoid biasing the fluxes. We inflated the uncertainties on the extracted spectra by $2.5\%$ to account for errors in the stellar flux calibration. The value of $2.5\%$ is estimated by comparing our empirically computed flux scaling factor with the value of $R^{2}/d^{2}$ of the star (using the radius from Peretti et al. 2019 and the Gaia parallax).
3.2.2 SPHERE IFS
HD 4747 B was observed on UT 2016 December 12 and 2017 September 28 with the Spectro-Polarimetric High-contrast Exoplanet Research (SPHERE; Beuzit et al. 2019). The SPHERE Integral Field Spectrograph (IFS) (Claudi et al., 2008) collects data in the $YH$ band from 0.95-1.6 $\mu$m ($R=29$). The extracted spectra was published in Peretti et al. (2019), but is not available. We therefore reduced the raw data using the SPHERE pipeline (Vigan, 2020), and performed a similar post-processing procedure with pyKLIP as described above for the GPI spectra. The only difference is that we needed to use ADI+SDI to perform PSF-subtraction for the SPHERE IFS data, which did not have enough parallactic angle rotation (only $\approx 0.2\degree$). For the SPHERE IFS data, flux calibration is based on unocculted observations of the host star. We chose to use the 2017 data for our analysis since it was taken under much better observing conditions and yields slightly higher spectral S/N than the 2016 data, despite shorter integration times.
3.2.3 Results and comparison with previous LRS
Our newly extracted GPI and SPHERE spectra are plotted in black circles and squares respectively in Fig. 2, and available in Appendix B (Table 4). The absolute flux scaling of our GPI spectrum agrees well with the published spectrum in gray from Crepp et al. (2018), but the uncertainties are much smaller. The shape of our new SPHERE spectrum also agrees well with that in Peretti et al. (2019). From the PSF-subtracted images, the brown dwarf companion is detected with a median S/N per wavelength bin of $\approx 61$ and $\approx 26$ in the GPI $H$ and $K_{1}$ bands, and $\approx 20$ in the SPHERE data. When comparing the extracted spectrum to that of field brown dwarfs from the SPEX library (Burgasser, 2014) in Fig. 2, we find that the newly extracted GPI spectrum is in better agreement compared to the previously published spectrum. As in Crepp et al. (2018) and Peretti et al. (2019), we find a spectral type near the L/T transition (the best matching spectra were from a L9.5 and T0 dwarf). The SPHERE IFS spectrum increases our wavelength coverage by a factor of $\sim 2$, which we find is important for constraining model atmosphere parameters in our fits to the LRS.
4 Spectral analysis
4.1 Forward modeling the KPIC high-resolution spectrum
Here, we briefly describe the framework to forward model and fit the HRS from KPIC, which follows Wang et al. (2021b). When a companion of interest is aligned with one of the KPIC fibers, the companion light and a fraction of light from the host star’s speckle field are injected into the fiber. At the projected separation of HD 4747 B ($\approx 0.6$″), we find the speckles are roughly the same brightness as our companion ($K_{s}\approx 14.4$ from Crepp et al. 2016). Furthermore, the light is transmitted through Earth’s atmosphere and modulated by the instrument optics. Thus, we build two forward models (one for the companion, one for speckles) and jointly fit them as a linear combination. Below we detail how we generate each of the model components.
The companion spectral templates are generated with petitRADTRANS. We shift the templates in wavelength space to fit for the radial velocity. Then, we rotationally broaden the templates by a projected rotation rate $v\sin{i}$ using the fastRotBroad function in PyAstronomy (Czesla et al., 2019), and convolve the templates with the instrumental LSF. The effect of limb darkening is included in petitRADTRANS by integrating intensities along multiple angles between the ray and atmospheric normal.
Next, we multiply the companion model by the telluric response function, which characterizes the atmospheric transmission as a function of wavelength and includes telluric absorption lines. The telluric model is calculated by dividing the spectrum of the standard star (HIP 6960) by a PHOENIX stellar model with matching properties ($T_{\rm eff}=9200$ and log($g$)=4.0).
To model the speckle contribution to the data, we use on-axis observations of the host star taken before the companion exposures. These observations are reduced in the same way as the companion spectra, but have much higher S/N. Unlike the companion models, the host star observations are already modulated by telluric transmission.
The last step is to remove continuum variations. The KPIC spectra are not flux-calibrated and contain a smoothly varying continuum due to stellar speckles and wavelength-dependent atmospheric refraction. Therefore, we apply high-pass filtering with a median filter of 100 pixels ($\sim 0.002~{}\mu$m) on both the data and models to subtract the continuum following Wang et al. (2021b). To determine the optimal filter size, we carried out a series of injection-recovery tests, and found that $\sim 100$ pixels is best at recovering weak molecular signals for our data set. Larger filter sizes (e.g. 200 pixels or more) do not remove the continuum sufficiently, and smaller filter sizes (50 pixels or less) tend to be overly aggressive at removing weak molecular signals.
Finally, we flux-normalize both the companion and stellar models and multiply them by different flux scaling factors, which are fitted parameters. The flux scales are in units of counts as measured by the NIRSPEC detector. After scaling, the companion and speckle models are added and the same high-pass filter is applied on the final model before fitting it to the data.
4.2 Atmospheric retrieval setup
We implement a ‘retrieval’ framework based on petitRADTRANS to model the data, which means that we freely retrieve the chemical abundances, vertical temperature structure, and cloud properties from the data. Previous studies have used retrievals to model HRS of self-luminous exoplanets and brown dwarfs (e.g. Burningham et al., 2017; Mollière et al., 2020), and show that it can be a powerful alternative to fitting self-consistent grid models, which solve for the abundances and temperature profiles with physical assumptions such as chemical equilibrium. The retrieval approach allows more flexibility to fit the data and can potentially provide much more detailed information about the atmospheric properties, with the caveat that it is important to check for physical plausibility of the models since retrievals need not be self-consistent.
In our main set of retrievals, we fit for the chemical abundances in terms of C/O and atmospheric metallicity [C/H]666We denote the atmospheric metallicity as [C/H] because we are only sensitive to C- and O-bearing molecules in this brown dwarf’s atmosphere. along with a quench pressure (where the chemical timescale of a certain reaction is equal to the mixing timescale) to allow for disequilibrium chemistry, the temperature profile (§ 4.2.1), the cloud structure (§ 4.2.2), and other parameters such as the radius. We denote these quenched chemistry retrievals to distinguish from free retrievals where the abundances of each gas species is fit independently. Each component of the model is described in the subsections below.
We use the correlated-k and line-by-line opacity sampling methods in petitRADTRANS for the low-resolution and high-resolution retrievals respectively. For high-resolution, we include opacities for CO, H${}_{2}$O, CH${}_{4}$, NH${}_{3}$, and CO${}_{2}$, and for low-resolution we additionally include Na and K. This is because the alkali lines have wings which affect the $\sim 1~{}\mu$m portion of the LRS, while their opacities are negligible over the portion of $K$ band covered by our HRS. We repeated our baseline HRS retrieval with Na and K included and found that the addition of these two species did not influence the results or improve the fit.
Because the native high-resolution opacities are at $R=10^{6}$, much higher than the resolution of our HRS resolution ($R\approx 35,000$), we down-sampled the opacity tables by a factor of six in order to speed up the retrievals (by roughly the same factor) and reduce the corresponding computational cost. We checked that the maximum deviation in synthetic spectra obtained by using the down-sampled opacities relative to the full-resolution opacity model is $<5\%$ of the minimum HRS error bars. In addition, we repeated our fiducial HRS retrieval with the native opacities ($R=10^{6}$) and found that it yielded the same results. We re-binned the correlated-k opacities to $R=200$ for our fits to the LRS, which has a maximum resolution of $66$. We also repeated our fiducial LRS retrieval at the native $R=1000$ opacities and found the results are fully consistent.
4.2.1 Temperature structure and chemistry
We retrieve the pressure-temperature (PT) profile of the brown dwarf between $P=10^{-4}-10^{3}$ bars, which sets the vertical extent of the atmosphere. We use the P-T profile parametrization from Mollière et al. (2020) which has six free parameters. The spatial coordinate is an optical depth $\tau=\delta P^{\alpha}$, where $\delta$ and $\alpha$ are the first two parameters. The atmosphere then consists of a high altitude region (top of atmosphere to $\tau=0.1$) fitted with three temperature points equi-distant in log pressure, a middle radiative region ($\tau=0.1$ to radiative-convective boundary) which uses the Eddington approximation with $T_{0}$ as the ‘internal temperature’, and a lower region (radiative-convective boundary to bottom of atmosphere), which is set to follow the moist adiabatic temperature gradient once the atmosphere becomes unstable to convection (Mollière et al., 2020). We ignore stellar irradiation as a source of heat because the total incident energy on HD 4747 B at periastron ($\approx 2.7$ au) is approximately four orders of magnitude less than its luminosity, which is dominated by the brown dwarf’s internal energy.
In our quenched chemistry retrievals, the C/O, [C/H], and P-T profile determine the equilibrium chemical abundances (mass fractions of molecules) as a function of pressure, by interpolating the chemical equilibrium table from Mollière et al. (2020). The opacities we include in the models are listed in § 4.2. In petitRADTRANS, the abundances of all metals except oxygen are assumed to scale together such that [C/H] = [Si/H] = [N/H], etc. Then, C/O and [C/H] are combined to set the oxygen abundance (Mollière et al., 2020). We use Asplund et al. (2009) as our reference for the solar metallicity in these models.
Finally, we include a quench pressure $P_{\rm quench}$ which fixes the abundances of H${}_{2}$O, CO, and CH${}_{4}$ where $P<P_{\rm quench}$ using the equilibrium values found at $P_{\rm quench}$ (Zahnle & Marley, 2014; Mollière et al., 2020). The inclusion of $P_{\rm quench}$ allows for the possibility of disequilibrium chemistry, which occurs where the atmospheric mixing timescale is shorter than the chemical reaction timescale. We only include a quench pressure for the net reaction between H${}_{2}$O, CO, and CH${}_{4}$ because these molecules are the only ones detectable in our KPIC HRS (see § 5.4 for the CH${}_{4}$ detection), and chemical kinetics modeling indicates that the abundances of these three molecules are closely linked to each other by a series of reactions (e.g. Moses et al., 2013). In summary, our quenched chemistry retrievals use C/O, [C/H] and $P_{\rm quench}$ to set the abundances of each gas species for a given P-T profile.
4.2.2 Clouds
Crepp et al. (2018) and Peretti et al. (2019) analyzed LRS for HD 4747 B and found evidence for a cloudy atmosphere. We summarize their results in Table 2 along with our new measurements. In this study, we consider both clear and cloudy models in order to explore the sensitivity of our retrieved abundances to the assumed cloud properties. For our cloudy model, we use the EddySed model from Ackerman & Marley (2001) as implemented in petitRADTRANS (Mollière et al., 2020). In this model, the cloud particles both absorb and scatter the outgoing photons from the atmosphere according to measured optical properties (Mollière et al., 2019). The cloud particles can be either crystalline or amorphous, and the opacities of the clouds are computed assuming either homogeneous and spherical particles, modeled with Mie theory, or irregularly-shaped cloud particles, modeled with the Distribution of Hollow Spheres (DHS) (Min et al., 2005; Mollière et al., 2019).
For HD 4747 B, we consider models with two different cloud species (MgSiO${}_{3}$ and Fe) and properties (amorphous or crystalline particles). We choose to focus on MgSiO${}_{3}$ and Fe for several reasons. First, the condensation curves of these two species intersect the thermal profile of a $T_{\rm eff}=1400$ K, log($g$)=5.5 object from the Sonora atmospheric model (Marley et al., 2021) at $\sim 10-50$ bars. While the Sonora model is cloudless, it provides a rough estimate of which cloud species are relevant. Second, recent theoretical work has shown that MgSiO${}_{3}$ is expected to be the most important cloud species for substellar objects with $T_{\rm eff}>950$ K due to its low nucleation energy barriers and the relatively high elemental abundances of Mg, Si, and O (Gao et al., 2020). Finally, studies using mid-IR spectroscopy from Spitzer have found direct evidence for a MgSiO${}_{3}$ absorption feature at $\sim 10~{}\mu$m in field brown dwarfs (Cushing et al., 2006; Luna & Morley, 2021), and specifically amorphous MgSiO${}_{3}$ (Burningham et al., 2021). Although MgSiO${}_{3}$ and Fe clouds do not have distinct features in the near-IR, they still impact the near-IR spectrum by contributing a wavelength-dependent opacity. Our baseline model uses amorphous MgSiO${}_{3}$ modeled with Mie theory (abbreviated MgSiO${}_{3}$, ‘am’) for the clouds. In addition, we also consider models with MgSiO${}_{3}$, ‘cd’, which assumes crystalline cloud particles modeled with DHS, as well as models with two cloud species (MgSiO${}_{3}$ + Fe) for the LRS.
4.2.3 Methane opacities
Given that HD 4747 B is located near the L/T transition for brown dwarfs, we might expect to observe methane in its atmosphere. Previous $L$ band studies have detected methane in field brown dwarfs with spectral types as early as mid-L, or up to $T_{\rm eff}\approx 1800$ K (Noll et al., 2000; Johnston et al., 2019). In this study, we adopted the HITEMP CH${}_{4}$ line list from Hargreaves et al. (2020), which we convert into opacities following the petitRADTRANS documentation. When cross-correlating a model generated with the HITEMP CH${}_{4}$ opacities with a late T dwarf, we obtained a CCF S/N of $\approx 15$, in comparison to $\approx 5$ when cross-correlating with a model generated from the default CH${}_{4}$ opacities from ExoMol (Yurchenko & Tennyson, 2014) in petitRADTRANS.
4.2.4 Additional fit parameters
petitRADTRANS computes the flux density as emitted at the surface of the object. For the LRS, we scale the model by the distance and companion radius, where the radius is another free parameter, and the distance is taken from the Gaia eDR3 parallax (Brown et al., 2021). For the HRS, we also fit the companion’s radial velocity and $v\sin{i}$, as well as an error multiple term to account for any underestimation in the data uncertainties.
Due to imperfect starlight subtraction in the spectral extraction process, we found that our LRS likely still contains correlated noise from the wavelength-dependence of speckles, as has been noted by several previous studies on high-contrast companions (e.g. De Rosa et al., 2016; Samland et al., 2017; Currie et al., 2018; Wang et al., 2020b, 2021c). This is evident in the residual frames, where we can see speckles at 5-20% of the companion intensity in the PSF-subtracted images. We therefore adopt a Gaussian process with a squared exponential kernel to empirically estimate the correlated noise in the GPI H, K and SPHERE YJH bands when fitting models to the data. Following Wang et al. (2020b), we assume that our extracted error bars contain a fraction $f_{\rm amp}$ of correlated noise, and $1-f_{\rm amp}$ of white noise, and fit for $f_{\rm amp}$ and the scale of correlation $l$. This adds $2\times 3=6$ additional parameters to the retrievals.
As an alternative model, we also tried fitting the LRS with error inflation terms and flux scaling factors for the SPHERE and GPI spectra along the lines of Mollière et al. (2020), but found that our results were very sensitive to our choice of prior for the flux scaling factor. We conclude that our GP model is better suited to account for correlated noise from speckles, and use it in all LRS fits presented in this work.
4.3 Priors
We adopt uniform or log-uniform priors for all model parameters except for the mass, for which we use a Gaussian prior of $67.2\pm 1.8~{}M_{\rm{Jup}}$ from the dynamical mass measurement (§ 2.2). For the parametric P-T profile parameters, we exclude profiles that contain temperature inversions, as the heat budgets of widely separated companions are dominated by their internal luminosities. For the companion’s radius, we use a uniform prior between 0.6 - 1.2 $R_{\rm{Jup}}$. When including a quench pressure, we use a log-uniform prior from $10^{-4}$ - $10^{3}$, which is the full pressure range of our models. The priors for all retrieval parameters are tabulated in Appendix C.
4.3.1 Model fitting with nested sampling
We use nested sampling as implemented by dynesty (Speagle, 2020) to find the posterior distributions for the model parameters. Specifically, we use 200 live points and adopt the stopping criterion that the estimated contribution of the remaining prior volume to the total evidence is less than 1%. We repeated a few retrievals using 1000 live points and found the evidence remains roughly the same, implying the fits have converged when using 200 live points.
One advantage of adopting nested sampling is that we can use the Bayesian evidence from each fit to calculate the Bayes factor $B$, which assesses the relative probability of model $M_{2}$ compared to $M_{1}$. We will use the Bayes factor to compare different models throughout this paper to determine whether a given $M_{2}$ is justified over $M_{1}$. In Table 2, we take a baseline model (MgSiO${}_{3}$, am) to be $M_{1}$ and compare other models to it. Based on Jeffreys (1983), a model with 100 times lower $B$ than the model with the highest $B$ can be ‘decisively’ rejected. $B$ of $\lesssim 10$ is considered weak evidence for preferring one model over the other. We first run retrievals with only the HRS (§ 5), only the LRS (§ 6), as well as joint retrievals with both HRS and LRS (§ 7).
5 High-resolution retrievals (KPIC)
5.1 Overview
From our HRS retrievals of HD 4747 B, we find that both clear and cloudy models yield consistent results for the atmospheric parameters (abundances, temperature structure, quenching) and bulk properties (radius, radial velocity, spin). A few selected parameters are plotted in Fig. 4 and tabulated in Table 2. The insensitivity of the HRS retrieval results to clouds, a major finding of this paper, is discussed in § 5.2. In Fig. 3, we plot the data, a best fit model, and residuals for the baseline HRS retrieval. We report values from this retrieval as the final results of this paper, with selected parameters shown in the first row of Table 2 and joint posterior distributions in Appendix C. We also plot the contribution from the planet and star separately at their best-fit flux levels. We compute the auto-correlation function of the residuals and find that there is no evidence for correlated noise or strong systematics. Unless otherwise specified, we quote results from the baseline EddySed cloud model (MgSiO${}_{3}$, am). See Appendix C for the posterior distributions of other parameters from our baseline model.
To make sure that we are fitting the correct signal, we check the RV and flux level of the companion. From our orbital posteriors for HD 4747 B, the expected RV shift on the night of our HRS observation is $15.0\pm 0.1$ km/s in the Earth’s reference frame, which is a combination of the system barycenter velocity, the Earth’s relative velocity with respect to HD 4747, and the companion’s orbital velocity. The fitted RV of $15.0\pm 0.4$ km/s agrees perfectly with this value (see Fig. 4). In addition, the companion flux level in the spectral orders from 2.29 to 2.49 $\mu$m is $85\pm 10$ counts, comparable to the speckle flux levels in these orders. Taking into account the difference in wavelengths and the difference in integration time (600 s for the companion, 60 s for the on-axis star), we estimate that our measured companion flux corresponds to $\Delta K_{s}=8.3\pm 0.3$ mag, which is within $3\sigma$ of the photometric $\Delta K_{s}=9.05\pm 0.14$ mag reported by Crepp et al. (2018). The agreement between these contrast values are reasonably good given the time-varying throughput of KPIC (Delorme et al., 2021), and the fact that we subtract out the continuum with high-pass filtering, effects which complicate a direct flux comparison.
Fig. 4 also shows the projected spin rate $v\sin{i}=13.2^{+1.4}_{-1.5}$ km/s, which is comparable to the rotation rates observed for field brown dwarfs with similar spectral types (e.g. Konopacky et al., 2012). We also plot the retrieved quench pressure $P_{\rm quench}$ in Fig. 4, which indicates that the chemical reaction timescale becomes longer than the vertical mixing timescale at pressures lower than $P_{\rm quench}$. Thus, disequilibrium chemistry is clearly affecting the atmosphere (see § 5.3 for details).
We compute $T_{\rm eff}$ by sampling from our posterior to generate low-resolution models over a large wavelength range (0.5 to 30 $\mu$m) and calculating the integrated flux. We then solve for $T_{\rm eff}$ using the Stefan-Boltzmann law. When computing $T_{\rm eff}$, we include opacities from Na and K, which are important sources of opacity near visible wavelengths. As shown in Table 2, the retrieved radius and $T_{\rm eff}$ from HRS have broad distributions, which reflect the relatively weak luminosity constraints from the HRS (log($L_{\rm bol}/L_{\odot}$)=$-4.33^{+0.23}_{-0.25}$). This is because the HRS is not flux-calibrated and we remove the continuum in our fits. Comparing to values of radius and $T_{\rm eff}$ from previous work based on LRS (Crepp et al., 2018; Peretti et al., 2019), our retrieved values from the HRS retrievals are consistent at the $1-2\sigma$ level (see Table 2). We discuss the constraints on these parameters from the LRS in § 6.2.
We compare our retrieved [C/H], [O/H], and C/O with that of the host star (see § 2.1) in Fig. 5. Our retrieved C abundance agrees well with the host star value, while the O abundance is lower by about $1\sigma$. This results in our retrieved C/O for the companion being higher by about $2\sigma$ compared to the stellar value. Here and elsewhere in the paper, we compute the ‘$\sigma$ difference’ between two measurements by dividing the difference in the two median values by the quadrature sum of the uncertainties from both measurements. We discuss the implications of our measured abundances for HD 4747 B in § 8.4.
5.2 Why are our KPIC HRS insensitive to clouds in HD 4747 B?
Clouds represent a significant source of uncertainty in many published models of substellar atmospheres (e.g. Burningham et al., 2017; Wang et al., 2020a). However, we find that the retrieved parameters from our KPIC HRS are insensitive to the choice of cloud model for HD 4747 B. As shown in Fig 4, the posteriors for radius, RV, $v\sin{i}$, C/O, [C/H], and quench pressure are nearly identical across the various models. The same is true for other parameters.
Table 2 shows that the different cloud models are indistinguishable for the KPIC HRS; the clear model fits as well as the cloudy models, with $B=0.61$, which does not pass the threshold of $B=10/0.1$ to be considered statistically favored/disfavored. This indicates that the data can be fitted adequately without clouds; indeed the cloud parameters for the EddySed models span their respective prior ranges almost uniformly as shown in Appendix C. As we will discuss in § 6, the LRS show that the atmosphere of HD 4747 B is cloudy. This implies that cloud opacity must be minimal at the pressures probed by our HRS.
To understand this, we plot in Fig. 6 the retrieved P-T profiles (black and blue lines), cloud condensation curves (dashed lines), and emission contribution functions. The left and right panels show results from the HRS and LRS retrievals, respectively. The emission contribution function for HRS shows that we are sensitive to pressures ranging from a few bars, where the continuum forms, up to $\approx 10^{-2}$ bars in the cores of individuals lines. Note that the contribution functions use the wavelength axes on the top. In the EddySed model, the cloud base is set at the intersection of the P-T profile and a given cloud condensation curve (dashed lines). For MgSiO${}_{3}$, this corresponds to a pressure of $\approx 10-20$ bars when using our HRS-retrieved P-T profile. As the cloud mass fraction drops exponentially above the cloud base in the EddySed model (controlled by $K_{\rm zz}$ and $f_{\rm sed}$), we find that the cloud opacity decreases to negligible levels by the time we reach pressures of a few bars where the continuum forms. For this reason, we do not consider models with Fe clouds in our HRS retrievals, since the Fe cloud base forms even deeper than that of MgSiO${}_{3}$.
Therefore, our KPIC HRS are insensitive to clouds because we cover both a relatively small wavelength range (2.29-2.49 $\mu$m) and a range where molecular opacities from H${}_{2}$O, CO, and CH${}_{4}$ are significant. The small wavelength range means that the cloud opacity is effectively constant in wavelength. The strong molecular opacity in HRS allows us to resolve many individual absorption lines and obtain good constraints on the atmospheric composition for molecules present in this region of the spectrum. The opacity of these molecules decrease at shorter wavelength due to decreasing excitation cross sections, so the continuum shifts to higher pressures (deeper down) at shorter wavelengths. This effect is visible in the LRS contribution function, where close to $1~{}\micron$, the emission originates from roughly the same pressure as the MgSiO${}_{3}$ cloud base, making the $y$ and $J$ bands particularly sensitive to clouds (see Fig. 10).
Could the KPIC HRS be affected by clouds at lower pressures (higher altitudes) than predicted by the EddySed model? Several studies have found that including clouds at lower pressures than predicted by EddySed produces better fits to mid-IR spectra of isolated brown dwarfs (e.g. Burningham et al., 2021; Luna & Morley, 2021). As shown in Fig. 6, our HRS P-T profiles show a nearly isothermal region between about 0.1-1 bars, which could suggest a degeneracy with clouds (Burningham et al., 2017). To check whether the P-T parameterization affects our results, we run a retrieval with a fixed P-T, namely the self-consistent profile over-plotted in gray. We find that all posteriors from this fixed P-T fit overlap within $1\sigma$ with those from our baseline retrieval. Thus, we conclude that the isothermal part of the P-T we retrieve is not biasing our conclusions. To further examine the possibility of clouds at lower pressures, we also run an opaque cloud model with infinite opacity below a retrieved pressure, and a gray cloud model that adds a constant cloud opacity at each pressure layer. When fitting the HRS with these more flexible cloud models, we also find consistent results with the baseline model. In the second model, the gray opacity is bounded to lie below $\sim 0.03~{}g/\rm{cm}^{-3}$, and the pressure of the infinitely opaque cloud is required to be deeper than $\sim 1$ bar. Therefore, even with these less constraining cloud parameterizations, we find that our HRS still prefers solutions with minimal cloud opacity.
5.3 Disequilibrium chemistry with deep quenching pressure
In our retrievals, we include a simple model for disequilibrium chemistry using the quench pressure prescription in petitRADTRANS, which is motivated by Zahnle & Marley (2014). Specifically, the abundances of CH${}_{4}$, CO, and H${}_{2}$O are held constant at atmospheric pressures lower than the retrieved $P_{\rm quench}$ parameter. We find that when including quenching, the goodness of fit increases drastically compared to fits with full chemical equilibrium. For example, between two clear retrievals with and without quenching, we find that $B\approx 380$ in favor of the quenched retrieval. From the Bayes factor interpretation of Benneke & Seager (2013), this represents a detection of quenching at $\approx 3.9\sigma$ significance. The quench pressure retrieved is also highly consistent between retrievals with different cloud models, with 1 and 2$\sigma$ intervals of $50-260$ and $14-836$ bars (Fig. 4). In this section, we explore reasons why the data prefer disequilibrium chemistry in the atmosphere of HD 4747 B. The physical implications of our retrieved quench pressure, including an estimate of the vertical diffusion coefficient ($K_{\rm zz}$), are discussed in § 8.2.
To understand why the data prefer a deep quench pressure, we plot lines of constant log(CO/CH${}_{4}$) volume mixing ratios (VMR) along with the best-fit P-T profile from our baseline HRS retrieval in Fig. 7. We calculate CO/CH${}_{4}$ from this quenched chemistry retrieval by finding the abundances of each molecule in the chemical grid, iterating over our posterior distribution of C/O, [C/H], and P-T profile. We find that CO/CH${}_{4}$=$13.6^{+5.8}_{-4.6}$. If the atmosphere was in chemical equilibrium, we repeat our calculation and find that we would expect CO/CH${}_{4}$=$1.35^{+0.21}_{-0.17}$, which is ten times smaller than our retrieved value in the quenched chemistry model. Thus, the relative under-abundance of CH${}_{4}$ relative to CO in our HRS leads our models to prefer a deep quench pressure. The value of CO/CH${}_{4}$ also determines our retrieved the quench pressure, whose $2\sigma$ interval is indicated by the thick blue region in Fig. 7. Because the P-T profile nearly overlaps the curve of CO/CH${}_{4}$ = 10 at $\sim 20$ bars and deeper, a broad range of quench pressures deeper than $\sim 20$ bars are consistent with the data.
As another way to visualize the detection of disequilibrium chemistry, we plot the molecular abundances in VMR as a function of pressure in Fig. 8. The solid lines show the VMRs for the HRS quenched chemistry retrieval, while dashed lines show the VMRs for the same model with quenching turned off manually. By comparing the solid and dashed lines, we see that at the pressures probed by our observations, the relative abundances of CO, CH${}_{4}$, and H${}_{2}$O differ by several orders of magnitude between the quenched model and expectations from chemical equilibrium.
5.4 Detection of methane in the HRS
In this section, we take a closer look at the relatively weak methane absorption signal in our HRS, which leads us to prefer quenched models where the CO/CH4 ratio is a factor of ten higher than predicted in models assuming chemical equilibrium. We confirm the presence of detectable levels of methane in the HRS by running a pair of free retrievals, one with only H${}_{2}$O and CO, and one with H${}_{2}$O, CO, and CH${}_{4}$. The results of these retrievals are listed in Table 3. In these free retrievals, we fit the abundances of each absorbing species independently and assume a constant abundance as a function of pressure. Although we also considered models that included NH${}_{3}$ and CO${}_{2}$, we only obtained upper limits on their abundances, and therefore excluded them from our fits in this section. Finally, given the insensitivity of the HRS to clouds, we carry out these tests with the clear model to save computation time.
We find that the data strongly prefer the model with CH${}_{4}$, with a Bayes factor of 84 ($3.4\sigma$ significance; Benneke & Seager 2013). As shown in Table 3, we obtain log(CH${}_{4}$)$=-4.82\pm 0.23$ from the free retrieval, and the CH${}_{4}$ posterior in Fig. 9 shows no strong covariance with the abundances of either CO or H${}_{2}$O. If the atmosphere was in chemical equilibrium, we would expect a CH${}_{4}$ VMR that is ten times higher than what we retrieve, according to the same calculation described in § 5.3. We note that the abundances from the free retrieval with CH${}_{4}$ also agree well with the corresponding VMRs from our quenched chemistry retrievals. This is not surprising given the deep quench pressure we retrieve, which makes the molecules abundances constant in the regions where our HRS is sensitive (see Fig. 8).
We separately visualize the CH${}_{4}$ detection in cross-correlation space by carrying out an analysis similar to that described in Zhang et al. (2021). First, we make a ‘pure CH${}_{4}$ template’ from the best-fit companion model with CH${}_{4}$, H${}_{2}$O, and CO by manually setting the abundances of H${}_{2}$O and CO to zero. If the model without CH${}_{4}$ is fitting poorly due to its inability to fit CH${}_{4}$ lines in the data, we would expect the residuals of this model, which we denote $R$ = (data - model without CH${}_{4}$), to contain CH${}_{4}$ lines. Therefore, we cross-correlate $R$ with the pure CH${}_{4}$ template, plotted as the blue CCF in Fig. 9. In addition, we plot the CCF of $R^{\prime}=$ (data - model with CH${}_{4}$) with the pure CH${}_{4}$ template in red for comparison. The blue CCF shows a peak at 0 km/s (solid gray line), where we expect a real signal to be since the models have been shifted by the best fit companion RV. If the residuals were dominated by telluric CH${}_{4}$ for example, the CCF peak would appear at the dotted gray line (negative of the RV, or -15 km/s). Thus, even though the height of the CH${}_{4}$ peak in the blue CCF is small compared to the surrounding structure, the fact that it is located at the companion RV is evidence of a real signal from CH${}_{4}$.
In our CCF framework, the y-axis is the estimated flux level (in counts) of the companion signal from a least-squares minimization. As shown in Fig. 9, we find a flux level of $\approx 50$ counts for CH${}_{4}$, which is an estimate of the companion flux in the residuals. Importantly, this value is consistent with the flux value found when we repeat the same CCF analysis with H${}_{2}$O (i.e. comparing a model with only CO and CH${}_{4}$ and the baseline model of CO, H${}_{2}$O, and CH${}_{4}$). For a molecule such as NH${}_{3}$, which we see no evidence of in the KPIC HRS, the flux value from the CCF becomes unbounded as the least-squares routine used for computing the CCF fails to converge.
Finally, we check for cross-talk between H${}_{2}$O and CH${}_{4}$ by cross-correlating $R$ with the pure water template and detect no CCF peaks. Furthermore, we note that in a retrieval with only CO and CH${}_{4}$ (no H${}_{2}$O), the retrieved CH${}_{4}$ abundance is consistent with the value from the full model including CO, H${}_{2}$O and CH${}_{4}$.
We therefore conclude that the data strongly favor the presence of detectable levels of methane in the HRS, with an abundance significantly lower than that predicted by equilibrium chemistry models. The detection of methane at log(CH${}_{4}$) = ${-4.82\pm 0.23}$ demonstrates the ability of KPIC to retrieve species that are more than an order of magnitude lower in VMR than the dominant molecular constituents in the data, in only 1 hour of integration time.
6 Low-resolution retrievals (GPI + SPHERE)
6.1 Overview
In this section, we present the results from our fits to the LRS and compare our retrieved parameters to those from the HRS fits. We fit the LRS using the same models as before. These include one clear model and four different implementations of the EddySed cloud model where we vary our assumptions about the unknown cloud properties. The cloudy models consist of two MgSiO${}_{3}$ retrievals with am and cd particles (explained in § 4.2.2), and two retrievals with MgSiO${}_{3}$ and Fe clouds (again, am and cd). In Fig. 10, we plot the data, best-fit cloudy and clear models, the residuals, and the GP models of the residuals. The posteriors for a few key parameters from these retrievals are plotted in Fig. 11 and tabulated in Table 2. See Appendix C for the posterior distributions of other parameters in the baseline model.
When comparing the clear and cloudy models in Fig. 10, we see that the data shortward of $\sim 1.2~{}\mu$m is poorly fit by models without clouds. This causes the clear model to have $B\approx 7\times 10^{-26}$; it is overwhelmingly ruled out compared to the baseline EddySed model. In addition, when we plot the models over a larger wavelength range in Fig. 12, we find that the cloudy models agree with the NIRC2 L band photometry from Crepp et al. (2016), while the clear model over-predicts the $L$ band flux by $\approx 2\sigma$. We did not include these photometric points in our retrievals.
Fig. 10 shows that the SPHERE J band data from $\approx 1.2-1.35~{}\mu$m is not well fit by even the cloudy model, which could either be caused by model mismatch or speckle contamination that artificially raises the flux. The GP model finds that $\sim 60\%$ of the SPHERE error bars and $\sim 90\%$ of the GPI error bars are from correlated noise, with correlation length scales of $\sim 6$ and $\sim 2$ wavelength channels, respectively. This confirms our initial intuition that the noise in the SPHERE and GPI images is likely dominated by correlated speckle noise based on visual inspection of the images. For the SPHERE data set, we estimate that the retrieved length scale is roughly equal to the number of steps that a speckle would move across the PSF for our brown dwarf’s separation; indeed, we see speckles moving across the companion PSF in the reduced images. Overall, the SPHERE spectrum is less reliable than that from GPI because only 4 exposures are available, compared to the $\sim 40$ exposures from GPI.
Finally, the P-T profile retrieved from our baseline LRS retrieval shows a bi-modal distribution (see Fig. 6). The degeneracy seen here may be related to issues with the LRS (see § 6.3).
6.2 Comparison with prior knowledge
Because the LRS is flux-calibrated, we can check whether our retrieved radii and effective temperatures are physical and consistent with prior knowledge for this benchmark companion. Using the known age and mass of HD 4747 B ($3\pm 2$ Gyr and $m=67.2\pm 1.8$ $M_{\rm{Jup}}$), we interpolate the COND evolutionary models (Baraffe et al., 2003) to find a model-predicted radius of $0.8^{+0.07}_{-0.03}R_{\rm{Jup}}$, and a predicted $T_{\rm eff}=1450^{+350}_{-180}$ K. Peretti et al. (2019) compared the SPHERE spectrum of HD 4747B to those of field brown dwarfs to derive a more tightly constrained $T_{\rm eff}=1350\pm 50$ K (see Table 2), which we adopt in the subsequent discussion.
We calculate the effective temperatures of the models in our retrievals by integrating the flux over 0.5-30 $\mu$m. For the baseline EddySed model, we find $T_{\rm eff}=1473^{+17}_{-20}$ K, and a radius of $0.70^{+0.05}_{-0.03}~{}R_{\rm{Jup}}$. Compared to prior expectations however, the radius retrieved is too small by $\approx 1.5\sigma$ while $T_{\rm eff}$ is too high by $\approx 2\sigma$. From substellar evolutionary models, the minimum possible radius of a brown dwarf should be $\approx 0.74R_{\rm{Jup}}$, which is imposed by electron degeneracy pressure (Chabrier et al., 2009). We find that $T_{\rm eff}$ and radius are correlated in the LRS retrievals, as shown in Fig. 11, which is expected as different combinations of these two parameters can produce the same total luminosity. However, our total luminosity agrees well with the luminosity predicted by evolutionary models.
Several previous retrieval studies have also found smaller than expected radius for L dwarfs, which may be attributed to the presence of heterogeneous surface features, such as patchy clouds, that are not captured in current 1-D retrieval frameworks (e.g. Kitzmann et al., 2020; Gonzales et al., 2020; Burningham et al., 2021). On the other hand, Gonzales et al. (2021) retrieved a radius consistent with evolutionary models for a seemingly cloudless L dwarf. Whether the radii from evolutionary models are correct is an assumption that is now being tested by a growing sample of transiting brown dwarfs from TESS (e.g. Carmichael et al., 2020).
In our retrievals with both MgSiO${}_{3}$ and Fe clouds, we retrieve slightly larger radii that are more consistent with the evolutionary model prediction. This could indicate that a single cloud model (MgSiO${}_{3}$) may be inadequate in attenuating the flux from the deep atmosphere. However, models with two cloud species do not improve the fit significantly ($B$ = 1.5-3 compared to the baseline model with MgSiO${}_{3}$ only). Furthermore, the MgSiO3, cd model actually has the largest retrieved radius, but our data cannot distinguish between crystalline and amorphous particles. We conclude that our retrieved radius is sensitive to aspects of the cloud models that are poorly constrained by the existing data for this object.
6.3 LRS at longer wavelengths could improve abundance and cloud constraints
While the LRS can provide tighter constraints on the cloud parameters and radius compared to the HRS, we find that many retrieved parameters, including the atmospheric abundances, are very sensitive to model choices. In Fig. 11, we overplot the posteriors distributions of a few parameters from our four EddySed models. The retrieved C/O and [C/H] have large uncertainties and can disagree at the $3\sigma$ level between models. The values also span a significant portion of the parameter space ($>$1 dex in metallicity), and show much stronger covariance compared those measured from the HRS (see Fig. 4). However, all cloudy models fit the LRS well, with Bayes factors within a factor of $\sim 3$ (see Table 2), so we cannot distinguish between them.
We note that Mollière et al. (2020) were able to obtain much better constraints on the composition of HR 8799e, which also has a cloudy atmosphere, using LRS data sets from 0.95-2.5 $\mu$m. Their LRS had SNR between 4-11 per wavelength bin, much lower than the SNR of our LRS (between 20-60 per wavelength bin). Unlike Mollière et al. (2020), however, our study does not have LRS in the second half of $K$ band ($2.2-2.5~{}\mu$m), which contains a strong CO bandhead as well as significant H${}_{2}$O and CH${}_{4}$ opacities. When we compute the CO abundances from our LRS retrievals, we find that they are not well constrained, with 1$\sigma$ intervals that are $\gtrsim 3$ wider than the CO constraint from HRS. In Fig. 12, we plot random draws of our baseline model color-coded by metallicity out to 5 $\mu$m. As shown, the models diverge quickly in the $2.2-2.5~{}\mu$m range. The fact that we miss this crucial wavelength region could explain why Mollière et al. (2020) obtain more robust constraints on atmospheric abundances, and P-T profiles that agree better with self-consistent models than we do, despite using data with a lower SNR.
Fig. 12 also shows a clear gradient in metallicity beyond 2.5 $\mu$m. In some of our cloudy LRS retrievals, we see a covariance between metallicity and cloud mass fraction, where lower metallicities correspond to higher cloud mass fractions, as well as larger, more physically consistent radii (see Fig. 11). The degeneracy between metallicity and cloud mass fraction might arise because both molecular opacities and clouds contribute opacity, and our data has insufficient wavelength coverage to probe more regions where the gas and cloud opacities are sufficiently different. From the LRS retrievals, we consistently find a factor of $\sim 2-3$ more CH${}_{4}$ and H${}_{2}$O than observed in the HRS, implying that the LRS retrievals could be compensating for our imperfect cloud models by increasing the gas opacities.
Using a more flexible cloud model might alleviate some of these issues. For example, Burningham et al. (2021) retrieved the 1-15 $\mu$m LRS of a field L dwarf and found the data preferred silicate clouds much higher up than the predicted cloud base locations from condensation curves. In addition, their retrieved cloud particles also have smaller sizes (sub-micron) than predicted by the EddySed model (a few microns). Similarly, Luna & Morley (2021) found that sub-micron cloud particles at lower pressures than predicted by EddySed are required to fit the mid-IR silicate feature ($\approx 8-10~{}\mu$m) of many L dwarfs. They found that the microphysical cloud model CARMA (Turco et al., 1979; Toon et al., 1988; Gao et al., 2018) allows them to fit their data much better and even place constraints on which cloud species are producing the observed absorption feature.
Both the above-mentioned studies benefited from data at $\sim 10~{}\mu$m that significantly help with constraining cloud properties. Thus, to obtain better abundance measurements with LRS, it is not only important to obtain full coverage in the near-IR (which we lack), but also to acquire data in the mid-IR. JWST can obtain low- and medium-resolution spectroscopy of brown dwarfs spanning the near- to mid-IR wavelengths using the NIRSpec and MIRI instruments. Future ground-based instruments such as SCALES at Keck (Stelter et al., 2020) will also provide LRS in the mid-IR.
7 Joint retrievals
In this section, we describe the results of joint retrievals to both the HRS and LRS for HD 4747 B. In practice, we set up two radiative transfer routines with petitRADTRANS using line-by-line (for HRS) and correlated-k (for LRS) opacity sampling respectively. The HRS and LRS models share the same atmospheric parameters and priors, but each has some unique parameters (e.g. RV and $v\sin{i}$ for HRS, GP kernel parameters for LRS). Within one nested sampling retrieval, we add the log likelihoods from the HRS and LRS components at each step of sampling to get the total log likelihood. We consider both clear and cloudy EddySed models for our joint retrievals.
Because the LRS prefer clouds, the cloudy model (MgSiO${}_{3}$, am) is overwhelmingly preferred in our joint retrieval, with a Bayes factor in excess of $10^{34}$ compared to the clear model. From the cloudy model, we retrieve C/O$=0.70\pm 0.03$ and [C/H]$=0.34\pm 0.07$. The retrieved uncertainties on these parameters are lower than in the HRS-only retrieval (which had C/O$=0.66\pm 0.04$ and [C/H]$=-0.10^{+0.18}_{-0.15}$). In addition, the C/O from our joint fit is consistent with the C/O from our HRS fit. This is not surprising, because the HRS places tight constraints on the relative line depths (and hence the relative abundance ratios) of CH${}_{4}$, H${}_{2}$O, and CO. However, the joint fit pushes the metallicity to higher values, which corresponds to increased gas abundances as shown in Fig. 13. The joint fit results translate to a $>4\sigma$ discrepancy in [C/H] between HD 4747 A&B, while there is no discrepancy if we take the results from the HRS fit. This implies that the joint fit might be compensating for inadequacies in modeling clouds by increasing the gas opacities, as discussed in § 6.3 for the LRS-only case. We ran additional joint retrievals where we varied the cloud parameters (e.g. adding Fe clouds) and found similar results.
If we compare the log likelihoods of the HRS part of the joint fit to that from the HRS-only fit, we find that the HRS is fit less well by $\approx e^{10}$ (which translates to $\approx 4\sigma$) in the joint fit, implying a trade-off between fitting the LRS and HRS. We can qualitatively compare the LRS S/N per wavelength bin to the CCF S/N of the HRS, which approximates the total constraining power of the HRS. When including all molecules in our model, we find a CCF S/N of $\approx 15$ for the HRS. For the LRS, the average S/N per wavelength bin is $\approx 20$ for the SPHERE $YJH$ data and $\approx 60/30$ between the GPI $H/K_{1}$ data. This explains why the joint fits prioritize fitting the LRS at the expense of the HRS.
As discussed in § 6, the LRS are very model-sensitive and additionally contaminated by correlated noise. For this reason, we adopt the HRS-only results as the best estimate of HD 4747 B’s atmospheric properties in this paper (see first row of Table 2). We leave it to future work, preferably aided by longer wavelength coverage in LRS, to achieve a more satisfactory joint retrieval.
8 Discussion
8.1 Next steps for high-resolution spectroscopy
Our KPIC HRS provide a better handle on the atmospheric abundances of HD 4747 B, and are less sensitive to model choices than our LRS. In fact, our $K$ band HRS are essentially agnostic to clouds in the brown dwarf’s atmosphere; all retrieved parameters are consistent independent of our chosen cloud model (Fig. 4). As discussed in § 5.2, this is because our HRS covers a wavelength region ($2.29-2.49~{}\mu$m) of high molecular opacity, and probes emission across a wide range of atmospheric pressures where cloud opacity is negligible (up to $10^{-2}$ bars in line cores). While clouds affect the continuum near $1~{}\mu$m in the LRS, they have little effect on the line depths across the wavelength range of our HRS. The relative lines depths are sensitive to relative molecular abundances, which directly constrains C/O. These results advocate for using HRS to measure atmospheric abundances.
In the future, it is important to explore whether these findings hold true for other substellar objects. In upcoming papers, we will present KPIC HRS retrievals of brown dwarf companions and giant planets spanning a range of effective temperatures and surface gravities. Ultimately, it would also be useful to constrain cloud properties with HRS. For transmission spectroscopy, Gandhi et al. (2020) found that their simulated near-IR HRS for warm Neptunes are more sensitive to molecular abundances than LRS for the same reasons highlighted in this study. While both clouds and metallicity affect the line depths in HRS, Gandhi et al. (2020) show that increasing the wavelength coverage (e.g. going from 0.9-1.7 $\mu$m to 0.9-2.5 $\mu$m) helps distinguish between clouds and metallicity and provide better constraints on both. Thus, if we wish to obtain constraints on clouds and abundances at the same time, it would be important to extend our current HRS to a broader range of wavelengths. KPIC Phase II will allow us to obtain $L$ band data ($\approx 3.4-4.1~{}\mu$m) to complement existing $K$ band data (Delorme et al., 2021), and future upgrades could benefit from including $H$ and $J$ bands as well.
In this study, we have assumed that the atmosphere of HD 4747 B is globally uniform. However, it would be important to examine the impact of 3-D effects, including non-uniform cloud coverage. Past studies with photometry or LRS show that many brown dwarfs exhibit clear rotational variability signals (e.g. Apai et al., 2013; Zhou et al., 2018; Biller et al., 2018; Manjavacas et al., 2021; Vos et al., 2022), which appear to be caused by inhomogeneities in their cloud properties. Therefore, time-resolved observations are important to understanding clouds and 3-D effects.
With HRS, we can use the time-varying line depth and shape to map the 2-D brightness distributions of these objects (e.g. Crossfield et al., 2014). In this paper, we used 1 hour of KPIC data for HD 4747 B. Given our measured $v\sin{i}$ and assuming a radius of $0.8R_{\rm{Jup}}$, we would expect a 5 or 7 hour rotation period if $i$ is equal to the orbital inclination or $i=90\degree$. Thus, it may be possible to sample a full rotation period within a single observing night, with the caveat that measurements of the true rotation period remain difficult for high-contrast companions (Biller et al., 2021).
8.2 Methane and the presence of disequilibrium chemistry
Our HRS retrievals indicate that the ratio of CO/CH4 (VMR) is $\approx 10$ times higher than expected by equilibrium chemistry (see § 5.3). To gain more physical intuition, we convert the quench pressure from our HRS retrievals to an estimate of the vertical diffusion coefficient, $K_{\rm zz}$. To do this, we match the chemical timescale of the CO-CH${}_{4}$ reaction from Zahnle & Marley (2014) with the mixing timescale $\tau_{\rm mix}=L^{2}/K_{\rm zz}$. While the length scale $L$ is typically taken to be the pressure scale height $H$ for lack of a better estimate, Smith (1998) show that this assumption is not valid across several reactions in the atmospheres of Jupiter and Neptune. In fact, Smith (1998) find that $L\approx 0.1H$, which changes the inferred $K_{\rm zz}$ by two orders of magnitude. Similarly, Ackerman & Marley (2001) also note that the mixing length is generally shorter than the pressure scale height $H$ in stable atmospheric regions. Due to the uncertainty in $L$, we adopt $L=\alpha H$, where $\alpha$ is a scaling factor, and $H=\frac{k_{B}T}{\mu mg}$ ($\mu$: mean molecular weight, $g$: surface gravity, $T$: the local temperature). For each value of quench pressure from our posteriors, we compute the necessary quantities to derive a posterior for $K_{\rm zz}$. For instance, if $\alpha=0.1$, we find $K_{\rm zz}=5\times 10^{8}-1\times 10^{12}~{}\rm{cm^{2}s^{-1}}$ ($1\sigma$). On the other hand, if $\alpha=1$, we obtain $K_{\rm zz}=5\times 10^{10}-1\times 10^{14}~{}\rm{cm^{2}s^{-1}}$.
There have been few quantitative measurements of $K_{\rm zz}$ for substellar companions. Miles et al. (2020) used M-band LRS to constrain the CO abundance and estimate the vertical diffusion coefficient, $K_{\rm zz}$, for seven field brown dwarfs. However, their objects have $T_{\rm eff}$ between 250-750 K, much colder than HD 4747 B. In terms of objects with $T_{\rm eff}\gtrsim 1000~{}$K, Barman et al. (2015) reported a detection of CH${}_{4}$ in HR 8799b ($T_{\rm eff}\sim 1000~{}$K) with Keck/OSIRIS data, which they used to estimate $K_{\rm zz}$ between $10^{6}-10^{8}~{}\rm{cm^{2}s^{-1}}$. However, this CH${}_{4}$ detection was not confirmed by an independent study (Petit dit de la Roche et al., 2018), and recently Ruffio et al. (2021) concluded that future higher-resolution follow up is needed to resolve the discrepant CH${}_{4}$ signal strengths found by different analyzes. Ruffio et al. (2021) point out that if the CH${}_{4}$ abundance was over-estimated by Barman et al. (2015), that would imply a larger $K_{\rm zz}$. Using LRS, Mollière et al. (2020) found a well-constrained quench pressure for HR 8799e ($T_{\rm eff}\sim 1100~{}$K) from petitRADTRANS retrievals, which could similarly be converted to a $K_{\rm zz}$ constraint. In summary, our finding HD 4747 B, which is $\sim 300-400~{}$K hotter than HR 8799b&e and much older (a few Gyr from § 2.1) than most directly imaged planets, represents an important new data point because hotter objects are expected to be closer to equilibrium, making chemical disequilibrium processes harder to detect (e.g. Moses et al., 2013).
Zahnle & Marley (2014) provide an upper limit for $K_{\rm zz}$ from mixing length theory (Gierasch & Conrath, 1985) assuming full convection. For HD 4747 B, their Equation 4 translates to an upper limit of $\approx 10^{9}~{}\rm{cm^{2}s^{-1}}$. Depending on $\alpha$, our retrieved $K_{\rm zz}$ either exceeds this upper limit by $\gtrsim 2\sigma$ (if $L=H$), or is very close to this limit (if $L=0.1H$). Together, this suggests that convection is driving the vertical mixing in HD 4747 B, and that the mixing efficiency is likely close to its predicted maximum. We check whether our inferred $K_{\rm zz}$ makes sense by comparing them to those predicted by self-consistent atmospheric models with disequilibrium chemistry from Karalidi et al. (2021) and Mukherjee et al. (2022, in prep). For an object with properties similar to HD 4747 B, our measured CH${}_{4}$ VMR is consistent with $K_{\rm zz}\sim 10^{8}-10^{12}$ in these models (with the assumption that $L=H$). These values of $K_{\rm zz}$ are roughly consistent with our estimate based on $P_{\rm quench}$, and also near the upper limit from Zahnle & Marley (2014). On the modeling front, it would be valuable to carry out 3-D hydrodynamical simulations (e.g. Zhang & Showman, 2018; Tan & Showman, 2021) of brown dwarf interiors to independently estimate $K_{\rm zz}$ (Tan, 2022) and compare the results to that inferred by our data. Such simulations could also reveal what physical processes might cause a discrepancy between mixing length theory and our observations.
8.3 Dynamical versus Spectroscopic Mass Constraints
For a majority of substellar companions observed by direct imaging, there are no dynamical mass constraints. To assess whether our mass prior plays an important role in the results, we repeat our HRS and LRS retrievals with the baseline cloud model but use uniform priors in mass from 10 to 100 $M_{\rm{Jup}}$ (‘free-mass’). For the HRS free-mass retrieval, we find that all parameters change by less than $1\sigma$ compared to the mass-prior retrieval. The mass itself shows a broad distribution (33-76 $M_{\rm{Jup}}$ at $1\sigma$) that encompasses the dynamical mass. Because our KPIC HRS are not flux calibrated, the radius is not well constrained. In this case, we get large uncertainties in the spectroscopic mass because mass is inferred from the retrieved surface gravity, which depends on the poorly-constrained radius.
Our LRS free-mass retrieval also yields posteriors for all parameters that are consistent between 1-2 $\sigma$ with the mass-prior retrieval. Furthermore, the mass retrieved by the LRS is $59^{+7}_{-8}~{}M_{\rm{Jup}}$, which agrees within about $1\sigma$ with the dynamical mass. This provides confidence that reasonable mass constraints can be placed on substellar objects from LRS. The radius retrieved is $0.77\pm 0.03~{}R_{\rm{Jup}}$, consistent with evolutionary model predictions and close to the radius from the mass-prior retrieval, suggesting the two retrievals find a similar surface gravity.
8.4 Atmospheric abundances of HD 4747 AB
We retrieve [C/H] and [O/H] values that are $1\sigma$ consistent with those of the host star, as discussed in § 5.1. Both companion and the star are mildly sub-solar in terms of their metal content. However, our retrieved C/O=$0.66\pm 0.04$ is higher by approximately $2\sigma$ than the stellar C/O=$0.48\pm 0.08$.
The question is whether the marginal discrepancy in C/O is from astrophysical or systematic reasons.
For example, Wang et al. (2022) carried out retrieval experiments on simulated HRS ($2.2-2.35~{}\mu$m, $R=35,000$) and found that their formal error bars are likely under-estimated due to systematic errors at the $\sim 0.15$ level in C/O. Using KPIC HRS from 2.23-2.33 $\mu$m, they found $\approx 1-1.5\sigma$ discrepancies between the [C/H] and [O/H] abundances of HR 7672 A and B, another benchmark brown dwarf system. On an earlier study of benchmark brown dwarfs, Line et al. (2015) quoted $1\sigma$ uncertainties of $0.2-0.3$ in their brown dwarf C/O (much larger than our formal C/O uncertainty of $0.04$), and concluded that a $2\sigma$ agreement between the stellar and companion C/O is sufficiently good given the caveats. It is also possible that the uncertainties on stellar abundances are under-estimated given non-LTE effects (Line et al., 2015).
Another factor that might contribute to the $2\sigma$ discrepancy in C/O is uncertainties in the chemistry of condensates. The chemical model of petitRADTRANS we use accounts for the equilibrium condensation of various species and reports the global (rather than gas phase) C and O abundances (Mollière et al., 2019). In particular, species such as MgSiO${}_{3}$ and Mg${}_{2}$SiO${}_{4}$ contain 3 or 4 oxygen atoms per molecule, and are expected to hold a significant portion of O (Line et al., 2015). From our HRS retrievals, we find that $\approx 18\%$ of O is condensed into solids such as MgSiO${}_{3}$. In order to decrease the global C/O of the brown dwarf by $\approx 0.1$ (therefore making the companion and stellar C/O agree at the $1\sigma$ level), we require a $\sim 20\%$ increase in the net O abundance. Keeping everything else unchanged, this means the MgSiO${}_{3}$ mass fraction, which is predicted by the chemical model to be $\sim 2\times 10^{-3}$ in our retrievals, needs to be doubled to $\sim 4\times 10^{-3}$. From the LRS retrievals, the cloud base MgSiO3 fraction can be as high as $10^{-2}$. Therefore, a factor of $\sim 2$ uncertainty in the abundance of MgSiO${}_{3}$ could make our C/O consistent at the $1\sigma$ level with the stellar value.
Given these caveats, we conclude that the $2\sigma$ difference between our retrieved C/O for HD 4747 B and the stellar value is not significant, and HD 4747 AB are consistent with being chemically homogeneous. Chemical homogeneity is expected by models where brown dwarf companions form via gravitational fragmentation in molecular clouds (e.g. Padoan & Nordlund, 2004) or massive protostellar disks (e.g. Stamatellos et al., 2007). Simulations suggest that brown dwarfs typically form as part of unstable, high-order multiple systems, which undergo chaotic interactions that reduce the multiplicity over time (e.g. Bate et al., 2002; Thies et al., 2010; Bate, 2012). With a semi-major axis of 10 au, HD 4747 B is unlikely to have been directly affected by such encounters, but its relatively high orbital eccentricity ($\approx 0.73$) could encode such a dynamically ‘hot’ past.
9 Conclusions
Using high-resolution spectra ($R\sim 35,000$) obtained by Keck/KPIC, we retrieve [C/H]=$-0.10^{+0.18}_{-0.15}$, [O/H]=$-0.18^{+0.18}_{-0.15}$, and C/O=$0.66\pm 0.04$ for the benchmark brown dwarf companion HD 4747 B (formal error bars). The C and O abundances are consistent with the stellar values to $\lesssim 1\sigma$, while the C/O ratio is consistent at the $2\sigma$ level, as expected for a binary-star like formation scenario. This shows that we can measure the atmospheric abundances for high contrast substellar companions to the $20\%$ level with KPIC and our current modeling framework, which Wang et al. (2022) also show for another benchmark brown dwarf. We outline some other key findings from our study below.
We measure precise abundances from the KPIC HRS ($2.29-2.49~{}\mu$m), which are insensitive to our choice of cloud model. Our abundance measurements suggest that HD 4747 B has a CO/CH${}_{4}$ ratio that is 10 times higher than predicted by equilibrium chemistry, corresponding to a quench pressure of $50-260$ bars ($1\sigma$). This translates to a high vertical diffusion coefficient $K_{\rm zz}$ which depends on the assumed length scale $L$. However, even if $L$ is ten times smaller than the pressure scale height, we get $K_{\rm zz}=5\times 10^{8}-1\times 10^{12}~{}\rm{cm^{2}s^{-1}}$, which implies a mixing strength that is at or above the upper limit predicted by mixing length theory.
The composition retrieved from our LRS (1-2.2 $\mu$m) are both sensitive to model choices, and can be biased by the presence of speckles. For this reason, HRS provides a more reliable picture of the atmospheric composition in the current data sets, although the LRS could be improved with additional observation at longer wavelengths including the $L$ and $M$ bands. Despite these challenges, the current LRS does provide a spectroscopic mass estimate that is $1\sigma$ consistent with the dynamical mass for the brown dwarf.
Although our joint retrieval results are likely biased by the limited LRS wavelength coverage, joint analyzes of LRS and HRS remain a promising avenue to constrain cloud properties and abundances simultaneously and provide a more complete picture of substellar atmospheres. When extended wavelength coverage is available, it would also be important to consider possible 3-D effects, including patchy clouds. These might be constrained by obtaining multiple spectra sampling a rotation period. Additional modeling work on condensation, chemistry, and vertical mixing rates are also important to inform future observational results.
We thank the referee for helpful comments that improved the manuscript. J.X. thanks Michael Zhang for advice in computing the CH${}_{4}$ opacities, Jonathan Fortney for discussions on chemical disequilibrium, and Konstantin Batygin for discussions on brown dwarf formation.
We wish to recognize and acknowledge the very significant cultural role and reverence that the summit of Maunakea has always had within the indigenous Hawaiian community. We are most fortunate to have the opportunity to conduct observations from this mountain. This research has benefitted from the SpeX Prism Library maintained by Adam Burgasser at http://www.browndwarfs.org/spexprism. Funding for KPIC has been provided by the California Institute of Technology, the Jet Propulsion Laboratory, the Heising-Simons Foundation, the Simons Foundation, and the United States National Science Foundation Grant No. AST-1611623. AV acknowledges funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme, grant agreements No. 757561 (HiRISE). HAK acknowledges support from the President’s and Director’s Research & Development Fund Program, which is jointly funded by the Jet Propulsion Laboratory and the California Institute of Technology under a contract with the National Aeronautics and Space Administration. The computations presented here were conducted in the Resnick High Performance Center, a facility supported by Resnick Sustainability Institute at the California Institute of Technology.
References
Ackerman & Marley (2001)
Ackerman, A. S., & Marley, M. S. 2001, The Astrophysical Journal, 556, 872,
doi: 10.1086/321540
Amarsi et al. (2019)
Amarsi, A. M., Nissen, P. E., & Skúladóttir, Á. 2019, Astronomy
and Astrophysics, 630, A104, doi: 10.1051/0004-6361/201936265
Apai et al. (2013)
Apai, D., Radigan, J., Buenzli, E., et al. 2013, The Astrophysical Journal,
768, 121, doi: 10.1088/0004-637X/768/2/121
Asplund et al. (2009)
Asplund, M., Grevesse, N., Sauval, A. J., & Scott, P. 2009, Annual Review of
Astronomy & Astrophysics, vol. 47, Issue 1, pp.481-522, 47, 481,
doi: 10.1146/annurev.astro.46.060407.145222
Baraffe et al. (2003)
Baraffe, I., Chabrier, G., Barman, T. S., Allard, F., & Hauschildt, P. H.
2003, A&A, 402, 701, doi: 10.1051/0004-6361:20030252
Barman et al. (2015)
Barman, T. S., Konopacky, Q. M., Macintosh, B., & Marois, C. 2015, ApJ, 804,
61, doi: 10.1088/0004-637X/804/1/61
Bate (2012)
Bate, M. R. 2012, MNRAS, 419, 3115, doi: 10.1111/j.1365-2966.2011.19955.x
Bate et al. (2002)
Bate, M. R., Bonnell, I. A., & Bromm, V. 2002, MNRAS, 332, L65,
doi: 10.1046/j.1365-8711.2002.05539.x
Benneke & Seager (2013)
Benneke, B., & Seager, S. 2013, The Astrophysical Journal, 778, 153,
doi: 10.1088/0004-637X/778/2/153
Beuzit et al. (2019)
Beuzit, J.-L., Vigan, A., Mouillet, D., et al. 2019, Astronomy and
Astrophysics, 631, A155, doi: 10.1051/0004-6361/201935251
Biller et al. (2018)
Biller, B. A., Vos, J., Buenzli, E., et al. 2018, AJ, 155, 95,
doi: 10.3847/1538-3881/aaa5a6
Biller et al. (2021)
Biller, B. A., Apai, D., Bonnefoy, M., et al. 2021, Monthly Notices of the
Royal Astronomical Society, 503, 743, doi: 10.1093/mnras/stab202
Bowler (2016)
Bowler, B. P. 2016, PASP, 128, 102001,
doi: 10.1088/1538-3873/128/968/102001
Brandt et al. (2021a)
Brandt, G. M., Michalik, D., Brandt, T. D., et al. 2021a,
arXiv:2109.06761 [astro-ph].
https://arxiv.org/abs/2109.06761
Brandt (2021)
Brandt, T. D. 2021, arXiv:2105.11662 [astro-ph].
https://arxiv.org/abs/2105.11662
Brandt et al. (2019)
Brandt, T. D., Dupuy, T. J., & Bowler, B. P. 2019, AJ, 158, 140,
doi: 10.3847/1538-3881/ab04a8
Brandt et al. (2021b)
Brandt, T. D., Dupuy, T. J., Li, Y., et al. 2021b,
arXiv:2105.11671 [astro-ph].
https://arxiv.org/abs/2105.11671
Brewer et al. (2016)
Brewer, J. M., Fischer, D. A., Valenti, J. A., & Piskunov, N. 2016, ApJS, 225,
32, doi: 10.3847/0067-0049/225/2/32
Brown et al. (2021)
Brown, A. G. A., Vallenari, A., Prusti, T., et al. 2021, A&A, 649, A1,
doi: 10.1051/0004-6361/202039657
Burgasser (2014)
Burgasser, A. J. 2014, ASI Conf. Ser., 11, 7.
https://ui.adsabs.harvard.edu/abs/2014ASInC..11....7B/abstract
Burningham et al. (2017)
Burningham, B., Marley, M. S., Line, M. R., et al. 2017, MNRAS, 470, 1177,
doi: 10.1093/mnras/stx1246
Burningham et al. (2021)
Burningham, B., Faherty, J. K., Gonzales, E. C., et al. 2021, Monthly Notices
of the Royal Astronomical Society, 506, 1944, doi: 10.1093/mnras/stab1361
Carmichael et al. (2020)
Carmichael, T. W., Quinn, S. N., Mustill, A. J., et al. 2020, The
Astronomical Journal, 160, 53, doi: 10.3847/1538-3881/ab9b84
Chabrier et al. (2009)
Chabrier, G., Baraffe, I., Leconte, J., Gallardo, J., & Barman, T.
2009, in American Institute of Physics Conference Series, Vol. 1094, 15th
Cambridge Workshop on Cool Stars, Stellar Systems, and the Sun, ed.
E. Stempels, 102–111, doi: 10.1063/1.3099078
Claudi et al. (2008)
Claudi, R. U., Turatto, M., Gratton, R. G., et al. 2008, in SPIE
Astronomical Telescopes + Instrumentation, ed. I. S. McLean & M. M.
Casali, Marseille, France, 70143E, doi: 10.1117/12.788366
Crepp et al. (2016)
Crepp, J. R., Gonzales, E. J., Bechter, E. B., et al. 2016, The Astrophysical
Journal, 831, 136, doi: 10.3847/0004-637X/831/2/136
Crepp et al. (2018)
Crepp, J. R., Principe, D. A., Wolff, S., et al. 2018, The Astrophysical
Journal, 853, 192, doi: 10.3847/1538-4357/aaa2fd
Crossfield et al. (2014)
Crossfield, I. J. M., Biller, B., Schlieder, J. E., et al. 2014, Nature, 505,
654, doi: 10.1038/nature12955
Currie et al. (2018)
Currie, T., Brandt, T. D., Uyama, T., et al. 2018, The Astronomical Journal,
156, 291, doi: 10.3847/1538-3881/aae9ea
Cushing et al. (2006)
Cushing, M. C., Roellig, T. L., Marley, M. S., et al. 2006, ApJ, 648, 614,
doi: 10.1086/505637
Cutri et al. (2003)
Cutri, R. M., Skrutskie, M. F., van Dyk, S., et al. 2003, 2MASS All Sky
Catalog of Point Sources. (NASA/IPAC Infrared Science Archive)
Czesla et al. (2019)
Czesla, S., Schröter, S., Schneider, C. P., et al. 2019, PyA:
Python astronomy-related packages.
http://ascl.net/1906.010
De Rosa et al. (2016)
De Rosa, R. J., Rameau, J., Patience, J., et al. 2016, The Astrophysical
Journal, 824, 121, doi: 10.3847/0004-637X/824/2/121
Delorme et al. (2021)
Delorme, J.-R., Jovanovic, N., Echeverri, D., et al. 2021, JATIS, 7, 035006,
doi: 10.1117/1.JATIS.7.3.035006
Foreman-Mackey et al. (2013)
Foreman-Mackey, D., Hogg, D. W., Lang, D., & Goodman, J. 2013, PASP, 125,
306, doi: 10.1086/670067
Gandhi et al. (2020)
Gandhi, S., Brogi, M., & Webb, R. K. 2020, Monthly Notices of the Royal
Astronomical Society, 498, 194, doi: 10.1093/mnras/staa2424
Gao et al. (2018)
Gao, P., Marley, M. S., & Ackerman, A. S. 2018, The Astrophysical Journal,
855, 86, doi: 10.3847/1538-4357/aab0a1
Gao et al. (2020)
Gao, P., Thorngren, D. P., Lee, G. K. H., et al. 2020, Nature Astronomy, 4,
951, doi: 10.1038/s41550-020-1114-3
Gierasch & Conrath (1985)
Gierasch, P. J., & Conrath, B. J. 1985, in Recent Advances in Planetary
Meteorology, ed. G. E. Hunt (Cambridge University Press), 121–146
Gonzales et al. (2020)
Gonzales, E. C., Burningham, B., Faherty, J. K., et al. 2020, The
Astrophysical Journal, 905, 46, doi: 10.3847/1538-4357/abbee2
Gonzales et al. (2021)
—. 2021, The Astrophysical Journal, 923, 19, doi: 10.3847/1538-4357/ac294e
Gravity Collaboration et al. (2020)
Gravity Collaboration, Nowak, M., Lacour, S., et al. 2020, Astronomy and
Astrophysics, 633, A110, doi: 10.1051/0004-6361/201936898
Hargreaves et al. (2020)
Hargreaves, R. J., Gordon, I. E., Rey, M., et al. 2020, The Astrophysical
Journal Supplement Series, 247, 55, doi: 10.3847/1538-4365/ab7a1a
Husser et al. (2013)
Husser, T.-O., Berg, S. W.-v., Dreizler, S., et al. 2013, A&A, 553, A6,
doi: 10.1051/0004-6361/201219058
Jeffreys (1983)
Jeffreys, Harold, S. 1983, Theory of probability (Oxford: Clarendon), 459
Johnston et al. (2019)
Johnston, V. D., Cushing, M. C., & Noll, K. S. 2019, Research Notes of the
AAS, 3, 52, doi: 10.3847/2515-5172/ab0e14
Karalidi et al. (2021)
Karalidi, T., Marley, M., Fortney, J. J., et al. 2021, The Astrophysical
Journal, 923, 269, doi: 10.3847/1538-4357/ac3140
Kaufer et al. (1997)
Kaufer, A., Wolf, B., Andersen, J., & Pasquini, L. 1997, The Messenger, 89, 1
Kitzmann et al. (2020)
Kitzmann, D., Heng, K., Oreshenko, M., et al. 2020, The Astrophysical
Journal, 890, 174, doi: 10.3847/1538-4357/ab6d71
Kolecki et al. (2021)
Kolecki, J. R., Wang, J., Johnson, J. A., et al. 2021, arXiv e-prints,
arXiv:2106.13251
Konopacky et al. (2012)
Konopacky, Q. M., Ghez, A. M., Fabrycky, D. C., et al. 2012, The
Astrophysical Journal, 750, 79, doi: 10.1088/0004-637X/750/1/79
Kotani et al. (2020)
Kotani, T., Kawahara, H., Ishizuka, M., et al. 2020, in Adaptive Optics
Systems VII, Vol. 11448 (International Society for Optics and Photonics),
1144878, doi: 10.1117/12.2561755
Lavie et al. (2017)
Lavie, B., Mendonça, J. M., Mordasini, C., et al. 2017, The Astronomical
Journal, 154, 91, doi: 10.3847/1538-3881/aa7ed8
Line et al. (2015)
Line, M. R., Teske, J., Burningham, B., Fortney, J. J., & Marley, M. S. 2015,
The Astrophysical Journal, 807, 183, doi: 10.1088/0004-637X/807/2/183
López et al. (2020)
López, R. A., Hoffman, E. B., Doppmann, G., et al. 2020, in Ground-Based
and Airborne Instrumentation for Astronomy VIII, Vol. 11447 (SPIE),
1436–1450, doi: 10.1117/12.2563075
Luna & Morley (2021)
Luna, J. L., & Morley, C. V. 2021, The Astrophysical Journal, 920, 146,
doi: 10.3847/1538-4357/ac1865
Madhusudhan (2012)
Madhusudhan, N. 2012, ApJ, 758, 36, doi: 10.1088/0004-637X/758/1/36
Madhusudhan et al. (2014)
Madhusudhan, N., Amin, M. A., & Kennedy, G. M. 2014, The Astrophysical
Journal, 794, L12, doi: 10.1088/2041-8205/794/1/L12
Mamajek & Hillenbrand (2008)
Mamajek, E. E., & Hillenbrand, L. A. 2008, The Astrophysical Journal, 687,
1264, doi: 10.1086/591785
Manjavacas et al. (2021)
Manjavacas, E., Karalidi, T., Vos, J. M., Biller, B. A., & Lew, B. W. P. 2021,
The Astronomical Journal, 162, 179, doi: 10.3847/1538-3881/ac174c
Marley & Robinson (2015)
Marley, M., & Robinson, T. 2015, Annu. Rev. Astron. Astrophys., 53, 279,
doi: 10.1146/annurev-astro-082214-122522
Marley et al. (2021)
Marley, M. S., Saumon, D., Visscher, C., et al. 2021, The Astrophysical
Journal, 920, 85, doi: 10.3847/1538-4357/ac141d
Martin et al. (2018)
Martin, E. C., Fitzgerald, M. P., McLean, I. S., et al. 2018, Proc. SPIE,
10702, 107020A, doi: 10.1117/12.2312266
Mawet et al. (2017)
Mawet, D., Ruane, G., Xuan, W., et al. 2017, ApJ, 838, 92,
doi: 10.3847/1538-4357/aa647f
Miles et al. (2020)
Miles, B. E., Skemer, A. J. I., Morley, C. V., et al. 2020, The Astronomical
Journal, 160, 63, doi: 10.3847/1538-3881/ab9114
Min et al. (2005)
Min, M., Hovenier, J. W., & de Koter, A. 2005, Astronomy and Astrophysics,
432, 909, doi: 10.1051/0004-6361:20041920
Mollière et al. (2019)
Mollière, P., Wardenier, J. P., van Boekel, R., et al. 2019, Astronomy
and Astrophysics, 627, A67, doi: 10.1051/0004-6361/201935470
Mollière et al. (2020)
Mollière, P., Stolker, T., Lacour, S., et al. 2020, Astronomy and
Astrophysics, 640, A131, doi: 10.1051/0004-6361/202038325
Mordasini et al. (2016)
Mordasini, C., van Boekel, R., Mollière, P., Henning, T., & Benneke, B.
2016, The Astrophysical Journal, 832, 41, doi: 10.3847/0004-637X/832/1/41
Moses et al. (2013)
Moses, J. I., Madhusudhan, N., Visscher, C., & Freedman, R. S. 2013, The
Astrophysical Journal, 763, 25, doi: 10.1088/0004-637X/763/1/25
Noll et al. (2000)
Noll, K. S., Geballe, T. R., Leggett, S. K., & Marley, M. S. 2000, The
Astrophysical Journal, 541, L75, doi: 10.1086/312906
Öberg & Bergin (2016)
Öberg, K. I., & Bergin, E. A. 2016, The Astrophysical Journal, 831, L19,
doi: 10.3847/2041-8205/831/2/L19
Öberg et al. (2011)
Öberg, K. I., Murray-Clay, R., & Bergin, E. A. 2011, ApJ, 743, L16,
doi: 10.1088/2041-8205/743/1/L16
Otten et al. (2021)
Otten, G. P. P. L., Vigan, A., Muslimov, E., et al. 2021, Astronomy &
Astrophysics, Volume 646, id.A150, $<$NUMPAGES$>$19$<$/NUMPAGES$>$
pp., 646, A150, doi: 10.1051/0004-6361/202038517
Padoan & Nordlund (2004)
Padoan, P., & Nordlund, Å. 2004, The Astrophysical Journal, 617, 559,
doi: 10.1086/345413
Pelletier et al. (2021)
Pelletier, S., Benneke, B., Darveau-Bernier, A., et al. 2021, arXiv
e-prints, arXiv:2105.10513
Peretti et al. (2019)
Peretti, S., Ségransan, D., Lavie, B., et al. 2019, A&A, 631, A107,
doi: 10.1051/0004-6361/201732454
Petit dit de la Roche et al. (2018)
Petit dit de la Roche, D. J. M., Hoeijmakers, H. J., & Snellen, I. a. G.
2018, Astronomy and Astrophysics, 616, A146,
doi: 10.1051/0004-6361/201833384
Piso et al. (2015)
Piso, A.-M. A., Öberg, K. I., Birnstiel, T., & Murray-Clay, R. A. 2015,
The Astrophysical Journal, 815, 109, doi: 10.1088/0004-637X/815/2/109
Pollack et al. (1996)
Pollack, J. B., Hubickyj, O., Bodenheimer, P., et al. 1996, Icarus, 124, 62,
doi: 10.1006/icar.1996.0190
Pueyo (2016)
Pueyo, L. 2016, ApJ, 824, 117, doi: 10.3847/0004-637X/824/2/117
Riello et al. (2021)
Riello, M., De Angeli, F., Evans, D. W., et al. 2021, Astronomy and
Astrophysics, 649, A3, doi: 10.1051/0004-6361/202039587
Rosenthal et al. (2021)
Rosenthal, L. J., Fulton, B. J., Hirsch, L. A., et al. 2021, arXiv:2105.11583
[astro-ph].
https://arxiv.org/abs/2105.11583
Ruffio (2019)
Ruffio, J.-B. 2019, Direct Detection and Characterization of Exoplanets,
Stanford University.
https://purl.stanford.edu/yq071tj0740
Ruffio et al. (2021)
Ruffio, J.-B., Konopacky, Q. M., Barman, T., et al. 2021, AJ, 162, 290,
doi: 10.3847/1538-3881/ac273a
Samland et al. (2017)
Samland, M., Mollière, P., Bonnefoy, M., et al. 2017, Astronomy and
Astrophysics, 603, A57, doi: 10.1051/0004-6361/201629767
Saumon & Marley (2008)
Saumon, D., & Marley, M. S. 2008, ApJ, 689, 1327, doi: 10.1086/592734
Skemer et al. (2014)
Skemer, A. J., Marley, M. S., Hinz, P. M., et al. 2014, ApJ, 792, 17,
doi: 10.1088/0004-637X/792/1/17
Smith (1998)
Smith, M. D. 1998, Icarus, 132, 176, doi: 10.1006/icar.1997.5886
Sneden (1973)
Sneden, C. 1973, The Astrophysical Journal, 184, 839, doi: 10.1086/152374
Soummer et al. (2012)
Soummer, R., Pueyo, L., & Larkin, J. 2012, The Astrophysical Journal, 755,
L28, doi: 10.1088/2041-8205/755/2/L28
Speagle (2020)
Speagle, J. S. 2020, Monthly Notices of the Royal Astronomical Society, 493,
3132, doi: 10.1093/mnras/staa278
Stamatellos et al. (2007)
Stamatellos, D., Hubber, D. A., & Whitworth, A. P. 2007, Monthly Notices of
the Royal Astronomical Society, 382, L30,
doi: 10.1111/j.1745-3933.2007.00383.x
Stelter et al. (2020)
Stelter, R. D., Skemer, A. J., Sallum, S., et al. 2020, in Ground-Based and
Airborne Instrumentation for Astronomy VIII, Vol. 11447 (SPIE),
1346–1369, doi: 10.1117/12.2562768
Tan (2022)
Tan, X. 2022, Monthly Notices of the Royal Astronomical Society, 511, 4861,
doi: 10.1093/mnras/stac344
Tan & Showman (2021)
Tan, X., & Showman, A. P. 2021, Monthly Notices of the Royal Astronomical
Society, 502, 678, doi: 10.1093/mnras/stab060
Thies et al. (2010)
Thies, I., Kroupa, P., Goodwin, S. P., Stamatellos, D., & Whitworth, A. P.
2010, The Astrophysical Journal, 717, 577,
doi: 10.1088/0004-637X/717/1/577
Toon et al. (1988)
Toon, O. B., Turco, R. P., Westphal, D., Malone, R., & Liu, M. 1988, Journal
of the Atmospheric Sciences, 45, 2123,
doi: 10.1175/1520-0469(1988)045<2123:AMMFAD>2.0.CO;2
Turco et al. (1979)
Turco, R. P., Hamill, P., Toon, O. B., Whitten, R. C., & Kiang, C. S. 1979,
Journal of the Atmospheric Sciences, 36, 699,
doi: 10.1175/1520-0469(1979)036<0699:AODMDA>2.0.CO;2
Vigan (2020)
Vigan, A. 2020, Astrophysics Source Code Library, ascl:2009.002
Vos et al. (2022)
Vos, J. M., Faherty, J. K., Gagné, J., et al. 2022, The Astrophysical
Journal, 924, 68, doi: 10.3847/1538-4357/ac4502
Vousden et al. (2016)
Vousden, W. D., Farr, W. M., & Mandel, I. 2016, Monthly Notices of the Royal
Astronomical Society, 455, 1919, doi: 10.1093/mnras/stv2422
Wang et al. (2017)
Wang, J., Mawet, D., Ruane, G., Hu, R., & Benneke, B. 2017, The Astronomical
Journal, 153, 183, doi: 10.3847/1538-3881/aa6474
Wang et al. (2020a)
Wang, J., Wang, J. J., Ma, B., et al. 2020a, The Astronomical
Journal, 160, 150, doi: 10.3847/1538-3881/ababa7
Wang et al. (2022)
Wang, J., Kolecki, J. R., Ruffio, J.-B., et al. 2022, The Astronomical
Journal, 163, 189, doi: 10.3847/1538-3881/ac56e2
Wang et al. (2021a)
Wang, J. J., Kulikauskas, M., & Blunt, S. 2021a, Astrophysics
Source Code Library, ascl:2101.003
Wang et al. (2015)
Wang, J. J., Ruffio, J.-B., De Rosa, R. J., et al. 2015, Astrophysics Source
Code Library, ascl:1506.001
Wang et al. (2020b)
Wang, J. J., Ginzburg, S., Ren, B., et al. 2020b, The
Astronomical Journal, 159, 263, doi: 10.3847/1538-3881/ab8aef
Wang et al. (2021b)
Wang, J. J., Ruffio, J.-B., Morris, E., et al. 2021b, The
Astronomical Journal, 162, 148, doi: 10.3847/1538-3881/ac1349
Wang et al. (2021c)
Wang, J. J., Vigan, A., Lacour, S., et al. 2021c, The
Astronomical Journal, 161, 148, doi: 10.3847/1538-3881/abdb2d
Xuan & Wyatt (2020)
Xuan, J. W., & Wyatt, M. C. 2020, Monthly Notices of the Royal Astronomical
Society, 497, 2096, doi: 10.1093/mnras/staa2033
Yurchenko & Tennyson (2014)
Yurchenko, S. N., & Tennyson, J. 2014, Monthly Notices of the Royal
Astronomical Society, 440, 1649, doi: 10.1093/mnras/stu326
Zahnle & Marley (2014)
Zahnle, K. J., & Marley, M. S. 2014, The Astrophysical Journal, 797, 41,
doi: 10.1088/0004-637X/797/1/41
Zhang & Showman (2018)
Zhang, X., & Showman, A. P. 2018, The Astrophysical Journal, 866, 1,
doi: 10.3847/1538-4357/aada85
Zhang et al. (2021)
Zhang, Y., Snellen, I. A. G., & Mollière, P. 2021, arXiv:2109.11569
[astro-ph].
https://arxiv.org/abs/2109.11569
Zhou et al. (2018)
Zhou, Y., Apai, D., Metchev, S., et al. 2018, The Astronomical Journal, 155,
132, doi: 10.3847/1538-3881/aaabbd
Appendix A Orbit fits for HD 4747 B
Our orbit fit for the HD 4747 system is shown in Fig. 14.
Appendix B Extracted Low-resolution spectrum and GPI astrometry
Our extracted spectrum for HD 4747 B based on observations with GPI (Crepp et al., 2018) and SPHERE (Peretti et al., 2019) are given in Table 4. Our relative astrometry measurements based on the GPI data are listed in Table 5.
\startlongtable
Appendix C Priors and posteriors for retrieval parameters
Here we list the priors on our retrieved parameters and include joint posterior distributions of selected parameters from our baseline HRS and LRS retrievals. |
Curves in characteristic $2$ with non-trivial
$2$-torsion
Wouter Castryck, Marco Streng, Damiano Testa
Abstract.
Cais, Ellenberg and Zureick-Brown recently observed that over finite fields of characteristic two,
almost all smooth plane projective curves of a given odd degree admit a non-trivial rational $2$-torsion point
on their Jacobian. We extend this observation to curves given
by Laurent polynomials with a fixed Newton polygon, provided that the polygon
satisfies a certain combinatorial property. This includes many classical families, such
as hyperelliptic curves of odd genus and $C_{a,b}$ curves. In the hyperelliptic case, we give an alternative
proof using an explicit description of the $2$-torsion subgroup.
1. Introduction
The starting point of this note is a recent theorem by Cais, Ellenberg and Zureick-Brown [CEZB, Thm. 4.2], asserting
that over a finite field $k$ of characteristic $2$, almost all smooth plane projective curves of a given odd degree $d\geq 3$ have a non-trivial $k$-rational $2$-torsion point on their Jacobian. Here, ‘almost all’ means that the corresponding proportion converges to $1$ as $\#k$ and/or $d$ tend to infinity.
The underlying observation is that such curves admit
•
a ‘geometric’ $k$-rational half-canonical divisor $\Theta_{\text{geom}}$: the canonical class of a smooth plane projective curve of degree $d$ equals $(d-3)H$, where $H$ is the class of hyperplane sections; if $d$ is odd then $\frac{1}{2}(d-3)H$ is half-canonical,
•
an ‘arithmetic’ $k$-rational half-canonical divisor $\Theta_{\text{arith}}$ (whose class is sometimes called the canonical theta characteristic), related to the fact that over a perfect field of characteristic $2$, the derivative of a Laurent series is always a square [Mum, p. 191].
The difference $\Theta_{\text{geom}}-\Theta_{\text{arith}}$ maps to a $k$-rational $2$-torsion point on the Jacobian. The proof of [CEZB, Thm. 4.2] then amounts to showing that, quite remarkably, this point is generically non-trivial.
There exist many classical families of curves admitting such a ‘geometric’ half-canonical divisor. Examples include hyperelliptic curves of odd genus $g$, whose canonical class is given by $(g-1)g^{1}_{2}$ (where $g^{1}_{2}$ denotes the hyperelliptic pencil), and smooth projective curves in $\mathbf{P}_{k}^{1}\times\mathbf{P}_{k}^{1}$ of even bidegree $(a,b)$ (both $a$ and $b$ even, that is), where the canonical class reads $(a-2)R_{1}+(b-2)R_{2}$ (here $R_{1},R_{2}$ are the two rulings of $\mathbf{P}_{k}^{1}\times\mathbf{P}_{k}^{1}$). The families mentioned so far are parameterized by sufficiently generic polynomials that are supported on the polygons
$d$$d$$0$smooth plane curvesof degree $d$$2g+2$$2$$0$hyperelliptic curvesof genus $g$$a$$b$$0$curves in $\mathbf{P}_{k}^{1}\times\mathbf{P}_{k}^{1}$of bidegree $(a,b)$,
respectively. The following lemma, which is an easy consequence of the theory of toric surfaces (see Section 2), gives a purely combinatorial reason for the existence of a half-canonical divisor in these cases.
Lemma 1.
Let $k$ be a perfect field and let $\Delta$ be a two-dimensional lattice polygon. For each edge $\tau\subset\Delta$, let $a_{\tau}X+b_{\tau}Y=c_{\tau}$ be its supporting line,
where $\gcd(a_{\tau},b_{\tau})=1$. Suppose that the system of congruences
(1)
$$\left\{\ a_{\tau}X+b_{\tau}Y\equiv c_{\tau}+1\pmod{2}\ \right\}_{\tau\text{ %
edge of }\Delta}$$
admits a solution in $\mathbf{Z}^{2}$. Then any sufficiently general Laurent polynomial $f\in k[x^{\pm 1},y^{\pm 1}]$ that is supported on $\Delta$ defines a curve carrying a $k$-rational half-canonical divisor
on its non-singular complete model.
In the proof of Lemma 1 below, where we describe this half-canonical divisor explicitly, we will be more precise on the meaning of ‘sufficiently general’.
Here again, when specializing to characteristic $2$, there is, in addition, an arithmetic $k$-rational half-canonical divisor. So it is natural to wonder whether the proof of [CEZB, Thm. 4.2] still applies in these cases. We will show that it usually does.
Theorem 2.
Let $\Delta$ be a lattice polygon satisfying the conditions of Lemma 1, where in addition we assume that $\Delta$ is not unimodularly equivalent to
$1$$3$$(3,1)$$k$$1$$0$for some $k\geq 1$or$1$$0$for some $0\leq k<\ell\geq 3$with $k$ even and $\ell$ odd.$(\ell,1)$$(k,2)$
Then there exists a non-empty Zariski open subset $S_{\Delta}/\mathbf{F}_{2}$ of the space of
Laurent polynomials that are supported on $\Delta$ having the following property.
For every perfect field $k$ of characteristic $2$ and every $f\in S_{\Delta}(k)$,
the Jacobian of the non-singular complete model of the
curve defined by $f$ has a non-trivial $k$-rational $2$-torsion point.
Right before the proof of Theorem 2 we will define the set $S_{\Delta}$ explicitly. As a consequence, if $k$
is a finite field of characteristic $2$, then the proportion of Laurent polynomials
that are supported on $\Delta$, which define a curve whose Jacobian has a
non-trivial $k$-rational $2$-torsion point, tends to $1$ as $\#k\rightarrow\infty$.
(See the end of
Section 3, where we also discuss asymptotics for
increasing dilations of $\Delta$, i.e. the analogue of $d\rightarrow\infty$ in the smooth plane curve case.)
This observation seems new even for hyperelliptic curves of odd genus.111In
view of the asymptotic consequences discussed in Section 3, this observation shows that
[CFHS, Principle 3] can fail for $g>2$.
In this case we can give an alternative proof yielding sufficient conditions having a more arithmetic flavor (such as being ordinary); see Section 4. Another interesting class of examples is given by the polygons
$a$$b$$0$
where $a$ and $b$ are not both even. The case $a=b$ corresponds to the smooth plane curves of odd degree considered in [CEZB]. The case $\gcd(a,b)=1$ corresponds to so-called $C_{a,b}$ curves. The
case $b=2$, $a=2g+1$ (a subcase of the latter) corresponds
to hyperelliptic curves having a prescribed $k$-rational Weierstrass point $P$. Note that in this case $g^{1}_{2}\sim 2P$, so there is indeed always a $k$-rational half-canonical divisor, regardless of the parity of $g$. Finally, the case $b=3$, $a\geq 4$ corresponds to trigonal curves having maximal Maroni invariant (that is trigonal curves for which the series $(h^{0}(ng^{1}_{3}))_{n\in\mathbf{Z}_{\geq 0}}$ starts increasing by steps of $3$ as late as the Riemann-Roch theorem allows it to do); if $a=6$, these are exactly the genus-$4$ curves having a unique $g^{1}_{3}$.
Remark 3.
This explains why Denef and Vercauteren had to allow a factor $2$ while generating
cryptographic hyperelliptic and $C_{a,b}$ curves in characteristic $2$; see Sections 6 of [DV1, DV2].
2. Half-canonical divisors from toric geometry
Let $k$ be a perfect field and let $f=\sum_{(i,j)\in\mathbf{Z}^{2}}c_{i,j}x^{i}y^{j}\,\in k[x^{\pm 1},y^{\pm 1}]$ be an absolutely
irreducible Laurent polynomial. Let
$$\Delta(f)=\textup{conv}\left\{\,\left.(i,j)\in\mathbf{Z}^{2}\,\right|\,c_{i,j}%
\neq 0\,\right\}$$
be the Newton polygon of $f$, which we assume to be two-dimensional. We say that $f$ is non-degenerate with respect to its Newton polygon if for every face $\tau\subset\Delta(f)$ (vertex, edge, or $\Delta(f)$ itself)
the system
$$f_{\tau}=\frac{\partial f_{\tau}}{\partial x}=\frac{\partial f_{\tau}}{%
\partial y}=0\qquad\text{with }f_{\tau}=\sum_{(i,j)\in\tau\cap\mathbf{Z}^{2}}c%
_{i,j}x^{i}y^{j}$$
has no solutions over an algebraic closure of $k$.
For a given Newton polygon $\Delta=\Delta(f)$, the condition of non-degeneracy is generically
satisfied, in the sense that it is characterized by the non-vanishing of
$$\rho_{\Delta}:=\text{Res}_{\Delta}\left(f,x\frac{\partial f}{\partial x},y%
\frac{\partial f}{\partial y}\right)\,\in\mathbf{Z}[c_{i,j}|(i,j)\in\Delta\cap%
\mathbf{Z}^{2}]$$
(where $\text{Res}_{\Delta}$ is the sparse resultant; $\rho_{\Delta}$ does not vanish identically in any characteristic [CV, §2]).
Let $C_{f}$ be the curve in $\mathbf{T}_{k}^{2}=\text{Spec}\,k[x^{\pm 1},y^{\pm 1}]$ cut out by $f$ and let $\text{Tor}_{k}(\Delta(f))$ be the toric surface corresponding to $\Delta(f)$; it is a compactification of $\mathbf{T}_{k}^{2}$ to which the self-action of $\mathbf{T}_{k}^{2}$ extends algebraically. There is a natural dimension-preserving bijection between the orbits of this extended action and the faces of $\Delta(f)$; for each face $\tau$, write $O(\tau)$ for the corresponding orbit. Then geometrically, the condition of non-degeneracy means that $C_{f}$ is a non-singular curve in $\mathbf{T}_{k}^{2}=O(\Delta(f))$ compactifying to a curve $C^{\prime}_{f}$ in $\text{Tor}_{k}(\Delta(f))$ that does not contain any of the zero-dimensional $O(\tau)$’s and that intersects the one-dimensional $O(\tau)$’s transversally.
$\Delta(f)$$\tau_{1}$$\tau_{2}$$\tau_{3}$$\tau_{4}$$\mathbf{R}^{2}$$\mathbf{T}^{2}$$C_{f}^{\prime}$$O(\tau_{2})$$O(\tau_{4})$$O(\tau_{1})$$O(\tau_{3})$$\text{Tor}_{k}(\Delta(f))$
In particular, since $\text{Tor}_{k}(\Delta(f))$ is normal, non-degeneracy implies that $C^{\prime}_{f}$ is a non-singular complete model of $C_{f}$. See [CDV, §2] for a more detailed description of this construction.
Example 4.
Assume that $\Delta(f)=\textup{conv}\{(0,0),(d,0),(0,d)\}$, i.e. $f$ is a degree-$d$ polynomial with non-zero coefficients at $1,x^{d},y^{d}$.
In this case $\text{Tor}_{k}(\Delta(f))$ is just the projective plane, and the toric orbits are $\mathbf{T}_{k}^{2}=O(\Delta(f))$,
the three coordinate points $(1:0:0)$, $(0:1:0)$ and $(0:0:1)$
(which are the orbits of the form $O(\text{vertex})$),
and the three coordinate axes from which the coordinate points are removed (these are the orbits of the form $O(\text{edge})$).
Then $f$ is non-degenerate with respect to its Newton polygon if and only if $C_{f}$ compactifies to a non-singular projective plane curve that
is non-tangent to any of the coordinate axes, and that does not contain any of the coordinate points.
Example 5.
Let $g\geq 2$ be an integer, and consider $f=y^{2}+h_{1}(x)y+h_{0}(x)$, where $\deg h_{1}\leq g+1$, $\deg h_{0}=2g+2$, and $h_{0}(0)\neq 0$. Then
$\Delta(f)=\textup{conv}\{(0,0),(2g+2,0),(0,2)\}$, and $\text{Tor}_{k}(\Delta(f))$ is the weighted projective plane $\mathbf{P}_{k}(1:g+1:1)$. Here again,
$f$ is non-degenerate with respect to its Newton polygon if and only if $C^{\prime}_{f}$ is a non-singular curve that is non-tangent to the coordinate axes
and does not contain any coordinate points. In this case $C^{\prime}_{f}$ is a hyperelliptic
curve of genus $g$ (cf. Remark 8).
Now for each edge $\tau\subset\Delta(f)$ let
$\nu_{\tau}\in\mathbf{Z}^{2}$ be the inward pointing primitive normal vector to $\tau$, let $p_{\tau}$ be any element of $\tau\cap\mathbf{Z}^{2}$, and let $D_{\tau}$ be the $k$-rational divisor on $C^{\prime}_{f}$ cut out by $O(\tau)$. If $f$ is non-degenerate with respect to its Newton polygon, then one can show
(2)
$$\text{div}\frac{dx}{xy\frac{\partial f}{\partial y}}=\sum_{\tau\text{ edge}}(-%
\langle\nu_{\tau},p_{\tau}\rangle-1)D_{\tau}.$$
Here $\langle\cdot,\cdot\rangle$ is the standard inner product on $\mathbf{R}^{2}$.
See [CDV, Cor. 3] for an elementary but elaborate proof of (2).
It is possible to give a more conceptual proof using adjunction theory.
Remark 6.
Using the theory of sparse resultants, one can show that $\partial f/\partial y$ does
not vanish identically,
so that the left-hand side of (2) makes sense.
Note also that $df=\frac{\partial f}{\partial x}dx+\frac{\partial f}{\partial y}dy=0$, so
we could as well have written
(3)
$$\text{div}\frac{dy}{xy\frac{\partial f}{\partial x}}.$$
Proof of Lemma 1.
Assume that $f$ is such that $\Delta(f)=\Delta$ and that it is non-degenerate with respect to its Newton polygon (which, as mentioned above, is a non-empty Zariski open condition).
Let $(i_{0},j_{0})\in\mathbf{Z}^{2}$ be a solution to
the given system of congruences. Then the translated polygon $(-i_{0},-j_{0})+\Delta$ is such that all corresponding $\langle\nu_{\tau},p_{\tau}\rangle$’s are odd. So by applying the above to $x^{-i_{0}}y^{-j_{0}}f$, we find that
$$\sum_{\tau\text{ edge}}\frac{-\langle\nu_{\tau},p_{\tau}\rangle-1}{2}D_{\tau}$$
is a $k$-rational half-canonical divisor on $C^{\prime}_{x^{-i_{0}}y^{-j_{0}}f}=C^{\prime}_{f}$.
∎
Remark 7.
Let $\Delta=\textup{conv}\{(0,0),(d,0),(0,d)\}$ with $d$ odd. Then the condition of non-degeneracy restricts our attention to smooth plane curves of degree $d$ that do not contain the coordinate points and that intersect the coordinate axes transversally. But of course any smooth plane curve of degree $d$ carries a $k$-rational half-canonical divisor. This shows that the non-degeneracy condition, even though it is generically satisfied,
is sometimes a bit stronger than needed.222The reader might want to note that there always exists an automorphism of $\mathbf{P}_{k}^{2}$ that puts our smooth plane curve in a non-degenerate position (at least if $\#k$ is sufficiently large). But for
more general instances of $\Delta$, the automorphism group of $\text{Tor}_{k}(\Delta)$ may be much smaller (e.g. the only automorphisms may be the ones coming from the $\mathbf{T}_{k}^{2}$-action), in which case it
might be impossible to resolve tangency to the one-dimensional toric orbits.
In general, the according weaker condition reads that $C_{f}$ compactifies to a non-singular curve in $\text{Tor}_{k}(\Delta)$.
Here we have to revisit Remark 6, however:
there do exist instances of absolutely irreducible Laurent polynomials $f\in k[x^{\pm 1},y^{\pm 1}]$ for which $C_{f}$ compactifies to a non-singular curve in $\text{Tor}_{k}(\Delta(f))$, yet for
which $\partial f/\partial y$ does vanish identically (example: $f=1+x^{2}y^{2}+x^{3}y^{2}$).
For these instances of $f$
the left-hand side of (2) does not make sense. But in that
case $\partial f/\partial x$ does not vanish identically (otherwise $C_{f}$ would have singularities), and one can prove that (2) holds with the left-hand side replaced by (3).
Remark 8.
We mention two other well-known features of non-degenerate Laurent polynomials, that can be seen as consequences to (2); see for instance [CV] and the references therein:
•
the geometric genus of $C_{f}$ equals $\#\left(\Delta(f)^{\circ}\cap\mathbf{Z}^{2}\right)$, where $\Delta(f)^{\circ}$ denotes the
interior of $\Delta(f)$, and
•
in case $\#\left(\Delta(f)^{\circ}\cap\mathbf{Z}^{2}\right)\geq 2$, $C_{f}$ is hyperelliptic if and only if $\Delta(f)^{\circ}\cap\mathbf{Z}^{2}$ is contained in a line.
3. Proof of the main result
Lemma 9.
Let $\Delta$ be a two-dimensional lattice polygon and
suppose as in Lemma 1 that (1) admits a solution in $\mathbf{Z}^{2}$. If $\Delta$ is not among the polygons excluded in the hypothesis of Theorem 2, then there is a solution of (1) contained in $\Delta\cap\mathbf{Z}^{2}$.
Proof.
Let us first classify all two-dimensional lattice polygons $\Delta$ for which the reduction-modulo-$2$ map $\pi_{\Delta}\colon\Delta\cap\mathbf{Z}^{2}\rightarrow\left(\mathbf{Z}/(2)%
\right)^{2}$ is not surjective.
If the interior lattice points of $\Delta$ lie on a line,
then surjectivity fails if and only if $\Delta$ is among
$(0,0)$$(0,1)$$(k,0)$,(for some $k\geq 1$)(a)$(0,0)$$(1,2)$$(2,0)$,(b)$(0,1)$$(1,2)$$(2,1)$,$(1,0)$,(c)$(0,1)$$(k,2)$$(\ell,1)$$(0,0)$($0\leq k\leq\ell\geq 3$, $k$ even)(d)
(up to unimodular equivalence). This assertion follows from Koelman’s classification; see [Koe, Ch. 4] or [Cas, Thm. 10]. As a consequence, any two-dimensional $\Delta$ for which $\pi_{\Delta}$ is not surjective must contain one of the polygons (a-d) as an ‘onion skin’, i.e. as a lattice polygon obtained by subsequently taking the convex hull of the interior lattice points. But using the criterion from [HS, Lem. 9-11] one sees that the only polygons in
(a-d) that can appear as
convex hull of the interior lattice points
of a strictly bigger lattice polygon $\Gamma$ are the polygons (a) with $k=1$ or $k=2$, the polygon (b) and the polygon (c). The only corresponding instance of $\Gamma$ for which $\pi_{\Gamma}$ is not surjective is
$(0,-1)$$(-1,2)$$(2,0)$(e)
which, again by [HS, Lem. 9-11], is an onion skin only of itself.
This ends the classification.
Now let $\Delta$ be a two-dimensional lattice polygon and suppose that (1) admits a solution in $\mathbf{Z}^{2}$. If
$\pi_{\Delta}$ is surjective, then it clearly also admits a solution in $\Delta\cap\mathbf{Z}^{2}$. So we may assume that
$\Delta$ is among (a-e). Then the lemma follows by noting that cases (b), (c) and (d) with $\ell$ even admit the solution $(1,1)\in\Delta\cap\mathbf{Z}^{2}$, and that cases (a), (e) and (d) with $\ell$ odd were excluded in the énoncé.
∎
Remark 10.
Because of Remark 8, the excluded polygons correspond to certain classes of smooth plane quartics, rational curves, and hyperelliptic curves, respectively.
We can now define the variety $S_{\Delta}$ mentioned in the statement of Theorem 2. Namely, we will prove the existence of a
non-trivial $k$-rational $2$-torsion point under the assumption that
•
$\Delta(f)=\Delta$ and $f$ is non-degenerate with respect to its Newton polygon (i.e. the genericity assumption from Lemma 1), and
•
for at least one solution $(i_{0},j_{0})\in\Delta\cap\mathbf{Z}^{2}$ to the system of congruences (1), the corresponding coefficient $c_{i_{0},j_{0}}$ is non-zero.
So we can let $S_{\Delta}$ be defined by $c_{i_{0},j_{0}}\rho_{\Delta}\neq 0$.
Remark 11.
Here again, one can weaken the non-degeneracy condition as described in Remark 7. When that stronger version is applied
to $\Delta=\textup{conv}\{(0,0),(d,0),(0,d)\}$ with $d$ odd, one exactly recovers [CEZB, Thm. 4.2].
Proof of Theorem 2.
By replacing $f$ with $x^{-i_{0}}y^{-j_{0}}f$
if needed, we assume that $(0,0)\in\Delta$ is a solution to the system of congruences (1) and that the constant term of $f$ is non-zero.
As explained in [Mum, p. 191], $C_{f}^{\prime}$ comes equipped with a $k$-rational divisor $\Theta_{\text{arith}}$ such that $2\Theta_{\text{arith}}=\text{div}\,dx$.
(Recall that the derivative of a Laurent series over $k$ is always a square, so the order of $dx$ at a point of $C_{f}^{\prime}$ is indeed even.)
On the other hand, Lemma 1 and its proof provide us with a $k$-rational divisor $\Theta_{\text{geom}}$ such that
$$2\Theta_{\text{geom}}=\text{div}\,\frac{dx}{xy\frac{\partial f}{\partial y}}.$$
In order to prove that $\Theta_{\text{geom}}\not\sim\Theta_{\text{arith}}$ (and hence that $\text{Jac}(C^{\prime}_{f})$ has a non-trivial $k$-rational $2$-torsion point), we
need to show that
$$xy\frac{\partial f}{\partial y}$$
is a non-square when considered as an element of the function field $k(C_{f})$. If it were a square, then there would exist Laurent polynomials $\alpha,G,H$ such that
(4)
$$H^{2}xy\frac{\partial f}{\partial y}+\alpha f=G^{2}\qquad\text{in $k[x^{\pm 1}%
,y^{\pm 1}]$},$$
where $f\nmid H$.
Taking derivatives with respect to $y$ yields
$$(\alpha+H^{2}x)\frac{\partial f}{\partial y}=\frac{\partial\alpha}{\partial y}f,$$
which together with (4) results in
$$\left((\alpha+H^{2}x)\alpha+H^{2}xy\frac{\partial\alpha}{\partial y}\right)f=(%
\alpha+H^{2}x)G^{2}.$$
Since $f$ is irreducible, it follows that $f\mid(\alpha+H^{2}x)$ or $f\mid G^{2}$. Using (4) and $f\nmid H$, the latter implies that $f\mid\frac{\partial f}{\partial y}$, which is
a contradiction (by the theory of sparse resultants, see Remark 6; one can alternatively repeat the argument using (3) if wanted).
So we know that $f\mid(\alpha+H^{2}x)$. Along with (4) we conclude that there exists a Laurent polynomial $\beta\in k[x^{\pm 1},y^{\pm 1}]$ such that
$$H^{2}x\left(y\frac{\partial f}{\partial y}+f\right)+\beta f^{2}=G^{2}.$$
Taking derivatives with respect to $x$ yields
$$H^{2}\left(f+x\frac{\partial f}{\partial x}+y\frac{\partial f}{\partial y}+xy%
\frac{\partial^{2}f}{\partial x\partial y}\right)+\frac{\partial\beta}{%
\partial x}f^{2}=0.$$
Since $f$ has a non-zero constant term, the large factor between brackets is non-zero. On the other hand, since $f\nmid H$, it must be a multiple of $f^{2}$.
Note that $\Delta(f^{2})=2\Delta(f)$, while $\Delta(f+\cdots+xy\partial^{2}f/(\partial x\partial y))\subset\Delta(f)$. This is a contradiction.
∎
We end this section by discussing some asymptotic consequences to Theorem 2.
Growing field size.
Let $\Delta$ be a two-dimensional
lattice polygon satisfying the conditions of Theorem 2.
Let $k$ be a finite field of characteristic $2$. Because
non-degeneracy is characterized by the non-vanishing of $\rho_{\Delta}$, the
proportion of Laurent polynomials $f\in k[x^{\pm 1},y^{\pm 1}]$ that are
non-degenerate with respect to their Newton polygon $\Delta(f)=\Delta$ (amongst all
Laurent polynomials that are supported on $\Delta$) converges to $1$
as $\#k\rightarrow\infty$. Then Theorem 2 implies:
$$\lim_{\#k\rightarrow\infty}\text{Prob}\left(\,\text{Jac}(C_{f}^{\prime})(k)[2]%
\neq 0\,\left|\,\text{$f\in k[x^{\pm 1},y^{\pm 1}]$ is non-degenerate
with \phantom{x}}\right.\right.$$
$$\left.\left.\text{\phantom{$\in k[x^{\pm 1},y^{\pm 1}]$} respect to its Newton%
polygon $\Delta(f)=\Delta$}\right.\right)=1.$$
As soon as $\#(\Delta^{\circ}\cap\mathbf{Z}^{2})\geq 2$ this is deviating statistical behavior: in view
of Katz-Sarnak-Chebotarev-type density theorems [KS, Theorem 9.7.13], for a general smooth proper family of
genus $g$ curves, one expects that the
probability of having a non-trivial
rational $2$-torsion point on the Jacobian
approaches the chance that a random matrix in $\text{GL}_{g}(\mathbf{F}_{2})$ satisfies
$\det(M-\text{Id})=0$, which is
$$-\sum_{r=1}^{g}\prod_{j=1}^{r}\frac{1}{1-2^{j}}$$
by [CFHS, Thm. 6]. For $g=1,2,3,4,\dots$, these probabilities are $1,\frac{2}{3},\frac{5}{7},\frac{32}{45},\dots$ (converging to about $0.71121$).
In the table below we denote by $\square_{i}$ the square $[0,i]^{2}$ (for $i=2,3,4$),
by $H_{g}$ the hyperelliptic polygon $\textup{conv}\{(0,0),(2g+2,0),(0,2)\}$ (for $g=7,8$), and
by $E$ the exceptional polygon $\textup{conv}\{(1,0),(3,1),(0,3)\}$ from the statement of Theorem 2.
Each entry
corresponds to a sample of $10^{4}$ uniformly randomly chosen
Laurent polynomials $f\in k[x^{\pm 1},y^{\pm 1}]$ that are supported on $\square_{2},\square_{3},\dots$
The table presents the proportion of $f$’s for which $\text{Jac}(C_{f}^{\prime})$ has a non-trivial $k$-rational $2$-torsion point,
among those $f$’s that are non-degenerate with
respect to their Newton polygon $\Delta(f)=\square_{2},\square_{3},\dots$ An entry ‘-’
means that we did not do the computation because it would take too long. The computations were
carried out using Magma [BCP].
$$k\phantom{i}$$
$$\begin{array}[]{c}\square_{2}\\
\text{\footnotesize{$(g=1)$}}\\
\end{array}$$
$$\begin{array}[]{c}\square_{3}\\
\text{\footnotesize{$(g=4)$}}\\
\end{array}$$
$$\begin{array}[]{c}\square_{4}\\
\text{\footnotesize{$(g=9)$}}\\
\end{array}$$
$$\begin{array}[]{c}H_{7}\\
\text{\footnotesize{$(g=7)$}}\\
\end{array}$$
$$\begin{array}[]{c}H_{8}\\
\text{\footnotesize{$(g=8)$}}\\
\end{array}$$
$$\begin{array}[]{c}E\\
\text{\footnotesize{$(g=3)$}}\\
\end{array}$$
$$\mathbb{F}_{2}\phantom{i}$$
$$0/0$$
0.370
0.958
0.995
0.670
0.143
$$\mathbb{F}_{4}\phantom{i}$$
0.750
0.621
1.000
1.000
0.795
0.449
$$\mathbb{F}_{8}\phantom{i}$$
0.884
0.654
1.000
1.000
0.852
0.591
$$\mathbb{F}_{16}\phantom{i}$$
0.940
0.697
1.000
1.000
0.872
0.661
$$\mathbb{F}_{32}\phantom{i}$$
0.968
0.704
1.000
1.000
0.877
0.696
$$\mathbb{F}_{64}\phantom{i}$$
0.986
-
1.000
1.000
0.880
0.694
$$\mathbb{F}_{128}\phantom{i}$$
0.992
-
1.000
1.000
0.889
0.708
$$\mathbb{F}_{256}\phantom{i}$$
0.996
-
1.000
1.000
0.888
-
$$\begin{array}[]{r}\text{\footnotesize{asymptotic}}\\
\text{\footnotesize{prediction}}\\
\end{array}$$
$$1$$
$$\frac{32}{45}\approx 0.711$$
$$1$$
$$1$$
$$\frac{8}{9}\approx 0.889$$
$$\frac{5}{7}\approx 0.714$$
Note that the conditions of Theorem 2 are
satisfied for $\square_{2}$, $\square_{4}$ and $H_{7}$. So here we proved that the proportion converges to $1$.
In the case of $H_{8}$, by the material in Section 4
(see Corollary 16) we know that the proportion converges
to $\frac{8}{9}$. In the other two cases $\square_{4}$ and $E$ we have no clue, so our best guess is that these follow
the $\text{GL}_{g}(\mathbf{F}_{2})$-model.
Growing polygon.
Let $k$ be a finite field of characteristic $2$. If $\Delta$ is a two-dimensional lattice polygon satisfying the conditions
of Lemma 1, then the same holds for each odd Minkowski multiple $(2n+1)\Delta$.
It seems reasonable to assume that the proportion of Laurent polynomials $f\in k[x^{\pm 1},y^{\pm 1}]$ that are
non-degenerate with respect to their Newton polygon $\Delta(f)=(2n+1)\Delta$ (amongst all
Laurent polynomials that are supported on $(2n+1)\Delta$) converges to
a certain strictly positive constant.
If $\text{Tor}_{k}(\Delta)$ is smooth then
this is certainly true for the larger proportion of Laurent polynomials $f$ satisfying
the weaker condition from Remark 7, namely that
$C_{f}$ compactifies to a non-singular curve in $\text{Tor}_{k}((2n+1)\Delta)$. Indeed,
using [Poo2, Thm. 1.1] one
can show that this proportion converges to
$$Z_{\text{Tor}_{k}(\Delta)}((\#k)^{-3})^{-1}=(1-(\#k)^{-1})(1-(\#k)^{-2})^{r(%
\Delta)-2}(1-(\#k)^{-3})$$
as $n\rightarrow\infty$; here $Z_{\text{Tor}_{k}(\Delta)}$ is the Hasse-Weil Zeta function of $\text{Tor}_{k}(\Delta)$, and $r(\Delta)$ is the number of vertices of $\Delta$.
On the other hand, the number of solutions to (1) inside $(2n+1)\Delta\cap\mathbf{Z}^{2}$ tends to infinity. So
the assumption would allow
one to conclude:
$$\lim_{n\rightarrow\infty}\text{Prob}\left(\,\text{Jac}(C_{f}^{\prime})(k)[2]%
\neq 0\,\left|\,\text{$f\in k[x^{\pm 1},y^{\pm 1}]$ is non-degenerate
with \phantom{x}}\right.\right.$$
$$\left.\left.\text{\phantom{$\in k[x^{\pm 1},y^{\pm 1}]$} respect to its Newton%
polygon $\Delta(f)=(2n+1)\Delta$}\right.\right)=1.$$
This is again deviating statistical behavior: in view of Cohen-Lenstra type heuristics, one naively expects a probability of about
$$1-\prod_{j=1}^{\infty}(1-2^{-j})\approx 0.71121;$$
see [CEZB] for some additional comments.
When applied to $\Delta=\textup{conv}\{(0,0),(1,0),(0,1)\}$ one recovers the claim made before [CEZB, Thm. 4.2]
(taking into account Remark 7).
4. Hyperelliptic curves of odd genus
Let $C$ be a hyperelliptic curve
of genus $g$
over a perfect field $k$.
Then $C$ has a smooth weighted projective plane model
(5)
$$C\colon\quad Y^{2}+H(X,Z)Y=F(X,Z),$$
where $H$ and $F$ in $k[X,Z]$ are homogeneous
of degrees $g+1$ and $2g+2$ respectively.
The Newton polygon of (the defining polynomial of) the corresponding
affine model $y^{2}+H(x,1)y-F(x,1)=0$
is contained in
a triangle
with vertices $(0,0)$, $(2g+2,0)$ and $(0,2)$,
and is generically equal to this triangle.
In particular, Theorem 2 implies
that if the characteristic of $k$ is $2$
and $C$ is sufficiently general of odd genus, then its Jacobian
has a non-trivial $k$-rational $2$-torsion point.
The purpose of this section is to show
that we can replace ‘sufficiently general’ by ‘ordinary’
in this case (Corollary 14), and to
give explicit examples of the rational half-canonical divisors
from the proof of Theorem 2.
Theorem 12.
Let $C/k$ be a hyperelliptic curve over a perfect
field $k$ of characteristic $2$
given by a smooth model (5).
The Jacobian of $C$ has no
rational point of order $2$
if and only if $H(X,Z)$
is a power of an irreducible odd-degree
polynomial in $k[X,Z]$.
Corollary 13.
Let $C/k$ be a hyperelliptic curve of odd $2$-rank
over a perfect field $k$ of characteristic $2$.
Then the Jacobian of $C$ has a $k$-rational point
of order $2$.
Corollary 14.
Let $C/k$ be an ordinary hyperelliptic curve of odd genus
over a perfect field $k$ of characteristic $2$.
Then the Jacobian of $C$ has a $k$-rational point
of order $2$.
Corollary 15.
Let $C/k$ be a hyperelliptic curve of genus $2^{m}-1$ over a perfect field
$k$ of characteristic $2$, for some integer $m\geq 2$.
If the Jacobian of $C$ has no $k$-rational point of order $2$,
then it has $2$-rank zero, but it is not supersingular.
Finally, for integers $g,r\geq 1$, let $c_{g,r}$ be the proportion of
equations (5) over $\mathbf{F}_{2^{r}}$
that define a curve of genus $g$ whose
Jacobian has at least one rational point of order $2$.
Corollary 16.
The limit $\lim_{r\rightarrow\infty}c_{g,r}$ exists and
we have
$$\lim_{r\rightarrow\infty}c_{g,r}=\begin{cases}1&\text{if $g$ is odd,}\\
g/(g+1)&\text{if $g$ is even}.\end{cases}$$
Proof of Theorem 12.
All we need to do is describe the two-torsion of the Jacobian
$\text{Jac}(C)$ of $C$. Since we were not able to find a ready-to-use statement in the literature,
we give a stand-alone treatment, even though what follows
is undoubtedly known to several experts in the field; for instance, it
is implicitly contained in [EP]. Let $\overline{k}$ be an algebraic closure
of $k$. Note that $C$ has a unique point $Q_{(a:b)}=(a:\sqrt{F(a,b)}:b)\in C(\overline{k})$
for every root $(a:b)\in\mathbf{P}^{1}_{\overline{k}}$ of $H=H(X,Z)$.
This gives $n$ points, where $n\in\{1,\dots,g+1\}$ is the number
of distinct roots of $H$.
Let $D$ be the divisor of zeroes of a vertical line,
so $D$ is effective of degree $2$.
All such divisors $D$ are linearly equivalent,
and are linearly equivalent to $2Q_{(a:b)}$ for each $(a:b)$.
In particular, if we let
$$A=\ker\left(\xymatrix{\displaystyle\bigoplus_{(a:b)}(\mathbf{Z}/2\mathbf{Z})%
\ar^{\sum}[r]&(\mathbf{Z}/2\mathbf{Z})}\right),$$
then we have a homomorphism
$$\displaystyle A$$
$$\displaystyle\quad\longrightarrow\quad\text{Jac}(C)(\overline{k})[2]$$
$$\displaystyle(c_{(a:b)}\ \mbox{mod}\ 2)_{(a:b)}$$
$$\displaystyle\quad\longmapsto\quad(\sum_{(a:b)}c_{(a:b)}Q_{(a:b)})-(\frac{1}{2%
}\sum_{(a:b)}c_{(a:b)})D.$$
In fact, this map is an isomorphism. Indeed, it is injective
because if the divisor of a function is invariant under the hyperelliptic involution, then so
is the function itself, i.e. it is contained in $\overline{k}(x)$. But at the points $Q_{(a:b)}$
such functions can only admit poles or zeroes having an even order.
Surjectivity follows from the fact that $\text{Jac}(C)(\overline{k})[2]$ is generated by divisors that are supported on
the Weierstrass locus of $C$. This can be seen using Cantor’s algorithm
[Kob, Appendix.§6-7], for the application of which one needs to
transform the curve to a so-called imaginary model; this is always possible over $\overline{k}$.
Alternatively,
surjectivity follows from the injectivity and the fact that $\#\text{Jac}(C)(\overline{k})[2]=2^{n-1}$ by
[EP, Thm. 1.3].
Then in particular, the rational $2$-torsion subgroup
$\text{Jac}(C)(k)[2]$
is isomorphic to the subgroup of elements of $A$
that are invariant under $\mathrm{Gal}(\overline{k}/k)$,
that is, to
$$A_{k}=\ker\left(\bigoplus_{P\mid H}(\mathbf{Z}/2\mathbf{Z})\rightarrow(\mathbf%
{Z}/2\mathbf{Z}):(c_{P})_{P}\mapsto\sum_{P}c_{P}\deg(P)\right)$$
where the sum is taken over the irreducible factors $P$ of $H$.
The only way for $A_{k}$ to be trivial is for $H$
to be the power of an irreducible factor $P$
of odd degree.
∎
Proof of Corollary 13.
Let $n$ be the degree of the radical $R$ of $H$.
The $2$-rank of $C$ equals $n-1$
(as in the proof of Theorem 12; see e.g. [EP, Thm. 1.3]).
So if the $2$-rank is odd, then $R$ has even degree, which
implies that $H$ is not a power of an odd-degree polynomial.
In particular, Theorem 12
implies that $C$ has a non-trivial $k$-rational $2$-torsion point.
∎
Proof of Corollary 14.
This is a special case of Corollary 13
since in characteristic $2$, the $2$-rank of an ordinary abelian variety equals
its dimension.
∎
Proof of Corollary 15.
If there is no rational point of order $2$, then $H$ is a power of
a polynomial of odd degree dividing $\deg(H)=g+1=2^{m}$.
In other words, it is a power
of a linear polynomial and hence the $2$ rank of $C$ is zero.
There are no supersingular hyperelliptic
curves of genus $2^{m}-1$ in characteristic $2$ by [SZ, Thm. 1.2].
∎
Proof of Corollary 16.
As $r\rightarrow\infty$, the proportion of equations (5)
for which $H$ is not separable becomes negligible.
By Theorem 12 it therefore
suffices to prove the corresponding limit for the proportion
of degree $g+1$ polynomials that are not irreducible of odd degree.
If $g$ is odd then this proportion is clearly $1$. If $g$ is even
then this
is the same as the proportion of reducible polynomials of degree $g+1$,
which converges to $1-(g+1)^{-1}$.
∎
Remark 17.
In Corollary 16, instead of working with the proportion of equations (5),
we can work with the corresponding proportion of $\mathbf{F}_{2^{r}}$-isomorphism classes of
hyperelliptic curves of genus $g$. This is because
the subset of equations (5) that define a hyperelliptic curve of genus $g$ whose
only non-trivial geometric automorphism
is the hyperelliptic involution (inside the affine space of all
equations of this form) is non-empty [Poo1], open, and defined over $\mathbf{F}_{2}$ (being invariant under
the $\text{Gal}(\overline{\mathbf{F}}_{2},\mathbf{F}_{2})$-action). See also [Zhu].
Theorem 2 proves Corollary 14
for sufficiently general curves.
We finish by showing that the $2$-torsion points
from both proofs are equal.
The proof of Theorem 2 provides
$\Theta_{\mathrm{arith}}$ and $\Theta_{\mathrm{geom}}$
with $2\Theta_{\mathrm{arith}}\sim 2\Theta_{\mathrm{geom}}$,
hence the class of
$T=\Theta_{\mathrm{arith}}-\Theta_{\mathrm{geom}}$
is two-torsion.
We have $2\Theta_{\mathrm{arith}}=\text{div}\,dx$.
To compute $2\Theta_{\mathrm{geom}}$, we need to take
an appropriate model as in the proof of
Lemma 1.
The bivariate polynomial $y^{2}+H(x,1)y+F(x,1)$
gives an affine model of our hyperelliptic curve $C$,
and if $g$ is odd, then
the system from Lemma 1 admits the solution $(1,1)$.
By the proof of that lemma, we should then look at the
toric model $C^{\prime}_{f}$ where
$$f=x^{-1}(y+H(x,1)+y^{-1}F(x,1)).$$
Then $\Theta_{\mathrm{geom}}$
is given by $2\Theta_{\mathrm{geom}}=\text{div}\frac{1}{xy\frac{\partial f}{\partial y}}dx$,
so
we compute
$$\frac{\partial f}{\partial y}=x^{-1}(1+y^{-2}F(x,1))=x^{-1}y^{-1}\ H(x,1).$$
We find
$$T=\Theta_{\mathrm{arith}}-\Theta_{\mathrm{geom}}=\frac{1}{2}\ \text{div}\,xy%
\frac{\partial f}{\partial y}=\frac{1}{2}\text{div}\,H(x,1),$$
where $\text{div}\,H(x,1)$ is twice the sum of all points
$P_{(a:b)}$ as $(a:b)$ ranges over the roots of $H(X,Z)$ in $\mathbf{P}^{1}_{\overline{k}}$ (with multiplicity),
minus $(g+1)$ times the divisor $D$ of degree $2$ at infinity.
So both proofs give the same $2$-torsion point in cases
where they both apply.
References
[BCP]
W. Bosma, J. Cannon, C. Playoust, The Magma algebra system. I. The user language,
J. Symbolic Comput. 24, pp. 235â265 (1997)
[CEZB]
B. Cais, J. Ellenberg, D. Zureick-Brown, Random Dieudonné modules, random $p$-divisible groups, and random curves over finite fields, to appear in J. Math. Inst. Jussieu
[Cas]
W. Castryck, Moving out the edges of a lattice polygon, Discrete and Computational Geometry 47(3), pp. 496-518 (2012)
[CDV]
W. Castryck, J. Denef, F. Vercauteren, Computing zeta functions of nondegenerate curves, Int. Math. Res. Pap. 2006, pp. 1-57 (2006)
[CFHS]
W. Castryck, A. Folsom, H. Hubrechts, A.V. Sutherland, The probability that the number of points on
the Jacobian of a genus $2$ curve is prime, Proc. London Math. Soc. 104(6), pp. 1235-1270 (2012)
[CV]
W. Castryck, J. Voight, On nondegeneracy of curves, Algebra & Number Theory 3(3), pp. 255-281 (2009)
[DV1]
J. Denef, F. Vercauteren, Computing zeta functions of hyperelliptic curves over finite fields of characteristic $2$, Proc.
of ‘Advances in Cryptology – CRYPTO 2002’, Lect. Not. Comp. Sc. 2442, pp. 308-323 (2002)
[DV2]
J. Denef, F. Vercauteren, Computing zeta functions of $C_{a,b}$ curves using Monsky-Washnitzer cohomology, Fin. Fields App. 12(1), pp. 78-102 (2006)
[EP]
A. Elkin, R. Pries, Ekedahl-Oort strata of hyperelliptic curves in characteristic $2$, to appear in Algebra & Number Theory
[HS]
C. Haase, J. Schicho, Lattice polygons and the number $2i+7$, American Mathematical Monthly 116(2), pp. 151-165 (2009)
[KS]
N. Katz, P. Sarnak, Random matrices, Frobenius eigenvalues, and monodromy,
American Mathematical Society (1999)
[Kob]
N. Koblitz, Algebraic aspects of cryptography, Algorithms and Computation in Mathematics 3, Springer (1999)
[Koe]
R. Koelman, The number of moduli of families of curves on toric surfaces,
Ph.D. thesis, Katholieke Universiteit Nijmegen (1991)
[Mum]
D. Mumford, Theta characteristics of an algebraic curve, Ann. Sci. de l’É.N.S. 4(2), pp. 181-192 (1971)
[Poo1]
B. Poonen, Varieties without extra automorphisms. II. Hyperelliptic curves,
Math. Res. Lett. 7 (1), pp. 77-82 (2000)
[Poo2]
B. Poonen, Bertini theorems over finite fields, Ann. Math. 160, pp. 1099-1127 (2004)
[SZ]
J. Scholten, H. Zhu, Hyperelliptic curves in characteristic $2$, Int. Math. Res. Not. 2002(17), pp. 905-917 (2002)
[Zhu]
H. Zhu, Hyperelliptic curves over $\mathbf{F}_{2}$ of every $2$-rank without extra automorphisms,
Proc. Amer. Math. Soc. 134(2), 323-331 (2006)
E-mail: [email protected].
Address: Departement Wiskunde, KU Leuven, Celestijnenlaan 200B, 3001 Leuven, Belgium.
E-mail: [email protected].
Address: Mathematisch Instituut, Universiteit Leiden, Postbus 9512, 2300 RA Leiden, The Netherlands.
E-mail: [email protected].
Address: Mathematics Institute, University of Warwick, Coventry CV4 7AL, United Kingdom. |
Higher Kazhdan projections, $\ell_{2}$-Betti numbers and Baum-Connes conjectures
Kang Li
Piotr W. Nowak
Sanaz Pooya
Abstract
We introduce higher-dimensional analogs of Kazhdan projections in matrix algebras over group $C^{*}$-algebras
and Roe algebras. These projections are constructed in the framework of cohomology with coefficients in unitary representations and
in certain cases give rise to non-trivial $K$-theory classes.
We apply the higher Kazhdan projections to establish a relation between
$\ell_{2}$-Betti numbers of a group and surjectivity of different Baum-Connes type assembly maps.
Kazhdan projections are certain idempotents whose existence in the maximal group $C^{*}$-algebra $C^{*}_{\max}(G)$ of a group $G$ characterizes Kazhdan’s property $(T)$
for $G$. They have found an important application in higher index theory, as the non-zero $K$-theory class represented by a Kazhdan projection in $K_{0}(C^{*}_{\max}(G))$ obstructs the Dirac-dual Dirac method of proving the Baum-Connes conjecture.
This idea has been used effectively in the coarse setting [higson, higson-lafforgue-skandalis], where Kazhdan projections were used to construct
counterexamples to the coarse Baum-Connes conjecture and to the Baum-Connes conjecture with coefficients.
However, Kazhdan projections related to property $(T)$ are not applicable to the classical Baum-Connes conjecture, as the image of a Kazhdan projection
in the reduced group $C^{*}$-algebra vanishes for any infinite group.
The goal of this work is to introduce
higher-dimensional analogs of Kazhdan projections with the motivation of applying them
to the $K$-theory of group $C^{*}$-algebra and to higher index theory.
These higher Kazhdan projections are defined in the context of higher cohomology with coefficients in unitary representations
and they are elements of matrix algebras over group $C^{*}$-algebras. Their defining feature is that their
image in certain unitary representations is the orthogonal projection onto the harmonic $n$-cochains.
We show that the higher Kazhdan projections exist in many natural cases and that they are non-zero.
We then investigate the classes represented by higher Kazhdan projection in the $K$-theory of group $C^{*}$-algebras.
This is done by establishing a connection between the properties of the introduced projections and $\ell_{2}$-Betti numbers of the
group. In particular, we can conclude that the higher Kazhdan projections over the reduced group $C^{*}$-algebra can be non-zero
and can give rise to non-zero $K$-theory
classes in $K_{0}(C^{*}_{r}(G))$.
To present the connections between the existence of higher Kazhdan projections and $\ell_{2}$-invariants recall that given
a finitely generated group $G$ the subring $\Lambda^{G}\subseteq\mathbb{Q}$ is generated
from $\mathbb{Z}$ by adjoining inverses of all orders of finite subgroups, as in [luck, Theorem 0.3].
The cohomological Laplacian in degree $n$ can be represented by a matrix with coefficients in the group ring $\mathbb{R}G$, denoted $\Delta_{n}$.
In the case of the regular representation $\lambda$ and the reduced group $C^{*}$-algebra $C^{*}_{r}(G)$
the operator $\Delta_{n}\in\mathbb{M}_{k}(C^{*}_{r}(G))$ is the cohomological Laplacian in degree $n$ in the $\ell_{2}$-cohomology of $G$.
We will say that $\Delta_{n}$ has a spectral gap in a chosen completion of $\mathbb{M}_{n}(\mathbb{C}G)$ if its spectrum is contained in $\{0\}\cup[\epsilon,\infty)$ for some $\epsilon>0$.
Recall that a group $G$ is of type $F_{n}$ if it has an Eilenberg-MacLane space
with a finite $n$-skeleton. Denote by $k_{n}$ the number of simplices in the chosen model of the Eilenberg-MacLane space.
Let $\beta_{(2)}^{n}(G)$ denote the $n$-th $\ell_{2}$-Betti number of $G$.
The following is a consequence of the existence of spectral gap for the Laplacian and [luck].
Proposition 1.
Let $G$ be of type $F_{n+1}$. Assume $\Delta_{n}\in\mathbb{M}_{k_{n}}(C^{*}_{r}(G))$ has a spectral gap.
If the Baum-Connes assembly map $K_{0}^{G}(\underline{E}G)\to K_{0}(C^{*}_{r}(G))$
is surjective
then
$$\beta_{(2)}^{n}(G)\in\Lambda^{G}.$$
In particular, if $G$ is torsion-free then $\beta_{(2)}^{n}(G)\in\mathbb{Z}.$
The above theorem illustrates a possible strategy
for finding counterexamples to the Baum-Connes conjecture: a group $G$ of type $F_{n+1}$ such that
$\beta_{(2)}^{n}(G)\notin\Lambda_{G}$ (in particular, if $\beta_{(2)}^{n}(G)$ is irrational) with 0 isolated in the spectrum of $\Delta_{n}$ will not satisfy the
Baum-Connes conjecture; more precisely, the Baum-Connes assembly map for $G$ will not be surjective.
We refer to [gomez-aparicio-etal, valette]
for an overview of the Baum-Connes conjecture. It is worth noting that such a counterexample, with bounded orders of finite subgroups, would also not satisfy the Atiyah conjecture, see e.g. [luck-book].
Our primary application is the construction of classes in the $K$-theory of the
Roe algebra of a box space of a residually finite group $G$ and their relation to the Lück approximation theorem.
Let $\{N_{i}\}$ be a family of finite index normal subgroups of a residually finite group $G$, satisfying $\bigcap N_{i}=\{e\}$.
Let $\lambda_{i}$ denote the associated quasi-regular representation of $G$ on $\ell_{2}(G/N_{i})$ and consider the family $\mathcal{N}=\{\lambda,\lambda_{1},\lambda_{2},\dots\}$ of unitary representations of $G$.
The $C^{*}$-algebra $C^{*}_{\mathcal{N}}(G)$ is the completion of the group ring $\mathbb{C}G$
with respect to the norm induced by the family $\mathcal{N}$ (see
Section 1.1 for a precise definition). We will denote
by $\beta^{n}(G)=\dim_{\mathbb{C}}H^{n}(G,\mathbb{C})$ the standard $n$-th Betti number of $G$.
Theorem 2.
Let $G$ be an exact, residually finite group of type $F_{n+1}$ and let $\{N_{i}\}$ and $\mathcal{N}$ be as above.
Assume that $\Delta_{n}\in\mathbb{M}_{k_{n}}(C^{*}_{\mathcal{N}}(G))$ has a spectral gap
and that the coarse Baum-Connes assembly map
$KX_{0}(Y)\to K_{0}(C^{*}(Y))$
for the box space
$Y=\coprod G/N_{i}$ of $G$
is surjective. Then
$$\beta_{(2)}^{n}(N_{i})=\beta^{n}(N_{i})$$
for all but finitely many $i$.
Note that the conclusion of the above theorem in fact is a strenghtening of Lück’s approximation theorem [luck-gafa].
Indeed, the expression in the formula in the above theorem can be rewritten as
$$[G:N_{i}]\left(\beta_{(2)}^{n}(G)-\dfrac{\beta^{n}(N_{i})}{[G:N_{i}]}\right)=0,$$
Theorem 2 thus
forces a much stronger equality of the involved Betti numbers.
On the other hand, as shown in [luck-approx-survey, Theorem 5.1] there are examples where the speed of convergence
of $\beta^{n}(N_{i})/[G:N_{i}]$ to $\beta_{(2)}^{n}(G)$ can be as slow as needed.
Theorem 2 in particular provides new strategies for contradicting the coarse Baum-Connes conjecture.
It is an important open question wether there exist such counterexamples that do not contain expander graphs.
However, it is also natural to conjecture that Theorem 2 could also lead to new counterexamples
to the Baum-Connes conjecture with coefficients
by embedding a counterexample to the coarse Baum-Connes conjecture, obtained through Theorem 2,
isometrically into a finitely generated group, as in [osajda].
Such constructions could lead to new counterexample to the Baum-Connes conjecture with coefficients, by applying arguments similar to the
ones in [higson-lafforgue-skandalis].
We remark that algebraic conditions implying the existence of gaps in the spectrum of the operators arising from the
cochain complexes with coefficients in unitary representations
were recently provided in
[bader-nowak]. Those conditions involve writing the elements $(\Delta_{n}^{+}-\lambda\Delta_{n}^{+})\Delta_{n}^{+}$
and $(\Delta_{n}^{-}-\lambda\Delta_{n}^{-})\Delta_{n}^{-}$, where $\Delta_{n}^{+}=d_{n}^{*}d_{n}$ and $\Delta_{n}^{-}=d_{n-1}d_{n-1}^{*}$ are
matrices over $\mathbb{R}G$ and the two summands of the Laplacian, as sums of squares in $\mathbb{M}_{n}(\mathbb{R}G)$.
Such conditions can be verified using computational methods.
Acknowledgments
We are grateful to Dawid Kielak, Roman Sauer, Thomas Schick and Alain Valette for helpful comments.
This project has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement no. 677120-INDEX).
Contents
1 Higher Kazhdan projections
1.1 Kazhdan projections related to property $(T)$
1.2 Kazhdan projections in higher degree
1.3 $K$-theory classes - a special case
1.4 The Baum-Connes assembly map
1.5 Examples
1.5.1 Free groups
1.5.2 Kähler groups
1.5.3 Lattices in $\operatorname{PGL}_{n+1}(\mathbb{Q}_{p})$
2 The coarse Baum-Connes conjecture
2.1 Roe algebras, box spaces and Laplacians
2.1.1 Roe algebras
2.1.2 Laplacians on box spaces
2.2 The lifting map
2.3 Surjectivity of the coarse Baum-Connes assembly map
3 Final remarks
1 Higher Kazhdan projections
Recall that a group has type $F_{n}$
if it admits an
Eilenberg-MacLane space $K(G,1)$ with a finite $n$-skeleton. The condition $F_{1}$ is equivalent to $G$ being finitely generated, and $F_{2}$ is equivalent
to
$G$ having a finite presentation.
1.1 Kazhdan projections related to property $(T)$
We begin with revisiting the classical notion of Kazhdan projections in the context of property $(T)$.
Let $G$ be a finitely generated group with $S=S^{-1}$ a fixed generating set. The real (respectively, complex) group ring of $G$ will be denoted
by $\mathbb{R}G$ (respectively, $\mathbb{C}G$).
Consider a family $\mathcal{F}$ of unitary representations and let the associated group $C^{*}$-algebra be the completion
$$C^{*}_{\mathcal{F}}(G)=\overline{\mathbb{C}G}^{\|\cdot\|_{\mathcal{F}}},$$
where
$$\|f\|_{\mathcal{F}}=\sup_{\pi\in\mathcal{F}}\|\pi(f)\|.$$
We will always assume that $\mathcal{F}$ is a faithful family (i.e. for every $0\neq f\in\mathbb{C}G$ we have $\pi(f)\neq 0$ for some $\pi\in\mathcal{F}$) to ensure $\|\cdot\|_{\mathcal{F}}$ is a norm.
In particular, we obtain the maximal group $C^{*}$-algebra $C^{*}_{\max}(G)$ if $\mathcal{F}$ is the family of all unitary representations,
and the reduced group $C^{*}$-algebra $C^{*}_{r}(G)$ if $\mathcal{F}=\{\lambda\}$, where $\lambda$ is the left regular representation.
The following is a classical characterization of property $(T)$ for $G$ due to Akemann and Walter [akemann-walter]. We will use this characterization
as a definition and refer to [bekka-etal] for a comprehensive overview of property $(T)$.
Theorem 3 (Akemann-Walter [akemann-walter]).
The group $G$ has property $(T)$ if and only if there exists a projection $p_{0}=p_{0}^{*}=p_{0}^{2}\in C^{*}_{\max}(G)$
with the property that for every unitary representation $\pi$ of $G$
the image $\pi(p_{0})$ is the orthogonal projection onto the subspace of invariant vectors of $\pi$.
See e.g. [higson-roe-book, Section 3.7] and [drutu-nowak] for a broader discussion of Kazhdan projections.
We will now interpret Kazhdan projections in the setting of group cohomology.
Let $\pi$ be a unitary representation of $G$ on a Hilbert space $\mathcal{H}$.
Denote by $d_{0}$ the $\#S\times 1$ matrix
$$\left[\begin{array}[]{c}1-s_{1}\\
\vdots\\
1-s_{n}\end{array}\right],$$
where $s_{i}$ runs through the elements of $S$, with coefficients in $\mathbb{R}G$
and the Laplacian $\Delta_{0}\in\mathbb{R}G$ is
$$\Delta_{0}=d_{0}^{*}d_{0}=2\left(\#S-\sum_{s\in S}s\right)\ \ \ \ \in\mathbb{R%
}G.$$
For any unitary representation $\pi$ of $G$ on a Hilbert space $\mathcal{H}$ we have
$$\ker\pi(\Delta_{0})=\ker\pi(d_{0})=\mathcal{H}^{\pi},$$
where $\mathcal{H}^{\pi}\subseteq\mathcal{H}$ denotes the closed subspace of invariant vectors of $\pi$.
Note also that the same space can be interpreted as reduced111Recall that in a setting where the cochain spaces in a cochain complex
$C^{n}\ _{\overrightarrow{\ \ d_{n}\ \ }}C^{n+1}$ are equipped with a Banach space structure the corresponding reduced cohomology is defined as the quotient $\overline{H}^{n}=\ker d_{n}/\overline{\operatorname{im}d_{n-1}}$, where $\overline{\operatorname{im}d_{n-1}}$ is the closure of the image of the codifferential $d_{n-1}$. Cohomology is said to be reduced in degree $n$ if $\overline{H}^{n}=H^{n}$.
cohomology in degree 0:
$$\ker\pi(\Delta_{0})=\mathcal{H}^{\pi}\simeq\overline{H}^{0}(G,\pi).$$
The image $\pi(p_{0})$ of the Kazhdan projection is the orthogonal projection
$$C^{0}(G,\pi)\,_{\overrightarrow{\ \ \ \ \ \pi(p_{0})\ \ \ \ \ }}\,\ker\pi(%
\Delta_{0}).$$
The Kazhdan projection $p_{0}$ exists in $C^{*}_{\mathcal{F}}(G)$ if and only if $\pi(\Delta_{0})$ has a uniform spectral gap for all $\pi\in\mathcal{F}$.
The projection $p_{0}$ is non-zero if at least one $\pi\in\mathcal{F}$ has a non-zero invariant vector.
1.2 Kazhdan projections in higher degree
We will now discuss a generalization of Kazhdan projections in the setting of higher group cohomology.
Let $G$ be a group of type $F_{n+1}$ with a chosen model $X$ of $K(G,1)$ with finite $n+1$-skeleton.
As in [bader-nowak] we can consider the cochain complex for cohomology of $G$ with coefficients in $\pi$, where
$$C^{n}(G,\pi)=\,\left\{f:X^{(n)}\to\mathcal{H}\right\},$$
By $\mathbb{M}_{m\times m^{\prime}}(\mathbb{R}G)=\mathbb{M}_{m\times m^{\prime}}%
\otimes\mathbb{R}G$ we denote the space of
$m\times m^{\prime}$ matrices with coefficients in the group ring $\mathbb{R}G$.
When $m=m^{\prime}$ the resulting algebra is denoted by $\mathbb{M}_{m}(\mathbb{R}G)$.
The codifferentials can be represented by elements
$$d_{n}\in\mathbb{M}_{k_{n+1}\times k_{n}}(\mathbb{R}G),$$
where $k_{i}$ denotes the number of $i$-simplices in $X$,
in the sense that for every unitary representation $\pi$ the codifferential is given by the operator
$$\pi(d_{n}):C^{n}(G,\pi)\to C^{n+1}(G,\pi),$$
where $\pi$ is applied to an element of $\mathbb{M}_{m\times m^{\prime}}(C^{*}_{\mathcal{F}}(G))$ entry-wise;
see e.g. [bader-nowak-deformations, bader-nowak].
The Laplacian element in degree $n$ is defined to be
$$\Delta_{n}=d_{n}^{*}d_{n}+d_{n-1}d_{n-1}^{*}\in\mathbb{M}_{k_{n}}(\mathbb{R}G),$$
and we denote the two summands by
$$\Delta_{n}^{+}=d_{n}^{*}d_{n},\ \ \ \ \ \Delta_{n}^{-}=d_{n-1}d_{n-1}^{*},$$
as in [bader-nowak].
In particular, $\Delta_{n}$, $\Delta_{n}^{+}$, and $\Delta_{n}^{-}$ are elements of any completion of $\mathbb{M}_{k_{n}}(\mathbb{R}G)$ such as
$\mathbb{M}_{k_{n}}(C^{*}_{\max}(G))$ or $\mathbb{M}_{k_{n}}(C^{*}_{r}(G)).$
For a unitary representation $\pi$ of $G$ we can now consider the operator $\pi(\Delta_{n})$. The kernel of $\pi(\Delta_{n})$
is the space of harmonic $n$-cochains for $\pi$, and we have the standard Hodge-de Rham isomorphism
$$\ker\pi(\Delta_{n})\simeq\overline{H}^{n}(G,\pi),$$
between that kernel and the reduced cohomology of $G$ with coefficients in $\pi$ (see e.g. [bader-nowak, Section 3]).
Similarly, the kernel of the projection $\pi(\Delta_{n}^{+})$ are the $n$-cocycles for $\pi$, and the kernel of $\pi(\Delta_{n}^{-})$ are the
$n$-cycles.
Definition 4 (Higher Kazhdan projections).
A Kazhdan projection in degree $n$ is a projection $p_{n}=p_{n}^{*}=p_{n}^{2}\in\mathbb{M}_{k_{n}}(C^{*}_{\mathcal{F}}(G))$
such that
for every unitary representation $\pi\in\mathcal{F}$ the projection $\pi(p_{n})$ is the orthogonal projection $C^{n}(G,\pi)\to\ker\pi(\Delta_{n})$.
A partial Kazhdan projection in degree $n$ is a projection $p_{n}^{+},p_{n}^{-}\in\mathbb{M}_{k_{n}}(C^{*}_{\mathcal{F}}(G))$, such
that for every unitary representation $\pi\in\mathcal{F}$ the projection $\pi(p_{n}^{+})$ (respectively, $\pi(p_{n}^{-})$) is the
orthogonal projection onto the kernel of $\pi(d_{n})$ (respectively, onto the kernel of $\pi(d_{n-1}^{*})$).
In degree 0 and when $\mathcal{F}$ is the family of all unitary representations the projection $p_{0}\in C^{*}_{\max}(G)$ is the classical Kazhdan projection.
Clearly, $p_{n}$ is non-zero if and only if for at least one $\pi\in\mathcal{F}$ we have $\pi(p_{n})\neq 0$.
From now on we will shorten the subscript $k_{n}$, denoting the number of $n$-simplices in the chosen $K(G,1)$, to $k$, as the dimension will
be clear from the context.
Similarly as in the case of property $(T)$, higher Kazhdan projections exist in the presence of a spectral gap.
Proposition 5.
Assume that $\Delta_{n}\in\mathbb{M}_{k}(C^{*}_{\mathcal{F}}(G))$ (respectively, $\Delta_{n}^{+}$, $\Delta_{n}^{-}$) has a spectral gap.
Then the Kazdhan projection $p_{n}$ (respectively, the partial Kazhdan projections $p_{n}^{+}$, $p_{n}^{-}$) exist in the $C^{*}$-algebra
$\mathbb{M}_{k_{n}}(C^{*}_{\mathcal{F}}(G))$.
Proof.
By assumption, the spectrum of $\Delta_{n}$ satisfies
$$\sigma(\pi(\Delta_{n}))\subseteq\{0\}\cup[\epsilon,\infty)$$
for every unitary representation $\pi\in\mathcal{F}$ and some $\epsilon>0$.
In particular, $\sigma(\Delta_{n})\subseteq\{0\}\cup[\epsilon,\infty)$
when $\Delta_{n}$ is viewed as an element of $\mathbb{M}_{k}(C^{*}_{\mathcal{F}}(G))$.
Due to the presence of the spectral gap we can apply continuous functional calculus and conclude that the limit of the heat semigroup
$$\lim_{t\to\infty}e^{-t\Delta_{n}}$$
is the spectral projection $p_{n}\ \in\mathbb{M}_{k}(C^{*}_{\mathcal{F}}(G))$.
By construction, $p_{n}$ has the required properties.
The proof for the partial projections is analogous.
∎
Note that the fact that the spectrum of $\Delta_{n}\in\mathbb{M}_{k}(C^{*}_{\mathcal{F}}(G))$ has a gap implies that both $\Delta_{n}^{+}$ and $\Delta_{n}^{-}$
also have spectral gaps and in particular all three higher Kazhdan projections exist in
$\mathbb{M}_{k}(C^{*}_{\mathcal{F}}(G))$.
We have the following decomposition for higher Kazhdan projections.
Lemma 6.
Assume that $\Delta_{n}\in\mathbb{M}_{k}(C^{*}_{\mathcal{F}}(G))$ has a spectral gap. Then
$$p_{n}=p_{n}^{+}p_{n}^{-}.$$
Proof.
It suffices to notice that $\Delta_{n}^{+}$ and $\Delta_{n}^{-}$ commute in $\mathbb{M}_{k}(C^{*}_{\mathcal{F}}(G))$ and satisfy
$$\Delta_{n}^{+}\Delta_{n}^{-}=\Delta_{n}^{-}\Delta_{n}^{+}=0,$$
using the property that $d_{i}d_{i-1}=0$ for every $i\in\mathbb{N}$. Consequently, the corresponding spectral projections commute
and their product is the projection onto the intersections of their kernels, which is precisely the kernel of $\Delta_{n}$.
∎
1.3 $K$-theory classes - a special case
In this section we consider the case when $\mathcal{F}=\{\lambda\}$, where $\lambda$ is the left regular
representation of $G$ on $\ell_{2}(G)$ and $C^{*}_{\mathcal{F}}(G)=C^{*}_{r}(G)$ is the reduced group $C^{*}$-algebra
of $G$. We refer to [higson-roe-book, valette] for background on $K$-theory and assembly maps.
The projection $p_{n}\in\mathbb{M}_{k}(C^{*}_{r}(G))$ exists if $\lambda(\Delta_{n})$ has a spectral gap at $0$.
In this case the kernel of the Laplacian $\lambda(\Delta_{n})$ is isomorphic to the reduced $\ell_{2}$-cohomology of $G$:
$$\ker\lambda(\Delta_{n})\simeq\ell_{2}\overline{H}^{n}(G)\simeq\ker\lambda(d_{n%
})\big{/}\,\overline{\operatorname{im}\lambda(d_{n-1})},$$
and it is non-trivial if and only the $n$-th $\ell_{2}$-Betti number $\beta_{(2)}^{n}(G)$ is non-zero.
The projection $p_{n}\in\mathbb{M}_{k}(C^{*}_{r}(G))$ gives rise to a $K$-theory class $[p_{n}]\in K_{0}(C^{*}_{r}(G))$.
We are interested in the properties of the classes defined by the higher Kazhdan projections $[p_{n}]\in K_{0}(C^{*}_{r}(G))$.
Recall that the canonical trace $\tau_{G}$ on $C^{*}_{r}(G)$ is given by
$$\tau_{G}(\alpha)=\langle\alpha\delta_{e},\delta_{e}\rangle,$$
where $\delta_{e}\in\ell_{2}(G)$ is the Dirac delta at the identity element $e\in G$. For an element $\alpha\in\mathbb{C}G$ we have $\tau(\alpha)=\alpha(e)$.
The induced trace on $\mathbb{M}_{k}(C^{*}_{r}(G))$ is defined as
$$\tau_{k,G}\left((a_{ij})\right)=\sum_{i=1}^{k}\tau_{G}(a_{ii}),$$
and there is an induced map
$$\tau_{*}:K_{0}(C^{*}_{r}(G))\to\mathbb{R}$$
defined by $\tau_{*}([p])=\tau_{k,G}(p)$,
for a class represented by a projection $p\in\mathbb{M}_{k}(C^{*}_{r}(G))$.
Proposition 7.
Assume that $\Delta_{n}$ has a spectral gap in $\mathbb{M}_{k}(C^{*}_{r}(G))$.
Then
$$\tau_{*}([p_{n}])=\beta_{(2)}^{n}(G).$$
In particular, if $\beta_{(2)}^{n}(G)\neq 0$ then
$[p_{n}]\neq 0$ in $K_{0}(C^{*}_{r}(G))$.
Proof.
By definition, the $n$-th $\ell_{2}$ Betti number of $G$ is the von Neumann dimension (over the group von Neumann algebra $L(G)$)
of the
orthogonal projection onto the kernel of the Laplacian. Under our assumptions the projection
is now an element of the smaller algebra $\mathbb{M}_{k}(C^{*}_{r}(G))\subseteq\mathbb{M}_{k}(L(G))$ and
$$\beta_{(2)}^{n}(G)=\tau_{k,G}(p_{n}).$$
Therefore, the induced trace on the group $K_{0}(C^{*}_{r}(G))$ satisfies $\tau_{*}([p_{n}])=\beta_{(2)}^{n}(G)$,
as claimed
∎
A similar argument shows non-vanishing of the classes represented by the partial projections.
Proposition 8.
Assume that $0$ is isolated in the spectrum of $\Delta_{n}^{+}$ (respectively, $\Delta_{n}^{-}$) in $\mathbb{M}_{k}(C^{*}_{r}(G))$.
If $\beta_{(2)}^{n}(G)\neq 0$ then
$[p_{n}^{+}]\neq 0$ (respectively, $[p_{n}^{-}]\neq 0$) in $K_{0}(C^{*}_{r}(G))$.
Proof.
By assumption $p_{n}^{+}$ exists in $\mathbb{M}_{k}(C^{*}_{r}(G))$ and $p_{n}$ exists in $\mathbb{M}_{k}(L(G))$ and we will compare the
values of the trace over the latter algebra.
Since $\tau_{k,G}(p_{n})\neq 0$ and
$$\operatorname{im}(p_{n})\subseteq\operatorname{im}(p_{n}^{+})$$
by monotonicity of the von Neumann dimension over $L(G)$
we have
$$\tau_{k,G}(p_{n}^{+})\geq\tau_{k,G}(p_{n})>0.$$
Consequently $\tau_{*}([p_{n}^{+}])\neq 0$ in $K_{0}(C^{*}_{r}(G))$.
∎
1.4 The Baum-Connes assembly map
The Baum-Connes assembly map is the map
$$\mu_{i}:K_{i}^{G}(\underline{E}G)\to K_{i}(C^{*}_{r}(G)),$$
$i=0,1$.
The trace conjecture of Baum and Connes and the modified trace conjecture by Lück [luck] predict the range of the composition
$\tau_{*}\circ\mu_{0}$.
In particular, the modified trace conjecture, formulated and proved by Lück [luck],
states that $\tau_{*}\circ\mu_{i}$ takes values in a subring $\Lambda^{G}\subseteq\mathbb{Q}$,
generated from $\mathbb{Z}$ by adjoining the inverses of cardinalities of finite subgroups.
Theorem 9 (Lück [luck, Theorem 0.3]).
If the Baum-Connes assembly map is surjective then the composition $\tau_{*}\circ\mu_{0}$ takes values in the ring
$\Lambda^{G}\subseteq\mathbb{Q}$.
We can now state the relation between the Baum-Connes conjecture and $\ell_{2}$-Betti numbers.
Proposition 1.
Let $G$ be of type $F_{n+1}$. Assume that $0$ is isolated in the spectrum of $\Delta_{n}\in\mathbb{M}_{k}(C^{*}_{r}(G))$.
If the Baum-Connes assembly map $K_{0}^{G}(\underline{E}G)\to K_{0}(C^{*}_{r}(G))$
is surjective
then
$$\beta_{(2)}^{n}(G)\in\Lambda^{G}.$$
In particular, if $G$ is torsion-free then $\beta_{(2)}^{n}(G)\in\mathbb{Z}.$
Remark 10 (The Euler class).
Consider an infinite group $G$ such that $K(G,1)$ can be chosen to be a finite complex.
Assume that $\lambda(\Delta_{n})$ has a spectral gap in $\mathbb{M}_{k}(C^{*}_{r}(G))$ for every $n\geq 0$
(in particular, $G$ is non-amenable).
Then we can define the Euler class
$$\Xi(G)=\sum_{i\geq 0}(-1)^{i}[p_{i}]$$
in the $K$-theory $K_{0}(C^{*}_{r}(G))$.
We have that
$$\tau_{*}(\Xi(G))=\chi_{(2)}(G)$$
is the $\ell_{2}$-Euler characteristic of the chosen $K(G,1)$.
Atiyah’s $L_{2}$-index theorem associates the $L_{2}$-index, analogous to the Fredholm index but defined via the trace on the von Neumann algebra,
to a lift $\widetilde{D}$ of an elliptic operator $D$ on a compact manifold $M$ to the universal cover of $M$.
Loosely speaking, in the presence of the spectral gap the $L_{2}$-indices of certain operators appearing in the Atiyah $L_{2}$-index theorem
are in fact traces of their Baum-Connes indices.
1.5 Examples
In case of an infinite group $G$ the projection $\lambda(p_{0})\in C^{*}_{r}(G)$ is always 0 since the kernel of $\lambda(\Delta_{0})$
consists precisely of constant functions in $\ell_{2}(G)$.
We will now discuss certain cases in which higher Kazhdan projections exist.
We will use the following fact, which is a reformulation of [bader-nowak, Proposition 16, 2) and 3)]
Lemma 11.
Let $\pi$ be a unitary representation of $G$.
The Laplacian $\pi(\Delta_{n})$ has a spectral gap around $0$ if and only if the cohomology groups $H^{n}(G,\pi)$ and $H^{n+1}(G,\pi)$ are both reduced.
Sketch of proof.
Using the notation of [bader-nowak], given a cochain complex
$$\dots\to C^{n-1}\to C^{n}\to C^{n+1}\to\dots,$$
where the $C^{n}$ have Hilbert space structures and the codifferentials $d_{n}:C^{n}\to C^{n+1}$ are bounded operators,
we have $\Delta_{n}=d_{n}^{*}d_{n}+d_{n-1}d_{n-1}^{*}$.
We have an orthogonal decomposition
$$C^{n}=C_{n}^{+}\oplus\ker\Delta_{n}\oplus C_{n}^{-},$$
where $C_{n}^{-}=\ker d_{n}\cap\ker(d_{n-1}^{*})^{\perp}$ and $C_{n}^{+}=\ker d_{n-1}^{*}\cap(\ker d_{n}^{*})^{\perp}$.
Then the restritictions of $d_{n}^{*}d_{n}$ and of $d_{n-1}d_{n-1}^{*}$ are invertible on $C_{n}^{+}$ and $C_{n}^{-}$, respectively if and only if
the cohomology groups $H^{n}$ and $H^{n+1}$ are reduced [bader-nowak]. Clearly, the first condition is equivalent to the fact that $0$ is isolated in
the spectrum of the Laplacian $\Delta_{n}$.
∎
1.5.1 Free groups
Consider the free group $F_{n}$ on $n\geq 2$ generators. The standard Eilenberg-MacLane space $K(F_{n},1)$ is the wedge $\bigvee_{n}S^{1}$ of $n$ circles,
whose universal cover is the tree, the Cayley graph of $F_{n}$. Then for $F_{n}$ the projection $p_{1}\in\mathbb{M}_{k}(C^{*}_{r}(F_{n}))$ exists and gives rise to a non-zero class in
$K$-theory $C^{*}_{r}(F_{n})$.
Indeed, since for $n\geq 2$, the free group $F_{n}$ is non-amenable we have that the $\ell_{2}$-cohomology
is reduced in degree 1 (i.e., the range of the codifferential into the 1-cochains is closed).
Since $K(F_{n},1)$ is 1-dimensional, the cohomology in degree 2 is reduced as there are no 2-cochains.
We thus have $d_{1}=0$, $p_{1}^{+}=\operatorname{I}$ and $p_{1}=p_{1}^{-}$.
These facts together with lemma
11 imply that the cohomological Laplacian $\Delta_{1}$ in degree 1 has a spectral gap.
The $\ell_{2}$-Betti number of a free group $F_{n}$ on $n$ generators is
$$\beta_{(2)}^{1}(F_{n})=n-1.$$
By the previous discussion,
$$[p_{1}]\neq 0\ \ \ \text{ in }\ \ \ K_{0}(C^{*}_{r}(F_{n})).$$
Moreover, we can identify precisely the class of $p_{1}$ in the $K$-theory group.
Indeed, $K_{0}(C^{*}_{r}(F_{n}))$ is isomorphic to $\mathbb{Z}$ and generated by $1\in C^{*}_{r}(F_{n})$. Because of this the value of the trace
determines the class of $p_{1}$ to be
$$[p_{1}]=n-1\ \ \ \ \in\ \ \ \ \mathbb{Z}\simeq K_{0}(C^{*}_{r}(F_{n})).$$
The corresponding Euler class (see Remark 10) in $K_{0}(C^{*}_{r}(F_{n}))$ is given by
$$\Xi(F_{n})=[p_{0}]-[p_{1}]=-[p_{1}].$$
More generally, the argument used in the above example together with [kahlerbook, Corollary 4.8] gives
the following.
Corollary 12.
Let $G$ be a finitely presented group with infinitely many ends such that $H^{2}(G,\ell_{2}(G))$ is reduced. Then the projection
$p_{1}(G)$ exists in $\mathbb{M}_{k_{n}}(C^{*}_{r}(G))$ and
$$[p_{1}]\neq 0\ \ \ \text{ in }\ \ \ K_{0}(C^{*}_{r}(G)).$$
1.5.2 Kähler groups
As shown in [efremov], [gromov-shubin, Remark 2], see also [luck-book, Theorem 2.68] or [luck-survey, Section 8],
the property that the Laplacian has a spectral gap in the analytic setting (i.e. $L_{2}$-cohomology defined in terms differential forms)
is equivalent to the existence of the spectral gap in the combinatorial setting.
This is most often phrased in terms of Novikov-Shubin invariants associated to spectral density functions: spectral gap is equivalent to
the Novikov-Shubin invariant being infinite, which is the same as the property that the spectral density function is constant in the neighborhood
of 0. Then it is shown that the combinatorial and analytic spectral density functions are dilation equivalent, and in particular, equal in a neighborhood of 0.
Consider now a closed Kähler hyperbolic manifold $M$ and denote $G=\pi_{1}(M)$.
As shown by Gromov [gromov-Kahler], see also [luck-book, Section 11.2.3], the Laplacian acting on $L_{2}$-differential forms on the universal cover
$\widetilde{M}$ has a spectral gap in every dimension. Gromov also showed that the $\ell_{2}$-cohomology of $M$
vanishes except for the middle dimension.
Consequently, if $G$ is a fundamental group of such a Kähler manifold then the assumptions of
Proposition 7
are satisfied and
the projection $p_{n}$ exists in $\mathbb{M}_{k_{n}}(C^{*}_{r}(G))$ in every dimension.
Moreover, the Euler class
$$\Xi(G)=\sum_{i\geq 0}(-1)^{i}[p_{i}]\ \ \ \ \in K_{0}(C^{*}_{r}(G))$$
exist and its trace is the $\ell_{2}$-Euler characteristic of $G$.
1.5.3 Lattices in $\operatorname{PGL}_{n+1}(\mathbb{Q}_{p})$
As shown in [bader-nowak, Proposition 19], given $n\geq 2$, a sufficiently large prime $p$ and a lattice $\Gamma\subseteq\operatorname{PGL}_{n+1}(\mathbb{Q}_{p})$ we have that
$H^{n}(\Gamma,\pi)$ is reduced for every unitary representation $\pi$. This fact is equivalent to $\pi(\Delta_{n}^{-})$ having a spectral gap for every unitary
representation $\pi$ of $\Gamma$ and it is easy to check that such a spectral gap must be then uniform; i.e., $\Delta_{n}^{+}$ has a spectral
gap in $\mathbb{M}_{k}(C^{*}_{\max}(\Gamma))$.
At the same time there exists a finite index subgroup $\Gamma^{\prime}\subseteq\Gamma$
such that $H^{n}(\Gamma^{\prime},\mathbb{C})\simeq H^{n}(\Gamma,\ell_{2}(\Gamma/\Gamma^{%
\prime}))\neq 0$.
As a consequence we obtain the following.
Proposition 13.
Let $\Gamma\subseteq\operatorname{PGL}_{n+1}(\mathbb{Q}_{p})$ be a lattice. Then the partial Kazhdan projection $p_{n}^{+}$
exists in $\mathbb{M}_{k}(C^{*}_{\max}(\Gamma))$ and is non-zero.
2 The coarse Baum-Connes conjecture
We will now prove our main results and describe the $K$-theory classes induced by the higher Kazhdan projections in Roe algebras and their
applications to the coarse Baum-Connes conjecture of a box space
of a residually finite group. The arguments we will use were introduced in [higson, higson-lafforgue-skandalis] and developed further in
[willett-yu1]. We also refer to [willett-yu-book, Chapter 13] for more details.
2.1 Roe algebras, box spaces and Laplacians
2.1.1 Roe algebras
The Roe algebra $C^{*}(X)$ of a discrete, bounded geometry metric space $X$ is the completion in $B(\ell_{2}(X;\mathcal{H}_{0}))$
of the $*$-algebra $\mathbb{C}[X]$
of finite propagation operators, which are locally compact; i.e. for a discrete space the
matrix coefficients are compact operators $T_{x,y}\in\mathcal{K}(\mathcal{H}_{0})$ on a fixed, infinite-dimensional Hilbert space $\mathcal{H}_{0}$.
If $G$ acts on $X$ then the equivariant Roe algebra is defined to be the closure of the subalgebra
$\mathbb{C}[X]^{G}\subseteq\mathbb{C}[X]$ of equivariant finite propagation operators, i.e. satisfying
$T_{x,y}=T_{gx,gy}$ for any $g\in G$ and $x,y\in X$.
The uniform Roe algebra $C^{*}_{u}(X)$ of $X$ is the completion in $B(\ell_{2}(X))$ of the $*$-algebra $\mathbb{C}_{u}[X]$ of
finite propagation operators. The equivariant uniform Roe algebra $\mathbb{C}_{u}^{*}(X)^{G}$ is defined as before by considering $G$-invariant operators in
$\mathbb{C}_{u}[X]$ and taking the closure inside $C^{*}_{u}(X)$.
See Definitions 3.2 and 3.6 in [willett-yu1] for a detailed description of the Roe algebra and the equivariant Roe algebra.
2.1.2 Laplacians on box spaces
Consider a finitely generated residually finite group $G$.
Let $\{N_{i}\}$ be a family of finite index normal subgroups of $G$ satisfying $\bigcap N_{i}=\{e\}$. Consider the space
$$Y=\coprod G/N_{i},$$
viewed as a box space with $d(G/N_{i},G/N_{j})\geq 2^{i+j}$.
Let $\lambda_{i}$ be the quasi-regular representation of $G$ on $\ell_{2}\left(G/N_{i}\right)$ given by pulling back the regular representation
of $G/N_{i}$ on $\ell_{2}(G/N_{i})$ to $G$ via the quotient map $G\to G/N_{i}$ and denote by $\mathcal{N}$ the family
of representations
$$\mathcal{N}=\{\lambda,\lambda_{1},\lambda_{2},\dots\}.$$
The standing assumption in this section will be that the cohomological $n$-Laplacian $\Delta_{n}$ has a spectral gap in
$\mathbb{M}_{k}(C^{*}_{\mathcal{N}}(G))$.
The cohomological Laplacian element $\Delta_{n}\in\mathbb{M}_{k}(\mathbb{C}G)$ in degree $n$ maps to $\lambda_{i}(\Delta_{n})$ in
$\mathbb{M}_{k}(\lambda_{i}(C^{*}_{\mathcal{N}}(G)))$ and we will denote
$$\Delta_{n}^{i}=\lambda_{i}(\Delta_{n}).$$
Define the Laplace element of the box space to be
$$D_{n}=\oplus_{i}\ \Delta_{n}^{i}\in\bigoplus\mathbb{M}_{k}(\mathbb{C}_{u}[G/N_%
{i}])\subseteq\mathbb{M}_{n}(\mathbb{C}_{u}[Y]).$$
Note that $D_{n}$ has a spectral gap in $C^{*}(Y)$ if and only if all the $\Delta_{n}^{i}$ have a spectral gap in $C^{*}_{\mathcal{N}}(G)$.
2.2 The lifting map
From now on we will assume the group $G$ satisfies the operator norm localization property.
As shown by Sako [sako] the operator norm localization property is equivalent to exactness for bounded geometry
metric spaces, in particular for finitely generated groups, so the assumptions of Theorem 2 guarantee this property
for $G$.
Note that we will only be using a special case of the setup described in [willett-yu1], as our
box space consists of a family of finite quotients of a single group $G$, which naturally serves as a covering space for all of the finite quotients.
This setting is the same as the one in [higson], however we do follow the detailed version of the arguments as in [willett-yu1].
Willett and Yu in [willett-yu1, Lemma 3.8] define a lifting map, a $*$-homomorphism
$$\phi:\mathbb{C}[Y]\to\dfrac{\prod_{i}\mathbb{C}[G]^{N_{i}}}{\bigoplus_{i}%
\mathbb{C}[G]^{N_{i}}},$$
into the algebra of sequences of elements of the $N_{i}$-invariant subspaces of $\mathbb{C}[G]$ modulo the relation of
being equal at all but finitely many entries.
We have
$$\mathbb{M}_{k}\left(\dfrac{\prod_{i}\mathbb{C}[G]^{N_{i}}}{\bigoplus_{i}%
\mathbb{C}[G]^{N_{i}}}\right)=\dfrac{\prod_{i}\mathbb{M}_{k}\left(\mathbb{C}[G%
]^{N_{i}}\right)}{\bigoplus_{i}\mathbb{M}_{k}\left(\mathbb{C}[G]^{N_{i}}\right)}$$
and for every $k\in\mathbb{N}$ the map $\phi$ induces a $*$-homomorphism
$$\phi^{(k)}:\mathbb{M}_{k}(\mathbb{C}[Y])\to\mathbb{M}_{k}\left(\dfrac{\prod%
\mathbb{C}[G]^{N_{i}}}{\bigoplus\mathbb{C}[G]^{N_{i}}}\right),$$
by applying $\phi$ entrywise.
We recall that in [willett-yu1, Lemma 3.12] it was shown that if $G$ has the operator norm localization property then the map
$\phi$ extends to a map
$$\phi:C^{*}(Y)\to\dfrac{\prod_{i}C^{*}(G)^{N_{i}}}{\bigoplus_{i}C^{*}(G)^{N_{i}%
}},$$
and similarly induces the corresponding map $\phi^{(n)}$ on the $k\times k$ matrices over these algebras.
Here, $\dfrac{\prod_{i}C^{*}(G)^{N_{i}}}{\bigoplus_{i}C^{*}(G)^{N_{i}}}$ is the algebra of sequences of elements of $C^{*}(G)^{N_{i}}$ modulo the relation of
being asymptotically equal.
The same formula as the one for $\phi$ in [willett-yu1, Lemma 3.8] gives a uniform version of the lifting map,
$$\phi_{u}:\mathbb{C}_{u}[Y]\to\dfrac{\prod_{i}\mathbb{C}_{u}[G]^{N_{i}}}{%
\bigoplus_{i}\mathbb{C}_{u}[G]^{N_{i}}},$$
which is in the same way a $*$-homomorphism.
If $G$ has the operator norm localization property (see e.g. [chen-etal, sako]) then it also has the uniform version of that property, as proved by Sako [sako].
This implies that $\phi_{u}$ extends to a $*$-homomorphism
$$\phi_{u}:C^{*}_{u}(Y)\to\dfrac{\prod_{i}C^{*}_{u}(G)^{N_{i}}}{\bigoplus_{i}C^{%
*}_{u}(G)^{N_{i}}}.$$
See 13.3.11 and 13.3.12 in [willett-yu-book] for the detailed arguments, which apply verbatim here.
Similarly, $\phi_{u}^{(n)}$ in both cases induces a map on matrices over the respective algebras.
2.3 Surjectivity of the coarse Baum-Connes assembly map
We will now focus on the lift of the Laplacian $D_{n}$.
In order to do this we will use an argument based on Lemma 5.6 in [willett-yu1].
Lemma 14.
Let $G$ be an exact group and assume that $\Delta_{n}$ has a spectral gap in $C^{*}_{\mathcal{N}}(G)$.
Denote by
$$P_{n}=\bigoplus\lambda_{i}(p_{n})\ \ \ \ \in\mathbb{M}_{k}(C_{u}^{*}(Y))$$
the spectral projection associated to $D_{n}\in\mathbb{M}_{k}(C_{u}^{*}(Y))$.
Then
$$\phi_{u}^{(k)}(P_{n})=\prod_{i=1}^{\infty}\lambda(p_{n}),$$
where $\lambda(p_{n})$ is the projection onto the harmonic $n$-cochains in the $\ell_{2}$-cohomology of $G$.
Proof.
The projection $P_{n}$ is the spectral projection of $D_{n}$.
The Laplacian element $D_{n}$ defined above lifts to an element
$$\phi_{u}^{(k)}(D_{n})\in\mathbb{M}_{k}\left(\dfrac{\prod_{i=1}^{\infty}\mathbb%
{C}_{u}[G]^{N_{i}}}{\bigoplus_{i=1}^{\infty}\mathbb{C}_{u}[G]^{N_{i}}}\right),$$
whose spectrum is contained in $\{0\}\cup[\epsilon,\infty)$ for some $\varepsilon>0$. Indeed, since the same
is true for $D_{n}\in C^{*}_{u}(Y)$, by assumption, and spectral gaps are preserved by homomorphisms of unital $C^{*}$-algebras.
It is easy to see that from the definition of $\phi$ in the proof of [willett-yu1, Lemma 3.8] that the lift of $D_{n}$
is represented by the constant sequence
$$\phi_{u}^{(k)}(D_{n})=\prod_{i=1}^{\infty}\Delta_{n},$$
in the algebra
$$\dfrac{\prod_{i=1}^{\infty}\mathbb{M}_{k}(\mathbb{C}G)}{\bigoplus_{i=1}^{%
\infty}\mathbb{M}_{k}(\mathbb{C}G)}\subseteq\mathbb{M}_{k}\left(\dfrac{\prod_{%
i=1}^{\infty}\mathbb{C}_{u}[G]^{N_{i}}}{\bigoplus_{i=1}^{\infty}\mathbb{C}_{u}%
[G]^{N_{i}}}\right).$$
Indeed, this follows from the fact that the size of the support of $\Delta_{n}^{i}$ and the formula for the lift $\phi_{(u)}$ are both independent
of the particular quotient $G/N_{i}$ for $i$ sufficiently large.
Consider now the spectral projection $\widetilde{P}_{n}$ associated to $\phi_{u}^{(k)}(D_{n})$. Then
$$\widetilde{P}_{n}=\lim_{t\to\infty}e^{-t\phi_{u}^{(k)}(D_{n})}=\lim_{t\to%
\infty}\phi_{u}^{(k)}\left(e^{-tD_{n}}\right)=\phi_{u}^{(k)}(P_{n}).$$
Thus the associated spectral projection is of the form $\prod_{i=1}^{\infty}\lambda(p_{n})$, where
$$\lambda(p_{n})\in\mathbb{M}_{k}(C^{*}_{r}(G))$$
is the
projection onto the harmonic $n$-cochains in the $\ell_{2}$-cochains of $G$, as claimed.
∎
We now define the higher Kazhdan projection of the box space $Y$ to be $P_{n}\otimes q\in\mathbb{M}_{n}(C^{*}(Y))$, where $q$ is any rank one projection
on $\mathcal{H}_{0}$. We consider the associated $K$-theory class $[P_{n}\otimes q]\in K_{0}(C^{*}(Y))$. We will show that under certain conditions
this class is non-zero and does not lie in the image of the coarse Baum-Connes conjecture for $Y$.
The first thing to notice is that the lift of the projection $P_{n}\otimes q$ can be described explicitly as an element of $C^{*}(G)^{G}$.
Lemma 15.
Let $q\in\mathcal{K}(\mathcal{H}_{0})$. Then
$$\phi(P_{n}\otimes q)=\phi_{u}(P_{n})\otimes q.$$
Proof.
Let $\alpha_{i}$ be a sequence of finite propagation operators such that
$P_{n}=\lim_{i}\alpha_{i}$.
Then
$$\phi(\alpha_{i}\otimes q)=\phi_{u}(\alpha_{i})\otimes q,$$
by the definition of $\phi$ ([willett-yu1, Lemma 3.8]).
Passing to the limit and using continuity of $\phi$ we obtain the claim.
∎
Following [higson, willett-yu1] define the map $d_{*}:K_{0}(C^{*}(Y))\to\dfrac{\prod\mathbb{Z}}{\bigoplus\mathbb{Z}}$ by
taking the map
$$d:C^{*}(Y)\to\dfrac{\prod\mathcal{K}(\ell_{2}(G/N_{i};\mathcal{H}_{0}))}{%
\bigoplus\mathcal{K}(\ell_{2}(G/N_{i};\mathcal{H}_{0}))},$$
defined by
$$A\mapsto\prod_{i=1}^{\infty}Q_{i}AQ_{i},$$
where $Q_{i}:\ell_{2}(Y)\to\ell_{2}(G/N_{i})$ is the projection on the $i$-th box $G/N_{i}$,
and considering the induced map on $K$-theory
$$d_{*}:K_{0}(C^{*}(Y))\to K_{0}\left(\dfrac{\prod\mathcal{K}(\ell_{2}(G/N_{i};%
\mathcal{H}_{0}))}{\bigoplus\mathcal{K}(\ell_{2}(G/N_{i};\mathcal{H}_{0}))}%
\right)\simeq\dfrac{\prod\mathbb{Z}}{\bigoplus\mathbb{Z}}.$$
The above map $d_{*}$ can recognize when our $K$-theory class is non-zero. The map $d_{*}$ applied to
the projection $P_{n}$ gives the sequence of dimensions of images of $\lambda_{i}(p_{n})$.
On the other hand, consider the trace $\tau_{i}$ on $C^{*}(G)^{N_{i}}\simeq C^{*}_{r}(N_{i})\otimes\mathcal{K}$
by considering the tensor product of the standard trace on $C^{*}_{r}(N_{i})$ with the canonical trace on $\mathcal{K}$.
These $\tau_{i}$ induce the map
$$T:K_{0}\left(\dfrac{\prod_{i}C^{*}(G)^{N_{i}}}{\bigoplus_{i}C^{*}(G)^{N_{i}}}%
\right)\to\dfrac{\prod\mathbb{R}}{\bigoplus\mathbb{R}}.$$
The following diagram was analyzed in [willett-yu1].
{tikzcd}
& K_0( ∏iC*(G)Ni⊕iC*(G)Ni )\arrow[rd, ”T”]&
KX_0(Y) \arrow[ru, ”~μ_c”] \arrow[rd, ”μ_c”] &&∏R⊕R
&K_0(C^*(Y)) \arrow[uu, ”ϕ_*” ] \arrow[ru, ”d_*”]
The next lemma was formulated in [willett-yu1] for projections in $C^{*}(X)$, however the proof applies equally to projections in
$\mathbb{M}_{n}(C^{*}(X))$.
Lemma 16 (Lemma 6.5 in [willett-yu1]).
If $p$ is a projection in $\mathbb{M}_{n}(C^{*}(Y))$ such that the class $[p]\in K_{0}(C^{*}(Y))$ is in the image of the coarse assembly map
$\mu_{0}:KX_{0}(Y)\to K_{0}(C^{*}(Y))$, then
$$d_{*}([p])=T(\phi_{*}([p]))\ \ \ \ \ \ \text{ in }\dfrac{\prod\mathbb{R}}{%
\bigoplus\mathbb{R}}.$$
These maps take values in $\prod\mathbb{R}/\oplus\mathbb{R}$ viewed as an object in the category of abelian groups; that is, the equality of the two traces is an equality
of coordinates for all but finitely many $i$ (see Section 13.3 in [willett-yu-book]).
The next statement allows to show that when it is non-compact, the projection $P_{n}$ gives rise to a non-zero $K$-theory class in
$K_{0}(C^{*}(Y))$.
Proposition 17.
Assume that $H^{n}(G,\ell_{2}(G/N_{i}))\neq 0$ for infinitely many $i\in\mathbb{N}$.
Then $d_{*}[P_{n}]\neq 0$.
Proof.
We have
$$\displaystyle d_{*}(P_{n})$$
$$\displaystyle=\prod_{i=1}^{\infty}\dim\lambda_{i}(p_{n})=\prod_{i=1}^{\infty}%
\dim H^{n}(G;\lambda_{i}).$$
Indeed, since
$$\mathbb{M}_{n}(C^{*}(Y))=C^{*}(\mathbb{Z}_{n}\times Y),$$
the map
$$d^{(n)}:\mathbb{M}_{n}(C^{*}(X))\to\dfrac{\prod\mathbb{M}_{n}(\mathcal{K}(\ell%
_{2}(G/N_{i};\mathcal{H}_{0})))}{\bigoplus\mathbb{M}_{n}(\mathcal{K}(\ell_{2}(%
G/N_{i};\mathcal{H}_{0})))}.$$
can be rewritten as
$$d:C^{*}(X\times\mathbb{Z}_{n})\to\dfrac{\prod\mathcal{K}(\ell_{2}(G/N_{i}%
\times\mathbb{Z}_{n};\mathcal{H}_{0}))}{\bigoplus\mathcal{K}(\ell_{2}(G/N_{i}%
\times\mathbb{Z}_{n};\mathcal{H}_{0}))}.$$
By the assumption on cohomology of $G$ with coefficients in $\lambda_{i}$, the projection $p$ is non-compact in $C^{*}(Y\times\mathbb{Z}_{n})$
and as shown in the proof of Theorem 6.1 on page 1407 in [willett-yu1], we have $d_{*}[p]\neq 0$.
∎
The next lemma shows that the trace of the lift is naturally related to $\ell_{2}$-Betti numbers.
Lemma 18.
$T(\phi_{*}([P_{n}]))=\prod_{i=1}^{\infty}[G:N_{i}]\beta_{(2)}^{n}(G)=\prod_{i=%
1}^{\infty}\,\beta_{(2)}^{n}(N_{i})$.
Proof.
For a finite index subgroup $N\subseteq G$ the trace $\tau_{N}$ on $C^{*}(G)^{N}$ is defined as the tensor product of the canonical trace $\operatorname{tr}_{N}$ on $C^{*}_{r}(N)$
with the the canonical (unbounded trace) $\operatorname{Tr}$ on the compact operators, via the isomorphism
$$\psi_{N}:{C^{*}(G)^{N}\ }_{\overrightarrow{\ \ \ \simeq\ \ \ \ }}\ C^{*}_{r}(N%
)\otimes\mathcal{K}.$$
The isomorphism $\psi_{N}$ is defined by considering a fundamental domain $D\subset G$ for $N\subseteq G$
and for $A\in C^{*}(G)^{N}$ identifying
$$A\mapsto\sum_{g\in N}u_{g}\otimes A^{(g)},$$
where $A^{(g)}\in\mathcal{K}(\ell_{2}(D,\mathcal{H}_{0}))$ is defined by the formula
$$A^{(g)}_{x,y}=A_{x,gy},$$
for $x,y\in D$.
With this identification we observe that for an element $A\in C^{*}(G)^{G}$ of the form $A=\alpha\otimes q$, where $\alpha\in C^{*}_{r}(G)$
and $q$ is a rank 1 projection on $\mathcal{H}_{0}$, we have
$$\displaystyle\tau_{N}(A)$$
$$\displaystyle=\tau_{N}\left(\sum_{g\in N}u_{g}\otimes A^{(g)}\right)$$
$$\displaystyle=\sum_{g\in N}\operatorname{tr}(u_{g})\operatorname{Tr}(A^{(g)})$$
$$\displaystyle=\operatorname{Tr}(A^{(e)})$$
In our case $A^{(e)}$ is a $D\times D$ matrix defined by restricting $A$ to $D$. Therefore
$$\operatorname{Tr}(A^{(e)})=\sum_{x\in D}\alpha_{x,x}^{(e)}\cdot\operatorname{%
rank}(q)=[G:N]\alpha_{e,e}^{(e)}.$$
The same relation passes to traces of matrices over the respective $C^{*}$-algebras.
In the case of the projection $P_{n}$ these formulas yield
$$\tau_{N}(P_{n})=[G:N]\beta_{(2)}^{n}(G),$$
as claimed.
∎
Before we summarize this discussion we will observe one more fact that will allow to relate our results to Lück’s approximation theorem.
Recall that the Betti number of a group $G$ is the number $\beta^{n}(G)=\dim_{\mathbb{C}}H^{n}(G,\mathbb{C})$.
Lemma 19.
For a finite index subgroup $N\subseteq G$ we have
$$\dim_{\mathbb{C}}H^{n}(G,\ell_{2}(G/N))=\beta^{n}(N).$$
Proof.
Since $N$ is of finite index in $G$ we have
$$\operatorname{CoInd}_{N}^{G}\mathbb{C}=\operatorname{Ind}_{N}^{G}\mathbb{C}=%
\ell_{2}(G/N),$$
see e.g. [brown-book, Proposition 5.9].
Applying Shapiro’s lemma [brown-book, Proposition 6.2] we obtain
$$H^{n}(G,\ell_{2}(G/N))\simeq H^{n}(G,\operatorname{CoInd}_{N}^{G}\mathbb{C})%
\simeq H^{n}(N,\mathbb{C}).$$
∎
We are now in the position to formulate the main theorem of this section.
Theorem 2.
Let $G$ be an exact, residually finite group of type $F_{n+1}$ and let $\{N_{i}\}$ and $\mathcal{N}$ be as above.
Assume that $\Delta_{n}\in\mathbb{M}_{k_{n}}(C^{*}_{\mathcal{N}}(G))$ has a spectral gap
and that the coarse Baum-Connes assembly map
$KX_{0}(Y)\to K_{0}(C^{*}(Y))$
for the box space
$Y=\coprod G/N_{i}$ of $G$
is surjective. Then
$$\beta_{(2)}^{n}(N_{i})=\beta^{n}(N_{i})$$
for all but finitely many $i$.
Proof.
The claim follows from the explicit computation of the values of both traces in
proposition 17 and
lemmas 18 and 19.
∎
Note that the Theorem 2
provides a strenghtening of Lück’s approximation theorem [luck-gafa] in the case described by the above theorem.
Indeed, we can rewrite the conclusion of Theorem 2 as vanishing of
$$[G:N_{i}]\left(\beta_{(2)}^{n}(G)-\dfrac{\beta^{n}(N_{i})}{[G:N_{i}]}\right)=0$$
for all but finitely many $i$.
Compare this with [luck-approx-survey, Theorem 5.1], where examples with slow speed of convergence have been constructed.
The speed of convergence of Betti numbers of finite quotients to the $\ell_{2}$-Betti number of a residually finite group
was also studied in [clair-whyte], however the techniques used there are different. In our case both the assumptions and the conclusions are stronger.
3 Final remarks
Question 20.
Is there a spectral gap characterization of the existence of higher Kazhdan projections in $C^{*}_{\mathcal{N}}(G)$?
As mentioned earlier, Kazhdan’s property $(T)$ for $G$ is characterized by either the vanishing of first cohomology of $G$ with every unitary coefficients,
or equivalently, by the fact that the first cohomology of $G$ with any unitary coefficients is always reduced.
Related higher-dimensional generalizations of property $(T)$ were discussed in [bader-nowak-deformations], see also [dechiffre-etal].
As pointed out in [bader-nowak], the generalizations of these two conditions to higher degrees
are not equivalent.
The existence of higher Kazhdan projections is related to the property that cohomology is reduced and to the existence of gaps in the spectrum
of the Laplacian rather than to vanishing of cohomology. Indeed, in our reasoning it is crucial that
cohomology does not vanish.
It would be interesting to determine if the existence of higher Kazhdan projections can be viewed as a higher-dimensional rigidity property.
Remark 21 (Higher Kazhdan projections and $K$-amenability).
Clearly, if for a particular $G$ we have at the same time that $\beta_{(2)}^{n}(G)=0$ and
$\pi(\Delta_{n})$ has a nontrivial kernel and a spectral gap for some $\pi\neq\lambda$,
then $G$ cannot be amenable.
Consider the map
$$K_{*}(C^{*}_{\max}(G))\to K_{*}(C^{*}_{r}(G)).$$
If $[p_{n}]\neq 0$ in the former, but $[p_{n}^{r}]=0$ in the latter then the map above cannot be an isomorphism.
In other words, if $\beta_{(2)}^{n}(G)=0$ and we could ensure condition that the class of $[p_{n}]$ is not 0, then the group
$G$ would not be $K$-amenable, as this last condition forces the above map in $K$-theory to be an isomorphism.
Remark 22 (Ghost projections).
We can extend the notion of a ghost operator to matrix algebras over the Roe algebra by defining an element $T\in\mathbb{M}_{n}(C^{*}(Y))$ to
be a ghost if $T_{i,j}\in C^{*}(Y)$ is a ghost for every $1\leq i,j\leq n$. It can be shown that the kernel of the lifting map $\phi^{(n)}$ consists precisely of
ghost operators. See [willett-yu-book, Corollary 13.3.14]. This observation provides a new cohomological tool to construct ghost projections
in the case when an $\ell_{2}$-Betti number of $G$ vanishes. The problem of constructing new examples
of ghost projection was posed by Willet and Yu [willett-yu1].
References
K. Li, P. W. Nowak and S. Pooya, Institute of Mathematics of the Polish Academy of Sciences, Śniadeckich 8, 00-656 Warszawa, Poland
[email protected], [email protected], [email protected] |
Metal-poor hypervelocity star candidates from the Sloan Digital Sky Survey
Yinbi Li11affiliation: Key Laboratory of Optical Astronomy, National
Astronomical Observatories, Chinese Academy of Sciences, Beijing
100012, China. 22affiliation: Graduate University of
Chinese Academy of Sciences, 19A Yuquan Road, Beijing 100049, China , Ali Luo11affiliation: Key Laboratory of Optical Astronomy, National
Astronomical Observatories, Chinese Academy of Sciences, Beijing
100012, China. , Gang Zhao11affiliation: Key Laboratory of Optical Astronomy, National
Astronomical Observatories, Chinese Academy of Sciences, Beijing
100012, China. ,
Youjun Lu11affiliation: Key Laboratory of Optical Astronomy, National
Astronomical Observatories, Chinese Academy of Sciences, Beijing
100012, China. , Juanjuan Ren11affiliation: Key Laboratory of Optical Astronomy, National
Astronomical Observatories, Chinese Academy of Sciences, Beijing
100012, China. 22affiliation: Graduate University of
Chinese Academy of Sciences, 19A Yuquan Road, Beijing 100049, China , Fang Zuo11affiliation: Key Laboratory of Optical Astronomy, National
Astronomical Observatories, Chinese Academy of Sciences, Beijing
100012, China.
[email protected]
Abstract
Hypervelocity stars are believed to be ejected out from the Galactic center through dynamical
interactions of (binary) stars with the central massive black hole(s). In this letter, we report
13 metal-poor F-type hypervelocity star candidates selected from 370,000 stars of the
data release 7 of the Sloan Digital Sky Survey. With a detailed analysis of the kinematics of
these stars, we find that seven of them were likely ejected from the Galactic center (GC)
or the Galactic disk, four neither originated from the GC nor the Galactic disk, and the
other two were possibly ejected from either the Galactic disk or other regions. Those candidates
which unlikely originated from the GC or the Galactic disk, may be explained by other mechanisms, like the
tidal disruption of the Milky Way’s dwarf galaxies in the Galactic potential, or the gravitational
interactions with a massive black hole at the center of M31 or M32.
galaxies: kinematics and dynamics — Galaxy:
structure — stars: kinematics and dynamics — stars: late-type — galaxies: individual
(M31, M32 and Leo A)
1 INTRODUCTION
The hypervelocity stars (HVSs), recently discovered in the Galactic halo (Brown et al., 2005; Hirsch et al., 2005; Edelmann et al., 2005),
are moving so fast that they may escape from the Galaxy. A natural explanation of these HVSs is that
they are ejected out from the Galactic center (GC) by interactions of stars with the massive black
hole (MBH) or the hypothetical binary MBHs as predicted by Hills (1988) and Yu & Tremaine (2003). The ejection mechanisms of HVSs can
be divided into three categories: tidal breakup of binary stars in the vicinity of a single MBH
(Hills, 1988; Yu & Tremaine, 2003; Bromley et al., 2006), and the binary stars are probably injected into the vicinity of the MBH from
the young stellar disk in the GC (e.g., Lu et al., 2010; Zhang et al., 2010) or from the Galactic bulge (Perets, 2009a, b);
single star encounters with a binary MBH (Yu & Tremaine, 2003; Sesana et al., 2006, 2007; Merritt, 2006);
or single star encounters with a cluster of stellar mass black holes around the MBH (O’Leary & Loeb, 2008).
More than 16 HVSs have been reported in the literature (Brown et al., 2009)(See also: Tillich et al., 2009; Kollmeier et al., 2009; Tillich et al., 2010),
most of them are 3–4 M${}_{\sun}$ late B-type stars. Assuming a Salpeter initial mass function (IMF),
the expected solar mass HVSs are about 10 times more abundant than 3–4 M${}_{\sun}$ HVSs (Brown et al., 2009). Kollmeier et al. (2009)
systematically searched for such low mass HVSs in about 290,000 spectra of the Sloan Digital Sky Survey (SDSS), however,
they found only 6 metal-poor stars that can be possibly taken as HVS candidates, which might suggest that the IMF of the parent
population of these HVSs is top heavy. A top heavy IMF of the HVS parent population is possibly consistent with the disk
origination (Bartko et al., 2010; Lu et al., 2010; Kollmeier et al., 2010). In order to distinguish the ejection mechanisms of HVSs and put constraints
on the origin of the parent population of HVSs, it is quite necessary to search for the low-mass HVSs.
In this letter, we aim to find F/G type low mass HVS candidates from the data release 7 (DR7) of SDSS, which provides a
large catalog of stars with precise multi-color photometry and medium resolution ($R\sim 1800$)spectra (York et al., 2000).
We report the finding of 13 old metal-poor F-type HVS candidates, and also discuss their possible origins by detailed
analysis of their kinematics. The letter is organized as follows. In Section 2, we introduce the searching process
of our low-mass HVS candidates. Section 3 discusses their possible
ejection mechanisms through kinematics. Finally, a brief conclusion is given in Section 4.
2 TARGET SELECTION
We analyze over 370,000 stars in the SDSS DR7 with their five-band photometry $ugriz$ and spectra, which are flux- and wavelength-calibrated in $3800-9200$ Å and are reduced by the automated SEGUE Stellar Parameter Pipeline (Lee et al., 2008) to produce reliable heliocentric radial velocities RV${}_{\sun}$ and atmospheric parameters.
2.1 High Velocity Objects
In order to find out the F/G type HVS candidates, we first select the high-velocity
objects from the Galactic radial velocity distribution of F/G type main sequence samples
with the following methods.
1.
Select F/G type main sequence samples with the five photometric criteria and log(g) $>$ 3.0 in
section 2.3.1 of Ivezić et al. (2008): [130775].
2.
Find objects at the high velocity end, which significantly deviate from the best fit Gaussian distribution
of line-of-sight velocities in the Galactic rest frame $V_{\rm rf}$: [369].
The numbers in brackets indicate the number of stars left after each selection step. The $V_{\rm rf}$
distribution of our 130775 samples, shown in the top panel of Figure 1, can be well fitted by
the two Gaussian distribution (red solid curve). The two Gaussian components have mean velocities
(and scatters) of 0 km s${}^{-1}$ (92 km s${}^{-1}$) and 124 km s${}^{-1}$ (55 km s${}^{-1}$) respectively.
Approximately 82% of the samples belong to the low velocity component, and 18% of them belong
to the high velocity component (two dashed green curves). These two components probably correspond to
the Galactic halo population (0 $\pm$ 100 km s${}^{-1}$) and thick disk population
(90 $\pm$ 45 km s${}^{-1}$; Williams et al., 2011) respectively. The normalized residual of the observation
from this two Gaussian distribution is shown in the bottom panel of the Figure 1, and at its
high velocity tail, high-significance deviations are found. So, we choose the high velocity objects with
$V_{\rm rf}>309$ km s${}^{-1}$ and $V_{\rm rf}<-285$ km s${}^{-1}$, which have normalized
residuals larger than 1 and are denoted by two red arrows.
2.2 HVS Candidates
To determine which high velocity objects are unbound to the Galaxy, we first obtain the phase space coordinates for them. The Galactic cartesian coordinate system adopted here is centered on the GC: the X axis points from the Sun to the GC with the Sun at $x$ = $-$8 kpc; the Y axis points in the direction of Galactic rotation; the Z axis points towards the Northern Galactic Pole. Assuming the motion of the local standard of rest (LSR) is 220 km s${}^{-1}$, and the velocity of the Sun with respect to the LSR is (11.1 km s${}^{-1}$,
12.24 km s${}^{-1}$, 7.25 km s${}^{-1}$, Schönrich et al., 2010), the phase-space coordinates to the GC for these high velocity objects can be derived by equatorial coordinates ($ra$, $dec$), galactic coordinates ($l$, $b$), $RV_{\sun}$, two components of the proper motion $\mu_{\alpha}\cos(\delta)$, $\mu_{\delta}$, and the heliocentric distances ($D_{\sun}$). Here, $D_{\sun}$ is equal to $\frac{1}{100}\times 10^{0.2~{}(r-M_{r})}$, where r is the deredden r-band apparent magnitude given by the DR7 and $M_{r}$ is the r-band absolute magnitude calculated by the photometric parallax relation of Ivezić et al. (2008).
Then, we adopt two different Galactic potential models, i.e., a spherically
symmetric model (Xue et al., 2008, hereafter Xue08) and a triaxial model (Gnedin et al., 2005, hereafter Gnedin05) to estimate the
escape velocities ($V_{esc}$). There are totally 13 objects (mpHVS1–mpHVS13, “mp” means metal-poor) as listed in Table 1, which are unbound to the Galaxy in the Xue08 model, but only three (mpHVS1–mpHVS3) are unbound in the Gnedin05 model. The rest stars in the 369 high velocity objects are all bound to the Galaxy in both models. We cross-check the $V_{esc}$ for these 13 unbound candidates using other Galactic potential models in the literature, e.g., a spherically symmetric model from Kenyon et al. (2008), two axisymmetric models from Koposov et al. (2010) and Paczynski (1990) with the same definition of unbound stars as in Kenyon et al. (2008), and find that the $V_{esc}$ from these divergent models are between the values obtained from the Gnedin05 and Xue08 models, and mpHVS1, mpHVS2 and mpHVS3 are unbound to the Galaxy in all models. Note here, a small fraction of the bound stars at the 2-$\sigma$ level could be actually unbound due to the velocity measurement error, but the possibilities are certainly much less than those of unbound candidates. The basic information of these 13 candidates are listed in Table 1– 3. Table 1 summarizes their basic parameters, i.e., the name and its notation, $\mu_{\alpha}\cos(\delta)$ and $\mu_{\delta}$, $RV_{\sun}$, $V_{\rm rf}$, g-band apparent magnitude g, and stellar atmospheric parameters $T_{\rm eff}$, log($g$), and [Fe/H]. Table 2 lists their 6D phase-space coordinates and heliocentric distances $D_{\sun}$. Except mpHVS13, the estimates of $D_{\sun}$ and velocities of other candidates are all valid, because their spectroscopic metallicities [Fe/H] are in the valid range of that given by Ivezić et al. (2008). Table 3 compares their Galactic total velocities $V_{\rm G}$ with $V_{\rm esc}$ of the Gnedin05 and Xue08 models at their Galactocentric distances $D_{\rm G}$, from which we can see whether they are unbound in a certain potential model. In addition, Kollmeier et al. (2009) found 6 analogous metal-poor HVS candidates, so we also check whether they are in our high velocity objects. Consequently, only SDSS J074557.31+181246.7 is selected out and belongs to our bound objects, while the other five candidates are not in our high velocity objects mainly because they are inconsistent with the selection criterion 1 in Section 2.1.
3 POSSIBLE ORIGINS OF 13 mpHVS CANDIDATES
To investigate the origins of these 13 mpHVS candidates, it is essential to study their kinematics. By varying the 3D velocities and positions within their measurement errors, we calculate 10000 tracing-back trajectories and the 1 $\sigma$ level intersection places with the Galactic disk for each of mpHVS1–3 with the Gnedin05 and Xue08 models, and those for each of mpHVS4–13 with the Xue08 model, because they are bound in the Gnedin05 model, which are shown in Figure 2. From the left panel of Figure 2, we can see:
(1) the trajectories of mpHVS1 and mpHVS2 do not intersect with the Galactic disk.
(2) the trajectories of mpHVS3 intersect with the Galactic disk about 12 and 15 Myr ago for the Gnedin05 and Xue08 models respectively, and the 1 $\sigma$ intersection places are respectively shown by red and green ellipses.
In the right panel:
(1) the trajectories of mpHVS4 and mpHVS13 do not intersect with the Galactic disk.
(2) the trajectories of mpHVS5, mpHVS7, mpHVS8, mpHVS10, mpHVS11 and mpHVS12 do intersect with the Galactic disk about 10 to 20 Myr ago, and the 1 $\sigma$ intersection regions are depicted by 6 ellipses with different colors.
(3) for mpHVS6 and mpHVS9, most trajectories do intersect with the Galactic disk, but most of the intersection regions are far away from the GC even to hundreds of kpc, so we just describe part of the trajectories for the two candidates, which are calculated with their present 6D phase-space coordinates, and do not show their intersection regions in the right panel.
The trajectories of mpHVS3 are consistent with being ejected out from the GC or the Galactic disk, while those of mpHVS5, mpHVS7, mpHVS8, mpHVS10, mpHVS11 and mpHVS12 are consistent with originating from the Galactic disk (see Figure 2). Therefore, a supernova explosion in a massive tight binary system (Blaauw, 1961) or dynamical interactions in dense stellar clusters (Poveda et al., 1967; Leonard, 1991; Gvaramadze et al., 2009; Gvaramadze & Gualandris, 2011) are possible originating mechanisms for these seven candidates, and dynamical interactions between (binary) stars and MBH(s) in the GC are also possible ejection mechanisms for mpHVS3. For mpHVS1, mpHVS2, mpHVS4 and mpHVS13, their trajectories are incompatible with the GC origin or the Galactic disk origin mechanisms (see Figure 2). However, their origins may be interpreted as following.
Sherwin et al. (2008) proposed that there are a large number of low mass ($\approx$ 1 M${}_{\sun}$) HVSs ejected by the MBH in the center of M31, and a fraction of them would be moving towards the Milky Way on the solar system side, with large Galactocentric radial approach velocities ($\sim-500$ km s${}^{-1}$) exceeding the local escape velocity. The Galactocentric radial velocities of mpHVS1, mpHVS2, mpHVS4 and mpHVS13 are $-476\pm$85 km s${}^{-1}$, $-597\pm$177 km s${}^{-1}$, $413\pm$96 km s${}^{-1}$ and $265\pm$126 km s${}^{-1}$ respectively, which means the approaching velocities of mpHVS1 and mpHVS2 are consistent well with the predicted $-500$ km s${}^{-1}$, but it seems unlikely that mpHVS4 and mpHVS13 originated from the central MBH of M31. The mpHVS1 and mpHVS2 are low mass stars ($\approx$ 1 M${}_{\sun}$) and on the side of the Sun (see Fig. 2), considering of these factors, they seem to have been ejected out from the central MBH of M31. The ejection rate of HVSs from M32 is even higher than that from M31 (Lu et al., 2007), and thus M32 is also a possible origin of the two mpHVS candidates.
Tidal disruptions of dwarf galaxies in the Milky Way can also produce high velocity stars as proposed by Abadi et al. (2009). And they suggested that the current detected HVSs surrounding the constellation Leo ($l\approx$ 230${}^{\circ}$, $b\approx$ 60${}^{\circ}$) may be a stream of stars from a dwarf galaxy that was recently tidally disrupted (but see a different explanation of this in Lu et al. (2010)), and Teyssier et al. (2009) further proposed that such tidal disruption process of the massive satellites or satellites on eccentric orbits can also generate a population of old isolated high speed “escaped” (unbound) or “wandering” (bound) stars 111The “wandering” stars may possibly be related to the bound stars in the 369 high velocity objects.. If mpHVS1, mpHVS2, mpHVS4 and mpHVS13 are stars in tidal stripping streams, they may originate from other star streams, as their loci ($l$: 58${}^{\circ}$$\sim$69${}^{\circ}$; $b$: -45${}^{\circ}$$\sim$33${}^{\circ}$) are far away from the Leo area. If they are the isolated old “escaped” tidal stripping stars, we can roughly estimate their stripping times since they only pass though our galaxy once and never come back. Their estimated tidal stripping occurred about $10^{8}$ yr ago, assuming the dwarfs lie at the virial radius. Such a short stripping time suggests that their parent dwarf galaxies or streams may be still detectable along their trajectories. Therefore, further investigations are needed to find such dwarf galaxies, but the current comparatively large data errors impede such studies.
From the results of numerical experiments of mpHVS6 and mpHVS9, they are possibly ejected out from either the Galactic disk by corresponding mechanisms described above, or other places. The Galactocentric radial velocities of mpHVS6 and mpHVS9 are $364\pm$223 km s${}^{-1}$ and $-273\pm$262 km s${}^{-1}$ respectively, so they are unlikely to be ejected by the central MBH of M31. The tidal interactions between satellite galaxies and our galaxy may be able to explain their origins, but they were surely not ejected out from star streams in the direction of constellation Leo because their loci ($l$:56${}^{\circ}$$\sim$185${}^{\circ}$; $b$: 26${}^{\circ}$$\sim$31${}^{\circ}$) are also far away from that region. More exact conclusions about the origins of these two candidates mainly depend on more accurate data measurements in the future.
Note here, the large velocities of these mpHVS candidates could partly be due to orbital motions,
if they were in compact binaries. Below, we give a simple estimation of the effect of binary orbital velocities
on the observed heliocentric radial velocities and the Galactic total velocities. If each of these mpHVS candidates
were in binaries, its companion could be a low mass main-sequence star, a neutron star or a black hole.
a.
a low mass ($\leq$ 1 M${}_{\sun}$) main-sequence companion: Rastegaev (2010)
calculated the distribution of orbital periods for 60 detected metal-poor F,G and early K type binaries, and found the
minimum orbital period was larger than 10 days, and correspondingly the minimum semi-major axis $\geq$ 0.1 AU.
If the mass and semi-major axis of the companion are 1 M${}_{\sun}$ and 0.1 AU respectively, the effects due to binary orbital
motion could be most significant. The average effect of the projected
orbital velocities on the line-of-sight is thus about 27 km s${}^{-1}$, and the maximum is about 67 km s${}^{-1}$,
which introduce only approximate 3 $\sim$ 7 km s${}^{-1}$ and 9 $\sim$ 31 km s${}^{-1}$ effect to the 3D velocities
of our candidates. In addition, if these candidates are really in such binaries, their heliocentric distances and
Galactic total velocities would be larger than our estimations, and the escape velocities would
be correspondingly smaller, so these binaries should more likely be able to escape from our Galaxy.
b.
a neutron star companion: The maximum theoretical mass of a neutron star is $\sim$ 3 M${}_{\sun}$ (Kiziltan et al., 2010),
and if its semi-major is 0.1 AU, the average line-of-sight projected orbital velocity is about 57 km s${}^{-1}$, and the
maximum value is about 141 km s${}^{-1}$. In this cases, approximate 11 $\sim$ 24 km s${}^{-1}$ and 49 $\sim$ 93 km s${}^{-1}$
effects on the total space velocities of these mpHVS candidates are introduced.
c.
a black hole companion: If the mass and semi-major axis of a black hole are 10 M${}_{\sun}$ and 0.1 AU respectively, the average
line-of-sight projected orbital velocity is about 115 km s${}^{-1}$, and the maximum velocity is about 284 km s${}^{-1}$,
which introduce approximate 35 $\sim$ 69 km s${}^{-1}$ and 150 $\sim$ 228 km s${}^{-1}$ effects on the total space velocities
of these candidates. In this cases, the binary effect appears significant, so multiple observations are important to investigate whether these HVS candidates are in binaries with black hole companions.
To close the discussion, we note here that only mpHVS1 is unbound with high statistical significance, although all 13 HVS candidates are unbound in some Galactic potential models. The proper motion estimates adopted in our analysis (see Table 1), given by the SDSS, may suffer some uncertainties as pointed out by Dong et al. (2011), which may effect our estimations on the 3D velocities of those mpHVS candidates. Applying the model of Dong et al. (2011) (see equation 6), we re-estimate statistically the true proper motions of those 13 HVS candidates at 3-$\sigma$ significance level. Considering of this correction, we find that the highest velocity candidate, mpHVS1, is still unbound to the Galaxy in both potential models though with a lower 3D velocity (877 km s${}^{-1}$). However, mpHVS4–13 of the other 12 candidates are bound to the Galaxy if adopting the Xue08 potential or mpHVS2–13 are bound if adopting the Gnedin05 potential. So, those mpHVS candidates, except mpHVS1, could be bound halo stars. If they were, it is possible to use them to set constraints on the Galactic potential (W.Brown 2011, private communication).
4 CONCLUSION
In this letter, we report 13 metal-poor F-type HVS candidates from over
370,000 stars in the SDSS DR7. Through the kinematic analysis, we find that seven of them are likely to have been
ejected from the GC or the Galactic disk. Two of them were possibly ejected out from either the Galactic disk
or other places. Meanwhile, the other four candidates were unlikely to originate from the GC or the
Galactic disk. Those candidates impossibly ejected from the GC or the Galactic disk may be explained by other
mechanisms, i.e., being ejected from the tidal break of the Milky Way’s dwarf galaxies (Abadi et al., 2009; Teyssier et al., 2009),
or from the center of M31 or M32 by the interactions of stars with the MBH (Sherwin et al., 2008). In order to understand
the origins of these mpHVS candidates well and solve the binary problem, second-epoch observations in future are needed.
We thank Yingchun Wei for useful discussions, Jeffrey A. Munn for a comment on the proper
motion provided by the SDSS, Željko Ivezić for a comment on evaluating the absolute
magnitude. This work makes use of data products from the SDSS, which is managed by the Astrophysical
Research Consortium for the Participating Institution. The work was funded by the National Science
Foundation of China (NSFC) under grant Nos. 10821061, 10973021, 10973017 and 11103030.
References
Abadi et al. (2009)
Abadi, M. G., Navarro, J. F., & Steinmetz, M. 2009, ApJ, 691, L63
Bartko et al. (2010)
Bartko, H., et al. 2010, ApJ, 708, 834
Blaauw (1961)
Blaauw, A. 1961, Bull. Astron. Inst. Netherlands, 15, 265
Bromley et al. (2006)
Bromley, B. C., Kenyon, S. J., Geller, M. J., Barcikowski, E., Brown, W. R., & Kurtz, M. J. 2006, ApJ, 653, 1194
Brown et al. (2009)
Brown, W. R., Geller, M. J., Kenyon, S. J., & Bromley, B. C. 2009, ApJ, 690, 1639
Brown et al. (2005)
Brown, W. R., Geller, M. J., Kenyon, S. J., & Kurtz, M. J. 2005, ApJ, 622, L33
Dong et al. (2011)
Dong, R. B., et al. 2011, AJ, 142, 116D
Edelmann et al. (2005)
Edelmann, H., Napiwotzki, R., Heber, U., Christlieb, N., & Reimers, D. 2005, ApJ, 634, L181
Gnedin et al. (2005)
Gnedin, O. Y., Gould, A., Miralda-Escudé, J., & Zentner, A. R. 2005, ApJ, 634, 344
Gould & Kollmeier (2004)
Gould, A., & Kollmeier, J. A. 2004, ApJS, 152, 103
Gvaramadze et al. (2009)
Gvaramadze V. V., Gualandris A., & Portegies Zwart S. 2009, MNRAS, 396, 570
Gvaramadze & Gualandris (2011)
Gvaramadze V. V. & Gualandris A. 2011, MNRAS, 410, 304.
Hills (1988)
Hills, J. G. 1988, Nature, 331, 687
Hirsch et al. (2005)
Hirsch, H. A., Heber, U., O’Toole, S. J., & Bresolin, F. 2005, A&A, 444, L61
Ivezić et al. (2008)
Ivezić, Ž., et al. 2008, ApJ, 684, 287
Kenyon et al. (2008)
Kenyon, S. J., Bromley, B. C., Geller, M. J., & Brown, W. R. 2008, ApJ, 680, 312
Kiziltan et al. (2010)
Kiziltan, B., Kottas, A., & Thorsett, S. E. 2010, astro-ph/10114291v1
Kollmeier et al. (2009)
Kollmeier, J. A., Gould, A., Knapp, G., & Beers, T. C. 2009, ApJ, 697, 1543
Kollmeier et al. (2010)
Kollmeier, J. A., et al. 2010, ApJ, 723, 812
Koposov et al. (2010)
Koposov, S. E., Rix, H. W., & Hogg, D. W. 2010, ApJ, 712, 260
Lee et al. (2008)
Lee, Y. S., et al. 2008, AJ, 136, 2022
Leonard (1991)
Leonard, P. J. T. 1991, AJ, 101, 562
Lu et al. (2010)
Lu, Y., Zhang, F., & Yu, Q. 2010, ApJ, 709, 1356
Lu et al. (2007)
Lu, Y., Yu, Q., & Lin, D. N. C. 2007, ApJ, 666, L89
Merritt (2006)
Merritt, D. 2006, ApJ, 648, 976
Munn et al. (2004)
Munn, J. A., et al. 2004, AJ, 127, 3034
O’Leary & Loeb (2008)
O’Leary, R. M., & Loeb, A. 2008, MNRAS, 383, 86
Paczynski (1990)
Paczynski, B. 1990, ApJ, 348, 485
Perets (2009a)
Perets, H. B. 2009a, ApJ, 690, 795
Perets (2009b)
Perets, H. B. 2009b, ApJ, 698, 1330
Poveda et al. (1967)
Poveda, A., Ruiz, J., & Allen, C. 1967, Bol. Obs. Tonantzintla Tacubaya, 4, 86
Rastegaev (2010)
Rastegaev, D. A. 2010, ApJ, 140, 2013
Schönrich et al. (2010)
Schönrich, R., Binney, J., & Dehnen, W. 2010, MNRAS, 403, 1829
Sesana et al. (2007)
Sesana, A., Haardt, F., & Madau, P. 2007, MNRAS, 379, 45
Sesana et al. (2006)
Sesana, A., Haardt, F., & Madau, P. 2006, ApJ, 651, 392
Sherwin et al. (2008)
Sherwin, B. D., Loeb, A., & O’Leary, R. M. 2008, MNRAS, 386, 1179
Teyssier et al. (2009)
Teyssier, M., Johnston, K. V., & Shara, M. M. 2009, ApJ, 707, L22
Tillich et al. (2009)
Tillich, A., Przybilla, N., Scholz, R. D, & Heber, U. 2009, A&A, 507, L37
Tillich et al. (2010)
Tillich, A., Przybilla, N., Scholz, R. D, & Heber, U. 2010, A&A, 517, A36
Williams et al. (2011)
Williams, M. E. K., et al. 2011, ApJ, 728, 102
Xue et al. (2008)
Xue, X. X., et al. 2008, ApJ, 684, 1143
York et al. (2000)
York, D. G., et al. 2000, AJ, 120, 1579
Yu & Tremaine (2003)
Yu, Q., & Tremaine, S. 2003, ApJ, 599, 1129
Zhang et al. (2010)
Zhang, F., Lu, Y., & Yu, Q. 2010, ApJ, 722, 1744 |
Some generalizations of the Aluthge transform of operators and their consequences
Mojtaba Bakherad${}^{1}$ and Khalid Shebrawi${}^{2}$
${}^{1}$Department of Mathematics, Faculty of Mathematics, University of Sistan and Baluchestan, Zahedan, I.R.Iran.
[email protected]; [email protected]
${}^{2}$Department of Mathematics, Al-Balqa’ Applied University, Salt, Jordan.
[email protected]; [email protected]
Abstract.
Let $A=U|A|$ be the polar decomposition of $A$. The
Aluthge transform of the operator $A$, denoted by $\tilde{A}$, is defined as
$\tilde{A}=|A|^{\frac{1}{2}}U|A|^{\frac{1}{2}}$.
In this paper, first we generalize the definition of Aluthge transform for
non-negative continuous functions $f,g$ such that $f(x)g(x)=x\,\,(x\geq 0)$. Then, by using of this definition, we get some numerical radius inequalities. Among other inequalities, it is shown that if $A$ is bounded linear operator on a complex Hilbert space ${\mathscr{H}}$, then
$$h\left(w(A)\right)\leq\frac{1}{4}\left\|h\left(g^{2}\left(\left|A\right|\right%
)\right)+h\left(f^{2}\left(\left|A\right|\right)\right)\right\|+\frac{1}{2}h%
\left(w\left(\tilde{A}_{f,g}\right)\right),$$
where $f,g$ are non-negative continuous functions such that $f(x)g(x)=x\,\,(x\geq 0)$, $h$ is a non-negative non-decreasing convex function on $[0,\infty)$ and $\tilde{A}_{f,g}=f(|A|)Ug(|A|)$.
Key words and phrases:Aluthge transform, Numerical radius, Operator matrices, Polar decomposition.
2010 Mathematics Subject Classification: Primary 47A12, Secondary 47A63, 47A30
1. Introduction
Let ${\mathbb{B}}(\mathscr{H})$ denote the $C^{*}$-algebra of all bounded linear operators on a complex Hilbert space ${\mathscr{H}}$ with an inner product $\langle\cdot,\cdot\rangle$ and the corresponding norm $\|\cdot\|$. In the case when ${\rm dim}{\mathscr{H}}=n$, we identify ${\mathbb{B}}({\mathscr{H}})$ with the matrix algebra $\mathbb{M}_{n}$ of all $n\times n$ matrices with entries in
the complex field.
For an operator $A\in{\mathbb{B}}(\mathscr{H})$, let $A=U|A|$ (U is a partial
isometry with $\textrm{ker}U=\textrm{rng}|A|^{\bot}$) be the polar decomposition of $A$. The
Aluthge transform of the operator $A$, denoted by $\tilde{A}$, is defined as
$\tilde{A}=|A|^{\frac{1}{2}}U|A|^{\frac{1}{2}}.$
In [12], Okubo introduced a more general notion called $t$-Aluthge transform which has later
been studied also in detail. This is defined for any $0<t\leq 1$ by
$\tilde{A}_{t}=|A|^{t}U|A|^{1-t}$. Clearly, for $t=\frac{1}{2}$ we obtain the usual Aluthge transform. As for the case $t=1$, the operator $\tilde{A}_{1}=|A|U$ is called the Duggal transform of $A\in{\mathbb{B}}(\mathscr{H})$. For $A\in{\mathbb{B}}(\mathscr{H})$, we generalize the Aluthge transform of the operator $A$ to the form
$$\displaystyle\tilde{A}_{f,g}=f(|A|)Ug(|A|),$$
in which $f,g$ are non-negative continuous functions such that $f(x)g(x)=x\,\,(x\geq 0).$ The numerical radius of $A\in{\mathbb{B}}({\mathscr{H}})$ is defined by
$$\displaystyle w(A):=\sup\{|\langle Ax,x\rangle|:x\in{\mathscr{H}},\|x\|=1\}.$$
It is well known that $w(\,\cdot\,)$ defines a norm on ${\mathbb{B}}({\mathscr{H}})$, which is equivalent to the usual operator norm $\|\cdot\|$. In fact, for any $A\in{\mathbb{B}}({\mathscr{H}})$,
$\frac{1}{2}\|A\|\leq w(A)\leq\|A\|$;
see [6]. Let $r(\cdot)$ denote to the spectral radius. It is well known that for every operator
$A\in{\mathbb{B}}({\mathscr{H}})$, we have $r(A)\leq w(A)$. An important inequality for $\omega(A)$ is the power inequality stating that $\omega(A^{n})\leq\omega(A)^{n}\,\,(n=1,2,\cdots)$. The quantity $w(A)$ is useful in studying perturbation, convergence and approximation problems as well as integrative method, etc. For more information see [3, 7, 8, 9] and references therein.
Let $A,B,C,D\in{\mathbb{B}}({\mathscr{H}})$. The operator matrices $\left[\begin{array}[]{cc}A&0\\
0&D\end{array}\right]$ and $\left[\begin{array}[]{cc}0&B\\
C&0\end{array}\right]$ are called the diagonal and off-diagonal parts of the operator matrix $\left[\begin{array}[]{cc}A&B\\
C&D\end{array}\right]$, respectively.
In [11], It has been shown that if $A$ is an operator in ${\mathbb{B}}({\mathscr{H}})$, then
$$\displaystyle w(A)\leq\frac{1}{2}\left(\|A\|+\|A^{2}\|^{\frac{1}{2}}\right).$$
(1.1)
Several refinements and generalizations of inequality (1.1) have been given; see [1, 4, 14, 15]. Yamazaki [15] showed that for $A\in{\mathbb{B}}({\mathscr{H}})$ and $t\in[0,1]$ we have
$$\displaystyle w(A)\leq\frac{1}{2}\left(\|A\|+w(\tilde{A}_{t})\right).$$
(1.2)
Davidson and Power [5] proved that if $A$ and $B$ are positive operators in ${\mathbb{B}}({\mathscr{H}})$, then
$$\displaystyle\|A+B\|\leq\max\{\|A\|,\|B\|\}+\|AB\|^{\frac{1}{2}}.$$
(1.3)
Inequality (1.3) has been generalized in [2, 13]. In [13], the author extended this inequality to the form
$$\displaystyle\|A+B^{*}\|\leq\max\{\|A\|,\|B\|\}+\frac{1}{2}\left(\left\||A|^{t%
}|B^{*}|^{1-t}\right\|+\left\||A^{*}|^{1-t}|B|^{t}\right\|\right),$$
(1.4)
in which $A,B\in{\mathbb{B}}({\mathscr{H}})$ and $t\in[0,1]$.
In this paper, by applying the generalized Aluthge transform of operators, we establish some inequalities involving the numerical radius. In particular, we extend inequality (1.2) and (1.4) for two non-negative continuous functions. We also show some upper bounds for the numerical radius of $2\times 2$ operators matrices.
2. main results
To prove our numerical radius inequalities, we need several known lemmas.
Lemma 2.1.
[1, Theorem 2.2]
Let $X,Y,S,T\in{\mathbb{B}}({\mathscr{H}})$. Then
$$\displaystyle r(XY+ST)\leq\frac{1}{2}\left(w(YX)+w(TS)\right)+\frac{1}{2}\sqrt%
{\left(w(YX)-w(TS)\right)^{2}+4\|YS\|\|TX\|}.$$
Lemma 2.2.
[15, 11]
Let $A\in{\mathbb{B}}({\mathscr{H}})$. Then
$(a)\,\,w(A)=\underset{\theta\in\mathbb{R}}{\max}\left\|\textrm{Re}\left(e^{i%
\theta}A\right)\right\|.$
$(b)\,\,w\left(\left[\begin{array}[]{cc}0&A\\
0&0\end{array}\right]\right)=\frac{1}{2}\|A\|.$
Polarization identity: For all $x,y\in{\mathscr{H}}$, we have
$$\left\langle x,y\right\rangle=\frac{1}{4}\sum_{k=0}^{3}\left\|x+i^{k}y\right\|%
^{2}i^{k}.$$
Now, we are ready to present our first result. The following theorem shows a generalization of inequality (1.2).
Theorem 2.3.
Let $A\in{\mathbb{B}}({\mathscr{H}})$ and $f,g$ be two non-negative
continuous functions on $[0,\infty)$ such that $f(x)g(x)=x\,\,(x\geq 0)$.
Then, for all non-negative non-decreasing convex function $h$ on $[0,\infty)$, we have
$$h\left(w(A)\right)\leq\frac{1}{4}\left\|h\left(g^{2}\left(\left|A\right|\right%
)\right)+h\left(f^{2}\left(\left|A\right|\right)\right)\right\|+\frac{1}{2}h%
\left(w\left(\tilde{A}_{f,g}\right)\right).$$
Proof.
Let $x$ be any unit vector. Then
$$\displaystyle\textrm{Re}\left\langle e^{i\theta}Ax,x\right\rangle$$
$$\displaystyle=$$
$$\displaystyle\textrm{Re}\left\langle e^{i\theta}U\left|A\right|x,x\right\rangle$$
$$\displaystyle=$$
$$\displaystyle\textrm{Re}\left\langle e^{i\theta}Ug\left(\left|A\right|\right)f%
\left(\left|A\right|\right)x,x\right\rangle$$
$$\displaystyle=$$
$$\displaystyle\textrm{Re}\left\langle e^{i\theta}f\left(\left|A\right|\right)x,%
g\left(\left|A\right|\right)U^{\ast}x\right\rangle$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{4}\left\|\left(e^{i\theta}f\left(\left|A\right|\right)+g%
\left(\left|A\right|\right)U^{\ast}\right)x\right\|^{2}-\frac{1}{4}\left\|%
\left(e^{i\theta}f\left(\left|A\right|\right)-g\left(\left|A\right|\right)U^{%
\ast}\right)x\right\|^{2}$$
$$\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\text{\ (by
polarization identity)}$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{4}\left\|\left(e^{i\theta}f\left(\left|A\right|\right)+g%
\left(\left|A\right|\right)U^{\ast}\right)x\right\|^{2}$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{4}\left\|\left(e^{i\theta}f\left(\left|A\right|\right)+g%
\left(\left|A\right|\right)U^{\ast}\right)\right\|^{2}$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{4}\left\|\left(e^{i\theta}f\left(\left|A\right|\right)+g%
\left(\left|A\right|\right)U^{\ast}\right)\left(e^{-i\theta}f\left(\left|A%
\right|\right)+Ug\left(\left|A\right|\right)\right)\right\|$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{4}\left\|g^{2}\left(\left|A\right|\right)+f^{2}\left(%
\left|A\right|\right)+e^{i\theta}\tilde{A}_{f,g}+e^{-i\theta}\left(\tilde{A}_{%
f,g}\right)^{\ast}\right\|$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{4}\left\|g^{2}\left(\left|A\right|\right)+f^{2}\left(%
\left|A\right|\right)\right\|+\frac{1}{4}\left\|e^{i\theta}\tilde{A}_{f,g}+e^{%
-i\theta}\left(\tilde{A}_{f,g}\right)^{\ast}\right\|$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{4}\left\|g^{2}\left(\left|A\right|\right)+f^{2}\left(%
\left|A\right|\right)\right\|+\frac{1}{2}\left\|\textrm{Re}\left(e^{i\theta}%
\tilde{A}_{f,g}\right)\right\|$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{4}\left\|g^{2}\left(\left|A\right|\right)+f^{2}\left(%
\left|A\right|\right)\right\|+\frac{1}{2}w\left(\tilde{A}_{f,g}\right).$$
Now, taking the supremum over all unit vector $x\in{\mathscr{H}}$ and applying Lemma 2.2 in the above inequality produces
$$\displaystyle w\left(A\right)\leq\frac{1}{4}\left\|g^{2}\left(\left|A\right|%
\right)+f^{2}\left(\left|A\right|\right)\right\|+\frac{1}{2}w\left(\tilde{A}_{%
f,g}\right).$$
Therefore,
$$\displaystyle h\left(w\left(A\right)\right)$$
$$\displaystyle\leq$$
$$\displaystyle h\left(\frac{1}{4}\left\|g^{2}\left(\left|A\right|\right)+f^{2}%
\left(\left|A\right|\right)\right\|+\frac{1}{2}w\left(\tilde{A}_{f,g}\right)\right)$$
$$\displaystyle=$$
$$\displaystyle h\left(\frac{1}{2}\left\|\frac{g^{2}\left(\left|A\right|\right)+%
f^{2}\left(\left|A\right|\right)}{2}\right\|+\frac{1}{2}w\left(\tilde{A}_{f,g}%
\right)\right)$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{2}h\left(\left\|\frac{g^{2}\left(\left|A\right|\right)+f%
^{2}\left(\left|A\right|\right)}{2}\right\|\right)+\frac{1}{2}h\left(w\left(%
\tilde{A}_{f,g}\right)\right)$$
$$\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad(\textrm{by the convexity of %
}\,h)$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{2}\left\|h\left(\frac{g^{2}\left(\left|A\right|\right)+f%
^{2}\left(\left|A\right|\right)}{2}\right)\right\|+\frac{1}{2}h\left(w\left(%
\tilde{A}_{f,g}\right)\right)$$
$$\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad(\textrm{by the functional %
calculus})$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{4}\left\|h\left(g^{2}\left(\left|A\right|\right)\right)+%
h\left(f^{2}\left(\left|A\right|\right)\right)\right\|+\frac{1}{2}h\left(w%
\left(\tilde{A}_{f,g}\right)\right)$$
$$\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad(\textrm{by the convexity of %
}\,h).$$
∎
Theorem 2.3 includes some special cases as follows.
Corollary 2.4.
Let $A\in{\mathbb{B}}({\mathscr{H}})$. Then, for all non-negative non-decreasing convex
function $h$ on $[0,\infty)$ and all $t\in\left[0,1\right]$, we have
$$h\left(w(A)\right)\leq\frac{1}{4}\left\|h\left(\left|A\right|^{2t}\right)+h%
\left(\left|A\right|^{2\left(1-t\right)}\right)\right\|+\frac{1}{2}h\left(w%
\left(\tilde{A}_{t}\right)\right).$$
(2.1)
Corollary 2.5.
Let $A\in{\mathbb{B}}({\mathscr{H}})$. Then, for all $t\in\left[0,1\right]$ and $r\geq 1$, we have
$$w^{r}(A)\leq\frac{1}{4}\|\left|A\right|^{2tr}+\left|A\right|^{2\left(1-t\right%
)r}\|+\frac{1}{2}w^{r}\left(\tilde{A}_{t}\right).$$
In particular,
$$w^{r}(A)\leq\frac{1}{2}\left(\|A\|^{r}+w^{r}\left(\tilde{A}\right)\right).$$
Proof.
The first inequality follows from
inequality (2.1) for the function $h\left(x\right)=x^{r}\,\,(r\geq 1)$. For the particular case, it is enough to put $t=\frac{1}{2}.$
∎
Theorem 2.3 gives the next result for the off-diagonal operator matrix $\left[\begin{array}[]{cc}0&A\\
B&0\end{array}\right]$.
Theorem 2.6.
Let $A,B\in{\mathbb{B}}({\mathscr{H}})$, $f,g$ be two non-negative
continuous functions on $[0,\infty)$ such that $f(x)g(x)=x\,\,(x\geq 0)$ and $r\geq 1$. Then
$$\displaystyle w^{r}\left(\left[\begin{array}[]{cc}0&A\\
B&0\end{array}\right]\right)$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{4}\max\left(\left\|g^{2r}\left(\left|A\right|\right)+f^{%
2r}\left(\left|A\right|\right)\right\|,\left\|g^{2r}\left(\left|B\right|\right%
)+f^{2r}\left(\left|B\right|\right)\right\|\right)$$
$$\displaystyle+\frac{1}{4}\big{(}\|f(|B|)g(|A^{\ast}|)\|^{r}+\|f(|A|)g(|B^{\ast%
}|)\|^{r}\big{)}.$$
Proof.
Let $A=U|A|$ and $B=V|B|$ be the polar decompositions of $A$ and $B$, respectively and let $T=\left[\begin{array}[]{cc}0&A\\
B&0\end{array}\right]$. It follows from the polar the composition of $T=\left[\begin{array}[]{cc}0&U\\
V&0\end{array}\right]\left[\begin{array}[]{cc}|B|&0\\
0&|A|\end{array}\right]$ that
$$\displaystyle\tilde{T}_{f,g}$$
$$\displaystyle=$$
$$\displaystyle f(|T|)\left[\begin{array}[]{cc}0&U\\
V&0\end{array}\right]g(|T|)$$
$$\displaystyle=$$
$$\displaystyle\left[\begin{array}[]{cc}f(|B|)&0\\
0&f(|A|)\end{array}\right]\left[\begin{array}[]{cc}0&U\\
V&0\end{array}\right]\left[\begin{array}[]{cc}g(|B|)&0\\
0&g(|A|)\end{array}\right]$$
$$\displaystyle=$$
$$\displaystyle\left[\begin{array}[]{cc}0&f(|B|)Ug(|A|)\\
f(|A|)Vg(|B|)&0\end{array}\right].$$
Using $|A^{\ast}|^{2}=AA^{\ast}=U|A|^{2}U^{\ast}$ and $|B^{\ast}|^{2}=BB^{\ast}=V|B|^{2}V^{\ast}$ we have $g(|A|)=U^{\ast}g(|A^{\ast}|)U$ and $g(|B|)=V^{\ast}g(|B^{\ast}|)V$ for every non-negative continuous function $g$ on $[0,\infty)$.
Therefore,
$$\displaystyle w\left(\tilde{T}_{f,g}\right)$$
$$\displaystyle=$$
$$\displaystyle w\left(\left[\begin{array}[]{cc}0&f(|B|)Ug(|A|)\\
f(|A|)Vg(|B|)&0\end{array}\right]\right)$$
(2.2)
$$\displaystyle\leq$$
$$\displaystyle w\left(\left[\begin{array}[]{cc}0&f(|B|)Ug(|A|)\\
0&0\end{array}\right]\right)+w\left(\left[\begin{array}[]{cc}0&0\\
f(|A|)Vg(|B|)&0\end{array}\right]\right)$$
$$\displaystyle=$$
$$\displaystyle w\left(\left[\begin{array}[]{cc}0&f(|B|)Ug(|A|)\\
0&0\end{array}\right]\right)+w\left(U^{*}\left[\begin{array}[]{cc}0&f(|A|)Vg(|%
B|)\\
0&0\end{array}\right]U\right)$$
$$\displaystyle=$$
$$\displaystyle w\left(\left[\begin{array}[]{cc}0&f(|B|)Ug(|A|)\\
0&0\end{array}\right]\right)+w\left(\left[\begin{array}[]{cc}0&f(|A|)Vg(|B|)\\
0&0\end{array}\right]\right)$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{2}\|f(|B|)Ug(|A|)\|+\frac{1}{2}\|f(|A|)Vg(|B|)\|$$
$$\displaystyle\qquad\qquad\qquad\qquad\qquad\textrm{(by Lemma \ref{kit1}(b))}$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{2}\|f(|B|)UU^{\ast}g(|A^{\ast}|)U\|+\frac{1}{2}\|f(|A|)%
VV^{\ast}g(|B^{\ast}|)V\|$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{2}\|f(|B|)g(|A^{\ast}|)\|+\frac{1}{2}\|f(|A|)g(|B^{\ast}%
|)\|,$$
where $U=\left[\begin{array}[]{cc}0&I\\
I&0\end{array}\right]$ is unitary. Applying Theorem 2.3 and inequality (2.2), we have
$$\displaystyle w^{r}\left(T\right)$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{4}\left\|g^{2r}\left(\left|T\right|\right)+f^{2r}\left(%
\left|T\right|\right)\right\|+\frac{1}{2}\left(w^{r}\left(\tilde{T}_{f,g}%
\right)\right)$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{4}\max\left(\left\|g^{2r}\left(\left|A\right|\right)+f^{%
2r}\left(\left|A\right|\right)\right\|,\left\|g^{2r}\left(\left|B\right|\right%
)+f^{2r}\left(\left|B\right|\right)\right\|\right)$$
$$\displaystyle+\frac{1}{2}\left[\frac{1}{2}\left(\|f(|B|)g(|A^{\ast}|)\|+\|f(|A%
|)g(|B^{\ast}|)\|\right)\right]^{r}$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{4}\max\left(\left\|g^{2r}\left(\left|A\right|\right)+f^{%
2r}\left(\left|A\right|\right)\right\|,\left\|g^{2r}\left(\left|B\right|\right%
)+f^{2r}\left(\left|B\right|\right)\right\|\right)$$
$$\displaystyle+\frac{1}{4}\|f(|B|)g(|A^{\ast}|)\|^{r}+\frac{1}{4}\|f(|A|)g(|B^{%
\ast}|)\|^{r}$$
$$\displaystyle\qquad\qquad(\textrm{by the convexity}\,h(x)=x^{r}).$$
∎
Corollary 2.7.
Let $A,B\in{\mathbb{B}}({\mathscr{H}})$. Then,
for all $t\in\left[0,1\right]$ and $r\geq 1$, we have
$$\displaystyle w^{\frac{r}{2}}\left(AB\right)$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{4}\max\left(\left\|\left|A\right|^{2tr}+\left|A\right|^{%
2\left(1-t\right)r}\right\|,\left\|\left|B\right|^{2tr}+\left|B\right|^{2\left%
(1-t\right)r}\right\|\right)$$
$$\displaystyle+\frac{1}{4}\left(\left\||A|^{t}\left|B^{\ast}\right|^{1-t}\right%
\|^{r}+\left\|\left|B\right|^{t}|A^{\ast}|^{1-t}\right\|^{r}\right).$$
Proof.
Applying the power inequality of the numerical radius, we have
$$\displaystyle w^{\frac{r}{2}}\left(AB\right)$$
$$\displaystyle\leq$$
$$\displaystyle\max\left(w^{\frac{r}{2}}\left(AB\right),w^{\frac{r}{2}}\left(BA%
\right)\right)$$
$$\displaystyle=$$
$$\displaystyle w^{\frac{r}{2}}\left(\left[\begin{array}[]{cc}AB&0\\
0&BA\end{array}\right]\right)$$
$$\displaystyle=$$
$$\displaystyle w^{\frac{r}{2}}\left(\left[\begin{array}[]{cc}0&A\\
B&0\end{array}\right]^{2}\right)$$
$$\displaystyle\leq$$
$$\displaystyle w^{r}\left(\left[\begin{array}[]{cc}0&A\\
B&0\end{array}\right]\right)$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{4}\max\left(\left\|\left|A\right|^{2tr}+\left|A\right|^{%
2\left(1-t\right)r}\right\|,\left\|\left|B\right|^{2tr}+\left|B\right|^{2\left%
(1-t\right)r}\right\|\right)$$
$$\displaystyle+\frac{1}{4}\left(\left\||A|^{t}\left|B^{\ast}\right|^{1-t}\right%
\|^{r}+\left\|\left|B\right|^{t}|A^{\ast}|^{1-t}\right\|^{r}\right)$$
$$\displaystyle\qquad\qquad\qquad\qquad\textrm{(by Theorem \ref{man-kal})}.$$
∎
Corollary 2.8.
Let $A,B\in{\mathbb{B}}({\mathscr{H}})$ be positive operators. Then,
for all $t\in\left[0,1\right]$ and $r\geq 1$, we have
$$\displaystyle\left\|A^{\frac{1}{2}}B^{\frac{1}{2}}\right\|^{r}$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{4}\max\left(\left\|A^{tr}+A^{\left(1-t\right)r}\right\|,%
\left\|B^{tr}+B^{2\left(1-t\right)r}\right\|\right)$$
$$\displaystyle+\frac{1}{4}\left(\left\|A^{t}B^{1-t}\right\|^{r}+\left\|B^{t}A^{%
1-t}\right\|^{r}\right).$$
Proof.
Since the spectral radius of any operator is dominated by its numerical
radius, then $r^{\frac{1}{2}}\left(AB\right)\leq w^{\frac{1}{2}}\left(AB\right).$ Applying a commutativity property of the spectral radius, we get
$$\displaystyle r^{\frac{r}{2}}\left(AB\right)$$
$$\displaystyle=$$
$$\displaystyle r^{\frac{r}{2}}\left(A^{\frac{1}{2}}A^{\frac{1}{2}}B^{\frac{1}{2%
}}B^{\frac{1}{2}}\right)$$
(2.3)
$$\displaystyle=$$
$$\displaystyle r^{\frac{r}{2}}\left(A^{\frac{1}{2}}B^{\frac{1}{2}}B^{\frac{1}{2%
}}A^{\frac{1}{2}}\right)$$
$$\displaystyle=$$
$$\displaystyle r^{\frac{r}{2}}\left(A^{\frac{1}{2}}B^{\frac{1}{2}}\left(A^{%
\frac{1}{2}}B^{\frac{1}{2}}\right)^{\ast}\right)$$
$$\displaystyle=$$
$$\displaystyle\left\|A^{\frac{1}{2}}B^{\frac{1}{2}}\left(A^{\frac{1}{2}}B^{%
\frac{1}{2}}\right)^{\ast}\right\|^{\frac{r}{2}}$$
$$\displaystyle=$$
$$\displaystyle\left\|A^{\frac{1}{2}}B^{\frac{1}{2}}\right\|^{r}.$$
Now, the result follows from Corollary 2.7.
∎
An important special case of Theorem 2.6, which refines inequality (1.4) can be stated as follows.
Corollary 2.9.
Let $A,B\in{\mathbb{B}}({\mathscr{H}})$ and $r\geq 1$. Then
$$\displaystyle\left\|A+B\right\|^{r}$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{2^{2-r}}\max\left(\left\|\left|A\right|^{2tr}+\left|A%
\right|^{2\left(1-t\right)r}\right\|,\left\|\left|B^{*}\right|^{2tr}+\left|B^{%
*}\right|^{2\left(1-t\right)r}\right\|\right)$$
$$\displaystyle+\frac{1}{2^{2-r}}\left(\left\||A|^{t}\left|B\right|^{1-t}\right%
\|^{r}+\left\|\left|B^{*}\right|^{t}|A^{\ast}|^{1-t}\right\|^{r}\right).$$
In particular, if $A$ and $B$ are normal, then
$$\displaystyle\left\|A+B\right\|^{r}\leq\frac{1}{2^{1-r}}\max\left(\left\|A%
\right\|^{r},\left\|B\right\|^{r}\right)+\frac{1}{2^{1-r}}\left\|AB\right\|^{%
\frac{r}{2}}.$$
Proof.
Applying Lemma 2.2 and Theorem 2.3, we have
$$\displaystyle\left\|A+B^{\ast}\right\|^{r}$$
$$\displaystyle=$$
$$\displaystyle\left\|T+T^{\ast}\right\|^{r}$$
$$\displaystyle\leq$$
$$\displaystyle 2^{r}\underset{\theta\in\mathbb{R}}{\max}\left\|\textrm{Re}\left%
(e^{i\theta}T\right)\right\|^{r}$$
$$\displaystyle=$$
$$\displaystyle 2^{r}w^{r}\left(T\right)$$
$$\displaystyle\leq$$
$$\displaystyle\frac{2^{r}}{4}\max\left(\left\|\left|A\right|^{2tr}+\left|A%
\right|^{2\left(1-t\right)r}\right\|,\left\|\left|B\right|^{2tr}+\left|B\right%
|^{2\left(1-t\right)r}\right\|\right)$$
$$\displaystyle+\frac{2^{r}}{4}\left(\left\||A|^{t}\left|B^{\ast}\right|^{1-t}%
\right\|^{r}+\left\|\left|B\right|^{t}|A^{\ast}|^{1-t}\right\|^{r}\right)$$
$$\displaystyle\textrm{(by Theorem \ref{man-kal})},$$
where $T=\left[\begin{array}[]{cc}0&A\\
B&0\end{array}\right].$ Now, the desired result follows by replacing $B$ by $B^{\ast}$.
For the particular case, since $A$ and $B$ are normal,
then $\left|B^{\ast}\right|=$ $\left|B\right|$ and $|A^{\ast}|=|A|.$ Applying equality $($2.3$)$ for the operators $\left|A\right|^{\frac{1}{2}}$ and $\left|B\right|^{\frac{1}{2}}$, we have
$$\displaystyle\left\|\left|A\right|^{\frac{1}{2}}\left|B\right|^{\frac{1}{2}}%
\right\|^{r}$$
$$\displaystyle=$$
$$\displaystyle r^{\frac{r}{2}}\left(\left|A\right|\left|B\right|\right)$$
$$\displaystyle\leq$$
$$\displaystyle\left\|\left|A\right|\left|B\right|\right\|^{\frac{r}{2}}$$
$$\displaystyle=$$
$$\displaystyle\left\|U^{\ast}AB^{\ast}V\right\|^{\frac{r}{2}}$$
$$\displaystyle=$$
$$\displaystyle\left\|AB^{\ast}\right\|^{\frac{r}{2}},$$
where $A=U|A|$ and $B=V\left|B\right|$ are the
polar decompositions of the operators $A$ and $B.$ This completes the proof
of the corollary.
∎
In the next result, we show another generalization of inequality (1.2).
Theorem 2.10.
Let $A\in{\mathbb{B}}({\mathscr{H}})$ and $f,g,h$ be non-negative non-decreasing continuous functions on $[0,\infty)$ such that $f(x)g(x)=x\,\,(x\geq 0)$. Then
$$\displaystyle h\left(w(A)\right)\leq\frac{1}{2}\Big{(}h\left(w\left(\tilde{A}_%
{f,g}\right)\right)+\left\|h(|A|)\right\|\Big{)}.$$
Proof.
Let $A=U|A|$ be the polar decomposition of A.
Then for every $\theta\in\mathbb{R}$, we have
$$\displaystyle\|\textrm{Re}\left(e^{i\theta}A\right)\|$$
$$\displaystyle=$$
$$\displaystyle r\left(\textrm{Re}\left(e^{i\theta}A\right)\right)$$
(2.4)
$$\displaystyle=$$
$$\displaystyle\frac{1}{2}r\left(e^{i\theta}A+e^{-i\theta}A^{\ast}\right)$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{2}r\left(e^{i\theta}U|A|+e^{-i\theta}|A|U^{\ast}\right)$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{2}r\left(e^{i\theta}Ug(|A|)f(|A|)+e^{-i\theta}f(|A|)g(|A%
|)U^{\ast}\right).$$
Now, if we put $X=e^{i\theta}Ug(|A|)$, $Y=f(|A|)$, $S=e^{-i\theta}f(|A|)$ and $T=g(|A|)U^{\ast}$ in Lemma 2.1, then we get
$$\displaystyle r\big{(}e^{i\theta}Ug(|A|)f(|A|)+e^{-i\theta}f(|A|)g(|A|)U^{\ast%
}\big{)}$$
(2.5)
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{2}\Big{(}w(f(|A|)Ug(|A|))+w(g(|A|)U^{\ast}f(|A|))\Big{)}$$
$$\displaystyle+\frac{1}{2}\sqrt{4\|e^{-i\theta}f(|A|)g(|A|)\|\|g(|A|)U^{\ast}e^%
{i\theta}Uf(|A|)\|}$$
$$\displaystyle\qquad\qquad\qquad\qquad\qquad\textrm{(by Lemma \ref{kit1})}$$
$$\displaystyle\leq$$
$$\displaystyle w(f(|A|)Ug(|A|))+\sqrt{\|f(|A|)\|\|f(|A|)\|\|g(|A|)\|\|g(|A|)\|}$$
$$\displaystyle=$$
$$\displaystyle w(f(|A|)Ug(|A|))+\sqrt{f(\|A\|)g(\|A\|)g(\|A\|)f(\|A\|)}$$
$$\displaystyle\qquad\qquad\qquad\qquad\qquad\textrm{(by the functional calculus)}$$
$$\displaystyle=$$
$$\displaystyle w(f(|A|)Ug(|A|))+\sqrt{\|A\|\|A\|}$$
$$\displaystyle=$$
$$\displaystyle w\left(\tilde{A}_{f,g}\right)+\|A\|.$$
Using inequalities (2.4), (2.5) and Lemma 2.2 we get
$$\displaystyle\omega(A)=\underset{\theta\in\mathbb{R}}{\max}\left\|\textrm{Re}%
\left(e^{i\theta}A\right)\right\|\leq\frac{1}{2}\Big{(}w\left(\tilde{A}_{f,g}%
\right)+\|A\|\Big{)}.$$
Hence
$$\displaystyle h\left(w(A)\right)$$
$$\displaystyle\leq$$
$$\displaystyle h\left(\frac{1}{2}\left[w\left(\tilde{A}_{f,g}\right)+\left\|A%
\right\|\right]\right)$$
$$\displaystyle\qquad\qquad\qquad(\textrm{by the monotonicity of}\,\,h)$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{2}h\left(w\left(\tilde{A}_{f,g}\right)\right)+\frac{1}{2%
}h\left(\left\|A\right\|\right)$$
$$\displaystyle\qquad\qquad\qquad(\textrm{by the convexity of}\,\,h)$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{2}h\left(w\left(\tilde{A}_{f,g}\right)\right)+\frac{1}{2%
}\left\|h(|A|)\right\|,$$
as required.
∎
Another proof for Theorem 2.3:
We can obtain Theorem 2.3 from Theorem 2.10.
To see this, first note that by the hypotheses of Theorem 2.3 we have
$$\displaystyle h(|A|)$$
$$\displaystyle=$$
$$\displaystyle h(g(|A|)f(|A|))$$
(2.6)
$$\displaystyle\leq$$
$$\displaystyle h\left(\frac{g^{2}(|A|)+f^{2}(|A|)}{2}\right)\qquad\textrm{(by %
the arithmetic-geometric inequality)}$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{2}\left(h\left(g^{2}(|A|)\right)+h\left(f^{2}(|A|)\right%
)\right)\qquad\textrm{(by the convexity of\,}h).$$
Hence, using Theorem 2.10 and inequality (2.6) we get
$$\displaystyle h\left(w(A)\right)$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{2}\Big{[}h\left(w\left(\tilde{A}_{f,g}\right)\right)+%
\left\|h(|A|)\right\|\Big{]}$$
$$\displaystyle\leq$$
$$\displaystyle\frac{1}{2}\Big{[}h\left(w\left(\tilde{A}_{f,g}\right)\right)+%
\frac{1}{2}\left\|h\left(g^{2}(|A|)\right)+h\left(f^{2}(|A|)\right)\right\|%
\Big{]}$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{2}h\left(w\left(\tilde{A}_{f,g}\right)\right)+\frac{1}{4%
}\left\|h\left(g^{2}(|A|)\right)+h\left(f^{2}(|A|)\right)\right\|.$$
Remark 2.11.
For the special case $f(x)=x^{t}$ and $g=x^{1-t}\,\,(t\in[0,1])$, we obtain the inequality (1.2)
$$\displaystyle w(A)\leq\frac{1}{2}\left(w\left(\tilde{A}_{t}\right)+\|A\|\right),$$
where $A\in{\mathbb{B}}({\mathscr{H}})$ and $t\in[0,1]$.
Using Theorem 2.10, we get the following result.
Corollary 2.12.
Let $A,B\in{\mathbb{B}}({\mathscr{H}})$ and $f,g$ be two non-negative non-decreasing continuous functions such that $f(x)g(x)=x\,\,(x\geq 0)$. Then
$$\displaystyle 2w^{r}\left(\left[\begin{array}[]{cc}0&A\\
B&0\end{array}\right]\right)\leq\max\{\|A\|^{r},\|B\|^{r}\}+\frac{1}{2}\big{(}%
\left\|f(|B|)g(|A^{\ast}|)\right\|^{r}+\left\|f(|A|)g(|B^{\ast}|)\right\|^{r}%
\big{)},$$
where $r\geq 1$.
Proof.
Using Theorem 2.10 and inequality (2.2), we have
$$\displaystyle 2w^{r}\left(\left[\begin{array}[]{cc}0&A\\
B&0\end{array}\right]\right)$$
$$\displaystyle\leq\left\|\left[\begin{array}[]{cc}0&A\\
B&0\end{array}\right]\right\|^{r}+w^{r}\left(\tilde{T}_{f,g}\right)$$
$$\displaystyle=\max\{\|A\|^{r},\|B\|^{r}\}+\left(\frac{1}{2}\big{[}\left\|f(|B|%
)g(|A^{\ast}|)\right\|+\|f(|A|)g(|B^{\ast}|)\|\big{]}\right)^{r}$$
$$\displaystyle\leq\max\{\|A\|^{r},\|B\|^{r}\}+\frac{1}{2}\big{(}\left\|f(|B|)g(%
|A^{\ast}|)\right\|^{r}+\|f(|A|)g(|B^{\ast}|)\|^{r}\big{)}$$
and the proof is complete.
∎
Corollary 2.13.
Let $A,B\in{\mathbb{B}}({\mathscr{H}})$ and $f,g$ be two non-negative non-decreasing continuous functions on $[0,\infty)$ such that $f(x)g(x)=x\,\,(x\geq 0)$. Then
$$\displaystyle\|A+B\|\leq\max\{\|A\|,\|B\|\}+\frac{1}{2}\Big{(}\big{\|}f(|B|)g(%
|A|)\big{\|}+\big{\|}f(|A^{\ast}|)g(|B^{\ast}|)\big{\|}\Big{)}.$$
Proof.
Let $T=\left[\begin{array}[]{cc}0&A\\
B&0\end{array}\right]$. Then
$$\displaystyle\left\|A+B^{{}^{\ast}}\right\|$$
$$\displaystyle=$$
$$\displaystyle\left\|T+T^{{}^{\ast}}\right\|$$
$$\displaystyle\leq$$
$$\displaystyle 2\underset{\theta\in\mathbb{R}}{\max}\left\|\textrm{Re}\left(e^{%
i\theta}T\right)\right\|$$
$$\displaystyle=$$
$$\displaystyle w(T)\qquad\qquad\qquad(\textrm{by Lemma \ref{1}})$$
$$\displaystyle\leq$$
$$\displaystyle\max\{\|A\|,\|B\|\}+\frac{1}{2}\Big{(}\big{\|}f(|B|)g(|A^{\ast}|)%
\big{\|}+\big{\|}f(|A|)g(|B^{\ast}|)\big{\|}\Big{)}$$
$$\displaystyle\qquad\qquad\qquad\qquad(\textrm{by Theorem \ref{main1}}).$$
If we replace $B$ by $B^{\ast}$, then we get the desired result.
∎
In the last results, we present some upper bounds for operator matrices. For this purpose, we need the following lemma.
Lemma 2.14.
[10, Theorem 1]
Let $A\in{\mathbb{B}}({\mathscr{H}})$ and $x,y\in{\mathscr{H}}$ be any vectors. If $f$, $g$ are non-negative continuous functions on $[0,\infty)$ which are satisfying the relation $f(x)g(x)=x\,(x\geq 0)$, then
$$\displaystyle|\left\langle Ax,y\right\rangle|^{2}\leq\left\langle f^{2}(|A|)x,%
x\right\rangle\,\left\langle g^{2}(|A^{*}|)y,y\right\rangle.$$
Theorem 2.15.
Let
$A,B,C,D\in{\mathbb{B}}({\mathscr{H}})$ and $f_{i},g_{i}\,\,(1\leq i\leq 4)$ be non-negative continues functions such that $f_{i}(x)g_{i}(x)=x\,\,(1\leq i\leq 4)$ for all $x\in[0,\infty)$. Then
$$\displaystyle\omega\left(\left[\begin{array}[]{cc}A&B\\
C&D\end{array}\right]\right)$$
$$\displaystyle\leq\max\left\{\left\|f_{1}^{2}(|A|)+g_{2}^{2}(|B^{*}|)+f_{3}^{2}%
(|C|)\right\|^{\frac{1}{2}},\left\|g_{4}(|D^{*}|)\right\|\right\}$$
$$\displaystyle\,\,+\max\left\{\left\|g_{1}(|A^{*}|)\right\|,\left\|f_{2}^{2}(|B%
|)+g_{3}^{2}(|C^{*}|)+f_{4}^{2}(|D|)\right\|^{\frac{1}{2}}\right\}.$$
Proof.
Let $T=\left[\begin{array}[]{cc}A&B\\
C&D\end{array}\right]$ and $\mathbf{x}=\left[\begin{array}[]{c}x_{1}\\
x_{2}\end{array}\right]$ be a unit vector (i.e., $\|x_{1}\|^{2}+\|x_{2}\|^{2}=1$). Then
$$\displaystyle\left|\left\langle T\mathbf{x},\mathbf{x}\right\rangle\right|$$
$$\displaystyle=$$
$$\displaystyle\left|\left\langle\left[\begin{array}[]{cc}A&B\\
C&D\end{array}\right]\left[\begin{array}[]{c}x_{1}\\
x_{2}\end{array}\right],\left[\begin{array}[]{c}x_{1}\\
x_{2}\end{array}\right]\right\rangle\right|$$
$$\displaystyle=$$
$$\displaystyle\left|\left\langle\left[\begin{array}[]{c}Ax_{1}+Bx_{2}\\
Cx_{1}+Dx_{2}\end{array}\right],\left[\begin{array}[]{c}x_{1}\\
x_{2}\end{array}\right]\right\rangle\right|$$
$$\displaystyle=$$
$$\displaystyle\left|\left\langle Ax_{1},x_{1}\right\rangle+\left\langle Bx_{2},%
x_{1}\right\rangle+\left\langle Cx_{1},x_{2}\right\rangle+\left\langle Dx_{2},%
x_{2}\right\rangle\right|$$
$$\displaystyle\leq$$
$$\displaystyle\left|\left\langle Ax_{1},x_{1}\right\rangle\right|+\left|\left%
\langle Bx_{2},x_{1}\right\rangle\right|+\left|\left\langle Cx_{1},x_{2}\right%
\rangle\right|+\left|\left\langle Dx_{2},x_{2}\right\rangle\right|$$
$$\displaystyle\left\langle f_{1}^{2}(|A|)x_{1},x_{1}\right\rangle^{\frac{1}{2}}%
\left\langle g_{1}^{2}(|A^{*}|)x_{1},x_{1}\right\rangle^{\frac{1}{2}}+\left%
\langle f_{2}^{2}(|B|)x_{2},x_{2}\right\rangle^{\frac{1}{2}}\left\langle g_{2}%
^{2}(|B^{*}|)x_{1},x_{1}\right\rangle^{\frac{1}{2}}$$
$$\displaystyle+\left\langle f_{3}^{2}(|C|)x_{1},x_{1}\right\rangle^{\frac{1}{2}%
}\left\langle g_{3}^{2}(|C^{*}|)x_{2},x_{2}\right\rangle^{\frac{1}{2}}+\left%
\langle f_{4}^{2}(|D|)x_{2}x_{2}\right\rangle^{\frac{1}{2}}\left\langle g_{4}^%
{2}(|D^{*}|)x_{2},x_{2}\right\rangle^{\frac{1}{2}}$$
$$\displaystyle\leq$$
$$\displaystyle\left(\left\langle f_{1}^{2}(|A|)x_{1},x_{1}\right\rangle+\left%
\langle g_{2}^{2}(|B^{*}|)x_{1},x_{1}\right\rangle+\left\langle f_{3}^{2}(|C|)%
x_{1},x_{1}\right\rangle+\left\langle g_{4}^{2}(|D^{*}|)x_{2},x_{2}\right%
\rangle\right)^{\frac{1}{2}}$$
$$\displaystyle\qquad\qquad\qquad\qquad\qquad\textrm{(\normalsize{by the Cauchy-%
Schwarz inequality})}$$
$$\displaystyle+\left(\left\langle g_{1}^{2}(|A^{*}|)x_{1},x_{1}\right\rangle+%
\left\langle f_{2}^{2}(|B|)x_{2},x_{2}\right\rangle+\left\langle g_{3}^{2}(|C^%
{*}|)x_{2},x_{2}\right\rangle+\left\langle f_{4}^{2}(|D|)x_{1},x_{1}\right%
\rangle\right)^{\frac{1}{2}}$$
$$\displaystyle=$$
$$\displaystyle\left(\left\langle\left(f_{1}^{2}(|A|)+g_{2}^{2}(|B^{*}|)+f_{3}^{%
2}(|C|)\right)x_{1},x_{1}\right\rangle+\left\langle g_{2}^{2}(|B^{*}|)x_{2},x_%
{2}\right\rangle\right)^{\frac{1}{2}}$$
$$\displaystyle+\left(\left\langle\left(f_{2}^{2}(|B|)+g_{3}^{2}(|C^{*}|)+f_{4}^%
{2}(|D|)\right)x_{2},x_{2}\right\rangle+\left\langle g_{1}^{2}(|A^{*}|)x_{1},x%
_{1}\right\rangle\right)^{\frac{1}{2}}$$
$$\displaystyle\leq$$
$$\displaystyle\left(\left\|f_{1}^{2}(|A|)+g_{2}^{2}(|B^{*}|)+f_{3}^{2}(|C|)%
\right\|\|x_{1}\|^{2}+\left\|g_{4}^{2}(|D^{*}|)\right\|\|x_{2}\|^{2}\right)^{%
\frac{1}{2}}$$
$$\displaystyle+\left(\left\|f_{2}^{2}(|B|)+g_{3}^{2}(|C^{*}|)+f_{4}^{2}(|D|)%
\right\|\|x_{2}\|^{2}+\left\|g_{1}^{2}(|A^{*}|)\right\|\|x_{1}\|^{2}\right)^{%
\frac{1}{2}}.$$
Let
$$\displaystyle\alpha=\left\|f_{1}^{2}(|A|)+g_{2}^{2}(|B^{*}|)+f_{3}^{2}(|C|)%
\right\|,\quad\beta=\left\|g_{4}^{2}(|D^{*}|)\right\|,$$
$$\displaystyle\mu=\left\|f_{2}^{2}(|B|)+g_{3}^{2}(|C^{*}|)+f_{4}^{2}(|D|)\right%
\|\quad\textrm{and}\,\,\lambda=\left\|g_{1}^{2}(|A^{*}|)\right\|.$$
It follows from
$$\underset{\|x_{1}\|^{2}+\|x_{2}\|^{2}=1}{\max}\left(\alpha\|x_{1}\|^{2}+\beta%
\|x_{2}\|^{2}\right)=\underset{\theta\in[0,2\pi]}{\max}(\alpha\sin^{2}\theta+%
\beta\cos^{2}\theta)=\max\{\alpha,\beta\}$$
and
$$\underset{\|x_{1}\|^{2}+\|x_{2}\|^{2}=1}{\max}\left(\lambda\|x_{1}\|^{2}+\mu\|%
x_{2}\|^{2}\right)=\underset{\theta\in[0,2\pi]}{\max}(\lambda\sin^{2}\theta+%
\mu\cos^{2}\theta)=\max\{\lambda,\mu\}$$
that
$$\displaystyle\left|\left\langle T\mathbf{x},\mathbf{x}\right\rangle\right|$$
$$\displaystyle\leq$$
$$\displaystyle\left(\left\|f_{1}^{2}(|A|)+g_{2}^{2}(|B^{*}|)+f_{3}^{2}(|C|)%
\right\|\|x_{1}\|^{2}+\left\|g_{4}^{2}(|D^{*}|)\right\|\|x_{2}\|^{2}\right)^{%
\frac{1}{2}}$$
$$\displaystyle+\left(\left\|f_{2}^{2}(|B|)+g_{3}^{2}(|C^{*}|)+f_{4}^{2}(|D|)%
\right\|\|x_{2}\|^{2}+\left\|g_{1}^{2}(|A^{*}|)\right\|\|x_{1}\|^{2}\right)^{%
\frac{1}{2}}$$
$$\displaystyle\leq$$
$$\displaystyle\max\left\{\left\|f_{1}^{2}(|A|)+g_{2}^{2}(|B^{*}|)+f_{3}^{2}(|C|%
)\right\|^{\frac{1}{2}},\left\|g_{4}^{2}(|D^{*}|)\right\|^{\frac{1}{2}}\right\}$$
$$\displaystyle+\max\left\{\left\|g_{1}^{2}(|A^{*}|)\right\|^{\frac{1}{2}},\left%
\|f_{2}^{2}(|B|)+g_{3}^{2}(|C^{*}|)+f_{4}^{2}(|D|)\right\|^{\frac{1}{2}}\right\}$$
$$\displaystyle=$$
$$\displaystyle\max\left\{\left\|f_{1}^{2}(|A|)+g_{2}^{2}(|B^{*}|)+f_{3}^{2}(|C|%
)\right\|^{\frac{1}{2}},\left\|g_{4}(|D^{*}|)\right\|\right\}$$
$$\displaystyle+\max\left\{\left\|g_{1}(|A^{*}|)\right\|,\left\|f_{2}^{2}(|B|)+g%
_{3}^{2}(|C^{*}|)+f_{4}^{2}(|D|)\right\|^{\frac{1}{2}}\right\}.$$
Taking the supremum over all unit vectors $\mathbf{x}$ we get the desired result.
∎
Corollary 2.16.
Let $A,B,C,D\in{\mathbb{B}}({\mathscr{H}})$. Then
$$\displaystyle\omega\left(\left[\begin{array}[]{cc}A&B\\
C&D\end{array}\right]\right)$$
$$\displaystyle\leq$$
$$\displaystyle\max\left\{\left\||A|^{2\alpha}+|B^{*}|^{2\gamma}+|C|^{2\mu}%
\right\|^{\frac{1}{2}},\left\||D^{*}|^{\omega}\right\|\right\}$$
$$\displaystyle+\max\left\{\left\||A^{*}|^{\beta}\right\|,\left\||B|^{2\zeta}+|C%
^{*}|^{2\nu}+|D|^{\kappa}\right\|\right\},$$
where $\alpha+\beta=\gamma+\zeta=\mu+\nu=\omega+\kappa=1$. In particular,
$$\displaystyle\omega\left(\left[\begin{array}[]{cc}A&B\\
0&0\end{array}\right]\right)\leq\max\left\{\left\||A^{*}|^{\beta}\right\|,%
\left\||B|^{\zeta}\right\|\right\}+\left\||A|^{2\alpha}+|B^{*}|^{2\gamma}%
\right\|^{\frac{1}{2}},$$
in which $\alpha+\beta=\gamma+\zeta=1$.
References
[1]
A. Abu-Omar and F. Kittaneh, A numerical radius inequality involving the
generalized Aluthge transform, Studia Math. 216 (2013), 69–75.
[2]
A. Abu-Omar and F. Kittaneh, Generalized spectral radius and norm inequalities for Hilbert space operators, International Journal of Mathematics Vol. 26, No. 11 (2015) 1550097 (9 pages).
[3]
O. Axelsson, H. Lu and B. Polman, On the numerical radius of matrices and its application to iterative solution methods, Linear Multilinear Algebra. 37 (1994), 225–238.
[4]
M. Bakherad and M.S. Moslehian, Complementary and refined inequalities of Callebaut inequality for operators, Linear Multilinear Algebra 63 (2015), no. 8, 1678–1692.
[5]
K. Davidson and S.C. Power, Best approximation in $C^{*}$-algebras, J. Reine Angew. Math. 368 (1986) 43-62.
[6]
K.E. Gustafson and D.K.M. Rao, Numerical Range, The Field of Values of Linear Operators and
Matrices, Springer, New York, 1997.
[7]
P.R. Halmos, A Hilbert Space Problem Book, 2nd ed., springer, New York, 1982.
[8]
O. Hirzallah, F. Kittaneh and K. Shebrawi,
Numerical radius inequalities for commutators of Hilbert space
operators, Numer. Funct. Anal. Optim. 32 (2011) 739–749.
[9]
O. Hirzallah, F. Kittaneh and K. Shebrawi,
Numerical radius inequalities for certain $2\times 2$,
operator matrices, Integral Equations Operator Theory, 71 (2011) 129–147.
[10]
F. Kittaneh, Notes on some inequalitis for Hilbert space operators,
Publ. Res. Inst. Math. Sci. 24 (2) (1988), 283–293.
[11]
F. Kittaneh, A numerical radius inequality and an estimate for the numerical radius of the Frobenius companion matrix, Studia Math. 158 (2003), 11–17.
[12]
K. Okubo, On weakly unitarily invariant norm and the Aluthge transformation, Linear Algebra Appl. 371 (2003) 369–375.
[13]
K. Shebrawi, Numerical radius inequalities for certain $2\times 2$ operator matrices $II$, Linear Algebra Appl. (2017), http://dx.doi.org/10.1016/j.laa.2017.02.019
[14]
K. Shebrawi and H. Albadawi, Numerical radius and operator norm inequalities, J. Math. Inequal. (2009) Article ID 492154, 11 pages.
[15]
T. Yamazaki, On upper and lower bounds of the numerical radius and an equality condition, Studia Math. 178 (2007), 83–89. |
Abstract
The leading contributions to the electron (or muon) electric dipole
moment due to CP violation in the charged Higgs sector are at the
two–loop level.
A careful model-independent analysis of the heavy fermion
contribution is provided.
We also consider some specific scenarios to
demonstrate how charged Higgs sector CP violation can naturally give
rise to large electric dipole moments.
Numerical results show that the
electron electric dipole moment in such models can lie at the
experimentally accessible level.
March 1997
hep-ph/9703435
NHCU-HEP-97-5
UICHEP-TH/97-3
Electron Electric Dipole Moment from
CP Violation in
the Charged Higgs Sector
David Bowser-Chao${}^{(1)}$,
Darwin Chang${}^{(2,3)}$,
and
Wai-Yee Keung${}^{(1)}$
${}^{(1)}$Physics Department, University of Illinois at Chicago,
IL 60607-7059, USA
${}^{(2)}$Physics Department,
National Tsing-Hua University, Hsinchu 30043, Taiwan, R.O.C.
${}^{(3)}$Institute of Physics, Academia Sinica, Taipei, R.O.C.
PACS numbers: 13.40.Fn, 11.30.Er, 12.15.Cc, 14.80.Dq
Introduction
Experiment has established that neither parity (P) nor charge
conjugation (C) are unbroken symmetries. Kaon
physics show that CP also fails to be an exact symmetry. The
CPT theorem then implies that time-reversal (T) is broken as
well, leading to expectation of a T-odd electric dipole moment (EDM)
for one or more of the elementary particles.
The Standard Model (SM) of electroweak interactions explains CP
violation in the $K-{\bar{K}}$ system as the result of a single complex
phase in the Cabibbo-Kobayashi-Maskawa (CKM) matrix. It also predicts an
electron EDM $d_{e}$ of about $8\times 10^{-41}{\rm e}\cdot{\rm cm}$
[1]
and a muon EDM of about $2\times 10^{-38}{\rm e}\cdot{\rm cm}$, while the neutron EDM (calculated from the up and down quark
EDMs) is estimated [2] to be less than $10^{-31}{\rm e}\cdot{\rm cm}$,
The experimental limits (given at 95% C.L.) are several orders of
magnitude above these predictions, with the limit on the electron EDM
$|d_{e}|<6.2\times 10^{-27}{\rm e}\cdot{\rm cm}$ [3]. The limit
on $d_{\mu}$ is even further removed from the SM prediction, with $|d_{\mu}|<1.1\times 10^{-18}{\rm e}\cdot{\rm cm}$ [4],
although there is a proposal to measure
the muon EDM down to $10^{-24}$ e$\cdot$cm [5].
The neutron EDM limit is $|d_{n}|<11\times 10^{-26}{\rm e}\cdot{\rm cm}$ [6]. Clearly, measurement of a non-zero electron, muon, or
neutron EDM close to current or proposed limits would point to physics
beyond the Standard Model.
New sources of CP violation can come from complex couplings or vacuum
expectation values (VEV) associated with the Higgs boson sector. A
significant EDM for elementary fermions can be
generated if CP violation is mediated by neutral Higgs-boson exchange
[7, 8, 9, 10, 11]. Dominant contributions
come from one-loop or two-loop diagrams. The one-loop terms are
proportional to $(m/v)^{3}$ (with one factor of $m/v$ due to an internal
mass insertion), while the two-loop terms are
proportional[9] to $m/v$, with $m$ being the fermion mass
and $v=246$ GeV the vacuum expectation value of the SM Higgs field.
The one-loop contributions are thus strongly suppressed relative to the
two-loop terms, by a factor of $(m/v)^{2}$.
Exhaustive studies[10, 12] have been carried out at the
two loop level on the electron EDM generated by CP violation in the neutral
Higgs-boson sector.
The contributions containing a heavy fermion or gauge boson loop
were considered[10]
and, within an $SU(2)\times U(1)$ theory, are more or less
model-independent. Additional, relatively more model-dependent,
contributions involving a physical charged
Higgs-boson loop (with a CP violating charged-Higgs-neutral-Higgs
coupling) have been considered in Ref.[12]. On the other hand,
the corresponding contribution to $d_{e}$ due solely
to CP violation in charged Higgs sector has not been studied in
the literature, even though this category of CP violating mechanism has been
emphasized for other phenomenological effects such as
top quark decay[13], the neutron electric dipole
moment[14, 15], and $\Gamma(b\to s\gamma)$[16].
With charged Higgs sector CP violation involving fermions, the one-loop
contribution is
suppressed as in the neutral Higgs case but suffers an additional
factor of $m_{\nu}/m_{e}$, where $m_{\nu}$ is the mass of the electron
neutrino.
If no right-handed neutrino exists, or the neutrino is
massless, the two-loop diagrams are unequivocally the leading
contribution. It is also important to note that the recent measurement
of the decay rate of $b\to s\gamma$ by the CLEO collaboration
[17] stringently constrains[18] the mass of the $H^{\pm}$
only in the 2HDM II. The constraint is easily evaded in 2HDM
III[16, 19, 20] and other extensions.
In this letter, we make a model-independent study of the two–loop
contribution to $d_{e}$ and $d_{\mu}$
from charged Higgs exchange between fermions.
Useful formulas are given. We
discuss specific models to see how this type of CP
nonconservation can arise.
Numerical results show the electron electric
dipole moment can naturally lie within reach of experiment.
We also
present results for the corresponding contributions to
$d_{n}$.
General Formalism
We first comment on the model-independence of our analysis,
which considers only those graphs that involve CP violation from
charged Higgs exchange between
fermions. There are other possibilities,
such as quartic charged Higgs vertices involving three or four distinct
charged Higgs bosons with complex coupling, but these are strongly
dependent on model details and less amenable to
model-independent parameterization.
Secondly, one may have several charged Higgs bosons,
but barring significant degeneracy,
cancellation among contributions from the various
charged Higgs should be mild. For simplicity we only consider
contributions from the lightest charged Higgs.
We shall ignore model dependent neutral
Higgs sector CP violation contributions and
CP violation involving both neutral and charged Higgs boson,
as our concern here is with exclusively
charged Higgs sector CP violation effects.
We shall also omit terms
where the neutral Higgs bosons participate only incidentally, serving as
a leg of an internal loop but not contributing a CP phase.
We expect
that these (perhaps non-negligible) contributions
should not strongly cancel with the part studied here; thus our
analysis should furnish a reasonable lower limit to $d_{e}$ for general
theories with CP violating charged Higgs exchange between fermions.
We parametrize the charged Higgs sector CP violation as follows:
$${\cal L}={g\over\sqrt{2}}\left({m_{t}c_{t}\over M_{W}}\bar{t}_{R}b_{L}H^{+}+{m%
_{e}c_{e}\over M_{W}}\bar{\nu}_{L}e_{R}H^{+}+\bar{t}_{L}\gamma^{\alpha}b_{L}W^%
{+}_{\alpha}+\bar{\nu}_{L}\gamma^{\alpha}e_{L}W^{+}_{\alpha}\right)$$
$$-m_{e}\bar{e_{L}}e_{R}-m_{t}\bar{t_{L}}t_{R}+{\rm H.c.}$$
(1)
Note that in our model-independent analysis, one needs
not specify the origin of the CP violation (e.g., explicit or
spontaneous CP violation).
We only illustrate the most important
contribution from the top-bottom generation; our study can easily be
generalized to the three generations case. The bottom quark mass $m_{b}$
is also set to zero. The
complex mixing parameters $c_{t}$ and $c_{e}$ signal deviations from the
2HDM II. If $c_{t}c_{e}^{*}$ has a non-zero imaginary part, the phase is
intrinsic to the lagrangian and cannot be rotated away by redefinition
of any or all of the fields in Eq.(1).
The two-loop charged Higgs contribution involves
Feynman diagrams such as the one shown in Fig. 1. We first present a
simple expression for the one-loop sub-diagram with fermion in the loop,
that is, the truncated three-point Green’s function $\Gamma^{\mu\nu}=\langle 0|[H^{-}(p)A^{\mu}(k){W^{+}}^{\nu}(-q)]_{+}|0\rangle$. We note
that $\Gamma^{\mu\nu}$ is the off-shell extension of the amplitude
(and Feynman diagrams) for
$H^{-}\to W^{-}\gamma$ given in Ref.[21], and that both
$\Gamma^{\mu\nu}$ and its charge conjugate contribute to $d_{e}$.
We consider only
the (gauge-invariant) set of terms
in $\Gamma^{\mu\nu}$ involving CP violation from
charged Higgs exchange between fermions.
We have verified that our results are gauge-independent,
but the calculation simplifies greatly
in the non-linear $R_{\xi}$ gauge[26]. Since the EDM is defined
in the soft photon limit, only the leading term in $k$ is kept. We also
work to lowest
order in $(m_{e}/M_{W})$, to which approximation the separation of
the calculation into $\Gamma^{\mu\nu}$ and its insertion in the full
two-loop graph is gauge-invariant. We thus obtain:
$$\Gamma^{\mu\nu}={3eg^{2}\over 16\pi^{2}}{c_{t}\over M_{W}}[(g^{\mu\nu}q\cdot k%
-q^{\mu}k^{\nu})S+iP\varepsilon^{\mu\nu\alpha\beta}p_{\alpha}q_{\beta}]\ ,$$
(2)
$$S=\int_{0}^{1}{q_{t}(1-y)^{2}+q_{b}y(1-y)\over 1-yq^{2}/m_{t}^{2}}dy\ ,\quad P%
=\int_{0}^{1}{q_{t}(1-y)+q_{b}y\over 1-yq^{2}/m_{t}^{2}}dy\ ,$$
(3)
and the quark charges are denoted $q_{t},q_{b}$.
The above vertex is further connected to the lepton propagator to
produce EDM (see Fig. 1).
$$d_{e}=\left({3eg^{2}\over 32\pi^{2}}\right)\left({g^{2}\over 32\pi^{2}M_{W}}%
\right)\left({m_{e}\over M_{W}}\right){\rm Im}(c_{t}^{*}c_{e})\left(q_{t}F_{t}%
+q_{b}F_{b}\right)\ .$$
(4)
Here the form factors $F_{\alpha}$ for $\alpha=t,b$ are given by
$$F_{\alpha}=\int_{0}^{\infty}\int_{0}^{1}{m_{t}^{2}f_{\alpha}(2-y)\;dy\,Q^{2}dQ%
^{2}\over(M_{H^{+}}^{2}+Q^{2})(m_{t}^{2}+yQ^{2})(M_{W}^{2}+Q^{2})}\ ;\quad f_{%
t}=(1-y)\ ,\quad f_{b}=y\ .$$
(5)
The integrations can be carried
through analytically,
$$T(z)={1-3z\over z^{2}}{\pi^{2}\over 6}-\left({1\over z}-{5\over 2}\right){\rm
ln%
\ }z-{1\over z}-\left(2-{1\over z}\right)\left(1-{1\over z}\right){\rm Sp}(1-z%
)\ ,$$
(6)
$$B(z)={1\over z}+{2z-1\over z^{2}}{\pi^{2}\over 6}+\left({3\over 2}-{1\over z}%
\right){\rm ln\ }z-\left(2-{1\over z}\right){1\over z}{\rm Sp}(1-z)\ ,$$
(7)
$$F_{t}={T(z_{H})-T(z_{W})\over z_{H}-z_{W}}\ ,\quad F_{b}={B(z_{H})-B(z_{W})%
\over z_{H}-z_{W}}\ ,$$
(8)
with
$z_{H}=M_{H^{+}}^{2}/m_{t}^{2}$ and $z_{W}=M_{W}^{2}/m_{t}^{2}$,
and the Spence function is defined by
${\rm Sp}(z)=-\int_{0}^{z}t^{-1}{\rm\ln}(1-t)dt$
with the normalization ${\rm Sp}(1)=\pi^{2}/6$.
Models
To illustrate how easily this
mechanism can give rise to a measurable electron EDM,
we consider some specific models.
We first note that the
charged Higgs contribution to $d_{e}$ vanishes to two loops in the
much-studied simple extension of the Standard Model —
the two Higgs-doublet Model (2HDM)[22], with the softly broken
discrete symmetry[23] imposed
to enforce natural flavor conservation (NFC).
At one loop level single charged
Higgs exchange between fermions does not
violate CP[8]. As for the two loop contribution, setting the
scalar doublets $\phi_{i}=(\phi^{+}_{i},\phi^{0}_{i}),i=1,2$ to have
purely real VEVs $\langle\phi^{0}_{i}\rangle_{0}=v_{i}$,
in the unitary gauge the charged Higgs propagators
$\langle\phi^{+}_{1}\phi^{+{\dagger}}_{2}\rangle_{0},\langle\phi^{+}_{1}\phi^{+%
{\dagger}}_{1}\rangle_{0},\langle\phi^{+}_{2}\phi^{+{\dagger}}_{2}\rangle_{0}$
are purely real[7, 8].
Ignoring
the CKM matrix, the only[8] complex coupling and thus
CP phase in the
lagrangian involving the charged Higgs appears in
${\rm Re}[h(\phi_{1}^{{\dagger}}\phi_{2}-v_{1}v_{2})^{2}]$;
thus charged Higgs CP violation in this model
necessarily also involves the pseudoscalar neutral Higgs.
There are, however, several other simple models which can easily contain
sufficient CP violation to produce an electron EDM at an
observable level, through exchange of a charged Higgs boson between fermions.
2HDM III.
In Model III with two Higgs doublets[24, 19], we can choose a basis
so that
$\langle\phi^{0}\rangle={v\over\sqrt{2}}$ and
$\langle\phi^{\prime 0}\rangle=0$.
Then $\phi$ mimics the SM Higgs, and $\phi^{\prime}$ produces new
physics beyond the SM; the physical charged Higgs $H^{+}$ is just
$\phi^{\prime+}$.
The Yukawa Lagrangian for Model III is
$$-{\cal L}_{Y}=\eta_{ij}^{U}\bar{Q}_{iL}\tilde{\phi}U_{jR}+\eta_{ij}^{D}\bar{Q}%
_{iL}\phi D_{jR}+\xi_{ij}^{U}\bar{Q}_{iL}\tilde{\phi^{\prime}}U_{jR}+\xi_{ij}^%
{D}\bar{Q}_{iL}\phi^{\prime}D_{jR}$$
(9)
$$\quad\quad+\eta_{ij}^{E}\bar{L}_{iL}\phi E_{jR}+\xi_{ij}^{E}\bar{L}_{iL}\phi^{%
\prime}E_{jR}\quad\;+{\rm H.c.}$$
Here $i,j$ are generation indices. Coupling matrices $\eta$ and
$\xi$ are, in general, non-diagonal.
$Q_{iL}$ and $L_{iL}$ are the
left-handed SU(2) doublets for quarks and leptons.
$U_{jR}$, $D_{jR}$, and $E_{R}$ are the right-handed SU(2) singlets for
up-type and down-type quarks and charged leptons respectively.
$\langle\phi\rangle$ generates all fermion mass matrices which are
diagonalized by bi-unitary transformations, e.g.
$M_{U}={\rm diag}(m_{u},m_{c},m_{t})={v\over\sqrt{2}}({\cal L}^{U})^{\dagger}%
\eta^{U}(\cal R^{U})$.
In terms of the mass eigenstates $U$, $D$, $E$, and $N$(neutrinos), the
relevant
charged Higgs interaction is given by
$${\cal L}_{H^{+}}=-H^{+}\bar{U}\biggr{[}V_{\rm KM}{\hat{\xi}}^{D}\hbox{$1\over 2%
$}(1+\gamma^{5})-{\hat{\xi}}^{U\dagger}V_{\rm KM}\hbox{$1\over 2$}(1-\gamma^{5%
})\biggr{]}D-H^{+}\bar{N}{\hat{\xi}}^{E}\hbox{$1\over 2$}(1+\gamma^{5})E+{\rm H%
.c.}\;\;,$$
(10)
with
the CKM matrix
$V_{\rm CKM}=({\cal L}_{U})^{\dagger}({\cal L}_{D})$, and
$\hat{\xi}^{P}=({\cal L}^{P})^{\dagger}\xi^{P}({\cal R}^{P})$ (for
$P=U,D,E$).
Tree-level flavor changing neutral currents (FCNC)
are implied by non-zero off-diagonal elements of the
matrices $\hat{\xi}^{U,D,E}$.
We adopt the simple ansatz[24]
$\hat{\xi}^{U,D,E}_{ij}=\lambda_{ij}{g\sqrt{m_{i}m_{j}}}/({\sqrt{2}M_{W}})\ .$
The mass hierarchy ensures that FCNC within the first two
generations are naturally suppressed by small quark masses, while
a larger freedom is allowed for FCNC involving the third
generations. Here $\lambda_{ij}$ can be
$O(1)$ and complex.
CP is already not a symmetry even if we restrict our attention to the
flavor conserving diagonal entries of $\lambda_{ii}$.
We consider only the third generation quark contribution and set $(V_{\rm CKM})_{tb}=1$.
The parameters $c_{t}$
and $c_{e}$ in Eq.(1) are given as,
$c_{e}=-\lambda_{ee}\ ,\quad c_{t}=\lambda_{tt}^{*}\ .$
While $\lambda_{tt}$ cannot be significantly larger than $O(1)$ without
producing strong coupling to the top quark, clearly $\lambda_{ee}$ can
be much larger than $O(1)$ so that ${\rm Im}(c_{t}^{*}c_{e})\gg 1$ is quite
allowed.
3HDM.
Charged Higgs sector CP violation can occur in the
three Higgs doublet model[25, 15].
The first two doublets $\phi_{1}$ and $\phi_{2}$ are responsible for the
masses of the $b$–like quarks and the $t$–like quarks respectively.
The charged leptons $e,\mu$ and $\tau$ only couple to $\phi_{1}$.
The last doublet $\phi_{3}$ does not couple to the known fermions.
This assignment naturally preserves NFC.
The mass eigenstates $H_{1}^{+}$ and $H_{2}^{+}$ together with the
unphysical charged Goldstone boson $H_{3}^{+}$ are linear combinations of
$\phi_{i}^{+}$:
$\phi_{i}^{+}=\sum_{j=1}^{3}U_{ij}H_{j}^{+}\quad(i=1,2,3)$.
As with the CKM matrix for three quark generations in the SM,
the mixing amplitude $U_{ij}$ matrix generally
contains a single non-zero complex phase,
which gives rise to CP nonconservation through the
Yukawa couplings,
$c_{t}=U_{2i}(v/v_{2}),\;c_{e}=U_{1i}(v/v_{1})$.
In the approximation that the lightest Higgs dominates, the index $i$
refers to the lightest charged Higgs. As with the 2HDM III, ${\rm Im}(c_{t}^{*}c_{e})$ can be much larger than one — which occurs here if
$v_{1}\ll v_{2}$ (the possibility $v_{2}\ll v_{1}$ is constrained by
maintaining perturbative coupling to the top quark).
Discussion
We have analyzed the heavy fermion contribution
to the electron EDM due to CP
violation in the charged Higgs sector. From the general structure of the
typical models discussed above, we have shown that the relevant CP
violating parameter Im$(c_{t}^{*}c_{e})$ can be of order one or larger. In
Fig. 2, we show the dependence of the electron EDM on $M_{H^{+}}$ for the
case Im$(c_{t}^{*}c_{e})=1$. The size of $d_{e}$ is naturally around $10^{-26}$
e$\cdot$cm, around the current limit.
As noted above, in any specific model there may be other
contributions to $d_{e}$, but in the absence of accidental
strong cancellation, we expect that the results
presented here reasonably estimate a lower limit on $d_{e}$
which applies to a very wide class of CP violating models.
In such case, Fig. 2 may
be used, for example, to rule out $m_{H^{+}}>200\,{\rm GeV}$ for
Im$(c_{t}^{*}c_{e})\geq 1$.
The muon EDM can be easily obtained by the replacements $m_{e}\to m_{\mu}$
and Im$(c_{t}^{*}c_{e})\to$ Im$(c_{t}^{*}c_{\mu})$. In this case, the analogous
case of
Im$(c_{t}^{*}c_{\mu})=1$ would lead to
an observable muon EDM,
assuming the proposed future sensitivity down to
$d_{\mu}=10^{-24}$ e$\cdot$cm.
Finally, we can carry over the calculation for $d_{e}$ to estimate
contributes to $d_{n}$, using SU(6) relations [27]:
$d_{n}={1\over 3}(4d_{d}-d_{u})$,
where we obtain the down and up quark EDMs with replacements as made
for $d_{\mu}$, but with an additional factor $\eta_{q}$ multiplying both
$d_{d}$ and $d_{u}$ coming from QCD evolution of the quark mass and the
quark EDM [28]:
$\eta_{q}=q(m_{t},m_{b})^{16/23}q(m_{b},m_{c})^{16/25}q(m_{c},\mu)^{16/27},$ with $q(m_{a},m_{b})=\alpha_{s}(m_{a}^{2})/\alpha_{s}(m_{b}^{2})$.
There are sizable uncertainties coming from the quark masses and the
extraction of $d_{n}$ from $d_{d}$ and $d_{u}$, but the resulting neutron EDM
should be $d_{n}\approx 10^{-27}\,(m_{d}/m_{e}){\rm Im}(c_{t}^{*}c_{d})\,{\rm e}\cdot{\rm
cm}$ for $m_{H^{+}}\approx 100$ GeV (ignoring the up quark
contribution). This contribution would reach the observable limit for
${\rm Im}(c_{t}^{*}c_{d})>6$. In contrast to the case of $d_{e}$ or $d_{\mu}$,
there is a sizable contribution from the charged Higgs boson through the
three-gluon operator [14, 29]. The relative magnitudes are highly
model-dependent: in 2HDM III, the three-gluon operator may vanish even
while the two-loop contributions presented here are non-zero, if $c_{t}$
is purely real but $c_{e}$ remains complex. In the 3HDM, however, the two
contributions to $d_{n}$ are either both zero or both non-zero. In any
case, barring strong cancellations, our result places a limit of
$m_{H^{+}}>100$ GeV for ${\rm Im}(c_{t}^{*}c_{d})=6$.
D. B.-C. and W.-Y. K. are supported by a grant from the Department of
Energy, and
D. C. by a grant from the National Science Council of R.O.C. We also wish to
thank I. Phillips, T-C. Yuan and C.-S. Li who were
involved in an earlier stage of this work.
Figures
Fig. 1.
A typical two-loop Feynman diagram for the electron EDM
due to charged Higgs sector CP violation. The other diagrams for the
one-loop subgraph $H^{-}\rightarrow W^{-}\gamma$ may be found
in Ref.[21].
Fig. 2.
Model-independent contributions to $d_{e}$ versus $M_{H^{+}}$ for
Im$(c_{t}^{*}c_{e})=1/2,\,1$. The horizontal line denotes the current
experimental limit.
References
[1]
D. Chang, W.–Y. Keung, and J. Liu Nucl. Phys. B355 295 (1991);
I.B. Khriplovich and M. Pospelov,
Sov. J. Nucl. Phys. 53, 638 (1991);
M. Booth, University of Chicago Report,
EFI-93-01, (1993), hep-ph/9301293.
[2]
E. P. Shabalin, Usp. Fiz. Nauk 139, 553 (1993)
[Sov. Phys. Usp. 26, 297 (1983)].
B. H. J. McKellar, S. R. Choudhury, X.-G. He, and S. Pakvasa,
Phys. Lett. B197, 556 (1987), and
references therein.
[3]
E. Commins et al., Phys. Rev. A 50, 2960 (1994);
K. Abdullah et al., Phys. Rev. Lett. 65, 2347 (1990).
[4]
J. Bailey et al., J. Phys. G4, 345 (1978);
J. Bailey et al., Nucl. Phys. B150, 1 (1979).
[5]
Y. Semertzidis et al.,
AGS Letter of Intent-Search for an Electric Dipole Moment of Muon
at the $10^{-24}$ level, submitted to the Program Advisory Committee
at BNL (1997).
[6]
I.S. Altarev et al., Phys. Lett. B 276, 242 (1992);
K.F. Smith et al., Phys. Lett. B 234, 191 (1990).
[7]
S. Weinberg, Phys. Rev. Lett. 63, 2333 (1989).
[8]
S. Weinberg, Phys. Rev. D 42, 860 (1990).
[9]
S. Barr and A. Zee, Phys. Rev. Lett. 65, 21 (1990);
and (E)-$ibid.$ 65, 2920 (1990).
[10]
R.G. Leigh, S. Paban and R.-M. Xu, Nucl. Phys. B352, 45 (1991);
J. Gunion and R. Vega, Phys. Lett. B251, 157 (1990);
D. Chang, W.-Y. Keung and T.C. Yuan, Phys. Rev. D 43, 14 (1991).
[11]
V. Barger, A. Das, and Chung Kao, hep-ph/9611344.
[12]
C. Kao and R.-M. Xu, Phys. Lett. B 296, 435 (1992).
[13]
R. Cruz, B. Grzadkowski, J.F. Gunion,
Phys. Lett. B 289, 440 (1992).
[14]
G. Boyd, A. Gupta, S. Trivedi, and M. Wise,
Phys. Lett. B241, 584 (1990);
D. Chang, W.–Y. Keung, C.S. Li and T.C. Yuan,
Phys. Lett. 241 589, (1990); see also
D. Dicus, Phys. Rev. D41, 999 (1990).
[15]
N. G. Deshpande and E. Ma, Phys. Rev. D 16, 1583 (1977).
[16]
P. Krawczyk and S. Pokorski, Nucl. Phys. B364, 10 (1991);
Wolfenstein and Y.L. Wu, Phys. Rev. Lett. 73, 2809 (1994).
[17]
CLEO Collaboration, M.S. Alam et al.,
Phys. Rev. Lett. 74, 2885 (1995).
[18]
R.G. Ellis, G.C. Joshi, M. Matsuda,
Phys. Lett. B 179, 119 (1986);
T. Rizzo, Phys. Rev. D 38, 820 (1988);
B. Grinstein and M. Wise, Phys. Lett. B 201, 274 (1988);
W.-S. Hou and R. Wiley, Phys. Lett. B 202, 591 (1988);
C.-Q. Geng and J. Ng, Phys. Rev. D 38, 2858 (1988);
V. Barger, J.L. Hewett and R.J.N. Phillips,
Phys. Rev. D 41, 3421 (1990).
[19]
G.C. Branco and M.N. Rebelo, Phys. Lett. B 160, 117 (1985);
J. Liu and L. Wolfenstein, Nucl. Phys. B289, 1 (1987).
[20]
D. Bowser-Chao, K. Cheung, and W.-Y. Keung, in preparation.
[21]
S. Raychaudhuri and A. Raychaudhuri, Phys. Lett. B 297, 159 (1992).
[22]
J.F. Donoghue and L.-F. Li, Phys. Rev. D 19, 945 (1979);
L. Hall and M. Wise, Nucl. Phys. B187, 397 (1981).
[23]
S.L. Glashow and S. Weinberg, Phys. Rev. D 15, 1958 (1977).
[24]
T.P. Cheng and M. Sher, Phys. Rev. D 35, 3484 (1987);
W.S. Hou, Phys. Lett. B296, 179 (1992);
L.J. Hall and S. Weinberg, Phys. Rev. D 48, R979 (1993);
D. Chang, W.S. Hou, and W.–Y. Keung, Phys. Rev. D 48, 217 (1993);
A. Das and C. Kao, Phys. Lett. B 296, 435 (1996);
D. Atwood, L. Reina, A. Soni, Phys. Rev. D 55, 3156 (1997).
[25]
S. Weinberg, Phys. Rev. Lett. 37, 657 (1976);
G. Branco, Phys. Rev. Lett. 44, 504 (1980).
[26]
K. Fujikawa, Phys. Rev. D 7, 393 (1973);
M.B. Gavela, G. Girardi, and C. Sorba,
Nucl. Phys. B193, 257 (1981);
J.C. Romão and A. Barroso, Phys. Rev. D 35, 2836 (1987).
[27]
X.-G. He, B. H. J. McKellar, and S. Pakvasa,
Int. J. Mod. Phys. A4, 5011 (1989);
D. Chang, in “The Standard Model and Beyond”,
Proceedings of the
Ninth Symposium on Theoretical
Physics, edited by J. E. Kim, World Scientific Co., Singapore, 1991.
[28]
J. Polchinski and M. B. Wise, Phys. Lett. B125 393 (1983).
[29]
Y. Grossman, Y. Nir, R. Rattazzi, to appear in the Review Volume
“Heavy Flavours II”, eds. A..J. Buras and M. Lindner, Advanced Series
on Directions in High Energy Physics, World Scientific Publishing Co.,
Singapore, hep-ph/9701231. |
An integer
construction of infinitesimals: Toward a theory of Eudoxus hyperreals
Alexandre Borovik
School of Mathematics, University
of Manchester, Oxford Street, Manchester, M13 9PL, United Kingdom
[email protected]
,
Renling Jin
Department of Mathematics, College of
Charleston, South Carolina 29424, USA
[email protected]
and
Mikhail G. Katz${}^{0}$
Department of Mathematics, Bar
Ilan University, Ramat Gan 52900 Israel
[email protected]
Abstract.
A construction of the real number system based on almost homomorphisms
of the integers ${\mathbb{Z}}$ was proposed by Schanuel, Arthan, and others. We
combine such a construction with the ultrapower or limit ultrapower
construction, to construct the hyperreals out of integers. In fact,
any hyperreal field, whose universe is a set, can be obtained by such
a one-step construction directly out of integers. Even the maximal
(i.e., On-saturated) hyperreal number system described by
Kanovei and Reeken (2004) and independently by Ehrlich (2012) can be
obtained in this fashion, albeit not in NBG. In NBG, it
can be obtained via a one-step construction by means of a definable
ultrapower (modulo a suitable definable class ultrafilter).
Key words and phrases:Eudoxus; hyperreals; infinitesimals; limit ultrapower;
universal hyperreal field
2000 Mathematics Subject Classification: Primary 26E35; Secondary 03C20
00footnotetext: Supported by the Israel Science Foundation grant
1517/12
Contents
1 From Kronecker to Schanuel
2 Passing it through an ultraproduct
3 Cantor, Dedekind, and Schanuel
4 Constructing an infinitesimal-enriched continuum
5 Formalization
6 Limit ultrapowers and definable ultrapowers
7 Universal and On-saturated Hyperreal Number Systems
1. From Kronecker to Schanuel
Kronecker famously remarked that, once we have the natural numbers in
hand, “everything else is the work of man” (see Weber [49]).
Does this apply to infinitesimals, as well?
The exposition in this section follows R. Arthan [7]. A
function $f$ from ${\mathbb{Z}}$ to ${\mathbb{Z}}$ is said to be an almost
homomorphism if and only if the function $d_{f}$ from ${\mathbb{Z}}\times{\mathbb{Z}}$
to ${\mathbb{Z}}$ defined by
$$d_{f}(p,q)=f(p+q)-f(p)-f(q)$$
has bounded
(i.e., finite) range, so that for a suitable integer $C$, we
have $|d_{f}(p,q)|<C$ for all $p,q\in{\mathbb{Z}}$. The set denoted
$${{\mathbb{Z}}}\rightarrow{{\mathbb{Z}}}$$
(1.1)
of all functions from ${\mathbb{Z}}$ to ${\mathbb{Z}}$ becomes an abelian group if we add
and negate functions pointwise:
$$(f+g)(p)=f(p)+g(p),\qquad(-f)(p)=-f(p).$$
It is easily checked that if $f$ and $g$ are almost homomorphisms then
so are $f+g$ and $-f$. Let $S$ be the set of all almost homomorphisms
from ${\mathbb{Z}}$ to ${\mathbb{Z}}$. Then $S$ is a subgroup of ${{\mathbb{Z}}}\rightarrow{{\mathbb{Z}}}$.
Let us write $B$ for the set of all functions from ${\mathbb{Z}}$ to ${\mathbb{Z}}$ whose
range is bounded. Then $B$ is a subgroup of $S$. The ‘‘Eudoxus
reals’’ are defined as follows.111The term “Eudoxus reals” has gained some
currency in the literature, see e.g., Arthan [7].
Shenitzer [46, p. 45] argues that Eudoxus anticipated 19th
century constructions of the real numbers. The attribution of such
ideas to Eudoxus, based on an interpretation involving Eudoxus,
Euclid, and Book 5 of The Elements, may be historically
questionable.
Definition 1.1.
The abelian group ${\mathbb{E}}$ of Eudoxus reals is the quotient group
$S/B$.
Elements of ${\mathbb{E}}$ are equivalence classes, $[f]$ say, where $f$ is an
almost homomorphism from ${\mathbb{Z}}$ to ${\mathbb{Z}}$, i.e., $f$ is a function from
${\mathbb{Z}}$ to ${\mathbb{Z}}$ such that $d_{f}(p,q)=f(p,q)-f(p)-f(q)$ defines a
function from ${\mathbb{Z}}\times{\mathbb{Z}}$ to ${\mathbb{Z}}$ whose range is bounded. We have
$[f]=[g]$ if and only if the difference $f-g$ has bounded range,
i.e., if and only if $|{f(p)-g(p)}|<C$ for some $C$ and all $p$ in
${\mathbb{Z}}$.
The addition and additive inverse in ${\mathbb{E}}$ are induced by the pointwise
addition and inverse of representative almost homomorphisms:
$$[f]+[g]=[f+g],\qquad-[f]=[-f]$$
where $f+g$ and $-f$ are defined by
$$(f+g)(p)=f(p)+g(p)$$
and
$$(-f)(p)=-f(p)$$
for all $p$ in ${\mathbb{Z}}$.
The group ${\mathbb{E}}$ of Eudoxus reals becomes an ordered abelian
group if we take the set $P\subset{\mathbb{E}}$ of positive elements to be
$$P=\left\{[f]\in{\mathbb{E}}:\sup_{m\in{\mathbb{N}}\subset{\mathbb{Z}}}f(m)=+%
\infty\right\}.$$
The multiplication on ${\mathbb{E}}$ is induced by composition of almost
homomorphisms. The multiplication turns ${\mathbb{E}}$ into a commutative ring
with unit. Moreover, this ring is a field. Even more
surprisingly, ${\mathbb{E}}$ is an ordered field with respect to the ordering
defined by $P$.
Theorem 1.2 (See Arthan [7]).
${\mathbb{E}}$ is a complete ordered field and is therefore isomorphic to the
field of real numbers ${\mathbb{R}}$.
The isomorphism ${\mathbb{R}}\to{\mathbb{E}}$ assigns to every real number $\alpha\in{\mathbb{R}}$, the class $[f_{\alpha}]$ of the function
$$\displaystyle f_{\alpha}:{\mathbb{Z}}$$
$$\displaystyle\longrightarrow$$
$$\displaystyle{\mathbb{Z}}$$
$$\displaystyle n$$
$$\displaystyle\mapsto$$
$$\displaystyle\lfloor\alpha n\rfloor,$$
where $\lfloor\,\cdot\,\rfloor$ is the integer part function.
In the remainder of the paper, we combine the above one-step
construction of the reals with the ultrapower or limit ultrapower
construction, to obtain hyperreal number systems directly out of the
integers. We show that any hyperreal field, whose universe is a set,
can be so obtained by such a one-step construction. Following this,
working in NBG (von Neumann-Bernays-Gödel set theory with the
Axiom of Global Choice), we further observe that by using a suitable
definable ultrapower, even the maximal (i.e., the
On-saturated)222Recall that a model $M$ is On-saturated if $M$ is
$\kappa$-saturated for any cardinal $\kappa$ in On. Here
On (or ON) is the class of all ordinals
(cf. Kunen [40, p. 17]). A hyperreal number system $\langle{\mathbb{R}},{}^{*}\!{\mathbb{R}},S\in\mathfrak{F}\rangle$ is On-saturated if it
satisfies the following condition: If $X$ is a set of equations and
inequalities involving real functions, hyperreal constants and
variables, then $X$ has a hyperreal solution whenever every finite
subset of $X$ has a hyperreal solution (see Ehrlich [17, section 9,
p. 34]).
hyperreal number system recently described by Ehrlich [17] can
be obtained in a one-step fashion directly from the integers.333Another version of such an On-saturated number system
was introduced by Kanovei and Reeken (2004, [25, Theorem
4.1.10(i)]) in the framework of axiomatic nonstandard analysis.
As Ehrlich [17, Theorem 20] showed, the ordered field
underlying an On-saturated hyperreal field is isomorphic to
J. H. Conway’s ordered field No, an ordered field Ehrlich describes as
the absolute arithmetic continuum (modulo NBG).
2. Passing it through an ultraproduct
Let now $\mathcal{Z}=\mathbb{Z}^{\mathbb{N}}$ be the ring of integer
sequences with operations of componentwise addition and
multiplication. We define a rescaling to be a sequence $\rho=\langle\rho_{n}:n\in\mathbb{N}\rangle$ of almost
homomorphisms $\rho_{n}:\mathbb{Z}\longrightarrow\mathbb{Z}$.
Rescalings are thought of as acting on $\mathcal{Z}$, hence the name.
A rescaling $\rho$ is called bounded if each of its
components, $\rho_{n}$, is bounded.
Rescalings factorized modulo bounded rescalings form a commutative
ring $\mathcal{E}$ with respect to addition and composition.
Quotients of $\mathcal{E}$ by its maximal ideals are hyperreal fields.
Thus, hyperreal fields are factor fields of the ring of rescalings of
integer sequences. This description is a tautological translation of
the classical construction, due to E. Hewitt [23], but it is
interesting for the sheer economy of the language used. We will give
further details in the sections below.
3. Cantor, Dedekind, and Schanuel
The strategy of Cantor’s construction of the real numbers444The construction of the real numbers as equivalence classes
of Cauchy sequences of rationals, usually attributed to Cantor, is
actually due to H. Méray (1869, [42]) who published three years
earlier than E. Heine and Cantor.
can be represented schematically by the diagram
$${\mathbb{R}}:=\left({\mathbb{N}}\to({\mathbb{Z}}\times{\mathbb{Z}})_{\alpha}%
\right)_{\beta}$$
(3.1)
where the subscript $\alpha$ evokes the passage from a pair of
integers to a rational number; the arrow $\to$ alludes to forming
sequences; and subscript $\beta$ reminds us to select Cauchy sequences
modulo equivalence. Meanwhile, Dedekind proceeds according to the
scheme
$${\mathbb{R}}:=\left(\mathcal{P}({\mathbb{Z}}\times{\mathbb{Z}})_{\alpha}\right%
)_{\gamma}$$
(3.2)
where $\alpha$ is as above, $\mathcal{P}$ alludes to the set-theoretic
power operation, and $\gamma$ selects his cuts. For a history of the
problem, see P. Ehrlich [16].
An alternative approach was proposed by Schanuel, and developed by
N. A’Campo [1], R. Arthan [6], [7],555Arthan’s Irrational construction of ${\mathbb{R}}$ from ${\mathbb{Z}}$
[6] describes a different way of skipping the rationals,
based on the observation that the Dedekind construction can take as
its starting point any Archimedean densely ordered commutative group.
The construction delivers a completion of the group, and one can
define multiplication by analyzing its order-preserving endomorphisms.
Arthan uses the additive group of the ring ${\mathbb{Z}}[\sqrt{2}]$, which can
be viewed as ${\mathbb{Z}}\times{\mathbb{Z}}$ with an ordering defined using a certain
recurrence relation.
T. Grundhöfer [22], R. Street [48], O. Deiser
[15, pp. 112-127], and others, who follow the formally simpler
blueprint
$${\mathbb{R}}:=({\mathbb{Z}}\to{\mathbb{Z}})_{\sigma}$$
(3.3)
where $\sigma$ selects certain almost homomorphisms from ${\mathbb{Z}}$ to
itself, such as the map
$$a\mapsto\lfloor ra\rfloor$$
(3.4)
for real $r$, modulo equivalence (think of $r$ as the ‘‘large-scale
slope’’ of the map).666One could also represent a real by a string based on its
decimal expansion, but the addition in such a presentation is highly
nontrivial due to carry-over, which can be arbitrarily long. In
contrast, the addition of almost homomorphisms is term-by-term.
Multiplication on the reals is induced by composition in ${\mathbb{Z}}\to{\mathbb{Z}}$,
see formula (1.1).
Such a construction has been referred to as the Eudoxus reals.777See footnote 1 for a discussion of the term.
The construction of ${\mathbb{R}}$ from ${\mathbb{Z}}$ by means of almost homomorphisms
has been described as “skipping the rationals ${\mathbb{Q}}$”.
We will refer to the arrow in (3.3) as the space
dimension, so to distinguish it from the time dimension
occurring in the following construction of an extension of ${\mathbb{N}}$:
$$({\mathbb{N}}\to{\mathbb{N}})_{\tau_{\!co\!f}}$$
(3.5)
where $\tau_{\!co\!f}$ identifies sequences
$f,g:{\mathbb{N}}\to{\mathbb{N}}$ which differ on a finite set of indices:
$$\{n\in{\mathbb{N}}:f(n)=g(n)\}\quad\hbox{is cofinite.}$$
(3.6)
Here the constant sequences induce an inclusion
$${\mathbb{N}}\to({\mathbb{N}}\to{\mathbb{N}})_{\tau_{\!co\!f}}.$$
Such a construction is closely related to (a version of)
the $\Omega$-calculus of Schmieden and Laugwitz [45]. The
resulting semiring has zero divisors. To obtain a model which
satisfies the first-order Peano axioms, we need to quotient it
further. Note that up to this point the construction has not used any
nonconstructive foundational material such as the axiom of choice or
the weaker axiom of the existence of nonprincipal ultrafilters.
4. Constructing an infinitesimal-enriched continuum
The traditional ultrapower construction of the hyperreals proceeds
according to the blueprint
$$({\mathbb{N}}\to{\mathbb{R}})_{\mathcal{U}}$$
where ${\mathcal{U}}$ is a fixed ultrafilter on ${\mathbb{N}}$. Replacing ${\mathbb{R}}$
by any of the possible constructions of ${\mathbb{R}}$ from ${\mathbb{Z}}$, one in
principle obtains what can be viewed as a direct construction of the
hyperreals out of the integers ${\mathbb{Z}}$. Formally, the most economical
construction of this sort passes via the Eudoxus reals.
An infinitesimal-enriched continuum can be visualized by means of an
infinite-magnification microscope as in Figure 1.
To construct such an infinitesimal-enriched field, we have to deal
with the problem that the semiring
$({\mathbb{N}}\to{\mathbb{N}})_{\tau_{\!co\!f}}$ constructed in the previous
section contains zero divisors.
To eliminate the zero divisors, we need to quotient the ring further.
This is done by extending the equivalence relation by means of a
maximal ideal defined in terms of an ultrafilter. Thus, we extend the
relation defined by (3.6) to the relation declaring $f$ and $g$
equivalent if
$$\{n\in{\mathbb{N}}:f(n)=g(n)\}\in\mathcal{U},$$
(4.1)
where $\mathcal{U}$ is a fixed ultrafilter on ${\mathbb{N}}$, and add
negatives. The resulting modification of (3.5), called an ultrapower, will be denoted
$$\mbox{I\!I\!I\!\!N}:=({\mathbb{N}}\to{\mathbb{N}})_{\tau}$$
(4.2)
and is related to Skolem’s construction in 1934 of a nonstandard model
of arithmetic [47]. We refer to the arrow in (4.2) as time to allude to the fact that a sequence that increases without
bound for large time will generate an infinite “natural” number
in IIIN. A “continuous” version of the ultrapower construction was
exploited by Hewitt in constructing his hyperreal fields in 1948
(see [23]).
The traditional ultrapower approach to constructing the hyperreals is
to start with the field of real numbers ${\mathbb{R}}$ and build the ultrapower
$$({\mathbb{N}}\to{\mathbb{R}})_{\tau}$$
(4.3)
where the subscript $\tau$ is equivalent to that of (4.2), see
e.g. Goldblatt [21]. For instance, relative to Cantor’s
procedure (3.1), this construction can be represented by the
scheme
$$({\mathbb{N}}\to({\mathbb{N}}\to({\mathbb{Z}}\times{\mathbb{Z}})_{\alpha})_{%
\beta})_{\tau}\;;$$
however, this construction employs needless intermediate procedures as
described above. Our approach is to follow instead the “skip the
rationals” blueprint
$$\text{\rm I\!I\!R}:=({\mathbb{N}}\to{\mathbb{Z}}^{\mathbb{N}})_{\sigma\tau}$$
(4.4)
where the image of each $a\in{\mathbb{N}}$ is the sequence $u^{a}\in{\mathbb{Z}}^{\mathbb{N}}$ with
general term $u_{n}^{a}$, so that $u^{a}=\langle u^{a}_{n}:n\in{\mathbb{N}}\rangle$.
Thus a general element of IIR is generated (represented) by the
sequence
$$\left\langle a\mapsto\left(n\mapsto u^{a}_{n}\right):a\in{\mathbb{N}}\right\rangle.$$
(4.5)
Here one requires that for each fixed element $n_{0}\in{\mathbb{N}}$ of the
exponent copy of ${\mathbb{N}}$, the map
$$a\mapsto u_{n_{0}}^{a}$$
is an almost homomorphism (space direction), while $\tau$
in (4.4) alludes to the ultrapower quotient in the
time-direction $n$. For instance, we can use almost homomorphisms of
type (3.4) with $r=\frac{1}{n}$. Then the sequence
$$\left\langle a\mapsto\left(n\mapsto\left\lfloor\tfrac{a}{n}\right\rfloor\right%
):a\in{\mathbb{N}}\right\rangle$$
(4.6)
generates an infinitesimal in IIR since the almost homomorphisms
are “getting flatter and flatter” for large time $n$.
Theorem 4.1.
Relative to the construction (4.4), we have a natural
inclusion ${\mathbb{R}}\subset\hbox{\rm\text{\rm I\!I\!R}}$. Furthermore, IIR
is isomorphic to the model ${}^{*}{\mathbb{R}}$ of the hyperreals obtained by
quotienting ${\mathbb{R}}^{\mathbb{N}}$ by the chosen ultrafilter, as in (4.3).
Proof.
Given a real number $r\in{\mathbb{R}}$, we choose the constant sequence given
by $u_{n}^{a}=\lfloor ra\rfloor$ (the sequence is constant in time $n$).
Sending $r$ to the element of IIR defined by the sequence
$$\left\langle a\mapsto\left(n\mapsto\left\lfloor ra\right\rfloor\right):a\in{%
\mathbb{N}}\right\rangle$$
yields the required inclusion ${\mathbb{R}}\hookrightarrow\text{\rm I\!I\!R}$. The
isomorphism $\text{\rm I\!I\!R}\to\,^{*}{\mathbb{R}}$ is obtained by letting
$$U_{n}=\lim\frac{u_{n}^{a}}{a}$$
for each $n\in{\mathbb{N}}$, and sending the element of IIR represented
by (4.5) to the hyperreal represented by the sequence $\langle U_{n}:n\in{\mathbb{N}}\rangle$.
∎
Denoting by $\Delta x$ the infinitesimal generated by the integer
object (4.6), we can then define the derivative of $y=f(x)$
at $x$ following Robinson as the real number $f^{\prime}(x)$ infinitely close
(or, in Fermat’s terminology, adequal)999See A. Weil [50, p. 1146].
to the infinitesimal ratio $\frac{\Delta y}{\Delta x}\in\text{\rm I\!I\!R}$.
Applications of infinitesimal-enriched continua range from aid in
teaching calculus ([18], [27], [28], [37],
[43]) to the Bolzmann equation (see L. Arkeryd [4, 5]); modeling of timed systems in computer science (see H. Rust
[44]); mathematical economics (see R. Anderson [3]);
mathematical physics (see Albeverio et al. [2]); etc.
A comprehensive re-appraisal of the historical antecedents of modern
infinitesimals has been undertaken in recent work by Błaszczyk et
al. [11], Borovik et al. [12], Bråting [13],
Kanovei [24], Katz & Katz [31, 29, 32, 30],
Katz & Leichtnam [33], Katz and Sherry [35, 36], and
others. A construction of infinitesimals by “splitting” Cantor’s
construction of the reals is presented in Giordano et al. [20].
5. Formalization
In this and the next sections we formalize and generalize the
arguments in the previous sections. We show that by a one-step
construction from ${\mathbb{Z}}$-valued functions we can obtain any given
(set) hyperreal field. We can even obtain a universal hyperreal field
which contains an isomorphic copy of every hyperreal field,
by a one-step construction from ${\mathbb{Z}}$-valued functions.
We assume that the reader is familiar with some basic concepts of
model theory. Consult [14] or [38] for concepts and
notations undefined here.
Let $\mathfrak{A}$ and $\mathfrak{B}$ be two models in a language
$\mathcal{L}$ with base sets $A$ and $B$, respectively. The model
$\mathfrak{B}$ is called an $\mathcal{L}$-elementary extension of the
model $\mathfrak{A}$, or $\mathfrak{A}$ is an $\mathcal{L}$-elementary
submodel of $\mathfrak{B}$, if there is an embedding $e:A\rightarrow B$, called an $\mathcal{L}$-elementary embedding, such that for any
first-order $\mathcal{L}$-sentence $\varphi(a_{1},a_{2},\ldots,a_{n})$ with
parameters $a_{1},a_{2},\ldots,a_{n}\in A$, $\varphi(a_{1},a_{2},\ldots,a_{n})$ is
true in $\mathfrak{A}$ if and only if
$\varphi(e(a_{1}),e(a_{2}),\ldots,e(a_{n}))$ is true in $\mathfrak{B}$.
Let $\mathfrak{A}$ and $\mathfrak{B}$ be two models in language $\mathcal{L}$
with base sets $A$ and $B$, respectively. Let
$$\mathcal{L}^{\prime}=\mathcal{L}\cup\{P_{R}:\exists m\in{\mathbb{N}},R%
\subseteq A^{m}\},$$
i.e., $\mathcal{L}^{\prime}$ is formed by adding to $\mathcal{L}$ an
$m$-dimensional relational symbol $P_{R}$ for each $m$-dimensional
relation $R$ on $A$ for any positive integer $m$. Let $\mathfrak{A}^{\prime}$ be
the natural $\mathcal{L}^{\prime}$-model with base set $A$, i.e., the
interpretation of $P_{R}$ in $\mathfrak{A}^{\prime}$ for each $R\subseteq A^{m}$ is
$R$. The model $\mathfrak{B}$ is called a complete elementary
extension of $\mathfrak{A}$ if $\mathfrak{B}$ can be expanded to an
$\mathcal{L}^{\prime}$-model $\mathfrak{B}^{\prime}$ with base set $B$ such that
$\mathfrak{B}^{\prime}$ is an $\mathcal{L}^{\prime}$-elementary extension of $\mathfrak{A}^{\prime}$.
It is a well-known fact that if $\mathfrak{B}$ is an ultrapower of
$\mathfrak{A}$ or a limit ultrapower of $\mathfrak{A}$,
then $\mathfrak{B}$ is a complete elementary extension of $\mathfrak{A}$.
In this section we always view the set ${\mathbb{R}}$ as the set of all Eudoxus
reals.
An ordered field is called a hyperreal field if it is a proper
complete elementary extension of ${\mathbb{R}}$. Let
$$\mathcal{L}^{\prime}=\{+,\cdot,\leqslant,0,1,P_{R}\}_{R\in\mathcal{R}}$$
where $\mathcal{R}$ is the collection of all finite-dimensional
relations on ${\mathbb{R}}$. We do not distinguish between ${\mathbb{R}}$ and the
$\mathcal{R}^{\prime}$-model
$\mathfrak{R}=({\mathbb{R}};+,\cdot,\leqslant,0,1,R)_{R\in\mathcal{R}}$. By
saying that ${}^{*}{\mathbb{R}}$ is a hyperreal field we will sometimes mean that
${}^{*}{\mathbb{R}}$ is the base set of the hyperreal field, but at other times we
mean that ${}^{*}{\mathbb{R}}$ is the hyperreal field viewed as an
$\mathcal{L}^{\prime}$-model. We will spell out the distinction when it
becomes necessary.
Recall that $S$ is the set of all bounded functions from ${\mathbb{Z}}$ to ${\mathbb{Z}}$.
For a pair of almost homomorphisms $f,g:{\mathbb{Z}}\rightarrow{\mathbb{Z}}$, we will
write $f\sim_{\sigma}g$ if and only if $f-g\in S$. Let $I$ be an
infinite set. If $F(x,y)$ is a two-variable function from ${\mathbb{Z}}\times I$ to ${\mathbb{Z}}$ and $i$ is a fixed element in $I$, we write $F(x,i)$ for
the one-variable function $F_{i}(x)=F(x,i)$ from ${\mathbb{Z}}$ to ${\mathbb{Z}}$.
Definition 5.1.
Let $I$ be any infinite set. We set
$$\displaystyle{\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})=$$
$$\displaystyle=\{F\in{\mathbb{Z}}^{{\mathbb{Z}}\times I}:\forall i\in I,\,F(x,i%
)\,\mbox{ is an almost homomorphism.}\}$$
Let $\mathcal{U}$ be a fixed non-principal ultrafilter on $I$. For a
pair of functions $f,g:I\rightarrow J$ for some set $J$, we set
$$f\sim_{\tau}g\,\mbox{ if and only if }\,\{i\in I:f(i)=g(i)\}\in\mathcal{U}.$$
Let $[f]_{\tau}=\{g\in I^{J}:g\sim_{\tau}f\}$.
Definition 5.2.
For any $F,G\in{\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})$ we will write
$$F\sim_{\sigma\tau}G\,\mbox{ if and only if }\,\{i\in I:F(x,i)\sim_{\sigma}G(x,%
i)\}\in\mathcal{U}.$$
It is easy to check that $\sim_{\sigma\tau}$ is an equivalence
relation on ${\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})$. For each $F\in{\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})$ let
$$[F]_{\sigma\tau}=\{G\in{\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}}):G\sim_%
{\sigma\tau}F\}.$$
For each $F(x,y)\in{\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})$ we can consider
$[F(\cdot,y)]_{\sigma}$ as a function of $y$ from $I$ to ${\mathbb{R}}$. Thus
the map
$$\Phi:{\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})/\!\sim_{\sigma\tau}%
\rightarrow{\mathbb{R}}^{I}/\mathcal{U}$$
such that $\Phi([F]_{\sigma\tau})=[[F]_{\sigma}]_{\tau}$ is an
isomorphism from ${\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})/\!\sim_{\sigma\tau}$ to
${\mathbb{R}}^{I}/\mathcal{U}$. Hence ${\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})/\!\sim_{\sigma\tau}$ can
be viewed as an ultrapower of ${\mathbb{R}}$. Therefore, the quotient
$$\text{\rm I\!I\!R}_{I}={\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})/\!\sim%
_{\sigma\tau}$$
is a hyperreal field constructed in one step from the set of
${\mathbb{Z}}$-valued functions ${\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})$.
If the set $I$ is ${\mathbb{N}}$, then ${\mathcal{A}}({\mathbb{Z}}\times{\mathbb{N}},{\mathbb{Z}})/\!\sim_{\sigma\tau}$
is exactly the hyperreal field IIR mentioned in the previous
sections. Since $I$ can be any infinite set, we can construct a
hyperreal field of arbitrarily large cardinality in one step from a
set of ${\mathbb{Z}}$-valued functions ${\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})$.
6. Limit ultrapowers and definable ultrapowers
If we consider a limit ultrapower instead of an ultrapower, we can
obtain any (set) hyperreal field by a one-step construction from a set
of ${\mathbb{Z}}$-valued functions ${\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})|\mathcal{G}$. The reader
could consult (Keisler [38]) for the notations, definitions, and
basic facts about limit ultrapowers. The main fact we need here is
the following theorem (see Keisler [38, Theorem 3.7]).
Theorem 6.1.
If $\mathfrak{A}$ and $\mathfrak{B}$ are two models of the same
language, then $\mathfrak{B}$ is a complete elementary extension of
$\mathfrak{A}$ if and only if $\mathfrak{B}$ is a limit ultrapower of
$\mathfrak{A}$.
Given any (set) hyperreal field ${}^{*}{\mathbb{R}}$, let $\mathcal{U}$ be the
ultrafilter on an infinite set $I$ and let $\mathcal{G}$ be the filter
on $I\times I$ such that ${}^{*}{\mathbb{R}}$ is isomorphic to the limit ultrapower
$({\mathbb{R}}^{I}/\mathcal{U})|\mathcal{G}$. We can describe the limit ultrapower
$({\mathbb{R}}^{I}/\mathcal{U})|\mathcal{G}$ in one step from the set of
${\mathbb{Z}}$-valued functions ${\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})|\mathcal{G}$. For each
$F\in{\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})$, let
$$eq(F)=\{(i,j)\in I\times I:[F(x,i)]_{\sigma}=[F(x,j)]_{\sigma}\}.$$
Let
$${\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})|\mathcal{G}=\{F\in{\mathcal{A%
}}({\mathbb{Z}}\times I,{\mathbb{Z}}):eq(F)\in\mathcal{G}\}.$$
Notice that ${\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})|\mathcal{G}$ is a subset of ${\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})$. Hence
$({\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})|\mathcal{G})/\!\sim_{\sigma\tau}$ can be viewed as
a subset of ${\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})/\!\sim_{\sigma\tau}$. Again for each
$$F\in{\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})/\!\sim_{\sigma\tau}$$
let
$$\Phi([F]_{\sigma\tau})=[[F]_{\sigma}]_{\tau}.$$
Then $\Phi$ is an isomorphism from
$$({\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})|\mathcal{G})/\!\sim_{\sigma\tau}$$
to $({\mathbb{R}}^{I}/\mathcal{U})|\mathcal{G}$. Therefore, $({\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})|\mathcal{G})/\!\sim_{\sigma\tau}$ as an elementary subfield of
${\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})/\!\sim_{\sigma\tau}$ is isomorphic to the hyperreal
field ${}^{*}{\mathbb{R}}$.
Theorem 6.2.
An isomorphic copy of any (set) hyperreal field ${}^{*}{\mathbb{R}}$ can be obtained
by a one-step construction from a set of ${\mathbb{Z}}$-valued functions
${\mathcal{A}}({\mathbb{Z}}\times I,{\mathbb{Z}})|\mathcal{G}$ for some filter $\mathcal{G}$ on
$I\times I$.
7. Universal and On-saturated Hyperreal Number Systems
We call a hyperreal field IIR universal if any hyperreal
field, which is a set or a proper class of NBG, can be
elementarily embedded in IIR. Obviously a universal hyperreal
field is necessarily a proper class. We now want to construct a
definable hyperreal field with the property that any definable
hyperreal field that can be obtained in NBG by a one-step
construction from a collection of ${\mathbb{Z}}$-valued functions can be
elementarily embedded in it. In a subsequent remark we point out that
in NBG we can actually construct a definable hyperreal field so
that every hyperreal field (definable or non-definable) can be
elementarily embedded in it. Moreover, the universal hyperreal field
so constructed is isomorphic to the On-saturated hyperreal
field described in [17].
Notice that NBG implies that there is a well order $\leqslant_{V}$
on $V$ where $V$ is the class of all sets. A class $X\subseteq V$ is
called $\Delta_{0}$-definable if there is a first-order formula
$\varphi(x)$ with set parameters in the language $\{\in,\leqslant\}$
such that for any set $a\in V$, $a\in X$ if and only if $\varphi(a)$
is true in $V$. Trivially, every set is $\Delta_{0}$-definable. We
work within a model of NBG with set universe $V$ plus all
$\Delta_{0}$-definable proper subclasses of $V$. By saying that a class
$A$ is definable we mean that $A$ is $\Delta_{0}$-definable.
Let $\Sigma$ be the class of all finite sets of ordinals, i.e.,
$\Sigma=On^{<\omega}$. Notice that $\Sigma$ is a definable proper class.
Let $\mathcal{P}$ be the
collection of all definable subclasses of $\Sigma$. Notice
that we can code $\mathcal{P}$ by a definable class.101010 This is true because each definable subclass of $\Sigma$
can be effectively coded by the Gödel number of a first-order
formula in the language of $\{\in,\leqslant_{V}\}$ and a set in $V$. By
the well order of $V$ we can determine a unique code for each
definable class in $\mathcal{P}$.
Using the global choice, we can form a non-principal definable
ultrafilter $\mathcal{F}\subseteq\mathcal{P}$ such that for each
$\alpha\in On$, the definable class
$$\hat{\alpha}=\{s\in\Sigma:\alpha\in s\}$$
is in $\mathcal{F}$. Again $\mathcal{F}$ can be coded by a definable
class. Let ${\mathcal{A}}_{0}({\mathbb{Z}}\times\Sigma,{\mathbb{Z}})$ be the collection of all
definable class functions $F$ from ${\mathbb{Z}}\times\Sigma$ to ${\mathbb{Z}}$ such that
for each $s\in\Sigma$, $F(x,s)$ is an almost homomorphism from ${\mathbb{Z}}$ to
${\mathbb{Z}}$. For any two functions $F$ and $G$ in ${\mathcal{A}}_{0}({\mathbb{Z}}\times\Sigma,{\mathbb{Z}})$,
we write $F\sim_{\sigma\tau}G$ if and only if the definable class
$$\{s\in\Sigma:F(x,s)-G(x,s)\in S\}$$
is in $\mathcal{F}$. Let $[F]_{\sigma\tau}$ be the collection of all
definable classes $G$ in ${\mathcal{A}}_{0}({\mathbb{Z}}\times\Sigma,{\mathbb{Z}})$ such that
$F\sim_{\sigma\tau}G$. Then $\sim_{\sigma\tau}$ is an equivalence
relation on ${\mathcal{A}}_{0}({\mathbb{Z}}\times\Sigma,{\mathbb{Z}})$. Let
$$\text{\rm I\!I\!R}_{0}={\mathcal{A}}_{0}({\mathbb{Z}}\times\Sigma,{\mathbb{Z}}%
)/\!\sim_{\sigma\tau}.$$
By the arguments employed before, we can show that $\text{\rm I\!I\!R}_{0}$ is
isomorphic to the definable ultrapower of ${\mathbb{R}}$ modulo
$\mathcal{F}$. Hence $\text{\rm I\!I\!R}_{0}$ is a complete elementary extension of
${\mathbb{R}}$. Therefore, $\text{\rm I\!I\!R}_{0}$ is a hyperreal field. By slightly modifying
the proof of [14, Theorem 4.3.12, p. 255] we can prove the
following theorem.
Theorem 7.1.
$\text{\rm I\!I\!R}_{0}$ is a class hyperreal field, and any definable hyperreal field
${}^{*}{\mathbb{R}}$ admits an elementary imbedding into $\text{\rm I\!I\!R}_{0}$.
Proof.
We only need to prove the second part of the theorem. For notational
convenience we view $\text{\rm I\!I\!R}_{0}$ as ${\mathbb{R}}^{\Sigma}/\mathcal{F}$ instead of
${\mathcal{A}}_{0}({\mathbb{Z}}\times\Sigma,{\mathbb{Z}})/\sim_{\sigma\tau}$ in this proof.
Given a definable hyperreal field ${}^{*}{\mathbb{R}}$, recall that $\mathcal{L}$ is
the language of ordered fields, and
$$\mathcal{L}^{\prime}=\mathcal{L}\cup\{P_{R}:R\,\mbox{ is a finite-dimensional
relation on }\,{\mathbb{R}}\}.$$
Let $\Lambda_{{}^{*}{\mathbb{R}}}$ be all quantifier-free $\mathcal{L}^{\prime}$-sentences
$\varphi(r_{1},r_{2},\ldots,r_{m})$ with parameters $r_{i}\in\,^{*}{\mathbb{R}}$ such that
$\varphi(r_{1},r_{2},\ldots,r_{m})$ is true in ${}^{*}{\mathbb{R}}$. Since ${}^{*}{\mathbb{R}}$ is a
definable class, $\Lambda_{{}^{*}{\mathbb{R}}}$ is a definable class (under a proper
coding). Let $\kappa$ be the size of $\Lambda_{{}^{*}{\mathbb{R}}}$, i.e., $\kappa$
is the cardinality of ${}^{*}{\mathbb{R}}$ if ${}^{*}{\mathbb{R}}$ is a set and $\kappa=On$ if
${}^{*}{\mathbb{R}}$ is a definable proper class. Let $j$ be a definable bijection
from $\kappa$ to $\Lambda_{{}^{*}{\mathbb{R}}}$.
For each $r\in\,^{*}{\mathbb{R}}$ we need to find a definable function
$F_{r}:\Sigma\rightarrow{\mathbb{R}}$ such that the map $r\mapsto[F_{r}]_{\tau}$ is an
$\mathcal{L}^{\prime}$-elementary embedding. We define these $F_{r}$ simultaneously.
Let $s\in\Sigma$. If $s\cap\kappa=\emptyset$ let $F_{r}(s)=0$. Suppose
$s\cap\kappa\not=\emptyset$ and let $s^{\prime}=s\cap\kappa$. Notice that if
$\kappa=On$, then $s=s^{\prime}$. Let
$\varphi_{s}(r_{1},r_{2},\ldots,r_{m})=\bigwedge_{\alpha\in s^{\prime}}j(\alpha)$.
Since
$$\exists x_{1},x_{2},\ldots,x_{m}\varphi_{s}(x_{1},x_{2},\ldots,x_{m})$$
is
true in ${}^{*}{\mathbb{R}}$, it is also true in ${\mathbb{R}}$. Let
$(a_{1},a_{2},\ldots,a_{m})\in{\mathbb{R}}^{m}$ be the $\leqslant_{V}$-least $m$-tuple
such that $\varphi_{s}(a_{1},a_{2},\ldots,a_{m})$ is true in ${\mathbb{R}}$. If
$r\not\in\{r_{1},r_{2},\ldots,r_{m}\}$, let $F_{r}(s)=0$. If $r=r_{i}$ for
$i=1,2,\ldots,m$, then let $F_{r}(s)=a_{i}$. Since $\varphi_{s}$ is
quantifier-free, the functions $F_{r}$ are definable classes in NBG.
We now verify that $\Phi:\,^{*}{\mathbb{R}}\rightarrow\text{\rm I\!I\!R}_{0}$ such that
$\Phi(r)=[F_{r}]_{\tau}$ is an $\mathcal{L}^{\prime}$-elementary embedding.
Let $\phi(r_{1},r_{2},\ldots,r_{m})$ be an arbitrary $\mathcal{L}^{\prime}$-sentence
with parameters
$r_{1},r_{2},\ldots,r_{m}\in\,^{*}{\mathbb{R}}$.
Suppose that $\phi(r_{1},r_{2},\ldots,r_{m})$ is true in ${}^{*}{\mathbb{R}}$. Since
$\phi(x_{1},x_{2},\ldots,x_{m})$ defines an $m$-ary relation $R_{\phi}$ on
${\mathbb{R}}$, we have that the $\mathcal{L}^{\prime}$-sentence
$$\eta=:\forall x_{1},x_{2},\ldots,x_{m}(\phi(x_{1},x_{2},\ldots,x_{m})%
\leftrightarrow R_{\phi}(x_{1},x_{2},\ldots,x_{m}))$$
is true in ${\mathbb{R}}$. Hence $\eta$ is true
in ${}^{*}{\mathbb{R}}$ and in $\text{\rm I\!I\!R}_{0}$. One of the consequences of this is that
$R_{\phi}(r_{1},r_{2},\ldots,r_{m})$ is true in ${}^{*}{\mathbb{R}}$, hence it is in
$\Lambda_{{}^{*}{\mathbb{R}}}$. Let $\alpha\in\kappa$ be such that
$j(\alpha)=R_{\phi}(r_{1},r_{2},\ldots,r_{m})$. If $\alpha\in s$, then
$$\varphi_{s}=R_{\phi}(r_{1},r_{2},\ldots,r_{m})\wedge\bigwedge_{\beta\in s^{%
\prime},\beta\not=\alpha}j(\beta).$$
Hence
$R_{\phi}(F_{r_{1}}(s),F_{r_{2}}(s),\ldots,F_{r_{m}}(s))$ is true in ${\mathbb{R}}$ by
the definition of the $F_{r}$’s. Since $\eta$ is true in ${\mathbb{R}}$, we have
that $\phi(F_{r_{1}}(s),F_{r_{2}}(s),\ldots,F_{r_{m}}(s))$ is true in
${\mathbb{R}}$. Thus
$$\{s\in\Sigma:\phi(F_{r_{1}}(s),F_{r_{2}}(s),\ldots,F_{r_{m}}(s))\,\mbox{
is true in }\,{\mathbb{R}}\}\supseteq\hat{\alpha}.$$
Since $\hat{\alpha}$ is a
member of $\mathcal{F}$, we have that
$\phi([F_{r_{1}}]_{\tau},[F_{r_{2}}]_{\tau},\ldots,[F_{r_{m}}]_{\tau})$ is
true in ${\mathbb{R}}^{\Sigma}/\mathcal{F}$.
Suppose that $\phi(r_{1},r_{2},\ldots,r_{m})$ is false in ${}^{*}{\mathbb{R}}$. Then
$\neg\phi(r_{1},r_{2},\ldots,r_{m})$ is true in ${}^{*}{\mathbb{R}}$. Hence by the same
argument we have that
$$\neg\phi([F_{r_{1}}]_{\tau},[F_{r_{2}}]_{\tau},\ldots,[F_{r_{m}}]_{\tau})\,%
\mbox{
is true in }\,{\mathbb{R}}^{\Sigma}/\mathcal{F}$$
which implies that
$$\phi([F_{r_{1}}]_{\tau},[F_{r_{2}}]_{\tau},\ldots,[F_{r_{m}}]_{\tau})\,\mbox{
is false in }\,{\mathbb{R}}^{\Sigma}/\mathcal{F}.$$
Hence $\Phi(r)=[F_{r}]_{\tau}$
is an $\mathcal{L}^{\prime}$-elementary embedding from ${}^{*}{\mathbb{R}}$ to $\text{\rm I\!I\!R}_{0}$.
∎
Remark 7.2.
We have shown that every definable hyperreal field can be elementarily
embedded into $\text{\rm I\!I\!R}_{0}$. If we want to show that $\text{\rm I\!I\!R}_{0}$ is
universal, we need to elementarily embed every (definable or
non-definable) hyperreal field ${}^{*}{\mathbb{R}}$ into the definable hyperreal
field $\text{\rm I\!I\!R}_{0}$. Notice that there are models of NBG with
non-definable classes. The proof of Theorem 7.1 may not
work when ${}^{*}{\mathbb{R}}$ is a non-definable class because the bijection $j$
may not be definable and $F_{r}$ may not be definable. If $F_{r}$ is not
definable, $[F_{r}]_{\tau}$ may not be an element in $\text{\rm I\!I\!R}_{0}$.
The idea of making every hyperreal field embeddable into $\text{\rm I\!I\!R}_{0}$ is
that we can make $\text{\rm I\!I\!R}_{0}$ On-saturated by selecting a definable
ultrafilter $\mathcal{F}$ more carefully. Notice that NBG
implies that every proper class has the same size On. Hence
${}^{*}{\mathbb{R}}$ can be expressed as the union of On-many sets. If we can
make sure that $\text{\rm I\!I\!R}_{0}$ is On-saturated, i.e.,
$\alpha$-saturated for any set cardinality $\alpha$, then ${}^{*}{\mathbb{R}}$ can
be elementarily embedded into $\text{\rm I\!I\!R}_{0}$ although such an embedding may
be non-definable.
The ultrafilter $\mathcal{F}$ used in the construction of $\text{\rm I\!I\!R}_{0}$ in
the proof of Theorem 7.1 is a definable version of a regular ultrafilter. In order to make sure that $\text{\rm I\!I\!R}_{0}$ is
On-saturated, we need to require that $\mathcal{F}$ be a
special kind of definable regular ultrafilter called a definable good ultrafilter. The definition of a (set) good ultrafilter can be
found in [14, p. 386]. The construction of an $\alpha^{+}$-good
ultrafilter can be found in either [14, Theorem 6.1.4] or Kunen
[39]. By the same idea of constructing $\mathcal{F}$ above we
can follow the steps in [39] or [14] to construct a
definable class good ultrafilter $\mathcal{F}$ on On. Now the
ultrapower $\text{\rm I\!I\!R}_{0}$ of ${\mathbb{R}}$ modulo the definable class good
ultrafilter $\mathcal{F}$ is On-saturated. The proof of this
fact is similar to that in [14]. However, since the definition
of a definable class good ultrafilter and the proof of the saturation
property of the ultrapower modulo a definable class good ultrafilter
are long and technical, and the ideas are similar to what we have
already presented above, we will not include them in this paper.
Another way of constructing a definable On-saturated hyperreal
field $\text{\rm I\!I\!R}_{0}$ is by taking the union of an On-long definable
elementary chain of set hyperreal fields $\left\{{}^{*}{\mathbb{R}}_{\alpha}:\alpha\in On\right\}$ with the property that ${}^{*}{\mathbb{R}}_{\alpha}$ is
$|\alpha|$-saturated. However, this construction cannot be easily
translated into a “one-step” construction. Moreover, if we allow
higher-order classes, we can express $\text{\rm I\!I\!R}_{0}$ as a one-step limit
ultrapower following the same idea as in the proof of [14, Theorem
6.4.10]. However, this is not possible in NBG since all
classes allowed in a model of NBG are subclasses of $V$.
On the other hand, as we indicated above, the process of constructing
$\text{\rm I\!I\!R}_{0}$ as a definable ultrapower can be carried out in NBG,
and done so in a “one-step” fashion.
Acknowledgments
We are grateful to R. Arthan, P. Ehrlich, and V. Kanovei for helpful
comments.
References
[1]
A’Campo, N.: A natural construction for the real numbers,
2003, see
http://arxiv.org/abs/math/0301015
[2]
Albeverio, S.; Høegh-Krohn, R.; Fenstad, J.; Lindstrøm, T.: Nonstandard methods in stochastic analysis and mathematical
physics. Pure and Applied Mathematics, 122. Academic
Press, Inc., Orlando, FL, 1986.
[3]
Anderson, R. M.: Infinitesimal Methods in Mathematical
Economics, Preprint, 2000. See
http://www.econ.berkeley.edu/$\sim$anderson/Book.pdf.
[4]
Arkeryd, L.: Intermolecular forces of infinite range
and the Boltzmann equation. Arch. Rational Mech. Anal.
77 (1981), no. 1, 11–21.
[5]
Arkeryd, L.: Nonstandard analysis. American
Mathematical Monthly 112 (2005), no. 10, 926-928.
[6]
Arthan, R.: An irrational construction of $\mathbb{R}$ from $\mathbb{Z}$. Theorem proving in higher order logics
(Edinburgh, 2001), 43–58, Lecture Notes in Computer Science
2152, Springer, Berlin, 2001.
[7]
Arthan, R.: The Eudoxus Real Numbers. See
http://arxiv.org/abs/math/0405454
[8]
Benci, V.; Di Nasso, M.: Numerosities of labelled sets: a
new way of counting. Adv. Math. 173 (2003), no. 1,
50–67.
[9]
Benci, V.; Di Nasso, M.: A purely algebraic
characterization of the hyperreal numbers.
Proc. Amer. Math. Soc. 133 (2005), no. 9, 2501–2505.
[10]
Biran, P.; Entov, M.; Polterovich, L.: Calabi
quasimorphisms for the symplectic ball. Commun. Contemp. Math. 6 (2004), no. 5, 793–802.
[11]
Błasczcyk, P.; Katz, M.; Sherry, D.: Ten
misconceptions from the history of analysis and their debunking. Foundations of Science, 2012. See
http://dx.doi.org/10.1007/s10699-012-9285-8
and http://arxiv.org/abs/1202.4153
[12]
Borovik, A.; Katz, M.: Who gave you the
Cauchy–Weierstrass tale? The dual history of rigorous calculus.
Foundations of Science 17 (2012), no. 3, 245-276. See
http://dx.doi.org/10.1007/s10699-011-9235-x
and http://arxiv.org/abs/1108.2885
[13]
Bråting, K.: A new look at E. G. Björling and the
Cauchy sum theorem. Arch. Hist. Exact Sci. 61
(2007), no. 5, 519–535.
[14]
Chang, C. C.; Keisler, H. J.: Model theory, third ed. (1990),
North-Holland Publishing Company, Amsterdam and New York.
[15]
Deiser, O.: Reelle Zahlen: Das klassische Kontinuum und
die natürlichen Folgen. Springer, 2007.
[16]
Ehrlich, P.: The rise of non-Archimedean mathematics
and the roots of a misconception. I. The emergence of non-Archimedean
systems of magnitudes. Arch. Hist. Exact Sci. 60
(2006), no. 1, 1–121.
[17]
Ehrlich, P.: The absolute arithmetic continuum and the
unification of all numbers great and small. Bulletin of Symbolic
Logic 18 (2012), no. 1, 1–45.
[18]
Ely, R.: Nonstandard student conceptions about
infinitesimals. Journal for Research in Mathematics Education
41 (2010), no. 2, 117-146.
[19]
Forti, M.; Di Nasso, M.; Benci, V.: Hausdorff
nonstandard extensions. Bol. Soc. Parana. Mat. (3) 20
(2002), no. 1-2, 9–20 (2003).
[20]
Giordano, P.; Katz, M.: Two ways of obtaining
infinitesimals by refining Cantor’s completion of the reals, 2011, see
http://arxiv.org/abs/1109.3553
[21]
Goldblatt, R.: Lectures on the hyperreals. An
introduction to nonstandard analysis. Graduate Texts in
Mathematics, 188. Springer-Verlag, New York, 1998.
[22]
Grundhöfer, T.: Describing the real numbers in terms
of integers. Arch. Math. (Basel) 85 (2005), no. 1,
79–81.
[23]
Hewitt, E.: Rings of real-valued continuous
functions. I. Trans. Amer. Math. Soc. 64 (1948),
45–99.
[24]
Kanovei, V.: Correctness of the Euler method of
decomposing the sine function into an infinite product. (Russian)
Uspekhi Mat. Nauk 43 (1988), no. 4 (262), 57–81, 255;
translation in Russian Math. Surveys 43 (1988), no. 4,
65–94.
[25]
Kanovei, V.G.; Reeken, M.: Nonstandard analysis,
axiomatically. Springer Monographs in Mathematics, Berlin: Springer,
2004.
[26]
Kanovei, V.; Katz, M.; Mormann, T.: Tools, Objects, and
Chimeras: Connes on the Role of Hyperreals in Mathematics.
Foundations of Science (to appear). See
http://dx.doi.org/s10699-012-9316-5
[27]
Katz, K.; Katz, M.: Zooming in on infinitesimal $1-.9..$
in a post-triumvirate era. Educational Studies in Mathematics
74 (2010), no. 3, 259-273.
See http://arxiv.org/abs/arXiv:1003.1501
[28]
Katz, K.; Katz, M.: When is $.999\ldots$ less than 1?
The Montana Mathematics Enthusiast 7 (2010), No. 1,
3–30.
[29]
Katz, K.; Katz, M.: Cauchy’s continuum.
Perspectives on Science 19 (2011), no. 4, 426-452.
See http://dx.doi.org/10.1162/POSC_a_00047
and http://arxiv.org/abs/1108.4201
[30]
Katz, K.; Katz, M.: Meaning in classical mathematics:
is it at odds with Intuitionism? Intellectica 56
(2011), no. 2, 223–302. See
http://arxiv.org/abs/1110.5456
[31]
Katz, K.; Katz, M.: A Burgessian critique of
nominalistic tendencies in contemporary mathematics and its
historiography. Foundations of Science 17 (2012),
no. 1, 51–89. See http://dx.doi.org/10.1007/s10699-011-9223-1
and http://arxiv.org/abs/1104.0375
[32]
Katz, K.; Katz, M.: Stevin numbers and reality,
Foundations of Science 17 (2012), no. 2, 109-123. See
http://arxiv.org/abs/1107.3688 and
http://dx.doi.org/10.1007/s10699-011-9228-9
[33]
Katz, M.; Leichtnam, E.: Commuting and non-commuting
infinitesimals. American Mathematical Monthly (to appear).
[34]
Katz, M.; Schaps, D.; Shnider, S.: Almost Equal: The
Method of Adequality from Diophantus to Fermat and Beyond.
Perspectives on Science 21 (2013), no. 3, to appear.
[35]
Katz, M.; Sherry, D.: Leibniz’s infinitesimals: Their
fictionality, their modern implementations, and their foes from
Berkeley to Russell and beyond. Erkenntnis (online first).
See http://dx.doi.org/10.1007/s10670-012-9370-y
and http://arxiv.org/abs/1205.0174
[36]
Katz, M.; Sherry, D.: Leibniz’s laws of continuity and
homogeneity. Notices of the American Mathematical Society
59 (2012), no. 11, to appear.
[37]
Katz, M.; Tall, D.: The tension between intuitive
infinitesimals and formal mathematical analysis. In: Bharath
Sriraman, Editor. Crossroads in the History of Mathematics and
Mathematics Education. The Montana Mathematics Enthusiast
Monographs in Mathematics Education 12, Information Age
Publishing, Inc., Charlotte, NC, 2012, pp. 71-89. See
http://arxiv.org/abs/1110.5747 and
http://www.infoagepub.com/products/Crossroads-in-the-History-of-Mathematics
[38]
Keisler, H. J.: Limit ultrapowers, Transactions of
the American Mathematical Society 107 (1963), 382–408.
[39]
Kunen, K.: Ultrafilters and independent sets.
Trans. Amer. Math. Soc. 172 (1972), 299–306.
[40]
Kunen, K.: Kunen, Kenneth Set theory. An introduction
to independence proofs. Studies in Logic and the Foundations of
Mathematics 102. North-Holland Publishing Co.,
Amsterdam-New York, 1980.
[41]
Lawvere, F. W.: Toward the description in a smooth topos
of the dynamically possible motions and deformations of a continuous
body. Third Colloquium on Categories (Amiens, 1980), Part I. Cahiers Topologie Géom. Différentielle 21 (1980),
no. 4, 377–392.
[42]
Méray, H. C. R.: Remarques sur la nature des
quantités définies par la condition de servir de limites à des
variables données, Revue des sociétiés savantes des
départments, Section sciences mathématiques, physiques et
naturelles, 4th ser., 10 (1869), 280–289.
[43]
Norton, A.; Baldwin M.: Does 0.999…really equal 1?
The Mathematics Educator 21 (2011/2012), no. 2, 58-67.
[44]
Rust, H.: Operational Semantics for Timed Systems.
Lecture Notes in Computer Science 3456 (2005), 23-29,
DOI: 10.1007/978-3-540-32008-1_4.
[45]
Schmieden, C.; Laugwitz, D.: Eine Erweiterung der
Infinitesimalrechnung. Math. Z. 69 (1958), 1–39.
[46]
Shenitzer, A.: A topics course in mathematics. The
Mathematical Intelligencer 9 (1987), no. 3, 44–52.
[47]
Skolem, T.: Über die Nicht-charakterisierbarkeit der
Zahlenreihe mittels endlich oder abzählbar unendlich vieler Aussagen
mit ausschliesslich Zahlenvariablen. Fundam. Math.
23, 150-161 (1934).
[48]
Street, R.: Update on the efficient reals, 2003, see
http://www.math.mq.edu.au/ street/reals.pdf
[49]
Weber, H.: Leopold Kronecker. Mathematische
Annalen 43 (1893), 1–25.
[50]
Weil, A.: Book Review: The mathematical career of Pierre
de Fermat. Bull. Amer. Math. Soc. 79 (1973), no. 6,
1138–1149. |
Almost partitioning 2-coloured complete 3-uniform hypergraphs into two monochromatic tight or loose cycles
Sebastián Bustamante${}^{,}$
All three authors acknowledge support by Millenium Nucleus Information and Coordination in Networks ICM/FIC RC130003.The first author was supported by CONICYT Doctoral Fellowship grant 21141116.
Department of Mathematical Engineering, University of Chile
Hiệp Hàn${}^{*,}$${}^{,}$
The second author was supported by Fondecyt Iniciación grant 11150913.
Instituto de Matemáticas, Pontificia Universidad Católica de Valparaíso
Maya Stein${}^{*,}$${}^{,}$
The third author received support by Fondecyt Regular grant 1140766.
Department of Mathematical Engineering, University of Chile
Abstract
We show that for every $\eta>0$ there exists an integer $n_{0}$ such that every $2$-colouring of the $3$-uniform complete hypergraph on $n\geq n_{0}$ vertices contains two disjoint monochromatic tight cycles of distinct colours that together cover
all but at most ${\eta n}$ vertices. The same result holds if we replace tight cycles with loose cycles.
1 Introduction
A certain type of Ramsey problems concerned with covering all vertices of the host graph instead of finding a small subgraph has gained popularity in recent years, both for graphs and hypergraphs. In particular, the problem of partitioning an edge-coloured complete (hyper-)graph into monochromatic cycles has received much attention. The recent surveys [5, 6] are a good starting point for the vast literature on the subject.
Central to the area is an old conjecture of Lehel for graphs (see [2]). It states that for every $n$, every two-colouring of the edges of the complete graph $K_{n}$ admits a
covering of the vertices of $K_{n}$ with at most two monochromatic vertex disjoint cycles of different colours. (For technical reasons, a single vertex or an edge count as a cycle.) This was confirmed for large $n$ in [13, 1], and for all $n$ by Bessy and Thomassé [3].
There is plenty of activity in determing the number of monochromatic cycles needed if $K_{n}$ is coloured with more than two colours. It is known that this number is independent of $n$, but the best known lower and upper bounds, $r+1$ and $100r\log r$, respectively, leave a considerable gap [15, 7].
For hypergraphs, much less is known. The problems transforms in the obvious way to hypergraphs, considering $r$-edge-colourings of the $k$-uniform $\mathcal{K}_{n}^{(k)}$ on $n$ vertices, the only question is which type of hypergraph cycles we would like to work with. Referring to [5] for other results, we concentrate here on loose and tight cycles (see the next section for their definitions).
For loose cycles the problem was studied in [8, 17] and the best bound, due to Sárközy [17], shows that at most $50rk\log(rk)$ disjoint loose monochromatic cycles are needed for partitioning $\mathcal{K}_{n}^{(k)}$, a number that this is independent of $n$.
In general even for the case $k=3$ the problem is wide open and it is for example unknown whether one can cover almost all vertices with two disjoint monochromatic loose cycles.
Concerning the more restrictive notion of tight cycles, the situation is even worse and to our best knowledge, nothing is known. Gyárfás [6] conjectures that there is $c=c(r,k)$ such that every $r$-coloured $\mathcal{K}_{n}^{(k)}$ has a partition into at most $c$ monochromatic tight cycles, but this is open even for the ‘easiest’ case of $3$-uniform hypergraphs and two colours.
Our main result establishes an approximate version for this case.
Theorem 1.1.
For every $\eta>0$ there exists $n_{0}$ such that if $n\geq n_{0}$ then every two-coloring of the edges of the complete 3-uniform hypergraph $\mathcal{K}_{n}^{(3)}$ admits two vertex-disjoint monochromatic tight cycles, of distinct colours,
which cover all but at most $\eta n$ vertices.
Moreover, we can choose the parity of the length of each of the cycles.
We might be interested in choosing the parity of the cycles for the following reason. If $\ell$ is even, then
any 3-uniform tight cycle on $\ell$ edges contains a loose cycle.
Hence, we can deduce that an analogue of Theorem 1.1 holds for loose cycles.
Corollary 1.2.
For every $\eta>0$ there exists $n_{0}$ such that if $n\geq n_{0}$ then every two-coloring of the edges of the complete 3-uniform hypergraph $\mathcal{K}_{n}^{(3)}$ admits two vertex-disjoint monochromatic loose cycles, of distinct colours,
which cover all but at most $\eta n$ vertices.
We believe that the error term $\eta n$ in Theorem 1.1 can be improved and that
every two-colouring of the edges of $\mathcal{K}_{n}^{(3)}$ admits two disjoint monochromatic tight cycles which cover all but at most a constant number $c$ of vertices (for some $c$ independent of $n$).
The proof of Theorem 1.1 is inspired by the work of Haxell et al. [10, 11] and relies on an application
of the hypergraph regularity lemma [4]. This reduces the problem at hand to that of finding, in any two-colouring of the edges of an almost complete 3-uniform hypergraph,
two disjoint monochromatic connected matchings which cover almost all vertices.
Here, as usual, a matching $\mathcal{M}$ in hypergraph $\mathcal{H}$ is a set of pairwise disjoint edges and $\mathcal{M}\subset\mathcal{H}$ is called connected
if between every pair $e,f\in\mathcal{M}$ there is a pseudo-path in $\mathcal{H}$ connecting $e$ and $f$, that is, there
is a sequence $(e_{1},\dots,e_{p})$ of not necessarily distinct edges of $\mathcal{H}$ such that $e=e_{1},f=e_{p}$ and $|e_{i}\cap e_{i+1}|=2$ for each $i\in[p-1]$.
(Note that these pseudo-paths may use vertices outside $V(\mathcal{M})$.) Now, we call a connected matching $\mathcal{M}$ in a $2$-coloured hypergraph a monochromatic connected matching if all edges in $\mathcal{M}$ and all edges on the connecting paths have the same colour.
So, our main contribution reduces to the following result, which might be of independent interest.
Theorem 1.3.
For every $\gamma>0$ there is $t_{0}$ such that the following holds for every $3$-uniform hypergraph $\mathcal{H}$ with $t>t_{0}$ vertices and
$(1-\gamma)\binom{t}{3}$ edges. Any two-colouring of the edges of $\mathcal{H}$ admits two disjoint monochromatic connected matchings covering at least $(1-240\gamma^{\frac{1}{6}})t$ vertices of $\mathcal{H}$.
We prove Theorem 1.3 in Section 2.
In Section 3, we introduce the regularity lemma for hypergraphs and state an embedding result from [11]. The proof of
Theorem 1.1 will be given in Section 4.
2 Monochromatic connected matchings
Before giving the proof of Theorem 1.3 we introduce some notation and auxiliary results.
Let $\mathcal{H}$ denote a $k$-uniform hypergraph, that is, a pair $\mathcal{H}=(V,E)$ with finite vertex set $V=V(\mathcal{H})$ and edge set $E=E(\mathcal{H})\subset\binom{V}{k}$, where $\binom{V}{k}$ denotes the set
of all $k$-element sets of $V$. Often $\mathcal{H}$ will be identified with its edges, that is, $\mathcal{H}\subset\binom{V}{k}$ and for an edge $\{x_{1},\dots,x_{k}\}\in\mathcal{H}$ we often omit brackets and write $x_{1}\dots x_{k}$ only.
A $k$-uniform hypergraph $\mathcal{C}$ is called an $\ell$-cycle
if there is a cyclic ordering of the vertices of $\mathcal{C}$ such that every edge consists of $k$ consecutive vertices, every vertex is contained in an edge and two consecutive edges
(where the ordering of the edges is inherited by the ordering of the vertices) intersect in exactly $\ell$ vertices. For $\ell=1$ we call the cycle loose whereas the cycle is called tight if $\ell=k-1$ (and we do not consider other values of $\ell$).
A tight path is a cycle from which one vertex and all incident edges are deleted. The length of a path, a pseudo-path or a cycle is the number of edges it contains.
As above, two edges in $\mathcal{H}$ are connected if there is a pseudo-path connecting them.
Connectedness is an equivalence relation on the edge set of $\mathcal{H}$
and the equivalence classes are called connected components.
All hypergraphs $\mathcal{H}$ considered from now on are $3$-uniform.
We will need the following result concerning the existence of perfect matchings in 3-uniform hypergraphs with high minimum vertex degree.
Theorem 2.1 ([9]).
For all $\eta>0$ there is a $n_{0}=n_{0}(\eta)$ such that for all $n>n_{0}$, $n\in 3\mathbb{Z}$, the following holds.
Suppose $\mathcal{H}$ is $3$-uniform hypergraph on $n$ vertices such that every vertex is contained in
at least $\left(\frac{5}{9}+\eta\right)\binom{n}{2}$ edges. Then $\mathcal{H}$ contains a perfect matching.
Denote by $\partial\mathcal{H}$ the shadow of $\mathcal{H}$, that is, the set of all pairs $xy$ for which there exists $z$ such that $xyz\in\mathcal{H}$. For a vertex $x$ in a hypergraph $\mathcal{H}$, let $N_{\mathcal{H}}(x)=\{y:xy\in\partial\mathcal{H}\}$. For two vertices $x,y$, let $N_{\mathcal{H}}(x,y)=\{z:xyz\in\mathcal{H}\}$. Note that if $y\in N_{\mathcal{H}}(x)$ (equivalently, $x\in N_{\mathcal{H}}(y)$) then $N_{\mathcal{H}}(x,y)\neq\emptyset$. We call all such pairs $xy$ of vertices active.
Lemma 2.2 ([10], Lemma 4.1).
Let $\gamma>0$ and let $\mathcal{H}$ be a $3$-uniform hypergraph on $t_{\mathcal{H}}$ vertices and at least $(1-\gamma)\binom{t}{3}$ edges.
Then $\mathcal{H}$ contains a subhypergraph $\mathcal{K}$ on $t_{\mathcal{K}}\geq(1-10\gamma^{1/6})t_{\mathcal{H}}$ vertices such that every vertex $x$ of $\mathcal{K}$ is in an active pair of $\mathcal{K}$ and for all active pairs $xy$ we have $|N_{\mathcal{K}}(x,y)|\geq(1-10\gamma^{1/6})t_{\mathcal{K}}$.
We are now ready to prove Theorem 1.3.
Proof of Theorem 1.3.
For given $\gamma>0$ let $\delta=10\gamma^{1/6}$ and apply Theorem 2.1 with $\eta=1/9$ to obtain $n_{0}$.
We choose $t_{0}=\max\left\{\frac{2}{\delta},\frac{n_{0}}{27\delta}\right\}$.
Suppose we are given a two-coloured $3$-uniform hypergraph $\mathcal{H}=\mathcal{H}_{\mathrm{red}}\cup\mathcal{H}_{\mathrm{blue}}$ on $t_{\mathcal{H}}>t_{0}$ vertices and
$(1-\gamma)\binom{t_{\mathcal{H}}}{3}$ edges. Apply Lemma 2.2 to $\mathcal{H}$ with parameter $\gamma$ to obtain
$\mathcal{K}$, $t:=t_{\mathcal{K}}$ with the properties stated in the lemma. We wish to find two monochromatic connected matchings covering all but at most $24\delta t\leq 24\delta t_{\mathcal{H}}$ vertices of $\mathcal{K}$.
Since every vertex is in an active pair in $\mathcal{K}$, we have
$$\displaystyle|N_{\mathcal{K}}(x)|\geq(1-\delta)t\quad\text{for all }x\in V(%
\mathcal{K}).$$
(1)
Let $\mathcal{K}=\mathcal{K}_{\mathrm{red}}\cup\mathcal{K}_{\mathrm{blue}}$ be the colouring of $\mathcal{K}$ inherited from $\mathcal{H}$.
Then a monochromatic component $\mathcal{C}$ of $\mathcal{K}$ is a connected component $\mathcal{K}_{\mathrm{red}}$ or $\mathcal{K}_{\mathrm{blue}}$.
Observation 2.3 ([11], Observation 8.2).
For every vertex $x\in V(\mathcal{K})$ there exists a monochromatic component $\mathcal{C}_{x}$ such that $|N_{\mathcal{C}_{x}}(x)|\geq(1-\delta)t.$
For each $x\in V(\mathcal{K})$ choose arbitrarily one component $\mathcal{C}_{x}$ as in Observation 2.3. Let $R=\{x\in V(\mathcal{K}):\mathcal{C}_{x}\text{ is red}\}$ and $B=\{x\in V(\mathcal{K}):\mathcal{C}_{x}\text{ is blue}\}$, and note that these two sets partition $V(\mathcal{K})$.
Observation 2.4 ([11], Observation 8.4).
If $|R|\geq 6\delta t$ (or $|B|\geq 6\delta t$, respectively), then there is a red component $\mathcal{R}$ (a blue component $\mathcal{B}$) such that $\mathcal{C}_{x}=\mathcal{R}$ ($\mathcal{C}_{x}=\mathcal{B}$) for all but at most $2\delta t$ vertices $x\in R$ ($x\in B$).
Set ${V}_{\mathrm{red}}:=\{x\in R:C_{x}=\mathcal{R}\}$ if $|{V}_{\mathrm{red}}|\geq 6\delta t$, and set ${V}_{\mathrm{blue}}:=\{x\in B:C_{x}=\mathcal{B}\}$ if $|B|\geq 6\delta t$. Otherwise let ${V}_{\mathrm{red}}$, or ${V}_{\mathrm{blue}}$, respectively, be the empty set. Our aim is to find two differently coloured disjoint connected matchings in $\mathcal{K}$ that together cover all but $12\delta t\leq 24\delta t-12\delta t$ vertices of ${V}_{\mathrm{red}}\cup{V}_{\mathrm{blue}}$.
We start by choosing a connected matching of maximal size in $\mathcal{R}\cup\mathcal{B}$. This matching decomposes into two disjoint monochromatic connected matchings, $\mathcal{M}_{\mathrm{red}}\subset\mathcal{R}$ and $\mathcal{M}_{\mathrm{blue}}\subset\mathcal{B}$, which together cover as many vertices as possible. Let ${V}_{\mathrm{red}}^{\prime}={V}_{\mathrm{red}}\setminus V(\mathcal{M}_{\mathrm%
{red}}\cup\mathcal{M}_{\mathrm{blue}})$ and ${V}_{\mathrm{blue}}^{\prime}={V}_{\mathrm{blue}}\setminus V(\mathcal{M}_{%
\mathrm{red}}\cup\mathcal{M}_{\mathrm{blue}})$. We may assume that ${V}_{\mathrm{red}}^{\prime}$ or ${V}_{\mathrm{blue}}^{\prime}$ has at least $12\delta t$ vertices, as otherwise we are done. By symmetry we may assume that
$$|{V}_{\mathrm{red}}^{\prime}|\geq 12\delta t.$$
(2)
Observe that there is no edge $xy$ with $x\in{V}_{\mathrm{red}}^{\prime}$ and $y\in{V}_{\mathrm{blue}}^{\prime}$ such that $xy\in\partial\mathcal{R}\cap\partial\mathcal{B}$. Indeed, any such edge $xy$ constitutes an active pair (by Lemma 2.2) and as $|{V}_{\mathrm{red}}^{\prime}|>\delta t+2$, there must be a vertex $z\in{V}_{\mathrm{red}}^{\prime}$ such that $xyz$ is an edge of $\mathcal{K}$. This contradicts the maximality of the matching $\mathcal{M}_{\mathrm{red}}\cup\mathcal{M}_{\mathrm{blue}}$.
Next, we claim that
$$|{V}_{\mathrm{blue}}^{\prime}|\leq 2\delta t.$$
(3)
Assume otherwise. Then, Observation 2.3 and the choice of the set ${V}_{\mathrm{red}}$ implies that the number of edges between ${V}_{\mathrm{red}}^{\prime}$ and ${V}_{\mathrm{blue}}^{\prime}$ that belong to $\partial\mathcal{R}$ is at least
$$|{V}_{\mathrm{red}}^{\prime}|\cdot(|{V}_{\mathrm{blue}}^{\prime}|-\delta t)%
\geq\frac{1}{2}|{V}_{\mathrm{red}}^{\prime}|\cdot|{V}_{\mathrm{blue}}^{\prime}|.$$
Similarly, there are at least $|{V}_{\mathrm{blue}}^{\prime}|\cdot(|{V}_{\mathrm{red}}^{\prime}|-\delta t)>%
\frac{1}{2}|{V}_{\mathrm{red}}^{\prime}|\cdot|{V}_{\mathrm{blue}}^{\prime}|$ edges between ${V}_{\mathrm{red}}^{\prime}$ and ${V}_{\mathrm{blue}}^{\prime}$ that belong to $\partial\mathcal{B}$. As there is no edge $xy$ with $x\in{V}_{\mathrm{red}}^{\prime}$ and $y\in{V}_{\mathrm{blue}}^{\prime}$ such that $xy\in\partial\mathcal{R}\cap\partial\mathcal{B}$, we have more than $|{V}_{\mathrm{red}}^{\prime}|\cdot|{V}_{\mathrm{blue}}^{\prime}|$ edges from ${V}_{\mathrm{red}}^{\prime}$ to ${V}_{\mathrm{blue}}^{\prime}$. This yields a contradiction and (3) follows.
Because of the maximality of $\mathcal{M}_{\mathrm{red}}\cup\mathcal{M}_{\mathrm{blue}}$, each edge having all its vertices in ${V}_{\mathrm{red}}^{\prime}$ is blue. Fix one such edge $xyz$. Obtain ${V}_{\mathrm{red}}^{\prime\prime}$ from ${V}_{\mathrm{red}}^{\prime}$ by deleting the at most $\delta t$ vertices $w$ with $wx\notin\partial\mathcal{R}$. Consider any edge $x^{\prime}y^{\prime}z^{\prime}$ with $x^{\prime},y^{\prime},z^{\prime}\in{V}_{\mathrm{red}}^{\prime\prime}$. As the pairs $xy$, $xx^{\prime}$, $x^{\prime}y^{\prime}$ are all active, and $|{V}_{\mathrm{red}}^{\prime\prime}|>3\delta t$, there is a vertex $v\in{V}_{\mathrm{red}}^{\prime\prime}$ that forms an edge with each of the three pairs, thus giving a pseudo-path in $\mathcal{K}[{V}_{\mathrm{red}}^{\prime\prime}]$ from $xyz$ to $x^{\prime}y^{\prime}z^{\prime}$. Denote by $\mathcal{B}^{\prime\prime}$ the blue component of $\mathcal{K}[{V}_{\mathrm{red}}^{\prime\prime}]$ that contains $xyz$, and let $\mathcal{B}^{\prime}$ be obtained from $\mathcal{B}^{\prime\prime}$ by deleting at most $2$ vertices and all incident edges, so that $|V[\mathcal{B}^{\prime}]|$ is a multiple of $3$. Then, by (2), we have
$$|V[\mathcal{B}^{\prime}]|\geq|{V}_{\mathrm{red}}^{\prime}|-\delta t-2\geq 10%
\delta t.$$
(4)
Let $x\in V[\mathcal{B}^{\prime}]$ be given. At least $|V[\mathcal{B}^{\prime}]|-\delta t$ vertices $y\in V[\mathcal{B}^{\prime}]$ are such that $xy\in\partial\mathcal{R}$, and, for each such $y$ there are at least $|V[\mathcal{B}^{\prime}]|-\delta t$ vertices $z\in V[\mathcal{B}^{\prime}]$ such that $xyz\in\mathcal{B}^{\prime}$. So, the total number of hyperedges of $\mathcal{B}^{\prime}$ that contain $x$ is at least
$$\frac{1}{2}(|V[\mathcal{B}^{\prime}]|-\delta t)^{2}\ \geq\ \frac{1}{2}\left(%
\frac{9}{10}|V[\mathcal{B}^{\prime}]|\right)^{2}\ \geq\ \frac{2}{3}{|V[%
\mathcal{B}^{\prime}]|\choose 2}.$$
Thus, Theorem 2.1 with $\eta=\frac{1}{9}$ yields a perfect matching $\mathcal{M}_{\mathrm{blue}}^{\prime}$ of $\mathcal{B}^{\prime}$.
At this point, we have three disjoint monochromatic connected matchings, one in red ($\mathcal{M}_{\mathrm{red}}\subset\mathcal{R}$) and two in blue ($\mathcal{M}_{\mathrm{blue}}\subset\mathcal{B}$ and $\mathcal{M}_{\mathrm{blue}}^{\prime}\subset\mathcal{B}^{\prime}$). Together, these matchings cover all but at most $3\delta t+2$ vertices of ${V}_{\mathrm{red}}\cup{V}_{\mathrm{blue}}$ (by (3) and by (4)). In particular, we can assume that
$\mathcal{M}_{\mathrm{blue}}$ contains at least two hyperedges,
as otherwise we can just forget about $\mathcal{M}_{\mathrm{blue}}$ and are done.
Our aim is now to dissolve the blue matching $\mathcal{M}_{\mathrm{blue}}$, and cover its vertices by new red edges, leaving at most $6\delta t$ vertices uncovered. In order to do so, let us first understand where the edges of $\mathcal{M}_{\mathrm{blue}}$ lie.
For convenience, let us call an edge in $\mathcal{K}$ good if two different pairs of its vertices $\{a,b\}$ and $\{c,d\}$ are such that $ab\in\partial\mathcal{R}$ and $cd\in\partial\mathcal{B}$. Notice that every good red edge is contained in $\mathcal{R}$ and every good blue edge is contained in $\mathcal{B}$.
First, we claim that for every edge $uvw\in\mathcal{M}_{\mathrm{blue}}$,
$$|\{u,v,w\}\cap{V}_{\mathrm{blue}}|\leq 1.$$
(5)
Indeed, otherwise there is an edge $uvw\in\mathcal{M}_{\mathrm{blue}}$ with $u,v\in{V}_{\mathrm{blue}}$. By the definition of $\mathcal{B}$, and by (2), there is an active edge $ua\in\partial\mathcal{B}$ with $a\in{V}_{\mathrm{red}}^{\prime}$. As $ua$ is an active pair, as $a$ has very large degree in $\partial\mathcal{R}$, and by (2), there is an edge $uab\in\mathcal{K}$ with $b\in{V}_{\mathrm{red}}^{\prime}$ such that $ab\in\partial\mathcal{R}$. Hence $uab$ is a good edge. Similarly, there
is a good edge $vcd$, with $c,d\in{V}_{\mathrm{red}}^{\prime}\setminus\{a,b\}$. Remove the edge $uvw$ from $\mathcal{M}_{\mathrm{blue}}$ and add edges $uab$ and $vcd$ to either $\mathcal{M}_{\mathrm{red}}$ or $\mathcal{M}_{\mathrm{blue}}$, according to their colour. The resulting matching covers more vertices than the matching $M_{A}\cup\mathcal{M}_{\mathrm{blue}}$, a contradiction. This proves (5).
Next, we claim that there is no edge $uvw\in\mathcal{M}_{\mathrm{blue}}$ with
$$|\{u,v,w\}\cap{V}_{\mathrm{blue}}|=1.$$
(6)
Assume otherwise. Then there is an edge $uvw\in\mathcal{M}_{\mathrm{blue}}$ with $u\in{V}_{\mathrm{blue}}$ and $v,w\in{V}_{\mathrm{red}}$. As in the proof of (5), we can cover $u$ with a good edge $uab$ such that $a,b\in{V}_{\mathrm{red}}^{\prime}$. Moreover, since $vw$ is an active pair, and $v$ has very large degree in $\partial\mathcal{R}$, there is an edge $vwc$ with $c\in{V}_{\mathrm{red}}^{\prime}\setminus\{a,b\}$ and $cv\in\partial R$.
Since $vw\in\partial\mathcal{B}$, the edge $vwc$ is good. So we can remove $uvw$ from $\mathcal{M}_{\mathrm{blue}}$ and add edges $uab$ and $vwc$ to $\mathcal{M}_{\mathrm{red}}\cup\mathcal{M}_{\mathrm{blue}}$, thus covering three additional vertices. This gives the desired contradiction to the choice of $\mathcal{M}_{\mathrm{red}}\cup\mathcal{M}_{\mathrm{blue}}$, and proves (6).
Putting (5) and (6) together, we know that
for every edge $uvw\in\mathcal{M}_{\mathrm{blue}}$ we have $u,v,w\in{V}_{\mathrm{red}}$. Consider any two edges $u_{1}v_{1}w_{1},u_{2}v_{2}w_{2}\in\mathcal{M}_{\mathrm{blue}}$. As before, there are vertices $a,b\in{V}_{\mathrm{red}}^{\prime}$ such that edges $v_{1}w_{1}a,v_{2}w_{2}b$ are good. Now, if there is a red edge $u_{1}u_{2}c$ with $c\in{V}_{\mathrm{red}}^{\prime}$ and $u_{1}c\in\partial\mathcal{R}$ then we can remove edges $u_{1}v_{1}w_{1},u_{2}v_{2}w_{2}$ and add the red edge $u_{1}u_{2}c$ to $\mathcal{M}_{\mathrm{red}}$ and edges $v_{1}w_{1}a,v_{2}w_{2}b$ to $\mathcal{M}_{\mathrm{red}}\cup\mathcal{M}_{\mathrm{blue}}$, according to their colour, contradicting the choice of $\mathcal{M}_{\mathrm{red}}\cup\mathcal{M}_{\mathrm{blue}}$. Therefore, for any choice of $u_{1}v_{1}w_{1},u_{2}v_{2}w_{2}\in\mathcal{M}_{\mathrm{blue}}$, we have that
all edges $$u_{1}u_{2}c$$ with $$c\in{V}_{\mathrm{red}}^{\prime}$$ and $$u_{1}c\in\partial\mathcal{R}$$ are blue.
(7)
Moreover, if there is a blue edge $u_{1}u_{2}x$ with $x\in\{v_{1},w_{1},v_{2},w_{2}\}$ then $u_{1}u_{2}$ is an active pair. In that case, we can calculate as before that an edge $u_{1}u_{2}c$ with $c\in{V}_{\mathrm{red}}^{\prime}$ and $u_{1}c\in\partial\mathcal{R}$ exists, and by (7), this edge is blue. The existence of the blue edge $u_{1}u_{2}x$ implies that we can link $u_{1}u_{2}c$ to $\mathcal{M}_{\mathrm{blue}}$ with a blue tight path. Thus, removing $u_{1}v_{1}w_{1}$ and $u_{2}v_{2}w_{2}$ from $\mathcal{M}_{\mathrm{blue}}$ and adding $v_{1}w_{1}a,v_{2}w_{2}b,u_{1}u_{2}c$ to $\mathcal{M}_{\mathrm{red}}\cup\mathcal{M}_{\mathrm{blue}}$ (where $a,b$ are as above), we obtain a contradiction to the choice of $\mathcal{M}_{\mathrm{red}}\cup\mathcal{M}_{\mathrm{blue}}$. So, for any choice of $u_{1}v_{1}w_{1},u_{2}v_{2}w_{2}\in\mathcal{M}_{\mathrm{blue}}$, we have that
all edges $$u_{1}u_{2}x$$ with $$x\in\{v_{1},w_{1},v_{2},w_{2}\}$$ are red.
(8)
We can now dissolve the edges of $\mathcal{M}_{\mathrm{blue}}$. For this, separate each hyperedge $uvw$ in $\mathcal{M}_{\mathrm{blue}}$ into an edge $uv$ and a single vertex $w$. Let $X$ be the set of all edges $uv$, and let $Y$ be the set of all vertices $w$ obtained in this way. Note that every $uv\in X$ is an active pair in $\mathcal{K}$, and therefore forms a hyperedge $uvw^{\prime}$ with all but at most $\delta t$ of the vertices $w^{\prime}\in Y$. Moreover, all but at most $\delta t$ of these hyperedges $uvw^{\prime}$ are such that $uw^{\prime}\in\partial\mathcal{R}$, because of the large degree $u$ has in $\partial\mathcal{R}$.
Consider the bipartite graph on $X\cup Y$ which has an edge between $uv\in X$ and $w^{\prime}$ whenever the hyperedge $uvw^{\prime}$ exists in $\mathcal{K}$ and $uw^{\prime}\in\partial\mathcal{R}$. Then
for each $X^{\prime}\subseteq X$ we have that $|N(X^{\prime})|\geq|Y|-2\delta t\geq|X^{\prime}|-2\delta t$. So by the defect form of Hall’s Theorem, there is a matching covering all but at most $2\delta t$ vertices of $|X^{\prime}|$. In $\mathcal{K}$, this corresponds to a matching $\mathcal{M}_{\mathrm{red}}^{\prime}$ covering all but at most $6\delta t$ vertices of $V(\mathcal{M}_{\mathrm{blue}})$. By (8), all hyperedges of $\mathcal{M}_{\mathrm{red}}^{\prime}$ are red. Furthermore, since we ensured that every hyperedge in $\mathcal{M}_{\mathrm{red}}$ contains a pair $uw^{\prime}$ that forms an ede of $\partial\mathcal{R}$, we know that $\mathcal{M}_{\mathrm{red}}$ and $\mathcal{M}_{\mathrm{red}}^{\prime}$ belong to the same red component of $\mathcal{K}$. In other words, $\mathcal{M}_{\mathrm{red}}\cup\mathcal{M}_{\mathrm{red}}^{\prime}$ and $\mathcal{M}_{\mathrm{blue}}^{\prime}$ are the two monochromatic connected matchings we had to find.
∎
3 Hypergraph regularity
In this section we introduce the regularity lemma for 3-uniform hypergraphs and state an embedding result from [11].
Graph regularity.
Let $G$ be a graph $G$ and let $X,Y\subseteq V(G)$ be disjoint.
The density of $(X,Y)$ is $d_{G}(X,Y)=\frac{e_{G}(X,Y)}{|X||Y|}$ where
$e_{G}(X,Y)$ denotes the number of edges of $G$ between $X$ and $Y$.
The bipartite graph $G$ on the partition classes $X$ and $Y$ is called $(d,\varepsilon)$-regular, if $|d_{G}(X^{\prime},Y^{\prime})-d|<\varepsilon$ holds for all
$X^{\prime}\subseteq X$ and $Y^{\prime}\subseteq Y$ of size $|X^{\prime}|>\varepsilon|X|$ and $|Y^{\prime}|>\varepsilon|Y|$.
If $d=d_{G}(X,Y)$ we say that $G$ is $\varepsilon$-regular.
Hypergraph regularity.
Let $\mathcal{H}$ be a $3$-uniform hypergraph. Let $P=P^{12}\cup P^{13}\cup P^{23}$ with $V(P)\subset V(\mathcal{H})$ be a tripartite graph which we also refer to as triad.
By $\mathcal{T}(P)$ denote the
$3$-uniform hypergraph on $V(P)$ whose edges are the triangles of $P$.
The density of $\mathcal{H}$ with respect to $P$ is
$$d_{\mathcal{H}}(P)=\frac{|\mathcal{H}\cap\mathcal{T}(P)|}{|\mathcal{T}(P)|}.$$
Similarly, for a tuple $\vec{\mathcal{Q}}=(Q_{1},\dots,Q_{r})$ of subgraphs of $P$, we define the density of $\mathcal{H}$ with respect to $\vec{\mathcal{Q}}$ as
$$d_{\mathcal{H}}(\vec{\mathcal{Q}})=\frac{|\mathcal{H}\cap\bigcup_{i\in[r]}%
\mathcal{T}(Q_{i})|}{|\bigcup_{i\in[r]}\mathcal{T}(Q_{i})|}.$$
Let $\alpha,\delta>0$ and let $r>0$ be an integer. We say that $\mathcal{H}$ is $(\alpha,\delta,r)$-regular with respect to $P$ if, for every $r$-tuple $\vec{\mathcal{Q}}=(Q_{1},\dots,Q_{r})$ of subgraphs of $P$ satisfying $|\bigcup_{i\in[r]}\mathcal{T}(Q_{i})|>\delta|\mathcal{T}(P)|$, we have $|d_{\mathcal{H}}(\vec{\mathcal{Q}})-\alpha|<\delta$. If $\alpha=d_{\mathcal{H}}(P)$ we say that $\mathcal{H}$ is $(\delta,r)$-regular with respect to $P$, and in the same situation, we say $P$ is $(\delta,r)$-regular (with respect to $\mathcal{H}$).
Moreover, if the bipartite graphs $P^{12},P^{13},P^{23}$ of an $(\alpha,\delta,r)$-regular $P=P^{12}\cup P^{13}\cup P^{23}$ are
$(1/\ell,\varepsilon)$-regular then
we say that the pair $(\mathcal{H},P)$ is an $(\alpha,\delta,\ell,r,\varepsilon)$-regular complex.
Finally, a partition
of $V$ into $V_{0}\cup V_{1}\cup\dots\cup V_{t}$ is called an equipartition if $|V_{0}|<t$ and $|V_{1}|=V_{2}|=\dots=|V_{t}|$.
We state the regularity lemma for $3$-uniform hypergraphs [4] as presented in [14].
Theorem 3.1 (Regularity Lemma for $3$-uniform Hypergraphs).
For all $\delta,t_{0}>0$, all integer-valued functions $r=r(t,\ell)$, and all decreasing sequences $\varepsilon(\ell)>0$ there exist constants $T_{0},L_{0}$ and $N_{0}$
such that every $3$-uniform hypergraph $\mathcal{H}$ with at least $N_{0}$ vertices admits a vertex equipartition
$$V(\mathcal{H})=V_{0}\cup V_{1}\cup\dots\cup V_{t}\qquad\text{ with }t_{0}\leq t%
<T_{0},$$
and, for each pair $i,j,$ $1\leq i<j\leq t,$ an edge partition of the complete bipartite graph
$$K(V_{i},V_{j})=\bigcup_{k\in[\ell]}P^{ij}_{k}\qquad\text{ with }1\leq\ell<L_{0}$$
such that
1.
all graphs $P_{k}^{ij}$ are $(1/\ell,\varepsilon(\ell))$-regular.
2.
$\mathcal{H}$ is $(\delta,r)$-regular with respect to all but at most $\delta\ell^{3}t^{3}$ tripartite graphs $P_{a}^{hi}\cup P_{b}^{hj}\cup P_{c}^{ij}$.
Note that the same partitions satisfy the conclusions of Theorem 3.1 for the complement of $\mathcal{H}$ as well.
Further, as noted in [11] by choosing a random index $k_{ij}\in[\ell]$ for each pair $(V_{i},V_{j})$ Markov’s inequality yields that
with positive probability there are less than $2\delta t^{3}$ chosen triads
which fail to be $(\delta,r)$-regular. Hence one obtains the following.
Observation 3.2.
In the partition produced by Theorem 3.1 there is a family $\mathcal{P}$ of bipartite graphs $P^{ij}=P^{ij}_{k_{ij}}$ with vertex classes $V_{i},V_{j}$, where $1\leq i<j\leq t$, such that $\mathcal{H}$ is $(\delta,r)$-regular with respect to all but at most $2\delta t^{3}$ tripartite graphs $P^{hi}\cup P^{hj}\cup P^{ij}$.
We end this section with a result from [16] and [11] which allows embedding tight paths in regular complexes.
In the following, an $S$-avoiding tight path is one which does not contain any vertex from $S$.
(Note that although Lemma 4.6 from [11] is stated slightly differently, its proof actually yields the version below.)
Lemma 3.3 ([11], Lemma 4.6).
For each $\alpha\in(0,1)$ there exist $\delta_{1}>0$ and sequences $r(\ell)$, $\varepsilon(\ell)$,
and $n_{1}(\ell)$, for $\ell\in\mathbb{N}$, with the following property.
For each $\ell\in\mathbb{N}$, and each $\delta\leq\delta_{1}$,
if $(\mathcal{H},P)$ is a $(d_{\mathcal{H}}(P),\delta,\ell,r(\ell),\varepsilon(\ell))$-complex with $d_{\mathcal{H}}(P)\geq\alpha$ and
all of the three vertex classes of $P$ have the same size $n>n_{1}(\ell)$, then there is a subgraph
$P_{0}$ on at most $27\sqrt{\delta}n^{2}/\ell$ edges of $P$ such that, for all ordered pairs of disjoint edges
$(e,f)\in(P\setminus P_{0})\times(P\setminus P_{0})$ there is $m=m(e,f)\in[3]$ such that the following holds. For every $S\subseteq V(\mathcal{H})\setminus(e\cup f)$ with
$|S|<n/(\log n)^{2}$, and for each $\ell$ with $3\leq\ell\leq(1-2\delta^{\frac{1}{4}})n$, there is an $S$-avoiding tight path from $e$ to $f$ of length $3\ell+m$ in $\mathcal{H}$.
4 Proof of Theorem 1.1
We follow a procedure suggested by Łuczak in [12] for graphs and used for tight cycles in $3$-uniform hypergraphs in [11].
Proof of Theorem 1.1.
For given $\eta>0$ we apply Lemma 1.3 with $\gamma=(\eta/480)^{6}$ to obtain $t_{0}$.
With foresight apply Lemma 3.3 with $\alpha=1/2$ to obtain $\delta_{1}$, and sequences $r(\ell)$, $\varepsilon(\ell)$,
and $n_{1}(\ell)$. Finally, apply Theorem 3.1 with $t_{0}$, $r(t,\ell)=r(\ell)$, $\varepsilon(\ell)$, $n_{1}(\ell)$ and $\delta=\min\{\delta_{1}/2,\gamma/48,(\eta/16)^{4}\}$
to obtain constants $T_{0}$, $L_{0}$ and $N_{0}$.
Given a two-colouring $\mathcal{K}_{n}=\mathcal{H}_{\mathrm{red}}\cup\mathcal{H}_{\mathrm{blue}}$ of the 3-uniform complete hypergraph $\mathcal{K}_{n}$ on $n>N_{0}$ vertices.
Apply Theorem 3.1 with the chosen constants to $\mathcal{H}_{\mathrm{red}}$ to
obtain partitions
$$V(\mathcal{K}_{n})=V_{0}\cup V_{1}\cup\dots\cup V_{t}\quad\text{and}\quad K(V_%
{i},V_{j})=\bigcup_{k\in[\ell]}P^{ij}_{k},1\leq i<j\leq t$$
with $t_{0}\leq t<T_{0}$, and $\ell<L_{0}$ which satisfy the properties detailed in Theorem 3.1.
The partitions satisfy the same properties for $\mathcal{H}_{\mathrm{blue}}$ as it is the complement hypergraph of $\mathcal{H}_{\mathrm{red}}$.
Observation 3.2 then yields a family of $(1/\ell,\varepsilon)$-regular bipartite graphs $P^{ij}=P^{ij}_{k_{ij}}$, one for each pair
$(V_{i},V_{j})$, $1\leq i<j\leq t$, such that $\mathcal{H}_{\mathrm{red}}$ (and thus also $\mathcal{H}_{\mathrm{blue}}$) is $(\delta,r)$-regular with respect to all but at most $2\delta t^{3}$ tripartite graphs $P^{ijk}=P^{ij}\cup P^{ik}\cup P^{jk}$.
We use this family to construct the reduced hypergraph $\mathcal{R}$ which has the vertex set $[t]$ and the edge set consisting of all triples $ijk$ such that $P^{ijk}$ is $(\delta,r)$-regular.
Further, colour the edge $ijk$ red if $d_{\mathcal{H}_{\mathrm{red}}}(P^{ijk})\geq 1/2$ and blue otherwise. Then
$\mathcal{R}$ has at least $\binom{t}{3}-2\delta t^{3}>(1-\gamma)\binom{t}{3}$ edges and as $d_{\mathcal{H}_{\mathrm{red}}}(P^{ijk})+d_{\mathcal{H}_{\mathrm{blue}}}(P^{ijk%
})=1$
we obtain a two-colouring of $\mathcal{R}=\mathcal{R}_{\mathrm{red}}\cup\mathcal{R}_{\mathrm{blue}}$.
Since $t\geq t_{0}$ Lemma 1.3 yields two disjoint monochromatic connected matchings
$\mathcal{M}_{\mathrm{red}}$ and $\mathcal{M}_{\mathrm{blue}}$
which cover all but at most $240\gamma^{\frac{1}{6}}t\leq\eta t/2$ vertices of $\mathcal{R}$ and in what follows we will
turn these connected matchings into disjoint monochromatic tight cycles in $\mathcal{K}_{n}$.
We start by choosing a red pseudo-path $\mathcal{Q}_{\mathrm{red}}=(e_{1},\dots,e_{p})\subset\mathcal{R}_{\mathrm{red}}$ which contains the matching $\mathcal{M}_{\mathrm{red}}$. This is possible since $\mathcal{M}_{\mathrm{red}}$ is a connected matching, and so, consecutive matching edges $g_{s},g_{s+1}\in\mathcal{M}_{\mathrm{red}}$ are connected by a red pseudo-path of length at most $\binom{t}{3}$.
The concatenation of these paths then forms a $\mathcal{Q}_{\mathrm{red}}$ as desired. In the same manner, choose a
blue pseudo-path $\mathcal{Q}_{\mathrm{blue}}=(e^{\prime}_{1},\dots,e^{\prime}_{q})\subset%
\mathcal{R}_{\mathrm{blue}}$ containing the matching $\mathcal{M}_{\mathrm{blue}}$. Note that although $\mathcal{M}_{\mathrm{red}}$ and $\mathcal{M}_{\mathrm{blue}}$ are disjoint, the two paths $\mathcal{Q}_{\mathrm{red}}$ and $\mathcal{Q}_{\mathrm{blue}}$
may have vertices in common.
For each edge $\{i,j,k\}=e_{s}\in\mathcal{Q}_{\mathrm{red}}$ let $P^{s}$ denote the triad $P^{ij}\cup P^{ik}\cup P^{jk}$ on the partition classes $V_{i}\cup V_{j}\cup V_{k}$ and recall that $P^{s}$ is $(\delta,r)$-regular (with respect to $\mathcal{H}_{\mathrm{red}}$).
Lemma 3.3 guarantees that one can find long tight paths in the complex $(\mathcal{H}_{\mathrm{red}},P^{s})$ for each matching edge $e_{s}\in\mathcal{M}_{\mathrm{red}}\subset\mathcal{Q}_{\mathrm{red}}$
which exhaust almost all vertices of $V_{i}\cup V_{j}\cup V_{k}$.
Using the connectedness of $\mathcal{Q}_{\mathrm{red}}$ and Lemma 3.3, we then want to
connect these long tight paths by short tight paths, hence obtain a tight cycle $\mathcal{C}_{\mathrm{red}}$ which covers almost all vertices of $\mathcal{K}_{n}$ spanned by $\mathcal{M}_{\mathrm{red}}$. We wish to do the same with $\mathcal{Q}_{\mathrm{blue}}$ to obtain
a tight cycle $\mathcal{C}_{\mathrm{blue}}$ which covers almost all vertices of $\mathcal{K}_{n}$ spanned by $\mathcal{M}_{\mathrm{blue}}$. The two cycles $\mathcal{C}_{\mathrm{red}}$ and $\mathcal{C}_{\mathrm{blue}}$ then exhaust most of the vertices of $\mathcal{K}_{n}$.
To keep the two cycles disjoint, however, the strategy will be slightly less straightforward. First, we will find two disjoint short tight cycles
$\mathcal{C}_{\mathrm{red}}^{\prime}$ and $\mathcal{C}_{\mathrm{blue}}^{\prime}$ in $\mathcal{K}_{n}$ visiting all triads $P^{s}$ corresponding to edges $e_{s}\in\mathcal{M}_{\mathrm{red}}$ and $P^{\prime s}$ corresponding to $e^{\prime}_{s}\in\mathcal{M}_{\mathrm{blue}}$, respectively.
Then, for each edge $e_{s}\in\mathcal{M}_{\mathrm{red}}$, and each edge $e^{\prime}_{s}\in\mathcal{M}_{\mathrm{blue}}$, we will replace parts of the cycles $\mathcal{C}_{\mathrm{red}}^{\prime}$, and of $\mathcal{C}_{\mathrm{blue}}^{\prime}$, i.e., paths corresponding to $e_{s}$ and $e^{\prime}_{s}$ by long tight paths as mentioned above.
We now give the details of this idea.
For each $s=1,\dots,p$, apply Lemma 3.3 to the complex $(\mathcal{H}_{\mathrm{red}},P^{s})$ to obtain the subgraph $P^{s}_{0}\subset P^{s}$ of “prohibited” edges
and let
$$B_{s}=(P^{s}\setminus P^{s}_{0})\cap(P^{s+1}\setminus P^{s+1}_{0})$$
which is a bipartite graph on the partition classes $V_{i}\cup V_{j}$ where $\{i,j\}=e_{s}\cap e_{s+1}$. We choose mutually distinct edges $f_{s},g_{s}\in B_{s}$, $s\in[p-1]$ which is possible due to the restriction on
$|P_{0}^{s}|$ provided $n$ is sufficiently large. Using Lemma 3.3 we then find a short tight cycle $\mathcal{C}_{\mathrm{red}}^{\prime}$ by
concatenating disjoint paths each of length at most 12 between $f_{1}$ and $f_{2}$, $f_{2}$ and $f_{3}$ $\dots$ between $f_{p-1}$ and $g_{p-1}$ and backwards between $g_{p-1}$ and $g_{p-2}$ $\dots$ and
finally between $g_{1}$ and $f_{1}$.
Note that the lemma allows the paths to be $S$-avoiding for any vertex set $S$ of size $|S|<n^{\prime}/(\log n^{\prime})^{2}$ where $n^{\prime}$ is the size of the partition classes.
Therefore, to guarantee the disjointness of the paths, we simply choose $S$ to be the vertices of the paths constructed so far which
has size at most $24p$, i.e., independent of $n^{\prime}>n/2t$. In the same way choose a short tight cycle $\mathcal{C}_{\mathrm{blue}}^{\prime}$ disjoint
from $\mathcal{C}_{\mathrm{red}}^{\prime}$ by including $V(\mathcal{C}_{\mathrm{red}}^{\prime})$ to $S$ in the applications of Lemma 3.3.
Let $S^{\prime}=V(\mathcal{C}_{\mathrm{red}}^{\prime})\cup V(\mathcal{C}_{\mathrm{%
blue}}^{\prime})$ which satisfies $|S^{\prime}|<n^{\prime}/(\log n^{\prime})^{2}$. It is easy to see that for each $e_{s}\in\mathcal{M}_{\mathrm{red}}$ there are two non-prohibited edges in $P^{s}$, connected by a subpath of $\mathcal{C}_{\mathrm{red}}^{\prime}$
which is entirely contained in $(\mathcal{H}_{\mathrm{red}},P^{s})$. Hence, by Lemma 3.3 we can replace this short path by an $S$-avoiding path in $(\mathcal{H}_{\mathrm{red}},P^{s})$ which covers all but at most $4\delta^{1/4}n^{\prime}$
vertices and having any desired parity. Doing this for all $e_{s}\in\mathcal{M}_{\mathrm{red}}$ and all $e^{\prime}_{s}\in\mathcal{M}_{\mathrm{blue}}$ and noting that $n^{\prime}\leq n/t$ we obtain two monochromatic disjoint tight cycles which cover all but at most
$$\displaystyle\big{(}|\mathcal{M}_{\mathrm{red}}|+|\mathcal{M}_{\mathrm{blue}}|%
\big{)}4\delta^{1/4}n^{\prime}+$$
$$\displaystyle\big{(}|V(\mathcal{R})|-|\mathcal{M}_{\mathrm{red}}|+|\mathcal{M}%
_{\mathrm{blue}}|\big{)}n^{\prime}+|V_{0}|$$
$$\displaystyle\leq\frac{1}{4}\eta n+\frac{1}{2}\eta n+t\leq\eta n$$
vertices of $\mathcal{K}_{n}$, and have any desired parity. This finishes the proof of the theorem.
∎
References
[1]
Allen, P.,
Covering two-edge-coloured complete graphs with two disjoint monochromatic cycles,
Combin. Probab. Comput., 17 (2008), 471–486.
[2]
Ayel, J.,
Sur l’existence de deux cycles supplémentaires unicolores, disjoints et de couleurs différentes dans un graphe complet bicolore,
Doctoral dissertation, Université scientifique et médicale de Grenoble (1979).
[3]
Bessy, S., & Thomassé, S.,
Partitioning a graph into a cycle and an anticycle, a proof of Lehel’s conjecture,
Journal of Combinatorial Theory, Series B, 100(2) (2010), 176–180.
[4]
Frankl, P., & Rödl, V.,
Extremal problems on set systems,
Random Structures & Algorithms, 20(2) (2002), 131–164.
[5]
Fujita, S., Liu, H. and Magnant, C.,
Monochromatic Structures in Edge-coloured Graphs and Hypergraphs - A survey,
Preprint 2015.
[6]
Gyárfás, A.,
Vertex covers by monochromatic pieces - A survey of results and problems,
Discrete Math., 339(7) (2016), 1970–1977.
[7]
Gyárfás, A., Ruszinkó, M., Sárközy, G., & Szemerédi, E.,
An improved bound for the monochromatic cycle partition number,
Journal of Combinatorial Theory, Series B, 96 (2006), 855–873.
[8]
Gyárfás, A., & Sárközy, G.,
Monochromatic path and cycle partitions in hypergraphs,
Electron. J. Combin., 20(1) (2013) P18.
[9]
Hàn, H., Person, Y., & Schacht, M.,
On perfect matchings in uniform hypergraphs with large minimum vertex degree,
SIAM Journal on Discrete Mathematics, 23(2) (2009), 732–748.
[10]
Haxell, P., Łuczak, T., Peng, Y., Rödl, V., Ruciński, A., Simonovits, M., & Skokan, J.,
The Ramsey number for hypergraph cycles I,
Journal of Combinatorial Theory, Series A, 113(1) (2006), 67–83.
[11]
Haxell, P., Łuczak, T., Peng, Y., Rödl, V., Ruciński, A., & Skokan, J.,
The Ramsey number for 3-uniform tight hypergraph cycles,
Combinatorics, Probability and Computing, 18(1-2) (2009), 165–203.
[12]
Łuczak, T.,
$R(C_{n},C_{n},C_{n})\leq(4+o(1))n$,
Journal of Combinatorial Theory, Series B, 75(2) (1999), 174–187.
[13]
Łuczak, T., Rödl, V., & Szemerédi, E.,
Partitioning two-coloured complete graphs into two monochromatic cycles,
Combinatorics, Probability and Computing, 7(04) (1998), 423–436.
[14]
Rödl, V., Ruciński, A., & Szemerédi, E.,
A Dirac-type theorem for 3-uniform hypergraphs,
Combinatorics, Probability and Computing, 15(1-2) (2006), 229–251.
[15]
Pokrovskiy, A.,
Partitioning edge-coloured complete graphs into monochromatic cycles and paths,
Journal of Combinatorial Theory, Series B, 106 (2014), 70–97.
[16]
Polcyn, J., Rödl, V., Ruciński, A., & Szemerédi, E.,
Short paths in quasi-random triple systems with sparse underlying graphs,
Journal of Combinatorial Theory, Series B, 96(4) (2006), 584–607.
[17]
Sárközy, G.,
Improved monochromatic loose cycle partitions in hypergraphs,
Discrete Math., 334 (2014), 52–62. |
$T$-dependence of the axion mass when the $U_{A}(1)$
and chiral symmetry breaking are tied ††thanks: Presented at Excited QCD 2019, Schladming, Austria,
Jan. 30 – Feb. 3, 2019.
Dubravko Klabučar
Davor Horvatić
Dalibor Kekez
Physics Department, Faculty of Science – PMF, University of Zagreb, Croatia
Rudjer Bošković Institute, Zagreb, Croatia
Abstract
Modulo the scale of spontaneous breaking of Peccei-Quinn symmetry, the
axion mass $m_{\rm a}(T)$ is given by the QCD topological susceptibility
$\chi(T)$ at all temperatures $T$. From an approach tying the $U_{A}(1)$ and
chiral symmetry breaking and getting good $T$-dependence of $\eta$ and
$\eta^{\prime}$ mesons, we get $\chi(T)$ for an effective Dyson-Schwinger
model of nonperturbative QCD. Comparison with lattice results for $\chi(T)$,
and thus also for $m_{\rm a}(T)$, shows good agreement for temperatures ranging
from zero up to the double of the chiral restoration temperature $T_{c}$.
\PACS
14.80.Mz , 14.80.Va , 12.38.Aw
1 Introduction
The fundamental theory of strong interactions, QCD, has the so-called Strong CP
problem. Namely, there is no experimental evidence of any CP-symmetry violation
in strong interactions, although there is in principle no reason why the QCD
Lagrangian should not include the so-called $\Theta$-term
${\cal L}^{\Theta}$, where gluon fields $F^{b}_{\mu\nu}(x)$ comprise the
CP-violating combination $Q(x)$,
$${\cal L}^{\Theta}(x)\,=\,\Theta\,\frac{g^{2}}{\,64\,\pi^{2}\,}\,\epsilon^{\mu%
\nu\rho\sigma}\,F^{b}_{\mu\nu}(x)\,F^{b}_{\rho\sigma}(x)\,\equiv\,\Theta\,Q(x)\,.$$
(1)
Admittedly, ${\cal L}^{\Theta}$ can be rewritten
as a total divergence, but, unlike in QED, this does not enable
discarding it in spite of the gluon fields vanishing sufficiently fast as
$|x|\to\infty$. This is because of nontrivial topological structures in QCD, such as instantons,
which are important for, e.g., solving of the $U_{A}(1)$ problem
and yielding the anomalously large mass of the $\eta^{\prime}$ meson.
Thus, there is no reason why the coefficient $\Theta$ of this term should be
of a very different magnitude from the coefficients of the other, CP-symmetric
terms comprising the usual CP-symmetric QCD Lagrangian. Nevertheless, the
experimental bound on the coefficient of the term is extremely low,
$|\Theta|<10^{-10}$ [1], and consistent with zero.
This is the mystery of the missing strong CP violation:
why is $\Theta$ so small?
Various proposed theoretical solutions stood the test of time with varying success.
A long-surviving solution, which is actually the preferred solution nowadays,
is a new particle beyond the Standard Model - the axion. Important is also that
axions turned out to be very interesting also for cosmology, as promising
candidates for dark matter. (See, e.g., [2, 3].)
2 Axion mass from the non-Abelian axial anomaly
Peccei and Quinn introduced [4, 5] a new axial global
symmetry $U(1)_{PQ}$ which is broken spontaneously at some scale $f_{\rm a}$.
This presumably huge [6] but otherwise presently unknown
scale is the key free parameter of axion theories, which determines the absolute
value of the axion mass $m_{\rm a}$. However, this constant cancels from
combinations such as $m_{\rm a}(T)/m_{\rm a}(0)$. Hence, useful insights and applications are possible in spite of $f_{\rm a}$ being not known.
We have often, including applications at $T>0$
[7, 8, 9, 10, 11],
employed a chirally well-behaved relativistic bound-state approach to modeling
nonperturbative QCD through Dyson-Schwinger equations (DSE) for Green’s functions
of the theory. (For reviews, see [12, 13, 14]
for example.)
Such calculations can yield model predictions on the QCD topological susceptibility,
including its temperature dependence $\chi(T)$, which are correctly related to the
QCD dynamical chiral symmetry breaking (DChSB) and restoration. It turns out that
$\chi(T)$ is precisely that factor in the axion mass $m_{\rm a}(T)$, which carries
the nontrivial $T$-dependence.
2.1 Axions as quasi-Goldstone bosons
The pseudoscalar axion field $\mbox{\large a}(x)$ arises as the
(would-be massless) Goldstone boson of the spontaneous breaking of the
Peccei-Quinn symmetry [15, 16].
The axion contributes to the total Lagrangian its kinetic term and its
interaction with the standard-model fermions. But what is important for
the resolution of the strong CP problem, is that the axion also couples
to the topological charge density operator operator $Q(x)$ in
Eq. (1). Then, the $\Theta$-term in the QCD Lagrangian
gets modified to
$${\cal L}^{\Theta\,+}_{\rm axion}\,=\,{\cal L}^{\Theta}\,+\frac{\mbox{\large a}%
(x)}{f_{\rm a}}\,Q(x)\,=\,\left(\,\Theta\,+\frac{\mbox{\large a}}{f_{\rm a}}%
\right)\,\frac{g^{2}}{64\pi^{2}}\,\epsilon^{\mu\nu\rho\sigma}F^{b}_{\mu\nu}\,F%
^{b}_{\rho\sigma}\,.$$
(2)
Through this coupling of the axion to gluons, the $U(1)_{PQ}$ symmetry is also broken
explicitly by the $U_{A}(1)$ non-Abelian, gluon axial anomaly, so that the axion
has a non-vanishing mass, $m_{\rm a}\neq 0$ [15, 16].
Gluons generate an effective axion potential, and its minimization leads to
the axion expectation value $\langle\mbox{\large a}\rangle$ such that the
modified coefficient, multiplying the topological charge density $Q(x)$,
should vanish:
$$\Theta\,+\frac{\langle\mbox{\large a}\rangle}{f_{\rm a}}\,=\,0\,.$$
(3)
The strong CP problem is thereby solved, irrespective of the initial
value of $\Theta$. Relaxation from any $\Theta$-value in
the early Universe towards the minimum at Eq. (3),
is known as misalignment production, and the the resulting axion oscillation energy
is a cold dark matter candidate (e.g., see [2, 3]).
2.2 Axion mass from anomalous $U_{A}(1)$ breaking driven by DChSB
A direct measure of the $U_{A}(1)$ symmetry breaking is the topological susceptibility
$\chi$, given by the convolution of the time-ordered product $\cal T$ of the
topological charge densities $Q(x)$ defined by Eq. (1) [or
Eq. (2)]:
$$\chi\,=\,\int d^{4}x\;\langle 0|\,{\cal T}\,Q(x)\,Q(0)\,|0\rangle\;.$$
(4)
The expansion of the effective axion
potential reveals in its quadratic term that the axion mass squared
(times $f_{\rm a}^{2}$) is equal111To a high level of accuracy, since
corrections to Eq. (5) are of the order $M_{\pi}^{2}/f_{\rm a}^{2}$
[20], where the pion mass $M_{\pi}$ is negligible.
to the QCD topological susceptibility.
This holds for all temperatures $T$:
$$m_{\rm a}^{2}(T)\,f_{\rm a}^{2}\,=\,\chi(T)\,.$$
(5)
On the other hand, in our study [11] of the $T$-dependence
of the $\eta$ and $\eta^{\prime}$ masses and the influence of the anomalous $U_{A}(1)$
breaking and restoration, we used the light-quark-sector result
[17, 18, 19]
$$\chi(T)\,=\,\frac{-\,1}{\,\,\,\frac{1}{\,m_{u}\,\langle{\bar{u}}u\,(T)\rangle}%
+\frac{1}{\,m_{d}\,\langle{\bar{d}}d\,(T)\rangle}+\frac{1}{\,m_{s}\,\langle{%
\bar{s}}s\,(T)\rangle}\,\,\,}\,+\,{\cal C}_{m}\,$$
(6)
where ${\cal C}_{m}$ is a very small correction term of higher orders in the
small current quark masses $m_{q}$ ($q=u,d,s$), and in the present context
we do not consider it further. Thus, the overwhelming part, namely the
leading term of $\chi$, is given by the quark condensates
$\langle{\bar{q}}q\rangle$ ($q=u,d,s$), which arise as order parameters
of DChSB. Their temperature dependence determines that of $\chi(T)$,
which in turn determines the $T$-dependence of the anomalous part of
the pseudoscalar meson masses in the $\eta$-$\eta^{\prime}$ complex.
This is the mechanism of Ref. [11], how DChSB and
chiral restoration drive, respectively, the anomalous breaking and
restoration of the $U_{A}(1)$ symmetry of QCD.
Now, Eqs. (5) and (6) show that
this mechanism determines also the $T$-dependence of the axion mass.
To describe $\eta^{\prime}$ and $\eta$ mesons, it is essential to include $U_{A}(1)$
symmetry breaking at least at the level of the masses. This could be done
simply [23, 24, 25], by adding the
anomalous contribution to isoscalar meson masses as a perturbation, thanks
to the fact that the $U_{A}(1)$ anomaly is suppressed in the limit of large
number of QCD colors $N_{c}$ [26, 27]. Concretely,
Ref. [25] adopted Shore’s equations [28],
where the $U_{A}(1)$-anomalous contribution to the light pseudoscalar masses
is expressed through the condensates of light quarks with non-vanishing
current masses. They are thus used also in $\chi(T)$ (6), since
this approach has recently been extended [11] to $T>0$.
This gave us our results for $\chi(T)$ depicted in Fig. 1.
Indeed, the now established smooth, crossover behavior around
the pseudocritical temperature $T_{c}$ for the chiral transition,
is obtained for the DChSB condensates of realistically massive
light quarks – i.e., the quarks with realistic
explicit chiral symmetry breaking [11].
In contrast, using in Eq. (6) the massless quark condensate
$\langle{\bar{q}}q\rangle_{0}$ (which drops sharply to zero at $T_{c}$)
instead of the “massive” ones,
would dictate a sharp transition of the second order at $T_{c}$
[10, 11] also for $\chi(T)$.
Obviously, this would imply that axions are massless for $T>T_{c}$.
In Fig. 1, we present (the leading term of) our model-calculated
[11] $\chi(T)^{1/4}$, depicted as the solid curve. Due
to Eq. (5), this is our model prediction for
$\sqrt{m_{\rm a}(T)\,f_{\rm a}}$. For temperatures
up to $T\approx 2.3\,T_{c}$, we compare it to the lattice results of Petreczky
et al., [21] and of Borsany et al.,
[22], rescaled to the relative temperature $T/T_{c}$.
3 Summary
The axion mass and its temperature dependence $m_{\rm a}(T)$
can be calculated in an effective model of nonperturbative QCD
(up to the constant scale parameter $f_{\rm a}$) as the square root
of the topological susceptibility $\chi(T)$. We obtained
it from the condensates of $u$-, $d$- and $s$-quarks and antiquarks
calculated in the SDE approach using a simplified nonperturbative
model interaction [11].
Our prediction on $m_{\rm a}(T)$ is thus supported by the fact
that our topological susceptibility also yields the $T$-dependence
of the $U_{A}(1)$ anomaly-influenced masses of $\eta^{\prime}$ and $\eta$
mesons which is consistent with experimental evidence [11].
Our result on $\chi(T)$ and the related axion mass is qualitatively
similar to the one obtained in the framework of the NJL model [29].
Our topological susceptibility is also qualitatively similar to the
pertinent lattice results [21, 22],
except that our dynamical model could so far access only much smaller
range of temperatures, $T<2.4\,T_{c}$.
On the other hand, the lattice supports the smooth crossover transition
of $\chi(T)$, which is, in our approach, the natural consequence of
employing the massive-quark condensates exhibiting crossover around the
chiral restoration temperature $T_{c}$. Hence, the (partial) $U_{A}(1)$
restoration observed in Ref. [11] must also be a
crossover, which in the present work, as well as in its longer
counterpart [30] containing a detailed analysis
of the model parameter dependence, translates into the corresponding
smooth $T$-dependence of the axion mass.
References
[1]
C. A. Baker et al.,
Phys. Rev. Lett. 97, 131801 (2006).
[2]
O. Wantz and E. P. S. Shellard,
Phys. Rev. D 82, 123508 (2010).
[3]
J. E. Kim, S. Nam and Y. K. Semetzidis,
Int. J. Mod. Phys. A 33, no. 03, 1830002 (2018).
[4]
R. D. Peccei and H. R. Quinn,
Phys. Rev. Lett. 38, 1440 (1977).
[5]
R. D. Peccei and H. R. Quinn,
Phys. Rev. D 16, 1791 (1977).
[6]
M. Tanabashi et al. [Particle Data Group],
Phys. Rev. D 98, no. 3, 030001 (2018).
[7]
D. Horvatić, D. Klabučar and A. Radzhabov,
Phys. Rev. D 76, 096009 (2007).
[8]
D. Horvatić, D. Blaschke, D. Klabučar and A. E. Radzhabov,
Phys. Part. Nucl. 39, 1033 (2008).
[9]
D. Horvatić, D. Blaschke, Y. Kalinovsky, D. Kekez and D. Klabučar,
Eur. Phys. J. A 38, 257 (2008).
[10]
S. Benić, D. Horvatić, D. Kekez and D. Klabučar,
Phys. Rev. D 84, 016006 (2011).
[11]
D. Horvatić, D. Kekez and D. Klabučar,
Phys. Rev. D 99, no. 1, 014007 (2019).
[12]
R. Alkofer and L. von Smekal,
Phys. Rept. 353, 281 (2001).
[13]
C. D. Roberts and S. M. Schmidt,
Prog. Part. Nucl. Phys. 45, S1 (2000).
[14]
C. S. Fischer,
J. Phys. G 32, R253 (2006).
[15]
S. Weinberg,
Phys. Rev. Lett. 40, 223 (1978).
[16]
F. Wilczek,
Phys. Rev. Lett. 40, 279 (1978).
[17]
P. Di Vecchia and G. Veneziano,
Nucl. Phys. B 171, 253 (1980).
[18]
H. Leutwyler and A. V. Smilga,
Phys. Rev. D 46, 5607 (1992).
[19]
S. Dürr,
Nucl. Phys. B 611, 281 (2001).
[20]
M. Gorghetto and G. Villadoro,
JHEP 1903, 033 (2019).
[21]
P. Petreczky, H. P. Schadler and S. Sharma,
Phys. Lett. B 762, 498 (2016).
[22]
S. Borsanyi et al.,
Nature 539, no. 7627, 69 (2016).
[23]
D. Klabučar and D. Kekez,
Phys. Rev. D 58, 096003 (1998).
[24]
D. Kekez and D. Klabučar,
Phys. Rev. D 73, 036002 (2006).
[25]
S. Benić, D. Horvatić, D. Kekez and D. Klabučar,
Phys. Lett. B 738, 113 (2014).
[26]
E. Witten,
Nucl. Phys. B 156, 269 (1979).
[27]
G. Veneziano,
Nucl. Phys. B 159, 213 (1979).
[28]
G. M. Shore,
Nucl. Phys. B 744, 34 (2006).
[29]
Z. Y. Lu and M. Ruggieri,
arXiv:1811.05102 [hep-ph].
[30]
D. Horvatić, D. Kekez and D. Klabučar,
Universe 5, no. 10, 208 (2019)
doi:10.3390/universe5100208
[arXiv:1909.09879 [hep-ph]]. |
KEK Preprint 2018-80
CHIBA-EP-234
Confinement of quarks in higher representations in view of dual superconductivity
Computing Research Center, High Energy Accelerator Research Organization (KEK)
SOKENDAI (The Graduate University for Advanced Studies), Tsukuba 305-0801, Japan
E-mail:
Ryutaro Matsudo
Department of Physics, Faculty of Science and Engineering,
Chiba University, Chiba 263-8522, Japan
E-mail:
[email protected]
Seikou Kato
Oyama National College of Technology, Oyama, Tochigi 323-0806, Japan
E-mail:
[email protected]
Kei-Ichi Kondo
Department of Physics, Graduate School of Science and Engineering, Chiba University,
Department of Physics, Graduate School of Science, Chiba University, Chiba 263-8522, Japan
E-mail:
[email protected]
Abstract:
Dual superconductor picture is one of the most promising scenarios for quark confinement.
We have proposed a new formulation of Yang-Mills theory on the lattice so that the so-called restricted field obtained from the gauge-covariant decomposition plays the dominant role in quark confinement.
This framework improves the Abelian projection in the gauge-independent manner.
For quarks in the fundamental representation, we have demonstrated some numerical evidences for the dual superconductivity.
However, it is known that the expected behavior of the Wilson loop in higher representations cannot be reproduced if the restricted part of the Wilson loop is extracted by adopting the Abelian projection or the field decomposition naively in the same way as in the fundamental representation.
In this talk, therefore, we focus on confinement of quarks in higher representations.
By virtue of the non-Abelian Stokes theorem for the Wilson loop operator, we propose suitable operators constructed from the restricted field only in the fundamental representation to reproduce the correct behavior of the original Wilson loop in higher representations.
Moreover, we perform lattice simulations to measure the static potential for quarks in higher representations using the proposed operators.
We find that the proposed operators well reproduce the expected behavior of the original Wilson loop average, which overcomes the problem that occurs in naively applying Abelian-projection to the Wilson loop operator for higher representations.
1 Introduction
The dual superconductor picture is one of the most promising scenarios for
quark confinement [1]. According to this picture, magnetic
monopoles causing the dual superconductivity are regarded as the dominant
degrees of freedom responsible for confinement. However, it is not so easy to
establish this hypothesis. Indeed, even the definition of magnetic monopoles
in the pure Yang-Mills theory is not obvious. If magnetic charges are naively
defined from electric ones by exchanging the role of the magnetic field and
electric one according to the electric-magnetic duality, one needs to
introduce singularities to obtain non-vanishing magnetic charges, as
represented by the Dirac monopole. For such configuration, however, the energy
becomes divergent.
There are two prescriptions avoiding this issue in defining magnetic
monopoles, i.e., the Abelian projection method and the
decomposition method, In the Abelian projection method [2],
the “diagonal component” of the Yang-Mills
gauge field is identified with the Abelian gauge field, and a magnetic
monopole is defined as the Dirac monopole. The energy density of this monopole
can be finite everywhere because the contribution from the singularity of a
Dirac monopole can be canceled by that of the off-diagonal components of the
gauge field. However, one needs to fix the gauge, because otherwise the
“diagonal component” is meaningless. In the
decomposition method, on the other hand, we need no more gauge fixing. The
gauge-covarint decomposition was first proposed by Cho, Duan and Ge [3]
for the $SU(2)$ Yang-Mills theory, and was extended to the $SU(N)$ Yang-Mills
theory (see [4] for review). The key ingredient in this
decomposition is the Lie-algebra valued field with a unit length which we call
the color field. Then, the monopoles can be defined by using the
gauge-invariant part proportional to the color field in the field strength
just like the Abelian field strength in the Abelian projection. Therefore, the
gauge invariance is explicitly maintained in the decomposition method.
It should be examined on the lattice whether or not these monopoles can
reproduce the expected infrared behavior of the original Wilson loop average.
In the preceding lattice studies for $SU(2)$ and $SU(3)$ Yang-Mills theory
within the Abelian projection method using the maximal Abelian (MA) gauge, it
was confirmed that (i) the string tension calculated from the diagonal part of
the original Yang-Mills field reproduces the full string tension calculated
from the original Yang-Mills fields [7, 9], and that (ii) the
string tension calculated from the magnetic monopole extracted from the
diagonal part mostly reproduce the full string tension [8, 9].
However, it is known that the resulting monopole contribution does not
reproduce the original Wilson loop average if the Abelian projection is
naively applied to the Wilson loop in higher representations [10].
This is because, in higher representations, the diagonal part of the Wilson
loop does not behave in the same way as the original Wilson loop. Poulis
heuristically found the correct way to extend the Abelian projection approach
for the adjoint representation in the $SU(2)$ Yang-Mills theory
[11].
In this talk, we propose a systematic prescription to extract the
“dominant” part of the Wilson loop average,
which can be applied to the Wilson loop operator in an arbitrary
representation of the $SU(N)$ Yang-Mills theory. To test this proposal, we
perform the numerical simulation on the lattice, and measure the Wilson loop
average in the higher representation for both the original Yang-Mills field
and the restricted field which is extracted as the dominant mode for quark
confinement by using the decomposition method, e.g., the Wilson loop averages
for the adjoint representation in the $SU(2)$ Yang-Mills theory, and for the
adjoint and sextet representations in the $SU(3)$Yang-Mills theory. We find
that the results support our claim.
2 The gauge field decomposition method and the non-Abelian Stokes
theorem
In this section, we give a brief review of the gauge field decomposition (
e.g., see [4]) and the non-Abelian Stokes theorem for the Wilson
loop in an arbitrary representation [5].
2.1 The gauge filed decomposition
We decompose the gauge link variable $U_{x,\mu}$ into the product of the two
variables $V_{x,\mu}$ and $X_{x,\mu}$ in such a way that the new variable
$V_{x,\mu}$, is transformed by the full $SU(N)$ gauge transformation
$\Omega_{x}$ as the gauge link variable $U_{x,\mu}$, while $X_{x,\mu}$
transforms as the site variable:
$$\displaystyle X_{x,\mu},U_{x,\mu}=X_{x,\mu}V_{x,\mu}\in G=SU(N),$$
(1a)
$$\displaystyle U_{x,\mu}\longrightarrow U_{x,\nu}^{\prime}=\Omega_{x}U_{x,\mu}%
\Omega_{x+\mu}^{{\dagger}},\text{ \ \ }V_{x,\mu}\longrightarrow V_{x,\nu}^{%
\prime}=\Omega_{x}V_{x,\mu}\Omega_{x+\mu}^{{\dagger}},\text{ \ }X_{x,\mu}%
\longrightarrow X_{x,\nu}^{\prime}=\Omega_{x}X_{x,\mu}\Omega_{x}^{{\dagger}}.$$
(1b)
From the physical point of view, $$V_{x,\mu}$$, which we call the restriced
field, could be the dominant mode for quark confinement, while $$X_{x,\mu}$$ is
the remainder part. The possible options of the decomposition are
discriminated by the stability subgroup of the gauge group. Here, we only
consider the maximal option.
The maximal option is obtained for the stability subgroup of the maximal torus
subgroup of $G$: $\tilde{H}=U(1)^{N-1}\subset SU(N).$ The resulting
decomposition is the gauge-invariant extension of the Abelian projection in
the maximal Abelian (MA) gauge. We introduce color fields as,
$$\bm{n}{}^{(k)}(x)=\Theta(x)H_{k}\Theta^{{\dagger}}(x)\in Lie[G/\tilde{H}]\text%
{
\ \ \ \ \ \ }(k=1,\ldots,N-1)\text{,}$$
(2)
which are expressed using a common $SU(N)$-valued field $\Theta(x)$ with the
Cartan generators $H_{k}$. The decomposition is obtained by solving the
defining equations:
$$\displaystyle D_{\mu}^{\epsilon}[V]\mathbf{n}_{x}^{(k)}$$
$$\displaystyle:=\frac{1}{\epsilon}\left[V_{x,\mu}\mathbf{n}_{x+\mu}^{(k)}-%
\mathbf{n}_{x}^{(k)}V_{x,\mu}\right]=0\text{ },$$
(3a)
$$\displaystyle\text{\ }g_{x}$$
$$\displaystyle:=e^{i2\pi q/N}\exp\left(-i\sum_{j=1}^{N-1}a_{x}^{(j)}\mathbf{n}_%
{x}^{(j)}\right)\text{,}$$
(3b)
where, the variable $$g_{x}$$ is the $$U(1)^{N}$$ part which is undetermined from
Eq.(3a) alone, $$a_{x}^{(j)}$$ are coeficients, and $$q$$ is an
integer. Note that the above defining equations correspond to the continuum
version ($$g_{x}=\mathbf{1}$$): $$D_{\mu}[\mathscr{V}]\mathbf{n}^{(k)}(x)=0$$ and $$\mathrm{tr}(\mathscr{X}_{\mu}(x)\mathbf{n}^{(k)}(x))$$ $$=0$$, respectively. These defining equations
can be solved exactly, and the solution is given by
$$\displaystyle X_{x,\mu}$$
$$\displaystyle=\widehat{K}_{x,\mu}^{{\dagger}}\det(\widehat{K}_{x,\mu})^{1/N}g_%
{x}^{-1},\text{ \ \ \ \ \ \ \ \ \ }V_{x,\mu}=X_{x,\mu}^{{\dagger}}U_{x,\mu},$$
$$\displaystyle\widehat{K}_{x,\mu}$$
$$\displaystyle:=\left(K_{x,\mu}K_{x,\mu}^{{\dagger}}\right)^{-1/2}K_{x,\mu},%
\text{ \ \ \ \ \ \ \ \ \ \ \ }K_{x,\mu}:=\bm{1}+2N\sum{}_{k=1}^{N-1}\bm{n}_{x}%
^{(k)}U_{x,\mu}\mathbf{n}_{x+\mu}^{(k)}U_{x,\mu}^{{\dagger}}.$$
(4)
In the naive continuum limit, we can reproduce the decomposition in the
continuum theory.
The color fields $\bm{n}{}^{(k)}$ are obtained by minimizing the functional $R_{\mathrm{MA}}$ for a
given configuration of the link variables $U_{x,\mu}$ with respect to the
gauge transformation
$$R_{\mathrm{MA}}[U,\{\bm{n}^{(k)}\}]=\sum_{x,\mu}\sum_{k=1}^{N-1}\operatorname{%
tr}[(D_{\mu}[U]\mathbf{n}_{x}^{(k)})^{{\dagger}}(D_{\mu}[U]\mathbf{n}_{x}^{(k)%
})].$$
(5)
If we choose this condition as the reduction condition, the definition of
monopoles is equivalent to that for the Abelian projection in the MA gauge. In
the present study for the $SU(3)$ gauge theory, we apply two additional
reduction conditions which are defined by minimizing following functionals
$$\displaystyle R_{n3}[U,\{\bm{n}^{(k)}\}]$$
$$\displaystyle=\sum_{x,\mu}\operatorname{tr}[(D_{\mu}[U]\bm{n}_{x}^{3})^{{%
\dagger}}(D_{\mu}[U]\bm{n}_{x}^{3})],\quad(\bm{n}_{x}^{3}:=\Theta_{x}T^{3}%
\Theta_{x}^{{\dagger}}),$$
(6)
$$\displaystyle R_{n8}[U,\{\bm{n}^{(k)}\}]$$
$$\displaystyle=\sum_{x,\mu}\operatorname{tr}[(D_{\mu}[U]\bm{n}_{x}^{8})^{{%
\dagger}}D_{\mu}[U]\bm{n}_{x}^{8})],\quad(\bm{n}_{x}^{8}:=\Theta_{x}T^{8}%
\Theta_{x}^{{\dagger}}).$$
(7)
2.2 Non-Abelian Stokes theorem
We can relate the decomposed field variables to a Wilson loop operator through
a version of the non-Abelian Stokes theorem (NAST) which was proposed by
Diakonov and Petrov [6]. In this version of the NAST, a Wilson loop
operator in a representation $R$ is rewritten into the surface integral form
by introducing a functional integral on the surface $S$ surrounded by the loop
$C$ as
$$\displaystyle W_{R}[\mathscr{A};C]$$
$$\displaystyle:=\int D\Omega\exp\left(ig\int_{S:\partial S=C}dS_{\mu\nu}\sum_{k%
=1}^{N-1}\Lambda_{k}F_{\mu\nu}^{(k)}\right)\text{
},$$
(8a)
$$\displaystyle F_{\mu\nu}^{(k)}$$
$$\displaystyle:=2\operatorname{tr}(\bm{n}^{(k)}\mathscr{F}_{\mu\nu}[\mathscr{V}%
]),$$
(8b)
where $$D\Omega$$ is the product of the Haar measure over the surface $$S$$,
$$\Lambda_{k}$$ is the $$k$$-th component of the highest-weight of the
representation $$R$$, the color fields are defined by $$\bm{n}^{(k)}=\Omega H_{k}\Omega^{{\dagger}}$$, and $$\mathscr{F}_{\mu\nu}[\mathscr{V}]$$ is the field
strength for the restricted field $$\mathscr{V}$$ in the continuous version.
Monopoles are defined in the same manner as the Dirac monopoles for the
Abelian-like gauge-invariant field strength $$F_{\mu\nu}^{(k)}$$. The resulting
monopoles are gauge invariant by construction. Thus we can relate the
restricted field to tied Wilson loop operator in the manifestly
gauge-invariant way.
In the actual lattice simulations, we do not follow this NAST directly.
Without performing the integration over the measure $D\Omega$, the argument of
the exponential is approximated by substituting the color field (as a
functional of the gauge field) which is obtained by solving the reduction
condition. Indeed, this approximation is commonly used for the Wilson loop in
the fundamental representation, e.g., the Abelian projection in the MA gauge
is equivalent to the field decomposition with the color fields determined by
minimizing the functional eq. 5.
The NAST can be applied not only to the fundamental representation but also to
any representation, which suggests the correct way to extract the dominant
part of the Wilson loop in higher representations as we explain in the next section.
3 Wilson loops in higher representations
In the preceding study, it was shown that the area law of the average of the
Wilson loop in the fundamental representation is reproduced by the monopole
contribution. However there is a possibility that the monopole contribution
accidentally coincides with the behavior of the original Wilson loop, and
therefore we should examine the other quantities. The Wilson loops in higher
representations are appropriate for this purpose because they have clear
physical meaning and show the characteristic behavior, e.g., the Casimir scaling.
It is known that if we adopt the Abelian projection naively to higher
representations, the monopole contributions do not reproduce the correct
behavior [10]. Thus, we have to find a more appropriate way to
extract the monopole contributions. The NAST (8a) suggests the
different operator as the dominant part of the Wilson loop in higher
representations. In the actual calculation, we approximate the NAST by
calculating the integrand of the NAST using the color fields satisfying the
reduction condition. In this approximation, the operator to be calculated is
not equivalent to the Wilson loop for the restricted field. From this point of
view, therefore, the distinct operator is suggested as the dominant part of
the Wilson loop. We can see the difference between the operator suggested by
the NAST and the Wilson loop for the restricted field by expressing it as the
surface integral form as
$$W_{R}[\mathscr{V};C]=\frac{1}{D_{R}}\sum_{\mathbf{\mu\in}\Delta_{R}}d_{\mathbf%
{\mu}}\exp\left(ig\sum_{k=1}^{N-1}\mu_{k}\int_{S}dS^{\alpha\beta}F_{\alpha%
\beta}^{(k)}\right),$$
(9)
where $D_{R}$ is the dimension of the representarion $R,$ $\Delta_{R}$ is set
of the weights of $R$, $\ \mu_{k}$ is the $k$-th component of the weight
$\mathbf{\mu,}$ and $d_{\mu}$ is the multiplicity of$\ \mathbf{\mu},$ which
satisfies $D_{R}=\sum_{\mathbf{\mu\in}\Delta_{R}}d_{\mathbf{\mu}}.$ In the
fundamental representation, all weights are equivalent to the highest weight
under the action of the Weyl group, but in higher representations there are
weights that are not equivalent to the highest weight. Therefore this
expression is not equivalent to the integrand of eq. 8a in higher
representations. For this reason we should modify eq. 9 so as to
include only the weights that is equivalent to the highest weights.
By using the untraced Wilson loop $V_{C}:=\prod_{\braket{x,\mu}\in C}V_{x,\mu}$ for the restricted field in the fundamental representation, we obtain the
Wilson loop in higher representations. For $SU(2)$, we propose the operator
for the spin-$j$ representation as
$$W_{[j]}^{SU(2)}[\mathscr{V};C]=\frac{1}{2}\operatorname{tr}(V_{C}^{2j}).$$
(10)
For $SU(3)$, we propose the operator for the representation with the Dynkin
index $[m_{1},m_{2}]$ as
$$W_{[m_{1},m_{2}]}^{SU(3)}[\mathscr{V};C]=\begin{cases}\frac{1}{6}\left(%
\operatorname{tr}(V_{C}^{m_{1}})\operatorname{tr}(V_{C}^{{\dagger}m_{2}})-%
\operatorname{tr}(V_{C}^{m_{1}}V_{C}^{{\dagger}m_{2}})\right)&\text{for}\quad m%
_{1},\text{ }m_{2}>0\\
\frac{1}{3}\operatorname{tr}(V_{C}^{m_{1}})&\text{for}\quad m_{2}=0\\
\frac{1}{3}\operatorname{tr}(V_{C}^{{\dagger}m_{2}})&\text{for}\quad m_{1}=0%
\end{cases}.$$
(11)
4 Numerical result
In order to support our claim that the dominant part of the Wilson loops in
higher representation is the operator suggested by the NAST, by using the
numerical simulation on the lattice, we check whether the string tension from
the Wilson loop in the representation $R$ of the restricted field $V$ can
reproduce that of the Yang-Mills field or not.
For this purpose, We set up the gauge configuration for the standard Wilson
action at $\beta=2.5$ on the $24^{4}$ lattice for $SU(2)$ case and at
$\beta=6.2$ on the $24^{4}$ lattice for $SU(3)$ case . For $SU(2)$ case, we
prepare $500$ configurations every $100$ sweeps after $3000$ thermalization by
using the heat bath method. For $SU(3)$ case, we prepare $1500$ configurations
every $50$ sweeps after $1000$ thermalization by using pseudo heat bath method
with over-relaxation algorithm ($20$ steps per sweep). In the measurement of
the Wilson loop average we apply the APE smearing technique for $SU(3)$ case
and the hyper-blocking for $SU(2)$ case to reduce noises and reduced the
exciting modes. The number of the smearing steps is determined so that the
ground state overlap is enhanced [12]. We have calculated the Wilson
loop average $W(R,T)$ for the rectangular loop with length $T$ and width $R$
to derive the potential $V(R,T)$ through the formula $V(R,T):=-\log(W(R,T+1)/W(R,T).$
In case of the $SU(2)$, we investigate the Wilson loop in the
adjoint representation $\mathbf{3}$ ($j=1$). The restricted field, the
extracted dominant mode for confinement, is obtained by using the
decomposition eq(4) for the color field which minimizes the
reduction condition Eq(5)($N=2).$ Figure 1 shows that
the string tension from the restricted and Yang-Mills filed in the adjoint
represent are in good agreement.
In case of $SU(3)$ we investigate the Wilson loop eq(11), i.e., (a) in
the fundamental representation $[0,1]=\mathbf{3}$, (b) in the adjoint
representation $[1,1]=\mathbf{8}$, and (c) in the sextet representation
$[0,2]=\mathbf{6}$. For each representation, we measure the Wilson loop
average for possible reduction conditions, eq(5) ($N=3)$,
eq(6), and eq(7). Figure 2 shows the static
potentials from the Wilson loop in the higher representations. Table
1 shows the string tensions which are extracted by fitting
the data with the linear potential for the infrared region. The string
tensions extracted from our proposed ”dominant” operators reproduce nearly
equal to or more than $80\%$ of the full string tension. These results
indicate that the proposed operators give actually the dominant part of the
Wilson loop average.
5 Conclusion
We have proposed a solution for the problem that the correct behavior of the
Wilson loop in higher representations cannot be reproduced if the restricted
part of the Wilson loop is naively extracted by adapting the Abelian
projection or the field decomposition in the same way as in the fundamental
representation. We have proposed the prescription to construct the operator
suitable for this purpose. We have performed numerical simulations to show
that this prescription works well in the adjoint rep. for $SU(2)$ color group,
and in the fundamental, adjoint , and sextet representations for $SU(3)$
color group. Further studies are needed in order to establish the magnetic
monopole dominance in the Wilson loop average for higher representations,
supplementary to the fundamental representation for which the magnetic
monopole dominance was established.
Acknowledgement
This work was supported by Grant-in-Aid for Scientific Research, JSPS KAKENHI
Grant Number (C) No.15K05042. R. M. was supported by Grant-in-Aid for JSPS
Research Fellow Grant Number 17J04780. The numerical calculations were in part
supported by the Large Scale Simulation Program No.16/17-20(2016-2017) of High
Energy Accelerator Research Organization (KEK), and were performed in part
using COMA(PACS-IX) at the CCS, University of Tsukuba.
References
[1]
Y. Nambu,
Phys. Rev. D10 (1974) 4262
; G. ’t Hooft,
in High Energy Physics,
Editorice Compositori, Bologna, 1975;
S. Mandelstam,
Phys. Rept. 23 (1976) 245
.
[2]
G. ’t Hooft,
Nucl. Phys. B 190 (1981) 455 .
[3]
Y. S. Duan, M. L. Ge,
Sinica Sci. 11 (1979) 1072; Y. M. Cho,
Phys. Rev. D 21 (1980) 1080.
[4]
K.-I. Kondo, S. Kato, A. Shibata and T. Shinohara,
Phys. Rept. 579 (2015) 1
[arXiv:1409.1599] .
[5]
R. Matsudo, K.-I. Kondo, Phys.Rev. D 92 (2015) no.12, 125038
[6]
D. Diakonov and V. Petrov,
Phys. Lett. B 224 (1989) 131
.
[7]
T. Suzuki and I. Yotsuyanagi,
Phys. Rev. D 42 (1990) 4257.
[8]
H. Shiba and T. Suzuki,
Phys. Lett. B 333 (1994) 461.
[9]
J. D. Stack, W. W. Tucker and R. J. Wensley,
Nucl. Phys. B 639 (2002) 203.
[10]
L. Del Debbio, M. Faber, J. Greensite and S. Olejnik,
Phys. Rev. D 53 (1996) 5891.
[11]
G. Poulis,
Phys. Rev. D 54 (1996) 6974.
[12]
G. Bali, C. Schlichter and K. Schilling, Phys. Rev. D
51 (1995) 5165.
[13]
M. N. Chernodub, K. Hashimoto and T. Suzuki, Phys. Rev. D
70 (2004) 014506. |
Top Higgs Yukawa Coupling Analysis from
$e^{+}e^{-}\rightarrow\bar{t}tH\rightarrow\bar{b}W^{-}bW^{+}\bar{b}b$
Hajrah Tabassam${}^{1}$ Victoria Martin${}^{2}$
1- Department of Physics
Quaid-i-Azam University Islamabad - Pakistan
2-School of Physics and Astronomy
University of Edinburgh - United Kigdom
Abstract
Electroweak symmetry breaking and the origin of mass of fermion and boson are fundamental questions to understand particle physics. The essential piece of this symmetry breaking, the Higgs boson, is the most probable particle to be discovered at the Large Hadron Collider (LHC).
If one, or more, Higgs particles are discovered,
precise measurements of all the properties of this Higgs will be very important including the measurement of Yukawa couplings of the fermions to the Higgs.
We present a study of $e^{+}e^{-}\rightarrow t\bar{t}H$ at a linear collider, with the aim of making a direct measurement of the the top-Higgs coupling, $g_{t\bar{t}H}$ using the semi-leptonic final state and $M_{H}$ of 120 GeV. We show that the top-Higgs coupling at the center of mass energy 500 GeV, can be measured with an accuracy of better than 28%.
1 Introduction
The ILC is a unique tool which could play an extremely important role in high-precision measurements. Data from the ILC will allow a determination of the profile of the Higgs bosons and their fundamental properties with a high level of confidence and will provide a unique opportunity to establish experimentally the mechanism that generates the particle masses.
The Higgs couplings to the fermions and the gauge bosons are related to the masses of these particles and the only free parameter of the model is the mass of the Higgs boson itself; there are, however, both experimental and theoretical constraints on the Higgs mass in the Standard Model.
Experimentally, the available direct information on the Higgs mass ($M_{H}$) is the lower limit $M_{H}\gtrsim 114.4$ GeV established at LEP [4].
The high accuracy of the electroweak data measured at LEP, SLC and Tevatron provides an indirect sensitivity to $M_{H}$: the Higgs boson contributes logarithmically, $\propto\log(M_{H}/M_{W})$, to the radiative corrections to the $W/Z$ boson propagators. A recent analysis, which uses the updated value of the top quark mass yields the value 85$\pm$25 GeV, with upper limit of $M_{H}\lesssim 165$ GeV [5].
Direct searches by CDF and D0 exclude a region at high mass between $156<M_{H}<177$ GeV [6].
ATLAS and CMS have also made direct searches for the Higgs boson. At the time of this workshop preliminary LHC results exclude the ranges 155-190 GeV [7] at ATLAS and 149-206 GeV [8] at CMS at 95% CL.
2 Higgs Production and Decay
In the Standard Model (SM), the profile of the Higgs particle is uniquely determined once its mass $M_{H}$ is fixed [1, 2]. The decay width, the branching ratios and the production cross sections are given by the strength of the Yukawa couplings to fermions and gauge bosons, the scale of which is set by the masses of these particles.
$$g_{fHf}\propto m_{f}$$
For a top quark mass of $M_{t}=175\;\mathrm{GeV}$, the Standard Model the Higgs Yukawa coupling is ${g_{t\bar{t}H}}=M_{t}/v=0.71$, where $v=246\;\mathrm{GeV}$ is the vacuum expectation value.
The aim of the work presented here is to determine how accurately ${g_{t\bar{t}H}}$ can be determined at the ILC through a measurement of $t\bar{t}H$ production.
There are two Higgs main production channels for $e^{+}e^{-}$ production at $\sqrt{s}=500\;\mathrm{GeV}$: the Higgs-strahlung process including $e^{+}e^{-}\rightarrow ZH$ and associated production with top pair $e^{+}e^{-}\rightarrow t\bar{t}H$ (figure 1) which we use for our analysis. The diagram where the Higgs boson is radiated from the $Z$ boson modifies only slightly the cross-section and thus are neglected in this study.
The signal has a cross section of $0.576\;\mathrm{fb}$ at $\sqrt{s}=500$ GeV due to phase space suppression but, at $\sqrt{s}=800$ GeV, it can reach the level of a few femtobarns.
3 Overview
References [11] and [12] have performed earlier feasibility studies of the top-Higgs Yukawa coupling via the process $e^{+}e^{-}\rightarrow t\bar{t}H$ at a linear collider and in [13], an analysis was performed at 800 GeV center-of-mass energy.
For the ILD detector, we perform the first complete analysis of the process for $M_{H}=120\;\mathrm{GeV}$ in the semileptonic channel where one $W$ boson from the top quark decay, $t\rightarrow bW$, decays as $W\rightarrow\ell\nu$ and the other as $W\rightarrow q\bar{q}$. A full reconstruction of the final state is conducted by reconstructing one hadronic and one leptonic $W$ decays. The top quarks and the Higgs boson are then reconstructed from $W^{\prime}$s and $b$-jets. Missing energy reconstruction has a large impact on the reconstruction of the semi-leptonic channel, and $b$-tagging has a crucial role in separating signal and background. The analysis is carried out for an integrated luminosity, $\mathcal{L}=1000\;\mathrm{fb}^{-1}$ and at $\sqrt{s}=500$ GeV. The main background processes are $t\bar{t}$ and $t\bar{t}Z$ production.
4 Monte Carlo Samples
The analysis was performed with the samples provided by the ILD optimisation group [9] with $M_{H}=120\;\mathrm{GeV}$ and $M_{t}=175\;\mathrm{GeV}$.
The samples were generated at SLAC using WHIZARD, and then simulated and reconstructed on the DESY Grid nodes. The GEANT4 based Mokka package was used for simulation in the ILD-00 detector model, which is the first simulated reference model of ILD. Reconstruction was performed using MarlinReco, PandoraPFA and LCFIVertex with the versions supplied in ILCInstall v01-07.
In table 1, the sample sizes used in the analysis are shown.
5 Event Reconstruction
We reconstruct the semi-leptonic final state of the signal process with one charged lepton (electron or muon) and a neutrino and six jets, accounting for
$\sim 20\%$ of the total cross section.
The final state in this channel has four $b$-jets, two light jets, $j$, one charged lepton, $\ell$, and missing energy for $\nu$:
$$e^{+}e^{-}\rightarrow t\bar{t}H\rightarrow W^{+}bW^{-}\bar{b}b\bar{b}%
\rightarrow\ell\nu\,2j\,4b$$
We select electrons and muons using selection criteria obtained from a single particle sample study.
Left figure 2 shows the reconstructed invariant mass of the charged lepton and the missing energy in the event.
We remove the electrons and muons from our sample and force the remaining particles into exactly six jets using the JETFinder algorithm. The LCFIVertex algorithm is run on the jets to obtain $b$-tagging information [10].
Jets passing the LCFIVertex reconstruction are sorted according to their $b$-tag value. The $b$-tag uses a neural net algorithm, producing a value between 0 and 1 for each jet: 1 being the most $b$-jet like, and 0 being the least $b$-jet like. To suppress backgrounds, and improve the reconstruction efficiency, we select events where at least four of the six jets have $b$-tag $>0.09$.
The four jets with the highest $b$-tag value are considered as the $b$-jets and the two with the lowest $b$-tag value are considered as the light jets. Hadronic $W$ candidates, shown in right figure 2, are created from pairs of light-jets in the event.
The remaining jets are considered as the $b$-jets and used to reconstruct $M_{bb},M_{\ell\nu b}$ and $M_{jjb}$.
Top-quark candidates are reconstructed using a tagged $b$-jet and a reconstructed $W$ candidate and the Higgs boson is reconstructed from a pair of tagged $b$-jets. All possible combinations of tagged $b$-jets are examined to reconstruct the Higgs boson and both top-quarks simultaneously.
To reduce combinatorial background, we define $\chi^{2}$ for each event:
$$\chi^{2}=\frac{(M_{\ell\nu b}-M_{t})^{2}}{\sigma_{\ell\nu b}^{2}}+\frac{(M_{%
jjb}-M_{t})^{2}}{\sigma_{jjb}^{2}}+\frac{(M_{bb}-M_{H})^{2}}{\sigma_{bb}^{2}}$$
(1)
where take the widths, $\sigma$, for each combinations from the reconstructed resolutions:
$$\sigma_{\ell\nu b}=20.7\pm 0.2\;\mathrm{GeV}\hskip 14.226378pt\sigma_{jjb}=23.%
5\pm 0.3\;\mathrm{GeV}\hskip 14.226378pt\sigma_{bb}=19.3\pm 0.2\;\mathrm{GeV}$$
The mass distributions of the lowest $\chi^{2}$ combination are given in figure 3.
6 Signal and Background Separation
The main backgrounds are from $t\bar{t}$ and $t\bar{t}Z$. We used the following variables to discriminate between signal and background by maximising ${S}/{\sqrt{S+B}}$ (where $S$ is signal and $B$ is background).
•
$E_{\mathrm{SumReco}}$: sum of the energies of all reconstructed particles in the event.
•
$P_{{\mathrm{Lepton}}}$: the reconstructed momentum of the identified lepton.
•
$P_{{\mathrm{miss}}}$: the reconstructed missing momentum.
•
$P_{{\mathrm{Jets}}}$: the momentum of each of the reconstructed jets.
•
$\chi^{2}$, as defined in equation (1).
•
$m_{\ell\nu}$: leptonic $W$ candidate mass
•
$m_{b\bar{b}}$: Higgs candidate mass.
•
Total reconstructed mass $m_{b\bar{b}}+m_{\ell\nu b}+m_{jjb}$.
Figure 4 shows some of these variables for signal and background.
The optimised selection on the above variables selects $(0.11\pm 0.01)\%$ of $t\bar{t}$, $(2.76\pm 0.12)\%$ of $t\bar{t}Z$, and $(7.57\pm 0.19)\%$ of signal events.
7 Measurement of $\mathbb{g_{t\bar{t}H}}$
In the Standard Model the Higgs Yukawa coupling is written in terms of the top quark mass ${g_{t\bar{t}H}}=M_{t}/v=0.71$.
To measure the top Higgs Yukawa coupling, we followed procedure given in [13].
The statistical and systematic uncertainties on the measurement of $g_{t\bar{t}H}$ can be expressed as:
$$\left(\frac{\Delta g_{t\bar{t}H}}{g_{t\bar{t}H}}\right)_{\mathrm{stat}}\approx%
\frac{1}{S_{\mathrm{stat}}(g^{2}_{t\bar{t}H})\sqrt{\epsilon\rho\mathcal{L}}}$$
$$\left(\frac{\Delta g_{t\bar{t}H}}{g_{t\bar{t}H}}\right)_{\mathrm{syst}}\approx%
\frac{1}{S_{\mathrm{syst}}(g^{2}_{t\bar{t}H})}\frac{1-\rho}{\rho}\frac{\Delta%
\sigma_{\mathrm{eff}}^{\mathrm{BG}}}{\sigma_{\mathrm{eff}}^{\mathrm{BG}}}\\
$$
Where $\epsilon=(7.57\pm 0.19)\%$ is the efficiency to select the data and $\rho=(12.5\pm 0.3)\%$ is the purity of the final selected sample. We use an integrated luminosity of $\mathcal{L}=1000\;\mathrm{fb}^{-1}$, equivalent three to four year’s running of the ILC.
The largest unknown comes from the knowledge of the background cross sections,
$\Delta\sigma_{\mathrm{eff}}^{\mathrm{BG}}/{\sigma_{\mathrm{eff}}^{\mathrm{BG}}}$, dominated by current knowledge of $t\bar{t}$ production. Following reference [13], we choose a relative uncertainty of 5%;
an ongoing study [16] will allow a better estimate of this uncertainty.
Ignoring the small contribution from Higgs radiation off the $Z$ (as shown in figure 1)
we can write:
$$\frac{d\sigma_{t\bar{t}H}}{d(g^{2}_{t\bar{t}H})}\approx\frac{\sigma_{t\bar{t}H%
}}{g^{2}_{t\bar{t}H}}=0.087\;\mathrm{fb}$$
with $\sigma_{t\bar{t}H}=0.044\;\mathrm{fb}$ being the effective cross section $7.57\%$ of the total cross section and ${g_{t\bar{t}H}}=0.71$.
Then the sensitivity factors $S_{\mathrm{stat}}$ and $S_{\mathrm{syst}}$ simply depend on cross section and on the squared coupling:
$$S_{\mathrm{stat}}(g^{2}_{t\bar{t}H})=\frac{1}{\sqrt{\sigma_{t\bar{t}H}}}\left|%
\frac{d\sigma_{t\bar{t}H}}{d(g^{2}_{t\bar{t}H})}\right|=\frac{\sqrt{\sigma_{t%
\bar{t}H}}}{g^{2}_{t\bar{t}H}}=1.50\;\mbox{fb}^{1/2}\hskip 28.452756pt$$
$$S_{\mathrm{syst}}(g^{2}_{t\bar{t}H})=\frac{1}{\sigma_{t\bar{t}H}}\left|\frac{d%
\sigma_{t\bar{t}H}}{d(g^{2}_{t\bar{t}H})}\right|=\frac{1}{g^{2}_{t\bar{t}H}}=1%
.98$$
The expected statistical and systematic uncertainties are therefore:
$$\left(\frac{\Delta g_{t\bar{t}H}}{g_{t\bar{t}H}}\right)_{\mathrm{stat}}=21.6\%%
\hskip 28.452756pt\left(\frac{\Delta g_{t\bar{t}H}}{g_{t\bar{t}H}}\right)_{%
\mathrm{sys}}=17.6\%$$
Combining the statistical and systematic uncertainties in quadrature we obtain an estimate on the overall uncertainty on ${g_{t\bar{t}H}}$ of 27.9%.
8 Results and Discussion
Figure 5 shows the distribution for the scaled signal and background samples. The main background after selection is due to the top-pair production.
The expected precision on measuring the top-Higgs Yukawa coupling is better than 28% for $M_{H}=120$ GeV,
if the knowledge of $t\bar{t}$ background normalisation is known at the 5% level.
Much of the uncertainty on the results of this analysis is due to the small cross section of the $t\bar{t}H$ signal process. Even though the signal cross section is very small at the $\sqrt{s}$ = 500 GeV, yet 28% accuracy is achieved.
Certainly there is room for improvement of the study presented.
In jet finding, we force particles into a jet which sometime don’t belong to that jet. It could affect the energy of the reconstructed jet. A scaling of the flavour tag input variables to the reconstructed jet energy can minimise these effects.
One of the areas of consideration is the improvement in the hadronic $W$ reconstruction.
A cut on the $b$-tag value of the jets is used to identify light jets which are then combined to reconstruct the $W$ boson. This cut was decided to accommodate the selection cut on $b$-tag of third and fourth jet. A re-examining of the cut on $b$-tag of fifth and sixth jet could recover the signal events with a $b$-tag $>0.09$ and hence, improve the efficiency and purity of the sample.
A cut based strategy is applied to discriminate the signal and background events. It is shown in previous studies that the use of neural network and likelihood methods can perform the signal background separation with better efficiency [12]. Hence, it is anticipated that a 2-3% improvement in the precision can be achieved by using neural network analysis.
For the current study, six fermion backgrounds are not included but reference [13] shows that the loss of precision on $g_{t\bar{t}H}$ measurement is negligible due to this background. An amendment in the analysis involves the $t\bar{t}$ background normalisation which is 5% in our analysis. If the exact background normalisation ${\Delta\sigma_{\mathrm{eff}}^{\mathrm{BG}}}/{\sigma_{\mathrm{eff}}^{\mathrm{BG%
}}}$ is used, our results might change. Once the results from the group working on measuring the cross section of top pair production [16] are ready, they can be included in the analysis.
References
[1]
J. R. Espinosa,
Surveys High Energ. Phys. 10 (1997) 279-311.
[hep-ph/9606316].
[2]
B. W. Lee, C. Quigg and H. B. Thacker,
Phys. Rev. D16, 1519, (1977).
[3]
A. Djouadi. The Anatomy of elecroweak symmetry breaking. I: The Higgs boson in the standard model.
Phys. Rept., 457:1-216, (2008)
[4]
R. Barate et al. [LEP Working Group for Higgs boson searches,
the ALEPH Collaboration, the DELPHI Collaboration, the L3 Collaboration and the OPAL Collaboration],
Phys. Lett. B 565 (2003) 61
[arXiv:hep-ex/0306033].
[5]
J. Erler,
Phys. Rev. D81, 051301 (2010).
[arXiv:1002.1320 [hep-ph]].
[6]
Combined CDF and D0 Upper Limits on Standard Model Higgs Boson Production with up to 8.6 fb${}^{1}$ of Data, FERMILAB-CONF-11-354-E [arXiv:1107.5518 [hep-ex]].
[7]
Combined Standard Model Higgs Boson Searches in pp Collisions at $\sqrt{s}=7$ TeV with the ATLAS Experiment at the LHC. 24 July 2011. ATLAS-CONF-2011-112.
[8]
Search for standard model Higgs boson in pp collisions at $\sqrt{s}=7$ TeV. 23 July 2011. CMS-PAS-HIG-11-011.
[9]
http://ilcsoft.desy.de/portal/data_samples
[10]
LCFI Collaboration, S. Hillert,
The LCFIVertex Package: vertex detector-based Reconstruction at the ILC. Nucl.Instrum.Meth.A610:573-589, (2009)
[11]
H.Baer, S. Dawson and L. Rein
Phys.Rev.D 61 (2000), 013002
[12]
A. Juste, G. Merino,
[hep-ph/9910301].
[13]
Arnaud Gay
Eur.Phys.J.C49:489-497,2007
[14]
K. Nakamura et al. (Particle Data Group) J. Phys. G37, 075021 (2010)
[15]
http://geant4.slac.stanford.edu/SLACTutorial09/ParticleGun.pdf
[16]
Philippe Doublet (LAL)
http://ilcagenda.linearcollider.org/conferenceDisplay.py?confId=4983 |
Uncertainty principle in a cavity at finite temperature
A. P. C. Malbouisson${}^{a}$
(a) Centro
Brasileiro de Pesquisas Físicas/MCTI, 22290-180, Rio de
Janeiro, RJ, Brazil
[email protected] (A.P.C. Malbouisson)
Abstract
We employ a dressed state approach to perform a study on the behavior of the uncertainty principle for a system in a heated cavity. We find, in a small cavity for a given temperature, an oscillatory behavior of the momentum–coordinate product, $(\Delta\,p)\,(\Delta\,q)$, which attains periodically finite absolute minimum (maximum) values, no matter
large is the elapsed time. This behavior is in a sharp contrast with
what happens in free space, in which case, the product $(\Delta\,p)\,(\Delta\,q)$ tends asymptotically, for each temperature, to a constant value, independent of time.
PACS number(s): 03.65.Ta
Introduction
An account on the subject of an interacting particle-environment system,
can be found in Refs. zurek ; paz ; haake ; caldeira ; livro , in which the environment
is considered as an infinite set of noninteractng oscilllators.
Here we consider a similar model, treated with a different approach.
Let us therefore begin with some words about the method we employ: From a general point of view,
apart from computer calculations in lattice field theory, the must currently used
method to treat the physics of interacting particles is perturbation theory, in which
the starting point are bare fields (particles) interacting by means of gauge fields.
Actually, as a matter of principle, the idea of
bare particles associated to bare matter fields and of a gauge particle mediating the
interaction among them, is in fact an artifact of perturbation
theory and, strictly speaking, is physically meaningless. A charged physical particle is always coupled to
the gauge field, it is always ”dressed” by a cloud of quanta of the gauge
field (photons, in the case of Electrodynamics). Exactly the same type of argument
applies mutatis mutandis to a particle-environment system,
in which case we may speak of a ”dressing”
of the particle by the thermal bath, the
particle being ”dressed” by a cloud of quanta of the environment.
This should be true in general for any system in which a material particle is coupled to a
field, no matter the specific nature of the field (environment) nor of the interaction involved.
We give a treatment to this kind of system using some dressed (or renormalized)
coordinates. In terms of these new coordinates dressed states are defined, which
allow to divide the coupled system into two parts, the dressed
particle and the dressed environment, which makes unnecessary to work
directly with the concepts of bare particle, bare environment and interaction
between them. A detailed exposition of our formalism and of its meaning, for both zero- and
finite temperature can be found in Refs. linhares ; termalizacao ; termica ; emaranho ; erico .
About the physical situation we deal with, on general grounds, very precise investigations have been done
on the fundamentals of quantum physics, in particular on the validity of the Heisenberg uncertainty relation. In clerk ,
it is reported that a great deal of
effort is being made to minimize external noising factors,
such as thermal fluctuations and electricity oscillations in experiments,
in order to verify the relation, in the spirit of zero-temperature quantum physics.
However, changes in the uncertainty principle induced by temperature is an idea already explored in the literature, in particular for open systems. In Abe , the authors study with a thermo-field-dynamics formalism, the relation between the sum of information-theoretic entropies in quantum mechanics with measurements of the position and momentum of a particle surrounded by a thermal environment. It is found that this quantity cannot be made arbitrarily small but has a universal lower bound dependent on the temperature. They also show that the Heisenberg uncertainty relation at finite temperature can be derived in this context. In Hu ; Hu1 it is obtained the uncertainty relation for a quantum open system consisting of a
Brownian particle interacting with an
ohmic bath of quantum oscillators at finite temperature.
These authors claim that this allows to get some insight into the physical mechanisms involved in the environment-induced decoherence process.
Also, modifications of the uncertainty principle have been proposed in, for instance,
a cosmological context. As remarked in amelino , in quantum gravity it seems to be
needed a generalized position–momentum uncertainty principle. The authors of
ref. amelino investigate a possible connection between the generalized uncertainty principle and
changes in the area–entropy black hole formula and the black hole evaporation process.
In this report, we study the behavior of the particle-environment system contained
in a cavity of arbitrary size, under the influence of a heated environment. The environment is
composed of an infinity of oscillators, and assumed to be at a given temperature, realized by taking
an appropriate thermal distribution for its modes.
This generalizes previous works for zero temperature, as for instance in Refs. adolfo3 ; adolfo4 ; adolfo5 ,
for both inhibition of spontaneous decay in cavities and the Brownian motion.
We study the time dependent mean value for the dressed
oscillator position operator taken in a dressed coherent state. We find that in both cases,
of the environment at zero temperature or of the heated environment,
these mean values are independent of the temperature and are given by the same expression.
On the other side, the mean squared error for both the particle position and momentum
are dependent on the temperature. From them we get the time and
temperature-dependent uncertainty Heisenberg
relation. We then investigate how heating
affects the uncertainty principle in a cavity of arbitrary size. This is particularly
interesting in a small cavity, where the result is not a trivially expected one.
The model
Our approach to the problem
makes use of the notion of dressed thermal states termica ,
in the context of a model already employed in the literature, of atoms, or
more generally material particles, in the harmonic approximation, coupled to
an environment modeled by an infinite set of point-like harmonic oscillators
(the field modes). The dressed thermal state approach is an extension of the
dressed (zero-temperature) formalism already used earlier adolfo1 ; adolfo3 ; adolfo4 ; adolfo5 .
We consider a bare particle (atom, molecule,…) approximated by a harmonic
oscillator described by the bare coordinate and momentum $q_{0},p_{0}$
respectively, having bare frequency $\omega_{0}$, linearly
coupled to a set of $N$ other harmonic oscillators (the environment)
described by bare coordinate and momenta $q_{k},p_{k}$ respectively,
with frequencies $\omega_{k}$, $k=1,2,\ldots,N$. The limit
$N\rightarrow\infty$ will be undesrstood. The whole system is
supposed to reside inside a perfectly reflecting spherical cavity of radius $R$ in
thermal equilibrium with the environment, at a temperature
$T=\beta^{-1}$. The system is described by the Hamiltonian
$$H=\frac{1}{2}\left[p_{0}^{2}+\omega_{0}^{2}q_{0}^{2}+\sum_{k=1}^{N}\left(p_{k}%
^{2}+\omega_{k}^{2}q_{k}^{2}\right)\right]-q_{0}\sum_{k=1}^{N}c_{k}q_{k},$$
(1)
The Hamiltonian (1) is transformed to principal
axis by means of a point transformation,
$q_{\mu}=\sum_{r=0}^{N}t_{\mu}^{r}Q_{r}\,,\,\,\,p_{\mu}=\sum_{r=0}^{N}t_{\mu}^{%
r}P_{r};\;\mu=(0,\{k\}),\;k=1,2,...,N;\;\;r=0,...N$,
performed by an orthonormal matrix $T=(t_{\mu}^{r})$. The subscript
$r$ refers to the normal
modes. In terms of normal momenta and coordinates, the transformed
Hamiltonian reads
$H=\frac{1}{2}\sum_{r=0}^{N}(P_{r}^{2}+\Omega_{r}^{2}Q_{r}^{2})$,
where the $\Omega_{r}$’s are the normal frequencies corresponding
to the collective stable oscillation modes of the coupled
system. Using the coordinate transformation in the
equations of motion and explicitly making use of the
normalization of the matrix $(t_{\mu}^{r})$, $\sum_{\mu=0}^{N}(t_{\mu}^{r})^{2}=1$,
we get the matrix elements $(t_{\mu}^{r})$ adolfo1 .
We take $c_{k}=\eta(\omega_{k})^{u}$, where $\eta$ is a constant
independent of $k$. In this case the environment is classified
according to $u>1$, $u=1$, or $u<1$, respectively as
supraohmic, ohmic or subohmic paz ; haake ;
we take, as in adolfo1 , $\eta=2\sqrt{g\Delta\omega/\pi}$,
where $\Delta\omega$ is the interval between two
neighboring field frequencies and $g$ a fixed constant
characterizing the strenght of the coupling particle-environment.
Restricting ourselves to an ohmic environment, we get an equation for the
$N+1$ eigenfrequencies $\Omega_{r}$,
corresponding to the
$N+1$ normal collective modes adolfo1 . In this
case the eigenfrequencies
equation contains a divergence for $N\rightarrow\infty$ and a
renormalization procedure is needed. This leads to
the renormalized frequency adolfo1 (this renormalization procedure was pioneered in Thirring ),
$$\bar{\omega}^{2}=\omega_{0}^{2}-\delta\omega^{2}=\lim_{N\rightarrow\infty}(%
\omega_{0}^{2}-N\eta^{2}).$$
(2)
where we have defined the counterterm
$\delta\omega^{2}=N\eta^{2}$.
We introduce dressed or renormalized coordinates $q_{0}^{\prime}$ and
$\{q_{k}^{\prime}\}$ for, respectively, the dressed atom
and the dressed field, defined by,
$$\sqrt{\bar{\omega}_{\mu}}q_{\mu}^{\prime}=\sum_{r}t_{\mu}^{r}\sqrt{\Omega_{r}}%
Q_{r},$$
(3)
where $\bar{\omega}_{\mu}=\{\bar{\omega},\;\omega_{k}\}$. In terms of these, we
define thermal dressed states, precisely described in termalizacao ; termica .
It is worthwhile to note that our renormalized coordinates are
objects different from both the bare coordinates, $q$, and the
normal coordinates $Q$. Also, our dressed states, although being
collective objects, should not be confused with the eigenstates
of the system footnote .
In terms of our renormalized coordinates and dressed states, we can find a natural
division of the system into the dressed (physically observed) particle and
the dressed environment. The dressed particle will contain automatically
all the effects of the environment on it.
A cavity of arbitrary size at finite temperature
To study the behavior of the system in a cavity of arbitrary size, we write the initial
physical state in terms of dressed coordinates. We assume that initially the system is described by the density operator,
$\hat{\rho}(0)=\hat{\rho}_{0}´\otimes\hat{\rho}_{\beta}^{\prime}\;,$
where $\hat{\rho}_{0}´$ is the density operator associated with the oscilator $q_{0}^{\prime}$, that
can be in general in a pure or mixed state. Also,
$\hat{\rho}_{\beta}^{\prime}$ is the dressed density operator associated with the dressed field
modes. We assume thermal equilibrium for these dressed modes, thus
$$\hat{\rho}_{\beta}^{\prime}=\frac{\bigotimes_{k}e^{-\beta\hat{H}_{k}^{\prime}}%
}{{\rm Tr}\bigotimes_{k}e^{-\beta\hat{H}_{k}^{\prime}}}\,\;{\rm{with}}\;\hat{H%
}_{k}^{\prime}=\frac{1}{2}\hat{p}_{k}^{\prime 2}+\frac{1}{2}\omega_{k}^{2}\hat%
{q}_{k}^{\prime 2}\;.$$
(4)
The time evolution of the density operator is given by the Liouville-von Newman equation, whose solution,
in the case of an entropic evolution, is given by
$\hat{\rho}(t)=e^{-\frac{i}{\hbar}\hat{H}t}\hat{\rho}(0)e^{\frac{i}{\hbar}\hat{%
H}t}\;.$
Then, the time evolution of the average thermal expectation value of an operator is given by,
$$\langle\hat{A}\rangle(t)={\rm Tr}\left(\hat{A}e^{-\frac{i}{\hbar}\hat{H}t}\hat%
{\rho}(0)e^{\frac{i}{\hbar}\hat{H}t}\right)\,={\rm Tr}\left(\hat{A}(t)\hat{%
\rho}(0)\right)\,$$
(5)
where the ciclic property of the trace has been used; above,
$\hat{A}(t)=e^{\frac{i}{\hbar}\hat{H}t}\hat{A}e^{-\frac{i}{\hbar}\hat{H}t}$ is the time-dependent operator
$\hat{A}$ in the
Heisenberg representation.
Dressed coherent states in a heated environment
Let us consider a Brownian
particle embedded in a heated environment as described above. In our language,
we speak of a dressed Brownian
particle and we use the dressed state formalism.
We assume, as usual, that initially the particle and the environment are
decoupled and that the coupling is turned on suddenly at some given time, that we choose at
$t=0$.
In our formalism, we define
$|\lambda\left.\right>$ as a dressed coherent state given by,
$$|\left.\lambda,n_{1}^{\prime},n_{2}^{\prime},..;t=0\right>={\rm e}^{-|\lambda|%
^{2}/2}\sum_{n_{0}^{\prime}=0}^{\infty}\frac{\lambda^{n_{0}^{\prime}}}{\sqrt{n%
_{0}^{\prime}!}}|\left.n_{0}^{\prime}n_{1}^{\prime},..\right>\;,$$
(6)
where $n_{0}^{\prime}$ stands for the occupation number of the dressed particle
and where $n_{1}^{\prime},n_{2}^{\prime},..$ are the occupation numbers of the field modes. For
zero temperature we have $n_{1}^{\prime}=n_{2}^{\prime},..=0$.
For finite temperature,
we will do computations taking $\hat{\rho}_{0}=|\lambda\rangle\langle\lambda|$. This means,
that at the initial time, the dressed particle oscillator is in a pure coherent state.
Keeping this in mind, we consider the quantity,
$\langle\hat{q}_{0}^{\prime}\rangle(t)$, that we denote by $q_{0}^{\prime}(t)$,
$q_{0}^{\prime}(t)={\rm Tr}\left(\hat{q}^{\prime}_{0}(t)\hat{\rho}(0)\right)$.
In order to evaluate the above expression we first compute $\hat{q}^{\prime}_{0}(t)$. Using the relation
between the dressed coordinates and the normal coordinates, Eq.(3),
the expression for $\hat{H}$ in terms of the normal coordinates,
and the Baker-Campbell-Hausdorff formulas, we
get, after some steps of calculation,
$$q_{0}^{\prime}(t)=\sqrt{\frac{\hbar}{2\bar{w}_{0}}}\left(\lambda f_{00}(t)+%
\lambda^{\ast}f_{00}^{\ast}(t)\right),$$
(7)
where $f_{00}(t)$ is one of the quantities $f_{\mu\nu}(t)=\sum_{s}t_{\mu}^{s}t_{\nu}^{s}e^{-i\Omega_{s}t}$ adolfo3 .
Note that the above expression is independent of the temperature and coincides with the one
obtained previously for the the zero temperature case adolfo3 . This is because
$\hat{\rho}_{\beta}$ has even parity in the dressed momentum and position operators.
Due to the same reason, we find an entirely similar
formula for $p_{0}^{\prime}(t)$.
The situation is different for the quantity $q_{0}^{\prime 2}(t)=\langle\hat{q}_{0}^{\prime 2}\rangle(t)$.
After performing similar computations as above,
we get,
$$\displaystyle q_{0}^{\prime 2}(t,\beta)=\frac{\hbar}{2\bar{\omega}}\left[\left%
(\lambda f_{00}(t)+\lambda^{\ast}f_{00}^{\ast}(t)\right)^{2}\right.$$
$$\displaystyle\left.+2\sum_{k}|f_{0k}(t)|^{2}n_{k}^{\prime}(\beta)+1\right]\;,$$
(8)
Then from Eqs. (7) and (8) we obtain for the mean square error,
$$\displaystyle(\Delta q^{\prime}_{0})^{2}(t,\beta)$$
$$\displaystyle=$$
$$\displaystyle\langle\hat{q}^{\prime 2}_{0}\rangle(t,\beta)-(\langle\hat{q}_{0}%
^{\prime}\rangle(t))^{2}$$
(9)
$$\displaystyle=$$
$$\displaystyle\frac{\hbar}{2\bar{\omega}}+\frac{\hbar}{\bar{\omega}}\sum_{k}|f_%
{0k}(t)|^{2}n_{k}^{\prime}(\beta)\;,$$
where $n_{k}^{\prime}(\beta)$ is given by the Bose-Einstein distribution,
$n_{k}^{\prime}(\beta)=1/(e^{\hbar\beta\omega_{k}}-1)$ termica .
Analogously we obtain
the momentum mean squared error,
$$\displaystyle(\Delta p_{0})^{2}(t,\beta)$$
$$\displaystyle=$$
$$\displaystyle p_{0}^{\prime 2}(t,\beta)-(p_{0}^{\prime}(t))^{2}$$
(10)
$$\displaystyle=$$
$$\displaystyle\frac{\hbar\bar{\omega}}{2}+\hbar\bar{\omega}\sum_{k}|f_{0k}(t)|^%
{2}n_{k}^{\prime}(\beta).$$
From Eqs. (9) and (10) we obtain the time and temperature-dependent Heisenberg relation,
$$\Delta q_{0}^{\prime}(t,\beta)\Delta p_{0}^{\prime}(t,\beta)=\frac{\hbar}{2}+%
\hbar\sum_{k}|f_{0k}(t)|^{2}n_{k}^{\prime}(\beta)\;.$$
(11)
Time behavior in a cavity with a heated environment
Let us consider the time evolution of the uncertainty relation $\Delta q_{0}^{\prime}(t,\beta)\Delta p_{0}^{\prime}(t,\beta)$
given in Eq. (11), in a finite (small) cavity, characterized by the dimensionless
parameter $\delta=gR/\pi c$ and take a
coupling regime defined by a relation between $g$ and the
emission frequency $\bar{\omega}$, $g=\alpha\,\bar{\omega}$ . For instance, if we consider $\delta=0.1$,
$\alpha=0.2$ and $\bar{\omega}\approx 10^{14}/{\rm{s}}$ (in the visible red),
this corresponds to a cavity radius $R\sim 10^{-6}{\rm{m}}$. We
measure the uncertainty relation in units of $\hbar$ and call it for simplicity
$\Delta q_{0}^{\prime}(t,\beta)\Delta p_{0}^{\prime}(t,\beta)=\Delta(t,\beta)$.
Then calculations can be performed in a similar
way as in Ref. termica and we obtain (there is no confusion between the variable $t$ describing time
and the matrix elements $t_{\mu}^{r}$),
$$\displaystyle\Delta(t,\beta)=\frac{1}{2}+\sum_{k=1}^{\infty}\frac{1}{e^{(\hbar%
\beta g/\delta)k}-1}\left[(t_{0}^{0})^{2}(t_{k}^{0})^{2}\right.$$
$$\displaystyle\left.+2\sum_{l=1}^{\infty}t_{0}^{0}t_{0}^{l}t_{k}^{0}t_{k}^{l}%
\cos(\Omega_{0}-\Omega_{l})t\right.$$
$$\displaystyle\left.+\sum_{l,n=1}^{\infty}t_{0}^{l}t_{0}^{n}t_{k}^{l}t_{k}^{n}%
\cos(\Omega_{l}-\Omega_{n})t\right]\,.$$
The matrix
elements $t_{\mu}^{r}$ in the above formulas are
evaluated in termica ,
$$\displaystyle t_{k}^{0}\approx\frac{k\,g^{2}\sqrt{2\delta}}{k^{2}g^{2}-\Omega_%
{0}^{2}\delta^{2}}\,;\;\;t_{k}^{l}\approx\frac{2k\delta}{k^{2}-(l+\epsilon_{l}%
)^{2}}\frac{1}{l}$$
$$\displaystyle(t_{0}^{k})^{2}\approx\frac{2gR}{\pi ck^{2}}=\frac{2\delta}{k^{2}%
}\,,\;\;(t_{0}^{0})^{2}\approx 1-\frac{\pi gR}{3c}=1-\frac{\pi^{2}\delta}{3},$$
where $\epsilon_{l}$ is a small quantity such that $0<\epsilon_{l}<1$. Actually, for a
small cavity ($\delta\ll 1$) $\epsilon_{l}\approx\delta/k$.
Comments
Eq. (LABEL:n0tbeta)
describes the time evolution of the uncertainty relation
$\Delta q_{0}^{\prime}(t,\beta)\Delta p_{0}^{\prime}(t,\beta)\equiv\Delta(t,\beta)$ in a small cavity.
A plot of this time evolution is given in Fig. 1
for some representative values of the temperature.
The thermal uncertainty function, $\Delta(t,\beta)$, is an oscillating
function which attains periodically an absolute minimum (maximum) value,
${\rm{Min}}[\Delta(t,\beta)]$ (${\rm{Max}}[\Delta(t,\beta)]$). Since
the periodical character of $\Delta(t,\beta)$ does not involve $\beta$, the
location of these extrema on the time axis
do not depend on the temperature. Indeed, we can see from Fig.1, that
the values of these minima and maxima depend on the temperature,
$\beta^{-1}$, but appears to be, for each temperature, the same for all values of time where they occur,
in other words the values of the absolute extrema appears to be independent of time.
In the detail of Fig.1, are plotted two neighboring absolute minimum and
maximum (corresponding to $t\approx 2.3$ and $t\approx 2.5$), as functions of temperature.
We find from these figures that raising the temperature increases the
amplitude of oscillation and the mean value of the uncertainty relation and that its
lower and upper bounds, also grows with temperature.
We infer from Fig.1 that for a small cavity, in all cases an oscillatory behavior is
present for $\Delta(t,\beta)$, with the amplitude of the oscillation depending on the temperature, $T$.
For larger values of $T$ the
amplitude of the oscillation, and both, its absolute minimum and maximum values,
are larger than for lower temperatures.
This behavior of the uncertainty principle is to be contrasted with
the case of an arbitrarily large cavity (free space). In this last case, the product $(\Delta\,p)\,(\Delta\,q)$
goes, asymptotically, as $t\rightarrow\infty$ for each temperature, to a constant value $\Delta(\beta)$.
This asymptotic value depends on the temperature and grows with it, but is independent of time.
Distinctly, for a small cavity, even for $t\rightarrow\infty$, the product $(\Delta\,p)\,(\Delta\,q)$ presents
oscillations, which have larger and larger amplitudes for higher temperatures.
The result above falls in a general context of the different behaviors of quantum systems confined in
cavities, as compared to free space, in both, zero or finite temperature, the system
being investigated using the formalism presented in this report.
In adolfo3 some of
us got the expected result
that the probability, $P(t)$, that a simple cold atom in free space, excited at $t=0$ remain excited
after an elapsed time $t$, decays monotonically, going to zero as $t\rightarrow\infty$, while
in a small cavity,
$P(t)$ has an oscillatory behavior, never reaching zero. For $\bar{\omega}\sim 10^{14}$ (in the visible red),
$R\approx 10^{-6}{\rm{m}}$, in a weak (of the order of electromagnetic) coupling regime, ${\rm{Min}}\,P(t)\approx 0.98$,
which is in agreement with experimental observations Haroche3 . At finite temperature it is obtained in termica ,
for a small cavity, that the occupation number of a simple atom in a heated environment has an oscillatory
behavior with time and that its mean value increases with increasing temperature. In emaranho ; erico the
behavior of an entangled bipartite system at zero and finite temperature is investigated; taking two measures of
entanglement, it results an oscillatory behavior for a small cavity (entanglement is preserved at all times),
while it disappears as $t\rightarrow\infty$ for free space.
We hope that the result presented in this report could have some usefulness in nanophysics or in quantum
information theory. At this moment we are not able to comment about these aspects; they will be subject of future studies.
Acknowledgements: The author thanks FAPERJ and CNPq (brazilian agencies) for partial financial support.
References
(1)
W.G. Unruh, W.H. Zurek, Phys. Rev. D, 40, 1071 (1989)
(2)
B.L. Hu, Juan Pablo Paz, Yuhong Zhang, Phys, Rev. D, 45, 2843 (1992)
(3)
F. Haake, R. Reibold, Phys. Rev. A, 32, 2462 (1982)
(4)
A.O. Caldeira, A.J. Legget, Ann. Phys. (N.Y) 149, 374 (1983)
(5)
F. C. Khanna, A. P. C. Malbouisson, J. M. C. Malbouisson and
A. E. Santana, Thermal Quantum Field Theory: Algebraic Aspects and
Applications (World Scientific, Singapore, 2009).
(6)
G. Flores-Hidalgo, C. A. Linhares, A. P. C. Malbouisson
and J. M. C. Malbouisson, J. Phys. A 41, 075404 (2008).
(7)
G. Flores-Hidalgo, A. P. C. Malbouisson, J. M. C.
Malbouisson, Y. W. Milla and A. E. Santana, Phys. Rev. A 79, 032105
(2009).
(8)
F. C. Khanna, A. P. C. Malbouisson, J.
M. C. Malbouisson and A.E. Santana, Phys. Rev. A 81, 032119
(2010).
(9)
E. R. Granhen, C. A. Linhares, A. P. C. Malbouisson and
J. M. C. Malbouisson, Phys. Rev. A 81, 053820 (2010); C. A. Linhares, A. P. C. Malbouisson and
J. M. C. Malbouisson, Phys. Rev. A 82, 055805 (2010).
(10)
E.G. Figueiredo, C. A. Linhares, A. P. C. Malbouisson
and J. M. C. Malbouisson, Phys. Rev. A 84, 045802 (2011).
(11)
Aashish Clerk, Nature Nanotechnology 4, 796 (2009)
(12)
Sumiyoshi Abe and Norikazu Suzuki , Phys. Rev. A 41, 4608 (1990)
(13)
B. L. Hu, Yuhong Zhang , Mod. Phys. Lett. A8 3575 (1993)
(14)
B. L. Hu, Yuhong Zhang , Int.J.Mod.Phys. A10 4537 (1995)
(15)
G. Amelino-Camelia, M. Arzano, Yi Ling and G. Mandanici
Class. and Quant Grav. 23, 2585 (2006)
(16)
N.P.Andion, A.P.C. Malbouisson and A. Mattos Neto, J.Phys.A34,
3735, (2001)
(17)
G. Flores-Hidalgo, A.P.C. Malbouisson, Phys. Rev. A66, 042118 (2002).
(18)
A.P.C. Malbouisson, Annals of Physics 308, 373
(2003).
(19)
G. Flores-Hidalgo and A.P.C. Malbouisson, Phys. Lett. A 337, 37 (2005).
(20)
W. Thirring, F. Schwabl, Ergeb. Exakt. Naturw. 36,
219 (1964)
(21)
C. Cohen-Tannoudji, Atoms in Electromagnetic Fields,
(World Scientific, Singapure, 1994).
(22)
T. Petrosky, G. Ordonez and I. Prigogine, Phys. Rev. A
68, 022107 (2003).
(23)
Notice that our dressed states are not the same as those employed
in studies involving the interaction of atoms and
electromagnetic fields bouquinCohen and in the study of the radiation damping of
classical systems petrosky .
(24)
W. Jhe, A. Anderson, E.A. Hinds, D. Meschede, L. Moi, S. Haroche,
Phys. Rev. Lett., 58, 666 (1987) |
Uncovering latent structure in valued graphs: A variational approach
[
[
[
[
[
[
AgroParisTech and INRA, AgroParisTech and INRA
and
INRA and University Bordeaux I
M. Mariadassou
S. Robin
UMR 518 AgroParisTech/INRA MIA
16, rue C. Bernard, F-75005 Paris
France
E-mail: e2
C. Vacher
UMR 1202 Univ. Bordeaux I/INRA BioGeCo
69, route d’Arcachon, F-33612
Cestas
France
(#1; 10 2008; 4 2010)
Abstract
As more and more network-structured data sets are available, the
statistical analysis of valued graphs has become common place.
Looking for a latent structure is one of the many strategies used
to better understand the behavior of a network. Several methods
already exist for the binary case.
We present a model-based strategy to uncover groups of nodes in
valued graphs. This framework can be used for a wide span of
parametric random graphs models and allows to include covariates.
Variational tools allow us to achieve approximate maximum
likelihood estimation of the parameters of these models. We
provide a simulation study showing that our estimation method
performs well over a broad range of situations. We apply this
method to analyze host–parasite interaction networks in forest
ecosystems.
,
host–parasite interactions,
latent structure,
mixture model,
random graph,
valued graph,
variational method,
doi: 10.1214/10-AOAS361
††volume: #1††issue: 2
4
2010
{aug}
A]Mahendra Mariadassoulabel=e1][email protected],
A]Stéphane Robinlabel=e2][email protected]
and C]Corinne Vacherlabel=e3][email protected]
Ecological networks
.
1 Introduction
Data sets presenting a network structure are increasingly studied in many
different domains such as sociology, energy, communication, ecology or
biology [Albert and Barabási (2002)]. Statistical tools are therefore needed to analyze
the structure of these networks, in order to understand their
properties or
behavior. A strong attention has been paid to the study of various
topological characteristics such as degree distribution, clustering
coefficient and diameter [see, e.g., Barabási and Albert (1999), Newman, Watts and Strogatz (2002)]. These
characteristics are useful to describe networks but not sufficient to
understand its whole structure.
A natural and intuitive way to capture an underlying structure is to look
for groups of edges having similar connection profiles [Getoor and Diehl (2004),
Newman, Watts and Strogatz (2002)], which is refereed to as community detection
[Girvan and Newman (2002), Newman (2004)]. This usually turns into an unsupervised
classification (or clustering) problem which requires efficient estimation
algorithms since the data set at hand is ever increasing.
Several attempts at community detection have been proposed in the
literature: greedy algorithms for community detection [Girvan and Newman (2002) and
Newman (2004)] and clustering based on spectral analysis of the adjacency
matrix of a graph [von Luxburg, Belkin and
Bousquet (2008)]. Greedy algorithms and spectral
clustering both assume that communities are determined by a strong
within connectivity opposed to a low between connectivity. This might
be true for so-called communities but need not be true for other
groups of nodes. For example, a group of nodes loosely connected to
each other but highly connected to a specific group of hubs have the
same connection profile and form a homogeneous group but do not form
a community. In addition, they do not offer an explicit generative
model nor a criterion to select the correct number of communities.
Model-based methods are appealing by contrast: explicit modeling of the
heterogeneity between nodes gives different groups an intuitive and
easy to
understand interpretation. Several probabilistic models exists for random
graphs [see Pattison and Robins (2007) for a complete review], ranging from the seminal
Erdös–Rényi (ER) model [Erdös and Rényi (1959)] to the sophisticated Stochastic
Block Model (SBM) [Nowicki and Snijders (2001)]. The ER model assumes independent and
identically distributed edges which entails that all nodes are structurally
equivalent and, thus, there is only one community, although a big one. The
$p_{1}$ model from Holland and Leinhardt (1981) extended the ER model by assuming independent
dyads instead of edges, allowing the breakthrough from
undirected to directed graphs. But again, all nodes are structurally
equivalent in the $p_{1}$ model. Fienberg and Wasserman (1981) and Fienberg, Meyer and Wasserman (1985) lifted these
constraints by assuming the nodes are distributed among $Q$ classes with
different connectivity profiles. In this model, groups are easily
interpreted as nodes belonging to the same class. Unfortunately,
Fienberg, Meyer and Wasserman (1985) assumes class assignments are perfectly well known, which
rarely happens. The state of the art in terms of graph modeling is the SBM,
inspired by Lorrain and White (1971) and introduced by Nowicki and Snijders (2001), which takes
advantage of mixture models and unknown latent variables to allow an easy
modeling of groups without requiring them to be known in advance.
In the SBM framework, community detection boils down to three crucial
steps: assignment of nodes to groups, estimation of the model parameter and
selection of the correct number of groups. Several authors offered their
method to solve these issues using Bayesian methods. Nowicki and Snijders (2001) work with
the original SBM model. Hofman and Wiggins (2008) work in a highly constrained version
of SBM in which heterogeneity is strictly limited to intra- and
inter-community connection and thus characterized by only two parameters,
against $Q^{2}$ in the unconstrained SBM. Airoldi et al. (2008) extend the SBM
framework by allowing nodes to exhibit multiple communities. By contrast,
Daudin, Picard and Robin (2008) use a frequentist approach to estimate the parameters of the
SBM. The frequentist approach is less computation intensive than its
Bayesian counterpart, whereas the Bayesian approach is supposed to better
account for the uncertainty. With the notable exception of Nowicki and Snijders (2001),
who use MCMC to estimate the model parameter, both lines of work make heavy
use of variational techniques: either Variational EM [Jaakkola (2000)] or
Variational Bayes [Attias (2000); Beal and Ghahramani (2003); Xing, Jordan and Russell (2003); Winn, Bishop and Jaakkola (2005)]. MCMC computational cost
is prohibitive, effectively leading to severe size limitations (around 200
nodes). Furthermore, because of the complex likelihood landscape in the
SBM, good mixing of the Markov Chain is hard to achieve and
monitor. Variational approximations, by contrast, replace the
likelihood by
a simple surrogate, chosen so that the error is minimal in some
sense. Frequentist and Bayesian approach then differ only in the use of
this surrogate likelihood: Bayesians combine it to a prior distribution of
the parameter (chosen from some suitable distribution), whereas frequentists
use it directly. In all these methods, the number of groups is fixed during
the estimation procedure and must be selected using some criterion. By
contrast, Kemp, Griffiths and Tenenbaum (2004) propose an original approach where the number of
groups changes and is selected during the estimation process. Both
Bayesian and frequentist estimations approaches give the same kind of
results: an optimal number of groups and a probabilistic assignment of
nodes to groups, depending on their connection profile. However, the
Bayesian estimation strategy leads to severe constraints on the choice of
prior and hyperprior distributions. The Daudin, Picard and Robin (2008) maximum likelihood
approach does not require any prior specification and is more efficient
than MCMC estimation [Picard et al. (2007)].
Previous models are all models for binary networks, for which the only
information is the presence or absence of an edge. Binary information
certainly describes the topology of a network but is a rather poor
description. It accounts neither for the intensity of the
interaction between two nodes nor for the specific features of an
edge. The intensity of an edge may typically indicate the amount of
energy transported from one node to another, the number of passengers
or the number of common features between two nodes, whereas the
specific feature of an edge may be the phylogenetic distance between
its two ending nodes. Many networks, such as power, communication,
social, ecological or biological networks, are naturally valued and
are somehow arbitrarily transformed to a binary graph. This
transformation sometimes conceals important results [Tykiakanis, Tscharntke and
Lewis (2007)].
Extending binary models and the associated estimation procedures to
valued graphs with specific features allows more complexity, and more
relevant information with it, to be processed while estimating the
structure of the network.
We are motivated by the search of a structure in valued graphs
describing the similarity between species within an assemblage
according to their biotic interactions. In ecology, an assemblage is
defined as a taxonomically related group of species that occurs in the
same geographic area [Ricklefs and Miller (2000)]. The species composing an
assemblage usually interact with many species belonging to other
assemblages and the nature of these interactions is often very diverse
(predator–prey interactions, host–parasite interactions, mutualistic
interactions, competitive interactions). One of the questions facing
ecologists is to understand what determines with whom a species
interact. Conventional wisdom is that within an assemblage, two
closely related species should share more interactions than two
evolutionary distant species because the range of interactions of a
species is constrained by its physiological, morphological and
behavioral attributes. In several cases, this conventional wisdom
is revealed to be true. Phylogenetically related plant species have been
shown to bear similar pathogens and herbivores [Brandle and Brandl (2006); Gilbert and Webb (2007)]
and the diet’s range of predators has been shown to be
phylogenetically constrained [Cattin et al. (2004)]. This tendency for
phylogenetically related species to resemble each other is called
phylogenetic signal [Blomberg and Garland (2002)]. In other cases, no phylogenetic
signal was detected [Rezende et al. (2007); Vacher, Piou and
Desprez-Loustau (2008)]. Selection pressures exerted
by the environment might account for this absence: species have to
adapt to varying environments to survive, diverging from close
relatives in their physiology, morphology and behavior, and possibly
developing novel interactions [Bersier and Kehrli (2008); Cattin et al. (2004)]. The valued graphs
under study have species as nodes and the number of shared
interactions as edges. We use a mixture model with phylogenetic
distance between species as covariate to measure the strength of the
phylogenetic signal. This latter is defined as the decrease in the
number of selected groups due to the inclusion of the covariate. Two
different assemblages are considered. The first assemblage is composed
of 51 tree species occurring in the French forests and the second is
composed of 153 parasitic fungal species also occurring in the French
forests. The interactions considered are host–parasite interactions.
We expect to find a lower phylogenetic signal in the host range of
parasitic fungal species [Bersier and Kehrli (2008); Rossberg et al. (2006); Vacher, Piou and
Desprez-Loustau (2008)]
than in the
vulnerability of tree species to parasites [Brandle and Brandl (2006); Gilbert and Webb (2007);
Vacher, Piou and
Desprez-Loustau (2008)].
In this paper we propose an extension to the stochastic block model,
introduced in Fienberg and Wasserman (1981); Fienberg, Meyer and Wasserman (1985); Nowicki and Snijders (2001), and the methods
of Airoldi and Carley (2005)
and Daudin, Picard and Robin (2008), that deals with valued graphs and accounts for
possible covariates. We use a general mixture model describing the
connection intensities between nodes spread among a certain number of
classes (Section 2). A variational EM approach to get an
optimal, in a sense to be defined, approximation of the likelihood is
then presented in Section 3. In Section
4 we give a general estimation algorithm and derive some
explicit formulas for the most popular distributions. The quality of
the estimates is studied on synthetic data in Section 5.
Finally, the model is used to elucidate the structure of host–parasite
interactions in forest ecosystems and results are discussed in Section
6.
2 Mixture model
We now present the general extension of SBM to valued graphs and
discuss the two particular modelings used for the tree species and
fungal species interaction networks.
2.1 Model and notation
Nodes.
Consider a graph with $n$ nodes, labeled in
$\{1,\ldots,n\}$. In our model the nodes are distributed among
$Q$ groups so that each node $i$ is associated to a random vector
$\mathbf{Z}_{i}=(Z_{i1},\ldots,Z_{iQ})$, with $Z_{iq}$ being $1$ if node
$i$ belongs to group $q$ and $0$ otherwise. The $\{\mathbf{Z}_{i}\}$ are
supposed to be independent identically distributed observations from
a multinomial distribution:
$$\{\mathbf{Z}_{i}\}_{i}\mbox{ i.i.d.}\sim\mathcaligr{M}(1;\bolds\alpha),$$
(1)
where $\bolds\alpha=(\alpha_{1},\ldots,\alpha_{Q})$ and $\sum_{q}\alpha_{q}=1$.
Edges.
Each edge from a node $i$ to a node $j$ is
associated to a random variable $X_{ij}$, coding for the strength of
the edge. Conditionally to the group of each node, or
equivalently knowing the $\{\mathbf{Z}_{i}\}$, the edges are supposed to be
independent. Knowing group $q$ of node $i$ and group $\ell$ of
node $j$, $X_{ij}$ is distributed as $f(\cdot,\theta_{q\ell}):=f_{q\ell}(\cdot)$, where $f_{\theta_{q\ell}}$ is a probability
distribution known up to a finite-dimensional parameter
$\theta_{q\ell}$:
$$X_{ij}|i\in q,j\in\ell\sim f(\cdot,\theta_{q\ell}):=f_{q\ell}(\cdot).$$
(2)
Up to a relabeling of the classes, the model is identifiable and
completely specified by both the mixture proportions $\bolds\alpha$ and
the connectivity matrix $\bolds\theta=(\theta_{q\ell})_{q,\ell=1,\ldots,Q}$. We denote $\bolds\gamma=(\bolds\alpha,\bolds\theta)$ the parameter of the
model.
Directed and undirected graphs
This modeling can be applied to both directed and undirected graphs.
In the directed version, the variables $X_{ij}$ and $X_{ji}$ are
supposed to be independent conditionally to the groups to which
nodes $i$ and $j$ belong. This hypothesis is not always realistic
since, for example, the traffic from $i$ to $j$ is likely to be
correlated to the traffic from $j$ to $i$. A way to account for such
a dependency is to consider a undirected graph with edges labeled
with the bivariate variables $\{(X_{ij},X_{ji})\}_{1\leq i<j\leq n}$. All the results presented in this paper are valid for
directed graphs. The results for undirected graphs can easily be
derived and are only briefly mentioned.
2.2 Modeling the number of shared hosts/parasites
In our tree interaction network, each edge is valued with the number
of common fungal species two tree species can host. Our purpose is to
understand the structure of this network and it is natural to model
the counts $X_{ij}$ as Poisson distributed. The mixture models aims
at explaining the heterogeneity of the $X_{ij}$. However, we would
also like to account for some factors that are known to be
influential. In our network, we expect two phylogenetically related
tree species $i$ and $j$ to share a high number $X_{ij}$ of parasitic
species. As such, their average number of shared parasitic species
$\mathbb{E}[X_{ij}]$ is expected to decrease with their phylogenetic
distance $y_{ij}$. We consider three alternatives, and compare two of
them.
Poisson mixture (PM):
In this mixture, we do not account for the
covariates and $X_{ij}$ only depends on the classes of $i$ and $j$:
$$X_{ij}|i\in q,j\in\ell\sim\mathcaligr{P}(\lambda_{q\ell}).$$
$\lambda_{q\ell}$ is then the mean number of common fungal species
(or mean interaction) between a tree species from group $q$ and one
from group $\ell$ and $\theta_{q\ell}=\lambda_{q\ell}$.
Poisson regression mixture with inhomogeneous effects (PRMI):
In
this mixture, we account for the covariates via a regression model that
is specific to the classes of $i$ and $j$:
$$X_{ij}|i\in q,j\in\ell\sim\mathcaligr{P}(\lambda_{q\ell}e^{\beta_{q\ell}^{%
\intercal}\mathbf{y}_{ij}}),$$
where $\mathbf{y}_{ij}$ is a vector of covariates and $\theta_{q\ell}=(\lambda_{q\ell},\beta_{q\ell})$.
Poisson regression mixture with homogeneous effects (PRMH):
In
this mixture, the effect of the covariates does not depend on the classes
of $i$ and $j$:
$$X_{ij}|i\in q,j\in\ell\sim\mathcaligr{P}(\lambda_{q\ell}e^{\beta^{\intercal}%
\mathbf{y}_{ij}}),$$
$\theta_{q\ell}=(\lambda_{q\ell},\beta)$.
We point out that models PRMI and PRMH have different purposes. In PRMI,
the link between the covariates and the edges is locally refined
within each class $(q,\ell),$ whereas in PRMH, the
covariates compete
globally with the group structure found by PM. In PRMH, the mixture looks
for remaining structure among the residuals of the regression model. If the
structure was completely explained by the covariates, the possibly many
components found using PM would reduce to a single component when using
PRMH. To a lesser extent, we expect the number of components to be smaller
with PRMH than with PM if the phylogenetic distance explains part of the
structure. As we look for structure beyond the one already explained by the
covariates, we consider only models PM and PRMH.
The same models are used for the fungal species interaction network.
In our examples, data consist in counts, but other types of data can be
handled with similar mixture and/or regression models (see Appendix
.6 for details).
3 Likelihood and variational EM
We now address the estimation of the parameter $\bolds\gamma=(\bolds\alpha,\bolds\theta)$. We show that the standard maximum likelihood
approach cannot be applied to our model and propose an alternative
strategy relying on variational tools, namely, variational EM.
3.1 Likelihoods
Let $\mathbf{X}$ denote the set of all edges, $\mathbf{X}=\{X_{ij}\}_{i,j=1,\ldots,n}$,
and $\mathbf{Z}$ the set of all indicator variables for nodes, $\mathbf{Z}=\{\mathbf{Z}_{i}\}_{i=1,\ldots,n}$. In the mixture model literature [McLahan and Peel (2000)]
$(\mathbf{X},\mathbf{Z})$ is referred to as the complete data set, while $\mathbf{X}$ is
referred to
as the incomplete data set. The conditional independence of the edges
knowing $\mathbf{Z}$ entails the decomposition $\log{\mathbb{P}(\mathbf{Z},\mathbf{X})}=\log{\mathbb{P}(\mathbf{Z})}+\log{%
\mathbb{P}(\mathbf{X}|\mathbf{Z})}$. It then follows from
(1) and (2) that the log-likelihood of the
complete data set is
$$\log{\mathbb{P}(\mathbf{Z},\mathbf{X})}=\sum_{i}\sum_{q}Z_{iq}\log{\alpha_{q}}%
+\sum_{i\neq j}\sum_{q,\ell}Z_{iq}Z_{j\ell}\log{f_{q\ell}(X_{ij})}.$$
(3)
The likelihood of the incomplete data set can be obtained by summing
$\mathbb{P}(\mathbf{Z},\mathbf{X})$ over all possible $\mathbf{Z}$’s: $\mathbb{P}(\mathbf{X})=\sum_{\mathbf{Z}}\mathbb{P}(\mathbf{Z},\mathbf{X})$. This summation involves $Q^{n}$ terms and quickly becomes
intractable. The popular E–M algorithm [Dempster, Laird and Rubin (1977)], widely used in
mixture problems, allows to maximize $\log{\mathbb{P}(\mathbf{X})}$ without explicitly
calculating it. The E-step relies on the calculation of the conditional
distribution of $\mathbf{Z}$ given $\mathbf{X}$: $\mathbb{P}(\mathbf{Z}|\mathbf{X})$.
Unfortunately, in
the case of network data, the strong dependency between edges makes this
calculation untractable.
Undirected graphs
The closed formula (3) still holds undirected graphs,
replacing the
sum over $i\neq j$ by a sum over $i<j$. This is also true for
equations (3.2) and (11) given below.
3.2 Variational EM
We propose to use an approximate maximum likelihood strategy based on
a variational approach [see Jordan et al. (1999) or the tutorial by
Jaakkola (2000)]. This strategy is also used in Govaert and Nadif (2005) for a
biclustering problem. We consider a lower bound of the log-likelihood
of the incomplete data set
$$\mathcaligr{J}(R_{\mathbf{X}},\bolds\gamma)=\log{\mathbb{P}(\mathbf{X};\bolds%
\gamma)}-\mathit{KL}(R_{\mathbf{X}}(\cdot),\mathbb{P}(\cdot|\mathbf{X};\bolds%
\gamma)),$$
(4)
where $\mathit{KL}$ denotes the Kullback–Leibler divergence and $R_{\mathbf{X}}$ stands for
some distribution on $\mathbf{Z}$. Classical properties of the Kullback–Leibler
divergence ensure that $\mathcaligr{J}$ has a unique maximum
$\log{\mathbb{P}(\mathbf{X};\bolds\gamma)}$, which is reached for $R_{\mathbf{X}}(\mathbf{Z})=\mathbb{P}(\mathbf{Z}|\mathbf{X})$. In other words, if $\mathbb{P}(\mathbf{Z}|\mathbf{X};\bolds\gamma)$ was
tractable, the maximization of $\mathcaligr{J}(R_{\mathbf{X}},\bolds\gamma)$ with respect to
$\bolds\gamma$ would be equivalent to the maximization of
$\log{\mathbb{P}(\mathbf{X};\bolds\gamma)}$. In our case, $\mathbb{P}(\mathbf{Z}|\mathbf{X};\bolds\gamma)$ is
untractable and we maximize $\mathcaligr{J}(R_{\mathbf{X}},\bolds\gamma)$ with respect to both
$R_{\mathbf{X}}$ and $\bolds\gamma$. Jaakkola (2000) shows that $\mathcaligr{J}(R_{\mathbf{X}},\bolds\gamma)$ can be
rewritten as
$$\mathcaligr{J}(R_{\mathbf{X}},\bolds\gamma)=\mathcaligr{H}(R_{\mathbf{X}})+%
\sum_{\mathbf{Z}}R_{\mathbf{X}}(\mathbf{Z})\log{\mathbb{P}(\mathbf{X},\mathbf{%
Z};\bolds\gamma)},$$
(5)
where $\mathcaligr{H}(\cdot)$ denotes the entropy of a distribution. The last term
of (5) can be deduced from (3):
$$\displaystyle\sum_{\mathbf{Z}}R_{\mathbf{X}}(\mathbf{Z})\log{\mathbb{P}(%
\mathbf{X},\mathbf{Z};\bolds\gamma)}$$
(6)
$$\displaystyle\quad=\sum_{i}\sum_{q}\mathbb{E}_{R_{\mathbf{X}}}(Z_{iq})\log{%
\alpha_{q}}+\sum_{i\neq j}\sum_{q,\ell}\mathbb{E}_{R_{\mathbf{X}}}(Z_{iq}Z_{j%
\ell})\log{f_{q\ell}(X_{ij})},$$
where $\mathbb{E}_{R_{\mathbf{X}}}$ denotes the expectation with respect to distribution
$R_{\mathbf{X}}$. Equation (3.2) requires only the knowledge of $\mathbb{E}_{R_{\mathbf{X}}}(Z_{iq})$
and $\mathbb{E}_{R_{\mathbf{X}}}(Z_{iq}Z_{j\ell})$ for all $i,j,q,\ell$. By contrast,
$\mathcaligr{H}(R_{\mathbf{X}})$ requires all order moments of $R_{\mathbf{X}}$ and is untractable in
general. Maximization of $\mathcaligr{J}(R_{\mathbf{X}},\bolds\gamma)$ in $R_{\mathbf{X}}$ can not be achieved
without some restrictions on $R_{\mathbf{X}}$. We therefore limit the search to the
class of completely factorized distributions:
$$R_{\mathbf{X}}(\mathbf{Z})=\prod_{i}h(\mathbf{Z}_{i},\bolds\tau_{i}),$$
(7)
where $h$ denotes the multinomial distribution and $\bolds\tau_{i}$ stands
for a
vector of probabilities, $\bolds\tau_{i}=(\tau_{i1},\ldots,\tau_{iQ})$ (with
$\sum_{q}\tau_{iq}=1$). In particular, $\mathbb{E}_{R_{\mathbf{X}}}(Z_{iq})=\tau_{iq}$
and $\mathbb{E}_{R_{\mathbf{X}}}(Z_{iq}Z_{j\ell})=\tau_{iq}\tau_{j\ell}$. In addition,
the entropy is additive over the coordinates for factorized distributions,
so that $\mathcaligr{H}(R_{\mathbf{X}})=\sum_{i}\mathcaligr{H}(h(\cdot,\bolds\tau_{i}))%
=-\sum_{i}\sum_{q}\tau_{iq}\log{\tau_{iq}}$. Wrapping everything together,
$$\displaystyle\mathcaligr{J}(R_{\mathbf{X}},\bolds\gamma)$$
$$\displaystyle=$$
$$\displaystyle-\sum_{i}\sum_{q}\tau_{iq}\log{\tau_{iq}}+\sum_{i}\sum_{q}\tau_{%
iq}\log{\alpha_{q}}$$
$$\displaystyle{}+\sum_{i\neq j}\sum_{q,\ell}\tau_{iq}\tau_{j\ell}\log{f_{q\ell}%
(X_{ij})}.$$
It is immediate from (3.2) that $\mathcaligr{J}(R_{\mathbf{X}},\bolds\gamma)$ is
tractable for distributions $R_{\mathbf{X}}$ of the form (7). The
$\bolds\tau_{i}$’s must be thought of as variational parameters to be
optimized so
that $R_{\mathbf{X}}(\mathbf{Z})$ fits $\mathbb{P}(\mathbf{Z}|\mathbf{X};\bolds\gamma)$ as well as
possible; they
depend on the observed data $\mathbf{X}$. Since $R_{\mathbf{X}}$ is restricted to be
of the
form (7), $\mathcaligr{J}(R_{\mathbf{X}},\bolds\gamma)$ is a lower bound of
$\log\mathbb{P}(\mathbf{X})$.
Discussion about tighter bounds
A fully factorized $R_{\mathbf{X}}$ is only one class of distributions we can
consider. Broader distribution classes should yield tighter bound of
$\mathcaligr{J}(R_{\mathbf{X}},\bolds\gamma)$. Unfortunately, for more general
distributions, the
entropy $\mathcaligr{H}(R_{\mathbf{X}})$ may not have a simple expression anymore
rendering the
exact calculation of $\mathcaligr{J}(R_{\mathbf{X}},\bolds\gamma)$ untractable: better
accuracy is
achieved at the expense of tractability. A solution to this issue is Bethe
free energy [Yedidia, Freeman and Weiss (2005)]. We did not consider it because it relies
on an
approximation of $\mathcaligr{H}(R_{\mathbf{X}})$ which disrupts the well-behaved
properties of
$\mathcaligr{J}(R_{\mathbf{X}},\bolds\gamma)$.
Another approach comes from Leisink and Kappen (2001) and Mariadassou (2006). Starting
from an
exponential inequality, they emphasize the strong connection between fully
factorized $R_{\mathbf{X}}$ and first order linear approximation of the exponential
function. Using a higher approximation of the exponential and some
distribution $S_{\mathbf{X}}$ in addition to $R_{\mathbf{X}}$, it is possible to
derive an
even tighter bound of $\mathcaligr{J}(R_{\mathbf{X}},\bolds\gamma)$. However, the estimation
algorithm is then of complexity $\mathcaligr{O}(n^{6}Q^{6})$ instead of
$\mathcaligr{O}(n^{2}Q^{2})$ for a gain which has the same order of magnitude as
the computer numerical precision.
4 Parameter estimation
We present here the two-steps algorithm used for the parameter estimation.
4.1 Estimation algorithm
As explained in Section 3.2, the maximum likelihood
estimator of $\bolds\gamma$ is
$$\widehat{\bolds\gamma}_{\mathit{ML}}=\operatorname{\arg\max}\limits_{\bolds%
\gamma}\log{\mathbb{P}(\mathbf{X};\bolds\gamma)}=\operatorname{\arg\max}%
\limits_{\bolds\gamma}\max_{R_{\mathbf{X}}}\mathcaligr{J}(R_{\mathbf{X}},%
\bolds\gamma).$$
In the variational framework, we restrict the last optimization
problem to factorized distributions. The estimate we propose is
hence
$$\widehat{\bolds\gamma}=\operatorname{\arg\max}\limits_{\bolds\gamma}\max_{R_{%
\mathbf{X}}\mathrm{\ factorized}}\mathcaligr{J}(R_{\mathbf{X}},\bolds\gamma).$$
The simultaneous optimization with respect to both $R_{\mathbf{X}}$ and
$\bolds\gamma$ is still too difficult, so we adopt the following
iterative strategy. Denoting by $R_{\mathbf{X}}^{(n)}$ and ${\bolds\gamma}^{(n)}$
the estimates after $n$ steps, we compute
$$\cases{R_{\mathbf{X}}^{(n+1)}=\operatorname{\arg\max}\limits_{R_{\mathbf{X}}%
\mathrm{\ factorized}}\mathcaligr{J}\bigl{(}R_{\mathbf{X}},{\bolds\gamma}^{(n)%
}\bigr{)},\cr{\bolds\gamma}^{(n+1)}=\operatorname{\arg\max}\limits_{\bolds%
\gamma}\mathcaligr{J}\bigl{(}R_{\mathbf{X}}^{(n+1)},\bolds\gamma\bigr{)}.}$$
(9)
The next two sections are dedicated to each of these steps.
Initialization step
The optimization procedure (9) only ensures the
convergence toward a local optimum, so the choice of the starting
point for $\bolds\gamma$ or $R_{\mathbf{X}}$ is crucial to avoid local optima. This
choice is difficult, but, to our experience, hierarchical clustering
seems to be a good strategy to get an initial value for $R_{\mathbf{X}}$.
4.2 Optimal approximate conditional distribution $R_{\mathbf{X}}$
We consider here the optimization of $\mathcaligr{J}$ with respect to $R_{\mathbf{X}}$.
For a
given value of $\bolds\gamma$, wedenote $\widehat{\bolds\tau}$ the variational
parameter defining the distribution $\widehat{R}_{\mathbf{X}}=\break{\arg\max}_{R_{\mathbf{X}}\mathrm{\ factorized}%
}\mathcaligr{J}(R_{\mathbf{X}},\bolds\gamma)$. This amounts to maximimizing $\mathcaligr{J}(R_{\mathbf{X}},\bolds\gamma)$,given in (3.2), under the condition that, for all $i$, the
$\tau_{iq}$’s must sum to 1. The derivative of $\mathcaligr{J}(R_{\mathbf{X}},\bolds\gamma)$ with
respect to $\tau_{iq}$ is
$$-\log{\tau_{iq}}-1+\log{\alpha_{q}}+\sum_{j\neq i}\sum_{\ell}\tau_{j\ell}[\log%
{f_{q\ell}(X_{ij})}+\log{f_{\ell q}(X_{ji})}]+L_{i},$$
where $L_{i}$ denotes the $i$th Lagrange multiplier. It results from the
previous equation that the optimal variational parameter $\widehat{\bolds\tau}$
satisfies the fixed point relation
$$\widehat{\tau}_{iq}\propto\alpha_{q}\prod_{j\neq i}\prod_{\ell}[f_{q\ell}(X_{%
ij})f_{\ell q}(X_{ji})]^{\widehat{\tau}_{j\ell}}.$$
(10)
The fixed point relation (10) can be related to a mean field
approximation [see Jaakkola (2000)]. We get $\widehat{\bolds\tau}$ simply by
iterating this relation until convergence.
Undirected graphs
For a undirected graph, $\widehat{\bolds\tau}$ satisfies
$$\widehat{\tau}_{iq}\propto\alpha_{q}\prod_{j\neq i}\prod_{\ell}[f_{q\ell}(X_{%
ij})]^{\widehat{\tau}_{j\ell}}.$$
4.3 Parameter estimates
We now have to maximize $\mathcaligr{J}$ with respect to $\bolds\gamma=(\bolds\alpha,\bolds\theta)$ for a given distribution $R_{\mathbf{X}}$. Again, this amounts to
maximizing $\mathcaligr{J}(R_{\mathbf{X}},\bolds\gamma)$, given in (3.2), under
the condition that $\sum_{q}\alpha_{q}=1$. Straightforward
calculations show
that the optimal $\bolds\alpha$ and $\bolds\theta$ are given by
$$\widehat{\alpha}_{q}=\frac{1}{n}\sum_{i}\tau_{iq},\qquad\widehat{\theta}_{q%
\ell}=\operatorname{\arg\max}\limits_{\theta}\sum_{i\neq j}\tau_{iq}\tau_{j%
\ell}\log{f(X_{ij};\theta)}.$$
(11)
Poisson models
Poisson models are of particular
interest for
our interaction networks. The optimal $\lambda_{q\ell}$ for model PM
presented in Section 2.2 is straightforward:
$$\widehat{\lambda}_{q\ell}=\sum_{i\neq j}\tau_{iq}\tau_{j\ell}X_{ij}\Big{/}\sum%
_{i\neq j}\tau_{iq}\tau_{j\ell}.$$
For models PRMH and PRMI presented in the same section, there is no closed
formula for $\lambda_{q\ell}$, $\beta_{q\ell}$ or $\beta$.
However, since
the Poisson regression model belongs to the exponential family, $\mathcaligr{J}$ is
only a weighted version of the log-likelihoods of the corresponding
generalized linear model. As such, standard optimization procedures can be
used.
Exponential family
The optimal $\bolds\theta$ is not
explicit in
the general case, but has a simpler form if the distribution $f$
belongs to
the exponential family. Namely, if $f$ belongs to an exponential family
with natural parameter $\theta$,
$$f(x;\theta)=\exp{[\bolds\Psi(x)^{\prime}\theta-A(\theta)]}.$$
According to (11), we look for $\widehat{\theta}={\arg\max}_{\theta}\sum_{i\neq j}\tau_{iq}\tau_{j\ell}\bolds%
\Psi(X_{ij})^{\prime}\theta-A(\theta)$. Maximizing this quantity in $\theta$ yields
$$\sum_{i\neq j}\tau_{iq}\tau_{j\ell}\bolds\Psi(X_{ij})-\nabla A(\theta)=\mathbf%
{0}.$$
If $\nabla A$ is invertible, the optimal $\theta$ is
$$\widehat{\theta}=(\nabla A)^{-1}\biggl{[}\sum_{i\neq j}\tau_{iq}\tau_{j\ell}%
\bolds\Psi(X_{ij})\biggr{]}.$$
(12)
4.4 Choice of the number of groups
In practice, the number of groups is unknown and should be estimated.
Many criterion have been proposed to select the dimensionality $Q$ of
the latent space, ranging from AIC to ICL. AIC, BIC and their variants
[Burnham and Anderson (1998)] are based on computing the likelihood of the observed
data $\mathbb{P}(\mathbf{X}|m_{Q})$ and penalizing it with some function of $Q$. But
the use of variational EM is precisely to avoid computation of
$\mathbb{P}(\mathbf{X}|m_{Q})$, which is untractable. Given a prior distribution
$\mathbb{P}(m_{Q})$ over models, and a prior distribution $\mathbb{P}(\bolds\gamma|m_{Q})$
for each model, variational Bayes [Beal and Ghahramani (2003)] works by selecting
the model with maximum posterior $\mathbb{P}(m_{Q}|\mathbf{X})$. Estimation of
$\mathbb{P}(\mathbf{X}|m_{Q})$ is then performed using variational EM and no
penalization is required, as complex models are already penalized by
diffuse prior $\mathbb{P}(\bolds\gamma|m_{Q})$. Extension of Deviance Information
Criterion (DIC) to finite mixture distributions via variational
approximations [McGrory and
Titterington (2007)] is even more straightforward:
choosing $Q^{*}$ larger than the expected number of components and
running the algorithm, extraneous classes become void as the
algorithm converges and the selected number of groups is just the
number of nonempty classes. In the context of unknown assignments,
Biernacki, Celeux and Govaert (2000) proposed the Integrated Classification Likelihood (ICL),
which is an approximation to the complete data likelihood
$\mathbb{P}(\mathbf{X},\mathbf{Z}|m_{Q})$. Variational Bayes, BIC and ICL can all be seen
as approximations to Bayes factors. Whereas Variational Bayes
integrates out the uncertainty about the parameter and the assignment
of nodes to groups, ICL replaces them by a point estimate, computed
thanks to variational EM. Traditional model selection essentially
involves a trade-off between goodness of fit and model complexity,
whereas ICL values both goodness of fit and classification sharpness.
Nowicki and Snijders (2001) do not propose any criterion to select the number of
groups. Hofman and Wiggins (2008) use McGrory’s method but in a very specific case of
the Stochastic Block Model. They also give no clue as to how to decide that
the algorithm has converged enough. Airoldi et al. (2008) use either a
modification to BIC (for small size networks) or cross-validation (for
large size networks) to select the number of groups. Daudin, Picard and Robin (2008) use a
modification to ICL criterion. Following along the same line as
Daudin, Picard and Robin (2008), we use a modification of ICL adapted to valued graphs to
select the number of classes.
ICL criterion:
For a model $m_{Q}$ with $Q$ classes where $\bolds\theta$ involves $P_{Q}$
independent parameters, the ICL criterion is
$$\displaystyle\mathit{ICL}(m_{Q})$$
$$\displaystyle=$$
$$\displaystyle\max_{\bolds\gamma}\log\mathbb{P}(\mathbf{X},\widetilde{\mathbf{Z%
}}|\bolds\gamma,m_{Q})$$
$$\displaystyle{}-\frac{1}{2}\{P_{Q}\log[n(n-1)]-(Q-1)\log(n)\},$$
where the missing data $\mathbf{Z}$ are replaced by their prediction
$\widetilde{\mathbf{Z}}$.
Note that the penalty term $-\frac{1}{2}\{P_{Q}\log[n(n-1)]-(Q-1)\log(n)\}$ is similar to the one of BIC, where the
$\log$ term refers to number of data. In the case of graphs, the
number of data is $n$ (i.e., the number of nodes) for the vector of
proportions $\bolds\alpha$ ($Q-1$ independent parameters), whereas it is
$n(n-1)$ (i.e., the number of edges) for parameter $\bolds\theta$ ($P_{Q}$
independent parameters). For the models PM, PRMI and PRMH (detailed in
Section 2.2), $P_{Q}$ is respectively ${Q(Q+1)}/{2}$,
$Q(Q+1)$ and $1+{Q(Q+1)}/{2}$.
5 Simulation study
5.1 Quality of the estimates
Simulation parameters
We considered undirected networks of size $n=100$ and $500$ with
$Q=3$ classes. To study balanced and unbalanced proportions, we set
$\alpha_{q}\propto a^{q}$, with $a=1,0.5,0.2$. $a=1$ gives
uniform proportions, while $a=0.2$ gives very unbalanced
proportions: $\bolds\alpha=(80.6\%,16.1\%,3.3\%)$. We finally
considered symmetric connection intensities $\lambda_{pq}$, setting
$\lambda_{pp}=\lambda^{\prime}$ for all $p$ and $\lambda_{pq}=\lambda^{\prime}\gamma$ for $p\neq q$. Parameter $\gamma$ controls the difference
between within class and between class connection intensities
($\gamma=0.1$, $0.5$, $0.9$, $1.5$), while $\lambda^{\prime}$ is set so
that the mean connection intensity $\lambda$ ($\lambda=2,5$)
depends neither on $\gamma$ nor $a$. $\gamma$ close to one makes the
distinction between the classes difficult. $\gamma$ larger than one
makes the within class connectivities less intense than the between
ones. We expect the fitting to be rather easy for the combination
$\{n=500,a=1,\lambda=5,\gamma=0.1\}$ and rather difficult for
$\{n=100,a=0.2,\lambda=2,\gamma=0.9\}$.
Simulations and computations
For each combination of the parameters, we simulated $S=100$
random graphs according to the corresponding mixture model. We
fitted the parameters using the algorithm described in
Section 4. To solve the identifiability problem of the
classes, we systematically ordered them in descending estimated
proportion order: $\widehat{\alpha}_{1}\geq\widehat{\alpha}_{2}\geq\widehat{\alpha}_{3}$. For each parameter, we calculated the estimated
Root Mean Squared Error (RMSE):
$$\displaystyle\mathit{RMSE}(\widehat{\alpha}_{p})$$
$$\displaystyle=$$
$$\displaystyle\sqrt{\frac{1}{S}\sum_{s=1}^{S}\bigl{(}\widehat{\alpha_{p}}^{(s)}%
-\alpha_{p}\bigr{)}^{2}},$$
$$\displaystyle\mathit{RMSE}(\widehat{\lambda}_{pq})$$
$$\displaystyle=$$
$$\displaystyle\sqrt{\frac{1}{S}\sum_{s=1}^{S}\bigl{(}\widehat{\lambda_{pq}}^{(s%
)}-\lambda_{pq}\bigr{)}^{2}},$$
where the superscript $(s)$ labels the estimates obtained in
simulation $s$. We also calculated the mean posterior entropy
$$H=\frac{1}{S}\sum_{s}\biggl{(}-\sum_{i}\sum_{q}\tau_{iq}^{(s)}\ln\tau_{iq}^{(s%
)}\biggr{)},$$
which gives us the degree of uncertainty of the classification.
Results
Figure 1 (resp. 2) gives
the RMSE for the proportion $\alpha_{q}$ (resp. connection intensities
$\lambda_{pq}$). As expected, the $\mathit{RMSE}$ is lower when $n$ is
larger. The parameters affecting the $\mathit{RMSE}$ are mainly $a$ and
$\gamma$, whereas $\lambda$ has nearly no effect. The departures
observed for $\alpha_{1}$ and $\alpha_{3}$ in the balanced case ($a=1.0$) are due to the systematic reordering of the proportions.
Since the graph is undirected, $\lambda_{pq}=\lambda_{qp}$, so
only nonredundant parameters are considered in
Figure 2. The overall quality of the
estimates is satisfying, especially for the diagonal terms
$\lambda_{qq}$. The within intensity parameter of the smallest class
$\lambda_{33}$ is the most difficult to estimate. The worst case
corresponds to a small graph ($n=100$) with very unbalanced classes
($a=0.2$) for parameter $\lambda_{12}$. In this case, the
algorithm is unable to distinguish the two larger classes (1 and 2),
so that the estimates extra-diagonal term $\widehat{\lambda}_{12}$
is close to the diagonal ones $\widehat{\lambda}_{11}$ and
$\widehat{\lambda}_{22}$, whereas its true value is up to ten times
smaller.
Figure 3 gives the mean entropy. Not
surprisingly, the most influential parameter is $\gamma$: when $\gamma$
is close to 1, the classes are almost indistinguishable. For small
graphs ($n=100$), the mean intensity $\lambda$ has almost no effect.
Because of the identifiability problem already mentioned, we did not
consider the classification error rate.
5.2 Model selection
We considered a undirected graph of size $n=50,\break 100,500$ and
$1000$ with $Q^{\star}=3$ classes. We considered the combination $\{a=0.5,\lambda=2,\gamma=0.5\}$ which turned out to be a medium
case (see Section 5.1) and computed ICL for
$Q$ ranging from $1$ to $10$ (from $1$ to $5$ for $n=1000$) before
selecting the $Q$ maximizing ICL. We repeated this for $S=100$
simulations.
Figure 4 gives ICL as a function of $Q$, while
Table 1 returns the frequency with which each
$Q$ is selected. As soon as $n$ is larger than $100$, ICL almost
always selects the correct number of classes; for smaller graphs
($n=50$), it tends to underestimate it. The proposed criterion is
thus highly efficient.
6 Uncovering the structure of host–parasite interactions in forest
ecosystems
Here we use mixture models to highlight the factors governing with
whom a species interact in an ecosystem. The factors which may account
for species interactions are introduced as covariates in the mixture
models. The explanatory power of each factor is measured as the
decrease in the number of groups selected. Our study focuses on
host–parasite interactions in forest ecosystems. We address the two
following questions: (1) Is similarity in the parasite assemblages of
two tree species explained by their phylogenetic relatedness rather
than by the degree of overlap of their distributional range? (2) Is
similarity in the host range of two parasitic fungal species explained
by their phylogenetic relatedness rather than their common nutritional
strategy? The explanatory power of phylogenetic relatedness is
subsequently called phylogenetic signal, as in the ecological
literature [Rezende et al. (2007); Vacher, Piou and
Desprez-Loustau (2008)].
6.1 Data
Host–parasite interaction records
We considered two undirected, valued networks having parasitic fungal
species ($n=154$) and tree species ($n=51$) as nodes, respectively. Edges
strength was defined as the number of shared host species and the
number of
shared parasitic species, respectively [Mariadassou, Robin and Vacher (2010)].
The methods used for collecting data on tree–fungus interactions are
fully described in Vacher, Piou and
Desprez-Loustau (2008). Fungal species names were checked
since then in the Index Fungorum database
(www.indexfungorum.
org): 17 names
were updated, yielding to 3
new species synonymies. The fusion of synonym species accounts for the
lower number of fungal species in the present study than in the
original publication.
Phylogenetic relatedness between species
In order to verify the existence of a phylogenetic signal in the
parasite assemblages of tree species, we estimated genetic distances
between all pairs of tree species. The maximally resolved seed plant
tree of the software Phylomatic2 [Webb and Donoghue (2005)] was used to produce a
phylogenetic tree for the 51 tree species included in our study. Then,
pairwise genetic distances (in million years) were extracted by using
the cophenetic.phylo function of the R ape package [Paradis, Claude and Strimmer (2004)].
Because the phylogenetic tree was loosely resolved for gymnosperms, we
also used taxonomic distances to estimate phylogenetic relatedness
between tree species.
Since all tree species included in the study belong to the phylum
Streptophyta, we used the finer taxonomic ranks of class, order,
family and genus to calculate pairwise taxonomic distances. Based on
the NCBI Taxonomy Browser
(www.ncbi.nlm.nih.gov/Taxonomy/),
we found that the
species are evenly distributed into two taxonomic classes
(Magnoliophyta and Conipherophyta) and further subdivided in 8 orders,
13 families and 26 genera. Following Poulin (2005), we considered that
the taxonomic distance is equal to $0$ if species are the same, $1$ if
they belong to the same genus, $2$ to the same family, $3$ to the same
order, $4$ to the same taxonomic class and $5$ if their only common
point lies in belonging to the phylum Streptophyta.
In order to investigate the existence of a phylogenetic signal in the
host range of parasitic fungal species, we estimated taxonomic
distances between all pairs of fungal species. Pairwise genetic
distances could not be calculated because genetic data were not
available for all the species. Since the 153 fungal species at hand
span a wider portion of the tree of life than the tree species, we had
to use the higher order rank of kingdom. The taxonomic distance for
fungal species thus ranges from $0$ to $6$ (kingdom level) when
compared to $0$ to $5$ for trees. The taxonomy was retrieved from
Index Fungorum (www.indexfungorum.org). All fungal species
included in the study belong to the Fungi kingdom, are divided in two
phyla (Ascomycota and Basidiomycota) and further subdivided in 9
taxonomic classes, 21 orders, 48 families and 107 genera. When pairs
included a species whose taxonomic is uncertain for a given taxonomic
rank, this rank was skipped and upper ranks were used to estimate
distance.
Other explanatory factors
Other factors than phylogenetic relatedness may account for pairwise
similarities in parasite assemblages between tree species. In
particular, two tree species having overlapping distributional range
are exposed to similar pools of parasitic species and may therefore
share more parasitic species than two tree species with
nonoverlapping distributions [Brandle and Brandl (2006)]. We tested this
hypothesis by calculating the geographical distance between all pairs
of tree species. The geographical distance is the Jaccard distance
[Jaccard (1901)] computed on the profiles of presence/absence in 309
geographical units covering the entire French territory.
In the case of fungal species, other factors may also account for
similarity in host range. Here we investigated whether fungal species
having similar nutritional strategies also have similar host ranges. Fungal
species were classified into ten nutritional strategies based on their
parasitic lifestyle (biotroph or necrotroph) and on the plant organs and
tissues attacked. Five strategies (strict foliar necrotroph parasites,
canker agents, stem decay fungi, obligate biotroph parasites and root decay
fungi) accounted for 87% of the fungal species. We considered that
nutritional distance between two species equals one if the strategies are
the same and 0 otherwise.
6.2 Identification of groups of species sharing similar
interactions
Model
For both networks, we used the mixture model to define groups of tree
species and fungal species having similar interactions We assumed
that, in each network, the edge intensities were Poisson distributed.
For both networks, we considered the PM and PRMH models (see Section
2.2) using pairwise distance between species
(genetic, taxonomic, geographic or nutritional) as a covariate.
PM model: No covariate
In the absence of covariates, the ICL criterion selected 7 groups of
tree species. Two groups of tree species (T2 and T5) were exclusively
composed of species belonging to the Magnoliophyta, whereas three other
groups (T1, T3 and T4) were exclusively composed of species belonging
to the Conipherophyta. The two last groups (T6 and T7) were mixed
(Table 2). According to the mean number of interactions
per species and the parameters estimates of the model
(Table 2), they were composed of tree species having few
parasitic species and sharing few of them with other tree species.
It is noteworthy that group T2 was composed of four species belonging
to the same order (Fagales) and also to the same family (Fagaceae).
Groups T1, T3 and T4 were also composed of species belonging to the
same family (Pinaceae) since the only three coniferous species
belonging to another family were classified in groups T6 and T7. These
results confirm that two plant species with a similar evolutionary
history are likely to share the same set of parasitic species
[Brandle and Brandl (2006), Gilbert and Webb (2007), Vacher, Piou and
Desprez-Loustau (2008)].
PRMH model: Accounting for phylogenetic relatedness
When accounting for taxonomy, ICL selected only $4$ groups of tree
species. The estimated regression coefficient was $\widehat{\beta}=-0.317$,
which means that, for the mean taxonomic distance $\overline{y}=3.82$,
the mean connexion intensity is reduced of $70\%$
($e^{\widehat{\beta}\overline{y}}=0.298$). The cross classification table
(Table 3) shows that the taxonomic distance reduces the
number of class by merging groups T1 and T2 with most of the trees of T4
and T5. T’3 essentially consists of T6, T’1 of T7 and T’2 is made of trees
from T3 completed with leftovers from other classes. Interestingly and
unlike the groups obtained with no covariates, no group has species
belonging exclusively to one or the other of the taxonomic classes
(Magnoliophyta or Conipherophyta): the association between group of trees
and taxonomy was cropped out by the covariate
(Table 4). The same results hold when using the genetic
distance as a covariate instead of the taxonomic distance (results not
shown).
Therefore, the inclusion of taxonomic (or genetic) distance as a
covariate shows that the phylogenetic relatedness between tree species
accounts for a large part of the structure of tree–parasitic fungus
interactions in forest ecosystems, but not for all the structure.
Indeed, even after controlling for the evolutionary history through
the taxonomic (or genetic) distance, ICL still finds $4$ groups of
trees, whereas we would expect only one group if the phylogeny was the
sole source of structure. Below we investigate whether the
distributional overlap between tree species is another source of
structure.
PRMH model: Accounting for distributional overlap
In contrast with the taxonomic and genetic distance, the geographical
distance between species does not reduce the number of groups (not
shown). This result suggests that the current distributional overlap
between tree species does not account for the similarity in their
parasite assemblages. This result is opposite to the conventional
wisdom in the field of community ecology, which favors ecological
processes, taking place over short time scale, over evolutionary
processes, taking place over longer time scales, as the main source of
biotic interaction diversity. Our findings point out that the relative
importance of these processes might be the other way round.
6.3 Factors accounting for the host ranges of parasitic fungal
species
PM model: No covariate
The ICL criterion selected 9 groups of parasitic fungal species. The
estimates intensities $\widehat{\lambda}_{q\ell}$ range from almost zero
($1.4\times 10^{-3}$) to $12.1$, while the group proportions
$\widehat{\alpha}_{q}$ range from $1.3\%$ to $40.2\%$ (Table 7).
PRMH model: Accounting for phylogenetic relatedness
Accounting for taxonomic distance does not reduce the number of groups
(not shown), indicating a lack of phylogenetic signal in the host
range of fungal species. These results parallel those obtained with
another clustering approach [Newman (2004)] for the same tree–fungus
network [Vacher, Piou and
Desprez-Loustau (2008)]. They are congruent with the results obtained
for other bipartite networks since asymmetries in the phylogenetic
signal have been found in numerous plant–animal mutualistic networks
[Rezende et al. (2007)] and in a host–parasite network between leaf-miner moths
and parasitoid insects [Ives and Godfray (2006)]. In the latter case, the authors
also observed a lack of signal through the parasite phylogeny. In the
case of the tree–fungus network, we proposed that the very early
divergence of the major fungal phyla may account for the asymmetric
influence of past evolutionary history [Vacher, Piou and
Desprez-Loustau (2008)]: the lack of
signal through the fungal phylogeny may be the result of parasitic
fungal species splitting into two groups when the Conipherophyta and
the Magnoliophyta diverged (both groups containing Ascomycota and
Basidiomycota species) and the subsequent coevolution of each set of
fungal species with its plant phylum. Stronger selection pressures on
parasitic species than on host species might also account for the
asymmetry of the signal [Bersier and Kehrli (2008); Rossberg et al. (2006)].
PRMH model: Accounting for nutritional strategies
Fungal
Correlation analysis showed an association between the 9 groups
selected with the PM model and the nutritional type.
In particular, two groups of fungal species (F2 and F3, see Appendix
.8) contained a high proportion of root decay fungi
(100% and 75%, respectively).
However, taking the nutritional strategy as a covariate does not
reduce the number of groups, indicating the lack of ‘nutritional
signal’ in the host range of parasitic fungal species.
6.4 Goodness of fit
Since no covariate decreases the number of mixture components in the
fungus interaction network, we assessed goodness of fit only for the
tree interaction network. The goodness is assessed in two ways: in
terms of likelihood with the ICL criterion and in terms of predictive
power for the strength of an interaction. The ICL criterion is
$-2876.6$ for the base model with no class. It jumps to $-1565.6$
($\Delta\mathit{ICL}=1212.8$) when allowing a mixture structure (with $7$
classes). It jumps again to $-1449.6$ ($\Delta\mathit{ICL}=116$) when adding
the taxonomic distance as a covariate in the model (with $4$ classes).
Interestingly, adding a covariate to the $4$ class mixture model
provides a gain in goodness of fit twice as big as the gain of adding
three additional classes ($\Delta\mathit{ICL}=214.2$ against $98.2$). But
adding a covariate only requires one additional parameter ($\beta$),
against $21$ for the three additional classes.
We also assessed goodness of fit in terms of predictive power. For the PRMH
model with 4 classes and taxonomic distance as a covariate, we can predict
both the weighted degree $K_{i}=\sum_{j\neq i}X_{ij}$ of node $i$ as
$\widehat{K}_{i}=\sum_{j}\sum_{q,\ell}\tau_{iq}\tau_{jl}\lambda_{ql}e^{\beta^{%
\intercal}\mathbf{y}_{ij}}$ and the value $X_{ij}$
of a
single edge fungal as $\widehat{X}_{ij}=\sum_{q,\ell}\tau_{iq}\tau_{j\ell}\lambda_{ql}e^{\beta^{%
\intercal}\mathbf{y}_{ij}}$. The prediction of
$K_{i}$ using
$\widehat{K}_{i}$ is pretty accurate (Figure 5 left,
$R^{2}=0.94$). The prediction of $X_{ij}$ using $\widehat{X}_{ij}$ is less
accurate, but the confidence region is still pretty good
(Figure 5 right, $R^{2}=0.56$).
6.5 Conclusion
The structure of host–parasite interactions in forest ecosystems is a
complex one. Some tree species share more parasites than others and
this variability is well captured by a mixture model. However and as
shown in Table 5, the naive mixture model deceptively
captures part of the variability readily explained by other factors,
such as the phylogenetic relatedness (measured either by taxonomic or
genetic distance) and artificially increases the number of groups in
the mixture. Accounting for relevant factors decreases the number of
groups selected.
Using group reduction as a yardstick (Table 5), we
conclude that similarity in the parasite assemblages of tree species
is explained by their phylogenetic relatedness rather than their
distributional overlap, indicating the importance of evolutionary
processes for explaining the current patterns of inter-specific
interactions. Our study is however inconclusive on the relative
contribution of phylogenetic relatedness and nutritional strategy to
the similarity in the host ranges of parasitic fungal species
parasites of two parasitic fungus (Table 8 in Appendix
.8). In either case, since the PRMH model still finds
$4$ (resp. $9$) classes for the tree species (resp. fungal species)
interaction network, a significant fraction of the variability remains
unexplained by our predictors.
Appendix
.6 Other mixture models
We examine here some other classical distributions which can be used
in our framework.
Bernoulli.
In some situations such as co-authorship or
social networks, the only available information is the presence or
absence of the edge. $X_{ij}$ is then supposed to be Bernoulli
distributed:
$$X_{ij}|i\in q,j\in\ell\sim\mathcaligr{B}(\pi_{q\ell}).$$
It is equivalent to the stochastic block model of Nowicki and Snijders (2001) or
Daudin, Picard and Robin (2008).
Multinomial.
In a social network, $X_{ij}$ may specify the
nature of the relationship: colleague, family, friend, etc.
The $X_{ij}$’s can then be modeled by multinomial variables:
$$X_{ij}|i\in q,j\in\ell\sim\mathcaligr{M}(1;\mathbf{p}_{q\ell}).$$
The parameter $\theta_{q\ell}$ to estimate is the vector of
probability $\mathbf{p}_{q\ell}=(p_{q\ell}^{1},\dots,p_{q\ell}^{m})$, $m$ being the number of possible labels.
In directed random graphs, this setting allows to account for some
dependency between symmetric edges $X_{ij}$ and $X_{ji}$. We only need
to consider the equivalent undirected graphs where edge
$(i,j)$ is labeled with the couple $(X_{ij},X_{ji})$. $m=4$
different labels can the be observed: $(0,0)$ if no edge exists,
$(1,0)$ for $i\rightarrow j$, $(1,1)$ for $i\leftarrow j$ and
$(1,1)$ for $i\leftrightarrow j$.
Gaussian.
Traffic networks describe the intensity of
the traffic between nodes. The airport network is a typical example
where the edges are valued according to the number of passengers
traveling from airport $i$ to airport $j$. The intensity $X_{ij}$
of the traffic can be assumed to be Gaussian:
$$X_{ij}|i\in q,j\in\ell\sim\mathcaligr{N}(\mu_{q\ell},\sigma^{2}_{q\ell}),%
\qquad\theta_{q\ell}=(\mu_{q\ell},\sigma^{2}_{q\ell}).$$
Bivariate Gaussian.
The correlation between symmetric edges
$X_{ij}$ and $X_{ji}$ can be accounted for, considering the
undirected valued graph where edge $(i,j)$ is valued by $(X_{ij},X_{ji})$, which is assumed to be Gaussian. Denoting $\mathbf{X}_{ij}=[X_{ij}X_{ji}]^{\prime}$,
$$\mathbf{X}_{ij}|i\in q,j\in\ell\sim\mathcaligr{N}(\bolds\mu_{q\ell},\bolds%
\Sigma_{q\ell}),\qquad\theta_{q\ell}=(\bolds\mu_{q\ell},\bolds\Sigma_{q\ell}).$$
Linear regression.
When covariates are available, the
linear model, either Gaussian for real valued edges or generalized for
integer valued (e.g., Poisson or Bernoulli) allows to include them. For
example, for Gaussian valued edges, denoting $\mathbf{y}_{ij}$ the $p\times 1$
vector of covariates describing edge $(i,j)$, we set
$$X_{ij}|i\in q,j\in\ell\sim\mathcaligr{N}({\bf\beta}_{q\ell}^{\intercal}.%
\mathbf{y}_{ij},\sigma_{q\ell}^{2}).$$
Simple linear regression.
A case of specific interest for
plant ecology is the simple linear homoskedastic regression with group
specific intercept $a_{q\ell}$ but constant regression coefficient
$b$. It is
particularly useful when controlling for the effect of geography, which
is assumed to be the same for all groups of plants. We then set
$$X_{ij}|i\in q,j\in\ell\sim\mathcaligr{N}(a_{q\ell}+by_{ij},\sigma^{2}).$$
The model can again be extended to Poisson or Bernoulli valued edges
using adequate link function.
.7 Parameter estimates for other distributions
Table 6 gives the parameter estimates for the
model listed in Section .6. The estimates of the
mean parameter for Gaussian ($\mu_{q\ell}$) distributions are the same
as the estimate of the probability $\pi_{q\ell}$ in the Bernoulli
case. The results displayed in this table are all straightforward.
Note that all estimates are weighted versions of the intuitive ones.
.8 Parameter estimates for the fungus interaction network
Acknowledgments
We thank the Département Santé des Forêts (DSF) of the French
Ministère de l’Agriculture et de la Pêche for allowing us to use
their database. We thank Dominique Piou and Marie-Laure
Desprez-Loustau for checking the data and for helpful comments on the
results.
{supplement}
[id=suppA]
\stitleInteraction network between tree and fungal species
\slink[doi]10.1214/07-AOAS361SUPP
\slink[url]http://lib.stat.cmu.edu/aoas/361/supplement.csv
\sdatatype.csv
\sdescription
This file contains:
•
The adjacency matrix of interactions between tree and fungal
species.
•
The list of the tree species.
•
The list of the fungal species.
•
The matrix of genetic distances between tree species.
•
The matrix of geographical distances between tree species.
•
The matrix of taxonomic distances between fungal species.
•
The matrix of nutritional type of the fungal species.
References
Airoldi and Carley (2005)
Airoldi, E. M. and Carley, K. M. (2005).
Sampling algorithms for pure network topologies.
ACM KDD Explorations 7 13–22.
Airoldi et al. (2008)
Airoldi, E. M., Blei, D. M., Fienberg,
S. E. and Xing, E. P. (2008).
Mixed membership stochastic blockmodels.
J. Mach. Learn. Res. 9 1981–2014.
Albert and Barabási (2002)
Albert, R. and Barabási, A. L. (2002).
Statistical mechanics of complex networks.
Rev. Modern Phys. 74 47–97.
\MR1895096
Attias (2000)
Attias, H. (2000).
A variational Bayesian framework for graphical models.
In Advances in Neural Information Processing Systems 12
209–215. MIT Press, Cambridge.
Barabási and Albert (1999)
Barabási, A. L. and Albert, R. (1999).
Emergence of scaling in random networks.
Science 286 509–512.
\MR2091634
Beal and Ghahramani (2003)
Beal, M. J. and Ghahramani, Z. (2003).
The variational Bayesian EM algorithm for incomplete data: With
application to scoring graphical model structures.
In Bayesian Statistics 7 (J. M. Bernardo et al., eds.) 543–552. Oxford Univ.
Press, Oxford.
\MR2003189
Bersier and Kehrli (2008)
Bersier, L. F. and Kehrli, P. (2008).
The signature of phylogenetic constraints on food-web structure.
Ecol. Complex. 5 132–139.
Biernacki, Celeux and Govaert (2000)
Biernacki, C., Celeux, G. and Govaert, G. (2000).
Assessing a mixture model for clustering with the integrated
completed likelihood.
IEEE Trans. Pattern Anal. Machine Intel. 22
719–725.
Blomberg and Garland (2002)
Blomberg, S. P. and Garland, T. J. (2002).
Tempo and mode in evolution: Phylogenetic inertia, adaptation and
comparative methods.
J. Evol. Biol. 15 899–910.
Brandle and Brandl (2006)
Brandle, M. and Brandl, R. (2006).
Is the composition of phytophagous insects and parasitic fungi among
trees predictable?
Oikos 113 296–304.
Burnham and Anderson (1998)
Burnham, K. P. and Anderson, R. A. (1998).
Model Selection and Inference: A Practical
Information-Theoretic Approach.
Wiley, New York.
Cattin et al. (2004)
Cattin, M. F., Bersier, L. F., Banasek-Richter, C. C., Baltensperger, R. and Gabriel, J. P. (2004).
Phylogenetic constraints and adaptation explain food-web structure.
Nature 427 835–839.
Daudin, Picard and Robin (2008)
Daudin, J.-J., Picard, F. and Robin, S. (2008).
A mixture model for random graphs.
Statist. Comput. 18 173–183.
\MR2390817
Dempster, Laird and Rubin (1977)
Dempster, A. P., Laird, N. M. and Rubin,
D. B. (1977).
Maximum likelihood from incomplete data via the EM algorithm.
J. Roy. Statist. Soc. B 39 1–38.
\MR0501537
Erdös and Rényi (1959)
Erdös, P. and Rényi, A. (1959).
On random graphs, i.
Publ. Math. 6 290–297.
\MR0120167
Fienberg and Wasserman (1981)
Fienberg, S. E. and Wasserman, S. (1981).
Categorical data
analysis of single sociometric relations. In
Sociological Methodology 1981 156–192.
Jossey-Bass, San Francisco.
Fienberg, Meyer and Wasserman (1985)
Fienberg, S. E., Meyer, M. M. and Wasserman, S. S. (1985).
Statistical analysis of multiple sociometric relations.
J. Amer. Statist. Assoc. 80 51–67.
Getoor and Diehl (2004)
Getoor, L. and Diehl, C. P. (2004).
Link mining: A survey.
SIGKDD Explor. 7 3–12.
Gilbert and Webb (2007)
Gilbert, G. S. and Webb, C. O. (2007).
Phylogenetic signal in plant pathogen-host range.
Proc. Natl. Acad. Sci. USA 104 4979–4983.
Girvan and Newman (2002)
Girvan, M. and Newman, M. E. J. (2002).
Community structure in social and biological networks.
Proc. Natl. Acad. Sci. USA 99
7821–7826.
\MR1908073
Govaert and Nadif (2005)
Govaert, G. and Nadif, M. (2005).
An EM algorithm for the block mixture model.
IEEE Trans. Pattern Anal. Machine Intel. 27
643–647.
Hofman and Wiggins (2008)
Hofman, J. M. and Wiggins, C. H. (2008).
A Bayesian approach to network modularity.
Phys. Rev. Lett. 100 258701.
Holland and Leinhardt (1981)
Holland, P. and Leinhardt, S. (1981).
An exponential family of probability distributions for directed
graphs.
J. Amer. Statist. Assoc. 76 33–50.
\MR0608176
Ives and Godfray (2006)
Ives, A. R. and Godfray, H. C. J. (2006).
Phylogenetic analysis of trophic associations.
Am. Nat. 16 E1–E14.
Jaakkola (2000)
Jaakkola, T. (2000).
Tutorial on variational approximation methods. In Advanced
Mean Field Methods: Theory and
Practice.
MIT Press, Cambridge.
Jaccard (1901)
Jaccard, P. (1901).
Tude comparative de la distribution florale dans une portion des
alpes et des jura.
Bullet. Soc. Vaud. Sci. Natur.
37 547–579.
Jordan et al. (1999)
Jordan, M. I., Ghahramani, Z., Jaakkola,
T. and Saul, L. K. (1999).
An introduction to variational methods for graphical models.
Mach. Learn. 37 183–233.
Kemp, Griffiths and Tenenbaum (2004)
Kemp, C., Griffiths, T. H. and Tenenbaum,
J. B. (2004).
Discovering latent classes in relational data.
Technical report, MIT Computer Science and Artificial Intelligence
Laboratory.
Leisink and Kappen (2001)
Leisink, M. A. R. and Kappen, H. J. (2001).
A tighter bound for graphical models.
Neural Comput. 13 2149–2171.
Lorrain and White (1971)
Lorrain, F. and White, H. C. (1971).
Structural equivalence of individuals in social networks.
J. Math. Soc. 1 49–80.
Mariadassou (2006)
Mariadassou, M. (2006).
Estimation paramétrique dans le modèle ERMG.
Master’s thesis, Univ. Paris XI/Ecole Nationale
Supèrieure.
Mariadassou, Robin and Vacher (2010)
Mariadassou, M., Robin, S. and Vacher, C. (2010).
Supplement to “Uncovering latent structure in valued graphs:
A variational approach.”
DOI: 10.1214/07-AOAS361SUPP.
McGrory and
Titterington (2007)
McGrory, C. A. and Titterington, D. M. (2007).
Variational approximations in Bayesian model selection for finite
mixture distributions.
Comput. Statist. Data Anal. 51
5352–5367.
\MR2370876
McLahan and Peel (2000)
McLahan, G. and Peel, D. (2000).
Finite Mixture Models.
Wiley, New York.
\MR1789474
Newman (2004)
Newman, M. E. J. (2004).
Fast algorithm for detecting community structure in networks.
Phys. Rev. E 69 066133.
Newman, Watts and Strogatz (2002)
Newman, M. E. J., Watts, D. J. and Strogatz, S. H. (2002).
Random graph models of social networks.
Proc. Natl. Acad. Sci. USA 99 2566–2572.
Nowicki and Snijders (2001)
Nowicki, K. and Snijders, T. A. B. (2001).
Estimation and prediction for stochastic block-structures.
J. Amer. Statist. Assoc. 96 1077–1087.
\MR1947255
Paradis, Claude and Strimmer (2004)
Paradis, E., Claude, J. and Strimmer, K. (2004).
Ape: Analyses of phylogenetics and evolution in R language.
Bioinformatics 20 289–290.
Pattison and Robins (2007)
Pattison, P. E. and Robins, G. L. (2007).
Probabilistic network theory.
In Handbook of Probability Theory with Applications.
Sage, Thousand Oaks, CA.
Picard et al. (2007)
Picard, F., Daudin, J.-J., Miele, V.,
Mariadassou, M. and Robin, S. (2007).
A novel framework for random graph models with heterogeneous
connectivity structure.
Submitted.
Poulin (2005)
Poulin, R. (2005).
Relative infection levels and taxonomic distances among the host
species used by a parasite: Insights into parasite specialization.
Parasitology 130 109–115.
Rezende et al. (2007)
Rezende, E. L., Lavabre, J. E., Guimaraes,
P. R., Jr., Jordano, P. and Bascompte, J. (2007).
Non-random coextinctions in phylogenetically structured mutualistic
networks.
Nature 448 925–928.
Ricklefs and Miller (2000)
Ricklefs, R. E. and Miller, G. L. (2000). Community
ecology. In
Ecology, 4th ed.
Freeman, San Francisco, CA.
Rossberg et al. (2006)
Rossberg, A. G., Ishii, R., Amemiya, T.
and Itoh, K. (2006).
Food webs: Experts consuming families of experts.
J. Theoret. Biol. 241 552–563.
\MR2254907
Tykiakanis, Tscharntke and
Lewis (2007)
Tykiakanis, J. M., Tscharntke, T. and Lewis, O. T. (2007).
Habitat modification alters the structure of tropical host-parasitoid
food webs.
Nature 51 202–205.
Vacher, Piou and
Desprez-Loustau (2008)
Vacher, C., Piou, D. and Desprez-Loustau,
M.-L. (2008).
Architecture of an antagonistic tree/fungus network: The asymmetric
influence of past evolutionary history.
PLoS ONE 3 1740.
von Luxburg, Belkin and
Bousquet (2008)
von Luxburg, U., Belkin, M. and Bousquet,
O. (2008).
Consistency of spectral clustering.
Ann. Statist. 36 555–586.
\MR2396807
Webb and Donoghue (2005)
Webb, C. O. and Donoghue, M. J. (2005).
Phylomatic: Tree assembly for applied phylogenetics.
Mol. Ecol. Notes 5 181–183.
Winn, Bishop and Jaakkola (2005)
Winn, J., Bishop, C. M. and Jaakkola, T. (2005).
Variational message passing.
J. Mach. Learn. Res. 6 661–694.
\MR2249835
Xing, Jordan and Russell (2003)
Xing, E., Jordan, M. and Russell, S. (2003).
A generalized mean field algorithm for variational inference in
exponential families.
In Proceedings of the 19th Annual Conference on Uncertainty in
Artificial Intelligence (UAI-03) 583–591. Morgan Kaufmann,
San Francisco, CA.
Yedidia, Freeman and Weiss (2005)
Yedidia, J. S., Freeman, W. T. and Weiss,
Y. (2005).
Constructing free-energy approximations and generalized belief
propagation algorithms.
IEEE Inform. Theory 15 2282–2312.
\MR2246363 |
A modular-invariant modified Weierstrass sigma-function as a
building block for lowest-Landau-level wavefunctions on the torus
F. D. M. Haldane
[email protected]
Department of Physics, Princeton
University, Princeton NJ 08544-0708, USA
(June 3, 2018, revised June 20, 2018)
Abstract
A “modified” variant of the Weierstrass sigma, zeta, and elliptic
functions is proposed whereby the zeta function is redefined by $\zeta(z)$ $\mapsto$
$\widetilde{\zeta}(z)$ $\equiv$ $\zeta(z)-\gamma_{2}z$, where $\gamma_{2}$ is
a lattice invariant
related to the almost-holomorphic modular invariant of the
quasi-modular-invariant weight-2 Eisenstein series. If $\omega_{i}$ is
a primitive half-period, $\widetilde{\zeta}(\omega_{i})$ = $\pi\omega_{i}^{*}/A$,
where $A$ is the area of the primitive cell of the lattice. The
quasiperiodicity of the modified sigma function is much simpler than
that of the original, and it becomes the building-block for the
modular-invariant formulation of lowest-Landau-level wavefunctions on
the torus. It is suggested that the “modified” sigma function is
more natural than the original Weierstrass form, which was formulated
before quasi-modular forms were understood. For the high-symmetry
(square and hexagonal) lattices, the modified and original sigma
functions coincide.
In 1862, Weierstrassweier , as part of his work on elliptic
functions,
introduced an odd meromorphic “zeta function”
$\zeta(z;\Lambda)$ which has simple poles with residue 1 on a lattice $\Lambda$ = $\{L\}$ =
$\{2m\omega_{1}+2n\omega_{2}\}$ in the complex plane, where $m$ and
$n$ are integers: $m,n$ $\in$ $\mathbb{Z}$.
Using the metric $d(z_{1},z_{2})$ = $|z_{1}-z_{2}|$, the lattice
has a primitive cell of area
$$A(\Lambda)=2|\omega_{1}^{*}\omega_{2}-\omega^{*}_{2}\omega_{1}|>0.$$
(1)
Here $\omega_{i}$ will mean any primitive half-period of the lattice, so
$2\omega_{i}$ $\in$ $\Lambda$, but
$2\omega_{i}/p$ $\not\in\Lambda$ for any integer $p>1$.
A pair of primitive half-periods $(\omega_{1},\omega_{2})$ is a basis of
$\Lambda$ if
$\omega_{1}+p\omega_{2}$ and $\omega_{2}+p\omega_{1}$ are primitive for
all integers $p$. Once a basis is chosen, it is useful to define
$\omega_{3}$ =$-(\omega_{1}+\omega_{2})$, which is also a primitive half-period.
The Weierstrass zeta function is defined by the convergent sum
$$\zeta(z;\Lambda)=\frac{1}{z}+{\sum_{L\neq 0}}^{\prime}\left(\frac{1}{z-L}+%
\frac{1}{L}+\frac{z}{L^{2}}\right)=\frac{1}{z}+{\sum_{L\neq 0}}^{\prime}\frac{%
z^{3}}{L^{2}(z^{2}-L^{2})},$$
(2)
where the primed sum is over all $L\in\Lambda$ except $L=0$.
Near $z$ = 0, it has a Laurent expansion
$$\zeta(z;\Lambda)=\frac{1}{z}-\sum_{k=2}^{\infty}\gamma_{2k}(\Lambda)z^{2k-1},$$
(3)
where
$\gamma_{2k}(\Lambda)$ are lattice invariants
$$\gamma_{2k}(\Lambda)={\sum_{L\neq 0}}^{\prime}\frac{1}{L^{2k}},\quad k\geq 2,$$
(4)
which are related to the Eisenstein series $G_{2k}(\omega_{2}/\omega_{1})$
= $(2\omega_{1})^{2}\gamma_{2k}(\Lambda)$.
The Weierstrass elliptic function $\wp(z;\Lambda)$ is (minus) the
derivative of the zeta function, and is an even doubly-periodic meromorphic
function of $z$ defined by
$$\wp(z;\Lambda)=-\zeta^{\prime}(z;\Lambda)\equiv-\frac{\partial}{\partial z}%
\zeta(z;\Lambda)=\frac{1}{z^{2}}+{\sum_{L\neq 0}}^{\prime}\left(\frac{1}{(z-L)%
^{2}}-\frac{1}{L^{2}}\right).$$
(5)
The elliptic function seem to have been Weierstrass’ main interest,
but the emphasis here will be on an application of the quasiperiodic
sigma function, defined in terms of the zeta function by
$$\sigma(z;\Lambda)=\lim_{\epsilon\rightarrow 0}\epsilon\exp\left(\int_{\epsilon%
}^{z}dz^{\prime}\zeta(z^{\prime};\Lambda)\right),$$
(6)
where the integral is independent of the path because the poles of the
zeta function have unit residue. The sigma function is an odd holomorphic
function with only simple zeroes which are at the lattice points
$z$ $\in$ $\Lambda$. Then
$$\zeta(z;\Lambda)=\frac{\sigma^{\prime}(z;\Lambda)}{\sigma(z;\Lambda)},\quad%
\sigma^{\prime}(0;\Lambda)=1.$$
(7)
A curious feature of the Weierstrass definitions
is the absence of a term
$O(z)$ in the Laurent expansion (3) for $\zeta(z;\Lambda)$.
Weierstrass presumably made
this choice because the definition (4) of $\gamma_{2k}(\Lambda)$ is ill-defined for $2k$
= 2, and he
was working long before Ramanujan investigated quasi-modular forms.
However, based on Ramanujan’s workram on such forms, there is a
“natural” definition of $\gamma_{2}(\Lambda)$ given by “Eisenstein summation”
$$\Gamma_{2}(\omega_{1},\Lambda)\equiv\sum_{m\neq 0}\frac{1}{(2m\omega_{1})^{2}}%
+\sum_{n\neq 0}\left(\sum_{m}\frac{1}{(2m\omega_{1}+2n\omega_{2})^{2}}\right)=%
\gamma_{2}(\Lambda)+\frac{\pi}{A(\Lambda)}\frac{\omega_{1}^{*}}{\omega_{1}},$$
(8)
where a sum over $n$ is interpreted as the limit as $N$
$\rightarrow$ $\infty$ of a sum over $n$ in the interval $[-N,N]$.
$\Gamma_{2}(\omega_{i},\Lambda)$ is related to the weight-2
quasi-modular-invariant
of the $2k$ = 2 Eisenstein series, and
$\gamma_{2}(\Lambda)$ is a lattice invariant related to the corresponding
“almost-holomorphic modular form”.
In fact knowledge of the invariant $\gamma_{2}(\Lambda)$
had already been available since 1847, fifteen years before
Weierstrass presented his work, in the
published (but apparently ignored) work of
Eisensteineis who died in 1851. This work was later briefly brought to
attention in 1891 in a lecture by Kronecker, who appears to have
belatedly recognized its
importance, but who unfortunately died just fifteen days after the lecture,
leaving extensive notes that remained unpublished, and which were believed lost until
recently rediscoveredems .
Some of Eisenstein’s results remained overlooked and forgotten
until they were reviewed in detail and publicized in a 1976 book by Weilweil .
Then defining “modified” Weierstrass functions by including the
$2k$ = $2$ term in the Laurent-series definition of
$\zeta(z;\Lambda)$ has the effect
$$\displaystyle\wp(z,\Lambda)$$
$$\displaystyle\mapsto\widetilde{\wp}(z,\Lambda)=\wp(z,\Lambda)+\gamma_{2}(%
\Lambda),$$
(9)
$$\displaystyle\zeta(z,\Lambda)$$
$$\displaystyle\mapsto\widetilde{\zeta}(z,\Lambda)=\zeta(z,\Lambda)-\gamma_{2}(%
\Lambda)z,$$
(10)
$$\displaystyle\sigma(z,\Lambda)$$
$$\displaystyle\mapsto\widetilde{\sigma}(z,\Lambda)=e^{-\frac{1}{2}\gamma_{2}(%
\Lambda)z^{2}}\sigma(z,\Lambda).$$
(11)
Note that $\gamma_{2}$ vanishes for both the hexagonal lattice ($\gamma_{4}$ = 0),
and the square lattice ($\gamma_{6}$ = 0), so in these high-symmetry
lattices, the “modified” and “original” Weierstrass functions
coincide. Specification of the pair of complex invariants $(\gamma_{4},\gamma_{6})$
(so long as they do not both vanish) uniquely parametrizes the lattice $\Lambda$.
For ease of notation, the dependence of these functions on the lattice
$\Lambda$ will now be left implicit, unless needed.
An important property of both the original and the modified Weierstrass
functions is that they depend only the lattice $\Lambda$, and do not
depend on a particular choice of basis $(\omega_{1},\omega_{2})$ (or on a
choice of orientation of the basis). The group $SL(2,\mathbb{Z})$ of
(orientation-preserving) changes of basis is known as the modular group,
and (with a slight abuse of terminology) the basis-independence of
the Weierstrass functions may be called “modular-invariance”.
The “modified” forms of the Weierstrass functions were
in fact already present in Eisenstein’s
workeis , as reviewed by Weilweil . In the
notation
used by Weil, $\widetilde{\zeta}(z)$ is “$E_{1}(z)$”, $\widetilde{\sigma}(z)$ is
“$\varphi(z)$”, and $\widetilde{\wp}(z)$ is “$E_{2}(z)$”.
A key result also found in Eisenstein’s workeis is that the
non-meromorphic function
$$\widehat{\zeta}(z,z^{*})=\widetilde{\zeta}(z)-\frac{\pi z^{*}}{A}$$
(12)
is a doubly-periodic modular-invariant
function (referred to as “$E_{2}^{*}(z)$” by Weilweil ).
In recent workrecent , there is renewed interest in this
function, which has been called
“Eisenstein’s (periodic) completion” of the
Weierstrass zeta function.
Generically, it has a lattice of simple poles
at $z$ $\in$ $\{L\}$, three lattices of
simple holomorphic zeroes
and two lattices of simple antiholomorphic zeroes.
Each lattice of holomorphic zeroes is invariant under inversion about
a pole, and they
are located at the points
$z$ $\in$ $\{L+\omega_{i}\}$, as a consequence of the relation
$$\zeta(\omega_{i})\equiv\eta_{i}=\gamma_{2}\omega_{i}+\frac{\pi\omega_{i}^{*}}{%
A};\quad\widetilde{\zeta}(\omega_{i})=\frac{\pi\omega_{i}^{*}}{A}.$$
(13)
The lattices of antiholomorphic zeroes are mapped into each other by
the inversion.
The six antiholomorphic zeroes
which are closest to a pole define the corners of a
six-sided unit cell centered on the pole, with holomorphic zeroes at
the centers of the sides. (This is similar to a “Wigner-Seitz” or
Voronoi cell, but direct inspection using the Weierstrass functions
implemented in Mathematica® shows that
the locations of the antiholomorphic zeroes are not in general given
by the Voronoi construction using the metric $d(z,z^{\prime})$ = $|z-z^{\prime}|$.)
In the limit of a simple-rectangular lattice, one pair of sides shrink
to points where two antiholomorphic zeroes combine with a holomorphic
one, leaving a rectangular unit cell with a simple antiholomorphic
zero at its corners.
Eisenstein’s periodic function (12) is a sum of
“holomorphic” (meromorphic) and antiholomorphic parts:
the “modified zeta function” is just the “holomorphic” part of
Eisenstein’s periodic completion of the zeta function.
A consequence of the relation (13) is the well-known property
$$\eta_{1}\omega_{2}-\eta_{2}\omega_{1}=\frac{\pi(\omega_{1}^{*}\omega_{2}-%
\omega_{1}^{*}\omega_{2})}{A}=\pm{\textstyle\frac{1}{2}}i\pi,$$
(14)
which is an orientation-independent version of a standard relation
between the $\eta_{i}$ and $\omega_{i}$. (Note that the formulation presented here
does not require the orientation of the basis to be chosen so
that the imaginary part of $\omega_{2}/\omega_{1}$ (or “$\tau$”) is
positive.) The explicit expression (13) involving
$\gamma_{2}(\Lambda)$
does not seem to be
as widely known in the literature on Weierstrass functions (it does
not seem to appear in the standard reference sources), but is
easily verified by
using a basis
$(\omega_{1},\omega_{2})$, with $\omega_{1}$ = $\omega_{i}$,
and evaluating the lattice sum (2) in the order used
in the definition (8) of $\Gamma_{2}(\omega_{1};\Lambda)$:
$$\eta_{1}=S(\omega_{1},\omega_{2})+\omega_{1}\Gamma_{2}(\omega_{1},\Lambda),$$
(15)
where the sum over $1/L$ is dropped because it vanishes by symmetry,
and
$$\displaystyle S(\omega_{1},\omega_{2})$$
$$\displaystyle=\sum_{n}\left(\sum_{m}\frac{1}{(2m+1)\omega_{1}+2n\omega_{2}}%
\right)=-\frac{\pi}{2\omega_{1}}\sum_{n}\tan(n\pi\omega_{2}/\omega_{1})=0.$$
(16)
The terms in the sum $S$ are odd under $n\mapsto-n$,
so it also vanishes. (This somewhat-heuristic symmetry-based argument can
be confirmed by more rigorous “$\epsilon\rightarrow 0$” treatments.)
The values of $\eta_{i}$ $\equiv$ $\zeta(\omega_{i};\Lambda)$
appear as a complication in
the expressions for the quasiperiodicity of the Weierstrass sigma function.
The equivalent quantity in the
“modified” sigma function has the much simpler form
$$\widetilde{\eta}_{i}\equiv\eta_{i}-\gamma_{2}\omega_{i}=\pi\omega_{i}^{*}/A,$$
(17)
so the modified values of the “strange numbers” $\eta_{i}$ that permeate various
sigma-function identities become the natural primitive “reciprocal half-lattice
vectors”
conjugate to the primitive “half-lattice vectors” $\omega_{i}$.
All the various Weierstrass sigma-function identities involving $\eta_{i}$ remain true
when the Weierstrass functions are replaced by their “modified”
forms and $\eta_{i}$ is replaced by $\pi\omega_{i}^{*}/A$.
In particular, for $q$ = $\exp(i\pi\tau)$ with $\tau$ = $\omega_{2}/\omega_{1}$,
$\sigma(z;\Lambda)$ has a rapidly-converging product representation
$$\sigma(z)=ze^{\eta_{1}z^{2}/2\omega_{1}}\left(\frac{\sin u}{u}\prod_{n=1}^{%
\infty}\frac{\left(q^{2n}+q^{-2n}-2\cos 2u\right)}{(q^{n}-q^{-n})^{2}}\right)%
\equiv ze^{\eta_{1}z^{2}/2\omega_{1}}\left(\frac{\vartheta_{1}(u|\tau)}{u%
\vartheta_{1}^{\prime}(0|\tau)}\right),\quad u=\frac{\pi z}{2\omega_{1}},$$
(18)
where $\vartheta_{1}(u|\tau)$ is the Jacobi theta-function with zeroes
at $\{m\pi+n\pi\tau\}$.
The modified form $\widetilde{\sigma}(z)$ is given by the same formula, with the replacement
$\eta_{1}\mapsto\widetilde{\eta}_{1}$ = $\pi\omega_{1}^{*}/A$.
Note that while the constructive definition of the original Weierstrass
sigma function (as an infinite product)
follows from the lattice sum (2), the constructive
definition of the modified sigma function is directly given by
(18) with
$\eta_{1}\mapsto\pi\omega^{*}_{1}/A$.
The expression (18) is not a constructive definition for the
original Weierstrass sigma function, as it requires a separate evaluation of
$\gamma_{2}(\Lambda)$ in order to provide $\eta_{1}$. Similarly, an
expression based on (2) does not provide a constructive
definition of the modified form, because it would then also require
a separate evaluation
of $\gamma_{2}(\Lambda)$.
The quasiperiodicity of the modified sigma function now has
the more “elegant” form
$$\widetilde{\sigma}(z+L)=\xi(L)e^{(\pi L^{*}/A)(z+\frac{1}{2}L)}\widetilde{%
\sigma}(z),$$
(19)
where $\xi(L)$ is the parity of $L$:
$\xi(L)=1$ if $\frac{1}{2}L\in\Lambda$, and $-1$ otherwise.
In the limit $|z|\rightarrow 0$, the holomorphic sigma functions (both the
modified and original forms) have the property
$$\lim_{z\rightarrow 0}\frac{\sigma(z)}{z}\rightarrow 1.$$
(20)
It is worth pointing out that the non-holomorphic function
$$\mathcal{Z}(z,z^{*})\equiv\widetilde{\sigma}(z)e^{-\frac{1}{2}\pi z^{*}z/A}$$
(21)
has the property that its absolute value $|\mathcal{Z}(z,z^{*})|$ is a function with the
lattice periodicity. This may be regarded as the nearest equivalent
to the linear function $f(z)$ = $z$ that can be made (quasi)periodic,
with
$$\lim_{z\rightarrow 0}\frac{\mathcal{Z}(L+z,L^{*}+z^{*})}{z}\rightarrow\xi(L).$$
(22)
This function is essentially an analog of the sigma function defined by using
(6) with Eisenstein’s periodic completion (12) as a
replacement for the zeta function.
For any primitive
$\omega_{i}$,
there is a symmetric variant of the sigma function defined by
$$\sigma_{i}(z)=\sigma_{i}(-z)=e^{-\eta_{i}z}\sigma(z+\omega_{i})/\sigma(\omega_%
{i}),$$
(23)
so $\sigma_{i}(\omega_{i}+L)$ = 0.
This is normalized so that $\sigma_{i}(0)$ = 1, and
there are just three
distinct functions $\sigma_{i}(z)$, $i$ = 1,2,3,
that can be labeled in a basis-dependent
way using $\omega_{1}$, $\omega_{2}$ and $\omega_{3}$.
The replacement of $\eta_{i}$ by $\widetilde{\eta}_{i}$ and $\sigma(z)$ by
$\widetilde{\sigma}(z)$ now
defines modified symmetric sigma function variants $\widetilde{\sigma}_{i}(z)$.
The symmetric variants have the same quasiperiodicity as the modified
sigma function under even translations:
$$\widetilde{\sigma}_{i}(z+2L)=e^{(\pi L^{*}/A)(2z+L)}\widetilde{\sigma}_{i}(z).$$
(24)
The product (18) provides a very efficient algorithm for
numerical computation of the modified sigma function in
finite-precision floating-point arithmetic, as terms in the product eventually
become numerically equal to unity above some $n$. An optimum
algorithm appears to involve choosing $\omega_{1}$ and $\omega_{2}$ so
that $\pm\omega_{1}$, $\pm\omega_{2}$ and $\pm\omega_{3}$, are on the
boundary of the Voronoi cell of the lattice point at the origin, with
$|\omega_{1}|$ $\leq$
$|\omega_{2}|$ $\leq$ $|\omega_{3}|$. Then use the quasiperiodicity
to reduce $z$ into this Voronoi cell, evaluate the product for
the reduced $z$, and use the quasiperiodicity to migrate the result
back to the original $z$. The accuracy of the migration process can
be enhanced if it is known that an integer multiple of $z$ is a lattice
point $L$.
In the view of this author, the various properties described above justify the view that
the modified functions are the natural ones that
Weierstrass should have defined,
and probably would have used, if knowledge of the theory of quasi-modular forms had been available
to him. Some evidence in favor of this assertion comes from an
application to the physics of electrons in a Landau level on
a two-dimensional plane that is compactified into a torus, where the
wavefunctions have a natural representation in terms of the modified
(as opposed to the original) sigma function.
This problem can be formulated as that of non-relativistic electrons with a dispersion
$\varepsilon(\bm{p})$
= $|\bm{p}|^{2}/2m$, with dynamical momentum $\bm{p}$ = $-i\hbar\bm{\nabla}-e\bm{A}(\bm{r})$, moving on a plane though which a uniform
magnetic flux density passes. This quantizes the kinetic energy to take discrete
values $(\hbar^{2}/m\ell^{2})(n+\frac{1}{2})$, $n=0,1,2,\ldots$, where $2\pi\ell^{2}$ is
the area through which one London quantum $\Phi_{0}$ = $2\pi\hbar/e$
of magnetic flux passes.
In the conventional treatment using the
so-called “symmetric gauge”, where the electromagnetic vector potential
$(A_{x},A_{y})$ is given by $A_{x}+iA_{y}$ =
$i\Phi_{0}z/(4\pi\ell^{2})$, and $z$ = $x+iy$ is a mapping from the
Euclidean plane to the complex plane, the lowest-Landau-level ($n=0$) wavefunctions have the form
$$\psi(z,z^{*})=f(z)e^{-\frac{1}{4}z^{*}z/\ell^{2}}$$
(25)
where $f(z)$ is a holomorphic function.
Quasiperiodic boundary conditions under a
Bravais lattice $\Lambda$ of translations with a
primitive region of area $A$ = $2\pi N_{\Phi}\ell^{2}$ (and integer
$N_{\Phi}$) can be imposed
(as a technical device to avoid the problems of working with an
infinite system), so that
$$\psi(z+L,z^{*}+L^{*})=\xi(L)^{N_{\Phi}}e^{K^{*}L-KL^{*}}e^{\frac{1}{4}(L^{*}z-%
Lz^{*})/\ell^{2}}\psi(z,z^{*}),$$
(26)
or
$$f(z+L)=e^{K^{*}L-KL^{*}}\left(\xi(L)e^{(\pi L^{*}/A)(z+\frac{1}{2}L)}\right)^{%
N_{\Phi}}f(z).$$
(27)
This is solved by wavefunctions with
$$f(z)\propto e^{K^{*}z}\left(\prod_{i=1}^{N_{\Phi}}\widetilde{\sigma}(z-w_{i})%
\right),\quad\sum_{i=1}^{N_{\Phi}}w_{i}=KA/\pi.$$
(28)
This wavefunction has $N_{\Phi}$
zeroes in the fundamental region.
The many-particle Slater-determinant
state with $N$ = $N_{\Phi}$ fermions that completely
fills the lowest Landau level is then given by
$$\Psi\propto e^{K^{*}Z}\widetilde{\sigma}(Z-W)\left(\prod_{i<j}\widetilde{%
\sigma}(z_{i}-z_{j})\right)\prod_{i}\left(e^{-\frac{1}{2}\pi z_{i}^{*}z_{i}/A}%
\right)^{N_{\Phi}},\quad Z=\sum_{i}z_{i},\quad W=KA/\pi.$$
(29)
Many other more-interesting model many-particle states on the torus (e.g., the
Laughlin state) can be written in an explicitly modular-invariant
way using the modified sigma function described here. These will be
described elsewhere.
The importance of these constructions using the modified sigma
function is that they are explicitly modular-invariant, unlike most
previous representations using the Jacobi elliptic function
$\vartheta_{1}(u|\tau)$, $u$ = $\pi z/2\omega_{1}$,
$\tau$ = $\omega_{2}/\omega_{1}$.
When constructing model wavefunctions
on the toroidal compactification of the plane,
an important physical requirement is that they do not depend on the
choice of basis of $\Lambda$, i.e., that they are
modular-invariant. Use of the modified sigma function ensures this property is explicit.
Whether or not the simple “modification” described here is useful in
other applications of the Weierstrass functions, it is now clear that the
modified sigma function is the variant of choice for describing lowest
Landau-level wavefunctions. However the simplicity of the formula (17),
and the appearance of “naturalness” of the “completion” by the inclusion of the
“missing” $2k$ = 2 term in the Laurent expansion (3),
suggests that all applications of the Weierstrass functions could
benefit
from the modification proposed here, which has antecedents in Eisenstein’s workeis ; weil .
Acknowledgements.
This work was supported by Department of Energy BES Grant DE-SC0002140.
References
(1)
K. Weierstrass, Mathematische Werke von Karl Weierstrass,
(Mayer & Müller, Berlin, 1895), Vol 2, p. 245.
(2)
S. Ramanujan, Trans. Cambridge Philos. Soc. 22, 159 (1916).
(3)
G. Eisenstein, J. Reine Angew. Math. 35, 137 (1847); reprinted in
G. Eisenstein, Mathematische Werke, (Chelsea, New York, 1989),
Vol. I, p. 357.
(4)
P. Charollois and R. Sczech, EMS Newsl. 9, 8 (2016).
(5)
A. Weil,
Elliptic functions according to Eisenstein and Kronecker,
2nd Edition (Springer-Verlag, Berlin, 1999). (1st Edition 1976.)
(6)
L. Rolen, Proc. Amer. Math. Soc. 144, 1453 (2016). |
The distribution of nearby stars in phase space mapped by
Hipparcos††thanks: Based on data from the Hipparcos astrometry
satellite
M. Crézé
1Centre de Données astronomique de Strasbourg, CNRS URA1280
11 rue de l’Université, F-67000 Strasbourg, France
12IUP de Vannes,
8 rue Montaigne, BP 561, 56017 Vannes Cedex, France
2
E. Chereul
1Centre de Données astronomique de Strasbourg, CNRS URA1280
11 rue de l’Université, F-67000 Strasbourg, France
1
O. Bienaymé
1Centre de Données astronomique de Strasbourg, CNRS URA1280
11 rue de l’Université, F-67000 Strasbourg, France
1
C.Pichon
3Astronomisches Institut Univ. Basel,
Venusstrasse 7,
CH-4102 Binningen, Switzerland 3
(Received / Accepted )
Abstract
Hipparcos data provide the first, volume limited and
absolute magnitude limited homogeneous tracer of stellar density and
velocity distributions in the solar neighbourhood. The density of
A-type stars more luminous than $M_{v}=2.5$ can be accurately mapped
within a sphere of 125 pc radius, while proper motions in galactic
latitude provide the vertical velocity distribution near the galactic
plane. The potential well across the galactic plane is traced
practically hypothesis-free and model-free. The local dynamical
density comes out as $\rho_{0}=0.076\pm 0.015~{}M_{\sun}~{}{\rm pc}^{-3}$ a
value well below all previous determinations leaving no room for any
disk shaped component of dark matter.
keywords: Hipparcos - Galaxy: kinematics and dynamics -
Galaxy: fundamental parameters -
Galaxy: halo -
solar neighbourhood -
Galaxy: structure -
dark matter -
\thesaurus
04 10.11.1; 10.06.2; 10.08.1; 10.19.1; 10.19.3; 12.04.1
\subtitle
I. The potential well and local dynamical mass
\offprints
M. Crézé
1 Introduction
All the data used here were collected by the Hipparcos satellite
([ESA, 1997]). Individual stellar distances within more than 125 pc
were obtained with an accuracy better than 10$\%$ for almost all
stars brighter than $m_{v}=8.$, together with accurate proper
motions. Based on these data, a unique opportunity is offered to
revisit stellar kinematics and dynamics; any subsample of sufficiently
luminous stars is completely included within well defined distance and
luminosity limits, providing a tracer of the local density-motion
equilibrium in the galaxy potential: a snapshot of the phase space. We
have selected a series of A-F dwarf samples ranging from $M_{v}=-1.0$
down to $M_{v}=4.5$. Completeness is fixed within 50 pc over the whole
magnitude range and within 125 pc at the luminous end ($M_{v}\leq 2.5$). In this series of papers we shall investigate such samples in
terms of density and velocity distribution small scale inhomogeneities
addressing the problem of cluster melting and phase mixing.
The expected first order departure to homogeneity is the potential
well across the galactic plane. This problem is well known in galactic
dynamics; it is usually referred to as “the $K_{z}$ problem”, where
$K_{z}$ means the force law perpendicular to the galactic plane. The
$K_{z}$ determination and subsequent derivation of the local mass
density $\rho_{0}$ has a long history, nearly comprehensive reviews
can be found in [Kerr & Lynden-Bell (1986)] covering the subject before 1984 and in
[Kuijken (1995)] since 1984. Early ideas were given by Kapteyn (1922),
while Oort (1932) produced the first tentative determination.
The essence of this determination is quite simple: the kinetic energy
of stellar motions in the z direction when stars cross the plane fixes
their capability to escape away from the potential well. Given a
stellar population at equilibrium in this well, its density law
$\nu(z)$ and velocity distribution at plane crossing $f(w_{0})$ are tied
to each other via the $K_{z}$ or the potential $\phi(z)$. Under quite
general conditions the relation that connects both distributions is
strictly expressed by Eq. (1).
$$\nu(\phi)=2\int_{\sqrt{2\phi}}^{\infty}{f(|w_{0}|)~{}w_{0}~{}{\rm d}w_{0}\over%
\sqrt{w_{0}^{2}-2\phi}}$$
(1)
This integral equation and its validity conditions are established and
discussed in detail by [Fuchs & Wielen (1993)] and [Flynn & Fuchs (1994)]. There is no
specification as to the form of distribution functions $f$ and $\phi$
except smoothness and separability of the z component.
Given $\nu(z)$ and $f(w_{0})$, $\phi(z)$ can be derived. Then according
to Poisson equation, the local dynamical density comes out as
$$\rho_{0}={1\over 4\pi G}({\rm d}^{2}\phi/{\rm d}z^{2})$$
(2)
Early determinations of $\rho_{0}$ ([Oort, 1960]) were as high as
$0.18\,M_{\sun}\,{\rm pc}^{-3}$ while the total mass density of stars
and interstellar matter would not exceed $0.08\,M_{\sun}\,{\rm pc}^{-3}$. The discrepancy between these two figures latter
re-emphasised by J. Bahcall and collaborators (1984ab, 1992), came
into the picture of the dark matter controversy to make it even
darker. While the local density of standard “dark halos” required
to maintain flat rotation curves of galaxies away from their center
would hardly exceed 0.01, such ten times larger densities could only
be accounted for in rather flat components which composition and
origin could not be the same. Following Bahcall’s claim, several
attempts were made to reduce the intrinsic inaccuracy of the dynamical
mass determinations in particular [Bienaymé et al (1987)], used a complete
modelisation of the galaxy evolution to tie the $K_{z}$ determination to
star counts, [Kuijken & Gilmore (1989)] used distant tracers of K type stars to
determine not the local volume density but the integrated surface
density below 1 kpc, [Flynn & Fuchs (1994)] pointed at difficulties in
relation with the underlying dynamical approach. Most concluded that
there was no conclusive evidence for a high dynamical density if all
causes of errors were taken into account.
The local mass density of observable components (the observed mass)
can hardly exceed $0.08M_{\sun}{\rm pc}^{-3}$ out of which a half
stands for stars which local luminosity function is reasonably well
known. The mass density of interstellar gas and dust is not so easily
accessed since the bulk of it, molecular hydrogen, can only be traced
via CO and the ratio CO/$\rm{H_{2}}$ is not so well established. Also
interstellar components are basically clumpy and distances poorly
known: it is not so easy to define a proper volume within which a
density makes sense. A conservative range for the observed mass
density should be $[0.06,0.10]$ including 0.04 for stars.
Nevertheless, so far the discrepancy used to be charged to the
dynamical density since the determination of this quantity involved
many difficulties, most poorly controlled :
•
Tracer homogeneity: first of all, the determination
require that a suitable tracer can be duly identified. A criterion
independent of velocity and distance should allow to detect and select
all stars matching the tracer definition. This was hardly achieved by
spectral surveys which completeness and selectivity degrade as
magnitude grows.
•
Stationarity: in order for the tracer to make dynamical
sense, it should be in equilibrium in the potential. Youngest stars
which velocities reflect the kinematics of the gas at their birth time
do not fulfill this requirement.
•
Undersampling: one searches for the bending of density
laws caused by gravitation along the z direction. This imposes not
only that tracer densities can be determined, but also that they can
be determined at scales well below the typical length of the
bending. This means that the tracer should be dense enough in the
region studied. Most determinations so far rely upon sample surveys in
the galactic polar caps, resulting in “pencil beam” samples quite
inappropriate to trace the density near the plane ([Crézé et al, 1989]). When
the bending is traced at large distances from the plane as in
[Kuijken & Gilmore (1989)], one gets information on the surface density below a
certain height. A comparison of this surface density with the local
volume density involves modeling hypotheses. Another associated effect
can be generated by a clumpy distribution of the tracer even
stationary if the tracer includes clusters or clumps the statistics of
densities and velocities may be locally biased. Other problems arise
from the velocity distribution sampling.
•
Systematic errors: out of necessity before Hipparcos,
all previous studies used photometric distances. Uncertainties in the
calibration of absolute magnitudes resulted in systematic errors in
density determinations. This is particularly true for red giants which
are not at all a physically homogeneous family and show all but normal
absolute magnitude distributions: it is striking that calibrations
based on different samples provide discrepant calibrations outside the
range permitted by formal errors. This produced both biased density
distributions and uncontrolled effects in selecting volume limited
samples.
•
Random errors: distances and velocities were also
affected by random measurement errors. Even perfectly calibrated,
standard errors of photometric distances based on intermediate band
photometry would hardly get below 20%.
Hipparcos data solve nearly all the problems quoted above: instead of
“pencil beam” tracers at high galactic latitudes which distribution
is only indirectly related to the local $K_{z}$, what we get here is a
dense probe inside the potential well.
The samples were preselected within magnitude limits inside the
completeness limit of the Hipparcos survey program
([Turon & Crifo, 1986]). It includes all stars brighter than $m_{v}$ = 8.0 (7.9
at low latitudes) with spectral types earlier than G0, stars with no
or poorly defined spectral types were included provisionally.
Within this coarse preselection, sample stars were eventually selected
on the basis of their Hipparcos derived distance, magnitude and
colour. So the only physical criterion is absolute magnitude while the
sample is distance limited on the basis of individual parallaxes.
Within 125 parsecs, Hipparcos parallaxes provide individual distances
with accuracies better than 10 percent for over ninety percent of the
sample. Distances are individual and free from calibration errors.
Samples are dense, typical inter-star distances are less than 10 pc.
This is suitable to trace density variations which typical scales are
of the order of hundred parsecs. Thanks to this high tracer density,
difficulties related to clumpiness can be monitored through
appropriate analysis of local density residuals. Stationarity
considerations can also be monitored since W velocities are available
for stars everywhere in the volume studied.
The main characteristics of the tracer samples are reviewed in section
2. In order to take full benefit of this new situation, an original
method has been developed to analyze the densities. The statistical
aspects of this method based on single star volumes (volume of the
sphere extending to the nearest neighbour) are described in section 3
yielding a model free view of the potential well. Then an estimate of
the local mass density of matter is produced (section 4).
Consequences in terms of galactic structure and dark matter
distribution are reviewed in conclusion (section 5).
2 The Hipparcos Sample
2.1 Sampling definition
The sample was pre-selected from the Hipparcos Input Catalogue
([ESA, 1992]), among the “Survey stars”. The Hipparcos Survey
delimitation is described in full detail in Turon & Crifo (1986). For
spectral types earlier than G5 the limiting magnitude is $m_{v}\leq~{}7.9+1.1sin|b|$. It is complete within those limits so
far as the apparent magnitudes were known. Inside the Hipparcos
Survey, the phase space tracer sample has been first given a coarse
limitation: spectral types later than A0 and earlier than G0,
luminosity class V through IV. Stars with uncertain or undefined
luminosity classes were kept in the sample at this stage. Within this
pre-selection, the final choice was based on Hipparcos magnitudes
($m_{v}<8.0$), colours ($-0.1~{}\leq~{}B-V~{}\leq~{}0.6$) and parallaxes
($\pi~{}\leq~{}8~{}{\rm mas}$). Then a series of subsamples were cut in
absolute magnitude and distance in such a way that each subsample is
complete for the adopted colour range within a well defined sphere.
The resulting sample characteristics are summarized in Tab.(1). The
sample designation system is straightforward: h125 means stars within
125 pc, under the apparent magnitude condition $m_{v}<8.0$, imposes
not to consider stars less luminous than $M_{v}\leq 2.5$. All stars of
sample h125 closer than 100 pc are included in sample H100 and so
on. Samples named k… come from the corresponding h… samples but
the identified cluster stars have been removed. The typical density of
each sample estimated as the inverse of the average single star volume
is given in the last column of Table 1. The corresponding
typical interstar distance of e.g. tracer h125 is 8.8 pc.
2.2 Completeness
With this method, the sample limits do not rely upon pre-Hipparcos
data except for the faintest stars next to the edge of each sphere,
some of those might have been missed due to the limited accuracy of
pre-Hipparcos magnitudes. The method may look a naive one since there
are now sophisticated algorithms to correct estimations for censored
data, (see for instance [Ratnatunga & Casertano, 1991]) but it is the only one
that does not make assumptions as to the absolute magnitude
distribution of the stars. All previous investigations did assume that
the absolute magnitudes of stars sharing a certain spectral or
photometric characteristic are normally distributed around a common
mean value.
Figures 1 and 2 give a quantitative
illustration of the risks involved: the absolute magnitude
distributions of dwarfs within four V-I ranges show an obvious
asymmetry while in the reddest part, the luminosity distribution spans
over more than two magnitudes. This means that however accurately the
apparent magnitude can be measured, the uncensoring correction would
be much more uncertain than other error factors.
With the approach adopted here, Fig. 3 shows the
$M_{v}$/distance distribution of the global sample, the neat border
follows simply the line $m_{v}=8$; the censoring effects are maximum
along this line. Our subsamples are cut along lines parallel to the
axes of this plot, so they reach the censoring line only at the bottom
right corner of each zone.
Since the accuracy of Hipparcos magnitudes is far beyond the
necessities of this study, the sampling biases can only result from
two effects: the parallax errors which, however unprecedently small
are still of the order of $10\%$ beyond 100 pc, and the stars lost at
the time of the early selection due to the inaccuracy of apparent
magnitudes available then. Concerning the second effect, it can only
result in stars being omitted near the edge of the sample. The residual
effect of apparent magnitude incompleteness can only affect the
faintest magnitudes at the edge of each sample sphere, since the
single star volume used in sections 3 and 4 as statistics for the
density estimation are local statistics, it is easy to keep such
effect under control in the study of residual, furthermore this kind
of incompleteness if important would bias the estimation towards
making the potential well even shallower than observed.
2.3 Distance errors
The other possible source of bias left is generated by distance
errors. Figure 4 shows the distribution of parallax
errors versus distance for sample h125 which, extending to 125 pc is
the most severely affected by distance errors. This is still a
moderate effect: one can see on the figure that only a handful of
parallax errors exceed 15$\%$. Parallaxe errors are Hipparcos formal
errors. All feasible external tests able to check the actual accuracy show
that these formal errors are if something a little pessimistic
([Arenou et al, 1995]) with respect to previous studies of the density
trends.
3 Mapping the tracer density and the potential well
The essential choice of this investigation is to stick as closely as
possible to observations. Whatever the velocity distribution $f(w_{0})$,
once an homogeneous tracer has been defined, that is once a selection
criterion has been defined uncensored with respect to velocities and
positions, Eq. (1) expresses the correspondence between the
tracer density law and the potential as a plain change of variables.
So if there is a potential well to be found in the data, it must
appear as a density peak. If we can map the density trend model free,
we’ll get a picture of the potential well.
3.1 Single star volume statistics
The tracer should be dense enough for the typical distances between
tracer stars to be small with respect to the typical scale of density
variations. Under this assumption, at a place were the tracer density
is $\nu$, the observed number of stars counted in any probe volume $V$
is a Poisson variate with expectation $V\cdot\nu$. Defining $v$ as
the single star volume around one specific star (that is the sphere
extending to the nearest neighbour), and introducing the quantity $x=v\cdot\nu$ we get a new variate which expectation is 1 and which
probability distribution is exponential (Eq. (3)). It is the
probability distribution of the distance between two events in a
Poisson process under unit density.
$${\rm d}P(x)=\exp(-x)~{}{\rm d}x$$
(3)
and the probability distribution of the single star volume $v$ is
$${\rm d}P(v)=\nu\cdot\exp(-\nu\cdot v)~{}{\rm d}v$$
(4)
so the expectation of the single star volume is $1/\nu$, which makes
it a suitable local statistics for the density. It is not a novel
approach to use statistics based on nearest neighbour distances to
investigate densities, but it is quite appropriate here since it turns
out to provide a parameter free Maximum Likelihood estimator for the
density by plain moving average of single star volumes. Also, single
star volumes being a local measure of the inverse density at the
position of each star, they offer the possibility to calculate
individual density residuals. Checks can be made that deviations from
average or model predicted values are randomly distributed and do not
show unmodeled systematic trends.
3.2 Maximum Likelihood estimator
In a constant density sample, the plain average of single star volumes
is a maximum likelihood statistics for the density: let $v_{i}$ be the
single star volume around star i, according to Eq. (4) the
log-likelihood of a sample of n stars $(i=1,n)$ under constant
density $\nu$ is simply given by
$$\log L=\sum_{i=1,n}(\log(\nu)-\nu\cdot v_{i})$$
(5)
and the Maximum Likelihood is reached for
$${\rm d}\log L/{\rm d}\,\nu=0\\
$$
(6)
which obvious solution is:
$$n/\nu=\sum_{i=1,n}v_{i}\\
$$
(7)
$$\nu=1/<v_{i}>$$
(8)
this makes the moving average of $v_{i}$ along a parameter a maximum
likelihood mapping of the density variations along this parameter,
provided only that variations along other parameters are negligible.
3.3 Residual statistics
Furthermore, given a model or an estimate of $\nu$ near star i the
quantities $x_{i}=\nu\cdot v_{i}$ are local “residuals” of the fit
and their distribution under the assumed model should be exactly
given, parameter-free, by Eq. (3). So the distribution of
residuals provides an immediate test of the validity of the model
including the fact that the density is smooth and do not include too
many clusters or voids. This distribution can be tested over the whole
sample or over any subset chosen to explore neglected parameters (e.g.
the completeness near the edge of the sampling volume, completeness in
apparent magnitude). In order to visualize the agreement (or
disagreement) of the observed distributions we have used two
representations: the histogram of the log residuals $y=log(x)$ which
distribution according to Eq. (3) must be given by
$${\rm d}P(y)=\exp[y-\exp(y)]{\rm d}y\\
$$
(9)
and the cumulative distribution of x:
$$H(X)=\int_{0}^{X}\exp(-x)\,{\rm d}x=1-\exp(-X)\\
$$
(10)
Let $n_{i}$ be the rank of residual $x_{i}$ in a sample of N values, then
according to Eq. (10),
$$n_{i}/N=1-\exp(-x_{i})\\
$$
(11)
$$-\log(1-n_{i}/N)=x_{i}\\
$$
(12)
In the following the residuals have been plotted as $-log(1-n_{i}/N)$
against $x_{i}$ producing what statisticians use to call exponential
probability plots. On such plot, observed residual distributions that
fail matching the expected exponential at any scale sign their
existence by departures from the diagonal.
3.4 Edge effect
Around stars close to the edge of the completeness sphere, the single
star volume is not anymore the full volume of the sphere to the
nearest neighbour. The volume must be corrected for the part of the
sphere outside the completeness volume (Fig. 5).
Defining $d_{nn}$ the distance to the nearest neighbour, $d$ the
Sun-star distance and $R$ the radius of the sample sphere, and
defining the two spheres, one centered at sun position with radius
$R$, the other at star position with radius $d_{nn}$, the corrected
volume is calculated as $V_{cor}$ according to the following formulae.
Angles $\theta_{1}$ and $\theta_{2}$ are the half apertures of the cones
intercepting the intersection of the two spheres (summit at sun
position for $\theta_{2}$, at star position for $\theta_{1}$).
$$\cos\theta_{2}=(d^{2}+R^{2}-d_{nn}^{2})/(2dR)$$
(13)
$$\sin\theta_{1}=(R/d_{nn})\sin\theta_{2}$$
(14)
$$V_{1}=(\pi/3)d_{nn}^{3}(2-3\cos\theta_{1}+\cos^{3}\theta_{1})$$
(15)
$$V_{2}=(\pi/3)R^{3}(2-3\cos\theta_{2}+\cos^{3}\theta_{2})$$
(16)
$$V_{cor}=(4/3)\pi d_{nn}^{3}-(V_{1}-V_{2})$$
(17)
3.5 Mapping the density
Single star volumes which are unbiased local estimates of the inverse
tracer density have been computed for all the tracer samples described
in Table 1. Defining x,y,z as a set of cartesian coordinates
centered at the sun position, with z positive pointing towards
$b=+90^{\circ}$, x positive towards the galactic center ($l=0^{\circ}$)
and y towards $l=90^{\circ}$, each series of values was sorted along
x, y and z successively and each resulting set smoothed by moving
average along the sort parameter. Single star volumes are averaged
over 101 neighbouring stars, producing parameter free maximum
likelihood inverse density profiles. Figures 6 through
8 show the resulting profiles along x, y and z for our
three most accurate and extended samples (namely samples h125, H100
and H80, according to definitions of section 2.1).
In these figures the x and y plots are given only as references: since
the sampling definition has spherical symmetry, any systematic effect along
the z axis which would trace a sampling artifact rather than a real trend
of the density would appear as well along the x and y axis.
At first glance only do the h125 and H100 profiles along the z axis
show significant bending. Expectedly the youngest subsamples with the
smallest velocity dispersions are also the most sensitive
tracers. This is a direct mapping of the potential well. The next
section will be dedicated to deriving quantitative consequences of
this observation. Beforehand the result should be scrutinized in the
light of two possible criticisms: biases in the inverse volume
estimates at the edges of the sample and clumpiness effects among
poorly mixed A stars.
Edge effects potentially related to completeness and/or to distance
errors can be easily controlled: would they be responsible for the
main part of the observed bending they would appear as well on the x
and y plots. There is no such thing visible even at marginal level. In
addition, any significant incompleteness at the edges would result in
a density decrease driving the single star volume statistics up at the
edges. As a result, the dynamical density derived below, once
corrected for this effect, would be even smaller.
A more serious problem comes from the clumpiness of the distribution:
young stars that dominate the most luminous samples are likely to be
partly concentrated in clusters or clumps on scales of a few parsecs
to a few ten parsecs. Such clumps must be suspected to distort the
density profile.
The fluctuations observed at small scales on figure 6 through 8 are not
significant in this respect, they are the result of random fluctuations
of single star volumes smoothed by the moving average process. The moving
average introduces a strong self correlation in the series. As a result
random fluctuations fake systematic ones on scales which are connected with
the z ranges spanned by 101 neighbouring stars, but do not reflect these
ranges in any simple way (it has something to do with the probability
of getting one, two, three … extreme peaks within adjacent strips).
Similar pictures are obtained on strictly uniform random simulations.
The only place where to check this effect is the statistics of residuals.
Real clumps which are not plain random fluctuations in a uniform density
are expected to sign their presence not by their local density which
may not be very high locally on the average, but by the frequency of
corresponding single star volume residuals (see section 3.3 for the
definition of these residuals. Such an overfrequency is
clearly visible on the bottom left wing of the histogram on figure 9a.
In terms of cumulative distribution it is even more clear on figure 9b.
A separate analysis of clustering has been performed by
means of wavelet analysis (Chereul et al. to appear as paper II of
this series) showing that no more than 8$\%$ out of stars in the h125
sample are involved in clumps. Removing only cluster members
identified through proper motions we get sample k125. The z inverse
density profile of this new sample is plotted on figure 6d, the
profile is not strongly modified by removing cluster stars. On the
residual plot (figure 9c) the residual anomaly at small scales
vanishes definitely. So the inverse density profiles turn out quite
robust to clumpiness effects. Furthermore we shall see in the following
(section 4.2 and figure 13) that local mass density estimates based on the
two samples (with or without cluster stars) are fully compatible. Also,
once a suitable density model is found, the residual statistics of
sample k125 (Figure 13b) do not show any significant feature that might signal
undetected clumpiness left.
This gives confidence that the bending is real and that figures 6cd, 7c
are the first ”image” of the galactic potential well. The remarkable
feature here is that this well is extremely shallow. We shall see in
the following that any substantial layer of dark matter contributing
significantly in the local mass density would unmistakably sign its
presence by deepening this well far beyond observed shapes. This is
true also for samples H80 and H100 which, being dominated by elder
stars are less affected by the clumpiness effect.
4 Estimating $\rho_{0}$
Solving Eqs. (1) and (2) for $\phi(z)$ can follow two
different ways: a parametric approach and a non parametric one.
Either were used on the above defined samples producing quite similar
results.
•
In the parametric approach, realistic mathematical forms
are adopted for both $\phi(z)$ and $f(w_{0})$. Free parameters for
$f(w_{0})$ can be determined by fitting the observed vertical velocity
distribution then adopting $f(w_{0})$ free parameters for the potential,
$\phi(z)$ can be derived from the observed density. The advantage of
this approach is simplicity in so far as manageable formulae can be
found to represent $f(w_{0})$. The cost is that results may be biased
by the choice of the model. An advantage may be that systematic
effects can be traced and individual residuals can be computed at the
position and velocity of each star. This approach is developed in
section 4.1.
•
At the other end one can try to produce a purely
numerical fit. Since this kind of fit is unstable to noise in the
data, some regularization should be applied. This non parametric
approach is developed in section 4.3.
4.1 Parametric approach: velocity analysis
The observed W vertical velocity distributions in the galactic plane
were derived from Hipparcos proper motions in galactic latitude and
parallaxes. For this analysis, samples were restricted to stars below
$b=\pm 10\degr$ so that $\mu_{b}/\pi$ simply reflects the W
tangential velocity.
W histograms of the four samples under consideration are shown on
Fig. 10.
All four distributions show reasonable symmetry and smoothness with
the following exceptions: h125 (Fig. 10a) shows pronounced
spikes corresponding to clusters (spikes vanish in the histogram of
k125) and a dissymmetry of large velocities (involving only 5 stars
out of 456). The plain average $W_{0}$ has been estimated for all
samples and appears quite stable suggesting that unsteady streaming
motions characteristics of very young stars do not dominate. Under
such conditions, the capability of the subsample of 536 low latitude
stars to be representative of the whole population within 125 pc is
questionable. Radial velocities have been compiled from the SIMBAD
database in order to calculate space velocities for as many stars as
possible. This resulted in a new velocity sample named h125W including
1366 stars spread in the whole sphere. The distribution of W velocity
components has been studied in the following along the same lines as
the other velocity sample.
A double gaussian centered model was fitted to each centered histogram
using a simple maximum likelihood scheme:
the model fitted is
$$\displaystyle{\rm d}P(w)=\{{\beta\over\sqrt{2\pi}~{}\sigma_{w1}}\exp[-{1\over 2%
}(w-<w>)^{2}/\sigma_{w1}^{2}]$$
$$\displaystyle+{1-\beta\over\sqrt{2\pi}~{}\sigma_{w2}}\exp[-{1\over 2}(w-<w>)^{%
2}/\sigma_{w2}^{2}]~{}\}~{}~{}{\rm d}w$$
(18)
and the maximum likelihood is searched for on $\beta$, $\sigma_{w1}$
and $\sigma_{w2}$.
There is no claim here that the kinematical mixture is isothermal. The
idea is simply that, if this mixture can be accurately represented by
a sum of gaussians, then there is a simple solution to the self
consistent dynamical equilibrium, the force law and the local volume
density can be solved for explicitly. The justification lays in the
quality of the representation. This is also justified a posteriori by
the fact that non parametric solutions do provide the same results.
Table 2 gives the maximum likelihood solutions for the four
main samples investigated here. Two tests of the quality of the fit
are given, one is the Kolmogorov-Smirnoff test: Pk is the observed
Kolmogorov maximum distance, Pk(0.8) and Pk(0.95) are the 20$\%$ and 5
$\%$ rejection thresholds for this parameter. Obviously no rejection
of the double gaussian representation can be supported at any
meaningful level.
The second test is given by probability plots (Fig. 11), velocity
ranks in each sample are plotted against the theoretical cumulative
probabilities of associated velocities: velocities distributed
according to the estimated distributions would produce a strictly
linear plot. So any deviation from linearity signs a failure of the
double gaussian to represent observations.
The likelihood surface is regular around each maximum, the dependence
in $\sigma_{w1}$ and $\sigma_{w2}$ follows a nice parabola in
log-likelihood indicating that the estimator behaves quite alike a
gaussian variate in this region. Standard errors follow. The
proportion $\beta$ is poorly determined although different values of
$\beta$ may produce quite different M.L. estimates for $\sigma_{w1}$
and $\sigma_{w2}$. But the Likelihood levels reached do not deviate
significantly from oneanother. A series of tests, not reported here,
show that adopting any of those solutions, nearly equivalent in
likelihood, do produce extremely close estimations of the dynamical
density. Expectedly, what matters here, is the quality of the
distribution representation, not the choice of specific functions.
Eventually this conclusion is comforted by the excellent agreement
between the results of the parametric and non parametric approach.
4.2 Parametric approach: density and potential analysis
Once we get models fitting W velocity distributions, which can be
represented by double gaussians, the self consistent solution of
Eqs. (1) and (2) can be expressed explicitly as
$$\displaystyle\ \nu(z)=\nu_{0}~{}[\,\beta~{}e^{-\phi(z)/\sigma_{w1}^{2}}+(1-%
\beta)~{}e^{-\phi(z)/\sigma_{w2}^{2}}\,]$$
(19)
Over the very small z range spanned by our 125 pc sphere, the
potential well can be approximated at any required accuracy by a plain
quadratic form
$$\phi(z)=\alpha\cdot(z-z_{0})^{2}$$
(20)
So, the density trend of a tracer $\nu(z)$ can be used to estimate
$\alpha$, $\nu_{0}$ and $z_{0}$, once $\beta$, $\sigma_{w1}$ and
$\sigma_{w2}$ have been derived from the velocities. The likelihood
of a set of single star volumes under the assumption that the density
trend is represented by Eqs. (19), (20), can be
expressed by Eq. (21) similar to Eq. (5).
$$\log L=\sum_{i=1,n}~{}(~{}\log\nu(z_{i})-\nu(z_{i})\cdot v_{i}~{})$$
(21)
A maximum likelihood can be searched for $\alpha$, $\nu_{0}$ and $z_{0}$.
According to the Poisson equation, the local dynamical density follows
from $\alpha$ as:
$$\rho_{0}={1\over 2\pi G}~{}~{}~{}\alpha$$
(22)
Where $1\,/(2\pi G)=37.004$ if $\alpha$ is in $({\rm km\,s}^{-1}\,{\rm pc}^{-1})^{2}$ and $\rho_{0}$ in $M_{\sun}\,{\rm pc}^{-3}$.
Table 3 summarizes the results of solutions based on the
three main samples (the separate solutions derived from samples
without identified cluster stars do not produce different results).
Three solutions are given for each sample: one is the 3-parameter
M.L. solution, the other two are M.L. solutions for $\alpha$,
$\nu_{0}$, while specific values have been forced for $z_{0}$.
All three parameter solutions produce unexpected positive values for
the z coordinate of the galactic plane. That is, the Sun would be
very slightly south of the plane, in contradiction with most accepted
values. We do not consider this result worth much consideration: for
one thing the estimated local density does not change appreciably if
$z_{0}$ is forced to 0 and even if it is pushed to a more classical $z_{0}=-10$, the estimated density decreases. For the other thing, there
is certainly some clumpiness among the youngest stars in this sample
which may blur the shape of the potential well the more so as this
well turns out shallower than expected. Since it is not possible to
evaluate accurately this effect we shall consider it as an intrinsic
limitation of this determination. Accordingly the adopted solution is
the weighted mean of M.L. solutions (weights proportional to inverse
square standard errors), but error bars are extended in order to
include all main solutions of Table 3.
$$\rho_{0}~{}=~{}0.076~{}\pm~{}0.015~{}M_{\sun}~{}{\rm pc}^{-3}$$
(23)
Based on this value and on the velocity distributions obtained in
section 4.1, predicted density models defined by Eq. (19) have
been overplotted on the z moving average profiles of all three tracer
samples (Fig. 12 a,b,c). For comparison, profiles
corresponding to arbitrary densities $\rho_{0}~{}=~{}0.10~{}M_{\sun}~{}{\rm pc}^{-3}$ and $\rho_{0}~{}=~{}0.15~{}M_{\sun}~{}{\rm pc}^{-3}$ are added. Obviously, even
under quite conservative hypotheses, none of the samples can be
reconciled with such local densities. The existence of any
substantial disk shaped dark matter would definitely push all three
curves outside the acceptable range.
The distributions of single star volume residuals under the three
density models have been scrutinized against possible unmodeled
systematic effects. Probability plots as defined in section 3.3 (Eq.
(11)) are given in Fig. 13.
The first two plots on top show the probability plot of residuals
for samples h125 and k125 . Clearly enough models based on local mass
densities higher than 0.08 generate a systematic distortion of the
cumulative distribution of residuals with respect to the expected one.
There is a small excess of low residuals even with our best fit model,
but this effect vanishes when cluster stars are removed (sample k125)
This correction does not change the conclusions on the local mass.
At this stage , the validity of the simplified assumption made on the
potential can also be tested. A square function in z has been adopted for the
potential, meaning a linear $K_{z}$. This means some homogeneity of the mass
distribution over a 100 parsecs scale. If a substantial fraction of the mass
happened to be concentrated in a very thin layer, this approximation would
not hold.
The Hipparcos samples are the first able to test this hypothesis
over such small scales. The test is provided by the statistics of single
star residuals. Would the $K_{z}$ slope change significantly within the limits
of the sample, then this statistics of subsamples at low z would not
match the expectation for the same assumed $\rho_{0}$. This test is
presented in the bottom parts of Fig. 13.
Samples h125 and H100 were split in different subsamples selected by
$z$ after restriction to a cylinder of radius $r$. Expectedly, the failure
of models based on $\rho_{0}=0.10$ and beyond is enhanced at high z
($|z|>69~{}{\rm pc}$).
But, even over the restricted range $|z|<69~{}{\rm pc}$ there is signal left in
the residuals to reject high density solutions. The low z part
of sample H100 ($|z|<55~{}{\rm pc}$) is unable to produce any discriminant signal.
These residual plots fully confirm the result that local total volume
densities beyond $0.10~{}M_{\sun}~{}{\rm pc}^{-3}$ cannot be reconciled
with the observed velocity-density distributions. Similar plots
restricted to subsamples selected by distance or magnitude or colour
do not show significant deviations that might sign the presence of
major selection effects in the sampling.
In terms of mass model the assumed linear $K_{z}$ implies that the total
mass density is constant close to the plane. Assuming that it is true
only for half of the mass, for instance the stellar distribution and
that the gas distribution follows a sharper gaussian distribution with
a scale height of 75 pc we may determine the extra forth order term in
the potential. Determining the amplitude of the potential on counts at
75 pc leads to a very small difference of 1/24 between the two
models. The resulting error on the derived $\rho_{0}$ is negligible with
respect to other uncertainties.
4.3 Non parametric approach
The non parametric approach does not rely upon any particular
mathematical model for the velocity distribution function nor the
height density profile ([Pichon & Thiébaut, 1997]). That is: ideally we should
try and solve self consistently Eqs. (1) and (2) for $\rho_{0}$, with
functions $f(|w_{0}|)$ and $\nu(\phi)$ defined on a purely numerical
basis (in the range investigated here, approximating $\phi(z)$ by a
simple square function is well justified). However a pure numerical
solution is unstable: a small departure in the measured data (e.g. due to noise) may produce drastically different solutions since these
solutions are dominated by artifacts due to the amplification of
noise. Some kind of trade off must therefore be found between the
level of smoothness imposed on the solution in order to deal with
these artifacts on the one hand, and the level of fluctuations
consistent with the amount of information in the signal on the other
hand. Finding such a balance is called the “regularization”
([Wahba & Wendelberger, 1979]) of the self consistency problem. Regularization allows
to avoid over-interpretation of the data. Here we chose to regularize
the non-parametric self consistency problem by requiring a smooth
solution. This criterion is tunable so that its importance may be
adjusted according to the noise level (i.e. the worse the data
quality is, or the sparser the sample is, the lower the informative
contents of the solution will be and the smoother the restored
distribution will be).
Under the assumption that the potential is separable, the proposed
method yields a unique mean density compatible with the measurement of
the velocity distribution perpendicular to the Galactic plane and the
corresponding measured density profile.
The discretized integral equation
From the knowledge of the level of errors in the measurements and
given the finite number of data points, it is possible to obtain
$f(w)$ by fitting the data with some model. A general approach
assumes that the solution can be described by its projection onto a
complete basis of functions with finite
support (here cubic splines) $\{e_{k}(w);k=1,\ldots,N\}$:
$$f(w)=\sum_{k=1}^{N}f_{k}\;e_{k}(w)\equiv{\mathbf{f}}\cdot{\mathbf{e}}\;.$$
(24)
where the $N$ parameters to fit are the weights $f_{k}$.
Real data correspond to discrete measurements $z_{i}$ and ${w}_{j}$ of $z$
and $w$ respectively. Following the non-parametric expansion
in Eq. (24), Eq. (1) becomes:
$$\rho_{i}\equiv\rho(z_{i})=\sum_{k=1}^{N}A_{i,k}(\alpha,z_{0})\;f_{k}\,,$$
(25)
with
$$A_{i,k}(\alpha,z_{0})=\int_{\sqrt{2\alpha z^{2}_{i}}}^{\infty}\frac{e_{k}(w)w}%
{\sqrt{w^{2}-2\alpha z^{2}_{i}}}\,{{\mathrm{d}}{w}}\,.$$
(26)
Since the relation between ${\mathbf{\tilde{\rho}}}\equiv\{\rho_{i}\}$
and ${\mathbf{f}}\equiv\{f_{k}\}$
is linear, Eq. (25) — the discretized
form of the integral equation (1) — can be written in a
matrix form:
$${\mathbf{\tilde{\rho}}}={\mathbf{A}}(\alpha,z_{0})\!\cdot\!{\mathbf{f}}$$
(27)
Here $z_{0}$ accounts for the relative offset in the position of the sun
with respect to the galactic plane as defined by the Hipparcos sample to
the measured density.
The Hipparcos sample gives access within a given sphere to a number of
stars for which the velocity, $w$, and/or the altitude, $z$, is known.
In other words the measurement yields a realization of the cumulative
height density distribution and the cumulative velocity distribution.
Indeed for each star at height $z$, within ${{\mathrm{d}}{z}}$ of the next star
(once they have been sorted in height), we can associate a volume
given by
$$\pi\,{{\mathrm{d}}{z}}\,R^{2}(1-z^{2}/R^{2})$$
(28)
where $R$ is the radius of the Hipparcos sphere; an alternative
estimated local density for this star is given by the inverse of this
volume. This density estimator yields a local density consistent
with that derived in section 3. An estimate of its cumulative
distribution is directly available from the measurements without
resorting to averaging nor binning. In order to project the modeled
distribution, $f$, into data space it is therefore best to model both
the cumulative density distribution, $R(z)$ and the cumulative
velocity distribution, $F(w)$, which satisfy respectively
$$R(z)=\int_{-\infty}^{z}\rho(z)\,{{\mathrm{d}}{z}}\,,\quad{\rm and}\quad F(w)=%
\int_{-\infty}^{w}f(w)\,{{\mathrm{d}}{w}}\,.$$
(29)
Now if $f(w)$ obeys Eq. (24), its cumulative distribution will satisfy
$$F(w)=\sum_{k=1}^{N}f_{k}\;E_{k}(w)\equiv{\mathbf{f}}\cdot{\mathbf{E}}\;,$$
(30)
where the basis of functions $\{E_{k}(w);k=1,\ldots,N\}$ is simply the
anti-derivative of $\{e_{k}(w)\}$ with respect to $w$.
Similarly the cumulative density distribution satisfies
$$\displaystyle R_{i}\equiv R(z_{i})=\sum_{k=1}^{N}\frac{1}{\sqrt{2\alpha}}C_{k}%
\left[\sqrt{2\alpha}(z_{i}-z_{0})\right]\;f_{k}\,,$$
$$\displaystyle\quad{\rm with}\quad C_{k}(Z)=\int_{0}^{Z}\int_{z}^{\infty}\frac{%
e_{k}(w)w}{\sqrt{w^{2}-z^{2}}}\,{{\mathrm{d}}{w}}\,{{\mathrm{d}}{z}}\,.$$
(31)
Note that for a cubic spline basis, the function $C_{k}$ is algebraic (though
somewhat clumsy). Eq. (30) and Eq. (31) provide
two non parametric means of relating the measurements, $R_{i}$ and $F_{i}$
to the underlying distribution, $f(w)$, with two extra parametres, $z_{0}$ and $\alpha$,
which one should adjust so that the distributions match.
Practical implementation
First we compute the best regularized fit of the cumulative velocity
distribution (we use $\chi^{2}$ fitting even though the noise is
Poissonian because it yields a simpler quadratic problem). We
therefore minimize
the quadratic form:
$$Q_{\mathrm{\mu}}({\mathbf{f}})={{(\tilde{{\mathbf{F}}}-{\mathbf{B}}\!\cdot\!{%
\mathbf{f}})}^{\bot}}\!\cdot\!{\mathbf{W}}\!\cdot\!(\tilde{{\mathbf{F}}}-{%
\mathbf{B}}\!\cdot\!{\mathbf{f}})+\mu\,{{{\mathbf{f}}}^{\bot}}\!\cdot\!{%
\mathbf{K}}\!\cdot\!{\mathbf{f}}\,,$$
(32)
where $\tilde{{\mathbf{F}}}$ is the measured cumulative velocity
distribution, ${\mathbf{K}}$ the discrete 2nd order differentiation
operator,
and ${\mathbf{B}}=\{E_{k}(w_{i})\}$ ; the
weight matrix ${\mathbf{W}}$ is the inverse of the covariance matrix of the
error in the
data and $\mu$ is the Lagrange parameter
tuning the level of regularization.
Error estimation is achieved via Monte Carlo simulation: the
regularization parameter is initialized at some fixed value (here
$\mu=10$); from this value a first cumulative distribution is
fitted. A draught corresponding to the same number of measured
velocity is generated and the procedure iterated. An estimate of the
error bars per point follows. The best fit ${\mathbf{f}}_{\mathrm{\mu}}$ which
minimizes $Q_{\mathrm{\mu}}$ in Eq. 32 is:
$${\mathbf{f}}_{\mathrm{\mu}}=({{{\mathbf{B}}}^{\bot}}\!\cdot\!{\mathbf{W}}\!%
\cdot\!{\mathbf{B}}+\mu\,{\mathbf{K}})^{-1}\!\cdot\!{{{\mathbf{B}}}^{\bot}}\!%
\cdot\!{\mathbf{W}}\!\cdot\!\tilde{{\mathbf{F}}}\,.$$
(33)
Next we
apply Eq. (31)
to project the distribution function
into cumulative density space. Calling ${\mathbf{C}}$ the matrix given
by its components $C_{i,k}=C_{k}\left[\sqrt{2\alpha}(z_{i}-z_{0})\right]/\sqrt{2\alpha}$,
we
then adjust the potential curvature $\alpha$
(together with the possible offset $z_{0}$) while minimizing
$$\displaystyle|{\mathbf{\tilde{R}}}-{\mathbf{C}}(\alpha,z_{0})\cdot{\mathbf{f}}%
_{\mu}|^{2}=$$
$$\displaystyle|{\mathbf{\tilde{R}}}-{\mathbf{C}}(\alpha,z_{0})\cdot({{{\mathbf{%
B}}}^{\bot}}\!\cdot\!{\mathbf{W}}\!\cdot\!{\mathbf{B}}+\mu\,{\mathbf{K}})^{-1}%
\!\cdot\!{{{\mathbf{B}}}^{\bot}}\!\cdot\!{\mathbf{W}}\!\cdot\!\tilde{{\mathbf{%
F}}}\,|^{2}\,,$$
(34)
with respect to $\alpha$ and $z_{0}$.
This
yields the local density of the
disk. Note that the last two steps could be carried simultaneously,
i.e. we could seek the triplet, $(\mu,\alpha,z_{0})$
which minimizes Eq. (34), or even better,
the unknown, $\mu,\alpha,z_{0}$ and $\{f_{k}\},\;k=1\cdots N$,
which minimize the quantity:
$$\displaystyle{{(\tilde{{\mathbf{F}}}-{\mathbf{B}}\!\cdot\!{\mathbf{f}})}^{\bot%
}}\!\cdot\!{\mathbf{W}}_{1}\!\cdot\!(\tilde{{\mathbf{F}}}-{\mathbf{B}}\!\cdot%
\!{\mathbf{f}})+$$
$$\displaystyle{{(\tilde{{\mathbf{R}}}-{\mathbf{C}}\!\cdot\!{\mathbf{f}})}^{\bot%
}}\!\cdot\!{\mathbf{W}}_{2}\!\cdot\!(\tilde{{\mathbf{R}}}-{\mathbf{C}}\!\cdot%
\!{\mathbf{f}})+\mu\,{{{\mathbf{f}}}^{\bot}}\!\cdot\!{\mathbf{K}}\!\cdot\!{%
\mathbf{f}}\,,$$
(35)
though
in practice the level of regularization, $\mu$, – within some range, only
affects the wings of the reconstructed density profile and this has no
consequence on the central density value, $\rho_{0}$ (a
supplementary difficulty
involved in minimizing Eq. (35)
would be to assess the relative weights between
errors in position and velocity).
Since our purpose was primarily to validate the parametric approach
described in sections 4.1 and 4.2, we also minimized
$$|{\mathbf{\tilde{\rho}}}-{\mathbf{A}}(\alpha,z_{0})\cdot{\mathbf{f}}_{\mu}|^{2}$$
(36)
instead of Eq. (34) using the nearest neighbour density estimator
described in section 3.
This lead to non parametric estimator for
sample h125 given by
$\rho_{0}=0.07-0.08~{}M_{\sun}~{}{\rm pc}^{-3}$ with $z_{0}=13\,\,{\rm pc}$.
The analysis of sample H100 gives
$\rho_{0}=0.08-0.10~{}M_{\sun}~{}{\rm pc}^{-3}$.
In practice, the parametric and non parametric analysis
therefore lead to
statistically consistent answer.
4.4 Radial force correction
The $K_{z}$ (vertical forces) determination constrains the local mass
density: precisely $\partial^{2}\Phi_{total}\over\partial z^{2}$.
A complete determination of the Poisson
equation should include the $K_{r}$ (radial force) component. This
second contribution is small and can be estimated from the A and B
Oort’s constants:
$$\Delta\Phi_{total}={\partial^{2}\Phi_{total}\over\partial z^{2}}+2(B^{2}-A^{2}%
)=4\pi G\rho_{total}$$
(37)
From the dynamical analysis of the present Hipparcos sample, we
have obtained 0.076 for the $\partial^{2}\Phi_{total}\over\partial z^{2}$ term. The term including A and B Oort’s constants ranges
between $\pm 0.004$; it is probably positive corresponding to a
rotation curve locally decreasing. This term is small and
we assume it is zero to simplify the discussion.
5 Conclusions
The consequence of estimations given in previous sections is to fix
the total mass density in the galactic disc in the neighbourhood of
the Sun in the range 0.065-0.10 $M_{\odot}~{}{\rm pc}^{-3}$ with a most
probable value by 0.076 $M_{\odot}~{}{\rm pc}^{-3}$. In conclusion we
review below the consequences of this very low value in terms of local
dynamical mass versus observed density, in terms of global
galaxy modeling and in terms of dark matter distribution.
5.1 Dynamical mass versus known matter
Previous dynamical determinations of the local mass density used to be
based on star counts a few hundred parsecs away from the galactic
plane providing constraints for the local column density integrated
over a substantial layer. Comparisons were tentatively made with the
observed density of known galactic components, namely visible stellar
populations, stellar remnants, ISM. So, on either side of the
comparison there were assumptions on the vertical distribution of
matter involved.
For the first time, the present determination is based on strictly
local data inside a sphere of more than 100 parsecs radius and
gives access directly to the total mass density around the Sun in
the galactic plane.
On the side of the known matter, the current situation is the
following: the main contributors are main sequence stars of the
disc, stellar remnants and the interstellar medium .
1) Recent determinations of the stellar disc luminosity function
have a maximum near $M_{v}$=12 and decreases quickly beyond.
Counting nearby stars Wielen et al. (1983) estimate the total mass
density of main sequence stars in the solar neighbourhood as
0.037 plus 0.008 in the form of giants and white dwarfs. Since then the
luminosity function below $M_{v}$ = 13 has been revised downwards (Reid et
al 1995, Gould et al 1996). Thus main sequence stars are expected to
contribute by 0.027-0.033$~{}M_{\odot}~{}{\rm pc^{-3}}$.
This is the most reliable part of the known matter, yet some extra
uncertainty might result from the non detection of low mass binaries.
Preliminary results based on Hipparcos parallaxes seem to imply
that stellar distances are on the average slightly larger than expected
which would mean an even smaller density (Jahreiss and Wielen, 1997).
2) Stellar remnants dominated by white dwarfs should contribute by
an additional $0.015\pm 0.005~{}M_{\odot}~{}{\rm pc^{-3}}$. This
figure is compatible with observational determination of the WD’s
luminosity function ([Sion & Liebert 1977], [Ishida et al 1982]). It
does not conflict with state of the art scenarios of star
formation.
3) The local mass density of the ISM estimated from HI and
$\rm{H_{2}}$ abundances. $\rm{H_{2}}$ is deduced from
$\rm{CO}$/$\rm{H_{2}}$ ratio. This estimation suffers very large
uncertainties. According to [Combes (1991)], a contribution about
0.04 is reasonable, but errors by a factor two or more are not
excluded. This is certainly the weakest link of this analysis.
So, 0.085 $M_{\odot}~{}{\rm pc^{-3}}$ is probably an acceptable
figure for the known mass. But the uncertainty is larger than
0.02. It does not make sense to overcomment about the observed
mass density being larger than the dynamical estimate: for one
thing the discrepancy is far below the error bars, for the other
thing there has been little effort dedicated so far to estimate
properly the smallest density compatible with observations, the
attention being always focused on upper bounds. For the moment the
only conclusion is that the density of known matter has a lower
limit of say 0.065 $M_{\odot}~{}{\rm pc^{-3}}$. In term of mass
discrepancy controversy, the main uncertainty is now on the
observation side.
5.2 Galactic mass model
In the following we adopt a simple, two component modelisation of
the galactic disc including stars and interstellar medium
(double-exponential and exponential-sech-square laws). The
formalism is standard, it is for instance the one adopted by
Gould et al (1996) (Eq. 1.1 in their paper).
A good convergence is now obtained on the main parameters of the
disc populations although there is no agreement as to the exact
decomposition in populations (see for example, [Reid & Gilmore 1983], Robin et
al 1986ab-1996, Haywood 1997ab, Gould et al 1996, etc..).
Based on the discussions quoted above the tentative structural
parameters of our model are close to Gould’s (see also [Sackett, 1997]):
stellar disc scale length 2.5 kpc, scale height 323 pc; thick disc
scale length 3.5 kpc, scale height (exponential) 656 pc. 20$\%$ of
the local stellar density is in the thick disc. The error on the
scale length has little effect on the disc mass $M_{d}$:
$M_{d}(2.5{\rm\,kpc})/M_{d}(3.5{\rm\,kpc})=1.34$ (assuming $R_{0}$=8.5 kpc)
and a factor 1.26 on the amplitude of the corresponding velocity
curves.
Adopting a total stellar mass density $\rho_{\star}=.043M_{\odot}\,{\rm pc}^{-3}$, the surface density at $R_{0}$ is then
$\Sigma_{0}=33.4~{}M_{\odot}\,{\rm pc}^{-2}$
The contributions of the bulge and stellar halo are small beyond
3 kpc of the galactic center. They are neglected in this discussion.
The interstellar matter is accounted for by a double exponential
disc 2500/80 pc its local density is set to 0.04 $M_{\odot}~{}{\rm pc}^{-3}$.
The total column density due to gas and stellar component is
$\Sigma_{0}=40~{}M_{\odot}~{}{\rm pc}^{-2}$.
A dark halo is necessary to maintain a flat rotation curve beyond
the solar galactic radius. It is modeled with a Miyamoto spheroid
that allows flattening.
This mass model is fitted to the rotation curve setting the solar
galactic radius to $R_{0}$=8.5 kpc and the velocity curve at $R_{0}$ to
$220\,{\rm km\,s^{-1}}$. The velocity curve is assumed flat beyond
5 kpc of the center to 20 kpc.
The first consequence is that the stellar disc (stars+remnants+gas)
does not contribute for more than half the mass implied by the
rotation curve at $R_{0}$=8.5 kpc: the galactic disc is far from
maximal. Even with $R_{0}$=7.5 kpc, a maximal disc would mean a
rotation curve plateau as low as 164 km s${}^{-1}$. This is excluded
by inner HI and CO rotation curves.
Filling the galactic mass distribution with a strictly spherical
dark matter halo required by the rotation curve, the dark matter
local density comes out as
$\rho_{dark\,halo}(R_{0},z=0)=0.007~{}M_{\odot}~{}{\rm pc}^{-3}$. Such a
local density of dark matter is well compatible with the above
quoted uncertainties of the dynamical and known mass. However, any
attempt to flatten the mass distribution of the dark component
result in an increase of the local halo density. Under the most
extreme hypotheses, that is considering the interstellar matter
contribution negligible and giving the stellar component a total
density $0.043~{}M_{\odot}~{}{\rm pc}^{-3}$ with the same scale length
and adopting our best dynamical estimate at its face value (0.076)
the acceptable halo scale height cannot go below 2150 pc.
So, there is just room for a spherical halo (local density:
$0.008~{}M_{\odot}~{}{\rm pc}^{-3}$). Extreme changes of the parameters
are needed to permit a significantly flatter dynamical halo. There is
no room for an important amount of dark matter in the disc. Dark
matter models assuming that the dark matter is in form of a flat
component related to a flat fractal ISM should be modified to
equivalent models where dark matter is in form of fractal structures
distributed in the halo (Pfenniger et al 1994ab, Gerhard & Silk
1996).
5.3 Summary of conclusions
1.
The potential well across the galactic plane has been traced
practically hypothesis-free and model-free, it turns out to be shallower
than expected.
2.
The local dynamical volume density comes out as
$0.076\pm 0.015~{}M_{\odot}~{}{\rm pc}^{-3}$ a value well compatible with
all existing observations of the known matter.
3.
Building a disc-thick disc mass model compatible with this
constraint as well as previous determinations of the local surface
density, it is shown that such a disc cannot be maximal, a massive
halo of dark matter is required.
4.
The dark halo should be spherical or nearly spherical in order
for its local density to remain within the range permitted between the
known matter density and the current determination of the dynamical
density.
5.
There is no room left for any disk shaped component of dark matter.
Acknowledgements.
Les travaux de MC à l’IUP de Vannes ont été rendus possibles
grâce à une aide spécifique de la Mission Scientifique et
Technique (DSPT3) et à l’assistance permanente de Robert Nadot.
CP thanks E. Thiébaut and O. Gerhard for valuable discussions and
acknowledges funding from the Swiss NF.
References
[Arenou et al, 1995]
Arenou F., Lindegren L., Froeschle F., Gomez A.E., Turon C., Perryman
M.A.C, Wielen R., 1995, A&A 304, 52.
[Bahcall, 1984a]
Bahcall J.N., 1984a, ApJ 276, 169
[Bahcall, 1984b]
Bahcall J.N., 1984b, ApJ 287, 926
[Bahcall et al, 1992]
Bahcall J.N., Flynn C., Gould A., 1992, ApJ 389, 234
[Bienaymé et al (1987)]
Bienaymé O., Robin A.C., Crézé M., 1987, A&A 180, 94
[Chereul et al, 1997]
Chereul E., Crézé M., Bienaymé O., 1997 (in preparation)
[Combes (1991)]
Combes F., 1991, ARAA 29, 195
[Crézé et al, 1989]
Crézé M., Robin A.C., Bienaymé O., 1989, A&A 211, 1
[ESA, 1992]
ESA, 1992,
The Hipparcos Input Catalogue, ESA SP-1136
[ESA, 1997]
ESA, 1997,
The Hipparcos Catalogue, ESA SP-1200
[Flynn & Fuchs (1994)]
Flynn C., Fuchs B., 1994, MNRAS 270, 471
[Fuchs & Wielen (1993)]
Fuchs B., Wielen R., 1993, Back to the Galaxy, eds Holt and Verter, 580
[Gerhard & Silk, 1996]
Gerhard O., Silk J., 1996, ApJ 472, 34
[Gould et al, 1996]
Gould A., Bahcall J.N., Flynn C., 1996, ApJ 465, 759
[Haywood et al, 1997a]
Haywood M., Robin A.C, Crézé M., 1997a, A&A 320, 428
[Haywood et al, 1997b]
Haywood M., Robin A.C, Crézé M., 1997b, A&A 320, 440
[Ishida et al 1982]
Ishida K., Mikami T., Nogushi T., Maehara H, 1982, PASJ, 34, 381
[Jahreiss & Wielen]
Jahreiss H., Wielen R., 1997,
’Presentation of The Hipparcos and Tycho Catalogues and first
astrophysical results of the Hipparcos astrometry mission’ (ESA) p.54
[Kapteyn (1922)]
Kapteyn J.C., 1922, ApJ 55, 302
[Kerr & Lynden-Bell (1986)]
Kerr F.J., Lynden-Bell D., 1986, MNRAS 221, 1023
[Kuijken & Gilmore (1989)]
Kuijken K., Gilmore G., 1989, MNRAS 239, 650
[Kuijken (1995)]
Kuijken K., 1995, Stellar populations,
IAU Symp. 164, 195
[Oort (1932)]
Oort J.H., 1932, BAN 6, 249
[Oort, 1960]
Oort J.H., 1960, BAN 15, 45
[Pichon & Thiébaut, 1997]
Pichon C., Thiébaut E., 1997, MNRAS (submitted)
[Pfenniger et al, 1994a]
Pfenniger D., Combes F., Martinet L., 1994a, A&A 285, 79
[Pfenniger et al, 1994b]
Pfenniger D., Combes F., 1994b, A&A 285, 94
[Ratnatunga & Casertano, 1991]
Ratnatunga K.U., Casertano S., 1991, AJ 101, 1075.
[Reid & Gilmore 1983]
Reid N., Gilmore G.,1983, MNRAS 202, 1025
[Reid et al (1995)]
Reid N., Hawley S. L., Gizis J.E., 1995, AJ 110(4), 1838
[Robin & Crézé, 1986a]
Robin A.C., Crézé M., 1986a, A&A 157, 71
[Robin & Crézé, 1986b]
Robin A.C., Crézé M., 1986b, A&AS 64, 53
[Robin et al, 1996]
Robin A.C., Haywood M., Crézé M., Ojha D.K., Bienaymé O.,1996, A&A 305, 125
[Turon & Crifo, 1986]
Turon, C., Crifo, F., 1986, in: ’IAU Highlights of Astronomy’, Vol. 7,
Swings J.P. (ed.), 683
[Sackett, 1997]
Sackett P.D., 1997, ApJ 483, 10
[Sion & Liebert 1977]
Sion E.M., Liebert J.W., 1977, ApJ 213, 468
[Wahba & Wendelberger, 1979]
Wahba G., Wendelberger J., 1979, Monthly Weather Review 108, 1122
[Wielen et al (1983)]
Wielen R., Jahreiss H., Kruger R., 1983, The nearby stars and the
stellar luminosity function, IAU Coll. 76, 163 |
An SU(2) gauge principle for the Cosmic Microwave Background: Perspectives on the Dark Sector of the Cosmological Model
Ralf Hofmann
Institut für Theoretische Physik, Universität Heidelberg,
Philosophenweg 16, D-69120 Heidelberg, Germany;
[email protected]
We review consequences for the radiation and dark sectors of the cosmological model arising
from the postulate that the Cosmic Microwave Background (CMB) is governed by an SU(2) rather than a U(1) gauge principle. We also speculate on the possibility of
actively assisted structure formation due to the de-percolation of
lump-like configurations of condensed ultralight axions with a Peccei-Quinn scale comparable to the Planck mass. The chiral-anomaly induced
potential of the axion condensate
receives contributions from SU(2)/SU(3) Yang-Mills factors of hierarchically
separated scales which act in a screened (reduced) way in confining phases.
1 Introduction
Judged by decently accurate agreement of cosmological parameter values extracted from (i) large-scale structure observing campaigns on galaxy- and
shear-correlation functions as well as redshift-space distortions towards a redshift of unity by photometric/spectroscopic surveys (sensitive
to Baryonic Accoustic Oscillations (BAO), the evolution of Dark Energy, and non-linear structure growth), for recent, ongoing, and future projects
see, e.g. [1, 2, 3, 4, 5],
(ii) fits of various CMB angular power spectra based on frequency-band optimised intensity and polarisation data collected by satellite
missions [6, 7, 8], and (iii) fits to cosmologically local luminosity-distance redshift data for Supernovae Ia (SNeIa)
[9, 10] the spatially flat standard Lambda Cold Dark Matter ($\Lambda$CDM) cosmological model is a good, robust
starting point to address the evolution of our Universe. About twenty years ago, the success of this model has
triggered a change in paradigm in accepting a present state of accelerated expansion induced by an essentially dark Universe
made of 70 % Dark Energy and 25 % Dark Matter.
While (i) and (ii) are anchored on comparably large cosmological standard-ruler co-moving distance scales, the sound horizons $r_{s,{\rm d}}$ and $r_{s,*}$,
which emerge at baryon-velocity freeze-out and recombination, respectively, during the epoch of CMB decoupling and therefore refer to very-high-redshift
physics ($z>1000$), (iii) is based on direct distance ladders linking to host galaxies at redshifts of say, $0.01<z<2.3$, which
do not significantly depend on the assumed model for Dark Energy [11] and are robust against sample variance,
local matter-density fluctuations, and directional bias [12, 13, 14, 15].
Thanks to growing data quality and
increased data-analysis sophistication to identify SNeIa [16] and SNeII [17] spectroscopically, to establish precise and independent
geometric distance indicators [18, 19] (e.g., Milky Way parallaxes, Large Magellanic Cloud (LMC) detached eclipsing binaries,
and masers in NGC 4258), tightly calibrated
period-to-luminosity relations for LMC cepheids and cepheids in SNeIa host galaxies [19], the use of
the Tip of the Red Giant Branch (TRGB) [20, 21] or the Asymptotic Giant Branch (AGB, Mira) [22] to connect to the distance ladder
independently of cepheids, the high value of the basic cosmological parameter
$H_{0}$ (Hubble expansion rate today) has (see [21], however), over the last decade and within $\Lambda$CDM,
developed a tension of up to $\sim 4.4\,\sigma$ [19] in SNeIa distance-redshift fits using the LMC geometrically
calibrated cepheid distance ladder compared to its low value
extracted from statistically and systematically accurate medium-to-high-l CMB angular power spectra [7, 8] and the BAO method
[3]. Both, the CMB and BAO probe the evolution of small radiation and matter density fluctuations in global cosmology: Fluctuations are
triggered by an initial, very-high-redshift (primordial) spectrum of scalar/tensor curvature
perturbations which, upon horizon entry, evolve linearly [23] up to late times when
structure formation generates non-linear contributions in the galaxy spectra [2, 3] and
subjects the CMB to gravitational lensing [24].
Recently, a cosmographic way of extracting $H_{0}$ through time delays of
strongly lensed high-redshift quasars, see e.g. [25, 26], almost matches the
precision of the presently most accurate SNeIa distance-redshift fits [19] – $H_{0}=71.9^{+2.4}_{-3.0}\,$ km s$\mbox{}^{-1}$Mpc$\mbox{}^{-1}$ vs. $H_{0}=(74.22\pm 1.82)\,$ km s$\mbox{}^{-1}$Mpc$\mbox{}^{-1}$ –, supporting
a high local value of the Hubble constant and rendering the local-global tension even
more significant [27]. The high local value of $H_{0}=(72\cdots 74)\,$ km s$\mbox{}^{-1}$Mpc$\mbox{}^{-1}$ [19, 27] (compared to the global value of
$H_{0}=(67.4\pm 0.5)\,$ km s$\mbox{}^{-1}$Mpc$\mbox{}^{-1}$ [8]) is stable to sources
of systematic uncertainty such as line-of-sight effects, peculiar motion (stellar kinematics), and assumptions made in the
lens model. There is good reason to expect that an improved localisation
of sources for gravitational-wave emission without an electromagnetic counterpart and the
increase of statistics in gravitational-wave events accompanied by
photon bursts (standard sirenes) within specific host galaxies will lead to luminosity-distance-high-redshift data
producing errors in $H_{0}$ comparable of those of [19],
independently of any distance-ladder calibration [28], see also [29].
Apart from the Hubble tension,
there are smaller local-global but possibly also tensions between BAO and the CMB
in other cosmological parameters such as the
amplitude of the density fluctuation power spectrum ($\sigma_{8}$)
and the matter content ($\Omega_{m}$), see [30]. Finally, there are persistent large-angle anomalies
in the CMB, already seen by the Cosmic Background Explorer (COBE) and strengthened by the Wilkinson Microwave Anisotropy Probe (WMAP) and Planck satellite,
whose ’atoms’ (a) lack of correlation in the TT two-point function, (b) a rather significant alignment of
the low multipoles ($p$-values of below 0.1 %), and (c) a dipolar modulation, which is independent of the multipole alignment (b),
indicate a breaking of statistical isotropy at low angular resolution, see [31] for a comprehensive
and complete review.
The present work intends to review and discuss a theoretical framework addressing a possibility to
resolve the above-sketched situation. The starting point is to subject the CMB to the thermodynamics of
an extended gauge principle in replacing the conventional group U(1) by SU(2). Motivated [32]
by explaining an excess in CMB line temperature at radio frequencies, see [33] and references therein,
this postulate implies a modified temperature-redshift relation which places CMB recombination to a redshift of
$\sim 1700$ rather than $\sim 1100$ and therefore significantly reduces the density of matter at that epoch.
To match $\Lambda$CDM at low $z$, one requires a release of Dark Matter from early Dark Energy
within the dark ages at a redshift $z_{p}$: lumps and vortices, formely tighly correlated within a condensate of ultralight axion particles, de-percolate into
independent pressureless (and selfgravitating) solitons due to cosmological expansion, thereby contributing actively to
non-linear structure formation at low $z$. A fit to CMB angular power spectra of a cosmological model, which incorporates these SU(2)
features on the perfect-fluid level but neglects low-$z$ radiative effects in SU(2) [34],
over-estimates TT power on large angular scales but generates a precise fit to the data for
$l\geq 30$. With $z_{p}\sim 53$ a locally favoured value of $H_{0}\sim 74\,$km s${}^{-1}$ Mpc${}^{-1}$ and a low baryon density
$\omega_{b,0}\sim 0.017$ are obtained. Moreover, the conventionally extracted near scale invariance
of adiabatic, scalar curvature perturbations comes out to be significantly
broken by an infrared enhancement of their power spectrum. Finally, the fitted redshift $\sim 6$ of
re-ionisation is compatible with the detection of the Gunn-Peterson-trough in the spectra of high-$z$
quasars [36] and distinctly differs from the high value
of CMB fits to $\Lambda$CDM [7, 8].
As of yet, there are loose ends to the SU(2) based scenario. Namely, the physics of
de-percolation requires extra initial conditions for matter density
fluctuations at $z_{p}$. In the absence of a precise modelling of the ’microscopics’ of the associated soliton
ensembles111Lump sizes could well match those of galactic dark-matter
halos, see Sec. 4. it is only a guess that
these fluctuations instantaneously follow the density fluctuations of primordial Dark Matter as assumed
in [34]. Moreover, it is necessary to investigate to what extent profiles of the axion
field (lumps of localised energy density) actively seed non-linear structure formation and whether their role in
galactic halo formation and baryon accretion meets the observational constraints (Tully-Fisher relation, etc.).
Also, radiative effects on thermal photon propagation at low
$z$ [37], which are expected to contribute or even explain the above-mentioned
’atoms’ (a), (b), and (c) of the CMB large-angle anomalies, see [38], and which could reduce the
excess in low-$l$ TT power of [34] to realistic levels, need to be incorporated into
the model. They require analyses in terms of Maxwell’s multipole-vector
formalism [31] and/or other robust and intuitive statistics to
characterise multipole alignment and the large-angle suppression of TT.
This review-type paper is organised as follows. In Sec. 2 we briefly discuss the main
distinguishing features between the conventional
$\Lambda$CDM model and cosmology based on a CMB which obeys deconfining SU(2) Yang-Mills thermodynamics. A presentation of
our recent results on angular-power-spectra fits to Planck data is carried out Sec. 3, including a discussion of
the parameters $H_{0}$, $n_{s}$, $\sigma_{8}$, $z_{\rm re}$, and $\omega_{\text{b,0}}$. In Sec. 4 we interpret the rough
characteristics of the Dark Sector employed in [34] to
match $\Lambda$CDM at low $z$. In particular, we point out that the value of $z_{p}$ seems to be consistent
with the typical dark-matter densities in the Milky Way. Finally, in
Sec. 5 we sketch what needs to be done to arrive at a solid, observationally well
backed-up judgement of whether SU(2)${}_{\rm CMB}$ based cosmology (and its extension
to SU(2) and SU(3) factors of hierarchically larger Yang-Mills scales including their nonthermal phase transitions) may provide a future
paradigm to connect local cosmology with the very early Universe. Throughout the article, super-natural units $\hbar=k_{B}=c=1$ are used.
2 SU(2)${}_{\rm CMB}$ vs. conventional CMB photon gas in $\Lambda$CDM
The introduction of an SU(2) gauge principle for the description of the CMB is motivated
theoretically by the fact that the deconfining thermodynamics of such a Yang-Mills theory
exhibits a thermal ground state, composed of densely packed (anti)caloron [39]
centers with overlapping peripheries [40, 41], which breaks SU(2) to U(1) in terms of an adjoint
Higgs mechanism [42]. Therefore, the spectrum of excitations consists
of one massless gauge mode, which can be identified with the CMB photon, and two
massive vector modes of a temperature-dependent mass on tree level: thermal
quasi-particle excitations. The interaction between these excitations is feeble [42].
This is exemplified by the one-loop polarisation tensor of the
massless mode [43, 37, 44]. As
a function of temperature, polarisation effects peak at about twice the critical temperature $T_{c}$ for
the deconfining-preconfining transition. As a function of increasing photon momentum,
there are regimes of radiative screening/antiscreening, the latter being subject to an exponential fall-off
[37]. At the phase boundary ($T\sim T_{c}$) electric monopoles [42], which occur as isolated and unresolved
defects deeply in the deconfining phase, become massless by virtue of screening due to transient dipoles [45] and therefore condense to endow the
formely massless gauge mode with a quasiparticle Meissner mass $m_{\gamma}$. This mass rises critically (with mean-field exponent)
as $T$ falls below $T_{c}$ [42]. Both, (i) radiative screening/antiscreening of massless modes and (ii) their Meissner effect are important
handles in linking SU(2)${}_{\textnormal{\tiny{cmb}}}$ to the CMB: While (i)
induces large-ange anomalies into the TT correlation [38] and
contributes dynamically to the CMB dipole [46, 47] (ii) gives rise to a nonthermal
spectrum of evanescent modes222That the deep Rayleigh-Jeans regime is indeed subject to classical wave
propagation is assured by the fact that wavelengths then are greater than the spatial scale $s\equiv\pi T|\phi|^{-2}$, separating a(n) (anti)caloron
center from its periphery where its (anti)selfdual gauge field is that of a dipole [41]. The expression for $s$ contains
the modulus $|\phi|=\sqrt{\Lambda_{\textnormal{\tiny{cmb}}}^{3}/(2\pi T)}$ of the emergent, adjoint Higgs field $\phi$ ($\Lambda_{\textnormal{\tiny{cmb}}}\sim 10^{-4}\,$eV the Yang-Mills scale
of SU(2)${}_{\textnormal{\tiny{cmb}}}$), associated with densely packed (anti)caloron centers,
and, explicitely, temperature $T$. for frequencies $\omega<m_{\gamma}$ once $T$ falls below $T_{c}$. This theoretical anomaly of the blackbody spectrum
in the Rayleigh-Jeans regime can be considered to explain the excess in CMB radio power below 1 GHz, see [33] and references therein,
thereby fixing $T_{c}=2.725\,$K and, as a
consequence of $\lambda_{c}=13.87=\frac{2\pi T_{c}}{\Lambda_{\textnormal{\tiny{cmb}}}}$ [42], the
Yang-Mills scale of SU(2)${}_{\textnormal{\tiny{cmb}}}$ to $\Lambda_{\textnormal{\tiny{cmb}}}\sim 10^{-4}\,$eV [32].
Having discussed the low-frequency deviations of SU(2)${}_{\rm CMB}$ from the conventional Rayleigh-Jeans
spectrum, which fix the Yang-Mills scale and associate with large-angle anomalies, we would now like to
review its implications for the cosmological model. Of paramount importance for the set-up of such a model
is the observtion that SU(2)${}_{\textnormal{\tiny{cmb}}}$ implies a modified temperature ($T$)-redshift
($z$) relation for the CMB which is derived from energy conservation of the
SU(2)${}_{\textnormal{\tiny{cmb}}}$ fluid in the deconfining phase in an FLRW
universe with scale factor $a$ normalised to unity today [48]. Denoting by $T_{c}=T_{0}=2.725\,$K [32]
the present CMB baseline temperature [6] and by $\rho_{{\rm SU(2)}_{\textnormal{\tiny{cmb}}}}$ and $P_{{\rm SU(2)}_{\textnormal{\tiny{cmb}}}}$ energy density and
pressure, respectively, of SU(2)${}_{\textnormal{\tiny{cmb}}}$, one has
$$a\equiv\frac{1}{z+1}=\exp\left(-\frac{1}{3}\log\left(\frac{s_{{\rm SU(2)}_{%
\textnormal{\tiny{cmb}}}}(T)}{s_{{\rm SU(2)}_{\textnormal{\tiny{cmb}}}}(T_{0})%
}\right)\right)\,,$$
(1)
where the entropy density $s_{{\rm SU(2)}_{\textnormal{\tiny{cmb}}}}$ is defined by
$$s_{{\rm SU(2)}_{\textnormal{\tiny{cmb}}}}\equiv\frac{\rho_{{\rm SU(2)}_{%
\textnormal{\tiny{cmb}}}}+P_{{\rm SU(2)}_{\textnormal{\tiny{cmb}}}}}{T}\,.$$
(2)
For $T\gg T_{0}$, Eq. (1) simplifies to
$$T=\left(\frac{1}{4}\right)^{1/3}T_{0}(z+1)\approx 0.63\,T_{0}(z+1)\,.$$
(3)
For arbitrary $T\geq T_{0}$, a multiplicative deviation $S(z)$ from linear scaling in $z+1$ can be introduced as
$${\cal S}(z)=\left(\frac{\rho_{{\rm SU(2)}_{\textnormal{\tiny{cmb}}}}(z=0)+P_{{%
\rm SU(2)}_{\textnormal{\tiny{cmb}}}}(z=0)}{\rho_{{\rm SU(2)}_{\textnormal{%
\tiny{cmb}}}}(z)+P_{{\rm SU(2)}_{\textnormal{\tiny{cmb}}}}(z)}\frac{T^{4}(z)}{%
T^{4}_{0}}\right)^{1/3}\,.$$
(4)
Therefore,
$$T={\cal S}(z)\,T_{0}(z+1)\,.$$
(5)
Fig. 1 depicts function ${\cal S}(z)$.
Amusingly, the asymptotic $T$-$z$ relation of Eq. (3) also holds for the relation between
$\rho_{{\rm SU(2)}_{\textnormal{\tiny{cmb}}}}(z)$ and the conventional CMB energy density $\rho_{\gamma}(z)$ in $\Lambda$CDM
(the energy density of a thermal U(1) photon gas, using the $T$-$z$ relation $T=T_{0}(z+1)$). Namely,
$$\rho_{{\rm SU(2)}_{\textnormal{\tiny{cmb}}}}(z)=4\,\left(\frac{1}{4}\right)^{4%
/3}\rho_{\gamma}(z)=\left(\frac{1}{4}\right)^{1/3}\rho_{\gamma}(z)\quad(z\gg 1%
)\,.$$
(6)
Therefore, the (gravitating) energy density of the CMB in SU(2)${}_{\textnormal{\tiny{cmb}}}$ is, at the same redshift $z\gg 1$,
by a factor of $\sim 0.63$ smaller than that of the $\Lambda$CDM model even though there are eight (two plus two times three)
gauge-mode polarisations in SU(2)${}_{\textnormal{\tiny{cmb}}}$ and only two such polarisations in the U(1) photon gas.
Not yet considering linear and next-to-linear perturbations
in SU(2)${}_{\textnormal{\tiny{cmb}}}$ to shape typical CMB large-angle anomalies in terms of
late-time screening/antiscreening effects [38], Eq. (5) has
implications for the Dark Sector if one wishes to maintain the successes of the standard cosmological
model at low $z$ where local cosmography and fits to distance-redshift curves produce a
consistent framework within $\Lambda$CDM. As it was shown in [34], the assumption
of matter domination at recombination, which is not unrealistic even for SU(2)${}_{\textnormal{\tiny{cmb}}}$ [34], implies that
$$z_{\Lambda{\rm CDM},*}\sim\left(\frac{1}{4}\right)^{1/3}z_{{\rm SU(2)}_{%
\textnormal{\tiny{cmb}}},*}\,$$
(7)
and, as a consequence,
$$\Omega_{\Lambda{\rm CDM},m,0}\approx 4\,\Omega_{{\rm SU(2)}_{\textnormal{\tiny%
{cmb}}},m,0}\,,$$
(8)
where $\Omega_{m,0}$ denotes today’s density parameter for nonrelativistic
matter (Dark Matter plus baryons), and $z_{*}$ is the redshift of CMB photon decoupling in either model. Since the matter sector of the
SU(2)${}_{\textnormal{\tiny{cmb}}}$ model, as roughly represented by Eq. (8), contradicts $\Lambda$CDM at low $z$ one needs to allow for a
transition between the two somewhere in the dark ages as $z$ decreases. In [34] a simple model, where the transition is sudden at a
redshift $z_{p}$ and maintains a small dark-energy residual, was introduced: a coherent axion field – a dark-energy like condensate of ultralight axion particles,
whose masses derive from U(1)${}_{A}$ anomalies [49, 50, 51, 52, 53] invoked by the topological
charges of (anti)caloron centers in the thermal ground state of SU(2)${}_{\rm CMB}$ and, in a screened way, SU(2)/SU(3) Yang-Mills
factors of higher scales – releases solitonic lumps by de-percolation due to the Universe’s expansion. Accordingly, we have
$$\Omega_{\rm ds}(z)=\Omega_{\Lambda}+\Omega_{\rm pdm,0}(z+1)^{3}+\Omega_{\rm edm%
,0}\left\{\begin{array}[]{lr}\left(z+1\right)^{3}\,,&z<z_{p}\\
\left(z_{p}+1\right)^{3}\,,&z\geq z_{p}\end{array}\right.\,.$$
(9)
Here $\Omega_{\Lambda}$ and $\Omega_{\rm pdm,0}+\Omega_{\rm edm,0}\equiv\Omega_{\rm cdm,0}$ represent today’s
density parameters for Dark Energy and Dark Matter, respectively, $\Omega_{\rm pdm,0}$ refers to primordial Dark Matter (that is, dark
matter that existed before the initial redshift of $z_{i}\sim 10^{4}$ used
in the CMB Boltzmann code) for all $z$ and $\Omega_{\rm edm,0}$ to emergent Dark Matter for $z<z_{p}$. In [34] the
initial conditions for the evolution
of density and velocity perturbations of the emergent Dark-Matter portion at $z_{p}$ are, up to a rescaling of order unity,
assumed to follow those of the primordial Dark Matter.
3 SU(2)${}_{\textnormal{\tiny{cmb}}}$ fit of cosmological parameters to Planck data
In [34] a simulation of the new cosmological model subject to SU(2)${}_{\textnormal{\tiny{cmb}}}$ and the
Dark Sector of Eq. (9) was performed using a modified version of the Cosmic Linear Anisotropy Solving System (CLASS) Boltzmann code
[54]. Best fits to the 2015 Planck data [8] on the angular power spectra of the two-point correlation functions
temperature – temperature (TT), electric-mode polarisation – electric-mode polarisation (EE), and temperature – electric-mode polarisation (TE), subject to typical
likelihood functions used by the Planck collaboration, were performed. Because temperature perturbations
can only be coherently propagated by (low-frequency) massless modes in SU(2)${}_{\textnormal{\tiny{cmb}}}$ [41] the propagation of the (massive)
vector modes was excluded in one version of the code (physically favoured). Also, entropy conservation in e${}^{+}$e${}^{-}$ annihilation invokes a slightly different
counting of relativistic degrees of freedom in SU(2)${}_{\textnormal{\tiny{cmb}}}$ for this epoch [55]. As a consequence, we have a $z$-dependent
density parameter for (massless) neutrinos given as [34]
$$\Omega_{\nu}(z)=\frac{7}{8}N_{\rm eff}\,\left(\frac{16}{23}\right)^{\frac{4}{3%
}}\Omega_{{\rm SU(2)}_{\textnormal{\tiny{cmb}}},\gamma}(z)\,,$$
(10)
where $N_{\rm eff}$ refers to the effective number of neutrino flavours (or any other extra relativistic, free-streaming, fermionic species), and $\Omega_{{\rm SU(2)}_{\textnormal{\tiny{cmb}}},\gamma}$ is the density parameter associated with the massless
mode only in the expression for $\rho_{{\rm SU(2)}_{\textnormal{\tiny{cmb}}}}(z)$ of Eq. (6). The value $N_{\rm eff}=3.046$ of the Planck result was
used as a fixed input in
[34]. The statistical goodness of the best fit was found to be comparable to the one obtained by the Planck collaboration [8],
see Figs. 2, Figs. 3, and Figs. 4 as well as the lower part of Tab. 1. There
is an excess of power in TT for $7\leq l\leq 30$, however.
This excess could be attributable to the omission of radiative effects in the low-$z$ propagation of the massless mode.
The consideration of the modified dispersion law into the CMB Boltzmann code presently is under way.
The following table was obtained in [34]:
We would like to discuss the following parameters (The value of $z_{p}$ will be discussed in Sec. 4.): (i) $H_{0}$, (ii) $n_{s}$, (iii) $\sigma_{8}$, (iv) $z_{\rm re}$,
and (v) $\omega_{\text{b,0}}$. (i): The fitted value of $H_{0}$ is in agreement with local-cosmology observation [19, 27] but
discrepant at the 4.5 $\sigma$ level with the value $H_{0}\sim(67.4\pm 0.5)$ km Mpc${}^{-1}$ s${}^{-1}$ in $\Lambda$CDM of global-cosmology fits as extracted by the
Planck collaboration [8],
a galaxy-custering survey – the Baryonic Oscillation
Spectroscopy Survey (BOSS) [3] – , and a galaxy-custering-weak-lensing survey – the Dark Energy Survey (DES) [2] –
whose distances are derived from inverse distance ladders using the sound horizons at CMB decoupling or
baryon drag as references. Such anchors assume
the validy of $\Lambda$CDM (or a variant thereof with variable Dark Energy) at high $z$, and the data analysis employs
fiducial cosmologies that are close to those of CMB fits. As a rule of thumb, the inclusion of extra relativistic degrees
of freedom within reasonable bounds [3, 2] but also sample variance, local matter-density fluctuations, or a directional bias in SNa Ia observations
[12, 13, 14, 15] cannot explain
the above-quoted tension in $H_{0}$. Note that the DES Y1 data on three two-point functions (cosmic shear auto correlation,
galaxy angular auto correlation, and galaxy-shear cross correlation in up to five redshift bins with $z\leq 1.3$), which are most
sensitive to $\Omega_{m}$ and $\sigma_{8}$ (or $S_{8}=\sigma_{8}/\Omega_{m}$),
cannot usefully constrain $H_{0}$ by itself. However, the combination of DES 1Y data with those of Planck (no lensing)
yields an increases in $H_{0}$ on the 1-2 $\sigma$ level compared
to the Planck (no lensing) data alone, and a similar tendency is seen when BAO data [3] are enriched with those of
DES Y1, see Table II in [2]. Loosely speaking, one thus may state a mild increase of $H_{0}$ with an increasing portion of late-time (local-cosmology) information in the data.
(ii): The index of the initial spectrum of adiabatic, scalar curvature perturbations $n_{s}$ is unusually low, expressing an enhancement
of infrared modes as compared to the ultraviolet ones (violation of scale invariance). As discussed in [34],
such a tilted spectrum is not consistent with single-field
slow-roll inflation and implies that the Hubble parameter during the inflationary
epoch has changed appreciably. (iii): There is a low value of $\sigma_{8}$ (initial amplitude of matter-density power spectrum at
comoving wavelength 8 h${}^{-1}$ Mpc with
$H_{0}\equiv h\,100$ km s${}^{-1}$Mpc${}^{-1}$) compared to CMB fits [8] and BAO [3] although the DES Y1 fit to cosmic shear data alone
would allow for a value $\sigma_{8}\sim 0.71$ within the 1-$\sigma$ margin [2]. This goes also for our
high value of today’s matter density parameter $\Omega_{\text{m,0}}$ (the ratio of physical matter density to the critical density).
(iv): The low value of the redshift to (instantaneous) re-ionisation $z_{\rm re}$ (and the correspondingly low optical depth $\tau_{\text{re}}$)
compared to values obtained in CMB fits [8] is consistent with the one extracted from the
observation of the Gunn-Peterson trough in the spectra of high-redshift quasars [36]. (v): The low value of
$\omega_{\text{b,0}}$ – today’s physical baryon density – could provide a theoretical solution of the missing baryon problem [34] (missing compared
to CMB fits to $\Lambda$CDM (CMB fits) and the primordial D/H (Deuterium-to-Hydrogen) ratio in Big-Bang Nucleosynthesis (BBN), the latter yielding
$\omega_{\text{b,0}}=0.02156\pm 0.00020$ [56] which is by 2.3 $\sigma$ lower than the Planck value
of $0.0225\pm 0.00016$ in Tab. 1). If this low value of $\omega_{\text{b,0}}$ could be
consolidated observationally than either the idea that BBN is determined by isolated nuclear reaction cross sections only or the
observation of a truly primordial D/H in metal-poor environments or both will have to be questioned in future ivestigations.
There is a recent claim, however, that the missing 30-40 % of baryons were
observed in highly ionized oxygen absorbers representing the warm-hot intergalactic medium when illuminated by the X-ray spectrum of a quasar with $z>0.4$
[57]. This results is disputed in [58]. In [59] the missing baryon problem is addressed by the
measurement of electron column density within the intergalactic medium using the dispersion of localised radio bursts with $z\leq 0.522$.
In projecting electron density
the measurement appeals to a flat $\Lambda$CDM CMB-fitted model (Planck collaboration [8]).
Presently, $\omega_{\text{b,0}}$ is extracted from the ASKAP data in [59]
to be consistent with CMB fits and BBN albeit subject to a 50 % error which is expected to decrease with the advent
of more powerful radio observatories such as SKA.
4 Axionic Dark Sector and galactic Dark-Matter densities
The model for the Dark Sector in Eq. (9) is motivated by the possibility that a
coherent condensate of ultra-light particles – an axion field [53] – forms selfgravitating
lumps or vortices in the course of nonthermal (Hagedorn) phase transitions due to SU(2)/SU(3)
Yang-Mills factors, governing the radiation and matter content of the very early Universe,
going confining. A portion of the thus created, large abundance of such
solitons percolates into a dark-energy like state: the contributions $\Omega_{\Lambda}$ and
$\Omega_{\rm edm,0}(z_{p}+1)^{3}$ in Eq. (9) of which the latter may de-percolate at $z=z_{p}$ into a dark-matter like
component $\Omega_{\rm edm,0}(z+1)^{3}$ for $z<z_{p}$, the expansion of the Universe increasing the average distance between the centers of
neighbouring solitons. By virtue of the U(1)${}_{A}$ anomaly, mediated by topological charge
density [49, 50, 51, 52], residing in turn within the ground state of an SU(2)/SU(3) Yang-Mills theory [42],
the mass $m_{a}$ of an axion particle due to a single such theory of Yang-Mills scale $\Lambda$ is given as [53]
$$m_{a}=\frac{\Lambda^{2}}{M_{P}}\,,$$
(11)
where we have assumed that the so-called Peccei-Quinn scale, which associates with a dynamical
chiral-symmetry breaking, was set equal to the Planck mass $M_{P}=1.221\times 10^{28}\,$eV, see [60, 61] for motivations.
Let (natural units: $c=\hbar=1$)
$$r_{c}\equiv 1/m_{a}$$
(12)
denote the Compton wavelength,
$$r_{B}\equiv\frac{M_{P}^{2}}{Mm_{a}^{2}}$$
(13)
the gravitational Bohr radius, where $M\sim 10^{12}\,M_{\odot}$ is the total dark mass of a typical (spiral) galaxy
like the Milky Way, and
$$d_{a}\equiv\left(\frac{m_{a}}{\rho_{\rm dm}}\right)^{1/3}$$
(14)
the interparticle distance where $\rho_{\rm dm}$ indicates the typical mean energy density in Dark Matter of a spiral. Following [63], we assume
$$\rho_{\rm dm}=(0.2\cdots 0.4)\,\mbox{GeV\,cm}^{-3}\,.$$
(15)
For the concept of a gravitational Bohr radius in a selfconsistent, non-relativistic potential model
to apply, the axion-particle velocity
$v_{a}$ in the condensate should be much small than unity. Appealing to the virial theorem at a distance to the gravitational center
of $r_{B}$, one has $v_{a}\sim\left(\frac{Mm_{a}}{M_{P}^{2}}\right)^{2}$ [62]. In [62], where a selfgravitating axion condensate
was treated non-relativistically by means of a non-linear and non-local Schrödinger equation to represent a typical
galactic dark-matter halo, the following citeria on the validity of such an approach were put forward (natural units: $c=\hbar=1$):
(i) $v_{a}\ll 1$. ii) $d_{a}\ll r_{c}$ is required
for the description of axion particles in terms of a coherent Bose
condensate to be realistic. (iii) $r_{B}$ should be the typical extent of a galactic Dark-Matter halo:
$r_{B}\sim 100-300\,$kpc. With $\Lambda_{\textnormal{\tiny{cmb}}}\sim 10^{-4}\,$eV one obtains $m_{a}=8.2\times 10^{-37}\,$eV, $d_{a}=(4.1\cdots 5.2)\times 10^{-34}\,$pc,
$r_{c}=3.2\times 10^{25}\,$pc, and
$r_{B}\sim 8\times 10^{24}\,$kpc. While (i) and (ii) are extremely well satified with $v_{a}\sim 10^{-28}$ and $\frac{d_{a}}{r_{c}}\sim 10^{-59}$
point (iii) is badly violated. ($r_{B}$ is about $2\times 10^{18}$
the size of the visible Universe).
Therefore, the Yang-Mills scale responsible for the axion mass that associates
with dark-matter halos of galaxies must be dramatically larger. Indeed, setting $\Lambda=10^{-2}\Lambda_{e}$ where $\Lambda_{e}=\frac{m_{e}}{118.6}$
is the Yang-Mills scale of an SU(2) theory
that could associate with the emergence of the electron of mass $m_{e}=511\,$keV [64, 65], one obtains
$m_{a}=1.5\times 10^{-25}\,$eV, $d_{a}=(2.3\cdots 2.9)\times 10^{-30}\,$pc, $r_{c}=1.7\times 10^{14}\,$pc, and
$r_{B}\sim 232\,$kpc. In addition to (i) and (ii) with $v_{a}\sim 10^{-4}$ and $\frac{d_{a}}{r_{c}}\sim 10^{-44}$ also
point (iii) is now well satisfied. If the explicit Yang-Mills scale of an SU(2) theory, which
is directly imprinted in the spectra of the excitations in the pre - and deconfining phases,
acts only in a screened way in the confining phase as far as the axial anomaly
is concerned – reducing its value by a factor of one hundred or so –, then the above axionic Dark-Sector
scenario would link the theory responsible for the emergence of the electron with galactic dark-matter halos! In addition, the axions of
$\text{SU(2)}_{\textnormal{\tiny{cmb}}}$ would provide the Dark-Energy density $\Omega_{\Lambda}$ of such a scenario.
Finally, we wish to point out that the de-percolation mechanism of axionic solitons (lumps forming out of former Dark-Energy density) in the cosmological model based on
SU(2)${}_{\textnormal{\tiny{cmb}}}$, which may be considered to underly the transition
in the Dark Sector at $z_{p}=53$ described by Eq. (9), is consistent with the Dark-Matter density in the Milky Way.
Namely, working with $H_{0}=74$ km s${}^{-1}$Mpc${}^{-1}$, see Table 1 and [19],
the total (critical) energy density $\rho_{c,0}$ of our spatially flat Universe
is at present
$$\rho_{c,0}=\frac{3}{8\pi}\,M_{P}^{2}H_{0}^{2}=1.75\times 10^{-9}\,{\rm eV}^{4}\,.$$
(16)
The portion of cosmological Dark Matter $\rho_{{\rm cdm},0}$ then is, see Table 1,
$$\rho_{{\rm cdm},0}\sim 0.35\,\rho_{c,0}$$
(17)
which yields a cosmological energy scale $E_{{\rm cdm},0}$ in association with Dark Matter of
$$E_{{\rm cdm},0}\equiv\rho^{1/4}_{{\rm DM},0}\sim 0.00497\,{\rm eV}\,.$$
(18)
On the other hand, we may imagine the percolate of axionic field profiles, which dissolves at $z_{p}$, to be associated
with densely packed dark-matter halos typical of today’s galaxies. Namely, scaling the typical Dark-Matter
energy density of the Milky Way $\rho_{\rm dm}$ of Eq. (15) from $z_{p}$ (at $z_{p}+0$
$\rho_{\rm dm}$ yet behaves like a cosmological constant) down to $z=0$ and allowing for a factor of
$(\omega_{\text{pdm,0}}+\omega_{\text{edm,0}})/\omega_{\text{edm,0}}=2.47$, see Table 1, one extracts the energy
scale for cosmological Dark Matter $E_{{\rm G},0}$ in association with galactic Dark-Matter halos and primordial dark matter as
$$E_{{\rm G},0}\equiv\left(\frac{2.47\rho_{\rm dm}}{(z_{p}+1)^{3}}\right)^{1/4}=%
(0.00879\cdots 0.0105)\,{\rm eV}\,.$$
(19)
A comparison of Eqs. (18) and (19) reveals that $E_{{\rm cdm},0}$ is smaller but comparable to
$E_{{\rm G},0}$. This could be due to the neglect of galactic-halo compactification through the missing pull
by neighbouring profiles in the axionic percolate and because of the omission of selfgravitation and
baryonic-matter accretion/gravitation during the evolution from $z_{p}=53$ to present. (That is, the use of the values
of Eq. (15) in Eq. (19) overestimates the homogeneous energy density in the percolate.) The de-percolation of axionic solitons at $z_{p}$, whose mean, selfgravitating
energy density $\rho_{\rm dm}$ in Dark Matter is nearly independent of cosmological
expansion but subject to local gravitation, could therefore be linked to cosmological Dark Matter
today within the Dark-Sector model of Eq. (9).
5 Conclusions
The present article’s goal was to address some tensions between local and global cosmology on the basis
of the $\Lambda$CDM standard model. Cracks in this model
could be identified during the last few years thanks to independent tests resting on precise
observational data and their sophisticated analysis. To reconcile these results, a change of
$\Lambda$CDM likely is required before the onset of the formation of non-linear, large-scale structure. Here, we have reviewed a proposal
made in [34], which assumes thermal photons to be governed by
an SU(2) rather than a U(1) gauge principle, and we have discussed the SU(2)${}_{\textnormal{\tiny{cmb}}}$-implied changes in cosmological
parameters and the structure of the Dark Sector. Noticeably, the tensions in $H_{0}$, the baryonic density, and the
redshift for re-ionisation are addressed in favour of local measurements. High-$z$ inputs to CMB and BAO simulations,
such as $n_{s}$ and $\sigma_{8}$, are sizeably reduced as compared to their fitted values in $\Lambda$CDM.
The Dark Sector now invokes a de-percolation of axionic
field profiles at a redshift of $z_{p}\sim 53$. This idea is roughly consistent with
typical galactic Dark-Matter halos today, such as the one of the Milky Way, being released from the percolate. Axionic
field profiles, in turn, appear to be compatible with Dark-Matter halos
in typical galaxies if (confining) Yang-Mills dynamics subject to much higher mass scales than that of SU(2)${}_{\textnormal{\tiny{cmb}}}$
is considered to produce the axion mass.
To consolidate such a scenario
two immediate fields of investigation suggest themselves: (i) A deep understanding of possible
selfgravitating profiles needs to be gained towards their role in actively assisted large-scale structure formation as well as in
quasar emergence, strong lensing, cosmic shear, galaxy clustering, and galaxy
phenomenology (Tully-Fisher, rotation curves, etc.), distinguishing spirals from ellipticals and satellites from hosts. (ii)
More directly, the CMB large-angle anomalies require an addressation in terms of radiative effects in SU(2)${}_{\textnormal{\tiny{cmb}}}$, playing out
at low redshifts, which includes a re-investigation of the CMB dipole.
Funding: None.
Conflicts of Interest: There is no conflict of interest.
References
[1]
Eisenstein, D. J. et al. SDSS-III Massive Spectroscopic Surveys of the Distant Universe, the Milky Way, and Extra-Solar Planetary Systems.
Astron. J. 2011, 142,3, 24.
[2]
Abbott, T. M. C. et al. Dark Energy Survey year 1 results: Cosmological constraints from galaxy clustering and weak lensing. Phys. Rev. D
2018, 98, 043526.
[3]
Alam, S. et al. The clustering of galaxies in the completed SDSS-III Baryon Oscillation Spectroscopic Survey:
cosmological analysis of the DR12 galaxy sample. Month. Not. Roy. Astron. Soc. 2017, 470, 2617.
[4]
Ivezić, Ž. et al. LSST: From Science Drivers to Reference Design and Anticipated Data Products. Astrophys. J. 2019, 873, 111.
[5]
Vargas-Magaña, M. et al. Unravelling the Universe with DESI. 2019, arXiv:1901.01581.
[6]
Mather, J. C.; Cheng, E. S.; Cottingham, D. A.; Eplee, R. E.,Jr.; Fixsen, D. J.; Hewagama, T.; Isaacman, R. B.; Jensen, K. A.; Meyer, S. S.;
Noerdlinger, P. D.; et al. Measurement of the cosmic microwave background spectrum by the COBE FIRAS instrument. Astrophys. J. 1994, 420, 439.
[7]
Hinshaw, G. et al. Nine-Year Wilkinson Microwave Anisotropy Probe (WMAP) Observations: Cosmological Parameter Results. Astrophys. J. Suppl. Ser. 2013,
208, 19.
[8]
Aghanim, N. et al. Planck 2018 results. VI Cosmological parameters. 2018, arXiv:1807.06209.
[9]
Riess, A. et al. Observational Evidence from Supernovae for an Accelerating Universe and a Cosmological Constant. Astron. J. 1998, 116,3, 1009.
[10]
Perlmutter, S. et al. Measurement of $\Omega$ and $\Lambda$ from 42 high-redshift supernovae. Astrophys. J. 1998, 517, 565.
[11]
Dhawan, S.; Brout, D.; Scolnik, D.; Goobar, A.; Riess, A. G.; Miranda, V. Cosmological model insensitivity of local $H_{0}$ from the Cepheid distance ladder.
2020, arXiv:2001.09260.
[12]
Marra, V. et al. Cosmic Variance and the Measurement of the Local Hubble Parameter. Phys. Rev. Lett. 2013, 110, 241305.
[13]
Oddersov, I.; Hannestad, S.; Haugbølle, T. On the local variation of the Hubble constant. JCAP 2014, 10, 028.
[14]
Oddersov, I.; Hannestad, S.; Brandbyge, J. The variance of the locally measured Hubble parameter explained with different estimators.
JCAP 2017, 03, 022.
[15]
Wu, H.-Y.; Huterer, D. Sample variance in the local measurement of the Hubble constant. Month. Not. Roy. Astron. Soc. 2017, 471,4, 4946.
[16]
Riess, A. G. et al. A 2.4% Determination of the Local Value of the Hubble Constant. Astrophys. J 2016, 826, 56.
[17]
de Jaeger, T.; Stahl, B. E.; Zheng, W.; Filippenko, A. V.; Riess, A. G.; Galbany, L. A measurment of the Hubble constant from Type II supernovae.
Month. Not. Roy. Astron. Soc. 2020, 496,3, 3402.
[18]
Riess, A. G. et al. New Parallaxes of Galacic Cepheids from Spatially Scanning the Hubble Space Telescope: Implications for the Hubble Constant.
Astrophys. J. 2018, 855, 136.
[19]
Riess, A. G. et al. Large Magellanic Cloud Cepheid Standards Provide a 1% Foundation for the Determination of the Hubble Constant
and Stronger Evidence for Physics beyond $\Lambda$CDM. Astrophys. J. 2019, 876, 85.
[20]
Jang, I. S.; Lee, M. G. The Tip of the Red Giant Branch Distances to Type Ia Supernova Host Galaxies. IV. Color Dependence and Zero-Point Calibration.
Astrophys. J. 2017, 835, 28.
[21]
Freedman, W. L. et al. The Carnegie-Chicago Hubble Program. VIII. An Independent Determination of the Hubble Constant Based on the
Tip of the Red Giant Branch. Astrophys. J. 2019, 882, 34.
[22]
Huang, C. D. et al. A Near-infrared Period-Luminosity Relation for Miras in NGC 4258, an Anchor for a New Distance Ladder.
Astrophys. J. 2018, 857, 67.
[23]
Ma, C.-P.; Bertschinger, E. Cosmological Perturbation Theory in the Synchronous and Conformal Newtonian Gauges. Astrophys. J. 1995, 455, 7.
[24]
Raghunathan, S. et al. Detection of CMB-Cluster Lensing using Polarization Data from SPTpol. Phys. Rev. Lett. 2019, 123, 181301.
[25]
Birrer, S.; Amara, A.; Refregier, A. The mass-sheet degeneracy and time-delay cosmography analysis of the strong lens RXJ1131-1231. JCAP 2016,
08, 020.
[26]
Bonvin, V. et al. HoLiCOW - V. New COSOMGRAIL time delays of HE 0435-1223: $H_{0}$ to 3.8 per cent precision from
strong lensing in a flat $\Lambda$CDM model. Month. Not. Roy. Astron. Soc. 2017, 465, 4914.
[27]
Wong, K. C. et al. HoLiCOW - XIII. A 2.4% measurement of $H_{0}$ from lensed quasars: 5.3 $\sigma$ tension between early and late-Universe probes.
Month. Not. Roy. Astron. Soc. 2020, https://doi.org/10.1093/mnras/stz3094.
[28]
Schutz, B. F. Determining he Hubble constant from gravitational wave observations. Nature 1986, 323, 310.
[29]
Chen, H.-Y.; Fishbach, M.; Holz, D. E. A two percent Hubble constant measurement from standard sirens within five years. Nature 2018, 562, 545.
[30]
Handley, W.; Lemos, P. Quantifying tensions in cosmological parameters: Interpreting the DES evidence ratio. Phys. Rev. D 2019, 100, 043504-1.
[31]
Schwarz, D. J.; Copi, C. J.; Huterer. D.; Starkman, G. D. CMB anomalies after Planck. Class. Quantum. Grav. 2016, 33, 184001.
[32]
Hofmann, R. Lowâfrequency line temperatures of the CMB (Cosmic Microwave Background). Annalen Phys.
2009, 18, 634.
[33]
Fixsen, D. J.; Kogut, A.; Levin, S.; Limon, M.; Lubin, P.; Mirel, P.; Seiffert, M.; Singal, J.; Wollack, E.; Villela, T. et al. ARCADE 2 measurement of the absolute sky brightness at 3-90 GHz. Astrophys. J. 2011, 734, 5.
[34]
Hahn, S.; Hofmann, R.; Kramer, D. SU(2)${}_{\rm CMB}$ and the cosmological model: Angular power spectra.
Mon. Not. Roy. Astron. Soc. 2019, 482, 4290.
[35]
Aghanim, N. et al. Planck 2015 results. XI. CMB power spectra, likelyhoods, and robustness of parameters. Astron. Astrophys. 2016, 594, A11.
[36]
Becker, R. H.; Fan, X.; White, R. L.; Strauss, M. A.; Narayanan, V. K.; Lupton, R. H.; Gunn, J. E.; Annis, J.; Bahcall, N. A.; Brinkmann, J.
Evidence for Reionization at $z\sim 6$:
Detection of a Gunn-Peterson Trough in a $z=6.28$ Quasar. Astrophys. J. 2001, 122, 2850.
[37]
Ludescher, J.; Hofmann, R. Thermal photon dispersion law and modified black-body spectra. Annalen Phys. 2009, 18,4, 271.
[38]
Hofmann, R. The fate of statistical isotropy. Nature Phys. 2013, 9, 686.
[39]
Harrington, B. J.; Shepard, H. K. Periodic Euclidean solutions and the finite-temperature Yang-Mills gas. Phys. Rev. D 1978, 17, 2122.
[40]
Herbst, U. and Hofmann, R. Emergent Inert Adjoint Scalar Field in SU(2) Yang-Mills Thermodynamics due to Coarse-Grained Topological Fluctuations.
ISRN High Energy Physics 2012, Article ID 373121.
[41]
Grandou, T.; Hofmann, R. Thermal ground state and nonthermal probes, Adv. Math. Phys. 2015, 197197.
[42]
Hofmann, R. The thermodynamics of quantum Yang–Mills theory: theory and application. World Scientific,
Singapore, 2012 (1st ed.) and 2016 (2nd ed.).
[43]
Schwarz, M.; Hofmann, R.; Giacosa, F. Radiative corrections to the pressure and the one-loop polarization tensor of
massless modes in SU(2) YangâMills thermodynamics. Int. J. Mod. Phys. A 2007, 22, 1213.
[44]
Falquez, C.; Hofmann, R.; Baumbach, T. Modification of blackâbody radiance at low temperatures and
frequencies. Ann. Phys. 2010, 522, 904.
[45]
Diakonov, D.; Gromov, N.; Petrov, V.; Slizovskiy, S. Quantum weights of dyons and of instantons with nontrivial holonomy.
Phys. Rev. D 2004, 70, 036003.
[46]
Szopa, M.; Hofmann, R. A Model for CMB anisotropies on large angular scales. JCAP 2008, 03, 001.
[47]
Ludescher, J; Hofmann, R. CMB dipole revisited. arXiv:0902.3898.
[48]
Hahn, S.; Hofmann, R. Exact determination of asymptotic CMB temperature-redshift relation. Mod. Phys. Lett. A 2018, 33, 1850029.
[49]
Adler, S. L. Axial-Vector Vertex in Spinor Electrodynamics. Phys. Rev. 1969, 177, 2426.
[50]
Adler, S. L.; Bardeen, W. A. Absence of Higer-Order Corrections in the Anomalous Axial-Vector Divergence Equation.
Phys. Rev. 1969, 182, 1517.
[51]
Bell, J. S.; Jackiw, R. A PCAC Puzzle: $\pi^{0}\to\gamma\gamma$ in the $\sigma$ model. Nuovo Cim. A 1969, 60, 47.
[52]
Fujikawa, K. Path-Integral Measure for Gauge-Invariant Fermion Theories. Phys. Rev. Lett. 1979, 42, 1195.
[53]
Peccei, R. D.; Quinn, H. R. Constraints imposed by CP conservation in the presence of pseudoparticles. Phys. Rev. D 1977, 16, 1791.
[54]
Blas, D.; Lesgourges, J.; Tram, T. The Cosmic Linear Anisotropy Solving System (CLASS). Part II: Approximation schemes. JCAP 2011, 07, 034.
[55]
Hofmann, R. Relic photon temperature versus redshift and the cosmic neutrino background. Annalen Phys. 2015, 527, 254.
[56]
Cooke, R. J.; Pettini, M.; Nollett, K. M.; Jorgenson, R. The primordial deuterium abundance of the most metal-poor damped Ly$\alpha$ system. Astrophys. J. 2016, 830, 148.
[57]
Nicastro F. et al. Observations of the missing baryons in the warm hot intergalactic medium. Nature 2018, 558, 406.
[58]
Johnson, S. D. et al. The Physical Origins of the Identified and Still Missing Components of the WarmâHot Intergalactic Medium: Insights from Deep Surveys in the Field of Blazar 1ES1553+113.
Astrophys. J. Lett. 2019, 884:L31, 1.
[59]
Macquart, J.-P. et al. A census of baryons in the Universe from localized fast radio bursts. Nature 2020, 581, 391.
[60]
Frieman, J. A.; Hill, C. T.; Stebbins, A.; Waga, I. Cosmology with Ultralight Pseudo Nambu-Goldstone Bosons. Phys. Rev. Lett. 1995, 75, 2077.
[61]
Giacosa, F.; Hofmann, R. A Planck-scale axion and SU(2) Yang-Mills dynamics: Present acceleration and the fate of the photon. Eur. Phys. J. C
2007, 50, 635.
[62]
Sin, S.-J. Late-time phase transition and the galactic halo as a Bose liquid. Phys. Rev. D 1994, 50, 3650.
[63]
Weber, M.; de Boer, W. Determination of the local dark matter density in our Galaxy. Astron. & Astrophys. 2010, 509, A25.
[64]
Hofmann, R. The isolated electron: De Broglie’s ”hidden” thermodynamics, SU(2) Quantum Yang-Mills theory, and a strongly perturbed BPS monopole. Entropy 2017, 19,11, 575.
[65]
Grandou, T.; Hofmann, R. On emergent particles and stable neutral plasma balls in SU(2) Yang-Mills thermodynamics. 2020, arXiv: 2007.08460. |
Axion field induces exact symmetry
P. [email protected]
Saha Institute of Nuclear Physics,
1/AF Bidhannagar
Calcutta 700064
()
Abstract
While no regularization is consistent with the anomalous chiral symmetry which
occurs for massless fermions,
the artificial axion-induced symmetry for massive fermions
is shown here to be consistent with a standard regularization,
even in curved spacetime,
so that it can be said to have no anomaly in gauge or gravitational fields.
Implications for $\theta$ terms are pointed out.
1 Introduction
Chiral symmetry, which is exact for massless fermions at the classical level
and approximate for light quarks, has been very useful in particle physics.
In analogy with this chiral symmetry,
an artificial chiral-like symmetry [1] was introduced
some time back for the strong interactions with massive quarks.
It implies the occurrence of a light
pseudoscalar particle, the axion [2], which has however not been detected in
spite of elaborate searches [3].
This symmetry has been suspected to be fraught with an anomaly because of the
involvement of a chiral transformation. However, this is not obvious
and has to be examined by careful regularization of the theory.
This is important because of implications for the symmetry of the theory.
While chiral symmetry of the action, corresponding to the transformation
$$\displaystyle\psi\rightarrow e^{i\alpha\gamma_{5}}\psi,\quad\bar{\psi}%
\rightarrow\bar{\psi}e^{i\alpha\gamma_{5}},$$
(1)
is broken by a non-vanishing quark mass term $m\bar{\psi}\psi$, with mass $m$,
the artificial chiral symmetry for massive fermions works by letting a new field
${\varphi}$ absorb the chiral transformation. The mass term is replaced
by
$$\displaystyle\bar{\psi}me^{i{\varphi}\gamma_{5}}\psi,$$
(2)
which is classically invariant under the transformation
$$\displaystyle\psi\rightarrow e^{i\alpha\gamma_{5}}\psi,\quad\bar{\psi}%
\rightarrow\bar{\psi}e^{i\alpha\gamma_{5}},\quad\varphi\rightarrow{\varphi}-2\alpha,$$
(3)
which is also a symmetry of the other terms of the action.
The original interaction introduced by Peccei and Quinn [1] was of the form
$$\displaystyle\bar{\psi}[\Phi{1+\gamma^{5}\over 2}+\Phi^{\dagger}{1-\gamma^{5}%
\over 2}]\psi,$$
(4)
where $\Phi$ is a complex scalar field with a symmetry breaking potential.
The artificial chiral symmetry transformation here is
$$\displaystyle\Phi\to e^{-2i\alpha}\Phi,\quad\psi\to e^{i\alpha\gamma_{5}}\psi,%
\quad\bar{\psi}\to\bar{\psi}e^{i\alpha\gamma_{5}}.$$
(5)
$\Phi$ may be taken to be of the form $\rho e^{i{\varphi}}$.
The amplitude $\rho$ of the scalar field acquires a vacuum expectation value
because of symmetry breaking, which provides a massive boson.
The phase ${\varphi}$ is the zero mode of the potential
and provides a Goldstone boson.
This is the axion, which should acquire
a mass because of the quark masses, but does not appear to exist.
The kinetic term of $\Phi$ yields
$\frac{1}{2}\rho_{0}^{2}\partial_{\mu}\varphi\partial^{\mu}\varphi$
as the kinetic term for $\varphi$, with $\rho_{0}$ the vacuum expectation
value of $\rho$.
The mass $m$ comes from $\rho_{0}$ and the coupling constant of $\Phi$.
The anomaly, which occurs for the usual chiral symmetry
and makes the divergence of the axial current $\bar{\psi}\gamma_{\mu}\gamma_{5}\psi$
nonvanishing,
appears when the singularities in the theory are handled by regularization.
Even the measure picture
of anomalies requires a regularization for actual calculations.
We shall therefore use an explicit
regularization for studying chiral transformations of fermions
in the presence of the axion coupling. This will tell us whether
the continuous symmetry (3)
introduced classically for massive quarks is anomalous
or not.
The new symmetry is expected to be broken
spontaneously by the vacuum because of the translation of the spinless field
by a c-number. The field must have a definite value in a fixed vacuum:
let it be denoted by $\varphi_{0}$. Then
$$\displaystyle\varphi^{\prime}=\varphi-\varphi_{0}$$
(6)
is the shifted field with vanishing vacuum expectation value. This is a
pseudoscalar field, so that $\bar{\psi}\varphi^{\prime}\gamma_{5}\psi$ and
$(\varphi^{\prime}{\rm~{}tr~{}}F^{\mu\nu}\tilde{F}_{\mu\nu})$ are scalars. On the other
hand, $(\varphi_{0}{\rm~{}tr~{}}F^{\mu\nu}\tilde{F}_{\mu\nu})$ is
a pseudoscalar,
like the $\theta$ term $(\theta{\rm~{}tr~{}}F^{\mu\nu}\tilde{F}_{\mu\nu})$.
If a $(\varphi_{0}{\rm~{}tr~{}}F^{\mu\nu}\tilde{F}_{\mu\nu})$ term is generated
when the fermion is integrated out, it
can modify the $\theta$ term. The possibility of this occurrence is related to
the question of an anomaly in the symmetry (3).
If the symmetry (3) is anomalous,
no regularization of the fermions will be consistent with
it. On the other hand, if any regularization is found to be consistent with
it, the symmetry has no anomaly in this quantization and
such symmetric regularizations are the ones to be preferred.
Pauli-Villars regularization will be studied in flat and in curved spacetime.
2 Pauli-Villars regularization
We recall first the simple form of the Pauli-Villars
regularization. For a fermion with mass $m$, the Lagrangian density
$$\displaystyle\bar{\psi}[i\not{D}-m]\psi+\bar{\chi}[i\not{D}-M]\chi$$
(7)
involves regulator spinor fields $\chi,\bar{\chi}$ which however are
assigned Bose statistics. The regulator mass $M$ is ultimately taken to
infinity when the regulator fields decouple. The regulator fields couple to the
gauge fields in the standard way. The regulator mass term breaks the
chiral symmetry and yields the chiral anomaly in the $M\to\infty$ limit.
This is the original form of the Pauli-Villars regularization. A more general
form with several species of regulator fields is also available [4]:
$$\displaystyle\bar{\psi}[i\not{D}-m]\psi+\sum_{j}\sum_{k=1}^{|c_{j}|}\bar{\chi}%
_{jk}[i\not{D}-M_{j}]\chi_{jk}.$$
(8)
Here $c_{j}$ are integers whose signs are related to the statistics assigned
to $\chi_{jk}$. They have to satisfy some conditions to ensure regularization
of the divergences [4]:
$$\displaystyle 1+\sum_{j}c_{j}=0,\quad m^{2}+\sum_{j}c_{j}M_{j}^{2}=0.$$
(9)
Chiral transformations work on $\chi$ as well as $\psi$. Hence
the axion coupling has to be introduced for both $\psi,\chi$,
like the gauge coupling. One has
$$\displaystyle\bar{\psi}[i\not{D}-me^{i\varphi\gamma_{5}}]\psi+\bar{\chi}[i\not%
{D}-Me^{i\varphi\gamma_{5}}]\chi.$$
(10)
This regularization (10) is invariant under the combined transformation
$$\displaystyle\psi\rightarrow e^{i\alpha\gamma_{5}}\psi,$$
$$\displaystyle\bar{\psi}\rightarrow\bar{\psi}e^{i\alpha\gamma_{5}},$$
$$\displaystyle\chi\rightarrow e^{i\alpha\gamma_{5}}\chi,$$
$$\displaystyle\bar{\chi}\rightarrow\bar{\chi}e^{i\alpha\gamma_{5}},$$
$$\displaystyle\varphi$$
$$\displaystyle\rightarrow$$
$$\displaystyle{\varphi}-2\alpha,$$
(11)
which is the Pauli-Villars extension of (3).
This means the symmetry survives when the regularization is
removed by taking the limit $M\to\infty$
and hence is not anomalous. Unlike the regularized axial current, which has
a pseudoscalar divergence arising from the masses,
the Noether current for the extended symmetry, namely
$$\displaystyle\bar{\psi}\gamma_{\mu}\gamma_{5}\psi+\bar{\chi}\gamma_{\mu}\gamma%
_{5}\chi+2\rho_{0}^{2}\partial\mu\varphi,$$
(12)
is conserved.
The non-vanishing divergence of the axial current is cancelled by that
of the $\varphi$ piece by virtue of the axion equation of motion.
It is interesting to note that
the $\varphi$ phases in (10) can be removed by a joint chiral
transformation of both the physical fermions and the regulator spinors:
$$\displaystyle\psi\to e^{-i\varphi\gamma_{5}/2}\psi,$$
$$\displaystyle\bar{\psi}\to\bar{\psi}e^{-i\varphi\gamma_{5}/2},$$
$$\displaystyle\chi\to e^{-i\varphi\gamma_{5}/2}\chi,$$
$$\displaystyle\bar{\chi}\to\bar{\chi}e^{-i\varphi\gamma_{5}/2}.$$
(13)
The derivative operators in the action produce derivatives of $\varphi$,
which appear after the transformation.
Now the transformation has a trivial Jacobian because the contribution of the
fermion field is cancelled by that of the bosonic
regulator field as in the latter case the determinant arising from
functional integration comes in the denominator. The argument for the
generalized Pauli-Villars regularization, with (10) replaced by
$$\displaystyle\bar{\psi}[i\not{D}-me^{i\varphi\gamma_{5}}]\psi+\sum_{j}\sum_{k=%
1}^{|c_{j}|}\bar{\chi}_{jk}[i\not{D}-M_{j}e^{i\varphi\gamma_{5}}]\chi_{jk},$$
(14)
involves a Jacobian with the factor $1+\sum_{j}c_{j}=0$
in the exponent [5]. Consequently the effective action does not contain
any $\varphi{\rm~{}tr~{}}F^{\mu\nu}\tilde{F}_{\mu\nu}$ in this regularization.
The conserved Noether current here is
$$\displaystyle\bar{\psi}\gamma_{\mu}\gamma_{5}\psi+\sum_{j}\sum_{k=1}^{|c_{j}|}%
\bar{\chi}_{jk}\gamma_{\mu}\gamma_{5}\chi_{jk}+2\rho_{0}^{2}\partial\mu\varphi,$$
(15)
Note that the argument may be easily extended to curved spacetime,
where the Dirac operator
$$\displaystyle i\not{D}=i\gamma^{l}e^{\mu}_{l}(\partial_{\mu}-iA_{\mu}-\frac{i}%
{2}A^{mn}_{\mu}\sigma_{mn})$$
(16)
comes with a tetrad $e^{\mu}_{l}$ and a spin connection $A^{mn}_{\mu}$ in
addition to the gauge field, but
it continues to anticommute with $\gamma^{5}$.
The regularized action is again invariant under the Pauli-Villars extension of
(3).
This means that there is not even an $R\tilde{R}$ anomaly [6] in this symmetry.
Similar results have also been found by others [7].
3 Conclusion
We have examined the chiral symmetry of the fermion action including
$\varphi$. This classical symmetry is preserved by the explicit regularization
(10).
When an acceptable regularization preserves a symmetry, one can
conclude that the quantum theory defined by the limit of that regularization
will also satisfy it. Hence, after the
fermion is integrated out, the effective action must be invariant under
translations of $\varphi_{0}$, i.e., independent of this variable.
This is indeed true of this regularization, as seen above, and there
is no $\varphi{\rm~{}tr~{}}F^{\mu\nu}\tilde{F}_{\mu\nu}$ term. This makes the
modification of a $\theta$ term by a contribution from $\varphi_{0}$ impossible.
The field $\varphi$ occurs only in the form of derivatives.
The extension of the argument to curved spacetime means that an existing
$R\tilde{R}$ term [8] too cannot be modified.
There is another way of seeing this.
The exact classical chiral symmetry could be expected to play a rôle like
the chiral symmetry that holds for massless fermions. With massless fermions,
the $\theta$ term can be modified by a chiral transformation which does not
alter the fermion action but produces an $F\tilde{F}$ term because of the chiral
anomaly. However,
the chiral symmetry which occurs in the presence of mass and $\varphi$ is not
anomalous, as we have seen.
So $\theta$ cannot be modified in that way.
It should be pointed out here that any
effect of $\theta$ terms can be removed by setting $\theta=0$ [5] or by making
it explicitly dynamical, which forces the topological charge to vanish.
A dynamical $\theta$-like term is essentially a Peccei-Quinn
$(\varphi_{0}{\rm~{}tr~{}}F^{\mu\nu}\tilde{F}_{\mu\nu})$ term without
an axion particle. However, as only fields are dynamical in field theory,
the field $\varphi$ was used in [1] and a direct
$(\varphi{\rm~{}tr~{}}F^{\mu\nu}\tilde{F}_{\mu\nu})$ term can remove $\theta$.
One may wonder how much freedom one has in choosing regularizations.
Different regularizations have been used in the past
and it is known that all do not lead to the same result. A key point is that
symmetries of the action are generally sought to be preserved by the
regularization. Thus one is always looking for Lorentz invariant and gauge
invariant regularizations for actions having such symmetries.
Regularizations which maintain all symmetries are technically natural.
We have seen that
the Pauli-Villars regularization
conveniently respects the symmetry (3) of the fermion action.
Non-symmetric regularizations are possible, but are not
preferable in any way.
Regularizations that break an
existing classical symmetry like (3) may to that extent be
termed technically unnatural.
The artificial chiral symmetry here is not anomalous
as a regularization consistent with it has been demonstrated to exist. In these
circumstances, working with regularizations inconsistent with it would be
a needless and avoidable violation.
We end with the hope that the exact chiral symmetry in the presence of axions
will be useful also in other calculations and even in curved spacetime.
REFERENCES
1.
R. Peccei, & H. Quinn,
Phys. Rev. Letters
38, 1440 (1977);
Phys. Rev. D16, 1791 (1977)
2.
S. Weinberg,
Phys. Rev. Letters 40, 223 (1978);
F. Wilczek, Phys. Rev. Letters 40, 279 (1978)
3.
C. Patrignani et al., Review of Particle Physics,
Chin. Phys. C40, 100001 (2016)
4.
L. D. Faddeev & A. A. Slavnov, Gauge Fields, (Benjamin-Cummings, Reading, 1982)
5.
H. Banerjee, D. Chatterjee & P. Mitra,
Phys. Letters B573, 109 (2003);
see also P. Mitra, Symmetries and symmetry breaking in field theory,
CRC Press, Florida (2014)
6.
T. Kimura, Prog. Theo. Phys. 42, 1191 (1969)
7.
G. Dvali and L. Funcke, arXiv:1608.08969 (2016)
8.
S. Deser, M. Duff and C. Isham, Phys. Lett. B93, 419 (1980) |
Implications of imaginary chemical potential for model building of QCD
Kouji Kashiwa
[email protected]
Yukawa Institute for Theoretical Physics,
Kyoto University, Kyoto 606-8502, Japan
Abstract
Properties of QCD at finite imaginary chemical potential are revisited to
utilize for the model building of QCD in low energy regimes.
For example, the electric holonomy which is closely related to the
Polyakov-loop drastically affects thermodynamic quantities beside
the Roberge-Weiss transition line.
To incorporate several properties at finite imaginary chemical
potential, it is important to introduce the
holonomy effects to the coupling constant of effective models.
This extension is possible by considering the entanglement vertex.
We show justifications of the entanglement vertex based on the
derivation of the effective four-fermi interaction in the
Nambu–Jona-Lasinio model and present its general form with the
local approximation.
To discuss how to remove model ambiguities in the entanglement
vertex, we calculate the chiral condensate with different
$\mathbb{Z}_{3}$ sectors and the dual quark condensate.
pacs: 11.30.Rd, 21.65.Qr, 25.75.Nq
††preprint: YITP-16-38
I Introduction
Understanding non-perturbative properties of Quantum Chromodynamics
(QCD) is one of the important and interesting subjects in nuclear and
elementary particle physics.
The lattice QCD simulation is a promising method to understand such
non-perturbative properties, but the sign problem obscures it at finite
real chemical potential ($\mu_{\mathrm{R}}$).
Therefore, low energy effective models of QCD are widely used to
investigate the QCD phase structure at finite $\mu_{\mathrm{R}}$.
In the modeling of QCD in low energy regimes, the chiral condensate
and the Polyakov-loop play an important role to describe the chiral
and confinement-deconfinement transitions.
The famous effective model is the Polyakov-loop extended
Nambu–Jona-Lasinio (PNJL) model Fukushima (2004).
In the PNJL model, the quark and gluon contributions are taken
into account through the one-loop level effective potential in usual.
There is the uncertainty how to introduce the gauge boson contribution.
Usually, the chiral part is described by using the Nambu–Jona-Lasinio
(NJL) model or similar quark models in effective models.
However, the gluon part is less clear than the quark part.
There are several effective models for the gluon contribution;
for example,
the logarithmic potential based on the strong coupling
expansion
Fukushima (2004); Hell et al. (2010),
the Ginzburg-Landau type potential
Ratti et al. (2006); Roessner et al. (2007),
the Meisinger-Miller-Ogilvie model
Meisinger et al. (2002),
the matrix model for the confinement-deconfinement transition
Dumitru et al. (2011); Kashiwa et al. (2012),
the effective potential from the Landau-gauge gluon and ghost propagators
Fukushima and Kashiwa (2013),
the holonomy potential Shuryak and Sulejmanpasic (2013) and so on.
In addition to individual parts of the quark and gluon contributions,
it is important how to control the strength of the correlation between
the quark and gluon contributions in effective models.
Importance of such correlation can be clearly seen in the
lattice QCD data at finite temperature ($T$) and the imaginary chemical
potential ($\mu_{\mathrm{I}}$) D’Elia and Sanfilippo (2009); Bonati et al. (2011).
At finite $\mu_{\mathrm{I}}$, there is the first-order transition line which is so
called the Roberg-Weiss (RW) transition and its endpoint which is so
called the RW endpoint Roberge and Weiss (1986).
In Ref. D’Elia and Sanfilippo (2009); Bonati et al. (2011), the pion mass ($m_{\pi}$)
dependence of the RW endpoint has been predicted by the lattice QCD
simulation.
At sufficiently large $m_{\pi}$, the order
of the RW endpoint is the first-order which is induced by the
first-order $\mathbb{Z}_{3}$ transition.
The first-order RW endpoint becomes the crossover by the
explicit breaking of the $\mathbb{Z}_{3}$ symmetry at moderate $m_{\pi}$.
However, there is the mysterious behavior of the RW endpoint at small
$m_{\pi}$; the order of the transition turns into the first-order!
This behavior cannot be reproduced by using simple effective model of
QCD, but it can be possible by considering the extension of the
coupling constant.
Recently, it has been proposed in
Ref. Kashiwa and Ohnishi (2015, 2016) that the
confinement-deconfinement transition can be described by using the
analogy of the topological order Wen (1990) motivated by the
recent progress done in Ref. Sato (2008) at $T=0$ QCD.
In the conjecture Kashiwa and Ohnishi (2015, 2016), the nontrivial
free-energy degeneracy at finite $\mu_{\mathrm{I}}$ plays a crucial role.
The non-trivial degeneracy is related with the drastic change of
the holonomy beside $\theta=(2k-1)k/N_{\mathrm{c}}$ in the deconfinement phase
where $\theta$ is the dimensionless chemical potential
$\theta\equiv\mu_{\mathrm{I}}/T$, $N_{\mathrm{c}}$ means the number of color and $k$ is any
integer.
Then, we can define the quantum order-parameter which is so called the
quark number holonomy on the manifold in all possible
boundary condition space which is equivalence to the $\theta$
space because the boundary condition and the
imaginary chemical potential have direct relation, see
Ref. Roberge and Weiss (1986); Kashiwa and Misumi (2013) as an example.
From these viewpoints, holonomy effects in the effective model is
crucial not only incorporating the confinement-deconfinement
transition into the model, but also understanding the
confinement-deconfinement transition itself.
In this paper, we revisit the imaginary chemical potential region
to obtain some constraints for the model building of QCD in the low
energy regime.
This paper is organized as follows.
In the next section, we summarize the RW transition and some related
topics on it at finite $\mu_{\mathrm{I}}$.
Section III explains the extension of the effective model based
on Ref. Sakai et al. (2010).
Then, we show justifications of the extension based on the derivation of
the effective four-fermi interaction.
In Sec. IV, we calculate the chiral condensate with different
$\mathbb{Z}_{3}$ sectors to discuss how to remove ambiguities of model
parameters.
Section V is devoted to summary.
II Imaginary chemical potential
It is well known that there is a special periodicity in the QCD
partition function ($Z$) as a function of $\theta$;
$$\displaystyle Z(\theta)$$
$$\displaystyle=Z\Bigl{(}\theta+\frac{2\pi k}{N_{\mathrm{c}}}\Bigr{)}.$$
(1)
This means that several quantities have the $2\pi/N_{\mathrm{c}}$
periodicity.
This periodicity is so called the Roberge-Weiss (RW) periodicity.
In the low and high $T$ region, origins of the RW periodicity is quite
different.
At low $T$, the RW periodicity is realized by only one global minimum on
the complex Polyakov-loop ($\Phi$) plane, but does not at high $T$.
The $\mathbb{Z}_{N_{\mathrm{c}}}$ images should be needed to realize the RW periodicity at high
$T$; for example, the $\mathbb{Z}_{3}$ images are $e^{i2\pi/3}$ and
$e^{i4\pi/3}$ for $\Phi=1$.
This difference of origins seems to be related with the confinement and
the deconfinement nature.
From the different realization of the RW periodicity at low and high
$T$, there should be the first-order transition and its endpoint at
$\theta=(2k-1)\pi/N_{\mathrm{c}}$.
Those are called the RW transition and the RW endpoint, respectively.
When we change $\theta$ with fixing $T$ in the confined phase, the
phase of the Polyakov-loop $\phi$ which is determined from
$$\displaystyle\Phi$$
$$\displaystyle=\frac{1}{N_{\mathrm{c}}}\mathrm{tr}{\cal P}\Bigl{[}\exp\Bigl{(}i\oint_{0}^{\beta}A_{4}(\tau,{\vec{x}})~{}d\tau\Bigr{)}\Bigr{]}=|\Phi|\hskip 1.42262pte^{i\phi},$$
(2)
is continuously rotated as the soliton solution which may be described
by the Jacobi elliptic function where ${\cal P}$ means the path ordering
operator, $g$ is the gauge coupling
constant and $\beta$ means the inverse temperature ($1/T$).
On the other hand, the Polyakov-loop phase is discontinuously
rotated in the deconfined phase.
The discontinuous point appears when we across the RW transition line.
The electric holonomy ($\nu$) is then discontinuously changed beside the RW
transition line and then we obtain
$$d\nu=\lim_{\epsilon\to 0}\Bigl{[}\nu(\theta_{\mathrm{RW}}-\epsilon)-\nu(\theta_{\mathrm{RW}}+\epsilon)\Bigr{]}\left\{\begin{array}[]{l}0~{}~{}~{}~{}~{}~{}~{}T<T_{\mathrm{RW}}\\[14.22636pt]
\dfrac{2\pi}{N_{\mathrm{c}}}~{}~{}~{}~{}T>T_{\mathrm{RW}},\\
\end{array}\right.$$
where we define $d\nu$ as $\{d\nu~{}|~{}0\leq d\nu\leq 2\pi\}$.
Just on the RW endpoint, $d\nu$ depends on its order:
If the RW endpoint is first-order, $d\nu$ should be $2\pi/N_{\mathrm{c}}$.
Because of this behavior, $\theta$-even quantities such as the entropy
density and pressure should have the cusp, but $\theta$-odd quantities
such as the quark number density have the gap at $\theta=(2k-1)\pi/N_{\mathrm{c}}$
above the RW endpoint.
The RW transition and endpoint have interesting phenomena as discussed above,
but it may have more mysterious behavior which is the triple-point
realization of the RW endpoint at small $m_{\pi}$ which is predicted by
the lattice QCD simulation D’Elia and Sanfilippo (2009); Bonati et al. (2011).
At sufficiently large $m_{\pi}$, the order
of the RW endpoint is the first-order.
This first-order RW endpoint becomes the crossover by the
explicit breaking of the $\mathbb{Z}_{3}$ symmetry when we set $m_{\pi}$ as
moderate values.
However, the order of the RW transition turns into the first-order again
at small $m_{\pi}$.
This behavior is considered as the consequence of the correlation between
the chiral and the confinement-deconfinement transition nature, but it is
still under debate.
By using simple effective models such as the standard PNJL model, this
behavior cannot be reproduced; for example, see
Ref. Sasaki et al. (2011).
To resolve this point, we need extension of the effective model.
Actually, by extending the coupling constant in the PNJL model, we can
reproduce the lattice QCD simulation Sakai et al. (2010), but the exact
form of the extended coupling constant is not discussed so far.
In the next section, we revisit this point.
III Entanglement vertex
In this study, we use the PNJL model.
The simplest two-flavor and three-color PNJL model Lagrangian density
becomes
$$\displaystyle{\cal L}$$
$$\displaystyle={\bar{q}}(i{\ooalign{\hfil/\hfil\crcr$D$}}-m_{0})q+G[({\bar{q}}q)^{2}+({\bar{q}}i\gamma_{5}{\vec{\tau}}q)^{2}]-\beta V{\cal U},$$
(5)
where $D^{\mu}=\partial^{\mu}+i\delta^{\nu}_{0}A^{\nu,a}\lambda_{a}/2$ with
Gell-Man matrices $\lambda_{a}$,
$m_{0}$ expresses the current quark mass, $G$ is the coupling
constant and $V$ does the
three-dimensional volume.
The term ${\cal U}$ expresses the gluon contribution.
Model parameters in the NJL part such as $m_{0}$, $G$ and the cutoff of
the three-dimensional momentum integration ($\Lambda$) is determined by
using the $m_{\pi}$ and pion decay constant.
The actual values used in this study are
$m_{0}=5.5$ MeV, $G=5.498$ GeV${}^{-2}$, $\Lambda=613.5$ MeV.
In this study, we use the mean-field approximation; see
Ref. Sakai et al. (2010) for the details of the calculation.
The actual thermodynamic potential can be obtained from the Lagrangian
density (5) as
$$\displaystyle\Omega$$
$$\displaystyle=-4\int\frac{d^{3}p}{(2\pi)^{3}}\Bigl{[}3E(p)+T(\ln f^{-}+\ln f^{+})\Bigr{]}$$
$$\displaystyle~{}~{}~{}+2G\sigma^{2}+{\cal U},$$
(6)
where
$$\displaystyle f^{-}$$
$$\displaystyle=1+3(\Phi+{\bar{\Phi}}e^{-\beta E^{-}(p)})e^{-\beta E^{-}(p)}+e^{-3\beta E^{-}(p)},$$
$$\displaystyle f^{+}$$
$$\displaystyle=1+3({\bar{\Phi}}+\Phi e^{-\beta E^{+}(p)})e^{-\beta E^{+}(p)}+e^{-3\beta E^{+}(p)},$$
(7)
here ${\bar{\Phi}}$ is the conjugate of $\Phi$ and
$E^{\mp}(p)=E(p)\mp\mu=\sqrt{p^{2}+M^{2}}\mp\mu$ with
$M=m_{0}-2G\sigma$ and the chemical potential $\mu=(\mu_{\mathrm{R}},\mu_{\mathrm{I}})$.
The quantity $\sigma$ becomes the chiral condensate in the present
model.
The gluon contribution ${\cal U}$ is taken into account via the
logarithmic Polyakov-loop effective potential Hell et al. (2010) as
$$\displaystyle\frac{{\cal U}}{T^{4}}$$
$$\displaystyle=-\frac{1}{2}b_{2}(T){\bar{\Phi}}\Phi$$
$$\displaystyle+b_{4}(T)\ln\Bigl{[}1-6{\bar{\Phi}}\Phi+4\Bigl{(}{\bar{\Phi}}^{3}+\Phi^{3}\Bigr{)}-3({\bar{\Phi}}\Phi)^{2}\Bigr{]},$$
(8)
where
$$\displaystyle b_{2}(T)$$
$$\displaystyle=a_{0}+a_{1}\Bigl{(}\frac{T_{0}}{T}\Bigr{)}+a_{2}\Bigl{(}\frac{T_{0}}{T}\Bigr{)}^{2}+a_{3}\Bigl{(}\frac{T_{0}}{T}\Bigr{)}^{3},$$
$$\displaystyle b_{4}(T)$$
$$\displaystyle=b_{4}\Bigl{(}\frac{T_{0}}{T}\Bigr{)}^{3},$$
(9)
with $a_{0}=3.51$, $a_{1}=-2.56$, $a_{2}=15.2$, $a_{3}=-0.62$ and
$b_{4}=-1.68$.
These parameters are determined to reproduce the lattice
QCD data such as the energy and the entropy density in the pure gauge
limit.
The parameter $T_{0}$ controls the confinement-deconfinement temperature
($T_{\mathrm{D}}$) in the pure gauge limit and thus we set $T_{0}=270$ MeV because we are interested in the quenched calculation.
In Ref. Sakai et al. (2010), the extension of the PNJL model coupling
constant has been proposed.
The extension has been done by the replacement
$$\displaystyle G\to G(\Phi)=G[1-\alpha_{1}{\bar{\Phi}}\Phi-\alpha_{2}({\bar{\Phi}}^{3}+\Phi^{3})],$$
(10)
where $\alpha_{1}$ and $\alpha_{2}$ are new parameters.
This extended coupling constant is so called the entanglement vertex.
In the extension, the $\mathbb{Z}_{3}$ symmetry of the vertex is assumed.
This interaction strongly correlates the chiral and the
confinement-deconfinement transition.
The entanglement nature is deeply related with the nonlocal properties
of the interaction which is usually hidden in the local PNJL model.
From the derivation of the NJL and also the PNJL model, the nonlocal
nature is a natural consequence.
The PNJL model can be obtained starting from QCD partition function by
using the color-current expansion and some ansatz for gluon
properties; details of the nonlocal PNJL model, for example, see
Ref. Van den Bossche (1997); Kondo (2010); Kashiwa et al. (2011) and
references therein.
The nonlocal interaction should feel effects coming from the
gluon two-point function since the interaction forms
$$\displaystyle{\cal L}_{\mathrm{int}}(x)$$
$$\displaystyle=g^{2}j_{a}^{\mu}(x)\int d^{4}y~{}W^{(2)ab}_{\mu\nu}j^{\nu}_{b}(y),$$
(11)
where
$j_{\mu}^{a}={\bar{q}}\frac{\lambda^{a}}{2}\gamma_{\mu}q$ is the
color-current, $W^{(2)}$ is the gluon two-point function without quark
loops because we use the color-current expansion around $j=0$.
Since the two-point function does not have quark loops, $W^{(2)}$ should
be $\mathbb{Z}_{N_{\mathrm{c}}}$ symmetric form.
To obtain the local version of the PNJL model, we should replace the
two-point function as the delta function since we assume $W^{(2)}$ is
very short range.
The gluon follows the adjoint representation of
$SU(N_{\mathrm{c}})$, and thus the entanglement vertex should obey the adjoint
representation nature.
To introduce the adjoint representation nature to the entanglement vertex,
we refer to the structure of the gluon one-loop effective potential and
the Bose distribution function with the electric holonomy.
Those manifest the $\mathbb{Z}_{3}$ symmetric form and other adjoint
representation natures.
So, more general form of the entanglement vertex for the color singlet
sector can be written as
$$\displaystyle G(\Phi)$$
$$\displaystyle=G_{\mathrm{N}}\Bigl{[}1+\Bigl{(}\sum_{n=1}^{8}\alpha^{\prime}_{n}C_{n}e^{-n\beta E}\Bigr{)}^{\gamma}\Bigr{]},$$
(12)
where $G_{\mathrm{N}}$, $\alpha^{\prime}_{n}$ and $\gamma$ are parameters,
$E=|{\vec{p}}|$ is the typical energy scale in the momentum space
and
$$\displaystyle C_{1}$$
$$\displaystyle=C_{7}=1-9{\bar{\Phi}}\Phi,$$
$$\displaystyle C_{2}$$
$$\displaystyle=C_{6}=1-27{\bar{\Phi}}\Phi+27({\bar{\Phi}}^{3}+\Phi^{3}),$$
$$\displaystyle C_{3}$$
$$\displaystyle=C_{5}=-2+27{\bar{\Phi}}\Phi-81({\bar{\Phi}}\Phi)^{2},$$
$$\displaystyle C_{4}$$
$$\displaystyle=2\Bigl{[}-1+9{\bar{\Phi}}\Phi-27({\bar{\Phi}^{3}+\Phi^{3}})+81({\bar{\Phi}}\Phi)^{2}\Bigr{]},$$
$$\displaystyle C_{8}$$
$$\displaystyle=1.$$
(13)
For details of functions $C_{n}$, see Ref. Sasaki and Redlich (2012).
Even if we assume the effective gluon mass such as the Gribov-Stingl
form at finite $T$ as
$$\displaystyle D^{(T,L)}_{T}$$
$$\displaystyle\propto\frac{d_{t,l}(p^{2}+d_{t,l}^{-1})}{(p^{2}+r^{2}_{t,l})^{2}},$$
(14)
we then obtain same expression of $C_{n}$ Fukushima and Kashiwa (2013),
where $D^{T}_{T}$ ($D^{L}_{T}$) are the tree-dimensional transverse
(longitudinal) gluon propagator.
Thus, we can expect that above expression is general form for the
entanglement vertex.
When we assume that the $W^{(2)}$ is short range, higher-order terms of
$e^{-\beta E}$ in Eq. (12) should be suppressed
and then only $C_{1}$ is relevant.
From here, we set $\gamma=1$ to make our discussion simple and it is not
an essential point in the level of present discussions.
Thus, we can effectively rewrite the coupling constant (12) as
the simple form
$$\displaystyle G(\Phi)$$
$$\displaystyle\sim G_{\mathrm{N}}\Bigl{(}1+\alpha{\bar{\Phi}}\Phi\Bigr{)},$$
(15)
where $G_{\mathrm{N}}$ is corresponding to $G$ in the original PNJL
model in this case because ${\bar{\Phi}}$ and $\Phi$ do not affect the
physics at $T=0$.
This form is similar to Eq. (10), but it is more general.
If we set $\alpha=-\alpha_{1}$ and $\alpha_{2}=0$ with $\gamma=1$, this
vertex is perfectly matched with the vertex (10).
From present representation, terms proportional to
$({\bar{\Phi}}^{3}+\Phi^{3})$
becomes the higher-order contribution and thus it can be safely
neglected in the level discussed here.
In the effective model construction, important operation is the
parameter fixing.
Details of parameter fixing are discussed in the next section.
Recently, it has been numerically shown that the $\mathbb{Z}_{3}$ factor
can affect the Landau-gauge gluon propagator in the deconfined phase by
using the pure gauge calculation in Ref. Silva and Oliveira (2016).
In the gluon propagator, the $\mathbb{Z}_{3}$ factors can appear via
${\cal A}_{4}^{ab}={\cal A}_{4}^{a}-{\cal A}_{4}^{b}$ when we consider the
background gauge field where ${\cal A}_{4}$ is the temporal component of the
field and $a,b$ run $1,\cdots,N_{\mathrm{c}}$ because the gluon obeys the adjoint
representation.
For $a=b$ diagonal sectors, the $\mathbb{Z}_{N_{\mathrm{c}}}$ factor cannot appear
because of the exact cancellation of the $\mathbb{Z}_{3}$ factors.
On the other hand, the non-diagonal $a\neq b$ sectors may feel the
effect via the effective gluon mass.
Also, in the confined and deconfined phases, the elementary degree of
freedom is drastically different; it is glueball in the confined phase,
but it is gluon in the deconfined phase.
Both reasons may lead to the lattice QCD predicted phenomena, but we
need more careful investigation for the lattice QCD data to conclude it.
In the short range limit of $W^{(2)}$, the color structure is extremely
simplified and thus above effect cannot appears in the local PNJL model.
To incorporate the effects, we should consider the nonlocal PNJL model,
but it seems a difficult task.
Even in the modern nonlocal PNJL
model Hell et al. (2009, 2010); Radzhabov et al. (2011); Pagura et al. (2012); Kashiwa et al. (2011); Carlomagno et al. (2013),
we implicitly use
the simplification of the color structure of the four-fermi interaction.
At present, we do not know how important this effects for the gauge
invariant observable, and thus
we assume that $W^{(2)}$ is short range to neglect the effect as a first
step.
This problem will be discussed in the nonlocal PNJL model in elsewhere.
IV Numerical results
In this section, we discuss how to remove ambiguities of the
parameters in the entanglement vertex by
using lattice QCD simulations with the quenched approximation.
The entanglement vertex correlates the chiral and
the confinement-deconfinement transition and thus the chiral condensate
with different $\mathbb{Z}_{3}$ sectors seems to be a promising quantity
to remove ambiguities of model parameters.
The existence of non-trivial $\mathbb{Z}_{3}$ sectors which is
characterized by $\phi=\frac{2\pi}{3}$ and $\frac{4\pi}{3}$ is not
trivial in the system with dynamical quarks.
Therefore, we here consider the quenched approximation since all
possible $\mathbb{Z}_{3}$ sectors are well defined.
The calculation of the chiral condensate with the different
$\mathbb{Z}_{3}$ sector was done in
Ref. Chandrasekharan and Christ (1996); Chandrasekharan and Huang (1996) and thus we
follow their calculation by using present extended model.
In the calculation Chandrasekharan and Huang (1996) , they introduce the
$T$-dependent coupling constant by hand, but this effect is automatically
introduced in the present entanglement PNJL model.
In this study, we first calculate $\Phi$ in the pure
gauge limit to consider the quenched calculation.
The Polyakov-loop with non-trivial $\mathbb{Z}_{3}$ sectors,
$\Phi_{\frac{2\pi}{3}}$ and $\Phi_{\frac{4\pi}{3}}$,
are obtained by the $\mathbb{Z}_{3}$ transformation from
the Polyakov-loop with the trivial $\mathbb{Z}_{3}$ sector, $\Phi_{0}$.
Figure 1 shows $\Phi$ with $\phi=0$ and
$\phi=\frac{2\pi}{3}$ in the pure gauge limit as a function of $T$.
After this calculation, we substitute those $\Phi$
to the PNJL model and then we minimize the thermodynamic
potential as a function of $\sigma$.
This procedure is corresponding to the quenched calculation of the PNJL
model.
Figure. 2 shows the chiral condensate with different
$\mathbb{Z}_{3}$ sectors at $\mu=0$ as a function of $T$.
The top and bottom panels of Fig. 2 show $\sigma/\sigma_{0}$
with the trivial $\mathbb{Z}_{3}$ sector $\phi=0$ and that with
$\phi=2\pi/3$ where $\sigma_{0}$ is $\sigma$ at $T=\mu=0$,
respectively.
It is natural to assume that the entanglement vertex is decreased with
increasing $T$ and thus the sign of $\alpha$ is negative
because of properties of the running coupling in QCD.
When $\alpha$ is small in the bottom panel, the chiral condensate
increases above $T_{\mathrm{D}}$, but it decreases in the large-$\alpha$
case.
There is the drastic change in the $T$-dependence of $\sigma$ and thus
we can determine the parameters in the entanglement
vertex by using precise quenched lattice QCD data.
At least in Ref. Chandrasekharan and Christ (1996); Chandrasekharan and Huang (1996),
the chiral condensate with nontrivial $\mathbb{Z}_{3}$ sectors are
decreased and thus $\alpha$ should be considerably large.
To conclude the detailed value of $\alpha$, we need more precise
quenched lattice QCD data.
Other promising quantity to remove the model ambiguities is the dual
quark condensate Bilgici et al. (2008, 2010); Bilgici .
The dual quark condensate is the order-parameter
of the $\mathbb{Z}_{N_{\mathrm{c}}}$ symmetry breaking as same
as the Polyakov-loop.
This quantity has been calculated by using the lattice QCD
simulation Bilgici et al. (2008, 2010); Bilgici ,
the Schwinger-Dyson equations Fischer (2009),
the effective model of QCD Kashiwa et al. (2009),
the functional renormalization group equations Braun et al. (2011)
and so on.
The dual quark condensate is defined as
$$\displaystyle\Sigma^{(n)}_{\sigma}$$
$$\displaystyle=-\int\frac{d\varphi}{2\pi}e^{-in\varphi}\sigma_{\varphi},$$
(16)
where $\varphi$ specifies the boundary condition of the temporal
direction for quarks, $n$ means the winding number for the temporal
direction and
$\sigma$ is the $\varphi$-dependent chiral condensate.
Usually, we take $n=1$.
The boundary condition and the imaginary chemical potential have
direct relation, $\theta=\varphi-\pi$.
In the system with dynamical quarks, there are some problems in the dual
quark condensate Benič (2013); Marquez et al. (2015); Zhang and Miao (2016), but
we can expect that the quenched calculations where the dual quark
condensate is well defined can provide the important
information for the effective model construction.
At present, quenched lattice QCD data for the dual quark condensate
are limited and thus we do not quantitatively compare our results with
the lattice data.
Figure 3 shows the $T$-dependence of $\Sigma_{\sigma}^{(1)}$
at $\mu=0$ with different values of $\alpha$.
We can find the visible $\alpha$-dependence in $\Sigma_{\sigma}^{(1)}$.
At least in quenched lattice QCD data we can
access Bilgici et al. (2008),
$\Sigma_{\sigma}^{(1)}$ increases with increasing $T$ and thus extremely
strong $\alpha$ may be excluded, but we need more quenched lattice QCD
data to conclude it.
The critical value ($\alpha_{c}$) which leads to the flat behavior of
$\Sigma_{\sigma}^{(1)}$ above $T_{\mathrm{D}}$ exists in the range
$-0.7<\alpha_{c}<-0.6$.
Finally, we propose a new quantity which can describe the
confinement-deconfinement transition based on the dual quark condensate.
The quantity is defined as
$$\displaystyle\Sigma^{(n)}_{\cal O}$$
$$\displaystyle=\int\frac{d\varphi}{2\pi}e^{-in\varphi}{\cal O}(\varphi),$$
(17)
where ${\cal O}$ is the $\theta$-even quantity.
If we consider that the dual quark condensate measures how
information of $\mathbb{Z}_{3}$ images are missed in the calculation of
Eq. (16) when the gauge configuration is fixed to the trivial
$\mathbb{Z}_{3}$ sector, there is no need to use $\sigma$ as the
integrand of Eq. (16).
It is related with the study based on the nontrivial free energy
degeneracy in Ref. Kashiwa and Ohnishi (2015, 2016).
Figure 4 shows $T$-dependence of the quantity (17)
with ${\cal O}=P$ and $n=1$ where $P$ is the pressure.
We can see that this quantity also describes the
confinement-deconfinement transition.
This behavior is, of course, trivial from the viewpoint of the RW
periodicity.
It is interesting to calculate it in the system with dynamical quarks
and some theories where we cannot easily calculate the chiral condensate.
V Summary
In this paper, we have discussed the extension of the effective model
such as the Polyakov-loop extended Nambu–Jona-Lasinio model.
Firstly, implications of QCD properties in the imaginary chemical
potential region to the effective model construction has been discussed.
Then the importance of the entanglement vertex Sakai et al. (2010)
has been explained from the derivation of the effective four-fermi
interaction from QCD partition function with some approximations and
ansatz.
Secondly, we have shown that the general form of the entanglement vertex
which should be needed to reproduce the lattice QCD prediction at finite
$\mu_{\mathrm{I}}$ D’Elia and Sanfilippo (2009); Bonati et al. (2011).
The general form is refer to the structure of the gluon one-loop
effective potential and the Bose distribution function and then it
naturally leads to the $\mathbb{Z}_{3}$ symmetric form.
Finally, the chiral condensate with different $\mathbb{Z}_{3}$ sectors
and the dual quark condensate with $n=1$
have been calculated by using the PNJL model with the entanglement
vertex.
The entanglement vertex has one parameter $\alpha$ when we assume
$\gamma=1$ at least in the next leading-order of $e^{-\beta E}$.
When $\alpha$ is mall, the chiral condensate with the non-trivial
$\mathbb{Z}_{3}$ sector increases even above $T_{\mathrm{D}}$, but it
decreases in the large-$\alpha$ case.
This result suggests that we can well remove the ambiguity of the
parameters by using the $T$-dependence of the chiral condensate
with the different $\mathbb{Z}_{3}$ sectors.
Also, we can find visible $\alpha$-dependence in the dual quark
condensate and thus this quantity should be the promising quantity to
remove the ambiguity of the parameters.
From qualitative behavior of quenched lattice QCD
data Chandrasekharan and Christ (1996); Chandrasekharan and Huang (1996); Bilgici et al. (2008),
we may exclude extremely strong and weak $\alpha$ values, but we need more
quenched lattice QCD data to conclude it.
In addition to above results, we propose new quantity which can
describe the confinement-deconfinement transition based on the dual
quark condensate.
Acknowledgements.
K.K. thanks A. Ohnishi and H. Kouno for useful comments.
K.K. is supported by Grants-in-Aid for Japan Society for the Promotion
of Science (JSPS) fellows No.26-1717.
References
Fukushima (2004)
K. Fukushima, Phys.Lett. B591, 277 (2004), arXiv:hep-ph/0310121 [hep-ph]
.
Hell et al. (2010)
T. Hell, S. Rossner,
M. Cristoforetti, and W. Weise, Phys. Rev. D81, 074034 (2010), arXiv:0911.3510 [hep-ph] .
Ratti et al. (2006)
C. Ratti, M. A. Thaler, and W. Weise, Phys.Rev. D73, 014019 (2006), arXiv:hep-ph/0506234 [hep-ph]
.
Roessner et al. (2007)
S. Roessner, C. Ratti, and W. Weise, Phys.Rev. D75, 034007 (2007), arXiv:hep-ph/0609281 [hep-ph]
.
Meisinger et al. (2002)
P. N. Meisinger, T. R. Miller, and M. C. Ogilvie, Phys.Rev. D65, 034009 (2002), arXiv:hep-ph/0108009 [hep-ph]
.
Dumitru et al. (2011)
A. Dumitru, Y. Guo,
Y. Hidaka, C. P. K. Altes, and R. D. Pisarski, Phys.Rev. D83, 034022
(2011), arXiv:1011.3820 [hep-ph] .
Kashiwa et al. (2012)
K. Kashiwa, R. D. Pisarski, and V. V. Skokov, Phys.Rev. D85, 114029 (2012), arXiv:1205.0545 [hep-ph] .
Fukushima and Kashiwa (2013)
K. Fukushima and K. Kashiwa, Phys.Lett. B723, 360 (2013), arXiv:1206.0685 [hep-ph] .
Shuryak and Sulejmanpasic (2013)
E. Shuryak and T. Sulejmanpasic, Phys. Lett. B726, 257 (2013), arXiv:1305.0796 [hep-ph] .
D’Elia and Sanfilippo (2009)
M. D’Elia and F. Sanfilippo, Phys. Rev. D80, 111501 (2009), arXiv:0909.0254 [hep-lat] .
Bonati et al. (2011)
C. Bonati, G. Cossu,
M. D’Elia, and F. Sanfilippo, Phys.Rev. D83, 054505
(2011), arXiv:1011.4515 [hep-lat] .
Roberge and Weiss (1986)
A. Roberge and N. Weiss, Nucl.Phys. B275, 734 (1986).
Kashiwa and Ohnishi (2015)
K. Kashiwa and A. Ohnishi, Phys. Lett. B750, 282 (2015), arXiv:1505.06799 [hep-ph] .
Kashiwa and Ohnishi (2016)
K. Kashiwa and A. Ohnishi, (2016), arXiv:1602.06037 [hep-ph] .
Wen (1990)
X. Wen, Int.J.Mod.Phys. B4, 239 (1990).
Sato (2008)
M. Sato, Phys.Rev. D77, 045013 (2008), arXiv:0705.2476 [hep-th] .
Kashiwa and Misumi (2013)
K. Kashiwa and T. Misumi, JHEP 1305, 042 (2013), arXiv:1302.2196 [hep-ph] .
Sakai et al. (2010)
Y. Sakai, T. Sasaki,
H. Kouno, and M. Yahiro, Phys.
Rev. D82, 076003
(2010), arXiv:1006.3648 [hep-ph] .
Sasaki et al. (2011)
T. Sasaki, Y. Sakai,
H. Kouno, and M. Yahiro, Phys.
Rev. D84, 091901
(2011), arXiv:1105.3959 [hep-ph] .
Van den Bossche (1997)
B. Van den Bossche, in Research Workshop on Deconfinement at Finite Temperature and Density Dubna,
Russia, October 1-25, 1997 (1997) arXiv:nucl-th/9807010 [nucl-th]
.
Kondo (2010)
K.-I. Kondo, Phys.Rev. D82, 065024 (2010), arXiv:1005.0314 [hep-th] .
Kashiwa et al. (2011)
K. Kashiwa, T. Hell, and W. Weise, Phys.Rev. D84, 056010 (2011), arXiv:1106.5025 [hep-ph] .
Sasaki and Redlich (2012)
C. Sasaki and K. Redlich, Phys.Rev. D86, 014007 (2012), arXiv:1204.4330 [hep-ph] .
Silva and Oliveira (2016)
P. J. Silva and O. Oliveira, (2016), arXiv:1601.01594 [hep-lat] .
Hell et al. (2009)
T. Hell, S. Roessner,
M. Cristoforetti, and W. Weise, Phys.Rev. D79, 014022 (2009), arXiv:0810.1099 [hep-ph] .
Radzhabov et al. (2011)
A. E. Radzhabov, D. Blaschke,
M. Buballa, and M. K. Volkov, Phys. Rev. D83, 116004 (2011), arXiv:1012.0664 [hep-ph] .
Pagura et al. (2012)
V. Pagura, D. Gomez Dumm,
and N. N. Scoccola, Phys. Lett. B707, 76 (2012), arXiv:1105.1739 [hep-ph] .
Carlomagno et al. (2013)
J. P. Carlomagno, D. Gómez Dumm, and N. N. Scoccola, Phys. Rev. D88, 074034 (2013), arXiv:1305.2969 [hep-ph]
.
Chandrasekharan and Christ (1996)
S. Chandrasekharan and N. H. Christ, Lattice ’95. Proceedings, International Symposium on Lattice Field Theory,
Melbourne, Australia, July 11-15, 1995, Nucl. Phys. Proc. Suppl. 47, 527 (1996), arXiv:hep-lat/9509095 [hep-lat]
.
Chandrasekharan and Huang (1996)
S. Chandrasekharan and S.-z. Huang, Phys. Rev. D53, 5100 (1996), arXiv:hep-ph/9512323 [hep-ph]
.
Bilgici et al. (2008)
E. Bilgici, F. Bruckmann,
C. Gattringer, and C. Hagen, Phys.Rev. D77, 094007 (2008), arXiv:0801.4051 [hep-lat] .
Bilgici et al. (2010)
E. Bilgici, F. Bruckmann,
J. Danzer, C. Gattringer, C. Hagen, E. M. Ilgenfritz, and A. Maas, Few Body Syst. 47, 125 (2010), arXiv:0906.3957 [hep-lat]
.
(33)
E. Bilgici, “Signatures of
confinement and chiral symmetry breaking in spectral quantities of lattice
Dirac operators,” University of Graz, 2009,
(http://physik.uni-graz.at/itp/files/bilgici/dissertation.pdf).
Fischer (2009)
C. S. Fischer, Phys.Rev.Lett. 103, 052003 (2009), arXiv:0904.2700 [hep-ph] .
Kashiwa et al. (2009)
K. Kashiwa, H. Kouno, and M. Yahiro, Phys.Rev. D80, 117901 (2009), arXiv:0908.1213 [hep-ph] .
Braun et al. (2011)
J. Braun, L. M. Haas,
F. Marhauser, and J. M. Pawlowski, Phys. Rev. Lett. 106, 022002 (2011), arXiv:0908.0008 [hep-ph] .
Benič (2013)
S. Benič, Phys.Rev. D88, 077501 (2013), arXiv:1305.6567 [hep-ph] .
Marquez et al. (2015)
F. Marquez, A. Ahmad,
M. Buballa, and A. Raya, Phys. Lett. B747, 529
(2015), arXiv:1504.06730 [nucl-th] .
Zhang and Miao (2016)
Z. Zhang and Q. Miao, Phys. Lett. B753, 670 (2016), arXiv:1507.07224 [hep-ph] . |
IDCS J1426.5+3508: Weak Lensing Analysis of a Massive Galaxy Cluster at $z=1.75$
Wenli Mo11affiliation: Department of Astronomy, University of Florida, Bryant Space Science Center, Gainesville, FL 32611, USA ,
Anthony Gonzalez11affiliation: Department of Astronomy, University of Florida, Bryant Space Science Center, Gainesville, FL 32611, USA ,
M. James Jee22affiliation: Department of Astronomy and Center for Galaxy Evolution Research, Yonsei University, 50 Yonsei-ro, Seoul 03722, Korea ,
Richard Massey33affiliation: Institute for Computational Cosmology, Durham University, South Road, Durham DH1 3LE, UK ,
Jason Rhodes44affiliation: Jet Propulsion Laboratory, California Institute of Technology, Pasadena, CA 91109, USA 55affiliation: California Institute of Technology, Pasadena, CA 91125, USA ,
Mark Brodwin66affiliation: Department of Physics and Astronomy, University of Missouri, 5110 Rockhill Road, Kansas City, MO 64110, USA ,
Peter Eisenhardt44affiliation: Jet Propulsion Laboratory, California Institute of Technology, Pasadena, CA 91109, USA ,
Daniel P. Marrone77affiliation: Steward Observatory, University of Arizona, 933 North Cherry Avenue, Tucson, AZ 85721, USA ,
S. A. Stanford88affiliation: Department of Physics, University of California, Davis, One Shields Avenue, Davis, CA 95616, USA ,
and
Gregory R. Zeimann99affiliation: Department of Astronomy & Astrophysics, The Pennsylvania State University, University Park, PA 16802, USA 1010affiliation: Institute for Gravitation and the Cosmos, The Pennsylvania State University, University Park, PA 16802, USA 1111affiliation: Department of Astronomy, University of Texas, Austin, TX 78712, USA
Abstract
We present a weak lensing study of the galaxy cluster IDCS J1426.5+3508 at $z=1.75$, which is the highest redshift strong lensing cluster known and the most distant cluster for which a weak lensing analysis has been undertaken. Using F160W, F814W, and F606W observations with the Hubble Space Telescope, we detect tangential shear at $2\sigma$ significance. Fitting a Navarro-Frenk-White mass profile to the shear with a theoretical median mass-concentration relation, we derive a mass $M_{200,\mathrm{crit}}=2.3^{+2.1}_{-1.4}\times 10^{14}$ M${}_{\odot}$. This mass is consistent with previous mass estimates from the Sunyaev-Zel’dovich (SZ) effect, X-ray, and strong lensing. The cluster lies on the local SZ-weak lensing mass scaling relation observed at low redshift, indicative of minimal evolution in this relation.
Subject headings: cosmology: observations – dark matter – galaxies: clusters: individual (IDCS J1426.5+3508) – gravitational lensing: weak.
1. Introduction
Massive, high-redshift galaxy clusters, though rare, provide valuable information about cosmological parameters, structure formation, and galaxy evolution. The redshift regime $z>1.5$ represents a critical era during which significant star formation occurs in cluster galaxies (e.g, Tran et al., 2010; Brodwin et al., 2013; Bayliss et al., 2014; Webb et al., 2015b; Alberts et al., 2014). Only a handful of clusters at redshift $z>1.5$ have been confirmed to date (e.g, Newman et al., 2014; Papovich et al., 2010; Zeimann et al., 2012; Muzzin et al., 2013; Tozzi et al., 2015; Webb et al., 2015a), thanks in part to large X-ray, infrared, and Sunyaev-Zel’dovich (SZ) effect surveys (e.g., Eisenhardt et al., 2008; Fassbender et al., 2011; Wylezalek et al., 2014; Bleem et al., 2015).
In this paper we focus upon the galaxy cluster IDCS J1426.5+3508. This cluster was first discovered in the IRAC Distant Cluster Survey (IDCS). Follow-up with the Hubble Space Telescope (HST) Wide Field Camera-3 (WFC3) grism and the Low Resolution Imaging Spectrometer (LRIS, Oke et al., 1995) on the W. M. Keck Observatory spectroscopically confirms a redshift of $z=1.75$ (Stanford et al., 2012). With $100$ ks of Chandra data, Brodwin et al. (2015) obtain mass estimates based upon the X-ray temperature, gas mass, and the product of core-excised X-ray temperature and gas mass of $M_{500,T_{X}}=3.3^{+5.7}_{-1.2}\times 10^{14}$ M${}_{\odot}$, $M_{500,M_{g}}=2.3^{+0.7}_{-0.5}\times 10^{14}$ M${}_{\odot}$, and $M_{500,\mathrm{Y}_{X}}=2.6^{+1.5}_{-0.5}\times 10^{14}$ M${}_{\odot}$, respectively. In comparison, SZ observations with the Combined Array for Research in Millimeter-wave Astronomy (CARMA) indicate $M_{500,\mathrm{SZ}}=2.6\pm 0.7\times 10^{14}$ M${}_{\odot}$ (Brodwin et al., 2012), implying $M_{200}=4.1\pm 1.1\times 10^{14}$ M${}_{\odot}$ with a Duffy et al. (2008) mass-concentration relation. These mass estimates establish IDCS J1426.5+3508 as the most massive galaxy cluster confirmed at redshift $z>1.5$. For comparison, another high-redshift cluster, XDCP J0044.0-2033 ($z=1.58$), has mass derived from the Vikhlinin et al. (2009) $Y_{X}-$M scaling relation $M_{500,Y_{X}}=2.2^{+0.5}_{-0.4}\times 10^{14}$ M${}_{\odot}$ (Tozzi et al., 2015).
Moreover, IDCS J1426.5+3508 is also the most distant strong lensing galaxy cluster. A giant gravitationally lensed arc associated with the cluster was discovered in HST/WFC3 and Advanced Camera for Surveys (ACS) imaging. An initial redshift estimate based on broadband photometry yielded $z\sim 3-6$ for the arc and a lower limit of $M_{200}>2.8\times 10^{14}$ M${}_{\odot}$ for the cluster via a strong-lensing analysis (Gonzalez et al., 2012). For $\Lambda$CDM, Gonzalez et al. (2012) calculate that the expected number of giant arcs at this redshift and brightness is vanishingly small, highlighting the long-standing arc statistics problem (see Meneghetti et al., 2013, for a recent review). One explanation suggested to explain the observed arc in IDCS J1426.5+3508 is that the cluster might have a substantially more concentrated density profile—and hence an enhanced lensing cross-section—than predicted for a cluster at this epoch.
Gravitational weak lensing determines the cluster mass via the distortion in shape of background galaxies due to the gravitational potential of the cluster. Unlike mass estimates derived from X-ray or SZ observations, weak lensing measures the mass distribution independent of the dynamical state or hydrostatic equilibrium of the cluster. Thus, weak lensing is also a powerful tool for calibrating X-ray or SZ mass estimates (e.g., Marrone et al., 2012; Hoekstra et al., 2012; Jee et al., 2011). Furthermore, the mass and concentration of the cluster can in principle be determined without a prescribed mass-concentration relationship.
However, weak lensing is observationally challenging, particularly for distant clusters. Because the lensing cluster is at high redshift, the number of lensed background galaxies is small, decreasing the lensing signal. Also, the extent of the distortions to be measured is on the scale of the observational point spread function (PSF; e.g., Rhodes et al., 2007; Jee et al., 2007). Deep, space-based observations with well-understood PSFs are therefore required to obtain high fidelity measurements of the weak lensing distortion.
In this work, we demonstrate the feasibility of weak lensing analyses at $z=1.75$. IDCS J1426.5+3508 is the highest redshift cluster to have joint X-ray, SZ, and weak lensing observations, and thus provides an opportunity to compare whether the scaling relations between these mass estimators, derived at low redshift, remain valid at this epoch. In Section 2 we describe the observations and data analysis, including the selection of source galaxies. We outline the measurement of galaxy shapes and the correction for PSF distortion in those measurements in Section 3. The weak lensing shear profile is presented in Section 4. We discuss the mass estimation and SZ-weak lensing scaling relation in Sections 5 and 6, respectively, and summarize our findings in Section 7.
Throughout the paper, we adopt the nine-year Wilkinson Microwave Anisotropy Probe (WMAP9) cosmological parameters of $\Omega_{\mathrm{M}}=0.287$, $\Omega_{\Lambda}=0.713$, and $H_{0}=69.32$ km s${}^{-1}$ Mpc${}^{-1}$ (Hinshaw et al., 2013) and define $h=H_{0}/100$. All magnitudes are in the AB system (Oke & Gunn, 1983). We take the F606W image for our shape measurement and the center of the cluster to be the location of the brightest cluster galaxy (BCG), as identified by Stanford et al. (2012). Unless otherwise stated, we report masses as overdensities relative to the critical density.
2. Observations
For our analysis, we use HST data from Cycle 20 taken on 2012 December 19, and 2013 April 17-19, with the ACS F606W (V-band), F814W (I-band), and WFC3 F160W (H-band) filters for a total of 21760, 8108, and 4947 seconds, respectively. We also use Cycle 17 observations from 2010 July 8 and 2010 November 7 with the F814W and F160W filters for 4513 and 2612 seconds, respectively. Since the F160W data are mosaicked, the effective exposure time for the central region is 7559 seconds while the minimum exposure time for the outer regions is 1212 seconds. The cluster is also among the targets in HST GO-program 13677 (PI Perlmutter) which will add observations in WFC3 F140W, F105W, and F814W filters for future studies of this cluster.
2.1. HST ACS Data Reduction
We performed basic processing of the ACS data with the CALACS pipeline. Our only modification to the standard procedure was to employ the charge transfer inefficiency (CTI) correction method developed by Massey et al. (2010, 2014) to correct for CTI degradation prior to running CALACS. CTI extends the shapes of galaxies and, if uncorrected, can imitate the effects of weak lensing. Astrometric shifts, geometric distortion correction, sky subtraction, cosmic ray removal, and final image stacking were handled by the standard TweakReg and Astrodrizzle packages. Images were drizzled to a $0.03\arcsec$ pixel scale with a Gaussian kernel and pixel fraction of 1.0. Rhodes et al. (2007) showed that changing the pixel scale from the ACS native scale of $0.05\arcsec$ to $0.03\arcsec$ optimizes our ability to correct for PSF effects while the results are largely insensitive to the choice of pixel fraction.
2.2. Source Galaxy Catalog
Starting with F606W as our detection image, we extracted a catalog of 4771 sources with SExtractor (Bertin & Arnouts, 1996) with a criteria of $5$ connecting pixels brighter than a sky rms of $2$. We then obtained matched photometry for these sources in F814W and F160W by running SExtractor in dual image mode. Figure 1 is the resulting color-magnitude diagram for the two bluest filters.
We identify the foreground and cluster galaxy contamination in our catalog via color selection. One benefit of employing an F606W-selected catalog is that most cluster red sequence galaxies are non-detections. Listed below are the sequence of color and size cuts applied to distinguish among foreground galaxies, cluster members, and background galaxies. We detect 3050 objects with F606W$<28.0$, the approximate $10\sigma$ depth of our observations. We first reject $295$ sources with either F606W$<24.0$ or FWHM$>0.9\arcsec$ to exclude bright and large objects that are probable foreground galaxies. Based on a Bruzual & Charlot (2003) model of a star-forming galaxy with $z_{\mathrm{form}}=6$, we next implement a color cut of F606W-F814W $>0.5$ to eliminate $786$ additional foreground galaxies in our sample. To reject cluster members on the red sequence, we rule out $20$ objects with F814W-F160W $>3.0$ (Stanford et al., 2012). We also cut $16$ objects with signal to noise $S/N<10$.
In addition, we exclude $922$ objects based on shape measurements constraints, further discussed in Section 3. The final catalog for the shear analysis consists of 1011 sources, equivalent to $89$ arcmin${}^{-2}$. As a cross-check, we construct a radial density profile for the remaining galaxies centered on the BCG. There is no central excess evident in the radial profile, indicating that residual contamination from blue cluster galaxies is minimal.
2.3. Redshift Distribution
Given the redshift of the cluster, compiling a clean catalog of background sources is not possible with the available photometric data. However, we can control the foreground contamination in our source catalog statistically. We estimate the redshift distribution of our source catalog using the Cosmic Assembly Near-infrared Deep Extragalactic Legacy Survey Ultra Deep Survey (CANDELS-UDS) multiwavelength catalog (Santini et al., 2014) which provides infrared-detected sources to an F160W magnitude limit of $27.45$ in a range of filters. Employing the selection criteria described in Section 2.2, we found 3512 sources, 46% of which were contaminants with $z\leq 1.75$. The distribution of CANDELS-UDS redshift for sources included and excluded by our selection criteria is illustrated in Figure 2.
We estimate the effective redshift of the source population by calculating the lensing efficiency, the mean of the ratio of angular diameter distances,
$$\langle\beta\rangle=\langle\mathrm{max}(0,{D_{ls}/D_{s}})\rangle,$$
(1)
where $D_{ls}$ and $D_{s}$ are the angular diameter distances between the lens and the source and the observer and the source, respectively. We set a floor of $\beta=0$ to account for foreground sources with $z\leq 1.75$. For the subset of CANDELS-UDS sources matching our selection criteria, we compute $\langle\beta\rangle=0.086$, corresponding to an effective redshift $z_{\mathrm{eff}}=2.05$, and width $\langle\beta^{2}\rangle=0.017$.
Despite CANDELS-UDS being the deepest catalog with data in our set of filters, 12% of the galaxies in our background source catalog are fainter in F160W than the CANDELS-UDS limiting magnitude. To investigate the effects of the mismatch in detection filter threshold between our data and the CANDELS-UDS catalog, we randomly select 114 objects from a sample of $1455$ CANDELS-UDS objects near the survey limiting magnitude ($27.0<\mathrm{F160W}<27.45$) and add their redshifts to the 3512 CANDELS-UDS sources matching our selection criteria. We choose these objects because they are most likely to mimic the redshift distribution of the sources with F160W$>27.45$ missing from the CANDELS-UDS catalog. We then repeat our calculation of $\beta$ using Equation 1 with the additional 12% of galaxies. Including the randomly selected redshift sample increases $\beta$ by $3.5\%$. This source of uncertainty for $\beta$ is a subdominant contributor to the total mass error budget.
3. Shape Measurement
3.1. RRG Method
We measure galaxy distortions by applying the RRG method developed by Rhodes et al. (2000), which calculates the second- and fourth-order Gaussian-weighted moments of each galaxy to determine ellipticity and correct for convolutions with the PSF. The weak lensing shear is then derived from the average ellipticity of the sources. We choose the weight function size of each object to be $w=\max(2\sqrt{ab},6)$ where $a$ and $b$ are the semi-major and -minor axes in pixels calculated from SExtractor.
For each object, the shear $\gamma$ is related to the ellipticity $e$ of that object
$$\gamma=C\frac{e}{G},$$
(2)
where $G=1.35$ is the shear susceptibility calculated from our data and $C=0.86^{-1}$ is a calibration factor determined from the analysis of simulated images containing a known shear (Leauthaud et al., 2007). The reduced shear is then
$$g=\frac{\gamma}{(1-\kappa)}$$
(3)
where $\kappa$ is the convergence. We further incorporate the width of the redshift distribution (Seitz & Schneider, 1997). The observed shear $g^{\prime}$ is then
$$g^{\prime}=(1+(\langle\beta^{2}\rangle/\langle\beta\rangle^{2}-1)\kappa)g=(1+1.55\kappa)g.$$
(4)
In addition to the criteria presented in Section 2.2, we introduce more selection criteria to our background source catalog due to shape measurement limitations. The RRG method cannot correct for the distortion on objects smaller than $\sim 1.5$ times the RRG weight function width of $0.18\arcsec$. Therefore, we reject $730$ galaxies with a SExtractor FWHM $<0.27\arcsec$ for which we were not able to apply a PSF correction. Also, we exclude $192$ objects that encountered centroiding errors during shape measurement or for which the PSF correction did not converge. Therefore, our final catalog for weak lensing analysis includes 1011 sources.
3.2. PSF Correction
The ACS PSF is known to be both temporally and spatially variable. Thermal breathing of the telescope, dependent on the 90-minute orbit of HST, and a slow deviation in the focus on the timescale of a few weeks further complicate the PSF. In the weak lensing regime, an uncorrected PSF distortion can drown out the signal, so precise characterization is critical.
We use the publicly available TinyTim software111http://tinytim.stsci.edu/ (Krist & Hook, 1997; Krist et al., 2011) to theoretically model the PSF distortion across our image. TinyTim simulates the PSF given a focus value and takes into account the instrument, chip number, chip position, filter, and spectrum of the object. Schrabback et al. (2010) notes additional variations in the PSF dependent on relative sun angle, but finds that $97\%$ of variations can be described solely by the focus parameter.
Our F606W image is a combination of 16 separate frames. We replicate the PSF in each frame individually by specifying a focus value determined by the HST focus model222http://focustool.stsci.edu/, applying the average focus during the exposure period of each individual frame.
Adopting a modified version of TinyTim333http://community.dur.ac.uk/r.j.massey/acs/PSF, we generate a distortion-corrected grid of PSFs across the entire image with pixel scale of $0.03\arcsec$. We create the PSF model for the full stacked F606W image by combining the 16 individual PSFs weighted by the exposure time. We then interpolate the second- and fourth-order moments of the PSF model to the arbitrary positions of galaxies throughout the field of view via a third-order polynomial in $x$ and $y$.
4. Weak Lensing Analysis
The weak lensing signal is expressed as the tangential component of the shear $g_{+}$. The cross component $g_{\times}$, defined as $g_{+}$ after the object is rotated by $45^{\circ}$ and represents the null statistic and noise of our measurements and should be consistent with zero in the absence of systematic errors.
We determine the shear as a function of cluster radius by radially binning our catalog objects and calculating the average $g^{\prime}_{+}$ and $g^{\prime}_{\times}$ of all objects in each bin. The cluster center is at the location of the BCG (RA=$14^{\mathrm{h}}26^{\mathrm{m}}32^{\mathrm{s}}.95$ and Dec=$+35^{\circ}08^{\prime}23~{}6\arcsec$). The shear profile is plotted in Figure 3 where the error bars shown for the shear measurements are the statistical uncertainty of the mean shear per bin. By adding in quadrature the ratio of mean shear and uncertainty per bin for the first six bins, we calculate a significance of $2.0$ for the detection. The Pearson’s correlation coefficient between the tangential and cross shear components is small ($r=-0.12$), indicating no correlation between the two components
As a test of robustness of this detection, we also employ an alternative shape measurement from the Jee et al. (2009, hereafter J09) and RRG methods. Using the same galaxy selection, we find that the shapes obtained using the J09 and RRG methods yield consistent results: the typical absolute difference between the mean measured values in each bin is within $0.8\sigma$ for the tangential shear and $1.3\sigma$ for the cross component.
The J09 approach is designed to extend to faint magnitudes. With this method, we also perform a 2-D mass reconstruction using a more relaxed set of selection criteria. Specifically, we allow a wider range in magnitudes ($22<$F606W$<29$ and $22<$F160W$<29$), a relaxed red sequence selection (F814W-F160W$<1.8$), and a decreased size selection (F606W half-light radius $r_{h}>1.2$). The centroid of the resulting 2-D mass reconstruction is close to the location of the BCG, providing further evidence that the shear is the result of the cluster weak lensing. The shear contours are overplotted on the HST image in Figure 4. Finally, we note that the tangential shear profile using this different selection remains consistent with Figure 3 but with smaller statistical uncertainties. We continue the analysis with our original shape measurement.
5. Mass Estimation
The tangential shear is related to the density profile of the cluster assuming a Navarro-Frenk-White (NFW) profile (Navarro et al., 1997) described by the virial radius $r_{200}$, defined as the radius inside which the mass density of the halo is equal to $200$ times the critical density, and the concentration $c=r_{200}/r_{s}$, where $r_{s}$ is the scale radius. For the radial dependence of the shear in an NFW model, we refer the reader to Equations $14-17$ of Wright & Brainerd (2000).
To fit the shear profile, we utilize the maximum likelihood estimation discussed in Schneider et al. (2000). The log-likelihood function is defined as
$$\ell_{\gamma}=\sum_{i=1}^{N\gamma}{\left[\frac{|e_{t,i}-g^{\prime}_{+}(\theta_{i})|^{2}}{\sigma^{2}[g^{\prime}_{+}(\theta_{i})]}\right]+2\ln\sigma[g^{\prime}_{+}(\theta_{i})]},$$
(5)
where $N_{\gamma}$ is the number of galaxies with measured ellipticity, $e_{t,i}$ is the tangential ellipticity component, $\theta_{i}$ is the position of the $i$-th galaxy, and $g^{\prime}_{+}(\theta_{i})$ is the observed tangential shear at $\theta_{i}$. The dispersion of observed ellipticities approximated as $\sigma[g^{\prime}_{+}(\theta_{i})]\approx\sigma_{e}(1-|g^{\prime}_{+}(\theta_{i})|^{2})$, the same as that defined in Equation 14 of Schneider et al. (2000), where $\sigma_{e}=0.3$ is the intrinsic ellipticity dispersion.
We lack the necessary signal to fit the mass and concentration simultaneously as free parameters. Thus, we derive the cluster mass with the concentration defined by the mass-concentration relation from Duffy et al. (2008),
$$c=5.71\Bigg{(}\frac{M_{200,\mathrm{crit}}}{2\times 10^{12}h^{-1}M_{\odot}}\Bigg{)}^{-0.084}(1+z)^{-0.47},$$
(6)
where the mass within $r_{200}$ is $M_{200,\mathrm{crit}}=200\rho_{\mathrm{crit}}(\frac{4}{3}\pi{r_{200}^{3}})$. In this case, $M_{200,\mathrm{crit}}$ is the only free parameter in the likelihood fit, and the resultant value can be applied to directly compare IDCS J1426.5+3508 with the SZ-lensing relation defined by low redshift clusters. All objects located at radii $15\arcsec<{r}<110\arcsec$ ($130<r<950$ kpc) are included in our fit (802 sources). We exclude the strong lensing region, defined by the radius of the known strongly-lensed arc (Gonzalez et al., 2012), and extend to a radius that best constrains the $1\sigma$ mass confidence interval. We derive $M_{200,\mathrm{crit}}=2.3^{+2.1}_{-1.4}\times 10^{14}$ M${}_{\odot}$ ($M_{500,\mathrm{crit}}=1.4^{+1.3}_{-0.9}\times 10^{14}~{}M_{\odot}$) which corresponds to $r_{200}=0.68^{+0.16}_{-0.18}$ Mpc ($r_{500}=0.42^{+0.10}_{-0.11}$ Mpc). Changing the center to that of the X-ray observations (Brodwin et al., 2015) yielded a mass within $1\sigma$ of the mass derived with BCG centering. The NFW profile described by our best-fit parameters is plotted in Figure 3.
IDCS J1426.5+3508 is the only galaxy cluster in this regime to have joint weak lensing, strong lensing, SZ, and X-ray observations. Brodwin et al. (2015) find good agreement between the mass estimations from X-ray, SZ, and strong lensing. Converting from $M_{500,\mathrm{crit}}$ to $M_{200,\mathrm{crit}}$ via the Duffy et al. (2008) relation, Brodwin et al. (2012) calculate an SZ mass $M_{200,\mathrm{crit}}^{\mathrm{SZ}}=4.1\pm 1.1\times 10^{14}$ M${}_{\odot}$, Brodwin et al. (2015) estimate a mass from the X-ray gas mass of $M_{200}^{\mathrm{M_{gas}}}=3.8^{+1.1}_{-0.8}\times 10^{14}$ M${}_{\odot}$, while Gonzalez et al. (2012) projected a lower limit of $M_{200}>2.8^{+1.0}_{-0.4}\times 10^{14}$ M${}_{\odot}$ from observations of the giant arc. Our weak lensing mass of $M_{200,\mathrm{crit}}=2.3^{+2.1}_{-1.4}\times 10^{14}$ M${}_{\odot}$ falls at the lower end of these mass estimates but is consistent.
6. SZ-Weak Lensing Scaling Relation
The SZ Compton parameter $Y$ scales with mass as $Y\propto{M^{5/3}/({D_{A}^{2}E(z)^{-2/3}})}$ where $D_{A}$ is the angular diameter distance and $E(z)$ is the evolution of the Hubble parameter. We combine our weak lensing-derived mass estimate with the spherically averaged dimensionless Comptonization parameter ($Y_{\mathrm{sph},500}=7.9\pm 3.2\times 10^{-12}$) from Brodwin et al. (2012) to compare IDCS J1426.5+3508 with the mass-$Y_{\mathrm{sph}}$ scaling relation derived by Marrone et al. (2012) using 18 Local Cluster Substructure Survey (LoCuSS) clusters in the range $z=[0.164,0.290]$. Figure 5 shows our results alongside the Marrone et al. (2012) sample and scaling relation. In this figure, we match the $\Lambda$CDM cosmology assumed by Marrone et al. (2012, $\Omega_{\mathrm{M}}=0.27$, $\Omega_{\Lambda}=0.73$, $H_{0}=73$ km s${}^{-1}$ Mpc${}^{-1}$). We also include the ACT-CL J0022.2-0036 at $z=0.81$ (Miyatake et al., 2013; Reese et al., 2012), a higher redshift cluster with comparable SZ and weak lensing observations, for comparison.
Our results for IDCS J1426.5+3508 are statistically consistent with the low redshift data, indicating minimal redshift evolution in the Marrone et al. (2012) scaling relation. To quantify the redshift evolution, we fit a regression of the form
$$\Big{(}\frac{1+z}{1+z_{\mathrm{M12}}}\Big{)}^{\alpha}=\Big{(}\frac{M_{500,\mathrm{WL}}}{10^{A}10^{14}M_{\odot}}\Big{)}\Big{(}\frac{10^{-5}\mathrm{Mpc}^{2}}{Y_{\mathrm{sph},500}D_{A}^{2}E(z)^{-2/3}}\Big{)}^{B}$$
(7)
where $z_{\mathrm{M12}}=0.24$ is the average redshift of the LoCuSS sample and $A=0.367$ and $B=0.44$ are regression coefficients from Marrone et al. (2012). Including the data for the $18$ LoCuSS clusters individually, ACT-CL J0022.2-0036, and IDCS J1426.5+3508, a least-squares fit yields $\alpha=-0.1\pm 1.0$, within $1\sigma$ of zero. Thus, we find no significant evidence for evolution in the Marrone et al. (2012) scaling relation.
7. Conclusions
We have presented the weak lensing analysis of IDCS J1426.5+3508. At $z=1.75$, this is the highest redshift galaxy cluster to be studied through weak lensing. We detect a tangential shear signal at a $2\sigma$-level significance. Assuming an NFW profile and a Duffy et al. (2008) mass-concentration relation, we fit for the cluster mass with a maximum-likelihood algorithm. Our mass estimate $M_{200,\mathrm{crit}}=2.3^{+2.1}_{-1.4}\times 10^{14}$ M${}_{\odot}$ is in agreement with estimates from X-ray, SZ, and strong lensing data. We also find that the SZ and weak lensing mass estimates agree well with the local scaling relation of Marrone et al. (2012), with negligible evolution in the relation.
Gonzalez et al. (2012) initially discovered the strong-lensed arc behind IDCS J1426.5+3508. If one makes standard assumptions for the galaxy cluster mass function, cosmological model, source galaxy redshift distribution, and cluster lensing cross section, then such an arc should not exist across the entire sky. The cluster lensing cross section increases with concentration, so one plausible explanation for this arc would be if the projected concentration of the cluster mass profile dramatically exceeds the typical value predicted for a cluster of this mass at this epoch. We are unable to determine a concentration as a free parameter for this cluster with the current data. However, a higher signal-to-noise weak lensing map and additional passbands to determine photometric redshifts would enable simultaneous fitting of the mass and concentration, improving Figure 3 and helping resolve the origin of the strong arc in this cluster.
Acknowledgments
The authors thank the anonymous referee and Daniel Stern for their insightful suggestions and Audrey Galametz for her help with the CANDELS-UDS data. Support for HST GO-program 11663, 12203, and 12994 was provided by NASA through a grant from the Space Telescope Science Institute, which is operated by the Association of Universities for Research in Astronomy, Inc., under NASA contract NAS 5-26555. We also acknowledge funding received from the National Science Foundation Graduate Research Fellowship under Grant No. DGE-1315138 (W.M.), NRF of Korea to CGER (M.J.J.), and Royal Society University Research Fellowship (R.M.). J.R. is supported by JPL, which is run by Caltech under a contract for NASA.
References
Alberts et al. (2014)
Alberts, S., Pope, A., Brodwin, M., et al. 2014, MNRAS, 437, 437
Bayliss et al. (2014)
Bayliss, M. B., Ashby, M. L. N., Ruel, J., et al. 2014, ApJ, 794, 12
Bertin & Arnouts (1996)
Bertin, E., & Arnouts, S. 1996, A&AS, 117, 393
Bleem et al. (2015)
Bleem, L. E., Stalder, B., de Haan, T., et al. 2015, ApJS, 216, 27
Brodwin et al. (2015)
Brodwin, M., McDonald, M., Gonzalez, A. H., et al. 2015, ArXiv
e-prints, arXiv:1504.01397
Brodwin et al. (2012)
Brodwin, M., Gonzalez, A. H., Stanford, S. A., et al. 2012, ApJ, 753,
162
Brodwin et al. (2013)
Brodwin, M., Stanford, S. A., Gonzalez, A. H., et al. 2013, ApJ, 779,
138
Bruzual & Charlot (2003)
Bruzual, G., & Charlot, S. 2003, MNRAS, 344, 1000
Duffy et al. (2008)
Duffy, A. R., Schaye, J., Kay, S. T., & Dalla Vecchia, C. 2008,
MNRAS, 390, L64
Eisenhardt et al. (2008)
Eisenhardt, P. R. M., Brodwin, M., Gonzalez, A. H., et al. 2008, ApJ,
684, 905
Fassbender et al. (2011)
Fassbender, R., Böhringer, H., Nastasi, A., et al. 2011, New
Journal of Physics, 13, 125014
Gonzalez et al. (2012)
Gonzalez, A. H., Stanford, S. A., Brodwin, M., et al. 2012, ApJ, 753,
163
Hinshaw et al. (2013)
Hinshaw, G., Larson, D., Komatsu, E., et al. 2013, ApJS, 208, 19
Hoekstra et al. (2012)
Hoekstra, H., Mahdavi, A., Babul, A., & Bildfell, C. 2012, MNRAS,
427, 1298
Jee et al. (2007)
Jee, M. J., Blakeslee, J. P., Sirianni, M., et al. 2007, PASP, 119,
1403
Jee et al. (2009)
Jee, M. J., Rosati, P., Ford, H. C., et al. 2009, ApJ, 704, 672
Jee et al. (2011)
Jee, M. J., Dawson, K. S., Hoekstra, H., et al. 2011, ApJ, 737, 59
Krist & Hook (1997)
Krist, J. E., & Hook, R. N. 1997, in The 1997 HST Calibration Workshop
with a New Generation of Instruments, ed. S. Casertano, R. Jedrzejewski,
T. Keyes, & M. Stevens, 192
Krist et al. (2011)
Krist, J. E., Hook, R. N., & Stoehr, F. 2011, in Society of
Photo-Optical Instrumentation Engineers (SPIE) Conference Series, Vol. 8127,
Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series
Leauthaud et al. (2007)
Leauthaud, A., Massey, R., Kneib, J.-P., et al. 2007, ApJS, 172, 219
Marrone et al. (2012)
Marrone, D. P., Smith, G. P., Okabe, N., et al. 2012, ApJ, 754, 119
Massey et al. (2010)
Massey, R., Stoughton, C., Leauthaud, A., et al. 2010, MNRAS, 401, 371
Massey et al. (2014)
Massey, R., Schrabback, T., Cordes, O., et al. 2014, MNRAS, 439, 887
Meneghetti et al. (2013)
Meneghetti, M., Bartelmann, M., Dahle, H., & Limousin, M. 2013, Space Sci. Rev.,
177, 31
Miyatake et al. (2013)
Miyatake, H., Nishizawa, A. J., Takada, M., et al. 2013, MNRAS, 429,
3627
Muzzin et al. (2013)
Muzzin, A., Wilson, G., Demarco, R., et al. 2013, ApJ, 767, 39
Navarro et al. (1997)
Navarro, J. F., Frenk, C. S., & White, S. D. M. 1997, ApJ, 490, 493
Newman et al. (2014)
Newman, A. B., Ellis, R. S., Andreon, S., et al. 2014, ApJ, 788, 51
Oke & Gunn (1983)
Oke, J. B., & Gunn, J. E. 1983, ApJ, 266, 713
Oke et al. (1995)
Oke, J. B., Cohen, J. G., Carr, M., et al. 1995, PASP, 107, 375
Papovich et al. (2010)
Papovich, C., Momcheva, I., Willmer, C. N. A., et al. 2010, ApJ, 716,
1503
Reese et al. (2012)
Reese, E. D., Mroczkowski, T., Menanteau, F., et al. 2012, ApJ, 751,
12
Rhodes et al. (2000)
Rhodes, J., Refregier, A., & Groth, E. J. 2000, ApJ, 536, 79
Rhodes et al. (2007)
Rhodes, J. D., Massey, R. J., Albert, J., et al. 2007, ApJS, 172, 203
Santini et al. (2014)
Santini, P., Ferguson, H. C., Fontana, A., et al. 2014, ArXiv e-prints,
arXiv:1412.5180
Schneider et al. (2000)
Schneider, P., King, L., & Erben, T. 2000, A&A, 353, 41
Schrabback et al. (2010)
Schrabback, T., Hartlap, J., Joachimi, B., et al. 2010, A&A, 516, A63
Seitz & Schneider (1997)
Seitz, C., & Schneider, P. 1997, A&A, 318, 687
Stanford et al. (2012)
Stanford, S. A., Brodwin, M., Gonzalez, A. H., et al. 2012, ApJ, 753,
164
Tozzi et al. (2015)
Tozzi, P., Santos, J. S., Jee, M. J., et al. 2015, ApJ, 799, 93
Tran et al. (2010)
Tran, K.-V. H., Papovich, C., Saintonge, A., et al. 2010, ApJ, 719,
L126
Vikhlinin et al. (2009)
Vikhlinin, A., Burenin, R. A., Ebeling, H., et al. 2009, ApJ, 692,
1033
Webb et al. (2015a)
Webb, T., Noble, A., DeGroot, A., et al. 2015a, ApJ, 809,
173
Webb et al. (2015b)
Webb, T., Muzzin, A., Noble, A., et al. 2015b, ArXiv
e-prints, arXiv:1508.07302
Wright & Brainerd (2000)
Wright, C. O., & Brainerd, T. G. 2000, ApJ, 534, 34
Wylezalek et al. (2014)
Wylezalek, D., Vernet, J., De Breuck, C., et al. 2014, ApJ, 786, 17
Zeimann et al. (2012)
Zeimann, G. R., Stanford, S. A., Brodwin, M., et al. 2012, ApJ, 756,
115 |
Theory of Non-Interacting Fermions and Bosons in the Canonical Ensemble
Hatem Barghathi
Department of Physics, University of Vermont, Burlington, VT 05405, USA
Department of Physics, Missouri University of Science and Technology, Rolla, MO 65409, USA
Jiangyong Yu
Department of Physics, University of Vermont, Burlington, VT 05405, USA
Adrian Del Maestro
Department of Physics, University of Vermont, Burlington, VT 05405, USA
Abstract
We present a self-contained theory for the exact calculation of particle number counting statistics of non-interacting indistinguishable particles in the canonical ensemble. This general framework introduces the concept of auxiliary partition functions, and represents a unification of previous distinct approaches with many known results appearing as direct consequences of the developed mathematical structure. In addition, we introduce a general decomposition of the correlations between occupation numbers in terms of the occupation numbers of individual energy levels, that is valid for both non-degenerate and degenerate spectra. To demonstrate the applicability of the theory in the presence of degeneracy, we compute energy level correlations up to fourth order in a bosonic ring in the presence of a magnetic field.
I Introduction
In quantum statistical physics, the analysis of a fixed number $N$ of indistinguishable particles is difficult, even in the non-interacting limit. In such a canonical ensemble, the constraint on the total number of particles gives rise to correlations between the occupation probability and the corresponding occupation numbers of the available energy levels. As a result, the canonical treatment of non-interacting fermions and bosons is in general avoided, and it is often completely absent in introductory texts addressing quantum statistical physics Kittel (1980); Landau and Lifshitz (1980); Pathria and Beale (2011). The standard approach is to relax the fixed $N$ constraint and instead describe the system using the grand canonical ensemble. While in most cases this approximation can be trusted in the large$-N$ limit, it can explicitly fail in the low-temperature regime Bedingham (2003); Mullin and Fernández (2003). The situation is even worse if the low-temperature system under study is mesoscopic, containing a relatively small number of particles. Recent advances in ultra-cold atom experiments Wenz et al. (2013); Parsons et al. (2015); Cheuk et al. (2015); Haller et al. (2015); Pegahan et al. (2019); Mukherjee et al. (2017); Hueck et al. (2018); Mukherjee et al. (2019); Onofrio (2016); Phelps et al. (2020) are making such conditions the rule rather than the exception, where the interactions between particles are often ignored, especially in the fermionic case Giorgini et al. (2008). Thus an accurate statistical representation of these systems should be canonical enforcing the existence of fixed $N$. In a different context, particle number conservation can reduce the amount of quantum information (entanglement) that can be extracted from a quantum state Horodecki et al. (2000); Bartlett and Wiseman (2003); Wiseman and Vaccaro (2003); Wiseman et al. (World Scientific, Singapore,
2004); Vaccaro et al. (2003); Schuch et al. (2004); Dunningham et al. (2005); Cramer et al. (2011); Klich and Levitov (2008) and thus the fixed $N$ constraint requires a canonical treatment. This is reflected in the calculation of the symmetry resolved entanglement which has recently been studied in a variety of physical systems Murciano et al. (2020a); Tan and Ryu (2020); Murciano et al. (2020b); Capizzi et al. (2020); Fraenkel and Goldstein (2020); Feldman and Goldstein (2019); Barghathi et al. (2019); Bonsignori et al. (2019); Barghathi et al. (2018); Kiefer-Emmanouilidis et al. (2020); Murciano et al. (2020c); Goldstein and Sela (2018).
Given its pedagogical and now practical importance, as well as the long history of the problem of non-interacting quantum particles, the last 50 years has provided a host of results, varying from general recursive relations that govern the canonical partition functions and the corresponding occupation numbers Schmidt (1989); Borrmann and Franke (1993); Borrmann et al. (1999); Pratt (2000); Mullin and Fernández (2003); Weiss and Wilkens (1997); Arnaud et al. (1999); Schönhammer (2017); Giraud et al. (2018); Tsutsui and Kita (2016); Zhou and Dai (2018), to approximate Denton et al. (1971, 1973); Schönhammer and Meden (1996); Bedingham (2003); Svidzinsky et al. (2018); Jha and Hirata (2020) and exact results for some special cases Chase et al. (1999); Schönhammer (2000); Kocharovsky and Kocharovsky (2010); Wang and Ma (2009); Akkelin and Sinyukov (2016); Magnus et al. (2017); Grabsch et al. (2018); Kruk et al. (2020). More recently, for the case of non-degenerate energy spectra, the exact decomposition of higher-order occupation number correlations in terms of the occupation numbers of individual bosonic and fermionic energy levels have been reported Schönhammer (2017); Giraud et al. (2018).
In this paper, we present a unified framework for the calculation of physical observables in the canonical ensemble for $N$ fermions or bosons that are applicable to general non-interacting Hamiltonians that may contain degenerate energy levels. We present an analysis of the mathematical structure of bosonic and fermionic canonical partition functions in the non-interacting limit that leads to a set of recursion relations for exactly calculating the energy level occupation probabilities and average occupation numbers. Using argument from linear algebra, we show how higher-order correlations between the occupation numbers can be factorized, allowing them to be obtained from the knowledge of the occupation numbers of the corresponding energy levels, a canonical generalization of Wick’s theorem Wick (1950). The key observation yielding simplification of calculations in the canonical ensemble is that occupation probabilities and occupation numbers can be expressed via auxiliary partition functions (APFs) – canonical fermionic or bosonic partition functions that correspond to a set of energy levels that are obtained from the full spectrum of the targeted system by making a subset of the energy levels degenerate (increasing its degeneracy if its already degenerate) or alternatively by excluding it from the spectrum. Results obtained via auxiliary partition functions are validated by demonstrating that a number of previously reported formulas can be naturally recovered in a straightforward fashion within this framework.
In a non-interacting system, observables such as the average energy and magnetization can be obtained solely from the knowledge of average occupation numbers, however, the calculation of the corresponding statistical fluctuations in such quantities, i.e., specific heat and magnetic susceptibility, requires knowledge of the fluctuations in occupation numbers and (in the canonical ensemble) correlations between them. Therefore, the factorization of correlations between occupation numbers in terms of the average occupation numbers of individual energy levels provides a simplified approach to calculate quantities such as specific heat and magnetic susceptibility of a given system. Here, we highlight a method to compute such correlations by considering a degenerate system of a finite periodic chain of non-interacting bosons that is influenced by an external uniform magnetic field.
The auxiliary partition function approach presented here provides a set of tools that can be used to analyze experimental data in low-density atomic gases where the number of particles is fixed. The application of the physically relevant canonical ensemble can eliminate errors introduced via the grand canonical approximation (especially in the inferred temperature) and lead to a more accurate interpretation of experimental results, including an improved diagnosis of the role of weak interaction effects.
It is hoped that the relative simplicity of the mathematical approach presented in this paper may encourage the inclusion of the interesting topic of the canonical treatment of Fermi and Bose gases in college-level textbooks.
In Sec.II we present the general formalism of the theory, where we introduce APFs and write occupation probability distributions, occupation numbers, and their correlations in terms of the APFs. In the same section, we also show how some of the previously known results could be recovered in a straightforward fashion. In Sec.III, considering fermionic and bosonic systems, we derive the decomposition of level higher-order occupation numbers correlations into individual energy levels occupation numbers for non-degenerate and degenerate energy spectra, alike. To illustrate the applicability of the theory, In Sec.IV, we consider a bosonic ring in the presence of a magnetic field. We conclude in Sec.V.
II Non-interacting Indistinguishable particles in the canonical ensemble
As the thermodynamic properties of a system of non-interacting particles are
governed by the single-particle spectrum and the underlying particle
statistics, we begin by considering the general one-particle spectrum $\epsilon_{i}$, with
$i\in\mathcal{S}=\{1,2,,\dots,M\}$. For an unbounded spectrum $M\to\infty$. In the canonical ensemble defined by fixing the total particle number $N$, the canonical partition function $Z_{N}\equiv Z_{N}(\mathcal{S})$ for $N$ indistinguishable particles is defined by:
$$Z_{N}=\sum_{\boldsymbol{n}|_{{{N}}}}X\quantity(\boldsymbol{n}|_{{{N}}}),$$
(1)
where
$$X(\boldsymbol{n}|_{N})=\prod_{i\in\mathcal{S}}\mathrm{e}^{-\beta\epsilon_{i}n_%
{i}}$$
(2)
are the Boltzmann factors at inverse temperature $\beta=1/k_{\rm B}T$ and the components $\{n_{i}\}$ of the vector $\boldsymbol{n}|_{N}=(n_{1},\dots,n_{M})|_{N}$ are the occupation numbers for the corresponding energy levels satisfying $\sum_{i\in\mathcal{S}}n_{i}=N$. The summation in Eq. (1) runs over all the possible occupation vectors $\boldsymbol{n}|_{N}$ which, in addition to conserving the total number of particles $N$, obeys occupation limits for each of the energy levels $\epsilon_{i}$: $n_{i}\leq n_{i}^{\max}$, thus $N\leq N_{\max}=\sum_{i}n_{i}^{\max}$.
Consider the spectrum defined by $\mathcal{S}$ as the union of disjoint subsets
(subspectra) $\mathcal{S}^{(1)}$ and $\mathcal{S}^{(2)}=\mathcal{S}\setminus S^{(1)}$, i.e., $\mathcal{S}=\mathcal{S}^{(1)}\cup\mathcal{S}^{(2)}$. Under
this decomposition the Boltzmann factors in $\mathcal{S}$ can be factorized as:
$X\quantity(\boldsymbol{n}|_{N})=X\quantity(\boldsymbol{n}^{(1)}|_{k})X%
\quantity(\boldsymbol{n}^{(2)}|_{N-k})$,
where $\boldsymbol{n}^{(1)}|_{k}$ and $\boldsymbol{n}^{(2)}|_{N-k}$ represent
occupation vectors of $k=\sum_{i\in\mathcal{S}^{(1)}}n_{i}^{(1)}$ and $N-k$ particles in $\mathcal{S}^{(1)}$ and $\mathcal{S}^{(2)}$, respectively. Summing $X\quantity(\boldsymbol{n}|_{N})$ over all possible $\boldsymbol{n}^{(1)}|_{k}$ and $\boldsymbol{n}^{(2)}|_{N-k}$ gives:
$$\sum_{\boldsymbol{n}^{(1)}|_{k}}\sum_{\boldsymbol{n}^{(2)}|_{N-k}}X\quantity(%
\boldsymbol{n}|_{N})=Z_{k}\quantity(\mathcal{S}^{(1)})Z_{N-k}\quantity(%
\mathcal{S}^{(2)}),$$
(3)
where we have introduced the APFs:
$$\displaystyle Z_{k}\quantity(\mathcal{S}^{(1)})$$
$$\displaystyle=\sum_{\boldsymbol{n}^{(1)}|_{k}}X\quantity(\boldsymbol{n}^{(1)}|%
_{k})$$
(4)
$$\displaystyle Z_{N-k}\quantity(\mathcal{S}^{(2)})$$
$$\displaystyle=\sum_{\boldsymbol{n}^{(2)}|_{N-k}}X\quantity(\boldsymbol{n}^{(2)%
}|_{N-k})\,.$$
(5)
For Eqs. (4) and (5) to satisfy the restrictions imposed by the per-energy level maximum occupancies $\{n_{i}^{\max}\}$ and the fixed $N$, $k$ must satisfy $\max(0,N-N^{\max}_{2})\leq k\leq\min(N,N^{\max}_{1})$, where $N^{\max}_{1}$ and $N^{\max}_{2}$ are the maximum numbers of particles allowed in $\mathcal{S}^{(1)}$ and $\mathcal{S}^{(2)}$, respectively. Additionally, any vector
$\boldsymbol{n}|_{N}$, under these constraints can be decomposed into two allowed vectors $\boldsymbol{n}^{(1)}|_{k}$ and $\boldsymbol{n}^{(2)}|_{N-k}$ and vice versa. Thus the sum in Eq. (1) can be similarly decomposed as: $\sum_{\boldsymbol{n}|_{N}}\equiv\sum_{k=k_{\min}}^{k_{\max}}\sum_{\boldsymbol{%
n}^{(1)}|_{k}}\sum_{\boldsymbol{n}^{(2)}|_{N-k}}$ where $k_{\min}=\max(0,N-N^{(2)}_{\max})$ and $k_{\max}=\min(N,N^{(1)}_{\max})$ yielding the full partition function
$$Z_{N}\equiv\sum_{k=k_{\min}}^{k_{\max}}Z_{k}\quantity(\mathcal{S}^{(1)})Z_{N-k%
}\quantity(\mathcal{S}^{(2)})\,.$$
(6)
Employing the convention that $Z_{N}(\mathcal{S})=0$ whenever $N$ is negative, or when it exceeds the maximum number of particles set by $\mathcal{S}$, the limits in the above summation can be simplified to $\sum_{k=0}^{N}$. The above notation can be made more explicit by specifying the subset of levels that are not included in the partition function
$$Z_{N}^{\setminus\mathcal{S}^{(1)}}\equiv Z_{N}\quantity(\mathcal{S}\setminus%
\mathcal{S}^{(1)})$$
(7)
and thus for $\mathcal{S}^{(1)}=\{j_{1},j_{2},\dots,j_{\ell}\}$ containing $\ell$ levels, Eq. (6) is equivalent to
$$Z_{N}\equiv\sum_{k=0}^{N}Z_{k}(\{j_{1},j_{2},\dots,j_{\ell}\})Z_{N-k}^{%
\setminus\{j_{1},j_{2},\dots,j_{\ell}\}}.$$
(8)
To obtain a physical interpretation of Eq. (8), recall that in the canonical ensemble, the likelihood of the $N$-particle system being in a microstate defined by the occupation vector $\boldsymbol{n}|_{N}$ is given by the ratio $X(\boldsymbol{n}|_{N})/Z_{N}$. Accordingly, for the subset of energy levels with indices $\{j_{1},j_{2},\dots,j_{\ell}\}$, the joint probability distribution of the corresponding occupation numbers $\mathcal{P}_{n_{j_{1}},n_{j_{2}},\dots,n_{j_{\ell}}}$ can be obtained by performing the summation $\sum_{\boldsymbol{n}^{(2)}|_{N-k}}$, where $k=\sum_{r=1}^{\ell}n_{j_{r}}$, yielding
$$\mathcal{P}_{n_{j_{1}},n_{j_{2}},\dots,n_{j_{\ell}}}=\frac{\mathrm{e}^{-\beta%
\sum_{r=1}^{\ell}\epsilon_{j_{r}}n_{j_{r}}}}{Z_{N}}Z_{N-k}^{\setminus\{j_{1},j%
_{2},\dots,j_{\ell}\}}.$$
(9)
It follows that the probability $P_{k}(\{j_{1},j_{2},\dots,j_{\ell}\})$ of finding $k$ particles in $\{j_{1},j_{2},\dots,j_{\ell}\}$ and, of course, $N-k$ particles in $\mathcal{S}\setminus\{j_{1},j_{2},\dots,j_{\ell}\}$, can be obtained by applying the summation $\sum_{\boldsymbol{n}^{(1)}|_{k}}$:
$$P_{k}(\{j_{1},j_{2},\dots,j_{\ell}\})=\frac{Z_{k}(\{j_{1},j_{2},\dots,j_{\ell}%
\})Z_{N-k}^{\setminus\{j_{1},j_{2},\dots,j_{\ell}\}}}{Z_{N}},$$
(10)
where the normalization of $P_{k}(\{j_{1},j_{2},\dots,j_{\ell}\})$ is guaranteed by
Eq.(8).
So far the analysis of the partition functions has been completely general and we have not specified what type of particles are being described. Now, let’s be more specific and consider a system that solely consists of either fermions or bosons.
II.1 An Inverted Analogy Between Fermionic and Bosonic Statistics
Having developed an intuition for the general structure of the canonical partition function under a bipartition into sub-spectra, we now observe how this can provide insights
into the relationship between fermionic ($n_{i}^{\max}=1$) and bosonic ($n_{i}^{\max}\to\infty$) statistics of $N$ non-interacting particles. To distinguish the two cases we introduce a new subscript on the partition function ($F$ for fermions and $B$ for bosons).
If the set $\mathcal{S}^{(1)}$ represent a single energy level with an index $j_{1}=j$ then using our convention we have:
$$Z_{\!{\scriptscriptstyle F}{\!,k}}(\{j\})=\begin{cases}\mathrm{e}^{-\beta%
\epsilon_{j}k}&0\leq k\leq 1\\
0&\text{otherwise}\end{cases}$$
(11)
for fermions and
$$Z_{\!{\scriptscriptstyle B}{\!,k}}(\{j\})=\begin{cases}\mathrm{e}^{-\beta%
\epsilon_{j}k}&k\geq 0\\
0&\text{otherwise}\end{cases}$$
(12)
for bosons. Substituting into Eq. (8) we immediately find:
$$\displaystyle Z_{\!{\scriptscriptstyle F}{\!,N}}$$
$$\displaystyle=Z_{\!{\scriptscriptstyle F}{\!,N}}^{\setminus\{{j}\}}+\mathrm{e}%
^{-\beta\epsilon_{j}}Z_{\!{\scriptscriptstyle F}{\!,N-1}}^{\setminus\{{j}\}}$$
(13)
$$\displaystyle Z_{\!{\scriptscriptstyle B}{\!,N}}$$
$$\displaystyle=\sum_{k=0}^{N}\mathrm{e}^{-\beta\epsilon_{j}k}Z_{\!{%
\scriptscriptstyle B}{\!,N-k}}^{\setminus\{{j}\}}\,.$$
(14)
The relations in Eq. (13) and (14) formally describe the procedure for generating the canonical partition function of the $N$ particle system after introducing an energy level $\epsilon_{j}$ to the preexisting spectrum $\mathcal{S}\setminus\{j\}$.
Examining the structure of Eq. (13) suggests a simple matrix form: $\boldsymbol{Z}=\mathsf{A}\boldsymbol{Z}^{\setminus\{j\}}$, where $\boldsymbol{Z}=(Z_{\!{\scriptscriptstyle F}{\!,0}},Z_{\!{\scriptscriptstyle F}%
{\!,1}},\dots)$, $\boldsymbol{Z}^{\setminus\{j\}}=(Z_{\!{\scriptscriptstyle F}{\!,0}}^{\setminus%
\{{j}\}},Z_{\!{\scriptscriptstyle F}{\!,1}}^{\setminus\{{j}\}},\dots)$ and the matrix $\mathsf{A}_{n,m}=\delta_{n,m}+\mathrm{e}^{-\beta\epsilon_{j}}\delta_{n,m+1}$ is bidiagonal and can be inverted such that:
$$Z_{\!{\scriptscriptstyle F}{\!,N}}^{\setminus\{{j}\}}=\sum_{k=0}^{N}(-1)^{k}%
\mathrm{e}^{-\beta\epsilon_{j}k}Z_{\!{\scriptscriptstyle F}{\!,N-k}}.$$
(15)
Comparing this expression with Eq. (14), we observe an identical structure apart from exchanging the factor $\mathrm{e}^{-\beta\epsilon_{j}}$ with $\quantity(-\mathrm{e}^{-\beta\epsilon_{j}})$. Thus we can obtain the inversion of Eq. (14) by replacing $\mathrm{e}^{-\beta\epsilon_{j}}$ with $\quantity(-\mathrm{e}^{-\beta\epsilon_{j}})$ in Eq. (13), i.e.,
$$Z_{\!{\scriptscriptstyle B}{\!,N}}^{\setminus\{{j}\}}=Z_{\!{\scriptscriptstyle
B%
}{\!,N}}-\mathrm{e}^{-\beta\epsilon_{j}}Z_{\!{\scriptscriptstyle B}{\!,N-1}}.$$
(16)
The relations Eq. (15) and Eq. (16) exemplify the elimination of an energy level as they represent the inverse of Eq. (13) and Eq. (14) respectively.
If we absorb the negative signs in Eqs. (15) by shifting the energy $\epsilon_{j}$ by $\pm i\pi/\beta$, we can write $Z_{\!{\scriptscriptstyle F}{\!,N}}^{\setminus\{{j}\}}=\sum_{k=0}^{N}\mathrm{e}%
^{-\beta\epsilon^{\prime}_{j}k}Z_{\!{\scriptscriptstyle F}{\!,N-k}}$, where $\epsilon^{\prime}_{j}=\epsilon_{j}\pm i\pi/\beta$. As the general bipartition into sub-spectra introduced in Eq. (6) holds for energy levels with mixed statistics, and doesn’t require real entries for the $\epsilon_{i}$, we can build the bosonic partition functions $Z_{\!{\scriptscriptstyle B}{\!,N}}^{\prime}(\{j_{1},j_{2},\dots,j_{\ell}\})$ using the shifted energies $\{\epsilon_{j_{1}}^{\prime},\epsilon_{j_{2}}^{\prime},\dots,\epsilon_{j_{\ell}%
}^{\prime}\}$ and then combine it with the $Z_{\!{\scriptscriptstyle F}{\!,N}}$ to generate a mixed-form:
$$Z_{\!{\scriptscriptstyle F}{\!,N}}^{\setminus\{{j_{1},j_{2},\dots,j_{\ell}}\}}%
=\sum_{k=0}^{N}Z_{\!{\scriptscriptstyle B}{\!,k}}^{\prime}(\{j_{1},j_{2},\dots%
,j_{\ell}\})Z_{\!{\scriptscriptstyle F}{\!,N-k}}\,.$$
(17)
The effect of shifting the single particle spectrum by a constant $\omega$ on the canonical partition function $Z_{N}$ is captured by a rescaling factor $\mathrm{e}^{-\beta\omega N}$ and the resulting $N$ particle partition function of the shifted spectrum is $\mathrm{e}^{-\beta\omega N}Z_{N}$. Using $Z_{\!{\scriptscriptstyle B}{\!,k}}^{\prime}(\{j_{1},j_{2},\dots,j_{\ell}\})={%
\rm{e}}^{\pm i\pi k}Z_{\!{\scriptscriptstyle B}{\!,k}}(\{j_{1},j_{2},\dots,j_{%
\ell}\})$ then yields
$$Z_{\!{\scriptscriptstyle F}{\!,N}}^{\setminus\{{j_{1},j_{2},\dots,j_{\ell}}\}}%
=\sum_{k=0}^{N}(-1)^{k}Z_{\!{\scriptscriptstyle B}{\!,k}}(\{j_{1},j_{2},\dots,%
j_{\ell}\})Z_{\!{\scriptscriptstyle F}{\!,N-k}}.$$
(18)
Starting from Eq. (16) and following the same argument, we obtain an equivalent expression for bosons
$$Z_{\!{\scriptscriptstyle B}{\!,N}}^{\setminus\{{j_{1},j_{2},\dots,j_{\ell}}\}}%
=\sum_{k=0}^{N}(-1)^{k}Z_{\!{\scriptscriptstyle F}{\!,k}}(\{j_{1},j_{2},\dots,%
j_{\ell}\})Z_{\!{\scriptscriptstyle B}{\!,N-k}}.$$
(19)
The last two equations can be seen as a generalization of Eqs. (15) and (16). However, they also recover the symmetry between fermionic and bosonic statistics. To this end, we show how to obtain the partition function of a given spectrum using the APFs of two complementary subsets of the spectrum through Eq. (8). Also, Eqs. (18) and (19) show that this relation can be inverted, i.e., we can calculate the APFs of a subset of energy level using the APFs of its complement with the opposite statistics and the partition functions of the full spectrum.
II.2 Energy Level Occupations and Correlations
For fermions, the Pauli exclusion principle restricts the number of particles occupying an energy level $\epsilon_{j}$ to $n_{j}=0$ or $1$. Equivalently, $n_{j}^{p}=n_{j}$ for any $p>0$, simplifying the calculation of energy level occupation numbers and the correlations between them, including higher moments. More specifically, the average $\expectationvalue{n_{j_{1}}^{p_{1}}n_{j_{2}}^{p_{2}}\dots n_{j_{\ell}}^{p_{%
\ell}}}_{\!{\scriptscriptstyle F}{\!,N}}=\expectationvalue{n_{j_{1}}n_{j_{2}}%
\dots n_{j_{\ell}}}_{\!{\scriptscriptstyle F}{\!,N}}$, where $p_{i}>0,\ i\in\{1,\dots,\ell\}$ and in general
$$\displaystyle\expectationvalue{n_{j_{1}}n_{j_{2}}\dots n_{j_{\ell}}}_{\!{%
\scriptscriptstyle F}{\!,N}}=\\
\displaystyle\sum_{\boldsymbol{n}(\{j_{1},j_{2},\dots,j_{\ell}\})}n_{j_{1}}n_{%
j_{2}}\dots n_{j_{\ell}}\mathcal{P}_{\!\!{\scriptscriptstyle F}{\!,n_{j_{1}},n%
_{j_{2}},\dots,n_{j_{\ell}}}}$$
(20)
using the probability:
$$\mathcal{P}_{\!\!{\scriptscriptstyle F}{\!,n_{j_{1}},n_{j_{2}},\dots,n_{j_{%
\ell}}}}=\frac{\mathrm{e}^{-\beta\sum_{r=1}^{\ell}\epsilon_{j_{r}}n_{j_{r}}}}{%
Z_{\!{\scriptscriptstyle F}{\!,N}}}Z_{\!{\scriptscriptstyle F}{\!,N-\sum_{r=1}%
^{\ell}n_{j_{r}}}}^{\setminus\{{j_{1},j_{2},\dots,j_{\ell}}\}},$$
(21)
defined in Eq. (9). The only term that survives in Eq. (20) has $n_{j_{1}}=n_{j_{2}}=\dots=n_{j_{\ell}}=1$ giving:
$$\expectationvalue{n_{j_{1}}n_{j_{2}}\dots n_{j_{\ell}}}_{\!{\scriptscriptstyle
F%
}{\!,N}}=\frac{\mathrm{e}^{-\beta\sum_{r=1}^{\ell}\epsilon_{j_{r}}}}{Z_{\!{%
\scriptscriptstyle F}{\!,N}}}Z_{\!{\scriptscriptstyle F}{\!,N-\ell}}^{%
\setminus\{{j_{1},j_{2},\dots,j_{\ell}}\}}.$$
(22)
Following the same procedure, Eq. (20) can be generalized to describe both correlations and anti-correlations between energy levels:
$$\displaystyle\expectationvalue{\prod_{r=1}^{\ell}{\quantity[n_{j_{r}}\gamma_{j%
_{r}}+(1-n_{j_{r}})(1-\gamma_{j_{r}})]}}_{\!{\scriptscriptstyle F}{\!,N}}=\\
\displaystyle\frac{1}{Z_{\!{\scriptscriptstyle F}{\!,N}}}\mathrm{e}^{-\beta%
\sum_{r=1}^{\ell}\epsilon_{j_{r}}\gamma_{j_{r}}}Z_{\!{\scriptscriptstyle F}{\!%
,N-\sum_{r=1}^{\ell}\gamma_{j_{r}}}}^{\setminus\{{j_{1},j_{2},\dots,j_{\ell}}%
\}},$$
(23)
where $\gamma_{j_{r}}=1,0$. For the latter with $\gamma_{j_{r}}=0$, the occupation numbers in Eq. (20) have been replaced with their complements, $1-n_{j_{r}}$.
Thus, we find that fermionic level occupations and correlations can be directly written in terms of APFs without resorting to the usual definition $\expectationvalue{n_{j_{1}}n_{j_{2}}\dots n_{j_{\ell}}}_{\!{\scriptscriptstyle
F%
}{\!,N}}=\frac{1}{Z_{\!{\scriptscriptstyle F}{\!,N}}}\frac{\partial^{\ell}Z_{%
\!{\scriptscriptstyle F}{\!,N}}}{\partial(-\beta\epsilon_{j_{1}})\dots\partial%
(-\beta\epsilon_{j_{\ell}})}$ which yields equivalent results.
When considering a single level $(\ell=1)$, the occupation probability of the $j^{\rm th}$ fermionic level immediately follows
$$\expectationvalue{n_{j}}_{\!{\scriptscriptstyle F}{\!,N}}=\frac{\mathrm{e}^{-%
\beta\epsilon_{j}}Z_{\!{\scriptscriptstyle F}{\!,N-1}}^{\setminus\{{j}\}}}{Z_{%
\!{\scriptscriptstyle F}{\!,N}}},$$
(24)
with the associated probability
$$\mathcal{P}_{\!\!{\scriptscriptstyle F}{\!,n_{j}}}=\frac{\mathrm{e}^{-\beta%
\epsilon_{j}n_{j}}}{Z_{\!{\scriptscriptstyle F}{\!,N}}}Z_{\!{%
\scriptscriptstyle F}{\!,N-n_{j}}}^{\setminus\{{j}\}}.$$
(25)
For bosons, the occupation numbers $\expectationvalue{n_{j}}_{\!{\scriptscriptstyle B}{\!,N}}$ can be calculated from the corresponding occupation probability distribution $\mathcal{P}_{n_{j}}$ which are obtained from Eq. (9):
$$\mathcal{P}_{\!\!{\scriptscriptstyle B}{\!,n_{j}}}=\frac{\mathrm{e}^{-\beta%
\epsilon_{j}n_{j}}}{Z_{\!{\scriptscriptstyle B}{\!,N}}}Z_{\!{%
\scriptscriptstyle B}{\!,N-n_{j}}}^{\setminus\{{j}\}},$$
(26)
and for the $\ell$-point correlations:
$$\mathcal{P}_{\!\!{\scriptscriptstyle B}{\!,n_{j_{1}},n_{j_{2}},\dots,n_{j_{%
\ell}}}}=\frac{\mathrm{e}^{-\beta\sum_{r=1}^{\ell}\epsilon_{j_{r}}n_{j_{r}}}}{%
Z_{\!{\scriptscriptstyle B}{\!,N}}}Z_{\!{\scriptscriptstyle B}{\!,N-\sum_{r=1}%
^{\ell}n_{j_{r}}}}^{\setminus\{{j_{1},j_{2},\dots,j_{\ell}}\}}.$$
(27)
However, unlike the fermionic case, such an approach requires performing the unrestricted summation $\sum_{\boldsymbol{n}(\{j_{1},j_{2},\dots,j_{\ell}\})}$.
An alternative method which avoids this difficulty can be developed by exploiting the inverted analogy between fermionic and bosonic statistics introduced in Sec.II.1. In the fermionic case, the occupation number of an energy level $\epsilon_{j}$ is proportional to the APF $Z_{\!{\scriptscriptstyle F}{\!,N-1}}^{\setminus\{{j}\}}$ (Eq. (24)) which corresponds to the actual spectrum of the system missing the energy level $\epsilon_{j}$.
This suggests a route forward for bosons via the analogous inversion of doubly including the energy level $\epsilon_{j}$ instead of removing it, i.e., we construct an APF where this level is twofold-degenerate. We denote the corresponding $N$-boson APF by $Z_{\!{\scriptscriptstyle B}{\!,N}}^{\cup\quantity{{j}}}$ and distinguish the two levels using the dressed indices $j^{\scriptscriptstyle(0)}$ and $j^{\scriptscriptstyle(1)}$ such that the resulting combined spectrum has level indices $\quantity{1,\dots j-1,j^{\scriptscriptstyle(0)},j^{\scriptscriptstyle(1)},j+1,%
\dots,M}$ where $\epsilon_{j}=\epsilon_{j^{\scriptscriptstyle(0)}}=\epsilon_{j^{%
\scriptscriptstyle(1)}}$.
Returning to the general definition of the canonical partition function in Eq. (1), we can write $Z_{\!{\scriptscriptstyle B}{\!,N}}^{\cup\quantity{{j}}}=\sum_{\boldsymbol{n}^{%
\prime}|_{N}}X(\boldsymbol{n}^{\prime}|_{N})$, where the occupation vectors $\boldsymbol{n}^{\prime}|_{N}$ have one extra component: $n_{j}$ in $\boldsymbol{n}|_{N}$ is replaced by $n_{j^{\scriptscriptstyle(0)}}$ and $n_{j^{\scriptscriptstyle(1)}}$. The modified Boltzmann factors are:
$$X(\boldsymbol{n}^{\prime}|_{N})=\mathrm{e}^{-\beta\epsilon_{j}[n_{j^{%
\scriptscriptstyle(0)}}+n_{j^{\scriptscriptstyle(1)}}]}\prod_{i\neq j^{%
\scriptscriptstyle(0)},j^{\scriptscriptstyle(1)}}\mathrm{e}^{-\beta\epsilon_{i%
}n_{i}}$$
(28)
and thus their value is dependent only on the total occupancy of the $j^{\rm th}$ level, $n_{j^{\scriptscriptstyle(0)}}+n_{j^{\scriptscriptstyle(1)}}$. As a result, $X(\boldsymbol{n}|_{N})=X(\boldsymbol{n}^{\prime}|_{N})$ for any occupation vectors $\boldsymbol{n}|_{N}$ and $\boldsymbol{n}^{\prime}|_{N}$ with all $i\neq j$ components equal as well as $[\boldsymbol{n}|_{N}]_{j}=n_{j}=n_{j^{\scriptscriptstyle(0)}}+n_{j^{%
\scriptscriptstyle(1)}}$. For fixed $\boldsymbol{n}|_{N}$, the number of vectors $\boldsymbol{n}^{\prime}|_{N}$ that satisfies the previous conditions is equal to $n_{j}+1$, or, the number of ways in which $n_{j}$ bosons can occupy two energy levels. The APF can then be written in terms of the original occupation vector $\boldsymbol{n}|_{N}$ by inserting a frequency factor to account for the extra level $j$:
$$\displaystyle Z_{\!{\scriptscriptstyle B}{\!,N}}^{\cup\quantity{{j}}}$$
$$\displaystyle=\sum_{\boldsymbol{n}|_{N}}(n_{j}+1)X(\boldsymbol{n}|_{N})$$
(29)
$$\displaystyle=Z_{\!{\scriptscriptstyle B}{\!,N}}\expectationvalue{n_{j}+1}_{\!%
{\scriptscriptstyle B}{\!,N}}$$
and thus we can write:
$$\expectationvalue{n_{j}}_{\!{\scriptscriptstyle B}{\!,N}}=Z_{\!{%
\scriptscriptstyle B}{\!,N}}^{\cup\quantity{{j}}}/Z_{\!{\scriptscriptstyle B}{%
\!,N}}-1.$$
(30)
Applying Eq. (16), gives $Z_{\!{\scriptscriptstyle F}{\!,N}}^{\cup\quantity{{j}}}=Z_{\!{%
\scriptscriptstyle B}{\!,N}}+\mathrm{e}^{-\beta\epsilon_{j}}Z_{\!{%
\scriptscriptstyle F}{\!,N-1}}^{\cup\quantity{{j}}}$ which can be substituted into Eq. (30) to arrive at
$$\expectationvalue{n_{j}}_{\!{\scriptscriptstyle B}{\!,N}}=\frac{\mathrm{e}^{-%
\beta\epsilon_{j}}Z_{\!{\scriptscriptstyle B}{\!,N-1}}^{\cup\quantity{{j}}}}{Z%
_{\!{\scriptscriptstyle B}{\!,N}}},$$
(31)
which is in the same form as Eq. (24) for fermions.
To generalize this expression to $\ell$-level correlations with $\ell>1$ we examine the numerator of Eq. (31), recalling that we have added an extra copy of energy level $j$ to the partition function for $N-1$ particles such that it now appears $m_{j}+1=2$ times in the associated spectrum:
$$\displaystyle\mathrm{e}^{-\beta\epsilon_{j}}Z_{\!{\scriptscriptstyle B}{\!,N-1%
}}^{\cup\quantity{{j}}}$$
$$\displaystyle=\mathrm{e}^{-\beta\epsilon_{j}}\sum_{\boldsymbol{n}^{\prime}|_{N%
-1}}X(\boldsymbol{n}^{\prime}|_{N-1})$$
$$\displaystyle=\!\!\sum_{\boldsymbol{n}|_{N-1}}\!\!\binom{\tilde{n}_{j}+(m_{j}+%
1)-1}{(m_{j}+1)-1}\mathrm{e}^{-\beta\epsilon_{j}}X(\boldsymbol{n}|_{N-1})$$
$$\displaystyle=\sum_{\boldsymbol{n}|_{N,n_{j}\geq 1}}n_{j}X(\boldsymbol{n}|_{N}%
)\,.$$
(32)
Here we obtain the second line using the same trick as in Eq. (29) to convert $X(\boldsymbol{n}^{\prime}|_{N-1})$ into $X(\boldsymbol{n}|_{N-1})$ by accounting for the degeneracy where $[\boldsymbol{n}|_{N-1}]_{j}=\tilde{n}_{j}\geq 0$ is the total number of particles occupying the level $j$ as it appears in the Boltzmann factor $X(\boldsymbol{n}|_{N-1})=\mathrm{e}^{-\beta\epsilon_{j}\tilde{n}_{j}}\prod_{i%
\neq j}\mathrm{e}^{-\beta\epsilon_{i}n_{i}}$. The complicated looking binomial coefficient
$\binom{\tilde{n}_{j}+(m_{j}+1)-1}{(m_{j}+1)-1}=\left(\kern-3.0pt\left(\genfrac%
{}{}{0.0pt}{}{\tilde{n}_{j}}{m_{j}+1}\right)\kern-3.0pt\right)$ is the multiset coefficient that counts the number of ways $\tilde{n}_{j}$ bosons can be distributed amongst the $m_{j}+1$ levels with energy $\epsilon_{j}$. Finally, the last line is obtained by using the
fact that $\mathrm{e}^{-\beta\epsilon_{j}}X(\boldsymbol{n}|_{N-1})=X(\boldsymbol{n}|_{N})$ where $[\boldsymbol{n}|_{N}]_{j}=n_{j}=\tilde{n}_{j}+1\geq 1$.
Eq. (32) can be immediately extended to the case where we add copies of not 1 but $\ell$ levels $\quantity{j_{1},\dots,j_{\ell}}$:
$$\mathrm{e}^{-\beta\sum_{r=1}^{\ell}\epsilon_{j_{r}}}Z_{\!{\scriptscriptstyle B%
}{\!,N-\ell}}^{\cup\quantity{{j_{1},\dots,j_{\ell}}}}=\sum_{\boldsymbol{n}|_{N%
}}\quantity(\prod_{r=1}^{\ell}{n}_{j_{r}})X(\boldsymbol{n}|_{N})$$
(33)
where the conditions $n_{j_{r}}\geq 1$ in the occupancy vector can be neglected as any $n_{j_{r}}=0$ terms do not contribute to the sum due to the multiplicative string. Finally, we can write the desired result:
$$\expectationvalue{n_{j_{1}}n_{j_{2}}\dots n_{j_{\ell}}}_{\!{\scriptscriptstyle
B%
}{\!,N}}=\frac{\mathrm{e}^{-\beta\sum_{r=1}^{\ell}\epsilon_{j_{r}}}}{Z_{\!{%
\scriptscriptstyle B}{\!,N}}}Z_{\!{\scriptscriptstyle B}{\!,N-\ell}}^{\cup%
\quantity{{j_{1},j_{2},\dots,j_{\ell}}}}.$$
(34)
An immediate extension of Eq. (34) will turn out to be useful, which introduces an APF with higher-order degeneracy. We consider $m_{j_{r}}$ extra copies of the $r^{\rm th}$ level $\epsilon_{j_{r}}$ with $r\in\{1,\dots\ell\}$ and find:
$$\displaystyle\expectationvalue{\prod_{r=1}^{\ell}{\binom{n_{j_{r}}-q_{j_{r}}+m%
_{j_{r}}}{m_{j_{r}}}}}_{\!{\scriptscriptstyle B}{\!,N}}=\\
\displaystyle\frac{1}{Z_{\!{\scriptscriptstyle B}{\!,N}}}\mathrm{e}^{-\beta%
\sum_{r=1}^{\ell}\epsilon_{j_{r}}q_{j_{r}}}Z_{\!{\scriptscriptstyle B}{\!,N-%
\sum_{r=1}^{\ell}q_{j_{r}}}}^{\cup\quantity{{j_{1}^{\scriptscriptstyle(1)},%
\dots,j_{1}^{\scriptscriptstyle(m_{j_{1}})},\dots j_{\ell}^{\scriptscriptstyle%
(1)},\dots,j_{\ell}^{\scriptscriptstyle(m_{j_{\ell}})}}}}$$
(35)
where the $q_{j_{r}}\leq m_{j_{r}}$ allow for the added freedom of choice of how many particles are associated with each of the degenerate levels and allow us to write the left-hand side in terms of the desired occupations $n_{j}=[\boldsymbol{n}|_{N}]_{j}$. Note: to simplify notation we only include a superscript on the levels in the bosonic APF and only if we are adding more than one extra copy per original level.
Eqs. (23) and (35) are the major results of this section, and demonstrate that for both fermions and bosons, $\ell$-level correlations can be written in terms of APFs for $N-\ell$ particles with $\ell$ energy levels removed (added) for fermions (bosons).
II.3 Recovering Known Results Via the APF Theory
In this section, we illustrate the utility of our auxiliary expressions in simplifying the derivation of known recursion relations that govern the canonical partition functions and occupation numbers for fermions and bosons.
Beginning with fermionic statistics, if we use the definition of $\expectationvalue{n_{j}}_{\!{\scriptscriptstyle F}{\!,N}}$ in Eq. (24) to substitute for $Z_{\!{\scriptscriptstyle F}{\!,N}}^{\setminus\{{j}\}}$ and $Z_{\!{\scriptscriptstyle F}{\!,N-1}}^{\setminus\{{j}\}}$ in Eq. (13),
we obtain the well-known recursion relation for occupation numbers Schmidt (1989); Borrmann et al. (1999)
$$\expectationvalue{n_{j}}_{\!{\scriptscriptstyle F}{\!,N+1}}=\frac{Z_{\!{%
\scriptscriptstyle F}{\!,N}}}{Z_{\!{\scriptscriptstyle F}{\!,N+1}}}\mathrm{e}^%
{-\beta\epsilon_{j}}\left(1-{\expectationvalue{n_{j}}_{\!{\scriptscriptstyle F%
}{\!,N}}}\right)\,.$$
(36)
$\expectationvalue{n_{j}}_{\!{\scriptscriptstyle F}{\!,N}}$ can be written explicitly in term of the partition functions by substituting for $Z_{\!{\scriptscriptstyle F}{\!,N-1}}^{\setminus\{{j}\}}$ in Eq. (24) using Eq. (15) as:
$$\expectationvalue{n_{j}}_{\!{\scriptscriptstyle F}{\!,N}}=\frac{1}{Z_{\!{%
\scriptscriptstyle F}{\!,N}}}\sum_{k=1}^{N}(-1)^{k-1}\mathrm{e}^{-\beta%
\epsilon_{j}k}Z_{\!{\scriptscriptstyle F}{\!,N-k}},$$
(37)
and using the canonical condition $\sum_{i}\expectationvalue{n_{i}}_{\!{\scriptscriptstyle F}{\!,N}}=N$ we find the original 1993 result of Borrmann and Franke Borrmann and Franke (1993):
$$Z_{\!{\scriptscriptstyle F}{\!,N}}=\frac{1}{N}\sum_{k=1}^{N}(-1)^{k-1}C_{k}Z_{%
\!{\scriptscriptstyle F}{\!,N-k}},$$
(38)
where $C_{k}=\sum_{j}\mathrm{e}^{-\beta\epsilon_{j}k}$.
Bosonic statistics can be treated in analogy to the fermionic case. Assigning the roles played by Eqs. (24), (13) and (15) to Eqs. (31), (16) and (14)111Here we replace $Z_{\!{\scriptscriptstyle B}{\!,x}}$ with $Z_{\!{\scriptscriptstyle B}{\!,x}}^{\cup\quantity{{j}}}$ and $Z_{\!{\scriptscriptstyle B}{\!,x}}^{\setminus\{{j}\}}$ with $Z_{\!{\scriptscriptstyle B}{\!,x}}$ in (16) and (14), respectively, we obtain the known bosonic equivalents Schmidt (1989); Borrmann et al. (1999); Borrmann and Franke (1993):
$$\displaystyle\expectationvalue{n_{j}}_{\!{\scriptscriptstyle B}{\!,N+1}}$$
$$\displaystyle=\frac{Z_{\!{\scriptscriptstyle B}{\!,N}}}{Z_{\!{%
\scriptscriptstyle B}{\!,N+1}}}\mathrm{e}^{-\beta\epsilon_{j}}\quantity(1+%
\expectationvalue{n_{j}}_{\!{\scriptscriptstyle B}{\!,N}})$$
(39)
$$\displaystyle\expectationvalue{n_{j}}_{\!{\scriptscriptstyle B}{\!,N}}$$
$$\displaystyle=\frac{1}{Z_{\!{\scriptscriptstyle B}{\!,N}}}\sum_{k=1}^{N}%
\mathrm{e}^{-\beta\epsilon_{j}k}Z_{\!{\scriptscriptstyle B}{\!,N-k}}$$
(40)
$$\displaystyle Z_{\!{\scriptscriptstyle B}{\!,N}}$$
$$\displaystyle=\frac{1}{N}\sum_{k=1}^{N}C_{k}Z_{\!{\scriptscriptstyle B}{\!,N-k%
}}\,.$$
(41)
Due to the Pauli exclusion principle for fermionic statistics, the occupation number $\expectationvalue{n_{j}}_{\!{\scriptscriptstyle F}{\!,N}}$ of an energy level $j$ gives us direct access to the occupation probability distribution $\mathcal{P}_{\!\!{\scriptscriptstyle F}{\!,n_{j}}}$ of the level. Despite the absence of any such simplification in the bosonic case, the occupation probability distribution $\mathcal{P}_{\!\!{\scriptscriptstyle B}{\!,n_{j}}}$ of a single energy level can be related to the corresponding partition functions as
$$\mathcal{P}_{\!\!{\scriptscriptstyle B}{\!,n_{j}}}=\mathrm{e}^{-\beta\epsilon_%
{j}n_{j}}\frac{Z_{\!{\scriptscriptstyle B}{\!,N-n_{j}}}}{Z_{\!{%
\scriptscriptstyle B}{\!,N}}}-\mathrm{e}^{-\beta\epsilon_{j}(n_{j}+1)}\frac{Z_%
{\!{\scriptscriptstyle B}{\!,N-n_{j}-1}}}{Z_{\!{\scriptscriptstyle B}{\!,N}}},$$
(42)
which is obtained by using Eq. (26) to substitute for $Z_{\!{\scriptscriptstyle B}{\!,N-n_{j}}}^{\setminus\{{j}\}}$ in Eq. (16), after replacing $N$ with $N-n_{j}$ Weiss and Wilkens (1997). In summary, within the unified framework of APFs, it is straightforward to obtain most of the well-known general relations in the fermionic and bosonic canonical ensemble that were previously derived using a host of different methods. This highlights the utility of this approach as a unifying framework when studying $N$ indistinguishable non-interacting particles.
II.4 General Expressions for Probabilities and Correlations
De facto, the APFs can also be used to generalize previous results, and in a form that is highly symmetric with respect to particle statistics. To accentuate this, let us introduce the notation:
$$\zeta=\begin{cases}+1&\Leftrightarrow B\Leftrightarrow\text{bosons}\\
-1&\Leftrightarrow F\Leftrightarrow\text{fermions}\end{cases}\,.$$
(43)
Then, the recursive relations for energy level correlations can be obtained using Eqs. (13) and (16) for fermions and bosons, respectively. The number of initial values of correlations needed is equal to the number of the involved points, e.g., for the two-level correlations we find:
$$\displaystyle\expectationvalue{n_{i}n_{j}}_{\!{\scriptscriptstyle\zeta}{\!,N+2}}$$
$$\displaystyle=\frac{Z_{\!{\scriptscriptstyle\zeta}{\!,N}}}{Z_{\!{%
\scriptscriptstyle\zeta}{\!,N+2}}}\mathrm{e}^{-\beta(\epsilon_{i}+\epsilon_{j}%
)}\quantity(1-\expectationvalue{n_{i}n_{j}}_{\!{\scriptscriptstyle\zeta}{\!,N}})$$
$$\displaystyle\;+\zeta\frac{Z_{\!{\scriptscriptstyle\zeta}{\!,N+1}}}{Z_{\!{%
\scriptscriptstyle\zeta}{\!,N+2}}}\quantity(\mathrm{e}^{-\beta\epsilon_{i}}+%
\mathrm{e}^{-\beta\epsilon_{j}})\expectationvalue{n_{i}n_{j}}_{\!{%
\scriptscriptstyle\zeta}{\!,N+1}}$$
(44)
Further, for the set of levels $\mathcal{S}_{\ell}=\{j_{1},\dots,j_{\ell}\}$, Eqs. (21) and (27), define the joint probability distributions of the occupation numbers in terms of the APFs $Z_{\!{\scriptscriptstyle F}{\!,k}}^{\setminus{\mathcal{S}_{\ell}}}$ and $Z_{\!{\scriptscriptstyle B}{\!,k}}^{\setminus{\mathcal{S}_{\ell}}}$, which can be re-expressed using Eq. (18) and Eq. (19) as
$$\mathcal{P}_{\!\!{\scriptscriptstyle\zeta}{\!,n_{j_{1}},\dots,n_{j_{\ell}}}}=%
\frac{\mathrm{e}^{-\beta E_{\rm{tot}}}}{Z_{\!{\scriptscriptstyle\zeta}{\!,N}}}%
\sum_{k=0}^{N-n_{\rm{tot}}}(-1)^{k}Z_{\!{\scriptscriptstyle-\zeta}{\!,k}}(%
\mathcal{S}_{\ell})Z_{\!{\scriptscriptstyle\zeta}{\!,N-n_{\rm{tot}}-k}}\\
$$
(45)
where, $E_{\rm{tot}}=\sum_{r=1}^{\ell}\epsilon_{j_{r}}n_{j_{r}}$ and $n_{\rm{tot}}=\sum_{r=1}^{\ell}n_{j_{r}}$. To avoid confusion we note again the convention in use that $Z_{\!{\scriptscriptstyle-\zeta}{\!,k}}(\mathcal{S_{\ell}})=0$ whenever $k$ exceeds the maximum number of particles set by $\mathcal{S_{\ell}}$ for fixed particle statistics.
Now, as the level occupation correlations of fermions and bosons are represented by the APFs $Z_{\!{\scriptscriptstyle F}{\!,k}}^{\setminus{\mathcal{S}_{\ell}}}$ and $Z_{\!{\scriptscriptstyle B}{\!,k}}^{\cup{\mathcal{S}_{\ell}}}$ in Eqs. (22) and (34), we can write
$$\expectationvalue{n_{j_{1}}\dots n_{j_{\ell}}}_{\!{\scriptscriptstyle\zeta}{\!%
,N}}=\frac{\mathrm{e}^{-\beta\sum_{r=1}^{\ell}\epsilon_{j_{r}}}}{Z_{\!{%
\scriptscriptstyle\zeta}{\!,N}}}\sum_{k=0}^{N-\ell}\zeta^{k}Z_{\!{%
\scriptscriptstyle B}{\!,k}}(\mathcal{S}_{\ell})Z_{\!{\scriptscriptstyle\zeta}%
{\!,N-\ell-k}}$$
(46)
where we note that the $k$-particle bosonic partition function for the levels $\mathcal{S}_{\ell}$ appears in both fermionic and bosonic correlations.
Alternatively, the APFs $Z_{\!{\scriptscriptstyle F}{\!,k}}^{\setminus{\mathcal{S}_{\ell}}}$, $Z_{\!{\scriptscriptstyle B}{\!,k}}^{\setminus{\mathcal{S}_{\ell}}}$ and $Z_{\!{\scriptscriptstyle B}{\!,k}}^{\cup{\mathcal{S}_{\ell}}}$ can be computed recursively using Eq. (38) and Eq. (41). This has the potential to simplify the calculation of the related full joint probability distribution, as it only requires calculating the corresponding APF with a number of particles in the range $k=0,\dots,N$.
II.4.1 Simplification for degenerate levels
The expressions derived in the previous section have a very simple form when the involved levels are degenerate, i.e., if we consider correlations between the set of levels $\{j^{(0)},\dots,j^{(\ell-1)}\}$, where $\epsilon_{j^{(s)}}=\epsilon_{j}$ for $s\in\{0,\dots,\ell-1\}$. The canonical partition function of $N$ bosons in $\ell$ degenerate energy levels is
$$Z_{\!{\scriptscriptstyle B}{\!,N}}(\{j^{(0)},\dots,j^{(\ell-1)}\})=\tbinom{N+%
\ell-1}{\ell-1}\mathrm{e}^{-\beta\epsilon_{j}N}\,,$$
(47)
and thus such correlations for fermions and bosons can be computed directly from Eq. (46) as
$$\expectationvalue{n_{j^{(0)}}\dots n_{j^{(\ell-1)}}}_{\!{\scriptscriptstyle%
\zeta}{\!,N}}=\frac{1}{Z_{\!{\scriptscriptstyle\zeta}{\!,N}}}\sum_{k=\ell}^{N}%
\zeta^{k-\ell}\tbinom{k-1}{\ell-1}\mathrm{e}^{-\beta\epsilon_{j}k}Z_{\!{%
\scriptscriptstyle\zeta}{\!,N-k}}$$
(48)
where we have shifted the summation. The numerical complexity of calculating Eq. (48) differs from that of Eq. (37) or Eq. (40) by the number of multiplications and additions needed to calculate the extra factor $\tbinom{k-1}{\ell-1}$, which can be viewed as a polynomial in $k$ of degree $\ell-1$. Thus calculating Eq. (48) requires additional $\approx(N-\ell)(\ell-1)$ multiplications, and a similar number of additions which still scales linearly with $N$ for a moderate value of $\ell$.
II.4.2 Expectation values of higher moments of degenerate levels
Let us now focus on the case of bosons and revisit Eq. (35) considering a single energy level $\epsilon_{j}$:
$$\expectationvalue{{\tbinom{n_{j}-q+m_{j}}{m_{j}}}}_{\!{\scriptscriptstyle B}{%
\!,N}}\!\!=\!\frac{1}{Z_{\!{\scriptscriptstyle B}{\!,N}}}\mathrm{e}^{-q\beta%
\epsilon_{j}}Z_{\!{\scriptscriptstyle B}{\!,N-q}}^{\cup\quantity{{j^{%
\scriptscriptstyle(1)},\dots,j^{\scriptscriptstyle(m_{j})}}}}.$$
(49)
Similarly, using Eq. (6), we can write the APF $Z_{\!{\scriptscriptstyle B}{\!,N-q}}^{\cup\quantity{{j^{\scriptscriptstyle(1)}%
,\dots,j^{\scriptscriptstyle(m_{j})}}}}$ in terms of the system partition function and the bosonic APF $Z_{\!{\scriptscriptstyle B}{\!,N-q}}(\{j^{\scriptscriptstyle(1)},\dots,j^{%
\scriptscriptstyle(m_{j})}\})$ of $m_{j}$ degenerate levels. As a result, we obtain
$$\expectationvalue{{\tbinom{n_{j}-q+m_{j}}{m_{j}}}}_{\!{\scriptscriptstyle B}{%
\!,N}}=\frac{1}{Z_{\!{\scriptscriptstyle B}{\!,N}}}\sum_{k=q}^{N}\tbinom{k+m_{%
j}-q-1}{m_{j}-1}\mathrm{e}^{-\beta\epsilon_{j}k}Z_{\!{\scriptscriptstyle B}{\!%
,N-k}},$$
(50)
where $q\leq m_{j}$. Note that if we set both of $m_{j}=\ell$ and $q=\ell$, then comparing with Eq. (48) we see that
$$\expectationvalue{{\binom{n_{j^{(s)}}}{\ell}}}_{\!{\scriptscriptstyle B}{\!,N}%
}=\expectationvalue{n_{j^{(0)}}\dots n_{j^{(\ell-1)}}}_{\!{\scriptscriptstyle B%
}{\!,N}},$$
(51)
for $s\in\{0,\dots,\ell-1\}$. This demonstrates that the correlations between degenerate levels can be expressed in terms of moments of the occupation number of any of the degenerate levels. This also helps to simplify the calculation of such moments, for example, we can write
$$\expectationvalue{n_{j}^{2}}_{\!{\scriptscriptstyle B}{\!,N}}=%
\expectationvalue{{\tbinom{n_{j}}{2}}}_{\!{\scriptscriptstyle B}{\!,N}}+%
\expectationvalue{{\tbinom{n_{j}+1}{2}}}_{\!{\scriptscriptstyle B}{\!,N}}$$
which can be simplified using Eq. (50) as
$$\expectationvalue{n_{j}^{2}}_{\!{\scriptscriptstyle B}{\!,N}}=\frac{1}{Z_{\!{%
\scriptscriptstyle B}{\!,N}}}\sum_{k=1}^{N}\quantity(2k-1)\mathrm{e}^{-\beta%
\epsilon_{j}k}Z_{\!{\scriptscriptstyle B}{\!,N-k}}.$$
(52)
In the same fashion, we can write
$$\expectationvalue{n_{j}^{3}}_{\!{\scriptscriptstyle B}{\!,N}}=%
\expectationvalue{{\tbinom{n_{j}}{3}}}_{\!{\scriptscriptstyle B}{\!,N}}+4%
\expectationvalue{{\tbinom{n_{j}+1}{3}}}_{\!{\scriptscriptstyle B}{\!,N}}+%
\expectationvalue{{\tbinom{n_{j}+2}{3}}}_{\!{\scriptscriptstyle B}{\!,N}}$$
and thus
$$\expectationvalue{n_{j}^{3}}_{\!{\scriptscriptstyle B}{\!,N}}=\frac{1}{Z_{\!{%
\scriptscriptstyle B}{\!,N}}}\sum_{k=1}^{N}\quantity(3k^{2}-3k+1)\mathrm{e}^{-%
\beta\epsilon_{j}k}Z_{\!{\scriptscriptstyle B}{\!,N-k}}.$$
(53)
III Decomposition of level correlations into occupation numbers
In the previous section, we illustrated that the joint probability distributions of the occupation numbers and corresponding level correlations can be represented by auxiliary partition functions. The APF is distinguished from the actual partition function of the $N$-particle system through either the inclusion or exclusion of a set of levels from/to the complete spectrum under study. The resulting complexity of performing an actual calculation thus depends on the size of the modified set as demonstrated by, e.g. Eq. (46).
It is known that the resulting complexity can be reduced by relating higher-order correlations between non-degenerate levels to the related level-occupation numbers Schönhammer (2017); Giraud et al. (2018), representing an approach similar to Wicks theorem which only holds in the grand canonical ensemble. In this section, we directly obtain many known results using the APF method and, more importantly, generalize them to deal with degenerate energy levels.
Before we introduce the general and systematic approach to this problem, let us return to Eqs. (23) and (35) and consider two specific examples of increasing difficulty.
III.1 Examples
III.1.1 Two-level correlations
Consider the expectation value of two bosonic energy levels $j_{1}$ and $j_{2}$ where
$\epsilon_{j_{1}}\neq\epsilon_{j_{2}}$: $\expectationvalue{n_{j_{1}}n_{j_{2}}}_{\!{\scriptscriptstyle B}{\!,N}}$. Employing Eq. (35) with $m_{j_{1}}=m_{j_{2}}=1$, $q_{j_{1}}=0$ and $q_{j_{2}}=1$, we find
$$\expectationvalue{\quantity(n_{j_{1}}+1)n_{j_{2}}}_{\!{\scriptscriptstyle B}{%
\!,N}}=\frac{\mathrm{e}^{-\beta\epsilon_{j_{2}}}}{Z_{\!{\scriptscriptstyle B}{%
\!,N}}}Z_{\!{\scriptscriptstyle B}{\!,N-1}}^{\cup\quantity{{j_{1},j_{2}}}},$$
(54)
and upon exchanging the values of $q_{j_{1}}$ and $q_{j_{2}}$, we have
$$\expectationvalue{n_{j_{1}}\quantity(n_{j_{2}}+1)}_{\!{\scriptscriptstyle B}{%
\!,N}}=\frac{\mathrm{e}^{-\beta\epsilon_{j_{1}}}}{Z_{\!{\scriptscriptstyle B}{%
\!,N}}}Z_{\!{\scriptscriptstyle B}{\!,N-1}}^{\cup\quantity{{j_{1},j_{2}}}}.$$
(55)
Next, eliminating the APF from the two equations yields our final result:
$$\expectationvalue{n_{j_{1}}n_{j_{2}}}_{\!{\scriptscriptstyle B}{\!,N}}=-\frac{%
\mathrm{e}^{\beta\epsilon_{j_{1}}}\expectationvalue{n_{j_{1}}}_{\!{%
\scriptscriptstyle B}{\!,N}}-\mathrm{e}^{\beta\epsilon_{j_{2}}}%
\expectationvalue{n_{j_{2}}}_{\!{\scriptscriptstyle B}{\!,N}}}{\mathrm{e}^{%
\beta\epsilon_{j_{1}}}-\mathrm{e}^{\beta\epsilon_{j_{2}}}}.$$
(56)
Similarly, if the levels are fermionic, we use Eq. (23) with $\quantity(\gamma_{j_{1}},\gamma_{j_{2}})=\quantity(0,1)$ and $\quantity(1,0)$ to find:
$$\expectationvalue{n_{j_{1}}n_{j_{2}}}_{\!{\scriptscriptstyle F}{\!,N}}=\frac{%
\mathrm{e}^{\beta\epsilon_{j_{1}}}\expectationvalue{n_{j_{1}}}_{\!{%
\scriptscriptstyle F}{\!,N}}-\mathrm{e}^{\beta\epsilon_{j_{2}}}%
\expectationvalue{n_{j_{2}}}_{\!{\scriptscriptstyle F}{\!,N}}}{\mathrm{e}^{%
\beta\epsilon_{j_{1}}}-\mathrm{e}^{\beta\epsilon_{j_{2}}}}.$$
(57)
These known results Schönhammer (2017); Giraud et al. (2018) are thus obtainable within the APF approach with a few lines of algebra by generating a set of independent equations. We now extend this idea to three energy levels.
III.1.2 Three-level correlations
The previous example for bosons is modified by adding a third level $j_{3}$ with $m_{j_{3}}=1$ and $\epsilon_{j_{3}}$ that is different than both of $\epsilon_{j_{1}}$ and $\epsilon_{j_{2}}$. Setting $\quantity(q_{j_{1}},q_{j_{2}},q_{j_{3}})=\quantity(1,0,1)$ and $\quantity(1,1,0)$ in Eq. (35) gives the two equations
$$\displaystyle\expectationvalue{n_{j_{1}}\quantity(n_{j_{2}}+1)n_{j_{3}}}_{\!{%
\scriptscriptstyle B}{\!,N}}$$
$$\displaystyle=\frac{\mathrm{e}^{-\beta\quantity(\epsilon_{j_{1}}+\epsilon_{j_{%
3}})}}{Z_{\!{\scriptscriptstyle B}{\!,N}}}Z_{\!{\scriptscriptstyle B}{\!,N-2}}%
^{\cup\quantity{{j_{1},j_{2},j_{3}}}}$$
(58)
$$\displaystyle\expectationvalue{n_{j_{1}}n_{j_{2}}\quantity(n_{j_{3}}+1)}_{\!{%
\scriptscriptstyle B}{\!,N}}$$
$$\displaystyle=\frac{\mathrm{e}^{-\beta\quantity(\epsilon_{j_{1}}+\epsilon_{j_{%
2}})}}{Z_{\!{\scriptscriptstyle B}{\!,N}}}Z_{\!{\scriptscriptstyle B}{\!,N-2}}%
^{\cup\quantity{{j_{1},j_{2},j_{3}}}}$$
(59)
respectively. Solving for $\expectationvalue{n_{j_{1}}n_{j_{2}}n_{j_{3}}}_{\!{\scriptscriptstyle B}{\!,N}}$ leads to
$$\expectationvalue{n_{j_{1}}n_{j_{2}}n_{j_{3}}}_{\!{\scriptscriptstyle B}{\!,N}%
}=-\frac{\mathrm{e}^{\beta\epsilon_{j_{2}}}\expectationvalue{n_{j_{1}}n_{j_{2}%
}}_{\!{\scriptscriptstyle B}{\!,N}}-\mathrm{e}^{\beta\epsilon_{j_{3}}}%
\expectationvalue{n_{j_{1}}n_{j_{3}}}_{\!{\scriptscriptstyle B}{\!,N}}}{%
\mathrm{e}^{\beta\epsilon_{j_{2}}}-\mathrm{e}^{\beta\epsilon_{j_{3}}}}.$$
(60)
which can be further broken down into single-level occupation numbers by application of Eq. (56).
A slightly modified approach can be used if two of the energy levels are degenerate,
$\epsilon_{j_{2}}=\epsilon_{j_{3}}\neq\epsilon_{j_{1}}$. As above, we denote degenerate levels via superscript and we relabel $j_{2}=j_{2}^{(0)}$ and $j_{3}=j_{2}^{(1)}$.
Then it is clear that $\expectationvalue{n_{j_{1}}\!\!n_{j_{2}^{(0)}}}_{\!{\scriptscriptstyle B}{\!,N%
}}=\expectationvalue{n_{j_{1}}\!\!n_{j_{2}^{(1)}}}_{\!{\scriptscriptstyle B}{%
\!,N}}$ and Eq. (60) is not immediately applicable. However, this can be resolved by replacing one of the choices of $\quantity(q_{j_{1}}\!\!\!\!,q_{j_{2}^{(0)}},q_{j_{2}^{(1)}})$, say $\quantity(1,1,0)$, with $\quantity(0,1,1)$, which gives
$$\displaystyle\expectationvalue{\!n_{j_{1}}\!\!\!n_{j_{2}^{(0)}}n_{j_{2}^{(1)}}%
\!}_{\!{\scriptscriptstyle B}{\!,N}}=\\
\displaystyle-\frac{\mathrm{e}^{\beta\epsilon_{j_{1}}}\!\expectationvalue{\!n_%
{j_{1}}\!\!n_{j_{2}^{(1)}}\!}_{\!{\scriptscriptstyle B}{\!,N}}\!-\mathrm{e}^{%
\beta\epsilon_{j_{2}}}\expectationvalue{n_{j_{2}^{(0)}}n_{j_{2}^{(1)}}\!\!}_{%
\!{\scriptscriptstyle B}{\!,N}}}{\mathrm{e}^{\beta\epsilon_{j_{1}}}-\mathrm{e}%
^{\beta\epsilon_{j_{2}}}}.$$
(61)
We now turn to the general decomposition of $\ell$-level correlations into functions of the occupation numbers of the energy levels for both the non-degenerate and degenerate spectra.
III.2 A Systematic Approach
In the following we show that Eqs. (6), (18) and (19) can be directly employed to systematically relate higher-order correlations to levels occupation numbers. We begin by considering the set of levels $\mathcal{S}_{\ell}=\{j_{1},j_{2},\dots,j_{\ell}\}$ and relate the corresponding $\ell$-point correlations to the $r$-point correlations for any nonempty subset of levels $\mathcal{S}_{r}=\{i_{1},i_{2},\dots,i_{r}\}\subset\mathcal{S}_{\ell}$.
Imposing fermionic level statistics on the spectrum $\mathcal{S}$, we can relate the correlations $\expectationvalue{n_{j_{1}}n_{j_{2}}\dots n_{j_{\ell}}}_{\!{\scriptscriptstyle
F%
}{\!,N}}$ to $\expectationvalue{n_{i_{1}}n_{i_{2}}\dots n_{i_{r}}}_{\!{\scriptscriptstyle F}%
{\!,N}}$ by relating their corresponding APFs, i.e., $Z_{\!{\scriptscriptstyle F}{\!,N-\ell}}^{\setminus{\mathcal{S}_{\ell}}}$ and $Z_{\!{\scriptscriptstyle F}{\!,N-r}}^{\setminus{\mathcal{S}_{r}}}$. This can be achieved by building $Z_{\!{\scriptscriptstyle F}{\!,N-1}}^{\setminus{\mathcal{S}_{r}}}$ via Eq. (6) to combine the APFs of $\mathcal{S}\setminus\mathcal{S}_{\ell}$ with that of $\mathcal{S}_{\ell}\setminus\mathcal{S}_{r}$ through
$$Z_{\!{\scriptscriptstyle F}{\!,N-r}}^{\setminus{\mathcal{S}_{r}}}=\sum_{k=0}^{%
\ell-r}Z_{\!{\scriptscriptstyle F}{\!,k}}^{\setminus{\mathcal{S}_{r}}}(%
\mathcal{S}_{\ell})Z_{\!{\scriptscriptstyle F}{\!,N-r-k}}^{\setminus{\mathcal{%
S}_{\ell}}}.$$
(62)
We note the upper limit in the previous summation is $\ell-r$, where in general it should be $k_{\max}=\min(\ell-r,N-r)$ (Eq. (6)). This is because the fermionic APF $Z_{\!{\scriptscriptstyle F}{\!,k}}^{\setminus{\mathcal{S}_{r}}}(\mathcal{S}_{%
\ell})$ cannot describe more than $\ell-r$ particles, as this is the number of levels that are in $\mathcal{S}_{\ell}\setminus\mathcal{S}_{r}$. Further, for $N<\ell$ the correlations $\langle n_{j_{1}}n_{j_{2}}\dots n_{j_{\ell}}\rangle_{N}=0$ for general particle statistics. As a result, we only consider $\ell$-points correlations with $N\geq\ell$. Moreover, we can obtain $Z_{\!{\scriptscriptstyle F}{\!,k}}^{\setminus{\mathcal{S}_{r}}}(\mathcal{S}_{%
\ell})$ by removing the contribution of the levels $\mathcal{S}_{r}$ from the APF of $\mathcal{S}_{\ell}$, using Eq. (18), as
$$Z_{\!{\scriptscriptstyle F}{\!,k}}^{\setminus{\mathcal{S}_{r}}}(\mathcal{S}_{%
\ell})=\sum_{m=0}^{k}(-1)^{m}Z_{\!{\scriptscriptstyle B}{\!,m}}(\mathcal{S}_{r%
})Z_{\!{\scriptscriptstyle F}{\!,k-m}}(\mathcal{S}_{\ell}).$$
(63)
If we isolate the last term in the summation in Eq. (62), i.e., $Z_{\!{\scriptscriptstyle F}{\!,\ell-r}}^{\setminus{\mathcal{S}_{r}}}(\mathcal{%
S}_{\ell})Z_{\!{\scriptscriptstyle F}{\!,N-\ell}}^{\setminus{\mathcal{S}_{\ell%
}}}$, while substituting for $Z_{\!{\scriptscriptstyle F}{\!,k}}^{\setminus{\mathcal{S}_{r}}}(\mathcal{S}_{%
\ell})$, using Eq. (63) in all other terms, we find
$$\displaystyle Z_{\!{\scriptscriptstyle F}{\!,N-r}}^{\setminus{\mathcal{S}_{r}}}=$$
$$\displaystyle\sum_{k=0}^{\ell-r-1}\sum_{m=0}^{k}(-1)^{m}Z_{\!{%
\scriptscriptstyle B}{\!,m}}(\mathcal{S}_{r})Z_{\!{\scriptscriptstyle F}{\!,k-%
m}}(\mathcal{S}_{\ell})Z_{\!{\scriptscriptstyle F}{\!,N-r-k}}^{\setminus{%
\mathcal{S}_{\ell}}}$$
$$\displaystyle+Z_{\!{\scriptscriptstyle F}{\!,\ell-r}}^{\setminus{\mathcal{S}_{%
r}}}(\mathcal{S}_{\ell})Z_{\!{\scriptscriptstyle F}{\!,N-\ell}}^{\setminus{%
\mathcal{S}_{\ell}}}.$$
(64)
After changing the order of the summations and shifting the indexes $k\to k-r+1$ and $m\to m-r$, we obtain the unwieldy expression
$$\displaystyle Z_{\!{\scriptscriptstyle F}{\!,N-r}}^{\setminus{\mathcal{S}_{r}}}$$
$$\displaystyle=(-1)^{r-1}\!\!\sum_{m=r}^{\ell-1}(-1)^{m+1}Z_{\!{%
\scriptscriptstyle B}{\!,m-r}}(\mathcal{S}_{r})$$
$$\displaystyle\times\sum_{k=m-1}^{\ell-2}\!\!\!\!Z_{\!{\scriptscriptstyle F}{\!%
,k-m+1}}(\mathcal{S}_{\ell})Z_{\!{\scriptscriptstyle F}{\!,N-k-1}}^{\setminus{%
\mathcal{S}_{\ell}}}$$
$$\displaystyle+Z_{\!{\scriptscriptstyle F}{\!,\ell-r}}^{\setminus{\mathcal{S}_{%
r}}}(\mathcal{S}_{\ell})Z_{\!{\scriptscriptstyle F}{\!,N-\ell}}^{\setminus{%
\mathcal{S}_{\ell}}}.$$
(65)
Next, we substitute for $Z_{\!{\scriptscriptstyle F}{\!,N-\ell}}^{\setminus{\mathcal{S}_{\ell}}}$ and $Z_{\!{\scriptscriptstyle F}{\!,N-r}}^{\setminus{\mathcal{S}_{r}}}$, using Eq. (22) and the fully occupied fermionic APF $Z_{\!{\scriptscriptstyle F}{\!,\ell-r}}^{\setminus{\mathcal{S}_{r}}}(\mathcal{%
S}_{\ell})=\mathrm{e}^{-\beta\sum_{j_{\nu}\in\mathcal{S}_{\ell}\setminus%
\mathcal{S}_{r}}\epsilon_{j_{\nu}}}$. Multiplying the result by $\mathrm{e}^{-\beta\sum_{i_{\nu}\in\mathcal{S}_{r}}\epsilon_{i_{\nu}}}/Z_{\!{%
\scriptscriptstyle F}{\!,N}}$ yields:
$$Y_{\!{\scriptscriptstyle F}{\!,0}}(\mathcal{S}_{\ell})+\sum_{m=r}^{\ell-1}A_{m%
}(\mathcal{S}_{r})Y_{\!{\scriptscriptstyle F}{\!,m}}(\mathcal{S}_{\ell})=b_{F}%
(\mathcal{S}_{r})$$
(66)
where
$$\displaystyle Y_{\!{\scriptscriptstyle F}{\!,0}}(\mathcal{S}_{\ell})$$
$$\displaystyle=\expectationvalue{n_{j_{1}}n_{j_{2}}\dots n_{j_{\ell}}}_{\!{%
\scriptscriptstyle F}{\!,N}}$$
(67)
$$\displaystyle Y_{\!{\scriptscriptstyle F}{\!,1\leq m\leq\ell-1}}(\mathcal{S}_{%
\ell})$$
$$\displaystyle=\frac{(-1)^{m+1}}{Z_{\!{\scriptscriptstyle F}{\!,N}}}\!\!\!\!%
\sum_{k=m-1}^{\ell-2}\!\!\!\!Z_{\!{\scriptscriptstyle F}{\!,k-m+1}}(\mathcal{S%
}_{\ell})Z_{\!{\scriptscriptstyle F}{\!,N-k-1}}^{\setminus{\mathcal{S}_{\ell}}}$$
(68)
are independent of $\mathcal{S}_{r}$. Therefore, for each choice of the subset
$\mathcal{S}_{r}$ we can write the linear nonhomogeneous equation
(66) in the $\ell$ variables $Y_{\!{\scriptscriptstyle F}{\!,m}}$ with coefficients:
$A_{0}=1$, $A_{0<m<r}=0$ and $A_{r\leq m\leq\ell-1}(\mathcal{S}_{r})=(-1)^{r-1}\mathrm{e}^{-\beta\sum_{i_{%
\nu}\in\mathcal{S}_{r}}\epsilon_{i_{\nu}}}Z_{\!{\scriptscriptstyle B}{\!,m-r}}%
(\mathcal{S}_{r})$. The homogeneity of the linear equation is violated by the term $b_{F}(\mathcal{S}_{r})=\expectationvalue{n_{i_{1}}n_{i_{2}}\dots n_{i_{r}}}_{%
\!{\scriptscriptstyle F}{\!,N}}$.
With this formulation, we observe that for any of the $2^{\ell}-2$ choices of $\mathcal{S}_{r}$, we can write a linear equation in the same $\ell$ variables $Y_{\!{\scriptscriptstyle F}{\!,m}}$, where $Y_{\!{\scriptscriptstyle F}{\!,0}}=\expectationvalue{n_{j_{1}}n_{j_{2}}\dots n%
_{j_{\ell}}}_{\!{\scriptscriptstyle F}{\!,N}}$ is the $\ell$-point correlation while the remaining $\ell-1$ variables are auxiliary, and depend symmetrically on the levels in $\mathcal{S}_{\ell}$. Also, the $r$-point correlation $\expectationvalue{n_{i_{1}}n_{i_{2}}\dots n_{i_{r}}}_{\!{\scriptscriptstyle F}%
{\!,N}}$ of the levels in $\mathcal{S}_{r}$ plays the role of the nonhomogeneous term in the linear equation and the coefficients $A_{m}$ of the equation can be determined by the bosonic APFs of $\mathcal{S}_{r}$. An analogous expression can be obtained for bosonic statistics, with the same coefficients $A_{m}$
$$Y_{\!{\scriptscriptstyle B}{\!,0}}(\mathcal{S}_{\ell})+\sum_{m=r}^{\ell-1}A_{m%
}(\mathcal{S}_{r})Y_{\!{\scriptscriptstyle B}{\!,m}}(\mathcal{S}_{\ell})=b_{B}%
(\mathcal{S}_{r}),$$
(69)
where, in this case, the variables are
$$\displaystyle Y_{\!{\scriptscriptstyle B}{\!,1\leq m\leq\ell-1}}(\mathcal{S}_{%
\ell})=\\
\displaystyle\frac{(-1)^{m+1}}{Z_{\!{\scriptscriptstyle B}{\!,N}}}\sum_{k=m-1}%
^{\ell-2}(-1)^{k}Z_{\!{\scriptscriptstyle F}{\!,k-m+1}}(\mathcal{S}_{\ell})Z_{%
\!{\scriptscriptstyle B}{\!,N-k-1}}^{\cup{\mathcal{S}_{\ell}}},$$
(70)
with $Y_{\!{\scriptscriptstyle B}{\!,0}}(\mathcal{S}_{\ell})=\quantity(-1)^{\ell-1}%
\expectationvalue{n_{j_{1}}n_{j_{2}}\dots n_{j_{\ell}}}_{\!{\scriptscriptstyle
B%
}{\!,N}}$ and $b_{B}(\mathcal{S}_{r})=(-1)^{r-1}\expectationvalue{n_{i_{1}}n_{i_{2}}\dots n_{%
i_{r}}}_{\!{\scriptscriptstyle B}{\!,N}}$ (see Appendix A for a complete derivation).
III.2.1 Non-degenerate levels
Consider the set $\mathcal{S}_{\ell}$ specifying a set of distinct energy levels and choose $\mathcal{S}_{r=1}$ such that it contains only one of the $\ell$ levels in $\mathcal{S}_{\ell}$. We can use Eq. (66) or (69) to construct a set of $\ell$ linear equations each corresponding to one level $j_{s}\in\mathcal{S}_{\ell}$ with energy $\epsilon_{j_{s}}$. For fermions the equations are
$$\expectationvalue{n_{j_{s}}}_{\!{\scriptscriptstyle F}{\!,N}}=\sum_{m=0}^{\ell%
-1}\mathrm{e}^{-m\beta\epsilon_{j_{s}}}Y_{\!{\scriptscriptstyle F}{\!,m}}(%
\mathcal{S}_{\ell}),$$
(71)
where the coefficients $A_{m}$ were obtained from the single-level bosonic APF $Z_{\!{\scriptscriptstyle B}{\!,m}}(\{j_{s}\})=\mathrm{e}^{-m\beta\epsilon_{j_{%
s}}}$ in Eq. (12). Therefore, using the set of the $\ell$ independent linear equations in $\ell$ variables defined by Eq. (71), we can solve for $Y_{\!{\scriptscriptstyle F}{\!,0}}(\mathcal{S}_{\ell})=\expectationvalue{n_{j_%
{1}}n_{j_{2}}\dots n_{j_{\ell}}}_{\!{\scriptscriptstyle F}{\!,N}}$ as
$$\displaystyle\expectationvalue{n_{j_{1}}\dots n_{j_{\ell}}}_{\!{%
\scriptscriptstyle F}{\!,N}}$$
$$\displaystyle=\frac{\begin{vmatrix}\expectationvalue{n_{j_{1}}}_{\!{%
\scriptscriptstyle F}{\!,N}}&\mathrm{e}^{-\beta\epsilon_{j_{1}}}&\dots&\mathrm%
{e}^{-\beta(\ell-1)\epsilon_{j_{1}}}\\
\vdots&\vdots&\ddots&\vdots\\
\expectationvalue{n_{j_{\ell}}}_{\!{\scriptscriptstyle F}{\!,N}}&\mathrm{e}^{-%
\beta\epsilon_{j_{\ell}}}&\dots&\mathrm{e}^{-\beta(\ell-1)\epsilon_{j_{\ell}}}%
\\
\end{vmatrix}}{\begin{vmatrix}1&\mathrm{e}^{-\beta\epsilon_{j_{1}}}&\dots&%
\mathrm{e}^{-\beta(\ell-1)\epsilon_{j_{1}}}\\
\vdots&\vdots&\ddots&\vdots\\
1&\mathrm{e}^{-\beta\epsilon_{j_{\ell}}}&\dots&\mathrm{e}^{-\beta(\ell-1)%
\epsilon_{j_{\ell}}}\end{vmatrix}}.$$
(72)
This result was recently obtained by Giraud, Grabsch and Texier Giraud et al. (2018), using the properties of the Schur functions and it can be simplified using Vandermonde determinants:
$$\displaystyle\expectationvalue{n_{j_{1}}n_{j_{2}}\dots n_{j_{\ell}}}_{\!{%
\scriptscriptstyle F}{\!,N}}$$
$$\displaystyle=\sum_{s=1}^{\ell}(-1)^{s-1}\expectationvalue{n_{j_{s}}}_{\!{%
\scriptscriptstyle F}{\!,N}}\mathrm{e}^{-\beta\sum_{j_{i}\neq j_{s}}\epsilon_{%
j_{i}}}$$
$$\displaystyle\quad\times\frac{\mathbb{V}^{\setminus\{j_{s}\}}\quantity(\mathrm%
{e}^{-\beta\epsilon_{j_{1}}},\dots,\mathrm{e}^{-\beta\epsilon_{j_{\ell}}})}{%
\mathbb{V}\quantity(\mathrm{e}^{-\beta\epsilon_{j_{1}}},\dots,\mathrm{e}^{-%
\beta\epsilon_{j_{\ell}}})},$$
(73)
where
$$\mathbb{V}\quantity(\alpha_{1},\dots,\alpha_{\ell})=\begin{vmatrix}1&\alpha_{1%
}&\dots&\alpha_{1}^{\ell-1}\\
\vdots&\vdots&\ddots&\vdots\\
1&\alpha_{\ell}&\dots&\alpha_{\ell}^{\ell-1}\end{vmatrix}=\prod_{i<j}\quantity%
(\alpha_{j}-\alpha_{i}).$$
(74)
Thus the fermionic $\ell$-level correlation can be simplified as:
$$\expectationvalue{n_{j_{1}}n_{j_{2}}\dots n_{j_{\ell}}}_{\!{\scriptscriptstyle
F%
}{\!,N}}=\sum_{r=1}^{\ell}\frac{\expectationvalue{n_{j_{r}}}_{\!{%
\scriptscriptstyle F}{\!,N}}}{\prod_{k\neq r}\quantity[1-\mathrm{e}^{\beta%
\quantity(\epsilon_{j_{k}}-\epsilon_{j_{r}})}]}.$$
(75)
This expression was also recently derived using an elegant second quantization schemeSchönhammer (2017).
An equivalent procedure can be performed for bosons, again using the set $\mathcal{S}_{r=1}$ to yield $\ell$ linearly independent equations
$$\expectationvalue{n_{j_{r}}}_{\!{\scriptscriptstyle B}{\!,N}}=\sum_{m=0}^{\ell%
-1}\mathrm{e}^{-m\beta\epsilon_{j_{r}}}Y_{\!{\scriptscriptstyle B}{\!,m}}(%
\mathcal{S}_{\ell}),$$
(76)
such that $\expectationvalue{n_{j_{1}}n_{j_{2}}\dots n_{j_{\ell}}}_{\!{\scriptscriptstyle
B%
}{\!,N}}=\quantity(-1)^{\ell-1}Y_{\!{\scriptscriptstyle B}{\!,0}}(\mathcal{S}_%
{\ell})$ can also be expressed in terms of determinants
leading toGiraud et al. (2018)
$$\expectationvalue{n_{j_{1}}n_{j_{2}}\dots n_{j_{\ell}}}_{\!{\scriptscriptstyle
B%
}{\!,N}}=\quantity(-1)^{\ell-1}\sum_{r=1}^{\ell}\frac{\expectationvalue{n_{j_{%
r}}}_{\!{\scriptscriptstyle B}{\!,N}}}{\prod_{k\neq r}\quantity[1-\mathrm{e}^{%
\beta\quantity(\epsilon_{j_{k}}-\epsilon_{j_{r}})}]}.$$
(77)
III.2.2 Degenerate levels
Up until this point we have considered $\ell$-point level correlations in two opposite regimes: (I) when all $\ell$-levels are degenerate Eq. (48) provides a direct route to the correlations through the determination of all partition functions up to $N$ particles, and (II) when all $\ell$ levels are distinct, an associated set of $\ell$ linear equations yields the correlations in terms of individual level occupation numbers. The independence of these linear equations, and thus the existence of a unique solution, is violated in the presence of degeneracy.
We now study the most general possible $\ell$-level correlation function defined by the set $\mathcal{S}_{\ell}$ which could include both degenerate and non-degenerate levels. Consider the subset of level indices $\mathcal{S}_{m_{i}}=\{i^{(0)},\dots,i^{(m_{i}-1)}\}\subset\mathcal{S}_{\ell}$ which contains $m_{i}>1$ degenerate levels (for $m_{i}=1$, we reproduce the non-degenerate analysis discussed above). As a result, the corresponding $m_{i}$ equations, out of the total set of $\ell$ linear equations defined by Eqs. (71) and (76) for fermions and bosons respectively are identical, as they are distinguished from each other only via the energies of the involved levels and their occupation numbers. Moreover, $\mathcal{S}_{\ell}$, could contain multiple subsets of degenerate energy levels, further complicating the problem. Eqs. (75) and (77) can not be applied in this case, as is apparent from their vanishing denominators whenever $\epsilon_{j_{k}}=\epsilon_{j_{r}}$.
To resolve the complication introduced by degenerate subsets, we generalize the procedure in Sec.III.1.2 to treat the case where a three-level correlation contained a subset of 2 degenerate levels. More explicitly, we relate the $\ell$-points correlations not only to the occupation numbers of the degenerate subsets, but to all of the $m_{i}$ distinct $r$-points correlations between the degenerate levels with $2\leq r\leq m_{i}$. This is useful as we have already introduced Eq. (48), which simplifies the calculation of the correlations between degenerate energy levels for fermionic and bosonic statistics. In addition, it will result in the generation of $m_{i}$ new independent equations that could be used to calculate $\ell$-points correlations.
Accordingly, for any choice of $\mathcal{S}_{r}=\{i^{(0)},\dots,i^{(r-1)}\}\subset\mathcal{S}_{m_{i}}\subset%
\mathcal{S}_{\ell}$, the corresponding coefficients in the constructed linear equations are $A_{0}=1$, $A_{0<m<r}=0$ and
$$A_{r\leq m\leq\ell-1}(\mathcal{S}_{r})=(-1)^{r-1}\binom{m-1}{r-1}\mathrm{e}^{-%
m\beta\epsilon_{i}},$$
(78)
where $\epsilon_{i}$ is the energy of all degenerate levels in $\mathcal{S}_{m_{i}}$ and
we have substituted for $Z_{\!{\scriptscriptstyle B}{\!,m-r}}(\{i^{(0)},\dots,i^{(r)}\})=\binom{m-1}{r-%
1}\mathrm{e}^{-(m-r)\beta\epsilon_{i}}$, a bosonic partition function of $r$ degenerate levels.
The non-homogeneities $b_{\zeta}(\mathcal{S}_{r})=(-\zeta)^{r-1}\expectationvalue{n_{i^{(1)}}n_{i^{(2%
)}}\dots n_{i^{(r-1)}}}_{\!{\scriptscriptstyle\zeta}{\!,N}}$ can be
calculated using Eq. (48), and thus, the original set of $m_{i}$ identical equations, can now be replaced with the following $m_{i}$ independent equations
$$Y_{\!{\scriptscriptstyle\zeta}{\!,0}}(\mathcal{S}_{\ell})+\sum_{m=r}^{\ell-1}(%
-1)^{r-1}\binom{m-1}{r-1}\mathrm{e}^{-m\beta\epsilon_{i}}Y_{\!{%
\scriptscriptstyle\zeta}{\!,m}}(\mathcal{S}_{\ell})=b_{\zeta}(\mathcal{S}_{r}),$$
(79)
for fermions $(\zeta=-1)$ and bosons $(\zeta=+1)$.
To illustrate how this works in practice, consider the $4$-point correlation of the levels $\{j_{1^{(0)}},j_{1^{(1)}},j_{1^{(2)}},j_{2}\!\!\}$, labeling distinct energies $\epsilon_{1}$ and $\epsilon_{2}$. The resulting new set of $\ell$ equations can be solved for both fermions and bosons to give:
$$\displaystyle\expectationvalue{n_{j_{1^{(0)}}}n_{j_{1^{(1)}}}n_{j_{1^{(2)}}}n_%
{j_{2}\!\!}}_{\!{\scriptscriptstyle\zeta}{\!,N}}$$
$$\displaystyle=(-\zeta)\frac{\begin{vmatrix}\expectationvalue{n_{j_{1^{(0)}}}n_%
{j_{1^{(1)}}}n_{j_{1^{(2)}}}}_{\!{\scriptscriptstyle\zeta}{\!,N}}&0&0&\mathrm{%
e}^{-3\beta\epsilon_{j_{1}}}\\
(-\zeta)\expectationvalue{n_{j_{1^{(0)}}}n_{j_{1^{(1)}}}}_{\!{%
\scriptscriptstyle\zeta}{\!,N}}&0&-\mathrm{e}^{-2\beta\epsilon_{j_{1}}}&-2%
\mathrm{e}^{-3\beta\epsilon_{j_{1}}}\\
\expectationvalue{n_{j_{1^{(0)}}}}_{\!{\scriptscriptstyle\zeta}{\!,N}}&\mathrm%
{e}^{-\beta\epsilon_{j_{1}}}&\mathrm{e}^{-2\beta\epsilon_{j_{1}}}&\mathrm{e}^{%
-3\beta\epsilon_{j_{1}}}\\
\expectationvalue{n_{j_{2}}\!\!}_{\!{\scriptscriptstyle\zeta}{\!,N}}&\mathrm{e%
}^{-\beta\epsilon_{j_{2}}}&\mathrm{e}^{-2\beta\epsilon_{j_{2}}}&\mathrm{e}^{-3%
\beta\epsilon_{j_{2}}}\\
\end{vmatrix}}{\begin{vmatrix}1&0&0&\mathrm{e}^{-3\beta\epsilon_{j_{1}}}\\
1&0&-\mathrm{e}^{-2\beta\epsilon_{j_{1}}}&-2\mathrm{e}^{-3\beta\epsilon_{j_{1}%
}}\\
1&\mathrm{e}^{-\beta\epsilon_{j_{1}}}&\mathrm{e}^{-2\beta\epsilon_{j_{1}}}&%
\mathrm{e}^{-3\beta\epsilon_{j_{1}}}\\
1&\mathrm{e}^{-\beta\epsilon_{j_{2}}}&\mathrm{e}^{-2\beta\epsilon_{j_{2}}}&%
\mathrm{e}^{-3\beta\epsilon_{j_{2}}}\\
\end{vmatrix}}.$$
(80)
IV Applications
To illustrate the applicability of our results for degenerate non-interacting systems of particles with fixed number, and highlight the practical usage of Eq. (80) we consider a one-dimensional tight-binding chain of $N$ spinfull bosons hopping over $L$ lattice sites. Code, scripts, and data used to produce all figures for the bosonic chain results can be found online rep (2020). The bosonic chain is subject to a static external magnetic field $B$ applied along the $z$-axis. Spin-$S$ bosons are described by the Hamiltonian
$$\widehat{H}=-t\sum_{\alpha,\sigma}\left(\hat{a}^{\dagger}_{\alpha+1,\sigma}%
\hat{a}_{\alpha,\sigma}+\text{h.c.}\right)-h\sum_{\alpha,\sigma,\sigma^{\prime%
}}\hat{a}^{\dagger}_{\alpha,\sigma}{S}^{z}_{\sigma,\sigma^{\prime}}\hat{a}_{%
\alpha,\sigma^{\prime}},$$
(81)
where $\hat{a}^{\dagger}_{\alpha,\sigma}$ and $\hat{a}_{\alpha,\sigma}$ are creation and annihilation operators for a boson at site $\alpha$ with $\sigma\in\{-S,\dots,0,\dots,S\}$ satisfying $[a_{\alpha,\sigma},a^{\dagger}_{\alpha^{\prime},\sigma^{\prime}}]=\delta_{%
\alpha,\alpha^{\prime}}\delta_{\sigma,\sigma^{\prime}}$ and $t$ measures the hopping amplitude. ${S}^{z}_{\sigma,\sigma^{\prime}}=\sigma\delta_{\sigma,\sigma^{\prime}}$ are the matrix elements of the diagonal $z$-projection of the spin-$S$ representation of the spin operator $\hat{\boldsymbol{S}}$. Here, $h=g{\rm{\mu_{B}}}B$, where $g$ is the corresponding spin-$S$ $g$-factor and $\mu_{B}$ is the Bohr magneton. We employ periodic boundary conditions, such that $\hat{a}_{L+1,\sigma}=\hat{a}_{1,\sigma}$, and to avoid having an unbalanced non-degenerate excited state, we fix the parity of $L$ to be odd.
The tight-binding Hamiltonian in Eq. (81) can be diagonalized
$$\widehat{H}=\sum_{j,\sigma}\epsilon_{j,\sigma}n_{j,\sigma},$$
(82)
where $\hat{n}_{j,\sigma}$ counts the number of bosons with energy
$$\epsilon_{j,\sigma}=-2t\cos\quantity(\frac{2\pi j}{L})-h\sigma,$$
(83)
and $j$ runs over the finite set $\mathcal{S}=\{-\frac{L-1}{2},\dots,0,\dots,\frac{L-1}{2}\}$ when $L$ is odd such that $\absolutevalue{\mathcal{S}}=L$. An examination of the single-particle spectrum shows that each energy level, except the ground state $\epsilon_{0,\sigma}=-2t-h\sigma$, is $2$-fold degenerate, where $\epsilon_{-j,\sigma}=\epsilon_{j,\sigma}$; a result of the right-left symmetry of the chain. Turning off the magnetic field and fixing $S>0$, gives rise to an extra degeneracy factor of $(2S+1)$ that affects all levels.
For all numerical results presented in this section, we fix $L=1001$, $N=1000$ and measure the inverse temperature $\beta=\frac{1}{k_{\rm B}T}$ in units of $1/t$, where $T$ is the absolute temperature and $k_{\rm{B}}$ is the Boltzmann constant.
IV.1 Spinless bosons $(S=0)$
We begin with the study of spinless bosons, where the model is insensitive to the applied magnetic field and we can drop the subscript $\sigma$ without loss of generality. Using the single-particle spectrum defined in Eq. (83) with $\sigma=0$ and $h=0$
in combination with Eq. (27), we calculate the joint probability distribution $\mathcal{P}_{\!\!{\scriptscriptstyle B}{\!,n_{0},n_{1}}}$ of the occupation numbers of the ground state and the first excited state, where, we choose the level $j=1$ out of the two degenerate levels $j=\pm 1$. Note that we do not bother to use the superscript notation to distinguish degenerate level indices $(1\equiv 1^{(0)},-1\equiv 1^{(1)})$ here as there are no ambiguities due to the sign of the index $j$.
The calculation proceeds by obtaining the APFs $Z_{\!{\scriptscriptstyle B}{\!,k}}^{\setminus\{{0,1}\}}$ for $0\leq k\leq N$ using the recursion relation Eq. (41), where the factors $C_{k}$ are calculated using the spectrum $\mathcal{S}\setminus\quantity{0,1}$. The resulting distribution is
$$\mathcal{P}_{\!\!{\scriptscriptstyle B}{\!,n_{0},n_{1}}}=\frac{\mathrm{e}^{-%
\beta\quantity(\epsilon_{0}n_{0}+\epsilon_{1}n_{1})}}{Z_{\!{\scriptscriptstyle
B%
}{\!,N}}}Z_{\!{\scriptscriptstyle B}{\!,N-n_{0}-n_{1}}}^{\setminus\{{0,1}\}}.$$
(84)
where $Z_{\!{\scriptscriptstyle B}{\!,N}}$ can be found by enforcing normalization.
We expect Eq. (84) to exhibit interesting features at low temperature where the particles are mostly occupying the ground state with some fluctuations amongst the low lying energy levels. To obtain an estimate of low in this context, we choose a value of the inverse temperature $\beta$ such that the ground state has a macroscopic occupation corresponding to 50% of the particles. We compare the ratio of the Boltzmann factors of having all particles in the ground state with that of having $N/2$ particles in the first excited state and the rest in the ground state. Setting the ratio of these factors $\mathrm{e}^{-\beta\quantity(\epsilon_{0}-\epsilon_{1})N/2}$ to $\sim 0.1$, suggests $\beta\sim 100/t$. The results are illustrated in the left panel of Fig. 1,
where the relative broadness of the distribution can be attributed to the degeneracy of the first exited level $j=\pm 1$.
If we now calculate the three-level joint probability distribution $\mathcal{P}_{\!\!{\scriptscriptstyle B}{\!,n_{0},n_{1},n_{-1}}}$ and consider the fixed slice with $n_{-1}=0$, as presented in the right panel of Fig. 1, we see that the distribution becomes significantly sharper, as blocking the level $j=-1$ makes the resulting non-normalized conditional distribution more sensitive to the conservation of the total number of particles.
We now turn to the calculation of the two-level connected correlation function for our bosonic system
$$\mathcal{C}(n_{i},n_{j})=\expectationvalue{n_{i}n_{j}}_{\!{\scriptscriptstyle B%
}{\!,N}}-\expectationvalue{n_{i}}_{\!{\scriptscriptstyle B}{\!,N}}%
\expectationvalue{n_{j}}_{\!{\scriptscriptstyle B}{\!,N}}.$$
(85)
The first step is to obtain the system partition function $Z_{\!{\scriptscriptstyle B}{\!,k}}$, recursively, using Eq. (41) starting from $Z_{\!{\scriptscriptstyle B}{\!,0}}$ up to $Z_{\!{\scriptscriptstyle B}{\!,N}}$. The occupation numbers $\expectationvalue{n_{j}}_{\!{\scriptscriptstyle B}{\!,N}}$ can then be easily calculated using Eq. (40). All that remains is to calculate the two-points correlations using Eq. (56) for the non-degenerate levels. For the correlations between degenerate levels ($\expectationvalue{n_{-j}n_{j}}_{\!{\scriptscriptstyle B}{\!,N}}$), we use Eq. (48), with $\zeta=+1$.
The results for $\mathcal{C}(n_{i},n_{j})$ for all levels $i$ and $j$ at inverse temperature $\beta=1/t$ are shown as a heat-map in the lower panel of Fig. 2.
Here, we chose temperature $\beta=1/t$ in order to distribute the correlations amongst higher energy levels. The upper panel of Fig. 2 shows $\mathcal{C}(n_{i},n_{j})$ as a function of $n_{i}$ for fixed $0\leq j\leq 500$ corresponding to horizontal cuts through the lower panel. The red open circles are correlations $\mathcal{C}(n_{-j},n_{j})$ for the degenerate levels
obtained from Eq. (48) demonstrating consistency with the rest of the graph. In the positive quadrant of the correlation heat-map, we use the values of $\expectationvalue{\tbinom{n_{j}}{2}}_{\!{\scriptscriptstyle B}{\!,N}}$ (marked with black circles) instead of $\expectationvalue{n_{j}^{2}}_{\!{\scriptscriptstyle B}{\!,N}}$, were the former is also consistent with surrounding data, as expected from Eq. (51), where $\expectationvalue{n_{-j}n_{j}}_{\!{\scriptscriptstyle B}{\!,N}}=%
\expectationvalue{\tbinom{n_{j}}{2}}_{\!{\scriptscriptstyle B}{\!,N}}$ and the symmetry $\mathcal{C}(n_{i},n_{j})=\mathcal{C}(n_{-i},n_{j})$ due to the degeneracy.
IV.2 Spin-$1$ bosons $(S=1)$
To illustrate the utility of auxiliary partitions functions in studying correlations in a highly degenerate spectrum, we consider the case of spin-$1$ bosons. In the absence of a magnetic field ($h=0$), each level picks up a degeneracy factor of $2S+1=3$, such that the ground state is three-fold degenerate and all of the excitation levels are six-fold degenerate.
Degeneracy effects are apparent in the two-level connected correlations $\mathcal{C}(n_{i,\sigma},n_{j,\sigma^{\prime}})$ at $\beta=1/t$, which we calculate for various values of $h$ as shown in Fig.3.
The top-left panel of the figure presents $\mathcal{C}(n_{i,\sigma},n_{j,\sigma^{\prime}})$ for $h=0$, where the choice of $\sigma$ and $\sigma^{\prime}$ matters only in the presence of a magnetic field. A comparison with the heat-map of Fig. 2, shows an overall broadening and a reduction of one order of magnitude in the maximum of $\mathcal{C}(n_{i,\sigma},n_{j,\sigma^{\prime}})$ for $S=1$, as compared to $S=0$ case.
Removing the energy-spin degeneracy by applying a strong magnetic field of $h=5t$, results in a splitting of the spectrum into three bands, each with bandwidth $4t$ and separated from each other via an energy bandgap of $t$. In this case, we first focus on correlations between the levels in the lower energy band ($\sigma=1$, bottom-left panel of Fig. 3), and see a partial recovery of the spinless bosons case (Fig. 2). Correlations that involve higher energy levels ($\sigma=0$ and $\sigma=-1$) are orders of magnitude weaker, at the considered temperatures, as shown in the right panels of Fig .3.
Finally, we turn to correlations between a set of levels that is partially degenerate, an interesting feature of spin-1 bosons. We consider the four-level (disconnected) correlations $\expectationvalue{n_{i,1}n_{j,1}n_{j,0}n_{j,-1}}_{\!{\scriptscriptstyle B}{\!,%
N}}$ between the levels $\epsilon_{i,1}$, $\epsilon_{j,1}$, $\epsilon_{j,0}$ and $\epsilon_{j,-1}$, where, in the absence of a magnetic field, the last three levels are degenerate for any $j$. We employ the bosonic version
of Eq. (80) with $\zeta=1$ with results shown in Fig. 4.
According to Eq. (35), the results that we obtain, in this case, also represent
$\expectationvalue{n_{i,\sigma}\tbinom{n_{j,\sigma^{\prime}}}{3}}_{\!{%
\scriptscriptstyle B}{\!,N}}$, for any $\sigma$ and $\sigma^{\prime}$ $\in\{1,0,-1\}$. For the fully degenerate case $i=-j$, we use Eq. (48). Once more, Eq. (35) guarantees that $\expectationvalue{n_{i,1}n_{-i,1}n_{-i,0}n_{-i,-1}}_{\!{\scriptscriptstyle B}{%
\!,N}}=\expectationvalue{\tbinom{n_{i,\sigma}}{4}}_{\!{\scriptscriptstyle B}{%
\!,N}}$. Therefore, we use $\expectationvalue{\tbinom{n_{i,\sigma}}{4}}_{\!{\scriptscriptstyle B}{\!,N}}$ instead of $\expectationvalue{n_{i,\sigma}^{4}}_{\!{\scriptscriptstyle B}{\!,N}}$, for the diagonal elements of the 4-level disconnected correlations presented in the lower panel of Fig. 4. The consistency of our calculations using different equations and methods described herein is demonstrated in the upper panel in analogy with Fig. 2.
V Discussion
In summary, we have presented a statistical theory of non-interacting identical quantum particles in the canonical ensemble, providing a unified framework that symmetrically captures both fermionic and bosonic statistics. Table 1 includes a listing of our most important results for fermions and bosons. We achieve this by: (1) Representing correlations (Eqs. (23) and (35)) and joint probability distributions ((21) and (27)) via auxiliary partition functions. (2) Deriving general relations between the canonical partition function of a given spectrum and that of the auxiliary partition function describing a spectral subset, as captured by Eqs. (8), (18) and (19).
These key equations can be manipulated to simplify the derivation of the known recursive relations for partition functions in the canonical ensemble and lead immediately to generalizations, and more importantly, provide useful formulas for calculating the correlations between degenerate energy levels and for calculating higher moments of the occupation numbers distribution. Also, Eqs. (23) and (35) can be used to reduce the complexity order of the desired correlations, or, to relate them to the occupation numbers of the involved levels and correlations between entirely degenerate levels (see Eq. (48)). Moreover, the ability to manipulate the way an auxiliary partition function is built out of other ones, allows us to construct a systematic approach towards the decomposition of many-energy level correlations in terms of individual level occupancies. This reflects the additional constraints between energy levels due to fixed $N$ even in the absence of interactions that are not present in a grand canonical description. Thus, we present an approach to working in the canonical ensemble that includes a generalization of Wick’s theorem, where we obtain previous results for non-degenerate levels Schönhammer (2017); Giraud et al. (2018) and extend them to the case of a degenerate spectrum.
Interestingly, despite the substantial difference between fermionic and bosonic statistics, the resulting formulas show evident similarity. If we compare Eqs. (36), (37) and (38). with Eqs. (39), (40) and (41), respectively, we see that the differences between the fermionic and the bosonic formulas can be captured by simple $\pm 1$ factors. In view of the current theory, such similarity is associated with the interplay between fermionic and bosonic auxiliary partition functions. The inverted symmetry between the two distinct statistics is apparent via a comparison of Eqs. (13) and (14) with Eqs. (15) and (16) reflective of the fact that adding a fermionic energy level to the partition function is similar to excluding a bosonic one and vice-versa.
The presented formulas for combining and resolving auxiliary partition functions allows for their construction via different routes which we have utilized to obtain exact expressions for the decomposition of correlations in terms of single-level occupation numbers. These different forms may also have value in overcoming the known numerical instabilities of the recursive formula for the fermionic partition function due to influence alternating signs Schönhammer (2017). In addition, the simplicity of the presented theory suggests a possible generalization to cover different energy-levels occupation-constraints beyond the fermionic and bosonic ones.
We envision the results presented herein could have applications in the computation of entanglement entropy in the presence of super-selection rules, as well as in modelling cold atom experiments. In the context of quantum information, the spectrum of the reduced density matrix corresponding to a mode bipartition of a state of conserved number $N$ of itinerant particles on a lattice can be associated with that of a fictional entanglement Hamiltonian. For non-interacting particles, the entanglement entropy can be obtained via the so-called correlation matrix method Peschel (2003); Peschel and Eisler (2009); Eisler and Peschel (2017); Peschel (2012) which requires the evaluation of the canonical partition function of the resulting non-interacting entanglement Hamiltonian. For trapped ultra-cold atoms at low densities where $N$ is fixed and interactions can be neglected, the analysis of experimental results in the physically correct canonical ensemble provides improved thermometry, especially for the case of fermions.
Finally, the ability to directly study level statistics in the canonical ensemble for bosons and fermions may have pedagogical value in the teaching of statistical mechanics, where the more physical concept of a fixed number of particles is quickly jettisoned and replaced with a grand canonical reservoir for the sake of simplifying derivations.
Acknowledgements.We thank D. Clougerty for bringing our attention to Ref. [Denton et al., 1973] and K. Schönhammer for discussions at an early stage of this work. This research was supported in part by the National Science Foundation (NSF) under awards DMR-1553991 (A.D.) and DMR-1828489 (H.B.). All computations were performed on the Vermont Advanced Computing Core supported in part by NSF award No. OAC-1827314.
Appendix A The derivation of Eq. (69)
Starting with the sets of levels $\mathcal{S}_{r}=\{i_{1},i_{2},\dots,i_{r}\}\subset\mathcal{S}_{\ell}=\{j_{1},j%
_{2},\dots,j_{\ell}\}$ and using Eq. (19) we have
$$Z_{\!{\scriptscriptstyle B}{\!,N-r}}^{\cup{\mathcal{S}_{r}}}=\sum_{k=0}^{\ell-%
r}(-1)^{k}Z_{\!{\scriptscriptstyle F}{\!,k}}^{\setminus{\mathcal{S}_{r}}}(%
\mathcal{S}_{\ell})Z_{\!{\scriptscriptstyle B}{\!,N-r-k}}^{\cup{\mathcal{S}_{%
\ell}}}.$$
(86)
Next, we substitute for $Z_{\!{\scriptscriptstyle F}{\!,k}}^{\setminus{\mathcal{S}_{r}}}(\mathcal{S}_{%
\ell})$, using Eq. (63), except for the last term $(-1)^{\ell-r}Z_{\!{\scriptscriptstyle F}{\!,\ell-r}}^{\setminus{\mathcal{S}_{r%
}}}(\mathcal{S}_{\ell})Z_{\!{\scriptscriptstyle B}{\!,N-\ell}}^{\cup{\mathcal{%
S}_{\ell}}}$ which we separate from the rest of the previous summation, thus we obtain
$$\displaystyle Z_{\!{\scriptscriptstyle B}{\!,N-r}}^{\cup{\mathcal{S}_{r}}}=$$
$$\displaystyle\sum_{k=0}^{\ell-r-1}\!\!\sum_{m=0}^{k}(-1)^{m+k}Z_{\!{%
\scriptscriptstyle B}{\!,m}}(\mathcal{S}_{r})Z_{\!{\scriptscriptstyle F}{\!,k-%
m}}(\mathcal{S}_{\ell})Z_{\!{\scriptscriptstyle B}{\!,N-r-k}}^{\cup{\mathcal{S%
}_{\ell}}}$$
$$\displaystyle+(-1)^{\ell-r}Z_{\!{\scriptscriptstyle F}{\!,\ell-r}}^{\setminus{%
\mathcal{S}_{r}}}(\mathcal{S}_{\ell})Z_{\!{\scriptscriptstyle B}{\!,N-\ell}}^{%
\cup{\mathcal{S}_{\ell}}}.$$
(87)
If we rearrange the summations and perform the indexes change $k\to k-r+1$ and $m\to m-r$, we get
$$\displaystyle Z_{\!{\scriptscriptstyle B}{\!,N-r}}^{\cup{\mathcal{S}_{r}}}\!\!=$$
$$\displaystyle\sum_{m=r}^{\ell-1}(-1)^{m+1}Z_{\!{\scriptscriptstyle B}{\!,m-r}}%
(\mathcal{S}_{r})$$
$$\displaystyle\times\sum_{k=m-1}^{\ell-2}\!\!\!\!(-1)^{k}Z_{\!{%
\scriptscriptstyle F}{\!,k-m+1}}(\mathcal{S}_{\ell})Z_{\!{\scriptscriptstyle B%
}{\!,N-k-1}}^{\cup{\mathcal{S}_{\ell}}}$$
$$\displaystyle+(-1)^{\ell-r}Z_{\!{\scriptscriptstyle F}{\!,\ell-r}}^{\setminus{%
\mathcal{S}_{r}}}(\mathcal{S}_{\ell})Z_{\!{\scriptscriptstyle B}{\!,N-\ell}}^{%
\cup{\mathcal{S}_{\ell}}}.$$
(88)
Now, using Eq. (34), we substitute for $Z_{\!{\scriptscriptstyle B}{\!,N-\ell}}^{\cup{\mathcal{S}_{\ell}}}$ and $Z_{\!{\scriptscriptstyle B}{\!,N-r}}^{\cup{\mathcal{S}_{r}}}$ as well as the APF $Z_{\!{\scriptscriptstyle F}{\!,\ell-r}}^{\setminus{\mathcal{S}_{r}}}(\mathcal{%
S}_{\ell})=\mathrm{e}^{-\beta\sum_{j_{\nu}\in\mathcal{S}_{\ell}\setminus%
\mathcal{S}_{r}}\epsilon_{j_{\nu}}}$. After multiplying the resulting equation by $\frac{(-1)^{r-1}\mathrm{e}^{-\beta\sum_{i_{\nu}\in\mathcal{S}_{r}}\epsilon_{i_%
{\nu}}}}{Z_{\!{\scriptscriptstyle B}{\!,N}}}$, we can write
$$Y_{\!{\scriptscriptstyle B}{\!,0}}+\sum_{m=r}^{\ell-1}A_{B,m}(\mathcal{S}_{r})%
Y_{\!{\scriptscriptstyle B}{\!,m}}=b_{B}(\mathcal{S}_{r}).$$
(89)
where
$$\displaystyle Y_{\!{\scriptscriptstyle B}{\!,1\leq m\leq\ell-1}}=\frac{(-1)^{m%
+1}}{Z_{\!{\scriptscriptstyle B}{\!,N}}}\!\!\!\!\sum_{k=m-1}^{\ell-2}\!\!\!\!(%
-1)^{k}Z_{\!{\scriptscriptstyle F}{\!,k-m+1}}(\mathcal{S}_{\ell})Z_{\!{%
\scriptscriptstyle B}{\!,N-k-1}}^{\cup{\mathcal{S}_{\ell}}},$$
(90)
$Y_{\!{\scriptscriptstyle B}{\!,0}}=(-1)^{\ell-1}\expectationvalue{n_{j_{1}}n_{%
j_{2}}\dots n_{j_{\ell}}}_{\!{\scriptscriptstyle B}{\!,N}}$, $A_{0}=1$, $A_{0<m<r}=0$ and $A_{r\leq m\leq\ell-1}(\mathcal{S}_{r})=(-1)^{r-1}\mathrm{e}^{-\beta\sum_{i_{%
\nu}\in\mathcal{S}_{r}}\epsilon_{i_{\nu}}}Z_{\!{\scriptscriptstyle B}{\!,m-r}}%
(\mathcal{S}_{r})$. Also, the term $b_{B}(\mathcal{S}_{r})=(-1)^{r-1}\expectationvalue{n_{i_{1}}n_{i_{2}}\dots n_{%
i_{r}}}_{\!{\scriptscriptstyle B}{\!,N}}$.
References
Kittel (1980)
C. Kittel, Thermal Physics, 2nd ed. (W.H. Freeman, San Francisco, 1980) pp. 121–225.
Landau and Lifshitz (1980)
L. D. Landau and E. M. Lifshitz, Statistical
Physics, 3rd ed., Part 1 (Elsevier, Burlington, MA, 1980) pp. 158–183.
Pathria and Beale (2011)
R. K. Pathria and P. D. Beale, Statistical
Mechanics, 3rd ed. (Elsevier, Boston, 2011) p. 148.
Bedingham (2003)
D. J. Bedingham, “Bose-Einstein condensation in the canonical ensemble,” Phys. Rev. D 68, 105007 (2003).
Mullin and Fernández (2003)
W. J. Mullin and J. P. Fernández, “Bose-Einstein condensation, fluctuations, and recurrence relations in
statistical mechanics,” Am. J. Phys. 71, 661 (2003).
Wenz et al. (2013)
A. N. Wenz, G. Zürn,
S. Murmann, I. Brouzos, T. Lompe, and S. Jochim, “From few to many: Observing the formation of a Fermi sea
one atom at a time,” Science 342, 457 (2013).
Parsons et al. (2015)
M. F. Parsons, F. Huber,
A. Mazurenko, C. S. Chiu, W. Setiawan, K. Wooley-Brown, S. Blatt, and M. Greiner, ‘‘Site-resolved imaging of fermionic ${}^{6}\mathrm{Li}$ in an
optical lattice,” Phys. Rev. Lett. 114, 213002 (2015).
Cheuk et al. (2015)
L. W. Cheuk, M. A. Nichols,
M. Okan, T. Gersdorf, V. V. Ramasesh, W. S. Bakr, T. Lompe, and M. W. Zwierlein, “Quantum-gas microscope for fermionic atoms,” Phys. Rev. Lett. 114, 193001 (2015).
Haller et al. (2015)
E. Haller, J. Hudson,
A. Kelly, D. A. Cotta, B. Peaudecerf, G. D. Bruce, and S. Kuhr, “Single-atom imaging of fermions in a quantum-gas
microscope,” Nat. Phys. 11, 738 (2015).
Pegahan et al. (2019)
S. Pegahan, J. Kangara,
I. Arakelyan, and J. E. Thomas, “Spin-energy correlation in degenerate
weakly interacting Fermi gases,” Phys.
Rev. A 99, 063620
(2019).
Mukherjee et al. (2017)
B. Mukherjee, Z. Yan,
P. B. Patel, Z. Hadzibabic, T. Yefsah, J. Struck, and M. W. Zwierlein, “Homogeneous atomic Fermi gases,” Phys. Rev. Lett. 118, 123401 (2017).
Hueck et al. (2018)
K. Hueck, N. Luick,
L. Sobirey, J. Siegl, T. Lompe, and H. Moritz, “Two-dimensional homogeneous Fermi gases,” Phys. Rev. Lett. 120, 060402 (2018).
Mukherjee et al. (2019)
B. Mukherjee, P. B. Patel, Z. Yan, R. J. Fletcher, J. Struck, and M. W. Zwierlein, “Spectral response and contact of the unitary
Fermi gas,” Phys. Rev. Lett. 122, 203402 (2019).
Onofrio (2016)
R. Onofrio, “Cooling and
thermometry of atomic Fermi gases,” Phys. Usp. 59, 1129
(2016).
Phelps et al. (2020)
G. A. Phelps, A. Hébert,
A. Krahn, S. Dickerson, F. Öztürk, S. Ebadi, L. Su, and M. Greiner, “Sub-second production of a quantum degenerate gas,” (2020), arXiv:2007.10807
.
Giorgini et al. (2008)
S. Giorgini, L. P. Pitaevskii, and S. Stringari, “Theory of
ultracold atomic Fermi gases,” Rev.
Mod. Phys. 80, 1215
(2008).
Horodecki et al. (2000)
M. Horodecki, P. Horodecki, and R. Horodecki, “Limits for
entanglement measures,” Phys. Rev. Lett. 84, 2014 (2000).
Bartlett and Wiseman (2003)
S. D. Bartlett and H. M. Wiseman, “Entanglement
constrained by superselection rules,” Phys. Rev. Lett. 91, 097903 (2003).
Wiseman and Vaccaro (2003)
H. M. Wiseman and J. A. Vaccaro, “Entanglement of
indistinguishable particles shared between two parties,” Phys. Rev. Lett. 91, 097902 (2003).
Wiseman et al. (World Scientific, Singapore,
2004)
H. M. Wiseman, S. D. Bartlett, and J. A. Vaccaro, “Ferreting out the fluffy bunnies:
Entanglement constrained by generalized superselection rules,” in Laser
Spectroscopy (World Scientific, Singapore, 2004) pp. 307–314.
Vaccaro et al. (2003)
J. A. Vaccaro, F. Anselmi, and H. M. Wiseman, “Entanglement of identical
particles and reference phase uncertainty,” Int.
J. Quantum. Inform. 01, 427 (2003).
Schuch et al. (2004)
N. Schuch, F. Verstraete,
and J. I. Cirac, “Nonlocal resources in the
presence of superselection rules,” Phys. Rev. Lett. 92, 087904 (2004).
Dunningham et al. (2005)
J. Dunningham, A. Rau, and K. Burnett, “From pedigree cats to
fluffy-bunnies,” Science 307, 872 (2005).
Cramer et al. (2011)
M. Cramer, M. B. Plenio,
and H. Wunderlich, “Measuring entanglement in
condensed matter systems,” Phys. Rev. Lett. 106, 020401 (2011).
Klich and Levitov (2008)
I. Klich and L. S. Levitov, “Scaling of
entanglement entropy and superselection rules,” (2008), arXiv:0812.0006
.
Murciano et al. (2020a)
S. Murciano, P. Ruggiero,
and P. Calabrese, “Symmetry resolved entanglement in
two-dimensional systems via dimensional reduction,” (2020a), arXiv:2003.11453 .
Tan and Ryu (2020)
M. T. Tan and S. Ryu, “Particle number
fluctuations, Rényi entropy, and symmetry-resolved entanglement entropy
in a two-dimensional Fermi gas from multidimensional bosonization,” Phys. Rev. B 101, 235169 (2020).
Murciano et al. (2020b)
S. Murciano, G. D. Giulio, and P. Calabrese, “Entanglement and
symmetry resolution in two dimensional free quantum field theories,”
(2020b), arXiv:2006.09069 .
Capizzi et al. (2020)
L. Capizzi, P. Ruggiero, and P. Calabrese, “Symmetry resolved entanglement
entropy of excited states in a CFT,” (2020), arXiv:2003.04670 .
Fraenkel and Goldstein (2020)
S. Fraenkel and M. Goldstein, “Symmetry
resolved entanglement: exact results in 1d and beyond,” J.
Stat. Mech.: Theor. Exp 2020, 033106 (2020).
Feldman and Goldstein (2019)
N. Feldman and M. Goldstein, “Dynamics of
charge-resolved entanglement after a local quench,” Phys. Rev. B 100, 235146 (2019).
Barghathi et al. (2019)
H. Barghathi, E. Casiano-Diaz, and A. Del Maestro, “Operationally accessible entanglement of one-dimensional spinless
fermions,” Phys. Rev. A 100, 022324 (2019).
Bonsignori et al. (2019)
R. Bonsignori, P. Ruggiero, and P. Calabrese, “Symmetry
resolved entanglement in free fermionic systems,” J.
Phys. A: Math. Theor. 52, 475302 (2019).
Barghathi et al. (2018)
H. Barghathi, C. M. Herdman, and A. Del Maestro, “Rényi
generalization of the accessible entanglement entropy,” Phys. Rev. Lett. 121, 150501 (2018).
Kiefer-Emmanouilidis et al. (2020)
M. Kiefer-Emmanouilidis, R. Unanyan, J. Sirker, and M. Fleischhauer, “Bounds on the entanglement
entropy by the number entropy in non-interacting fermionic systems,” SciPost Phys. 8, 83 (2020).
Murciano et al. (2020c)
S. Murciano, G. D. Giulio, and P. Calabrese, “Symmetry
resolved entanglement in gapped integrable systems: a corner transfer matrix
approach,” SciPost Phys. 8, 46 (2020c).
Goldstein and Sela (2018)
M. Goldstein and E. Sela, “Symmetry-resolved
entanglement in many-body systems,” Phys. Rev. Lett. 120, 200602 (2018).
Schmidt (1989)
H. Schmidt, “A simple
derivation of distribution functions for Bose and Fermi statistics,” Am. J. Phys. 57, 1150
(1989).
Borrmann and Franke (1993)
P. Borrmann and G. Franke, “Recursion
formulas for quantum statistical partition functions,” J. Chem.
Phys. 98, 2484 (1993).
Borrmann et al. (1999)
P. Borrmann, J. Harting,
O. Mülken, and E. R. Hilf, “Calculation of thermodynamic properties
of finite Bose-Einstein systems,” Phys.
Rev. A 60, 1519
(1999).
Pratt (2000)
S. Pratt, “Canonical and
microcanonical calculations for Fermi systems,” Phys. Rev. Lett. 84, 4255 (2000).
Weiss and Wilkens (1997)
C. Weiss and M. Wilkens, “Particle number
counting statistics in ideal Bose gases,” Opt.
Express 1, 272 (1997).
Arnaud et al. (1999)
J. Arnaud, J. M. Boé,
L. Chusseau, and F. Philippe, “Illustration of the Fermi-Dirac
statistics,” Am. J. Phys. 67, 215 (1999).
Schönhammer (2017)
K. Schönhammer, “Deviations from Wick’s theorem in the canonical ensemble,” Phys. Rev. A 96, 012102 (2017).
Giraud et al. (2018)
O. Giraud, A. Grabsch, and C. Texier, ‘‘Correlations of occupation
numbers in the canonical ensemble and application to a Bose-Einstein
condensate in a one-dimensional harmonic trap,” Phys.
Rev. A 97, 053615
(2018).
Tsutsui and Kita (2016)
K. Tsutsui and T. Kita, “Quantum correlations of
ideal Bose and Fermi gases in the canonical ensemble,” J. Phys.
Soc. Jpn. 85, 114603
(2016).
Zhou and Dai (2018)
C.-C. Zhou and W.-S. Dai, “Canonical partition
functions: ideal quantum gases, interacting classical gases, and interacting
quantum gases,” J. Stat. Mech.: Theor. Exp 2018, 023105 (2018).
Denton et al. (1971)
R. Denton, B. Mühlschlegel, and D. J. Scalapino, “Electronic heat capacity and susceptibility of small metal particles,” Phys. Rev. Lett. 26, 707 (1971).
Denton et al. (1973)
R. Denton, B. Mühlschlegel, and D. J. Scalapino, “Thermodynamic properties of electrons in small metal particles,” Phys. Rev. B 7, 3589
(1973).
Schönhammer and Meden (1996)
K. Schönhammer and V. Meden, “Fermion-boson
transmutation and comparison of statistical ensembles in one dimension,” Am. J. Phys. 64, 1168
(1996).
Svidzinsky et al. (2018)
A. Svidzinsky, M. Kim,
G. Agarwal, and M. O. Scully, “Canonical ensemble ground state and
correlation entropy of Bose–Einstein condensate,” New J. Phys. 20, 013002 (2018).
Jha and Hirata (2020)
P. K. Jha and S. Hirata, “Finite-temperature many-body
perturbation theory in the canonical ensemble,” Phys. Rev. E 101, 022106 (2020).
Chase et al. (1999)
K. C. Chase, A. Z. Mekjian,
and L. Zamick, “Canonical and microcanonical
ensemble approaches to Bose-Einstein condensation: The thermodynamics of
particles in harmonic traps,” Eur. Phys. J. B 8, 281 (1999).
Schönhammer (2000)
K. Schönhammer, ‘‘Thermodynamics and occupation numbers of a Fermi gas in the canonical
ensemble,” Am. J. Phys. 68, 1032 (2000).
Kocharovsky and Kocharovsky (2010)
V. V. Kocharovsky and V. V. Kocharovsky, “Analytical
theory of mesoscopic Bose-Einstein condensation in an ideal gas,” Phys. Rev. A 81, 033615 (2010).
Wang and Ma (2009)
J.-h. Wang and Y.-l. Ma, “Thermodynamics and
finite-size scaling of homogeneous weakly interacting Bose gases within an
exact canonical statistics,” Phys. Rev. A 79, 033604 (2009).
Akkelin and Sinyukov (2016)
S. V. Akkelin and Y. M. Sinyukov, “Quantum
canonical ensemble and correlation femtoscopy at fixed multiplicities,” Phys. Rev. C 94, 014908 (2016).
Magnus et al. (2017)
W. Magnus, L. Lemmens, and F. Brosens, “Quantum canonical ensemble:
A projection operator approach,” Physica A 482, 1
(2017).
Grabsch et al. (2018)
A. Grabsch, S. N. Majumdar, G. Schehr, and C. Texier, “Fluctuations of observables
for free fermions in a harmonic trap at finite temperature,” SciPost Phys. 4, 14 (2018).
Kruk et al. (2020)
M. Kruk, M. Łebek, and K. Rzążewski, “Statistical properties of cold bosons in a ring
trap,” Phys. Rev. A 101, 023622 (2020).
Wick (1950)
G. C. Wick, “The Evaluation
of the Collision Matrix,” Phys. Rev. 80, 268 (1950).
(62)
Here we replace $Z_{\tmspace-{.1667em}{\scriptscriptstyle B}{\tmspace-{.1667em},x}}$
with $Z_{\tmspace-{.1667em}{\scriptscriptstyle B}{\tmspace-{.1667em},x}}^{\cup%
\quantity{{j}}}$ and
$Z_{\tmspace-{.1667em}{\scriptscriptstyle B}{\tmspace-{.1667em},x}}^{\setminus%
\{{j}\}}$
with $Z_{\tmspace-{.1667em}{\scriptscriptstyle B}{\tmspace-{.1667em},x}}$ in (16) and
(14).
rep (2020)
(2020), Code, scripts
and data for this work are posted on a GitHub repository
https://github.com/DelMaestroGroup/CanonicalEnsembleTheory-paper-code
.
Peschel (2003)
I. Peschel, “Calculation of
reduced density matrices from correlation functions,” J.
Phys. A: Math. Gen. 36, L205 (2003).
Peschel and Eisler (2009)
I. Peschel and V. Eisler, “Reduced density
matrices and entanglement entropy in free lattice models,” J. Phys. A: Math. Theor. 42, 504003 (2009).
Eisler and Peschel (2017)
V. Eisler and I. Peschel, “Analytical
results for the entanglement hamiltonian of a free-fermion chain,” J. Phys. A: Math. Theor. 50, 284003 (2017).
Peschel (2012)
I. Peschel, “Special review:
Entanglement in solvable many-particle models,” Braz.
J. Phys 42, 267
(2012). |
Collisions of Main-Sequence Stars and the
Formation of Blue Stragglers in Globular Clusters
James C. Lombardi, Jr.11affiliation: Center for Radiophysics and Space Research, Cornell
University,
Ithaca, NY 14853. Email: [email protected]. 22affiliation: Department of Astronomy, Cornell University. ,
Frederic A. Rasio33affiliation: Department of Physics, MIT 6-201, Cambridge, MA 02139.
Email: [email protected]. and
Stuart L. Shapiro44affiliation: Center for Astrophysics & Relativity,
326 Siena Drive, Ithaca, NY 14850
Abstract
We report the results of new SPH calculations of parabolic collisions
between two main-sequence stars in a globular cluster. Such
collisions are directly relevant to the formation of blue stragglers.
In particular, we consider parent stars of mass $M/M_{TO}=0.2,0.5,0.75$,
and $1$, where $M_{TO}$ is the cluster turnoff mass (typically about
$0.8\,M_{\odot}$). Low-mass stars (with $M=0.2M_{TO}$ or $0.5M_{TO}$) are
assumed to be fully convective and are therefore modeled as
$n=1.5$ polytropes. Stars at the turnoff (with $M=M_{TO}$) are
assumed to be mostly radiative and are modeled as $n=3$ polytropes.
Intermediate-mass stars (with $M=0.75M_{TO}$) are modeled as composite
polytropes consisting
of a radiative core with polytropic index $n=3$ and a convective envelope with
$n=1.5$.
We focus our study on the question of hydrodynamic mixing of helium
and hydrogen, which plays a crucial role in determining the observable
characteristics of blue stragglers. In all cases we find that
there is negligible hydrodynamic mixing of helium into the outer
envelope of the merger remnant. The amount of hydrogen mixed into the core of
the merger depends
strongly on the entropy profiles of the two colliding stars. For two
stars with nearly equal masses (and hence entropy profiles) very little
hydrodynamic mixing occurs at all, especially if they are close to the
turnoff point.
This is because the hydrogen-rich material from
both stars maintains, on average, a higher specific entropy than the
helium-rich material. If the two parent stars are close to turnoff,
very little hydrogen is present at the center of the merger remnant
and the main-sequence lifetime of the blue straggler could be very short.
In contrast, during a collision between two stars of
sufficiently different masses (mass ratio $q\lesssim 0.5$),
the hydrogen-rich material originally in the smaller star maintains,
on average, a lower specific
entropy than that of the more massive star and therefore settles
preferentially near the center of the merger remnant.
Through this process, moderately massive blue stragglers
(with masses $M_{TO}\lesssim M_{BS}\lesssim 1.5M_{TO}$) can obtain a significant supply
of fresh hydrogen fuel, thereby extending their main-sequence lifetime.
Based on our results we conclude that, in contrast to what has been
done in previous studies, blue stragglers formed by direct stellar
collisions should not necessarily be assumed to have initially homogeneous
composition profiles. However, we also demonstrate that the final merged
configurations produced by our hydrodynamic calculations, although very close
to hydrostatic equilibrium,
are usually far from thermal equilibrium. Therefore, it is possible
that convective or rotationally-induced mixing could occur on a thermal
timescale, as the merger remnant recontracts to the main-sequence.
celestial mechanics, stellar dynamics – globular clusters:
general – hydrodynamics – stars: blue stragglers – stars: evolution
– stars: interiors – stars: rotation
1 Introduction and Motivation
Blue stragglers are stars that appear along an extension of the main-sequence
(hereafter MS), beyond the turnoff point in the color-magnitude diagram of
a cluster.
It is generally believed that they are more massive objects (mass
$M_{BS}>M_{TO}$)
formed by the merger of two MS stars (each of mass $<M_{TO}$). Merging
can occur through a physical collision, or following the coalescence of the
two components of a close binary system (Leonard 1989; Livio 1993; Stryker
1993;
Bailyn & Pinsonneault 1995). Clear evidence for
binary coalescence has been found in the form of contact binaries among
blue stragglers in the low-density globular clusters NGC 5466 (Mateo et
al. 1990) and M71 (Yan & Mateo 1994), as well as in open clusters
(Kalużny & Ruciński 1993; Milone & Latham 1994; Jahn, Kalużny
& Ruciński 1995). Evidence for stellar collisions
comes from recent detections by HST of large numbers of blue stragglers
concentrated in the cores of some of the densest clusters,
such as M15 (De Marchi & Paresce 1994; Guhathakurta et al. 1995)
and M30 (Yanny et al. 1994), and from the apparent
lack of binaries in such dense systems (Shara et al. 1995).
Collisions can happen directly between two single stars only in the cores
of the densest clusters, but even in somewhat lower-density clusters they
can also happen indirectly, during resonant interactions involving primordial
binaries (Sigurdsson, Davies, & Bolte 1994; Sigurdsson & Phinney 1995;
Davies & Benz 1995). Observational evidence for the existence of primordial
binaries in globular clusters is now well established (Hut et al. 1992;
Cote et al. 1994).
Benz & Hills (1987) performed the first
three-dimensional calculations of direct collisions between two MS
stars. An important conclusion of their pioneering study was that stellar
collisions could lead to thorough mixing of the fluid.
The mixing of fresh hydrogen fuel into the core
of the merger would reset the nuclear clock of the blue straggler,
allowing it to remain on the MS for up to $\sim 10^{9}\,$yr after its
formation.
In subsequent work it was generally assumed that blue stragglers
resulting from stellar collisions were nearly homogeneous, therefore starting
their life close to the zero-age MS, but with an
anomalously high helium abundance coming from the (evolved) parent stars.
In contrast, little hydrodynamic mixing would be expected to occur during the
much gentler process of binary coalescence, which could take place on a
stellar evolution time scale rather than on a dynamical time scale
(Mateo et al. 1990; Bailyn 1992; but see Rasio & Shapiro 1995, and Rasio
1995).
On the basis of these ideas, Bailyn (1992) suggested a way of distinguishing
observationally between the two possible formation processes. The helium
abundance in the envelope of a blue straggler, which reflects the degree
of mixing during its formation process, can affect its observed position in
a color-magnitude diagram. Blue stragglers made from collisions
would have a higher helium abundance in their outer layers than those made
from binary mergers, and this would generally make them appear somewhat
brighter
and bluer. The analysis was carried out by Bailyn & Pinsonneault (1995)
who performed detailed stellar evolution calculations for blue stragglers
assuming various initial profiles. To represent the collisional case, they
again assumed chemically homogeneous initial
profiles with enhanced helium abundances, calculating the total helium mass
from
the age of the cluster and the masses of the parent stars.
In this paper we re-examine the question of mixing in stellar collisions.
We improve on the previous work of Benz & Hills (1987)
by adopting more realistic stellar models,
and by performing numerical calculations with increased spatial resolution.
We use the smoothed particle hydrodynamics (SPH) method, with $N=3\times 10^{4}$
particles for most of our calculations (Benz & Hills 1987 also used SPH, but
with
only $N=1024$ particles).
The colliding stars in our calculations are modeled as polytropes or
composite polytropes (Chandrasekhar 1939;
Rappaport, Verbunt, & Joss 1983; Ruciński 1988), and we adopt a
simple ideal gas equation of state. The polytropic index $n$
relates the pressure and density profiles in the star according to $P\propto\rho^{1+1/n}$. The adiabatic index $\Gamma_{1}={5\over 3}$ for an ideal
gas and we write the equation of state $P=A\rho^{\Gamma_{1}}$. Here
$A$ is a physical parameter related to the local specific entropy $s$
according to $A\propto\exp((\Gamma_{1}-1)s/k)$, where $k$ is Boltzmann’s
constant.
When $\Gamma_{1}\neq 1+1/n$ the quantity $A$, and hence the entropy $s$, has a
non-zero gradient.
Benz & Hills (1987) used $n=1.5$, $\Gamma_{1}={5\over 3}$ polytropic
models to represent MS stars. Unfortunately, such models
apply only to very low-mass MS stars with large convective envelopes.
For Population II MS stars, the effective polytropic index (defined in
terms of the degree of central mass concentration) is close to $n=1.5$
only for a mass $M\lesssim 0.4\,M_{\odot}$ (see Lai, Rasio, & Shapiro 1994,
Table 3). The object formed by a merger of two such low-mass stars would
hardly be recognizable as a blue straggler, since it would lie below,
or not far above, the MS turnoff point (typically $M_{TO}\simeq 0.8M_{\odot}$)
in a color-magnitude diagram.
Stars near the MS turnoff point have very shallow convective envelopes and are
much better described by $n=3$, $\Gamma_{1}={5\over 3}$ polytropes
(Eddington’s “standard model”, see, e.g., Clayton 1983).
These stars have a density profile much more centrally concentrated
than that of an $n=1.5$ polytropes, and this fact has important
consequences for the hydrodynamics of collisions.
Population II MS stars with masses in the intermediate range
$0.4\,M_{\odot}\lesssim M\lesssim 0.8\,M_{\odot}$ can be modeled by composite polytropes
with a polytropic index of $n=3$ for the radiative core and
$n=1.5$ for the convective envelope.
Stars close to the MS turnoff point in a cluster are the most relevant to
consider for stellar collision calculations, for two reasons.
First, as the cluster evolves via two-body relaxation,
the more massive MS stars will tend to concentrate
in the dense cluster core, where the collision rate is highest
(see, e.g., Spitzer 1987). Second, collision rates can be increased
dramatically
by the presence of a significant fraction of primordial binaries in the
cluster,
and the more massive MS stars will preferentially tend to be exchanged into
such a binary, or collide with another star, following a dynamical interaction
between two binaries or between a binary and a single star
(Sigurdsson & Phinney 1995).
Lai, Rasio & Shapiro (1993) have calculated collisions between MS stars
modeled by $n=3$, $\Gamma_{1}=5/3$ polytropes, but they focused on
high-velocity (hyperbolic) collisions more relevant to galactic nuclei
than to globular clusters.
The velocity dispersion of globular cluster stars is typically
$\sim 10{\,{\rm km}\,{\rm s}^{-1}}$, which is much smaller than the escape velocity from the stellar
surface. For example, a star of mass $M=0.8\,M_{\odot}$
and radius $R=R_{\odot}$ has an escape velocity $(2GM/R)^{1/2}=552{\,{\rm km}\,{\rm s}^{-1}}$. For this reason, we consider only parabolic collisions in this
paper, i.e., all initial trajectories are assumed to have zero orbital energy.
Our paper is organized as follows. In §2 we describe our
implementation of the SPH method and the numerical setup of our
calculations. In §3, we present the models used for MS stars,
detailing their assumed structure and chemical composition profiles.
We also describe the initial ($t=0$) configuration of the trajectory.
Our results are presented in §4. After describing the results for two
typical collisions in detail, we then characterize the rotation states
and the final profiles of all our merger remnants. We also present a
general method, which does not depend on our particular choice of
initial chemical composition profiles, for calculating the final
profile of any passively advected quantity in the merger remnant. We
conclude our results with an analysis of the numerical accuracy of our
simulations. Finally, in §5, we discuss the astrophysical
implications of our results as well as directions for future work.
2 Numerical Method and Conventions
2.1 The SPH Code
Our numerical calculations are done using the smoothed particle
hydrodynamics (SPH) method (see Monaghan 1992 for a recent review). We
used a modified version of the code developed by Rasio (1991)
specifically for the study of stellar interactions (see Rasio &
Shapiro 1995 and references therein). Since SPH is a Lagrangian method,
in which particles are used to represent fluid elements,
it is ideally suited for the study
of hydrodynamic mixing. Indeed, with the assumption that the gas remains
fully ionized throughout the dynamical evolution,
chemical abundances are passively advected quantities.
Therefore, the chemical
composition in the final fluid configuration can be determined after
the completion of a calculation simply by noting the original and final
positions of all SPH particles and by assigning particle abundances
according to an initial profile.
Associated with each SPH particle
$i$ is its position ${\bf r}_{i}$, velocity ${\bf v}_{i}$, mass $m_{i}$ and a
purely numerical “smoothing length” $h_{i}$ specifying the local
spatial resolution. An estimate of the fluid density at ${\bf r}_{i}$ is
calculated from the masses, positions, and smoothing lengths of
neighboring particles as a local weighted average,
$$\rho_{i}=\sum_{j}m_{j}W_{ij},$$
(1)
where the symmetric weights $W_{ij}=W_{ji}$ are calculated from the
method of Hernquist and Katz (1989), as
$$W_{ij}={1\over 2}\left[W(|{\bf r}_{i}-{\bf r}_{j}|,h_{i})+W(|{\bf r}_{i}-{\bf r%
}_{j}|,h_{j})\right].$$
(2)
Here $W(r,h)$ is an interpolation kernel, for which we use the second-order
accurate form of Monaghan and Lattanzio (1985),
$$W(r,h)={1\over\pi h^{3}}\cases{1-{3\over 2}\left({r\over h}\right)^{2}+{3\over
4%
}\left({r\over h}\right)^{3},&$0\leq{r\over h}<1$,\cr{1\over 4}\left[2-\left({%
r\over h}\right)\right]^{3},&$1\leq{r\over h}<2$,\cr 0,&${r\over h}\geq 2$.\cr}$$
(3)
In addition to passively advected scalar quantities (such as the hydrogen and
helium mass fractions $X_{i}$ and $Y_{i}$)
each particle $i$ also carries the local entropy variable
$A_{i}$. The specific entropy $s_{i}$ at ${\bf r}_{i}$ is
related to $A_{i}$ by
$$s_{i}-s_{o}={k\over\Gamma_{1}-1}\ln A_{i},$$
(4)
where $k$ is Boltzmann’s constant and $s_{o}$ is a fiducial constant. Neglecting
radiation pressure, the
pressure at ${\bf r}_{i}$ is estimated as
$$p_{i}=A_{i}\,\rho_{i}^{\Gamma_{1}},$$
(5)
where $\Gamma_{1}={5\over 3}$ is the ratio of specific heats for a fully
ionized ideal gas. As a self-consistency test, we check that
throughout the dynamical evolution the vast majority of particles have
$p_{i}$ remaining much larger than the radiation pressure ${1\over 3}aT_{i}^{4}$, where $a$ is the radiation constant and $T_{i}$ is the local
temperature (approximated by assuming an ideal gas).
An SPH code must solve the equations of motion of a large number $N$ of
Lagrangian fluid particles. Particle positions are updated according
to
$$\dot{\bf r}_{i}={\bf v}_{i}.$$
(6)
The velocity of particle $i$ is updated according to
$$m_{i}\dot{\bf v}_{i}={\bf F}^{(Grav)}_{i}+{\bf F}^{(SPH)}_{i}$$
(7)
where ${\bf F}^{(Grav)}_{i}$ is the gravitational force calculated by a
particle-mesh convolution algorithm (Hockney and Eastwood 1988, Wells
et al. 1990) based on Fast Fourier Transforms (FFT) on a $128^{3}$
grid, and
$${\bf F}^{(SPH)}_{i}=-\sum_{j}m_{i}m_{j}\left[\left({p_{i}\over\rho_{i}^{2}}+{p%
_{j}\over\rho_{j}^{2}}\right)+\Pi_{ij}\right]\nabla_{i}W_{ij}.$$
(8)
Here $\Pi_{ij}$ is an artificial viscosity term, while the rest of
equation (8) represents one of many possible SPH-estimators for the
local pressure-gradient force $-m_{i}(\nabla p/\rho)_{i}$ (see, e.g., Monaghan
1985).
For the artificial viscosity, a symmetrized version of the form
proposed by Monaghan (1989) is adopted,
$$\Pi_{ij}={-\alpha\mu_{ij}c_{ij}+\beta\mu_{ij}^{2}\over\rho_{ij}},$$
(9)
where $\alpha$ and $\beta$ are constant parameters,
$c_{ij}=(c_{i}+c_{j})/2$, and
$$\mu_{ij}=\cases{{({\bf v}_{i}-{\bf v}_{j})\cdot({\bf r}_{i}-{\bf r}_{j})\over h%
_{ij}(|{\bf r}_{i}-{\bf r}_{j}|^{2}/h_{ij}^{2}+\eta^{2})},&when $({\bf v}_{i}-%
{\bf v}_{j})\cdot({\bf r}_{i}-{\bf r}_{j})<0$,\cr 0,&when $({\bf v}_{i}-{\bf v%
}_{j})\cdot({\bf r}_{i}-{\bf r}_{j})\geq 0$,\cr}$$
(10)
with $h_{ij}=(h_{i}+h_{j})/2$. We have used $\alpha=1$, $\beta=2$ and
$\eta^{2}=0.01$, which provides a good description of shocks (Monaghan
1989, Hernquist and Katz 1989).
To complete the evolution equations of the fluid, $A_{i}$ is evolved
according to a discretized version of the first law of thermodynamics:
$${dA_{i}\over dt}={\gamma-1\over 2\rho_{i}^{\gamma-1}}\,\sum_{j}m_{j}\,\Pi_{ij}%
\,\,({\bf v}_{i}-{\bf v}_{j})\cdot\nabla_{i}W_{ij}.$$
(11)
Equation (11) has the advantage that the total entropy is strictly
conserved in the absence of shocks (ie. when $\Pi_{ij}=0$), and the
disadvantage that the total energy is only approximately conserved
(Rasio 1991; Hernquist 1993). Both total energy and angular momentum
conservation are monitored throughout the integrations as a measure of
numerical accuracy, and these quantities are conserved typically at the
percent level.
The dynamical equations are integrated using a second-order explicit
leap-frog scheme. Such a low order scheme is appropriate because
pressure gradient forces are subject to numerical noise. We calculate
the timestep as $\Delta t=C_{N}\,{\rm Min}(\Delta t_{1},\Delta t_{2})$ where
$\Delta t_{1}={\rm Min}_{i}\,(h_{i}/\dot{\bf v}_{i})^{1/2}$, $\Delta t_{2}={\rm Min}_{i}(h_{i}/(c_{i}^{2}+v_{i}^{2})^{1/2})$ and the Courant number $C_{N}=0.8$.
Other details of our implementation, as well as a number of test-bed
calculations using our SPH code are presented in Lombardi, Rasio,
& Shapiro (1995b).
Twenty of the twenty-three calculations reported here employ
$N=3\times 10^{4}$ equal-mass particles, while the remaining three
calculations (cases U, V and W in Table 2 below) use $N=1.8\times 10^{4}$
equal-mass particles. Unequal-mass SPH particles, sometimes used to
allow for higher resolution in low density regions, tend to settle
spuriously to preferred regions in the gravitational potential due to
numerical discreteness effects. Therefore, calculations with
equal-mass particles are best suited for studying mixing. In all
cases, time-dependent, individual particle smoothing lengths $h_{i}$
insure that the spatial resolution remains acceptable throughout the
dynamical evolution and that each particle interacts with a constant
number of neighbors $N_{N}\simeq 64$. With these resources, the numerical
integration of the SPH equations typically takes about 2 CPU hours per
time unit (eq. [12]) on an IBM SP-2 supercomputer.
2.2 Choice of Units
Throughout this paper, numerical results are given in units where
$G=M_{TO}=R_{TO}=1$, where $G$ is Newton’s gravitational constant and
$M_{TO}$ and $R_{TO}$ are the mass and radius of a terminal-age MS
(TAMS) star at the cluster turnoff point. The units of time, velocity, and
density are then
$$\displaystyle t_{u}$$
$$\displaystyle=$$
$$\displaystyle\left({R_{TO}^{3}\over GM_{TO}}\right)^{1/2}=1782\,{\rm s}\times%
\left({M_{TO}\over 0.8M_{\odot}}\right)^{-1/2}\left({R_{TO}\over R_{\odot}}%
\right)^{3/2}$$
(12)
$$\displaystyle v_{u}$$
$$\displaystyle=$$
$$\displaystyle\left({GM_{TO}\over R_{TO}}\right)^{1/2}=391\,{\rm km}\,{\rm s}^{%
-1}\times\left({M_{TO}\over 0.8M_{\odot}}\right)^{1/2}\left({R_{TO}\over R_{%
\odot}}\right)^{-1/2}$$
(13)
$$\displaystyle\rho_{u}$$
$$\displaystyle=$$
$$\displaystyle{M_{TO}\over R_{TO}^{3}}=5.90\,{\rm g}\,{\rm cm}^{-3}\times\left(%
{M_{TO}\over 0.8M_{\odot}}\right)\left({R_{TO}\over R_{\odot}}\right)^{-3}.$$
(14)
Furthermore, the units of temperature and specific entropy are chosen to be
$$\displaystyle T_{u}$$
$$\displaystyle=$$
$$\displaystyle{GM_{TO}m_{H}\over kR_{TO}}=1.85\times 10^{7}\,{\rm K}\times\left%
({M_{TO}\over 0.8M_{\odot}}\right)\left({R_{TO}\over R_{\odot}}\right)^{-1}$$
(15)
$$\displaystyle s_{u}$$
$$\displaystyle=$$
$$\displaystyle{k\over M_{TO}}=8.68\times 10^{-50}\,{\rm erg}\,{\rm K}^{-1}\,{%
\rm g}^{-1}\times\left({M_{TO}\over 0.8M_{\odot}}\right)^{-1}$$
(16)
where $m_{H}$ is the mass of hydrogen and $k$ is Boltzmann’s constant.
2.3 Determination of the Bound Mass and Termination of the Calculation
The iterative procedure used to determine the total amount of
gravitationally bound mass $M$ of a merger remnant is the same as in
Rasio (1991). Namely, particles with negative specific
enthalpy with respect to the bound fluid’s center of mass are
considered bound. During all of the stellar collisions we considered,
only a small fraction (typically a few percent)
of the mass is ejected and becomes gravitationally unbound.
Some SPH particles, although bound, are ejected so
far away from the system center of mass that it would take many
dynamical times for them to rain back onto the central remnant and settle
into equilibrium. Rather than wait for those particles (which would
allow for more spurious diffusion in the central region,
see §4.5), we terminate the calculation once we are confident that
at least the inner $95\%$ of the mass has settled into equilibrium. We
confirm this by two stability tests. First, we check that the
specific entropy $s$ increases from the center to the surface of the
merger remnant, a sufficient (and necessary for non-rotating stars)
condition for convective stability (see the discussion surrounding
eq. [18]). For rotating merger remnants we also check another
dynamical stability criteria, namely that the specific angular momentum
increases from the poles to the equator along surfaces of constant
entropy (Tassoul 1978).
3 Initial Data
We consider parent MS stars of masses $M=0.2,0.5,0.75$ and $1M_{TO}$.
The stellar radii are taken from the results of evolution calculations
for Population II stars by D’Antona (1987).
Table 1 lists the values we adopt, as well as the radii enclosing
mass fractions $0.9$ and $0.95$.
For the $0.5M_{TO}$ star, we adopt the value $0.29R_{TO}$ for the radius
of the interface between the radiative and convective zones.
These values correspond to an interpolation of the
results of D’Antona (1987) for an age $t=15\,$Gyr.
The $M=0.2$ and $0.5M_{TO}$ stars are modeled as $n=1.5$ polytropes,
whereas the $M=M_{TO}$ stars are $n=3$ polytropes. The $0.75M_{TO}$
stars are modeled as composite polytropes consisting of a radiative
core with index $n=3$ and a convective envelope with $n=1.5$. Figure 1
shows the specific entropy profiles of these models. The convective
regions have constant specific entropy. Note that the
specific entropy in the $M=0.2$ and $0.5M_{TO}$ stars is
everywhere smaller than the minimum specific entropy in the two more
massive stars. This fact plays a central role in understanding the dynamics
of the merger involving either an $M=0.2$ or $0.5M_{TO}$ star with a
more massive star.
We have used the stellar evolution code developed by Sienkiewicz and
collaborators (cf. Sienkiewicz, Bahcall, & Paczyński 1990) to
compute the chemical composition profile in the radiative zones of the
$M=0.75$ and $1M_{TO}$ parent stellar models. We evolved MS stars of
total mass $M=0.6$ and $0.8\,M_{\odot}$, primordial helium abundance
$Y=0.25$, and metallicity $Z=0.001$ for a time $t\simeq 15\,$Gyr. This
brought the $M=0.8\,M_{\odot}$ star to the point of hydrogen exhaustion
at the center. In the convective regions of our $M=0.5,0.75$ and $1M_{TO}$ parent stars, we set a constant helium abundance $Y=0.25$. For
the $M=0.2M_{TO}$ star, we set $Y=0.24$ everywhere. Figure 2 shows
the resulting profiles, which are used to assign the helium abundance to all
the SPH particles in the calculations. The final column in Table 1
gives the total mass fraction of helium in each of the parent stars.
Although the composition profiles do not affect the hydrodynamics of
a collision in any way, they are needed to determine the chemical composition
profile of the merger remnant. In §4.4, we present a method for
applying our results to arbitrary initial composition profiles.
The stars are initially non-rotating and separated by at least 4 times
the radius of the larger star, which allows tidal effects to be
neglected in the initial configuration. The initial velocities are
calculated by approximating the stars as point masses on an orbit with
zero orbital energy and a pericenter separation $r_{p}$. The Cartesian
coordinate system is chosen such that these hypothetical point masses
of mass $M_{1}$ and $M_{2}$ would reach pericenter at positions
$x_{i}=(-1)^{i}(1-M_{i}/(M_{1}+M_{2}))r_{p}$, $y_{i}=z_{i}=0$, where $i=1,2$ and
$i=1$ refers to the more massive star. The orbital plane
is chosen to be $z=0$. With these choices, the center of mass
resides at the the origin.
4 Results
Table 2 lists the values of the most important initial parameters and
final results for all the calculations we performed.
The first column gives the label by which the calculation is
referred to in this paper. The second and third columns give the
masses $M_{1}$ and $M_{2}$ of the colliding stars, in units of
$M_{TO}\simeq 0.8\,M_{\odot}$. Column 4 gives the ratio $r_{p}/(R_{1}+R_{2})$,
where $r_{p}$ is the pericenter separation for the initial orbit and
$R_{1}+R_{2}$ is the sum of the two (unperturbed) stellar radii. This ratio
has the value $0$ for a head-on collision, and $1$ for a grazing encounter.
Note, however, that an encounter with $r_{p}/(R_{1}+R_{2})\gtrsim 1$ can still
lead to a direct collision in the outer envelopes of the two stars
because of the large tidal deformations near pericenter. We did not
attempt to perform any calculations for $r_{p}/(R_{1}+R_{2})>1$ here, for
reasons discussed in §5.
Column 5 gives the initial separation $r_{0}$ in units of $R_{TO}$.
Column 6 gives the final time $t_{f}$ at which the calculation was terminated,
in the unit of equation (12); see §2.3 for a discussion of how the values
of $t_{f}$ were obtained.
Column 7 gives the number $n_{p}$ of successive pericenter interactions that
the stars experience before merging. In general, $n_{p}$ increases with $r_{p}$,
and it is only for very nearly head-on collisions that the two stars merge
immediately after the first impact ($n_{p}=1$ in that case).
Column 8 gives the mass-loss fraction $1-M/(M_{1}+M_{2})$, where $M$
is the mass of the bound fluid in the final merged configuration.
Column 9 gives the ratio $T/|W|$
of rotational kinetic energy to gravitational binding energy of the (bound)
merger remnant in its center-of-mass frame at time $t_{f}$.
Columns 10 and 11 give the velocity
components $V_{x}$ and $V_{y}$ in the units of equation (13) for the merger
remnant’s center of mass at
time $t_{f}$ in the system’s center-of-mass frame.
Since the amount of mass ejected during a parabolic collision is
very small, the merger remnant
never acquires a large recoil velocity. The largest
value of 0.035 in our calculations occurs for case M
and corresponds to a physical speed of about
$14{\,{\rm km}\,{\rm s}^{-1}}$ (for $M_{TO}=0.8\,M_{\odot}$ and $R_{TO}=R_{\odot}$).
This may be large enough to eject the object from the cluster core,
but not to eject it from the entire cluster.
4.1 Discussion of the Results for Two Typical Cases
One of our calculations involving two TAMS stars (Case C) has already
been described by Lombardi, Rasio & Shapiro (1995a).
In this section we discuss in some detail the results of two other
representative cases (E and G).
Figure 3 illustrates the dynamical evolution for Case E:
a TAMS star ($M_{1}=M_{TO}$) collides with a slightly less massive star
($M_{2}=0.75\,M_{TO}$). The initial separation is $r_{0}=5R_{TO}$ and the
parabolic trajectory has a pericenter separation $r_{p}=0.25(R_{1}+R_{2})$.
The first collision at time $t\simeq 4$ disrupts the outer layers
of the two stars,
but leaves their inner cores essentially undisturbed. The two
components withdraw to apocenter at $t\simeq 7$, and by $t\simeq 10$
are colliding for the second, and final, time ($n_{p}=2$).
The merger remnant undergoes some large-amplitude oscillations which
damp away quickly due to shock dissipation.
The final ($t=41$) equilibrium configuration (see
Figure 4) is an axisymmetric, rapidly rotating object ($T/|W|=0.07$).
Figure 5 shows SPH-particle values of the angular velocity $\Omega$ as
a function of radius $r$ in the equatorial plane. We see clearly that
the large envelope of the merger remnant is differentially rotating. The
uniformly rotating core contains only about 15% of the mass.
The angular velocity drops to half its central value
near $r=1.1R_{TO}$, and 80% of mass is enclosed
within the isodensity surface with this equatorial radius. Only about
2% of the total mass is ejected during this collision, and the ejection
is nearly isotropic. As a result, the final recoil
velocity of the merger remnant in the orbital plane is only about $0.007$.
Figure 6 displays the thermal energy $U$, kinetic energy $T$,
gravitational potential energy $W$ and total energy $E=U+T+W$ as a
function of time $t$ for case E. The total energy is conserved to within
$2\%$. Dips in the potential energy $W$ correspond either to
a collision of the two components before final merging or to a maximum
contraction during the subsequent oscillations of the merger remnant.
The criterion we use to distinguish
collisions (which should be included in the number of interactions
$n_{p}$ before the stars merge) from oscillations is that the first local
maximum of $W$ which is lower than the previous local maximum occurs
immediately after the final merging.
The idea behind this criterion is that a collision without merger
ultimately tends to increase the system’s gravitational potential
energy, whereas a merger will decrease the potential energy. For
example, in Figure 6, the local maximum of $W$ at $t\simeq 11$ is
lower than the one at $t\simeq 6$, so that the dips in $W$ at $t\simeq 4$ and $11$ account for the number $n_{p}=2$ of interactions given in Table 2 for
case E. The remaining dips at $t\simeq 12$ and $15$ correspond to the
peak contraction of the merger remnant during oscillations. The value
$n_{p}=2$ obtained here in this way agrees with what one gets
simply by direct visual inspection of the system at various times.
In some cases, however, visual inspection can be
subjective since it is often difficult to recognize two
components connected by a bridge of high-density material
just prior to final merging.
Figure 7 illustrates the dynamical evolution for case G, which involves
a TAMS star ($M_{1}=M_{TO}$) and a low-mass MS star with
$M_{2}=0.5\,M_{TO}$ on a head-on parabolic trajectory with initial
separation $r_{0}=5$. The initial collision occurs at time $t\simeq 4$,
and the stars never separate again. The resulting isodensity surfaces of
the final equilibrium configuration are essentially spherically
symmetric (Figure 8). About 6% of the total mass becomes
gravitationally unbound following the collision, and it is ejected
preferentially in the $+x$-direction. Of this ejected material, 95%
originated in the more massive ($M=M_{TO}$) star.
Figures 9(a) and (b) show the entropy profiles for the final
configurations in cases E and G. Except over the outer few percent of
the mass, where equilibrium has not yet been reached (see §2.3), the specific entropy $s$ is an increasing function of the
interior mass fraction $m/M$. Here $m$ is
the mass inside an isodensity surface, and $M$ is the total bound mass
of the merger remnant. The scatter of the points in Figure 9 is real, since
isodensity surfaces and surfaces of constant entropy do not coincide. The
especially small scatter in Figure 9(b)
demonstrates that the entropy does tend towards spherical symmetry in
non-rotating merger remnants, despite the strong angular dependence of
the shock-heating due to the geometry of the collision.
Even though the density and entropy profiles of both the merger remnant
and parent stars are spherically symmetric in case G, this does not
imply that the chemical composition must also share this symmetry.
Indeed, the effects of anisotropic shock-heating are always evident in
the final spatial distribution of the chemical composition.
On a constant-entropy surface in the final configuration,
particles which have been
shock heated the most necessarily had the lowest entropy prior to the
collision. Since lower entropy material generally has higher helium
abundance (see Figures 1 and 2), shock-heated regions tend to have
higher helium abundances. Generally, fluid elements which reside in the
orbital plane, and especially those which lie along the collision axis in
$r_{p}=0$ cases, are
shielded the least from the shock.
Figure 10
displays the angular distribution of the helium
abundance for the merger remnants of cases E and G, near the interior
mass fractions $m/M=0.25,0.5$ and $0.75$.
The helium abundance
$Y$ peaks in Figure 10(b) and (d) at the polar angle $\theta={\pi\over 2}$ (the equatorial plane), as well as in Figure 10(c) at $\phi=0$ (the
collision axis).
In the $r_{p}\neq 0$ cases, shear in the
differentially rotating merger remnant tends to make the profiles
axisymmetric (see Fig. 10(a)). However, no dynamical motions exist
to circulate the fluid along the meridional directions, and
consequently, on an isodensity surface, the fractional helium abundance
increases from the poles to the equators for both rotating and non-rotating
merger
remnants (cf. Fig. 10(b and d)). Meridional circulation will smooth out these
deviations from compositional spherical symmetry over a timescale much longer
than that treatable by our purely dynamical
code (see related discussion in §5). As a
practical concern, we note that stellar evolution codes, which can use our
merger remnants as initial data, usually assume
spherical symmetry. For these reasons, we often average out the
angular dependence when presenting composition, and other, profiles.
Figures 11(a) and (b) show the helium mass fraction $Y$ as a function of
the interior mass fraction $m/M$ for the final merged configuration in
cases E and G, respectively. The points correspond to the final SPH
particle values, with the long-dashed curve representing their
average. The spread in the points is due to the mixing of the fluid as
well as the fact that the final profiles are not spherically
symmetric. Only a small amount of the observed mixing is due to the
spurious diffusion of SPH particles, i.e., diffusion which is purely a
numerical artifact of the SPH scheme (see §4.5). In case E
(Figure 11(a)), there is a small amount of hydrogen in the core, with the
innermost 1% of the mass being 85% helium and the inner 25% being
60% helium. For both cases E and G, it is immediately apparent that
the helium enrichment in the outer layers is minimal since the fractional
helium abundance is just barely above $Y=0.25$, the value in the outer
layers of the parent stars.
The horizontal line at the bottom of Figure 11(b) corresponds to the
particles in case G which originated in the less massive parent star, star 2,
all of which have a helium abundance $Y=0.25$. Although these
particles are spread over the entire range $0<m/M<1$ in the merger
remnant, they are found preferentially near the center. Of all the
particles which originated in star 2, 69% ultimately end up with
$m/M<0.25$, while only 6% end in the range $0.75<m/M<1$. Essentially,
the entire star 2 has sunk to the center of the merger remnant,
displacing the material in
star 1 and leaving only a small amount of shock heated gas in the
remnant’s outer envelope. Consequently, the hydrogen enrichment in the core
is quite pronounced; all of the innermost 22% of the mass originated
in star 2 and is therefore 75% hydrogen. Furthermore, the helium
abundance jumps to a maximum average abundance exceeding $Y=0.7$ near
$m/M=0.3$. The subsequent stellar evolution of an object with such an
atypical chemical abundance profile could be quite peculiar.
4.2 Rotational Properties of the Merger Remnants
The collisions with $r_{p}\neq 0$ result in rapidly, differentially
rotating merger remnants. Rotating fluid configurations with
$T/|W|\gtrsim 0.14$ are secularly
unstable, and those with $T/|W|\gtrsim 0.26$ are dynamically unstable
(Chandrasekhar 1969; Shapiro and Teukolsky 1983, Chap. 7).
The final merged configurations
are, by definition, dynamically stable, but they could in principle be
secularly unstable. Although some of our calculations produce
merger remnants close to
the secular instability limit, none of them exceed it.
However, extrapolation of our results to larger values of $r_{p}$
suggests that secular instabilities could well develop
in some merger remnants.
Table 3 lists the values of the central angular velocity $\Omega_{0}$ in the
equatorial plane as well as other quantities characterizing the rotation in
the outer layers of the merger remnants. Specifically, for the two mass
fractions $m/M=0.9$ and $m/M=0.95$, we give the values of
the polar and equatorial radii $r_{p}$ and $r_{e}$, the angular velocity
$\Omega$ in the equatorial plane, and the ratio $\Omega^{2}r_{e}/g$ of
centrifugal to gravitational acceleration in the equatorial plane.
We see that $\Omega^{2}r_{e}/g$ can be a significant fraction of unity,
indicating that some configurations are rotating near break-up.
The central angular velocity $\Omega_{0}$ is typically an order of magnitude
larger than the angular velocity $\Omega$ at $m/M=0.95$.
Figure 12 shows contours of the specific angular momentum $\Omega\varpi^{2}$, where $\varpi$ is the cylindrical radius measured from the
rotation axis, in the vertical $(x,z)$ plane for several representative
cases. The outermost bounding curves correspond to
$m/M=0.95$. Clearly, the merger remnants are not barotropic since the condition
$d\Omega/dz=0$ is not satisfied everywhere. The
implications of this result will be discussed in §5.
4.3 Interior Structure of the Merger Remnants
Figures 13(a)–(g) show the variation of the density $\rho$, relative specific
entropy $s-s_{o}$, helium fraction $Y$ and temperature $T$ as a function of
$m/M$ for all merger remnants. The density and entropy profiles are
fundamental in the sense that they do not depend on the assumed initial
helium profiles. The entropy and helium profiles have been averaged
over isodensity surfaces. The temperature profile is calculated from
the entropy and helium profiles by setting the pressure $p=\rho kT/\mu$ equal
to
$p=A\rho^{\Gamma_{1}}$, solving for $T$ and using equation (4). Here the mean
molecular weight $\mu$ is given by
$$\mu=m_{H}(2X+{3\over 4}Y+{1\over 2}Z)^{-1},$$
(17)
where $m_{H}$ is the mass of hydrogen and $X$, $Y$ and $Z$ are the
fractional abundances of hydrogen, helium and metals. For Population II
stars, $Z\sim 10^{-4}-10^{-3}$ and the precise value does not
significantly affect the calculated temperature profiles.
Note the peculiar shapes of some of the temperature and helium profiles
in Figure 13. For example, often the temperature or helium abundance
reaches its maximum value somewhere other than the center of the star.
Although these configurations are very close to hydrostatic equilibrium,
it is clear that they are not in thermal equilibrium (see related discussion
in §5). These unusual profiles suggest that we look at the condition for
convective stability more carefully. For a non-rotating star, this condition
can be written simply as
$${ds\over dr}>0,$$
(18)
where $s$ is the local specific entropy (see, eg., Landau & Lifshitz
1957, §4). When written in terms of temperature and composition
gradients, equation (18) becomes, for an ideal gas,
$${1\over T}{dT\over dr}>{1\over T}\left({dT\over dr}\right)_{ad}+{1\over\mu}{d%
\mu\over dr}$$
(19)
which is known as the Ledoux criterion (Kippenhahn & Weigert 1990, Chap. 6).
Here the subscript $ad$ denotes that the derivative is to be taken at
constant entropy. Most of our merger remnants
have composition gradients, and it is in the regions where $d\mu/dr>0$
that equation (19) requires $dT/dr>0$ for stability. For chemically homogeneous
stars,
the second term on the right-hand side of equation (19) vanishes, and
the familiar Schwarzschild criterion results.
Although equation (18) is quite general, it does require slight
modification for rotating stars (see Tassoul 1978, Chap. 7).
Figure 13 (a, d, and f) demonstrate that merger remnants formed from
equal mass parent stars have composition profiles which mimic those of
the parents, as can be seen by comparing the resulting helium profiles
to the corresponding parent profiles in Figure 2. In Figure 13(f), all
of the merger remnants have $Y=0.25$ for all $m/M$, which is simply
because the fully convective parents stars in these cases had $Y=0.25$
everywhere.
We see from Figure 13(c) that the central specific entropy of the
merger remnants increases with $r_{p}$, which can explain the
qualitatively different shapes of the corresponding helium abundance
profiles. This increase occurs because the number of interactions
$n_{p}$, and hence the level of shock heating in star 2 (the smaller
star), increases with $r_{p}$. The shock-heating in the central region
of star 1 is less sensitive to $n_{p}$, since the outer envelope absorbs
the brunt of the shock. For case G (solid line), $n_{p}=1$ and much of
star 2 is able to maintain a lower specific entropy than the minimum
value in star 1. Since low-entropy material tends to sink to the
bottom of the gravitational potential well, the merger remnant’s core
consists entirely of fluid originally from star 2 and therefore with a
helium abundance $Y=0.25$. For case H (long-dashed line), $n_{p}=2$ and,
although the central fractional helium abundance is still $0.25$, there
is enough shock heating for the fluid at small $m/M$ to be affected by
contributions from both stars. For case I (short-dashed line), $n_{p}=3$
and the additional shock heating is sufficient to prevent most of the
fluid from star 2 from reaching the center of the remnant, which
consequently is not significantly replenished with hydrogen.
Figure 13(g) displays the profiles for merger remnants resulting from
collisions between two stars of masses $M_{1}=M_{TO}$ and $M_{2}=0.2M_{TO}$.
In the head-on case, the less massive star (star 2) plummets so
quickly to the center that there is significant
shock-heating in the core of star 1, where the highest fractional
helium abundance resides. This
causes the helium-rich material to be spread throughout a larger region of
the merger remnant, and the resulting helium profile is not as sharply
peaked as in cases V and W.
4.4 Calculating Final Profiles
In order to keep our results useful for future applications,
we now present a simple and general method for constructing the
final composition profiles in our calculated merger remnants,
corresponding to any assumed initial composition profiles.
From our results, we extract some simple functions
which can be applied to transform, for example, any given initial
helium abundance profiles into a final helium abundance profile for
the merger remnant. Indeed, these transfer functions allow one to find
the final profile of any passively advected quantity,
provided only that the profiles of that quantity in the parent stars are both
known and spherically symmetric.
Table 4 and Figure 14 establish a correlation between the initial and
final mass fractions of a fluid particle. They also demonstrate that
the details of how the fluid elements are mixed during a collision can
be quite complicated. Table 4 treats six of the collisions between
equal-mass stars. The parent stars and merger remnant are partitioned
into zones according to interior mass fractions. For every zone in the
final configuration, we list the fraction of particles which originated
in each of the initial zones. Although there is definitely a preferred
final mass fraction $m/M$ for a given initial mass fraction $m_{i}/M_{i}$,
there is always a range of $m/M$ obtainable. In Figure 14, which is
for case G, this range of mass fractions is evident in the spread of
points around a preferred average. The lower band of points
surrounding the solid line correspond to particles which originated in
star 1, while the upper band surrounding the dashed line correspond to
particles which originated in star 2. The lines correspond to the
average initial mass fraction $\langle m_{i}/M_{i}\rangle$ for stars
$i=1,2$ as a function of the final mass fraction $m/M$, obtained by
binning values of $m/M$. In contrast, note that if the parent stars
were completely mixed by the collision then the points would be
distributed uniformly over the entire plot with an average initial mass
fraction $\langle m_{i}/M_{i}\rangle={1\over 2}$ for all $m/M$.
Let us define $p_{1}=p_{1}(m/M)$ to be the probability that a particle with
final mass fraction $m/M$ originated in star 1. Obviously, $1-p_{1}$ is
then the probability that the particle originated in star 2. With
this definition we can approximate the final profile of any passively
advected quantity $Q$ according to
$$Q\left({m\over M}\right)\simeq p_{1}\left({m\over M}\right)Q_{1}\left(\left.%
\left\langle{m_{1}\over M_{1}}\right\rangle\right|_{m\over M}\right)+\left(1-p%
_{1}\left({m\over M}\right)\right)Q_{2}\left(\left.\left\langle{m_{2}\over M_{%
2}}\right\rangle\right|_{m\over M}\right),$$
(20)
where $Q_{i}$ are the initial (spherically symmetric) profiles for that
quantity in the parent stars $i=1,2$. The quantities $\left.\langle m_{i}/M_{i}\rangle\right|_{m/M}$ which appear in equation (20) are the average
initial mass fractions, such as the ones in Figure 14, evaluated at the
final mass fraction $m/M$. If all particles at $m/M$ came from a
single value of $m_{i}/M_{i}$, then equation (20) would be exact. In addition,
if the initial profiles are linear over the range of $m_{i}/M_{i}$ which
contributes to the abundance at $m/M$, then the above relationship is
exact.
Figures 15(a)–(g) give the average mass fractions $\langle m_{i}/M_{i}\rangle$ as a function of $m/M$ for all of our collisions, while
Figure 16 gives the function $p_{1}=p_{1}(m/M)$ for the 12 collisions
involving parent stars of unequal mass. For collisions involving two
identical stars we necessarily have $\langle m_{1}/M_{1}\rangle=\langle m_{2}/M_{2}\rangle$ and $p_{1}={1\over 2}$ for all $m/M$. The solid, long
dashed and short-dashed lines correspond to pericenter separations
$r_{p}$ of $0,0.25$, and $0.5(R_{1}+R_{2})$, respectively; in Figure 15(f) the
dot-dashed and dotted lines refer to case S ($r_{p}=0.75(R_{1}+R_{2})$) and
case T ($r_{p}=0.95(R_{1}+R_{2})$), respectively. Note that the horizontal
line $\langle m_{i}/M_{i}\rangle={1\over 2}$ would correspond to the fluid
of star $i$ being completely mixed throughout the merger remnant, which
is not the case for any of our calculations. For collisions involving
equal-mass stars, if there were no shock heating and no mass loss then
every particle would have identical initial and final mass fractions
(i.e., $m_{i}/M_{i}=m/M$) so that the merger remnant’s helium profile would
be the same as in the parent stars. In the equal-mass cases A, B,
C, J, K and L, we do find $\langle m_{i}/M_{i}\rangle\simeq m/M$ and the
final helium profiles are indeed quite similar to the parent profile, as shown
in §4.3.
Along with equation (20), the functions of Figures 15 and 16 provide
the means for approximating the final profile of any passively advected
quantity. As a concrete example of how to use this method, we will now
calculate the fractional helium abundance at $m/M=0.28$ in the merger
remnant of case G, using the same initial profiles as shown in
Figure 2. From the solid lines corresponding to case G in
Figures 15(c) and 16, we find that$\left.\langle m_{1}/M_{1}\rangle\right|_{0.28}=0.05$, $\left.\langle m_{2}/M_{2}\rangle\right|_{0.28}=0.65$ and $p_{1}(0.28)=0.62$. Therefore,
$$Y(0.28)\simeq 0.62\times Y_{1}(0.05)+(1-0.62)\times Y_{2}(0.65)=0.70,$$
(21)
where we have used $Y_{1}(0.05)=0.97$ and $Y_{2}(0.65)=0.25$, obtained from
the solid and short-dashed lines of Figure 2, respectively. By
repeating this calculation for other values of $m/M$, we construct the
approximate helium profile shown as the dashed line in Figure 17. Also
shown for comparison is the “exact” profile (solid line) constructed
by considering the individual helium abundance carried by each particle
(the same curve which appears in Figure 11(b)). We consider the
agreement to be quite good, given the simplicity of the approximation
scheme and the fact that it does not require access to large data files
containing information on all $N=3\times 10^{4}$ particles.
4.5 Effects of Spurious Diffusion on Observed Mixing
In all SPH calculations, numerical noise can lead to spurious, or
artificial, diffusion of SPH particles. In order to estimate how much
of the observed mixing is in fact caused by particle diffusion, we have
performed a series of systematic tests to evaluate quantitatively the
effects of spurious transport in SPH calculations (Lombardi, Rasio, &
Shapiro 1995b). In particular these tests measure, as a function of
the neighbor number $N_{N}$ and local noise level $v_{rms}$ (the root
mean square particle velocity deviation from the local mean), dimensionless
spurious diffusion coefficients defined by
$$D\equiv{n^{1/3}\over c_{s}}{d\Delta r_{s}^{2}\over dt}$$
(22)
where $\Delta r_{s}$ is the distance traveled by a particle due to
spurious diffusion,
$n$ is the local number density of SPH particles, and $c_{s}$ is
the local sound speed.
Once measured, these diffusion coefficients can be applied to each
particle in our simulations by monitoring its local values of
$v_{rms}$, $n$ and $c_{s}$ and then estimating how far
that particle has spuriously diffused by numerically integrating
$$\Delta r_{s}^{2}=\int{D{n^{-1/3}\over c_{s}}dt}.$$
(23)
From the local density gradient at the particle’s final position we can
then estimate the equivalent mass fraction corresponding to this
displacement according to
$$\Delta m_{s}\simeq{\Delta r_{s}\over 3^{1/2}}|{\bf\nabla}\rho|.$$
(24)
By repeating this procedure for all the particles, we arrive at an
average spurious diffusion distance $\langle\Delta r_{s}\rangle$ and
mass fraction $\langle|\Delta m_{s}|\rangle$, as well as a
root-mean-square spurious displacement $\langle\Delta r_{s}^{2}\rangle^{1/2}$ and mass fraction $\langle\Delta m_{s}^{2}\rangle^{1/2}$.
This method of estimating spurious diffusion distances will be referred
to as Method I.
In the case of a head-on collision, another method (which we call Method II)
can be used, which exploits the axisymmetry around the collision axis
(the x-axis).
If we make the assumption that the entire dynamical evolution remains
axisymmetric (this would not be the case if, e.g., Rayleigh-Taylor
instabilities were to develop), then
a particle should always remain in
the plane containing the collision axis and the particle’s initial
position. The extent to which a particle diffuses in the direction
perpendicular to this plane provides an estimate of the spurious diffusion
distance. Assuming the diffused amount is the same in all three directions,
the total
spurious diffusion distance is then estimated simply by multiplying by
$3^{1/2}$. We finally convert the spurious diffusion distance to
an equivalent mass fraction exactly as in Method I (see eq. [24]).
The results of the two methods applied to our calculations are given
in Table 5. When two numbers are given, the second one is
calculated by Method II. Included in this table are the average
spurious diffusion distance $\langle\Delta r_{s}\rangle$, the root-mean-square
diffusion distance $\langle\Delta r_{s}^{2}\rangle^{1/2}$, the
average equivalent mass fraction $\langle|\Delta m_{s}|\rangle/M$,
and the root-mean-square mass fraction $\langle\Delta m_{s}^{2}\rangle^{1/2}/M$.
It is clear that the two methods are generally in good agreement,
and that, when expressed in terms of $m/M$, the effects of
spurious diffusion are always small.
Also listed in Table 5
are the observed average ($\langle|\Delta m_{o}|\rangle/M$)
and root-mean-square ($\langle\Delta m_{o}^{2}\rangle^{1/2}/M$) total deviation
in final mass fraction.
The last column then subtracts from the total square deviation
$\langle\Delta m_{o}^{2}\rangle$ the contribution $\langle\Delta m_{s}^{2}\rangle$
from spurious diffusion.
For example, in case G, we observe a root-mean-square change in the
interior mass fraction of ${\langle\Delta m_{o}^{2}\rangle^{1/2}/M}=0.09$ over the entire calculation. Using Method I, we estimate that
the root-mean-square change in interior mass fraction due to spurious
diffusion is ${\langle\Delta m_{s}^{2}\rangle^{1/2}/M}\simeq 0.036$,
while Method II gives an estimate of $0.057$ for this quantity. We
therefore believe that the physical root-mean-square spread (i.e., the
spread in a calculation free of spurious diffusion) would be
approximately $0.08$ or $0.07$, depending on whether Method I
or Method II is more accurate.
5 Summary and Discussion
The main results of this paper can be summarized as follows. We have
demonstrated that the typical merger remnants produced by collisions
are rapidly and differentially rotating, and are far from chemically
homogeneous, with composition profiles that can be rather peculiar in
certain cases. For example, it often happens that the maximum helium
abundance does not occur at the center of the remnant
(cf. Fig. 11(b)). The merger remnants produced by our dynamical
calculations, although very close to hydrostatic equilibrium, are
usually far from thermal equilibrium, as discussed below. In
particular, we have shown that the merger remnants are not barotropes
(ie., the condition $d\Omega/dz=0$ is generally not satisfied), and
that their temperature profiles can have positive gradients ($dT/dr>0$)
in certain regions.
At a qualitative level, many of our results can be understood very
simply in terms of the requirement of convective stability of the final
merger remnant. If entropy production in shocks could be neglected
(which may be reasonable for parabolic collisions, especially in the
head-on case), then one could predict the qualitative features of the
remnant’s composition profile simply by observing the composition and
entropy profiles of the parent stars. Convective stability requires
that the specific entropy $s$ increase from the center to the surface
($ds/dr>0$) in the final hydrostatic equilibrium configuration.
Therefore, in the absence of shock-heating, fluid elements conserve
their entropy and the final composition profile of a merger remnant
could be determined simply by combining mass shells in order of
increasing entropy, from the center to the outside. Many of our results
follow directly. For example, in the case of a collision between two
identical stars, it is obvious why the composition profile of the
merger remnant remains very similar to that of the parent stars. For
two stars of very different masses, the much lower-entropy material of
the lower-mass star tends to concentrate at the center of the final
configuration, leading to the unusual composition and temperature
profiles seen in Figure 13 (c, e, and g).
Regions where the dynamical stability criterion $ds/dr>0$ (eq. [18]) is
satisfied can nevertheless be thermally, or secularly, unstable. The
small vertical oscillations (at the local Brunt-Väisälä frequency
$\Omega_{BV}\propto[ds/dr]^{1/2}$) of a fluid element in such a region
have amplitudes that grow slowly, and mixing will occur on a timescale
comparable to the local radiative damping time (see, e.g., Kippenhahn
& Weigert, Chap. 6). The thermal instability can be of two types.
When $d\mu/dr>0$ and $dT/dr>0$ (as in Figure 13 (c, e, and g)), a
so-called thermohaline instability can develop, allowing fingers of
the high-$\mu$ material to penetrate down into the lower-$\mu$, colder
material below (see, e.g., Ulrich 1972). When such mixing occurs in the
stellar core, it tends
to increase the central helium abundance and therefore decrease the
time that the merger remnant can remain on the MS.
When $d\mu/dr<0$ but $dT/dr<(dT/dr)_{ad}$, so-called semiconvection
occurs (Spruit 1992). In terms of easily computed SPH variables, this
criterion is equivalent to
$$0<{1\over A}{dA\over dr}<-{\Gamma_{1}\over\mu}{d\mu\over dr}$$
(25)
where $A$ is related to specific entropy by equation (4). We have
tested our merger remnants formed from head-on collisions and found that
this instability is typically present. Figure 18 shows, as a function
of the final mass fraction $m/M$, the fraction $f_{sc}$ of gas which is
semiconvective for six of our merger remnants. In all cases no
semiconvective instability exists in the outer $\sim$20% of the mass,
so that we do not expect this mixing mechanism can increase the helium
abundance of the outer layers. Figure 18 does demonstrate, however,
that some merger remnants (those of cases A, D and J) have a unstable
region which extends to the center, and these remnants therefore have a
means of mixing hydrogen into their cores. For instance in case A, we
see that the inner $\sim$40% could be significantly affected. As
semiconvection slowly attempts to decrease the central helium fraction,
it must compete against hydrogen burning. In addition, the right hand
side of equation (25) changes as the fluid mixes, so that the details
of this complicated process can only be followed numerically with a
stellar evolution code.
For a rotating, chemically homogeneous star, stable thermal equilibrium
requires $d\Omega/dz=0$, where $\Omega$ is the angular velocity and $z$
is measured along the rotation axis (the Goldreich-Schubert stability
criterion; see, e.g., Tassoul 1978, Chap. 7).
From the representative set of specific angular momentum
contours presented in Figure 12, it is therefore evident that the
merger remnants of cases Q, R, S and T (which are chemically
homogeneous, since their parent stars were fully mixed) cannot be in
thermal equilibrium. In chemically inhomogeneous stars, regions with a
sufficiently large and stabilizing composition gradient ($d\mu/dr<0$)
can in principle still be thermally stable even with $d\Omega/dz\neq 0$. However, it seems unlikely that the composition profiles generated
dynamically by a collision would conspire to keep the remnants
everywhere thermally stable.
Deupree (1990) has shown that stars with rapidly rotating cores and
slowly rotating envelopes can have their MS lifetime extended beyond
that of their non-rotating counterparts. The fact that much of the
angular momentum is hidden deep in the remnant’s interior suggests a
possible explanation for why observations of blue stragglers in open
clusters such as M67 find no signs of rapid rotation (Peterson, Carney,
& Latham 1984; Mathys 1991). Recently, Leonard & Livio (1995) have
argued that the spin-down timescale of blue stragglers due to magnetic
breaking should be on the order of only $10^{5}$ years, which is much
less than the thermal timescale of approximately $10^{7}$ years, so that
initially rapidly rotating merger remnants may not be a problem for the
collisional formation scenario.
Stellar encounters with separations $r_{p}$ larger than those considered
in this paper are difficult to compute directly with SPH since the
amount of energy $\Delta E$ dissipated during the first interaction is
then so small that the integration time until the next pericenter
passage can be several orders of magnitude larger than the hydrodynamic
time. We believe that our results can be safely extrapolated all the
way to values of $r_{p}\simeq 1.2(R_{1}+R_{2})$. For instance, from our
results for collisions of equal-mass stars it seems very likely that
the helium profile of the merger remnant will always mimic that of the
parent stars. In Figure 13(f), note that the density, entropy and
temperature profiles also seem to be converging onto a fixed profile,
and that the profiles for the $r_{p}=0.5,0.75$ and $0.95(R_{1}+R_{2})$ cases
all look very similar. For $r_{p}\gtrsim 1.2(R_{1}+R_{2})$ the encounter is
better described as a tidal capture than a collision, i.e., the amount
of energy dissipated is sufficient to form a bound system, but no
direct collision occurs, even in the outer layers of the stars. The
maximum value $r_{p}=r_{cap}$ for tidal capture can be calculated
accurately from linear perturbation theory (Press & Teukolsky 1977;
McMillan, McDermott, & Taam 1987). For two identical $0.8\,M_{\odot}$ MS
stars and a relative velocity at infinity $v_{\infty}=10{\,{\rm km}\,{\rm s}^{-1}}$, McMillan
et al. (1987) find $r_{cap}/(R_{1}+R_{2})\simeq 1.4$, which leaves little
room for “clean” tidal captures. In addition, the long-term evolution
of a tidal-capture binary may well lead to merging of the two stars
even if the initial interaction is in the linear regime (for recent
discussions, see Mardling 1995a,b and Kumar & Goodman 1995).
It must be stressed that the amount of mixing determined by SPH
calculations is always an upper limit. Indeed, some of the mixing
observed in a calculation will always be a numerical artifact.
Low-resolution SPH calculations in particular tend to be very noisy and
the noise can lead to spurious diffusion of SPH particles, independent
of any real physical mixing of fluid elements. In §4.5 we introduced
two simple methods to evaluate quantitatively the effects of spurious
diffusion in our calculations. The results suggest that spurious
diffusion does not significantly corrupt our results. This is seen for
example in the near agreement of numbers in the last two columns of
Table 5. However, both methods are approximate. The diffusion
coefficients used in Method I have been measured in the absence of
artificial viscosity. The presence of artificial viscosity in our
collision calculations may slightly change the effective values of
these coefficients. Method II assumes that we can neglect any
nonaxisymmetric instabilities in a head-on collision and that spurious
diffusion is isotropic (which may not be true in the presence of strong
entropy gradients).
Nevertheless, the reasonable agreement between the two methods
(cf. Table 5) gives us confidence that these simplifying assumptions
are generally satisfied to a good approximation. The general question
of spurious transport in SPH calculations will be addressed in a
separate paper (Lombardi et al. 1995b).
Benz & Hills (1987) performed the first fully three-dimensional
calculations of collisions between two identical MS stars. In contrast
to the present work, they considered only $n=1.5$ polytropes, which are
mostly relevant for collisions of very low-mass stars ($M_{1}=M_{2}\lesssim 0.4M_{\odot}$). Their calculations for parabolic collisions indicated a
higher level of mixing than we find in similar cases (cases P, Q, R, S
and T in our Table 1). We have also computed a number of collisions
using $N=1024$ SPH particles (as in Benz & Hills 1987, rather than
$N=3\times 10^{4}$ particles as in most of our other calculations), and
found substantially higher levels of spurious diffusion. We conclude
that a significant part of the mixing observed by Benz & Hills (1987)
was a numerical artifact. However, we should note that our
higher-resolution calculations for $n=1.5$ polytropes (cases P, Q, R,
S and T) do exhibit a generally higher level of mixing than observed
for other models (e.g., cases A, B and C). This is not surprising since
we expect parent stars of constant entropy, which are only marginally
stable against convection, to be easier to mix than those with
significant positive entropy gradients (stable stratifications). In
addition, the more homogeneous density profile of $n=1.5$ polytropes
leads to a better distribution of the impact energy throughout the
entire mass of fluid. Benz & Hills (1992) have performed calculations
of collisions between $n=1.5$ polytropes with a mass ratio
$M_{2}/M_{1}=0.2$, using $3500$ SPH particles per star. These calculations
are not directly relevant to blue straggler formation, given the very
low masses of the MS stars involved ($M_{1}\lesssim 0.4\,M_{\odot}$, hence
$M_{2}<0.1\,M_{\odot}$), and, since none of our calculations model such low
masses, no direct comparison is possible.
There are a number of ways by which our results could be improved or
extended. For instance, the equation of state could be extended to
include radiation, partial ionization and electron degeneracy
corrections. We do not expect these corrections to be significant.
More accurate initial composition profiles could be also used.
Note, however, that our results can be applied to
arbitrary initial profiles by the method of §4.4.
Incidentally, profiles of ${}^{7}$Li would be particularly interesting to
consider since this element is destroyed at
temperatures $T\gtrsim 10^{6}K$ and is therefore an
observationally measurable indicator of mixing (see, e.g., Hobbs &
Mathieu 1991 and Pritchet & Glaspey 1991).
Our dynamical calculations and the determination of hydrodynamical
mixing are only the first step in modeling blue straggler formation.
The merger remnants, which are much larger than normal equilibrium MS
stars of the same mass, will recontract to the MS on a thermal
timescale ($\sim 10^{7}\,$yr). As they evolve, other mixing processes
such as meridional circulation and convection may well be important.
Calculations of this thermal relaxation phase using the results of
dynamical calculations, such as those presented in this paper, as
initial conditions, will be necessary in order to make detailed
predictions for the observable parameters of blue stragglers.
Recently, Sills, Bailyn and Demarque (1995, hereafter SBD) have begun
investigating the consequences of blue stragglers being born unmixed.
To create their unmixed initial model of a blue straggler formed by the
collision of two TAMS stars, SBD relaxes a non-rotating TAMS star whose
mass has been artificially doubled but which is otherwise unchanged.
The subsequent stellar evolution is contrasted to that of a fully mixed
(ie. chemically homogeneous) blue straggler. SBD finds that the high
central helium concentration in the unmixed models causes the time
spent on the MS ($\sim 5\times 10^{7}$ yr) to be drastically shorter than
for the fully mixed counterparts ($\sim 5\times 10^{8}$ yr), making it
difficult to account for the observed numbers blue stragglers in the
core of NGC 6397. In addition, such unmixed blue stragglers are
neither bright nor blue enough to explain the observations. A blue
straggler population consisting purely of non-rotating, unmixed merger
remnants of two TAMS parent stars is therefore not sufficient to
explain the core blue stragglers in NGC 6397.
Future work which follows the approach of SBD would clearly be
beneficial. A number of factors need to be considered in more detail.
For instance, it is unrealistic to expect that all collisional blue
stragglers are born only from TAMS parent stars. Instead, attention
must also be given to collisions between unequal mass parent stars,
which form merger remnants with profiles that are neither homogeneous
nor like that of the parents. That is, such remnants are neither fully
mixed nor unmixed. Since these blue stragglers have an enhanced
hydrogen abundance in their cores, they will presumably remain on the
MS for a longer time and have a different position on a color magnitude
diagram than their unmixed counterparts. Moreover, most blue
stragglers will be formed rapidly rotating, especially in the stellar
core, which acts to extend their postponed residence on the MS (Deupree
1990). In addition, the density and specific entropy profiles shown in
Figure 13 can be used to specify the structure of the blue straggler
and improve upon SBD’s somewhat ad hoc profiles obtained by
artificially scaling the mass of an equilibrium star.
Support for this work was provided by NSF Grant AST 91–19475,
NASA Grant NAG 5–2809, and NASA Grant HF-1037.01-92A.
Computations were performed at the Cornell Theory Center, which
receives major funding from the NSF and IBM Corporation, with
additional support from the New York State Science and Technology
Foundation and members of the Corporate Research Institute.
References
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15)
(16)
(17)
(18)
(19)
(20)
(21)
(22)
(23)
(24)
(25)
(26)
(27)
(28)
(29)
(30)
(31)
(32)
(33)
(34)
(35)
(36)
(37)
(38)
(39)
(40)
(41)
(42)
(43)
(44)
(45)
(46)
(47)
(48)
(49)
(50)
(51)
(52)
(53)
(54)
(55)
(56)
(57)
(58)
(59)
(60)
(61)
(62)
(63)
(64) |
Writing and Reading antiferromagnetic Mn${}_{2}$Au: Néel spin-orbit torques and large anisotropic magnetoresistance.
S. Yu. Bodnar
Institut für Physik, Johannes Gutenberg-Universität, Staudinger Weg 7, 55128 Mainz, Germany
L. Šmejkal
Institut für Physik, Johannes Gutenberg-Universität, Staudinger Weg 7, 55128 Mainz, Germany
Institute of Physics, Academy of Sciences of the Czech Republic, Cukrovarnicka 10, 162 00 Praha 6, Czech Republic
Faculty of Mathematics and Physics, Charles University, Department of Condensed Matter Physics, Ke Karlovu 5,
12116 Praha 2, Czech Republic
I. Turek
Faculty of Mathematics and Physics, Charles University, Department of Condensed Matter Physics, Ke Karlovu 5,
12116 Praha 2, Czech Republic
T. Jungwirth
Institute of Physics, Academy of Sciences of the Czech Republic, Cukrovarnicka 10, 162 00 Praha 6, Czech Republic
School of Physics and Astronomy, University of Nottingham, University Park, Nottingham NG7 2RD, United Kingdom
O. Gomonay
Institut für Physik, Johannes Gutenberg-Universität, Staudinger Weg 7, 55128 Mainz, Germany
J. Sinova
Institut für Physik, Johannes Gutenberg-Universität, Staudinger Weg 7, 55128 Mainz, Germany
A.A. Sapozhnik
Institut für Physik, Johannes Gutenberg-Universität, Staudinger Weg 7, 55128 Mainz, Germany
H.-J. Elmers
Institut für Physik, Johannes Gutenberg-Universität, Staudinger Weg 7, 55128 Mainz, Germany
M. Kläui
Institut für Physik, Johannes Gutenberg-Universität, Staudinger Weg 7, 55128 Mainz, Germany
M. Jourdan
[email protected]
Institut für Physik, Johannes Gutenberg-Universität, Staudinger Weg 7, 55128 Mainz, Germany
(December 7, 2020)
††preprint:
Antiferromagnets are magnetically ordered materials which exhibit no net moment and thus are insensitive to magnetic fields. Antiferromagnetic spintronics Jun16 aims to take advantage of this insensitivity for enhanced stability, while at the same time active manipulation up to the natural THz dynamic speeds of antiferromagnets Kam11 is possible, thus combining exceptional storage density and ultra-fast switching. However, the active manipulation and read-out of the Néel vector (staggered moment) orientation is challenging. Recent predictions have opened up a path based on a new spin-orbit torque Zel14 , which couples directly to the Néel order parameter. This Néel spin-orbit torque was first experimentally demonstrated in a pioneering work using semimetallic CuMnAs Wad16 . Here we demonstrate for Mn${}_{2}$Au, a good conductor with a high ordering temperature suitable for applications, reliable and reproducible switching using current pulses and read-out by magnetoresistance measurements. The symmetry of the torques agrees with theoretical predictions and a large read-out magnetoresistance effect of more than ${\bf\simeq 6}$ ${\bf\%}$ is reproduced by ab initio transport calculations.
For the key application operations of reading and writing in antiferromagnets, different approaches have been previously put forward. Initial experiments on spin-valve structures with an antiferromagnet (AFM) as the active layer manipulated the Néel vector by an exchange-spring effect with a ferromagnet (FM) and read-out via Tunneling-Anisotropic Magnetoresistance (T-AMR) measurements Par11 . Other related experiments were based on the same effect Fin14 , or on a FM to AFM phase transition Mar14 . However, the most promising approach is to use current induced spin-orbit torques for switching the Néel vector. It exhibits superior scaling and its counterpart in ferromagnets is already established and considered among the most efficient switching mechanisms for memory applications. Gam11 ; Bra14 .
Only two compounds, CuMnAs and Mn${}_{2}$Au, are known to provide at room temperature the collinear commensurate antiferromagnetic order and specific crystal structure, which is predicted to result in the staggered spin accumulation in the sublattice structure, leading to bulk Néel spin-orbit torques allowing for current induced switching of the Néel vector Zel14 .
Semimetallic CuMnAs was grown previously by molecular beam epitaxy (MBE) with a Néel temperature of $\simeq 500$ K Wad13 and current-induced switching of these samples was recently demonstrated for the first time Wad16 ; Grz17 . However, for spintronics applications the compound Mn${}_{2}$Au provides several advantages, as it is a good metallic conductor and does not contain toxic components. Furthermore, its magnetic ordering temperature is well above 1000 K Bar13 , providing the necessary thermal stability for applications. Mn${}_{2}$Au shows a simple antiferromagnetic structure with the collinear magnetic moments in the (001)-plane Shi10 ; Bar13 ; Bar16 . Thin film samples were previously grown in (101)-orientation by MBE Wu12 and Fe/Mn${}_{2}$Au(101) bilayers showed AMR effects of up $2.5$ $\%$ in a 14 T rotating magnetic field Wu16 .
While Mn${}_{2}$Au was the first compound for which current-induced internal staggered spin-orbit torques were predicted Zel14 , corresponding experimental evidence has been missing. Here we report current induced Néel vector switching in Mn${}_{2}$Au(001) epitaxial thin films, which is is easily read-out by a large AMR.
Our Al${}_{2}$O${}_{3}$/Ta(10 nm)/Mn${}_{2}$Au(75 nm)/Ta(3 nm) samples were prepared by sputtering as described elsewhere Jou15 and patterned into a star-structure as shown in Fig. 1.
This geometry allows for electric writing of the Néel vector orientation by pulsing currents along the two perpendicular directions $I_{pulse}^{1}$ and $I_{pulse}^{2}$ and for electric read-out by measuring either the transversal resistivity $\rho_{xy}$, i. e. the Planar Hall Effect (PHE), or the longitudinal resistivity $\rho_{long}$, corresponding to the AMR of the samples. Depending on the orientation of the patterned structure, the pulse currents can be sent along different crystallographic directions.
The AMR of a single domain sample is given by
$${\rm AMR_{hkl}}=\frac{\rho_{long}(\phi=0^{\rm o})-\rho_{long}(\phi=90^{\rm o})%
}{\bar{\rho}_{long}}=\frac{\Delta\rho_{long}}{\bar{\rho}_{long}},$$
(1)
where $\rho_{long}$ is longitudinal resistivity, $\phi$ is the angle between the Néel vector and current direction and [hkl] is the Néel vector orientation in the basis of the tetragonal conventional unit cell.
The PHE usually observed in ferromagnetic materials scales with the AMR and shows a dependence on the angle $\phi$ given by Tom75 ; See11 :
$$\rho_{xy}=\Delta\rho_{long}\,\sin\phi\cos\phi$$
(2)
Thus also in antiferromagnets $\rho_{xy}$ has its maximum value and changes sign if $\phi$ switches from $+45^{\rm o}$ to $-45^{\rm o}$.
To study the switching, trains of $100$ current pulses with a pulse length of $1$ ms and a delay between the pulses of 10 ms were applied. As after a pulse train thermal relaxation behaviour on a time scale of $1$ s after was observed, the read-out was performed with a delay of $10$ s.
Fig. 2 shows the transversal resistivity $\rho_{xy}$ versus the number of applied pulse trains.
First, a pulse current density of $1.4\times 10^{7}$ A/cm${}^{2}$ was applied along the [1$\bar{1}$0] direction, resulting in a small change of the corresponding Hall voltage. Without reaching saturation after $50$ pulse trains the pulse current direction was switched to [110], resulting in a reversal of the corresponding change of the transversal resistivity. This sequence could be reproduced several times. Increased pulse current densities of $1.7\times 10^{7}$ A/cm${}^{2}$ and $1.8\times 10^{7}$ A/cm${}^{2}$ resulted in larger changes of the corresponding Hall voltages. By increasing the number of pulse trains applied along the [110] direction to $500$, a trend towards saturation of the Hall voltage was obtained.
Internal field like spin-orbit torques are expected to generate reversible switching between distinct stable states if the current is injected along biaxial easy directions Zel14 ; Roy16 . However, we observed reversible switching to stable states for pulse currents along both the crystallographic [110] and [100] axes (rotated star-pattern). Thus we conclude that the in-plane magnetic anisotropy of our Mn${}_{2}$Au thin films is weak. This is consistent with our calculations of the magnetocrystalline anisotropy energy (MAE), which is almost negligible within the ab-plane (see Methods).
An example of the resulting changes of the transversal and longitudinal resistivities generated by pulse currents along the [100] and [010] directions is displayed in Fig. 3.
The upper panel (Fig. 3a) shows the longitudinal resistivity probed after each of the first 1600 pulse trains consisting of 100 pulses each with a current density $1.0\times 10^{7}$ A/cm${}^{2}$. For the first sequences only small variations of the longitudinal resistivity were observed. However, with the application of subsequent pulse trains the magnitude of the effect increased. This training-like behaviour may be associated with the motion and pinning of AFM domain walls in the sample. After 1600 pulse trains a constant resistance change of $\Delta\rho_{long}/{\bar{\rho}}=2.5$ % induced by 100 pulse trains was reached, which is an order of magnitude higher than what is observed for CuMnAs Wad16 .
To check the origin of these changes, the transversal resistivity of the sample was measured and also showed reproducible pulse current induced changes (Fig. 3b). The increase of the transversal resistivity induced by 100 pulse trains amounted to $\rho_{xy}=0.5$ ${\rm\mu\Omega cm}$.
Based on these numbers the identification of the longitudinal and transversal resistivities with the AMR and PHE can be verified: If both effects originate from the same anisotropic electron scattering, they have to be related by equation (2). We assume a switching of the Néel-vector in parts of the sample corresponding to a change of $\phi$ in equation (2) from $+45$ ${}^{\rm o}$ to $-45$ ${}^{\rm o}$, i. e.
$$\Delta\rho_{xy}=\rho_{xy}(+45^{\rm o})-\rho_{xy}(-45^{\rm o})=\Delta\rho_{long},$$
(3)
Thus we find that $\rho_{xy}=0.5$ ${\rm\mu\Omega cm}$ corresponds again to $\Delta\rho_{long}/{\bar{\rho}}=2.5$ %. This consistency of the longitudinal and transversal resistivities provides strong evidence for an intrinsic electronic origin of the pulse current induced changes of the magnetoresistance signals.
After two more pulse current direction reversals reproducing the previous behaviour of the sample, the pulse current direction was kept along [100] for 800 additional pulse trains. This resulted in a sign reversal of the PHE. Although a small offset of the transversal voltage due to e. g. imperfections of the patterned structure is possible, the magnitude of the PHE indicates that the Néel vector now switched in the majority of the sample. After about 300 pulse trains along the [100] direction a beginning saturation of the PHE resistivity appeared, but was not completed when after 500 additional pulse trains the sample broke. A maximum transversal resistivity of $\rho_{xy}=1.25$ ${\rm\mu\Omega cm}$ was reached, which based on equation (3) corresponds to an AMR of $6.25$ $\%$. This is one of the largest found in metallic magnetic thin films, and its size bodes well for easy read-out of the antiferromagnetic state as necessary for device applications. While small variations exist between samples, we observe consistently larger AMR effects for pulse currents along the [100] than for the [110] directions.
To understand the origin of the magnetoresistance effects, we calculated the AMR of single domain Mn${}_{2}$Au, assuming a complete $90^{\rm o}$ switching of the Néel vector (see Methods). In general, AMR originates from the effects of spin-orbit coupling on the band structure and from scattering from an extrinsic disorder potential Ran08 . When incorporating the effects of realistic disorder in the calculations (see Methods), two types were considered: Off-stoichiometry and inter-site swapping between Mn and Au atoms. Experimentally, the former was analyzed by energy dispersive x-ray spectroscopy (EDX) of $500$ nm thick Mn${}_{2}$Au films resulting in a stoichiometry of $66.2\pm 0.3$ $\%$ Mn and $33.8\pm 0.3$ $\%$ of Au, which indicates a slight Au excess. Also a small degree of inter-site disorder can be expected, but its quantification is experimentally not accessible. Thus we simulated a slight excess of Au randomly distributed over the Mn sites and random Mn - Au swapping.
We calculated the AMR for two crystal directions of the Néel vector, AMR${}_{100}$ and AMR${}_{110}$. Fig. 4a, shows the results for different degrees of disorder in Mn${}_{2}$Au.
Large AMR values consistent with our experiments were obtained for small degrees of disorder reaching a maximum value of $6.3$ % for 0.5 % excess of Au. Moreover, we obtain $AMR_{100}>AMR_{110}$, which reproduces the experimentally observed trend for the two crystalline directions. The corresponding calculated residual resistivities as shown in Fig. 4b are consistent with the experimentally obtained values ($\simeq 8$ ${\rm\mu\Omega cm}$ Jou15 ), corroborating the relevance of the simulated type of disorder.
In summary, in-plane switching of the Néel vector in the antiferromagnetic metal Mn${}_{2}$Au by current pulses was realized using intrinsic spin-orbit torques. Consistent measurements of the AMR and PHE showed pulse current direction dependent reversible changes, providing direct evidence for Néel vector switching. Easy read-out of the switching is provided by a large amplitude of the AMR of more than 6%, which is more than an order of magnitude higher than previously observed for other antiferromagnetic systems and one of the highest AMR amplitudes found for metallic magnetic thin films. We can reproduce the magnitude of the effect theoretically by including realistic disorder and, in particular, find the same dependence of the amplitude on the crystallographic directions in the experiment as in the calculation. With the basic principles of writing and read-out demonstrated, combined with a theoretical understanding of the underlying spin-orbit torques, and the large magnetoresistive effects, the metallic compound Mn${}_{2}$Au is a prime candidate to enable future AFM spintronics.
Acknowledgments: This work is supported by the German Research Foundation (DFG) through the Transregional Collaborative Research Center SFB/TRR173 Spin+X, Projects A03 and A05. J.S., L.S., O.G., and T.J. acknowledge the support of the Alexander von Humboldt Foundation, the ERC Synergy Grant SC2 (No. 610115), the Ministry of Education of the Czech Republic Grant No. LM2015087, and the Grant Agency of the Czech Republic grant no. 14-37427G, L.S. acknowledges support from the Grant Agency of the Charles University, no. 280815. Access to computing and storage facilities owned by parties and projects contributing to the National Grid Infrastructure MetaCentrum provided under the program ”Projects of Large Research, Development, and Innovations Infrastructures” (CESNET LM2015042), is greatly appreciated. The work of I.T. was supported by the Czech Science Foundation (Grant No. 14-37427G).
Author contributions
S.Yu.B. and M.J. mainly wrote the paper and performed the transport measurements; L.S. performed the AMR and anisotropy calculations and wrote the corresponding part of the manuscript, I.T. developed the codes for the transport calculations, S.Yu.B. and A.A.S. prepared the samples, H.-J.E, M.K., O.G., T.J., I.T, and J.S. discussed the results and contributed to the writing of the manuscript; M.J. coordinated the project.
Methods
Measurement Procedure
The current pulse trains were generated by a Keithley 2430 Source Meter. After each pulse train a delay of $10$ s for thermal relaxation was followed by a measurement of the transversal or longitudinal voltage across the central part of the patterned structure resulting from a probe current density of $10^{4}$ A/cm${}^{2}$. Typically this procedure was repeated several times before the pulse trains were sent along the perpendicular direction of the cross structure, keeping the probing contacts unchanged.
Based on time resolved resistivity measurements during the application of pulse trains and temperature dependent resistivity measurements of Mn${}_{2}$Au thin films the local temperature of the relevant sample region was estimated: Values of up to $600$ K in the stable regime and $\simeq 800$ K at current densities, which finally destroyed the samples, were obtained. The local temperature rapidly decays after the application of a current pulse train, therefore thermoelectric voltages can be neglected.
MAE and AMR calculation
The MAE was calculated using the FLAPW (Full Potential Linearized Augmented Plane Wave) method in combination with the GGA (Generalized Gradiend Approximation). We found the MAE in line with previous reports Shi00 $\simeq 1$ ${\rm\mu}$eV per formula unit, which is at the resolution limit of our method based on the magnetic force theorem.
To calculate the AMR in Mn${}_{2}$Au ab initio we employed the fully relativistic Dirac tight binding-linear muffin-tin orbital plus coherent potential approximation (FRD-TB-LMTO+CPA) method in combination with the Kubo formula Tur02 ; Tur14 . We used the s-, p-, and d-type orbitals in the basis and the LSDA (Local Spin Density Approximation) Vosko-Wilk-Nusair exchange-correlation potential parametrization Vos80 . The ground-state magnetization and density of states was reproduced consistently with a previous reports Shi10 ; Zel14 . In the transport calculation we used up to $10^{8}$ k-points in the Brillouin zone and for the residual resistivity calculations we set the imaginary part of the complex energy to $\Gamma=10^{-5}$ Ry.
References
(1)
Jungwirth, T., Marti, X., Wadley, P. & Wunderlich, J. Antiferromagnetic spintronics. Nat. Nanotech. 11, 231 (2016).
(2)
Kampfrath, T. et al. Coherent terahertz control of antiferromagnetic spin waves. Nat. Phot. 5, 31 (2011).
(3)
Zelezny, J. et al. Relativistic Neel-order fields induced by electrical current in antiferromagnets. Phys. Rev. Lett. 113, 157201 (2014).
(4)
Wadley, P. et al. Electrical switching of an antiferromagnet. Science 351, 587 (2016).
(5)
Park, B. G., et al. A spin-valve-like magnetoresistance of an antiferromagnet-based tunnel junction. Nat. Mat. 10, 347 (2011).
(6)
Fina, I. et al. Anisotropic magnetoresistance in an antiferromagnetic semiconductor. Nat. Commun. 5, 4671 (2014).
(7)
Marti, X. et al. Room-temperature antiferromagnetic memory resistor. Nat. Mat. 13, 367 (2014).
(8)
Gambardella, P., and Miron, I. M. Current-induced spin-orbit torques. Phil. Trans. R. Soc. A 369 3175, (2011).
(9)
Brataas, A. & Hals, K. M. D. Spin-orbit torques in action. Nat. Nanotech. 9, 86 (2014).
(10)
Wadley, P. et al. Tetragonal phase of epitaxial room-temperature antiferromagnet CuMnAs. Nat. Commun. 4, 2322 (2013).
(11)
Grzybowski, M. J. et al. Imaging Current-Induced Switching of Antiferromagnetic Domains in CuMnAs. Phys. Rev. Lett. 118, 057701 (2017).
(12)
Shick, A. B., Khmelevskyi, S., Mryasov, O. N., Wunderlich, J. & Jungwirth, T. Spin-orbit coupling induced anisotropy effects in bimetallic antiferromagnets: A route towards antiferromagnetic spintronics. Phys. Rev B 81, 212409 (2010).
(13)
Barthem, V. M. T. S., Colin, C. V., Mayaffre, H., Julien, M.-H.
& D. Givord, Revealing the properties of Mn${}_{2}$Au for antiferromagnetic spintronics. Nat. Commun. 4, 2892 (2013)
(14)
Barthem, V. M. T. S. et al. Easy moment direction and antiferromagnetic domain wall motion in Mn2Au. J. Magn. Magn. Mater. 406, 289 (2016).
(15)
Han-Chun Wu et al. Mn${}_{2}$Au: Body-Centered-Tetragonal Bimetallic Antiferromagnets Grown by Molecular Beam Epitaxy. Adv. Mat. 24, 6374 (2012).
(16)
Han-Chun Wu et al. Anomalous Anisotropic Magnetoresistance of Antiferromagnetic Epitaxial Bimetallic Films: Mn${}_{2}$Au and Mn${}_{2}$Au/Fe Bilayers. Adv. Funct. Mat. 201601348 (2016)
(17)
Jourdan, M. et al. Epitaxial Mn${}_{2}$Au thin films for antiferromagnetic spintronics. J. Phys. D: Appl. Phys. 48, 385001 (2015).
(18)
Thompson, D. A., Romankiw, L. T., & Mayadas, A. F. Thin film magnetoresistors in memory, storage, and related applications. IEEE Trans. Magn. 11, 1039 (1975).
(19)
Seemann, K. M. et al. Origin of the planar Hall effect in nanocrystalline Co60Fe20B20. Phys. Rev. Lett. 107, 086603 (2011).
(20)
Roy, P. E., Otxoa, R. M. & Wunderlich, J. Robust picosecond writing of a layered antiferromagnet by staggered spin-orbit fields. Phys. Rev. B 94, 014439 (2016).
(21)
Turek, I., Kudrnovsky, J. & Drchal, V. Ab initio theory of galvanomagnetic phenomena in ferromagnetic metals and disordered alloys. Phys. Rev. B 86, 014405 (2012).
(22)
De Ranieri, E. et al. Lithographically and electrically controlled strain effects on anisotropic magnetoresistance in (Ga,Mn)As. New J. Phys., 10, 065003, (2008).
(23)
Shick, A. B., Khmelevskyi, S., Mryasov, O. N., Wunderlich, J. & Jungwirth, T. Spin-orbit coupling induced anisotropy effects in bimetallic antiferromagnets: A route towards antiferromagnetic spintronics, Phys. Rev. B, 81, 212409 (2010).
(24)
Turek, I., Kudrnovsky, J., Drchal, V., Szunyogh, L. & Weinberger, P. Interatomic electron transport by semiempirical and ab initio tight-binding approaches. Phys. Rev. B 65, 125101, (2002).
(25)
Turek, I., Kudrnovsky, J. & Drchal, V. Fermi sea term in the relativistic linear muffn-tin-orbital transport theory for random alloys. Phys. Rev. B 89, 064405, (2014).
(26)
Vosko, S. H., Wilk, L. and Nusair, M. Accurate spin-dependent electron liquid correlation energies for local spin density calculations: a critical analysis. Can. J. Phys., 58, 1200 (1980). |
WIS/02/16-FEB-DPPA
IPMU-16-0022
UT-16-9
S-folds and 4d $\mathcal{N}{=}\,3$ superconformal field theories
Ofer Aharony${}^{1}$ and Yuji Tachikawa${}^{2,3}$
1.
Department of Particle Physics and Astrophysics,
Weizmann Institute of Science, Rehovot 7610001, Israel
2.
Department of Physics, Faculty of Science,
University of Tokyo, Bunkyo-ku, Tokyo 113-0033, Japan
3.
Kavli Institute for the Physics and Mathematics of the Universe,
University of Tokyo, Kashiwa, Chiba 277-8583, Japan
S-folds are generalizations of orientifolds in type IIB string theory, such that the geometric identifications are accompanied by non-trivial S-duality transformations.
They were recently used by García-Etxebarria and Regalado to provide the first construction of four dimensional ${\cal N}{=}3$ superconformal theories.
In this note, we classify the different variants of these ${\cal N}{=}3$-preserving S-folds, distinguished by an analog of discrete torsion,
using both a direct analysis of the different torsion classes and the compactification of the S-folds to three dimensional M-theory backgrounds.
Upon adding D3-branes, these variants lead to different
classes of ${\cal N}{=}3$ superconformal field theories. We also analyze the holographic
duals of these theories, and in particular clarify the role of discrete gauge and global
symmetries in holography.
Contents
1 Introduction and summary of results
2 Preliminary analysis
2.1 Three general properties of ${\cal N}=3$ SCFTs
2.2 D3-branes on S-folds and their discrete symmetries
3 Holographic dual
3.1 AdS duals of CFTs with discrete symmetries
3.2 Discrete torsion
3.3 Generalized Pfaffians
3.4 Realization of the $\mathbb{Z}_{k}$ gauge theory in F-theory
4 Comparison to M-theory
4.1 Discrete fluxes in M-theory on $(\mathbb{C}^{3}\times T^{2})/\mathbb{Z}_{k}$
4.2 M2-brane and D3-brane charges
5 Special cases and $\mathcal{N}{=}\,4$ enhancement
1 Introduction and summary of results
Field theories with superconformal symmetry are useful laboratories for learning about the behavior of quantum field theories in general, and strongly coupled field theories in particular. This is because the superconformal symmetry allows many computations to be performed in these theories, using methods such as localization, integrability, and the superconformal bootstrap.
The field theories (above two dimensions) about which the most is known are ${\cal N}=4$ superconformal field theories (SCFTs) in four dimensions, that have been called the ‘harmonic oscillator of quantum field theories’. These theories have an exactly marginal deformation, and it is believed that they are all gauge theories with some gauge group $G$, such that the exactly marginal deformation is the gauge coupling constant. In these theories many observables have already been computed as functions (trivial or non-trivial) of the coupling constant, and there is a hope that they can be completely solved.
Four dimensional theories with ${\cal N}=2$ superconformal symmetry have also been extensively studied. Some of these theories have exactly marginal deformations and corresponding weak coupling limits (at least for some sector of the theory), while others do not. We do not yet have a full classification of ${\cal N}=2$ SCFTs (for rank-1 theories, see [1, 2, 3]),
though a large class of theories, called class S, has been constructed following [4].
Many observables can be computed also in ${\cal N}=2$ SCFTs, at least if they have a weak coupling limit, they are connected by renormalization group flows to theories that have such limits, or they are one of the theories of class S.
Four dimensional ${\cal N}=3$ theories should naively provide an intermediate class of theories, that is more general than ${\cal N}=4$, but such that more computations can be made than in general ${\cal N}=2$ theories. ${\cal N}=3$ SCFTs (that are not also ${\cal N}=4$ SCFTs) have no exactly marginal deformations [5, 6] and thus no weak coupling limits that would aid in classifying and performing computations in these theories. Until recently no ${\cal N}=3$ SCFTs were known, but recently a class of such theories was constructed by García-Etxebarria and Regalado in [7]. Their construction uses a generalization of orientifolds in string theory.
Orientifold 3-planes (the generalization to other dimensions is straightforward) are defined in type IIB string theory as planes in space-time, such that in the transverse space $y\in\mathbb{R}^{6}$ to these orientifolds, there is an identification between the points $y$ and $(-y)$, but with opposite orientations for strings (or, equivalently, with an opposite value for the $B_{2}$ and $C_{2}$ 2-form potentials of type IIB). This means that we identify configurations related by a $\mathbb{Z}_{2}$ symmetry that involves a spatial reflection in the transverse $SO(6)$, and also a transformation $(-I)$ in the $SL(2,\mathbb{Z})$ S-duality group of type IIB string theory. This breaks half of the supersymmetry, preserving a four dimensional ${\cal N}=4$ supersymmetry. In particular, putting $N$ D3-branes on the orientifold (these do not break any extra supersymmetries) gives at low energies four dimensional ${\cal N}=4$ SCFTs.
In [7] this was generalized to identifying configurations related by a $\mathbb{Z}_{k}$ symmetry, that acts both by a $(2\pi/k)$ rotation in the three transverse coordinates in $\mathbb{C}^{3}=\mathbb{R}^{6}$, and by an element of $SL(2,\mathbb{Z})$ whose $k$’th power is the identity. We will call the fixed planes of such transformations S-folds111This term, first coined in [8], generalizes the term T-folds that is used to describe identifications by elements of the T-duality group. S-folds of string theory involving S-duality twists together with shifts along a circle were studied, for instance, in [9, 10, 11, 12, 13], and similar S-folds were studied in the 4d $\mathcal{N}{=}\,4$ SYM theory in [14, 15, 16].; for $k=2$ they are the same as the usual orientifolds. Viewing $SL(2,\mathbb{Z})$ as the modular group of a torus, the $\mathbb{Z}_{k}$ S-duality transformation may be viewed as a rotation of the torus by an angle $(2\pi/k)$; such a rotation maps the torus to itself if and only if $k=3,4,6$ and its modular parameter is $\tau=e^{2i\pi/k}$, so S-folds of this type exist only for these values of $k$ and $\tau$. It is natural to define such an identification using F-theory [17], in which the $SL(2,\mathbb{Z})$ S-duality group is described as adding an extra zero-size torus whose modular parameter is the coupling constant $\tau$ of type IIB string theory; in this language the S-folds of [7] are the same as F-theory on $(\mathbb{C}^{3}\times T^{2})/\mathbb{Z}_{k}$. This specific $\mathbb{Z}_{k}$ identification preserves a four dimensional ${\cal N}=3$ supersymmetry. So, putting $N$ D3-branes on the S-fold gives at low energies theories with ${\cal N}=3$ superconformal symmetry. Note that D3-branes sitting on the S-fold are invariant under all the transformations discussed above, and, in particular, the D3-brane charge of the S-fold is well-defined.
In this note we analyze further the theories constructed in [7]. We focus on asking what are the extra parameters associated with S-folds, analogous to discrete torsion for orbifolds and orientifolds (namely, to non-trivial configurations of the various $p$-forms of type IIB string theory in the presence of the S-fold). S-folds with different parameters can carry different D3-brane charges. Upon putting $N$ D3-branes on the S-fold, these discrete parameters label different ${\cal N}=3$ SCFTs. Our main result is a classification of these extra parameters (following a preliminary discussion in [7]); we show that there are two variants of S-folds with $k=3,4$ and just a single variant with $k=6$. Each variant leads to different ${\cal N}=3$ SCFTs, with different central charges and chiral operators.
In some cases discrete global symmetries play an interesting role. In a specific type of orientifold (the $O3^{-}$ plane), the theory of $N$ D3-branes on the orientifold is an $\mathrm{O}(2N)$ gauge theory, and this may be viewed as an $\mathrm{SO}(2N)$ gauge theory in which a discrete global $\mathbb{Z}_{2}$ symmetry is gauged. Both the $\mathrm{SO}(2N)$ and the $\mathrm{O}(2N)$ theories exist as ${\cal N}=4$ SCFTs, and a few of their properties are different. We will see that a similar phenomenon happens also for D3-branes on S-folds, and discuss its realization in the AdS/CFT dual of these theories, in terms of discrete gauge symmetries in the bulk. This dual is obviously given by F-theory on $AdS_{5}\times(S^{5}\times T^{2})/\mathbb{Z}_{k}$.
We begin in section 2 with a naive analysis of the possible discrete parameters, by considering the possible discrete identifications on the moduli space of $N$ D3-branes, and the effects of discrete symmetries. In section 3 we discuss the holographic duals of the ${\cal N}=3$ SCFTs, and the realization of the discrete symmetries there. We show that holography suggests that only some of the possibilities found in the naive analysis are consistent. The compactification of F-theory on $(\mathbb{C}^{3}\times T^{2})/\mathbb{Z}_{k}$ on a circle gives M-theory on $(\mathbb{C}^{3}\times T^{2})/\mathbb{Z}_{k}$, with each S-fold splitting into several $\mathbb{C}^{4}/\mathbb{Z}_{l}$ singularities in M-theory. In section 4 we show that the consistency of this reduction implies that indeed only the possibilities found in section 3 are consistent. In section 5 we discuss the fact that for specific values of $k$ and $N$ some variants of the ${\cal N}=3$ theories have enhanced ${\cal N}=4$ supersymmetry, with gauge groups $SU(3)$, $SO(5)$ and $G_{2}$, and check the consistency of this. This provides new brane constructions and new (strongly coupled) AdS duals for these specific ${\cal N}=4$ SCFTs.
In this paper we discuss only the S-folds which give rise to four dimensional ${\cal N}=3$ supersymmetric theories. It would be interesting to study S-folds that preserve different amounts of supersymmetry, and that have different dimensions. A particularly interesting case, which should have many similarities to our discussion, is S-folds preserving four dimensional ${\cal N}=2$ supersymmetry. We leave the study of these theories to the future.
2 Preliminary analysis
In this paper we study the S-folds introduced by
García-Etxebarria and Regalado in [7], which are equivalent to F-theory on $(\mathbb{C}^{3}\times T^{2})/\mathbb{Z}_{k}$ for $k=3,4,6$, generalizing the standard orientifold 3-planes that arise for $k=2$.
Just as the orientifold 3-planes have four variants, $O3^{-}$, $\widetilde{O3}{}^{-}$, $O3^{+}$ and $\widetilde{O3}{}^{+}$, that differ by discrete fluxes, we expect that these new S-folds could also have a few variants.
In this section we perform a preliminary analysis of the possible variants,
from the viewpoint of the moduli space of the $\mathcal{N}{=}\,3$ superconformal field theories realized on $N$ D3-branes probing these S-folds. In the following sections we will discuss additional constraints on the possible variants.
2.1 Three general properties of ${\cal N}=3$ SCFTs
We start from the following three properties of theories of this type, that are generally believed to hold:
•
The conformal anomalies (central charges) $a$ and $c$ are equal in any $\mathcal{N}{=}\,3$ SCFT [5]:
$$a=c.$$
(2.1)
The value of $a=c$ determines also all anomalies of the $SU(3)_{R}\times U(1)_{R}$ symmetries in these theories.
•
The geometry of the gravitational dual of $N$ D3-branes sitting on a singularity involving a $\mathbb{Z}_{k}$ identification of their transverse space, and carrying $\epsilon$ units of D3-brane charge, is AdS${}_{5}\times S^{5}/\mathbb{Z}_{k}$, with $(N+\epsilon)$ units of 5-form flux. This implies that the large $N$ central charge of the corresponding CFTs is
$$a\sim c\sim k(N+\epsilon)^{2}/4+O(N^{0}).$$
(2.2)
•
In any $\mathcal{N}{=}\,2$ theory, the Coulomb branch operators (chiral operators whose expectation values label the Coulomb branch of the moduli space, namely the space of vacuum expectation values of the scalars in the vector multiplets) form a ring generated by $n$ operators, whose expectation values are all independent without any relation. Furthermore, up to a caveat mentioned below, there is a relation
$$2a-c=\sum_{i=1}^{n}(2\Delta_{i}-1)/4,$$
(2.3)
where $\Delta_{i}$ are the scaling dimensions of the generators of the Coulomb branch operators.
The property (2.2) can be derived easily following the analysis of [18, 19, 20].
Essentially, the curvature of the AdS space is supported by the total energy of the five-form field, and the volume of the $\mathbb{Z}_{k}$ quotient is $1/k$ of the volume before the quotient, changing the value of Newton’s constant on $AdS_{5}$ by a factor of $k$.
The equation (2.3), originally conjectured in [21], was given a derivation that applies to a large subclass of $\mathcal{N}{=}\,2$ theories in [22] (though it is not clear that this subclass includes the theories we discuss here). More precisely, this relation applies only to gauge theories whose gauge group has no disconnected parts.
To illustrate this, consider the $\mathcal{N}{=}\,4$ super Yang-Mills theories with gauge groups $\mathrm{U}(1)$ and $\mathrm{O}(2)$.
They have the same central charges $2a-c=1/4$, coming from a single free vector multiplet, but the gauge-invariant Coulomb branch operator has dimension $1$ for the former and $2$ for the latter,
so that property 3 only holds for the former. This is because the scalars in the vector multiplet change sign under the disconnected component of the $\mathrm{O}(2)$ gauge group. The $\mathrm{O}(2)$ theory is obtained by gauging a $\mathbb{Z}_{2}$ global symmetry of the $\mathrm{U}(1)=\mathrm{SO}(2)$ theory, where the $\mathbb{Z}_{2}$ symmetry is generated by the quotient $\mathrm{O}(2)/\mathrm{SO}(2)$; this operation does not change the central charges.
This means that the $\mathrm{O}(2)$ theory includes a $\mathbb{Z}_{2}$ gauge theory, implying that it has a non-trivial $\mathbb{Z}_{2}$ 2-form global symmetry in the sense of [23], as we further discuss below.
We assume in the following that the relation (2.3) holds whenever the theory does not have any non-trivial 2-form global symmetry, even if the theory does not have a gauge theory description.
2.2 D3-branes on S-folds and their discrete symmetries
Let us now see what the three properties recalled above tell us about the properties of S-folds.
Let us probe the $\mathbb{Z}_{k}$ S-fold by $N$ D3-branes. The moduli space of ${\cal N}=3$ theories is described by the expectation values of scalars in ${\cal N}=3$ vector multiplets, which are identical to ${\cal N}=4$ vector multiplets; viewing an ${\cal N}=3$ theory as an ${\cal N}=2$ theory, these decompose into a vector multiplet and a hypermultiplet.
Let us choose a specific $\mathcal{N}{=}\,2$ subgroup of the $\mathcal{N}{=}\,3$ symmetry, and consider the component of the moduli space which is a Coulomb branch from the point of view of this $\mathcal{N}{=}\,2$ subgroup. This implies that all the D3-branes
lie on a particular $\mathbb{C}/\mathbb{Z}_{k}$ within $\mathbb{C}^{3}/\mathbb{Z}_{k}$.
Denote by $z_{i}$ $(i=1,\ldots,N)$ the positions on $\mathbb{C}$ of these D3-branes.
Since the D3-branes are identical objects, the identifications that are imposed on the moduli space by the geometry are
$$\displaystyle z_{i}$$
$$\displaystyle\leftrightarrow z_{j},$$
$$\displaystyle\text{all other $z_{a}$ fixed},$$
$$\displaystyle z_{i}$$
$$\displaystyle\mapsto\gamma z_{i},$$
$$\displaystyle\text{all other $z_{j}$ fixed},$$
(2.4)
where $\gamma\equiv\exp(2\pi\mathrm{i}/k)$. For the purpose of the moduli space identifications we can ignore the additional S-duality transformation.
From these identifications we can see that the gauge-invariant Coulomb branch operators are the symmetric polynomials of $z_{i}{}^{k}$, and their generators are ($\sum_{i=1}^{N}z_{i}{}^{jk}$) for $j=1,\cdots,N$, with dimensions
$$k,2k,\cdots,Nk.$$
(2.5)
From (2.1) and (2.3), this naively implies that
$$4a=4c=kN^{2}+(k-1)N.$$
(2.6)
Applying (2.2), we find that the D3-brane charge of the $\mathbb{Z}_{k}$ S-fold is $\epsilon=(k-1)/(2k)$.
As discussed above, we know that even for $k=2$ this is not the whole story, due to the possibility of discrete torsion and discrete symmetries.
Let us review in detail this $k=2$ case, in a language that will be useful for our later analysis.
Among the known O3-planes, the above analysis only applies to $\widetilde{O3}{}^{-}$, $O3^{+}$ and $\widetilde{O3}{}^{+}$, for which $\epsilon=(k-1)/(2k)=+1/4$ is the correct value.
When the orientifold is $O3^{-}$, the $\mathrm{O}(2N)$ gauge group arising on $N$ D3-branes can be viewed as a $\mathbb{Z}_{2}$ gauging of the $\mathrm{SO}(2N)$ gauge theory, and thus, to compute $a$ and $c$, we need to use the $\mathrm{SO}(2N)$ theory instead of the $\mathrm{O}(2N)$ theory. In the $\mathrm{SO}(2N)$ theory
the residual gauge symmetries on the Coulomb branch, acting on the eigenvalues of a matrix in the adjoint representation of $\mathrm{SO}(2N)$, are generated by
$$\displaystyle z_{i}$$
$$\displaystyle\leftrightarrow z_{j},$$
$$\displaystyle\text{all other $z_{a}$ fixed},$$
$$\displaystyle(z_{i},z_{j})$$
$$\displaystyle\mapsto(-z_{i},-z_{j}),$$
$$\displaystyle\text{all other $z_{a}$ fixed}.$$
(2.7)
The dimensions of the independent gauge-invariant operators are then
$$2,4,\cdots,2(N-1);N,$$
(2.8)
where the last invariant is the Coulomb branch expression for the Pfaffian of an adjoint matrix of $SO(2N)$,
$z_{1}z_{2}\ldots z_{N}$.
From this list we can compute $a$ and $c$ again using (2.1) and (2.3) and find
$$4a=4c=2N^{2}-N,$$
(2.9)
from which we find $\epsilon=-1/4$ from (2.2).
This is indeed the correct value for $O3^{-}$.
The $\mathrm{SO}(2N)$ theory has a discrete $\mathbb{Z}_{2}$ global symmetry, corresponding to gauge transformations by elements of $\mathrm{O}(2N)$ which are not elements of $\mathrm{SO}(2N)$. Note that in the $\mathrm{SO}(2N)$ theory these are global symmetries rather than gauge symmetries. On the moduli space one of these transformations acts as $z_{1}\to-z_{1}$, and with this extra identification the group generated by (2.7) becomes the same as the group (2.4) acting on the D3-branes. In particular, this identification projects out the Pfaffian operator of dimension $N$, such that after it we obtain the Coulomb branch operators with dimensions (2.5). So, for this specific orientifold plane, the theory on the D3-branes is described by a $\mathbb{Z}_{2}$ quotient of some ‘parent’ theory which has a different group of identifications (2.7), and correspondingly different Coulomb branch operators. Only after gauging a $\mathbb{Z}_{2}$ global symmetry of this ‘parent’ theory do we get the theory of the D3-branes on the $O3^{-}$ orientifold.
We stress here that this gauging of $\mathbb{Z}_{2}$ symmetry does not change the central charges.
It is natural to ask how we could know directly from the theory of the D3-branes on the $O3^{-}$ plane that its central charges are not the same as the naive ones, because it arises as a $\mathbb{Z}_{2}$ gauging of some other theory. As discussed in [24, 23], when we have a discrete $\mathbb{Z}_{p}$ global symmetry, we have local operators that transform under this symmetry, as well as 3-plane operators that describe domain walls separating vacua that differ by a $\mathbb{Z}_{p}$ transformation. When we gauge the $\mathbb{Z}_{p}$ global symmetry, these local and 3-plane operators disappear from the spectrum. Instead we obtain new 2-plane operators (that may be viewed as worldvolumes of strings), characterized by having a $\mathbb{Z}_{p}$ gauge transformation when we go around them. These 2-plane operators are charged under a 2-form $\mathbb{Z}_{p}$ global symmetry in the language of [23]. So whenever we have a theory with a 2-form $\mathbb{Z}_{p}$ global symmetry, it is natural to expect that it arises by gauging a $\mathbb{Z}_{p}$ global symmetry of some ‘parent’ theory. And indeed, the analysis of [23] implies that this ‘parent’ theory can be obtained by gauging the 2-form $\mathbb{Z}_{p}$ global symmetry. Thus, whenever we have a theory with a $\mathbb{Z}_{p}$ 2-form global symmetry, we expect that its central charges would not be given by (2.3), but rather by those of its ‘parent’ theory222Gauging a discrete $\mathbb{Z}_{p}$ global symmetry does not change the dynamics on $\mathbb{R}^{4}$, but it does change the spectrum of local and non-local operators..
Suggested by this analysis for $k=2$, we expect that also for $k=3,4,6$, different versions of S-folds will be characterized by different 2-form $\mathbb{Z}_{p}\subset\mathbb{Z}_{k}$ global symmetries for the corresponding theories on the D3-branes, that will imply that these theories arise from ‘parent’ theories with $\mathbb{Z}_{p}$ global symmetries. The analysis above suggests that the identifications on the moduli spaces of these ‘parent’ theories should be subgroups of the group generated by (2.4), such that the ring of invariants is polynomial without any relation, and such that adding an extra $\mathbb{Z}_{p}$ generator produces the group (2.4).
Luckily such groups are already classified and are known as Shephard-Todd complex reflection groups; the fact that the ring of invariants is polynomial without any relation if and only if the group acting on $\mathbb{C}^{N}$ is a complex reflection group is known as the theorem of Chevalley, Shephard and Todd.333See [25] for a recent use of Shephard-Todd complex reflection groups in four-dimensional $\mathcal{N}{=}\,2$ theories. For more on complex reflection groups, see [26].
In our case, the available groups are known as $G(N,p,k)$, and are generated by the following elements:
$$\displaystyle z_{i}$$
$$\displaystyle\leftrightarrow z_{j},$$
$$\displaystyle\text{all other $z_{a}$ fixed},$$
$$\displaystyle(z_{i},z_{j})$$
$$\displaystyle\mapsto(\gamma z_{i},\gamma^{-1}z_{j}),$$
$$\displaystyle\text{all other $z_{a}$ fixed},$$
$$\displaystyle z_{i}$$
$$\displaystyle\mapsto\gamma^{p}z_{i},$$
$$\displaystyle\text{all other $z_{j}$ fixed},$$
(2.10)
where $p$ is a divisor of $k$. By adding another $\mathbb{Z}_{p}$ generator acting as $z_{i}\mapsto\gamma z_{i}$, all these groups become the groups (2.4).
Denoting $\ell=k/p$, we see that if there is a ‘parent’ theory with the identifications (2.10) on its moduli space, then its gauge-invariant Coulomb branch operators would be generated by the symmetric polynomials of $z_{i}^{k}$ and by $(z_{1}z_{2}\cdots z_{N})^{\ell}$, and therefore the dimensions of the Coulomb branch generators would be given by
$$k,2k,\ldots,(N-1)k;N\ell.$$
(2.11)
From (2.1) and (2.3) we then find that the central charges of the ‘parent’ theory, and also of its $\mathbb{Z}_{p}$ quotient that would describe the theory of $N$ D3-branes on the corresponding S-fold, are given by
$$4a=4c=kN^{2}+(2\ell-k-1)N.$$
(2.12)
Note that even though the $\mathbb{Z}_{p}$-gaugings of the theories with different values of $\ell$ all have the same moduli space, they are distinct theories with different central charges. When $p$ is not prime, one can also gauge subgroups of $\mathbb{Z}_{p}$, giving rise to additional theories, which again are not equivalent to the theories for other values of $\ell$ (even though they may have the same moduli space).
Therefore, it seems at this stage, that each $\mathbb{Z}_{k}$ S-fold with $k=2,3,4,6$ can have variants labeled by $\ell$ which is an integer dividing $k$.
The D3-brane charges $\epsilon_{k,\ell}$ of these S-folds can be easily found using (2.2) and (2.12):
$$\begin{array}[]{c||c|c|c|c|ccc}&\ell=1&\ell=2&\ell=3&\ell=4&\ell=6\\
\hline\cr\hline\cr k=2&\pagecolor{black!10}-1/4&\pagecolor{black!10}+1/4&&\\
k=3&\pagecolor{black!10}-1/3&&\pagecolor{black!10}+1/3&&\\
k=4&\pagecolor{black!10}-3/8&-1/8&&\pagecolor{black!10}+3/8&\\
k=6&\pagecolor{black!10}-5/12&-1/4&-1/12&&+5/12\end{array}$$
(2.13)
Note that just from the analysis above it is far from clear that S-folds and ${\cal N}=3$ theories corresponding to such identifications actually exist. However, we expect any S-fold to fall into one of these categories.
In the next two sections, we will see that the S-fold variants that really exist are only those shaded in the table (2.13) above, first by carefully studying the holographic dual and then by comparing with M-theory. Note also that there could be more than one theory with the same identifications. For $k=2$ there are three orientifold-types with the same $\ell=2$ identifications, though they all give rise to the same theory on D3-branes because they are all related by S-duality. We do not know the corresponding situation in our case. We also cannot rule out the existence of ${\cal N}=3$ theories having Coulomb branch operators (2.11) also for the values of $\ell$ and $k$ that do not come from S-folds. It would be interesting to shed further light on these questions, perhaps by a conformal bootstrap analysis of ${\cal N}=3$ SCFTs, or by analyzing the two dimensional chiral rings of the corresponding theories [27, 28].
3 Holographic dual
3.1 AdS duals of CFTs with discrete symmetries
Theories of quantum gravity are not expected to have any global symmetries (see, for instance, [24]). There are very strong arguments that this is the case for continuous symmetries, and it is believed to be true also for discrete symmetries. In the AdS/CFT correspondence, this implies that any global symmetry in the CFT should come from a gauge symmetry in anti-de Sitter (AdS) space.
For continuous global symmetries it is known that indeed they come from gauge fields in the bulk, with a boundary condition that sets their field strength to zero on the boundary. In some cases (like AdS${}_{4}$ [29]) there is also another consistent boundary condition in which the field strength does not go to zero on the boundary, and its boundary value becomes a gauge field in the CFT. Changing the boundary condition may be interpreted as gauging the global symmetry in the CFT (in cases where this gauging still gives a CFT).
In this section we discuss the analogous statements for discrete symmetries. We focus on the $AdS_{5}\leftrightarrow$ CFT${}_{4}$ case, but the generalization to other dimensions is straightforward. The discussion here is a special case of the discussion in appendix B of [23], generalized to five dimensions, but as far as we know its implications for the AdS/CFT correspondence were not explicitly written down before (see also [30]).
Consider a $\mathbb{Z}_{k}$ gauge symmetry on $AdS_{5}$ (thought of as a sector of a full theory of quantum gravity on $AdS_{5}$, that is dual to a 4d CFT). A universal way to describe such a gauge symmetry in five space-time dimensions is by a topological theory of a 1-form $A$ and a 3-form $C$, with an action
$${\cal L}={{ik}\over{2\pi}}A\wedge dC.$$
(3.1)
$A$ can be thought of as the gauge field for a $\mathbb{Z}_{k}$ symmetry; for example, the action above arises from a $U(1)$ gauge
symmetry that is spontaneously broken to $\mathbb{Z}_{k}$.
The forms $A$ and $C$ are both gauge fields, whose field strengths $dA$ and $dC$ vanish by the equations of motion. There are gauge transformations that shift the integrals of $A$ and $C$ over closed cycles by one. A gauge symmetry is just a redundancy in our description, but an invariant property of this theory is that it has line and 3-surface operators, given by $e^{i\oint A}$ and $e^{i\oint C}$, with $A$ and $C$ integrated over closed cycles. And, it has a 1-form $\mathbb{Z}_{k}$ global symmetry that multiplies the line operators $e^{i\oint A}$ by $e^{2\pi i/k}$, and a similar 3-form global symmetry.
When we put such a theory on $AdS_{5}$, we need to choose boundary conditions; the possible boundary conditions for such topological gauge theories were discussed in [23]. The variational principle implies that we need to set to zero $A\wedge C$ along the boundary. If we set to zero $C$ along the boundary, then the boundary value of $A$ gives a gauge field in the dual CFT, corresponding to a $\mathbb{Z}_{k}$ gauge symmetry in this CFT. With this boundary condition line operators in the bulk are allowed to approach the boundary and to become line operators in the CFT, while 3-surface operators cannot approach the boundary, but can end on the boundary, giving 2-surface operators in the CFT. This is as expected for a $\mathbb{Z}_{k}$ gauge symmetry in four space-time dimensions.
On the other hand, if we set $A$ to zero on the boundary, we obtain a $\mathbb{Z}_{k}$ global symmetry in the dual CFT. The line operators ending on the boundary now give local operators in the CFT, and the 1-form global symmetry in the bulk becomes a standard $\mathbb{Z}_{k}$ global symmetry in the CFT, under which these local operators are charged. The 3-surface operators going to the boundary give 3-surface operators in the CFT, which are domain walls between different vacua related by $\mathbb{Z}_{k}$.
Thus, whenever we have a $\mathbb{Z}_{k}$ gauge symmetry on $AdS_{5}$, there are two natural boundary conditions. One of them gives a conformal field theory with a $\mathbb{Z}_{k}$ global symmetry, and the other gives a conformal field theory with a $\mathbb{Z}_{k}$ gauge symmetry (and a $\mathbb{Z}_{k}$ 2-form global symmetry). The second theory is related to the first one by gauging its $\mathbb{Z}_{k}$ global symmetry, and similarly the first one arises from the second by gauging its $\mathbb{Z}_{k}$ 2-form global symmetry.
When $k$ is not prime, there are also additional possible boundary conditions, corresponding to gauging subgroups of $\mathbb{Z}_{k}$.
In the context of our discussion in the previous section, this implies that the ‘parent’ theories and their $\mathbb{Z}_{k}$-gaugings should arise from the same holographic dual, just with different boundary conditions for the $\mathbb{Z}_{k}$ gauge fields.
We will see in Sec. 3.4 how the $\mathbb{Z}_{k}$ gauge fields of (3.1) arise in F-theory on $AdS_{5}\times(S^{5}\times T^{2})/\mathbb{Z}_{k}$ from integrals of the type IIB five-form field $F_{5}$ on discrete cycles.
Note that all this applies already to the $AdS_{5}\times S^{5}/\mathbb{Z}_{2}$ case discussed in [31]; typically in that case only the option of having a global $\mathbb{Z}_{2}$ symmetry, leading to the $\mathrm{SO}(2N)$ gauge theory, is discussed.
Finally, note that a very similar story occurs already in type IIB string theory on $AdS_{5}\times S^{5}$, which includes a topological sector in the bulk corresponding to a 1-form $\mathbb{Z}_{N}$ gauge symmetry, with an action ${iN\over 2\pi}B_{2}\wedge dC_{2}$ (where $B_{2}$ and $C_{2}$ are the 2-form fields of type IIB string theory) [32, 33]; this topological theory was discussed in section 6 of [23]. In this case both simple boundary conditions give rise to a 1-form $\mathbb{Z}_{N}$ global symmetry in the dual CFT, and the resulting theories are $SU(N)$ and $SU(N)/\mathbb{Z}_{N}$ gauge theories [34, 35, 23]. In this specific case there is also an option of coupling these theories to a continuous $U(1)$ gauge symmetry, leading to a $U(N)$ theory [36, 37]. One can also quantize the topological theory in ways that do not lead to local field theories [38].
3.2 Discrete torsion
In [31], Witten showed how to characterize the variants of O3-planes by studying the discrete torsion on $S^{5}/\mathbb{Z}_{2}$.
In this section we generalize this to $S^{5}/\mathbb{Z}_{k}$ S-folds, for $k=3,4,6$.
For $k=2$, the discrete torsion of the NSNS and of the RR three-form field strengths takes values in $H^{3}(S^{5}/\mathbb{Z}_{2},\tilde{\mathbb{Z}})$, where the tilde over $\mathbb{Z}$ means that the coefficient system is multiplied by $(-1)$ when we go around the $\mathbb{Z}_{2}$ torsion 3-cycle of $S^{5}/\mathbb{Z}_{2}$.
For $k=3,4,6$, we have a $\mathbb{Z}_{k}$ torsion 3-cycle in $S^{5}/\mathbb{Z}_{k}$, and we have
an action of an element $\rho\in\mathrm{SL}(2,\mathbb{Z})$ on the NS-NS and R-R field strengths in $\mathbb{Z}\oplus\mathbb{Z}$
when we go around the $\mathbb{Z}_{k}$ torsion cycle of $S^{5}/\mathbb{Z}_{k}$. We can choose a specific form for this
$\rho$, given, say, by $\rho=\begin{pmatrix}-1&0\\
0&-1\end{pmatrix}$ for $k=2$, $\rho=\begin{pmatrix}-1&-1\\
1&0\end{pmatrix}$ for $k=3$, $\rho=\begin{pmatrix}0&-1\\
1&0\end{pmatrix}$ for $k=4$, and $\rho=\begin{pmatrix}0&-1\\
1&1\end{pmatrix}$ for $k=6$. Note that the eigenvalues of these matrices are $\gamma$ and $\gamma^{-1}$.
The discrete torsion is then given by $H^{3}(S^{5}/\mathbb{Z}_{k},(\mathbb{Z}\oplus\mathbb{Z})_{\rho})$, and its
computation is standard in mathematics444The method to compute cohomologies with twisted coefficients is explained in e.g. Hatcher [39] Chapter 3.H ; the cell decomposition of generalized lens spaces, of which $S^{5}/\mathbb{Z}_{k}$ is one, is given in the same book, Example 2.43. Section 5.2.1 of Davis-Kirk [40] was also quite helpful. .
In general, $H^{*}(S^{2n-1}/\mathbb{Z}_{k},A)$ where $A$ is a $\mathbb{Z}_{k}$-module is given by the cohomology of the complex
$$C^{0}\xrightarrow{1-t}C^{1}\xrightarrow{1+t+\cdots+t^{k-1}}C^{2}\xrightarrow{1-t}\cdots\xrightarrow{1+t+\cdots+t^{k-1}}C^{2n-2}\xrightarrow{1-t}C^{2n-1}$$
(3.2)
where all $C^{i}\simeq A$, $t$ is the generator of $\mathbb{Z}_{k}$, and the differential $d$ is alternately given by the multiplication by $1-t$ or by $1+t+\cdots+t^{k-1}$.
It is easy to see that $d^{2}=(1-t)(1+t+\cdots+t^{k-1})=1-t^{k}=0$.
When $k=2$, $t$ is just the multiplication by $-1$. Then $1+t=0$ and $1-t=2$, from which we conclude $H^{3}(S^{5}/\mathbb{Z}_{2},(\mathbb{Z}\oplus\mathbb{Z})_{\rho})=\mathbb{Z}_{2}\oplus\mathbb{Z}_{2}$, reproducing four types of O3-planes.
When $k=3,4,6$, the action $\rho$ of the generator of $\mathbb{Z}_{k}$ on $\mathbb{Z}\oplus\mathbb{Z}$ obeys
$1+\rho+\cdots+\rho^{k-1}=0$, and $\det(1-\rho)=3,2,1$ for $k=3,4,6$, respectively, and therefore
$$H^{3}(S^{5}/\mathbb{Z}_{k},(\mathbb{Z}\oplus\mathbb{Z})_{\rho})=\begin{cases}\mathbb{Z}_{3}&(k=3),\\
\mathbb{Z}_{2}&(k=4),\\
\mathbb{Z}_{1}&(k=6).\end{cases}$$
(3.3)
This gives the discrete torsion groups arising from the 3-form fields of type IIB string theory on these S-folds.
For $k=3$, we have three different possibilities, but
two non-trivial elements of $\mathbb{Z}_{3}$ are related by conjugation in $\mathrm{SL}(2,\mathbb{Z})$, so up to S-duality transformations there are just two types of S-folds with $k=3$ (in the same sense that up to S-duality there are just two types of O3-planes that give different theories for the D3-branes on them).
For $k=4$, the cohomology is $\mathbb{Z}_{2}$, and therefore we expect two types of S-folds.
Similarly, for $k=6$, there is only one type of S-fold.
3.3 Generalized Pfaffians
In the holographic duals of the S-folds, the discrete torsion described above corresponds to discrete 3-form fluxes on the $S^{3}/\mathbb{Z}_{k}$ discrete 3-cycle in $S^{5}/\mathbb{Z}_{k}$, and it affects the spectrum of wrapped branes on this 3-cycle. We can use this to match the discussion of the previous subsection to our analysis of Section 2.
For $k=2$ this was analyzed in [31]. There is a $\mathbb{Z}_{2}$-torsion three-cycle of the form $S^{3}/\mathbb{Z}_{2}$ within $S^{5}/\mathbb{Z}_{2}$,
and only when the discrete torsion is zero, i.e. when the 3-plane is $O3^{-}$, we can wrap a D3-brane on this cycle.
By analyzing the properties of this wrapped D3-brane we find that it corresponds to a dimension $N$ operator in the dual CFT,
which can be naturally identified with the Pfaffian operator, that only exists in the $\mathrm{SO}(2N)$ theory but not in $\mathrm{SO}(2N+1)$ or $\mathrm{Sp}(N)$ theories. Note that, according to the discussions above, this operator exists in the ‘parent’ $\mathrm{SO}(2N)$ theory, but not after we gauge the $\mathbb{Z}_{2}$ to get the $\mathrm{O}(2N)$ theory. So in the AdS dual it exists when we choose the boundary condition for the $\mathbb{Z}_{2}$ gauge theory in the bulk that gives a $\mathbb{Z}_{2}$ global symmetry, corresponding to the ‘parent’ theory, but not for the other boundary condition, that corresponds to the theory on the D3-branes.
The obstruction to wrap D3-branes on $S^{3}/\mathbb{Z}_{2}$ can be understood as follows.
The NSNS 3-form flux $G$ can in general be in a non-trivial cohomological class.
But when pulled-back to the worldvolume of a single D3-brane, $G$ is the exterior derivative of a gauge-invariant object $B-F$, where $F$ is the gauge field on the D3-brane,
and therefore the cohomology class $[G]$ should be trivial. The argument for the RR flux is the S-dual of this.
For all $k=3$, $4$, $6$, there is a $\mathbb{Z}_{k}$-torsion three-cycle in $S^{5}/\mathbb{Z}_{k}$ of the form $S^{3}/\mathbb{Z}_{k}$.
When the discrete torsion is zero, there is no obstruction to wrapping a D3-brane on this cycle.
The scaling dimension of this wrapped D3-brane can be easily found to be $kN/k=N$.
This matches the scaling dimensions we found for the $\ell=1$ variants of Section 2. So we identify
the S-fold with no discrete torsion with the $\ell=1$ case (either the ‘parent’ theory, or its $\mathbb{Z}_{k}$ gauging that gives the theory on the D3-branes, depending on the boundary conditions).
For $k=6$, there is nothing more to discuss, since $H^{3}(S^{5}/\mathbb{Z}_{k},(\mathbb{Z}\oplus\mathbb{Z})_{\rho})$ itself is trivial, so we do not find any variant except $\ell=1$.
For $k=3$, when the discrete torsion in $H^{3}(S^{5}/\mathbb{Z}_{k},(\mathbb{Z}\oplus\mathbb{Z})_{\rho})=\mathbb{Z}_{3}$ is non-trivial, we cannot wrap a D3-brane on this cycle. So
this should correspond to the $\ell=3$ variant (with no discrete symmetries).
The most subtle is the $k=4$ case, when the discrete torsion is given by an element in $H^{3}(S^{5}/\mathbb{Z}_{k},(\mathbb{Z}\oplus\mathbb{Z})_{\rho})=\mathbb{Z}_{2}$. We claim
that the non-trivial element of this discrete torsion gives the $\ell=4$ variant of Section 2, not the $\ell=2$ variant.
To see this, it is instructive to recall why in the case $k=2$ we can wrap two D3-branes on $S^{3}/\mathbb{Z}_{2}$ even with the discrete torsion.
Note that it is not just that for $N$ D3-branes wrapping on the same cycle, the triviality of $N\cdot[G]$ suffices. For one thing, if we have two D3-branes wrapping on the same locus with at least $\mathrm{U}(1)\times\mathrm{U}(1)$ unbroken, then $B-F_{1}$ and $B-F_{2}$ are both gauge-invariant (where $F_{1,2}$ are the $\mathrm{U}(1)$ gauge fields coming from the first and the second Chan-Paton factor), and therefore $[G]$ still needs to be zero.
To wrap two D3-branes on $S^{3}/\mathbb{Z}_{2}$ with discrete torsion consistently, one needs to require that the Chan-Paton indices 1,2 are interchanged when we go around the $\mathbb{Z}_{2}$ cycle. This means that in fact there is a single connected D3-brane of the shape $S^{3}$, which is wrapped on $S^{3}/\mathbb{Z}_{2}$ using a 2:1 quotient map. In this particular setting, we know how to judge the consistency of wrapping: on $S^{3}$, $(B-F)$ is a gauge-invariant object, so $[G]$ should vanish there. What needs to be checked is then to pull back the spacetime $G$ using the map
$$S^{3}\to S^{3}/\mathbb{Z}_{2}\to S^{5}/\mathbb{Z}_{2},$$
(3.4)
where the first is the 2:1 projection and the second is the embedding. The result is zero, from the trivial fact that $H^{3}(S^{3},\mathbb{Z})=\mathbb{Z}$ does not have non-zero two-torsion elements.
Thus we see that we can wrap two D3-branes in this way.
Now, let us come back to the $k=4$ case, and try to wrap two D3-branes on $S^{3}/\mathbb{Z}_{4}$ with discrete torsion, which should naively be possible since the discrete torsion lies in $\mathbb{Z}_{2}$. Again, if the Chan-Paton structure is trivial, we cannot wrap them. If we try to do the analogue of the $k=2$ case above, we can try wrapping two D3-branes such that the Chan-Paton indices 1 and 2 are exchanged. Just as above, this is equivalent to wrapping one D3-brane on $S^{3}/\mathbb{Z}_{2}$, covering $S^{3}/\mathbb{Z}_{4}$ with a 2:1 quotient map.
To test the consistency of the wrapping, one needs to pull back the spacetime obstruction on $S^{5}/\mathbb{Z}_{4}$ via
$$S^{3}/\mathbb{Z}_{2}\to S^{3}/\mathbb{Z}_{4}\to S^{5}/\mathbb{Z}_{4}.$$
(3.5)
One finds that the pull-back is non-trivial,
showing that this embedding is inconsistent.555In general, the pull-back map from $S^{n}/\mathbb{Z}_{ab}$ to $S^{n}/\mathbb{Z}_{b}$ is a multiplication map by $1+t+t^{2}+...+t^{a-1}$, where $t$ is the generator of the $\mathbb{Z}_{a}$ action that divides $S^{n}/\mathbb{Z}_{b}$ to $S^{n}/\mathbb{Z}_{ab}$.
It is clear that we can wrap four D3-branes on the discrete cycle, so this implies that the theory with this discrete torsion should be the $\ell=4$ variant discussed in Section 2, and that we cannot have S-folds with the $\ell=2$ variant.
This does not prove, we admit, that there are no other non-Abelian configurations on two D3-branes on $S^{3}/\mathbb{Z}_{4}$ that still allow the wrapping.
But we will see that the choice $\ell=4$ for this discrete torsion variant matches with the M-theory computation below.
3.4 Realization of the $\mathbb{Z}_{k}$ gauge theory in F-theory
After these discussions we can understand how the $\mathbb{Z}_{k}$ gauge theory of Section 3.1 arises in F-theory on $S^{5}/\mathbb{Z}_{k}$, when the discrete torsion in $H^{3}(S^{5}/\mathbb{Z}_{k},(\mathbb{Z}\oplus\mathbb{Z})_{\rho})$ is zero.
Recall that $H_{n}(S^{5}/\mathbb{Z}_{k},\mathbb{Z})$ is $\mathbb{Z}_{k}$ for $n=1$ and $3$, and its generator is $S^{1}/\mathbb{Z}_{k}$ and $S^{3}/\mathbb{Z}_{k}$, respectively.
We can wrap a D3-brane on these cycles to obtain a world 3-cycle and a worldline on $AdS_{5}$.
(Note that the latter is the generalized Pfaffian particle discussed above.)
Clearly, they both carry $\mathbb{Z}_{k}$ charges.
We are then naturally led to identify the worldline as coupling to the 1-form $A$ and the world 3-cycle to the 3-form $C$ in (3.1).
For this identification to make sense, it should be the case that if we rotate the generalized Pfaffian around an $S^{1}$ that has a unit linking number with the world 3-cycle of the D3-brane wrapped on $S^{1}/\mathbb{Z}_{k}$,
there should be a non-trivial $\mathbb{Z}_{k}$ holonomy of unit strength.
This can be seen as follows.
A D3-brane wrapped on $S^{1}/\mathbb{Z}_{k}$ creates an $F_{5}$ flux given by the Poincaré dual of its worldvolume.
In this case this is given by $s\otimes c$, where $s$ is the generator of $H^{1}(S^{1},\mathbb{Z})$ where this $S^{1}$ is linked in $AdS_{5}$ around the worldvolume of the D3-brane, and $c$ is the generator of $H^{4}(S^{5}/\mathbb{Z}_{k},\mathbb{Z})=\mathbb{Z}_{k}$.
Now, we use the cohomology long exact sequence associated to the short exact sequence $0\to\mathbb{Z}\to\mathbb{R}\to\mathrm{U}(1)\to 0$ to conclude that there is a natural isomorphism
$$H^{3}(S^{5}/\mathbb{Z}_{k},\mathrm{U}(1))\simeq H^{4}(S^{5}/\mathbb{Z}_{k},\mathbb{Z})=\mathbb{Z}_{k}.$$
(3.6)
In physics terms, this means that the discrete $\mathbb{Z}_{k}$ field strength (which is $\mathbb{Z}$-valued) with four legs along $S^{5}/\mathbb{Z}_{k}$ can be naturally identified with the discrete $\mathbb{Z}_{k}$ holonomy (which is $\mathrm{U}(1)$-valued) with three legs along $S^{5}/\mathbb{Z}_{k}$.
From this we see that we have a holonomy of the type IIB $C_{4}$ field given by the element
$$s\otimes c\in H^{1}(S^{1},\mathbb{Z})\otimes H^{3}(S^{5}/\mathbb{Z}_{k},\mathrm{U}(1)),$$
(3.7)
which can be naturally integrated on the cycle $S^{1}\times S^{3}/\mathbb{Z}_{k}$ to give $\exp(2\pi i/k)$.
This means that the generalized Pfaffian particle wrapped on $S^{3}/\mathbb{Z}_{k}$, when carried around the $S^{1}$ linking the worldvolume of the D3-brane wrapped on $S^{1}/\mathbb{Z}_{k}$, experiences this holonomy.
This is indeed the behavior we expect for the objects charged under the $\mathbb{Z}_{k}$ 1-form $A$ and the $\mathbb{Z}_{k}$ 3-form $C$ in (3.1).
4 Comparison to M-theory
In this section, we consider M-theory configurations on $(\mathbb{C}^{3}\times T^{2})/\mathbb{Z}_{k}$. We know that we obtain such configurations from any of our S-folds on a circle, using the standard relation between type IIB theory on a circle and M-theory on a torus. However, the opposite is not true, since when we translate some configuration of discrete fluxes in M-theory on $(\mathbb{C}^{3}\times T^{2})/\mathbb{Z}_{k}$ to F-theory on a circle, we could also have some non-trivial action of the shift around the F-theory circle, corresponding to a ‘shift-S-fold’
where there is a rotation on the transverse $\mathbb{C}^{3}$ when we go around the compactified $S^{1}$.
By analyzing all possible discrete charges in M-theory and translating them to F-theory,
we learn about all possible variants of S-folds.
4.1 Discrete fluxes in M-theory on $(\mathbb{C}^{3}\times T^{2})/\mathbb{Z}_{k}$
Let us start by analyzing the possible discrete fluxes in M-theory, which come from 4-form fluxes.
For a given $k$, $(\mathbb{C}^{3}\times T^{2})/\mathbb{Z}_{k}$ has several fixed points of the form $\mathbb{C}^{4}/\mathbb{Z}_{\ell_{i}}$,
each of which has an associated $H^{4}(S^{7}/\mathbb{Z}_{\ell_{i}},\mathbb{Z})=\mathbb{Z}_{\ell_{i}}$ in M-theory [41].
From this viewpoint, the possible discrete charges are given by the orbifold actions on the fixed points,
$$\bigoplus H^{3}(\mathbb{C}^{4}/\mathbb{Z}_{\ell_{i}},\mathbb{Z})=\begin{cases}\mathbb{Z}_{2}\oplus\mathbb{Z}_{2}\oplus\mathbb{Z}_{2}\oplus\mathbb{Z}_{2}&(k=2),\\
\mathbb{Z}_{3}\oplus\mathbb{Z}_{3}\oplus\mathbb{Z}_{3}&(k=3),\\
\mathbb{Z}_{4}\oplus\mathbb{Z}_{4}\oplus\mathbb{Z}_{2}&(k=4),\\
\mathbb{Z}_{6}\oplus\mathbb{Z}_{3}\oplus\mathbb{Z}_{2}&(k=6).\end{cases}$$
(4.1)
We can alternatively measure the same charges by considering $H^{4}$ of the ‘asymptotic infinity’ of $(\mathbb{C}^{3}\times T^{2})/\mathbb{Z}_{k}$, which has the form $(S^{5}\times T^{2})/\mathbb{Z}_{k}$.
This is a $T^{2}$ bundle over $S^{5}/\mathbb{Z}_{k}$, and as such, one can apply the Leray-Serre spectral sequence666For an introduction on the Leray-Serre spectral sequence, see e.g. [42].
In our case, the computation goes as follows.
The second page $E^{p,q}_{2}$ of the spectral sequence is given by $E^{p,q}_{2}=H^{p}(S^{5}/\mathbb{Z}_{k},H^{q}(T^{2},\mathbb{Z}))$, with the differentials $d_{2}^{p,q}:E^{p,q}_{2}\to E^{p+2,q-1}_{2}$. The third page is given by $E^{p,q}_{3}=\mathrm{Ker}\ d^{p,q}_{3}/\mathrm{Im}\ d^{p-2,q+1}$, and has the differentials $d_{3}^{p,q}:E^{p,q}_{3}\to E^{p+3,q-2}_{3}$.
The second page and the part relevant for us of the third page are then given by
$\begin{array}[]{c|cccccc}2&\mathbb{Z}&0&\mathbb{Z}_{k}&0&\mathbb{Z}_{k}&\mathbb{Z}\\
1&0&X&0&X&0&X\\
0&\mathbb{Z}&0&\mathbb{Z}_{k}&0&\mathbb{Z}_{k}&\mathbb{Z}\\
\hline\cr&0&1&2&3&4&5\end{array},\qquad\begin{array}[]{c|cccccc}2&&0&\mathbb{Z}_{k}&&&\\
1&&&&X&&\\
0&&&&&\mathbb{Z}_{k}&\mathbb{Z}\\
\hline\cr&0&1&2&3&4&5\end{array}$
where $X=\mathbb{Z}_{2}\oplus\mathbb{Z}_{2},\mathbb{Z}_{3},\mathbb{Z}_{2},\mathbb{Z}_{1}$ for $k=2,3,4,6$, respectively.
From this we conclude that in fact $F^{p,q}/F^{p+1,q-1}=E^{p,q}_{\infty}\simeq E^{p,q}_{2}$ for $p+q=4$., that says that it has the filtration
$$H^{4}((S^{5}\times T^{2})/\mathbb{Z}_{k},\mathbb{Z})=F^{2,2}\supset F^{3,1}\supset F^{4,0},$$
(4.2)
where
$$\displaystyle F^{2,2}/F^{3,1}$$
$$\displaystyle=H^{2}(S^{5}/\mathbb{Z}_{k},H^{2}(T^{2},\mathbb{Z}))=H^{2}(S^{5}/\mathbb{Z}_{k},\mathbb{Z}),$$
(4.3)
$$\displaystyle F^{3,1}/F^{4,0}$$
$$\displaystyle=H^{3}(S^{5}/\mathbb{Z}_{k},H^{1}(T^{2},\mathbb{Z}))=H^{3}(S^{5}/\mathbb{Z}_{k},(\mathbb{Z}\oplus\mathbb{Z})_{\rho}),$$
(4.4)
$$\displaystyle F^{4,0}$$
$$\displaystyle=H^{4}(S^{5}/\mathbb{Z}_{k},H^{0}(T^{2},\mathbb{Z}))=H^{4}(S^{5}/\mathbb{Z}_{k},\mathbb{Z}).$$
(4.5)
The different subgroups here correspond in a sense to the ‘number of legs along the base $S^{5}/\mathbb{Z}_{k}$ and the fiber $T^{2}$’.
As we are taking the $\mathbb{Z}$-valued cohomology, we only have a filtration and it is not guaranteed that the group (4.2) is a direct sum of (4.3)–(4.5).
We can easily compute (4.3)–(4.5) to obtain
$$F^{2,2}/F^{3,1}=\mathbb{Z}_{k},\quad F^{3,1}/F^{4,0}=\begin{cases}\mathbb{Z}_{2}\oplus\mathbb{Z}_{2}&(k=2)\\
\mathbb{Z}_{3}&(k=3)\\
\mathbb{Z}_{2}&(k=4)\\
\mathbb{Z}_{1}&(k=6)\end{cases},\quad F^{4,0}=\mathbb{Z}_{k}.$$
(4.6)
From the standard F-theory / M-theory mapping, we see that each piece (4.3)–(4.5)
has the following effect in the F-theory language:
•
The piece $F^{2,2}/F^{3,1}$ becomes a component of the metric of F-theory on $S^{1}$.
Concretely, it specifies the amount of the rotation on the transverse space while we go around $S^{1}$.
Therefore, the system is a plain $S^{1}$ compactification if the flux in this piece is zero, whereas it is a shift S-fold if it is non-zero.
•
The piece $F^{3,1}/F^{4,0}$ becomes the discrete RR and NSNS 3-form fluxes around the S-fold.
Indeed, the coefficient system $H^{1}(T^{2},\mathbb{Z})$ of the fiber is exactly the one $(\mathbb{Z}\oplus\mathbb{Z})_{\rho}$ that we discussed in the previous section, and there is a natural isomorphism of $F^{3,1}/F^{4,0}$ with the discrete torsion (3.3).
•
The piece $F^{4,0}$ becomes an $F_{5}$ flux having one leg along the $S^{1}$. As discussed in Sec. 3.4, we have a natural identification $H^{4}(S^{5}/\mathbb{Z}_{k},\mathbb{Z})=H^{3}(S^{5}/\mathbb{Z}_{k},\mathrm{U}(1))$.
So, this piece can also be regarded as specifying a $C_{4}$ holonomy having one leg along the $S^{1}$.
Note that
$$(F^{2,2}/F^{3,1})\oplus(F^{3,1}/F^{4,0})\oplus F^{4,0}=\begin{cases}\mathbb{Z}_{2}\oplus\mathbb{Z}_{2}\oplus\mathbb{Z}_{2}\oplus\mathbb{Z}_{2}&(k=2),\\
\mathbb{Z}_{3}\oplus\mathbb{Z}_{3}\oplus\mathbb{Z}_{3}&(k=3),\\
\mathbb{Z}_{4}\oplus\mathbb{Z}_{4}\oplus\mathbb{Z}_{2}&(k=4),\\
\mathbb{Z}_{6}\oplus\mathbb{Z}_{3}\oplus\mathbb{Z}_{2}&(k=6),\end{cases}$$
(4.7)
and we see that (4.7) and (4.1) are the same as abstract groups.
What remains is to figure out the precise mapping between the two.
This is a purely geometrical question since it is just the relation of the $H^{4}$ of $(\mathbb{C}^{3}\times T^{2})/\mathbb{Z}_{k}$ computed at the origin and at the asymptotic infinity.
As we have not yet done this computation, we will use guesswork, and then verify that it leads to consistent results for the M2-brane and D3-brane charges of the various singularities.
The $k=2$ case can be worked out using the known properties of the orientifolds. In this case the map
$$a:\bigoplus H^{3}(\mathbb{C}^{4}/\mathbb{Z}_{\ell_{i}})=(\mathbb{Z}_{2})^{4}\to F^{2,2}/F^{3,1}=\mathbb{Z}_{2}$$
(4.8)
is given by the sum of the four $\mathbb{Z}_{2}$’s, and the map
$$b:F^{4,0}=\mathbb{Z}_{2}\to\bigoplus H^{3}(\mathbb{C}^{4}/\mathbb{Z}_{\ell_{i}})=(\mathbb{Z}_{2})^{4}$$
(4.9)
is the diagonal embedding.
Note that $a\circ b$ is indeed a zero map.
We assume that a similar relation holds also in the other cases. Namely, we assume that
$$a:\bigoplus H^{3}(\mathbb{C}^{4}/\mathbb{Z}_{\ell_{i}})\to F^{2,2}/F^{3,1}=\mathbb{Z}_{k}$$
(4.10)
is the ‘sum’ and that
$$b:F^{4,0}=\mathbb{Z}_{k}\to\bigoplus H^{3}(\mathbb{C}^{4}/\mathbb{Z}_{\ell_{i}})$$
(4.11)
is the ‘diagonal embedding’. For $k=4$ and $k=6$, we need to use natural maps such as $\mathbb{Z}_{2}\to\mathbb{Z}_{6}$ and $\mathbb{Z}_{6}\to\mathbb{Z}_{2}$. We just choose a multiplication by 3 in the former, and the mod-2 map in the latter, and similarly for $k=4$.
We see that the composition $a\circ b$ is indeed a zero map, which gives a small check of our assumptions.
4.2 M2-brane and D3-brane charges
Using the discussions above, let us test our identification by working out the M2-brane and D3-brane charges of our various configurations.
We use the formula of [43] for the charge of $\mathbb{C}^{4}/\mathbb{Z}_{k}$ orbifolds with discrete torsion,
$$\text{(M2-brane charge)}=-\frac{1}{24}(k-\frac{1}{k})+\frac{\ell(k-\ell)}{2k}.$$
(4.12)
The cases we use are
$$\begin{array}[]{cr|ccccccc}&&\ell=0&\ell=1&\ell=2&\ell=3&\ell=4&\ell=5\\
\hline\cr k=2:&-\frac{1}{16}+&0&\frac{1}{4}\\[10.00002pt]
k=3:&-\frac{1}{9}+&0&\frac{1}{3}&\frac{1}{3}\\[10.00002pt]
k=4:&-\frac{5}{32}+&0&\frac{3}{8}&\frac{1}{2}&\frac{3}{8}\\[10.00002pt]
k=6:&-\frac{35}{144}+&0&\frac{5}{12}&\frac{2}{3}&\frac{3}{4}&\frac{2}{3}&\frac{5}{12}\end{array}.$$
(4.13)
A long list of tables analyzing all possible cases follows.
The end result is that every possible M-theory configuration can be interpreted as a plain $S^{1}$ compactification or a shift-S-fold on $S^{1}$ of precisely the S-folds that we discussed in the previous section (the types shaded in (2.13)),
namely the S-folds with
$$(k,\ell)=(2,1),(2,2),(3,1),(3,3),(4,1),(4,4),\ \text{and}\ (6,1).$$
(4.14)
In all cases we compute the D3-brane charge using the sum of the M2-brane charges of the orbifold singularities (4.12) and successfully compare it with our expectation (2.13).
4.2.1 $k=2$
Let us denote the discrete charges of the four fixed points as elements in $\begin{pmatrix}\mathbb{Z}_{2}&\mathbb{Z}_{2}\\
\mathbb{Z}_{2}&\mathbb{Z}_{2}\end{pmatrix}$. The piece $F_{4,0}=\mathbb{Z}_{2}$ is generated by $\begin{pmatrix}0&0\\
0&0\end{pmatrix}$ and $\begin{pmatrix}1&1\\
1&1\end{pmatrix}$. The piece $F^{2,2}/F^{3,1}$ is given by the sum of the four entries.
In the interpretations below, we consider the columns of the matrices to correspond to O2-planes, when we interpret our
M-theory configuration in type IIA by shrinking one cycle of the torus (as we can do for $k=2$).
Of course everything should work out correctly in this $k=2$ case, and has been already worked out in [44, 45].
We reproduce the analysis here since it is a useful warm-up for $k=3,4,6$.
Shift $=0/2\in F^{2,2}/F^{3,1}$:
$$\begin{array}[]{c|ccccccc}\hline\cr\text{label}&&&\begin{pmatrix}0&0\\
0&0\end{pmatrix}&\begin{pmatrix}1&1\\
1&1\end{pmatrix}\\
\text{IIA}&&&O2{}^{-}+O2{}^{-}&\widetilde{O2}{}^{-}+\widetilde{O2}{}^{-}\\
\text{\#D3}&&&-\frac{1}{4}&-\frac{1}{4}+1\\
\hline\cr\text{label}&\begin{pmatrix}0&1\\
0&1\end{pmatrix}&\begin{pmatrix}1&0\\
1&0\end{pmatrix}&&&\begin{pmatrix}0&0\\
1&1\end{pmatrix}&\begin{pmatrix}1&1\\
0&0\end{pmatrix}\\
\text{IIA}&O2{}^{-}+\widetilde{O2}{}^{-}&\widetilde{O2}{}^{-}+O2{}^{-}&&&O2{}^{+}+O2{}^{+}&\widetilde{O2}{}^{+}+\widetilde{O2}{}^{+}\\
\text{\#D3}&+\frac{1}{4}&+\frac{1}{4}&&&+\frac{1}{4}&+\frac{1}{4}\\
\hline\cr\text{label}&&&\begin{pmatrix}0&1\\
1&0\end{pmatrix}&\begin{pmatrix}1&0\\
0&1\end{pmatrix}\\
\text{IIA}&&&O2{}^{+}+\widetilde{O2}{}^{+}&\widetilde{O2}{}^{+}+{O2}{}^{+}\\
\text{\#D3}&&&+\frac{1}{4}&+\frac{1}{4}\\
\hline\cr\end{array}$$
The first two come from $O3{}^{-}$, but the latter of the two has one additional mobile D3-brane stuck at the origin, due to a non-trivial Wilson line around $S^{1}$ in the component of $\mathrm{O}(2N)$ disconnected from the identity.
The rest all come from the other three O3-planes wrapped on $S^{1}$.
Shift $=1/2\in F^{2,2}/F^{3,1}$:
$$\begin{array}[]{c|ccccccc}\hline\cr\text{label}&&&\begin{pmatrix}1&0\\
0&0\end{pmatrix}&\begin{pmatrix}0&1\\
1&1\end{pmatrix}\\
\text{IIA}&&&\widetilde{O2}{}^{+}+O2{}^{-}&{O2}{}^{-}+\widetilde{O2}{}^{-}\\
\text{\#D3}&&&0&0+\frac{1}{2}\\
\hline\cr\text{label}&\begin{pmatrix}1&1\\
0&1\end{pmatrix}&\begin{pmatrix}0&0\\
1&0\end{pmatrix}&&&\begin{pmatrix}1&0\\
1&1\end{pmatrix}&\begin{pmatrix}0&1\\
0&0\end{pmatrix}\\
\text{IIA}&\widetilde{O2}{}^{+}+\widetilde{O2}{}^{-}&{O2}{}^{+}+O2{}^{-}&&&\widetilde{O2}{}^{-}+O2{}^{+}&{O2}{}^{-}+\widetilde{O2}{}^{+}\\
\text{\#D3}&0+\frac{1}{2}&0&&&0&0+\frac{1}{2}\\
\hline\cr\text{label}&&&\begin{pmatrix}1&1\\
1&0\end{pmatrix}&\begin{pmatrix}0&0\\
0&1\end{pmatrix}\\
\text{IIA}&&&\widetilde{O2}{}^{-}+\widetilde{O2}{}^{+}&{O2}{}^{-}+{O2}{}^{+}\\
\text{\#D3}&&&0+\frac{1}{2}&0\\
\hline\cr\end{array}$$
They are
$\mathbb{Z}_{2}$ shift-orientifolds around $S^{1}$.
More precisely, the geometry is a $\mathbb{C}^{3}$ fibration over $S^{1}$ such that when we go around $S^{1}$, we have a multiplication by $(-1)$ on $\mathbb{C}^{3}$.
The ones with D3-brane charge $0$ are an empty shift-orientifold, and the ones with charge $0+1/2$ have one D3-brane wrapped around $S^{1}$.
Note that the charge $1/2$ we are seeing here reflects the fact that the $T^{2}$ of M-theory is fibered over $\mathbb{C}^{3}/\mathbb{Z}_{2}$.
In the type IIB frame, a fiber over a particular point on $S^{1}$ is $\mathbb{C}^{3}$, and therefore the D3-brane charge is 1 if we integrate over the asymptotic infinity of this $\mathbb{C}^{3}$.
4.2.2 $k=3$
Let us denote the discrete charges of the three fixed points as elements in $(\mathbb{Z}_{3},\mathbb{Z}_{3},\mathbb{Z}_{3})$. The piece $F_{4,0}=\mathbb{Z}_{3}$ consists of $(0,0,0)$, $(1,1,1)$, $(2,2,2)=-(1,1,1)$.
The piece $F^{2,2}/F^{3,1}=\mathbb{Z}_{3}$ corresponding to the shift is given by the sum of the entries.
There are 27 choices in total.
Shift $=0/3\in F^{2,2}/F^{3,1}$:
$$\begin{array}[]{c|ccc}\hline\cr\text{label}&(0,0,0)&(1,1,1)&(2,2,2)\\
\text{\#D3}&-\frac{1}{3}&-\frac{1}{3}+1&-\frac{1}{3}+1\\
\hline\cr\text{label}&(0,1,2)&(1,2,0)&(2,0,1)\\
\text{\#D3}&+\frac{1}{3}&+\frac{1}{3}&+\frac{1}{3}\\
\hline\cr\text{label}&(0,2,1)&(1,0,2)&(2,1,0)\\
\text{\#D3}&+\frac{1}{3}&+\frac{1}{3}&+\frac{1}{3}\\
\hline\cr\end{array}$$
Among the first three, the first entry has the right D3-brane charge to be the $(k=3,\ell=1)$ S-fold, see (2.13). The other two have one more mobile D3-brane, stuck at the origin through a non-trivial $\mathbb{Z}_{3}$ background holonomy around $S^{1}$.
The others all have the right D3-charge to be the $(k=3,\ell=3)$ S-fold, see (2.13) again.
The second three and the third three have opposite $\mathbb{Z}_{3}$ charge characterizing the type of the S-fold.
Shift $=1/3\in F^{2,2}/F^{3,1}$:
$$\begin{array}[]{c|ccc}\hline\cr\text{label}&(1,0,0)&(2,1,1)&(0,2,2)\\
\text{\#D3}&0&\frac{2}{3}&\frac{1}{3}\\
\hline\cr\text{label}&(1,1,2)&(2,2,0)&(0,0,1)\\
\text{\#D3}&\frac{2}{3}&\frac{1}{3}&0\\
\hline\cr\text{label}&(1,2,1)&(2,0,2)&(0,1,0)\\
\text{\#D3}&\frac{2}{3}&\frac{1}{3}&0\\
\hline\cr\end{array}$$
They are the $\mathbb{Z}_{3}$ shift-S-folds, with the rotation angle $2\pi/3$, of the flat background on $S^{1}$, with 0, 1, 2 D3-brane(s) wrapped around at the origin.
Shift $=2/3\in F^{2,2}/F^{3,1}$:
$$\begin{array}[]{c|ccc}\hline\cr\text{label}&(2,0,0)&(0,1,1)&(1,2,2)\\
\text{\#D3}&0&\frac{2}{3}&\frac{1}{3}\\
\hline\cr\text{label}&(2,1,2)&(0,2,0)&(1,0,1)\\
\text{\#D3}&\frac{2}{3}&0&\frac{1}{3}\\
\hline\cr\text{label}&(2,2,1)&(0,0,2)&(1,1,0)\\
\text{\#D3}&\frac{2}{3}&0&\frac{1}{3}\\
\hline\cr\end{array}$$
They are again the $\mathbb{Z}_{3}$ shift-S-folds, but with the rotation angle $4\pi/3$, of the flat background on $S^{1}$, with 0, 1, 2 mobile D3-brane(s) wrapped around at the origin.
4.2.3 $k=4$
We denote the charges at the fixed points as elements in $(\mathbb{Z}_{4},\mathbb{Z}_{4};\mathbb{Z}_{2})$.
The piece $F_{4,0}=\mathbb{Z}_{4}$ consists of $(0,0;0)$, $(1,1;1)$, $(2,2;0)$, $(3,3;1)$.
Shift $=0/4\in F^{2,2}/F^{3,1}$:
$$\begin{array}[]{c|cccc}\hline\cr\text{label}&(0,0;0)&(1,1;1)&(2,2;0)&(3,3;1)\\
\text{\#D3}&-\frac{3}{8}&-\frac{3}{8}+1&-\frac{3}{8}+1&-\frac{3}{8}+1\\
\hline\cr\text{label}&(0,2;1)&(1,3;0)&(2,0;1)&(3,1;0)\\
\text{\#D3}&+\frac{3}{8}&+\frac{3}{8}&+\frac{3}{8}&+\frac{3}{8}\\
\hline\cr\end{array}$$
Among the first four, the first has the right D3-brane charge to be the $(k=4,\ell=1)$ S-fold, see (2.13).
The other three have one more mobile D3-brane, stuck at the origin through non-trivial $\mathbb{Z}_{4}$ background holonomy around $S^{1}$.
All the second four have the right D3-charge to be the $(k=4,\ell=4)$ S-fold, see (2.13).
Shift $=1/4\in F^{2,2}/F^{3,1}$:
$$\begin{array}[]{c|cccc}\hline\cr\text{label}&(0,1;0)&(1,2;1)&(2,3;0)&(3,0;1)\\
\text{\#D3}&0&+\frac{3}{4}&+\frac{1}{2}&+\frac{1}{4}\\
\hline\cr\text{label}&(0,3;1)&(1,0;0)&(2,1;1)&(3,2;0)\\
\text{\#D3}&+\frac{1}{4}&0&+\frac{3}{4}&+\frac{1}{2}\\
\hline\cr\end{array}$$
These are $\mathbb{Z}_{4}$ shift-S-folds of flat space, with rotation angle $\pi/2$,
and with $0$, $1$, $2$ or $3$ mobile D3-brane(s) around $S^{1}$.
Shift $=2/4\in F^{2,2}/F^{3,1}$:
$$\begin{array}[]{c|cccc}\hline\cr\text{label}&(0,0;1)&(1,1;0)&(2,2;1)&(3,3;0)\\
\text{\#D3}&-\frac{1}{8}&-\frac{1}{8}+\frac{1}{2}&-\frac{1}{8}+1&-\frac{1}{8}+\frac{1}{2}\\
\hline\cr\text{label}&(0,2;0)&(1,3;1)&(2,0;0)&(3,1;1)\\
\text{\#D3}&+\frac{1}{8}&+\frac{1}{8}+\frac{1}{2}&+\frac{1}{8}&+\frac{1}{8}+\frac{1}{2}.\\
\hline\cr\end{array}$$
These are the $\mathbb{Z}_{2}$ shift-S-folds on $S^{1}$ of the O3-planes.
Among the first four, the first has the right D3-brane charge to be the $\mathbb{Z}_{2}$ shift-S-fold of $O3{}^{-}$.
Recall that the base of the M-theory configuration is $\mathbb{C}^{3}/\mathbb{Z}_{4}$, but the in the type IIB description, the fiber at a particular point on $S^{1}$ is $\mathbb{C}^{3}/\mathbb{Z}_{2}$. The D3-brane charge of $O3{}^{-}$ is $-1/4$, and therefore the charge as seen from the M-theory configuration is $1/2\cdot(-1/4)=-1/8$.
The others have one or two additional mobile D3-brane(s) trapped at the origin.
Among the second four, the first and the third have the right D3-brane charge to be the $\mathbb{Z}_{2}$ shift-S-fold of $\widetilde{O3}{}^{+}$.
The second and the fourth have one additional mobile D3-brane trapped at the origin.
Note that we cannot take the $\mathbb{Z}_{2}$ shift S-fold of $\widetilde{O3}{}^{-}$ or $O3{}^{+}$,
since they are exchanged by this $\mathbb{Z}_{2}$ operation.
Shift $=3/4\in F^{2,2}/F^{3,1}$:
$$\begin{array}[]{c|cccc}\hline\cr\text{label}&(0,1;1)&(1,2;0)&(2,3;1)&(3,0;0)\\
\text{\#D3}&+\frac{1}{4}&+\frac{1}{2}&+\frac{3}{4}&0\\
\hline\cr\text{label}&(0,3;0)&(1,0;1)&(2,1;0)&(3,2;1)\\
\text{\#D3}&0&+\frac{1}{4}&+\frac{1}{2}&+\frac{3}{4}.\\
\hline\cr\end{array}$$
These are $\mathbb{Z}_{4}$ shift-S-folds of flat space, with rotation angle $3\pi/2$,
and with $0$, $1$, $2$ or $3$ mobile D3-brane(s) around $S^{1}$.
4.2.4 $k=6$
We denote the charges at the fixed points as elements in $(\mathbb{Z}_{6},\mathbb{Z}_{3},\mathbb{Z}_{2})$.
The piece $F_{4,0}=\mathbb{Z}_{6}$ consists of $(0,0,0)$, $(1,1,1)$, $(2,2,0)$, $(3,0,1)$, $(4,1,0)$, $(5,2,1)$.
Shift $=0/6\in F^{2,2}/F^{3,1}$:
$$\begin{array}[]{c|cccccc}\hline\cr\text{label}&(0,0,0)&(1,1,1)&(2,2,0)&(3,0,1)&(4,1,0)&(5,2,1)\\
\text{\#D3}&-\frac{5}{12}&-\frac{5}{12}+1&-\frac{5}{12}+1&-\frac{5}{12}+1&-\frac{5}{12}+1&-\frac{5}{12}+1\\
\hline\cr\end{array}$$
The first has the right D3-brane charge to be the $(k=6,\ell=1)$ S-fold, see (2.13).
The other five have one more mobile D3-brane, stuck at the origin through a non-trivial $\mathbb{Z}_{6}$ background holonomy around $S^{1}$.
Shift $=1/6\in F^{2,2}/F^{3,1}$:
$$\begin{array}[]{c|cccccc}\hline\cr\text{label}&(1,0,0)&(2,1,1)&(3,2,0)&(4,0,1)&(5,1,0)&(0,2,1)\\
\text{\#D3}&0&\frac{5}{6}&\frac{2}{3}&\frac{1}{2}&\frac{1}{3}&\frac{1}{6}\\
\hline\cr\end{array}$$
This is the $\mathbb{Z}_{6}$ shift-S-fold of flat space, with zero to five D3-branes stuck at the origin. The rotation angle is $2\pi/6$.
Shift $=2/6\in F^{2,2}/F^{3,1}$:
$$\begin{array}[]{c|cccccc}\hline\cr\text{label}&(2,0,0)&(3,1,1)&(4,2,0)&(5,0,1)&(0,1,0)&(1,2,1)\\
\text{\#D3}&-\frac{1}{12}+\frac{1}{3}&-\frac{1}{12}+1&-\frac{1}{12}+\frac{2}{3}&-\frac{1}{12}+\frac{1}{3}&-\frac{1}{12}&-\frac{1}{12}+\frac{2}{3}\\
\hline\cr\end{array}$$
The fifth has the correct charge to be the $\mathbb{Z}_{3}$ shift-S-fold of the standard $O3{}^{-}$-plane.
Since the $O3{}^{-}$-plane itself has an identification by the angle $\pi$, its $\mathbb{Z}_{3}$ quotient involves the rotation by $\pi/3$ of the transverse space.
Note also that the $O3{}^{-}$-plane has the D3-brane charge $-1/4$, therefore we see $1/3\cdot(-1/4)=-1/12$ in M-theory.
The others have one or two additional D3-brane(s) at the origin.
Shift $=3/6\in F^{2,2}/F^{3,1}$:
$$\begin{array}[]{c|cccccc}\hline\cr\text{label}&(3,0,0)&(4,1,1)&(5,2,0)&(0,0,1)&(1,1,0)&(2,2,1)\\
\text{\#D3}&-\frac{1}{6}+\frac{1}{2}&-\frac{1}{6}+1&-\frac{1}{6}+\frac{1}{2}&-\frac{1}{6}&-\frac{1}{6}+\frac{1}{2}&-\frac{1}{6}+1\\
\hline\cr\end{array}$$
The fourth has the correct charge to be the $\mathbb{Z}_{2}$ shift-S-fold of the $(k=3,\ell=1)$ S-fold. The others have one or two additional mobile D3-brane(s) on top.
Shift $=4/6\in F^{2,2}/F^{3,1}$:
$$\begin{array}[]{c|cccccc}\hline\cr\text{label}&(4,0,0)&(5,1,1)&(0,2,0)&(1,0,1)&(2,1,0)&(3,2,1)\\
\text{\#D3}&-\frac{1}{12}+\frac{1}{3}&-\frac{1}{12}+\frac{2}{3}&-\frac{1}{12}&-\frac{1}{12}+\frac{1}{3}&-\frac{1}{12}+\frac{2}{3}&-\frac{1}{12}+1\\
\hline\cr\end{array}$$
The third has the correct charge to be the $\mathbb{Z}_{3}$ shift-S-fold, of rotation angle $2\pi/3$, of the standard $O3{}^{-}$-plane. The others have one or two additional D3-brane(s) at the origin.
Shift $=5/6\in F^{2,2}/F^{3,1}$:
$$\begin{array}[]{c|cccccc}\hline\cr\text{label}&(5,0,0)&(0,1,1)&(1,2,0)&(2,0,1)&(3,1,0)&(4,2,1)\\
\text{\#D3}&0&\frac{1}{6}&\frac{1}{3}&\frac{1}{2}&\frac{2}{3}&\frac{5}{6}\\
\hline\cr\end{array}$$
This is the $\mathbb{Z}_{6}$ shift-S-fold of flat space, with zero to five D3-branes stuck at the origin. The rotation angle is $2\pi\cdot 5/6$.
Comments:
Note that in the cases with shift $2,4\in\mathbb{Z}_{6}$, we only find $\mathbb{Z}_{3}$ shift-S-folds of ${O3}{}^{-}$, but we
do not have $\mathbb{Z}_{3}$ shift-S-folds of $\widetilde{O3}{}^{-}$, $O3{}^{+}$ and $\widetilde{O3}{}^{+}$. This is as it should be, because these three types of O3-planes are permuted by the $\mathbb{Z}_{3}$ action.
Similarly, there is no $\mathbb{Z}_{2}$ shift-orientifold of the $(k=3,\ell=3)$ orientifold, since the $\mathbb{Z}_{2}$ action exchanges the two subtly different versions that we denoted by $(1,1,1)$ and (2,2,2) in Sec. 4.2.2.
5 Special cases and $\mathcal{N}{=}\,4$ enhancement
So far, we saw that $N$ D3-branes probing various variants of the S-folds give rise to $\mathcal{N}{=}\,3$ superconformal field theories characterized by
$$(k,\ell)=(3,1),(3,3),(4,1),(4,4),(6,1).$$
(5.1)
When $\ell<k$ we expect to have both ‘parent’ theories and their discrete gaugings. In this section we discuss some interesting special cases, including cases where the ${\cal N}=3$ supersymmetry is enhanced to ${\cal N}=4$.
As discussed in [5], an enhancement of supersymmetry to $\mathcal{N}{=}\,4$ occurs if and only if there is a Coulomb branch operator of dimension 1 or 2, since $\mathcal{N}{=}\,3$ supersymmetry then dictates the presence of extra supercharges.
The dimensions of the Coulomb branch operators of our ‘parent’ $\mathcal{N}{=}\,3$ theories were given in (2.11).
For $k=2$ we always have such an enhancement, but for $k=3,4,6$ we see that it happens just for $\ell=1$ and $N=1,2$.
Note that the theory with the lowest central charge which does not have any enhancement is the $N=1$, $k=\ell=3$ theory, whose only Coulomb branch operator has dimension 3. The central charges of this theory are the same as those of five vector multiplets.
Since the Coulomb branch operators of $\mathcal{N}{=}\,3$ theories must be integers as shown in Section 3.1 of [28], then this must be the ‘minimal’ $\mathcal{N}{=}\,3$ SCFT, assuming the general validity of the formula (2.3). It would be interesting to test this by a superconformal bootstrap analysis, generalizing the ones in [46, 47].
Going back to theories with $\mathcal{N}{=}\,4$ supersymmetry, the case $N=1$ is rather trivial: we just have a Coulomb branch operator of dimension one, so the moduli space is just $\mathbb{C}^{3}$, and we get the $\mathcal{N}{=}\,4$ super Yang-Mills theory with gauge group $\mathrm{U}(1)$.
So let us discuss the $N=2$ cases.
The spectrum of the Coulomb branch operators of the ‘parent’ theory is given by
$$\begin{cases}2,3&(k=3),\\
2,4&(k=4),\\
2,6&(k=6).\end{cases}$$
(5.2)
These spectra agree with those of an $\mathcal{N}{=}\,4$ super Yang-Mills theory with gauge group $\mathrm{SU}(3)$, $\mathrm{SO}(5)$ and $G_{2}$, respectively.
Below we give evidence that indeed, the ‘parents’ of these S-fold configurations give rise to these $\mathcal{N}{=}\,4$ super Yang-Mills theories, realized in a somewhat unusual manner. Note that ${\cal N}=4$ theories always have an exactly marginal deformation, sitting in the same multiplet as the dimension two Coulomb branch operator, and our conjectured relation implies that for our $N=2$ theories this is the gauge coupling of these ${\cal N}=4$ gauge theories. Our discussion in the previous sections implies that in the AdS dual of these ‘parent’ theories, this marginal deformation corresponds to a scalar field coming from a D3-brane wrapped on the torsion 3-cycle; for this specific case this wrapped D3-brane gives rise to a massless field.
Again let us limit ourselves to the points on the moduli space corresponding to a Coulomb branch from the point of view of an ${\cal N}=2$ description of our SCFTs. In the $N=2,\ell=1$ theories, this subspace is parameterized by $z_{1,2}\in\mathbb{C}$, with
the gauge symmetry (2.10)
$$(z_{1},z_{2})\mapsto(\gamma^{n}z_{2},\gamma^{-n}z_{1}),$$
(5.3)
where $\gamma=e^{2\pi i/k}$ and $n$ is any integer.
The charges in a basis that is natural from this point of view can be written as $(e_{1},m_{1};e_{2},m_{2})$.
Naively, one would expect the electric charges $e_{1},e_{2}$ to correspond to electric charges of the corresponding ${\cal N}=4$ theory, but this cannot be the case because of the non-trivial $Sp(4,\mathbb{Z})$ action on these charges, induced by the $\mathrm{SL}(2,\mathbb{Z})$ transformation that accompanies the identification (5.3). So instead we consider the
rank-2 sublattice containing charges of the form $(Q;\overline{Q})$ where we regard $(e,m)\in\mathbb{Z}\oplus\mathbb{Z}$ as a complex number $Q=e+m\gamma$.
Two charges from this sublattice are local with respect to each other. To see this, note that given $Q=e+m\gamma$ and $Q^{\prime}=e^{\prime}+m^{\prime}\gamma$, their Dirac pairing is
$$em^{\prime}-e^{\prime}m=(\overline{Q}Q^{\prime}-\overline{Q}{}^{\prime}Q)/(\gamma-\overline{\gamma}).$$
(5.4)
Then the Dirac pairing between the two charges $(Q;\overline{Q})$ and $(Q^{\prime};\overline{Q}{}^{\prime})$ is clearly zero.
This means that we can take these charges to be the “electric charges” in the $\mathcal{N}{=}\,4$ description. We provide some consistency checks for this below.
Now, note that the $\mathrm{SL}(2,\mathbb{Z})$ element associated to the $\mathbb{Z}_{k}$ orbifold then acts on $Q$ just by multiplication by $\gamma$.
Then the gauge transformation (5.3) acts on this variable $Q$ as
$$Q\mapsto\gamma^{n}\overline{Q},$$
(5.5)
which is a reflection of the complex plane along the line $e^{\pi in/k}\mathbb{R}$.
This makes it clear that the group generated by (5.3) for $k=3,4,6$ is the Weyl group of $\mathrm{SU}(3)$, $\mathrm{SO}(5)$ and $G_{2}$, respectively.
Let us test our identification by looking at half-BPS particles. There’s no string connecting $z_{1}$ and $\gamma^{n}z_{1}$, since we know nothing happens when $z_{1}=0$ for $\ell=1$. So there are only strings connecting $\gamma^{n}z_{1}$ and $\gamma^{m}z_{2}$.
Using (5.3) we can always restrict $z_{1}$ to have a phase between 0 and $2\pi/k$.
Then we just have to consider all $(p,q)$-strings connecting $z_{1}$ and $\gamma^{n}z_{2}$ for $n=0,\ldots,k-1$.
Since the IIB coupling constant is $\tau=\gamma$, the central charges of half-BPS particles are given by
$$(p+q\gamma)(z_{1}-\gamma^{n}z_{2}).$$
(5.6)
We conjectured that “electric” states have the charge $(Q;\overline{Q})$, and then their central charges are given by $Qz_{1}-\overline{Q}z_{2}$. Comparing with (5.6), we see that “electric” objects have the following $Q$:
•
For $k=3$, $Q=\omega^{n}$ and $Q=(1+\omega)\omega^{n}=-\omega^{n-1}$ where $\omega=e^{2\pi i/3}$,
•
For $k=4$, $Q=i^{n}$ and $Q=(1+i)i^{n}$,
•
For $k=6$, $Q=\gamma^{n}$ and $Q=(1+\gamma)\gamma^{n}$.
Clearly they can be identified with the roots of $\mathrm{SU}(3)$, $\mathrm{SO}(5)$, and $G_{2}$, respectively.
The metric on the moduli space is also correctly mapped to that on the Cartan subalgebra of these groups.
The original metric is $dz_{1}d\overline{z}_{1}+dz_{2}d\overline{z}_{2}$ on $\mathbb{C}^{2}$,
and we choose a real subspace $\mathbb{R}^{2}$ of the form $Qz_{1}-\overline{Q}z_{2}$.
Then, two vectors $Qz_{1}-\overline{Q}z_{2}$ and $Q^{\prime}z_{1}-\overline{Q}^{\prime}z_{2}$ in $\mathbb{R}^{2}$ have the induced inner product $(2\mathrm{Re}(Q\overline{Q}^{\prime}))$.
Using this, we can easily check that the vectors listed above have the same inner products as the root vectors of $\mathrm{SU}(3)$, $\mathrm{SO}(5)$, and $G_{2}$, in the normalization that the short roots have length squared 2.
Finally, recall that the dyons of $\mathcal{N}{=}\,4$ SYM have central charges of the form
$$(p+q\tau_{YM})(\alpha_{s}\cdot\phi),\qquad(p+q\frac{\tau_{YM}}{r})(\alpha_{l}\cdot\phi),$$
(5.7)
where $\alpha_{s,l}$ are short and long roots, and $r$ is the length squared of the long roots divided by that of the short roots.
We can check that the spectrum (5.6) can be matched with (5.7) with the identification of the roots given above, if we take $\tau_{YM}=-1/(1+\gamma)$, uniformly for $k=3,4,6$.
As a further check, note that for $k=6$ this is exactly the value of $\tau_{YM}$ for which the $G_{2}$ ${\cal N}=4$ theory has a discrete $\mathbb{Z}_{6}$ symmetry [48].
Acknowledgements
The authors would like to thank I. García-Etxebarria, D. Harlow, Z. Komargodski, T. Nishinaka, H. Ooguri and N. Seiberg for useful discussions.
The work of OA is supported in part by an Israel Science Foundation center for excellence grant, by the I-CORE program of the Planning and Budgeting Committee and the Israel Science Foundation (grant number 1937/12), by the Minerva foundation with funding from the Federal German Ministry for Education and Research, by a Henri Gutwirth award from the Henri Gutwirth Fund for the Promotion of Research, and by the ISF within the ISF-UGC joint research program framework (grant no. 1200/14). OA is the Samuel Sebba Professorial Chair of Pure and Applied Physics.
The work of YT is partially supported in part by JSPS Grant-in-Aid for Scientific Research No. 25870159,
and by WPI Initiative, MEXT, Japan at IPMU, the University of Tokyo.
References
[1]
P. Argyres, M. Lotito, Y. Lü, and M. Martone, “Geometric constraints on
the space of ${\mathcal{N}}{=}\,2$ SCFTs I: physical constraints on relevant
deformations,”
arXiv:1505.04814 [hep-th].
[2]
P. C. Argyres, M. Lotito, Y. Lü, and M. Martone, “Geometric constraints
on the space of ${\mathcal{N}}{=}\,2$ SCFTs II: Construction of special
Kähler geometries and RG flows,”
arXiv:1601.00011 [hep-th].
[3]
P. C. Argyres, M. Lotito, Y. Lü, and M. Martone, “Expanding the landscape
of $\mathcal{N}$=2 rank 1 SCFTs,”
arXiv:1602.02764 [hep-th].
[4]
D. Gaiotto, “${\mathcal{N}}\!=2$ Dualities,”
JHEP 1208
(2012) 034,
arXiv:0904.2715 [hep-th].
[5]
O. Aharony and M. Evtikhiev, “On Four Dimensional ${\mathcal{N}}\!=3$
Superconformal Theories,”
arXiv:1512.03524 [hep-th].
[6]
C. Cordova, T. T. Dumitrescu, and K. Intriligator, “Deformations of
Superconformal Theories,”
arXiv:1602.01217 [hep-th].
[7]
I. García-Etxebarria and D. Regalado, “${\mathcal{N}}{=}\,3$ four
dimensional field theories,”
arXiv:1512.06434 [hep-th].
[8]
C. M. Hull, “A Geometry for Non-Geometric String Backgrounds,”
JHEP 10 (2005) 065,
arXiv:hep-th/0406102.
[9]
A. Dabholkar and C. Hull, “Duality Twists, Orbifolds, and Fluxes,”
JHEP 09 (2003) 054,
arXiv:hep-th/0210209.
[10]
C. M. Hull and A. Catal-Ozer, “Compactifications with S Duality Twists,”
JHEP 10 (2003) 034,
arXiv:hep-th/0308133.
[11]
C. M. Hull and R. A. Reid-Edwards, “Flux Compactifications of M-theory on
Twisted Tori,” JHEP 10 (2006) 086,
arXiv:hep-th/0603094.
[12]
A. Catal-Ozer, “Duality Twists on a Group Manifold,”
JHEP 10 (2006) 072,
arXiv:hep-th/0606278.
[13]
R. A. Reid-Edwards, “Geometric and Non-Geometric Compactifications of IIB
Supergravity,” JHEP 12 (2008) 043,
arXiv:hep-th/0610263.
[14]
O. J. Ganor and Y. P. Hong, “Selfduality and Chern-Simons Theory,”
arXiv:0812.1213 [hep-th].
[15]
O. J. Ganor, Y. P. Hong, and H. S. Tan, “Ground States of S-Duality Twisted
${\mathcal{N}}\!=4$ Super Yang-Mills Theory,”
JHEP 03
(2011) 099,
arXiv:1007.3749 [hep-th].
[16]
O. J. Ganor, Y. P. Hong, R. Markov, and H. S. Tan, “Static Charges in the
Low-Energy Theory of the S-Duality Twist,”
JHEP 04
(2012) 041,
arXiv:1201.2679 [hep-th].
[17]
C. Vafa, “Evidence for F Theory,”
Nucl. Phys.
B469 (1996) 403–418,
arXiv:hep-th/9602022.
[18]
A. Fayyazuddin and M. Spałinski, “Large $N$ Superconformal Gauge Theories
and Supergravity Orientifolds,”
Nucl. Phys.
B535 (1998) 219–232,
arXiv:hep-th/9805096.
[19]
O. Aharony, A. Fayyazuddin, and J. M. Maldacena, “The Large $N$ Limit of
${\mathcal{N}}\!=2$,1 Field Theories from Three- Branes in F-Theory,” JHEP 07 (1998) 013,
arXiv:hep-th/9806159.
[20]
O. Aharony and Y. Tachikawa, “A Holographic Computation of the Central
Charges of $d=4$, ${\mathcal{N}}\!=2$ SCFTs,”
JHEP 0801 (2008) 037,
arXiv:0711.4532 [hep-th].
[21]
P. C. Argyres and J. R. Wittig, “Infinite Coupling Duals of
${\mathcal{N}}\!=2$ Gauge Theories and New Rank 1 Superconformal Field
Theories,” JHEP 01 (2008) 074,
arXiv:0712.2028 [hep-th].
[22]
A. D. Shapere and Y. Tachikawa, “Central Charges of ${\mathcal{N}}\!=2$
Superconformal Field Theories in Four Dimensions,”
JHEP 0809 (2008) 109,
arXiv:0804.1957 [hep-th].
[23]
D. Gaiotto, A. Kapustin, N. Seiberg, and B. Willett, “Generalized Global
Symmetries,” JHEP
02 (2015) 172,
arXiv:1412.5148 [hep-th].
[24]
T. Banks and N. Seiberg, “Symmetries and Strings in Field Theory and
Gravity,” Phys.
Rev. D83 (2011) 084019,
arXiv:1011.5120 [hep-th].
[25]
S. Cecotti and M. Del Zotto, “Higher S-Dualities and Shephard-Todd Groups,”
JHEP 09
(2015) 035,
arXiv:1507.01799 [hep-th].
[26]
G. I. Lehrer and D. E. Taylor, Unitary Reflection Groups, vol. 20 of Australian Mathematical Society Lecture Series.
Cambridge University Press, 2009.
[27]
C. Beem, M. Lemos, P. Liendo, W. Peelaers, L. Rastelli, et al.,
“Infinite Chiral Symmetry in Four Dimensions,”
arXiv:1312.5344 [hep-th].
[28]
T. Nishinaka and Y. Tachikawa, “On 4D Rank-One ${\mathcal{N}}\!=3$
Superconformal Field Theories,”
arXiv:1602.01503 [hep-th].
[29]
E. Witten, “SL(2,Z) action on three-dimensional conformal field theories with
Abelian symmetry,”
arXiv:hep-th/0307041.
[30]
D. Harlow and H. Ooguri. to appear.
[31]
E. Witten, “Baryons and Branes in Anti De Sitter Space,” JHEP
07 (1998) 006,
arXiv:hep-th/9805112.
[32]
D. J. Gross and H. Ooguri, “Aspects of large N gauge theory dynamics as seen
by string theory,” Phys. Rev. D58 (1998) 106002,
arXiv:hep-th/9805129.
[33]
O. Aharony and E. Witten, “Anti-de Sitter Space and the Center of the Gauge
Group,” JHEP
11 (1998) 018,
arXiv:hep-th/9807205.
[34]
O. Aharony, N. Seiberg, and Y. Tachikawa, “Reading between the lines of
four-dimensional gauge theories,”
JHEP 08
(2013) 115,
arXiv:1305.0318 [hep-th].
[35]
A. Kapustin and N. Seiberg, “Coupling a QFT to a TQFT and Duality,”
JHEP 04
(2014) 001,
arXiv:1401.0740 [hep-th].
[36]
J. M. Maldacena, G. W. Moore, and N. Seiberg, “D-Brane Charges in Five-Brane
Backgrounds,” JHEP 10 (2001) 005,
arXiv:hep-th/0108152.
[37]
D. Belov and G. W. Moore, “Conformal Blocks for $\mathrm{AdS}_{5}$
Singletons,”
arXiv:hep-th/0412167.
[38]
E. Witten, “AdS/CFT Correspondence and Topological Field Theory,” JHEP 12 (1998) 012,
arXiv:hep-th/9812012.
[39]
A. Hatcher, Algebraic topology.
Cambridge University Press, 2002.
[40]
J. F. Davis and P. Kirk, Lecture Notes in Algebraic Topology, vol. 35 of
Graduate Studies in Mathematics.
American Mathematical Society, 2001.
[41]
O. Aharony, O. Bergman, and D. L. Jafferis, “Fractional M2-Branes,”
JHEP 11 (2008) 043,
arXiv:0807.4924 [hep-th].
[42]
A. Hatcher, Spectral Sequences in algebraic topology.
available at
http://www.math.cornell.edu/$\sim$hatcher/SSAT/SSATpage.html.
[43]
O. Aharony, A. Hashimoto, S. Hirano, and P. Ouyang, “D-Brane Charges in
Gravitational Duals of 2+1 Dimensional Gauge Theories and Duality
Cascades,” JHEP
01 (2010) 072,
arXiv:0906.2390 [hep-th].
[44]
A. Hanany and B. Kol, “On Orientifolds, Discrete Torsion, Branes and M
Theory,” JHEP 06 (2000) 013,
arXiv:hep-th/0003025.
[45]
A. Hanany and J. Troost, “Orientifold Planes, Affine Algebras and Magnetic
Monopoles,” JHEP 08 (2001) 021,
arXiv:hep-th/0107153.
[46]
C. Beem, L. Rastelli, and B. C. van Rees, “The ${\mathcal{N}}\!=4$
Superconformal Bootstrap,”
arXiv:1304.1803 [hep-th].
[47]
P. Liendo, I. Ramirez, and J. Seo, “Stress-Tensor OPE in $\mathcal{N}=2$
Superconformal Theories,”
JHEP 02
(2016) 019,
arXiv:1509.00033 [hep-th].
[48]
P. C. Argyres, A. Kapustin, and N. Seiberg, “On S-Duality for
Non-Simply-Laced Gauge Groups,”
JHEP 06 (2006) 043,
arXiv:hep-th/0603048. |
MIT-CTP/4956
\@open@row |
11institutetext: Université Laval, Québec, Canada 22institutetext: Depix, Montréal, Canada
https://lvsn.github.io/EditableIndoorLight/
Editable indoor lighting estimation
Henrique Weber
11
0000-0001-7103-7220
Mathieu Garon
22
0000-0003-1811-4156
Jean-François Lalonde
11
0000-0002-6583-2364
Abstract
We present a method for estimating lighting from a single perspective image of an indoor scene. Previous methods for predicting indoor illumination usually focus on either simple, parametric lighting that lack realism, or on richer representations that are difficult or even impossible to understand or modify after prediction. We propose a pipeline that estimates a parametric light that is easy to edit and allows renderings with strong shadows, alongside with a non-parametric texture with high-frequency information necessary for realistic rendering of specular objects. Once estimated, the predictions obtained with our model are interpretable and can easily be modified by an artist/user with a few mouse clicks. Quantitative and qualitative results show that our approach makes indoor lighting estimation easier to handle by a casual user, while still producing competitive results.
Keywords: lighting estimation, virtual object insertion, HDR
1 Introduction
Mixing virtual content realistically with real imagery is required in an increasing range of applications, from special effects to image editing and augmented reality (AR). This has created the need for capturing the lighting conditions of a scene with ever increasing accuracy and flexibility. In his seminal work, Debevec [6] suggested to capture the lighting conditions with a high dynamic range light probe. While it has been improved over the years, this technique, dubbed image-based lighting, is still at the heart of lighting capture for high end special effects in movies nowadays111See https://www.fxguide.com/fxfeatured/the-definitive-weta-digital-guide-to-ibl/.. Since the democratization of virtual object insertion for consumer image editing and AR, capturing light conditions with light probes restricts non professional users to have access to the scene and to use specialized equipment. To circumvent those limitations, approaches for automatically estimating the lighting conditions directly from images have been proposed.
In this line of work, the trend has been to estimate more and more complex lighting representations. This is exemplified by works such as Lighthouse [25], which propose to learn a multi-scale volumetric representation from an input stereo pair. Similarly, Li et al. [19] learn a dense 2D grid of spherical gaussians over the image plane. Wang et al. [27] propose to learn a 3D volume of similar spherical gaussians. While these lighting representations have been shown to yield realistic and spatially-varying relighting results, they have the unfortunate downside of being hard to understand: they do not lend themselves to being easily editable by a user. This quickly becomes a source of limitation when erroneous automatic results need to be corrected for improved accuracy or when creative freedom is required.
In this work, we depart from this trend and propose a simple, interpretable, and editable lighting representation (fig. 1).
But what does it mean for a lighting representation to be editable? We argue that an editable lighting representation must: 1) disentangle various components of illumination; 2) allow an intuitive control over those components; and, of course, 3) enable realistic relighting results.
Existing lighting representations in the literature do not possess all three properties. Environment maps [11, 24, 17] can be rotated but they compound light sources and environment textures together such that one cannot, say, easily increase the intensity of the light source without affecting everything else.
Rotating the environment map inevitably rotates the entire scene, turning walls into ceilings, etc., when changing the elevation.
Dense and/or volumetric representations [12, 19, 25, 27] are composed of 2D (or 3D) grids containing hundreds of parameters, which would have to be modified in a consistent way to achieve the desired result, an unachievable task for most. Parametric representations [10] model individual light sources with a few intuitive parameters, which can be modified independently of the others, but cannot generate realistic reflections.
Our proposed representation is the first to offer all three desired properties and is composed of two parts: 1) a parametric light source for modeling shading in high dynamic range; and 2) a non-parametric texture to generate realistic reflections off of shiny objects.
Our representation builds on the hypothesis (which we validate) that most indoor scenes can accurately be modeled by a single, dominant directional light source. We model this in high dynamic range with a parametric representation [10] that explicitly models the light source intensity, size, and 3D position. This representation is intuitive and can easily be edited by a user simply by moving the light source around in 3D.
This light source is complemented with a spatially-varying environment map texture, mapped onto a coarse 3D representation of the indoor scene. For this, we rely on a layout estimation network, which estimates a cuboid-like model of the scene from the input image. In addition, we also use a texture estimation network, whose output is conditioned on a combination of the input image, the scene layout and the parametric lighting representation. By explicitly tying the appearance of the environment texture with the position of the parametric light source, modifying the light source parameters (e.g. moving around the light) will automatically adjust the environment in a realistic fashion.
While our representation is significantly simplified, we find that it offers several advantages over the previous approaches. First, it renders both realistic shading (due to the high dynamic range of the estimated parametric light) and reflections (due to the estimated environment map texture). Second, it can efficiently be trained on real images, thereby alleviating any domain gap that typically arise when approaches need synthetic imagery for training [25, 19, 27]. Third—and perhaps most importantly—it is interpretable and editable. Since all automatic approaches are bound to make mistakes, it is of paramount importance in many scenarios that their output be adjustable by a user. By modifying the light parameters and/or the scene layout using simple user interfaces, our approach bridges the gap between realism and editability for lighting estimation.
2 Related work
For succinctness, we focus on single-image indoor lighting estimation methods in the section below, and refer the reader to the recent survey on deep models for lighting estimation for a broader overview [8].
2.0.1 Lighting estimation
Gardner et al. [11] proposed the first deep learning-based lighting estimation method for indoor scenes, and predicted an HDR environment map (equirectangular image) from a single image. This representation was also used in [17] for both indoors and outdoors, in [24] to take into account the object insertion position, in [23] which presented a real-time on-device approach, in [22] for scene decomposition, and in [3] which exploited the front and back cameras in current mobile devices. Finally, [28] propose to learn the space of indoor lighting using environment maps on single objects.
Other works explored alternative representations, such as spherical harmonics [12, 20, 34] that are useful for real-time rendering but are typically unsuitable for modeling high-frequency lighting (such as bright light sources) and are not ideal for non diffuse object rendering. [10] proposed to estimate a set of 3 parametric lights, which can easily be edited. However, that representation cannot generate realistic reflections. EMlight [33] propose a more expressive model by predicting gaussians on a spherical model. Similar to us, GMlight [31] back-projects the spherical gaussians to an estimated 3D model of the scene. This is further extended in [1] by the use of graph neural networks, and in [32] through the use of spherical wavelets dubbed “needlets”.
Recently, methods have attempted to learn volumetric lighting representations from images. Of note, Lighthouse [25] learns multi-scale volumetric lighting from a stereo pair, [19] predicts a dense 2D grid of spherical gaussians which is further extended into a 3D volumetric representation by Wang et al. [27]. While these yield convincing spatially-varying results, these representations cannot easily be interacted by a user.
2.0.2 Scene decomposition
Holistic scene decomposition [2] is deeply tied to lighting estimation as both are required to invert the image formation process. Li et al. [19] proposes to extract the scene geometry and the lighting simultaneously. Similarly, [7] extract only the geometry of the scene by estimating the normal and depth of the scene. These geometric representations are however non-parametric and thus difficult to edit or comprehend.
[16] proposes a simplified parametric model where a room layout is recovered in the camera field of view. Similarly, [35] presents a method to estimate the layout given a panoramic image of an indoor scene. We use the method of [16] to estimate a panoramic layout given a perspective image, thus providing a simple cuboid representation that allows for spatially varying textured lighting representation.
3 Editable indoor lighting representation
We begin by presenting our hybrid parametric/non-parametric lighting representation which aims at bridging the gap between realism and editability. We also show how that representation can be fitted to high dynamic range panoramas to obtain a training dataset, and conclude by presenting how it can be used for virtual object relighting.
3.1 Lighting representation
Our proposed light representation, shown in fig. 2, is composed of two main components: an HDR parametric light source $\mathbf{p}$; and an LDR textured cuboid $\mathcal{C}$.
3.1.1 Light source
As in [10], the light source parameters $\mathbf{p}$ are defined as
$$\mathbf{p}=\{\mathbf{l},d,s,\mathbf{c},\mathbf{a}\}\,,$$
(1)
where $\mathbf{l}\in\mathbb{R}^{3}$ is a unit vector specifying the light direction in XYZ coordinates, $d$ is the distance in meters, $s$ the radius (in meters), $\mathbf{c},\mathbf{a}\in\mathbb{R}^{3}$ are the light source and ambient colors in RGB, respectively. Here, $\mathbf{l}$, $d$ and $s$ are defined with respect to the camera. In contrast with [10], we use a single light source.
3.1.2 Textured cuboid
The cuboid $\mathcal{C}=\{\mathbf{T},\mathbf{L}\}$ is represented by a texture $\mathbf{T}\in\mathbb{R}^{2H\times H\times 3}$, which is an RGB spherical image of resolution $2H\times H$ stored in equirectangular (latitude-longitude) format, and a scene layout $\mathbf{L}\in\mathbb{R}^{2H\times H}$. The layout is a binary image of the same resolution, also in equirectangular format, indicating the intersections of the main planar surfaces in the room (walls, floor, ceiling) as an edge map [9].
3.2 Ground truth dataset
The ground truth is derived from the Laval Indoor HDR Dataset [11], which contains 2,100 HDR panoramas (with approximate depth labels from [10]). We extract $\mathbf{p}$ and $\mathcal{C}$ from each panorama using the following procedure. First, the HDR panorama is clipped to LDR (we re-expose such that the 90th-percentile is 0.8 then clip to [0, 1]) and directly used as the texture $\mathbf{T}$. Then the intersection between the main surfaces are manually labelled to define the layout $\mathbf{L}$. Lastly, we extract a dominant parametric light source from the HDR panorama. In order to determine the main light source, the $N=5$ brightest individual light sources are first detected using the region-growing procedure in [10]. A test scene (9 diffuse spheres arranged in a $3\times 3$ grid on a diffuse ground plane, seen from top as in fig. 4b) is rendered with each light source independently by masking out all other pixels—the brightest render determines the strongest light source.
An initial estimate of the light parameters $\mathbf{p}$ are obtained by the following. The distance $d$ is approximated by using the average depth of the region, direction $\mathbf{l}$ as the region centroid, the angular size from the major and minor axes of an ellipse fitted to the same region. Finally, the light color $\mathbf{c}$ and ambient term $\mathbf{a}$ are initialized with a least-squares fit to a rendering of the test scene using the HDR panorama. From the initial parameters, $\mathbf{p}$ is further refined:
$$\mathbf{p}^{*}=\operatorname*{arg\,min}_{\mathbf{p}}||\mathcal{R}(\mathbf{p})-\mathcal{R}(\tilde{\mathbf{P}})||_{2}\,.$$
(2)
$\mathcal{R}(x)$ is a differentiable rendering operator (implemented with Redner [18]) that renders a test scene using $\mathbf{p}$. The optimization is performed using gradient descent with Adam [15]. Finally, the texture map $\mathbf{T}$ is rescaled with the estimated ambient term $\mathbf{a}^{*}$ to ensure that the texture yields the same average RGB color.
3.3 Virtual object rendering
To render a virtual object using our lighting representation, we employ the Cycles rendering engine222Available within Blender at https://www.blender.org.. A scene, as shown in fig. 2, is composed of a 3D emissive sphere for the parametric light $\mathbf{p}$ and the textured cuboid mesh $\mathcal{C}$. The cuboid mesh is derived by detecting the cuboid corners from the layout using high pass filters. We use the following geometric constraints to simplify the back-projection of the scene corners to 3D. First, the shape is limited to a cuboid, meaning that opposing faces are parallel. Second, the panorama layouts were trained using a camera elevation of $0^{\circ}$ (pointing at the horizon) and height of 1.6 meter above the ground. Using these constraints, the bottom corners can easily be projected on the ground plane, and the top corners can be used to compute the ceiling height (averaged from the 4 corners). A texture map can then be computed using every planar surfaces of the cuboid. Finally, the parametric light and the texture are rendered in two rendering passes. After rendering, the relit virtual object can be composited into the image using differential rendering [6].
4 Approach
Our approach, illustrated in fig. 3, is composed of three main networks: light, layout, and texture which are combined together to estimate our light representation (c.f., sec. 3) from an image. We assume that the layout of the input image is available, in practice this is obtained with an off-the-shelf solution [30].
4.0.1 Light network
A “light” network is trained to learn the mapping from input image $\mathbf{I}\in\mathbb{R}^{128\times 128\times 3}$ to estimated lighting parameters $\mathbf{p}$ (sec. 3) using a similar approach to [10]. Specifically, the light network is composed of a headless DenseNet-121 encoder [14] to produce a 2048-dimensional latent vector, followed by a fully-connected layer (512 units), and ultimately with an output layer producing the light source parameters $\mathbf{p}$.
The light network is trained on light parameters fitted on panoramas from the Laval Indoor HDR Dataset [11] using the procedure described in sec. 3.2. To generate the input image from the panorama, we follow [11] and extract rectified crops from the HDR panoramas. The resulting images are converted to LDR by re-exposing to make the median intensity equal to 0.45, clipping to 1, and applying a $\gamma=\nicefrac{{1}}{{2.4}}$ tonemapping. The same exposure factor is subsequently applied to the color $\mathbf{c}$ and ambient $a$ light parameters to ensure consistency. Note that the training process is significantly simplified compared to [11] as the network predicts only a single set of parameters.
We employ individual loss functions on each of the parameters independently: L2 for direction $\mathbf{l}$, depth $d$, size $s$, and ambient color $a$, and L1 for light color $\mathbf{c}$. In addition, we also employ an angular loss for both the ambient and light colors $a$ and $\mathbf{c}$ to enforce color consistency. The weights for each term were obtained through a Bayesian optimization on the validation set (see supp. mat.).
4.0.2 Layout network
The mapping from the input RGB image $\mathbf{I}$ and its layout (obtained with [30]) to the estimated scene layout $\hat{\mathbf{L}}$ (sec. 3) is learned by the “layout” network whose architecture is that of pix2pixHD [26]. Both inputs are concatenated channel-wise. The layout network is trained on both the Laval and the Zillow Indoor Dataset [5], which contains 67,448 LDR indoor panoramas of 1575 unfurnished residences along with their scene layouts. To train the network, a combination of GAN, feature matching and perceptual losses are employed [26]. The same default weights as in [26] are used in training.
4.0.3 Texture network
Finally, the estimated environment texture $\hat{\mathbf{T}}$ is predicted by a “texture” network whose architecture is also that of pix2pixHD [26]. It accepts as input a channel-wise concatenation of three images: the input RGB image $\mathbf{I}$, the estimated light parameters $\hat{\mathbf{p}}$ projected in an equirectangular format, and the estimated scene layout $\hat{\mathbf{L}}$. The equirectangular images are vertically concatenated to the input image. Note that the $\hat{\mathbf{p}}$ projection is performed using a subset of all parameters (direction $\mathbf{l}$ and size $s$ only).
The texture network is also trained on both Laval and Zillow datasets. To obtain the required light source position from the Zillow dataset, we detect the largest connected component whose intensity is above the 98th percentile over the upper half of the panorama. To convert the Laval HDR panoramas to LDR, first a scale factor is found such as the crop taken from that panorama has its 90th percentile mapped to 0.8. This scale factor is then applied to the panorama such as its scale matches the one of the crop. The texture network is trained with the same combination of losses as the layout network.
5 Experiments
5.1 Validation of our 1-light approximation
We test our hypothesis that most indoor scenes are well-approximated by a single dominant light source with an ambient term. We render a scene with the ground truth environment map, and compare it with the renders obtained from the parametric lighting optimization procedure described in sec. 3. Fig. 4a shows the cumulative distribution of the contribution of the strongest light with respect to the entire lighting of the scene. Note that the strongest light source contributes to more than 95%/80%/60% of the total lighting for 25%/50%/75% of the images in our test set. Fig. 4b shows example images for each of these scenarios. We find that even if we expect indoor scenes to have multiple light sources, the vast majority can accurately be represented by a single dominant light.
5.2 Light estimation comparison
We now evaluate our method and compare it with recent state-of-the-art light estimation approaches. We first validate that our model performs better on quantitative metrics evaluated on physic-based renders of a scene using a test set provided by [11]. For each of the 224 panoramas in the test split, we extract 10 images using the same sampling distribution as in [11], for a total of 2,240 images for evaluation. We also show renders in various scenes to demonstrate how our solution is visually more appealing.
5.2.1 Quantitative comparison
To evaluate the lighting estimates, we render a test scene composed of an array of spheres viewed from above (sec. 3) and compute error metrics on the resulting rendering when compared to the ground truth obtained with the original HDR panorama. We report RMSE, si-RMSE [13], PSNR, and RGB angular error [17]. We also compute the FID444Implementation taken from https://pypi.org/project/pytorch-fid/. on the resulting environment maps to evaluate the realism of reflections (similar to [23]).
We evaluate against the following works.
First, two versions of [10] are compared: the original (3) where 3 light sources are estimated, and a version (1) trained to predict a single parametric light. Second, we also compare to Lighthouse [25], which expects a stereo pair as input. As a substitute, we generate a second image with a small baseline using Synsin [29] (visual inspection confirmed this yields reasonable results). For [12], we select the coordinates of the image center for the object position. For [23], we implemented their proposed “Cluster ID loss” and tonemapping (eq. 1 in [23]) but used pix2pixHD as backbone. Finally, we also compare against [33]. Results for each metrics are reported in tab. 1, which shows that despite our model being simple, it achieves the best score in every metric. We argue that it is because of its simplicity that we can achieve competitive results. Our approach can be trained on real data (as opposed to [12, 25] and does not require an elaborate 2-stage training process (compared to [10]).
We also demonstrate a significantly lower FID score than other methods thus bridging the gap between representation realism and HDR accuracy.
5.2.2 Qualitative comparison
We also present qualitative results in fig. 5 where predictions are rendered on 3 spheres with varying reflectance properties (diffuse, mirror, glossy). In addition, a tonemapped equirectangular view of the estimated light representation is provided under each render. We show an example from each error percentiles according to the RMSE metric. Our proposed method is perceptually better on the mirror spheres as other methods do not model high frequency details from the scene. We also notice accurate shadow and shading from all the spheres. We show objects realistically composited into photographs in fig. 6. Note how the reflections of the virtual objects and their cast shadows on the ground plane perceptually match the input photograph. Finally, we also compare against [27] in fig. 7.
5.3 Ablation study on input layout
One may consider that requiring the image layout as input may make our method sensitive to its estimation. To show this is not the case, we perform an experiment where we provide a black layout as input to the layout network (equivalent to no layout estimate). As can be seen in fig. 8, providing a black layout as input simply results in a different layout prediction where the texture still remains coherent with the RGB input and estimated light direction. The FID of the generated panoramas with no input layout is 88.68 (compared to 89.58 from tab. 1), showing that this essentially has no impact.
5.4 Ablation study on the texture network
We also tested different configurations for the texture network in order to validate our design choices. More specifically, we trained the texture network providing as input: (1) only the RGB crop (FID of 167.39), (2) RGB crop and parametric light (FID of 97.13), and (3) RGB crop and layout (FID of 151.04). In contrast, our full approach obtained an FID of 89.57 (see tab. 1).
6 Editing the estimated lighting
Because of its intuitive nature, it is simple and natural for a user to edit our estimated lighting representation, should the estimate not perfectly match the background image or simply for artistic purposes. Fig. 9 shows that our approach simultaneously disentangles various components of illumination, allows an intuitive control over those components, and enables realistic relighting results. First, fig. 9a shows that a user can rotate the light source about its azimuth angle. Note how the estimated texture (second row) is consistent with the desired light position (third row), while preserving the same overall structure. The renders (first row) exhibit realistic reflections and shadows that correspond to the desired lighting directions. A similar behaviour can be observed in figs 9b and 9c when the elevation angle and size are modified, respectively. In fig. 9d, we show that it is also possible to edit the scene layout and obtain an estimated texture map $\hat{\mathbf{T}}$ that is consistent with the users request. We also show results of compositing virtual objects directly into a scene in fig. 6. As shown in fig. 1, realistic rendering results can intuitively be edited to achieve the desired look.
To the best of our knowledge, the only other method which allows intuitive editing of indoor lighting estimate is that of Gardner et al. [10]. Unfortunately, realistic renders are limited to diffuse objects and cannot be extended to reflective objects as shown in fig. 5.
7 Discussion
This paper proposes a lighting estimation approach which produces an intuitive, user-editable lighting representation given a single indoor input image. By explicitly representing the dominant light source using a parametric model, and the ambient environment map using a textured cuboid, our approach bridges the gap between generating realistic shading (produced by HDR light sources) and reflections (produced by textured environment maps) on rendered virtual objects. We demonstrate, through extensive experiments, that our approach provides competitive quantitative performance when compared to recent lighting estimation techniques. In particular, when compared to the only other approach which can be user-edited [10], our approach yields significant improved results.
Limitations and future work
While our proposed approach estimates a 3D representation of the surrounding lighting environment, it does not reason about light occlusions in the scene as opposed to other techniques such as [12, 19, 27]. Incorporating these higher-order interactions while maintaining interpretability and editability of the output representation is an interesting direction for future research. In addition, the estimated environment textures were shown to produce realistic reflections on shiny objects, but a close inspection reveals that they are low resolution and contain some visual artifacts. It is likely that more recent image-to-image translation architectures [4, 21] could be used to improve realism.
Acknowledgements
This research was supported by MITACS and the NSERC grant RGPIN-2020-04799. The authors thank Pascal Audet for his help.
References
[1]
Bai, J., Guo, J., Wan, C., Chen, Z., He, Z., Yang, S., Yu, P., Zhang, Y., Guo,
Y.: Deep graph learning for spatially-varying indoor lighting prediction.
arXiv preprint arXiv:2202.06300 (2022)
[2]
Barron, J.T., Malik, J.: Shape, illumination, and reflectance from shading.
IEEE TPAMI 37(8), 1670–1687 (2014)
[3]
Cheng, D., Shi, J., Chen, Y., Deng, X., Zhang, X.: Learning scene illumination
by pairwise photos from rear and front mobile cameras. Computer Graphics
Forum 37(7), 213–221 (2018)
[4]
Choi, Y., Uh, Y., Yoo, J., Ha, J.W.: Stargan v2: Diverse image synthesis for
multiple domains. In: CVPR (2020)
[5]
Cruz, S., Hutchcroft, W., Li, Y., Khosravan, N., Boyadzhiev, I., Kang, S.B.:
Zillow indoor dataset: Annotated floor plans with 360º panoramas and 3d room
layouts. In: CVPR (2021)
[6]
Debevec, P.: Rendering synthetic objects into real scenes: Bridging traditional
and image-based graphics with global illumination and high dynamic range
photography. In: Proceedings of the 25th Annual Conference on Computer
Graphics and Interactive Techniques. pp. 189–198. SIGGRAPH (1998)
[7]
Eigen, D., Fergus, R.: Predicting depth, surface normals and semantic labels
with a common multi-scale convolutional architecture. In: Proceedings of the
IEEE international conference on computer vision. pp. 2650–2658 (2015)
[8]
Einabadi, F., Guillemaut, J.Y., Hilton, A.: Deep neural models for illumination
estimation and relighting: A survey. Computer Graphics Forum 40(6),
315–331 (2021)
[9]
Fernandez-Labrador, C., Facil, J.M., Perez-Yus, A., Demonceaux, C., Civera, J.,
Guerrero, J.J.: Corners for layout: End-to-end layout recovery from 360
images. IEEE Robotics and Automation Letters 5(2), 1255–1262
(2020)
[10]
Gardner, M.A., Hold-Geoffroy, Y., Sunkavalli, K., Gagne, C., Lalonde, J.F.:
Deep parametric indoor lighting estimation. In: ICCV (2019)
[11]
Gardner, M.A., Sunkavalli, K., Yumer, E., Shen, X., Gambaretto, E., Gagné,
C., Lalonde, J.F.: Learning to predict indoor illumination from a single
image. ACM TOG 36(6) (2017)
[12]
Garon, M., Sunkavalli, K., Hadap, S., Carr, N., Lalonde, J.F.: Fast
spatially-varying indoor lighting estimation. In: CVPR (2019)
[13]
Grosse, R., Johnson, M.K., Adelson, E.H., Freeman, W.T.: Ground truth dataset
and baseline evaluations for intrinsic image algorithms. In: 2009 IEEE 12th
International Conference on Computer Vision. pp. 2335–2342. IEEE (2009)
[14]
Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K.Q.: Densely connected
convolutional networks. In: CVPR (2017)
[15]
Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. arXiv
preprint arXiv:1412.6980 (2014)
[16]
Lee, C.Y., Badrinarayanan, V., Malisiewicz, T., Rabinovich, A.: Roomnet:
End-to-end room layout estimation. In: ICCV (2017)
[17]
LeGendre, C., Ma, W.C., Fyffe, G., Flynn, J., Charbonnel, L., Busch, J.,
Debevec, P.: Deeplight: Learning illumination for unconstrained mobile mixed
reality. In: CVPR (2019)
[18]
Li, T.M., Aittala, M., Durand, F., Lehtinen, J.: Differentiable monte carlo ray
tracing through edge sampling. ACM TOG 37(6), 1–11 (2018)
[19]
Li, Z., Shafiei, M., Ramamoorthi, R., Sunkavalli, K., Chandraker, M.: Inverse
rendering for complex indoor scenes: Shape, spatially-varying lighting and
svbrdf from a single image. In: CVPR (2020)
[20]
Mandl, D., Yi, K.M., Mohr, P., Roth, P., Fua, P., Lepetit, V., Schmalstieg, D.,
Kalkofen, D.: Learning lightprobes for mixed reality illumination. In: ISMAR
(2017)
[21]
Park, T., Efros, A.A., Zhang, R., Zhu, J.Y.: Contrastive learning for unpaired
image-to-image translation. In: ECCV (2020)
[22]
Sengupta, S., Gu, J., Kim, K., Liu, G., Jacobs, D.W., Kautz, J.: Neural inverse
rendering of an indoor scene from a single image. In: ICCV (2019)
[23]
Somanath, G., Kurz, D.: Hdr environment map estimation for real-time augmented
reality. In: CVPR (2021)
[24]
Song, S., Funkhouser, T.: Neural illumination: Lighting prediction for indoor
environments. In: CVPR (2019)
[25]
Srinivasan, P.P., Mildenhall, B., Tancik, M., Barron, J.T., Tucker, R.,
Snavely, N.: Lighthouse: Predicting lighting volumes for spatially-coherent
illumination. In: CVPR (2020)
[26]
Wang, T.C., Liu, M.Y., Zhu, J.Y., Tao, A., Kautz, J., Catanzaro, B.:
High-resolution image synthesis and semantic manipulation with conditional
gans. In: CVPR (2018)
[27]
Wang, Z., Philion, J., Fidler, S., Kautz, J.: Learning indoor inverse rendering
with 3d spatially-varying lighting. In: ICCV (2021)
[28]
Weber, H., Prévost, D., Lalonde, J.F.: Learning to estimate indoor lighting
from 3d objects. In: 3DV (2018)
[29]
Wiles, O., Gkioxari, G., Szeliski, R., Johnson, J.: Synsin: End-to-end view
synthesis from a single image. In: CVPR (2020)
[30]
Yang, C., Zheng, J., Dai, X., Tang, R., Ma, Y., Yuan, X.: Learning to
reconstruct 3d non-cuboid room layout from a single rgb image. In: Wint.
Conf. App. Comp. Vis. (2022)
[31]
Zhan, F., Yu, Y., Wu, R., Zhang, C., Lu, S., Shao, L., Ma, F., Xie, X.:
Gmlight: Lighting estimation via geometric distribution approximation. IEEE
TIP (2022)
[32]
Zhan, F., Zhang, C., Hu, W., Lu, S., Ma, F., Xie, X., Shao, L.: Sparse needlets
for lighting estimation with spherical transport loss. In: ICCV (2021)
[33]
Zhan, F., Zhang, C., Yu, Y., Chang, Y., Lu, S., Ma, F., Xie, X.: Emlight:
Lighting estimation via spherical distribution approximation. In: AAAI (2021)
[34]
Zhao, Y., Guo, T.: Pointar: Efficient lighting estimation for mobile augmented
reality. In: ECCV (2020)
[35]
Zou, C., Colburn, A., Shan, Q., Hoiem, D.: Layoutnet: Reconstructing the 3d
room layout from a single rgb image. In: CVPR (2018) |
A high-granularity calorimeter insert based on SiPM-on-tile technology
at the future Electron-Ion Collider
Miguel Arratia
[email protected]
Kenneth Barish
Liam Blanchard
Huan Z. Huang
Zhongling Ji
Bishnu Karki
Owen Long
Ryan Milton
Ananya Paul
Sebouh J. Paul
Sean Preins
Barak Schmookler
Oleg Tsai
Zhiwan Xu
Department of Physics and Astronomy, University of California, Riverside, CA 92521, USA
Department of Physics and Astronomy, University of California, Los Angeles, CA 90095, USA
Abstract
We present a design for a high-granularity calorimeter insert for future experiments at the Electron-Ion Collider (EIC). The sampling-calorimeter design uses scintillator tiles read out with silicon photomultipliers. It maximizes coverage close to the beampipe, while solving challenges arising from the beam-crossing angle and mechanical integration. It yields a compensated response that is linear over the energy range of interest for the EIC. Its energy resolution meets the requirements set in the EIC Yellow Report even with a basic reconstruction algorithm. Moreover, this detector will provide 5D shower data (position, energy, and time), which can be exploited with machine-learning techniques. This detector concept has the potential to unleash the power of imaging calorimetry at the EIC to enable measurements at extreme kinematics in electron-proton and electron-nucleus collisions.
Contents
1 Introduction
2 Proposed Design
2.1 External constraints
2.2 Sampling calorimeter layout
2.3 Scintillator cells
2.4 Silicon photo-multipliers
2.5 SiPM-carrying boards and backplane board
2.6 Accessibility for maintenance
2.7 Geometric acceptance
2.8 Total calorimeter thickness
3 Simulation and performance
3.1 Geometry and simulation setup with DD4HEP
3.2 Shower-shape examples
3.3 Hit distributions
3.4 Linearity for hadrons and electrons and $e/h$ ratio
3.5 Non-Gaussian tails due to leakage
3.6 Single-hadron energy resolution
3.7 Pseudorapidity dependence of performance
4 Summary and conclusions
1 Introduction
One of the key requirements for detectors at the future Electron-Ion Collider (EIC) is to have tracking and full calorimetry with 2$\pi$ azimuthal coverage over a large range in pseudorapidity—nominally $|\eta|<4.0$ (1). This would ensure a “$4\pi$, general-purpose detector” that could deliver the original EIC scientific goals (2) and much beyond.
While both the ATHENA (3) and ECCE (4) detector designs contemplate close-to-full coverage with full calorimetry, the specifics on how to implement it in the forward region remain undefined. Covering the region $3<|\eta|<4$ is challenging due to the EIC beam-crossing angle, which is 25 mrad (1).
As illustrated in Fig. 1, the beampipe envelope crosses the region where the forward hadronic calorimeter (HCal) would be located at an angle that approximately corresponds to the proton direction. Any detector covering this region needs to simultaneously fill a complex volume and keep clearance from the beampipe, while fitting the other calorimeters without needing additional support structures.
In addition, the detector in the $3<\eta<4$ region needs to be well matched to the particle densities and energies expected at the EIC. At the highest energy setting of 18 GeV electron beam and 275 GeV proton beam, the jet spectra reach close to the proton-beam energy with a high rate at nominal luminosity (1). Furthermore, single-particle measurements up to 60 GeV are set as a requirement (1). Given the large energies involved, an HCal becomes crucial since poor tracking performance is expected at forward angles (5; 1).
Instrumenting the $3<\eta<4$ region demands a relatively small detector area ($\approx 0.6\times 0.6$ m${}^{2}$ for a detector located at $z=3.8$ m) but also high granularity to yield a reasonable $\eta$ resolution and disentangle nearby particles. Small cells can also tessellate the complex geometry near the beampipe in an effective manner.
Good angular coverage and resolution would be beneficial for measurements of jets and their substructure, which will enable a wide variety of studies at the EIC, and inclusive deep-inelastic scattering at high-$x$ (1; 6).
The concept of high-granularity calorimetry has been developed for the last two decades or so (7), driven by the CALICE collaboration, and motivated by the particle-flow paradigm (8). More recently, this approach has been adopted for the CMS High-Granularity Calorimeter (HGCAL) upgrade (9), and is envisioned for experiments at several proposed $e^{+}e^{-}$ colliders such as ILC (10), CLIC (11), and CEPC (12; 13; 14).
Extensive R$\&$D efforts (7) and rapid progress in silicon-photomultiplier (SiPM) technology (15) have culminated in the “SiPM-on-tile” concept for hadronic calorimetry. In this approach, the active elements of the calorimeter are small scintillator tiles ($\approx$ 3$\times$ 3 cm${}^{2}$) air coupled to a SiPM at their center (16; 17), without using wavelength-shifting fibers that were used in earlier designs (18; 19; 20; 21).
The advent of high-granularity calorimeters has boosted studies of machine-learning algorithms (22) that could exploit the benefits of “imaging calorimetry” for a variety of tasks (23; 24; 25; 26; 25; 27; 28; 29; 30; 31; 32; 33; 34; 35; 36; 37).
Some early proposals for EIC detectors, such as TOPSIDE, considered imaging calorimetry over the entire solid angle (38; 1), including hadronic calorimetry similar to the CALICE design, although those options were not further pursued111In particular, they were not in any of the official EIC proposals (3; 4; 39). On the other hand, an imaging electromagnetic calorimeter based on silicon-pixel sensors was included in the ATHENA proposal (3). Imaging calorimetry with silicon sensors is also envisioned for the zero-degree calorimeters.. Instead, the forward HCals for both the ATHENA (3) and ECCE (40) proposals included a more traditional steel-scintillator design with longitudinal granularity that varies between 4 and 7 segments, which is similar to that used in the ZEUS (41) and H1 (42) detectors at HERA.
The SiPM-on-tile technology represents an attractive option to instrument the $3<\eta<4$ region at the EIC for various reasons. First, the expected energy range for jets largely overlaps with that of proposed $e^{+}e^{-}$ Higgs factories, so the vast CALICE literature and test-beam data (18; 19; 20; 43; 44; 45; 46; 47; 48; 49; 50; 51; 52; 53; 54) could inform designs and validate simulations. Second, the radiation level expected at the EIC is orders of magnitude smaller than the LHC so constraints from SiPM radiation damage are not a limiting factor. Third, the forward geometry limits the detector volume, numbers of channels, and cost, while still covering a significant range in $\eta$–$\phi$ space.
In this paper, we present a concept for a high-granularity calorimeter insert (HG-CALI) based on the SiPM-on-tile technology that achieves excellent acceptance near the beampipe, while satisfying the mechanical constraints imposed by integration with the rest of the detector. It is designed to be compensated ($e/h\approx 1$) and meets the requirements set in the EIC Yellow Report (1) with a basic reconstruction algorithm. Moreover, it will open up potential for improvements based on machine-learning techniques that exploit the 5D measurement of showers. It would be complemented with a high-granularity electromagnetic calorimeter insert based on tungsten powder/scintillating fiber technology (55), which will be described in a separate publication.
2 Proposed Design
2.1 External constraints
Two main factors constrain the design of the proposed calorimeter insert. First, it must integrate into the larger calorimeter endcap without additional support structures that would create dead areas. Second, it must surround the beampipe within some minimal acceptable clearance in order to maximize its coverage.
In the ATHENA design (3), the forward HCal design is based on the technology previously used in the STAR experiment (56; 57) and consists of alternating layers of steel absorbers (consisting of $96\times 98\times 20$ mm blocks) and scintillator tiles with a thickness of 3 mm. The absorber blocks of both the endcap and the insert are held in place using dowel pins which go into the holes on the top of each absorber block, through holes of the 1.9 mm thick “master plates”, and then into the holes in the blocks above them. This is illustrated in Fig. 2, which shows the high-granularity insert (described in more detail in the following sections) surrounded by the full HCal endcap (left), and zoomed-in with some of the adjacent structures within the endcap (right).
The endcap HCal is designed to open up by splitting in half vertically for maintenance, as it will rest on a support structure with a rail system (1). Therefore, the insert is designed to split into two parts along with the endcap.
The current design for the beampipe (58) is conically shaped and has an outer radius of 11.15 cm at the upstream face of the HCal ($z=380$ cm) and 13.40 cm at the front face of the last layer of the HCal ($z=498$ cm). Furthermore, the beampipe axis is at a small angle ($-$24.3 mrad) relative to the lab-frame $z$ axis, which is defined parallel to the solenoidal magnetic field axis and to the electron beampipe. The proton beam axis is at $-$25 mrad relative to the electron axis. Distances and angles defined with respect to the proton axis are labelled with an asterisk superscript throughout this paper. We assume a clearance of about 4 cm along the entire volume of the HCal insert.
Based on these requirements, the proposed HG-CALI will cover the equivalent area of 6$\times$6 blocks of the HCal endcap, or approximately 60$\times$60 cm. However, since the beampipe is offset in the negative $x$ direction, the part of the insert on the left side when viewed from upstream (positive $x$) covers the width of two HCal endcap blocks, plus the gaps between them, (19.6 cm), while the right side (negative $x$) of the insert covers the width of four HCal endcap blocks and the gaps between them (39.6 cm). Together with the 4 mm gap between the two halves of the HCal, and the 4 mm of readout backplanes on either side, the total width of the proposed insert is 60.4 cm.
To account for the beampipe, we include a hole in each layer of the insert, as shown in the left side of Fig. 3.
The shape of this hole is circular for the most part. However, the right side of the insert is shaped like a capital letter D, with flat extensions on the top and bottom to ensure clearance even while the calorimeter is sliding open.
The radii and position of the centers of the circles vary from layer to layer, as shown in the top image of Fig. 4, because of the conical shape and tilt (respectively) of the beampipe. This approach leads to an optimal layout, in the sense that it covers all the available volume, being limited only by the clearance from the beampipe. The dimensions of the insert are shown in the bottom image of Fig. 4.
Although we have shown a specific way to implement the HG-CALI with the ATHENA HCal design (3), this could be adapted to fit the ECCE HCal (40) design, which also uses a modular assembly structure based on rectilinear towers that run parallel to the $z$ axis. The HG-CALI could also be incorporated into the second EIC detector (59); for example, it is also compatible with the CORE design (39). It could also be used in future experiments at the Chinese EicC, which considers a crossing-angle of 50 mrad (60).
2.2 Sampling calorimeter layout
The HG-CALI consists of alternating absorber blocks and scintillator assemblies. Each scintillator assembly consist of a plastic scintillator sandwiched between a 2 mm cover on the upstream side of it and a printed-circuit board (PCB) with SiPMs for readout on the downstream side. It has 50 layers, each of which is 2.34 cm thick, followed by a downstream absorber-only layer. Figure 5 shows the longitudinal dimensions of the absorbers and the different components of the scintillator assemblies.
The HG-CALI is divided into two sets of layers with different absorber materials. In the first, tungsten is used; in the second, high-strength steel is used. The use of tungsten is motivated for two reasons: first, it allows for a compensated design; second, its lower radiation length and nuclear-interaction lengths compared to steel ($X_{0}=$ 0.35 cm, $\lambda_{n}$= 9.95 cm for W and $X_{0}=$ 1.76 cm, $\lambda_{n}$= 16.8 cm for Fe) helps in reducing leakage of the core of hadronic showers to the beampipe.
The use of magnetic steel is required to contain magnetic-field flux. The EIC poses a strict requirement on fringe fields, mainly driven by the need of accurate polarization of the beams. The amount of steel required in this region is still under study and no formal requirement has been established. We have assumed a total thickness of 33.81 cm of steel (the last 21 layers) for the simulations presented in Sec. 3.
2.3 Scintillator cells
A single scintillator “megatile” is used for each half-layer, which has a slightly different shape to match the corresponding absorber shape. The scintillator panels are tessellated with hexagonal cells defined by grooves.
Each cell will have a dimple in its center above the SiPMs that are hosted on the PCB. The grooves are filled with a mixture of epoxy and titanium dioxide pigment, following procedures similar to the STAR electromagnetic calorimeter (61). An alternative could be to paint the cells’ edges with reflective paint, such as Saint Gobain BC-621. The latter approach was used in the CMS HCAL (62).
To fully tessellate each layer, some of the cells on the edges will be cropped or extended. This is not necessarily an issue given that the design allows for a cell-to-cell calibration based on minimum-ionizing particles.
The megatile is covered on the front and back with reflective foil like the VM2000 enhanced specular reflector (ESR) to increase the light yield and to improve uniformity across the cell area. The ESR foil on the downstream side of the megatiles will include holes for the dimples in each cell. The ESR foil will not be wrapped to facilitate assembly; instead, the edges of the scintillator megatiles will be painted.
The dimples in the center of each cell will have the same dimensions used in the CMS HGCAL test beam (63), which define a spherical dome with a radius of curvature of 3.8 mm and a depth of 1.6 mm, as shown in Fig. 6. The dimples improve signal uniformity across the area of the cell by alleviating hot-spot effects near the SiPM (16; 17).
The megatile concept has several advantages over a design in which each tile is wrapped individually, such as the ones used in the CALICE prototype (20) and the CMS HGCAL (9), especially for a small and complicated geometry such as that of HG-CALI. The megatile design would simplify assembly and reduce cell-to-cell variations caused by wrapping with ESR foil (64).
Previous experience with calorimeters that use megatiles such as the STAR electromagnetic calorimeter (61) suggest that optical cross-talk can be reduced to about 2$\%$ when using a mixture of epoxy and titanium dioxide to fill the grooves that define the cells. This level of optical crosstalk is acceptable for this application, given that hadronic showers encompass a large number of hits over a wide area.
The megatile concept provides us with the flexibility to define the transverse granularity of the detector, and thus optimize it. We envision that the area of these cells will vary from layer to layer, for example as shown in Fig. 7, with finer granularity in the furthest upstream layers and coarser granularity in the downstream. The granularity is yet to be optimized, but we estimate that a reasonable choice would yield a total of less than 5k channels.
Test-beam results by the CMS HGCAL collaboration (63) showed that the light yield of 3 mm thick scintillator cells (Eljen
Technology EJ-200) is about 32 photo-electrons for 9 cm${}^{2}$ cells; about 26 photo-electrons for 12 cm${}^{2}$ cells, and about 20 photoelectrons for 36 cm${}^{2}$ cells.
2.4 Silicon photo-multipliers
Both the CALICE and CMS collaborations use 1.3$\times$1.3 mm${}^{2}$ SiPMs (63; 9). Larger-area SIPMs were considered, e.g., by Jiang et al. in the context of R$\&$D for the CEPC (65). Unlike the applications for ILC or CEPC, which will use millions of channels, or the CMS HGCAL, which uses about 200k channels, our design will have just a few thousand channels. Therefore, the total cost of SiPM would still be reasonable even if we choose models with larger areas.
We anticipate that larger area SiPMs will make the calorimeter more robust against non-uniformity created by SiPM misalignment (64), as well as saturation effects. The photoelectron yield increases linearly with the area of the SiPM, so choosing a 9 mm${}^{2}$ SiPM could increase the light yield by roughly a factor of 4.5 with respect to the CMS HGCAL (63).
As we will show in Sec. 3, the maximum hit energy reaches $\approx 200$ MeV or 250 MIPs. We anticipate an operation for 30 photo-electrons per MIP for the $3\times 3$ cm${}^{2}$ cells. This high light yield would create a buffer to cover for the reduced signal-to-noise ratio after radiation damage.
The Hamamatsu model S13360-3025PE has 14.4k pixels, which translates into a dynamic range of more than 400 MIPs, so no saturation effect is expected for $3\times 3$ cm${}^{2}$ cells.
The neutron fluence expected at the EIC near the insert region will not exceed $10^{12}$ neutrons/cm${}^{2}$ per year from the electron-proton collisions at the top luminosity $(10^{34}$ cm${}^{-2}$s${}^{-1})$ (1). The corresponding ionizing radiation dose is 25 Gy per year.
The expected SiPM radiation damage would be classified as “modest” in Garutti’s review (66) on the topic. It would mainly lead to increased dark current that can affect the ability to identify single-photon peaks. This level of damage can partially be recovered with annealing (66), which could be done after each run (as explained in Sec. 2.6).
For reference, CMS uses SiPM-on-tile technology for regions where the total dose is less than 3000 Gy and fluences less than 8$\times$10${}^{13}$ $n_{\mathrm{eq}}$/cm${}^{2}$, although the SiPMs will be operated at $-$30 C to reduce the impact of the increased noise (9).
The STAR experiment at RHIC is currently operating a steel-scintillator calorimeter readout with SiPMs (56) that is expected to receive fluences that are similar to those expected at the EIC. Detailed data on radiation-induced noise taken with STAR will guide future iterations of the HG-CALI design.
2.5 SiPM-carrying boards and backplane board
The SiPMs will be hosted in a PCB with no active elements. The PCB will also host LEDs next to each SiPM in the dimple area to be used for monitoring and calibration purposes, similarly to the CMS HGCAL design (9). The SiPM-carrying board will host a high-density connector to connect to the backplane board. The SiPMs will not be actively cooled.
On either side of the insert are 4 mm PCB backplanes used for readout, which the PCB for each scintillator assemblies connect into like a traditional crate. Figure 8 shows the location of the PCB backplanes.
The SiPM-carrying boards will be plugged in and unplugged one-at-a-time during assembly and maintenance. The backplane board will also be used to power the SiPMs and control the LED system for calibration and monitoring.
2.6 Accessibility for maintenance
The scintillator assemblies could be removed for maintenance by sliding them out laterally from between the absorbers when the HCal is opened, as illustrated in Fig. 9. This will be possible because the interaction point will have a rail system that will be integrated to the support structure (1).
This feature ensures that the SiPMs boards could be annealed at high-temperature after each run to minimize radiation damage, and that SiPMs and scintillator tiles could be replaced.
2.7 Geometric acceptance
The overall effects of the EIC crossing angle and other beam parameters are discussed in Ref. (67). Events in the lab frame, where the beams have a non-zero crossing angle, can be Lorentz transformed into a frame where the beams are anti-parallel and have energies very close to those in the lab frame. This “minimally-transformed, collinear frame” is a natural choice to extract many physics observables and the accompanying particles’ angular distributions.
In the hadron- (electron-) going direction, scattering angles defined in the lab frame relative to the hadron (electron) beam will be equivalent to the scattering angle in the minimally-transformed collinear frame (68). In particular, for a detector in the hadron-going direction, its angular acceptance should be defined relative to the hadron beam.
To estimate the geometrical acceptance of the insert, we determined where cones of constant $\eta^{*}$ (that is, pseudorapidity with respect to the $z^{*}$ axis – the proton beam axis) would intersect the detector.
Figure 10 shows projections of these $\eta^{*}$ cones superimposed on a $yz^{*}$ (vertical vs. longitudinal) cross-section of the insert. Within this slice, the $\eta^{*}=4.0$ cone skims the inner edge of the detector, while those at lower $\eta^{*}$ hit the face of the detector.
To determine the geometrical acceptance over the full azimuthal range we projected these cones of constant $\eta^{*}$ onto the faces of the scintillators planes of the insert, resulting in elliptical rings of constant $\eta^{*}$. An example is shown in the inset of Fig. 11 for $\eta^{*}$=3.7, which covers the full azimuthal acceptance on all layers of the detector.
We show in Fig. 11 the cross sections of these cones for several values of $\eta^{*}$ at the first layer (top left panel), one of the middle layers (top right panel), and the last layer (bottom left panel).
While there is some acceptance up to $\eta^{*}=4.0$, this is limited to one side of the detector, and it barely skims the first layer, causing a degradation of the signal, as we quantify in Sec. 3.
Since the final detector design may differ from the baseline design presented in this work, we discuss possible alternatives: with a clearance parameter of 3 cm instead of 4 cm, the entire $\eta^{*}=3.8$ ring fits entirely within the first scintillator layer. If the clearance parameter were increased to 5 cm, the full $\eta^{*}=3.7$ ring would still be within the acceptance of the detector, but it would come very close to the edge, resulting in some degradation.
The relevant scale to compare various clearance options has to be the effective radiation length of the insert, which in our design is minimized using mostly tungsten as an absorber. This is because the relevant metric is whether or not the core of the hadronic showers can be well contained, as we describe more in Sec. 3.
Given that even seemingly small differences of clearance definitions can have a large impact in extending the $\eta^{*}$ coverage of the detector, careful engineering design and studies are required to minimize the clearance as much as possible.
2.8 Total calorimeter thickness
We determined the amount of material (in nuclear interaction lengths, $\lambda_{n}$) that a particle travelling in a straight line would pass through in the HG-CALI and the beampipe as modeled in Ref. (58), as a function of its polar angle $\theta^{*}$ with respect to the $z^{*}$ axis. We averaged this with respect to the azimuthal angle $\phi^{*}$ and show the results in the left panel of Fig. 12.
The HG-CALI covers polar angles from about 2${}^{\circ}$ to 6${}^{\circ}$, with up to about 7 nuclear interaction lengths. Between 3${}^{\circ}$ and 6${}^{\circ}$, there is overlap between the HG-CALI and the HCal endcap. The HCal endcap has a somewhat lower number of nuclear interaction lengths (about 6) than the insert. There is a small amount of material traversed by particles passing through the sides of the beampipe (about 1$\lambda_{n}$), which may allow the detection of showering at very small $\theta^{*}$.
We show in the right panel of Fig. 12 the dependence of the amount of material traversed in the HCal insert on $\phi^{*}$ and $\eta^{*}$. The amount of material is nearly uniform in $\phi^{*}$ in the $3.5\leq\eta^{*}\leq 3.8$ range. Since the insert’s outer shape is (approximately) a square, the value of $\eta^{*}$ at the edge of the insert depends non-trivially on $\phi^{*}$. At about $3.0<\eta^{*}<3.3$ (depending on $\phi^{*}$) particles may skim the edge of the insert, creating a signal in both the insert and the HCal. Particles with $3.3<\eta^{*}<3.9$ encounter at least two nuclear interaction lengths of material in the insert at all azimuthal angles.
The HG-CALI can complement the HCal endcap at angles where the effective thickness is not maximal by providing a measurement of the shower-start position and incident-particle angle, both of which will be better measured in the insert.
3 Simulation and performance
3.1 Geometry and simulation setup with DD4HEP
We used the DD4HEP framework (69) to define the geometry and ran the Geant4 framework (70) to evaluate the performance of the HG-CALI. The version of Geant4 used is 10.7.p03.
Table 1 shows the materials, thicknesses, and numbers of interaction lengths for each component of a single layer of the insert (see Fig. 5). The previously described layer design is followed: 30 layers W/Sc, 20 layers steel/Sc, and 1 final layer of steel.
For simplicity, the granularity of the insert was assumed to be 3$\times$3 cm${}^{2}$ for all the layers. Additionally, the model did not include the dimples and grooves in the scintillator layers, nor the SiPMs in the PCB. The model included a realistic beampipe geometry and material description (58).
We also included in the simulation an HCal model similar to the STAR/ATHENA design (56; 3), assuming a granularity of 10$\times$10 cm${}^{2}$, 20 mm thick absorbers, and 3 mm thick scintillators.
The solenoidal magnetic field is expected to be low in the forward region relevant for the insert, but is not precisely known at this time so it was not included in the simulation.
For the simulation of the particle shower in the material, we used the FTFP_BERT_HP physics list of Geant4, which well describes the CALICE test-beam data (51), including the time evolution of hadronic showers in both tungsten and steel absorbers (49). To model saturation effects in the scintillator response we applied Birks’ law (71) with a constant 0.126 mm/MeV.
The particle origin position was assumed to be at $z=0$ and the first layer of the HCal insert was placed at $z=380$ mm.
No noise was included in the simulation, as a realistic implementation is not available at this time, but its effect was accounted for with an energy threshold as described in the next section. For the digitization step, a 12-bit dynamic range for the ADC and maximum value set to 200 MeV.
Each cell was assigned a timestamp dictated by the particle with the earliest time of arrival to the cell. The effect of the time resolution was included as a cell-by-cell smearing with a Gaussian with $\mu=0$, $\sigma=$1 ns. Unlike in Ref. (51), no SiPM-saturation effect was simulated since this effect is expected to be negligible with the model chosen as baseline222For reference, the SiPM used in the CALICE W-AHCAL prototype contained about 1k channels whereas the one used in our design has 14.4k channels..
The hits were also required to have $E_{\mathrm{hit}}>0.1E_{\mathrm{MIP}}$, where $E_{\mathrm{MIP}}$ is the most-probable energy deposited by a minimum-ionizing particle (MIP), which is about 0.8 MeV as obtained from simulations of high-energy muons. This energy cut was assumed to reduce the noise to a negligible level, as it will correspond to a cut of about 3 photoelectrons. Our selection was more relaxed than the CALICE studies (20) that used an older generation of SiPMs with significantly larger noise, after-pulsing, and cross-talk.
We used a time cut of $t<50$ ns, where $t=0$ is defined when the generated particle hit the front of the insert. As we will show in Sec. 3.4, this selection yields a compensated response with a basic reconstruction algorithm. Much larger integration times might be possible given the EIC bunch structure and absence of pileup. Such time information can be exploited with advanced algorithms, such as those proposed in Refs. (72; 29; 73).
3.2 Shower-shape examples
Figure 13 shows some example shower shapes for a 50 GeV $\pi^{-}$ with $\eta^{*}=3.7$ and a random $\phi^{*}$. As expected, the $\pi^{-}$ showers can be roughly described as having a narrow core from electromagnetic sub-showers and a halo from hadronic sub-showers. The core typically contains cells with large energy deposits, whereas the halo is composed of mainly low-energy cells. The latter are dominated by the copious production of neutrons produced in tungsten. Similar observations have been made with CALICE prototypes (7).
The hadronic-shower shapes, although irregular, provide a hint on the performance of the insert. The production of neutrons, which drives the hadronic halo, occurs over wide angles. Some of the neutrons will be emitted in the direction of the beampipe hole, but often these appear in the other side of the calorimeter instead of being lost altogether. On the other hand, the core of the shower is electromagnetic in nature so its size is driven by the radiation length, which is short in tungsten.
3.3 Hit distributions
Figure 14 shows the hit-energy and multiplicity distributions for a 50 GeV $\pi^{-}$ with $\eta^{*}=3.7$. We show the distributions separately for all hits above the 0.1 MIP threshold, all hits that have more energy than the average of all the hits in their respective showers, and those with more than 5 MIPs of energy.
The average hit energy was found to be about 1.6 MeV, with 17% of the hits containing energy above 1.6 MeV and 7.6% containing energy above 5 MIPs (4 MeV).
The maximum hit energy was about 200 MeV. The hits-per-shower distribution shows a broad peak around 275 hits, whereas the hits above average show a narrower distribution centered around 50 hits, and those above 5 MIPs show an average of 20 hits.
3.4 Linearity for hadrons and electrons and $e/h$ ratio
Figure 15 shows the energy distributions of showers at $\eta^{*}$= 3.7 both for $\pi^{-}$ and $e^{-}$ with various energies. The total reconstructed energy was computed as:
$$E_{total}=\frac{E_{Insert}}{SF_{Insert}}+\frac{E_{HCal}}{SF_{HCal}}$$
(1)
where $E_{Insert}$ and $E_{HCal}$ are the sum of hit energies in the insert and HCal respectively, and $SF_{Insert}$ and $SF_{HCal}$ are the sampling fractions for the insert and HCal, respectively. The sampling fraction is computed by taking the ratio of the output energy to the input energy in the calorimeter. The sampling fraction is 0.9$\%$ for the insert and 2.2$\%$ for the HCal.
The response for pions in the insert is mostly Gaussian but it shows a low-energy tail, which is much less pronounced with the combined measurement using the HCal insert and the HCal (this is quantified in Sec. 3.5). This suggests that the low-energy tail is due to leakage in the transverse direction from the HCal insert towards the HCal.
The response for electrons is mostly Gaussian. This also suggests that that the electromagnetic core of the hadronic showers is well contained due to short radiation length of tungsten.
Figure 16 shows the mean energy deposited in the HCal insert for electrons and pions, along with the $e/h$ ratio, which is found to be about 0.97 on average.
3.5 Non-Gaussian tails due to leakage
Leakage is quantified as the fraction of events in which the shower energy is below $\mu-3\sigma$, where $\mu$ and $\sigma$ are obtained from the Gaussian fit to the distribution. Figure 17 shows the leakage for $\pi^{-}$ and $e^{-}$ at different energies and $\eta^{*}$=3.7. The leakage in the insert alone for $\pi^{-}$ is about 10% for the entire energy range but relatively small (about 1% or lower) for electrons. After considering both the HCal and the HCal insert, the leakage for $\pi^{-}$ showers drops to about $1\%$, suggesting most of the leakage at $\eta^{*}$=3.7 is in the transverse direction with a minimal amount of it leaking through the beampipe. This indicates that neutrons, which are produced over a large solid angle, traverse the hole and are measured rather than being lost in the small solid angle of the beampipe exit (as illustrated in Fig. 13).
3.6 Single-hadron energy resolution
Figure 18 shows that the resolution, defined as ratio of the width ($\sigma$) to the mean value of the fitted energy distribution, for charged pions at $\eta^{*}$=3.7. The HCal insert meets the requirements set in the EIC Yellow Report (1), even beyond the pseudorapidity range for which it was set.
The expected resolution for the HG-CALI is similar to the CALICE test-beam data (51). We have also implemented the CALICE geometry in DD4HEP for both the steel and tungsten prototypes and were able to reproduce the linearity and the resolutions reported in Ref. (51) to within 10$\%$, indicating that the similarity that we see in Fig. 18 is not coincidental.
3.7 Pseudorapidity dependence of performance
Figure 19 shows the distribution of shower energies at a constant pion energy of 50 GeV but with different $\eta^{*}$. The measured distributions using both the HCal insert and HCal can be well described by a Gaussian fit, but the fit quality deteriorates with increasing $\eta^{*}$. For $\eta^{*}>$3.9, the fraction of events in the low-energy tail increases due to losses to the beampipe hole. This stems from the worse geometrical acceptance of the HCal insert here (see Sec. 2.7).
Figure 20 shows the leakage for 50 GeV $\pi^{-}$ and $e^{-}$ at different $\eta^{*}$. Taking both the HCal endcap and the HCal insert into account removes most of the leakage, which is in the transverse direction and can be recovered, as shown by solid markers. At the high $\eta^{*}$ the leakage increases sharply but is still only about 5$\%$ at $\eta^{*}=3.9$ for $\pi^{-}$ and about 2$\%$ for $e^{-}$.
Figure 21 shows the energy resolutions obtained with the HCal insert and the HCal as a function of $\eta^{*}$ for 50 GeV $\pi^{-}$. The resolution obtained with HG-CALI only is constant at about $9\%$ over the region $3.4<\eta^{*}<3.8$ and it degrades at lower and higher $\eta^{*}$. Including the HCal improves the resolution only at low $\eta^{*}$ but not at high $\eta^{*}$. The horizontal line represents the EIC Yellow Report requirement (1), which is defined up to $\eta=3.5$, beyond which only “degraded performance” is specified. The HG-CALI resolution meets the YR requirement up to $\eta^{*}=3.8$, and gets slightly worse at the highest $\eta^{*}$ region due to the poorer acceptance as discussed in Sec. 2.7.
4 Summary and conclusions
We have presented a design for a high-granularity calorimeter insert (HG-CALI) for the forward hadronic calorimeter of the EIC detector. Its sampling design is based on the CALICE “SiPM-on-tile” concept and uses scintillator megatiles that are grooved to define hexagonal cells. Each layer is distinctly shaped to adapt to the conical beampipe shape that accommodates a beam-crossing angle of 25 mrad.
The HG-CALI represents the first concrete proposal to instrument the $3<\eta^{*}<4$ range, which is challenging due to the EIC crossing angle. It offers an optimal acceptance coverage that is limited only by the clearance from the beampipe. It thus represents a way to maximize acceptance up to nearly $\eta^{*}=4.0$, which is a key requirement set in the EIC Yellow Report (1).
The HG-CALI is designed to integrate well with the rest of the endcap calorimeters and will enable easy access to the scintillator layers and SiPMs for maintenance and upgrades. Such a feature will enable frequent annealing of the SiPMs, which will keep radiation damage within manageable levels.
The energy resolution of the HG-CALI meets the EIC Yellow Report requirements even with a basic reconstruction algorithm. Moreover, this device will provide 5D shower information (energy, time, and 3D position) that can be exploited with machine-learning techniques.
The HG-CALI will complement the forward HCal in the region $3.0<\eta^{*}<3.4$ or so by providing an accurate estimate of the shower start position and incident angle over some range in azimuth. Such information could be used with machine-learning approaches to account for shower fluctuations and thus improve the energy resolution of the combined system.
This design will be complemented by a high-granularity electromagnetic calorimeter insert based on tungsten-powder/scintillating fiber with SiPM readout. The complete insert system will be compensated, yielding the prospect of precise and accurate measurements.
The HG-CALI has the potential to unleash the promise of imaging calorimetry at the EIC, which could support a rich physics program centered on forward jets and their substructure in polarized electron-proton and electron-nucleus collisions.
Model availability
The Google Sketchup model, as well as the CAD file in .step format, for the proposed detector can be found in Ref. (74).
The DD4HEP model can be found in Ref. (75).
Acknowledgements
We thank members of the EIC Detector-1 proto-collaboration for useful discussions about the insert, as well as help from their Software Team on matters related to the DD4Hep framework. We thank Oskar Hartbrich for pointing out the importance of the timing cuts and the choice of resolution metric to reproduce CALICE results. We also thank Gerard Visser for feedback on the readout scheme. We also thank members of the California EIC consortium for their feedback on our design. This work was supported by MRPI program of the University of California Office of the President, award number 00010100. M.A, A.P and S.P acknowledge support from DOE grant award number DE-SC0022324. M.A, K.B, and B.K acknowledge support from DOE grant award number DE-SC0022355.
Bibliography
(1)
R. Abdul Khalek, et al., Science Requirements and Detector Concepts for the
Electron-Ion Collider: EIC Yellow Report (3 2021).
arXiv:2103.05419.
(2)
A. Accardi, et al., Electron Ion Collider: The Next QCD Frontier:
Understanding the glue that binds us all, Eur. Phys. J. A 52 (9) (2016)
268.
arXiv:1212.1701,
doi:10.1140/epja/i2016-16268-9.
(3)
ATHENA Collaboration, ATHENA
Detector Proposal - A Totally Hermetic Electron Nucleus Apparatus proposed
for IP6 at the Electron-Ion Collider (Dec. 2021).
doi:10.5281/zenodo.6539707.
URL https://doi.org/10.5281/zenodo.6539707
(4)
ECCE consortium, EIC Comprehensive Chromodynamics Experiment Collaboration
Detector Proposal (May 2022).
doi:10.5281/zenodo.6537588.
(5)
J. Arrington, et al., EIC Physics from An All-Silicon Tracking Detector (2
2021).
arXiv:2102.08337.
(6)
R. Abdul Khalek, et al., Snowmass 2021 White Paper: Electron Ion Collider for
High Energy Physics, in: 2022 Snowmass Summer Study, 2022.
arXiv:2203.13199.
(7)
F. Sefkow, A. White, K. Kawagoe, R. Pöschl, J. Repond, Experimental Tests of
Particle Flow Calorimetry, Rev. Mod. Phys. 88 (2016) 015003.
arXiv:1507.05893,
doi:10.1103/RevModPhys.88.015003.
(8)
M. A. Thomson, Particle Flow Calorimetry and the PandoraPFA Algorithm, Nucl.
Instrum. Meth. A 611 (2009) 25–40.
arXiv:0907.3577,
doi:10.1016/j.nima.2009.09.009.
(9)
The Phase-2 Upgrade of the CMS Endcap Calorimeter, Tech. rep., CERN, Geneva
(Nov 2017).
doi:10.17181/CERN.IV8M.1JY2.
(10)
H. Abramowicz, et al., International Large Detector: Interim Design Report (3
2020).
arXiv:2003.01116.
(11)
A. C. Abusleme Hoffman, et al., Detector Technologies for CLIC 1/2019 (5
2019).
arXiv:1905.02520,
doi:10.23731/CYRM-2019-001.
(12)
M. Dong, et al., CEPC Conceptual Design Report: Volume 2 - Physics &
Detector (11 2018).
arXiv:1811.10545.
(13)
L. Li, Q. Shan, W. Jia, B. Yu, Y. Liu, Y. Shi, T. Hu, J. Jiang, Optimization
of the CEPC-AHCAL scintillator detector cells, JINST 16 (03) (2021) P03001.
doi:10.1088/1748-0221/16/03/P03001.
(14)
Y. Duan, et al., Scintillator tile batch test of CEPC AHCAL, JINST 17 (05)
(2022) P05006.
arXiv:2111.03660,
doi:10.1088/1748-0221/17/05/P05006.
(15)
F. Simon, Silicon photomultipliers in particle and nuclear physics, Nucl.
Instrum. Meth. A 926 (2019) 85–100.
arXiv:1811.03877,
doi:10.1016/j.nima.2018.11.042.
(16)
G. Blazey, et al., Directly Coupled Tiles as Elements of a Scintillator
Calorimeter with MPPC Readout, Nucl. Instrum. Meth. A 605 (2009) 277–281.
doi:10.1016/j.nima.2009.03.253.
(17)
F. Simon, C. Soldner, Uniformity Studies of Scintillator Tiles directly
coupled to SiPMs for Imaging Calorimetry, Nucl. Instrum. Meth. A 620 (2010)
196–201.
arXiv:1001.4665,
doi:10.1016/j.nima.2010.03.142.
(18)
V. Andreev, et al., A high granularity scintillator hadronic-calorimeter with
SiPM readout for a linear collider detector, Nucl. Instrum. Meth. A 540
(2005) 368–380.
doi:10.1016/j.nima.2004.12.002.
(19)
V. Andreev, et al., A high-granularity plastic scintillator tile hadronic
calorimeter with APD readout for a linear collider detector, Nucl. Instrum.
Meth. A 564 (2006) 144–154.
doi:10.1016/j.nima.2006.04.044.
(20)
C. Adloff, et al., Construction and Commissioning of the CALICE Analog Hadron
Calorimeter Prototype, JINST 5 (2010) P05004.
arXiv:1003.2662,
doi:10.1088/1748-0221/5/05/P05004.
(21)
F. Simon, C. Soldner, Uniformity studies of scintillator tiles directly coupled
to SiPMs for imaging calorimetry, Nuclear Instruments and Methods in
Physics Research Section A: Accelerators, Spectrometers, Detectors and
Associated Equipment 620 (2-3) (2010) 196–201.
doi:10.1016/j.nima.2010.03.142.
(22)
M. Feickert, B. Nachman, A Living Review of Machine Learning for Particle
Physics (2 2021).
arXiv:2102.02770.
(23)
M. Paganini, L. de Oliveira, B. Nachman, CaloGAN : Simulating 3D high energy
particle showers in multilayer electromagnetic calorimeters with generative
adversarial networks, Phys. Rev. D 97 (1) (2018) 014021.
arXiv:1712.10321,
doi:10.1103/PhysRevD.97.014021.
(24)
M. Paganini, L. de Oliveira, B. Nachman, Accelerating Science with Generative
Adversarial Networks: An Application to 3D Particle Showers in Multilayer
Calorimeters, Phys. Rev. Lett. 120 (4) (2018) 042003.
arXiv:1705.02355,
doi:10.1103/PhysRevLett.120.042003.
(25)
D. Belayneh, et al., Calorimetry with deep learning: particle simulation and
reconstruction for collider physics, Eur. Phys. J. C 80 (7) (2020) 688.
arXiv:1912.06794,
doi:10.1140/epjc/s10052-020-8251-9.
(26)
S. R. Qasim, J. Kieseler, Y. Iiyama, M. Pierini, Learning representations of
irregular particle-detector geometry with distance-weighted graph networks,
Eur. Phys. J. C 79 (7) (2019) 608.
arXiv:1902.07987,
doi:10.1140/epjc/s10052-019-7113-9.
(27)
F. A. Di Bello, S. Ganguly, E. Gross, M. Kado, M. Pitt, L. Santi, J. Shlomi,
Towards a Computer Vision Particle Flow, Eur. Phys. J. C 81 (2) (2021) 107.
arXiv:2003.08863,
doi:10.1140/epjc/s10052-021-08897-0.
(28)
E. Buhmann, S. Diefenbacher, E. Eren, F. Gaede, G. Kasieczka, A. Korol,
K. Krüger, Getting High: High Fidelity Simulation of High Granularity
Calorimeters with High Speed, Comput. Softw. Big Sci. 5 (1) (2021) 13.
arXiv:2005.05334,
doi:10.1007/s41781-021-00056-0.
(29)
N. Akchurin, C. Cowden, J. Damgov, A. Hussain, S. Kunori, On the use of neural
networks for energy reconstruction in high-granularity calorimeters, JINST
16 (12) (2021) P12036.
arXiv:2107.10207,
doi:10.1088/1748-0221/16/12/P12036.
(30)
J. Pata, J. Duarte, J.-R. Vlimant, M. Pierini, M. Spiropulu, MLPF: Efficient
machine-learned particle-flow reconstruction using graph neural networks,
Eur. Phys. J. C 81 (5) (2021) 381.
arXiv:2101.08578,
doi:10.1140/epjc/s10052-021-09158-w.
(31)
C. Neubüser, J. Kieseler, P. Lujan, Optimising longitudinal and lateral
calorimeter granularity for software compensation in hadronic showers using
deep neural networks, Eur. Phys. J. C 82 (1) (2022) 92.
arXiv:2101.08150,
doi:10.1140/epjc/s10052-022-10031-7.
(32)
N. Akchurin, C. Cowden, J. Damgov, A. Hussain, S. Kunori, Perspectives on the
Calibration of CNN Energy Reconstruction in Highly Granular Calorimeters (8
2021).
arXiv:2108.10963.
(33)
E. Buhmann, S. Diefenbacher, D. Hundhausen, G. Kasieczka, W. Korcari, E. Eren,
F. Gaede, K. Krüger, P. McKeown, L. Rustige, Hadrons, better, faster,
stronger, Mach. Learn. Sci. Tech. 3 (2) (2022) 025014.
arXiv:2112.09709,
doi:10.1088/2632-2153/ac7848.
(34)
G. R. Khattak, S. Vallecorsa, F. Carminati, G. M. Khan, Fast simulation of a
high granularity calorimeter by generative adversarial networks, Eur. Phys.
J. C 82 (4) (2022) 386.
arXiv:2109.07388,
doi:10.1140/epjc/s10052-022-10258-4.
(35)
M. Chadeeva, S. Korpachev, Machine-learning-based prediction of parameters of
secondaries in hadronic showers using calorimetric observables (5 2022).
arXiv:2205.12534.
(36)
S. R. Qasim, N. Chernyavskaya, J. Kieseler, K. Long, O. Viazlo, M. Pierini,
R. Nawaz, End-to-end multi-particle reconstruction in high occupancy imaging
calorimeters with graph neural networks (4 2022).
arXiv:2204.01681.
(37)
V. Mikuni, B. Nachman, Score-based Generative Models for Calorimeter Shower
Simulation (6 2022).
arXiv:2206.11898.
(38)
J. Repond, Detector Concepts of the Electron-Ion Collider, PoS High-pT2019
(2020) 015.
doi:10.22323/1.355.0015.
(39)
CORE Proto-Collaboration, CORE
- a COmpact detectoR for the EIC (Dec. 2021).
doi:10.5281/zenodo.6536630.
URL https://doi.org/10.5281/zenodo.6536630
(40)
F. Bock, et al., Design and Simulated Performance of Calorimetry Systems for
the ECCE Detector at the Electron Ion Collider (7 2022).
arXiv:2207.09437.
(41)
U. Behrens, et al., Test of the ZEUS Forward Calorimeter Prototype, Nucl.
Instrum. Meth. A 289 (1990) 115–138.
doi:10.1016/0168-9002(90)90253-3.
(42)
B. Andrieu, et al., The H1 liquid argon calorimeter system, Nucl. Instrum.
Meth. A 336 (1993) 460–498.
doi:10.1016/0168-9002(93)91257-N.
(43)
C. Adloff, et al., Electromagnetic response of a highly granular hadronic
calorimeter, JINST 6 (2011) P04003.
arXiv:1012.4343,
doi:10.1088/1748-0221/6/04/P04003.
(44)
C. Adloff, J. Blaha, J. J. Blaising, C. Drancourt, A. Espargiliere, R. Galione,
N. Geffroy, Y. Karyotakis, J. Prast, G. Vouters, Tests of a particle flow
algorithm with CALICE test beam data, JINST 6 (2011) P07005.
arXiv:1105.3417,
doi:10.1088/1748-0221/6/07/P07005.
(45)
C. Adloff, et al., Hadronic energy resolution of a highly granular
scintillator-steel hadron calorimeter using software compensation
techniques, JINST 7 (2012) P09017.
arXiv:1207.4210,
doi:10.1088/1748-0221/7/09/P09017.
(46)
F. Simon, C. Soldner, L. Weuste, T3B — an experiment to measure
the time structure of hadronic showers, JINST 8 (2013) P12001.
arXiv:1309.6143,
doi:10.1088/1748-0221/8/12/P12001.
(47)
C. Adloff, et al., Track segments in hadronic showers in a highly granular
scintillator-steel hadron calorimeter, JINST 8 (2013) P09001.
arXiv:1305.7027,
doi:10.1088/1748-0221/8/09/P09001.
(48)
C. Adloff, et al., Validation of GEANT4 Monte Carlo Models with a Highly
Granular Scintillator-Steel Hadron Calorimeter, JINST 8 (2013) 07005.
arXiv:1306.3037,
doi:10.1088/1748-0221/8/07/P07005.
(49)
C. Adloff, et al., The Time Structure of Hadronic Showers in highly granular
Calorimeters with Tungsten and Steel Absorbers, JINST 9 (2014) P07022.
arXiv:1404.6454,
doi:10.1088/1748-0221/9/07/P07022.
(50)
B. Bilki, et al., Pion and proton showers in the CALICE scintillator-steel
analogue hadron calorimeter, JINST 10 (04) (2015) P04014.
arXiv:1412.2653,
doi:10.1088/1748-0221/10/04/P04014.
(51)
M. Chefdeville, et al., Shower development of particles with momenta from 15
GeV to 150 GeV in the CALICE scintillator-tungsten hadronic calorimeter,
JINST 10 (12) (2015) P12006.
arXiv:1509.00617,
doi:10.1088/1748-0221/10/12/P12006.
(52)
G. Eigen, et al., Hadron shower decomposition in the highly granular CALICE
analogue hadron calorimeter, JINST 11 (06) (2016) P06013.
arXiv:1602.08578,
doi:10.1088/1748-0221/11/06/P06013.
(53)
J. Repond, et al., Hadronic Energy Resolution of a Combined High Granularity
Scintillator Calorimeter System, JINST 13 (12) (2018) P12022.
arXiv:1809.03909,
doi:10.1088/1748-0221/13/12/P12022.
(54)
F. Sefkow, F. Simon, A highly granular SiPM-on-tile calorimeter prototype, J.
Phys. Conf. Ser. 1162 (1) (2019) 012012.
arXiv:1808.09281,
doi:10.1088/1742-6596/1162/1/012012.
(55)
O. D. Tsai, et al., Results of R&D on a new construction technique for
W/ScFi calorimeters, J. Phys. Conf. Ser. 404 (2012) 012023.
doi:10.1088/1742-6596/404/1/012023.
(56)
O. D. Tsai, et al., Development of a forward calorimeter system for the STAR
experiment, J. Phys. Conf. Ser. 587 (1) (2015) 012053.
doi:10.1088/1742-6596/587/1/012053.
(57)
E.-C. Aschenauer, et al., The RHIC Cold QCD Plan for 2017 to 2023: A Portal to
the EIC (2 2016).
arXiv:1602.03922.
(58)
EIC Detector Managerie: Step file for the detector chamber,
https://physdiv.jlab.org/EIC/Menagerie/CAD/Beam%20Pipe/Detector%20chamber%20210222.stp.
(59)
EICUG, Maximizing the scientific
output of the eic (Apr. 2022).
doi:10.5281/zenodo.6422182.
URL https://doi.org/10.5281/zenodo.6422182
(60)
D. P. Anderle, et al., Electron-ion collider in China, Front. Phys. (Beijing)
16 (6) (2021) 64701.
arXiv:2102.09222,
doi:10.1007/s11467-021-1062-0.
(61)
M. Beddo, et al., The STAR barrel electromagnetic calorimeter, Nucl. Instrum.
Meth. A 499 (2003) 725–739.
doi:10.1016/S0168-9002(02)01970-8.
(62)
The CMS hadron calorimeter project: Technical Design Report, Technical design
report. CMS, CERN, Geneva, 1997.
(63)
A. Belloni, et al., Test beam study of SiPM-on-tile configurations, JINST
16 (07) (2021) P07022.
arXiv:2102.08499,
doi:10.1088/1748-0221/16/07/P07022.
(64)
L. M. S. de Silva, F. Simon, Effects of misalignment on response uniformity of
SiPM-on-tile technology for highly granular calorimeters, JINST 15 (06)
(2020) P06030.
arXiv:2004.05066,
doi:10.1088/1748-0221/15/06/P06030.
(65)
J. Jiang, S. Zhao, Y. Niu, Y. Shi, Y. Liu, D. Han, T. Hu, B. Yu, Study of SiPM
for CEPC-AHCAL, Nucl. Instrum. Meth. A 980 (2020) 164481.
doi:10.1016/j.nima.2020.164481.
(66)
E. Garutti, Y. Musienko, Radiation damage of SiPMs, Nucl. Instrum. Meth. A
926 (2019) 69–84.
arXiv:1809.06361,
doi:10.1016/j.nima.2018.10.191.
(67)
J. Adam, et al., Accelerator and
beam conditions critical for physics and detector simulations for the
Electron-Ion Collider (2021).
doi:10.5281/zenodo.6514605.
URL https://doi.org/10.5281/zenodo.6514605
(68)
B. Schmookler, et al., High $Q^{2}$
electron-proton elastic scattering at the future Electron-Ion Collider
(2022).
doi:10.48550/ARXIV.2207.04378.
URL https://arxiv.org/abs/2207.04378
(69)
M. Frank, F. Gaede, C. Grefe, P. Mato, DD4hep: A Detector Description Toolkit
for High Energy Physics Experiments, J. Phys. Conf. Ser. 513 (2014) 022010.
doi:10.1088/1742-6596/513/2/022010.
(70)
S. Agostinelli, et al., GEANT4–a simulation toolkit, Nucl. Instrum. Meth. A
506 (2003) 250–303.
doi:10.1016/S0168-9002(03)01368-8.
(71)
J. B. Birks, Scintillations from Organic Crystals: Specific Fluorescence and
Relative Response to Different Radiations, Proc. Phys. Soc. A 64 (1951)
874–877.
doi:10.1088/0370-1298/64/10/303.
(72)
C. Graf, F. Simon, Time-assisted energy reconstruction in a highly-granular
hadronic calorimeter (3 2022).
arXiv:2203.01317.
(73)
S. R. Qasim, K. Long, J. Kieseler, M. Pierini, R. Nawaz, Multi-particle
reconstruction in the High Granularity Calorimeter using object condensation
and graph neural networks, EPJ Web Conf. 251 (2021) 03072.
arXiv:2106.01832,
doi:10.1051/epjconf/202125103072.
(74)
S. J. Paul, M. Arratia, O. Tsai,
Model for a Calorimeter Insert
for the Electron-Ion Collider (7 2022).
doi:10.5281/zenodo.6950495.
URL https://doi.org/10.5281/zenodo.6950495
(75)
R. Milton, B. Schmookler, Z. Ji, B. Karki, S. Paul, M. Arratia,
EIC pEndcap Insert in DD4hep
(7 2022).
doi:10.5281/zenodo.6836014.
URL https://doi.org/10.5281/zenodo.6836014 |
Application of belief functions to medical image segmentation: A review
Ling Huang
Su Ruan
Heudiasyc, CNRS, Université de technologie de Compiègne, France
Quantif, LITIS, University of Rouen Normandy, France
Abstract
The investigation of uncertainty is of major importance in risk-critical applications, such as medical image segmentation. Belief function theory, a formal framework for uncertainty analysis and multiple evidence fusion, has made significant contributions to medical image segmentation, especially since the development of deep learning. In this paper, we provide an introduction to the topic of medical image segmentation methods using belief function theory. We classify the methods according to the fusion step and explain how information with uncertainty or imprecision is modeled and fused with belief function theory. In addition, we discuss the challenges and limitations of present belief function-based medical image segmentation and propose orientations for future research. Future research could investigate both belief function theory and deep learning to achieve more promising and reliable segmentation results.
keywords:
Belief function theory, Medical image segmentation, Uncertainty quantification, Information fusion, Evidence fusion, Reliability, Deep learning
1 Introduction
The precise delineation of the target lesion on the medical image is essential for optimizing disease treatment. In clinical routine, this segmentation is performed manually by physicians and contains shortcomings. First, the pixel or voxel-level segmentation process is time-consuming, especially with 3D images. Second, the segmentation performances are limited by the quality of medical images, the difficulty of the disease, and the domain knowledge of the experts. Thus, physicians always diagnose a disease by summarising multiple sources of information. The advances in medical imaging machines and technology now allow us to obtain medical images in several modalities, such as Magnetic Resonance Imaging (MRI)/Positron Emission Tomography(PET), multi-sequence MRI, or PET/ Computed Tomography (CT). The different modalities provide different information about cancer and other abnormalities in the human body. Combining these modalities makes it possible to segment the organ and lesion region in a reliable way.
Former medical image segmentation approaches batenburg2009adaptive ; kimmel2003fast ; onoma2014segmentation ; salvador2004determining ; suykens1999least ; StroblBias ; kleinbaum1996logistic focus on low-level feature analysis, e.g., gray and textual features, or hand-crafted features. Those approaches have limitations in segmentation accuracy, efficiency, and reliability, which make it a great gap between experimental performance and clinical application. More recently, the success of deep learning in the medical domain brought a lot of contributions to medical image segmentation tasks ronnebergerconvolutional ; myronenko20183d ; isensee2018nnu ; bahdanau2014neural ; vaswani2017attention ; carion2020end ; han2021transformer . It first solves the segmentation efficiency problem that enables large-scale medical image segmentation and then contributes greatly to its accuracy.
Despite the excellent performance of deep learning-based medical image segmentation methods, the doubt about the reliability of the segmentation results remains hullermeier2021aleatoric , which has been the obstacle between experimental performance and the application of therapeutic decision-making for complex oncological cases. A reliable segmentation model should be well calibrated with accuracy and confidence to be an identity function $y=x$. Figure 1 shows an example of an over-confidence segmentation model that outputs a result with high segmentation accuracy and low confidence. Therefore, a trustworthy representation of uncertainty is desirable and should be considered a key feature of any deep learning method, especially in safety-critical application domains, e.g., medical image segmentation. In general, deep models have two sources of uncertainty: aleatoric uncertainty and epistemic uncertainty hora1996aleatory ; der2009aleatory . Aleatoric uncertainty refers to the notion of randomness, i.e., the variability in an experiment’s outcome due to inherently random effects. In contrast, epistemic uncertainty refers to uncertainty caused by a lack of knowledge (ignorance) about the best model, i.e., the ignorance of the learning algorithm or decision-maker. As opposed to uncertainty caused by randomness, uncertainty caused by ignorance can be reduced based on additional information or the design of a suitable learning algorithm.
The medical image segmentation, the uncertainty can be summarized into three levels lakshminarayanan2017simple : pixel/voxel-level uncertainty, instance-level uncertainty, and subject-level uncertainty. Pixel/voxel-level uncertainty is useful for interaction with physicians by providing additional guidance for correction segmentation results. Instance-level is the uncertainty aggregate by pixel/voxel-level uncertainty, which can be used to reduce the false discovery rate. Subject-level uncertainty offers information on whether the segmentation model is reliable. Early approaches to quantify the segmentation uncertainty quantification were based on Bayesian theory hinton1993keeping ; mackay1992practical .
The popularity of deep segmentation models has revived the research on model uncertainty estimation and has given rise to specific methods such as variational dropout gal2016dropout ; tran2018bayesian , and model ensembles lakshminarayanan2017simple ; rupprecht2017learning . While the probabilistic segmentation models capture knowledge in terms of a single probability distribution and can not distinguish between aleatoric and epistemic uncertainty, which limits the further study of uncertainty and reliability. In this paper, we focus on a different uncertainty quantification approach based on belief function theory (BFT) dempster1967upper ; shafer1976mathematical ; denoeux20b with the specific focus on medical image segmentation. It can model the epistemic uncertainty directly and enable researchers further explore the possibility of improving the model’s reliability based on the existing uncertainty. Researchers from the medical image segmentation community have been actively involved in the research of BFT in handling uncertain information modeling, and fusion and have been shown to be very promising for the segmentation task for imperfect medical images lian2018joint ; ghesu2021quantifying ; huang2021belief .
The rest of the paper will provide an overview of the BFT-based medical image segmentation methods. Recent review work for medical image segmentation focus on deep learning-based methods razzak2018deep ; hesamian2019deep ; liu2021review , with a specific focus on deep feature extraction framework siddique2021u , multimodal information fusion zhou2019review , embracing imperfect dataset ramesh2021review , etc. This paper will focus on discussing:
(1)
how imperfect medical images can be modeled by assigning them the degree of belief and uncertainty directly,
(2)
how multiple sources of evidence (with conflict) can be fused by BFT,
(3)
how the above advantages can be merged with the popular deep learning methods to get an accurate and reliable deep segmentation model.
To the best of our knowledge, this is the first review that summarizes the existing medical image segmentation methods with uncertainty quantification and multiple evidence fusion. We hope this review will raise the community’s awareness of the existing solutions for imperfect medical image data segmentation and further inspire researchers to explore the possibility of taking the benefit of both BFT and deep learning to make automatic segmentation methods reliable and interpretable.
We organize this paper as follows: Section 2 summarizes the development of existing medical image segmentation methods and their limitations. Section 3 introduces the fundamentals of BFT. Section 4 introduces the BFT-based methods to model uncertainty, with Sections 4.1 and 4.2 present supervised and unsupervised methods to model uncertainty, respectively. Section 5 gives an overview of BFT-based medical image segmentation methods with Sections 5.1 and 5.2 present the BFT-based medical image segmentation with, respectively, single and multiple classifiers or clustering algorithms. Section 6 concludes this review and give some potential research directions.
Search criterion. To identify related contributions, we mainly retrieved papers containing “medical image segmentation” and “belief function theory” or “Dempster-Shafer theory” or “evidence theory” in the title or abstract from IEEE, Springer, PubMed, Google Scholar, and ScienceDirect. Additionally, conference proceedings for NIPS, CVPR, ECCV, ICCV, MICCAI, and ISBI were searched based on the titles of papers. Papers that do not primarily focus on medical image segmentation problems were excluded.
2 Medical image segmentation
Medical image segmentation involves the extraction of regions of interest (ROI) from 2D/3D image data (e.g., pathological or optical imaging with color mages, MRI, PET, and CT scans. The main goal of segmenting medical images is to identify areas of cancer and other abnormalities in the human body, for example, brain tissue and tumor segmentation, the interior of the human body (such as lung, spinal canal, and vertebrae), skin and cell lesion segmentation with optical imaging with color. Previously, segmenting medical images was time-consuming for physicians, while recent advances in machine learning techniques, especially deep learning, make it easier to perform routine tasks.
2.1 Traditional approaches to medical images segmentation
Early image segmentation methods used the information provided by the image itself, e.g., gray, textual, contrast and histogram features, and segmenting ROI based threshold batenburg2009adaptive , edge detection kimmel2003fast , graph partitioning onoma2014segmentation , and clustering salvador2004determining . More recently, researchers are interested in hard-crafted features, e.g., Scale Invariant Feature Transform (SIFT) lowe1999object , Features from Accelerated Segment Test (FAST) rosten2006machine and Geometric hashing mian2006three , and segmenting ROI using machine learning-based methods such as support vector machine (SVM) suykens1999least , random forest (RF) StroblBias and logistic regression (LR) kleinbaum1996logistic . Those methods have attracted great interest for a while, but their accuracy cannot meet clinical application requirements.
2.2 Deep learning-based methods
Long et al. long2015fully were the first authors to show that a fully convolutional network (FCN) could be trained end-to-end for semantic segmentation, exceeding the state-of-the-art when the paper was published in 2015. UNet ronnebergerconvolutional , a successful modification and extension of FCN, has become the most popular model for medical image segmentation in recent years. Based on UNet, research for deep learning-based medical image segmentation can be summarized in two major directions: feature representation and model optimization.
To better represent input information, well-designed DNNs with encoder-decoder architecture (e.g., 3D UNetcciccek20163d , Res-UNet xiao2018weighted , Dense-UNet guan2019fully , MultiResUNet ibtehaz2020multiresunet , Deep3D-UNet zhu2018anatomynet , V-Netmilletari2016v , nnUNetisensee2018nnu ) were proposed and have achieved good performance. Recently, the popular ideas e.g., residual connection he2016deep , attention mechanism bahdanau2014neural and transformer mechanism vaswani2017attention ; carion2020end ; han2021transformer , have achieved promising performance with DNNs, which also lead to the contributions in medical image domain, e.g., Residual-UNet ronnebergerconvolutional ; xiao2018weighted , Attention UNet oktay2018attention ; trebing2021smaat and Transformer-UNet cao2021swin ; hatamizadeh2022unetr ; hatamizadeh2022swin .
As for the optimization of the DNN-based medical image segmentation models, researchers in the medical image segmentation domain tend to use Dice Loss or a combination of dice loss and cross entropy as a total loss milletari2016v instead of the classical loss functions (e.g., Cross-Entropy Loss, Mean Square Loss) to handle the unbalanced label distribution problem, with more focuses on mining the foreground area. Based on Dice loss, some variants such as GeneralizedDiceLoss sudre2017generalised , TverskyLoss salehi2017tversky and ContrastiveLoss chen2020simple , have been proposed to further solve the unbalanced label problem.
Dice score, Sensitivity, and Precision are the most commonly used evaluation criteria to assess the quality of deep medical image segmentation methods, which are defined as follows:
$$\textsf{Dice}(P,T)=\frac{2\times TP}{FP+2\times TP+FN},$$
(1)
$$\textsf{Sensitivity}(P,T)=\frac{TP}{TP+FN},$$
(2)
$$\textsf{Precision}(P,T)=\frac{TP}{TP+FP},$$
(3)
where $TP$, $FP$, and $FN$ denote the numbers of true positive, false positive, and false negative voxels, respectively. (See Figure 2). The Dice score is a global measure of segmentation performance. Sensitivity is the proportion, among actual tumor voxels, of voxels correctly predicted as tumors. Precision is the proportion, among predicted tumor voxels, of voxels that actually belong to the tumor region; it is, thus, an estimate of the probability that the model is correct when it predicts that a voxel is in a lymphoma region. These two criteria, thus, have to be considered jointly. A more comprehensive introduction to the evaluation criteria for medical image segmentation can be found in taha2015metrics .
2.3 The limitations
The above research focuses on improving the accuracy of segmentation performance under the assumption of adequate and perfect input information and accurate and appropriate prior knowledge. While in reality, especially in the medical image segmentation domain, both the input information and prior knowledge are imperfect and contain a degree of uncertainty. Figure 3a shows the example of a information with uncertainty. Taking a brain tumor segmentation task as an example. Let $X$ be the type of tumor of a voxel, and $\Omega=\{ED,ET,NCR,Others\}$, corresponding to the possibilities: edema, enhancing tumor, necrotic core, and others. Assuming that a specialist provides the information $X\in\{ED,ET\}$, while there is a probability $p=0.1$ that the information is unreliable. How to represent this situation by a probability function becomes a challenging problem. Another situation is if we have multiple information sources containing uncertainty, i.e., Figure 3b; how can we model that kind of uncertainty and fusion the evidence based on that? Furthermore, if the multiple information sources are in conflict and contain uncertainty as well, i.e., Figure 3c, it is difficult to represent and summarize that information by probabilistic models. Thanks to BFT, these challenges can be addressed by designing new frameworks for modeling, reasoning, and fusing imperfect (uncertain, imprecise) information. Next section we will give a brief introduction to BFT.
3 Fundamentals of belief function theory
BFT is a generalization of Bayesian theory, but it is more flexible than the Bayesian approach and suitable for weaker conditions sun2018multi , i.e., imperfect (uncertain, imprecise, partial) information. Figure 4 shows the difference between a probabilistic model and a BFT model when applied to a three-class classification task ($\Omega=\{a,b,c\}$). For input $x$, the probabilistic model outputs the probability that $x$ belongs to classes a, b, and c as 0.4, 0.5, and 0.1, respectively. While the BFT model can generate the belief that $x$ belongs to any subset of $\Omega$, e.g., $\{a,b\}$, $\{b,c\}$. Compared with the probabilistic model, the BFT model has more degrees of freedom to represent its uncertainty directly, i.e., it models the uncertainty into the subset of $\Omega$ and the ignorance set $\Omega$.
In the past decades, BFT has generated considerable interests and has had a great success in diverse fields, including uncertain reasoning smets1990combination ; yager1987dempster ; dubois1988representation ; denoeux2008conjunctive ; xiao2020generalization , classification denoeux1995k ; denoeux2000neural ; xiao2022negation and clustering denoeux2004evclus ; masson2009recm , etc. It was first originated by Dempster dempster1967upper in the context of statistic inference in 1968 and was formalized by Shafer shafer1976mathematical as a theory of evidence in 1976. In 1986, Dubois and Parde proposed an approach to the computerized processing of uncertainty dubois2012possibility . In 1978, Yager proposed a new combination rule of the belief function framework yager1987dempster . In 1990, BFT was further popularized and developed by Smets smets1990combination as the ’Transferable Belief Model’ with pignistic transformation for decision making. Since then, booming developments have been made. More detailed information about the development of BFT in 40 years can be found in denzux201640 .
We will first briefly introduce the basic notions of BFT in Section 3.1, which includes evidence representation (mass functions, belief, and plausibility functions). Second, we introduce Dempster’s combination rule to explain the operations of multiple sources of evidence in Section 3.2. Third, we introduce a discounting operation for unreliable sources in Section 3.3. Fourth, we introduce some commonly used decision-making methods in Section 3.4.
3.1 Representation of evidence
Let $\Omega=\{\omega_{1},\omega_{2},...,\omega_{C}\}$ be a finite set of all possible hypotheses about some problem, called a frame of discernment. Evidence about a variable $\omega$ taking values in $\Omega$ can be represented by a mass function $m$, from the power set $2^{\Omega}$ to $[0,1]$, such that
$$\displaystyle\sum_{A\subseteq\Omega}m(A)=1,$$
(4a)
$$\displaystyle m(\emptyset)=0.$$
(4b)
which can also be called basic belief assignment (BBA). The BBA methods used in medical image segmentation to generate mass functions will be introduced in Section 4 Each subset $A\subseteq\Omega$ and $m(A)$ is called a focal set of $m$. The total ignorance of the problem can be represented as $m(\Omega)$. If all focal sets are singletons, then $m$ is said to be Bayesian. It is equivalent to a probability distribution. The information provided by a mass function $m$ can be represented by a belief function $Bel$ or a plausibility function $Pl$ from $2^{\Omega}$ to $[0,1]$, respectively, as
$$Bel(A)=\sum_{B\subseteq A}m(B)$$
(5)
and
$$Pl(A)=\sum_{B\cap A\neq\emptyset}m(B)=1-Bel(\bar{A}),$$
(6)
for all $A\subseteq\Omega$. The quantity $Bel(A)$ can be interpreted as a degree of support to $A$, while $Pl(A)$ can be interpreted as a measure of lack of support given to the complement of $A$ (or we can say $Pl(A)$ represents the belief that might be committed to $A$).
3.2 Dempster’s rule
In BFT, the belief about a certain item is elaborated by aggregating different belief functions over the same frame of discernment. Given two mass functions $m_{1}$ and $m_{2}$ derived from two independent items of evidence, the final belief that supports $A$ can be obtained by combining $m_{1}$ and $m_{2}$ with Dempster’s rule shafer1976mathematical , which is defined as
$$(m_{1}\oplus m_{2})(A)=\frac{1}{1-\kappa}\sum_{B\cap D=A}m_{1}(B)m_{2}(D),$$
(7)
for all $A\subseteq\Omega,A\neq\emptyset$, and $(m_{1}\oplus m_{2})(\emptyset)=0$. The coefficient $\kappa$ is the degree of conflict between $m_{1}$ and $m_{2}$, it is defined as
$$\kappa=\sum_{B\cap D=\emptyset}m_{1}(B)m_{2}(D).$$
(8)
3.3 Discounting
In (8), if $m_{1}$ and $m_{2}$ are logically contradictory and, consequently, then we can not use Dempster’s rule to combine the two pieces of evidence. Discounting strategies thus can be used to combine evidence under high contradiction shafer1976mathematical ; mercier2008refined ; denoeux2019new . Let $m$ be a mass function on $\Omega$ and $\xi$ a coefficient in $[0,1]$. The discounting operation shafer1976mathematical with discount rate $\xi$ transforms $m$ into a weaker, less informative mass function defined as follows:
$${}^{\xi}m(A)=(1-\xi)m(A),\quad\forall A\subset\Omega,$$
(9a)
$${}^{\xi}m(\Omega)=(1-\xi)m(\Omega)+\xi,$$
(9b)
coefficient $1-\xi$ represents the degree of belief that the source $m$ is reliable. In the medical domain, the discounting function is widely used in multimodal evidence fusion by considering source information reliability and discounts their outputs using these precalculated reliability factors.
3.4 Decision-making
After combining all the available evidence in the form of a mass function, it is necessary to decide the belief to be assigned into focal sets. In this section, we introduce some classical BFT-based decision-making methods.
Upper and lower expected utilities
Let $u$ be a utility function. The lower and upper expectations of $u$ with respect to $m$ are defined, respectively, as the averages of the minima and the maxima of $u$ within each focal set of $m$:
$$\displaystyle\underline{E}_{m}(u)=\sum_{A\subseteq\Omega}m(A)\min_{\omega\in A}u(\omega),$$
(10a)
$$\displaystyle\overline{E}_{m}(u)=\sum_{A\subseteq\Omega}m(A)\max_{\omega\in A}u(\omega).$$
(10b)
When $m$ is Bayesian, $\underline{E}_{m}(u)=\overline{E}_{m}(u)$. If $m$ is logical with focal set $A$, then $\underline{E}_{m}(u)$ and $\overline{E}_{m}(u)$ are, respectively, the belief and plausibility functions as we mentioned in Section 3.1. Both lower or upper expectations can be chosen for the final decision according to the given task.
Pignistic criterion
In 1990, Smets proposed a pignistic transformation smets1990combination that each mass of belief is distributed equally among the elements for all, which is defined as
$$BetP(\omega)=\sum_{\omega\in A}\frac{m(A)}{\mid A\mid},\quad\forall\omega\in\Omega,$$
(11)
where $\left|A\right|$ denotes the cardinality of $A\subseteq\Omega$.
Besides the above methods, there are various decision-making methods proposed for BFT, such as Generalized OWA criterion yager1992decision , Generalized minimax regretYAGER2004109 , Generalized divergence xiao2022generalized , etc. More details about decision-making with BFT can be found in the review paper denoeux2019decision .
4 BBA methods to generate mass functions
To model segmentation uncertainty, the first step is to generate mass functions. In this section, we introduce the BBA methods applied to medical image segmentation. Figure 5 is an overview of BBA methods. In general, those methods can be separated into supervised and unsupervised methods according to if the annotations are used.
4.1 Supervised BBA methods
The supervised BBA methods can be classified into two categories. One is the likelihood-based methods, such as Shafer’s model shafer1976mathematical and Appriou’s model appriou1999multisensor ; appriou200501 . The other is the distance-based methods, such as evidential KNN denoeux1995k , evidential neural network classifier denoeux2000neural and Radial basis function network denoeux2019decision . It should be noted that the distance-based BBA methods are easy to be merged with popular deep segmentation models and have shown promising results huang2021belief ; huang2022lymphoma .
4.1.1 The likelihood-based BBA methods
Shafer’s model
In shafer1976mathematical , Shafer proposed a likelihood-based evidential model to calculate mass functions. Assuming that conditional density functions $f(x\mid\omega_{c})$ are known, then the conditional likelihood associated with the pattern $X$ is defined by $\ell(\omega_{c}\mid x)=f(x\mid\omega_{c})$. The mass functions are defined according to the knowledge of all hypotheses $\omega_{1},...,\omega_{C}$. Firstly, the plausibility of a simple hypothesis $\omega_{c}$ is proportional to its likelihood. The plausibility is thus given by
$$Pl(\{\omega_{c}\})=\hbar\cdot\ell(\omega_{c}\mid x),\quad\forall\omega_{c}\in\Omega,$$
(12)
where $\hbar$ is a normalization factor with $\hbar=1\max_{\omega\in\Omega}\ell(\omega|x)$. The plausibility of a set $A$ is thus given by
$$Pl(A)=\hbar\cdot\underset{\omega_{c}\in A}{\max}\ell(\omega_{c}\mid x).$$
(13)
Appriou’s model
Same as Shafer’s one, Appriou appriou1999multisensor ; appriou200501 proposed two likelihood-based models to calculate mass functions with the frame of discernment $\Omega=\{\omega_{c},\neg{\omega_{c}}\}$. For the first model, the mass functions are defined by
$$\displaystyle m(\{\omega_{c}\})=0,$$
(14a)
$$\displaystyle m(\{\neg{\omega_{c}}\})=\alpha_{c}({1-\hbar\cdot\ell(\omega_{c}\mid x)}),$$
(14b)
$$\displaystyle m(\Omega)=1-\alpha_{c}{(1-\hbar\cdot\ell(\omega_{c}\mid x))},$$
(14c)
where $\alpha_{c}$ is a reliability factor depending on the hypothesis $\omega_{c}$ and on the source information. The second model is defined as
$$\displaystyle m(\{\omega_{c}\})=\alpha_{c}\cdot\hbar\cdot\ell(\omega_{c}\mid x)/{(1+\hbar\cdot\ell(\omega_{c}\mid x))},$$
(15a)
$$\displaystyle m(\{\neg{\omega_{c}}\})=\alpha_{c}/{(1+\hbar\cdot\ell(\omega_{c}\mid x))},$$
(15b)
$$\displaystyle m(\Omega)=1-\alpha_{c}.$$
(15c)
4.1.2 The distance-based BBA methods
Evidential KNN (EKNN) rule
In denoeux1995k , Denoeux proposed a distance-based KNN classifier for classification tasks. Let $N_{K}(x)$ denote the set of the $K$ nearest neighbors of $x$ in learning set $Z$. Each $x_{i}\in N_{K}(x)$ is considered as a piece of evidence regarding the class label of $x$. The strength of evidence decreases with the distance between $x$ and $x_{i}$. The evidence of $(x_{i},y_{i})$ support class $c$ is represented by
$$\displaystyle m_{i}(\{\omega_{c}\})=\varphi_{c}(d_{i})y_{ic},\quad 1\leq c\leq C,$$
(16a)
$$\displaystyle m_{i}(\Omega)=1-\varphi_{c}(d_{i}),$$
(16b)
where $d_{i}$ is the distance between $x$ and $x_{i}$, which can be Euclidean or other distance function; and $y_{ic}=1$ if $y_{i}=\omega_{c}$ and $y_{ic}=0$ otherwise. The parameter $\varphi_{c}$ is a decreasing function with
$$\varphi_{c}(d)=\alpha\exp(-\gamma d^{2}),$$
(17)
where $\alpha$ and $\gamma$ are two tuning parameters. The evidence of the $K$ nearest neighbors of $x$ is fused by Dempster’s rule:
$$m=\bigoplus_{x_{i}\in N_{K}(x)}m_{i}.$$
(18)
The final decision is made according to maximum plausibility. The detailed optimization of these parameters is described in zouhal1998evidence . Based on the first work, Denoeux et al. proposed contextual discounting evidential KNN denoeux2019new with partially supervised learning to address the annotation limitation problem.
Evidential neural network (ENN)
The success of machine learning encouraged the exploration of applying belief function theory with learning methods. In denoeux2000neural , Denœux proposed an ENN classifier in which mass functions are computed based on distances to prototypes.
The ENN classifier is composed of an input layer of $H$ neurons, two hidden layers, and an output layer. The first input layer is composed of $I$ units, whose weights vectors are prototypes $p_{1},\ldots,p_{I}$ in input space. The activation of unit $i$ in the prototype layer is
$$s_{i}=\alpha_{i}\exp(-\gamma_{i}d_{i}^{2}),$$
(19)
where $d_{i}=\left\|x-p_{i}\right\|$ is the Euclidean distance between input vector $x$ and prototype $p_{i}$, $\gamma_{i}>0$ is a scale parameter, and $\alpha_{i}\in[0,1]$ is an additional parameter. The second hidden layer computes mass functions $m_{i}$ representing the evidence of each prototype $p_{i}$, using the following equations:
$$\displaystyle m_{i}(\{\omega_{c}\})$$
$$\displaystyle=u_{ic}s_{i},\quad c=1,...,C$$
(20a)
$$\displaystyle m_{i}(\Omega)$$
$$\displaystyle=1-s_{i},$$
(20b)
where $u_{ic}$ is the membership degree of prototype $i$ to class $\omega_{c}$, and $\sum_{c=1}^{C}u_{ic}=1$. Finally, the third layer combines the $I$ mass functions $m_{1},\ldots,m_{I}$ using Dempster’s rule. The output mass function $m=\bigoplus_{i=1}^{I}m_{i}$ is a discounted Bayesian mass function that summarizes the evidence of the $I$ prototypes.
Radial basis function (RBF) network
In huang2022lymphoma , Huang et al. confirmed the RBF network can be an alternative approach to ENN that is based on the aggregation of weights of evidence. Consider an RBF network with $I$ prototype (hidden) units. The activation of hidden unit $i$ is
$$s_{i}=\exp(-\gamma_{i}d_{i}^{2}),$$
(21)
where, $d_{i}=\left\|\boldsymbol{x}-\boldsymbol{p}_{i}\right\|$ is the Euclidean distance between input vector $\boldsymbol{x}$ and prototype $\boldsymbol{p}_{i}$, and $\gamma_{i}>0$ is a scale parameter. Here we only show an example of binary classification task $C=2$ and $\Omega=\{\omega_{1},\omega_{2}\}$ (The multi-class example can be found in denoeux19d ). Let $v_{i}$ be the weight of the connection between hidden unit $i$ and the output unit, and let $w_{i}=s_{i}v_{i}$ be the product of the output of unit $i$ and weight $v_{i}$. The quantities $w_{i}$ can be interpreted as weights of evidence for class $\omega_{1}$ or $\omega_{2}$, depending on the sign of $v_{i}$.
To each prototype $i$ can be associated with the following simple mass function:
$$m_{i}=\{\omega_{1}\}^{w_{i}^{+}}\oplus\{\omega_{2}\}^{w_{i}^{-}},$$
where $w_{i}^{+}$ and $w_{i}^{-}$ denote, respectively, the positive and negative parts of $w_{i}$. Combining the evidence of all prototypes in favor of $\omega_{1}$ or $\omega_{2}$ by Dempster’s rule, we get the mass function $m=\bigoplus_{i=1}^{I}m_{i}$ that has the following expression:
$$\displaystyle m(\{\omega_{1}\})$$
$$\displaystyle=\frac{[1-\exp(-w^{+})]\exp(-w^{-})}{1-\kappa},$$
(22a)
$$\displaystyle m(\{\omega_{2}\})$$
$$\displaystyle=\frac{[1-\exp(-w^{-})]\exp(-w^{+})}{1-\kappa},$$
(22b)
$$\displaystyle m(\Omega)$$
$$\displaystyle=\frac{\exp(-w^{+}-w^{-})}{1-\kappa}=\frac{\exp(-\sum_{i=1}^{I}|w_{i}|)}{1-\kappa},$$
(22c)
where $\kappa$ is the degree of conflict between $\{\omega_{1}\}^{w^{+}}$ and $\{\omega_{2}\}^{w^{-}}$ with
$$\kappa=[1-\exp(-w^{+})][1-\exp(-w^{-})].$$
(23)
4.2 Unsupervised BBA methods
For unsupervised BBA methods, the goal is to find an algorithm to generate mass functions without any label information. In earlier BBA studies, Fuzzy C-means (FCM) dunn1973fuzzy was the most popular algorithm used to generate membership values (MVs). Based on MVs, the authors can obtain mass functions according to some domain knowledge, e.g., threshold zhu2002automatic , or user-specific parameters ghasemi2012brain . Sigmoid and the one-side Gaussian function can also be used to generate MVs safranek1990evidence . Credal partition denoeux2004evclus , the extension of fuzzy partition, enables us to generate mass functions directly masson2008ecm . Besides the two popular BBA methods, mass functions can also be generated from the Gaussian distribution of the input to the cluster center chen2012manifold .
4.2.1 MVs-based BBA methods
FCM
Considering that there are some BFT-based methods that use FCM to generate MVs, thus we introduce it here to offer a basic view for readers. With FCM, any $\mathbf{x}$ has a set of coefficients giving the degree of being in the $k$-th cluster $w_{k}(x)$. The centroid of a cluster is the mean of all points, weighted by their degree of belonging to the cluster,
$$c_{k}={{\sum_{x}{w_{k}(x)}^{m}x}\over{\sum_{x}{w_{k}(x)}^{m}}},$$
(24)
where $m$ is the hyper-parameter that controls how fuzzy the cluster will be. The higher it is, the fuzzier. Given a finite set of data, the FCM algorithm returns a list of cluster centres $P=\{c_{1},...,c_{C}\}$ and a partition matrix $W=\omega_{ij},i=1,...N,j=1,...,C$,
$$w_{{ij}}={\frac{1}{\sum_{{k=1}}^{{C}}\left({\frac{\left\|{\mathbf{x}}_{i}-{\mathbf{c}}_{j}\right\|}{\left\|{\mathbf{x}}_{i}-{\mathbf{c}}_{c}\right\|}}\right)^{{{\frac{2}{m-1}}}}}},$$
(25)
where $w_{ij}$, is the the degree of $\mathbf{x}_{i}$ belongs to cluster $\mathbf{c}_{j}$. The objective function is defined as
$$\underset{P}{arg\max}\sum_{i=1}^{N}\sum_{j=1}^{C}w_{ij}^{m}\left\|\mathbf{x}_{i}-\mathbf{c}_{j}\right\|.$$
(26)
Zhu’s model
In zhu2002automatic , Zhu et al. proposed a method to determine mass functions using FCM and neighborhood information. Mass of simple hypothesis $\{\omega_{c}\}$ is directly obtained from the filtered membership functions $\upsilon_{c}(g)$ of the gray level $g(x,y)$ to cluster $c$ as $m(\{\omega_{c}\})=\upsilon_{c}(g)$, indicates that for a given gray level, the piece of evidence of belonging to the cluster $c$ is directly given by its degree of membership to the same cluster. If there is a high ambiguity in assigning a gray level $g(x,y)$ to cluster $c$ or $c+1$, that is, $\left|\upsilon_{c}(g)-\upsilon_{c+1}(g)\right|<\varepsilon$, where $\varepsilon$ is a thresholding value, then a double hypothesis is formed. The value of the threshold $\varepsilon$ is chosen following the applications in question. The authors suggested fixing $\varepsilon$ at 0.1. Once the double hypotheses are formed, their associated mass is calculated according to the following formula:
$$m(\{\omega_{c},\omega_{c+1}\})=\frac{S[\upsilon_{c}(g),\upsilon_{c+1}(g)]}{2S_{max}},$$
(27)
where $S$ represents a surface of a triangle and $S_{max}$ is the maximum of ambiguity. The surface of such a triangle depends
both on the degrees of the membership functions of $g(x,y)$ to clusters $c$ and $c+1$ and on the conflicts between these MVs. Figure 6a shows how the triangle is constructed and how the mass of double hypotheses $\{\omega_{c},\omega_{c+1}\}$ is derived from the surface of the triangle. The vertical axis of Figure 6a represents the MVs. The surface of the two dotted triangles represents two so-called triangular membership functions corresponding to classes $c$ and $c+1$. The two triangles are isosceles and have the same length for their bases. The heights of the triangles are equal to $m(\{\omega_{c}\})$ and $m(\{\omega_{c+1}\})$, respectively. The overlapping surface $S$ of the two triangles represents the MV to the double hypothesis $\{\omega_{c},\omega_{c+1}\}$. Therefore, the mass value attributed to the double hypothesis $\{\omega_{c},\omega_{c+1}\}$ can be directly calculated from the surface $S$. Figure 6b shows the condition of the maximum ambiguity case.
Ratio MV (RMV) transformation
In ghasemi2012brain , Ghasemi et al. proposed a ratio membership value transformation method to calculate mass functions. The FCM algorithm was first used to generate MVs $f_{\omega_{c}}$ for each pixel. Then the MVs are used to build the mass functions. For this purpose, the three ratios of the available MVs are calculated, corresponding to three situations: no-uncertainty (NU), semi-uncertainty (SU), and perfect-uncertainty (PU). First, PU is a critical situation in which the RMVs are smaller than $\alpha$, then the mass function is calculated as $m(\{\omega_{1}\})=m(\{\omega_{2}\})=m(\Omega)=(f_{\omega_{1}}+f_{\omega_{2}})/3$. Second, two thresholds $\alpha$ and $\beta$ with $\alpha=1.5$ and $\beta=3$ are selected to control the boundary between SU and PU, and between NU and SU, separately. For example, with
$$f_{\omega_{1}}=0.18,\\
f_{\omega_{2}}=0.81,\\
RMV=f_{\omega_{1}}/f_{\omega_{2}}=4.5,RMV>\beta,$$
the two MVs fall in the NU category. If
$$f_{\omega_{1}}=0.25,\\
f_{\omega_{2}}=0.65,\\
RMV=f_{\omega_{1}}/f_{\omega_{2}}=2.6,\\
\alpha<RMV<\beta,$$
the two MVs are in the SU category. The mass functions are calculated using
$$\displaystyle m(\{\omega_{1}\})=f_{\omega_{1}}-\frac{\lambda_{\omega_{1},\omega_{2}}}{2},$$
(28a)
$$\displaystyle m(\{\omega_{2}\})=f_{\omega_{2}}-\frac{\lambda_{\omega_{1},\omega_{2}}}{2},$$
(28b)
$$\displaystyle m(\Omega)=\lambda_{\omega_{1},\omega_{2}},$$
(28c)
where $\lambda$ is an uncertainty distance value that defined as $\lambda_{\omega_{1},\omega_{2}}=\frac{\left|f_{\omega_{1}}-f_{\omega_{2}}\right|}{\beta-\alpha}$.
4.2.2 Evidential C-means (ECM)
In denoeux2004evclus , Denoeux et al. proposed an evidential clustering algorithm that extends the notion of fuzzy partition with credal partition, which extends the existing concepts of hard, fuzzy (probabilistic), and possibilistic partition by allocating each object a ’mass of belief,’ not only to single clusters but also to any subsets of $\Omega=\{\omega_{1},...,\omega_{C}\}$.
Credal partition
Assuming there is a collection of five objects for two classes, mass functions for each source are given in Table 1. They represent different situations: the class of source 1 indicates strong evidence that the class of source 1 does not lie in any focal set; the class of source 2 is completely unknown, and the class of source 3 is known with certainty; the cases of sources 4 and 5 correspond to situations of partial knowledge ($m_{5}$ is Bayesian). We can then assign basic beliefs to each partition with the given objective functions with the credal partition.
Evidential C-Means (ECM)
Based on the credal partition, another BBA method, ECM masson2008ecm , was introduced to generate mass functions. In ECM, a cluster is represented by a prototype $p_{c}$. For each non-empty set $A_{j}\subseteq\Omega$, a prototype $\bar{p_{j}}$ is defined as the center of mass of the prototypes $p_{c}$ such that $\omega_{c}\in A_{j}$. Then the non-empty focal set is defined as $F=\{A_{1},...,A_{f}\}\subseteq 2^{\Omega}\setminus\left\{\emptyset\right\}$. Deriving a credal partition from object data implies determining, for each object $x_{i}$, the quantities $m_{i}j=m_{i}(A_{j}),A_{i}\neq\emptyset,A_{j}\subseteq\Omega$. In such a way that $m_{i}j$ is low (resp. high) when the distance between $x_{i}$ and the focal set $\bar{p_{j}}$ is high (resp. low). The distance between an object and any nonempty subset of $\Omega$ has thus to be defined by
$$d_{ij}^{2}=\left\|x_{i}-\bar{p_{j}}\right\|^{2}.$$
(29)
4.2.3 Gaussian distribution (GD)-based model
Besides the FCM-based and credal partition-based BBA method, the mass functions can also be generated from GD chen2012manifold .
The mass of simple hypotheses $\{\omega_{c}\}$ can be obtained from the assumption of GD according to the information $x_{i}$ of a pixel from an input image to cluster $c$ as follows:
$$m(\{\omega_{c}\})=\frac{1}{\sigma_{c}\sqrt{2\pi}}\exp{\frac{-(x_{i}-\mu_{c})^{2}}{2\sigma_{c}^{2}}},$$
(30)
where $\mu_{c}$ and $\sigma^{2}_{c}$ represent, respectively, the mean and the variance of the cluster $c$, which can be estimated by
$$\mu_{c}=\frac{1}{n_{c}}\sum_{i=1}^{n_{c}}x_{i},$$
(31)
$$\sigma_{c}^{2}=\frac{1}{n_{c}}\sum_{i=1}^{n_{c}}(x_{i}-\mu_{c})^{2},$$
(32)
where $n_{c}$ is the number of pixels in the cluster $c$. The mass of double hypotheses involves several sets $\{\omega_{1},\omega_{2},...,\omega_{T}\}$ is determined as
$$m(\{\omega_{1},\omega_{2},...,\omega_{T}\})=\frac{1}{\sigma_{t}\sqrt{2\pi}}\exp{\frac{-(x_{i}-\mu_{t})^{2}}{2\sigma_{t}^{2}}},$$
(33)
where $\mu_{t}=\frac{1}{T}\sum_{i=1}^{T}\mu_{i}$, $\sigma_{t}=\max(\sigma_{1},\sigma_{2},...,\sigma_{T}),2\leq T\leq C$, $C$ is the number of clusters.
4.2.4 Binary frames of Discernment (BFOD)
Under the assumption that the membership value is available, Safranek et al. introduced a BFOD-based BBA method safranek1990evidence to transform membership value into mass functions. The BFOD is constructed as $\Omega=\{\omega,\neg\omega\}$ with a function $cf(\nu)$, $cf(\nu)\in\left[0,1\right]$ that indicates confidence factor. Sigmoid and the one-sided Gaussian function were the most appropriate functions for $cf(\nu)$ reported by the authors. Once a confidence value is obtained, the transformation for mass functions can be accomplished by defining appropriate transfer functions with parameters A and B.
$$\displaystyle m(\{\omega\})=\frac{B}{1-A}cf(\nu)-\frac{AB}{1-A},$$
(34a)
$$\displaystyle m(\{\neg\omega\})=\frac{-B}{1-A}cf(\nu)+B,$$
(34b)
$$\displaystyle m(\Omega)=1-m(\{\omega\})-m(\{\neg\omega\}),$$
(34c)
where $A$ and $B$ are user-specific parameters. The left-hand side stays clamped at zero when the right-hand side goes negative. The parameter $A$ is the confidence-factor axis intercept of the curve that depicts the dependence of $m(\{\omega\})$ on confidence factors, and $B$ is the maximum support value assigned to $m(\{\omega\})$ or $m(\{\neg\omega\})$.
4.3 Discussion
If segmentation labels are available, we suggest using the supervised BBA methods to generate mass functions, especially the distance-based BBA methods, because the distance-based BBA methods can be merged easily with the popular deep learning methods and construct a learnable end-to-end deep segmentation model. The detailed merging example can be found in huang2021evidential , where the authors first apply the idea of merging the ENN or RBF to the deep medical image segmentation model by mapping feature vectors into mass functions instead of using Softmax to map them into probabilities.
If the segmentation labels are not available or only partially available, researchers can still generate mass functions using the unsupervised BBA methods, e.g., using ECM for tumor segmentation lian2018unsupervised . To our best knowledge, there is no research that works on merging unsupervised BBA methods into deep segmentation models even though some studies already make the end-to-end neural network-based evidential clustering model possible, e.g., denoeux2021nn . We believe it is a good choice for researchers who work on unlabeled datasets and wish to directly quantify segmentation uncertainty, they can consider applying unsupervised BBA methods to deep segmentation models.
5 BFT-based medical image segmentation methods
To summarize the BFT-based medical image segmentation methods, we can either classify them by the input modality of the images or by the specific clinical application. Figure 7a shows the proportion of types of medical images applied in the segmentation task and Figure 7b the proportions of application in the medical image segmentation task.
We presented the common BBA methods to generate mass functions for medical images in Section 4, the most interesting way to analyze those methods is to classify them according to the stage at which they fuse mass functions. Figure 8 gives an overview of the BFT-based medical image segmentation methods where the fusion operation is performed.
We first classify them by the number of classifiers/clusterers, i.e., if the fusion of mass functions is performed at the classifiers/clusterers level. Then we classify them by the number of input modalities, i.e., if the fusion of mass functions is performed at the modality level. It should be noted that the medical image segmentation with single modal input and single classifier/clusterer can also have a fusion operation, which is performed at the pixel/voxel level. Figure 9 gives the proportions of the BFT-based segmentation methods if multiple classifiers/clusterers or multimodal medical images are used. 76% of the methods use a single classifier or cluster. Among these methods, 18% take single modal medical images as input, and 58% take multimodal medical images as input. The remaining 24% of the methods use several classifiers or clusterers Among those methods, 21% take single modal medical images as input, and 3% take multimodal medical images as inputs. Sections 5.1 and 5.2 will go for more details for the fusion operations and discuss their performances.
5.1 BFT-based medical image segmentation with a single classifier or clusterer
The medical image segmentation methods we introduced in this section focus on using a single classifier or clusterer. To simplify the introduction, we classify them into single modal and multimodal inputs and introduce them in Sections 5.1.1 and 5.1.2, respectively.
5.1.1 Single modal evidence fusion
Figure 10 shows the framework of single modal evidence fusion with a single classifier (we only take the classifier as an example in the rest of our paper to simplify the explanation). The inputs of the framework are single-modal medical images. The segmentation process comprises three steps: masses calculation (including feature extraction and BBA), feature-level masses fusion, and decision-making. Since decision-making is not the focus of this paper, we will not go into details about it and readers can refer to Ref. denoeux2019decision for a recent review of decision methods based on belief functions.
1)
First, the mass calculation step assigns each pixel/voxel $K$ mass functions based on the evidence of $K$ different BBA methods, or $K$ kinds of input features, or $K$ nearest neighbors, or $K$ prototype centers.
2)
Dempster’s rule is then used for each pixel/voxel to fuse the feature-level mass functions.
3)
Finally, decision-making is done to obtain the final segmentation results.
The feature extraction method here is used to generate MVs (corresponding to traditional medical image segmentation methods) or deep features (corresponding to deep learning-based medical image segmentation methods). The feature extraction method could be intensity-based methods such as threshold intensity, probabilistic-based methods such as SVM, or fuzzy set-based methods such as FCM, etc. The BBA methods are various, therefore we introduce them in specific tasks as follows. In general, the methods introduced in this section focus on feature-level evidence fusion. Table 2 shows the segmentation methods with a single modal input and classifier/cluster that focus on feature-level evidence fusion.
In SuhKnowledge , Suh et al. developed a knowledge-based endocardial and epicardial boundary detection and segmentation system with cardiac MR image sequences. Pixel and location information were mapped into mass functions by Shafer’s model shafer1976mathematical . The two source mass functions were fused by using Dempster’s rule. Experiments were applied to cardiac short-axis images and obtained an excellent success rate ($>90\%$). However, this work only focused on cardiac boundary detection and did not discuss the details of the heart. In 1509843 , Vauclin et al. proposed a BFT-based lung and spinal canal segmentation model. The k-means clustering algorithm was first used to perform a pre-segmentation. Then a 3D filter exploits the results of the pre-segmentation to compute the MVs from spatial neighbors using Shafer’s model. Segmentation results showed the credal partition permits the reduction of the connection risks between the trachea and the lung when they are very close and between the left and right lungs at the anterior or posterior junctions.
In gerig2000exploring , Gerig et al. presented a method for automatic segmentation and characterization of object changes in time series of 3D MR images. A set of MVs was derived from time series according to brightness changes. BFOD transformation safranek1990evidence was used here to map the obtained features into mass functions. Then the set of evidence was combined by using Dempster’s rule. Experiments visually compared with results from alternative segmentation methods revealed an excellent sensitivity and specificity performance in the brain lesion region. The author also pointed out that better performance could be obtained with multimodal and multiple time-series evidence fusion. Simulation experimental results showed that about 80% of the implanted voxels could be detected for most generated lesions.
In vannoorenberghe2006belief , Vannoorenberghe et al. presented a BFT-based thoracic image segmentation method. First, a K-means algorithm performed coarse segmentation on the original CT images. Second, the EKNN rule denoeux1995k was applied by considering spatial information and calculating feature-level mass functions. The authors claimed that using the motioned segmentation scheme leads to a complementary approach combining region segmentation and edge detection. Experimental results showed promising results on 2D and 3D CT images for lung segmentation.
In derraz2013globally , Derraz et al. proposed an active contours (AC)-based chan2000active global segmentation method for vector-valued image that incorporated both probability and mass functions. All features issued from the vector-valued image were integrated with inside/outside descriptors to drive the segmentation results by maximizing the Maximum-Likelihood distance between foreground and background. Appriou’s second model appriou200501 was used to calculate the imprecision caused by low contrast and noise between inside and outside descriptors issued from the multiple channels. Then the fast algorithm based on Split Bregman goldstein2009split was used for final segmentation by forming a fast and accurate minimization algorithm for the Total Variation (TV) problem. Experiments were conducted on color biomedical images (eosinophil, lymphocyte, eosinophil, monocyte, and neutrophil cell mohamed2012enhanced ) and have around 6% improvements by using F-score on five image groups. In the same year, Derraz et al. proposed a new segmentation method derraz2013image based on Active Contours (AC) for the vector-valued image that incorporates Bhattacharyya distance michailovich2007image . The only difference is that the authors calculate the probability functions by Bhattacharyya distance instead of the Maximum-Likelihood distance in this paper. The performance of the proposed algorithm was demonstrated on the retinopathy dataset quellec2008optimal niemeijer2009retinopathy and with an increase of 3% in F-score compared with the best-performed methods.
In lian2017spatial , Lian et al. introduced a tumor delineation method in fluorodeoxyglucose positron emission tomography (FDG-PET) images by using spatial ECM lelandais2014fusion with adaptive distance metric. The authors proposed the adaptive distance metric to extract the most valuable features, and spatial ECM was used to calculate mass functions. Compared with ECM and spatial ECM, the proposed method showed a 14% and 10% increase in Dice score when evaluated on the FDG-PET images of non-small cell lung cancer (NSCLC) patients, showing good performance.
In huang2021evidential , Huang et al. proposed a 3D PET/CT lymphoma segmentation framework with BFT and deep learning. This is the first work that applied BFT with a deep neural network for medical image segmentation. In this paper, the PET and CT images were concatenated as a signal modal input method and transferred into UNet to get high-level semantic features. Then the ENN classifier was used to map high-level semantic features into mass functions by fusing the contribution of $K$ prototypes. Moreover, the segmentation uncertainty was considered in this paper with an uncertainty loss during training. The reported quantitative and qualitative results showed that the proposal outperforms the state-of-the-art methods. Based on the first work, Huang et al. verified the similarity of RBF and ENN in uncertainty quantification and merged them with the deep neural network (UNet) for lymphoma segmentation huang2022lymphoma . The segmentation performance confirmed that RBF is an alternative approach of ENN denoeux19d to act as an evidential classifier and showed that the proposal outperforms the baseline method UNet and the state-of-the-art both in accuracy and reliability.
Before 2020, the BFT-based medical image segmentation methods with single modal medical used low-level image features, e.g., grayscale and shape features, to generate mass functions, which limits the segmentation accuracy. Moreover, None of them discussed the reliability of the segmentation results. Huang et al. first merge BFT with a deep segmentation model (UNet) and learn the representation of mass functions with some learning algorithms huang2021evidential . Based on this, Huang et al further discuss the relationship between segmentation accuracy and reliability in huang2022lymphoma , which takes a new direction to study reliable medical image segmentation methods and bridge the gap between experimental results and clinical application.
5.1.2 Multimodal evidence fusion
Single-modality medical images often do not contain enough information to present the information about the disease and are often tainted with uncertainty. In addition to feature-level evidence fusion, the fusion of multimodal evidence is also important to achieve accurate and reliable medical image segmentation performance. Approaches for modality-level evidence fusion can be summarized into three main categories according to the way they calculate the evidence: probabilistic-based fusion, fuzzy set-based fusion, and BFT-based fusion. The development of convolution neural networks (CNNs) further contributes to the probabilistic-based fusion methods zhou2019review , which can be summarized into image-level fusion (e.g., data concatenation peiris2021volumetric ), feature-level fusion (e.g., attention mechanism concatenation zhou2020fusion zhou2022tri ), and decision-level fusion (e.g., model ensembles kamnitsas2017ensembles ). While none of those methods consider the conflict of source evidence, i.e., the modality-level uncertainty is not well studied, which limits the reliability and explainability of the performance.
This section focus on the BFT-based segmentation methods with modality-level evidence fusion. Figure 11 shows an example framework of multimodal evidence fusion with a single classifier. We separate the segmentation process into four steps: masses calculation, feature-level evidence fusion (alternative), modality-level evidence fusion, and decision-making. Compared with single-modal evidence fusion 5.2.1, multimodal evidence fusion focuses here not only on feature-level but also on modality-level evidence fusion (It should be noted that feature-level evidence fusion is not necessary in this case). multimodal evidence fusion is the most popular application for BFT in the medical image segmentation domain. Therefore we classify those methods according to their input modal for better analysis. Table 3 summarizes the segmentation methods with multimodal inputs and a single classifier/cluster with the main focus on modality-level evidence fusion.
Fusion of multimodal MR images
In bloch1996some , Bloch first proposed a BFT-based dual-echo MR pathological brains tissue segmentation model with uncertainty and imprecision quantification. The author assigned mass functions based on a reasoning approach that uses grey-level histograms provided by each image to choose focal elements. After defining mass functions, Dempster’s rule combined the mass from dual-echo MR images for each pixel. Based on the first work with BFT, in bloch2008defining , Bloch proposed to use fuzzy mathematical morphology bloch1995fuzzy , i.e., erosion and dilation, to generate mass functions by introducing imprecision in the probability functions and estimating compound hypotheses. Then Dempster’s rule is used to fuse mass functions from multimodal images. It should be noted that this paper is under the strong assumption that it is possible to represent imprecision by a fuzzy set, also called the structuring element. Application examples on dual-echo MR image fusion showed that the fuzzy mathematical morphology operations could represent either spatial domain imprecision or feature space imprecision (i.e., grey levels features). The visualized brain tissue segmentation results yield the robustness of the proposed method.
As we motioned in Section 4.2.1, Zhu et al. proposed modeling mass functions in BFT using FCM and spatial neighborhood information for image segmentation. The visualized segmentation results on MR brain images showed that the fusion-based segmented regions are relatively homogeneous, enabling accurate measurement of brain tissue volumes compared with single modal input MR image input.
In ghasemi2012brain , Ghasemi et al. presented a brain tissue segmentation approach based on FCM and BFT. The authors used the FCM to model two different input features: pixel intensity and spatial information, as membership values (MVs). Then for each pixel, the RMV transformation ghasemi2012brain was used to map MVs into mass functions. Last, the authors used Dempster’s rule to fuse intensity-based and spatial-based mass functions to get final segmentation results. Compared with FCM, the authors reported an increase in Dice and accuracy. As an extension of ghasemi2012brain , Ghasemi et al. proposed an unsupervised brain tumor segmentation method that modeled pixel intensity and spatial information into mass functions with RMV transformation and fused the two mass functions with Dempster’s rule in ghasemi2013novel .
In Rui2013Lesion , Wang et al. proposed a lesion segmentation method for infarction and cytotoxic brain edema. The authors used a method similar to Zhu’s model to define simple and double hypotheses. FCM bezdek1984fcm was used first to construct the mass functions of simple hypotheses $\{a\}$ and $\{b\}$. Then the mass function of double hypotheses was defined by
$m(\{a,b\})=\frac{1}{4}\times\frac{min(m(\{a\}),m(\{b\}))}{m(\{a\})+m(\{b\})}$. Then the authors used Dempster’s rule for modality-level evidence fusion. The results showed that infarction and cytotoxic brain edema around the infarction could be well segmented by final segmentation.
In makni2014introducing , Makni et al. introduced a multi-parametric MR image segmentation model by using spatial neighborhood in ECM for prostate zonal anatomy. The authors extended ECM with neighboring voxels information to map multi-parametric MR images into mass functions. Then the authors used prior knowledge related to defects in the acquisition system to reduce uncertainty and imprecision. Finally, the authors used Dempster’s rule to fuse the mass functions from the multi-parametric MR images. The method achieved good performance on prostate multi-parametric MR image segmentation.
In tavakoli2018brain , Tavakoli et al. proposed a segmentation method based on the evidential fusion of different modalities (T1, T2, and Proton density (PD)) for brain tissue. The authors used FCM to get MVs and used the RMV transformation ghasemi2012brain to transform the clustering MVs into mass functions. The authors first formed the belief structure for each modal image and used Dempster’s rule to fuse the three modalities’ mass functions of T1, T2, and PD. Compared with FCM, this proposal achieved a 5% improvement in the Dice score. Based on Tavakoli’s method tavakoli2018brain , Lima et al. proposed a modified brain tissue segmentation method in lima2019modified . The authors tested their method with four modality inputs: T1, T2, PD, and Flair. The reported results outperformed both the baseline method FCM and Tavakoli’s method with four modality evidence fusions.
In xiao2017vascular , Xiao et al. proposed an automatic vascular segmentation algorithm, which combines the grayscale and shape features of blood vessels and extracts 3D vascular structures from the head phase-contrast MR angiography dataset. First, grayscale and shape features are mapped into mass functions by using the GD-based method. Second, a new reconstructed vascular image was established according to the fusion of vascular grayscale and shape features by Dempster’s rule. Third, a segmentation ratio coefficient was proposed to control the segmentation result according to the noise distribution characteristic. Experiment results showed that vascular structures could be detected when both grayscale and shape features are robust. Compared with traditional grayscale feature-based or shape feature-based methods, the proposal showed better performance in segmentation accuracy with the decreased over-segmentation and under-segmentation ratios by fusing two sources of information.
Since Bloch’s early work fully demonstrated the advantages of BFT in modeling uncertain and imprecision, introducing partial or global ignorance, and fusing conflicting evidence in a multimodal MR images segmentation task, researchers in this domain have gone further to explore the advantages of BFT in multi MR image fusion. Among these research works, FCM is the most popular clustering algorithm to map input information into MVs. Ration MV transformation or Zhu’s model is usually used to generate mass functions from MVs. Apart from these two-step methods, the GD-based model can also be used to generate mass functions directly.
Fusion of RGB channels
In vannoorenberghe1999dempster , Vannoorenberghe et al. pointed out that taking R, G, and B channels as three independent information sources can be limited and weak for medical image tasks and proposed a color image segmentation method based on BFT. They calculated the degree of evidence by mapping R, G, and B channel’s intensity into mass functions using the Gaussian distribution information (similar to GD-based model chen2012manifold ) with an additional discounting operation). Then three pieces of evidence were fused with Dempster’s rule. The proposed segmentation method was applied to biomedical images to detect skin cancer (melanoma). Experiments showed a significant part of the lesion is correctly extracted from the safe skin. The segmentation performance is limited by feature representation, e.g., some regions correspond to pixels that cannot be classified as either the safe skin or the lesion because only the pixel-level feature is insufficient for hard-example segmentation.
In chaabane2009relevance Chaabane et al. proposed a color medical image segmentation method based on fusion operation. Compared to vannoorenberghe1999dempster , the authors first modeled the probability of each region belongs an image by a Gaussian distribution, then used Appriou’s second model to map probability into mass functions. Dempster’s rule then combines three color channel evidence. Compared with single-channel segmentation results, the fused results achieved a 10% increase in segmentation sensitivity.
Different from the methods vannoorenberghe1999dempster and chaabane2009relevance that decompose color images into R, G, B channels, Harrabi et al. harrabi2012color presented a color image segmentation method that represents the color image with 6 color spaces (RGB, HSV, YIQ, XYZ, LAB, LUV). The segmentation operation is based on multi-level thresholding and evidence fusion techniques. First, the authors identify the most significant peaks of the histogram by multi-level thresholding with the two-stage Otsu optimization approach. Second, the GD-based model was used to calculate the mass functions for each color space. Then the authors used Dempster’s rule to combine six sources of information. Compared with single color spaces, such as RGB and HSV, the fused result of six color spaces has a great increase in segmentation sensitivity, for example, an increase of 4% and 7% in RGB and HSV, respectively.
In trabelsi2015skin , Trabelsi et al. applied BFT in optical imaging with color to improve skin lesion segmentation performance. The authors decomposed color images into R, G, and B channels and applied the FCM method on each channel to get probability functions for pixel $x$ in each color space. In this paper, the author takes the probability functions as mass functions and calculates the orthogonal sum of the probability functions for the three-channel images. Even experiments showed about 10% improvements in segmentation accuracy compared with single-channel results, further study could be done on this work to generate uncertainty from probability functions to get achieve performance.
In general, the BFT-based RGB medical image segmentation approaches are used to generate mass functions by possibility distributions, e.g., Gaussian distribution and Possibility C-means distribution, and fuse them by Dempster’s rule. Though the authors claimed they could get better performance compared with single color input, the segmentation performance is limited by features because gray-scale and intensity features are not robust and efficient in representing image information. Further research could take both feature extraction and uncertainty quantification into consideration, e.g., a deep feature extraction model with an evidential classifier, to improve the performance.
Fusion of PET/CT
In lelandais2012segmentation , Lelandais et al. proposed a BFT-based multi-trace PET images segmentation method to segment biological target volumes. First, the authors used a modified FCM algorithm with the discounting algorithm to determine mass functions. The modification integrated a disjunctive combination of neighboring voxels inside the iterative process. Second, the operation of reduction of imperfect data was conducted by fusing neighboring voxels using Dempster’s rule. Based on this first work, Lelandais et al. proposed an ECM-based fusion model lelandais2014fusion for biological target volume segmentation with multi-tracer PET images. The segmentation method introduced in this paper is similar to the one introduced in makni2014introducing with a different application.
In derraz2015joint , Derraz et al. proposed a multimodal tumor segmentation framework for PET and CT inputs. Different from Lelandais et.al.’s work that uses ECM or optimized ECM to generate mass functions directly, the authors get mass functions in two steps. They first proposed a NonLocal Active Contours (NLAC) based variational segmentation framework to get probability results. Then, similar to the authors’ previous work derraz2013globally ; derraz2013image , they used Appriou’s second model appriou200501 to map MVs into mass functions. Then Dempster’s rule is used to fuse the mass functions from PET and CT modalities. The framework was evaluated on a lung tumor segmentation task. Compared with the state-of-the-art methods, this framework yielded the highest Dice score for tumor segmentation.
Based on Lelandais et al.’s work, Lian et al. proposed a tumor segmentation method lian2017accurate by using Spatial ECM lelandais2014fusion with Adaptive Distance Metric lian2017spatial in FDG-PET images with the guidance of CT images. Based on the first work, Lian et al. proposed a co-segmentation method of lung tumor segmentation lian2018joint ; lian2018unsupervised . They take PET and CT as independent inputs and use ECM to generate mass functions. At the same time, an adaptive distance metric is used to quantify clustering distortions and spatial relationships during the evidential clustering procedure. Then Dempster’s rule is used to fuse mass functions from PET and CT modalities. The quantitative and qualitative evaluation results showed superior performance compared with single modal segmentation results with an increase of 1% and 58% in PET and CT in Dice scores, respectively.
ECM is the most common BBA method to generate mass functions for PET/CT medical image segmentation approaches. Similar to the BFT-based RGB medical image segmentation methods, the segmentation performance here is limited by feature extraction methods. Further research could build on the study of deep feature representation and combine ECM with deep neural networks to learn mass feature representation. A good example of applying the evidential cluster in a neural network can be found in denoeux2021nn .
5.2 BFT-based medical image segmentation with several classifiers or clusterers
It is normal to see that two or more physicians cooperate together for disease diagnosis, which can minimize the impact of physicians’ misjudgments. Therefore, combining the results from multiple decision mechanisms as well as addressing the conflicts is critical to achieving a more reliable diagnosis. This section introduces the BFT-based medical image segmentation methods with several classifiers or clusterers. We follow the same way in Section 5.1 to separate them into a single modal and multimodal evidence fusion and introduce them in Sections 5.2.1 and 5.2.2, respectively.
5.2.1 Single modal evidence fusion
Compared with the methods presented in Section 5.2.1, the methods introduced in this section focus on feature-level (alternative) and classifier-level evidence fusion, which aims to minimize the impact of misjudgments caused by the model’s inner shortcomings. Figure 12 shows an example of a medical image segmentation framework with single modal input and several classifiers. We separate the segmentation process into four steps: masses calculation, feature-level evidence fusion (alternative), classifier-level evidence fusion, and decision-making. Table 4 summarizes the segmentation methods with single modal inputs and several single classifiers/clusterers with the main focus on classifier/clusterer level evidence fusion.
In capelle2002segmentation , Capelle et al. proposed a segmentation scheme for MR images based on BFT. The authors followed the same idea of the Evidential K-NN rule denoeux1995k to map image features into mass functions. Then, the authors applied the evidential fusion process to classify the voxels. Based on the first work, Capelle et al. proposed an evidential segmentation scheme of multimodal MR images for brain tumor detection in capelle2004evidential . This work focused on analyzing different evidential modeling techniques and on the influence of the introduction of spatial information to find the most effective brain tumor segmentation method. Three different BFT-based models: the distance-based BFT model (EKNN) denoeux2000neural , the likelihood function-based BFT method (Shafer’s model shafer1976mathematical , and Appriou’s first model (appriou1999multisensor ) were used to model information, and Dempster’s rule is used to fuse the three mass functions. This study concluded that neighborhood information increases the evidence of class membership for each voxel, thus making the final decision more reliable. Experimental results showed better segmentation performance compared with the state-of-the-art methods when the paper was published.
In taleb2002information , Taleb-Ahmed proposed a segmentation method of MR sequences of vertebrae in the form of images of their multiple slices with BFT. The authors used three different classifiers to calculate three kinds of mass functions. Firstly, the authors used grey-level intensity and standard deviation information to calculate two pixel-level mass functions with two fixed thresholds. Then the distance between two matching contours of consecutive slices ($P$ and $Q$) is used to calculate contour-level mass functions with the following definition:
$$\displaystyle m(\{S_{PQ}\})=1-e^{-\eta\left|d(P_{i},Q_{i})-\beta\right|},\quad d(P_{i},Q_{i})\in[\rho..\beta],$$
(35a)
$$\displaystyle m(\overline{\{S_{PQ}\}})=1-e^{-\eta\left|d(P_{i},Q_{i})-\beta\right|},\quad d(P_{i},Q_{i})\in[\beta..\propto],$$
(35b)
$$\displaystyle m(\Omega)=e^{-\eta\left|d(P_{i},Q_{i})-\beta\right|},\quad\forall d(P_{i},Q_{i}),$$
(35c)
where $P_{i}$ and $Q_{i}$ are two matching points of the slices $P$ and $Q$, $d(P_{i},Q_{i})$ is corresponding distance; $\Omega=\{S_{PQ},\overline{S_{PQ}}\}$, $S_{PQ}$ means two points $P_{i}$ and $Q_{i}$ are both belonging to cortical osseous. Parameter $\beta$ represents the tolerance which the expert associates with the value $d(P_{i},Q_{i})$, $\rho$ is the inter-slice distance and $\eta$ makes it possible to tolerate a greater inaccuracy in the geometrical resemblance of two consecutive contours. Then Dempster’s rule was used to combine the three mass functions for final segmentation.
In barhoumi2007collaborative , Barhoumi et al. introduced a new collaborative computer-aided diagnosis system for skin malignancy tracking. First, two diagnosis opinions were produced by perceptron neural network classification and content-based image retrieval (CBIR) schemes. Then Dempster’s rule was used to fuse the two opinions to achieve a final malignancy segmentation. Simulation results showed that the proposed BFT-based combination could generate accurate diagnosis results. The frame of discernment in this paper is composed of two singletons, and the mass functions in this paper are said to be Bayesian.
In guan2011study , Guan et al. proposed a human brain tissue segmentation method with BFT. The authors first used Markov random forest (MRF) li2009markov for spatial information segmentation and then used a Two-Dimensional Histogram (TDH) method of fuzzy clustering duan2008multi to get a vague segmentation. Then a redundancy image was generated, representing the difference between the MRF and TDH methods. Then Zhu’s model zhu2002automatic was used to calculate mass functions. Finally, Dempster’s rule fused the two segmentation results and the generated redundancy image to handle controversial pixels. The visual segmentation results showed that this method has higher segmentation accuracy compared with the state-of-the-art.
As discussed in Section 3.2, Dempster’s rule becomes numerically unstable when combining highly conflicting mass functions. In this condition, the fused results will be improper, and any tiny changes in mass functions can bring sharp changes to the fusion results. Researchers in the medical domain has also awarded this limitation. In ketout2012improved , Ketout et al. proposed a modified mass computation method to break this limitation and apply their proposal for endocardial contour detection. First, the output of each active contour model (ACM) kass1988snakes was regarded as mass functions. Second, a threshold was proposed to check if the evidence $m$ conflicts with others. If there was conflict, a modifying operation was used on the conflicting evidence $m$: $m=m+1/l$ with modified pace $1/l$. The parameter $l$ is related to the number of hypothesis $W$ in $\Omega$, generally, $l=2W$. Then the results of edge set-based segmentation li2005level and region set-based segmentation chan2001active were fused by using the improved BFT xin2005improved to get a more accurate contour of the left ventricle. Experimental results showed that the fused contour is closer to the ground truth than the contour from the edge or region. And false detection of the two contours was suppressed in the resulting image by rejecting the conflicting events by the fusion algorithm. Meanwhile, the proposed method could detect the edges of the endocardial borders even in low-contrast regions.
In wen2018improved , Wen et al. proposed an improved MR image segmentation method based on FCM and BFT. First, the authors did a image-level fusion for two modality images $A$ and $B$ with the function $F(x,y)=w_{1}\times g_{A}(x,y)+w_{2}\times g_{B}(x,y)$, where $x$ and $y$ are image pixels and $w_{1}$ and $w_{2}$ are the weighted coefficients which are used to adjust the importance of different images for the final fusion result with $w_{1}+w_{2}=1$. Second, the authors calculated the MV by FCM and calculated the mass functions of simple and double hypotheses by Zhu’s model zhu2002automatic without the neighboring pixel information. Third, the authors generated another mass functions by weighting those of its neighboring pixels with the GD-based model and used Zhu’s model zhu2002automatic again to construct simple and double hypotheses mass functions. Finally, the authors used Dempster’s rule to complete the fusion of the two mass functions to achieve the final segmentation. Compared with the FCM-based method, the proposed method can better decrease the conflict in multiple sources to achieve easy convergence and significant improvement by using Dempster’s rule for classifier-level evidence fusion.
Besides, with the development of CNNs, the research community used Dempster’s rule for the fusion of multiple CNN classifiers. In george2020breast , George et al. proposed a breast cancer detection system using transfer learning and BFT. This first work first applied BFT in multiple evidence fusion with deep learning. Image features were represented in high-level features using a convolutional neural network for feature extraction, such as ResNet-18, ResNet-50, ResNet-101, GoogleNet, and AlexNet. A patch-based feature extraction strategy was used to avoid wrong segmentation of the boundaries and provide features with good discriminative power for classification. The extracted features were classified into benign and malignant classes using a support vector machine (SVM). A discounting operation was applied to transfer probability-based segmentation maps into mass functions. The discounted outputs from the different CNN-SVM frameworks are now combined using Dempster’s rule. This work takes advantage of deep learning and BFT and has achieved good performance. Compared with majority voting-based fusion methods, BFT-based fusion showed superior segmentation accuracy. Compared with a single classifier, such as ResNet-101, the fused framework achieves the accuracy of sensitivity, specificity, and AUC with an increase of 1%, 0.5%, 3%, and 2%, respectively. Also, the authors compared their results with the state-of-the-art method and achieved comparable segmentation accuracy.
Apart from using BFT to combine the discounted probabilities from the CNN classifiers george2020breast , another solution is to construct a deep evidential segmentation framework directly. In huang2021belief , Huang et al. proposed a BFT-based semi-supervised learning framework (EUNet) for brain tumor segmentation. This work applied BFT in a deep neural network to quantify segmentation uncertainty directly. During training, two kinds of evidence were obtained: the segmentation probability functions and mass functions generated by UNet and EUNet, respectively. Dempster’s rule was then used to fuse the two pieces of evidence. Experimental results showed that the proposal has better performance than state-of-the-art methods. It achieved around 10%, 4%, and 2% increase in Dice score, Positive predictive value, and sensitivity, respectively, compared with the baseline (UNet). Moreover, the authors showed how BFT exceeds probability-based methods in addressing uncertainty boundary problems. This is the first work embeds BFT into CNN and achieves an end-to-end deep evidential segmentation model.
The approaches introduced in this section use several classifiers/clusterers to generate different mass functions and fuse them by Dempster’s rule. Among those approaches, George et al. george2020breast first applied Dempster’s rule to combine the discounted probabilities from different deep segmentation models. Huang et al. huang2021belief first merged ENN with UNet to construct an end-to-end segmentation model and fuse two kinds of evidence by Dempster’s rule. Compared to George et al.’s approach george2020breast , Huang et al.’s approach huang2021belief can generate mass functions directly from a deep segmentation model, which is more promising.
5.2.2 Multimodal evidence fusion
Figure 13 shows an example of a medical image segmentation framework with several classifiers and multimodal inputs, which is more complex than the frameworks we introduced in Sections 5.2.2, 5.2.2, and 5.1.1. The segmentation process comprises five steps: masses calculation, feature-level evidence fusion (alternative), classifier-level evidence fusion, modality-level evidence fusion, and decision-making. Pixels at the same position from different modalities are transferred into different classifiers and different BBA methods to get pixel-level mass functions. Dempster’s rule is used first to fuse feature-level evidence fusion, then to fuse classifier-level evidence, and last to fuse modality-level evidence. Table 5 summarizes the segmentation methods with multimodal inputs and several classifiers/clusterers focusing on classifier/clusterer fusion and modality-level evidence fusion
In gautier2000belief , Gautier et al. proposed a method for helping physicians monitor the spinal column’s diseases with multimodal MR images. At first, an initial segmentation was applied with active contour lai1994deformable . Then several mass functions were obtained from experts’ opinions on constructing the frame of discernment. Thus, the mass functions here are human-defined. Finally, Dempster’s rule was then used to fuse the mass functions from different experts. The proposal got the most reliable segmentation maps with the experimental illustration results when the paper was published.
Based on the first work of multimodal medical image fusion with a single cluster chaabane2009relevance , Chaabane et al. presented another BFT-based segmentation method with several clustererschaabane2011new . First, possibilistic C-means clustering bezdek2013pattern was used on R, G, and B channels to get three MVs. Then the MVs were mapped into the mass functions of simple and double hypotheses using Zhu’s model zhu2002automatic . Dempster’s rule is used first to fuse three mass functions from three corresponding color spaces. Based on the initial segmentation results, another mass function was calculated associated with each pixel and its neighboring pixels for each color space. Finally, the authors used Dempster’s rule again to fuse the two mass functions from two corresponding clusterers. Experimental segmentation performance in cell images highlights the effectiveness of the proposed method. Compared with the FCM-based segmentation method, the proposal increased by 15% in sensitivity segmentation.
In lajili2018two , Lajili et al. proposed a two-step evidential fusion approach for breast region segmentation. The first evidential segmentation results were obtained by a grey-scale-based K-means clustering method. A sigmoid function was used to define the mass functions for each class $C_{x}(x=1...,k)$:
$$\displaystyle m_{z\in C_{x}}(\{Br\})=\frac{1}{1+\exp^{2-x}},$$
(36a)
$$\displaystyle m_{z\in C_{x}}(\{Ba\})=\frac{1}{1+\exp^{-2-x}},$$
(36b)
$$\displaystyle m_{z\in C_{x}}(\Omega)=1-m(\{Br\})-m(\{Ba\}),$$
(36c)
where $x$ denotes the index of the class $C_{x}$ for pixel $z$, $Br$ represents breast, and $Ba$ represents background. For local-homogeneity-based segmentation, the authors modeled the uncertainty with a threshold value $\alpha$, by defining $m(\{Br\})=1-\alpha$, $m(\{Ba\})=0$, $m(\{Br,Ba\})=\alpha$, where $\alpha$ represents the belief mass affected by the uncertainty on the membership of the pixel $z$. A final fusion strategy with Dempster’s rule was applied to combine evidence from the two mass functions. Experiments were conducted on two breast datasets suckling1994mammographic bowyer1996digital . The proposed segmentation approach yielded more accurate results than the best-performed method. It extracted the breast region with correctness equal to 0.97, which is 9% higher than the best-performed method.
In huang2021deep , Huang et al. proposed to take PET and CT as two modal inputs and use to UNet model to segment lymphoma separately. Then the two segmentation masks are fused by Dempster’s rule. Though this is the first work that applied BFT in multimodal evidence fusion with several deep segmentation models, the shortcoming of this work is that only probability functions other than mass functions are used for evidence fusion. To improve this first work, Huang et al. proposed a multimodal evidence fusion framework with contextual discounting for brain tumor segmentation huang2022contextual . In this work. With four modules of feature extraction and one module of evidential segmentation, the method assigns to each voxel a mass function. In addition, a contextual discounting layer is designed to take into account the reliability of the method when classifying different classes. Finally, the Dempster rule is used to combine the discounted evidence to obtain a final segmentation. The authors also claimed that this method could be used together with any state-of-the-art segmentation module to improve the final performance.
There are only a few studies that take multimodal medical images as independent inputs and use independent classifiers to generate mass functions. The performance of this kind of approach is limited by the representation of image features and the ability to quantify the model’s uncertainty. Huang et al. huang2022contextual first merge ENN with UNet for the fusion of multimodal MR images with contextual discounting. It enables the model to generate a learned reliability metric from input modalities during different segmentation tasks, which is good potential to make the results more explainable.
5.3 Discussion
The choice of using single modal or multimodal depends on the dataset. Generally, the more source data we have, the more reliable segmentation results we will get. The choice of a single or several classifiers/clusterers depends on the limitation of the computation source and the requirement of computation efficiency.
Considering the main advantages of using multiple classifiers for multimodal medical images, as shown in the work of huang2022contextual , a potential research direction based on this approach could be promising.
Prior to 2020, BFT-based medical image segmentation methods limited segmentation accuracy due to the use of low-level image features, such as grayscale and shape features, to generate mass functions. Moreover, none of them considered the segmentation reliability. Since the application of deep learning in medical image segmentation has been very successful, the use of BFT in deep neural networks should be a promising research direction, in particular, to quantify the uncertainty and reliability of the segmentation results.
6 Conclusion and future work
6.1 Conclusion
The main challenges we have for medical image segmentation tasks are: 1) imperfect medical image resolution, 2) noisy ground truth, and 3) unreliable segmentation models due to imperfect data and ground truths. This leads to epistemic segmentation uncertainty and has been the gap between experimental performance and clinical application for a while. For example, traditional medical image segmentation methods are limited by the use of low-level image features for decision-making kimmel2003fast ; salvador2004determining ; onoma2014segmentation , leading to poor segmentation accuracy. The success of machine learning, especially deep learning, has made promising contributions to segmentation accuracy thanks to its powerful ability in high-level semantic feature representation, such as UNet ronnebergerconvolutional and its variants trebing2021smaat ; cao2021swin ; hatamizadeh2022unetr . In contrast, a clinically applicable deep segmentation model depends not on the segmentation accuracy but also the reliability, which is critical for the high-risk task.
The systematic and comprehensive discussion of uncertainty and reliability of machine learning methods started in 2020 hullermeier2021aleatoric , which also attracted researchers’ attention in the medical image segmentation domain to study both accurate and reliable segmentation methods kwon2020uncertainty ; mehrtash2020confidence ; ghoshal2021estimating ; zhou2022trustworthy . This review provides a comprehensive overview of the BFT-based medical image segmentation methods. It first shows how imperfect medical image could be modeled by the degree of belief and ignorance(uncertainty) by mass functions; and then shows how unreliable or conflicting source information can be modeled to reach reliable fusion results. The application of segmentation methods covers a range of segmenting different human tissues (i.e., heart, lung) and tumors (i.e., brain, breast, and skin). In particular, we first briefly introduce medical image segmentation methods in Section 2 by introducing the traditional and deep learning medical segmentation methods and by discussing their shortcomings and limitations. Then we presented the history and the fundamental knowledge of BFT in Section 3. Section 4 comprehensively introduced the BBA methods used to represent uncertainty (generate mass functions) in medical image segmentation tasks. Sections 5 introduced the BFT-based medical image segmentation methods in detail by grouping them according to the number of input modalities and classifiers used to generate mass functions.
In certain cases, the BFT-based methods could surpass the probability-based methods by modeling information more effectively (with uncertainty or ignorance quantification in focal sets) and fusing multiple evidence in different stages, such ad feature-level, modality-level, and classifier-level evidence fusion. We hope this review could provide an intuitive understanding of BFT-based techniques that significantly contribute to medical image segmentation, especially for uncertain/imprecise multimodal medical image segmentation.
6.2 Future work
Although medical image segmentation with belief function theory has advantages in expressing segmentation uncertainty and improving the reliability of the performance, limitations of existing methods remain and can be further investigated.
First, most of the existing BFT-based medical image segmentation methods still stay at the stage of using low-level features and do not fully use the advantages of deep learning. Applying BFT in deep segmentation networks should be a potential research direction for getting both accurate and reliable segmentation models, particularly in medical image segmentation because of the availability of medical knowledge a priori easily modeled by mass functions. Ref george2020breast is the first successful application of Dempster’s rule for the fusion of different CNNs. Ref huang2022lymphoma confirmed the similarity of ENN and RBF when acting as an evidential classifier and integrated both the two classifiers within a deep segmentation network to improve the segmentation accuracy and reliability. We believe that more promising results will be obtained by cooperating BFT with the existing powerful deep medical image segmentation models trebing2021smaat ; cao2021swin ; hatamizadeh2022unetr .
Second, even though the BFT-based fusion methods have considered the conflict of multiple sources of evidence, there is a problem in that the multimodal medical images may have different reliability when segmenting different regions. For example, for the BRATS brain tumor dataset 111http://braintumorsegmentation.org/, the domain knowledge from the physicians points that areas describe the enhancing tumor with visually faint and enhancement on T1Gd MR images, and edema is defined by the abnormal hyperintense signal envelope on the FLAIR modality baid2021rsna . The multimodal medical image fusion operation may fail if some source information is unreliable or partially reliable when doing different segmentation tasks. In huang2022contextual , Huang et al. proposed a new method to learn reliability coefficients conditionally on different segmentation tasks in a deep neural network. Another interesting future work, therefore, is to obtain task-specific reliability, as well as to explain the results with deep evidential neural networks using contextual discounting for multimodal or cross-modal medical image segmentation tasks.
Third, acquiring big labeled training data is particularly challenging for medical image segmentation tasks, which has become the bottleneck of learning-based segmentation performance. The successful application of using unsupervised BBA methods for medical image segmentation points out another new direction to address the short annotation problem. As far as we know, there is no research work on the study of unsupervised BBA methods in a deep learning network. Evidential Clustering with neural network (NN-EVCLUS) denoeux21b and EKNN with partially supervised learning denoeux2019new are two successful applications of generating mass functions with unsupervised and semi-supervised learning. These works provide insights into how to use unsupervised/semi-supervised learning to quantify segmentation uncertainty with unannotated or partially annotated data sets.
We hope this review could increase awareness of the challenges of existing BFT-based medical image segmentation methods and call for future contributions to bridge the gap between experimental performance and clinical application, as well as achieve accurate, reliable, and explainable deep segmentation models.
Acknowledgements
The authors like to thank Thierry Denoeux for helpful remarks and comments on the content of this paper. We are also grateful for the insight suggestions made by the reviewers. The first author received funding from the China Scholarship Council (No.201808331005).
This work was carried out in the framework of the Labex MS2T, which was funded by the French Government through the program “Investments for the future” managed by the National Agency for Research (Reference ANR-11-IDEX-0004-02)
References
(1)
J. Batenburg, K., J. Sijbers, Adaptive thresholding of tomograms by projection
distance minimization, Pattern Recognition 42 (10) (2009) 2297–2305.
(2)
R. Kimmel, Fast edge integration, in: Geometric Level Set Methods in Imaging,
Vision, and Graphics, Springer, New York, USA, 2003, pp. 59–77.
(3)
D. Onoma, S. Ruan, S. Thureau, L. Nkhali, R. Modzelewski, G. Monnehan, P. Vera,
I. Gardin, Segmentation of heterogeneous or small fdg pet positive tissue
based on a 3d-locally adaptive random walk algorithm, Computerized Medical
Imaging and Graphics 38 (8) (2014) 753–763.
(4)
S. Salvador, P. Chan, Determining the number of clusters/segments in
hierarchical clustering/segmentation algorithms, in: 16th IEEE international
conference on tools with artificial intelligence, IEEE, Boca Raton, USA, Nov,
2004, pp. 576–584.
(5)
J. Suykens, J. Vandewalle, Least squares support vector machine classifiers,
Neural processing letters 9 (3) (1999) 293–300.
(6)
C. Strobl, A.-L. Boulesteix, A. Zeileis, T. Hothorn, Bias in random forest
variable importance measures: Illustrations, sources and a solution, BMC
bioinformatics 8 (1) (2007) 25.
(7)
D. Kleinbaum, D. Hedeker, Logistic regression: a self-learning text,
Statistical Methods in Medical Research 5 (1) (1996) 103.
(8)
O. Ronneberger, P. Fischer, T.-n. Brox, Convolutional networks for biomedical
image segmentation, in: International Conference on Medical Image Computing
and Computer-Assisted Intervention, Munich, Germany, Oct, 2015.
(9)
A. Myronenko, 3d mri brain tumor segmentation using autoencoder regularization,
in: International MICCAI Brainlesion Workshop, Springer, 2018, pp. 311–320.
(10)
F. Isensee, J. Petersen, . Klein, D. Zimmerer, nnu-net: Self-adapting framework
for u-net-based medical image segmentation, arXiv preprint arXiv:1809.10486
(2018).
(11)
D. Bahdanau, K. Cho, Y. Bengio, Neural machine translation by jointly learning
to align and translate, arXiv preprint arXiv:1409.0473 (2014).
(12)
A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez,
Ł. Kaiser, I. Polosukhin, Attention is all you need, Advances in neural
information processing systems 30 (2017).
(13)
N. Carion, F. Massa, G. Synnaeve, N. Usunier, A. Kirillov, S. Zagoruyko,
End-to-end object detection with transformers, in: European conference on
computer vision, Springer, 2020, pp. 213–229.
(14)
K. Han, A. Xiao, E. Wu, J. Guo, C. Xu, Y. Wang, Transformer in transformer,
Advances in Neural Information Processing Systems 34 (2021) 15908–15919.
(15)
E. Hüllermeier, W. Waegeman, Aleatoric and epistemic uncertainty in machine
learning: An introduction to concepts and methods, Machine Learning 110 (3)
(2021) 457–506.
(16)
S. C. Hora, Aleatory and epistemic uncertainty in probability elicitation with
an example from hazardous waste management, Reliability Engineering & System
Safety 54 (2-3) (1996) 217–223.
(17)
A. Der Kiureghian, O. Ditlevsen, Aleatory or epistemic? does it matter?,
Structural safety 31 (2) (2009) 105–112.
(18)
B. Lakshminarayanan, A. Pritzel, C. Blundell, Simple and scalable predictive
uncertainty estimation using deep ensembles, Advances in neural information
processing systems 30 (2017).
(19)
G. E. Hinton, D. Van Camp, Keeping the neural networks simple by minimizing the
description length of the weights, in: Proceedings of the sixth annual
conference on Computational learning theory, 1993, pp. 5–13.
(20)
D. J. MacKay, A practical bayesian framework for backpropagation networks,
Neural computation 4 (3) (1992) 448–472.
(21)
Y. Gal, Z. Ghahramani, Dropout as a bayesian approximation: Representing model
uncertainty in deep learning, in: international conference on machine
learning, PMLR, 2016, pp. 1050–1059.
(22)
D. Tran, M. Dusenberry, M. van der Wilk, D. Hafner, Bayesian layers: A module
for neural network uncertainty, Advances in neural information processing
systems 32 (2019).
(23)
C. Rupprecht, I. Laina, R. DiPietro, M. Baust, F. Tombari, N. Navab, G. D.
Hager, Learning in an uncertain world: Representing ambiguity through
multiple hypotheses, in: Proceedings of the IEEE international conference on
computer vision, 2017, pp. 3591–3600.
(24)
A.-P. Dempster, Upper and lower probability inferences based on a sample from a
finite univariate population, Biometrika 54 (3-4) (1967) 515–528.
(25)
G. Shafer, A mathematical theory of evidence, Princeton university press, 1976.
(26)
T. Denœux, D. Dubois, H. Prade, Representations of uncertainty in
artificial intelligence: Probability and possibility, in: A Guided Tour of
Artificial Intelligence Research, Springer, 2020, pp. 69–117.
(27)
C. Lian, S. Ruan, T. Denœux, H. Li, P. Vera, Joint tumor segmentation in
pet-ct images using co-clustering and fusion based on belief functions, IEEE
Transactions on Image Processing 28 (2) (2018) 755–766.
(28)
F. Ghesu, B. Georgescu, A. Mansoor, Y. Yoo, E. Gibson, R. Vishwanath,
A. Balachandran, J. Balter, Y. Cao, R. Singh, Quantifying and leveraging
predictive uncertainty for medical image assessment, Medical Image Analysis
68 (2021) 101855.
(29)
L. Huang, S. Ruan, T. Denœux, Belief function-based semi-supervised
learning for brain tumor segmentation, in: 2021 IEEE 18th International
Symposium on Biomedical Imaging (ISBI), IEEE, 2021, pp. 160–164.
(30)
M. I. Razzak, S. Naz, A. Zaib, Deep learning for medical image processing:
Overview, challenges and the future, Classification in BioApps (2018)
323–350.
(31)
M. H. Hesamian, W. Jia, X. He, P. Kennedy, Deep learning techniques for medical
image segmentation: achievements and challenges, Journal of digital imaging
32 (4) (2019) 582–596.
(32)
X. Liu, L. Song, S. Liu, Y. Zhang, A review of deep-learning-based medical
image segmentation methods, Sustainability 13 (3) (2021) 1224.
(33)
N. Siddique, S. Paheding, C. P. Elkin, V. Devabhaktuni, U-net and its variants
for medical image segmentation: A review of theory and applications, Ieee
Access 9 (2021) 82031–82057.
(34)
T. Zhou, S. Ruan, S. Canu, A review: Deep learning for medical image
segmentation using multi-modality fusion, Array 3 (2019) 100004.
(35)
K. Ramesh, G. K. Kumar, K. Swapna, D. Datta, S. S. Rajest, A review of medical
image segmentation algorithms, EAI Endorsed Transactions on Pervasive Health
and Technology 7 (27) (2021) e6–e6.
(36)
D. G. Lowe, Object recognition from local scale-invariant features, in:
Proceedings of the seventh IEEE international conference on computer vision,
Vol. 2, Ieee, 1999, pp. 1150–1157.
(37)
E. Rosten, T. Drummond, Machine learning for high-speed corner detection, in:
European conference on computer vision, Springer, 2006, pp. 430–443.
(38)
A. S. Mian, M. Bennamoun, R. Owens, Three-dimensional model-based object
recognition and segmentation in cluttered scenes, IEEE transactions on
pattern analysis and machine intelligence 28 (10) (2006) 1584–1601.
(39)
J. Long, E. Shelhamer, T. Darrell, Fully convolutional networks for semantic
segmentation, in: Proceedings of the IEEE conference on computer vision and
pattern recognition, Boston, USA, Jun, 2015, pp. 3431–3440.
(40)
Ö. Çiçek, A. Abdulkadir, S. S. Lienkamp, T. Brox,
O. Ronneberger, 3D u-net: learning dense volumetric segmentation from
sparse annotation, in: International conference on medical image computing
and computer-assisted intervention, Springer, Athens, Greece, Oct, 2016, pp.
424–432.
(41)
X. Xiao, S. Lian, Z. Luo, S. Li, Weighted res-unet for high-quality retina
vessel segmentation, in: 9th International Conference on Information
Technology in Medicine and Education (ITME), IEEE, Hangzhou, China, Oct,
2018, pp. 327–331.
(42)
S. Guan, A. A. Khan, S. Sikdar, P. V. Chitnis, Fully dense unet for 2-d sparse
photoacoustic tomography artifact removal, IEEE journal of biomedical and
health informatics 24 (2) (2019) 568–576.
(43)
N. Ibtehaz, M.-S. Rahman, Multiresunet: Rethinking the u-net architecture for
multimodal biomedical image segmentation, Neural Networks 121 (2020) 74–87.
(44)
W. Zhu, Y. Huang, Z. Tang, H.and Qian, W. Du, N.and Fan, X. Xie, Anatomynet:
Deep 3D squeeze-and-excitation u-nets for fast and fully automated
whole-volume anatomical segmentation, bioRxiv (2018) 392969.
(45)
F. Milletari, N. Navab, S.-A. Ahmadi, V-net: Fully convolutional neural
networks for volumetric medical image segmentation, in: 2016 fourth
international conference on 3D vision, IEEE, 2016, pp. 565–571.
(46)
K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition,
in: Proceedings of the IEEE conference on computer vision and pattern
recognition, Las Vegas, USA, Jun, 2016, pp. 770–778.
(47)
O. Oktay, J. Schlemper, L. Folgoc, M. Lee, M.and Heinrich, K. Misawa, K. Mori,
S. McDonagh, N. Hammerla, B. Kainz, Attention u-net: Learning where to look
for the pancreas, arXiv preprint arXiv:1804.03999 (2018).
(48)
K. Trebing, T. Stanczyk, S. Mehrkanoon, Smaat-unet: Precipitation nowcasting
using a small attention-unet architecture, Pattern Recognition Letters 145
(2021) 178–186.
(49)
H. Cao, Y. Wang, J. Chen, D. Jiang, X. Zhang, Q. Tian, M. Wang, Swin-unet:
Unet-like pure transformer for medical image segmentation, arXiv preprint
arXiv:2105.05537 (2021).
(50)
A. Hatamizadeh, Y. Tang, V. Nath, D. Yang, A. Myronenko, B. Landman, H. R.
Roth, D. Xu, Unetr: Transformers for 3d medical image segmentation, in:
Proceedings of the IEEE/CVF Winter Conference on Applications of Computer
Vision, 2022, pp. 574–584.
(51)
A. Hatamizadeh, V. Nath, Y. Tang, D. Yang, H. Roth, D. Xu, Swin unetr: Swin
transformers for semantic segmentation of brain tumors in mri images, arXiv
preprint arXiv:2201.01266 (2022).
(52)
C. H. Sudre, W. Li, T. Vercauteren, S. Ourselin, M. Jorge Cardoso, Generalised
dice overlap as a deep learning loss function for highly unbalanced
segmentations, in: Deep learning in medical image analysis and multimodal
learning for clinical decision support, Springer, 2017, pp. 240–248.
(53)
S. S. M. Salehi, D. Erdogmus, A. Gholipour, Tversky loss function for image
segmentation using 3d fully convolutional deep networks, in: International
workshop on machine learning in medical imaging, Springer, 2017, pp.
379–387.
(54)
T. Chen, S. Kornblith, M. Norouzi, G. Hinton, A simple framework for
contrastive learning of visual representations, in: International conference
on machine learning, PMLR, 2020, pp. 1597–1607.
(55)
A. A. Taha, A. Hanbury, Metrics for evaluating 3d medical image segmentation:
analysis, selection, and tool, BMC medical imaging 15 (1) (2015) 1–28.
(56)
L. Sun, Y. Wang, A multi-attribute fusion approach extending
Dempster-Shafer theory for combinatorial-type evidences, Expert Systems
with Applications 96 (2018) 218–229.
(57)
P. Smets, The combination of evidence in the transferable belief model, IEEE
Transactions on pattern analysis and machine intelligence 12 (5) (1990)
447–458.
(58)
R.-R. Yager, On the Dempster-Shafer framework and new combination rules,
Information sciences 41 (2) (1987) 93–137.
(59)
D. Dubois, H. Prade, Representation and combination of uncertainty with belief
functions and possibility measures, Computational intelligence 4 (3) (1988)
244–264.
(60)
T. Denœux, Conjunctive and disjunctive combination of belief functions
induced by nondistinct bodies of evidence, Artificial Intelligence 172 (2-3)
(2008) 234–264.
(61)
F. Xiao, Generalization of dempster–shafer theory: A complex mass function,
Applied Intelligence 50 (10) (2020) 3266–3275.
(62)
T. Denœux, A k-nearest neighbor classification rule based on
Dempster-Shafer theory, IEEE Transactions on Systems, Man, and
Cybernetics 25 (5) (1995) 804–813.
doi:10.1109/21.376493.
(63)
T. Denœux, A neural network classifier based on Dempster-Shafer theory,
IEEE Transactions on Systems, Man, and Cybernetics-Part A: Systems and Humans
30 (2) (2000) 131–150.
(64)
F. Xiao, W. Pedrycz, Negation of the quantum mass function for multisource
quantum information fusion with its application to pattern classification,
IEEE Transactions on Pattern Analysis and Machine Intelligence (2022).
(65)
T. Denœux, M.-H. Masson, Evclus: evidential clustering of proximity data,
IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics)
34 (1) (2004) 95–109.
(66)
M.-H. Masson, T. Denœux, Recm: Relational evidential c-means algorithm,
Pattern Recognition Letters 30 (11) (2009) 1015–1026.
(67)
D. Dubois, H. Prade, Possibility theory: an approach to computerized processing
of uncertainty, Springer Science & Business Media, 2012.
(68)
T. Denóeux, 40 years of dempster-shafer theory, International Journal of
Approximate Reasoning 79 (C) (2016) 1–6.
(69)
D. Mercier, B. Quost, T. Denœux, Refined modeling of sensor reliability in
the belief function framework using contextual discounting, Information
fusion 9 (2) (2008) 246–258.
(70)
T. Denœux, O. Kanjanatarakul, S. Sriboonchitta, A new evidential k-nearest
neighbor rule based on contextual discounting with partially supervised
learning, International Journal of Approximate Reasoning 113 (2019) 287–302.
(71)
R. Yager, Decision making under Dempster-Shafer uncertainties,
International Journal of General System 20 (3) (1992) 233–245.
(72)
R. Yager, Decision making using minimization of regret, International Journal
of Approximate Reasoning 36 (2) (2004) 109–128.
doi:https://doi.org/10.1016/j.ijar.2003.10.003.
(73)
F. Xiao, J. Wen, W. Pedrycz, Generalized divergence-based decision making
method with an application to pattern classification, IEEE Transactions on
Knowledge and Data Engineering (2022).
(74)
T. Denœux, Decision-making with belief functions: A review, International
Journal of Approximate Reasoning 109 (2019) 87–110.
(75)
A. Appriou, Multisensor signal processing in the framework of the theory of
evidence, Application of mathematical signal processing techniques to mission
systems 32 (1999).
(76)
A. Appriou, 01-approche générique de la gestion de l’incertain dans
les processus de fusion multisenseur, Traitement du signal (2005).
(77)
L. Huang, S. Ruan, P. Decazes, T. Denoeux, Lymphoma segmentation from 3d
pet-ct images using a deep evidential network, International Journal of
Approximate Reasoning 149 (2022) 39–60.
doi:https://doi.org/10.1016/j.ijar.2022.06.007.
(78)
L. M. Zouhal, T. Denœux, An evidence-theoretic k-nn rule with parameter
optimization, IEEE Transactions on Systems, Man, and Cybernetics, Part C
(Applications and Reviews) 28 (2) (1998) 263–271.
(79)
T. Denœux, Logistic regression, neural networks and Dempster-Shafer
theory: A new perspective, Knowledge-Based Systems 176 (2019) 54–67.
(80)
J. Dunn, A fuzzy relative of the isodata process and its use in detecting
compact well-separated clusters, Journal of Cybernetics 3 (3) (1973) 32–57.
doi:10.1080/01969727308546046.
(81)
Y. Zhu, L. Bentabet, O. Dupuis, V. Kaftandjian, D. Babot, M. Rombaut, Automatic
determination of mass functions in Dempster-Shafer theory using fuzzy
C-means and spatial neighborhood information for image segmentation,
Optical Engineering 41 (4) (2002) 760–770.
(82)
J. Ghasemi, M. R. K. Mollaei, R. Ghaderi, A. Hojjatoleslami, Brain tissue
segmentation based on spatial information fusion by Dempster-Shafer
theory, Journal of Zhejiang University SCIENCE C 13 (7) (2012) 520–533.
(83)
R. Safranek, S. Gottschlich, A. Kak, Evidence accumulation using binary frames
of discernment for verification vision, IEEE Transactions on Robotics and
Automation 6 (4) (1990) 405–417.
(84)
M.-H. Masson, T. Denœux, Ecm: An evidential version of the fuzzy c-means
algorithm, Pattern Recognition 41 (4) (2008) 1384–1397.
(85)
X. Chen, D. Zou, Q. Zhao, P. Tan, Manifold preserving edit propagation, ACM
Transactions on Graphics (TOG) 31 (6) (2012) 1–7.
(86)
L. Huang, S. Ruan, P. Decazes, T. Denœux, Evidential segmentation of 3d
pet/ct images, in: International Conference on Belief Functions, Springer,
2021, pp. 159–167.
(87)
C. Lian, H. Li, P. Vera, S. Ruan, Unsupervised co-segmentation of tumor in
pet-ct images using belief functions based fusion, in: 2018 Ieee 15th
International Symposium On Biomedical Imaging (Isbi 2018), IEEE, 2018, pp.
220–223.
(88)
T. Denœux, Nn-evclus: Neural network-based evidential clustering,
Information Sciences 572 (2021) 297–330.
(89)
D. Suh, R. Eisner, R. Mersereau, R. Pettigrew, Knowledge-based system for
boundary detection of four-dimensional cardiac magnetic resonance image
sequences, IEEE transactions on medical imaging 12 (1) (1993) 65–72.
(90)
G. Gerig, D. Welti, R. Guttmann, C., C. Colchester, A., G. Székely,
Exploring the discrimination power of the time domain for segmentation and
characterization of active lesions in serial mr data, Medical Image Analysis
4 (1) (2000) 31–42.
(91)
S. Vauclin, P. Zhang, I. Gardin, O. O. Gallocher, P. Vannoorenberghe,
Segmentation of thoracic computed tomography images, in: International
Symposium on Signals, Circuits and Systems, Vol. 1, Iasi, Romania, Jul, 2005,
pp. 31–34.
doi:10.1109/ISSCS.2005.1509843.
(92)
P. Vannoorenberghe, G. Flouzat, A belief-based pixel labeling strategy for
medical and satellite image segmentation, in: 2006 IEEE International
Conference on Fuzzy Systems, IEEE, Vancouver, BC, Canada, Jul, 2006, pp.
1093–1098.
(93)
F. Derraz, M. Boussahla, L. Peyrodie, Globally segmentation using active
contours and belief function, in: International Conference on Advanced
Concepts for Intelligent Vision Systems, Springer, Poznań, Poland, Oct,
2013, pp. 546–554.
(94)
A. Derraz, F.and Pinti, M. Boussahla, L. Peyrodie, H. Toumi, Image segmentation
using active contours and evidential distance, in: Iberoamerican Congress on
Pattern Recognition, Springer, Havana, Cuba, Nov, 2013, pp. 472–479.
(95)
C. Lian, S. Ruan, T. Denœux, H. Li, P. Vera, Spatial evidential clustering
with adaptive distance metric for tumor segmentation in fdg-pet images, IEEE
Transactions on Biomedical Engineering 65 (1) (2017) 21–30.
(96)
C. Lian, T. Ruan, S.and Denœux, H. Li, P. Vera, Tumor delineation in
fdg-pet images using a new evidential clustering algorithm with spatial
regularization and adaptive distance metric, in: 2017 IEEE 14th International
Symposium on Biomedical Imaging (ISBI 2017), IEEE, Melbourne, Australia, Apr,
2017, pp. 1177–1180.
(97)
T. Chan, B. Sandberg, L. Vese, Active contours without edges for vector-valued
images, Journal of Visual Communication and Image Representation 11 (2)
(2000) 130–141.
(98)
T. Goldstein, S. Osher, The split bregman method for l1-regularized problems,
SIAM journal on imaging sciences 2 (2) (2009) 323–343.
(99)
M. Mohamed, B. Far, An enhanced threshold based technique for white blood cells
nuclei automatic segmentation, in: 14th International Conference on e-Health
Networking, Applications and Services (Healthcom), IEEE, Beijing, China, Oct,
2012, pp. 202–207.
(100)
O. Michailovich, Y. Rathi, A. Tannenbaum, Image segmentation using active
contours driven by the bhattacharyya gradient flow, IEEE Transactions on
Image Processing 16 (11) (2007) 2787–2801.
(101)
G. Quellec, M. Lamard, P. Josselin, G. Cazuguel, B. Cochener, C. Roux, Optimal
wavelet transform for the detection of microaneurysms in retina photographs,
IEEE transactions on medical imaging 27 (9) (2008) 1230–1241.
(102)
M. Niemeijer, B. Van Ginneken, M. Cree, A. Mizutani, G. Quellec,
C. Sánchez, B. Zhang, R. Hornero, M. Lamard, C. Muramatsu, Retinopathy
online challenge: automatic detection of microaneurysms in digital color
fundus photographs, IEEE transactions on medical imaging 29 (1) (2009)
185–195.
(103)
B. Lelandais, S. Ruan, T. Denœux, P. Vera, I. Gardin, Fusion of
multi-tracer pet images for dose painting, Medical image analysis 18 (7)
(2014) 1247–1259.
(104)
H. Peiris, M. Hayat, Z. Chen, G. Egan, M. Harandi, A volumetric transformer for
accurate 3d tumor segmentation, arXiv preprint arXiv:2111.13300 (2021).
(105)
T. Zhou, S. Canu, S. Ruan, Fusion based on attention mechanism and context
constraint for multi-modal brain tumor segmentation, Computerized Medical
Imaging and Graphics 86 (2020) 101811.
(106)
T. Zhou, S. Ruan, P. Vera, S. Canu, A tri-attention fusion guided multi-modal
segmentation network, Pattern Recognition 124 (2022) 108417.
(107)
K. Kamnitsas, W. Bai, E. Ferrante, S. McDonagh, M. Sinclair, N. Pawlowski,
M. Rajchl, M. Lee, B. Kainz, D. Rueckert, Ensembles of multiple models and
architectures for robust brain tumour segmentation, in: International MICCAI
Brainlesion Workshop, Springer, Cham, 2017, pp. 450–462.
(108)
P. Vannoorenberghe, O. Colot, D. De Brucq, Dempster-Shafer’s theory as an
aid to color information processing. application to melanoma detection in
dermatology, in: Proceedings 10th International Conference on Image Analysis
and Processing, IEEE, Venice, Italy, Sept, 1999, pp. 774–779.
(109)
I. Bloch, Some aspects of Dempster-Shafer evidence theory for
classification of multi-modality medical images taking partial volume effect
into account, Pattern Recognition Letters 17 (8) (1996) 905–919.
(110)
A. Taleb-Ahmed, L. Gautier, On information fusion to improve segmentation of
mri sequences, Information Fusion 3 (2) (2002) 103–117.
(111)
S. Chaabane, F. Fnaiech, M. Sayadi, E. Brassart, Relevance of the
Dempster-Shafer evidence theory for image segmentation, in: 3rd
International Conference on Signals, Circuits and Systems (SCS), IEEE,
Medenine, Tunisia, Nov, 2009, pp. 1–4.
(112)
I. Bloch, Defining belief functions using mathematical morphology–application
to image fusion under imprecision, International journal of approximate
reasoning 48 (2) (2008) 437–465.
(113)
S.-B. Chaabane, F. Sayadi, M.and Fnaiech, E. Brassart, F. Betin, A new method
for the estimation of mass functions in the Dempster-Shafer’s evidence
theory: application to colour image segmentation, Circuits, Systems, and
Signal Processing 30 (1) (2011) 55–71.
(114)
R. Harrabi, E. B. Braiek, Color image segmentation using multi-level
thresholding approach and data fusion techniques: application in the breast
cancer cells images, EURASIP Journal on Image and Video Processing 2012 (1)
(2012) 11.
(115)
B. Lelandais, I. Gardin, L. Mouchard, P. Vera, S. Ruan, Segmentation of
biological target volumes on multi-tracer pet images based on information
fusion for achieving dose painting in radiotherapy, in: International
Conference on Medical Image Computing and Computer-Assisted Intervention,
Springer, Nice, France, Oct, 2012, pp. 545–552.
(116)
R. Wang, X. Shen, Y. Li, Y. Zhu, C. Hui, S. Zhang, Lesion segmentation in acute
cerebral infarction based on Dempster-Shafer theory, in: International
Conference on Wavelet Analysis and Pattern Recognition, IEEE, Zurich,
Switzerland, Jul, 2013, pp. 209–214.
(117)
J. Ghasemi, R. Ghaderi, M.-K. Mollaei, S. Hojjatoleslami, A novel fuzzy
Dempster-Shafer inference system for brain mri segmentation, Information
Sciences 223 (2013) 205–220.
(118)
N. Makni, N. Betrouni, O. Colot, Introducing spatial neighbourhood in
evidential c-means for segmentation of multi-source images: Application to
prostate multi-parametric MRI, Information Fusion 19 (2014) 61–72.
(119)
F. Derraz, A. Pinti, L. Peyrodie, M. Bousahla, H. Toumi, Joint variational
segmentation of ct/pet data using non-local active contours and belief
functions, Pattern Recognition and Image Analysis 25 (3) (2015) 407–412.
(120)
O. Trabelsi, L. Tlig, M. Sayadi, F. Fnaiech, Skin lesion segmentation using the
ds evidence theory based on the fcm using feature parameters, in: 12th
International Multi-Conference on Systems, Signals & Devices (SSD15), IEEE,
Sfax, Tunisia, Nov, 2015, pp. 1–5.
(121)
R. Xiao, H. Ding, F. Zhai, T. Zhao, W. Zhou, G. Wang, Vascular segmentation of
head phase-contrast magnetic resonance angiograms using grayscale and shape
features, Computer Methods and Programs in Biomedicine 142 (2017) 157–166.
(122)
C. Lian, S. Ruan, T. Denœux, Y. Guo, P. Vera, Accurate tumor segmentation
in fdg-pet images with guidance of complementary ct images, in: International
Conference on Image Processing (ICIP), IEEE, Beijing, China, Sept, 2017, pp.
4447–4451.
(123)
F. Tavakoli, J. Ghasemi, Brain mri segmentation by combining different mri
modalities using Dempster-Shafer theory, IET Image Processing 12 (8)
(2018) 1322–1330.
(124)
S.-A. Lima, M.-R. Islam, A modified method for brain mri segmentation using
Dempster-Shafer theory, in: 22nd International Conference on Computer and
Information Technology, IEEE, Helsinki, Finland, Aug, 2019, pp. 1–6.
(125)
H. Bloch, I.and Maître, Fuzzy mathematical morphologies: a comparative
study, Pattern recognition 28 (9) (1995) 1341–1387.
(126)
J. C. Bezdek, R. Ehrlich, W. Full, Fcm: The fuzzy c-means clustering algorithm,
Computers & Geosciences 10 (2-3) (1984) 191–203.
(127)
A.-S. Capelle, O. Colot, C. Fernandez-Maloigne, Segmentation of multi-modality
mr images by means of evidence theory for 3d reconstruction of brain tumors,
in: Proceedings. International Conference on Image Processing, Vol. 2, IEEE,
New York, USA, Sept, 2002, pp. II–II.
(128)
A.-S. Capelle, O. Colot, C. Fernandez-Maloigne, Evidential segmentation scheme
of multi-echo mr images for the detection of brain tumors using neighborhood
information, Information Fusion 5 (3) (2004) 203–216.
(129)
W. Barhoumi, S. Dhahbi, E. Zagrouba, A collaborative system for pigmented skin
lesions malignancy tracking, in: 2007 IEEE International Workshop on Imaging
Systems and Techniques, IEEE, Cracovia, Poland, May, 2007, pp. 1–6.
(130)
Y.-H. Guan, Y. Luo, T. Yang, L. Qiu, J. Li, Study on the application of MRF
and fuzzy clustering as well as the DS theory to image fusion segmentation
of the human brain, in: 4th International Conference on Intelligent Networks
and Intelligent Systems, IEEE, Washington, USA, Nov, 2011, pp. 57–60.
(131)
H. Ketout, J. Gu, G. Horne, Improved Dempster and Shafer theory to fuse
region and edge based level set for endocardial contour detection, in:
Proceedings of the 10th World Congress on Intelligent Control and Automation,
IEEE, Beijing, China, Jul, 2012, pp. 5013–5018.
(132)
J. Wen, Y. Tian, S. Yehang, T. Yongchuan, H. Weiwei, Improved evidential fuzzy
c-means method, Journal of Systems Engineering and Electronics 29 (1) (2018)
187–195.
(133)
K. George, S. Faziludeen, P. Sankaran, Breast cancer detection from biopsy
images using nucleus guided transfer learning and belief based fusion,
Computers in Biology and Medicine 124 (2020) 103954.
(134)
S. Li, Markov random field modeling in image analysis, Springer Science &
Business Media, 2009.
(135)
R. Duan, Y.-H. Guan, Multi-threshold value segmentation approach for medical
images [j], Journal of Computer Applications 28 (S2) (2008) 196–197.
(136)
M. Kass, D. Witkin, A.and Terzopoulos, Snakes: Active contour models,
International journal of computer vision 1 (4) (1988) 321–331.
(137)
C. Li, C. Xu, C. Gui, M.-D. Fox, Level set evolution without re-initialization:
a new variational formulation, in: IEEE computer society conference on
computer vision and pattern recognition, Vol. 1, IEEE, San Diego, California,
Jun, 2005, pp. 430–436.
(138)
T.-F. Chan, L.-A. Vese, Active contours without edges, IEEE Transactions on
image processing 10 (2) (2001) 266–277.
(139)
G. Xin, Y. Xiao, H. You, An improved Dempster-Shafer algorithm for
resolving the conflicting evidences, International Journal of Information
Technology 11 (12) (2005) 68–75.
(140)
L. Gautier, A. Taleb-Ahmed, M. Rombaut, J.-G. Postaire, H. Leclet, Belief
function in low level data fusion: application in MRI images of vertebra,
in: Proceedings of the Third International Conference on Information Fusion,
Vol. 1, IEEE, Paris, France, Jul, 2000, pp. TUC5–3.
(141)
R. Lajili, K. Kalti, A. Touil, B. Solaiman, N. E. B. Amara, Two-step evidential
fusion approach for accurate breast region segmentation in mammograms, IET
Image Processing 12 (11) (2018) 1972–1982.
(142)
L. Huang, T. Denœux, D. Tonnelet, P. Decazes, S. Ruan, Deep pet/ct fusion
with dempster-shafer theory for lymphoma segmentation, in: International
Workshop on Machine Learning in Medical Imaging, Springer, 2021, pp. 30–39.
(143)
L. Huang, T. Denoeux, P. Vera, S. Ruan, Evidence fusion with contextual
discounting for multi-modality medical image segmentation, in: L. Wang,
Q. Dou, P. T. Fletcher, S. Speidel, S. Li (Eds.), Medical Image Computing and
Computer Assisted Intervention – MICCAI 2022, Springer Nature Switzerland,
Cham, 2022, pp. 401–411.
(144)
K. Lai, Deformable contours: Modeling, extraction, detection and
classification, Ph.D. thesis, University of Wisconsin–Madison (1994).
(145)
J. Bezdek, Pattern recognition with fuzzy objective function algorithms,
Springer Science & Business Media, 2013.
(146)
J. Suckling, The mammographic image analysis society digital mammogram
database, Digital Mammo (1994) 375–386.
(147)
K. Bowyer, D. Kopans, W. Kegelmeyer, R. Moore, M. Sallam, K. Chang, K. Woods,
The digital database for screening mammography, in: Third international
workshop on digital mammography, Vol. 58, Mammography, Chicago, Jun, 1996,
p. 27.
(148)
Y. Kwon, J.-H. Won, B. J. Kim, M. C. Paik, Uncertainty quantification using
bayesian neural networks in classification: Application to biomedical image
segmentation, Computational Statistics & Data Analysis 142 (2020) 106816.
(149)
A. Mehrtash, W. M. Wells, C. M. Tempany, P. Abolmaesumi, T. Kapur, Confidence
calibration and predictive uncertainty estimation for deep medical image
segmentation, IEEE transactions on medical imaging 39 (12) (2020) 3868–3878.
(150)
B. Ghoshal, A. Tucker, B. Sanghera, W. Lup Wong, Estimating uncertainty in deep
learning for reporting confidence to clinicians in medical image segmentation
and diseases detection, Computational Intelligence 37 (2) (2021) 701–734.
(151)
K. Zou, X. Yuan, X. Shen, M. Wang, H. Fu, Tbrats: Trusted brain tumor
segmentation, in: L. Wang, Q. Dou, P. T. Fletcher, S. Speidel, S. Li (Eds.),
Medical Image Computing and Computer Assisted Intervention – MICCAI 2022,
Springer Nature Switzerland, Cham, 2022, pp. 503–513.
(152)
U. Baid, S. Ghodasara, S. Mohan, M. Bilello, E. Calabrese, E. Colak,
K. Farahani, J. Kalpathy-Cramer, F. C. Kitamura, S. Pati, et al., The
rsna-asnr-miccai brats 2021 benchmark on brain tumor segmentation and
radiogenomic classification, arXiv preprint arXiv:2107.02314 (2021).
(153)
T. Denœux, NN-EVCLUS: Neural network-based evidential clustering,
Information Sciences 572 (2021) 297–330. |
Braided diagram groups and local similarity groups
Daniel S. Farley
Department of Mathematics and Statistics
Miami University
Oxford, OH 45056 U.S.A.
[email protected]
and
Bruce Hughes
Department of Mathematics
Vanderbilt University
Nashville, TN 37240 U.S.A.
[email protected]
(Date:: November 22, 2020)
Abstract.
Hughes defined a class of groups that act as local similarities on compact ultrametric spaces. Guba and Sapir had previously defined braided diagram groups over semigroup presentations. The two classes of groups share some common characteristics: both act properly by isometries on CAT(0) cubical complexes, and certain
groups in both classes have type $F_{\infty}$, for instance.
Here we clarify the relationship between these families of groups: the braided diagram groups over tree-like semigroup presentations are precisely
the groups that act on compact ultrametric spaces via small similarity structures. The proof can be considered a generalization of the proof that Thompson’s group $V$
is a braided diagram group over a tree-like semigroup presentation.
We also prove that certain additional groups, such as the Houghton groups
$H_{n}$, and $QAut(T_{2,c})$, lie in both classes.
The second-named author was supported in part by NSF Grant DMS–0504176.
Contents
1 Introduction
2 Braided diagram groups
3 Groups defined by finite similarity structures
4 Braided diagram groups and groups determined by finite similarity structures
2000 Math. Subject Class. Primary 20F65, 54E45
Contents
1 Introduction
2 Braided diagram groups
3 Groups defined by finite similarity structures
4 Braided diagram groups and groups determined by finite similarity structures
1. Introduction
In [7], Hughes described a class of groups that act as homeomorphisms on compact ultrametric spaces. Fix a compact ultrametric space $X$.
The essence of the idea
was to associate to $X$ a finite similarity structure, which is a function that associates to each ordered pair of balls $B_{1},B_{2}\subseteq X$ a finite set ${\rm Sim}_{X}(B_{1},B_{2})$ of surjective similarities from $B_{1}$ to $B_{2}$. (A similarity is a map that stretches or contracts distances by a fixed constant.) The finite
sets ${\rm Sim}_{X}(B_{1},B_{2})$ are assumed to have certain desirable closure properties (such as closure under composition). A homeomorphism $h:X\rightarrow X$ is said to be locally determined by ${\rm Sim}_{X}$ if each $x\in X$ has a ball neighborhood $B$ with the property that $h(B)$
is a ball and
the restriction of $h$ to $B$ agrees with one of the local similarities $\sigma\in{\rm Sim}_{X}(B,h(B))$. The collection of all homeomorphisms that are locally determined by
${\rm Sim}_{X}$ forms a group under composition. We will call such a group an FSS group (finite similarity structure group) for short. Hughes [7] proved
that each FSS group has the Haagerup property, and even acts properly on a CAT(0) cubical complex. In [5], the authors described a class of FSS groups that
have type $F_{\infty}$. That class includes Thompson’s group $V$, and the main theorem of [5] is best understood as a generalization of [2],
where Brown originally showed that $V$ has type $F_{\infty}$.
In earlier work, Guba and Sapir [6] had sketched a theory of braided diagram groups over semigroup presentations, and proved that Thompson’s group $V$
is a braided diagram group over the semigroup presentation $\langle x\mid x=x^{2}\rangle$. Farley [4] showed that braided diagram groups over semigroup presentations
act properly on CAT(0) cubical complexes.
The class $\mathcal{F}$ of FSS groups and the class $\mathcal{B}$ of braided diagram groups therefore have a common origin, as generalizations of Thompson’s group $V$. Both classes also share other features in common (as noted above). It is therefore natural to wonder to what extent the two classes are the same. The main goal of this note is to prove Theorem 4.12, which says that the FSS groups determined by small similarity structures (Definition 4.6)
are precisely the same as the braided diagram groups determined by tree-like semigroup presentations (Definition 4.1). It is even possible that Theorem 4.12 describes the precise
extent of the overlap between $\mathcal{F}$ and $\mathcal{B}$, but we do not know how to prove this.
We include all relevant definitions, and our treatment is fairly self-contained as a result. A precise definition of braided diagram groups is given in
Section 2, the precise definition of FSS groups appears in Section 3, and the main theorem is proved in Section
4. Along the way, we give additional examples in the class $\mathcal{F}\cap\mathcal{B}$, including the Houghton groups $H_{n}$ and
a certain group $QAut(T_{2,c})$ of quasi-automorphisms of the infinite binary tree. (These are Examples 4.3 and 4.4, respectively.)
This note has been adapted from the longer preprint [5]. The first part of the latter preprint (including roughly the first six sections) will be published elsewhere. The first author would like to thank the organizers of the Durham Symposium (August 2013) for the opportunity to speak. Example 4.3 first appeared as part of the
first author’s lecture. The idea of Example 4.4 occurred to the first author after listening to Collin Bleak’s lecture at the Symposium.
2. Braided diagram groups
In this section, we will recall the definition of braided diagram groups over semigroup presentations.
Note that the theory of braided diagram groups was first sketched by Guba and Sapir [6]. A more extended introduction to braided diagram groups
appears in [4].
Definition 2.1.
Let $\Sigma$ be a set, called an alphabet. The free semigroup on $\Sigma$, denoted $\Sigma^{+}$, is
the collection of all positive non-empty strings formed from $\Sigma$, i.e.,
$$\Sigma^{+}=\{u_{1}u_{2}\ldots u_{n}\mid n\in\mathbb{N},u_{i}\in\Sigma\text{ %
for }i\in\{1,\ldots,n\}\}.$$
The free monoid on $\Sigma$, denoted $\Sigma^{\ast}$, is the union $\Sigma^{+}\cup\{1\}$, where $1$ denotes the
empty string. (Here we assume that $1\not\in\Sigma$ to avoid ambiguity.) The operations in $\Sigma^{+}$ and $\Sigma^{\ast}$ are concatenation.
We write $w_{1}\equiv w_{2}$ if $w_{1}$ and $w_{2}$ are equal as words in $\Sigma^{\ast}$.
Definition 2.2.
A semigroup presentation $\mathcal{P}=\langle\Sigma\mid\mathcal{R}\rangle$ consists of an alphabet $\Sigma$ and
a set $\mathcal{R}\subseteq\Sigma^{+}\times\Sigma^{+}$. The elements of $\mathcal{R}$ are called relations.
Remark 2.3.
A relation $(w_{1},w_{2})\in\mathcal{R}$ can be viewed as an equality between the words $w_{1}$ and $w_{2}$.
We use ordered pairs to describe these equalities because we will occasionally want to make a distinction between the
left and right sides of a relation.
A semigroup presentation $\mathcal{P}$ determines a semigroup $S_{\mathcal{P}}$, just as a group presentation determines a group. We will, however,
make essentially no use of this semigroup $S_{\mathcal{P}}$. Our interest is in braided diagrams over $\mathcal{P}$ (see below).
Definition 2.4.
(Braided Semigroup Diagrams)
A frame is a homeomorphic copy of $\partial([0,1]^{2})=(\{0,1\}\times[0,1])\cup([0,1]\times\{0,1\})$. A frame
has a top side, $(0,1)\times\{1\}$, a bottom side, $(0,1)\times\{0\}$, and left and right sides,
$\{0\}\times[0,1]$ and $\{1\}\times[0,1]$, respectively. The top and bottom of a frame have obvious left to right orderings.
A transistor is a homeomorphic copy of $[0,1]^{2}$. A transistor has top, bottom, left, and right sides, just as a frame does. The
top and bottom of a transistor also have obvious left to right orderings.
A wire is a homeomorphic copy of $[0,1]$. Each wire has a bottom $0$ and a top $1$.
Let $\mathcal{P}=\langle\Sigma\mid\mathcal{R}\rangle$ be a semigroup presentation. Let $\mathcal{T}(\Delta)$ be a finite
(possibly empty) set of transistors. Let $\mathcal{W}(\Delta)$ be a finite, nonempty set of wires. We let $F(\Delta)=\partial([0,1]^{2})$ be
a frame. We let $\ell_{\Delta}:\mathcal{W}(\Delta)\rightarrow\Sigma$ be an arbitrary function, called the labelling function.
For each wire $W\in\mathcal{W}(\Delta)$, we choose a point $t(W)$ on the bottom of a transistor, or on the top of the frame, and a point $b(W)$
on the top of a transistor, or on the bottom of the frame. The points $t(W)$ and $b(W)$ are called the top and bottom contacts of $W$, respectively.
We attach the top of each wire $W$ to $t(W)$ and the bottom of $W$ to $b(W)$. The resulting topological space $\Delta$ is called a braided diagram over
$\mathcal{P}$ if the following additional conditions are satisfied:
(1)
If $W_{i}$, $W_{j}\in\mathcal{W}(\Delta)$, $t(W_{i})=t(W_{j})$ only if $W_{i}=W_{j}$, and $b(W_{i})=b(W_{j})$ only if $W_{i}=W_{j}$.
In other words, the disjoint union of all of the wires maps injectively into the quotient.
(2)
We consider the top of some transistor $T\in\mathcal{T}(\Delta)$. Reading from left to right, we find contacts
$$b(W_{i_{1}}),b(W_{i_{2}}),\ldots,b(W_{i_{n}}),$$
where $n\geq 0$. The word $\ell_{t}(T)=\ell(W_{i_{1}})\ell(W_{i_{2}})\ldots\ell(W_{i_{n}})$ is called the top label of $T$.
Similarly, reading from left to right along the bottom of $T$, we find contacts
$$t(W_{j_{1}}),t(W_{j_{2}}),\ldots,t(W_{j_{m}}),$$
where $m\geq 0$. The word $\ell_{b}(T)=\ell(W_{j_{1}})\ell(W_{j_{2}})\ldots\ell(W_{j_{m}})$ is called the bottom label of $T$.
We require that, for any $T\in\mathcal{T}(\Delta)$, either $(\ell_{t}(T),\ell_{b}(T))\in\mathcal{R}$ or $(\ell_{b}(T),\ell_{t}(T))\in\mathcal{R}$.
(We emphasize that it is not sufficient for $\ell_{t}(T)$ to be equivalent to $\ell_{b}(T)$ modulo the relation $\sim$ determined by $\mathcal{R}$. Note also that this condition implies that
$\ell_{b}(T)$ and $\ell_{t}(T)$ are both non-empty, since $\mathcal{P}$ is a semigroup presentation. In particular, each transistor has wires attached to its top and bottom faces.)
(3)
We define a relation $\preceq$ on $\mathcal{T}(\Delta)$ as follows. Write $T_{1}\preceq T_{2}$ if there is some wire $W$ such that
$t(W)\in T_{2}$ and $b(W)\in T_{1}$. We require that the transitive closure $\dot{\preceq}$ of $\preceq$ be a strict partial order on $\mathcal{T}(\Delta)$.
Definition 2.5.
Let $\Delta$ be a braided diagram over $\mathcal{P}$. Reading from left to right across the top of the frame $F(\Delta)$,
we find contacts
$$t(W_{i_{1}}),t(W_{i_{2}}),\ldots,t(W_{i_{n}}),$$
for some $n\geq 1$. The word $\ell(W_{i_{1}})\ell(W_{i_{2}})\ldots\ell(W_{i_{n}})=\ell_{t}(\Delta)$
is called the top label of $\Delta$. We can similarly define the bottom label of $\Delta$, $\ell_{b}(\Delta)$.
We say that $\Delta$ is a braided $(\ell_{t}(\Delta),\ell_{b}(\Delta))$-diagram over $\mathcal{P}$.
Remark 2.6.
One should note that braided diagrams, despite the name, are not truly braided. In fact, two braided diagrams are equivalent
(see Definition 2.10) if there is a certain type of marked homeomorphism between them. Equivalence therefore does not depend
on any embedding into a larger space. Braided diagram groups (as defined in Theorem 2.13) also seem to have little in common
with Artin’s braid groups.
Example 2.7.
Let $\mathcal{P}=\langle a,b,c\mid ab=ba,ac=ca,bc=cb\rangle$. Figure 1 shows an example of a braided $(aabc,acba)$-diagram
over the semigroup presentation $\mathcal{P}$. The frame is the box formed by the dashed line. The wires that appear to cross in the figure do not
really touch, and it is unnecessary to specify which wire passes over the other one. See Remark 2.6.
Definition 2.8.
(Concatenation of braided diagrams)
Let $\Delta_{1}$ and $\Delta_{2}$ be braided diagrams over $\mathcal{P}$. We suppose that $\Delta_{1}$
is a $(w_{1},w_{2})$-diagram and $\Delta_{2}$ is a $(w_{2},w_{3})$-diagram. We can multiply $\Delta_{1}$ and $\Delta_{2}$ by stacking them. More explicitly,
we remove the bottom of the frame of $\Delta_{1}$ and the top of the frame of $\Delta_{2}$, and then glue together the wires in order from left to right. This gluing is compatible with the labeling of the wires, since the bottom label of $\Delta_{1}$ is the same as the top label of $\Delta_{2}$. The result
is a braided diagram $\Delta_{1}\circ\Delta_{2}$, called the concatenation of $\Delta_{1}$ and $\Delta_{2}$.
Definition 2.9.
(Dipoles)
Let $\Delta$ be a braided semigroup diagram over $\mathcal{P}$. We say that the transistors $T_{1},T_{2}\in\mathcal{T}(\Delta)$, $T_{1}\,\preceq\,T_{2}$,
form a dipole if:
(1)
the bottom label of $T_{1}$ is the same as the top label of $T_{2}$, and
(2)
there are wires $W_{i_{1}},W_{i_{2}},\ldots,W_{i_{n}}(n\geq 1)$ such that the bottom contacts $T_{2}$, read from left to right, are
precisely
$$t(W_{i_{1}}),t(W_{i_{2}}),\ldots,t(W_{i_{n}})$$
and the top contacts of $T_{1}$, read from left to right, are precisely
$$b(W_{i_{1}}),b(W_{i_{2}}),\ldots,b(W_{i_{n}}).$$
Define a new braided diagram as follows. Remove the transistors $T_{1}$ and $T_{2}$ and all of the wires $W_{i_{1}},\ldots,W_{i_{n}}$ connecting the top
of $T_{1}$ to the bottom of $T_{2}$. Let $W_{j_{1}},\ldots,W_{j_{m}}$ be the wires attached (in that order)
to the top of $T_{2}$, and let $W_{k_{1}},\ldots,W_{k_{m}}$ be the wires attached to the bottom of $T_{1}$. We glue the bottom of
$W_{j_{\ell}}$ to the top of $W_{k_{\ell}}$. There is a natural well-defined labelling function on the resulting wires, since
$\ell(W_{j_{\ell}})=\ell(W_{k_{\ell}})$ by our assumptions. We say that the new diagram $\Delta^{\prime}$ is obtained from $\Delta$ by reducing the dipole
$(T_{1},T_{2})$. The inverse operation is called inserting a dipole.
Definition 2.10.
(Equivalent Diagrams)
We say that two diagrams $\Delta_{1}$, $\Delta_{2}$ are equivalent if there is a homeomorphism
$\phi:\Delta_{1}\rightarrow\Delta_{2}$ that preserves the labels on the wires, restricts to a homeomorphism
$\phi_{|}:F(\Delta_{1})\rightarrow F(\Delta_{2})$, preserves the tops and bottoms of the transistors and frame, and preserves
the left to right orientations on the transistors and the frame. We write $\Delta_{1}\equiv\Delta_{2}$.
Definition 2.11.
(Equivalent Modulo Dipoles; Reduced Diagram)
We say that $\Delta$ and $\Delta^{\prime}$ are equivalent modulo dipoles
if there is a sequence $\Delta\equiv\Delta_{1}\equiv\Delta_{2}\equiv\ldots\equiv\Delta_{n}\equiv\Delta%
^{\prime}$,
where $\Delta_{i+1}$ is obtained from $\Delta_{i}$ by either inserting or removing a dipole, for $i\in\{1,\ldots,n-1\}$. We write $\Delta=\Delta^{\prime}$.
(The relation of equivalence modulo dipoles is indeed an equivalence relation – see [4].)
A braided diagram $\Delta$ over a semigroup presentation is called reduced if it contains no dipoles. Each equivalence class modulo dipoles
contains a unique reduced diagram [4].
Example 2.12.
In Figure 2, we have two braided diagrams over the semigroup presentation $\mathcal{P}=\langle a,b,c\mid ab=ba,ac=ca,bc=cb\rangle$.
The two rightmost transistors in the diagram on the left form a dipole, and the diagram on the right is the result of reducing that dipole.
Theorem 2.13.
[4]
Let $\mathcal{P}=\langle\Sigma\mid\mathcal{R}\rangle$ be a semigroup presentation, and let $w\in\Sigma^{+}$. We let $D_{b}(\mathcal{P},w)$ denote the set of equivalence classes of braided $(w,w)$-diagrams modulo dipoles. The operation of concatenation induces a well-defined group operation on $D_{b}(\mathcal{P},w)$.
This group $D_{b}(\mathcal{P},w)$ is called the braided diagram group over $\mathcal{P}$ based at $w$.
3. Groups defined by finite similarity structures
3.1. Review of ultrametric spaces and finite similarity structures
We now give a quick review of finite similarity structures on compact ultrametric spaces, as defined in Hughes [7]. Most of this subsection is taken directly from
[5].
Definition 3.1.
An ultrametric space is a metric space $(X,d)$ such that $d(x,y)\leq\max\{d(x,z),d(z,y)\}$ for all
$x,y,z\in X$.
Lemma 3.2.
Let $X$ be an ultrametric space.
(1)
Let $B_{r}(x)$ be an open metric ball in $X$. If $y\in B_{r}(x)$, then $B_{r}(x)=B_{r}(y)$.
(2)
If $B_{1}$ and $B_{2}$ are open metric balls in $X$, then either the balls are disjoint, or one is contained in the other.
(3)
Every open ball in $X$ is a closed set, and every closed ball in $X$ is an open set.
(4)
If $X$ is compact, then each open ball $B$ is contained in at most finitely many distinct open balls of $X$.
(5)
If $X$ is compact, then each open ball in $X$ is a closed ball (possibly of a different radius), and each closed ball is an open ball.
(6)
If $X$ is compact and $x$ is not an isolated point, then each open ball $B_{r}(x)$ is partitioned by its maximal proper open subballs, which are finite in number.
∎
Convention 3.3.
We assume for the rest of the section that $X$ is a compact ultrametric space. By Lemma 3.2(5), open balls are closed balls, and closed balls are open balls,
so we can refer to both simply as balls, and we will follow this practice from now on.
Definition 3.4.
Let $f:X\rightarrow Y$ be a function between metric spaces. We say that $f$ is a similarity if there is a constant $C>0$ such that
$d_{Y}(f(x_{1}),f(x_{2}))=Cd_{X}(x_{1},x_{2})$, for all $x_{1}$ and $x_{2}$ in $X$.
Definition 3.5.
A
finite similarity structure for $X$ is a function
${\rm Sim}_{X}$ that assigns to
each ordered pair $B_{1},B_{2}$ of balls in $X$
a (possibly empty)
set ${\rm Sim}_{X}(B_{1},B_{2})$ of surjective similarities
$B_{1}\to B_{2}$ such that whenever
$B_{1},B_{2},B_{3}$ are balls in $X$, the following properties
hold:
(1)
(Finiteness) ${\rm Sim}_{X}(B_{1},B_{2})$ is a finite set.
(2)
(Identities) $\mathrm{id}_{B_{1}}\in{\rm Sim}_{X}(B_{1},B_{1})$.
(3)
(Inverses) If $h\in{\rm Sim}_{X}(B_{1},B_{2})$, then $h^{-1}\in{\rm Sim}_{X}(B_{2},B_{1})$.
(4)
(Compositions)
If $h_{1}\in{\rm Sim}_{X}(B_{1},B_{2})$ and $h_{2}\in{\rm Sim}_{X}(B_{2},B_{3})$, then
$h_{2}h_{1}\in{\rm Sim}_{X}(B_{1},B_{3})$.
(5)
(Restrictions)
If $h\in{\rm Sim}_{X}(B_{1},B_{2})$ and $B_{3}\subseteq B_{1}$,
then
$$h|B_{3}\in{\rm Sim}_{X}(B_{3},h(B_{3})).$$
Definition 3.6.
A homeomorphism $h\colon X\to X$ is locally determined by ${\rm Sim}_{X}$ provided that for every $x\in X$, there exists a ball
$B^{\prime}$ in $X$ such that $x\in B^{\prime}$, $h(B^{\prime})$ is a ball in $X$, and $h|B^{\prime}\in{\rm Sim}(B^{\prime},h(B^{\prime}))$.
Definition 3.7.
The finite similarity structure
(FSS) group
$\Gamma({\rm Sim}_{X})$ is the set of all homeomorphisms $h\colon X\to X$ such that $h$ is locally determined by
${\rm Sim}_{X}$.
Remark 3.8.
The fact that $\Gamma({\rm Sim}_{X})$ is a group under composition is due to Hughes [7].
3.2. A description of the homeomorphisms determined by a similarity structure
In this subsection, we offer a somewhat simpler description
of the elements in the groups $\Gamma({\rm Sim}_{X})$ (Proposition 3.11), which shows that elements $\gamma\in\Gamma({\rm Sim}_{X})$ can be described in a manner reminiscent of the tree pair representatives for elements in Thompson’s group $V$ (see [3]).
Definition 3.9.
We define the standard partitions of $X$ inductively as follows.
(1)
$\{X\}$ is a standard partition.
(2)
If $\mathcal{P}=\{\widehat{B}_{1},\ldots,\widehat{B}_{n}\}$ is a standard partition, and $\{B_{1},\ldots,B_{m}\}$ is the partition of $\widehat{B}_{i}$ into maximal proper subballs, then
$(\mathcal{P}-\{\widehat{B}_{i}\})\cup\{B_{1},\ldots,B_{m}\}$ is also a standard partition.
Clearly, each standard partition is a partition of $X$ into balls.
Lemma 3.10.
Every partition $\mathcal{P}$ of $X$ into balls is standard.
Proof.
We prove this by induction on $|\mathcal{P}|$. It is clearly true if $|\mathcal{P}|=1$. We note that compactness implies that each
partition $\mathcal{P}$ of $X$ into balls must be finite.
For an arbitrary ball $B\subseteq X$, we define the depth of $B$, denoted $d(B)$, to be the number of distinct balls
of $X$ that contain $B$. (This definition is similar to Definition 3.19 from [5].) We note that $d(B)$ is a positive integer by Lemma 3.2(4), and $d(X)=1$.
Now we suppose that a partition $\mathcal{P}$ is given to us. We assume inductively that all partitions with smaller numbers of
balls are standard. By finiteness of $\mathcal{P}$,
there is some ball $B$ having maximum depth $m$, where we can assume that $m\geq 2$. Let $\widehat{B}$ denote the ball
containing $B$ as a maximal proper subball. Clearly, $d(\widehat{B})=m-1$. We let $\{B_{0},\ldots,B_{k}\}$ be the collection
of maximal proper subballs of $\widehat{B}$, where $B=B_{0}$ and $k\geq 1$.
We claim that $\{B_{0},B_{1},B_{2},\ldots,B_{k}\}\subseteq\mathcal{P}$. Choose $x\in B_{i}$. Our assumptions
imply that $x$ is in some ball $B^{\prime}$ of $\mathcal{P}$ such that $d(B^{\prime})\leq m$. The only such balls are $B_{i}$, $\widehat{B}$, and
any balls that contain $\widehat{B}$. (This uses an appeal to Lemma 3.2(2).) Since $\widehat{B}\cap B_{0}\neq\emptyset$ and $B_{0}=B\in\mathcal{P}$,
the only possibility is that $B^{\prime}=B_{i}$, since $\mathcal{P}$ is a partition. This proves that $\{B_{0},\ldots,B_{k}\}\subseteq\mathcal{P}$.
Now we consider the partition $\mathcal{P}^{\prime}=(\mathcal{P}-\{B_{0},\ldots,B_{k}\})\cup\{\widehat{B}\}$.
This partition is standard by the induction hypothesis, and it follows directly that $\mathcal{P}$ itself is standard.
∎
Proposition 3.11.
Let ${\rm Sim}_{X}$ be a finite similarity structure on $X$, and let $\gamma\in\Gamma({\rm Sim}_{X})$. There exist standard partitions $\mathcal{P}_{1}=\{B_{1},\ldots,B_{n}\}$ and $\mathcal{P}_{2}$ of $X$, a bijection $\phi:\mathcal{P}_{1}\rightarrow\mathcal{P}_{2}$, and elements $\sigma_{i}\in{\rm Sim}(B_{i},\phi(B_{i}))$
such that $\gamma_{\mid B_{i}}=\sigma_{i}$, for $i=1,\ldots,n$.
Moreover, we can arrange that the balls $B_{i}$ are maximal in the sense that if $B\subseteq X$ and $\gamma(B)$ are balls such
that $\gamma_{\mid B}\in{\rm Sim}_{X}(B,\gamma(B))$, then $B\subseteq B_{i}$, for some $i\in\{1,\ldots,n\}$.
Proof.
Since $\gamma$ is locally determined by ${\rm Sim}_{X}$, we can find an open cover of $X$ by balls such that the restriction of $\gamma$ to each ball is a local similarity in the
${\rm Sim}_{X}$-structure. By compactness of $X$, we can pass to a finite subcover. An application of Lemma 3.2(2) allows us to pass to a subcover that is also
a partition. We call this partition $\mathcal{P}_{1}$. We can then set $\mathcal{P}_{2}=\gamma(\mathcal{P}_{1})$. Both partitions are standard by Lemma 3.10.
The final statement is essentially Lemma 3.7 from [7].
∎
4. Braided diagram groups and groups determined by finite similarity structures
4.1. Braided diagram groups over tree-like semigroup presentations
Definition 4.1.
A semigroup presentation $\mathcal{P}=\langle\Sigma\mid\mathcal{R}\rangle$ is tree-like if,
(1)
every relation $(w_{1},w_{2})\in\mathcal{R}$ satisfies $|w_{1}|=1$ and $|w_{2}|>1$;
(2)
if $(a,w_{1}),(a,w_{2})\in\mathcal{R}$, then $w_{1}\equiv w_{2}$.
Example 4.2.
The generalized Thompson’s groups $V_{d}$ are isomorphic to the braided diagram groups $D_{b}(\mathcal{P},x)$,
where $\mathcal{P}=\langle x\mid(x,x^{d})\rangle$ is a tree-like semigroup presentation.
This fact was already proved in [6] and [4], and it is also a consequence of Theorem 4.12.
Example 4.3.
Consider the graph $G_{n}$ made up from a disjoint union of $n$ rays: $G_{n}=\{1,\ldots,n\}\times[0,\infty)$.
We assume that each ray is given the standard CW-complex structure with a vertex at
each integer. We define the Houghton group $H_{n}$ to be the set of bijections $h$ of the vertices $G_{n}^{0}$ such that
(1)
$h$ preserves adjacency, with at most finitely many exceptions, and
(2)
$h$ preserves ends: that is, for each $i\in\{1,\ldots,n\}$, there are $i_{1},i_{2}\in\mathbb{N}$ such that $h(\{i\}\times[i_{1},\infty))=\{i\}\times[i_{2},\infty)$.
Ken Brown [2] showed that $H_{n}$ is a group of type $F_{n-1}$ but not of type $F_{n}$.
We will sketch a proof that each Houghton group $H_{n}$ is a braided diagram group over a tree-like semigroup presentation. (It follows, in particular, that each of these
groups is $\Gamma({\rm Sim}_{X})$, for an appropriate compact ultrametric space $X$ and finite similarity structure ${\rm Sim}_{X}$, by Theorem 4.12.)
For $n\geq 2$, consider the semigroup presentation
$$\mathcal{P}_{n}=\langle a,r,x_{1},\ldots,x_{n}\mid(r,x_{1}x_{2}x_{3}\ldots x_{%
n}),(x_{1},ax_{1}),(x_{2},ax_{2}),\ldots,(x_{n},ax_{n})\rangle.$$
Similarly, we can define $\mathcal{P}_{1}=\langle a,r\mid(r,ar)\rangle$.
We claim that $D_{b}(\mathcal{P}_{n},r)$ is isomorphic to $H_{n}$. We sketch the proof for $n\geq 2$; the proof in case $n=1$ is very similar.
The elements of $D_{b}(\mathcal{P}_{n},r)$ can be expressed in the form $\Delta_{2}\circ\Delta_{1}^{-1}$, where each transistor in $\Delta_{i}$ is “positive”; i.e., the top label of the transistor is the left side of a relation in $\mathcal{P}_{n}$, and the bottom label is the right side. (This is proved as part of the proof of Theorem 4.12.)
We can think of each diagram $\Delta_{i}$ as a recipe for separating $G_{n}$ into connected components. The wires running between $\Delta_{1}$ and $\Delta_{2}$
in the concatenation $\Delta_{2}\circ\Delta_{1}^{-1}$ describe how these connected components should be matched by the bijection $h\in H_{n}$. To put it more explicitly,
the relations represent the following operations:
(1)
the relation $(r,x_{1}x_{2}\ldots x_{n})$ describes the initial configuration $G_{n}$ of $n$ disjoint rays. The letters $x_{i}$ ($i\in\{1,\ldots,n\}$) represent the isomorphism types of the different rays.
The different subscripts prevent different ends of $G_{n}$ from being permuted nontrivially. (If we wish to remove the end-preserving condition above, we can simply replace the
$n$ distinct symbols $x_{1}$, $x_{2}$, $\ldots$, $x_{n}$ by the single symbol $x$.)
(2)
the relation $(x_{i},ax_{i})$ (for $i\in\{1,\ldots,n\}$) represents the action of breaking the initial vertex away from the ray of isomorphism type $x_{i}$. The initial vertex of the ray gets the label $a$, and the new ray retains the label $x_{i}$ (since it is of the same combinatorial type as the original ray, and the new ray is a permissible target for the original ray
under the action of the Houghton group). The letter $a$ thus represents a single floating vertex. Any two such vertices can be matched by an element of the Houghton group, which
is why we use a single label for all of these vertices.
We illustrate how an $(r,r)$-diagram over $\mathcal{P}_{2}$ represents an element of $H_{2}$. Figure 3 depicts an $(r,r)$-diagram $\Delta_{h}$ over $\mathcal{P}_{2}$.
This diagram $\Delta_{h}$ represents the element $h\in H_{2}$ that sends: $(2,n)$ to $(2,n-1)$, for each $n\geq 1$, $(1,n)$ to $(1,n+1)$ (for all $n$), and $(2,0)$ to $(1,0)$.
Note that the bottom portion of the diagram represents a subdivision of the domain, and the top portion represents a subdivision of the range.
It is straightforward to check that the indicated function does not change if we insert or remove dipoles.
Example 4.4.
The same principle can be used to exhibit the group $\Gamma=\mathrm{QAut}(\mathcal{T}_{2,c})$ as a braided diagram group over a tree-like
semigroup presentation. Here $\Gamma$ (as defined in [1]) is the group of self-bijections $h$ of the vertices of the infinite ordered rooted binary tree $T$ such that
(1)
$h$ preserves adjacency, with at most finitely many exceptions, and
(2)
$h$ preserves the left-right ordering of the edges incident with and below a given vertex, again with at most finitely many exceptions.
Consider the semigroup presentation $\mathcal{P}=\langle a,x\mid(x,xax)\rangle$. We claim that $D_{b}(\mathcal{P},x)$ is isomorphic to $\Gamma$. Much of the discussion from the previous example carries over identically.
We will simply indicate how the single relation allows us to simulate breaking the binary tree into pieces. (Such a dissection of $T$ would be represented by
a positive diagram, as above. The wires connecting the bottoms of the positive diagrams $\Delta_{1}$ and $\Delta_{2}$ would again represent how the resulting pieces are matched by
a bijection.)
The letter $x$ represents the isomorphism type of the binary tree $T$. The relation represents breaking the tree at the root. The result of this operation yields a floating vertex
(represented by the letter $a$), and two new rooted binary trees (both represented by $x$). The first $x$ in $xax$ represents the left branch, and the second $x$ represents the
right. The description of the isomorphism of $D_{b}(\mathcal{P},x)$ with $\Gamma$ now follows the general pattern of the previous example.
Remark 4.5.
Example 4.3 shows that the $F_{\infty}$ result of [5] cannot be extended to all groups determined by finite similarity structures (as defined in Section 3).
All of the groups in the above examples act properly on CAT(0) cubical complexes by a construction of [4].
We note also that the representation of the above groups as braided diagram groups suggests a method for producing embeddings into other groups, such as (perhaps most notably) Thompson’s
group $V$. For instance, the group from Example 4.4 can be embedded into $V$ as follows. Given a braided diagram $\Delta$ over $\mathcal{P}$, systematically
replace each $a$ label with an $x$. The result is a braided diagram over the semigroup presentation $\mathcal{P}^{\prime}=\langle x\mid(x,x^{3})\rangle$. The indicated function
$\phi:D_{b}(\mathcal{P},x)\rightarrow D_{b}(\mathcal{P}^{\prime},x)$ is
easily seen to be a homomorphism, and $\phi$ is injective since it sends reduced diagrams to reduced diagrams. We can now appeal to the fact that
$D_{b}(\mathcal{P}^{\prime},x)\cong V_{3}$ (the $3$-ary version of Thompson’s group $V$), and the latter group embeds in $V$ itself.
The group from Example 4.4 was previously known to embed in $V$ by a result of [1].
4.2. Groups determined by small similarity structures
Definition 4.6.
Let $X$ be a compact ultrametric space. We say that the finite similarity structure ${\rm Sim}_{X}$ is small
if, for every pair of balls $B_{1},B_{2}$ in $X$, $|{\rm Sim}_{X}(B_{1},B_{2})|\leq 1$.
Definition 4.7.
Let $X$ be a compact ultrametric space endowed with a small similarity structure ${\rm Sim}_{X}$. If $B\subseteq X$ is a ball in
$X$ that is not an isolated point, then a local ball order at $B$ is an assignment of a linear order $<$ to the set $\{\widehat{B}_{1},\ldots,\widehat{B}_{n}\}$
of maximal proper subballs of $B$. A ball order on $X$ is an assignment of such a linear order to each ball $B\subseteq X$ that is not a singleton. The ball order is
compatible with ${\rm Sim}_{X}$ if each $h\in{\rm Sim}_{X}(B_{1},B_{2})$ induces an order-preserving bijection of the maximal proper subballs of $B_{1}$ and $B_{2}$,
for all choices of $B_{1}$ and $B_{2}$.
Lemma 4.8.
Let $X$ be a compact ultrametric space endowed with a small similarity structure. There exists a ball order on $X$ that is compatible with ${\rm Sim}_{X}$.
Proof.
We recall a definition from [5].
Let $B\subseteq X$ be a metric ball. Let
$[B]=\{B^{\prime}\subseteq X\mid{\rm Sim}_{X}(B,B^{\prime})\neq\emptyset\}$; $[B]$ is called the ${\rm Sim}_{X}$-class of $B$.
From a given ${\rm Sim}_{X}$-class of balls, choose a ball $B$. If $B$ is not a singleton, then there exists a collection of maximal proper subballs
$B_{1}$, $\ldots$, $B_{n}$ of $B$. Choose a linear order on this collection of balls; without loss of generality, $B_{1}<B_{2}<\ldots<B_{n}$. If $B^{\prime}$ is another ball
in $[B]$, then we can let $h$ denote the unique element of ${\rm Sim}_{X}(B,B^{\prime})$. This $h$ carries the maximal proper subballs of $B$ into maximal proper subballs of $B^{\prime}$, and thereby
induces an order $h(B_{1})<h(B_{2})<\ldots<h(B_{n})$ on the maximal proper subballs of $B^{\prime}$. This procedure gives a local ball order to each ball $B^{\prime}\in[B]$.
We repeat this procedure for each ${\rm Sim}_{X}$-class of balls. The result is a ball order on $X$ that is compatible with ${\rm Sim}_{X}$.
∎
Remark 4.9.
A ball order on $X$ also determines a linear order on any given collection of pairwise disjoint balls in $X$. For let $\mathcal{C}$
be such a collection, and let $B_{1},B_{2}\in\mathcal{C}$. There is a unique smallest ball $B\subseteq X$ that contains
both $B_{1}$ and $B_{2}$, by Lemma 3.2(4). Let $\{\widehat{B}_{1},\ldots,\widehat{B}_{n}\}$
be the collection of maximal proper subballs of $B$. By minimality of $B$,
we must have that $B_{1}$ and $B_{2}$ are contained in distinct
maximal proper subballs of $B$; say $B_{1}\subseteq\widehat{B}_{1}$ and $B_{2}\subseteq\widehat{B}_{2}$. We write
$B_{1}<B_{2}$ if and only if $\widehat{B}_{1}<\widehat{B}_{2}$. This defines a linear order on $\mathcal{C}$.
The verification is straightforward.
Definition 4.10.
Let $X$ be a compact ultrametric space with a small similarity structure ${\rm Sim}_{X}$ and a compatible ball order.
Define a semigroup presentation $\mathcal{P}_{{\rm Sim}_{X}}=\langle\Sigma\mid\mathcal{R}\rangle$
as follows. Let
$$\Sigma=\{[B]\mid B\text{ is a ball in }X\}.$$
If $B\subseteq X$ is a ball, let $B_{1},\ldots,B_{n}$ be the maximal proper subballs
of $B$, listed in order. If $B$ is a point, then $n=0$. We set
$$\mathcal{R}=\{([B],[B_{1}][B_{2}]\ldots[B_{n}])\mid n\geq 1,B\text{ is a ball %
in }X\}.$$
Remark 4.11.
We note that $\mathcal{P}_{{\rm Sim}_{X}}$ will always be a tree-like semigroup presentation, for any choice of
compact ultrametric space $X$, small similarity structure ${\rm Sim}_{X}$, and compatible ball order.
4.3. The main theorem
Theorem 4.12.
If $X$ is a compact ultrametric space with a small similarity structure ${\rm Sim}_{X}$ and compatible ball order,
then
$$\Gamma({\rm Sim}_{X})\cong D_{b}(\mathcal{P}_{{\rm Sim}_{X}},[X]).$$
Conversely, if $\mathcal{P}=\langle\Sigma\mid\mathcal{R}\rangle$ is a tree-like semigroup presentation,
and $x\in\Sigma$, then there is a compact ultrametric space $X_{\mathcal{P}}$, a small
finite similarity structure ${\rm Sim}_{X_{\mathcal{P}}}$, and a compatible ball order such that
$$D_{b}(\mathcal{P},x)\cong\Gamma({\rm Sim}_{X_{\mathcal{P}}}).$$
Proof.
If $\gamma\in\Gamma({\rm Sim}_{X})$, then, by Proposition 3.11, there are standard
partitions $\mathcal{P}_{1}$, $\mathcal{P}_{2}$ of $X$ into balls,
and a bijection $\phi:\mathcal{P}_{1}\rightarrow\mathcal{P}_{2}$ such that, for any $B\in\mathcal{P}_{1}$,
$\gamma(B)=\phi(B)$ and $\gamma|_{B}\in{\rm Sim}_{X}(B,\gamma(B))$.
Since $|{\rm Sim}_{X}(B,\gamma(B))|\leq 1$, the triple $(\mathcal{P}_{1},\mathcal{P}_{2},\phi)$ determines $\gamma$ without
ambiguity. We call $(\mathcal{P}_{1},\mathcal{P}_{2},\phi)$ a defining triple for $\gamma$. Note that a given $\gamma$
will usually have many defining triples. Let $\mathcal{D}$ be the set of all defining triples, for $\gamma$ running over all of
$\Gamma({\rm Sim}_{X})$.
We will now define a map $\psi:\mathcal{D}\rightarrow D_{b}(\mathcal{P}_{{\rm Sim}_{X}},[X])$. To a partition $\mathcal{P}$
of $X$ into balls, we first assign a braided diagram $\Delta_{\mathcal{P}}$ over $\mathcal{P}_{{\rm Sim}_{X}}$. There is a transistor
$T_{B}\in\mathcal{T}(\Delta_{\mathcal{P}})$ for each ball $B$ which properly contains some ball of $\mathcal{P}$. There is a wire
$W_{B}\in\mathcal{W}(\Delta_{\mathcal{P}})$ for each ball $B$ which contains a ball of $\mathcal{P}$. The wires are attached as follows:
(1)
If $B=X$, then we attach the top of $W_{B}$ to the top of the frame.
If $B\neq X$, then the top of the wire $W_{B}$ is attached to the bottom of the transistor $T_{\widehat{B}}$, where $\widehat{B}$ is the (unique) ball
that contains $B$ as a maximal proper subball.
Moreover, we attach the wires in an “order-respecting” fashion. Thus, if $\widehat{B}$ is a ball properly containing balls of $\mathcal{P}$, we let
$B_{1},B_{2},\ldots,B_{n}$ be the collection of maximal proper subballs of $\widehat{B}$, listed in order. We attach the wires $W_{B_{1}},W_{B_{2}},\ldots,W_{B_{n}}$ so that $t(W_{B_{i}})$ is to the left of $t(W_{B_{j}})$ on the bottom of $T_{\widehat{B}}$ if $i<j$.
(2)
The bottom of the wire $W_{B}$ is attached to the top of $T_{B}$ if $B$ properly contains a ball of $\mathcal{P}$. If not (i.e., if $B\in\mathcal{P}$),
then we attach the bottom of $W_{B}$ to the bottom of the frame. We can arrange, moreover, that the wires are attached in an order-respecting manner to the
bottom of the frame. (Thus, if $B_{1}<B_{2}$ ($B_{1},B_{2}\in\mathcal{P}$), we have that $b(W_{B_{1}})$ is to the left of $b(W_{B_{2}})$.)
The labelling function $\ell:\mathcal{W}(\Delta_{\mathcal{P}})\rightarrow\Sigma$ sends $W_{B}$ to $[B]$. It is straightforward to check that the
resulting $\Delta_{\mathcal{P}}$ is a braided diagram over $\mathcal{P}_{{\rm Sim}_{X}}$. The top label of $\Delta_{\mathcal{P}}$ is $[X]$.
Given a bijection $\phi:\mathcal{P}_{1}\rightarrow\mathcal{P}_{2}$, where $\mathcal{P}_{1}$ and $\mathcal{P}_{2}$ are partitions of $X$ into balls
and $[B]=[\phi(B)]$, we can define a braided diagram $\Delta_{\phi}$ over $\mathcal{P}_{{\rm Sim}_{X}}$ as follows. We let
$\mathcal{T}(\Delta_{\phi})=\emptyset$, and $\mathcal{W}(\Delta_{\phi})=\{W_{B}\mid B\in\mathcal{P}_{1}\}$.
We attach the top of each wire to the frame in such a way that $t(W_{B_{1}})$ is to the left of $t(W_{B_{2}})$ if $B_{1}<B_{2}$.
(Here $<$ refers to the ordering from Remark 4.9.) We attach the
bottom of each wire to the bottom of the frame in such a way that $b(W_{B_{1}})$ is to the left of $b(W_{B_{2}})$ if $\phi(B_{1})<\phi(B_{2})$.
Now, for a defining triple $(\mathcal{P}_{1},\mathcal{P}_{2},\phi)\in\mathcal{D}$, we set
$\psi((\mathcal{P}_{1},\mathcal{P}_{2},\phi))=\Delta_{\mathcal{P}_{2}}\circ%
\Delta_{\phi^{-1}}\circ\Delta_{\mathcal{P}_{1}}^{-1}\in\mathcal{D}_{b}(%
\mathcal{P}_{{\rm Sim}_{X}},[X])$.
We claim that any two defining triples $(\mathcal{P}_{1},\mathcal{P}_{2},\phi)$, $(\mathcal{P}_{1}^{\prime},\mathcal{P}_{2}^{\prime},\phi^{\prime})$
for a given $\gamma\in\Gamma({\rm Sim}_{X})$
have the same image
in $D_{b}(\mathcal{P}_{{\rm Sim}_{X}},[X])$, modulo dipoles. We begin by proving an intermediate statement. Let $(\mathcal{P}_{1},\mathcal{P}_{2},\phi)$
be a defining triple. Let $B\in\mathcal{P}_{1}$, and let $\widehat{B}_{1},\ldots,\widehat{B}_{n}$ be the collection of maximal proper subballs of $B$, listed
in order. We let $B^{\prime}=\phi(B)$ and let $\widehat{B}^{\prime}_{1},\ldots,\widehat{B}^{\prime}_{n}$ be the collection of maximal proper subballs of $B^{\prime}$. (Note that
$[B^{\prime}]=[B]$ by our assumptions, so both have the same number of maximal proper subballs, and in fact $[\widehat{B}_{i}]=[\widehat{B}^{\prime}_{i}]$ for $i=1,\ldots,n$,
since $\gamma|_{B}\in{\rm Sim}_{X}(B,B^{\prime})$ and the elements of ${\rm Sim}_{X}(B,B^{\prime})$ preserve order.) We set
$\widehat{\mathcal{P}}_{1}=(\mathcal{P}_{1}-\{B\})\cup\{\widehat{B}_{1},\ldots,%
\widehat{B}_{n}\}$,
$\widehat{\mathcal{P}}_{2}=(\mathcal{P}_{2}-\{B^{\prime}\})\cup\{\widehat{B}^{%
\prime}_{1},\ldots,\widehat{B}^{\prime}_{n}\}$,
and $\widehat{\phi}|_{\mathcal{P}_{1}-\{B\}}=\phi|_{\mathcal{P}_{1}-\{B\}}$,
$\widehat{\phi}(\widehat{B}_{i})=\widehat{B}^{\prime}_{i}$. We say that $(\widehat{\mathcal{P}}_{1},\widehat{\mathcal{P}}_{2},\widehat{\phi})$
is obtained from $(\mathcal{P}_{1},\mathcal{P}_{2},\phi)$ by subdivision at $(B,B^{\prime})$.
A straightforward argument shows that $\psi((\widehat{\mathcal{P}}_{1},\widehat{\mathcal{P}}_{2},\widehat{\phi}))$ is in fact obtained from
$\psi((\mathcal{P}_{1},\mathcal{P}_{2},\phi))$ by inserting a dipole. We omit the details, which rely on the fact that each
element of the ${\rm Sim}_{X}$-structure preserves the local ball order.
Now suppose that $(\mathcal{P}_{1},\mathcal{P}_{2},\phi)$ and $(\mathcal{P}^{\prime}_{1},\mathcal{P}^{\prime}_{2},\phi^{\prime})$ are defining triples
for the same element $\gamma\in\Gamma({\rm Sim}_{X})$. We can find a common refinement $\mathcal{P}^{\prime\prime}_{1}$ of $\mathcal{P}_{1}$ and $\mathcal{P}^{\prime}_{1}$.
Using the fact that all partitions of $X$ into balls are standard (Lemma
3.10),
we can pass from $(\mathcal{P}_{1},\mathcal{P}_{2},\phi)$ to $(\mathcal{P}^{\prime\prime}_{1},\widehat{P}_{2},\widehat{\phi})$ by repeated subdivision
(for some partition $\widehat{\mathcal{P}}_{2}$ of $X$ into balls and some bijection $\widehat{\phi}:\mathcal{P}^{\prime\prime}_{1}\rightarrow\widehat{\mathcal{P}}_%
{2}$).
Since subdivision does not change the values of $\psi$ modulo dipoles, $\psi((\mathcal{P}_{1},\mathcal{P}_{2},\phi))=\psi((\mathcal{P}^{\prime\prime}_%
{1},\widehat{\mathcal{P}}_{2},\widehat{\phi}))$ modulo dipoles. Similarly, we can subdivide
$(\mathcal{P}^{\prime}_{1},\mathcal{P}^{\prime}_{2},\phi^{\prime})$ repeatedly in order to obtain $(\mathcal{P}^{\prime\prime}_{1},\widehat{\mathcal{P}}^{\prime}_{2},\widehat{%
\phi}^{\prime})$, where
$\psi((\mathcal{P}^{\prime}_{1},\mathcal{P}^{\prime}_{2},\phi^{\prime}))=\psi((%
\mathcal{P}^{\prime\prime}_{1},\widehat{\mathcal{P}}^{\prime}_{2},\widehat{%
\phi}^{\prime}))$ modulo dipoles.
Both $(\mathcal{P}^{\prime\prime}_{1},\widehat{\mathcal{P}}^{\prime}_{2},\widehat{%
\phi}^{\prime})$
and $(\mathcal{P}^{\prime\prime}_{1},\widehat{\mathcal{P}}_{2},\widehat{\phi})$ are defining triples for $\gamma$, so we are forced to have
$\widehat{\phi}=\widehat{\phi}^{\prime}$ and $\widehat{\mathcal{P}}_{2}=\widehat{\mathcal{P}}^{\prime}_{2}$.
It follows that $\psi((\mathcal{P}_{1},\mathcal{P}_{2},\phi))=\psi((\mathcal{P}^{\prime}_{1},%
\mathcal{P}^{\prime}_{2},\phi^{\prime}))$, so $\psi$ induces a function
from $\Gamma({\rm Sim}_{X})$ to $D_{b}(\mathcal{P}_{{\rm Sim}_{X}},[X])$. We will call this function $\widehat{\psi}$.
Now we will show that $\widehat{\psi}:\Gamma({\rm Sim}_{X})\rightarrow D_{b}(\mathcal{P}_{{\rm Sim}_{%
X}},[X])$ is a homomorphism. Let $\gamma,\gamma^{\prime}\in\Gamma({\rm Sim}_{X})$. After subdividing as necessary, we
can choose defining triples $(\mathcal{P}_{1},\mathcal{P}_{2},\phi)$ and $(\mathcal{P}^{\prime}_{1},\mathcal{P}^{\prime}_{2},\phi^{\prime})$ for
$\gamma$ and $\gamma^{\prime}$ (respectively) in such a way that $\mathcal{P}_{2}=\mathcal{P}^{\prime}_{1}$. It follows easily that
$(\mathcal{P}_{1},\mathcal{P}^{\prime}_{2},\phi^{\prime}\phi)$ is a defining triple for $\gamma^{\prime}\gamma$. Therefore,
$\widehat{\psi}(\gamma^{\prime}\gamma)=\Delta_{\mathcal{P}^{\prime}_{2}}\circ%
\Delta_{(\phi^{\prime}\phi)^{-1}}\circ\Delta^{-1}_{\mathcal{P}_{1}}$.
Now
$$\displaystyle\widehat{\psi}(\gamma^{\prime})\circ\widehat{\psi}(\gamma)$$
$$\displaystyle=\Delta_{\mathcal{P}^{\prime}_{2}}\circ\Delta_{(\phi^{\prime})^{-%
1}}\circ\Delta^{-1}_{\mathcal{P}^{\prime}_{1}}\circ\Delta_{\mathcal{P}_{2}}%
\circ\Delta_{\phi^{-1}}\circ\Delta^{-1}_{\mathcal{P}_{1}}$$
$$\displaystyle=\Delta_{\mathcal{P}^{\prime}_{2}}\circ\Delta_{(\phi^{\prime})^{-%
1}}\circ\Delta_{\phi^{-1}}\circ\Delta^{-1}_{\mathcal{P}_{1}}$$
$$\displaystyle=\Delta_{\mathcal{P}^{\prime}_{2}}\circ\Delta_{(\phi^{\prime}\phi%
)^{-1}}\circ\Delta^{-1}_{\mathcal{P}_{1}}$$
Therefore, $\widehat{\psi}$ is a homomorphism.
We now show that $\widehat{\psi}:\Gamma({\rm Sim}_{X})\rightarrow D_{b}(\mathcal{P}_{{\rm Sim}_{%
X}},[X])$ is injective. Suppose that $\widehat{\psi}(\gamma)=1$.
Using the final statement of Proposition 3.11, we choose a defining triple $(\mathcal{P}_{1},\mathcal{P}_{2},\phi)$ for $\gamma$ with the property that, if $B\subseteq X$ is a ball, $\gamma(B)$ is a ball,
and $\gamma|_{B}\in{\rm Sim}_{X}(B,\gamma(B))$, then $B$ is contained in some ball of $\mathcal{P}_{1}$. We claim that
$\psi((\mathcal{P}_{1},\mathcal{P}_{2},\phi))$ is a reduced diagram. If there were a dipole $(T_{1},T_{2})$, then we would have
$T_{1}\in\mathcal{T}(\Delta_{\mathcal{P}_{1}}^{-1})$ and $T_{2}\in\mathcal{T}(\Delta_{\mathcal{P}_{2}})$, since it is impossible for $\Delta_{\mathcal{P}}$
to contain any dipoles, for any partition $\mathcal{P}$ of $X$ into balls. Thus $T_{1}=T_{B_{1}}$ and $T_{2}=T_{B_{2}}$, where $[B_{1}]=[B_{2}]$ and
the wires from the bottom of $T_{B_{2}}$ attach to the top of $T_{B_{1}}$, in order.
This means that, if $\widehat{B}_{1},\ldots,\widehat{B}_{n}$ are the maximal proper subballs of $B_{1}$,
and $\widehat{B}^{\prime}_{1},\ldots,\widehat{B}^{\prime}_{n}$ are the maximal proper subballs of $B_{2}$, then $\gamma(\widehat{B}_{i})=\widehat{B}^{\prime}_{i}$, where the latter
is a ball, and $\gamma|_{\widehat{B}_{i}}\in{\rm Sim}_{X}(\widehat{B}_{i},\widehat{B}^{\prime}%
_{i})$.
Now, since $[B_{1}]=[B_{2}]$, there is $h\in{\rm Sim}_{X}(B_{1},B_{2})$. Since ${\rm Sim}_{X}$ is closed under restrictions and $h$ preserves order,
we have $h_{i}\in{\rm Sim}_{X}(\widehat{B}_{i},\widehat{B}^{\prime}_{i})$ for $i=1,\ldots,n$, where $h_{i}=h|_{\widehat{B}_{i}}$. It follows that
$\gamma|_{\widehat{B}_{i}}=h_{i}$, so, in particular, $\gamma|_{B_{1}}=h$.
Since $B_{1}$ properly contains some ball in $\mathcal{P}_{1}$, this is a contradiction. Thus, $\psi((\mathcal{P}_{1},\mathcal{P}_{2},\phi))$
is reduced.
We claim that $\psi((\mathcal{P}_{1},\mathcal{P}_{2},\phi))$ contains no transistors (due to the condition $\widehat{\psi}(\gamma)=1$).
We have shown that $\psi((\mathcal{P}_{1},\mathcal{P}_{2},\phi))$ is a reduced diagram in the same class as the identity
$1\in D_{b}(\mathcal{P}_{{\rm Sim}_{X}},[X])$. The identity can be represented as the (unique) $([X],[X])$-diagram $\Delta_{1}$ with only a single wire,
$W_{X}$, and no transistors. We must have $\psi((\mathcal{P}_{1},\mathcal{P}_{2},\phi))\equiv\Delta_{1}$. Thus,
there is no ball that properly contains a ball of $\mathcal{P}_{1}$. It can only be that $\mathcal{P}_{1}=\{X\}$, so we must have
$\gamma\in{\rm Sim}_{X}(X,X)$. This forces $\gamma=1$, so $\widehat{\psi}$ is injective.
Finally we must show that $\widehat{\psi}:\Gamma({\rm Sim}_{X})\rightarrow D_{b}(\mathcal{P}_{{\rm Sim}_{%
X}},[X])$ is surjective. Let
$\Delta$ be a reduced $([X],[X])$-diagram over $\mathcal{P}_{{\rm Sim}_{X}}$. A transistor $T\in\mathcal{T}(\Delta)$ is called positive
if its top label is the left side of a relation in $\mathcal{P}_{{\rm Sim}_{X}}$, otherwise (i.e., if the top label is the right side of a relation
in $\mathcal{P}_{{\rm Sim}_{X}}$) the transistor $T$ is negative. It is easy to see that the sets of positive and negative transistors partition
$\mathcal{T}(\Delta)$. We claim that, if $\Delta$ is reduced, then we cannot have $T_{1}\dot{\preceq}T_{2}$ when $T_{1}$ is positive and $T_{2}$
is negative. If we had such $T_{1}\dot{\preceq}T_{2}$, then we could find $T^{\prime}_{1}\preceq T^{\prime}_{2}$, where $T^{\prime}_{1}$ is positive and $T^{\prime}_{2}$ is negative.
Since $T^{\prime}_{1}$ is positive, there is only one wire $W$ attached to the top of $T^{\prime}_{1}$. This wire must be attached to the bottom of $T^{\prime}_{2}$, since
$T^{\prime}_{1}\preceq T^{\prime}_{2}$, and it must be the only wire attached to the bottom of $T^{\prime}_{2}$, since $T^{\prime}_{2}$ is negative and $\mathcal{P}_{{\rm Sim}_{X}}$
is a tree-like semigroup presentation by Remark 4.11. Suppose that $\ell(w)=[B]$. By the definition of $\mathcal{P}_{{\rm Sim}_{X}}$, $[B]$ is the left
side of exactly one relation, namely $([B],[B_{1}][B_{2}]\ldots[B_{n}])$, where the $B_{i}$ are maximal proper subballs of $B$, listed in order. It follows
that the bottom label of $T^{\prime}_{1}$ is $[B_{1}][B_{2}]\ldots[B_{n}]$ and the top label of $T^{\prime}_{2}$ is $[B_{1}][B_{2}]\ldots[B_{n}]$. Therefore
$(T^{\prime}_{1},T^{\prime}_{2})$ is a dipole. This proves the claim.
A diagram over $\mathcal{P}_{{\rm Sim}_{X}}$ is positive if all of its transistors are positive, and negative if all of its transistors are negative.
We note that $\Delta$ is positive if and only if $\Delta^{-1}$ is negative, by the description of inverses in the proof of Theorem 2.13.
The above reasoning shows that any reduced $([X],[X])$-diagram over $\mathcal{P}_{{\rm Sim}_{X}}$ can be written
$\Delta=\Delta^{+}_{1}\circ\left(\Delta_{2}^{+}\right)^{-1}$, where $\Delta^{+}_{i}$ is a positive diagram for $i=1,2$.
We claim that any positive diagram $\Delta$ over $\mathcal{P}_{{\rm Sim}_{X}}$ with top label $[X]$ is $\Delta_{\mathcal{P}}$ (up to a reordering of the bottom
contacts), where $\mathcal{P}$ is some partition of $X$. There is a unique wire $W\in\mathcal{W}(\Delta)$ making a top contact with the frame. We call
this wire $W_{X}$. Note that its label is $[X]$ by our assumptions. The bottom contact of $W_{X}$ lies either on the bottom of the frame, or on top of some
transistor. In the first case, we have $\Delta=\Delta_{\mathcal{P}}$ for $\mathcal{P}=\{X\}$ and we are done. In the second, the bottom contact of
$W_{X}$ lies on top of some transistor $T$, which we call $T_{X}$. Since the top label of $T_{X}$ is $[X]$, the bottom label must be $[B_{1}]\ldots[B_{k}]$,
where $B_{1},\ldots,B_{k}$ are the maximal proper subballs of $X$. Thus there are wires $W_{1},\ldots,W_{k}$ attached to the bottom of $T_{X}$, and we
have $\ell(W_{i})=[B_{i}]$, for $i=1,\ldots,k$. We relabel each of the wires $W_{B_{1}},\ldots,W_{B_{k}}$, respectively. Note that
$\{B_{1},\ldots,B_{k}\}$ is a partition of $X$ into balls. We can continue in this way, inductively labelling each wire with a ball $B\subseteq X$.
If we let $\overline{B}_{1},\ldots,\overline{B}_{m}$ be the resulting labels of the wires which make bottom contacts with the frame, then
$\{\overline{B}_{1},\ldots,\overline{B}_{m}\}=\mathcal{P}$ is a partition of $X$ into balls, and $\Delta=\Delta_{\mathcal{P}}$ by construction,
up to a reordering of the bottom contacts.
We can now prove surjectivity of $\widehat{\psi}$. Let $\Delta\in D_{b}(\mathcal{P}_{{\rm Sim}_{X}},[X])$ be reduced. We can write
$\Delta=\Delta_{2}^{+}\circ\left(\Delta_{1}^{+}\right)^{-1}$, where $\Delta_{i}^{+}$ is positive, for $i=1,2$. It follows that
$\Delta_{i}^{+}=\Delta_{\mathcal{P}_{i}}\circ\sigma_{i}$, for $i=1,2$, where $\mathcal{P}_{i}$ is a partition of $X$ into balls and $\sigma_{i}$ is
diagram containing no transistors. Thus, $\Delta=\Delta_{\mathcal{P}_{2}}\circ\sigma_{2}\circ\sigma_{1}^{-1}\circ\Delta_%
{\mathcal{P}_{1}}^{-1}=\psi((\mathcal{P}_{1},\mathcal{P}_{2},\phi))$, where $\phi:\mathcal{P}_{1}\rightarrow\mathcal{P}_{2}$ is a bijection determined by
$\sigma_{2}\circ\sigma_{1}^{-1}$. Therefore, $\widehat{\psi}$ is surjective.
Now we must show that if $\mathcal{P}=\langle\Sigma\mid\mathcal{R}\rangle$ is a tree-like semigroup presentation, $x\in\Sigma$, then there
is a compact ultrametric space $X_{\mathcal{P}}$, a small similarity structure ${\rm Sim}_{X_{\mathcal{P}}}$, and a compatible ball order, such that
$D_{b}(\mathcal{P},x)\cong\Gamma({\rm Sim}_{X_{\mathcal{P}}})$.
Construct a labelled ordered simplicial tree $T_{(\mathcal{P},x)}$ as follows. Begin with a vertex $\ast$, the root, labelled by $x\in\Sigma$.
By the definition of tree-like semigroup presentation (Definition 4.1), there is at most one relation in $\mathcal{R}$ having
the word $x$ as its left side. Let us suppose first that $(x,x_{1}x_{2}\ldots x_{k})\in\mathcal{R}$, where $k\geq 2$.
We introduce $k$ children of the root, labelled $x_{1},\ldots,x_{k}$ (respectively), each connected to the root by an edge.
The children are ordered from left to right in such a way that we read the
word $x_{1}x_{2}\ldots x_{k}$ as we read the labels of the children from left to right. If, on the other hand, $x$ is not the left side of any relation
in $\mathcal{R}$, then the tree terminates – there is only the root. We continue similarly: if $x_{i}$ is the left side of some relation
$(x_{i},y_{1}y_{2}\ldots y_{m})\in\mathcal{R}$ ($m\geq 2$),
then this relation is unique and we introduce a labelled ordered collection of children, as above. If $x_{i}$ is not the left side of any relation
in $\mathcal{R}$, then $x_{i}$ has no children. This builds a labelled ordered tree $T_{(\mathcal{P},x)}$. We note that if a vertex $v\in T_{(\mathcal{P},x)}$ is labelled by $y\in\Sigma$, then the subcomplex $T_{v}\leq T_{(\mathcal{P},x)}$ spanned by $v$ and all of its descendants is isomorphic
to $T_{(\mathcal{P},y)}$, by a simplicial isomorphism which preserves the labelling and the order.
We let $\mathrm{Ends}(T_{(\mathcal{P},x)})$ denote the set of all edge-paths $p$ in $T_{(\mathcal{P},x)}$ such that:
i) $p$ is without backtracking; ii) $p$ begins at the root; iii) $p$ is either infinite, or $p$ terminates at a vertex without children.
We define a metric on $\mathrm{Ends}(T_{(\mathcal{P},x)})$ as follows. If $p,p^{\prime}\in\mathrm{Ends}(T_{(\mathcal{P},x)})$ and $p,p^{\prime}$ have exactly
$m$ edges in common, then we set $d(p,p^{\prime})=e^{-m}$. This metric makes $\mathrm{Ends}(T_{(\mathcal{P},x)})$ a compact ultrametric space, and a ball order is given
by the ordering of the tree. We can describe the
balls in $\mathrm{Ends}(T_{(\mathcal{P},x)})$ explicitly. Let $v$ be a vertex of $T_{(\mathcal{P},x)}$. We set $B_{v}=\{p\in\mathrm{Ends}(T_{(\mathcal{P},x)})\mid v\text{ lies on }p\}$. Every such set is a ball, and every ball in $\mathrm{Ends}(T_{(\mathcal{P},x)})$
has this form. We can now describe a finite similarity structure ${\rm Sim}_{X_{\mathcal{P}}}$ on $\mathrm{Ends}(T_{(\mathcal{P},x)})$. Let $B_{v}$ and $B_{v^{\prime}}$ be the balls
corresponding to the vertices $v,v^{\prime}\in T_{(\mathcal{P},x)}$. If $v$ and $v^{\prime}$ have different labels, then we set ${\rm Sim}_{X_{\mathcal{P}}}(B_{v},B_{v^{\prime}})=\emptyset$.
If $v$ and $v^{\prime}$ have the same label, say $y\in\Sigma$, then there is label- and order-preserving simplicial isomorphism $\psi:T_{v}\rightarrow T_{v^{\prime}}$.
Suppose that $p_{v}$ is the unique edge-path without backtracking connecting the root to $v$. Any point in $B_{v}$ can be expressed in the form
$p_{v}q$, where $q$ is an edge-path without backtracking in $T_{v}$. We let $\widehat{\psi}:B_{v}\rightarrow B_{v^{\prime}}$ be defined by the
rule $\widehat{\psi}(p_{v}q)=p_{v^{\prime}}\psi(q)$. The map $\widehat{\psi}$ is easily seen to be a surjective similarity.
We set ${\rm Sim}_{X_{\mathcal{P}}}(B_{v},B_{v^{\prime}})=\{\widehat{\psi}\}$. The resulting assignments give a small similarity structure ${\rm Sim}_{X_{\mathcal{P}}}$ on
the compact
ultrametric space $\mathrm{Ends}(T_{(\mathcal{P},x)})$ that is compatible with the ball order.
Now we can apply the first part of the theorem: setting $X_{(\mathcal{P},x)}=\mathrm{Ends}(T_{(\mathcal{P},x)})$, we
have $\Gamma({\rm Sim}_{X_{(\mathcal{P},x)}})\cong D_{b}(\mathcal{P}_{{\rm Sim}_{X_{%
(\mathcal{P},x)}}},[X_{(\mathcal{P},x)}])\cong D_{b}(\mathcal{P},x)$, where the first isomorphism follows from the forward direction of the theorem, and the second isomorphism follows from the canonical identification of the semigroup presentation $\mathcal{P}_{{\rm Sim}_{X_{(\mathcal{P},x)}}}$
with $\mathcal{P}$.
∎
References
[1]
Collin Bleak, Francesco Matucci, and Max Neunhöffer.
Embeddings into Thompson’s group $\mathrm{V}$ and $co\mathcal{CF}$
groups.
arXiv:1312.1855.
[2]
Kenneth S. Brown.
Finiteness properties of groups.
J. Pure Appl. Algebra, 44(1-3):45–75, 1987.
[3]
J. W. Cannon, W. J. Floyd, and W. R. Parry.
Introductory notes on Richard Thompson’s groups.
Enseign. Math. (2), 42(3-4):215–256, 1996.
[4]
Daniel S. Farley.
Actions of picture groups on CAT(0) cubical complexes.
Geom. Dedicata, 110:221–242, 2005.
[5]
Daniel S. Farley and Bruce Hughes.
Finiteness properties of some groups of local similarities.
arXiv:1206.2692.
[6]
Victor Guba and Mark Sapir.
Diagram groups.
Mem. Amer. Math. Soc., 130(620):viii+117, 1997.
[7]
Bruce Hughes.
Local similarities and the Haagerup property, with an appendix by
Daniel S. Farley.
Groups Geom. Dyn., 3:299–315, 2009. |
Spatial properties of entangled photon pairs generated in nonlinear layered structures
Jan Peřina, Jr.
Palacký University, RCPTM, Joint Laboratory of
Optics, 17. listopadu 12, 771 46 Olomouc, Czech Republic
[email protected]
Abstract
A spatial quantum model of spontaneous parametric down-conversion
in nonlinear layered structures is developed expanding the
interacting vectorial fields into monochromatic plane waves. A
two-photon spectral amplitude depending on the signal- and
idler-field frequencies and propagation directions is used to
derive transverse profiles of the emitted fields as well as their
spatial correlations. Intensity spatial profiles and their spatial
correlations are mainly determined by the positions of
transmission peaks formed in these structures with photonic bands.
A method for geometry optimization of the structures with respect
to efficiency of the nonlinear process is suggested. Several
structures composed of GaN/AlN layers are analyzed as typical
examples. They allow the generation of photon pairs correlated in
several emission directions. Photon-pair generation rates
increasing better than the second power of the number of layers
can be reached. Also structures efficiently generated photon pairs
showing anti-bunching and anti-coalescence can be obtained. Three
reasons for splitting the correlated area in photonic-band-gap
structures are revealed: zig-zag movement of photons inside the
structure, spatial symmetry and polarization-dependent properties.
Also spectral splitting can be observed in these structures.
nonlinear photonic-band-gap structure, nonlinear layered
structure, entangled photon pair, correlated area
pacs: 42.65.Lm,42.70.Qs,42.50.Dv
I Introduction
The first temporal correlations between the signal and idler
photons generated in the process of spontaneous parametric
down-conversion (SPDC) were observed already more than thirty
years ago Hong et al. (1987). They manifested quantum entanglement in
the common state of both photons generated in one quantum event of
the spontaneous nonlinear process Mandel and Wolf (1995). Since then
the understanding of properties of such photon pairs has grown
enormously. Polarization properties and namely polarization
entanglement between the signal and idler photons originating in
tensorial character of the nonlinear susceptibility have attracted
the greatest attention from the very beginning. The reason lies in
the simplicity of such states described in the Hilbert space with
dimension $2\times 2$. This made these states easily
experimentally accessible. Despite their simplicity, such states
have allowed to reveal many substantial features of quantum
physics related to correlations between subsystems
Bou (2000). Violation of the Bell inequalities that
ruled out neoclassical theories Peřina et al. (1994), the effect of
collapse of a wave-function Peřina et al. (1994) and the ability of
teleportation of a quantum state Bouwmeester et al. (1997) belong to
the most important.
In the time domain, both photons occur in a very narrow temporal
window Hong et al. (1987) because they are emitted at one instant
after the annihilation of a pump photon. Moreover the conservation
law of energy dictates entanglement among monochromatic components
of the signal and idler fields Mandel and Wolf (1995). This
entanglement leads to a typical finite entanglement time between
the detection instants of the signal and idler photons observable
in a Hong-Ou-Mandel interferometer Hong et al. (1987) or
sum-frequency generation of the paired photons
Harris (2007); Brida et al. (2009). These properties have been
extensively studied namely in connection with the generation by
pulsed pump fields Keller and Rubin (1997); Peřina Jr.
et al. (1999) that allows
precise synchronization of photons from different photon pairs.
Spatial properties of photon pairs have attracted attention last.
Correlations in the transverse planes of the signal and idler
photons occur here due to geometric properties of a photon-pair
source and pump-beam spatial profile. They originate in the
necessity of reaching good spatial phase matching of the
interacting fields Mandel and Wolf (1995) that results in an efficient
nonlinear process. For example, the sum of wave vectors of the
signal and idler fields has to approximately give the wave vector
of the pump field to observe an efficient photon-pair generation
in an extended bulk crystal using collimated pumping. This results
in strong correlations in emission directions of the signal and
idler photons Joobeur et al. (1994, 1996); Vallone et al. (2007). These
correlations may even be exploited to ’transfer’ spatial
properties of the pump beam into spatial correlations of the
signal and idler beams Monken et al. (1998); Walborn et al. (2004). We may
observe analogy between spatial and spectral correlations
(entanglement) of two photons in a pair. Similarly as spectral
correlations may be tailored by a pump-field spectral profile,
spatial correlations may be controlled by a pump-beam spatial
profile. Also signal and idler electric-field phase variations in
the transverse plane belong to important characteristics of photon
pairs. They can be quantified in terms of eigenstates of the
angular orbital momentum operator Law and Eberly (2004). Under certain
conditions, entanglement between such states in the signal and
idler fields has been observed Mair et al. (2001); Oemrawsingh et al. (2005).
Experimentally, spatial correlations can be conveniently measured
by moving a fiber tip in the transverse plane, as it has been
done, e.g., in Molina-Terriza
et al. (2005). The use of an
intensified CCD camera has provided a more elegant way for the
experimental investigations of spatial correlations
Jost et al. (1998); Haderka et al. (2005). Spatial correlations between the
down-converted beams have also been extensively used for quantum
ghost imaging Rubin and Shih (2008). We note that spatial correlations
occur not only in spontaneous regime, they have been found also in
the regime with prevailing stimulated emission
Brambilla et al. (2004); Jedrkiewicz et al. (2004).
All these forms of entanglement are potentially interesting both
for fundamental physical experiments and practical applications
including metrology Migdall (1999), quantum cryptography
Bruß and Lütkenhaus (2000) and quantum-information processing
Bou (2000). In principle, all these forms of
entanglement can occur simultaneously depending on the source of
photon pairs. However, usual sources of photon pairs are
constructed such that only one or two forms of entanglement (e.g.,
polarization and spectral) are found and may efficiently be
experimentally exploited.
The effort to generate entangled photon pairs extended over as
many basis states as possible (and defined in several degrees of
freedom) and to modify ’the structure of entanglement’ belong
together with the effort to enhance quantum efficiency of the
generation process to the leitmotifs of the development in this
field. Bulk nonlinear crystals that were nearly exclusively used
at the beginning have been gradually replaced by more complex and
efficient nonlinear structures including poled nonlinear materials
Harris (2007); Kitaeva (2007); Svozilík and
Peřina Jr. (2009, 2010), nonlinear
waveguides URen et al. (2004); Spillane et al. (2007); Chen et al. (2009) and nonlinear
photonic structures Ber (2001). Nonlinear photonic
structures are extraordinarily interesting because they allow
efficient photon-pair generation owing to enhanced electric-field
amplitudes observed inside structures with photonic band-gaps
Yablonovitch (1987); John (1987) on one side, they also allow
relatively wide tailoring of properties of the emitted photon
pairs on the other side
Vamivakas et al. (2004); Centini et al. (2005); Peřina Jr.
et al. (2006). Structured
nonlinear fibers that rely on four-wave mixing
Li et al. (2005); Fulconis et al. (2005); Fan et al. (2005) represent a typical example. Or
waveguides with Bragg mirrors Abolghasem et al. (2009); Svozilík et al. (2011)
can be mentioned as perspective sources.
In the article, we consider systems composed of parallel nonlinear
layers. Back-scattering of the fields on the boundaries between
layers provides suitable conditions for the enhancement of
electric-field amplitudes under certain conditions Yeh (1988).
Moreover, spatial properties of photon pairs can be efficiently
tailored using parameters of these systems (e.g., the number of
layers). Nonlinear layered structures have already been studied in
the simplified geometry both in the framework of classical
Centini et al. (2005) and quantum Peřina Jr.
et al. (2006) descriptions.
Here, we generalize the quantum model present in
Peřina Jr.
et al. (2006) to the real spatial geometry including
vectorial character of the interacting fields. This allows us to
study transverse intensity profiles of the down-converted beams as
well as correlated areas of the signal and idler photons
considering typical layered structures made of GaN/AlN. We note
that GaN/AlN structures with random layers’ lengths have already
been studied as sources of photon pairs with ultra-narrow spectral
widths that originated in an optical analog of Anderson
localization of the down-converted fields
Peřina Jr.
et al. (2009a, b). It has also been shown that
GaN/AlN nonlinear layers allow the generation of photon-pair
states antisymmetric with respect to the exchange of the signal-
and idler-field frequencies that exhibit anti-bunching and
anti-coalescence Peřina Jr.
et al. (2007a).
The investigated layered structures as sources of photon pairs can
be compared with other photon-pair sources with respect to
intensity transverse profile, correlated area and efficiency of
photon-pair generation as follows. The vast majority of
photon-pair sources including wave-guiding structures and
nonlinear crystals with/withour poling are designed such that the
signal (and similarly the idler) photon is generated into one
compact and small emission area. In some sources, bulk crystals
are cut in such a way that the spatial symmetry allows to generate
photons around the whole cone surface Vallone et al. (2007). In this
case the generated state is entangled also in the wave vectors of
the signal and idler photons. As the Hilbert space corresponding
to this degree of freedom has a greater number of independent
states we obtain a higher-dimensional entangled state. In case of
layered structures, the signal (and idler) photon can even be
generated around several concentric cone surfaces depending on
complexity of the structure. This even enlarges the number of
independent states in the transverse area making the generated
state suitable for ’parallel processing’ of quantum information in
the transverse plane. As for correlated areas, they are described
by compact profiles (usually of an elliptical shape) in all common
photon-pair sources with cw pumping. Splitting of correlated areas
into several well-separated parts is a distinguished property of
nonlinear layered structures. Even three different mechanisms
leading to this splitting exist in layered structures: zig-zag
movement of photons inside the structure, spatial symmetry and
polarization-dependent properties. The last two mechanisms may in
principle occur also in other photon-pair sources. However,
enhancement of the fields’ amplitudes inside a structure is needed
to make them significant. Finally, periodically-poled nonlinear
crystals provide the greatest photon-pair fluxes. On the other
hand, wave-guiding structures including nonlinear planar
wave-guides and nonlinear structured fibers have the greatest
nonlinear conversion efficiencies due to the transverse
confinement of the interacting fields. However, they allow for
only moderate pumping intensities because of a possible material
damage and competing nonlinear processes. Nonlinear layered
structures lie in the middle in this comparison. They provide
greater nonlinear conversion efficiencies than bulk crystals owing
to enhanced electric-field amplitudes along the propagation
direction. Compared to wave-guiding structures, their nonlinear
conversion efficiencies are lower because the ’fields’
confinement’ is only in one dimension. On the other hand, similar
pumping intensities as those used for bulk crystals can be
applied.
The paper is organized as follows. A spatial vectorial quantum
model of SPDC in layered media is present in Sec. II and provides
quantities characterizing the emitted photon pairs. A systematic
method for designing efficient nonlinear structures is described
in Sec. III. Transverse intensity profiles and correlated areas
are discussed in Sec. IV using structures with different numbers
of layers. Conclusions are drawn in Sec. V.
II Spatial quantum model of spontaneous parametric
down-conversion
Nonlinear Hamiltonian $\hat{H}_{\rm int}$ characterizing SPDC in
a nonlinear medium of volume $\cal V$ at time $t$ can be
written in the form Mandel and Wolf (1995):
$$\displaystyle\hat{H}_{\rm int}(t)$$
$$\displaystyle=$$
$$\displaystyle\epsilon_{0}\int_{\cal V}d{\bf r}$$
(1)
$$\displaystyle{\bf d}({\bf r}):\left[{\bf E}_{p}^{(+)}({\bf r},t)\hat{\bf E}_{s%
}^{(-)}({\bf r},t)\hat{\bf E}_{i}^{(-)}({\bf r},t)+{\rm h.c.}\right].$$
In Eq. (1), ${\bf d}$ stands for a third-order tensor of
nonlinear coefficients and the symbol $:$ means shorthand of the
tensor ${\bf d}$ with respect to its three indices. A strong
pump field is described by the positive-frequency part ${\bf E}_{p}^{(+)}({\bf r},t)$ of its electric-field amplitude vector.
A signal [idler] field at single-photon level is characterized by
the negative-frequency part $\hat{\bf E}_{s}^{(-)}({\bf r},t)$
[$\hat{\bf E}_{s}^{(-)}({\bf r},t)$] of its electric-field
operator amplitude. Symbol $\epsilon_{0}$ denotes permittivity of
vacuum whereas ${\rm h.c.}$ stands for the Hermitian conjugate
term.
The positive-frequency amplitudes ${\bf E}_{m}^{(+)}({\bf r},t)$
of the interacting fields ($m=p,s,i$) can be in general
decomposed into plane waves with wave vectors ${\bf k}_{m}$ and
amplitudes ${\bf E}_{m}^{(+)}({\bf k}_{m})$:
$${\bf E}_{m}^{(+)}({\bf r},t)=\frac{1}{(\sqrt{2\pi})^{3}}\int d^{3}{\bf k}_{m}{%
\bf E}_{m}^{(+)}({\bf k}_{m})\exp(i{\bf k}_{m}{\bf r}-i\omega_{m}t);$$
(2)
$\omega_{m}$ is the frequency of field $m$ determined in
accordance with dispersion relations.
Considering the incident and un-depleted pump field, we assume
that its temporal spectrum ${\cal E}_{p}(\omega_{p})$ as well as
spatial spectrum ${\cal E}_{p}^{\rm tr}(k_{px},k_{py})$ in the
transverse plane are given. In this case, the decomposition of
amplitude ${\bf E}_{p}^{(+)}({\bf r},t)$ in Eq. (2) can be
rewritten as:
$$\displaystyle{\bf E}_{p}^{(+)}({\bf r},t)$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{(\sqrt{2\pi}c)^{3}}\int_{-\pi/2}^{\pi/2}\sin(\vartheta_{%
p})d\vartheta_{p}\int_{-\pi/2}^{\pi/2}d\psi_{p}$$
(3)
$$\displaystyle\int_{0}^{\infty}\omega_{p}^{2}d\omega_{p}\;{\cal E}_{p}(\omega_{%
p}){\cal E}_{p}^{\rm tr}\left[k_{p,x}({\bf\Omega_{p}}),k_{p,y}({\bf\Omega_{p}}%
)\right]$$
$$\displaystyle\mbox{}\times\exp\left[ik_{p,x}({\bf\Omega_{p}})x+ik_{p,y}({\bf%
\Omega_{p}})y\right]\sum_{\alpha={\rm TE,TM}}{\bf E}_{p,\alpha}^{(+)}(z,{\bf%
\Omega_{p}})$$
$$\displaystyle\mbox{}\times\exp(-i\omega_{p}t);$$
using the vector ${\bf\Omega}_{p}\equiv(\omega_{p},\vartheta_{p},\psi_{p})$ of ’spherical coordinates’ $\omega_{p}$, $\vartheta_{p}$ and $\psi_{p}$. Speed of light in
vacuum is denoted as $c$. Assuming for simplicity air around the
structure the $x$ and $y$ components of wave vector ${\bf k}_{p}$ in front of the structure are given as:
$$\displaystyle k_{p,x}({\bf\Omega}_{p})$$
$$\displaystyle=$$
$$\displaystyle-\frac{\omega_{p}\sin(\psi_{p})\sin(\vartheta_{p})}{c},$$
$$\displaystyle k_{p,y}({\bf\Omega}_{p})$$
$$\displaystyle=$$
$$\displaystyle\frac{\omega_{p}\cos(\psi_{p})\sin(\vartheta_{p})}{c}.$$
(4)
The decomposition of pump-field amplitude ${\bf E}_{p}^{(+)}$
into TE and TM waves as given in Eq. (3) is done with
respect to the plane of incidence of the wave with wave vector ${\bf k}_{p}$ propagating through the layered structure (see
Fig. 1 for a scheme of the structure). We note that
projections of ${\bf k}$ vectors into the planes of boundaries
are conserved through the structure. The pump-field amplitudes ${\bf E}_{p,\alpha}^{(+)}$ introduced in Eq. (3) describe
field evolution along the $z$ axis where the field undergoes
back-scattering at the boundaries.
Considering a structure with $N$ layers and boundaries
perpendicular to the $z$ axis localized at positions $z_{n}$, $n=0,\ldots,N$, the pump-field amplitude ${\bf E}_{p,\alpha}^{(+)}$ can be written in the form (for more
details, see Peřina Jr.
et al. (2006)):
$$\displaystyle{\bf E}_{p,\alpha}^{(+)}(z,{\bf\Omega}_{p})$$
$$\displaystyle=$$
$$\displaystyle{\rm rect}_{-\infty,z_{0}}(z)\sum_{a=F,B}A_{p_{a},\alpha}^{(0)}({%
\bf\Omega}_{p})$$
(5)
$$\displaystyle\mbox{}\times{\bf e}_{p_{a},\alpha}^{(0)}({\bf\Omega}_{p})\exp%
\left[iK_{p_{a},z}^{(0)}({\bf\Omega}_{p})(z-z_{0})\right]$$
$$\displaystyle\mbox{}\hskip-8.535827pt+\sum_{l=1}^{N}{\rm rect}_{z_{l-1},z_{l}}%
(z)\sum_{a=F,B}A_{p_{a},\alpha}^{(l)}({\bf\Omega}_{p})$$
$$\displaystyle\mbox{}\hskip-28.452756pt\times{\bf e}_{p_{a},\alpha}^{(l)}({\bf%
\Omega}_{p})\exp\left[iK_{p_{a},z}^{(l)}({\bf\Omega}_{p})(z-z_{l-1})\right]$$
$$\displaystyle\mbox{}\hskip-8.535827pt+{\rm rect}_{z_{N},\infty}(z)\sum_{a=F,B}%
A_{p_{a},\alpha}^{(N+1)}({\bf\Omega}_{p})$$
$$\displaystyle\mbox{}\hskip-28.452756pt\times{\bf e}_{p_{a},\alpha}^{(N+1)}({%
\bf\Omega}_{p})\exp\left[iK_{p_{a},z}^{(N+1)}({\bf\Omega}_{p})(z-z_{N})\right];$$
$$\displaystyle\alpha={\rm TE,TM}.$$
The function ${\rm rect}_{z_{a},z_{b}}(z)$ equals one for $z_{a}\leq z\leq z_{b}$ and is zero otherwise. Symbols ${\bf e}^{(l)}_{p_{F},\alpha}$ and ${\bf e}^{(l)}_{p_{B},\alpha}$ denote
polarization vectors of $\alpha$ waves for forward- and
backward-propagating fields with respect to the $+z$ axis,
respectively.
The $z$ component $K_{p_{a},z}^{(l)}({\bf\Omega}_{p})$ of wave
vector in an $l$th layer belonging to wave $a$ with frequency
$\omega_{p}$ and propagating in direction ($\vartheta_{p},\psi_{p}$) in front of the structure can be expressed as:
$$K_{p_{a},z}^{(l)}({\bf\Omega}_{p})=\pm\frac{n_{p}^{(l)}(\omega_{p})\omega_{p}}%
{c}\cos(\vartheta_{p}^{(l)}),$$
(6)
where the sign $+$ ($-$) is appropriate for a forward-
(backward-) propagating wave. Index of refraction of the pump
field in an $l$th layer is denoted as $n_{p}^{(l)}$. As we
assume that the layered structure is surrounded by air, we have $n_{p}^{(0)}=n_{p}^{(N+1)}=1$. The propagation angle $\vartheta_{p}^{(l)}$ in the $l$th layer is then derived from
Snell’s law:
$$n_{p}^{(0)}\sin(\vartheta_{p}^{(0)})=n_{p}^{(l)}\sin(\vartheta_{p}^{(l)}),%
\hskip 14.226378ptl=1,\ldots,N+1;$$
(7)
$\vartheta_{p}^{(0)}\equiv\vartheta_{p}$.
The coefficients $A_{p_{F},\alpha}^{(l)}({\bf\Omega}_{p})$ and $A_{p_{B},\alpha}^{(l)}({\bf\Omega}_{p})$ introduced in Eq. (5)
determine amplitudes of $\alpha$ waves with frequency $\omega_{p}$ propagating forward and backward, respectively, in the ($\vartheta_{p},\psi_{p}$) direction. These coefficients are derived
from Fresnel’s relations at the boundaries using, e.g., the
transfer-matrix formalism Yeh (1988). Values of the
coefficients $A_{p_{F},\alpha}^{(0)}({\bf\Omega}_{p})$ and $A_{p_{B},\alpha}^{(N+1)}({\bf\Omega}_{p})$ for $\alpha={\rm TE,TM}$ characterize the pump fields incident on the structure from
both sides and represent the boundary conditions. The
transfer-matrix formalism has been elaborated in detail in
Peřina Jr.
et al. (2006) for layered structures. The coefficients $A^{(l)}$ of the fields with wave vectors lying in the plane of
incidence given by an angle $\psi_{p}$ are given by Eq. (23) of
Ref. Peřina Jr.
et al. (2006).
The signal and idler fields are at single-photon level in the
spontaneous process and so they have to be treated using quantum
theory Mandel and Wolf (1995). Their electric-field operator amplitudes
in the layered structure can be decomposed into plane waves using
Eq. (1) and subsequently described in analogy to the pump
field. Using the ’spherical coordinates’ $\omega_{m}$ $,\vartheta_{m}$ and $\psi_{m}$ and defining ${\bf\Omega}_{m}=(\omega_{m},\vartheta_{m},\psi_{m})$, $m=s,i$, we can express the
positive-frequency part $\hat{E}_{m}^{(+)}$ of the
electric-field operator amplitude of mode $m$ as:
$$\displaystyle\hat{\bf E}_{m}^{(+)}({\bf r},t)$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{c^{3}}\int_{-\pi/2}^{\pi/2}\sin(\vartheta_{m})d\vartheta%
_{m}\int_{-\pi/2}^{\pi/2}d\psi_{m}$$
(8)
$$\displaystyle\int_{0}^{\infty}\omega_{m}^{2}d\omega_{m}\exp\left[ik_{m,x}({\bf%
\Omega}_{m})x+ik_{m,y}({\bf\Omega}_{m})y\right]$$
$$\displaystyle\mbox{}\times\sum_{\alpha={\rm TE,TM}}\sqrt{\frac{\hbar\omega_{m}%
}{16\pi^{3}\epsilon_{0}}}\;\hat{\bf a}_{m,\alpha}(z,{\bf\Omega}_{m})\exp(-i%
\omega_{m}t);$$
$\hbar$ stands for the reduced Planck constant and $\hat{\bf E}_{m}^{(-)}=\hat{\bf E}_{m}^{(+)\dagger}$ . The expression $\sqrt{\hbar\omega_{m}/(16\pi^{3}\epsilon_{0})}$ gives an
electric-field amplitude per one photon with energy $\hbar\omega_{m}$ propagating at speed $c$. The $x$ and $y$
components of wave vectors ${\bf k}_{m}$ of the signal and idler
fields ($m=s,i$) are defined outside the structure:
$$\displaystyle k_{m,x}({\bf\Omega}_{m})$$
$$\displaystyle=$$
$$\displaystyle-\frac{\omega_{m}\sin(\psi_{m})\sin(\vartheta_{m})}{c},$$
$$\displaystyle k_{m,y}({\bf\Omega}_{m})$$
$$\displaystyle=$$
$$\displaystyle\frac{\omega_{m}\cos(\psi_{m})\sin(\vartheta_{m})}{c}.$$
(9)
The operator amplitudes $\hat{\bf a}_{m,\alpha}(z,{\bf\Omega}_{m})$ introduced in Eq. (8) can be derived in the considered
layered structure in the form:
$$\displaystyle\hat{\bf a}_{m,\alpha}(z,{\bf\Omega}_{m})$$
$$\displaystyle=$$
$$\displaystyle{\rm rect}_{-\infty,z_{0}}(z)\sum_{a=F,B}\hat{a}_{m_{a},\alpha}^{%
(0)}({\bf\Omega}_{m})$$
(10)
$$\displaystyle\mbox{}\times{\bf e}_{m_{a},\alpha}^{(0)}({\bf\Omega}_{m})\exp%
\left[iK_{m_{a},z}^{(0)}({\bf\Omega}_{m})(z-z_{0})\right]$$
$$\displaystyle\mbox{}\hskip-8.535827pt+\sum_{l=1}^{N}{\rm rect}_{z_{l-1},z_{l}}%
(z)\sum_{a=F,B}\hat{a}_{m_{a},\alpha}^{(l)}({\bf\Omega}_{m})$$
$$\displaystyle\mbox{}\hskip-28.452756pt\times{\bf e}_{m_{a},\alpha}^{(l)}\exp%
\left[iK_{m_{a},z}^{(l)}({\bf\Omega}_{m})(z-z_{l-1})\right]$$
$$\displaystyle\mbox{}\hskip-8.535827pt+{\rm rect}_{z_{N},\infty}(z)\sum_{a=F,B}%
\hat{a}_{m_{a},\alpha}^{(N+1)}({\bf\Omega}_{m})$$
$$\displaystyle\mbox{}\hskip-28.452756pt\times{\bf e}_{m_{a},\alpha}^{(N+1)}({%
\bf\Omega}_{m})\exp\left[iK_{m_{a},z}^{(N+1)}({\bf\Omega}_{m})(z-z_{N})\right];$$
$$\displaystyle m=s,i;\hskip 14.226378pt\alpha={\rm TE,TM}.$$
Here, the symbols ${\bf e}^{(l)}_{m_{F},\alpha}$ and ${\bf e}^{(l)}_{m_{B},\alpha}$ stand for polarization vectors of $\alpha$ wave of field $m$ propagating forward and backward,
respectively. The annihilation operators $\hat{a}_{m_{a},\alpha}^{(l)}({\bf\Omega}_{m})$ are defined at the
end of an $l$th layer for the $\alpha$ wave with frequency $\omega_{m}$ of field $m$ propagating along the direction ($\vartheta_{m},\psi_{m}$) either forward ($a=F$) or backward ($a=B$).
The $z$ component $K_{m_{a},z}^{(l)}({\bf\Omega}_{m})$ of wave
vector in the $l$th layer is determined as:
$$\displaystyle K_{m_{a},z}^{(l)}({\bf\Omega}_{m})=\pm\frac{n_{m}^{(l)}(\omega_{%
m})\omega_{m}}{c}\cos(\vartheta_{m}^{(l)}),$$
$$\displaystyle m=s,i;\hskip 14.226378pta=F,B.$$
(11)
The sign $+$ ($-$) stands for a forward- (backward-)
propagating wave and $n_{m}^{(l)}$ gives index of refraction of
field $m$ in the $l$th layer. Snell’s law provides the
propagation angles $\vartheta_{m}^{(l)}$ needed in the $l$th
layer:
$$\displaystyle n_{m}^{(l)}(\omega_{m})\sin(\vartheta_{m}^{(l)})=n_{m}^{(N+1)}(%
\omega_{m})\sin(\vartheta_{m}^{(N+1)}),$$
$$\displaystyle \hskip 2.845276ptl=0,\ldots,N;$$
(12)
$\vartheta_{m}^{(N+1)}\equiv\vartheta_{m}$.
The operators $\hat{a}_{m_{a},\alpha}^{(l)}({\bf\Omega}_{m})$ for $l=0,\ldots,N+1$ and fixed $m$, $\alpha$, $\omega_{m}$, $\vartheta_{m}$, and $\psi_{m}$ are connected by unitary
transformations at boundaries (Fresnel’s relations) and free-field
propagation transformations. This means that the usual boson
commutation relations obeyed by the incident fields are
’transferred’ through the structure and the only nonzero
commutation relations Vogel et al. (2001) are the following ones:
$$\displaystyle[\hat{a}_{m_{a},\alpha}^{(l)}({\bf\Omega}_{m}),\hat{a}_{{m^{%
\prime}}_{a^{\prime}},\alpha^{\prime}}^{(l^{\prime})\dagger}({\bf\Omega}^{%
\prime}_{m^{\prime}})]$$
$$\displaystyle=$$
$$\displaystyle\frac{c^{2}}{|\sin(\vartheta_{m})|\omega_{m}^{2}}\delta_{m,m^{%
\prime}}\delta_{a,a^{\prime}}$$
(13)
$$\displaystyle\delta_{\alpha,\alpha^{\prime}}\delta_{l,l^{\prime}}\delta(\omega%
_{m}-\omega^{\prime}_{m^{\prime}})\delta(\vartheta_{m}-\vartheta^{\prime}_{m^{%
\prime}})\delta(\psi_{m}-\psi^{\prime}_{m^{\prime}}).$$
The transfer-matrix formalism allows to express the operators $\hat{a}_{m_{a},\alpha}^{(l)}({\bf\Omega}_{m})$ in terms of the
operators $\hat{a}_{m_{F},\alpha}^{(N+1)}({\bf\Omega}_{m})$ and $\hat{a}_{m_{B},\alpha}^{(0)}({\bf\Omega}_{m})$ describing the
outgoing fields. The appropriate relations valid for fields with
wave vectors lying in the plane of incidence given by an angle $\psi_{m}$ are given in Eq. (21) of Ref. Peřina Jr.
et al. (2006).
The generation of a photon pair in the process of SPDC governed by
the Hamiltonian $\hat{H}_{\rm int}$ in Eq. (1) is
described by a perturbation solution of the Schrödinger
equation to the first order at time $t\rightarrow\infty$
assuming an incident vacuum state $|{\rm vac}\rangle$ in the
signal and idler fields at time $t\rightarrow-\infty$. The
resulting state $|\psi\rangle_{s,i}^{\rm out}$ can be derived in
the form:
$$\displaystyle|\psi\rangle_{s,i}^{\rm out}$$
$$\displaystyle=$$
$$\displaystyle|{\rm vac}\rangle-\frac{i}{2c^{8}}\sum_{l=1}^{N}\sum_{a,b,g=F,B}%
\sum_{\alpha,\beta,\gamma={\rm TE,TM}}$$
(14)
$$\displaystyle\prod_{m=p,s,i}\int_{-\pi/2}^{\pi/2}\sin(\vartheta_{m})d\vartheta%
_{m}\int_{-\pi/2}^{\pi/2}d\psi_{m}\int_{0}^{\infty}\omega_{m}^{2}d\omega_{m}$$
$$\displaystyle\sqrt{\omega_{s}\omega_{i}}\;{\cal E}_{p}(\omega_{p}){\cal E}_{p}%
^{\rm tr}(k_{p,x},k_{p,y})\delta(\omega_{p}-\omega_{s}-\omega_{i})$$
$$\displaystyle\mbox{}\times\delta\left[k_{p,x}-k_{s,x}-k_{i,x}\right]\delta%
\left[k_{p,y}-k_{s,y}-k_{i,y}\right]$$
$$\displaystyle\mbox{}\times{\bf d}^{(l)}:{\bf e}_{p_{a},\alpha}^{(l)}({\bf%
\Omega}_{m}){\bf e}_{s_{b},\beta}^{(l)*}({\bf\Omega}_{s}){\bf e}_{i_{g},\gamma%
}^{(l)*}({\bf\Omega}_{i})$$
$$\displaystyle\mbox{}\times\exp\left[\frac{i}{2}\Delta K_{p_{a}s_{b}i_{g},z}^{(%
l)}({\bf\Omega}_{p},{\bf\Omega}_{s},{\bf\Omega}_{i})L_{l}\right]$$
$$\displaystyle\mbox{}\times L_{l}{\rm sinc}\left[\frac{1}{2}\Delta K_{p_{a}s_{b%
}i_{g},z}^{(l)}({\bf\Omega}_{p},{\bf\Omega}_{s},{\bf\Omega}_{i})L_{l}\right]$$
$$\displaystyle\mbox{}\times A_{p_{a},\alpha}^{(l)}({\bf\Omega}_{m})\hat{a}_{s_{%
b},\beta}^{(l)\dagger}({\bf\Omega}_{s})\hat{a}_{i_{g},\gamma}^{(l)\dagger}({%
\bf\Omega}_{i})|{\rm vac}\rangle.$$
The functions $\Delta K_{p_{a}s_{b}i_{g},z}^{(l)}({\bf\Omega}_{p},{\bf\Omega}_{s},{\bf\Omega}%
_{i})=K_{p_{a},z}^{(l)}({\bf\Omega}_{p})-K_{s_{b},z}^{(l)}({\bf\Omega}_{s})-K_%
{i_{g},z}^{(l)}({\bf\Omega}_{i})$ stand for phase mismatches in the $l$th layer. Symbol $L_{l}$
means the length of $l$th layer ($L_{l}=z_{l}-z_{l-1}$). The
transverse wave vectors $k_{m,x}({\bf\Omega}_{m})$ and $k_{m,y}({\bf\Omega}_{m})$ for $m=p,s,i$ are defined in
Eqs. (4) and (9) and characterize the fields outside
the structure. We note that the approach based on the solution of
Schrödinger equation does not take into account surface SPDC
that generates additional photon pairs
Peřina Jr.
et al. (2009c, d).
The phase-matching conditions in the transverse plane $xy$ are
described by two $\delta$ functions in Eq. (14), that
determine the emission direction ($\vartheta_{i},\psi_{i}$) of an
idler photon provided that the signal-photon emission direction ($\vartheta_{s},\psi_{s}$) is given and the pump-field is in the form
of a plane wave propagating in direction ($\vartheta_{p},\psi_{p}$).
Simple geometric considerations provide the following formulas:
$$\displaystyle\psi_{i}$$
$$\displaystyle=$$
$$\displaystyle\psi_{p}+\arctan\left[\frac{\omega_{s}\sin(\vartheta_{s})\sin(%
\psi_{p}-\psi_{s})}{\omega_{p}\sin(\vartheta_{p})-\omega_{s}\sin(\vartheta_{s}%
)\cos(\psi_{p}-\psi_{s})}\right],$$
$$\displaystyle\vartheta_{i}$$
$$\displaystyle=$$
$$\displaystyle\arcsin\left[\frac{\omega_{p}\sin(\vartheta_{p})}{\omega_{i}\cos(%
\psi_{p}-\psi_{i})}\right.$$
(15)
$$\displaystyle\mbox{}\hskip 56.905512pt\left.-\frac{\omega_{s}\cos(\psi_{p}-%
\psi_{s})}{\omega_{i}\cos(\psi_{p}-\psi_{i})}\sin(\vartheta_{s})\right].$$
If the pump beam is focused, the strict phase-matching conditions
in the transverse plane formulated in Eq. (15) are relaxed
and we arrive at correlation areas with finite spreads.
The expression in Eq. (14) for the state $|\psi\rangle_{si}^{\rm out}$ can be rewritten into the form
containing only the outgoing creation operators $\hat{a}_{m_{F},\beta}^{(N+1)\dagger}({\bf\Omega}_{m})$ and $\hat{a}_{m_{B},\beta}^{(0)\dagger}({\bf\Omega}_{m})$ for $m=s,i$
and $\beta={\rm TE,TM}$ using the formulas in Eqs. (21) and
(23) of Ref. Peřina Jr.
et al. (2006). The outgoing operators can be
finally transformed into the polarization basis of the detector
using a suitable unitary transformation. We assume that the
detection plane is perpendicular to the field propagation
direction ($\vartheta,\psi$) and its $s$-polarization
(denoted as $\perp$) is parallel to the horizontal plane $xz$;
$p$-polarization (denoted as $\|$) is defined by the
orthogonality conditions. Assuming field $m$ ($m=s,i$) at
frequency $\omega_{m}$ and propagating along the angles $\vartheta_{m}$ and $\psi_{m}$ the needed $\vartheta_{m}$- and $\psi_{m}$-dependent unitary transformation can be written in the
form:
$$\displaystyle\left[\matrix{\hat{a}_{m_{F},{\rm TE}}^{(N+1)}({\bf\Omega}_{m})%
\cr\hat{a}_{m_{F},{\rm TM}}^{(N+1)}({\bf\Omega}_{m})}\right]$$
$$\displaystyle=$$
$$\displaystyle\left[\matrix{\cos(\zeta_{m})&\sin(\zeta_{m})\cr-\sin(\zeta_{m})&%
\cos(\zeta_{m})}\right]\left[\matrix{\hat{a}_{m_{F},\perp}({\bf\Omega}_{m})\cr%
\hat{a}_{m_{F},\|}({\bf\Omega}_{m})}\right],$$
$$\displaystyle\left[\matrix{\hat{a}_{m_{B},{\rm TE}}^{(0)}({\bf\Omega}_{m})\cr%
\hat{a}_{m_{B},{\rm TM}}^{(0)}({\bf\Omega}_{m})}\right]$$
$$\displaystyle=$$
$$\displaystyle\left[\matrix{\cos(\zeta_{m})&\sin(\zeta_{m})\cr-\sin(\zeta_{m})&%
\cos(\zeta_{m})}\right]\left[\matrix{\hat{a}_{m_{B},\perp}({\bf\Omega}_{m})\cr%
\hat{a}_{m_{B},\|}({\bf\Omega}_{m})}\right],$$
$$\displaystyle\zeta_{m}(\vartheta_{m},\psi_{m})$$
$$\displaystyle=$$
$$\displaystyle\arccos\left[\frac{\cos(\psi_{m})}{\sqrt{1+\sin^{2}(\psi_{m})\tan%
^{2}(\theta_{m})}}\right]$$
(16)
$$\displaystyle\mbox{}\times{\rm sign}(\psi_{m}),\hskip 28.452756ptm=s,i.$$
The function ${\rm sign}$ gives the sign of its argument. The
newly introduced annihilation operators $\hat{a}_{m_{b},\alpha}({\bf\Omega}_{m})$, $m=s,i$, $b=F,B$, $\alpha=\|,\perp$, describe the signal and idler fields in the
polarization bases connected with the detectors.
The terms of state $|\psi\rangle^{\rm out}_{s,i}$ in
Eq. (14) describing the created photon pair can be
decomposed into four groups according to the propagation
directions of the signal and idler photons with respect to the $+z$ axis ($FF$, $FB$, $BF$, $BB$). Inside these groups
there occur four contributions that differ in signal- and
idler-photon polarization directions ($\|\,\|$, $\|\,\perp$, $\perp\,\|$, $\perp\,\perp$). Each contribution can be
written in the form:
$$\displaystyle|\psi\rangle_{s_{a},i_{b}}^{\alpha,\beta}({\bf r},t)$$
$$\displaystyle=$$
$$\displaystyle\prod_{m=s,i}\int_{-\pi/2}^{\pi/2}\sin(\vartheta_{m})d\vartheta_{%
m}\int_{-\pi/2}^{\pi/2}d\psi_{m}$$
(17)
$$\displaystyle\int_{0}^{\infty}d\omega_{m}\phi_{ab}^{\alpha,\beta}({\bf\Omega}_%
{s},{\bf\Omega}_{i})\hat{a}_{s_{a},\alpha}^{\dagger}({\bf\Omega}_{s})\hat{a}_{%
i_{b},\beta}^{\dagger}({\bf\Omega}_{i})|{\rm vac}\rangle$$
$$\displaystyle\mbox{}\hskip-65.441339pt\times\exp[-i({\bf k}_{s_{a}}^{\rm out}+%
{\bf k}_{i_{b}}^{\rm out}){\bf r}]\exp[i(\omega_{s}+\omega_{i})t];$$
$$\displaystyle a,b=F,B;\hskip 14.226378pt\alpha,\beta=\|,\perp.$$
The wave vectors ${\bf k}_{s_{a}}^{\rm out}$ and ${\bf k}_{i_{b}}^{\rm out}$ characterize free-field evolution of the
emitted signal and idler fields, respectively, outside the
structure. The functions $\phi_{ab}^{\alpha,\beta}({\bf\Omega}_{s},{\bf\Omega}_{i})$ introduced in Eq. (17)
characterize completely properties of the generated photon pair.
It gives the probability amplitude of having an $\alpha$-polarized signal photon at frequency $\omega_{s}$ propagating
along direction ($\vartheta_{s},\psi_{s}$) together with its $\beta$-polarized idler photon at frequency $\omega_{i}$
propagating along direction ($\vartheta_{i},\psi_{i}$) at the
output $ab$ of the structure.
Intensity spatial and spectral properties of a photon pair
Peřina Jr.
et al. (2006) can be conveniently derived from a density $n_{ab}^{\alpha,\beta}$ of the mean photon-pair numbers belonging
to the state $|\psi\rangle_{s_{a},i_{b}}^{\alpha,\beta}$. The
density $n_{ab}^{\alpha,\beta}$ is defined along the expression
$$\displaystyle n_{ab}^{\alpha,\beta}({\bf\Omega}_{s},{\bf\Omega}_{i})={}_{s_{a}%
,i_{b}}^{\alpha,\beta}\langle\psi|\hat{n}_{s_{a},\alpha}({\bf\Omega}_{s})\hat{%
n}_{i_{b},\beta}({\bf\Omega}_{i})|\psi\rangle_{s_{a},i_{b}}^{\alpha,\beta}.$$
(18)
The photon-number density operator $\hat{n}_{m_{a},\alpha}({\bf\Omega}_{m})$ is expressed as
$$\displaystyle\hat{n}_{m_{a},\alpha}({\bf\Omega}_{m})$$
$$\displaystyle=$$
$$\displaystyle\hat{a}_{m_{a},\alpha}^{\dagger}({\bf\Omega}_{m})\hat{a}_{m_{a},%
\alpha}({\bf\Omega}_{m}).$$
(19)
Using Eq. (17) the formula for density $n_{ab}^{\alpha,\beta}$ of mean photon-pair numbers in Eq.
(18) attains a simple form:
$$n_{ab}^{\alpha,\beta}({\bf\Omega}_{s},{\bf\Omega}_{i})=|\phi_{ab}^{\alpha,%
\beta}({\bf\Omega}_{s},{\bf\Omega}_{i})|^{2}.$$
(20)
A density $n_{s,ab}^{\alpha,\beta}({\bf\Omega}_{s})$ of mean
signal-photon numbers in the state $|\psi\rangle_{s_{a},i_{b}}^{\alpha,\beta}$ can easily be determined
from the density $n_{ab}^{\alpha,\beta}$ defined in
Eq. (18) and using the relation in Eq. (20):
$$\displaystyle n_{s,ab}^{\alpha,\beta}({\bf\Omega}_{s})$$
$$\displaystyle=$$
$$\displaystyle\int_{-\pi/2}^{\pi/2}\sin(\vartheta_{i})d\vartheta_{i}\int_{-\pi/%
2}^{\pi/2}d\psi_{i}\int_{0}^{\infty}d\omega_{i}$$
(21)
$$\displaystyle|\phi_{ab}^{\alpha,\beta}({\bf\Omega}_{s},{\bf\Omega}_{i})|^{2}.$$
If spectral resolution in detection of the signal-field transverse
profile is not available, a spatial density $n_{s,ab}^{{\rm tr},\alpha,\beta}(\vartheta_{s},\psi_{s})$ of mean signal-photon
numbers emitted in the direction ($\vartheta_{s},\psi_{s}$) is a
useful characteristic. It can be derived using the density $n_{s,ab}^{\alpha,\beta}$ of mean signal-photon numbers given in
Eq. (21):
$$\displaystyle n_{s,ab}^{{\rm tr},\alpha,\beta}(\vartheta_{s},\psi_{s})$$
$$\displaystyle=$$
$$\displaystyle\int_{0}^{\infty}d\omega_{s}n_{s,ab}^{\alpha,\beta}({\bf\Omega}_{%
s}).$$
(22)
The above densities can be analogously defined also for the idler
field.
Spatial correlations between the signal- and idler-field photon
numbers in their transverse planes can be quantified in terms of
the fourth-order correlation functions $n_{ab}^{{\rm cor},\alpha,\beta}(\vartheta_{s},\psi_{s},\vartheta_{i},\psi_{i})$ that
give joint densities of photon-pair numbers such that a signal
photon propagates along the ($\vartheta_{s},\psi_{s}$) direction and
the idler twin propagates along the ($\vartheta_{i},\psi_{i}$)
direction:
$$\displaystyle n_{ab}^{{\rm cor},\alpha,\beta}(\vartheta_{s},\psi_{s},\vartheta%
_{i},\psi_{i})$$
$$\displaystyle=$$
$$\displaystyle\int_{0}^{\infty}d\omega_{s}\int_{0}^{\infty}d\omega_{i}n_{ab}^{%
\alpha,\beta}({\bf\Omega}_{s},{\bf\Omega}_{i}).$$
Provided that the signal-photon propagation direction ($\vartheta_{s}^{0},\psi_{s}^{0}$) is given, the joint density $n_{ab}^{{\rm cor},\alpha,\beta}(\vartheta_{s}^{0},\psi_{s}^{0},\vartheta_{i},%
\psi_{i})$
remains a function of the idler-field emission angles $\vartheta_{i}$ and $\psi_{i}$ and its profile defines a correlated
area. The correlated area of an idler photon determines an area in
the idler-field transverse plane where an idler photon can be
expected provided that its signal twin has been detected in the ($\vartheta_{s}^{0},\psi_{s}^{0}$) direction.
Finally, an overall mean photon-pair number $N_{ab}^{\alpha,\beta}$ related to the state $|\psi\rangle_{s_{a},i_{b}}^{\alpha,\beta}$ can be found using the
following relation:
$$\displaystyle N_{ab}^{\alpha,\beta}$$
$$\displaystyle=$$
$$\displaystyle\prod_{m=s,i}\int_{-\pi/2}^{\pi/2}\sin(\vartheta_{m})d\vartheta_{%
m}\int_{-\pi/2}^{\pi/2}d\psi_{m}\int_{0}^{\infty}d\omega_{m}$$
(24)
$$\displaystyle n_{ab}^{\alpha,\beta}({\bf\Omega}_{s},{\bf\Omega}_{i}).$$
An important feature of layered structures is an increase of the
efficiency of nonlinear process due to the enhanced electric-field
amplitudes caused by interference along the $z$ axis. This
increase can be quantified with respect to a certain reference
structure which fully exploits the nonlinearity, but does not rely
on interference. In this reference structure, there occurs no
back-scattering of the propagating fields and also the nonlinear
process is assumed to be fully phase matched. The orientations of
nonlinear layers and polarizations of the interacting fields are
such that the most intense nonlinear effect occurs. This reference
structure provides an ideal photon pair with a signal photon
emitted in an arbitrary direction ($\vartheta_{s},\psi_{s}$) and an
idler photon in the corresponding ($\vartheta_{i},\psi_{i}$)
direction. Fixing $\vartheta_{s}$ and $\psi_{s}$, the emitted pair
can be described by the following output state $|\psi\rangle_{s,i}^{\rm ref}$ [compare Eq. (14)]:
$$\displaystyle|\psi\rangle_{s,i}^{\rm ref}$$
$$\displaystyle=$$
$$\displaystyle-\frac{i}{2c^{8}}\int_{0}^{\infty}\,\omega_{s}^{2}d\omega_{s}\int%
_{0}^{\infty}\,\omega_{i}^{2}d\omega_{i}\sqrt{\omega_{s}\omega_{i}}$$
(25)
$$\displaystyle\mbox{}{\cal E}_{p}(\omega_{s}+\omega_{i})\sum_{l=1}^{N}\max({\bf
d%
}^{(l)})L_{l}\hat{a}_{s}^{\dagger}(\omega_{s})\hat{a}_{i}^{\dagger}(\omega_{i}%
)|{\rm vac}\rangle,$$
where $\hat{a}_{s}^{\dagger}(\omega_{s})$ [$\hat{a}_{i}^{\dagger}(\omega_{i})$] stands for a signal- [idler-]
field creation operator outside the reference structure. The
function $\max$ used in Eq. (25) gives the maximum value
among the elements of tensor ${\bf d}^{(l)}$.
Using the reference structure a relative density $\eta_{s,ab}^{\alpha,\beta}({\bf\Omega}_{s})$ of mean signal-photon
numbers belonging to the state $|\psi\rangle_{s_{a},i_{b}}^{\alpha,\beta}$ can naturally be defined
as
$$\eta_{s,ab}^{\alpha,\beta}({\bf\Omega}_{s})=\frac{n_{s,ab}^{\alpha,\beta}({\bf%
\Omega}_{s})}{n_{s}^{\rm ref}(\omega_{s})},$$
(26)
where the signal-field photon-number density $n_{s,ab}^{\alpha,\beta}$ is written in Eq. (21). The
reference signal-field photon-number density $n_{s}^{\rm ref}$
characterizes the state $|\psi\rangle_{s,i}^{\rm ref}$ in
Eq. (25) and does not depend on the propagation angles $\vartheta_{s}$ and $\psi_{s}$.
Properties of photon pairs in the time domain are complementary to
those found in the spectral domain and belong to important
characteristics of photon pairs. We can mention signal- and
idler-field photon fluxes or coincidence-count interference
patterns in different kinds of interferometers as examples. These
photon-pair properties can be investigated, e.g., using the
formulas contained in Sec. IIC of Ref. Peřina Jr.
et al. (2006) even in
this spatial vectorial model.
Numerical calculations that follow are performed for a cw pump
field with a Gaussian transverse profile described as follows:
$$\displaystyle{\cal E}_{p}(\omega_{p})$$
$$\displaystyle=$$
$$\displaystyle\xi_{p}\delta(\omega_{p}-\omega_{p}^{0}),$$
$$\displaystyle{\cal E}_{p}^{\rm tr}(k_{x},k_{y})$$
$$\displaystyle=$$
$$\displaystyle\frac{r_{p}}{\sqrt{2\pi}}\exp\left[\frac{r^{2}_{p}(k_{x}^{2}+k_{y%
}^{2})}{4}\right].$$
(27)
In Eq. (27), $\xi_{p}$ gives an amplitude of the pump field
with the carrying frequency $\omega_{p}^{0}$; $r_{p}$ denotes the
width of amplitude transverse profile. Normalization of the
function ${\cal E}_{p}^{\rm tr}$ is such that $\int dk_{x}\int dk_{y}|{\cal E}_{p}^{\rm tr}(k_{x},k_{y})|^{2}=1$. In case of cw
pumping, there occur formulas in the above equations that contain
a formal expression $\delta^{2}(\omega)$. This expression has to
be replaced by the expression $2T/(2\pi)\delta(\omega)$, where
the detection interval extends over $(-T,T)$ and the limit $T\rightarrow\infty$ may be considered.
III Design of an efficient layered structure
Here we consider structures with odd numbers $N$ of layers made
of two kinds of materials. Layers of material $b$ of length $l_{b}$ are sandwiched by layers of material $a$ having lengths $l_{a}$. Also pumping at a defined carrying frequency $\omega_{p}^{0}$
and impinging on the structure at normal incidence is assumed. The
down-converted signal and idler fields are assumed to have nearly
degenerate frequencies.
We suggest a method for designing an efficient layered structure
from the point of view of, in general, three-mode nonlinear
interaction. It is based on two observations:
•
An efficient nonlinear process occurs provided that all three
nonlinearly interacting fields lie inside their transmission
peaks. This behavior originates in the fact that the
electric-field amplitudes of monochromatic fields with frequencies
in transmission peaks are enhanced inside the structure owing to
constructive interference of back-scattered light. It follows from
the band-gap theory that the closer the transmission peak to a
band gap, the greater the electric-field amplitudes.
•
As numerical calculations have revealed, the overlap integral over the
amplitudes of three interacting fields giving the strength of the
effective nonlinear interaction [see Eq. (1)] vanishes if
the signal- and idler-field amplitudes along the structure are the
same. This means that photon-pair states degenerate in
frequencies, emitted in symmetric directions and having the same
polarizations cannot be generated.
We note that these facts have been found crucial in designing
layered structures efficient for second-harmonic generation
Scalora et al. (1997). Considering a collinear interaction, the
requirements are even more strict because of only one propagation
direction. This requires specific approaches that rely, e.g., on
tuning the frequencies of transmission peaks by changing the index
of refraction of one type of the layers Scalora et al. (1997).
Alternatively, non-collinear second-harmonic generation has been
considered. In this case, there exists one free parameter (radial
emission angle) that can be varied in order to find an efficient
structure. This represents an equivalent problem to that
considered here and can be treated by the developed systematic
approach.
Returning back to the considered layered structures they are
characterized by three parameters: number $N$ of layers and
lengths $l_{a}$ and $l_{b}$ of these layers. A detailed inspection
has revealed that the number $N$ of layers significantly
determines the number of generated photon pairs as well as angular
extensions of the emission areas of photons in a pair (see
Fig. 3 bellow). The greater the number $N$ of layers
the greater the number of generated photon pairs and also the
smaller the angular extensions of emission areas. From practical
point of view, the number $N$ of layers is approximately fixed
considering these dependencies. Having the number $N$ of layers
fixed, there remain two adjustable parameters - layers’ lengths $l_{a}$ and $l_{b}$. However, these two lengths cannot be chosen
arbitrarily because the pump field at the carrying frequency $\omega_{p}^{0}$ has to be in a transmission peak.
Let us fix the lengths $l_{a}$ and $l_{b}$ (together with the
number $N$ of layers) for a moment and determine the spectral
intensity transmission $T(\omega)$ along the $+z$ direction
(of the pump-field propagation) using, e.g., the transfer-matrix
formalism Yeh (1988). Increasing the frequency $\omega$
there occur forbidden bands one following the other. The
difference in central frequencies of the neighbor forbidden bands
is roughly the same in accord with the band-gap theory. This may
be convenient for spectrally nearly degenerate SPDC provided that
the process of SPDC can be tuned such that the pump-field
frequency lies inside a transmission peak close to the second
forbidden band whereas the signal- and idler-field frequencies are
inside the transmission peaks near to the first forbidden band. We
note that all three interacting fields can accommodate themselves
either to the transmission peaks above or bellow the forbidden
bands. This helps to obey the quite strong requirements of the
three-field interaction by varying the signal-field emission angle
$\vartheta_{s}$. However, suitable conditions can only be revealed
numerically.
Now we return back to the problem in which the pump-field carrying
frequency $\omega_{p}^{0}$ is fixed. As follows from the
considerations of the previous paragraph, useful structures are
those that have the first upper or the first lower transmission
peak near to the second forbidden band at the pump-field carrying
frequency $\omega_{p}^{0}$. Inspection of the behavior of structures
with different layers’ lengths $l_{a}$ and $l_{b}$ has shown that
there exists a system of curves in the plane spanned by layers’
lengths $l_{a}$ and $l_{b}$ that provides the required
transmission peaks at the frequency $\omega_{p}^{0}$. One curve
corresponds to the first lower transmission peak of the second
forbidden band. Similarly, another curve is associated with the
first upper transmission peak of the second forbidden band. They
can be revealed as follows using the scaling property of
diffraction phenomena in optics.
We consider a fictitious dispersion-free structure with the
indexes of refraction appropriate for the pump-field frequency $\omega_{p}^{0}$ and define the corresponding optical lengths $l_{a}^{\rm opt}$ and $l_{b}^{\rm opt}$. It can be shown that the
ratio $L=l_{b}^{{\rm opt}}/l_{a}^{{\rm opt}}$ of optical lengths
represents a suitable variable for parametrization of these
curves. Suitable optical lengths $l_{a}^{\rm opt}$ and $l_{b}^{\rm opt}$ for a given value of the ratio $L$ can be revealed easily
using the scaling property of diffraction phenomena. We fix the
value of length $l_{a}^{\rm opt}$ to, e.g., $l_{a}^{{\rm opt},0}=\lambda_{p}^{0}/2=\pi c/\omega_{p}^{0}$. This gives the unit length of
diffraction phenomena that can be, in principle, chosen
arbitrarily. As the ratio $L$ is given, the optical length $l_{b}^{{\rm opt},0}$ is derived as $l_{b}^{{\rm opt},0}=Ll_{a}^{{\rm opt},0}$ and we can calculate the spectral intensity
transmission $T_{p}(\omega_{p})$ for this structure. We further
identify the frequency $\omega_{p}^{\rm max}$ of the first lower
(or upper) transmission peak of the second forbidden band. Then we
have to ’transfer’ the actual frequency $\omega_{p}^{\rm max}$ of
the transmission peak to the required frequency $\omega_{p}^{0}$
using the scaling property. The scaling property provides the
layers’ optical lengths in the form: $l_{a}^{\rm opt}=l_{a}^{{\rm opt},0}\omega_{p}^{0}/\omega_{p}^{\rm max}$ and $l_{b}^{\rm opt}=Ll_{a}^{\rm opt}$. We note that the obtained lengths differ for the
lower and the upper transmission peaks.
In the next step we move along the obtained two curves (for the
upper and the lower transmission peaks) parameterized by the ratio
$L$ and numerically analyze the structures. The maximum $\eta_{s}^{\rm max}$ of relative density $\eta_{s}(\omega_{s},\vartheta_{s},\psi_{s}^{0})$ of mean signal-photon
numbers taken over the signal-field frequency $\omega_{s}$ and
radial emission angle $\vartheta_{s}$ assuming the fixed
signal-field azimuthal emission angle $\psi_{s}^{0}$ has been found
a suitable quantity for monitoring efficiency of the nonlinear
process. The signal- and idler-field polarizations are assumed to
be fixed. The greater the value of maximum $\eta_{s}^{\rm max}$ of
relative density the closer the signal- and idler-field
transmission peaks to the first forbidden band. A curve giving the
dependence of maximum $\eta_{s}^{\rm max}$ of relative density on
the ratio $L$ is thus a good indicator for choosing suitable
layers’ lengths. We note that the curve depends on polarization
properties of the signal and idler fields as well as the azimuthal
signal-field emission angle $\psi_{s}^{0}$. It is also possible to
monitor another quantity in this procedure, e.g., the overall mean
photon-pair number $N$ given in Eq. (24). Or the density
of mean signal-photon numbers emitted for a fixed azimuthal
emission angle $\psi_{s}^{0}$ and determined as $\int_{-\pi/2}^{\pi/2}\sin(\vartheta_{s})d\vartheta_{s}n_{s}^{\rm tr}(\vartheta%
_{s},\psi_{s})$ using the formula in Eq. (22) may
be considered.
As an example, the dependence of maximum $\eta_{s}^{\perp,\|,\rm max}$ of relative density of two photons propagating along the $+z$ axis on the ratio $L$ for the first lower and the first
upper pump-field transmission peak of the second forbidden band of
the structure composed of 11 and 101 layers, respectively, is
plotted in Fig. 2 (for details, see below). The curve in
Fig. 2(a) appropriate for the structure with 11 layers is
continuous and rather flat. This means that there exists a whole
continuous set of structures giving an efficient nonlinear
process. On the other hand, efficient nonlinear structures with $N=101$ layers are found only in peaks of the curve in
Fig. 2(b). This is a consequence of complex interference
of back-scattered light inside these structures. We note that the
appropriate transmission peaks in all three interacting fields do
not necessarily exist for all values of the ratio $L$. Also the
greatest values of maximum $\eta_{s}^{\perp,\|,\rm max}$ of the
relative density are found for the values of ratio $L$ in a
certain restricted region. As the graph in Fig. 2(b)
shows, the greatest values of maximum $\eta_{s}^{\perp,\|,\rm max}$ occur around $L=0.5$.
The curves in Fig. 2 demonstrate a ’weak’ increase of
values of the maximum $\eta_{s}^{\perp,\|,\rm max}$ of relative
signal-field density with the increasing values of the number $N$ of layers. As the densities $n_{s}^{\rm ref}$ of the mean
signal-photon numbers of the reference structure are linearly
proportional to the second power of the length of nonlinear
material inside the structure, the densities $n_{s}$ of the mean
signal-photon numbers increase better than $N^{2}$. This is very
important, because the mean number of emitted photon pairs
increases less than the second power of the length of nonlinear
material in other structures producing photon pairs like bulk
crystals, periodically-poled crystals or wave-guiding structures
because of ’longitudinal’ phase-matching conditions. Nearly linear
dependence of the mean photon-pair numbers on the length of a
nonlinear structure is commonly observed.
IV Intensity transverse profiles and correlated areas
We consider three examples of layered structures made of GaN/AlN
that demonstrate typical features of the emitted photon pairs.
They generate photon pairs with different polarizations of the
signal and idler photons. They are designed in such a way that
three different mechanisms of splitting the correlated area are
observed. The pump field propagating at normal incidence at the
carrying wavelength $\lambda_{p}^{0}=400$ nm is assumed. The
signal and idler fields are emitted around the degenerate
wavelengths $\lambda_{s}^{0}=\lambda_{i}^{0}=800$ nm in
non-collinear geometry. The GaN and AlN layers are positioned such
that their optical axes are perpendicular to the boundaries. The
considered structures differ in the number $N$ of layers ($N=11$, 51 and 101). This results in different signal-field intensity
transverse profiles plotted in Fig. 3. Whereas the
shortest structure with $N=11$ layers has only one emission
area, the longest one with $N=101$ layers already forms five
emission rings. We note that transverse profiles of the
down-converted fields can also be efficiently tailored by
modifying the pump-beam profile Walborn et al. (2004).
The analyzed structures have certain properties in common. Namely,
they cannot generate proton pairs in nearly collinear geometries
because of the symmetry that cancels the overlap integral. As for
correlated areas, their azimuthal spreads $\Delta\psi_{i}$ depend
on the width $r_{p}$ of transverse pump-field profile
Molina-Terriza
et al. (2005); Hamar et al. (2010). The radial spreads $\Delta\vartheta_{i}$ depend also on the geometry of the layered
structure; the greater the number $N$ of layers, the smaller the
correlated area. In more detail, the considered structures behave
as follows.
IV.1 Structure with 11 layers
The structure is composed of 6 nonlinear GaN layers 90.14 nm long
and 5 linear AlN layers 74.92 nm long. Despite a relatively small
number of layers, photonic band gaps are already formed though the
bottoms of forbidden bands have intensity transmissions around
0.3. For $\psi_{s}=0$ deg, an efficient SPDC occurs for
polarizations (TE,TM,TE) and (TE,TE,TM) for the
(pump,signal,idler) fields due to the nonlinear coefficient $d(1,1,3)$. The fields are tuned to the first lower transmission
peaks of the first and second forbidden bands, so the efficiency
of SPDC is nearly optimal. The signal and idler transmission peaks
are spectrally broad and, as a consequence, we observe one large
emission area in the signal-field transverse plane that extends
from cca $\vartheta_{s}=20$ deg to 60 deg [see Fig. 3(a)].
The relative density $\eta_{s}$ of mean signal-photon numbers as
plotted in Fig. 4 reveals that the emitted photons have
nearly degenerate frequencies and relatively broad spectra. The
shape of relative density $\eta_{s}$ as shown in Fig. 4
is caused by a different radial dependence of intensity
transmissions $T_{\rm TE}$ and $T_{\rm TM}$. In the area
around $\psi_{s}=\pm 90$ deg there occurs no photon-pair
generation because of geometric reasons (the propagation of signal
and idler fields as TM waves is not supported). Compared to one
GaN layer $6\times 90.14$ nm long, the structure gives cca 2
times greater relative densities $\eta_{s}$.
The correlated area of an idler photon for a collimated pump beam
is composed of several ’islands’ localized around different
idler-field radial emission angles $\vartheta_{i}$ [see
Fig. 5(a)]. This breaking is caused by the zig-zag
movement of two photons inside the structure after being generated
in one layer. The comparison of correlated areas for collimated
and focused pump beams using, e.g., the graphs in
Fig. 5(a) and (b) reveals that the spread of correlated
area along the azimuthal emission angle $\psi_{i}$ depends
strongly on the amount of focusing the pump beam. On the other
hand, the shape of correlated area in the radial emission angle $\vartheta_{i}$ is blurred by a focused pump beam keeping the
overall spread of correlated area roughly the same.
IV.2 Structure with 51 layers
The second structure contains 26 nonlinear GaN layers 106.87 nm
long and 25 linear AlN layers 65.99 nm long. This structure has
already well-formed forbidden bands. As Fig. 3(b) shows
there occurs an efficient SPDC in two concentric rings in the
transverse plane. The structure is designed such that an efficient
nonlinear interaction for fields’ polarizations (TM,TM,TM) occurs
along the direction $\psi_{s}=0$ deg using the nonlinear
coefficient $d(2,2,3)$. The pump field lies in the first upper
transmission peak of the second forbidden band and the signal and
idler fields are in the first and second lower transmission peaks
of the first forbidden band. The relative density $\eta_{s}$ of
mean signal-photon numbers as plotted in Fig. 6
demonstrates that the signal-field (and also idler-field)
intensity spectra are composed of two symmetric peaks. There
occurs no photon-pair generation for degenerate signal- and
idler-field frequencies because of symmetry. This indicates the
generation of photon pairs in the state that is anti-symmetric
with respect to the exchange of the signal- and idler-field
frequencies. Anti-bunching of photons in a pair and
anti-coalescence (observed in a Hong-Ou-Mandel interferometer) are
the distinguished features of this state Peřina Jr.
et al. (2007a). The
graph in Fig. 3(b) reveals that an efficient SPDC occurs
also for azimuthal emission angles $\psi_{s}$ around $\pm 90$ deg. In this case, the signal photon is generated as TE wave and
the idler photon as TM wave or vice versa. The structure gives cca
50 times greater relative densities $\eta_{s}$ in comparison with
one GaN layer $26\times 106.87$ nm long.
As a consequence of two-peak structure of the signal- and
idler-field spectra there occurs splitting of the correlated area
of an idler photon, as documented in Fig. 7(a). The
overall correlated area is composed of two symmetric parts.
Because this splitting arises from symmetry, it survives even for
focused pump beams, as demonstrated in Fig. 7(b).
IV.3 Structure with 101 layers
The last structure is composed of 51 nonlinear GaN layers
106.42 nm long and 50 linear AlN layers 65.71 nm long. The pump
field lies in the first upper transmission peak of the second
forbidden band. A detailed comparison of the relative density $\eta_{s}$ of mean signal-photon numbers (see Fig. 8) with
intensity transmission spectra $T_{\rm TE}$ and $T_{\rm TM}$
valid for TE and TM waves, respectively, and obtained for $\psi=0$ deg (see Fig. 9) reveals the following. An efficient
SPDC occurs at intersections of $j$th and $(j+1)$th lower
transmissions peaks of the first forbidden band for $j=2,3,4,5,6$. This results in five concentric rings in the density $n_{s}^{\rm tr}$ of mean signal-photon numbers clearly visible in
Fig. 3(c). We note that the nonlinear coefficient $d(1,1,3)$ is exploited here and (TE,TM,TE) and (TE,TE,TM)
[(TM,TM,TM)] polarizations are suitable for the angles around $\psi_{s}=0$ deg [$\psi_{s}=\pm 90$ deg]. The relative density $\eta_{s}$ of mean signal-photon numbers indicates that the
signal-field intensity spectra are composed of two peaks of
different weights for $\psi_{s}=0$ deg. One peak is related to a
TE signal-field wave, whereas the second one arises from a TM
signal-field wave. Comparison with one GaN layer $51\times 106.42$ nm long reveals that the enhancement of optical fields inside
the structure results in an increase of the relative density $\eta_{s}$ cca 330 times.
A two-peak spectral structure leads to splitting of the correlated
area of the idler photon into two parts Peřina Jr.
et al. (2007b), as
shown in Fig. 10(a). However, two parts are not symmetric
in this case and they can also merge together provided that the
pump beam is sufficiently focused [see Fig. 10(b)]. The
distance between two parts in the idler-photon radial emission
angle $\vartheta_{i}$ increases with the increasing signal-photon
radial emission angle $\vartheta_{s}^{0}$. Whereas this distance
equals cca 0.6 deg for $\vartheta_{s}^{0}=23$ deg (the first
ring), it equals already cca 4 deg for $\vartheta_{s}^{0}=66$ deg
(the fifth ring). Idler photons found in different parts of the
correlated area differ in their polarizations.
The above investigated structures demonstrate main features of
photon pairs generated in nonlinear layered structures. These
structures allow to generate photon pairs entangled in
frequencies, polarizations as well as in emission directions.
Moreover, they allow tailoring of properties of photon pairs
varying namely the number of layers. A small amount of nonlinear
material inside them is compensated by an increase of
electric-field amplitudes originating in fields’ back-scattering.
V Conclusions
We have developed a spatial vectorial quantum model of spontaneous
parametric down-conversion in nonlinear layered structures. Photon
pairs generated in these structures can be entangled in
frequencies, polarizations as well as emission directions. Namely
entanglement in emission directions is important because it can be
efficiently tailored varying the number of layers. A method for
designing efficient layered structures has been suggested. Its
efficiency has been demonstrated suggesting three typical
structures. It has been shown that the number of generated
photon-pairs increases greater than the second power of the number
of layers. Signal-field intensity profiles and correlated areas in
the transverse plane have been analyzed in the designed
structures. Intensity profiles are typically composed of
concentric rings. The greater the number of layers, the greater
the number of rings. Correlated areas may be broken into several
parts because of three possible reasons: i) Zig-zag movement of
photons inside the structure, ii) Necessity to obey geometric
symmetries, and iii) Polarization-dependent properties. Splitting
of the correlated area arising from the geometric symmetry
survives even for a focused pump beam. Also states of photon pairs
showing anti-bunching and anti-coalescence can efficiently be
generated in layered structures. We believe that nonlinear layered
structures are potentially interesting namely as efficient sources
of photon-pair fields entangled in propagation directions which
might be useful, e.g., in ’parallel processing’ of quantum
information.
Acknowledgements.Support by projects IAA100100713 of GA AV ČR
and COST OC 09026, 1M06002 and Operational Program Research and
Development for Innovations - European Social Fund (project
CZ.1.05/2.1.00/03.0058) of the Ministry of Education of the Czech
Republic is acknowledged. The author thanks M. Centini for useful
discussions.
References
Hong et al. (1987)
C. K. Hong,
Z. Y. Ou, and
L. Mandel,
Phys. Rev. Lett. 59,
2044 (1987).
Mandel and Wolf (1995)
L. Mandel and
E. Wolf,
Optical Coherence and Quantum Optics
(Cambridge Univ. Press, Cambridge,
1995).
Bou (2000)
in The Physics of Quantum Information, edited by
D. Bouwmeester,
A. Ekert, and
A. Zeilinger
(Springer, Berlin, 2000).
Peřina et al. (1994)
J. Peřina,
Z. Hradil, and
B. Jurčo,
Quantum Optics and Fundamentals of Physics
(Kluwer, Dordrecht, 1994).
Bouwmeester et al. (1997)
D. Bouwmeester,
J. W. Pan,
K. Mattle,
M. Eibl,
H. Weinfurter,
and
A. Zeilinger,
Nature 390,
575 (1997).
Harris (2007)
S. E. Harris,
Phys. Rev. Lett. 98,
063602 (2007).
Brida et al. (2009)
G. Brida,
M. V. Chekhova,
I. P. Degiovanni,
M. Genovese,
G. K. Kitaeva,
A. Meda, and
O. A. Shumilkina,
Phys. Rev. Lett. 103,
193602 (2009).
Keller and Rubin (1997)
T. E. Keller and
M. H. Rubin,
Phys. Rev. A 56,
1534 (1997).
Peřina Jr.
et al. (1999)
J. Peřina Jr.,
A. V. Sergienko,
B. M. Jost,
B. E. A. Saleh,
and M. C. Teich,
Phys. Rev. A 59,
2359 (1999).
Joobeur et al. (1994)
A. Joobeur,
B. E. A. Saleh,
and M. C. Teich,
Physical Review A 50,
3349 (1994).
Joobeur et al. (1996)
A. Joobeur,
B. E. A. Saleh,
T. S. Larchuk,
and M. C. Teich,
Physical Review A 53,
4360 (1996).
Vallone et al. (2007)
G. Vallone,
E. Pomarico,
P. Mataloni,
F. De Martini,
and V. Berardi,
Phys. Rev. Lett. 98,
180502 (2007).
Monken et al. (1998)
C. H. Monken,
P. H. Souto Ribeiro,
and S. Padua,
Phys. Rev. A 57,
3123 (1998).
Walborn et al. (2004)
S. P. Walborn,
A. N. de Oliveira,
R. S. Thebaldi,
and C. H.
Monken, Phys. Rev. A
69, 023811
(2004).
Law and Eberly (2004)
C. K. Law and
J. H. Eberly,
Phys. Rev. Lett. 92,
127903 (2004).
Mair et al. (2001)
A. Mair,
A. Vaziri,
G. Weihs, and
A. Zeilinger,
Nature 412,
313 (2001).
Oemrawsingh et al. (2005)
S. S. R. Oemrawsingh,
X. Ma,
D. Voigt,
A. Aiello,
E. R. Eliel,
G. W. t Hooft,
and J. P.
Woerdman, Phys. Rev. Lett.
95, 240501
(2005).
Molina-Terriza
et al. (2005)
G. Molina-Terriza,
S. Minardi,
Y. Deyanova,
C. I. Osorio,
M. Hendrych, and
J. P. Torres,
Phys. Rev. A 72,
065802 (2005).
Jost et al. (1998)
B. M. Jost,
A. V. Sergienko,
A. F. Abouraddy,
B. E. A. Saleh,
and M. C. Teich,
Opt. Express 3,
81 (1998).
Haderka et al. (2005)
O. Haderka,
J. Peřina Jr.,
and M. Hamar,
J. Opt. B: Quantum Semiclass. Opt.
7, S572 (2005).
Rubin and Shih (2008)
M. H. Rubin and
Y. H. Shih,
Phys. Rev. A 78,
033836 (2008).
Brambilla et al. (2004)
E. Brambilla,
A. Gatti,
M. Bache, and
L. A. Lugiato,
Phys. Rev. A 69,
023802 (2004).
Jedrkiewicz et al. (2004)
O. Jedrkiewicz,
Y. K. Jiang,
E. Brambilla,
A. Gatti,
M. Bache,
L. A. Lugiato,
and
P. Di Trapani,
Phys. Rev. Lett. 93,
243601 (2004).
Migdall (1999)
A. Migdall,
Physics Today 41,
1 (1999).
Bruß and Lütkenhaus (2000)
D. Bruß and
N. Lütkenhaus, in
Applicable Algebra in Engineering, Communication
and Computing, Vol. 10 (Springer, Berlin,
2000), p. 383.
Kitaeva (2007)
G. K. Kitaeva,
Phys. Rev. A 76,
043841 (2007).
Svozilík and
Peřina Jr. (2009)
J. Svozilík
and
J. Peřina Jr.,
Phys. Rev. A 80,
023819 (2009).
Svozilík and
Peřina Jr. (2010)
J. Svozilík
and
J. Peřina Jr.,
Opt. Express 18,
27130 (2010).
URen et al. (2004)
A. B. URen,
C. Silberhorn,
K. Banaszek, and
I. A. Walmsley,
Phys. Rev. Lett. 93,
093601 (2004).
Spillane et al. (2007)
S. M. Spillane,
M. Fiorentino,
and R. G.
Beausoleil, Opt. Express
15, 8770 (2007).
Chen et al. (2009)
J. Chen,
A. J. Pearlman,
A. Ling,
J. Fan, and
A. Migdall,
Opt. Express 17,
6727 (2009).
Ber (2001)
in Nanoscale Linear and Nonlinear Optics, AIP Vol.
560, edited by
M. Bertolotti,
C. M. Bowden,
and C. Sibilia
(AIP, Melville, 2001).
Yablonovitch (1987)
E. Yablonovitch,
Phys. Rev. Lett. 58,
2059 (1987).
John (1987)
S. John, Phys.
Rev. Lett. 58, 2486
(1987).
Vamivakas et al. (2004)
A. N. Vamivakas,
B. E. A. Saleh,
A. V. Sergienko,
and M. C. Teich,
Phys. Rev. A 70,
043810 (2004).
Centini et al. (2005)
M. Centini,
J. Peřina Jr.,
L. Sciscione,
C. Sibilia,
M. Scalora,
M. J. Bloemer,
and
M. Bertolotti,
Phys. Rev. A 72,
033806 (2005).
Peřina Jr.
et al. (2006)
J. Peřina Jr.,
M. Centini,
C. Sibilia,
M. Bertolotti,
and M. Scalora,
Phys. Rev. A 73,
033823 (2006).
Li et al. (2005)
X. Li,
P. L. Voss,
J. E. Sharping,
and P. Kumar,
Phys. Rev. Lett. 94,
053601 (2005).
Fulconis et al. (2005)
J. Fulconis,
O. Alibart,
W. Wadsworth,
P. Russell, and
J. Rarity,
Opt. Express 13,
7572 (2005).
Fan et al. (2005)
J. Fan,
A. Migdall, and
L. J. Wang,
Opt. Lett. 30,
3368 (2005).
Abolghasem et al. (2009)
P. Abolghasem,
M. Hendrych,
X. Shi,
J. P. Torres,
and A. S. Helmy,
Opt. Lett. 34,
2000 (2009).
Svozilík et al. (2011)
J. Svozilík,
M. Hendrych,
A. S. Helmy, and
J. P. Torres,
Opt. Express 19,
3115 (2011).
Yeh (1988)
P. Yeh,
Optical Waves in Layered Media
(Wiley, New York, 1988).
Peřina Jr.
et al. (2009a)
J. Peřina Jr.,
M. Centini,
C. Sibilia, and
M. Bertolotti,
J. Russ. Laser Res. 30,
508 (2009a).
Peřina Jr.
et al. (2009b)
J. Peřina Jr.,
M. Centini,
C. Sibilia, and
M. Bertolotti,
Phys. Rev. A 80,
033844 (2009b).
Peřina Jr.
et al. (2007a)
J. Peřina Jr.,
M. Centini,
C. Sibilia,
M. Bertolotti,
and M. Scalora,
Phys. Rev. A 75,
013805 (2007a).
Vogel et al. (2001)
W. Vogel,
D. G. Welsch,
and
S. Walentowicz,
Quantum Optics (Wiley-VCH,
Weinheim, 2001).
Peřina Jr.
et al. (2009c)
J. Peřina Jr.,
A. Lukš,
O. Haderka, and
M. Scalora,
Phys. Rev. Lett. 103,
063902 (2009c).
Peřina Jr.
et al. (2009d)
J. Peřina Jr.,
A. Lukš, and
O. Haderka,
Phys. Rev. A 80,
043837 (2009d).
Scalora et al. (1997)
M. Scalora,
M. J. Bloemer,
A. S. Manka,
J. P. Dowling,
C. M. Bowden,
R. Viswanathan,
and J. W. Haus,
Phys. Rev. A 56,
3166 (1997).
Hamar et al. (2010)
M. Hamar,
J. Peřina Jr.,
O. Haderka, and
V. Michálek,
Phys. Rev. A 81,
043827 (2010).
Peřina Jr.
et al. (2007b)
J. Peřina Jr.,
M. Centini,
C. Sibilia,
M. Bertolotti,
and M. Scalora,
in Conference on Coherence and Quantum Optics
(Optical Society of America,
2007b), p. CSuA6. |
Supersymmetric quantum mechanics on the lattice:
III. Simulations and algorithms
David Baumgartner and Urs Wenger
Albert Einstein Center for Fundamental Physics,
Institute for Theoretical Physics, University of Bern,
Sidlerstrasse 5, CH–3012 Bern, Switzerland
()
Abstract
In the fermion loop formulation the contributions to the partition
function naturally separate into topological equivalence classes
with a definite sign. This separation forms the basis for an
efficient fermion update algorithm using a fluctuating open fermion
string. It guarantees sufficient tunnelling between the topological
sectors, and hence provides a solution to the fermion sign problem
affecting systems with broken supersymmetry. Moreover, the algorithm
shows no critical slowing down even in the massless limit and can
hence handle the massless Goldstino mode emerging in the
supersymmetry broken phase. In this paper – the third in a series
of three – we present the details of the simulation algorithm and
demonstrate its efficiency by means of a few examples.
1 Introduction
The reformulation of supersymmetric quantum mechanics on the lattice
in terms of bosonic and fermionic bonds as derived in the first paper
of our series [1] provides a perfect setup for
Monte Carlo simulations. First of all, the reduction in complexity by
going from continuous to discrete variables is enormous. More
specifically though, expressing the Grassmann fields in terms of
fermionic bonds avoids the expensive calculation of the fermion
determinant and allows the use of special algorithms for which
critical slowing down is essentially absent
[2, 3] and simulations are possible even
in the massless limit [4]. This is of particular
importance for systems with broken supersymmetry, since the physics of
those is driven by the massless Goldstino mode. In the present paper
– the last in a series of three – we describe in detail such an
algorithm and demonstrate its efficiency. Since the model can be
solved exactly at finite lattice spacing by means of transfer
matrices, as discussed in the second paper of our series
[5], there is in principle no need for
numerical simulations. Hence, the present paper rather constitutes a
feasibility study to test the practicability and efficiency of the
proposed simulation algorithm for the quantum mechanical system in the
bond formulation. In that sense it also serves as a preparation for
the application of the algorithm, in particular the fermionic part, in
more complex situations, such as in supersymmetric Yang-Mills quantum
mechanics [6], in the ${\cal N}=1$ Wess-Zumino
model [7, 8, 9]
or in the supersymmetric nonlinear $O(N)$ sigma model
[10]. The advantage of the application of the
algorithm in the quantum mechanical model presented here is of course
the fact that the correctness of the algorithm can be crosschecked
with the exact results from the transfer matrix approach, and that the
algorithm can hence be validated in detail.
There is another rather pedagogical reason which motivates to consider
a new simulation algorithm for quantum mechanics in the bond
formulation. Often, simple quantum mechanical systems such as the
harmonic and anharmonic oscillator are used to introduce the path
integral approach. Similarly, the systems also provide a pedagogical
context in which various Monte Carlo simulation algorithms can be
illustrated and discussed, see for example [11] for an
early example. However, it turns out that the standard Metropolis
algorithms and even more advanced algorithms such as the
overrelaxation or heat bath algorithm become extremely inefficient
towards the continuum limit. This has to do with the usual critical
slowing down of the simulations towards that limit, and for the
anharmonic oscillator also with the suppressed tunnelling at small
lattice spacing. The algorithms presented here do not suffer from
these deficiencies, because they eliminate critical slowing down. In
addition, in the bond formulation the $\mathbb{Z}_{2}$-symmetry $\phi\rightarrow-\phi$ is exactly maintained for each bond configuration.
Last but not least, the numerical simulations presented here serve as
a test of the practicability of the solution of the fermion sign
problem proposed in [4] and discussed further in
the first paper of our series [1]. The solution
is based on two ingredients. Firstly, the lattice regulates the
vanishing Witten index and therefore also the sign problem. Secondly,
the fermion loop formulation provides a tool to handle the fluctuating
sign, because it naturally separates the contributions to the
partition function into topological equivalence classes, each
possessing a definite sign. Nevertheless, it is a priori not clear
whether the lattice artefacts and the statistical fluctuations can be
kept under sufficient control in a practical simulation. The
statistical fluctuations of the sign are essentially determined by the
amount of tunnelling between the topological sectors, i.e., between
the fermionic and bosonic vacuum. In order for the fermion update
algorithm to be a true solution to the sign problem, it must guarantee
a sufficiently efficient tunnelling rate. The results in this paper
demonstrate that this is indeed the case. Not surprisingly, the open
fermion string algorithm discussed here has proven to be extremely
successful in the ${\cal N}=1$ Wess-Zumino model
[9] in which the fermion sign problem is
prevailing.
Of course, supersymmetric quantum mechanics has already been simulated
on the lattice in various setups using standard algorithms, cf. for
example
[12, 13, 14, 15, 16, 17, 18, 19, 20]. However,
the bond formulation together with the simulation algorithm presented
here brings the numerical nonperturbative calculations to a new,
unprecedented level of accuracy. In that sense, the results presented
here and partly in [4] serve as a benchmark
against which new formulations or simulation algorithms can be tested.
The present paper is organised as follows. In Section
2 we construct in detail an algorithm designed for
updating the bosonic and fermionic bond configurations. The
discussion includes the explicit update steps and the derivation of
the corresponding acceptance ratios. Their evaluation requires the
calculation of site weight ratios which turn out to become numerically
unstable for large site occupation numbers. Therefore, in Section
3 we present a computational strategy which allows to
evaluate the ratios for arbitrarily large occupation numbers. In
Section 4, we then present the results obtained using
the proposed algorithm. The simulations are for the same
discretisation schemes and superpotentials we used in the previous two
papers [1, 5]. Since this
section is merely meant as a validation of the algorithm, the
discussion of the physics behind the results is kept short and we
refer to the exact results in [5] for a more
throrough discussion.
2 Simulation algorithm
We start our discussion from the partition function of supersymmetric
quantum mechanics on the lattice written as a sum over all allowed,
possibly constrained bond configurations $\mathcal{C}=\{n^{b}_{i}(x),n^{f}(x)\}$ in
the configuration space $\mathcal{Z}$,
$$Z=\sum_{\mathcal{C}\subset\mathcal{Z}}W_{F}(\mathcal{C})\,,$$
(1)
where the fermion number $F=0,1$ is determined by the fermionic bond
configuration $\{n^{f}(x)\}$ with $n^{f}(x)=F,\,\forall x$, and the weight $W_{F}(\mathcal{C})$
of a configuration is given by
$$W_{F}(\mathcal{C})=\prod_{x}\left(\prod_{i}\frac{w_{i}^{n^{b}_{i}(x)}}{n^{b}_{%
i}(x)!}\right)\prod_{x}Q_{F}(N(x))\,.$$
(2)
Here, $x$ denotes the sites of the lattice and $i$ labels the various
types of bosonic bonds $b_{i}$ with $i\in\{j\rightarrow k\,|\,j,k\in\mathbbm{N}\}$. The corresponding bosonic bond weights are denoted
by $w_{i}$ and $n_{i}^{b}(x)\in\mathbbm{N}_{0}$ is the occupation number of
the bond $b_{i}$ connecting the sites $x$ and $x+1$. The site weight
$Q_{F}$ depends on the site occupation number, i.e., the total number of
bosonic bonds connected to site $x$,
$$N(x)=\sum_{j,k}\left(j\cdot n^{b}_{j\rightarrow k}(x)+k\cdot n^{b}_{j%
\rightarrow k}(x-1)\right)$$
(3)
and is given by
$$Q_{F}(N)=\int_{-\infty}^{\infty}d\phi\ \phi^{N}\mathrm{e}^{-V(\phi)}M(\phi)^{1%
-F}\,.$$
(4)
In Section 3 we will discuss in detail the
computational strategy necessary to reliably evaluate ratios of these
integrals for arbitrary and possibly large site occupation
numbers. The type of bonds $b_{i}$, the weights $w_{i}$ as well as the
potential $V(\phi)$ and the monomer term $M(\phi)$ in eq.(4) depend on the specifics of the chosen discretisation and
the superpotential $P(\phi)$. We refer to the appendix of our first
paper [1] for a compilation of the
discretisations and superpotentials considered in our series.
As mentioned above, the bond configurations $\mathcal{C}=\{n^{b}_{i}(x),n^{f}(x)\}$
are possibly constrained. In particular we have the local fermionic
constraints
$$n^{f}(x-1)=n^{f}(x)\,$$
(5)
while the local bosonic constraints
$$N(x)=0\mod 2\,$$
(6)
may or may not be present depending on the bosonic symmetries of the system.
The challenge of updating constrained bond configurations lies
precisely in the difficulty to maintain the constraints while moving
efficiently through the configuration space $\mathcal{Z}$. In
[21] Prokof’ev and Svistunov proposed to extend the
constrained bosonic bond configuration space by introducing local
sources which explicitly violate the constraints. The so-called worm
algorithm then probes the extended configuration space by moving the
local violations around the lattice, thereby sampling directly
the bosonic correlation function corresponding to the sources
introduced. The contact with the original configuration space $\mathcal{Z}$ is
established when the violations annihilate each other, e.g. when
moving to the same site on the lattice, such that the bond
configuration fulfills again all constraints.
In [2] the idea has been extended to fermionic
systems expressed in terms of fermionic bonds. The fermionic
constraint in eq.(5) allows only either an
empty or a completely filled fermion bond configuration. The
difficulty for the direct application of the worm idea to the
fermionic system lies in the fact that the introduction of the fermionic
source term $\psi_{x}\overline{\psi}_{x}$ is incompatible with the presence of
the fermion loop at site $x$. A simple solution is to allow the
unphysical situation of the site $x$ being occupied by a propagating
fermion and two additional sources. Such a configuration violates the
Pauli exclusion principle and does not contribute to any physical
observable. In the Grassmann path integral such a configurations
indeed vanishes trivially.
In order to be more explicit it is necessary to introduce the bond
configuration spaces of bosonic and fermionic two-point correlation
functions, $\mathcal{G}^{b}_{F}$ and $\mathcal{G}^{f}$, respectively, following the notation
in our first paper [1]. Bond configurations in
$\mathcal{G}^{b}_{F}$ contribute to the non-normalised bosonic two-point function
according to
$$g^{b}_{F}(x_{1}-x_{2})\equiv\langle\!\langle\phi_{x_{1}}\phi_{x_{2}}\rangle\!%
\rangle_{F}=\sum_{\mathcal{C}\subset\mathcal{G}^{b}_{F}}\left(\prod_{x}\frac{Q%
_{F}(N(x)+\delta_{x,x_{1}}+\delta_{x,x_{2}})}{Q_{F}(N(x))}\right)\cdot W_{F}(%
\mathcal{C})\,,$$
(7)
while the configurations in $\mathcal{G}^{f}$ contribute to the non-normalised
fermionic two-point function as
$$g^{f}(x_{1}-x_{2})\equiv\langle\!\langle\psi_{x_{1}}\overline{\psi}_{x_{2}}%
\rangle\!\rangle=\sum_{\mathcal{C}\subset\mathcal{G}^{f}}\left[\prod_{x\in%
\phantom{\notin}\!\!\!\!\mathcal{F}}\frac{Q_{1}(N(x))}{Q_{0}(N(x))}\right]%
\cdot W_{0}(\mathcal{C})\,,$$
(8)
where $\cal{F}$ denotes the set of lattice sites belonging to the open
fermion string associated with the fermionic correlation function. The
key point of the bosonic and fermionic updating algorithm is that the
bond configurations for $g^{b}_{F}(0)$, $g^{f}(0)$ and $Z_{F}$ have identical
bond elements. As a consequence, statistics for $g^{b,f}$ and $Z$ can
be accummulated in the same Monte Carlo process. If the bosonic
constraints in eq.(6) are not present,
e.g. for superpotentials with broken supersymmetry, the equivalence of
bond configurations even extends to $g^{b}_{F}(x)$, i.e., $\mathcal{Z}_{F}=\mathcal{G}^{b}_{F}$. The movements from one configuration space to the other are
induced by introducing or removing bosonic or fermionic sources
according to the scheme given in figure 8 of our first paper
[1]. For convenience we reproduce it here in
figure 1.
In the following we will now discuss in detail the various updating
steps which establish explicitly the connection between the bond
configuration spaces $\mathcal{G}^{f},\mathcal{Z}_{F},\mathcal{G}^{b}_{F}$ and in addition move the
system within $\mathcal{G}^{f}$ and $\mathcal{G}^{b}_{F}$. The moves are generated by a Monte
Carlo process with probabilities given by the weights of the
configurations in eqs.(2), (7)
and (8). In particular, we derive the transition
probabilities $P_{X}(\mathcal{C}\rightarrow\mathcal{C}^{\prime})$ for the transition $X$
from bond configuration $\mathcal{C}$ to $\mathcal{C}^{\prime}$, which is then accepted by the
usual Metropolis prescription
$$P_{acc}(\mathcal{C}\rightarrow\mathcal{C}^{\prime})=\mathop{\rm min}\{1,P_{X}(%
\mathcal{C}\rightarrow\mathcal{C}^{\prime})\}.$$
(9)
In order to simplify the discussion we select the update from $\mathcal{Z}_{F}$
to $\mathcal{G}^{b}_{F}$ or $\mathcal{G}^{f}$ with equal probability which is balanced by
corresponding proposal probabilities to select between moving in
$\mathcal{G}^{f}$ and $\mathcal{G}^{b}_{F}$ or returning to $\mathcal{Z}_{F}$.
2.1 Updating the fermionic bond configuration
Here we discuss the various update steps which moves the system within
$\mathcal{G}^{f}$ and relate the bond configurations spaces $\mathcal{Z}_{0}$ and $\mathcal{Z}_{1}$ via
$\mathcal{G}^{f}$.
Moves within $\mathcal{G}^{f}$ are induced by shifting $\overline{\psi}$ by one lattice
spacing from site $x$ to site $x+1$, and vice versa, while keeping the
other source $\psi$ fixed. Such an update step is graphically
illustrated in figure 2 and is called ‘shift’
update step.
A shift in forward direction, $x\rightarrow x+1$, automatically
involves the removal of the fermionic bond $b^{f}(x)$, whereas a shift
in backward direction, $x+1\rightarrow x$, requires the addition of a
new fermionic bond $b^{f}(x)$. Both directions are proposed with equal
probability $1/2$ and are hence balanced against each other as long as
the new site does not coincide with the position of the source
$\psi$. The formula in eq.(8) provides us with the
acceptance ratios
$$\displaystyle P_{\mathrm{sh}}(x+1\rightarrow x)$$
$$\displaystyle=$$
$$\displaystyle\displaystyle\frac{Q_{1}(N(x))}{Q_{0}(N(x))}\,,$$
(10)
$$\displaystyle P_{\mathrm{sh}}(x\rightarrow x+1)$$
$$\displaystyle=$$
$$\displaystyle\displaystyle\frac{Q_{0}(N(x))}{Q_{1}(N(x))}\,.$$
(11)
Now let us consider the case when we propose to shift the source
$\overline{\psi}$ forward to the site $x$ where the source $\psi$ is present,
as depicted in the upper half of figure 3.
The forward shift update step $x-1\rightarrow x$ is balanced with the
backward shift update step $x\rightarrow x-1$. This backward shift,
however, is proposed with probability 1 instead of probability 1/2
since the shift of $\overline{\psi}$ from $x\rightarrow x+1$ would involve
the creation of an open fermion string around the entire lattice. The
asymmetry in the proposition probabilities is balanced by the choice
of the probability $p_{\mathrm{rm}}=1/2$ to remove the sources
$\psi\overline{\psi}$, such that we find the acceptance ratio for a shift in
backward direction $x\rightarrow x-1$ to be the same as in
eq.(10), namely
$$P_{\mathrm{sh}}(x\rightarrow x-1)=\frac{Q_{1}(N(x-1))}{Q_{0}(N(x-1))}.$$
(12)
The shift step is balanced with the corresponding one in forward
direction with acceptance ratio as given in eq.(11),
$$P_{\mathrm{sh}}(x-1\rightarrow x)=\displaystyle\frac{Q_{0}(N(x-1))}{Q_{1}(N(x-%
1))}$$
(13)
The step from $\mathcal{G}^{f}$ to $\mathcal{Z}_{0}$ and vice versa is induced by
introducing or removing a pair of fermionic sources $\psi\overline{\psi}$
at site $x$, respectively. It is called ‘put/remove’ update step and
is graphically illustrated in the lower half of figure
3.
The removal of the fermionic sources is suggested with probability
$p_{\mathrm{rm}}=1/2$ and is balanced on one side by the probability
to add bosonic sources, and on the other by the probability to shift
one of the sources and hence move within $\mathcal{C}^{f}$. Because the
‘put/remove’ update step does not alter the fermionic bond
configuration, we have $\mathcal{Z}_{0}=\mathcal{G}^{f}(0)$. On the other hand it adds or
removes a fermionic monomer term $M(\phi)$ at site $x$. The relative
weight of the configurations with or without this term is given by
$Q_{1}/Q_{0}$ and the acceptance ratios on a lattice with $L_{t}$ sites
become
$$\displaystyle P_{\mathrm{rm}}(x\rightarrow\varnothing)$$
$$\displaystyle=$$
$$\displaystyle\frac{2}{L_{t}}\displaystyle\frac{Q_{0}(N(x))}{Q_{1}(N(x))},$$
(14)
$$\displaystyle P_{\mathrm{put}}(\varnothing\rightarrow x)$$
$$\displaystyle=$$
$$\displaystyle\frac{L_{t}}{2}\displaystyle\frac{Q_{1}(N(x))}{Q_{0}(N(x))}\,.$$
(15)
The factor $L_{t}$ compensates for the proposition probability to choose
lattice site $x$ out of $L_{t}$ possibilities when putting the sources,
while the factor 2 compensates for the asymmetric shift proposal
probability when moving $\overline{\psi}$ from $x$ to $x-1$, since the
shift of $\overline{\psi}$ from $x$ to $x+1$ is not allowed.
Next we consider the shift update step for the case when the source
$\overline{\psi}$ at site $x+1$ is shifted backwards to site $x$ which is
already occupied by the sink $\psi$. The step is graphically
illustrated in the upper half of figure
4.
While the resulting fermion bond configuration is a valid one (it
belongs to $\mathcal{Z}_{1}$), the whole fermion configuration including the
source and the sink represents an unphysical situation, and in fact
does not contribute to any physical observable, as discussed
before. Therefore, such a backward shift from $\mathcal{G}^{f}$ to $\mathcal{Z}_{1}$,
essentially closing the open fermion string, automatically induces the
removal of the fermionic source and sink pair $\psi\overline{\psi}$ from site
$x$ as illustrated in the lower half of figure
4. Such a step is called a hybrid
‘shift/remove’ update step. Of course, the step is balanced with a
hybrid ‘put/shift’ update step when the additional fermionic sink and
source variables are put on a closed fermion loop at the site $x$. As
usual, the acceptance ratios for the hybrid update steps can be read
off from the weights of the configurations involved and yield
$$\displaystyle P_{\mathrm{sh/rm}}(x+1\rightarrow x\rightarrow\varnothing)$$
$$\displaystyle=$$
$$\displaystyle\frac{2}{L_{t}},$$
(16)
$$\displaystyle P_{\mathrm{put/sh}}(\varnothing\rightarrow x\rightarrow x+1)$$
$$\displaystyle=$$
$$\displaystyle\frac{L_{t}}{2}.$$
(17)
The factor $L_{t}$ compensates for the proposition probability to choose
the same lattice site $x$ when putting the sources $\psi\overline{\psi}$ back
on the lattice, whereas the factor $2$ compensates for the proposition
probability to shift in forward or backward direction when the fermion
string is still open. Note that there are no ratios of $Q$-weights
involved, since no monomer term is added or removed by the hybrid
shift/remove update step.
To complete our discussion of the fermionic bond update, we note that
the algorithm provides improved estimators for the fermionic two-point
function $g^{f}(x)$ and the partition functions $Z_{F}$. Because the
algorithm samples directly the configuration space $\mathcal{G}^{f}$, every open
fermion string configuration contributes unity to the stochastic Monte
Carlo estimator for $g^{f}(x)$. To be precise we have
$$g^{f}(x|\mathcal{C}\in\mathcal{G}^{f})=\delta_{x,x_{1}-x_{2}}\,,$$
(18)
where $x_{1}$ and $x_{2}$ are the end and starting point of the open
fermion string, i.e., the positions of the sink $\psi$ and the source
$\overline{\psi}$, respectively. Similarly, every bond configuration in
$\mathcal{Z}_{F}$ is generated with its proper weight and hence contributes unity
to the stochastic estimator for $Z_{F}$, i.e., the Monte Carlo
estimator for $Z_{F}$ is simply
$$Z_{F}(\mathcal{C}\in\mathcal{Z}_{F})=1\,.$$
(19)
Finally we note that the factors of $L_{t}$ appearing in the acceptance
ratios above may become inconvenient in practice, especially towards
the continuum limit when $L_{t}\rightarrow\infty$. The factors only
occur when contact between $\mathcal{Z}_{F}$ and $\mathcal{G}^{f}$ is made, i.e., they are
responsible for getting the relative normalisation between $Z_{F}$ and
$g^{f}$ right. However, since we make use of translational invariance in
eq.(18) the factors of $L_{t}$ are in fact cancelled
and can hence be omitted.
2.2 Updating the bosonic bond configuration
In this section we now discuss the update steps which relate the bond
configuration spaces $\mathcal{Z}_{F}$ and $\mathcal{G}^{b}_{F}$ for a fixed fermionic bond
configuration with fermion number $F=0,1$.
We point out that for an arbitrary superpotential there are in general
no restrictions on the bosonic bond configurations. This is for
example the case for the superpotential $P_{b}$ which we consider in our
series of papers, cf. eq.(53). In contrast,
the superpotential $P_{u}$ in eq.(54) yields
the local constraint $N(x)=0\mod 2$ on the site occupation number,
due to the parity symmetry $\phi\rightarrow-\phi$. In the following
discussion, we always present the generic case first, and then specify
the modifications or simplifications due to the constraint. In
analogy to the fermionic bond update, the ‘put/remove’ and the ‘shift’
updates are the main steps for updating the bosonic bond configurations.
The ‘put/remove’ step introduces or removes one or two sources $\phi$,
while the ‘shift’ step shifts the sources by one lattice spacing. If
there are no restrictions on the bond configuration, we are free to
decide for each Monte Carlo step whether to proceed by a ‘remove’ update
or a ‘shift’ update. With probability $p_{\mathrm{rm}}$, we propose to
remove the sources from the lattice, while the proposition to continue
the worm update with a ‘shift’ step is chosen with probability $1-p_{\mathrm{rm}}$.
The step from $\mathcal{G}_{F}^{b}$ to $\mathcal{Z}_{F}$ (and vica versa) is induced by
removing (or introducing) a bosonic source $\phi$ at sites $x_{1}$ and
$x_{2}$, with $x_{1}=x_{2}$ not excluded. The step does not alter the bond
configuration, but only the site occupation numbers at sites $x_{1}$ and
$x_{2}$. Thus, only the ratios of the site weights $Q_{F}$ are involved in
the acceptance probability for e.g. the ‘remove’ step,
$$\displaystyle P_{\mathrm{rm}}(x_{1},x_{2}\rightarrow\varnothing)$$
$$\displaystyle=$$
$$\displaystyle\!\!\!\left\{\begin{array}[]{ll}\displaystyle\frac{1}{p_{\mathrm{%
rm}}L_{t}^{2}}\frac{Q_{F}(N(x_{1})-2)}{Q_{F}(N(x_{1}))}&\mathrm{if}\ x_{1}=x_{%
2},\\
\displaystyle\frac{1}{p_{\mathrm{rm}}L_{t}^{2}}\frac{Q_{F}(N(x_{1})-1)}{Q_{F}(%
N(x_{1}))}\frac{Q_{F}(N(x_{2})-1)}{Q_{F}(N(x_{2}))}&\mathrm{if}\ x_{1}\neq x_{%
2}.\end{array}\right.$$
(20)
The prefactor $1/(p_{\mathrm{rm}}L_{t}^{2})$ is motivated as follows. The
factor $1/L_{t}^{2}$ balances the probability for the proposition of putting
the bosonic sources at the sites $x_{1}$ and $x_{2}$ when re-entering the
configuration space $\mathcal{G}^{b}_{F}$, while the factor $1/p_{\mathrm{rm}}$
balances the proposition probability for the choice of proceeding by
the shift update instead of the remove update, as discussed above.
The acceptance ratios for re-entering the configuration space $\mathcal{G}^{b}_{F}$
from $\mathcal{Z}_{F}$ are given by
$$\displaystyle P_{\mathrm{put}}(\varnothing\rightarrow x_{1},x_{2})$$
$$\displaystyle=$$
$$\displaystyle\left\{\begin{array}[]{ll}\displaystyle p_{\mathrm{rm}}L_{t}^{2}%
\frac{Q_{F}(N(x_{1})+2)}{Q_{F}(N(x_{1}))}&\mathrm{if}\ x_{1}=x_{2},\\
\displaystyle p_{\mathrm{rm}}L_{t}^{2}\frac{Q_{F}(N(x_{1})+1)}{Q_{F}(N(x_{1}))%
}\frac{Q_{F}(N(x_{2})+1)}{Q_{F}(N(x_{2}))}&\mathrm{if}\ x_{1}\neq x_{2}.\end{%
array}\right.$$
(21)
Two remarks are in order. Firstly, if there are no constraints on the
bond configuration, one can in principle introduce just a single
source $\phi$ which subsequently is shifted around. In effect, the
algorithm then samples the one-point function which in this situation
is indeed nonvanishing. Secondly, we note that if the constraint $N=0\mod 2$ is in place, the two sources can only be placed or removed
when $x_{1}=x_{2}$. As a consequence, only the first of the two
acceptance ratios in eq.(20) and
eq.(21) are relevant, while the second
ones are zero by definition.
Next, we discuss the bosonic ‘shift’ update. With this step we now
change the bosonic bond configuration. Shifting the source from site
$x$ to a next neighbouring site $y$ is always associated with an
increase or a decrease of the bosonic bond occupation number between
the sites $x$ and $y$ by one. Whether or not the occupation number is
increased or decreased is decided with probability $1/2$. Similarly,
the source can move forward or backward, and we propose both
directions with equal probability $1/2$. In addition, when there are
several types of bosonic bonds $b_{i}$ with $i\in\{j\rightarrow k|j,k\in\mathbb{N}\}$, we need to decide in each step which
bond is updated. We do so by choosing the proposition probabilities
$p_{j\rightarrow k}$ with $\sum_{j,k}p_{j\rightarrow k}=1$. However, because the proposals are completely symmetric, these
probabilities do not affect the acceptance ratios. In the following,
we will use the shorthand notation
$$\displaystyle n^{j\rightarrow k}_{xy}$$
$$\displaystyle=$$
$$\displaystyle\left\{\begin{array}[]{ll}n^{b}_{j\rightarrow k}(x)&\mathrm{if}\ %
y=x+1,\\
n^{b}_{j\rightarrow k}(y)&\mathrm{if}\ y=x-1,\end{array}\right.$$
(22)
for the occupation number of the bosonic bonds $b_{j\rightarrow k}$
between the sites $x$ and $y$. The shifts $x\rightarrow y$ and
$n^{j\rightarrow k}_{xy}\rightarrow n^{j\rightarrow k}_{xy}+1$ are
balanced with shifts $y\rightarrow x$ and $n^{j\rightarrow k}_{xy}\rightarrow n^{j\rightarrow k}_{xy}-1$, which gives the acceptance
ratios
$$\displaystyle P_{\mathrm{sh}}(x\rightarrow y,n^{j\rightarrow k}_{xy}%
\rightarrow n^{j\rightarrow k}_{xy}+1)\\
\displaystyle=\!\left\{\!\!\!\begin{array}[]{ll}\displaystyle\frac{w_{j%
\rightarrow k}}{n^{j\rightarrow k}_{xy}+1}\frac{Q_{F}(N(x)+j-1)}{Q_{F}(N(x))}%
\cdot\frac{Q_{F}(N(y)+k+1)}{Q_{F}(N(y))}&\mathrm{if}\ y=x+1,\\
\displaystyle\frac{w_{j\rightarrow k}}{n^{j\rightarrow k}_{xy}+1}\frac{Q_{F}(N%
(x)+k-1)}{Q_{F}(N(x))}\cdot\frac{Q_{F}(N(y)+j+1)}{Q_{F}(N(y))}&\mathrm{if}\ y=%
x-1,\\
\end{array}\right.$$
(23)
$$\displaystyle P_{\mathrm{sh}}(x\rightarrow y,n^{j\rightarrow k}_{xy}%
\rightarrow n^{j\rightarrow k}_{xy}-1)\\
\displaystyle=\!\left\{\!\!\!\begin{array}[]{ll}\displaystyle\frac{n^{j%
\rightarrow k}_{xy}}{w_{j\rightarrow k}}\frac{Q_{F}(N(x)-j-1)}{Q_{F}(N(x))}%
\cdot\frac{Q_{F}(N(y)-k+1)}{Q_{F}(N(y))}&\mathrm{if}\ y=x+1,\\
\displaystyle\frac{n^{j\rightarrow k}_{xy}}{w_{j\rightarrow k}}\frac{Q_{F}(N(x%
)-k-1)}{Q_{F}(N(x))}\cdot\frac{Q_{F}(N(y)-j+1)}{Q_{F}(N(y))}&\mathrm{if}\ y=x-%
1.\end{array}\right.$$
(24)
Of course, these generic ratios simplify considerably for the specific
bonds $b_{i},i\in\{1\rightarrow 1,1\rightarrow 2,1\rightarrow 3\}$ relevant for the superpotentials considered in our series of
papers. For example, the acceptance ratios for updating the bond
$b_{1\rightarrow 1}$ read
$$\displaystyle P_{\mathrm{sh}}(x\rightarrow y,n^{1\rightarrow 1}_{xy}%
\rightarrow n^{1\rightarrow 1}_{xy}+1)$$
$$\displaystyle=$$
$$\displaystyle\frac{w_{1\rightarrow 1}}{n^{1\rightarrow 1}_{xy}+1}\cdot\frac{Q_%
{F}(N(y)+2)}{Q_{F}(N(y))},$$
(25)
$$\displaystyle P_{\mathrm{sh}}(x\rightarrow y,n^{1\rightarrow 1}_{xy}%
\rightarrow n^{1\rightarrow 1}_{xy}-1)$$
$$\displaystyle=$$
$$\displaystyle\frac{n^{1\rightarrow 1}_{xy}}{w_{1\rightarrow 1}}\cdot\frac{Q_{F%
}(N(x)-2)}{Q_{F}(N(x))}\,.$$
(26)
Because the bond is symmetric, there is no need to distinguish whether
$y=x+1$ or $y=x-1$.
To complete the discussion of the bosonic bond update, we point out
that the algorithm again provides improved estimators for the bosonic
two-point function $g^{b}_{F}(x)$ and the partition functions $Z_{F}$. As in
the fermionic case, the algorithm samples directly the configuration
space $\mathcal{G}^{b}_{F}$ with the correct weighting when the sources are
present. Therefore, every configuration contributes unity to the
stochastic Monte Carlo estimator for $g^{b}_{F}(x)$, and we have
$$g_{F}^{b}(x|\mathcal{G}^{b}_{F})=\delta_{x_{1}-x_{2},x}\,,$$
(27)
where $x_{1}$ and $x_{2}$ are the positions of the two sources. Whenever
the bosonic update decides to remove the sources, we have a
configuration in $\mathcal{Z}_{F}$ and hence a contribution of unity to the
stochastic estimator for $Z_{F}$, that is, we have
$$Z_{F}(\mathcal{C}\in\mathcal{Z}_{F})=1\,.$$
(28)
In complete analogy to the fermionic update we note that the factors
of $L_{t}$ appearing in the acceptance ratios of the ‘put/remove’ step
can be compensated by adjusting the overall normalisation of the
two-point function, e.g. by making use of translational invariance.
3 Calculation of the site weight ratios
In order to calculate the weight of a bond configuration, it is
necessary to know the site weights
$$Q_{F}(n)=\int_{-\infty}^{\infty}d\phi\ \phi^{n}\mathrm{e}^{-V(\phi)}M(\phi)^{1%
-F},$$
(29)
where $V(\phi)$ and $M(\phi)$ depend on the superpotential and the
discretisation employed, and $F=0,1$ is the fermion number, for
arbitrary values of the site occupation number $n$. The values of $n$
required in practice are usually limited to $\mathcal{O}(10^{3})$. However, it
turns out that even for moderate values of $n$ of order $\mathcal{O}(100)$ the
site weights $Q_{F}(n)$ can quickly grow larger than $10^{100}$ or
more. As a consequence, the calculation of the site weights quickly
becomes numerically unstable for growing $n$. In fact, even for simple
potentials when the weights can be calculated analytically in terms of
confluent hypergeometric functions, the numerical evaluation of these
functions is difficult for large $n$, and even specialised libraries
such as the ones available in Wolfram’s Mathematica
[22] appear not to be accurate enough.
Fortunately, for the Monte Carlo simulations we only need ratios of
the site weights, such as $Q_{F}(n+1)/Q_{F}(n),Q_{F}(n+2)/Q_{F}(n)$ and
$Q_{1}(n)/Q_{0}(n)$, and these ratios usually do not become larger than
$\mathcal{O}(10)$ even for large $n$. In addition, also the transfer matrix
elements can be rewritten in terms of these ratios as discussed in th
appendix of our second paper of the series [5].
Therefore, we now present a numerically stable computational strategy
to calculate the site weight ratios reliably for arbitrary values of
the site occupation numbers.
We start by defining an arbitrary polynomial superpotential
$$P(\phi)=\sum_{i=0}^{p}p_{i}\phi^{i},$$
(30)
and the corresponding bosonic self-interaction potential $V(\phi)$ as
well as the monomer weight $M(\phi)$,
$$V(\phi)=\sum_{i=0}^{2(p-1)}k_{i}\phi^{i},\quad\quad M(\phi)=\sum_{i=0}^{p-2}m_%
{i}\phi^{i}.$$
(31)
Explicitly, the weights in each sector are then given by
$$Q_{1}(n)=\int_{-\infty}^{\infty}d\phi\ \phi^{n}\mathrm{e}^{-V(\phi)}$$
(32)
and
$$Q_{0}(n)=\sum_{i=0}^{p-2}m_{i}Q_{1}(n+i).$$
(33)
For convenience we also define the ratios of the site weights
$Q_{F}(n)$,
$$\displaystyle R^{\prime}_{F}(n)$$
$$\displaystyle=$$
$$\displaystyle\frac{Q_{F}(n+1)}{Q_{F}(n)},$$
(34)
$$\displaystyle R_{F}(n)$$
$$\displaystyle=$$
$$\displaystyle\frac{Q_{F}(n+2)}{Q_{F}(n)},$$
(35)
$$\displaystyle R_{m}(n)$$
$$\displaystyle=$$
$$\displaystyle\frac{Q_{0}(n)}{Q_{1}(n)}$$
(36)
which are used for the acceptance ratios in the Monte Carlo
simulations. In principle, only the ratios $R_{1}^{\prime}(n)$ need to be
calculated since all other ratios can be derived from those. For
example, $R_{1}(n)$ can be expressed in terms of $R_{1}^{\prime}(n)$ as
$$R_{1}(n)=R_{1}^{\prime}(n+1)R_{1}^{\prime}(n)\,,$$
(37)
but since in some cases $Q_{1}(n\,\text{odd})=0$ the introduction of
$R_{1}(2n)$ is nevertheless necessary. $R_{m}(n)$ can be expressed via the
ratios $R_{1}(n)$ and $R_{1}^{\prime}(n)$ and appropriate products thereof,
$$R_{m}(n)=m_{0}+R_{1}^{\prime}(n)\left(m_{1}+R_{1}(n+2)\left(m_{3}+\ldots\right%
)\right)+R_{1}(n)\left(m_{2}+R_{1}(n+2)\left(m_{4}+\ldots\right)\right),$$
(38)
and the ratios $R_{0}^{\prime}(n)$ and $R_{0}(n)$ via $R_{m}(n)$, $R_{1}(n)$ and $R_{1}^{\prime}(n)$ by
$$\displaystyle R_{0}(n)$$
$$\displaystyle=$$
$$\displaystyle\frac{R_{m}(n+2)}{R_{m}(n)}R_{1}(n),$$
(39)
$$\displaystyle R_{0}^{\prime}(n)$$
$$\displaystyle=$$
$$\displaystyle\frac{R_{m}(n+1)}{R_{m}(n)}R_{1}^{\prime}(n).$$
(40)
First, we now discuss how to gain numerical stability for the special
case of an even superpotential $P(\phi)$. In a second step we will
then adapt the idea to treat the somewhat more subtle case of an
arbitrary superpotential.
3.1 Even superpotential
Unbroken supersymmetric quantum mechanics requires a superpotential
$P(\phi)$ with $\deg(P(\phi))$ = 0 mod 2. In particular, in our series
of papers we investigate the superpotential
$$P(\phi)=p_{2}\phi^{2}+p_{4}\phi^{4}$$
(41)
which is symmetric w.r.t. the parity transformation $\phi\rightarrow-\phi$. As a consequence of the symmetry, $Q_{F}(n\,\text{odd})=0$
for both $F=0,1$ and the ratios $R_{F}^{\prime}(n)$ need not be considered –
instead, it is sufficient to determine $R_{1}(2n)$ with $n\in\mathbb{N}_{0}$ only.
For the the potential $V(\phi)$ we then have the form
$$V(\phi)=k_{2}\phi^{2}+k_{4}\phi^{4}+k_{6}\phi^{6}\,,$$
(42)
consistent with both the standard discretisation and the $Q$-exact
one. To keep the integrals numerically under control, for fixed $n$
we apply a variable transformation $\phi\rightarrow\phi/\widetilde{\phi}$ to obtain rescaled weights $\widetilde{Q}_{1}(2n)$ as
$$Q_{1}(2n)=\widetilde{\phi}\,^{2n+1}\widetilde{Q}_{1}(2n).$$
(43)
Since we have $Q_{1}(2n)\geq 0$, we can choose the rescaling factor to
be $\widetilde{\phi}=Q_{1}(2n)^{1/(2n+1)}$ and the rescaled weight
becomes $\widetilde{Q}_{1}(2n)=1$. Calculating the ratio of rescaled
weights as
$$\widetilde{R}_{1}(2n)=\frac{\widetilde{Q}_{1}(2n+2)}{\widetilde{Q}_{1}(2n)}=%
\widetilde{Q}_{1}(2n+2),$$
(44)
where both integrals $\widetilde{Q}_{1}(2n+2)$ and
$\widetilde{Q}_{1}(2n)$ are rescaled with the same factor
$\widetilde{\phi}=Q_{1}(2n)^{1/(2n+1)}$, we find that
$$R_{1}(2n)=\widetilde{\phi}\,^{2}\,\widetilde{R}_{1}(2n)\,.$$
(45)
In addition, the rescaled weight $\widetilde{Q}_{1}(2n+2)$ is now of
$\mathcal{O}(1)$ and can be evaluated reliably via numerical integration. So
if we start by integrating directly the numerically stable site
weights $Q_{1}(0)$ and $Q_{1}(2)$, we can recursively generate ratios
$R_{1}(2n)$ with higher and higher $n$. Note that after each calculation
of a ratio $R_{1}(2n)$, one needs to update the rescaling factor
$\widetilde{\phi}\rightarrow\widetilde{\phi}\,^{\prime}$. This can be
achieved most easily via
$$\widetilde{\phi}\,^{\prime}=\widetilde{\phi}\,^{\frac{2n+1}{2n+3}}\,R_{1}(2n)^%
{\frac{1}{2n+3}}.$$
(46)
Our procedure guarantees that all involved quantities are of
$\mathcal{O}(1)$. Once all ratios $R_{1}(2n)$ are known, one can calculate the
ratios $R_{m}(2n)$, noting that for the specific superpotential we
consider, eq.(38) simplifies to
$$R_{m}(2n)=m_{0}+m_{2}R_{1}(2n).$$
(47)
The calculation of the ratios $R_{0}(2n)$ as given in eq.(39) is
then straightforward.
3.2 Arbitrary Superpotential
In the context of broken supersymmetric quantum mechanics, one
encounters superpotentials with $\deg(P(\phi))$ = 1 mod 2. Therefore,
we now adapt the procedure from above to superpotentials of this
form. For simplicity, we restrict ourselves to the odd superpotential
we consider as the example in our series of papers,
$$P(\phi)=\sum_{i=1}^{3}p_{i}\phi^{i}.$$
(48)
If at least one of the coefficients $p_{1}$ and $p_{2}$ is nonzero, which
is always the case for the superpotentials we use, $V(\phi)$ reads
$$V(\phi)=k_{1}\phi+k_{2}\phi^{2}+k_{3}\phi^{3}+k_{4}\phi^{4},$$
(49)
and at least one of the coefficients $k_{1}$ and $k_{3}$ is nonzero
either. This has a two important consequences. Firstly, the moments
defined in eq.(32) are nonzero for $n$ odd, from which it
follows that the ratios $R_{F}^{\prime}(n)$ defined in eq.(36) have to
be calculated as well. Secondly, the weights $Q_{1}(n)$ are no longer
necessarily positive. It turns out, however, that for all pratical
purposes it does not affect the simulations. We will discuss this
further in Section 4.
For the evaluation of the integrals, we apply the same variable
transformation $\phi\rightarrow\phi/\widetilde{\phi}$ as before,
such that we have rescaled weights $\widetilde{Q}_{1}(n)$ given by
$$Q_{1}(n)=\widetilde{\phi}\,^{n+1}\,\widetilde{Q}_{1}(n).$$
(50)
We now choose $\widetilde{\phi}=|Q_{1}(n)|^{1/(n+1)}\cdot\mathop{\rm sgn}(Q_{1}(n))$. Then, the integral becomes $\widetilde{Q}_{1}(n)=1$
again as before. Furthermore, defining the rescaled ratios
$\widetilde{R}_{1}^{\prime}(n)$ to be
$$\widetilde{R}_{1}^{\prime}(n)=\frac{\widetilde{Q}_{1}(n+1)}{\widetilde{Q}_{1}(%
n)}=\widetilde{Q}_{1}(n+1),$$
(51)
where both integrals $\widetilde{Q}_{1}(n+1)$ and $\widetilde{Q}_{1}(n)$
are rescaled with the same factor $\widetilde{\phi}=|Q_{1}(n)|^{1/(n+1)}\cdot\mathop{\rm sgn}(Q_{1}(n))$, we find $R_{1}^{\prime}(n)=\widetilde{\phi}\,\widetilde{R}_{1}^{\prime}(n)$. We proceed analogously to the case of the even
superpotential by recursive iteration, with the only exception that we
generate the ratios $R_{1}^{\prime}(n)$ instead of the ratios $R_{1}(n)$. The
update for the rescaling factor $\widetilde{\phi}\rightarrow\widetilde{\phi}\,^{\prime}$ is done via
$$\widetilde{\phi}\,^{\prime}=|\widetilde{\phi}|^{\frac{n+1}{n+2}}\,|R_{1}^{%
\prime}(n)|^{\frac{1}{n+2}}\cdot\mathop{\rm sgn}(R_{1}^{\prime}(n)).$$
(52)
Once all the ratios $R_{1}^{\prime}(n)$ are known, one can calculate the ratios
$R_{1}(n)$ via eq.(37), the ratios $R_{m}(n)$ via eq.(38),
and the ratios $R_{0}(n)$ and $R_{0}^{\prime}(n)$ via eq.(39) and
(40), respectively.
4 Results of the Monte Carlo simulations
The results in this section are merely thought of as a proof of the
feasibility of the algorithm and as a test of its efficiency.
Comparing the Monte Carlo results with the exact solution of the
system at finite lattice spacing provided in our second paper
[5] of course also serves as a validation for
the algorithm. We refer to that paper for a thorough discussion and
physical interpretation of the results.
For the following Monte Carlo simulations, we consider the same
superpotentials and discretisations as in the previous two papers. In
particular, we simulate the system using the action with counterterm
for both unbroken and broken supersymmetry as well as the $Q$-exact
action for unbroken supersymmetry. Details for the various actions can
be found in the first paper of our series. Here we only give the
details of the superpotentials for broken and unbroken supersymmetry,
respectively,
$$\displaystyle P_{b}(\phi)$$
$$\displaystyle=-\frac{\mu^{2}}{4\lambda}\phi+\frac{1}{3}\lambda\phi^{3}\,,$$
(53)
$$\displaystyle P_{u}(\phi)$$
$$\displaystyle=\frac{1}{2}\mu\phi^{2}+\frac{1}{4}g\phi^{4}\,,$$
(54)
and we recall that the continuum limit is taken by fixing the
dimensionful parameters $\mu,g,\lambda$ and $L$ while taking the
lattice spacing $a\rightarrow 0$. In practice, the dimensionless
ratios $f_{u}=g/\mu^{2}$ and $f_{b}=\lambda/\mu^{3/2}$ fix the couplings
and $\mu L$ the extent of the system in units of $\mu$, while $a\mu$
and $a/L$ are subsequently sent to zero. In analogy to the number of
sweeps for a standard Monte Carlo simulation, we count the number of
times the algorithm is in either one of the two configuration spaces
$\mathcal{Z}_{F},F=0,1$. The statistics for a simulation is therefore given by
$Z_{0}+Z_{1}=Z_{a}$.
First, we consider the standard discretisation with the superpotential
$P_{u}$ such that supersymmetry is unbroken. As a first observable, we
show the results for the bosonic and fermionic correlation functions
for $\mu L=10$, $L/a=60$ and $f_{u}=1$ for $Z_{a}=10^{7}$ in
figure 5.
This is essentially the same plot as figure 10(b) in our second paper
[5], but now with the additional data from the
Monte Carlo simulation and plotted on a logarithmic scale. Note that
we use the notation $x=t$ in accordance with
[5]. The simulation indeed reproduces the exact
result within very small statistical errors which demonstrate the
efficiency of the algorithm. The exponential error reduction is due to
the use of the improved estimators for the two-point function which
are available in the context of the worm algorithms. The improvement
is particularly impressive for the fermionic correlator where the
error reduction allows to follow the correlator over more than seven
orders of magnitude without loss of statistical significance. In fact
the relative error for the lowest value of the fermionic correlator is
still only $4\%$.
As a second example, we show the mass gaps for different $\mu L$ at a
coupling $f_{u}=1$ with statistics of $Z_{a}=10^{6}$ in figure
6.
The $\mu L$ considered are in the region where thermal effects are
negligible and essentially only $Z_{0}$ contributes to the total
partition function, such that $Z_{a}\simeq Z_{0}$. We extract the
masses from the asymptotic behaviour of the correlation functon at
large $t$, i.e., we extract the lowest energy gap. Because of the
extremely good signal-to-noise ratio the asymptotic behaviour can be
truly reached and, in doing so, systematic errors from contributions
of excited states are essentially excluded. Of course, we know from
our exact results that the overlap of the simple operators we use to
construct the two-point function is close to maximal. This is clearly
visible in figure 5 where we observe an almost
purely exponential decay for all $t/L$. Because the energy gaps are
independent of $\mu L$, they are expected to fall on top of each other
for all values of $\mu L$ at fixed lattice spacing $a\mu$. This is
indeed the case within our numerical accuracy, and the extracted
masses, when expressed in units of $\mu$, indeed extrapolate to the
correct zero-temperature continuum limit. The inset of figure
6 shows a detailed comparison of the
simulation results with our exact solution from
[5] represented by the dashed line and we
observe a beautiful agreement even very close to the continuum.
Next, we consider the action with counterterm and the superpotential
$P_{b}$ for which the supersymmetry is broken. In this case we encounter
an issue concerning the potential non-positivity of the weights which
we already mentioned in Section 3.2. This potentially dangerous sign problem is not of
fermionic origin, but is instead related to the bond formulation of
the bosonic degrees of freedom. As a matter of fact it occurs already
in the purely bosonic system, independent of the dimensionality of the
system. However, negative weights only occur in a region of parameter
space which becomes irrelevant towards the continuum limit. In that
sense, the sign problem is a lattice artefact and can be avoided
straightforwardly. Nevertheless, in order to eliminate any systematic
error we deal with this bosonic sign problem by incorporating the sign
of the configuration into the observables, even though it has no
practical consequences.
As a first observable in the broken case, we show the bosonic and
fermionic two-point functions, $\langle\phi_{t}\phi_{0}\rangle$ and
$\langle\psi_{t}\overline{\psi}_{0}\rangle$, for periodic and antiperiodic
b.c. for $\mu L=10$ at fixed coupling $f_{b}=1$ in figure
7 for a statistics of $Z_{a}=10^{8}$.
The exact results from [5] are shown as dashed
lines. The simulation yields results which agree with the exact
results within the very small statistical errors on the level of
1$\permil$. Note that the correlators for periodic and antiperiodic
b.c. are constructed a posteriori from the simulation results in the
bosonic and fermionic sectors $Z_{0}$ and $Z_{1}$, respectively, and it is
crucial to sample the relative weight between the two sectors
correctly in order to get the final values right. The relative
sampling is solely in the responsibility of the fermion simulation
algorithm. Our results in figure 7 show that the
open fermion string algorithm indeed transits sufficiently well
between the two sectors.
This statement can be made more quantitative by looking at the ratio
$Z_{p}/Z_{a}$ which represents the Witten index in our field theoretic
setup. From our exact results in [5] we expect
a nonzero Witten index at finite lattice spacing which however
extrapolates to zero in the continuum limit. So the behaviour of the
algorithm towards the continuum limit is particularly interesting,
because for vanishing lattice spacing the would-be Goldstino at finite
lattice spacing turns into a true, massless Goldstino. In such a
situation one usually encounters critical slowing down of the
simulation algorithms, such that the errors on the results grow large
and the results become unreliable. The massless Goldstino is directly
related to the tunnelling between the bosonic and the fermionic sector,
and the reproduction of a Witten index $W=0$ in the continuum with
small errors is hence a true demonstration of the efficiency of the
open fermion string algorithm to transit between the bosonic and
fermionic sector. In addition, we know from [5]
that the lattice artefacts are exponentially enhanced towards zero
temperature and it is interesting to see how the simulation algorithm
handles this situation at coarse lattice spacing.
In figure 8 we show the ratio $Z_{p}/Z_{a}$
as a function of the lattice spacing $a\mu$ for different values of
$\mu L$ at fixed coupling $f_{b}=1$.
For this quantity, too, the simulation yields results which agree with
the exact results within the small statistical errors. Moreover, the
efficiency of the algorithm does not appear to deteriorate towards the
continuum limit or for small values of $\mu L$ where the Witten index
is very close to zero. This can for example be seen from the fact that
the errors obtained with fixed statistics essentially remain constant
towards the continuum limit and are also independent of the system
size. This nicely demonstrates the efficiency of the algorithm also
for a system with broken supersymmetry.
The last system we investigate with the worm algorithm is unbroken
supersymmetry formulated with the $Q$-exact action111For Monte
Carlo simulations using the $Q$-exact action for broken
supersymmetry, we encounter the very same problems we ran into in
the transfer matrix approach. The bond occupation number grows
extremely large even on small lattices and for coarse lattice
spacings such that the generation of reliable results turns out to
be impossible.. We first consider the ratio of partition functions
$Z_{p}/Z_{a}$ which in the limit of $\mu L\rightarrow\infty$ yields
the Witten index. From a simulational point of view, the ratio
essentially calculates the fraction of configurations in sector $\mathcal{Z}_{0}$
versus the ones in $\mathcal{Z}_{1}$. For unbroken supersymmetry the system is
almost exclusively in the bosonic sector, and hence the ratio is very
close to one except when the size of the system becomes very small,
i.e., in the high temperature limit. Moreover, from our exact results
in [5] we know that the lattice artefacts in
this quantity are very small and the continuum limit is not very
interesting. For these reasons, we consider in figure
9 the dependence of the ratio $Z_{p}/Z_{a}$ on
$\mu L$ for different values of the lattice spacing $a/L$ with a
statistics of $Z_{a}=10^{8}$.
Also for this quantity, we find that the results agree with the exact
result within the very small statistical errors. Again, the open
fermion string algorithm proves to be very efficient even close to
$\mu L\simeq 0$ where the tunnelling from the bosonic to the fermionic
sector and vice versa becomes important and dominates the behaviour of
the system. Thus, even in this somewhat extreme situation of very high
temperature, the algorithm does not show any signs of critical slowing
down despite the fact that there is a quasi-zero mode in the system.
Note that the algorithm is capable of handling negative bare masses
independent of the discretisation used and
fig.9 is simply also an illustration of this
fact.
The last quantity we calculate are the lowest bosonic and fermionic
mass gaps for different $\mu L$ at fixed coupling of $f_{u}=1$ from a
statistics of $Z_{a}=10^{6}$. The mass gaps are extracted from the
two-point correlation functions exactly in the same way as before for
the standard action, and in figure 10 we show
the results of this analysis.
As expected, the masses for the boson and the fermion are indeed
indistinguishable within statistical errors. The degeneracy of the
masses at finite lattice spacing due to the $Q$-exactness of the
action emerges also for the results from Monte Carlo simulations. Note
that the chosen values for $\mu L$ lie well within the region where
thermal effects are negligible and the masses extrapolate nicely to
the correct zero-temperature continuum limit. The inset in figure
10 shows a detailed comparison with our exact
results from [5] and we again observe beautiful
agreement.
5 Conclusions
In this paper we present an algorithm for simulating ${\cal N}=2$
supersymmetric quantum mechanics on the lattice. The algorithm is
based on the reformulation of the system in terms of bosonic and
fermionic bonds, and in essence represents an efficient Monte Carlo
scheme for updating fermionic and bosonic bond configurations. The
updating of the fermionic degrees of freedom is of specific interest,
because this in in general the most challenging part of a
simulation. This is particularly true for systems with broken
supersymmetry, where standard simulation algorithms suffer from
critical slowing down due to the massless Goldstino mode. In addition,
these systems inevitably also suffer from a sign problem related to
the Goldstino and the vanishing Witten index.
In contrast, the fermion simulation algorithm proposed in
[2] eliminates critical slowing down by directly
sampling the fermionic two-point correlation function. It is based on
introducing a fluctuating open fermion string which efficiently
updates the bond configurations on all length scales up to the
correlation length associated with the fermionic correlation
function. As a consequence, the fermion string induces frequent
tunnellings between the bosonic and fermionic vacuum when that
correlation length becomes large. Since the two vacua contribute to
the partition function with opposite signs, the frequent tunnelling
guarantees sufficiently small statistical fluctuations for the average
sign, and hence a solution to the fermion sign problem. In fact, the
more severe the sign problem gets towards the continuum limit, the
more efficiently the algorithm tunnels between the bosonic and
fermionic sectors. This is of course due to the growing correlation
length associated with the vanishing Goldstino mass. The bosonic
degrees of freedom can be expressed in terms of bonds as
well. Therefore, we also give the details of an updating algorithm for
the bosonic bond configurations. Since we consider $Q$-exact
discretisations in addition to the standard one, the algorithm
involves updating generic types of bonds.
The simulation algorithm requires the calculation of the site weights
$Q_{F}(N)$. Their numerical evaluation, however, turns out to be
numerically unstable for large site occupation numbers $N$. Hence, in
Section 3, we devise a computational strategy which
allows to reliably evaluate the ratios of weights for arbitrarily
large occupation numbers. Since this is a generic problem occuring in
the bond formulation of field theories with real scalar fields, such
an computational scheme is useful also in other situations.
Finally, we present a selection of results obtained using the open
fermion string algorithm. We concentrate on two specific realisations
of supersymmetric quantum mechanics, one with broken and one with
unbroken supersymmetry. In addition, we consider both the standard and
the $Q$-exact discretisation. Since exact results are available at
finite lattice spacing from our investigation in
[5], we can benchmark our stochastic results
and directly validate them. The calculation of the bosonic and
fermionic correlation functions shows that they can be determined very
accurately over several orders of magnitude. This allows for a very
precise computation of the boson and fermion masses, the latter in
many cases with a smaller error than the former. In general, a
precision of 1‰ can be reached with a very modest
computational effort. In systems with broken supersymmetry it is
crucial that the simulation algorithm efficiently samples the relative
weights between the bosonic and fermionic sectors. Our results for the
partition function ratio $Z_{p}/Z_{a}$, i.e., the Witten index, show that
this is indeed the case. For fixed statistics, the errors do not grow
towards the continuum limit. In that limit the index gets very close
to zero and the sign problem would therefore be most
severe. Similarly, the error is essentially independent of the system
size, which shows that the sign problem is truly solved.
References
[1]
D. Baumgartner and U. Wenger, Supersymmetric quantum mechanics on the
lattice: I. Loop formulation, Nucl.Phys. B894 (2015) 223–253,
[arXiv:1412.5393].
[2]
U. Wenger, Efficient simulation of relativistic fermions via vertex
models, Phys.Rev. D80 (2009) 071503,
[arXiv:0812.3565].
[3]
U. Wenger, Simulating Wilson fermions without critical slowing down,
PoS LAT2009 (2009) 022,
[arXiv:0911.4099].
[4]
D. Baumgartner and U. Wenger, Simulation of supersymmetric models on the
lattice without a sign problem, PoS LAT2010 (2011)
[arXiv:1104.0213].
[5]
D. Baumgartner and U. Wenger, Supersymmetric quantum mechanics on the
lattice: II. Exact results, arXiv:1503.0523. Accepted for publication in Nucl.Phys. B.
[6]
K. Steinhauer and U. Wenger, Loop formulation of supersymmetric
Yang-Mills quantum mechanics, JHEP 1412 (2014) 044,
[arXiv:1410.0235].
[7]
D. Baumgartner, K. Steinhauer, and U. Wenger, Supersymmetry breaking on
the lattice: the N=1 Wess-Zumino model, PoS LATTICE2011 (2011)
253, [arXiv:1111.6042].
[8]
D. Baumgartner, K. Steinhauer, and U. Wenger, Spontaneous supersymmetry
breaking in the 2d N=1 Wess-Zumino model, PoS LATTICE2012
(2012) 043, [arXiv:1311.5089].
[9]
K. Steinhauer and U. Wenger, Spontaneous supersymmetry breaking in the
two-dimensional N=1 Wess-Zumino model, Phys.Rev.Lett. 113
(2014) 231601, [arXiv:1410.6665].
[10]
K. Steinhauer and U. Wenger, Loop formulation of the supersymmetric
nonlinear O(N) sigma model, PoS LATTICE2013 (2013) 092,
[arXiv:1311.5403].
[11]
M. Creutz and B. Freedman, A statistical approach to quantum mechanics,
Annals of Physics 132 (1981), no. 2 427 – 462.
[12]
S. Catterall and E. Gregory, A Lattice path integral for supersymmetric
quantum mechanics, Phys.Lett. B487 (2000) 349–356,
[hep-lat/0006013].
[13]
J. Giedt, R. Koniuk, E. Poppitz, and T. Yavin, Less naive about
supersymmetric lattice quantum mechanics, JHEP 0412 (2004)
033, [hep-lat/0410041].
[14]
A. Kirchberg, J. Lange, and A. Wipf, From the Dirac operator to
Wess-Zumino models on spatial lattices, Annals Phys. 316
(2005) 357–392, [hep-th/0407207].
[15]
G. Bergner, T. Kaestner, S. Uhlmann, and A. Wipf, Low-dimensional
Supersymmetric Lattice Models, Annals Phys. 323 (2008)
946–988, [arXiv:0705.2212].
[16]
T. Kaestner, G. Bergner, S. Uhlmann, A. Wipf, and C. Wozar, Supersymmetric lattice models in one and two dimensions, PoS LAT2007 (2007) 265, [arXiv:0709.0822].
[17]
C. Wozar and A. Wipf, Supersymmetry breaking in low dimensional models,
Annals Phys. 327 (2012) 774–807,
[arXiv:1107.3324].
[18]
I. Kanamori, H. Suzuki, and F. Sugino, Euclidean lattice simulation for
dynamical supersymmetry breaking, Phys.Rev. D77 (2008) 091502,
[arXiv:0711.2099].
[19]
I. Kanamori, F. Sugino, and H. Suzuki, Observing dynamical supersymmetry
breaking with Euclidean lattice simulations, Prog.Theor.Phys. 119 (2008) 797–827, [arXiv:0711.2132].
[20]
I. Kanamori, A Method for Measuring the Witten Index Using Lattice
Simulation, Nucl.Phys. B841 (2010) 426–447,
[arXiv:1006.2468].
[21]
N. V. Prokof’ev and B. V. Svistunov, Worm algorithms for classical
statistical models, Phys.Rev.Lett. 87 (2001) 160601,
[cond-mat/0103146].
[22]
Wolfram Research Inc., Mathematica Version 9.0, 2012. |
Solitons and geometrical structures
in a perfect fluid spacetime
Adara M. Blaga
()
Abstract
Geometrical aspects of a perfect fluid spacetime are described in terms of different curvature tensors and $\eta$-Ricci and $\eta$-Einstein solitons in a perfect fluid spacetime are determined. Conditions for the Ricci soliton to be steady, expanding or shrinking are also given. In a particular case when the potential vector field $\xi$ of the soliton is of gradient type, $\xi:=grad(f)$, we derive from the soliton equation a nonlinear second order PDE satisfied by $f$.
†† 2010 Mathematics Subject Classification.
53B50, 53C44, 53C50, 83C02.
†† Key words and phrases.
Ricci soliton, Einstein soliton, perfect fluid, Lorentz space.
1 Introduction
Lorentzian manifolds form a special subclass of pseudo-Riemannian manifolds of great importance in general relativity, where spacetime can be modeled as a $4$-dimensional Lorentzian manifold of signature (3,1) or, equivalently, (1,3).
Relativistic fluid models are of great interest in different branches of astrophysics, plasma physics, nuclear physics etc. Perfect fluids are often used in general relativity to model idealized distributions of matter, such as the interior of a star or an isotropic universe. Einstein’s gravitational equation can describe the behavior of a perfect fluid inside of a spherical object and the Friedmann-Lemaître-Robertson-Walker equations are used to describe the evolution of the universe. In general relativity, the source for the gravitational field is the energy-momentum tensor.
A perfect fluid can be completely characterized by its rest frame mass density and isotropic pressure. It has no shear stresses, viscosity, nor heat conduction and is characterized by an energy-momentum tensor of the form:
(1)
$$T(X,Y)=pg(X,Y)+(\sigma+p)\eta(X)\eta(Y),$$
for any $X$, $Y\in\chi(M)$, where $p$ is the isotropic pressure, $\sigma$ is the energy-density, $g$ is the metric tensor of Minkowski spacetime, $\xi:=\sharp(\eta)$ is the velocity vector of the fluid and $g(\xi,\xi)=-1$. If $\sigma=-p$,
the energy-momentum tensor is Lorentz-invariant ($T=-\sigma g$) and in this case we talk about the vacuum. If $\sigma=3p$, the medium is a radiation fluid.
The field equations governing the perfect fluid motion are Einstein’s gravitational equations:
(2)
$$kT(X,Y)=S(X,Y)+(\lambda-\frac{scal}{2})g(X,Y),$$
for any $X$, $Y\in\chi(M)$, where $\lambda$ is the cosmological constant,
$k$ is the gravitational constant (which can be taken $8\pi G$, with $G$ the universal gravitational constant), $S$ is the Ricci tensor and $scal$ is the scalar curvature of $g$. They are obtained from Einstein’s equations by adding a cosmological constant in order to get a static universe, according to Einstein’s idea. In modern cosmology, it is considered as a candidate for dark energy, the cause of the acceleration of the expansion of the universe.
Replacing $T$ from (1) we obtain:
(3)
$$S(X,Y)=-(\lambda-\frac{scal}{2}-kp)g(X,Y)+k(\sigma+p)\eta(X)\eta(Y),$$
for any $X$, $Y\in\chi(M)$. Recall that a manifold having the property that the Ricci tensor $S$ is a functional combination of $g$ and $\eta\otimes\eta$, for $\eta$ a $1$-form $g$ dual to a unitary vector field, is called quasi-Einstein [4]. Quasi-Einstein manifolds arose during the study of exact solutions of Einstein field equations. For example, the Robertson-Walker spacetime are quasi-Einstein manifolds [11]. They also can be taken as a
model of the perfect fluid spacetime in general relativity [9], [10].
Ricci flow and Einstein flow are intrinsec geometric flows on a pseudo-Riemannian manifold, whose fixed points are solitons. In our paper, we are interested in a generalized version of the following two types of solitons:
1.
Ricci solitons [12], which generates self-similar solutions to the Ricci flow:
(4)
$$\frac{\partial}{\partial t}g=-2S,$$
2.
Einstein solitons [3], which generate self-similar solutions to the Einstein flow:
(5)
$$\frac{\partial}{\partial t}g=-2(S-\frac{scal}{2}g).$$
Perturbing the equations that define these types of solitons by a multiple of a certain $(0,2)$-tensor field $\eta\otimes\eta$, we obtain two slightly more general notions, namely, $\eta$-Ricci solitons and $\eta$-Einstein solitons, which we shall consider in a perfect fluid spacetime, i.e. in a $4$-dimensional pseudo-Riemannian manifold $M$ with a Lorentzian metric $g$ whose content is a perfect fluid.
2 Basic properties of a perfect fluid spacetime
Let $(M,g)$ be a general relativistic perfect fluid spacetime satisfying (3).
Consider $\{E_{i}\}_{1\leq i\leq 4}$ an orthonormal frame field i.e. $g(E_{i},E_{j})=\varepsilon_{ij}\delta_{ij}$, $i$, $j\in\{1,2,3,4\}$ with $\varepsilon_{11}=-1$, $\varepsilon_{ii}=1$, $i\in\{2,3,4\}$, $\varepsilon_{ij}=0$, $i,j\in\{1,2,3,4\}$, $i\neq j$. Let $\xi=\sum_{i=1}^{4}\xi^{i}E_{i}$. Then
$$-1=g(\xi,\xi)=\sum_{1\leq i,j\leq 4}\xi^{i}\xi^{j}g(E_{i},E_{j})=\sum_{i=1}^{4%
}\varepsilon_{ii}(\xi^{i})^{2}$$
and
$$\eta(E_{i})=g(E_{i},\xi)=\sum_{j=1}^{4}\xi^{j}g(E_{i},E_{j})=\varepsilon_{ii}%
\xi^{i}.$$
Contracting (3) and taking into account that $g(\xi,\xi)=-1$, we get:
(6)
$$scal=4\lambda+k(\sigma-3p).$$
Therefore:
(7)
$$S(X,Y)=(\lambda+\frac{k(\sigma-p)}{2})g(X,Y)+k(\sigma+p)\eta(X)\eta(Y),$$
for any $X$, $Y\in\chi(M)$.
Example 2.1.
A radiation fluid has constant scalar curvature equal to $4\lambda$.
If we denote by $\nabla$ the Levi-Civita connection associated to $g$, then for any $X$, $Y$, $Z\in\chi(M)$:
$$(\nabla_{X}S)(Y,Z):=X(S(Y,Z))-S(\nabla_{X}Y,Z)-S(Y,\nabla_{X}Z)=$$
$$=k(\sigma+p)\{\eta(Y)g(\nabla_{X}\xi,Z)+\eta(Z)g(\nabla_{X}\xi,Y)\}=$$
(8)
$$=k(\sigma+p)\{\eta(Y)(\nabla_{X}\eta)Z+\eta(Z)(\nabla_{X}\eta)Y\}.$$
Imposing different conditions on the covariant differential of $S$, we have that:
i) $M$ is Ricci symmetric if $\nabla S=0$;
ii) $S$ is Codazzi tensor if $(\nabla_{X}S)(Y,Z)=(\nabla_{Y}S)(X,Z)$, for any $X$, $Y$, $Z\in\chi(M)$;
iii) $S$ is $\alpha$-recurrent if $(\nabla_{X}S)(Y,Z)=\alpha(X)S(Y,Z)$, for any $X$, $Y$, $Z\in\chi(M)$, for $\alpha$ a nonzero $1$-form;
iv) $S$ is weakly pseudo Ricci symmetric if $S$ is not identically zero and $(\nabla_{X}S)(Y,Z)=\alpha(X)S(Y,Z)+\alpha(Y)S(Z,X)+\alpha(Z)S(X,Y)$, for any $X$, $Y$, $Z\in\chi(M)$, for $\alpha$ a nonzero $1$-form;
v) $S$ is pseudo Ricci symmetric if $S$ is not identically zero and $(\nabla_{X}S)(Y,Z)=2\alpha(X)S(Y,Z)+\alpha(Y)S(Z,X)+\alpha(Z)S(X,Y)$, for any $X$, $Y$, $Z\in\chi(M)$, for $\alpha$ a nonzero $1$-form.
Proposition 2.2.
Let $(M,g)$ be a general relativistic perfect fluid spacetime satisfying (7).
1.
If $M$ is Ricci symmetric or $S$ is Codazzi tensor, then $p=-\sigma$ or $\nabla\xi=0$.
2.
If $S$ is $\alpha$-recurrent, then $p=-\sigma=\frac{\lambda}{k}$ or $\nabla\xi=0$.
3.
If $S$ is (weakly) pseudo Ricci symmetric, then $p=\frac{2}{3}(\frac{\lambda}{k})-\frac{\sigma}{3}$. In this cases, $\xi$ is torse-forming (in particular, irrotational and geodesic) vector field and $\eta$ is closed (and Codazzi) $1$-form.
Proof..
1.
If $\nabla S=0$, writing (2) for $Y=Z$ we get $(\sigma+p)\eta(Y)g(\nabla_{X}\xi,Y)=0$, for any $X$, $Y\in\chi(M)$. It follows $\nabla\xi=0$ or $p=-\sigma$.
If $(\nabla_{X}S)(Y,Z)=(\nabla_{Y}S)(X,Z)$, for any $X$, $Y$, $Z\in\chi(M)$, using (2) for $Y=Z:=\xi$ we get $(\sigma+p)g(\nabla_{\xi}\xi,X)=0$, for any $X\in\chi(M)$. It follows $\nabla_{\xi}\xi=0$ or $p=-\sigma$. If $\nabla_{\xi}\xi=0$, using (2) for $X:=\xi$ we get $(\sigma+p)g(\nabla_{Y}\xi,Z)=0$, for any $Y$, $Z\in\chi(M)$. It follows $\nabla\xi=0$ or $p=-\sigma$.
2.
If $(\nabla_{X}S)(Y,Z)=\alpha(X)S(Y,Z)$, for any $X$, $Y$, $Z\in\chi(M)$, using (7) and (2) and writing the obtained relation for $Y=Z:=\xi$ we get
$$(\lambda-\frac{k(\sigma+3p)}{2})\alpha(X)=0,$$
for any $X\in\chi(M)$. It follows $p=\frac{2\lambda-k\sigma}{3k}$. Writing the same relation for $Z:=\xi$ we get
$$-k(\sigma+p)g(\nabla_{X}\xi,Y)=(\lambda-\frac{k(\sigma+3p)}{2})\alpha(X)\eta(Y%
)=0,$$
for any $X$, $Y\in\chi(M)$. It follows $\nabla\xi=0$ or $p=-\sigma$ (so $p=-\sigma=\frac{\lambda}{k}$).
3.
If $(\nabla_{X}S)(Y,Z)=\alpha(X)S(Y,Z)+\alpha(Y)S(Z,X)+\alpha(Z)S(X,Y)$, for any $X$, $Y$, $Z\in\chi(M)$, using (7) and (2) and writing the obtained relation for $Y=Z:=\xi$ we get
$$(\lambda-\frac{k(\sigma+3p)}{2})[\alpha(X)-2\alpha(\xi)\eta(X)]=0,$$
for any $X\in\chi(M)$. Writing the same relation for $X=Y=Z:=\xi$ we get
$$(\lambda-\frac{k(\sigma+3p)}{2})\alpha(\xi)=0.$$
If $\lambda-\frac{k(\sigma+3p)}{2}\neq 0$ follows $\alpha(\xi)=0$ and $\alpha(X)=\alpha(\xi)\eta(X)=0$, for any $X\in\chi(M)$, which contradicts the fact that $\alpha$ is nonzero. Therefore, $\lambda-\frac{k(\sigma+3p)}{2}=0$, so $p=\frac{2\lambda-k\sigma}{3k}$.
Writing now the obtained relation for $Z:=\xi$ we get
$$k(\sigma+p)\{g(\nabla_{X}\xi,Y)+\alpha(\xi)[g(X,Y)+\eta(X)\eta(Y)]\}=0,$$
for any $X$, $Y\in\chi(M)$.
If $\sigma+p=0$ follows $p=-\sigma=\frac{\lambda}{k}$ and from (7), $S(X,Y)=0$, for any $X$, $Y\in\chi(M)$, which contradicts the fact that $S$ is nonzero. Therefore:
$$g(\nabla_{X}\xi,Y)+\alpha(\xi)[g(X,Y)+\eta(X)\eta(Y)]=0,$$
for any $X$, $Y\in\chi(M)$ which implies
$$\nabla_{X}\xi=-\alpha(\xi)[X+\eta(X)\xi],$$
for any $X\in\chi(M)$ i.e. $\xi$ is a torse-forming vector field. Then
$$(curl(\xi))(X,Y):=g(\nabla_{X}\xi,Y)-g(\nabla_{Y}\xi,X)=0,$$
for any $X$, $Y\in\chi(M)$ i.e. $\xi$ is irrotational. Also
$$g(\nabla_{\xi}\xi,X)=g(\nabla_{X}\xi,\xi)=\frac{1}{2}X(g(\xi,\xi))=0,$$
for any $X\in\chi(M)$ i.e. $\xi$ is geodesic.
Concerning $\eta$, notice that
$$(\nabla_{X}\eta)Y-(\nabla_{Y}\eta)X:=X(\eta(Y))-\eta(\nabla_{X}Y)-Y(\eta(X))+%
\eta(\nabla_{Y}X):=(d\eta)(X,Y).$$
Also
$$X(\eta(Y))-\eta(\nabla_{X}Y)-Y(\eta(X))+\eta(\nabla_{Y}X)=g(\nabla_{X}\xi,Y)-g%
(\nabla_{Y}\xi,X)=0$$
which implies $d\eta=0$.
If $(\nabla_{X}S)(Y,Z)=2\alpha(X)S(Y,Z)+\alpha(Y)S(Z,X)+\alpha(Z)S(X,Y)$, for any $X$, $Y$, $Z\in\chi(M)$, following the steps of computations above, we get the same conclusions.
∎
Remark the following facts:
1) If in the general relativistic perfect fluid spacetime $\xi$ is not $\nabla$-parallel and it has $\alpha$-recurrent Ricci tensor field, we have the vacuum case. The energy-momentum tensor is $T_{vac}=\frac{\lambda}{k}g$.
2) $S(X,\xi)=-S(\xi,\xi)\eta(X)=[\lambda-\frac{k(\sigma+3p)}{2}]g(X,\xi)$, for any $X\in\chi(M)$ which shows that $\lambda-\frac{k(\sigma+3p)}{2}$ is the eigenvalue of $Q$ corresponding to the eigenvector $\xi$, where $Q$ is defined by $g(QX,Y):=S(X,Y)$, $X$, $Y\in\chi(M)$. From Proposition 2.2 we deduce that if $\xi$ is not $\nabla$-parallel and $M$ is Ricci symmetric or $S$ is Codazzi tensor, then $(M,g)$ is Einstein and $S(X,\xi)=(\lambda+k\sigma)\eta(X)$, for any $X\in\chi(M)$, and if $S$ is (weakly) pseudo Ricci symmetric, then $S(X,\xi)=0$, for any $X\in\chi(M)$, hence $\xi\in\ker Q$; also, $div(\xi):=\sum_{i=1}^{4}\varepsilon_{ii}g(\nabla_{E_{i}}\xi,E_{i})=-3\alpha(\xi)$. Notice that in all these cases, if $\sigma>-\frac{\lambda}{k}$, the scalar curvature $scal:=\sum_{i=1}^{4}\varepsilon_{ii}S(E_{i},E_{i})=4\lambda+k(\sigma-3p)>3(%
\lambda-kp)$ is positive.
3) Considering Plebanski energy conditions $\sigma\geq 0$ and $-\sigma\leq p\leq\sigma$ for perfect fluids, when $S$ is (weakly) pseudo Ricci symmetric, the energy-density is lower bounded by $\max\{-\frac{\lambda}{k},\frac{\lambda}{2k}\}$. It was observed that a positive cosmological constant $\lambda$ acts as repulsive gravity, explaining the accelerating universe. The observations of Edwin Hubble confirmed that the universe is expanding, therefore, this seems to be the real case, so $\sigma\geq\frac{\lambda}{2k}$.
4) Let $hX:=X+\eta(X)\xi$ be the projection tensor, $X\in\chi(M)$. If $S$ is (weakly) pseudo Ricci symmetric and $\alpha(\xi)\neq 0$, then $(h,\xi,\eta,g)$ is a Lorentzian concircular structure [17]. In the particular case $\alpha(\xi)=-1$ it becomes LP-Sasakian structure [13]. In the other case, if $\alpha(\xi)=0$, then $\xi$ and $\eta$ are $\nabla$-parallel, $\xi$ is divergence free, hence harmonic vector field.
Applying the covariant derivative to (2) we obtain:
(9)
$$k(\nabla_{X}T)(Y,Z)=(\nabla_{X}S)(Y,Z)-\frac{1}{2}X(scal)g(Y,Z),$$
for any $X$, $Y$, $Z\in\chi(M)$.
Proposition 2.3.
Let $(M,g)$ be a general relativistic perfect fluid spacetime satisfying (7).
1.
If $T$ is covariantly constant, then the scalar curvature is constant.
2.
If $T$ is Codazzi tensor, then the scalar curvature satisfies $grad(scal)+(dscal)(\xi)\xi+2k(\sigma+p)\nabla_{\xi}\xi=0$.
3.
If $T$ is $\alpha$-recurrent, then the scalar curvature satisfies $dscal+[4\lambda-scal-k(\sigma+3p)]\alpha=0$.
Proof..
1.
If $T$ is covariantly constant i.e. $\nabla T=0$ follows $(\nabla_{X}S)(Y,Z)=\frac{1}{2}X(scal)g(Y,Z)$, for any $X$, $Y$, $Z\in\chi(M)$. For $Y=Z:=\xi$ we have:
(10)
$$2S(\nabla_{X}\xi,\xi)=\frac{1}{2}X(scal),$$
for any $X\in\chi(M)$.
Replacing $S$ from (7) we obtain $X(scal)=0$, for any $X\in\chi(M)$, which implies that the scalar curvature is constant. Therefore, $\nabla S=0$ (i.e. $M$ is Ricci symmetric) and from Proposition 2.2 follows $p=-\sigma$ (the vacuum case) or $\nabla\xi=0$. If $\xi$ is not $\nabla$-parallel:
(11)
$$S=-(\lambda-\frac{scal}{2}-kp)g$$
and from (6) we get:
(12)
$$p=\frac{4\lambda-scal}{4k},$$
so the pressure and the energy-density are constant.
2.
The condition $(\nabla_{X}T)(Y,Z)=(\nabla_{Y}T)(X,Z)$, for any $X$, $Y$, $Z\in\chi(M)$ is equivalent to:
$$(\nabla_{X}S)(Y,Z)-(\nabla_{Y}S)(X,Z)=\frac{1}{2}[X(scal)g(Y,Z)-Y(scal)g(Z,X)],$$
for any $X$, $Y$, $Z\in\chi(M)$. Replacing $\nabla S$ from (2) we get:
(13)
$$2k(\sigma+p)[\eta(Y)g(\nabla_{X}\xi,Z)+\eta(Z)g(\nabla_{X}\xi,Y)-\eta(X)g(%
\nabla_{Y}\xi,Z)-\eta(Z)g(\nabla_{Y}\xi,X)]=$$
$$=X(scal)g(Y,Z)-Y(scal)g(Z,X),$$
for any $X$, $Y$, $Z\in\chi(M)$. For $Y=Z:=\xi$ and taking into account that $g(\nabla_{X}\xi,\xi)=0$, for any $X\in\chi(M)$, the above relation becomes:
$$2k(\sigma+p)g(\nabla_{\xi}\xi,X)+X(scal)+\xi(scal)\eta(X)=0,$$
for any $X\in\chi(M)$. Also $X(scal)=(dscal)(X)=g(grad(scal),X)$ and we obtain:
$$2k(\sigma+p)\nabla_{\xi}\xi+grad(scal)+(dscal)(\xi)\xi=0.$$
If the scalar curvature is constant, then $\nabla_{\xi}\xi=0$ or $p=-\sigma$. If $\nabla_{\xi}\xi=0$, taking $X:=\xi$ in (13) we get
$(\sigma+p)g(\nabla_{Y}\xi,Z)=0$ which implies $p=-\sigma$ or $\nabla\xi=0$.
If $\xi$ is geodesic vector field, then $grad(\sigma)+(d\sigma)(\xi)\xi=3[grad(p)+(dp)(\xi)\xi]$ and the gradient of the scalar curvature is collinear with $\xi$.
3.
The condition $(\nabla_{X}T)(Y,Z)=\alpha(X)T(Y,Z)$, for any $X$, $Y$, $Z\in\chi(M)$, with $\alpha$ a nonzero $1$-form, is equivalent to:
$$(\nabla_{X}S)(Y,Z)-\frac{X(scal)}{2}g(Y,Z)=\alpha(X)S(Y,Z)+\alpha(X)(\lambda-%
\frac{scal}{2})g(Y,Z),$$
for any $X$, $Y$, $Z\in\chi(M)$. Replacing $S$ from (7) and $\nabla S$ from (2) we get:
(14)
$$k(\sigma+p)[\eta(Y)g(\nabla_{X}\xi,Z)+\eta(Z)g(\nabla_{X}\xi,Y)-\alpha(X)\eta(%
Y)\eta(Z)]=$$
$$=\frac{1}{2}\{X(scal)+\alpha(X)[4\lambda-scal+k(\sigma-p)]\}g(Y,Z),$$
for any $X$, $Y$, $Z\in\chi(M)$. For $Y=Z:=\xi$ and taking into account that $g(\nabla_{X}\xi,\xi)=0$, for any $X\in\chi(M)$, the above relation becomes:
$$X(scal)=\alpha(X)[-4\lambda+scal+k(\sigma+3p)],$$
for any $X\in\chi(M)$.
If the scalar curvature is constant, then $\sigma=0$ and the pressure is constant $p=\frac{4\lambda-scal}{3k}$.
∎
Remark the following facts:
1) If the energy-momentum tensor $T$ of a general relativistic perfect fluid spacetime is covariantly constant, then we have the vacuum case or $\xi$ is $\nabla$-parallel (in particular, geodesic).
2) If $T$ is Codazzi and the scalar curvature is not constant but $\xi$ is a geodesic vector field, then the gradient of the scalar curvature is collinear with $\xi$. It was proved that for a Codazzi energy-momentum tensor, the Ricci tensor $S$ is conserved [2].
3) For an $\alpha$-recurrent energy-momentum tensor, if the scalar curvature is constant, then the pressure is constant ($p=\frac{4\lambda-scal}{3k}$), but the energy-density vanishes.
3 Perfect fluid spacetime with torse-forming vector field $\xi$
We shall treat the special case when $\xi$ is a torse-forming vector field [20] of the form:
(15)
$$\nabla\xi=I_{\chi(M)}+\eta\otimes\xi.$$
Then $\nabla_{\xi}\xi=\xi+\eta(\xi)\xi=0$, for any $X\in\chi(M)$ i.e. $\xi$ is geodesic, $g(\nabla_{X}\xi,\xi)=0$ and $(d\eta)(X,Y):=X(\eta(Y))-Y(\eta(X))-\eta([X,Y])=X(g(Y,\xi))-Y(g(X,\xi))-g(%
\nabla_{X}Y,\xi)+g(\nabla_{Y}X,\xi)=g(\nabla_{X}\xi,Y)-g(\nabla_{Y}\xi,X)=0$. We also have:
(16)
$$R(X,Y)\xi=\eta(Y)X-\eta(X)Y,$$
(17)
$$\eta(R(X,Y)Z)=-\eta(Y)g(X,Z)+\eta(X)g(Y,Z),$$
for any $X$, $Y$, $Z\in\chi(M)$.
In this case, we shall see which are the consequences of certain conditions imposed to different types of curvatures of this space, namely, when the curvature satisfies conditions of the type $\mathcal{T}(\xi,\cdot)\cdot S=0$ and $S(\xi,\cdot)\cdot\mathcal{T}=0$, where
$\mathcal{T}$ is the Riemann curvature tensor $R$, the projective curvature tensor $\mathcal{W}$ [14], the concircular curvature tensor $\mathcal{P}$ [19], the conformal curvature tensor $\mathcal{C}$ [1] and the conharmonic curvature tensor $\mathcal{H}$ [16] defined as follows:
(18)
$$\mathcal{W}(X,Y)Z:=R(X,Y)Z+\frac{1}{\dim(M)-1}[g(Z,X)QY-g(Y,Z)QX]$$
(19)
$$\mathcal{P}(X,Y)Z:=R(X,Y)Z+\frac{scal}{\dim(M)(\dim(M)-1)}[g(Z,X)Y-g(Y,Z)X]$$
$$\mathcal{C}(X,Y)Z:=R(X,Y)Z+\frac{1}{\dim(M)-2}\{-\frac{scal}{\dim(M)-1}[g(Z,X)%
Y-g(Y,Z)X]+$$
(20)
$$+g(Z,X)QY-g(Y,Z)QX+S(Z,X)Y-S(Y,Z)X\}$$
(21)
$$\mathcal{H}(X,Y)Z:=R(X,Y)Z+\frac{1}{\dim(M)-2}[g(Z,X)QY-g(Y,Z)QX+S(Z,X)Y-S(Y,Z%
)X]$$
for any $X$, $Y$, $Z\in\chi(M)$.
Remark that for an empty gravitational field characterized by vanishing Ricci tensor, the curvature tensors $R$, $\mathcal{W}$ and $\mathcal{H}$ coincide.
Let us denote by $\mathcal{T}$ a curvature tensor of type $(1,3)$ and ask for certain Ricci-semisymmetry curvature conditions, namely, $\mathcal{T}(\xi,\cdot)\cdot S=0$ and $S(\xi,\cdot)\cdot\mathcal{T}=0$, where by $\cdot$ we denote the derivation of the tensor algebra at each point of the tangent space:
•
$(\mathcal{T}(\xi,X)\cdot S)(Y,Z):=((\xi\wedge_{\mathcal{T}}X)\cdot S)(Y,Z):=S(%
(\xi\wedge_{\mathcal{T}}X)Y,Z)+S(Y,(\xi\wedge_{\mathcal{T}}X)Z)$, for $(X\wedge_{\mathcal{T}}Y)Z:=\mathcal{T}(X,Y)Z$;
•
$(S(\xi,X)\cdot\mathcal{T})(Y,Z)W:=((\xi\wedge_{S}X)\cdot\mathcal{T})(Y,Z)W:=%
\mathcal{T}((\xi\wedge_{S}X)Y,Z)W+\linebreak\mathcal{T}(Y,(\xi\wedge_{S}X)Z)W+%
\mathcal{T}(Y,Z)(\xi\wedge_{S}X)W$, for $(X\wedge_{S}Y)Z:=S(Y,Z)X-S(X,Z)Y$.
3.1 Perfect fluid spacetime satisfying $\mathcal{T}(\xi,\cdot)\cdot S=0$
The condition $\mathcal{T}(\xi,\cdot)\cdot S=0$ is equivalent to
(22)
$$S(\mathcal{T}(\xi,X)Y,Z)+S(Y,\mathcal{T}(\xi,X)Z)=0,$$
for any $X$, $Y$, $Z\in\chi(M)$ and from (7) with:
$$(\lambda+\frac{k(\sigma-p)}{2})[g(\mathcal{T}(\xi,X)Y,Z)+g(Y,\mathcal{T}(\xi,X%
)Z)]+$$
$$+k(\sigma+p)[\eta(\mathcal{T}(\xi,X)Y)\eta(Z)+\eta(Y)\eta(\mathcal{T}(\xi,X)Z)%
]=0,$$
for any $X$, $Y$, $Z\in\chi(M)$.
Theorem 3.1.
Let $(M,g)$ be a general relativistic perfect fluid spacetime satisfying (3) with torse-forming vector field $\xi$.
1.
If $R(\xi,\cdot)\cdot S=0$, then $p=-\sigma$.
2.
If $\mathcal{W}(\xi,\cdot)\cdot S=0$, then $p=-\sigma$ or $p=\frac{2\lambda-3}{2k}$.
3.
If $\mathcal{P}(\xi,\cdot)\cdot S=0$, then $p=-\sigma$ or $p=\frac{4\lambda+k\sigma-12}{3k}$.
4.
If $\mathcal{C}(\xi,\cdot)\cdot S=0$, then $p=-\sigma$ or $p=\frac{2\lambda-k\sigma-6}{3k}$.
5.
If $\mathcal{H}(\xi,\cdot)\cdot S=0$, then $p=-\sigma$ or $p=\frac{\lambda-1}{k}$.
Proof..
1.
From the symmetries of $R$ and (16) and (17) we obtain:
(23)
$$k(\sigma+p)[\eta(Z)g(X,Y)+\eta(Y)g(Z,X)+2\eta(X)\eta(Y)\eta(Z)]=0,$$
for any $X$, $Y$, $Z\in\chi(M)$. Take $Z:=\xi$ and (23) becomes:
(24)
$$k(\sigma+p)[g(X,Y)+\eta(X)\eta(Y)]=0,$$
for any $X$, $Y\in\chi(M)$ and we obtain $p=-\sigma$.
2.
From (18), from the symmetries of $R$ and (16) and (17) we obtain:
(25)
$$k(\sigma+p)(2\lambda-2kp-3)[\eta(Z)g(X,Y)+\eta(Y)g(Z,X)+2\eta(X)\eta(Y)\eta(Z)%
]=0,$$
for any $X$, $Y$, $Z\in\chi(M)$. Take $Z:=\xi$ and (25) becomes:
(26)
$$k(\sigma+p)(2\lambda-2kp-3)[g(X,Y)+\eta(X)\eta(Y)]=0,$$
for any $X$, $Y\in\chi(M)$ and we obtain $p=-\sigma$ or the pressure is constant $p=\frac{2\lambda-3}{2k}$.
3.
From (19), from the symmetries of $R$ and (16) and (17) we obtain:
(27)
$$k(\sigma+p)[4\lambda+k(\sigma-3p)-12][\eta(Z)g(X,Y)+\eta(Y)g(Z,X)+2\eta(X)\eta%
(Y)\eta(Z)]=0,$$
for any $X$, $Y$, $Z\in\chi(M)$. Take $Z:=\xi$ and (27) becomes:
(28)
$$k(\sigma+p)[4\lambda+k(\sigma-3p)-12][g(X,Y)+\eta(X)\eta(Y)]=0,$$
for any $X$, $Y\in\chi(M)$ and we obtain $p=-\sigma$ or $p=\frac{4\lambda+k\sigma-12}{3k}$.
4.
From (3), from the symmetries of $R$ and (16) and (17) we obtain:
(29)
$$k(\sigma+p)[2\lambda-k(\sigma+3p)-6][\eta(Z)g(X,Y)+\eta(Y)g(Z,X)+2\eta(X)\eta(%
Y)\eta(Z)]=0,$$
for any $X$, $Y$, $Z\in\chi(M)$. Take $Z:=\xi$ and (32) becomes:
(30)
$$k(\sigma+p)[2\lambda-k(\sigma+3p)-6][g(X,Y)+\eta(X)\eta(Y)]=0,$$
for any $X$, $Y\in\chi(M)$ and we obtain $p=-\sigma$ or $p=\frac{2\lambda-k\sigma-6}{3k}$.
5.
From (21), from the symmetries of $R$ and (16) and (17) we obtain:
(31)
$$k(\sigma+p)(\lambda-kp-1)[\eta(Z)g(X,Y)+\eta(Y)g(Z,X)+2\eta(X)\eta(Y)\eta(Z)]=0,$$
for any $X$, $Y$, $Z\in\chi(M)$. Take $Z:=\xi$ and (32) becomes:
(32)
$$k(\sigma+p)(\lambda-kp-1)[g(X,Y)+\eta(X)\eta(Y)]=0,$$
for any $X$, $Y\in\chi(M)$ and we obtain $p=-\sigma$ or $p=\frac{\lambda-1}{k}$.
∎
3.2 Perfect fluid spacetime satisfying $S(\xi,\cdot)\cdot\mathcal{T}=0$
Denote by
$$A=\lambda+\frac{k(\sigma-p)}{2},\ \ B=k(\sigma+p),$$
$$a=\frac{1}{3},\ \ b=\frac{scal}{12}(=\frac{4\lambda+k(\sigma-3p)}{12}),\ \ c=-%
\frac{scal}{6}(=-\frac{4\lambda+k(\sigma-3p)}{6}),\ \ d=\frac{1}{2}$$
and the curvature tensors can be written:
(33)
$$S(X,Y)=Ag(X,Y)+B\eta(X)\eta(Y)$$
(34)
$$\mathcal{W}(X,Y)Z=R(X,Y)Z+a[g(Z,X)QY-g(Y,Z)QX]$$
(35)
$$\mathcal{P}(X,Y)Z=R(X,Y)Z+b[g(Z,X)Y-g(Y,Z)X]$$
$$\mathcal{C}(X,Y)Z=R(X,Y)Z+c[g(Z,X)Y-g(Y,Z)X]+$$
(36)
$$+d[g(Z,X)QY-g(Y,Z)QX]+d[S(Z,X)Y-S(Y,Z)X]$$
(37)
$$\mathcal{H}(X,Y)Z=R(X,Y)Z+d[g(Z,X)QY-g(Y,Z)QX]+d[S(Z,X)Y-S(Y,Z)X]$$
for any $X$, $Y$, $Z\in\chi(M)$.
The condition $S(\xi,\cdot)\cdot\mathcal{T}=0$ is equivalent to
$$S(X,\mathcal{T}(Y,Z)W)\xi-S(\xi,\mathcal{T}(Y,Z)W)X+S(X,Y)\mathcal{T}(\xi,Z)W-$$
$$-S(\xi,Y)\mathcal{T}(X,Z)W+S(X,Z)\mathcal{T}(Y,\xi)W-S(\xi,Z)\mathcal{T}(Y,X)W+$$
(38)
$$+S(X,W)\mathcal{T}(Y,Z)\xi-S(\xi,W)\mathcal{T}(Y,Z)X=0,$$
for any $X$, $Y$, $Z$, $W\in\chi(M)$.
Taking the inner product with $\xi$, the relation (3.2) becomes:
$$-S(X,\mathcal{T}(Y,Z)W)-S(\xi,\mathcal{T}(Y,Z)W)\eta(X)+$$
$$+S(X,Y)\eta(\mathcal{T}(\xi,Z)W)-S(\xi,Y)\eta(\mathcal{T}(X,Z)W)+S(X,Z)\eta(%
\mathcal{T}(Y,\xi)W)-$$
(39)
$$-S(\xi,Z)\eta(\mathcal{T}(Y,X)W)+S(X,W)\eta(\mathcal{T}(Y,Z)\xi)-S(\xi,W)\eta(%
\mathcal{T}(Y,Z)X)=0,$$
for any $X$, $Y$, $Z$, $W\in\chi(M)$.
Theorem 3.2.
Let $(M,g)$ be a general relativistic perfect fluid spacetime satisfying (3) with torse-forming vector field $\xi$.
1.
If $S(\xi,\cdot)\cdot R=0$, then $p=\frac{\lambda}{k}$.
2.
If $S(\xi,\cdot)\cdot\mathcal{W}=0$, then $p_{1,2}=\frac{\lambda-6+3(\lambda+k\sigma)\pm 2\sqrt{3[(\lambda+k\sigma)^{2}-3%
(\lambda+k\sigma)+3]}}{k}$.
3.
If $S(\xi,\cdot)\cdot\mathcal{P}=0$, then $p=\frac{\lambda}{k}$ or $p=\frac{4\lambda+k\sigma-12}{3k}$.
4.
If $S(\xi,\cdot)\cdot\mathcal{C}=0$, then $p=\frac{\lambda}{k}$ or $p=\frac{2\lambda-k\sigma-6}{3k}$.
5.
If $S(\xi,\cdot)\cdot\mathcal{H}=0$, then $p=\frac{\lambda}{k}$ or $p=\frac{\lambda-1}{k}$.
Proof..
1.
From (3.2) and (33) we obtain:
$$-Ag(X,R(Y,Z)W)+A[g(X,Z)g(Y,W)-g(X,Y)g(Z,W)]+$$
$$+2A[\eta(X)\eta(Z)g(Y,W)-\eta(X)\eta(Y)g(Z,W)]+$$
(40)
$$+B[\eta(Y)\eta(W)g(X,Z)-\eta(Z)\eta(W)g(X,Y)]=0,$$
for any $X$, $Y$, $Z$, $W\in\chi(M)$. Taking $Z=W:=\xi$ we get:
(41)
$$(2A-B)[\eta(X)\eta(Y)+g(X,Y)]=0,$$
for any $X$, $Y\in\chi(M)$. It follows $p=\frac{\lambda}{k}$.
2.
From (3.2), (33) and (34) we obtain:
$$-Ag(X,R(Y,Z)W)+A(1+aB-2aA)[g(X,Z)g(Y,W)-g(X,Y)g(Z,W)]+$$
$$+A(2+aB-2aA)[\eta(X)\eta(Z)g(Y,W)-\eta(X)\eta(Y)g(Z,W)]+$$
(42)
$$+B(1+aB-aA)[\eta(Y)\eta(W)g(X,Z)-\eta(Z)\eta(W)g(X,Y)]=0,$$
for any $X$, $Y$, $Z$, $W\in\chi(M)$. Taking $Z=W:=\xi$ we get:
(43)
$$[2A-B+a(2AB-2A^{2}-B^{2})][\eta(X)\eta(Y)+g(X,Y)]=0,$$
for any $X$, $Y\in\chi(M)$. It follows $p_{1,2}=\frac{\lambda-6+3(\lambda+k\sigma)\pm 2\sqrt{3[(\lambda+k\sigma)^{2}-3%
(\lambda+k\sigma)+3]}}{k}$.
3.
From (3.2), (33) and (35) we obtain:
$$-Ag(X,R(Y,Z)W)+A(1-2b)[g(X,Z)g(Y,W)-g(X,Y)g(Z,W)]+$$
$$+2A(1-b)[\eta(X)\eta(Z)g(Y,W)-\eta(X)\eta(Y)g(Z,W)]+$$
(44)
$$+B(1-b)[\eta(Y)\eta(W)g(X,Z)-\eta(Z)\eta(W)g(X,Y)]=0,$$
for any $X$, $Y$, $Z$, $W\in\chi(M)$. Taking $Z=W:=\xi$ we get:
(45)
$$(2A-B)(1-b)[\eta(X)\eta(Y)+g(X,Y)]=0,$$
for any $X$, $Y\in\chi(M)$. It follows $p=\frac{\lambda}{k}$ or $p=\frac{4\lambda+k\sigma-12}{3k}$.
4.
From (3.2), (33) and (3.2) we obtain:
$$-Ag(X,R(Y,Z)W)+A(1-2c+dB-4dA)[g(X,Z)g(Y,W)-g(X,Y)g(Z,W)]+$$
$$+A(2-2c+dB-4dA)[\eta(X)\eta(Z)g(Y,W)-\eta(X)\eta(Y)g(Z,W)]+$$
(46)
$$+B(1-c+dB-3dA)[\eta(Y)\eta(W)g(X,Z)-\eta(Z)\eta(W)g(X,Y)]=0,$$
for any $X$, $Y$, $Z$, $W\in\chi(M)$. Taking $Z=W:=\xi$ we get:
(47)
$$(2A-B)[1-c-d(2A-B)][\eta(X)\eta(Y)+g(X,Y)]=0,$$
for any $X$, $Y\in\chi(M)$. It follows $p=\frac{\lambda}{k}$ or $p=\frac{2\lambda-k\sigma-6}{3k}$.
5.
From (3.2), (33) and (37) we obtain:
$$-Ag(X,R(Y,Z)W)+A(1+dB-4dA)[g(X,Z)g(Y,W)-g(X,Y)g(Z,W)]+$$
$$+A(2+dB-4dA)[\eta(X)\eta(Z)g(Y,W)-\eta(X)\eta(Y)g(Z,W)]+$$
(48)
$$+B(1+dB-3dA)[\eta(Y)\eta(W)g(X,Z)-\eta(Z)\eta(W)g(X,Y)]=0,$$
for any $X$, $Y$, $Z$, $W\in\chi(M)$. Taking $Z=W:=\xi$ we get:
(49)
$$(2A-B)[1-d(2A-B)][\eta(X)\eta(Y)+g(X,Y)]=0,$$
for any $X$, $Y\in\chi(M)$. It follows $p=\frac{\lambda}{k}$ or $p=\frac{\lambda-1}{k}$.
∎
Remark the following facts:
1) If for a general relativistic perfect fluid spacetime $(M,g)$ satisfying (3) with torse-forming vector field $\xi$ the projective curvature tensor $\mathcal{W}$ or the conharmonic curvature tensor $\mathcal{H}$ satisfy $\mathcal{W}(\xi,\cdot)\cdot S=0$, respectively $\mathcal{H}(\xi,\cdot)\cdot S=0$, then we have the vacuum case or the pressure is constant, but $R(\xi,\cdot)\cdot S=0$ leads only to the vacuum case.
2) Under the same assumptions, the condition $S(\xi,\cdot)\cdot R=0$ or $S(\xi,\cdot)\cdot\mathcal{H}=0$ implies a constant pressure of the fluid.
4 Solitons in a perfect fluid spacetime
4.1 $\eta$-Ricci solitons
Consider the equation:
(50)
$$\mathcal{L}_{\xi}g+2S+2ag+2b\eta\otimes\eta=0,$$
where $g$ is a pseudo-Riemannian metric, $S$ is the Ricci curvature, $\xi$ is a vector field, $\eta$ is a $1$-form and $a$ and $b$ are real constants. The data $(g,\xi,a,b)$ which satisfy
the equation (50) is said to be an $\eta$-Ricci soliton in $M$ [6]; in particular, if $b=0$, $(g,\xi,a)$ is a Ricci soliton [12] and it is called shrinking, steady or expanding according as $a$ is negative, zero or positive, respectively [7].
Writing explicitly the Lie derivative $\mathcal{L}_{\xi}g$ we get $(\mathcal{L}_{\xi}g)(X,Y)=g(\nabla_{X}\xi,Y)+g(X,\nabla_{Y}\xi)$ and from (50) we obtain:
(51)
$$S(X,Y)=-ag(X,Y)-b\eta(X)\eta(Y)-\frac{1}{2}[g(\nabla_{X}\xi,Y)+g(X,\nabla_{Y}%
\xi)],$$
for any $X$, $Y\in\chi(M)$.
Contracting (51) we get:
(52)
$$scal=-a\dim(M)+b-div(\xi).$$
Let $(M,g)$ be a general relativistic perfect fluid spacetime and $(g,\xi,a,b)$ be an $\eta$-Ricci soliton in $M$. From (7) and (51) we obtain:
(53)
$$[\lambda+\frac{k(\sigma-p)}{2}+a]g(X,Y)+[k(\sigma+p)+b]\eta(X)\eta(Y)+\frac{1}%
{2}[g(\nabla_{X}\xi,Y)+g(X,\nabla_{Y}\xi)]=0,$$
for any $X$, $Y\in\chi(M)$.
Consider $\{E_{i}\}_{1\leq i\leq 4}$ an orthonormal frame field and let $\xi=\sum_{i=1}^{4}\xi^{i}E_{i}$. We have shown in the previous section that $\sum_{i=1}^{4}\varepsilon_{ii}(\xi^{i})^{2}=-1$ and $\eta(E_{i})=\varepsilon_{ii}\xi^{i}$.
Multiplying (53) by $\varepsilon_{ii}$ and summing over $i$ for $X=Y:=E_{i}$, we get:
(54)
$$4a-b=-4\lambda-k(\sigma-3p)-div(\xi).$$
Writing (53) for $X=Y:=\xi$, we obtain:
(55)
$$a-b=-\lambda+\frac{k(\sigma+3p)}{2}.$$
Therefore:
(56)
$$\left\{\begin{array}[]{ll}a=-\lambda-\frac{k(\sigma-p)}{2}-\frac{div(\xi)}{3}%
\\
b=-k(\sigma+p)-\frac{div(\xi)}{3}\end{array}\right..$$
Theorem 4.1.
Let $(M,g)$ be a $4$-dimensional pseudo-Riemannian manifold and
$\eta$ be the $g$-dual $1$-form of the gradient vector field $\xi:=grad(f)$ with $g(\xi,\xi)=-1$. If (50) defines an $\eta$-Ricci soliton in $M$, then the Laplacian equation satisfied by $f$ becomes:
(57)
$$\Delta(f)=-3[b+k(\sigma+p)].$$
Remark 4.2.
If $b=0$ in (50), we obtain the Ricci soliton with $a=-\lambda+\frac{k(\sigma+3p)}{2}$ which is steady if $p=\frac{2}{3}(\frac{\lambda}{k})-\frac{\sigma}{3}$, expanding if $p>\frac{2}{3}(\frac{\lambda}{k})-\frac{\sigma}{3}$ and shrinking if $p<\frac{2}{3}(\frac{\lambda}{k})-\frac{\sigma}{3}$. In these cases, $div(\xi)=-3k(\sigma+p)$. From Plebanski energy conditions for perfect fluids we deduce that $\sigma\geq\max\{-\frac{\lambda}{k},\frac{\lambda}{2k}\}$ for the steady case, $\sigma>\frac{\lambda}{2k}$ and $\sigma>-\frac{\lambda}{k}$ for the expanding and shrinking case, respectively.
Example 4.3.
An $\eta$-Ricci soliton $(g,\xi,a,b)$ in a radiation fluid is given by
$$\left\{\begin{array}[]{ll}a=-\lambda-kp-\frac{div(\xi)}{3}\\
b=-4kp-\frac{div(\xi)}{3}\end{array}\right..$$
4.2 $\eta$-Einstein solitons
Consider the equation:
(58)
$$\mathcal{L}_{\xi}g+2S+(2a-scal)g+2b\eta\otimes\eta=0,$$
where $g$ is a pseudo-Riemannian metric, $S$ is the Ricci curvature, $scal$ is the scalar curvature, $\xi$ is a vector field, $\eta$ is a $1$-form and $a$ and $b$ are real constants. The data $(g,\xi,a,b)$ which satisfy
the equation (58) is said to be an $\eta$-Einstein soliton in $M$; in particular, if $b=0$, $(g,\xi,a)$ is an Einstein soliton [3].
Writing explicitly the Lie derivative $\mathcal{L}_{\xi}g$ we get $(\mathcal{L}_{\xi}g)(X,Y)=g(\nabla_{X}\xi,Y)+g(X,\nabla_{Y}\xi)$ and from (58) we obtain:
(59)
$$S(X,Y)=-(a-\frac{scal}{2})g(X,Y)-b\eta(X)\eta(Y)-\frac{1}{2}[g(\nabla_{X}\xi,Y%
)+g(X,\nabla_{Y}\xi)],$$
for any $X$, $Y\in\chi(M)$.
Contracting (59) we get:
(60)
$$\frac{2-\dim(M)}{2}scal=-a\dim(M)+b-div(\xi).$$
Let $(M,g)$ be a general relativistic perfect fluid spacetime and $(g,\xi,a,b)$ be an $\eta$-Einstein soliton in $M$. From (6), (7) and (59) we obtain:
(61)
$$(\lambda-kp-a)g(X,Y)-[k(\sigma+p)+b]\eta(X)\eta(Y)-\frac{1}{2}[g(\nabla_{X}\xi%
,Y)+g(X,\nabla_{Y}\xi)]=0,$$
for any $X$, $Y\in\chi(M)$.
Consider $\{E_{i}\}_{1\leq i\leq 4}$ an orthonormal frame field and let $\xi=\sum_{i=1}^{4}\xi^{i}E_{i}$. We have shown in the previous section that $\sum_{i=1}^{4}\varepsilon_{ii}(\xi^{i})^{2}=-1$ and $\eta(E_{i})=\varepsilon_{ii}\xi^{i}$.
Multiplying (61) by $\varepsilon_{ii}$ and summing over $i$ for $X=Y:=E_{i}$, we get:
(62)
$$4a-b=4\lambda+k(\sigma-3p)-div(\xi).$$
Writing (61) for $X=Y:=\xi$, we obtain:
(63)
$$a-b=\lambda+k\sigma.$$
Therefore:
(64)
$$\left\{\begin{array}[]{ll}a=\lambda-kp-\frac{div(\xi)}{3}\\
b=-k(\sigma+p)-\frac{div(\xi)}{3}\end{array}\right..$$
Theorem 4.4.
Let $(M,g)$ be a $4$-dimensional pseudo-Riemannian manifold and
$\eta$ be the $g$-dual $1$-form of the gradient vector field $\xi:=grad(f)$ with $g(\xi,\xi)=-1$. If (58) defines an $\eta$-Einstein soliton in $M$, then the Laplacian equation satisfied by $f$ becomes:
(65)
$$\Delta(f)=-3[b+k(\sigma+p)].$$
Remark 4.5.
If $b=0$ in (58), we obtain the Einstein soliton with $a=\lambda+k\sigma$ which is steady if $p=\frac{2}{3}(\frac{\lambda}{k})-\frac{\sigma}{3}$, expanding if $p>\frac{2}{3}(\frac{\lambda}{k})-\frac{\sigma}{3}$ and shrinking if $p<\frac{2}{3}(\frac{\lambda}{k})-\frac{\sigma}{3}$. In these cases, $div(\xi)=-3k(\sigma+p)$. From Plebanski energy conditions for perfect fluids we deduce that $\sigma\geq\max\{-\frac{\lambda}{k},\frac{\lambda}{2k}\}$ for the steady case, $\sigma>\frac{\lambda}{2k}$ and $\sigma>-\frac{\lambda}{k}$ for the expanding and shrinking case, respectively.
Example 4.6.
An $\eta$-Einstein soliton $(g,\xi,a,b)$ in a radiation fluid is given by
$$\left\{\begin{array}[]{ll}a=\lambda-kp-\frac{div(\xi)}{3}\\
b=-4kp-\frac{div(\xi)}{3}\end{array}\right..$$
Remark the following facts:
1) From Examples 4.3 and 4.6 we deduce that the Ricci soliton in a radiation fluid is steady if $p=\frac{\lambda}{3k}$,
expanding if $p>\frac{\lambda}{3k}$ and shrinking if $p<\frac{\lambda}{3k}$.
2) In a general relativistic perfect fluid spacetime, if the vector field $\xi$ is torse-forming i.e. $\nabla_{X}\xi=\alpha[X+\eta(X)\xi]$, for any $X\in\chi(M)$ with $\alpha$ a nonzero real number, then $div(\xi)=3\alpha$. In this case, the existence of a Ricci soliton given by (50) for $b=0$, from Plebanski energy conditions implies $-2k\sigma\leq\alpha<0$ (precisely, $\alpha=-k(\sigma+p)$).
3) If the vector field $\xi$ is conformally Killing i.e. $L_{\xi}g=rg$ with $r$ a nonzero real number, then the existence of a Ricci soliton given by (50) for $b=0$, implies the vacuum case. Moreover, the soliton is steady if $p=\frac{\lambda}{k}+\frac{r}{2k}$, expanding if $p>\frac{\lambda}{k}+\frac{r}{2k}$ and shrinking if $p<\frac{\lambda}{k}+\frac{r}{2k}$.
4) The existence of a steady Ricci soliton or a (weakly) pseudo Ricci symmetric Ricci tensor field imply the same condition on the pressure of the fluid, a surprisingly fact being that the $1$-form $\alpha$, arbitrary chosen, does not appear effectively.
References
[1]
Adati, A. and Miyazawa, T.: On a Riemannian Space with recurrent conformal curvature, Tensor N.S. 18, 355–342 (1967).
[2]
Ahsan, Z. and Ali, M.: Curvature Tensor for the Spacetime of General Relativity: Int. J. Geom. Methods Mod. Phys. 14, 1750078 (2017) [13 pages].
[3]
Catino, G. and Mazzieri, L.: Gradient Einstein solitons, Nonlinear Analysis: Theory, Methods and Applications 132, 66–94 (2016).
[4]
Chaki, M. C. and Maity, R. K.: On quasi Einstein manifolds, Publ. Math. Debrecen 57, 297–306 (2000).
[5]
Chaki, M. C. and Ray, S.: Space-times with covariant-constant energy momentum tensor, Int. J. Theoret. Phys. 35, 1027–1032 (1996).
[6]
Cho, J. T. and Kimura, M.: Ricci solitons and real hypersurfaces in a complex space form, Tohoku Math. J. 61, no. 2, 205–212 (2009).
[7]
Chow, B., Lu, P., Ni, L.: Hamilton’s Ricci Flow, Graduate Studies in Mathematics, 77, AMS, Providence, RI, USA (2006).
[8]
De, U. C. and Gazi, A. K.: On pseudo Ricci symmetric manifolds, An. Şt. Univ. Al. I. Cuza Iaşi, tom LVIII, f. 1, 209–222 (2012).
[9]
De, U. C. and Ghosh, G. C.: On quasi-Einstein and special quasi-Einstein manifolds, Proc.
of the Int. Conf. of Mathematics and its Applications, Kuwait University, April 5-7, 2004, 178–191.
[10]
De, U. C. and Ghosh, G. C.: On quasi-Einstein manifolds, Periodica Math. Hungarica 48, no. 12, 223–231 (2004).
[11]
Deszcz, R., Hotlos, M. and Senturk, Z.: On curvature properties of quasi-Einstein hypersurfaces
in semi-Euclidean spaces, Soochow J. Math. 27, 375–389 (2001).
[12]
Hamilton, R. S.: The Ricci flow on surfaces, Math. and general relativity (Santa Cruz, CA, 1986), 237–262, Contemp. Math. 71, AMS (1988).
[13]
Matsumoto, K.: On Lorentzian paracontact manifolds, Bull. of Yamagata Univ., Nat. Sci. 12, 151–156 (1989).
[14]
Mishra, R. S.: A Course in Tensors with Applications to Riemannian Geometry, Pothishala Private 1st edition (1965).
[15]
Neill, O’.: Semi-Riemannian Geometry with Applications to Relativity, Pure and Applied Math. 103, Academic Press, New York (1983).
[16]
Pokhariyal, G. P. and Mishra, R. S.: Curvature tensors and their relativistics significance, Yokohama Math. J. 18, 105–108 (1970).
[17]
Shaikh, A. A.: On Lorentzian almost paracontact manifolds with a structure of the concircular type, Kyungpook J. Math. 43, 305–314 (2003).
[18]
Stephani, H.: General Relativity-An Introduction to the Theory of Gravitational Field, Cambridge University Press, Cambridge (1982).
[19]
Yano, K.: Concircular geometry I. Concircular transformations, Proc. Imp. Acad. Tokyo 16, 195–200 (1940).
[20]
Yano, K.: On torse forming direction in a Riemannian space, Proc. Imp. Acad. Tokyo 20, 340–345 (1944).
Department of Mathematics
West University of Timişoara
Bld. V. Pârvan nr. 4, 300223, Timişoara, România
[email protected] |
Can We Assess Mental Health through Social Media and Smart Devices? Addressing Bias in Methodology and Evaluation
Adam Tsakalidis
1University of Warwick, UK
1{a.tsakalidis, m.liakata, t.damoulas, a.i.cristea}@warwick.ac.uk
2The Alan Turing Institute, UK
2
Maria Liakata
1University of Warwick, UK
1{a.tsakalidis, m.liakata, t.damoulas, a.i.cristea}@warwick.ac.uk
2The Alan Turing Institute, UK
2
Theo Damoulas
1University of Warwick, UK
1{a.tsakalidis, m.liakata, t.damoulas, a.i.cristea}@warwick.ac.uk
2The Alan Turing Institute, UK
2
Alexandra I. Cristea
1University of Warwick, UK
1{a.tsakalidis, m.liakata, t.damoulas, a.i.cristea}@warwick.ac.uk
3Durham University, UK3
Abstract
Predicting mental health from smartphone and social media data on a longitudinal basis has recently attracted great interest, with very promising results being reported across many studies [3, 9, 13, 26]. Such approaches have the potential to revolutionise mental health assessment, if their development and evaluation follows a real world deployment setting. In this work we take a closer look at state-of-the-art approaches, using different mental health datasets and indicators, different feature sources and multiple simulations, in order to assess their ability to generalise. We demonstrate that under a pragmatic evaluation framework, none of the approaches deliver or even approach the reported performances. In fact, we show that current state-of-the-art approaches can barely outperform the most naïve baselines in the real-world setting, posing serious questions not only about their deployment ability, but also about the contribution of the derived features for the mental health assessment task and how to make better use of such data in the future.
Keywords:mental health bias evaluation wellbeing natural language processing smartphones sensors social media
\newfloatcommand
capbtabboxtable[][\FBwidth]
1 Introduction
Establishing the right indicators of mental well-being is a grand challenge posed by the World Health Organisation [7]. Poor mental health is highly correlated with low motivation, lack of satisfaction, low productivity and a negative economic impact [20]. The current approach is to combine census data at the population level [19], thus failing to capture well-being on an individual basis. The latter is only possible via self-reporting on the basis of established psychological scales, which are hard to acquire consistently on a longitudinal basis, and they capture long-term aggregates instead of the current state of the individual.
The widespread use of smart-phones and social media offers new ways of assessing mental well-being, and recent research [1, 2, 3, 5, 9, 10, 13, 14, 22, 23, 26] has started exploring the effectiveness of these modalities for automatically assessing the mental health of a subject, reporting very high accuracy. What is typically done in these studies is to use features based on the subjects’ smart phone logs and social media, to predict some self-reported mental health index (e.g., “wellbeing”, “depression” and others), which is provided either on a Likert scale or on the basis of a psychological questionnaire (e.g., PHQ-8 [12], PANAS [29], WEMWBS [25] and others).
Most of these studies are longitudinal, where data about individuals is collected over a period of time and predictions of mental health are made over a sliding time window. Having such longitudinal studies is highly desirable, as it can allow fine-grained monitoring of mental health. However, a crucial question is what constitutes an appropriate evaluation framework, in order for such approaches to be employable in a real world setting. Generalisation to previously unobserved users can only be assessed via leave-N-users-out cross-validation setups, where typically, N is equal to one (LOUOCV, see Table 1). However, due to the small number of subjects that are available, such generalisation is hard to achieve by any approach [13]. Alternatively, personalised models [3, 13] for every individual can be evaluated via a within-subject, leave-N-instances-out cross-validation (for N=1, LOIOCV), where an instance for a user $u$ at time $i$ is defined as a {$X_{ui}$, $y_{ui}$} tuple of {features(u, i), mental-health-score(u, i)}. In a real world setting, a LOIOCV model is trained on some user-specific instances, aiming to predict her mental health state at some future time points. Again however, the limited number of instances for every user make such models unable to generalize well. In order to overcome these issues, previous work [2, 5, 9, 10, 22, 26] has combined the instances {$X_{u_{j}i}$, $y_{u_{j}i}$} from different individuals $u_{j}$ and performed evaluation using randomised cross validation (MIXED). While such approaches can attain optimistic performance, the corresponding models fail to generalise to the general population and also fail to ensure effective personalised assessment of the mental health state of a single individual.
In this paper we demonstrate the challenges that current state-of-the-art models face, when tested in a real-world setting. We work on two longitudinal datasets with four mental health targets, using different features derived from a wide range of heterogeneous sources. Following the state-of-the-art experimental methods and evaluation settings, we achieve very promising results, regardless of the features we employ and the mental health target we aim to predict. However, when tested under a pragmatic setting, the performance of these models drops heavily, failing to outperform the most naïve – from a modelling perspective – baselines: majority voting, random classifiers, models trained on the identity of the user, etc. This poses serious questions about the contribution of the features derived from social media, smartphones and sensors for the task of automatically assessing well-being on a longitudinal basis. Our goal is to flesh out, study and discuss such limitations through extensive experimentation across multiple settings, and to propose a pragmatic evaluation and model-building framework for future research in this domain.
2 Related Work
Research in assessing mental health on a longitudinal basis aims to make use of relevant features extracted from various modalities, in order to train models for automatically predicting a user’s mental state (target), either in a classification or a regression manner [1, 2, 3, 9, 10, 13, 26]. Examples of state-of-the-art work in this domain are listed in Table 2, along with the number of subjects that was used and the method upon which evaluation took place. Most approaches have used the “MIXED” approach to evaluate models [1, 2, 5, 9, 10, 22, 26], which, as we will show, is vulnerable to bias, due to the danger of recognising the user in the test set and thus simply inferring her average mood score. LOIOCV approaches that have not ensured that their train/test sets are independent are also vulnerable to bias in a realistic setting [3, 13]. From the works listed in Table 2, only Suhara et al. [23] achieves unbiased results with respect to model generalisability; however, the features employed for their prediction task are derived from self-reported questionnaires of the subjects and not by automatic means.
3 Problem Statement
We first describe three major problems stemming from unrealistic construction and evaluation of mental health assessment models and then we briefly present the state-of-the-art in each case, which we followed in our experiments.
P1
Training on past values of the target variable: This issue arises when the past $N$ mood scores of a user are required to predict his/her next mood score in an autoregressive manner. Since such an approach would require the previous N scores of past mood forms, it would limit its ability to generalise without the need of manual user input in a continuous basis. This makes it impractical for a real-world scenario. Most importantly, it is difficult to measure the contribution of the features towards the prediction task, unless the model is evaluated using target feature ablation. For demonstration purposes, we have followed the experimental setup by LiKamWa et al. [13], which is one of the leading works in this field.
P2
Inferring test set labels: When training personalised models (LOIOCV) in a longitudinal study, it is important to make sure that there are no overlapping instances across consecutive time windows. Some past works have extracted features {$f(t-N)$, …, $f(t)$} over $N$ days, in order to predict the $score_{t}$ on day $N+1$ [3, 13]. Such approaches are biased if there are overlapping days of train/test data. To illustrate this problem we have followed the approach by Canzian and Musolesi [3], as one of the pioneering works on predicting depression with GPS traces, on a longitudinal basis.
P3
Predicting users instead of mood scores: Most approaches merge all the instances from different subjects, in an attempt to build user-agnostic models in a randomised cross-validation framework [2, 9, 10, 26]. This is problematic, especially when dealing with a small number of subjects, whose behaviour (as captured through their data) and mental health scores differ on an individual basis. Such approaches are in danger of “predicting” the user in the test set, since her (test set) features might be highly correlated with her features in the training set, and thus infer her average well-being score, based on the corresponding observations of the training set. Such approaches cannot guarantee that they will generalise on either a population-wide (LOUOCV) or a personalised (LOIOCV) level. In order to examine this effect in both a regression and a classification setting, we have followed the experimental framework by Tsakalidis et al. [26] and Jaques et al. [9].
3.1 P1: Training on past values of the target (LOIOCV, LOUOCV)
LiKamWa et al. [13] collected smartphone data from $32$ subjects over a period of two months. The subjects were asked to self-report their “pleasure” and “activeness” scores at least four times a day, following a Likert scale (1 to 5), and the average daily scores served as the two targets. The authors aggregated various features on social interactions (e.g., number of emails sent to frequently interacting contacts) and routine activities (e.g., browsing and location history) derived from the smartphones of the participants. These features were extracted over a period of three days, along with the two most recent scores on activeness and pleasure. The issue that naturally arises is that such a method cannot generalise to new subjects in the LOUOCV setup, as it requires their last two days of self-assessed scores. Moreover, in the LOIOCV setup, the approach is limited in a real world setting, since it requires the previous mental health scores by the subject to provide an estimate of her current state. Even in this case though, the feature extraction should be based on past information only – under LOIOCV in [13], the current mood score we aim at predicting is also used as a feature in the (time-wise) subsequent two instances of the training data.
Experiments in [13] are conducted under LOIOCV and LOUOCV, using Multiple Linear Regression (LR) with Sequential Feature Selection (in LOUOCV, the past two pairs of target labels of the test user are still used as features). In order to better examine the effectiveness of the features for the task, the same model can be tested without any ground-truth data as input. Nevertheless, a simplistic model predicting the per-subject average outperforms their LR in the LOUOCV approach, which poses the question of whether the smartphone-derived features can be used effectively to create a generalisable model that can assess the mental health of unobserved users. Finally, the same model tested in the LOIOCV setup achieves the lowest error; however, this is trained not only on target scores overlapping with the test set, but also on features derived over a period of three days, introducing further potential bias, as discussed in the following.
3.2 P2: Inferring Test Labels (LOIOCV)
Canzian and Musolesi [3] extracted mobility metrics from $28$ subjects to predict their depressive state, as derived from their daily self-reported PHQ-8 questionnaires. A 14-day moving average filter is first applied to the PHQ-8 scores and the mean value of the same day (e.g. Monday) is subtracted from the normalised scores, to avoid cyclic trends. This normalisation results into making the target score $s_{t}$ on day $t$ dependent on the past {$s_{t-14},...,s_{t-1}$} scores. The normalised PHQ-8 scores are then converted into two classes, with the instances deviating more than one standard deviation above the mean score of a subject being assigned to the class “1” (“0”, otherwise). The features are extracted over various time windows (looking at $T_{HIST}=\{0,...,14\}$ days before the completion of a mood form) and personalised model learning and evaluation are performed for every $T_{HIST}$ separately, using a LOIOCV framework.
What is notable is that the results improve significantly when features are extracted from a wider $T_{HIST}$ window. This could imply that the depressive state of an individual can be detected with a high accuracy if we look back at her history. However, by training and testing a model on instances whose features are derived from the same days, there is a high risk of over-fitting the model to the timestamp of the day in which the mood form was completed. In the worst-case scenario, there will be an instance in the train set whose features (e.g. total covered distance) are derived from the 14 days, 13 of which will also be used for the instance in the test set. Additionally, the target values of these two instances will also be highly correlated due to the moving average filter, making the task artificially easy for large $T_{HIST}$ and not applicable in a real-world setting.
While we focus on the approach in [3], a similar approach with respect to feature extraction was also followed in LiKamWa et al. [13] and Bogomolov et al. [2], extracting features from the past 2 and 2 to
5 days, respectively.
3.3 P3: Predicting Users (LOUOCV)
Tsakalidis et al. [26] monitored the behaviour of $19$ individuals over four months. The subjects were asked to complete two psychological scales [25, 29] on a daily basis, leading to three target scores (positive, negative, mental well-being); various features from smartphones (e.g., time spent on the preferred locations) and textual features (e.g., ngrams) were extracted passively over the 24 hours preceding a mood form timestamp. Model training and evaluation was performed in a randomised (MIXED) cross-validation setup, leading to high accuracy ($R^{2}=0.76$). However, a case demonstrating the potential user bias is when the models are trained on the textual sources: initially the highest $R^{2}$ ($0.22$) is achieved when a model is applied to the mental-wellbeing target; by normalising the textual features on a per-user basis, the $R^{2}$ increases to $0.65$. While this is likely to happen because the vocabulary used by different users is normalised, there is also the danger of over-fitting the trained model to the identity of the user. To examine this potential, the LOIOCV/LOUOCV setups need to be studied alongside the MIXED validation approach, with and without the per-user feature normalisation step.
A similar issue is encountered in Jaques et al. [9] who monitored $68$ subjects over a period of a month. Four types of features were extracted from survey and smart devices carried by subjects. Self-reported scores on a daily basis served as the ground truth. The authors labelled the instances with the top $30\%$ of all the scores as “happy” and the lowest $30\%$ as “sad” and randomly separated them into training, validation and test sets, leading to the same user bias issue. Since different users exhibit different mood scores on average [26], by selecting instances from the top and bottom scores, one might end up separating users and convert the mood prediction task into a user identification one. A more suitable task could have been to try to predict the highest and lowest scores of every individual separately, either in a LOIOCV or in a LOUOCV setup.
While we focus on the works of Tsakalidis et al. [26] and Jaques et al. [9], similar experimental setups were also followed in [10], using the median of scores to separate the instances and performing five-fold cross-validation, and by Bogomolov et al. in [2], working on a user-agnostic validation setting on 117 subjects to predict their happiness levels, and in [1], for the stress level classification task.
4 Experiments
4.1 Datasets
By definition, the aforementioned issues are feature-, dataset- and target-independent (albeit the magnitude of the effects may vary). To illustrate this, we run a series of experiments employing two datasets, with different feature sources and four different mental health targets.
Dataset 1: We employed the dataset obtained by Tsakalidis et al. [26], a pioneering dataset which contains a mix of longitudinal textual and mobile phone usage data for 30 subjects. From a textual perspective, this dataset consists of social media posts (1,854/5,167 facebook/twitter posts) and private messages (64,221/132/47,043 facebook/twitter/ SMS messages) sent by the subjects. For our ground truth, we use the {positive, negative, mental well-being} mood scores (in the ranges of [10-50], [10-50], [14-70], respectively) derived from self-assessed psychological scales during the study period.
Dataset 2: We employed the StudentLife dataset [28], which contains a wealth of information derived from the smartphones of 48 students during a 10-week period. Such information includes samples of the detected activity of the subject, timestamps of detected conversations, audio mode of the smartphone, status of the smartphone (e.g., charging, locked), etc. For our target, we used the self-reported stress levels of the students (range [0-4]), which were provided several times a day. For the approach in LiKamWa et al. [13], we considered the average daily stress level of a student as our ground-truth, as in the original paper; for the rest, we used all of the stress scores and extracted features based on some time interval preceding their completion, as described next, in 4.3111For P3, this creates the P2 cross-correlation issue in the MIXED/LOIOCV settings. For this reason, we ran the experiments by considering only the last entered score in a given day as our target. We did not witness any major differences that would alter our conclusions..
4.2 Task Description
We studied the major issues in the following experimental settings (see Table 3):
P1: Using Past Labels:
We followed the experimental setting in [13] (see section 3.1): we treated our task as a regression problem and used Mean Squared Error (MSE) and classification accuracy222Accuracy is defined in [13] as follows: 5 classes are assumed (e.g., [0, …, 4]) and the squared error $e$ between the centre of a class halfway towards the next class is calculated (e.g., 0.25). If the squared error of a test instance is smaller than $e$, then it is considered as having been classified correctly. for evaluation. We trained a Linear Regression (LR) model and performed feature selection using Sequential Feature Selection under the LOIOCV and LOUOCV setups; feature extraction is performed over the previous 3 days preceding the completion of a mood form. For comparison, we use the same baselines as in [13]: Model A always predicts the average mood score for a certain user (AVG); Model B predicts the last entered scores (LAST); Model C makes a prediction using the LR model trained on the ground-truth features only (-feat). We also include Model D, trained on non-target features only (-mood) in an unbiased LOUOCV setting.
P2: Inferring Test Labels: We followed the experimental setting presented in [3]. We process our ground-truth in the same way as the original paper (see section 3.2) and thus treat our task as a binary classification problem. We use an SVM${}_{RBF}$ classifier, using grid search for parameter optimisation, and perform evaluation using specificity and sensitivity. We run experiments in the LOIOCV and LOUOCV settings, performing feature extraction at different time windows ($T_{HIST}=\{1,...,14\}$). In order to better demonstrate the problem that arises here, we use the previous label classifier (LAST) and the SVM classifier to which we feed only the mood timestamp as a feature (DATE) for comparison. Finally, we replace our features with completely random data and train the same SVM with $T_{HIST}=14$ by keeping the same ground truth, performing 100 experiments and reporting averages of sensitivity and specificity (RAND).
P3: Predicting Users: We followed the evaluation settings of two past works (see section 3.3), with the only difference being the use of 5-fold CV instead of a train/dev/test split that was used in [9]. The features of every instance are extracted from the past day before the completion of a mood form. In Experiment 1 we follow the setup in [26]: we perform 5-fold CV (MIXED) using SVM (SVR${}_{RBF}$) and evaluate performance based on $R^{2}$ and $RMSE$. We compare the performance when tested under the LOIOCV/LOUOCV setups, with and without the per-user feature normalisation step. We also compare the performance of the MIXED setting, when our model is trained on the one-hot-encoded user id only. In Experiment 2 we follow the setup in [9]: we label the instances as “high” (“low”), if they belong to the top-30% (bottom-30%) of mood score values (“UNIQ” – for “unique” – setup). We train an SVM classifier in 5-fold CV using accuracy for evaluation and compare performance in the LOIOCV and LOUOCV settings. In order to further examine user bias, we perform the same experiments, this time by labelling the instances on a per-user basis (‘‘PERS’’ -- for ‘‘personalised’’ -- setup), aiming to predict the per-user high/low mood days333In cases where the lowest of the top-30% scores ($s$) was equal to the highest of the bottom-30% scores, we excluded the instances with score $s$..
4.3 Features
For Dataset 1, we first defined a ‘‘user snippet’’ as the concatenation of all texts generated by a user within a set time interval, such that the maximum time difference between two consecutive document timestamps is less than 20 minutes. We performed some standard noise reduction steps (converted text to lowercase, replaced URLs/user mentions and performed language identification444https://pypi.python.org/pypi/langid and tokenisation [6]). Given a mood form and a set of snippets produced by a user before the completion of a mood form, we extracted some commonly used feature sets for every snippet written in English [26], which were used in all experiments. To ensure sufficient data density, we excluded users for whom we had overall fewer than $25$ snippets on the days before the completion of the mood form or fewer than $40$ mood forms overall, leading to $27$ users and $2,368$ mood forms. For Dataset 2, we extracted the features presented in Table 4. We only kept the users that had at least 10 self-reported stress questionnaires, leading to $44$ users and $2,146$ instances. For our random experiments used in P2, in Dataset 1 we replaced the text representation of every snippet with random noise ($\mu=0,\sigma=1$) of the same feature dimensionality; in Dataset 2, we replaced the actual inferred value of every activity/audio sample with a random inference class; we also replaced each of the detected conversation samples and samples detected in a dark environment/locked/charging, with a random number ($<$100, uniformly distributed) indicating the number of pseudo-detected samples.
5 Results
5.1 P1: Using Past Labels
Table 5 presents the results on the basis of the methodology by LiKamWa et al. [13], along with the average scores reported in [13] – note that the range of the mood scores varies on a per-target basis; hence, the reported results of different models should be compared among each other when tested on the same target.
As in [13], always predicting the average score (AVG) for an unseen user performs better than applying a LR model trained on other users in a LOUOCV setting. If the same LR model used in LOUOCV is trained without using the previously self-reported ground-truth scores (Model D, -mood), its performance drops further. This showcases that personalised models are needed for more accurate mental health assessment (note that the AVG baseline is, in fact, a personalised baseline) and that there is no evidence that we can employ effective models in real-world applications to predict the mental health of previously unseen individuals, based on this setting.
The accuracy of LR under LOIOCV is higher, except for the “stress” target, where the performance is comparable to LOUOCV and lower than the AVG baseline. However, the problem in LOIOCV is the fact that the features are extracted based on the past three days, thus creating a temporal cross-correlation in our input space. If a similar correlation exists in the output space (target), then we end up in danger of overfitting our model to the training examples that are temporally close to the test instance. This type of bias is essentially present if we force a temporal correlation in the output space, as studied next.
5.2 P2: Inferring Test Labels
The charts in Fig. 1 (top) show the results by following the LOIOCV approach from Canzian and Musolesi [3]. The pattern that these metrics take is consistent and quite similar to the original paper: specificity remains at high values, while sensitivity increases as we increase the time window from which we extract our features. The charts on the bottom in Fig. 1 show the corresponding results in the LOUOCV setting. Here, such a generalisation is not feasible, since the increases in sensitivity are accompanied by sharp drops in the specificity scores.
The arising issue though lies in the LOIOCV setting. By training and testing on the same days (for $T_{HIST}>1$), the kernel matrix takes high values for cells which are highly correlated with respect to time, making the evaluation of the contribution of the features difficult. To support this statement, we train the same model under LOIOCV, using only on the mood form completion date (Unix epoch) as a feature. The results are very similar to those achieved by training on $T_{HIST}=14$ (see Table 6). We also include the results of another naïve classifier (LAST), predicting always the last observed score in the training set, which again achieves similar results. The clearest demonstration of the problem though is by comparing the results of the RAND against the FEAT classifier, which shows that under the proposed evaluation setup we can achieve similar performance if we replace our inputs with random data, clearly demonstrating the temporal bias that can lead to over-optimistic results, even in the LOIOCV setting.
5.3 P3: Predicting Users
Experiment 1:
Table 3 shows the results based on the evaluation setup of Tsakalidis et al. [26]. In the MIXED cases, the pattern is consistent with [26], indicating that normalising the features on a per-user basis yields better results, when dealing with sparse textual features (positive, negative, wellbeing targets).
The explanation of this effect lies within the danger of predicting the user’s identity instead of her mood scores. This is why the per-user normalisation does not have any effect for the stress target, since for that we are using dense features derived from smartphones: the vocabulary used by the subjects for the other targets is more indicative of their identity. In order to further support this statement, we trained the SVR model using only the one-hot encoded user id as a feature, without any textual features. Our results yielded $R^{2}$={$0.64$, $0.50$, $0.66$} and $RMSE$={$5.50$, $5.32$, $6.50$} for the {positive, negative, wellbeing} targets, clearly demonstrating the user bias in the MIXED setting.
The RMSEs in LOIOCV are the lowest, since different individuals exhibit different ranges of mental health scores. Nevertheless, $R^{2}$ is slightly negative, implying again that the average predictor for a single user provides a better estimate for her mental health score. Note that while the predictions across all individuals seem to be very accurate (see Fig. 3), by separating them on a per-user basis, we end up with a negative $R^{2}$.
In the unbiased LOUOCV setting the results are, again, very poor. The reason for the high differences observed between the three settings is provided by the $R^{2}$ formula itself ($1-(\sum_{i}(pred_{i}-y_{i})^{2})/(\sum_{i}(y_{i}-\bar{y})^{2}))$. In the MIXED case, we train and test on the same users, while $\bar{y}$ is calculated as the mean of the mood scores across all users, whereas in the LOIOCV/LOUOCV cases, $\bar{y}$ is calculated for every user separately. In MIXED, by identifying who the user is, we have a rough estimate of her mood score, which is by itself a good predictor, if it is compared with the average predictor across all mood scores of all users. Thus, the effect of the features in this setting cannot be assessed with certainty.
Experiment 2:
Table 7 displays our results based on Jaques et al. [9] (see section 3.3). The average accuracy on the “UNIQ” setup is higher by $14\%$ compared to the majority classifier in MIXED. The LOIOCV setting also yields very promising results (mean accuracy: $81.17\%$). As in all previous cases, in LOUOCV our models fail to outperform the majority classifier. A closer look at the LOIOCV and MIXED results though reveals the user bias issue that is responsible for the high accuracy. For example, $33\%$ of the users had all of their “positive” scores binned into one class, as these subjects were exhibiting higher (or lower) mental health scores throughout the experiment, whereas another $33\%$ of the subjects had $85\%$ of their instances classified into one class. By recognising the user, we can achieve high accuracy in the MIXED setting; in the LOIOCV, the majority classifier can also achieve at least $85\%$ accuracy for $18$/$27$ users.
In the “PERS” setup, we removed the user bias, by separating the two classes on a per-user basis. The results now drop heavily even in the two previously well-performing settings and can barely outperform the majority classifier. Note that the task in Experiment 2 is relatively easier, since we are trying to classify instances into two classes which are well-distinguished from each other from a psychological point of view. However, by removing the user bias, the contribution of the user-generated features to this task becomes once again unclear.
6 Proposal for Future Directions
Our results emphasize the difficulty of automatically predicting individuals’ mental health scores in a real-world setting and demonstrate the dangers due to flaws in the experimental setup. Our findings do not imply that the presented issues will manifest themselves to the same degree in different datasets – e.g., the danger of predicting the user in the MIXED setting is higher when using the texts of 27 users rather than sensor-based features of more users [1, 2, 9, 22]. Nevertheless, it is crucial to establish appropriate evaluation settings to avoid providing false alarms to users, if our aim is to build systems that can be deployed in practice. To this end, we propose model building and evaluation under the following:
•
LOUOCV: By definition, training should be performed strictly on features and target data derived from a sample of users and tested on a completely new user, since using target data from the unseen user as features violates the independence hypothesis. A model trained in this setting should achieve consistently better results on the unseen user compared to the naïve (from a modelling perspective) model that always predicts his/her average score.
•
LOIOCV: By definition, the models trained under this setting should not violate the iid hypothesis. We have demonstrated that the temporal dependence between instances in the train and test set can provide over-optimistic results. A model trained on this setting should consistently outperform naïve, yet competitive, baseline methods, such as the last-entered mood score predictor, the user’s average mood predictor and the auto-regressive model.
Models that can be effectively applied in any of the above settings could revolutionise the mental health assessment process while providing us in an unbiased setting with great insights on the types of behaviour that affect our mental well-being. On the other hand, positive results in the MIXED setting cannot guarantee model performance in a real-world setting in either LOUOCV or LOIOCV, even if they are compared against the user average baseline [4].
Transfer learning approaches can provide significant help in the LOUOCV setting. However, these assume that single-domain models have been effectively learned beforehand – but all of our single-user (LOIOCV) experiments provided negative results. Better feature engineering through latent feature representations may prove to be beneficial. While different users exhibit different behaviours, these behaviours may follow similar patterns in a latent space. Such representations have seen great success in recent years in the field of natural language processing [15], where the aim is to capture latent similarities between seemingly diverse concepts and represent every feature based on its context. Finally, working with larger datasets can help in providing more data to train on, but also in assessing the model’s ability to generalise in a more realistic setting.
7 Conclusion
Assessing mental health with digital media is a task which could have great impact on monitoring of mental well-being and personalised health. In the current paper, we have followed past experimental settings to evaluate the contribution of various features to the task of automatically predicting different mental health indices of an individual. We find that under an unbiased, real-world setting, the performance of state-of-the-art models drops significantly, making the contribution of the features impossible to assess. Crucially, this holds for both cases of creating a model that can be applied in previously unobserved users (LOUOCV) and a personalised model that is learned for every user individually (LOIOCV).
Our major goal for the future is to achieve positive results in the LOUOCV setting. To overcome the problem of having only few instances from a diversely behaving small group of subjects, transfer learning techniques on latent feature representations could be beneficial. A successful model in this setting would not only provide us with insights on what types of behaviour affect mental state, but could also be employed in a real-world system without the danger of providing false alarms to its users.
Acknowledgements
The current work was supported by the EPSRC through the University of Warwick’s CDT in Urban Science and Progress (grant EP/L016400/1) and through The Alan Turing Institute (grant EP/N510129/1). We would like to thank the anonymous reviewers for their detailed feedback and the authors of the works that were analysed in our paper (N. Jaques, R. LiKamWa, M. Musolesi) for the fruitful discussions over several aspects of the presented challenges.
References
[1]
Bogomolov, A., Lepri, B., Ferron, M., Pianesi, F., Pentland, A.S.: Pervasive
Stress Recognition for Sustainable Living. In: Pervasive Computing
and Communications Workshops (PERCOM Workshops), 2014 IEEE International
Conference on. pp. 345–350. IEEE (2014)
[2]
Bogomolov, A., Lepri, B., Pianesi, F.: Happiness Recognition from Mobile
Phone Data. In: Social Computing (SocialCom), 2013 International
Conference on. pp. 790–795. IEEE (2013)
[3]
Canzian, L., Musolesi, M.: Trajectories of Depression: Unobtrusive
Monitoring of Depressive States by Means of Smartphone Mobility
Traces Analysis. In: Proceedings of the 2015 ACM International Joint
Conference on Pervasive and Ubiquitous Computing. pp. 1293–1304. ACM (2015)
[4]
DeMasi, O., Kording, K., Recht, B.: Meaningless Comparisons Lead to False
Optimism in Medical Machine Learning. PLoS One 12(9),
e0184604 (2017)
[5]
Farhan, A.A., Yue, C., Morillo, R., Ware, S., Lu, J., Bi, J., Kamath, J.,
Russell, A., Bamis, A., Wang, B.: Behavior vs. Introspection: Refining
Prediction of Clinical Depression via Smartphone Sensing Data.
In: Wireless Health. pp. 30–37 (2016)
[6]
Gimpel, K., Schneider, N., O’Connor, B., Das, D., Mills, D., Eisenstein, J.,
Heilman, M., Yogatama, D., Flanigan, J., Smith, N.A.: Part-of-speech
Tagging for Twitter: Annotation, Features, and Experiments. In:
Proceedings of the 49th Annual Meeting of the Association for Computational
Linguistics: Human Language Technologies: Short Papers-Volume 2. pp.
42–47. Association for Computational Linguistics (2011)
[7]
Herrman, H., Saxena, S., Moodie, R., et al.: Promoting Mental Health:
Concepts, Emerging Evidence, Practice: A Report of the World
Health Organization, Department of Mental Health and Substance
Abuse in Collaboration with the Victorian Health Promotion
Foundation and the University of Melbourne. World Health Organization
(2005)
[8]
Hu, M., Liu, B.: Mining and Summarizing Customer Reviews. In: Proceedings
of the 10th ACM SIGKDD International Conference on Knowledge Discovery and
Data Mining. pp. 168–177. ACM (2004)
[9]
Jaques, N., Taylor, S., Azaria, A., Ghandeharioun, A., Sano, A., Picard, R.:
Predicting Students’ Happiness from Physiology, Phone, Mobility,
and Behavioral Data. In: Affective Computing and Intelligent Interaction
(ACII), 2015 International Conference on. pp. 222–228. IEEE (2015)
[10]
Jaques, N., Taylor, S., Sano, A., Picard, R.: Multi-task, Multi-kernel
Learning for Estimating Individual Wellbeing. In: Proc. NIPS Workshop
on Multimodal Machine Learning, Montreal, Quebec (2015)
[11]
Kiritchenko, S., Zhu, X., Mohammad, S.M.: Sentiment Analysis of Short
Informal Texts. Journal of Artificial Intelligence Research 50,
723–762 (2014)
[12]
Kroenke, K., Strine, T.W., Spitzer, R.L., Williams, J.B., Berry, J.T., Mokdad,
A.H.: The PHQ-8 as a Measure of Current Depression in the General
Population. Journal of Affective Disorders 114(1), 163–173
(2009)
[13]
LiKamWa, R., Liu, Y., Lane, N.D., Zhong, L.: Moodscope: Building a Mood
Sensor from Smartphone Usage Patterns. In: Proceeding of the 11th
Annual International Conference on Mobile Systems, Applications, and
Services. pp. 389–402. ACM (2013)
[14]
Ma, Y., Xu, B., Bai, Y., Sun, G., Zhu, R.: Daily Mood Assessment based on
Mobile Phone Sensing. In: Wearable and Implantable Body Sensor Networks
(BSN), 2012 9th International Conference on. pp. 142–147. IEEE (2012)
[15]
Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed
Representations of Words and Phrases and their Compositionality. In:
Advances in Neural Information Processing Systems. pp. 3111–3119 (2013)
[16]
Mohammad, S.: #Emotional Tweets. In: *SEM 2012: The 1st Joint Conference
on Lexical and Computational Semantics – Volume 1: Proceedings of the Main
Conference and the Shared Task, and Volume 2: Proceedings of the 6th
International Workshop on Semantic Evaluation (SemEval 2012). pp. 246–255.
Association for Computational Linguistics (2012)
[17]
Mohammad, S., Dunne, C., Dorr, B.: Generating High-Coverage Semantic
Orientation Lexicons from Overtly Marked Words and a Thesaurus.
In: Proceedings of the 2009 Conference on Empirical Methods in Natural
Language Processing: Volume 2. pp. 599–608. Association for Computational
Linguistics (2009)
[18]
Nielsen, F.Å.: A new ANEW: Evaluation of a Word List for
Sentiment Analysis in Microblogs. In: Workshop on’Making Sense of
Microposts: Big Things Come in Small Packages. pp. 93–98 (2011)
[19]
OECD: How’s Life? 2013: Measuring Well-being (2013),
http://dx.doi.org/10.1787/9789264201392-en
[20]
Olesen, J., Gustavsson, A., Svensson, M., Wittchen, H.U., Jönsson, B.: The
Economic Cost of Brain Disorders in Europe. European Journal of
Neurology 19(1), 155–162 (2012)
[21]
Preoţiuc-Pietro, D., Volkova, S., Lampos, V., Bachrach, Y., Aletras, N.:
Studying User Income through Language, Behaviour and Affect in
Social Media. PloS One 10(9), e0138717 (2015)
[22]
Servia-Rodríguez, S., Rachuri, K.K., Mascolo, C., Rentfrow, P.J., Lathia,
N., Sandstrom, G.M.: Mobile Sensing at the Service of Mental
Well-being: a Large-scale Longitudinal Study. In: Proceedings of the
26th International Conference on World Wide Web. pp. 103–112. International
World Wide Web Conferences Steering Committee (2017)
[23]
Suhara, Y., Xu, Y., Pentland, A.: DeepMood: Forecasting Depressed Mood
based on Self-reported Histories via Recurrent Neural Networks. In:
Proceedings of the 26th International Conference on World Wide Web. pp.
715–724. International World Wide Web Conferences Steering Committee (2017)
[24]
Tang, D., Wei, F., Yang, N., Zhou, M., Liu, T., Qin, B.: Learning
Sentiment-specific Word Embedding for Twitter Sentiment
Classification. In: Proceedings of the 52nd Annual Meeting of the
Association for Computational Linguistics (Volume 1: Long Papers). vol. 1,
pp. 1555–1565 (2014)
[25]
Tennant, R., Hiller, L., Fishwick, R., Platt, S., Joseph, S., Weich, S.,
Parkinson, J., Secker, J., Stewart-Brown, S.: The Warwick-Edinburgh
Mental Well-being Scale (WEMWBS): Development and UK
Validation. Health and Quality of life Outcomes 5(1), 63 (2007)
[26]
Tsakalidis, A., Liakata, M., Damoulas, T., Jellinek, B., Guo, W., Cristea,
A.I.: Combining Heterogeneous User Generated Data to Sense
Well-being. In: Proceedings of the 26th International Conference on
Computational Linguistics. pp. 3007–3018 (2016)
[27]
Wang, R., Aung, M.S., Abdullah, S., Brian, R., Campbell, A.T., Choudhury, T.,
Hauser, M., Kane, J., Merrill, M., Scherer, E.A., et al.: CrossCheck:
Toward Passive Sensing and Detection of Mental Health Changes
in People with Schizophrenia. In: Proceedings of the 2016 ACM
International Joint Conference on Pervasive and Ubiquitous Computing. pp.
886–897. ACM (2016)
[28]
Wang, R., Chen, F., Chen, Z., Li, T., Harari, G., Tignor, S., Zhou, X.,
Ben-Zeev, D., Campbell, A.T.: StudentLife: Assessing Mental Health,
Academic Performance and Behavioral Trends of College Students
using Smartphones. In: Proceedings of the 2014 ACM International Joint
Conference on Pervasive and Ubiquitous Computing. pp. 3–14. ACM (2014)
[29]
Watson, D., Clark, L.A., Tellegen, A.: Development and Validation of Brief
Measures of Positive and Negative Affect: The PANAS Scales.
Journal of Personality and Social Psychology 54(6), 1063
(1988)
[30]
Zhu, X., Kiritchenko, S., Mohammad, S.M.: NRC-Canada-2014: Recent
Improvements in the Sentiment Analysis of Tweets. In: Proceedings of
the 8th International Workshop on Semantic Evaluation (SemEval 2014). pp.
443–447. Citeseer (2014) |
Quillen–Suslin Theory for Classical
Groups: Revisited over Graded Rings
Rabeya Basu
and
Manish Kumar Singh
(Date:: )
Corresponding Author: Rabeya Basu – [email protected]
2010 Mathematics Subject Classification:
11E57, 13A02, 13B25, 13B99, 13C10, 13C99, 15B99.
Key words: Bilinear forms, Symplectic and
Orthogonal forms, Graded rings.
Abstract: In this paper we deduce a graded version of Quillen–Suslin’s Local-Global Principle for the traditional classical groups, viz. general linear,
symplectic and orthogonal groups and establish its equivalence of the normality property of the respective elementary subgroups. This
generalizes previous result of Basu–Rao–Khanna; cf. [3]. Then, as an application,
we establish an analogue Local-Global Principle for the commutator subgroups of the special linear and symplectic groups.
Finally, by using Swan-Weibel’s homotopy trick, we establish
graded analogue of the Local-Global Principle
for the transvection subgroups of the full automorphism groups of the linear, symplectic and orthogonal modules. This generalizes
the previous result of Bak–Basu–Rao; cf. [5].
1. Introduction
In [3], the first author with Ravi A. Rao and Reema Khanna revisited Quillen-Suslin’s Local-Global Principle
for the linear, symplectic and orthogonal groups to show the equivalence of the
Local-Global Principle and Normality property of their Elementary subgroups. A “relative version”
is also established recently by the same authors; cf. [6]. In [3], the authors have also deduced a Local-Global Principle for
the commutator subgroups of the special linear groups, and symplectic groups. In this article we aim to revisit those results over commutative $\mathbb{N}$-graded
rings with identity. We follow the line of proof described in [3].
For the linear case, the graded version of the Local-Global Principle was studied by Chouinard in [7], and by J. Gubeladze in [8], [9].
Though the analogue results are expected for the symplectic and orthogonal groups, according to our best knowledge,
that is not written explicitly in any existing literatures. By deducing the equivalence, we establish the graded version of the
Local-Global Principle for the above two types of classical groups.
To generalize the existing results, from the polynomial rings to the graded rings, one has to use the line of proof of Quillen–Suslin’s patching argument
(cf. [12], [13]),
and Swan–Weibel’s homotopy trick.
For a nice exposition we refer to [8] by J. Gubeladze.
Though, we are writing this article for commutative graded rings, one may also
consider standard graded algebras which are finite over its center. For a commutative $\mathbb{N}$-graded ring with 1,
we establish:
$(1)$ (Theorem 3.8) Normality of the elementary subgroups is equivalent to the graded
Local-Global Principle for the linear, symplectic and orthogonal groups.
$(2)$ (Theorem 4.8) Analogue of Quillen–Suslin’s Local-Global
Principle for the commutator subgroups of the special linear and symplectic groups over graded rings.
In [5], the following was established by the first author with
A. Bak and R.A. Rao:
(Local-Global Principle for the Transvection Subgroups)
An analogue
of Quillen–Suslin’s Local-Global Principle for the transvection subgroup of
the automorphism group of projective,
symplectic and orthogonal modules of global rank at least 1 and
local rank at least 3, under the assumption that the projective module has
constant local rank and that the symplectic and orthogonal modules are locally
an orthogonal sum of a constant number of hyperbolic planes.
In this article, we observe that by using Swan-Weibel’s homotopy trick,
one gets an analogue statement for the graded case. More precisely, we deduce the following fact:
$(3)$ (Theorem 5.4) Let $A=\bigoplus_{i=0}^{\infty}A_{i}$ be a graded ring and $Q\simeq P\oplus A_{0}$ be a
projective $A_{0}$-module ($Q\simeq P\oplus\mathbb{H}(A_{0})$ for symplectic and orthogonal modules). If an automorphism of $Q\otimes_{A_{0}}A$
is locally in the transvection subgroup, then it is globally in the transvection subgroup.
2. Definitions and Notations
Let us start by recalling the following well-known
fact: Given a ring $R$ and a subring $R_{0}$, one can express $R$ as a
direct limit of subrings which are finitely generated over $R_{0}$ as
rings. Considering $R_{0}$ to be the minimal subring (i.e. the
image of $\mathbb{Z}$), it follows that every ring is a direct limit of
Noetherian rings. Hence we may consider $R$ to be Noetherian (cf. Pg 271 [11]).
Throughout this paper we assume $A$ to be a Noetherian, commutative graded ring with identity $1$. We shall write
$A=A_{0}\oplus A_{1}\oplus A_{2}\oplus\cdots$. As we know the multiplication in a graded ring satisfies the following property:
For all $i,j$, $A_{i}A_{j}\subset A_{i+j}$. An element $a\in A$ will be denoted by $a=a_{0}+a_{1}+a_{2}+\cdots$, where $a_{i}\in A_{i}$ for each $i$, and
and all but finitely many $a_{i}$ are zero. Let $A_{+}=A_{1}\oplus A_{2}\oplus\cdots$.
Graded structure of $A$ induces a graded structure on ${\rm M}_{n}(A)$ (ring of $n\times n$ matrices).
Let $S$ be a multiplicatively closed subset of $A_{0}$.
Then for a non-zero divisor $s\in S$ we shall denote the
localization of a matrix $\alpha\in{\rm M}_{n}(A)$ to ${\rm M}_{n}(A_{s})$ as $\alpha_{s}$.
Otherwise, $\alpha_{i},\;i\in\mathbb{Z}$ will represent the $i$-th component of $\alpha$.
We shall use standard notations ${\rm GL}_{n}(A)$ and ${\rm SL}_{n}(A)$ to denote the group of invertible matrices and its subgroup of all
invertible matrices with determinant $1$ respectively.
We recall the well-known “Swan-Weibel’s homotopy trick”, which is the main ingredient to handle the graded case.
Definition 2.1.
Let $a\in A_{0}$ be a fixed element. We fix an element $b=b_{0}+b_{1}+\cdots$ in $A$ and define a ring homomorphism
$\epsilon:A\rightarrow A[X]$ given by
$$\epsilon(b)=\epsilon(b_{0}+b_{1}+\cdots)\;=\;b_{0}+b_{1}X+b_{2}X^{2}+\cdots+b_%
{i}X^{i}+\cdots.$$
Then we evaluate the polynomial $\epsilon(b)(X)$ at $X=a$ and denote the image by $b^{+}(a)$ i.e. $b^{+}(a)=\epsilon(b)(a)$.
Note that $\big{(}b^{+}(x)\big{)}^{+}(y)=b^{+}(xy)$. Observe, $b_{0}=b^{+}(0)$.
We shall use this fact frequently.
The above ring homomorphism $\epsilon$ induces a group homomorphism at the ${\rm GL}_{n}(A)$ level for every $n\geq 1$, i.e. for
$\alpha\in{\rm GL}_{n}(A)$ we get a map
$$\epsilon:{\rm GL}_{n}(A)\rightarrow{\rm GL}_{n}(A[X])\text{ defined by}$$
$$\alpha=\alpha_{0}\oplus\alpha_{1}\oplus\alpha_{2}\oplus\cdots\mapsto\alpha_{0}%
\oplus\alpha_{1}X\oplus\alpha_{2}X^{2}\cdots,$$
where $\alpha_{i}\in{\rm GL}_{n}(A_{i})$.
As above for $a\in A_{0}$, we define $\alpha^{+}(a)$ as
$$\alpha^{+}(a)=\epsilon(\alpha)(a).$$
Now we are going to recall the definitions of the traditional classical groups, viz. the general linear groups, the symplectic and orthogonal groups
(of even size) and their
type subgroups, viz. elementary (symplectic and orthogonal elementary resp.) subgroups.
Let $e_{ij}$ be the matrix with $1$ in the $ij$-position and $0$’s elsewhere.
The matrices of the form $\{{\rm E}_{ij}(\lambda):\lambda\in A\mid i\neq j\},$ where
$${\rm E}_{ij}(\lambda)={\rm I}_{n}+\lambda e_{ij}$$
are called the elementary generators.
Definition 2.2.
The subgroup generated by the set $\{{\rm E}_{ij}(\lambda):\lambda\in A\mid i\neq j\}$, is called the
elementary subgroup of the general linear group and is denoted by ${\rm E}_{n}(A)$.
Observe that ${\rm E}_{n}(A)\subseteq{\rm SL}_{n}(A)\subseteq{\rm GL}_{n}(A)$.
Let $\sigma$ be a permutation defined by: For $i\in\{1,\ldots,2m\}$
$$\sigma(2i)=2i-1\text{ and }\sigma(2i-1)=2i.$$
With respect to this permutation, we define two $2m\times 2m$ forms (viz.) $\psi_{m}$ and $\widetilde{\psi}_{m}$ as follows: For $m>1$, let
$$\psi_{1}=\begin{bmatrix}0&1\\
-1&0\end{bmatrix}\text{ and for $m>1$},\,\,\psi_{m}=\begin{bmatrix}\psi_{m-1}&%
0\\
0&{\rm I}_{2}\end{bmatrix},$$
$$\widetilde{\psi}_{1}=\begin{bmatrix}0&1\\
1&0\end{bmatrix}\text{ and for $m>1$},\,\,\widetilde{\psi}_{m}=\begin{bmatrix}%
\widetilde{\psi}_{m-1}&0\\
0&{\rm I}_{2}\end{bmatrix}.$$
Using above two forms, we define the following traditional classical groups:
Definition 2.3.
A matrix $\alpha\in{\rm GL}_{2m}(A)$ is called symplectic if it fixes $\psi_{m}$ under the action of conjugation, i.e.
$$\alpha^{t}\psi_{m}\alpha=\psi_{m}.$$
The group generated by all symplectic matrices is called the symplectic group and is denoted by ${\rm Sp}_{2m}(A)={\rm Sp}_{n}(A)$, where $n=2m$.
Definition 2.4.
A matrix $\alpha\in{\rm GL}_{2m}(A)$ is called orthogonal if it fixes $\widetilde{\psi}_{m}$ under the action of conjugation, i.e.
$$\alpha^{t}\widetilde{\psi}_{m}\alpha=\widetilde{\psi}_{m}.$$
The group generated by all orthogonal matrices is called orthogonal group and is denoted by ${\rm O}_{2m}(A)={\rm O}_{n}(A)$, where $n=2m$.
Definition 2.5.
The matrices of the form
$$\{{se}_{ij}(z)\in{\rm GL}_{2m}(A):z\in A\mid i\neq j\},$$
where
$$se_{ij}(z)={\rm I}_{2m}+ze_{ij}\text{ if }i=\sigma(j)$$
or
$$se_{ij}(z)={\rm I}_{2m}+ze_{ij}-(-1)^{i+j}ze_{\sigma(j)\sigma(i)}\text{ if }i%
\neq\sigma(j)\text{ and }i<j.$$
are called symplectic elementary generators.
The subgroup generated by symplectic elementary generators is called symplectic elementary group.
Definition 2.6.
The matrices of the form $\{{oe}_{ij}(z):z\in A\mid i\neq j\},$ where
$$oe_{ij}(z)={\rm I}_{2m}+ze_{ij}-ze_{\sigma(j)\sigma(i)},\text{ if }i\neq\sigma%
(j)\text{ and }i<j.$$
are called orthogonal elementary generators.
The subgroup generated by orthogonal elementary generators is called orthogonal elementary group.
Remark 2.7.
It is a well known fact that the elementary subgroups are normal subgroups of the respective classical groups; for $n\geq 3$ in the
linear case, $n=2m\geq 4$ in the symplectic case, and $n=2m\geq 6$ in the orthogonal case.
The linear case was due to A. Suslin (cf. [15]), symplectic was proved by V. Kopeiko (cf. [15]).
Finally the orthogonal case was proved by Suslin–Kopeiko (cf. [14]).
We use the following notations
to treat above three groups uniformly:
$${\rm G}(n,A)\text{ will denote }{\rm GL}_{n}(A)\text{ or }{\rm Sp}_{2m}(A)%
\text{ or }{\rm O}_{2m}(A).$$
$${\rm E}(n,A)\text{ will denote }{\rm E}_{n}(A)\text{ or }{\rm EO}_{2m}(A)\text%
{ or }{\rm ESp}_{2m}(A).$$
$${\rm S}(n,A)\text{ will denote }{\rm SL}_{n}(A)\text{ or }{\rm SO}_{2m}(A),%
\text{ and }$$
$$ge_{ij}\text{ will denote the elementary generators of }{\rm E}_{n}(A)\text{ %
or }{\rm ESp}_{n}(A)\text{or }{\rm EO}_{n}(A).$$
Here ${\rm SO}_{n}(A)$ is the subgroup of ${\rm O}_{n}(A)$ with determinant $1$. Throughout the paper we will assume $n=2m$
to treat the non-linear cases. We shall assume $n\geq 3$ while treating the linear case, and
$n\geq 6$; i.e. $m\geq 3$ while treating the symplectic and orthogonal cases.
Let
$$\widetilde{v}=\begin{cases}v\psi_{m}\text{ in the symplectic case, and }\\
v\widetilde{\psi}_{m}\text{ in the orthogonal case. }\end{cases}$$
Definition 2.8.
Let $v,w\in A^{n}$ be vectors of length $n$, then we define the inner product $\langle v,w\rangle$ as follows:
(1)
$\langle v,w\rangle$ = $v^{t}w$ in the linear case,
(2)
$\langle v,w\rangle$ = $\widetilde{v}w$ in the symplectic and orthogonal cases.
Definition 2.9.
Let $v,w\in A$, we define the map $M:A^{n}\times A^{n}\rightarrow M_{n}(A)$ as follows:
(1)
$M(v,w)=vw^{t}$ in the linear case,
(2)
$M(v,w)=v\widetilde{w}+w\widetilde{v}$ in the symplectic case,
(3)
$M(v,w)=v\widetilde{w}-w\widetilde{v}$ in the orthogonal case.
By $[G,G]$, we denote the commutator subgroup of $G$ i.e. group generated by elements of the form $ghg^{-1}h^{-1}$, for $g,h\in G$.
Definition 2.10.
A row $(v_{1},\ldots,v_{n})\in A^{n}$ is called unimodular row of length $n$
if there exist $b_{1},\ldots,b_{n}\in A$ such that
$\Pi_{i=1}^{n}v_{i}b_{i}=1$; i.e. If $I$ is the ideal generated by
$\langle v_{1},\ldots,v_{n}\rangle$, then $(v_{1},\ldots,v_{n})$ is unimodular
if and only if $I=A$. In the above case we say $A$ has $n$-cover.
3. Local-Global Principle for classical groups
Before discussing the main theorem we recall following standard facts.
Lemma 3.1 (Splitting Lemma).
The elementary generators of $n\times n$
matrices satisfy the following property:
$$ge_{ij}(x+y)=ge_{ij}(x)\;ge_{ij}(y)$$
for all $x,y\in A$
and for all $i,j=1,\ldots,n$ with $i\neq j$.
Proof.
Standard.
(cf. [3], $3, Lemma $3.2$).
∎
Lemma 3.2.
Let $G$ be a group, and $a_{i},b_{i}\in G$ for $i=1,\ldots,r$.
Let $J_{k}=\underset{j=1}{\overset{k}{\Pi}}a_{j}$.
Then
$$\underset{i=1}{\overset{r}{\huge\Pi}}a_{i}b_{i}=\underset{i=1}{\overset{r}{\Pi%
}}J_{i}b_{i}J_{i}^{-1}\underset{i=1}{\overset{r}{\Pi}}a_{i}.$$
Proof.
cf. [3], $3, Lemma $3.4.$
∎
Following structural lemma pays a key role in the proof of main theorem.
It is well-known for the polynomial rings; cf. [3] ($3, Lemma $3.6$).
Here, we deduce the analogue for the graded rings.
Lemma 3.3.
Let ${\rm G}(n,A,A_{+})$ denote the subgroup of ${\rm G}(n,A)$ which
is equal to ${\rm I}_{n}$ modulo $A_{+}$.
Then the group ${\rm G}(n,A,A_{+})\cap{\rm E}(n,A)$ is generated by the elements of the type
$\epsilon\,ge_{ij}(A_{+})\,\epsilon^{-1}$ for some $\epsilon\in{\rm E}(n,A_{0})$.
Proof.
Let $\alpha\in{\rm E}(n,A)\cap{\rm G}(n,A,A_{+})$. Then we can write
$$\alpha=\underset{k=1}{\overset{r}{\Pi}}ge_{i_{k}j_{k}}(a_{k})$$
for some elements $a_{k}\in A$, $k=1,\ldots,r$.
As $a_{k}=({a_{0}})_{k}+({a_{+}})_{k}$ for some $(a_{0})_{k}\in A_{0}$ and $(a_{+})_{k}\in A_{+}$. Using the splitting lemma (Lemma 3.1) we can rewrite the expression as:
$$\alpha=\underset{k=1}{\overset{r}{\Pi}}\big{(}ge_{i_{k}j_{k}}{(a_{0})}_{k}\big%
{)}\,\big{(}ge_{i_{k}j_{k}}{(a_{+})}_{k}\big{)}.$$
We write $\epsilon_{t}=\Pi_{k=1}^{t}ge_{i_{k}j_{k}}\big{(}{(a_{0})}_{k}\big{)}$ for $t\in\{1,2,\ldots,r\}$. Observe that $\epsilon_{r}={\rm I}_{n}$, as $\alpha\in{\rm G}(n,A,A_{+})$.
Then
$$\alpha=\Big{(}\underset{k=1}{\overset{r}{\Pi}}\epsilon_{k}ge_{i_{k}j_{k}}\big{%
(}{(a_{+})}_{k}\big{)}\epsilon_{k}^{-1}\Big{)}\Big{(}\underset{k=1}{\overset{r%
}{\Pi}}ge_{i_{k}j_{k}}\big{(}{(a_{0})}_{k}\big{)}\Big{)}=\mathcal{A}\mathcal{B%
}\text{ (say) },$$
where $\mathcal{A}=\Big{(}\underset{k=1}{\overset{r}{\Pi}}\epsilon_{k}ge_{i_{k},j_{k}%
}\big{(}{(a_{+})}_{k}\big{)}\epsilon_{k}^{-1}\Big{)}$
and $\mathcal{B}=\Big{(}\underset{k=1}{\overset{r}{\Pi}}ge_{i_{k}j_{k}}\big{(}{(a_{%
0})}_{k}\big{)}\Big{)}.$
Now go modulo $A_{+}$. Let bar (–) denote the quotient ring modulo $A_{+}$. Then,
$$\overline{\alpha}=\overline{{\rm I}}_{n}=\overline{\mathcal{A}}\overline{%
\mathcal{B}}=\overline{{\rm I}}_{n}\overline{\mathcal{B}}=\bar{{\rm I}}_{n}%
\implies\overline{\mathcal{B}}=\overline{{\rm I}}_{n},$$
as $\alpha\in{\rm G}(n,A,A_{+})$. Since entries of $\mathcal{B}$
are in $A_{0}$, it follows that $\mathcal{B}={\rm I}_{n}$.
$$\alpha=\underset{k=1}{\overset{r}{\Pi}}\epsilon_{k}\,\big{(}ge_{i_{k}j_{k}}%
\big{(}{(a_{+})}_{k}\big{)}\big{)}\,\epsilon_{k}^{-1};$$
as desired.
∎
Now we prove a variant of “Dilation Lemma” mentioned in the statement of $(3)$ of Theorem 3.8.
Lemma 3.4.
Assume the “Dilation Lemma” $($Theorem 3.8 - $($4$)$$)$
to be true. Let $\alpha_{s}\in{\rm E}(n,A_{s}),\;$ with $\alpha^{+}(0)={{\rm I}_{n}}$. Then one gets
$$\alpha^{+}(b+d)\alpha^{+}(d)^{-1}\in{\rm E}(n,A)$$
for some $s,d\in A_{0}$ and $b=s^{l},l\gg 0$.
Proof.
We have $\alpha_{s}\in{\rm E}(n,A_{s})$.
Hence $\alpha_{s}^{+}(X)\in{\rm E}(n,A_{s}[X])$. Let $\beta^{+}(X)=\alpha^{+}(X+d){\alpha^{+}(d)}^{-1}$, where $d\in A_{0}$.
Then $\beta^{+}_{s}(X)\in{\rm E}(n,A_{s}[X])$ and $\beta^{+}(0)={{\rm I}_{n}}$.
Hence by Theorem 3.8 - (4) their exists
$\widetilde{\beta}(X)\in{\rm E}(n,A[X])$ such that $\widetilde{\beta}_{s}(X)=\beta_{s}^{+}(bX)$. Putting $X=1$, we get the required result.
∎
Following is a very crucial result we need for our method of proof. There are many places we use this fact in a very subtle way.
In particular, we mainly use this lemma for the step $(4)\Rightarrow(3)$ of 3.8.
Lemma 3.5.
(cf.[10], Lemma 5.1) Let $R$ be a Noetherian ring and $s\in R$.
Then there exists a natural number $k$ such that the homomorphism
${\rm G}(n,s^{k}R)\rightarrow{\rm G}(n,R_{s})$ $($induced by localization
homomorphism $R\rightarrow R_{s})$ is injective. Moreover, it follows that
the induced map ${\rm E}(n,R,s^{k}R)\rightarrow{\rm E}(n,R_{s})$ is
injective.
The next two lemma’s will be used in intermediaries to prove the equivalent conditions mentioned in Theorem 3.8. We state it without proof.
Lemma 3.6.
Let $v=(v_{1},v_{2},\ldots,v_{n})$ be a unimodular row over over a
commutative semilocal ring $R$. Then the row $(v_{1},v_{2},\ldots,v_{n})$ is completable; i.e. $(v_{1},v_{2},\ldots,v_{n})$
elementarily equivalent to the row $(1,0,\ldots,0)$;
i.e. their exists $\epsilon\in{\rm E}_{n}(R)$ such that $(v_{1},v_{2},\ldots,v_{n})\epsilon=(1,0,\ldots,0)$.
Proof.
cf. [4], Lemma $1.2.21$.
∎
Lemma 3.7.
Let $A$ be a ring and $v\in{\rm E}(n,A)e_{1}$. Let $w\in A^{n}$ be a
column vector such that $\langle v,w\rangle=0$. Then ${\rm I}_{n}+M(v,w)\in E(n,A)$.
Proof.
cf. [4], Lemma $2.2.4$.
∎
Theorem 3.8.
The followings are equivalent for any graded ring $A=\oplus_{i=0}^{\infty}A_{i}$ for
$n\geq 3$ in the linear case and $n\geq 6$ otherwise.
(1)
(Normality):
${\rm E}(n,A)$ is a normal subgroup of ${\rm G}(n,A)$.
(2)
If $v\in{\rm Um}_{n}(A)$ and $\langle v,w\rangle=0$, then ${\rm I}_{n}+M(v,w)\in{\rm E}(n,A)$.
(3)
(Local-Global Principle):
Let $\alpha\in{\rm G}(n,A)$ with $\alpha^{+}(0)={\rm I}_{n}$. If
for every maximal ideal $\mathfrak{m}$ of $A_{0}$
$\alpha_{\mathfrak{m}}\in{\rm E}(n,A_{\mathfrak{m}})$,
then $\alpha\in{\rm E}(n,A)$.
(4)
(Dilation Lemma):
Let $\alpha\in{\rm G}(n,A)$ with $\alpha^{+}(0)={\rm I}_{n}$ and $\alpha_{s}\in{\rm E}(n,A_{s})$ for some non-zero divisor $s\in A_{0}$.
Then there exists $\beta\in{\rm E}(n,A)$ such that
$$\beta_{s}^{+}(b)=\alpha_{s}^{+}(b)$$
for some $b=s^{l}$; $l\gg 0$. i.e.
$\alpha^{+}_{s}(s^{l})$ will be defined over $A$ for $l\gg 0$.
(5)
If $\alpha\in{\rm E}(n,A)$, then $\alpha^{+}(a)\in{\rm E}(n,A)$, for every $a\in A_{0}$.
(6)
If $v\in{\rm E}(n,A)e_{1},$ and $\langle v,w\rangle=0$, then ${\rm I}_{n}+M(v,w)\in{\rm E}(n,A)$.
Proof.
(6) $\Rightarrow$ (5):
Let $\alpha=\underset{k=1}{\overset{t}{\Pi}}\big{(}{\rm I}_{n}+a\;M(e_{i_{k}},e_{j_%
{k}})\big{)}\text{ where }a\in A$, and $t\geq 1$, a positive integer.
Then
$$\alpha^{+}(b)=\underset{k=1}{\overset{t}{\Pi}}\big{(}{\rm I}_{n}+a^{+}(b)\;M(e%
_{i_{k}},e_{j_{k}})\big{)},$$
where
$b\in A_{0}$. Take $v=e_{i}$ and $w=a^{+}(b)e_{j}$, Then $\alpha={\rm I}_{n}+M(v,w)$ and $\langle v,w\rangle=0$. Indeed,
Linear case:
$$\displaystyle{\rm M}(v,w)=$$
$$\displaystyle vw^{t}=e_{i}(a^{+}(b)e_{j})^{t}=a^{+}(b)e_{i}e_{j}^{t}=a^{+}(b)%
\;{\rm M}(e_{i},e_{j}),$$
$$\displaystyle\langle v,w\rangle=$$
$$\displaystyle v^{t}w=e_{i}^{t}(a^{+}(b)e_{j})^{t}=a^{+}(b)\;e_{i}^{t}e_{j}=a^{%
+}(b)\langle e_{i},e_{j}\rangle=0.$$
Symplectic case:
$$\displaystyle{\rm M}(v,w)=$$
$$\displaystyle vw^{t}\psi_{m}+wv^{t}\psi_{m}=e_{i}(a^{+}(b)e_{j})^{t}\psi_{m}+a%
^{+}(b)e_{j}e_{i}^{t}\psi_{m}$$
$$\displaystyle=$$
$$\displaystyle a^{+}(b)e_{i}e_{j}^{t}\psi_{m}+a^{+}(b)e_{j}e_{i}^{t}\psi_{m}=a^%
{+}(b)\;{\rm M}(e_{i},e_{j}),$$
$$\displaystyle\langle e_{i},e_{j}\rangle=e_{i}^{t}\psi_{m}e_{j}=(e_{i}^{t}\psi_%
{m}e_{i})(e_{i}^{t}e_{j})=\psi_{m}(e_{i}^{t}e_{j})=0\text{ hence, }$$
$$\displaystyle\langle v,w\rangle=\;v^{t}\psi_{m}w=e_{i}^{t}\psi_{m}a^{+}(b)e_{j%
}=a^{+}(b)e_{i}^{t}e_{j}=a^{+}(b)\langle e_{i},e_{j}\rangle=0.$$
Orthogonal case:
$$\displaystyle{\rm M}(v,w)=$$
$$\displaystyle v.w^{t}\widetilde{\psi}_{m}-w.v^{t}\widetilde{\psi}_{m}=e_{i}.(a%
^{+}(b)e_{j})^{t}\widetilde{\psi}_{m}-a^{+}(b)e_{j}.e_{i}^{t}\widetilde{\psi}_%
{m}$$
$$\displaystyle=$$
$$\displaystyle a^{+}(b)e_{i}.e_{j}^{t}\widetilde{\psi}_{m}-a^{+}(b)e_{j}.e_{i}^%
{t}\widetilde{\psi}_{m}=a^{+}(b)\;{\rm M}(e_{i},e_{j}),$$
$$\displaystyle\langle e_{i},e_{j}\rangle=e_{i}^{t}\widetilde{\psi}_{m}e_{j}=(e_%
{i}^{t}\widetilde{\psi}_{m}e_{i})(e_{i}^{t}e_{j})=\widetilde{\psi}_{m}(e_{i}^{%
t}e_{j})=0\text{ hence, }$$
$$\displaystyle\langle v,w\rangle=\;v^{t}\widetilde{\psi}_{m}w=e_{i}^{t}%
\widetilde{\psi}_{m}a^{+}(b)e_{j}=a^{+}(b)e_{i}^{t}e_{j}=a^{+}(b)\langle e_{i}%
,e_{j}\rangle=0.$$
Hence applying (6) over ring $A$, we have $\alpha^{+}(b)\in{\rm E}(n,A)$. Therefore $\alpha^{+}(b)\in{\rm E}(n,A)$ for $b\in A_{0}$.
(5) $\Rightarrow$ (4):
Since $\alpha_{s}\in{\rm E}(n,A_{s})$ with $({\alpha_{0}}_{s})={\rm I}_{n}$, the diagonal entries of $\alpha$
are of the form $1+g_{ii}$, where $g_{ii}\in({A_{+}})_{s}$ and off-diagonal entries are of the
form $g_{ij}$, where $i\neq j$ and $g_{ij}\in({A_{+})}_{s}$.
We choose $l$ to be large enough such that $s^{l}$ is greater than the common denominator of all
$g_{ii}$ and $g_{ij}$. Then using (5), we get
$$\alpha_{s}^{+}(s^{l})\in{\rm E}(n,A_{s}).$$
Since that $\alpha^{+}(s^{l})$ permits a natural pullback (as denominators are cleared), we have $\alpha^{+}(s^{l})\in{\rm E}(n,A)$.
(4) $\Rightarrow$ (3):
Since $\alpha_{\mathfrak{m}}\in{\rm E}(n,A_{\mathfrak{m}})$, we have an element $s\in A_{0}-\mathfrak{m}$ such that
$\alpha_{s}\in{\rm E}(n,A_{s})$. Let $s_{1},s_{2},\ldots,s_{r}\in A_{0}$ be non-zero divisors with $s_{i}\in A_{0}-\mathfrak{m}_{i}$
such that $\langle s_{1},s_{2},\ldots,s_{r}\rangle=A$. From (4) we have $\alpha^{+}(b_{i})\in{\rm E}(n,A),$
for some $b_{i}={{s_{i}}^{l_{i}}}$ with $b_{1}+\cdots+b_{r}=1$. Now consider $\alpha_{s_{1}s_{2}\ldots s_{r}}$,
which is the image of $\alpha$ in $A_{s_{1}\cdots s_{r}}$. Due to Lemma 3.5, $\alpha\mapsto\alpha_{s_{1}s_{2}\cdots s_{r}}$
is injective and hence we can perform our calculation in $A_{s_{1}\cdots s_{r}}$ and then pull it back to $A$.
$$\displaystyle\alpha_{s_{1}s_{2}\ldots s_{r}}=$$
$$\displaystyle\alpha_{s_{1}s_{2}\ldots s_{r}}^{+}(b_{1}+b_{2}\cdots+b_{r})$$
$$\displaystyle=$$
$$\displaystyle{{\big{(}(\alpha_{s_{1}})}_{s_{2}s_{3}\ldots}}\big{)}^{+}(b_{1}+%
\cdots+b_{r}){{\big{(}(\alpha_{s_{1}})}_{s_{2}s_{3}\ldots}}\big{)}^{+}(b_{2}+%
\cdots+b_{r})^{-1}\cdots$$
$$\displaystyle{\big{(}(\alpha_{s_{i}})}_{s_{1}\ldots\hat{s_{i}}\ldots s_{r}}%
\big{)}^{+}(b_{i}+\cdots+b_{r}){\big{(}(\alpha_{s_{i}})}_{s_{1}\ldots\hat{s_{i%
}}\ldots s_{r}}\big{)}^{+}(b_{i+1}+\cdots+b_{r})^{-1}$$
$$\displaystyle\cdots{{\big{(}(\alpha_{s_{r}})}_{s_{1}s_{2}\ldots s_{r-1}}}\big{%
)}^{+}(b_{r}){{\big{(}(\alpha_{s_{r}})}_{s_{1}s_{2}\ldots s_{r-1}}}\big{)}^{+}%
(0)^{-1}$$
Observing that each
$${\big{(}(\alpha_{s_{i}})}_{s_{1}s_{2}\ldots\hat{s_{i}}\ldots s_{r}}\big{)}^{+}%
(b_{i}+\cdots+b_{r}){\big{(}(\alpha_{s_{i}})}_{s_{1}\ldots\hat{s_{i}}\ldots s_%
{r}}\big{)}^{+}(b_{i+1}+\cdots+b_{r})^{-1}\in{\rm E}(n,A)$$
due to Lemma 3.4 (here $\hat{s}_{i}$ means we omit $s_{i}$ in the product $s_{1}\cdots\hat{s_{i}}\cdots s_{r}$),
we have $\alpha_{s_{1}\cdots s_{r}}\in{\rm E}(n,A_{s_{1}\cdots s_{r}})$ and hence $\alpha\in{\rm E}(n,A)$.
Remark 3.9.
Following is a commutative diagram (here we are assuming $\langle s_{i},s_{j}\rangle=A$):
\UseComputerModernTips
$$\xymatrix{A\ar[r]\ar[d]&A_{s_{i}}\ar[d]\\
A_{s_{j}}\ar[r]&A_{s_{i}s_{j}}}$$
Let $\theta_{i}=\alpha_{s_{i}}^{+}(b_{i}+\cdots+b_{r})\alpha_{s_{i}}^{+}(b_{i-1}+%
\cdots+b_{r})\in A_{s_{i}}$ and $\theta_{ij}$
be the image of $\theta_{i}$ in $A_{s_{i}s_{j}}$ and similarly $\pi_{j}=\alpha_{s_{j}}^{+}(b_{j}+\cdots+b_{r})\alpha_{s_{j}}^{+}(b_{j-1}+%
\cdots+b_{r})\in A_{s_{j}}$
and $\pi_{s_{i},s_{j}}$ be its image in $A_{s_{i}s_{j}}$. Then due to Lemma 3.5 the product $\big{(}\theta_{s_{i}s_{j}}\big{)}\big{(}\pi_{s_{i}s_{j}}\big{)}$
can be identified with the product $\theta_{i}\pi_{j}$.
(3) $\Rightarrow$ (2):
Since polynomial rings are special case of graded rings, the result follows by using $3\Rightarrow 2$ in [3], $\lx@sectionsign 3$.
(2) $\Rightarrow$ (1) $\Rightarrow$ (6):
The proof goes as in [3]. ($$3$, $2\Rightarrow 1\Rightarrow 7\Rightarrow 6$).
∎
4. Local-Global Principle for commutator subgroup
In this section we deduce the analogue of Local-Global Principle for the commutator subgroup of the linear and the symplectic group over graded rings.
Unless mentioned otherwise, we assume $n\geq 3$ for the linear case and $n\geq 6$ for the symplectic case.
Let us begin with the following well-known fact for semilocal rings.
Lemma 4.1.
Let $A$ be a semilocal commutative ring with identity. Then for $n\geq 2$ in the linear case and
$n\geq 4$ in the symplectic case, one gets
$${\rm S}(n,A)={\rm E}(n,A).$$
Proof.
cf. Lemma $1.2.25$ in [4].
∎
Remark 4.2.
If $\alpha=(\alpha_{ij}/1)\in{\rm G}(n,A_{s})$, then $\alpha$ has a natural pullback $\beta=(\alpha_{ij})\in{\rm G}(n,A)$ such that $\beta_{s}=\alpha$.
If $\alpha_{s}\in{\rm S}(n,A_{s})$ such that it admits a natural pullback $\beta\in{\rm G}(n,A)$,
then $\beta\in{\rm S}(n,A)$.
The next lemma deduces an analogue result of “Dilation Lemma” (Theorem 3.8-($3$)) for ${\rm S}(n,A)$; the special linear (resp. symplectic) group.
Lemma 4.3.
Let $\alpha=(\alpha)_{ij}\in{\rm S}(n,A)$. Then $\alpha^{+}(a)\in{\rm S}(n,A),$
where $a\in A_{0}$. Hence if $\alpha\in G(n,A)$ with $\alpha^{+}(0)={{\rm I}_{n}}$ and $\alpha_{s}\in{\rm S}(n,A_{s})\text{ then }\alpha_{s}^{+}(b)\in{\rm S}(n,A)$ $($after identifying $\alpha_{s}^{+}(b)$ with its pullback by using Lemma 3.5$)$,
where $s\in A_{0}$ is a non-zero divisor and $b=s^{l}$ with $l\gg 0$, for $n\geq 1$ in linear case and $n\geq 2$ in symplectic case.
Proof.
Since $\text{ det }:A\rightarrow A^{*}\text{ (units of }A)\subset A_{0},$ and $A_{i}A_{j}\subset A_{i+j}$,
the non-zero component of $\alpha$ doesn’t contribute for the value of the determinant and hence
$$\text{det }\alpha=\text{det }\alpha^{+}(0).$$
Therefore, if $\beta=\alpha^{+}(a)$, then
$$\text{det }\beta=\text{det }\beta^{+}(0)=\text{det }\big{(}\alpha^{+}(a)\big{)%
}^{+}(0)=\text{det }\alpha^{+}(a\cdot 0)=\text{det }\alpha^{+}(0)=1.$$
Hence $\alpha^{+}(a)\in{\rm S}(n,A)$.
Since $\alpha_{s}\in{\rm S}(n,A_{s})$ with $\alpha^{+}(0)={{\rm I}_{n}}$, the diagonal entries are of the form $1+g_{ii}$, and
off diagonal entries are $g_{ij}$, where $g_{ij}\in{(A_{+}})_{s}$ for all $i,j$.
We choose $b\in(s)$ such that $b=s^{l}$ with $l\gg 0$, so that $b$ can dilute
the denominator of each entries. Then $\alpha_{s}^{+}(b)\in{\rm S}(n,A)$.
∎
The next lemma gives some structural information about commutators.
Lemma 4.4.
Let $\alpha,\beta\in{\rm S}(n,A)$ and $A_{0}$ be a commutative semilocal ring.
Then the commutator subgroup
$$[\alpha,\beta]\in[\alpha\alpha_{0}^{-1},\beta\beta_{0}^{-1}]\;{\rm E}(n,A).$$
Proof.
Since $A_{0}$ is semilocal, we have ${\rm S}(n,A_{0})={\rm E}(n,A_{0})$ by Lemma 4.1. Hence $\alpha^{+}(0),\beta^{+}(0)\in{\rm E}(n,A_{0})$. Let $a=\alpha{\alpha^{+}(0)}^{-1}$ and $b=\beta{\beta^{+}(0)}^{-1}$. Then
$$\displaystyle[\alpha,\beta]$$
$$\displaystyle=[\alpha{\alpha^{+}(0)}^{-1}\alpha^{+}(0),\beta{\beta^{+}(0)}^{-1%
}\beta^{+}(0)]$$
$$\displaystyle=a\alpha^{+}(0)b\beta^{+}(0){\alpha^{+}(0)}^{-1}a^{-1}{\beta^{+}(%
0)}^{-1}b^{-1}$$
$$\displaystyle=\big{(}aba\!^{-1}b\!^{-1}\big{)}\big{(}bab\!^{-1}\alpha^{+}(0)ba%
\!^{-1}b\!^{-1}\big{)}\big{(}ba\beta^{+}(0){\alpha^{+}(0)}\!^{-1}a\!^{-1}b^{-1%
}\big{)}\big{(}b{\beta^{+}(0)}^{-1}b\!^{-1}\big{)}.$$
Since ${\rm E}(n,A)$ is a normal subgroup of ${\rm S}(n,A)$, the elements $bab^{-1}\alpha^{+}(0)ba^{-1}b^{-1}$,
$ba\beta^{+}(0){\alpha^{+}(0)}^{-1}a^{-1}b^{-1}$ and $b{\beta^{+}(0)}^{-1}b^{-1}$ are in ${\rm E}(n,A).$
∎
Corollary 4.5.
Let $\alpha\in[{\rm S}(n,A),{\rm S}(n,A)]$ with $\alpha^{+}(0)={\rm I}_{n},$ and
let $A_{0}$ be a semilocal commutative ring. Then using the normality property of the elementary $($resp. elementary symplectic$)$ subgroup
$\alpha$ can be written as
$$\underset{k=1}{\overset{t}{\Pi}}\;[\beta_{k},\gamma_{k}]\;\epsilon,$$
for some $t\geq 1$, and $\beta_{k},\gamma_{k}\in{\rm S}(n,A)$, with $\beta_{k}^{+}(0)=\gamma_{k}^{+}(0)={\rm I}_{n}$, and $\epsilon\in{\rm E}(n,A)$ with $\epsilon^{+}(0)={\rm I}_{n}$.
Proof.
Since $\alpha\in[{\rm S}(n,A),{\rm S}(n,A)]$, $\alpha=\underset{k=1}{\overset{t}{\Pi}}[a_{k},b_{k}]$ for some $t\geq 1$.
Using Lemma 4.4 we identify $\beta_{k}$ with $a_{k}{{a_{k}}^{+}(0)}^{-1}$ and
$\gamma_{k}$ with $b_{k}{{b_{k}}^{+}(0)}^{-1}$. This gives
$$\alpha=\underset{k=1}{\overset{t}{\Pi}}\;[\beta_{k},\gamma_{k}]\;\epsilon.$$
Then it follows that $\epsilon^{+}(0)={\rm I}_{n}$, as $\alpha^{+}(0)={\rm I}_{n}$.
∎
The next lemma gives a variant of “Dilation Lemma” (Lemma 4.3)
for ${\rm S}(n,A)$; the special linear (resp. symplectic) group.
Lemma 4.6.
If $\alpha\in{\rm S}(n,A_{s})$ with $\alpha^{+}(0)={{\rm I}_{n}}$, then
$$\alpha^{+}(b+d){\alpha^{+}(d)}^{-1}\in{\rm S}(n,A),$$
where $s$ is a non-zero divisor and $b,d\in A_{0}$ with $b=s^{i}$ for some $i\gg 0$.
Proof.
Let $\alpha^{+}(X)\in{\rm G}(n,A[X])$, then $\alpha_{s}^{+}(X)\in{\rm E}(n,A_{s}[X])$.
Let $\beta^{+}(X)=\alpha^{+}(X+d){\alpha^{+}(d)}^{-1}$. Then $\beta_{s}^{+}(X)\in{\rm S}(n,A_{s}[X])$.
Hence by Lemma 4.3, $\beta^{+}(bX)\in{\rm S}(n,A)$. Putting $X=1$,
we get result.
∎
The following lemma makes use of Lemma 4.3 to deduce “Dilation Lemma” for the commutator subgroup $[{\rm S}(n,A),{\rm S}(n,A)]$.
Lemma 4.7.
Let $\alpha_{s}=\underset{i=1}{\overset{t}{\Pi}}[{\beta_{i}}_{s},{\gamma_{i}}_{s}]%
\epsilon_{s}$ for some $t\geq 1$,
such that ${\beta_{i}}_{s}$ and ${\gamma_{i}}_{s}\in{\rm S}(n,A_{s})$ and $\epsilon_{s}\in{\rm E}(n,A_{s})$
with $\gamma_{i_{s}}^{+}(0)=\beta_{i_{s}}^{+}(0)={\epsilon_{s}}^{+}(0)={\rm I}_{n}$.
Then $($by identifying with its pullback by using Lemma 3.5$)$
$$\alpha_{s}^{+}(b+d)\alpha_{s}^{+}(d)^{-1}\in[{\rm S}(n,A),{\rm S}(n,A)],$$
where $b,d\in A_{0}$ with $b=s^{l}$ for some $l\gg 0$.
Proof.
Without loss of generality (as we would conclude by the end of the proof),
we can assume
$$\alpha_{s}=[\beta_{s},\gamma_{s}]\epsilon_{s}=\beta_{s}\gamma_{s}{\beta_{s}}^{%
-1}{\gamma_{s}}^{-1}\epsilon_{s}.$$
Since $\beta_{s}^{+}(0)={\rm I}_{n}$, by Lemma 4.3 ${\beta_{s}}^{+}(b),{\gamma_{s}}^{+}(b)\in{\rm S}(n,A)$ for some $b=s^{l}$ with $l\gg 0$. Also,
${\epsilon_{s}}^{+}(b)\in{\rm E}(n,A)$ by (5) of Theorem 3.8.
Hence
$$\displaystyle\alpha_{s}^{+}(b+d){\alpha_{s}^{+}(d)}^{-1}=$$
$$\displaystyle\;\beta_{s}^{+}(b+d)\gamma_{s}^{+}(b+d){\beta_{s}^{+}(b+d)}^{-1}{%
\gamma_{s}^{+}(b+d)}^{-1}$$
$$\displaystyle\epsilon_{s}^{+}(b+d){\epsilon_{s}^{+}(d)}^{-1}\gamma_{s}^{+}(d)%
\beta_{s}^{+}(d){\gamma_{s}^{+}(d)}^{-1}{\beta_{s}^{+}(d)}^{-1}.$$
Since ${\rm E}(n,A_{s})$ and ${\rm S}(n,A_{s})$ are normal subgroups in ${\rm G}(n,A_{s})$, by rearranging (using $\epsilon_{s}^{+}(b+d){\epsilon_{s}^{+}(d)}^{-1}$
as intermediary) we can consider
$\gamma_{s}^{+}(b+d){\gamma_{s}^{+}(d)}^{-1}$ and $\beta_{s}^{+}(b+d){\beta_{s}^{+}(d)}^{-1}$ together. Now by using Lemma 4.6
we get $\gamma_{s}^{+}(b+d){\gamma_{s}^{+}(d)}^{-1}$ and $\beta_{s}^{+}(b+d){\beta_{s}^{+}(d)}^{-1}\in{\rm S}(n,A)$.
Hence $($by identifying with its pullback by using Lemma 3.5$)$ it follows that
$$\alpha_{s}^{+}(b+d){\alpha_{s}^{+}(d)}^{-1}\in[{\rm S}(n,A),{\rm S}(n,A)].$$
∎
Now we deduce the graded version of the Local-Global Principle for the commutators subgroups.
Theorem 4.8.
Let $\alpha\in{\rm S}(n,A)$ with $\alpha^{+}(0)={\rm I}_{n}$.
If
$$\alpha_{\mathfrak{p}}\in[{\rm S}(n,A_{\mathfrak{p}}),{\rm S}(n,A_{\mathfrak{p}%
})]\text{ for all }\mathfrak{p}\in{\rm Spec(A_{0})},$$
then
$$\alpha\in[{\rm S}(n,A),{\rm S}(n,A)].$$
Proof.
Since $\alpha_{\mathfrak{p}}\in[{\rm S}(n,A_{\mathfrak{p}}),{\rm S}(n,A_{\mathfrak{p}%
})]\subset[{\rm S}(n,A_{\mathfrak{m}}),{\rm S}(n,A_{\mathfrak{m}})]$,
we have for a maximal ideal $\mathfrak{m}\supseteq\mathfrak{p}$, $s\in A_{0}-\mathfrak{m}$
such that
$$\alpha_{s}\in[{\rm S}(n,A_{s}),{\rm S}(n,A_{s})],$$
hence $\alpha_{s}$ can be decomposed as
$$\alpha_{s}=\underset{i=1}{\overset{t}{\Pi}}[{\beta_{i_{s}}},{\gamma_{i_{s}}}]%
\epsilon_{s}$$
by Corollary
4.5, where $\beta_{i_{s}},\gamma_{i_{s}}\in{\rm S}(n,A_{s})$ and $\epsilon_{s}\in{\rm E}(n,A_{s})$.
Now using the “Dilation Lemma” (Lemma 4.3 for ${\rm S}(n,A_{s})$ and theorem 3.8 $(4)$ for
${\rm E}(n,A_{s})$) on each of these elements, we have for $b=s^{i},i\gg 0$
$$\alpha_{s}^{+}(b)=\underset{i=1}{\overset{t}{\Pi}}[\beta_{i_{s}}^{+}(b),\gamma%
_{i_{s}}^{+}(b)]\epsilon_{s}^{+}(b)\in[{\rm S}(n,A),{\rm S}(n,A)]E(n,A).$$
Since ${\rm E}(n,A_{s})\subseteq[{\rm S}(n,A_{s}),{\rm S}(n,A_{s})]$, we have $\alpha_{s}\in[{\rm S}(n,A_{s}),{\rm S}(n,A_{s})]$.
Let $s_{1},\ldots,s_{r}\in A_{0}$ non-zero divisors such that $s_{i}\in A_{0}-\mathfrak{m}_{i}$ and $\langle s_{1},\ldots,s_{r}\rangle=A_{0}$. Then
it follows that $b_{1}+\cdots+b_{r}=1$ for suitable (as before) $b_{i}\in(s_{i})$; and $i=1,\ldots r$. Now
$$\displaystyle\alpha_{s_{1}s_{2}\ldots s_{r}}=$$
$$\displaystyle\alpha_{s_{1}s_{2}\ldots s_{r}}^{+}(1)$$
$$\displaystyle=$$
$$\displaystyle\alpha_{s_{1}s_{2}\ldots s_{r}}^{+}(b_{1}+b_{2}+\cdots+b_{r})$$
$$\displaystyle=$$
$$\displaystyle{{\big{(}(\alpha_{s_{1}})}_{s_{2}s_{3}s_{4}\ldots}}\big{)}^{+}(b_%
{1}+\cdots+b_{r}){{\big{(}(\alpha_{s_{1}})}_{s_{2}s_{3}s_{4}\ldots}}\big{)}^{+%
}(b_{2}+\cdots+b_{r})^{-1}\!\!\cdots$$
$$\displaystyle{\big{(}(\alpha_{s_{i}})}_{s_{1}\ldots\hat{s_{i}}\ldots s_{r}}%
\big{)}^{+}\!\!(b_{i}+\cdots+b_{r}){\big{(}(\alpha_{s_{i}})}_{s_{1}\ldots\hat{%
s_{i}}\ldots s_{r}}\big{)}^{+}\!\!(b_{i+1}+\cdots+b_{r})^{-1}$$
$$\displaystyle\cdots{{\big{(}(\alpha_{s_{r}})}_{s_{1}s_{2}\ldots s_{r-1}}}\big{%
)}^{+}\!\!(b_{r}){{\big{(}(\alpha_{s_{r}})}_{s_{1}s_{2}\ldots s_{r-1}}}\big{)}%
^{+}\!\!(0)^{-1}.$$
Using Lemma 3.5 the product is well defined (see Remark 3.9) and using Lemma 4.7, we conclude that $\alpha\in[{\rm S}(n,A),{\rm S}(n,A)].$
∎
5. Auxiliary Result for Transvection subgroup
Let $R$ be a commutative ring with $1$. We recall that a finitely generated projective $R$-module $Q$ has a unimodular element
if their exists $q\in Q$ such that $qR\simeq R$ and $Q\simeq qR\oplus P$ for some projective $R$-module $P$.
In this section, we shall consider three types of (finitely generated) classical modules; viz. projective,
symplectic and orthogonal modules over graded rings. Let $A=\bigoplus_{i=0}^{\infty}{A_{i}}$ be a commutative graded Noetherian ring with identity.
For definitions and related facts we refer
[5]; $1, $2. In that paper, the first author with A. Bak and R.A. Rao has established an
analogous Local-Global Principle for the elementary transvection subgroup of the
automorphism group of projective, symplectic and orthogonal modules of global
rank at least 1 and local rank at least 3. In this article we deduce an analogous statement for the above classical groups over graded rings.
We shall assume for every maximal ideal $\mathfrak{m}$ of $A_{0}$, the symplectic and orthogonal module $Q_{\mathfrak{m}}\simeq(A_{0}^{2m})_{\mathfrak{m}}$
with the standard form (for suitable integer $m$).
Remark 5.1.
By definition the global rank or simply rank of a finitely generated projective $R$-module
(resp. symplectic or orthogonal $R$-module) is the largest integer $k$ such that $\oplus^{k}R$
(resp. $\perp^{k}$ $\mathbb{H}(R)$) is a direct summand (resp. orthogonal summand) of the module. Here
$\mathbb{H}(R)$ denotes the hyperbolic plane.
Let $Q$ denote a projective, symplectic or orthogonal $A_{0}$-module of global rank
$\geq 1$, and total (or local)
rank $r+1$ in the linear case and $2r+2$ otherwise and let $Q_{1}=Q\otimes_{A_{0}}A$.
We use the following notations to deal with the above three classical modules uniformly.
(1)
${\rm G}(Q_{1})$ := the full automorphism group of $Q_{1}$.
(2)
${\rm T}(Q_{1})$ := the subgroup generated by transvections of $Q_{1}$.
(3)
${\rm ET}(Q_{1})$ := the subgroup generated by elementary-transvections of $Q_{1}$.
(4)
$\widetilde{\rm E}(r,A)$ := ${\rm E}(r+1,A)$ for linear case and ${\rm E}(2r+2,A)$ otherwise.
In [5], the first author with Ravi A. Rao and A. Bak established the “Dilation Lemma” and the “Local-Global Principle” for
the transvection subgroups of the automorphism
group over polynomial rings. In this section, we generalize their results over graded rings. For the statements over polynomial rings
we request the reader to look at Proposition 3.1 and Theorem 3.6 in [5].
Following is the graded version of the “Dilation Lemma”.
Proposition 5.2 (Dilation Lemma).
Let $Q$ be a projective $A_{0}$-module, and $Q_{1}=Q\otimes_{A_{0}}A$.
Let $s$ be a non-nilpotent element in $A_{0}$. Let $\alpha\in G(Q_{1})$ with $\alpha^{+}(0)={\rm I}_{n}$. Suppose
$$\alpha_{s}\in\begin{cases}{\rm E}\big{(}r+1,(A_{0})_{s}\big{)}&\text{ in the %
linear case,}\\
{\rm E}\big{(}2r+2,(A_{0})_{s}\big{)}&\text{ otherwise. }\end{cases}$$
Then there exists $\widetilde{\alpha}\in{\rm ET}(Q_{1})$ and $l\gg 0$ such that $\widetilde{\alpha}$
localizes at $\alpha^{+}(b)$ for some $b=(s^{l})$ and $\widetilde{\alpha}^{+}(0)={\rm I}_{n}$.
Proof.
. Let $\alpha\in{\rm G}(Q_{1})$ and $\alpha(X):=\alpha^{+}(X)\in{\rm G}(Q_{1}[X])$.
Since $\alpha_{s}\in\widetilde{{\rm E}}(r,A_{s})$, $\alpha_{s}(X)\in\widetilde{{\rm E}}(r,A_{s}[X])$. Since $\alpha(0):=\alpha^{+}(0)={\rm I}_{n}$, we can apply “Dilation Lemma” for the ring $A[X]$ (cf.[5], Proposition 3.1) and hence there exists
$\widetilde{\alpha}(X)\in{\rm ET}(Q_{1}[X])$ such that $\widetilde{\alpha}_{s}(X)=\alpha_{s}^{+}(bX)$ for some
$b\in(s^{l})$, $l\gg 0$. Substituting $X=1$, we get $\widetilde{\alpha}\in{\rm ET}(Q_{1})$ and
$\widetilde{\alpha}_{s}:=\widetilde{\alpha}_{s}(1)=\alpha_{s}(b)=\alpha_{s}^{+}%
(b)$. This proves the proposition.
∎
Lemma 5.3.
If $\alpha\in{\rm G}(Q_{1})$ with $\alpha^{+}(0)={\rm I}_{n}$ such that $\alpha_{s}\in{\rm E}(n,A_{s})$,
then $\alpha^{+}(b+d)\alpha^{+}(d)^{-1}\in{\rm ET}(Q_{1})$ for some $b\in(s^{l})$ with $l\gg 0$, and $d\in A_{0}$.
Proof.
Let $\beta=\alpha_{s}^{+}(b+d){\alpha_{s}^{+}(d)}^{-1}$. Since $\alpha_{s}\in{\rm E}(n,A_{s})$
implies $\alpha^{+}(a)\in{\rm E}(n,A_{s})$ for $a\in A_{s}$ and hence $\beta\in{\rm E}(n,A_{s})$.
Hence by Theorem 5.2, their exists a $\widetilde{\beta}\in{\rm ET}(Q_{1})$ such that $\widetilde{\beta}_{s}^{+}(b)=\beta$. Hence the lemma follows.
∎
Theorem 5.4 (Local-Global Principle).
Let $Q$ be a projective $A_{0}$-module, and $Q_{1}=Q\otimes_{A_{0}}A$.
Suppose $\alpha\in G(Q_{1})$ with $\alpha^{+}(0)={\rm I}_{n}$. If
$$\alpha_{\mathfrak{m}}\in\begin{cases}{\rm E}\big{(}r+1,(A_{0})_{\mathfrak{m}}%
\big{)}&\text{ in the linear case, }\\
{\rm E}\big{(}2r+2,(A_{0})_{\mathfrak{m}}\big{)}&\text{ otherwise. }\end{cases}$$
for all $\mathfrak{m}\in$ ${\rm max}\;(A_{0})$, then $\alpha\in{\rm ET\big{(}Q_{1}\big{)}}\subseteq{\rm T\big{(}Q_{1}\big{)}}$.
Proof.
Let $\alpha\in{\rm G}(Q_{1})$ with $\alpha^{+}(0)={\rm I}_{n}$. Since $\alpha_{\mathfrak{m}}\in\widetilde{{\rm E}}(r,A_{\mathfrak{m}})$.
Hence their exists a non-nilpotent $s\in A_{0}-\mathfrak{m}$, such that $\alpha_{s}\in\widetilde{{\rm E}}(r,A_{s})$. Hence by the above “Dilation Lemma”
there exists $\widetilde{\alpha}\in{\rm ET}(Q_{1})$ such that $\widetilde{\alpha}_{s}=\alpha_{s}^{+}(b)$, for some $b=s^{l}$ with $l\gg 0$.
For each maximal ideal $\mathfrak{m}_{i}$ we can find a suitable $b_{i}$. Since $A$ is Noetherian, it follows that $b_{1}+\cdots+b_{r}=1$ for some positive integer $r$.
Now we observe that $\alpha^{+}(b_{i}+\cdots+b_{r}){\alpha^{+}(b_{i+1}+\cdots+b_{r})}^{-1}\in{\rm ET%
}(Q_{1})$ and hence
calculating in the similar manner as we did it in Remark 3.9, we get
$$\displaystyle\alpha=\alpha^{+}(1)=$$
$$\displaystyle\alpha^{+}(b_{1}+\cdots+b_{r}){\alpha^{+}(b_{2}+\cdots+b_{r})}^{-%
1}\cdots\alpha^{+}(b_{i}+\cdots+b_{r})$$
$$\displaystyle{\alpha^{+}(b_{i+1}+\cdots+b_{r})}^{-1}\cdots\alpha^{+}(0)^{-1}%
\in{\rm ET}(Q_{1})\subseteq{\rm T}(Q_{1});$$
as desired.
∎
Acknowledgement
The second author expresses his sincere thanks to IISER Pune for allowing him to work on this project and
providing him their infrastructural facilities.
References
[1]
Bak, Anthony, and Nikolai Vavilov; ”Structure of hyperbolic unitary groups I:
elementary subgroups.” In Algebra Colloquium, vol. 7, no. 2, pp. 159-196. Springer-Verlag, 2000
[2]
H. Bass; Algebraic $K$-Theory. W . A. Benjamin,
Inc., New York, Amsterdam (1968).
[3]
Rabeya. Basu, Ravi A. Rao, Reema Khanna; On
Quillen’s Local-Global Principle, Commutative algebra and algebraic geometry,
17–30, Contemp. Math., 390, Amer. Math. Soc., Providence, RI, 2005.
[4]
Rabeya Basu; Results in Classical Algebraic
${K}$-theory, Ph.D. thesis, Tata Institute of Fundamental Research
(2007), pp 70.
[5]
A. Bak, R. Basu, R.A. Rao;
Local-Global Principle for transvection groups. Proceedings of The American Mathematical Society 138 (2010), no. 4,
1191–1204.
[6]
Rabeya Basu, Ravi A. Rao, Reema Khanna; On the equivalence of the relative
Quillen-Suslin Local-Global Principle and normality of the relative elementary subgroup, (to appear)
Proceeding of the conference Leavitt-path algebras and $K$-theory (CUSAT), Indian Statistical Institute series, Springer.
[7]
Chouinard L.G.; L.G., Projective modules over Krull semigroups, Michigan Math.J.29, 143-148, 1982.
[8]
J. Gubeladze; Classical algebraic K-theory
of monoid algebras, K-theory and homological algebra, Lecture Notes in Mathematics 137,
1990, 36–94.
[9]
J. Gubeladze; Anderson’s conjecture and the maximal monoid class over which projective modules are free, Math. USSR-Sb. 63 (1998), 165–180.
[10]
Roozbeh Hazrat, Nikolai Vavilov; ${K_{1}}$ of
Chevalley groups are nilpotent, Journal of Pure and Applied Algebra
179 (2003), no. 1-2, 99–116.
[11]
H. Matsumura; Commutative ring theory. Translated from the Japanese by M. Reid. Second edition.
Cambridge Studies in Advanced Mathematics, 8. Cambridge University Press, Cambridge, 1989.
[12]
D. Quillen; Projective Modules over
Polynomial Rings, Invent. Math. 36 (1976), 167–171.
[13]
A.A. Suslin; On the structure of special
linear group over polynomial rings, Math. USSR. Izv. 11 (1977),
221–238.
[14]
A.A. Suslin, V.I. Kopeiko; Quadratic
modules and Orthogonal groups over polynomial rings, Nauchn. Sem.,
LOMI 71 (1978), 216–250.
[15]
M.S. Tulenbaev; Schur multiplier of a group of elementary matrices of finite order,
Zapiski Nauchnykh Seminarov Leningradskogo Otdeleniya Matematicheskogo Instuta im V.A. Steklova Akad. Nauk
SSSR, Vol. 86 (1979), 162–169.
[16]
L.N. Vaserstein; On the normal subgroups of
$GL_{n},$ over a ring, Algebraic $K$-Theory, Evanston (1980) (Proc. Conf.,
Northwestern Univ., Evanston Ill., (1980)), pp. 456–465, Lecture
Notes in Math., 854, Springer, Berlin-New York, (1981).
Indian Institute of Science Education and Research $($IISER Pune$)$, India,
email: [email protected], [email protected]
Indian Institute of Science Education and Research $($IISER Pune$)$, India,
email: [email protected] |
Ab-initio coupled-cluster effective interactions for the shell model:
Application to neutron-rich oxygen and carbon isotopes
G. R. Jansen
Physics Division, Oak Ridge National
Laboratory, Oak Ridge, TN 37831, USA
Department of Physics and Astronomy, University of Tennessee,
Knoxville, TN 37996, USA
J. Engel
Deptartment of Physics and Astronomy, University
of North Carolina, Chapel Hill, NC, 27516-3255, USA
G. Hagen
Physics Division, Oak Ridge National Laboratory,
Oak Ridge, TN 37831, USA
Department of Physics and Astronomy, University of Tennessee,
Knoxville, TN 37996, USA
P. Navratil
TRIUMF, 4004 Wesbrook Mall, Vancouver,
British Columbia, V6T 2A3 Canada
A. Signoracci
Physics Division, Oak Ridge National
Laboratory, Oak Ridge, TN 37831, USA
Department of Physics and Astronomy, University of Tennessee,
Knoxville, TN 37996, USA
Abstract
We derive and compute effective valence-space shell-model
interactions from ab-initio coupled-cluster theory and apply them to
open-shell and neutron-rich oxygen and carbon isotopes. Our
shell-model interactions are based on nucleon-nucleon and
three-nucleon forces from chiral effective-field theory. We compute
the energies of ground and low-lying states, and find good agreement
with experiment. In particular our calculations are consistent with
the $N=14,16$ shell closures in ${}^{22,24}$O, while for ${}^{20}$C the
corresponding $N=14$ closure is weaker. We find good agreement
between our coupled-cluster effective-interaction results with those
obtained from standard single-reference coupled-cluster calculations
for up to eight valence neutrons.
pacs: 21.30.Fe, 21.60.Cs, 21.60.De, 21.10.-k
Introduction. – The nuclear shell model is the
foundation on which our understanding of nuclei is built.
One of the most important problems in nuclear structure
today is to understand how shell structure changes with
neutron-to-proton ratio throughout the nuclear chart. Shell structure
influences the locations of the neutron and proton drip lines and the
stability of matter. Examples of changes in shell structure are the
appearance of new magic numbers $N=14$ and $N=16$ in the neutron-rich
oxygen isotopes Stanoiu et al. (2004); Kanungo et al. (2009), and the emergence of
an $N=34$ sub-shell closure in ${}^{54}$Ca
Holt et al. (2012); Hagen et al. (2012a); Steppenbeck
et al. (2013a, b).
Phenomenological shell-model Hamiltonians such as the $sd$ Hamiltonian of Brown
and Wildenthal Wildenthal (1984); Brown and Wildenthal (1988) (abbreviated USD) and
the $p$-$sd$ Hamiltonian of Warturburton and Brown Warburton and Brown (1992) (abbreviated
WBP), have successfully described properties of nuclei with proton number $Z$
and neutron number $N$ less than about 20. To understand the origin of shell
structure, however, researchers are now trying to derive the shell model from
realistic nucleon-nucleon (NN) and three-nucleon forces (3NFs), without
further phenomenology Hjorth-Jensen et al. (1995); Otsuka et al. (2010); Holt et al. (2012). Within the
last few years, for example, Otsuka et al. (2010) showed that 3NFs play a
pivotal role in placing the drip line (correctly) in the oxygen isotopes at
${}^{24}$O, and Holt et al. (2012) showed that the inclusion of 3NFs are
necessary to explain the high $2^{+}$ state in ${}^{48}$Ca .
Until recently, all work to compute effective shell-model interactions was
perturbative. Lately, however, nonperturbative calculations have become
possible. Some have been based on the ab-initio no-core shell model
Navrátil et al. (2009); Barrett et al. (2013), via a valence-cluster expansion
Navrátil et al. (1997); Lisetskiy et al. (2008); Shukla et al. (2011), and others on the in-medium
similarity renormalization group Tsukiyama et al. (2012). In this Letter we use a
third approach, the ab-initio coupled-cluster method
Coester (1958); Coester and Kümmel (1960); Čížek (1966, 2007); Kümmel et al. (1978); Hagen et al. (2013), to construct
effective shell-model interactions for use in open-shell and neutron-rich
nuclei. Starting from NN interactions and 3NFs generated by chiral
effective-field-theory, we compute the ground- and excited-state energies of
neutron-rich carbon and oxygen isotopes with up to eight neutrons in the valence
space. Intense theoretical and experimental interest surround the structure of
both these isotope chains, and particularly the neutron-rich carbon isotopes.
Separation energies, spin assignments for low-lying states, the energies of
$2^{+}$ states, and transition rates in these isotopes all depend on the
locations of shell gaps.
Kondo et al. (2009); Strongman et al. (2009); Ozawa et al. (2011); Kobayashi et al. (2012); Petri et al. (2012); Voss et al. (2012); Petri et al. (2011); Forssén et al. (2013).
At present there is no evidence for a shell closure at the $N=14$ nucleus
${}^{20}$C Stanoiu et al. (2008), despite the $N=14$ shell closure at ${}^{22}$O.
Furthermore, Efimov physics may be at play in ${}^{22}$C
Tanaka et al. (2010); Ershov et al. (2012). This Letter takes the first steps towards an
ab-initio shell-model description of the neutron-rich carbon isotopes, and
addresses the role of 3NFs in these isotopes.
Hamiltonian and model space. – Our coupled-cluster calculations start
from the intrinsic $A$-nucleon Hamiltonian,
$$\hat{H}=\sum_{i<j}\left({({\bf p}_{i}-{\bf p}_{j})^{2}\over 2mA}+\hat{V}_{NN}^%
{(i,j)}\right)+\sum_{i<j<k}\hat{V}_{\rm 3N}^{(i,j,k)}.$$
(1)
Here the intrinsic kinetic energy (the first term) depends on the mass number
$A\equiv Z+N$. The potential $\hat{V}_{NN}$ denotes the chiral NN interaction
at next-to-next-to-next-to leading order Entem and Machleidt (2003); Machleidt and Entem (2011) (with
cutoff $\Lambda=500$ MeV), and $\hat{V}_{\rm 3N}$ is the 3NF that enters at
next-to-next-to leading order with a local regulator Navrátil (2007) (with
cutoff $\Lambda_{\rm 3N}=400$ MeV). The low-energy constants of the 3NF are
given by $c_{E}=0.098$ and $c_{D}=-0.2$. These were initially determined from a
fit to the triton half-life and binding energy with a cutoff $\Lambda_{\rm 3N}=500$ MeV Gazit et al. (2009), and then, with $\Lambda_{\rm 3N}=400$ MeV, $c_{E}$
was re-adjusted to reproduce the ${}^{4}$He binding energy while $c_{D}$ was kept
fixed Roth et al. (2012). To achieve faster convergence as model space size is
increased, we use the similarity renormalization group (SRG) to evolve
$\hat{V}_{\rm NN}$ and $\hat{V}_{\rm 3N}$ to the lower momentum scale
$\lambda_{\rm SRG}=2.0\ \mathrm{fm}^{-1}$ Jurgenson et al. (2009). For the
coupled-cluster calculations themselves, we use a Hartree-Fock basis, with each
orbital expanded in another basis containing up to $N_{\rm max}+1=13$ major
shells from a harmonic-oscillator potential with frequency $\hbar\omega=20$MeV. We limit the number of 3NF matrix elements through the additional cut
$N_{1}+N_{2}+N_{3}\leq 14$. The resulting model-space is sufficient to obtain
well converged results for energies of the states reported in this Letter. We
use the normal-ordered two-body approximation for the 3NF
Hagen et al. (2007); Roth et al. (2012), which has been shown to work well in light-
and medium mass nuclei Roth et al. (2012). By using a Hartree-Fock basis we
minimize the oscillator-frequency dependence arising from the neglect of
residual 3NFs.
Formalism. – To derive an effective shell-model Hamiltonian in a valence
space from ab-initio coupled-cluster theory, we use the valence-cluster
expansion first applied in the no-core shell model
Navrátil and Barrett (1996); Lisetskiy et al. (2008). We expand the Hamiltonian in
Eq. (1) in a form suitable for the shell model:
$$H^{A}_{\rm CCEI}=H_{0}^{A,A_{c}}+H_{1}^{A,A_{c}+1}+H_{2}^{A,A_{c}+2}+\ldots,$$
(2)
where CCEI stands for coupled-cluster effective interaction, $A$ is
the mass of the nucleus we wish to treat, and $A_{c}$ is the mass of the
nucleus with a closed core below the valence space. In
Eq. (2), $H_{0}^{A,A_{c}}$ is the Hamiltonian for the core,
$H_{1}^{A,A_{c}+1}$ is the valence one-body Hamiltonian, and
$H_{2}^{A,A_{c}+2}$ is the additional two-body piece. In this work we
limit ourselves to one- and two-body terms in the valence-space
shell-model Hamiltonian. To solve for the ground-state of the core
nucleus $A_{c}$ we use the coupled-cluster method in the
singles-and-doubles approximation (CCSD) with the $\Lambda$-triples
correction treated perturbatively ($\Lambda$-CCSD(T))
Kucharski and Bartlett (1998); Taube and Bartlett (2008). To obtain the ground and excited
states for the $A_{c}+1$ and $A_{c}+2$ nuclei we use the one- and
two-particle-attached equation-of-motion (EOM) coupled-cluster methods
Gour et al. (2006); Jansen et al. (2011); Jansen (2013). For the particle-attached
EOM we truncate at one-particle and two-particle-one-hole excitations,
and for the two-particle-attached EOM we truncate at two-particle and
three-particle-one-hole excitations. In coupled-cluster theory the
basic ingredient is the similarity-transformed Hamiltonian
$\overline{H}=e^{-T}He^{T}$, which is inherently non-Hermitian
Bartlett and Musiał (2007), thus we need to solve for the left- and the right
eigenstates to obtain a complete bi-orthogonal set of
states. From the left and right eigen-states we can write
Eq. (2) in a spectral representation.
The valence space representation of Eq. (2) consists of a core
energy term (which we compute from $H_{0}^{A,A_{c}}$ using
$\Lambda$-CCSD(T)), a one-body term (built from the particle-attached
eigenvalues of $H_{1}^{A,A_{c}+1}$), and a two-body term. The two-body
term is computed using the Okubo-Lee-Suzuki similarity transformation
Okubo (1954); Suzuki and Lee (1980); Suzuki (1982); Kvaal (2008) by projecting the
two-particle attached EOM eigen states onto two-body valence-space
states. The Okubo-Lee-Suzuki
projection of $H_{2}^{A,A_{c}+2}$ onto the model-space is
Navrátil and Barrett (1996); Kvaal (2008),
$$\langle\alpha_{P}|\overline{H_{\rm eff}^{A}}|\alpha_{P^{\prime}}\rangle=\sum_{%
k=1}^{d}\langle\alpha_{P}|R_{k}^{A,A_{c}+2}\rangle e_{k}\overline{\langle%
\alpha_{P^{\prime}}|R_{k}^{A,A_{c}+2}\rangle}.$$
(3)
Here the $|R_{k}^{A,A_{c}+2}\rangle$ are the two-particle attached
EOM eigenstates with eigenvalue $e_{k}$ for $A_{c}+2$ (with mass $A$ of
the target nucleus in the kinetic energy), $|\alpha_{P}\rangle$
are the model-space states, the sum is over the $k$ two-particle
attached eigenstates that have the largest overlap with the model
space. The $\langle\alpha_{P}|R_{k}^{A,A_{c}+2}\rangle$ are the
matrix elements of the unitary operator ${\bf U}$. $\overline{\langle\alpha_{P^{\prime}}|R_{k}^{A,A_{c}+2}\rangle}$denote the matrix elements
of the inverse of ${\bf U}$. To obtain the effective two-body
shell-model interaction, we subtract the one-body part from
Eq. (3) to avoid double counting.
Note, that we could also construct an effective Hamiltonian using the
corresponding left eigenvectors. This introduces an ambiguity in the
construction of $\overline{H_{\rm eff}^{A}}$. We have verified,
however, that in this work the matrix elements of the effective
operator are almost identical for either choice. The effective
Hamiltonian in Eq. (3) is not Hermitian. Current
shell-model codes require Hermitian matrices. To obtain a Hermitian
representation of the effective shell-model Hamiltonian we construct
the metric operator $S^{\dagger}S$ where $S$ is a matrix that
diagonalizes ${H}^{A}_{\rm CCEI}$; the Hermitian shell-model
Hamiltonian is then $\left[S^{\dagger}S\right]{H}^{A}_{\rm CCEI}\left[S^{\dagger}S\right]^{-1/2}$ Scholtz et al. (1992); Navrátil and Barrett (1996).
Results. – Here we report our CCEI results for ground and low-lying
states in oxygen and carbon isotopes. We choose ${}^{16}$O in oxygen and ${}^{14}$C
in carbon as the closed-shell cores. We then project the one- and
two-particle-attached coupled-cluster wave functions onto the one- and
two-particle model space states in the valence space — the
$d_{5/2},s_{1/2},d_{3/2}$ shell — and proceed to use the resulting
shell-model Hamiltonians in heavier isotopes.
We would like to gauge the accuracy of our CCEI approach by comparing
with full $\Lambda$-CCSD(T) calculations, the results of which we
refer to as reference values. Figure 1 shows the
ground-state energies of all oxygen isotopes ${}^{16-24}$O computed with
the CCEI (red squares), experimental ground-state energies (black
circles), and the $\Lambda$-CCSD(T) ground-state energies in
${}^{21,22,23,24}$O. Our $\Lambda$-CCSD(T) calculations use the model
space mentioned earlier, while the calculations that determine our
CCEI use $N_{\rm max}=12$ and $N_{1}+N_{2}+N_{3}=12$. We believe that
our CCEI results are converged to within $\sim 100$ keV. Both our
$\Lambda$-CCSD(T) and CCEI results are in good agreement with
experimental binding energies. Our CCEI and $\Lambda$-CCSD(T)
calculations also agree well with a variety of recent calculations in
the oxygen isotopes that start with the same Hamiltonian
Hergert et al. (2013); Cipollone et al. (2013).
If we look more closely, we see that the reference $\Lambda$-CCSD(T) results in
${}^{21,22}$O are in excellent agreement with our CCEI results. In ${}^{23,24}$O
the CCEI results start to deviate from the $\Lambda$-CCSD(T) reference values.
In ${}^{24}$O the CCEI ground-state is less bound by about $3.5$ MeV than
obtained with $\Lambda$-CCSD(T). The difference indicates that effective
three-body interactions induced by the Okubo-Lee-Suzuki transformation (which
we neglect) start to play a role in the CCEI approach when the number of
valence nucleons gets too large. The problem can be remedied by including these
interactions or by increasing the valence space size.
Next, we compare low-lying CCEI excited-state energies in ${}^{22}$O
with an EOM coupled-cluster calculation that includes singles and
doubles excitations Stanton and Bartlett (1993). EOM-CCSD can accurately
describe low-lying states that are dominated by one-particle-one-hole
excitations Bartlett and Musiał (2007), and we therefore choose those states
for comparison. In ${}^{22}$O we obtain low-lying $2^{+}$ and $3^{+}$
states with $2.5$ MeV and $3.5$ MeV of excitation energy. The CCEI
result for the same states is $2.7$ MeV and $4.0$ MeV, though the CCEI
result for the $3^{+}$ state in ${}^{22}$O is not yet converged; it moves
down by $\sim 150$ keV when we increase the model space size from
$N=10$ to $N=12$ oscillator shells. The $2^{+}$ state changes only by
$\sim 5$ keV indicating that it, by contrast, is well
converged. Standard EOM-CCSD works well for states that are dominated
by one-particle-one-hole excitations. In our CCEI calculations,
correlations between all particles in the valence space are treated
exactly. Therefore, we expect to see some differences in the computed
spectra. For example, in CCEI we are able to compute the second $0^{+}$
state in ${}^{22}$O, which is dominated by two-particle-two-hole
excitations from the ground-state.
We turn now to carbon. The $\Lambda$-CCSD(T) ground-state energies of
${}^{14,15,16}$C are $-104.0$ MeV, $-104.2$ MeV, and $-106.6$ MeV,
respectively. In ${}^{14}$C the result agrees well with the
experimental ground-state energy of $-105.3$ MeV, but for ${}^{15,16}$C
our particle-attached and two-particle-attached EOM results are
$2.3$ MeV and $4.2$ MeV under-bound with respect to experimental data.
The under-binding persists throughout the chain of carbon isotopes in
our CCEI calculations.
Figures 2 and 3 summarize our CCEI results for
the excited states in the neutron-rich oxygen and carbon isotopes. Figure
2 shows the excitation spectra for ${}^{19-24}$O. Our results are
overall in very good agreement with the experimental excited-state energies.
And without any adjustment of parameters we obtain spectra that are
qualitatively similar to that produced by the phenomenological USD
Hamiltonian. Our $2^{+}$ states in the even oxygen isotopes are on
target and consistent with $N=14$ sub-shell closure in ${}^{22}$O and the $N=16$
shell closure in ${}^{24}$O.
Figure 3 shows excitation spectra in ${}^{17-22}$C. Here
the right column contains results produced by the phenomenological
$p$-$sd$ WBP shell-model Hamiltonian Warburton and Brown (1992) with
$2\hbar\omega$ excitations. As in the oxygen isotopes we agree very
well overall with experiment, and our $2^{+}$ states are remarkably
close to the data. Our $2^{+}$ energy in ${}^{20}$C is at $1.72$ MeV while
the corresponding $2^{+}$ state in ${}^{22}$O lies at $2.78$ MeV; thus our
results are consistent with a weaker $N=14$ sub-shell closure in the
carbon isotopes.
In the odd isotopes ${}^{17,19,21}$C we get the $1/2^{+}$ state in the wrong
position. Our calculations, however, rely on an underlying harmonic-oscillator
basis and therefore do not account for the particle continuum. The $1/2^{+}$
state is dominated by $s$-waves and is located close to the particle emission
threshold, where continuum effects are obviously important
Hagen et al. (2012b, a). The $3/2^{+}$ and the $5/2^{+}$ states are dominated
by $d$-waves, which couple somewhat less to the continuum because of the $l=2$
centrifugal barrier Hagen et al. (2012b). Overall, we expect continuum effects to
be quite significant for the ${}^{17,19,21}$C isotopes. Preliminary calculations
within the no-core shell model with continuum Baroni et al. (2013) for ${}^{17}$C,
with the same chiral NN+3N interaction used here, show that the $1/2^{+}$ state
(unbound in our calculation) gains about 2 MeV in energy and becomes bound. At
the same time, the $3/2^{+}$ and $5/2^{+}$ states are lowered in energy by more
than 1 MeV. We anticipate similar or even stronger continuum effects ${}^{19}$C
and ${}^{21}$C; these would most likely make the $1/2^{+}$ states the ground
states, as they are in reality.
Summary. – We have used coupled cluster theory to derive
shell-model Hamiltonians that depend on no parameters other than those
in the initial NN and $3N$ chiral interactions. We have reproduced
ground and excitation energies with good accuracy in carbon and oxygen
isotopes. The results demonstrate both the predictive power of
Hamiltonians from chiral EFT and the accuracy of the coupled cluster
framework. Finally, our shell model calculations can and will be
systematically improved (e.g. by including induced three-body
interactions), extended to include effective operators other than the
Hamiltonian, and applied to heavier nuclei, where accurate
phenomenological Hamiltonians are harder to obtain.
Acknowledgements.
We thank Scott Bogner, Morten Hjorth-Jensen and Thomas Papenbrock
for useful discussions. This work was supported by the Office of
Nuclear Physics, U.S. Department of Energy (Oak Ridge National
Laboratory), under DE-FG02-96ER40963 (University of Tennessee),
DE-FG02-97ER41019 (University of North Carolina), DE-SC0008499
(NUCLEI SciDAC collaboration), NERRSC Grant No. 491045-2011, and
the Field Work Proposal ERKBP57 at Oak Ridge National Laboratory.
Computer time was provided by the Innovative and Novel Computational
Impact on Theory and Experiment (INCITE) program. TRIUMF receives
funding via a contribution through the National Research Council
Canada. This research used resources of the Oak Ridge Leadership
Computing Facility located in the Oak Ridge National Laboratory,
which is supported by the Office of Science of the Department of
Energy under Contract No. DE-AC05-00OR22725, and used computational
resources of the National Center for Computational Sciences, the
National Institute for Computational Sciences, and the Notur project
in Norway.
Note added. – Very recently Bogner et al
Bogner et al. (2014) applied the in-medium
similarity-renormalization-group method to construct non-perturbative
shell-model interactions and applied it to neutron-rich oxygen
isotopes.
References
Stanoiu et al. (2004)
M. Stanoiu,
F. Azaiez,
Z. Dombrádi,
O. Sorlin,
B. A. Brown,
M. Belleguic,
D. Sohler,
M. G. Saint Laurent,
M. J. Lopez-Jimenez,
Y. E. Penionzhkevich,
et al., Phys. Rev. C
69, 034312
(2004),
URL http://link.aps.org/doi/10.1103/PhysRevC.69.034312.
Kanungo et al. (2009)
R. Kanungo,
C. Nociforo,
A. Prochazka,
T. Aumann,
D. Boutin,
D. Cortina-Gil,
B. Davids,
M. Diakaki,
F. Farinon,
H. Geissel,
et al., Phys. Rev. Lett.
102, 152501
(2009),
URL http://link.aps.org/doi/10.1103/PhysRevLett.102.152501.
Holt et al. (2012)
J. D. Holt,
T. Otsuka,
A. Schwenk, and
T. Suzuki,
Journal of Physics G: Nuclear and Particle Physics
39, 085111
(2012),
URL http://stacks.iop.org/0954-3899/39/i=8/a=085111.
Hagen et al. (2012a)
G. Hagen,
M. Hjorth-Jensen,
G. R. Jansen,
R. Machleidt,
and
T. Papenbrock,
Phys. Rev. Lett. 109,
032502 (2012a),
URL http://link.aps.org/doi/10.1103/PhysRevLett.109.032502.
Steppenbeck
et al. (2013a)
D. Steppenbeck,
S. Takeuchi,
N. Aoi,
P. Doornenbal,
J. Lee,
M. Matsushita,
H. Wang,
H. Baba,
N. Fukuda,
S. Go, et al.,
Journal of Physics: Conference Series
445, 012012
(2013a),
URL http://stacks.iop.org/1742-6596/445/i=1/a=012012.
Steppenbeck
et al. (2013b)
D. Steppenbeck,
S. Takeuchi,
N. Aoi,
P. Doornenbal,
M. Matsushita,
H. Wang,
H. Baba,
N. Fukuda,
S. Go,
M. Honma,
et al., Nature
502, 207
(2013b),
URL http://dx.doi.org/10.1038/nature12522.
Wildenthal (1984)
B. Wildenthal,
Progress in Particle and Nuclear Physics
11, 5 (1984),
ISSN 0146-6410,
URL http://www.sciencedirect.com/science/article/pii/0146641084900115.
Brown and Wildenthal (1988)
B. A. Brown and
B. H. Wildenthal,
Annual Review of Nuclear and Particle Science
38, 29 (1988),
URL http://www.annualreviews.org/doi/abs/10.1146/annurev.ns.38.120188.000333.
Warburton and Brown (1992)
E. K. Warburton
and B. A. Brown,
Phys. Rev. C 46,
923 (1992),
URL http://link.aps.org/doi/10.1103/PhysRevC.46.923.
Hjorth-Jensen et al. (1995)
M. Hjorth-Jensen,
T. T. Kuo, and
E. Osnes,
Physics Reports 261,
125 (1995),
URL http://www.sciencedirect.com/science/article/pii/0370157395000126.
Otsuka et al. (2010)
T. Otsuka,
T. Suzuki,
J. D. Holt,
A. Schwenk, and
Y. Akaishi,
Phys. Rev. Lett. 105,
032501 (2010),
URL http://link.aps.org/doi/10.1103/PhysRevLett.105.032501.
Navrátil et al. (2009)
P. Navrátil,
S. Quaglioni,
I. Stetcu, and
B. R. Barrett,
Journal of Physics G: Nuclear and Particle Physics
36, 083101
(2009),
URL http://stacks.iop.org/0954-3899/36/i=8/a=083101.
Barrett et al. (2013)
B. R. Barrett,
P. Navrátil,
and J. P. Vary,
Progress in Particle and Nuclear Physics
69, 131 (2013),
URL http://www.sciencedirect.com/science/article/pii/S0146641012001184.
Navrátil et al. (1997)
P. Navrátil,
M. Thoresen, and
B. R. Barrett,
Phys. Rev. C 55,
R573 (1997),
URL http://link.aps.org/doi/10.1103/PhysRevC.55.R573.
Lisetskiy et al. (2008)
A. F. Lisetskiy,
B. R. Barrett,
M. K. G. Kruse,
P. Navratil,
I. Stetcu, and
J. P. Vary,
Phys. Rev. C 78,
044302 (2008),
URL http://link.aps.org/doi/10.1103/PhysRevC.78.044302.
Shukla et al. (2011)
D. Shukla,
J. Engel, and
P. Navratil,
Phys. Rev. C 84,
044316 (2011),
URL http://link.aps.org/doi/10.1103/PhysRevC.84.044316.
Tsukiyama et al. (2012)
K. Tsukiyama,
S. K. Bogner,
and A. Schwenk,
Phys. Rev. C 85,
061304 (2012),
URL http://link.aps.org/doi/10.1103/PhysRevC.85.061304.
Coester (1958)
F. Coester,
Nuclear Physics 7,
421 (1958),
URL http://www.sciencedirect.com/science/article/pii/0029558258902803.
Coester and Kümmel (1960)
F. Coester and
H. Kümmel,
Nuclear Physics 17,
477 (1960),
URL http://www.sciencedirect.com/science/article/pii/0029558260901401.
Čížek (1966)
J. Čížek,
The Journal of Chemical Physics
45, 4256 (1966),
URL http://link.aip.org/link/?JCP/45/4256/1.
Čížek (2007)
J. Čížek,
On the Use of the Cluster Expansion and the Technique
of Diagrams in Calculations of Correlation Effects in Atoms and Molecules
(John Wiley & Sons, Inc., 2007), pp.
35–89, ISBN 9780470143599,
URL http://dx.doi.org/10.1002/9780470143599.ch2.
Kümmel et al. (1978)
H. Kümmel,
K. H. Lührmann,
and J. G.
Zabolitzky, Physics Reports
36, 1 (1978),
URL http://www.sciencedirect.com/science/article/pii/0370157378900819.
Hagen et al. (2013)
G. Hagen,
T. Papenbrock,
M. Hjorth-Jensen,
and D. J.
Dean, ArXiv e-prints
(2013), eprint 1312.7872.
Kondo et al. (2009)
Y. Kondo,
T. Nakamura,
Y. Satou,
T. Matsumoto,
N. Aoi,
N. Endo,
N. Fukuda,
T. Gomi,
Y. Hashimoto,
M. Ishihara,
et al., Phys. Rev. C
79, 014602
(2009),
URL http://link.aps.org/doi/10.1103/PhysRevC.79.014602.
Strongman et al. (2009)
M. J. Strongman,
A. Spyrou,
C. R. Hoffman,
T. Baumann,
D. Bazin,
J. Brown,
P. A. DeYoung,
J. E. Finck,
N. Frank,
S. Mosby,
et al., Phys. Rev. C
80, 021302
(2009),
URL http://link.aps.org/doi/10.1103/PhysRevC.80.021302.
Ozawa et al. (2011)
A. Ozawa,
Y. Hashizume,
Y. Aoki,
K. Tanaka,
T. Aiba,
N. Aoi,
H. Baba,
B. A. Brown,
M. Fukuda,
K. Inafuku,
et al., Phys. Rev. C
84, 064315
(2011),
URL http://link.aps.org/doi/10.1103/PhysRevC.84.064315.
Kobayashi et al. (2012)
N. Kobayashi,
T. Nakamura,
J. A. Tostevin,
Y. Kondo,
N. Aoi,
H. Baba,
S. Deguchi,
J. Gibelin,
M. Ishihara,
Y. Kawada,
et al., Phys. Rev. C
86, 054604
(2012),
URL http://link.aps.org/doi/10.1103/PhysRevC.86.054604.
Petri et al. (2012)
M. Petri,
S. Paschalis,
R. M. Clark,
P. Fallon,
A. O. Macchiavelli,
K. Starosta,
T. Baugher,
D. Bazin,
L. Cartegni,
H. L. Crawford,
et al., Phys. Rev. C
86, 044329
(2012),
URL http://link.aps.org/doi/10.1103/PhysRevC.86.044329.
Voss et al. (2012)
P. Voss,
T. Baugher,
D. Bazin,
R. M. Clark,
H. L. Crawford,
A. Dewald,
P. Fallon,
A. Gade,
G. F. Grinyer,
H. Iwasaki,
et al., Phys. Rev. C
86, 011303
(2012),
URL http://link.aps.org/doi/10.1103/PhysRevC.86.011303.
Petri et al. (2011)
M. Petri,
P. Fallon,
A. O. Macchiavelli,
S. Paschalis,
K. Starosta,
T. Baugher,
D. Bazin,
L. Cartegni,
R. M. Clark,
H. L. Crawford,
et al., Phys. Rev. Lett.
107, 102501
(2011),
URL http://link.aps.org/doi/10.1103/PhysRevLett.107.102501.
Forssén et al. (2013)
C. Forssén,
R. Roth, and
P. Navrátil,
Journal of Physics G: Nuclear and Particle Physics
40, 055105
(2013),
URL http://stacks.iop.org/0954-3899/40/i=5/a=055105.
Stanoiu et al. (2008)
M. Stanoiu,
D. Sohler,
O. Sorlin,
F. Azaiez,
Z. Dombrádi,
B. A. Brown,
M. Belleguic,
C. Borcea,
C. Bourgeois,
Z. Dlouhy,
et al., Phys. Rev. C
78, 034315
(2008),
URL http://link.aps.org/doi/10.1103/PhysRevC.78.034315.
Tanaka et al. (2010)
K. Tanaka,
T. Yamaguchi,
T. Suzuki,
T. Ohtsubo,
M. Fukuda,
D. Nishimura,
M. Takechi,
K. Ogata,
A. Ozawa,
T. Izumikawa,
et al., Phys. Rev. Lett.
104, 062701
(2010),
URL http://link.aps.org/doi/10.1103/PhysRevLett.104.062701.
Ershov et al. (2012)
S. N. Ershov,
J. S. Vaagen,
and M. V.
Zhukov, Phys. Rev. C
86, 034331
(2012),
URL http://link.aps.org/doi/10.1103/PhysRevC.86.034331.
Entem and Machleidt (2003)
D. R. Entem and
R. Machleidt,
Phys. Rev. C 68,
041001 (2003),
URL http://link.aps.org/doi/10.1103/PhysRevC.68.041001.
Machleidt and Entem (2011)
R. Machleidt and
D. Entem,
Physics Reports 503,
1 (2011),
URL http://www.sciencedirect.com/science/article/pii/S0370157311000457.
Navrátil (2007)
P. Navrátil,
Few-Body Systems 41,
117 (2007),
URL http://dx.doi.org/10.1007/s00601-007-0193-3.
Gazit et al. (2009)
D. Gazit,
S. Quaglioni,
and
P. Navrátil,
Phys. Rev. Lett. 103,
102502 (2009),
URL http://link.aps.org/doi/10.1103/PhysRevLett.103.102502.
Roth et al. (2012)
R. Roth,
S. Binder,
K. Vobig,
A. Calci,
J. Langhammer,
and
P. Navrátil,
Phys. Rev. Lett. 109,
052501 (2012),
URL http://link.aps.org/doi/10.1103/PhysRevLett.109.052501.
Jurgenson et al. (2009)
E. D. Jurgenson,
P. Navrátil,
and R. J.
Furnstahl, Phys. Rev. Lett.
103, 082501
(2009),
URL http://link.aps.org/doi/10.1103/PhysRevLett.103.082501.
Hagen et al. (2007)
G. Hagen,
T. Papenbrock,
D. J. Dean,
A. Schwenk,
A. Nogga,
M. Włoch, and
P. Piecuch,
Phys. Rev. C 76,
034302 (2007),
URL http://link.aps.org/doi/10.1103/PhysRevC.76.034302.
Navrátil and Barrett (1996)
P. Navrátil and
B. R. Barrett,
Phys. Rev. C 54,
2986 (1996),
URL http://link.aps.org/doi/10.1103/PhysRevC.54.2986.
Kucharski and Bartlett (1998)
S. A. Kucharski
and R. J.
Bartlett, The Journal of Chemical Physics
108, 5243 (1998),
URL http://scitation.aip.org/content/aip/journal/jcp/108/13/10.1063/1.475961.
Taube and Bartlett (2008)
A. G. Taube and
R. J. Bartlett,
The Journal of Chemical Physics
128, 044110
(pages 13) (2008),
URL http://link.aip.org/link/?JCP/128/044110/1.
Gour et al. (2006)
J. R. Gour,
P. Piecuch,
M. Hjorth-Jensen,
M. Włoch, and
D. J. Dean,
Phys. Rev. C 74,
024310 (2006),
URL http://link.aps.org/doi/10.1103/PhysRevC.74.024310.
Jansen et al. (2011)
G. R. Jansen,
M. Hjorth-Jensen,
G. Hagen, and
T. Papenbrock,
Phys. Rev. C 83,
054306 (2011),
URL http://link.aps.org/doi/10.1103/PhysRevC.83.054306.
Jansen (2013)
G. R. Jansen,
Phys. Rev. C 88,
024305 (2013),
URL http://link.aps.org/doi/10.1103/PhysRevC.88.024305.
Bartlett and Musiał (2007)
R. J. Bartlett and
M. Musiał,
Rev. Mod. Phys. 79,
291 (2007),
URL http://link.aps.org/doi/10.1103/RevModPhys.79.291.
Okubo (1954)
S. Okubo,
Progress of Theoretical Physics
12, 603 (1954),
eprint http://ptp.oxfordjournals.org/content/12/5/603.full.pdf+html,
URL http://ptp.oxfordjournals.org/content/12/5/603.abstract.
Suzuki and Lee (1980)
K. Suzuki and
S. Lee,
Progress of Theoretical Physics
64, 2091 (1980).
Suzuki (1982)
K. Suzuki,
Progress of Theoretical Physics
68, 246 (1982),
eprint http://ptp.oxfordjournals.org/content/68/1/246.full.pdf+html,
URL http://ptp.oxfordjournals.org/content/68/1/246.abstract.
Kvaal (2008)
S. Kvaal,
Phys. Rev. C 78,
044330 (2008),
URL http://link.aps.org/doi/10.1103/PhysRevC.78.044330.
Scholtz et al. (1992)
F. Scholtz,
H. Geyer, and
F. Hahne,
Annals of Physics 213,
74 (1992), ISSN 0003-4916,
URL http://www.sciencedirect.com/science/article/pii/000349169290284S.
Hergert et al. (2013)
H. Hergert,
S. Binder,
A. Calci,
J. Langhammer,
and R. Roth,
Phys. Rev. Lett. 110,
242501 (2013),
URL http://link.aps.org/doi/10.1103/PhysRevLett.110.242501.
Cipollone et al. (2013)
A. Cipollone,
C. Barbieri, and
P. Navrátil,
Phys. Rev. Lett. 111,
062501 (2013),
URL http://link.aps.org/doi/10.1103/PhysRevLett.111.062501.
Stanton and Bartlett (1993)
J. F. Stanton and
R. J. Bartlett,
The Journal of Chemical Physics
98, 7029 (1993),
URL http://scitation.aip.org/content/aip/journal/jcp/98/9/10.1063/1.464746.
Hagen et al. (2012b)
G. Hagen,
M. Hjorth-Jensen,
G. R. Jansen,
R. Machleidt,
and
T. Papenbrock,
Phys. Rev. Lett. 108,
242501 (2012b),
URL http://link.aps.org/doi/10.1103/PhysRevLett.108.242501.
Baroni et al. (2013)
S. Baroni,
P. Navrátil,
and
S. Quaglioni,
Phys. Rev. C 87,
034326 (2013),
URL http://link.aps.org/doi/10.1103/PhysRevC.87.034326.
Bogner et al. (2014)
S. K. Bogner,
H. Hergert,
J. D. Holt,
A. Schwenk,
S. Binder,
A. Calci,
J. Langhammer,
and R. Roth,
ArXiv e-prints (2014),
eprint 1402.1407. |
Abstract
It follows from the famous Fagin’s theorem that all problems in NP are expressible in
existential second-order logic ($\exists$SO), and vice versa. Indeed, there are well-known
$\exists$SO characterizations of NP-complete problems such as $3$-colorability, Hamiltonicity and clique.
Furthermore, the $\exists$SO sentences that characterize those problems are simple and elegant.
However, there are also NP problems that do not seem to possess equally simple and elegant $\exists$SO characterizations.
In this work, we are mainly interested in this latter class of problems.
In particular, we characterize in second-order logic the class of hypercube graphs and the classes SATQBF${}_{k}$ of satisfiable quantified Boolean formulae with $k$ alternations of quantifiers. We also provide detailed descriptions of the strategies followed to obtain the corresponding nontrivial second-order sentences. Finally, we sketch a third-order logic sentence that defines the
class SATQBF = $\bigcup_{k\geq 1}$SATQBF${}_{k}$. The sub-formulae
used in the construction of these complex second- and third-order logic sentences, are good candidates to form part of a library of formulae. Same as libraries
of frequently used functions simplify the writing of complex
computer programs, a library of formulae could
potentially simplify the writing of complex second- and third-order queries,
minimizing the probability of error.
Expressing Properties in Second and Third Order Logic: Hypercube Graphs and SATQBF 111Pre-print of article submitted to an special issue of the Logic Journal of the IGPL with selected papers from the 16th Brazilian Logic Conference.
\@afterheading
School of Information Management, Victoria University of Wellington,
PO Box 600, Wellington 6140, New Zealand. E-mail: [email protected]
School of Engineering and Advanced Technology, Massey University,
Private Box 756, Wellington 6140, New Zealand. E-mail: [email protected]
ICTIC, Universidad de la Cuenca del Plata, Corrientes, Argentina and
Department of Informatics, Universidad Nacional de San Luis,
Ejercito de Los Andes 950, D5700HHW, San Luis, Argentina.
E-mail: [email protected]
Keywords: second-order logic, third-order logic, quantified Boolean formulae, queries, finite
model theory, hypercube graphs
1 Introduction
Examples of second-order formulae expressing different properties of
graphs are fairly common in the literature. Classical
examples are $3$-colorability,
Hamiltonicity, and clique (see [8, 10]
among others). These properties can be expressed by simple and
elegant second-order formulae. Likewise, there are graph properties
that can be expressed by simple and elegant third-order
formulae. One of those properties is that of being a hypercube graph
(see [5]).
An $n$-hypercube graph ${\bf Q}_{n}$, also called
an $n$-cube, is
an undirected graph whose vertices are binary $n$-tuples.
Two vertices of ${\bf Q}_{n}$ are adjacent iff they differ in exactly one bit.
The expressive power of third-order
logic is not actually required to characterize hypercube graphs, since
they can be recognized in nondeterministic polynomial
time. Recall that by Fagin’s theorem [4], $\exists$SO captures NP.
Thus there are formulae in existential second-order logic
($\exists$SO) which can express this property. Nevertheless,
to define the class of hypercube graphs in
second-order logic is certainly more challenging than to define it
in third-order logic.
From an applied perspective, this indicates that it makes sense to investigate higher-order quantifiers in the context of database query languages. Despite the fact that most of the queries commonly used in the industry are in P, the use of higher-order quantifiers can potentially simplify the way in which many of those queries are expressed.
Let SATQBF${}_{k}$ denote the class of satisfiable quantified Boolean formulae with $k$ alternating blocks of quantifiers. From Fagin-Stockmeyer characterization of the polynomial-time hierarchy [13] and the fact that SATQBF${}_{k}$ is complete for the level $\Sigma^{p}_{k}$ of that hierarchy [14], it follows that for every $k\geq 1$, SATQBF${}_{k}$ can be defined by a formula in the prenex fragment $\Sigma^{1}_{k}$ of second-order logic with $k$ alternating blocks of quantifiers. SATQBF${}_{k}$ provides a prime example of a property (or query) whose expression in the language of second-order logic is possible but challenging. Indeed, it is not a trivial task to write a second-order logic sentence that evaluates to true precisely on those word models that represent sentences in SATQBF${}_{k}$. As usual in finite model theory [3], the term word model refers here to a finite relational structure formed by a binary relation and a finite number of unary relations. By contrast, if we restrict our attention to quantified Boolean formulae in which the quantified free part is in conjunctive normal form and has exactly three Boolean variables in each conjunct, then the problem is expressible in monadic second-order logic provided that the formulae are encoded using a different kind of finite relational structures which include ternary relations (see [10]).
Thus, on the one hand there are well-known NP-complete problems such as
$3$-colorability, Hamiltonicity and clique, that have corresponding well-known characterizations
in $\exists$SO which are simple and elegant. Those characterizations
have in common that the existential second-order quantifiers
can be identified with the guessing stage of the NP algorithm, and that the
remaining first-order formula corresponds to the polynomial time
deterministic verification stage. On the other hand, there are well-known problems such as
hypercube graph (which can also be characterized in $\exists$SO) and SATQBF${}_{k}$ (which can
be characterized in $\Sigma^{1}_{k}$) that do not appear to have a straightforward characterization in second-order logic,
even if we consider the full second-order language.
This observation prompted us to write second-order characterizations of hypercube graph and SATQBF${}_{k}$.
The resulting second-order sentence for hypercube graph can be found in [11]. The corresponding sentence for
SATQBF${}_{k}$ was included in [12]. Both sentences are complex and several pages long.
In this article we present a detailed description of the strategies followed to write these
sentences. The sub-formulae used for the
implementation of these strategies could be part of a future library of
second-order formulae.
Same as libraries
of frequently used functions simplify the writing of complex
computer programs, a library of formulae could
potentially simplify the writing of complex second-order queries,
minimizing the probability of error.
The minimization of the probability of error constitutes an important objective in the context of this work,
since given a query $q$ and a second-order formula $\varphi$, it is not possible to formally prove whether $\varphi$ expresses $q$.
For this reason, we make use of full second-order logic to present the characterizations of hypercube graph and SATQBF${}_{k}$, even though its $\exists$SO and $\Sigma^{1}_{k}$ fragments, respectively, already have the expressive power required for these tasks.
This has permitted us to write relatively clear and intuitive formulae as well as to follow a top-down strategy, similar to that commonly used in the development of computer programs, to further reduce the chance of error.
If we consider the whole class SATQBF = $\bigcup_{k\geq 1}$SATQBF${}_{k}$ of satisfiable quantified Boolean formulae, then the problem becomes PSPACE-complete. Since PSPACE can be captured by second-order logic extended with a transitive closure operator, and furthermore this logic is widely conjectured to be strictly more expressive than the standard second-order logic, the existence of a second-order logic characterization of this problem is unlikely. Thus, we decided to look for a characterization in third-order logic. Note that it is a well-known fact that third-order logic is powerful enough as to characterize every problem in PSPACE.
We conclude the paper presenting a sketch of a third-order logic sentence that defines the
class SATQBF. That is, we present a strategy to write a third-order sentence
that evaluates to true precisely on those word models that represent sentences in SATQBF.
We strongly believe that in many respects the descriptive approach
to Complexity is more convenient than the classical one. That is,
using formulae of some logic to study upper bounds in the time or
space complexity of a given problem, instead of Turing machines.
There are many different measures which can be taken on the formulae
that express a given problem such as quantifier rank, quantifier
blocks alternation, number of variables, number of binary
connectives, and arity of quantified relation variables. It has been proved
that bounds on those measures impact on the
expressive power of logics over finite models (see
[10], [3], [8]). Furthermore, it is
rather obvious that all those measures are decidable, in contrast to
the use of Turing machines, where the usual measures relevant to
computation power such as time, space, treesize, and number of
alternations, are clearly undecidable. Regarding lower bounds there
are also several well studied and powerful techniques in Descriptive
Complexity which proved to be extremely useful in the last decades,
such as Ehrenfeucht-Fraisse games and their variations (see
[9] in particular) and 0-1 Laws (again see [10],
[3], [8]).
Hence, it is important to learn how to build
formulae which are large, but still intuitive and clearly
understandable in a top down approach, in the same way that this is
important in the construction of algorithms in the classical
approach to Complexity, which are also clear and intuitive no matter
their size. The work reported in this article is to the authors’
knowledge one of the first steps in that direction.
In the next section, we introduce the necessary notation and
formally describe by means of a third-order logic sentence, the
class of hypercube graphs. In Section 3, we define in second-order
logic the basic arithmetic operations that we need for this work. We
describe the strategy used to characterize the class of hypercube
graphs in the language of second-order logic in Section 4. In
Section 5 we formally describe the problems SATQBF${}_{k}$ and SATQBF,
and we consider their complexity. In Section 6, we explain in full
detail how to build for each $k\geq 1$, a second-order sentence that
expresses $SATQBF_{k}$. In Section 7 we explain how to build a third-order
logic sentence which expresses SATQBF, and we give a sketch of
such formula. Finally in Section 8, we present some final considerations.
2 Background
We assume that the reader is acquainted with the basic concepts
and the framework of Finite Model Theory [3, 10]. We use the notation from [10].
We work on the vocabulary $\sigma=\{E\}$ of
graphs. An undirected graph ${\bf G}$ is a finite relational structure of vocabulary
$\sigma$ satisfying $\varphi_{1}\equiv\forall xy(E(x,y)\rightarrow E(y,x))$ and $\varphi_{2}\equiv\forall x(\neg E(x,x))$. If we do not require
${\bf G}$ to satisfy neither $\varphi_{1}$ nor $\varphi_{2}$, then we speak of a directed graph (or
digraph). We denote as $V$ the domain of the structure ${\bf G}$, i.e., the set of vertices
of the graph ${\bf G}$. The edge relation of ${\bf G}$ is denoted as $E^{\bf G}$.
By second-order logic we refer to the logic
that is obtained when first-order
logic is extended with second-order variables which range over subsets and
relations defined over the domain, and quantification over such variables. As usual, we use uppercase
letters $X,Y,Z,\ldots$ to denote second-order variables and lower case
letters $x,y,z,\ldots$ to denote first-order variables. The arity of the
second-order variables that we use in our formulae is always clear from the context.
See [10] or [3] for a formal definition of second-order logic in the
context of finite model theory. We include an example of a second-order
formula that defines a simple graph property instead.
Example 1
An undirected graph ${\bf G}$ is regular if
all its vertices have the same degree. It is well known that the class of
regular graphs is not definable in first-order logic [3, 8]. In
second-order logic, this class can be defined as follows:
$\exists A\big{(}\forall x\big{(}\exists B(\mathrm{A1}\wedge\mathrm{A2})\big{)}%
\big{)}$ where
•
A1 expresses “$B$ is the set of vertices which are adjacent to $x$”.
$\mathrm{A1}\equiv\forall z\big{(}B(z)\leftrightarrow E(x,z)\big{)}$
•
A2 expresses “the sets $A$ and $B$ have the same cardinality”
with a formula stating that there is a bijection $F$ from $A$ to $B$.
$\mathrm{A2}\equiv\exists F\forall xyz\big{(}\mathrm{A2.1}\wedge\mathrm{A2.2}%
\wedge\mathrm{A2.3}\wedge\mathrm{A2.4}\wedge\mathrm{A2.5}\big{)}$ where
–
A2.1 means “$F$ is a subset of $A\times B$”.
$\mathrm{A2.1}\equiv\big{(}F(x,y)\rightarrow A(x)\wedge B(y)\big{)}$
–
A2.2 means “$F$ is a function”.
$\mathrm{A2.2}\equiv\big{(}F(x,y)\wedge F(x,z)\rightarrow y=z\big{)}$
–
A2.3 means “$F$ is total”.
$\mathrm{A2.3}\equiv\big{(}A(x)\rightarrow\exists y(F(x,y))\big{)}$
–
A2.4 means “$F$ is injective”.
$\mathrm{A2.4}\equiv\big{(}F(x,z)\wedge F(y,z)\rightarrow x=y\big{)}$
–
A2.5 means “$F$ is surjective”.
$\mathrm{A2.5}\equiv\big{(}B(y)\rightarrow\exists x(F(x,y))\big{)}$
We say that a sentence $\varphi$ expresses a Boolean
query $q$ (or property) over finite relational structures of vocabulary
$\sigma$, if for every finite relational structure ${\bf G}$ of vocabulary
$\sigma$, $q({\bf G})=\mathrm{true}$ iff
${\bf G}\models\varphi$. For instance the sentence in Example 1 expresses
the Boolean query: Is ${\bf G}$ a regular graph?
We denote by
$Mod(\varphi)$ the class of finite $\sigma$-structures ${\bf G}$ such
that ${\bf G}\models\varphi$. A class of finite $\sigma$-structures
${\cal C}$ is definable in a logic $\cal L$, if ${\cal C}=Mod(\varphi)$ for some ${\cal L}$-sentence $\varphi$ of vocabulary $\sigma$. For instance
the class of regular graphs is definable in second-order logic, as shown by the formula given in Example 1.
Next, we define the class of hypercube graphs using a relatively simple and
elegant formula in third-order logic. This logic extends second-order
logic with third-order variables which range over subsets and relations
defined over the powerset of the domain, and quantification over such variables. We use uppercase
calligraphic letters ${\cal X},{\cal Y},{\cal Z},\ldots$ to denote
third-order variables. A formal definition of higher-order logics in the
context of finite model theory can be found in [7] among others.
Example 2
An $n$-hypercube (or $n$-cube for short) ${\bf Q}_{n}$ can be
defined as an undirected graph whose vertices are all the binary $n$-tuples.
Two vertices of ${\bf Q}_{n}$ are adjacent iff they differ in exactly one bit.
A $1$-cube ${\bf Q}_{1}$, a $2$-cube ${\bf Q}_{2}$ and a $3$-cube ${\bf Q}_{3}$ are displayed in Figure 2.1.
We can build an $(n+1)$-cube ${\bf Q}_{n+1}$ starting with two
isomorphic copies of an $n$-cube ${\bf Q}_{n}$ and adding edges between
corresponding vertices. Using this fact, we can define in
third-order logic the so called class of hypercube graphs, as
follows:
$\exists{\cal C}\exists{\cal O}\big{(}\mathrm{A1}\wedge\mathrm{A2}\wedge\forall
G%
_{1}\forall G_{2}\big{(}({\cal C}(G_{1})\wedge{\cal C}(G_{2})\wedge\mathrm{A3}%
)\rightarrow\mathrm{A4}\big{)}\wedge\mathrm{A5}\wedge\mathrm{A6}\big{)}$ where
•
A1 expresses “${\cal C}$ is a class of undirected graphs”.
•
A2 expresses “${\cal O}$ is a total order on ${\cal C}$”.
•
A3 expresses “$G_{1}$ is the immediate predecessor of $G_{2}$ in the order ${\cal O}$”.
•
A4 expresses “$G_{2}$ can be built from two isomorphic copies of $G_{1}$ by adding edges between the corresponding vertices”.
•
A5 expresses “the first graph in the order ${\cal O}$ is a $Q_{1}$”.
•
A6 expresses “the last graph in the order ${\cal O}$ is the input graph”.
In turn, we can express A4 as follows:
$\exists F_{1}\exists F_{2}\big{(}\mathrm{A4.1}\wedge\mathrm{A4.2}\wedge\mathrm%
{A4.3}\wedge\forall x(x\in dom(G_{1})\rightarrow\mathrm{A4.4})\wedge$
$\neg\exists xy(x,y\in dom(G_{1})\wedge x\neq y\wedge\mathrm{A4.5})\big{)}$ where
•
A4.1 expresses “$F_{1}$ and $F_{2}$ are injective and total functions from $dom(G_{1})$ to $dom(G_{2})$”.
•
A4.2 expresses “the ranges of $F_{1}$ and $F_{2}$ form a partition of $dom(G_{2})$”.
•
A4.3 expresses “$F_{1}$ and $F_{2}$ are isomorphisms from $G_{1}$ to the sub-graphs of $G_{2}$ induced by the ranges of $F_{1}$ and $F_{2}$, respectively”.
•
A4.4 expresses “there is an edge in $G_{2}$ which connects $F_{1}(x)$ and $F_{2}(x)$”.
•
A4.5 expresses “there is an edge in $G_{2}$ which connects $F_{1}(x)$ and $F_{2}(y)$”.
Note that, if there is an edge $(a,b)$ in $G_{2}$ such that $a$ belongs to the
range of $F_{1}$ and $b$ belongs to the range of $F_{2}$, or vice versa, then either $F^{-1}_{1}(a)=F^{-1}_{2}(b)$ or $F^{-1}_{1}(b)=F^{-1}_{2}(a)$.
The missing logic formulae in this example are left as an exercise for the reader.
The property of a graph being an $n$-cube for some $n$, is known to be in NP.
A nondeterministic Turing machine can decide in polynomial time whether an
input structure ${\bf G}$ of the vocabulary $\sigma$ of graphs is an
hypercube, by simply computing the following steps:
i.
Compute the logarithm in base $2$ of the size $n$ of the domain of the input
structure ${\bf G}$ which must be a positive integer;
ii.
Guess a sequence $s_{1},\ldots,s_{n}$ of $n$ binary strings, each of length $\log_{2}n$;
iii.
Check in polynomial time that all binary strings are unique, that the sequence contains all binary strings of length $\log_{2}n$ and that,
for some ordering $a_{s_{1}},\ldots,a_{s_{n}}$ of the nodes in $V$, a string $s_{i}$ differs
from a string $s_{j}$ in exactly $1$ bit iff there is an edge
$(a_{s_{i}},a_{s_{j}})\in E^{\bf G}$.
Thus, as we mentioned in the introduction, the full expressive power of third-order
logic is not actually needed to characterize the class of hypercube graphs. In fact, there is a
formula in $\exists$SO which can express this property. Recall that by Fagin’s
theorem [4], $\exists$SO captures $NP$. However, it is very
unlikely that there is a formula in second-order logic, not to mention in
$\exists$SO, that expresses the property
in a way which is as intuitive and simple as in the example above.
3 Arithmetic in Second-Order Logic
We define in this section the basic arithmetic operations of addition,
multiplication and exponentiation in second-order logic over finite structures. We encode initial
segments of natural number as finite relational structures by using linear
digraphs. Let $\bf G$ be a linear digraph. The first (root) element of the domain in the order determined by
the edge relation $E^{\bf G}$ represents the $0$, the second element in this
order represents the $1$, the third element represents the $2$ and so on.
Since in a linear digraph, $E^{\bf G}$ is the successor relation, for clarity we use $\mathrm{succ}(x,y)$ to denote $E(x,y)$. We also use $x=n$ where $n>0$ to denote the formula of the form
$$\exists y(\mathrm{succ}(y,x)\wedge\exists x(\mathrm{succ}(x,y)\wedge\exists y(%
\mathrm{succ}(y,x)\wedge\cdots\wedge\varphi)\cdots))$$
with $n$ nested quantifiers and $\varphi\equiv\neg\exists x(\mathrm{succ}(x,y))$ if $n$ is odd or $\varphi\equiv\neg\exists y(\mathrm{succ}(y,x))$ if $n$ is even. Likewise, $x=0$ denotes $\neg\exists y(\mathrm{succ}(y,x))$.
We assume a total order
$\leq$ of the nodes in $V$ such that $x\leq y$ iff there is a path from
$x$ to $y$ in $\bf G$ or $x=y$. This total order is easily definable in
second-order logic.
Let us start by defining the operation of addition. The strategy is depicted in
Figure 3.1 in which we show the result $z$ of
adding $x$ and $y$ along a linear graph.
The predicate $\mathrm{sum}(x,y,z)$, which is true iff $z=x+y$, can be defined in second-order logic as follows.
$\big{(}x=0\wedge z=y\big{)}\vee\big{(}y=0\wedge z=x\big{)}\vee$
$\big{(}x\neq 0\wedge y\neq 0\wedge\exists F\big{(}\mathrm{A1}\wedge F(z,y)%
\wedge\exists x^{\prime}y^{\prime}(\hskip 2.845276pt\mathrm{succ}(x,x^{\prime}%
)\wedge F(x^{\prime},y^{\prime})\wedge y^{\prime}=1)\wedge$
$\forall x^{\prime}y^{\prime}x^{\prime\prime}y^{\prime\prime}((\hskip 2.845276%
pt\mathrm{succ}(x^{\prime},y^{\prime})\wedge F(x^{\prime},x^{\prime\prime})%
\wedge F(y^{\prime},y^{\prime\prime}))\rightarrow\hskip 2.845276pt\mathrm{succ%
}(x^{\prime\prime},y^{\prime\prime}))\big{)}\big{)}$
where A1 expresses “$F$ is an injective function with domain $\{n\in V\hskip 2.845276pt|\hskip 2.845276pt\mathrm{succ}(x)\leq n\leq z\}$”.
It is an easy and supplementary task to write the actual formula corresponding to A1. For the sake of clarity, we avoid this kind of supplementary details from now on.
The next
arithmetic operation that we define is multiplication. The strategy is
depicted in Figure 3.2 in which we show the result $z$ of $x$ times
$y$. Each of the nodes in the subset $S=\{2,\ldots,x\}$ can be considered as a root of a different ordered tree in a forest. Each
root node in the forest has $y$ children and the result $z$ is the last child
of node $x$.
The predicate $\mathrm{times}(x,y,z)$, which is true if $z=x\times y$, can be defined in second-order logic as follows.
$(x=1\wedge y\neq 0\wedge z=y)\vee(y=1\wedge x\neq 0\wedge z=x)\vee((x=0\vee y=%
0)\wedge z=0)\vee$
$\big{(}x\neq 0\wedge y\neq 0\wedge x\neq 1\wedge y\neq 1\wedge$
$\exists S\big{(}\forall u\big{(}\big{(}(2\leq u\wedge u\leq x)\rightarrow[%
\exists y^{\prime}(S(u,y^{\prime}))\wedge$
$\forall x^{\prime}y^{\prime}((x^{\prime}\leq y^{\prime}\wedge x^{\prime}\neq y%
^{\prime}\wedge S(u,x^{\prime})\wedge S(u,y^{\prime}))\rightarrow$
$\neg\exists z^{\prime}(x^{\prime}\leq z^{\prime}\wedge z^{\prime}\leq y^{%
\prime}\wedge\neg S(u,z^{\prime})))\wedge$
$\exists F(\mathrm{A1})\wedge\mathrm{A2}\wedge\mathrm{A3}\wedge\mathrm{A4}]\big%
{)}\wedge$
$\mathrm{A5}\wedge\forall uv(S(u,v)\rightarrow(2\leq u\wedge u\leq x))\big{)}%
\big{)}$ where
•
A1 expresses “$F$ is a bijection from $\{n\in V\hskip 2.845276pt|\hskip 2.845276ptS(u,n)\}$ to $\{n\in V\hskip 2.845276pt|\hskip 2.845276pt1\leq n\leq y\}$, which means that the output degree of $u$ is $y$”.
•
A2 expresses “if $u=2$ then the first child of $u$ is $\mathrm{succ}(y)$”.
•
A3 expresses “if $u=x$ then the last child of $u$ is $z$”.
•
A4 expresses “if $u\neq 2$ then $\mathrm{succ}(c_{u-1},c_{u})$ for $c_{u-1}$ the last child of $u-1$ and $c_{u}$ the first child of $u$”.
•
A5 expresses “the input degree of every node in $S$ is $\leq 1$”.
Finally, we need to define the arithmetic operation of exponentiation in second-order logic. In this case, the strategy is depicted in
Figure 3.3. Note that, the first node in the linear digraph is
$x^{1}$, the second node is $x^{2}$, and so on till node $y$-th (the final
node) which is $x^{y}$.
The predicate $\mathrm{exp}(x,y,z)$, which is true if $z=x^{y}$, can be
defined in second-order logic as follows.
$(x\neq 0\wedge y=0\wedge z=1)\vee(y=1\wedge z=x)\vee(x=1\wedge z=1)\vee$
$\big{(}x\geq 2\wedge y\geq 2\wedge\exists V^{\prime}E^{\prime}\big{(}\mathrm{A%
1}\wedge\exists F(\mathrm{A2})\wedge$
$\forall u(\neg V^{\prime}(u)\vee(u=x\vee\exists x^{\prime}(E^{\prime}(x^{%
\prime},u)\wedge\mathrm{times}(x,x^{\prime},u))))\big{)}\big{)}$ where
•
A1 expresses “$(V^{\prime},E^{\prime})$ is a linear digraph whose first (root) node is $x$ and whose last (leaf) node is $z$”.
•
A2 expresses “$F$ is a bijection from $V^{\prime}$ to $\{1,\ldots,y\}$, i.e., $|V^{\prime}|=y$”.
4 Hypercube Graph in Second-Order Logic
We describe in this section two different strategies to define in second-order logic the
class of hypercube graphs. The first strategy is based in the usual definition of Hypercube graph which identifies the nodes of the graph with binary strings. This definition was explained and expressed by means of a third-order logic formula in Example 2. The second strategy is based in the following definition: An $n$-hypercube graph is a graph with $2^{n}$ nodes, which correspond to the subsets of a set with $n$ elements. Two nodes labelled by subsets $S_{i}$ and $S_{j}$ are joined by an edge if and only if $S_{i}$ can be obtained from $S_{j}$ by adding or removing a single element. The first strategy resulted in a more cumbersome formula than the formula produced by the second strategy. However, the descriptive complexity of the formula produced by this latter strategy is higher.
4.1 First Strategy
The idea is to use binary encodings to represent
each node in the graph, and then to compare the binary encodings of two
connected nodes to identify whether they differ exactly in $1$ bit.
Following a top down approach to the problem, we start with a very general
schema of the formula and then we explore the main sub-formulae involved
in the solution. We aim for a good balance between
level of detail and clarity of presentation. Consequently,
we leave out of the presentation some trivial sub-formulae which are not central to the general strategy.
Let $\bf G$ be an undirected graph with $|V|=n$. The following second-order formula is satisfied by $\bf G$ iff $\bf G$ is an $m$-hypercube graph for some $m$.
$$\displaystyle\varphi_{1}\equiv\exists\leq\big{(}\mathrm{A1}\wedge\exists F\,%
\exists m\big{(}\mathrm{A2}\wedge\forall xy(E(x,y)\leftrightarrow\mathrm{A3})%
\wedge\mathrm{A4}\big{)}\big{)}\text{ where}$$
•
A1 expresses “$\leq$ is a total order of the domain $V$ of $\bf G$”.
•
A2 expresses “$F$ is a bijection on $V$”.
•
A3 expresses “The binary encodings of $F(x)$ and $F(y)$ have both length $m$ and differ exactly in one bit”.
•
A4 expresses “There is a node whose binary encoding contains no zeros”.
The total order $\leq$ is used to identify each individual node of $V$. Thus,
we can assume that $V=\{0,\ldots,n-1\}$. This is needed for the
binary encoding of the nodes in $V$, as it will become clear latter on.
It should be clear how to express A1 and A2 in the language of second-order
logic. Thus we concentrate our effort in explaining the strategy to express A3.
Finally, note that A4 means that all binary encodings (of length $m$) correspond to some node in $V$, which implies that the number of nodes of ${\bf G}$ is a power of $2$, and also that $m=\log_{2}n$. A sub-formula that expresses A4 can be easily built by using the same ideas that we use for A3 below. That is, we can existentially quantify for some node $z$, a linear digraph $(V_{z},E_{z})$ and a Boolean assignment $B_{z}$ which assigns $1$ to each node, and such that the binary string represented by $(V_{z},E_{z},B_{z})$ is the binary encoding of $F(z)$.
The following formula expresses A3.
$$\displaystyle\exists V_{x}E_{x}V_{y}E_{y}B_{x}B_{y}\big{(}$$
$$\displaystyle\mathrm{A3.1}$$
$$\displaystyle\wedge\mathrm{A3.2}\wedge\mathrm{A3.3}\wedge\mathrm{A3.4}\wedge%
\mathrm{A3.5}\wedge$$
$$\displaystyle\exists G\big{(}$$
$$\displaystyle\mathrm{A3.6}\wedge$$
$$\displaystyle\forall uv((E_{x}(u,v)\rightarrow\exists u^{\prime}v^{\prime}(G(u%
,u^{\prime})\wedge G(v,v^{\prime})\wedge E_{y}(u^{\prime},v^{\prime})))\wedge$$
$$\displaystyle\quad\quad(E_{y}(u,v)\rightarrow\exists u^{\prime}v^{\prime}(G(u^%
{\prime},u)\wedge G(v^{\prime},v)\wedge E_{x}(u^{\prime},v^{\prime}))))\wedge$$
$$\displaystyle\exists v\forall v^{\prime}((\mathrm{A3.7}\rightarrow v^{\prime}%
\neq v)\wedge$$
$$\displaystyle\quad\quad\quad(\mathrm{A3.8}\rightarrow v^{\prime}=v))\big{)}%
\big{)}\text{ where}$$
•
A3.1 expresses “$(V_{x},E_{x})$ and $(V_{y},E_{y})$ are linear digraphs”.
•
A3.2 expresses “$B_{x}$ is a function from $V_{x}$ to $\{0,1\}$”.
•
A3.3 expresses “$B_{y}$ is a function from $V_{y}$ to $\{0,1\}$”.
•
A3.4 expresses “$(V_{x},E_{x},B_{x})$ is the binary encoding of $F(x)$”.
•
A3.5 expresses “$(V_{y},E_{y},B_{y})$ is the binary encoding of $F(y)$”.
•
A3.6 expresses “$G$ is a bijection from $V_{x}$ to $V_{y}$”.
•
A3.7 expresses “$B_{x}(v^{\prime})=B_{y}(G(v^{\prime}))$”.
•
A3.8 expresses “$B_{x}(v^{\prime})\neq B_{y}(G(v^{\prime}))$”.
To complete the picture, we need to explain how to write A3.4 and A3.5 in the language of second-order logic.
Since both can be expressed in second-order logic in a similar way, we only show the formula for A3.4. Let $x_{i}$ be the $i$-th node in the
linear graph $(V_{x},E_{x})$ defined in the previous formula. We say that $(V_{x},E_{x},B_{x})$ is the binary
encoding of $F(x)$ if
$$F(x)=b_{1}\times 2^{m-1}+b_{2}\times 2^{m-2}+\cdots+b_{m}\times 2^{0}\text{, %
where $b_{i}=B_{x}(x_{i})$.}$$
In second-order logic, we use a
function $W_{x}$ which
assigns to each node $x_{i}$ in $V_{x}$ its corresponding value $b_{i}\times 2^{m-i}$ in the encoding. This function is depicted in
Figure 4.1. The following formula defines the encoding.
$\exists W_{x}\,I_{x}\,n_{x}\,v\,w\,\forall x^{\prime}\big{(}\mathrm{A3.4.1}\wedge$
$\forall s\,s^{\prime}\,q\,q^{\prime}((E_{x}(s,q)\wedge I_{x}(s,s^{\prime})%
\wedge I_{x}(q,q^{\prime}))\rightarrow\mathrm{succ}(s^{\prime},q^{\prime}))\wedge$
$\mathrm{A3.4.2}\wedge\mathrm{A3.4.3}\wedge\,I_{x}(w,m)\wedge$
$\exists V^{\prime}\,E^{\prime}\,v_{1}\,v_{2}\,w[\mathrm{A3.4.4}\wedge\mathrm{A%
3.4.5}\wedge$
$\forall u\big{(}\neg V^{\prime}(u)\vee((u=v_{1}\rightarrow u=0)\wedge(u=v_{2}%
\rightarrow u=1)\wedge(u=w\rightarrow\mathrm{A3.4.6})\wedge$
$((u\neq v_{1}\wedge u\neq v_{2})\rightarrow\exists y^{\prime}(E^{\prime}(y^{%
\prime},u)\wedge\mathrm{A3.4.7})))\big{)}\wedge$
$V_{x}(x^{\prime})\rightarrow[(\mathrm{A3.4.8})\vee(\mathrm{A3.4.9}\,\wedge%
\exists t(W_{x}(x^{\prime},t)\,\wedge\,\mathrm{A3.4.10}))]\wedge$
$\mathrm{A3.4.11}\wedge\mathrm{A3.4.12}]\big{)}$ where
•
A3.4.1 expresses “$I_{x}$ is a bijection from $V_{x}$ to $\{1,\ldots,m\}$”.
•
A3.4.2 expresses “$v$ and $w$ are the first and last nodes of $(V_{x},E_{x})$, respectively”.
•
A3.4.3 expresses “$I_{x}(v,1)$”.
•
A3.4.4 expresses “$(V^{\prime},E^{\prime})$ is a linear graph”.
•
A3.4.5 expresses “$v_{1},v_{2}$ and $w$ are the $1$-st, $2$-nd and last nodes in $(V^{\prime},E^{\prime})$, respectively”.
•
A3.4.6 expresses “$\mathrm{exp}(2,m-1,u)$”.
•
A3.4.7 expresses “$\mathrm{times}(2,y^{\prime},u)$”.
•
A3.4.8 expresses “$B_{x}(x^{\prime})=0$” $\wedge$ “$W_{x}(x^{\prime})=0$”.
•
A3.4.9 expresses “$B_{x}(x^{\prime})=1$” $\wedge$ “$\mathrm{sum}(n_{x},I_{x}(x^{\prime}),m)$”.
•
A3.4.10 expresses “$\mathrm{exp}(2,n_{x},t)$”
•
A3.4.11 expresses “$F(x)=W_{x}(x_{1})+W_{x}(x_{2})+\cdots+W_{x}(x_{m})$ for $x_{i}$ the $i$-th node in $(V_{x},E_{x})$”.
•
A3.4.12 expresses “$W_{x}$ is a function from $V_{x}$ to $V^{\prime}$”.
Finally, we note that A3.4.11 can be expressed as follows.
$\exists U_{x}\big{(}\mathrm{A3.4.11.1}\wedge\forall x^{\prime}\big{(}\neg V_{x%
}(x^{\prime})\vee(\mathrm{A3.4.11.2}\wedge\mathrm{A3.4.11.3}\wedge$
$(\mathrm{A3.4.11.4}\rightarrow\exists x^{\prime\prime}(E_{x}(x^{\prime\prime},%
x^{\prime})\wedge\mathrm{A3.4.11.5})))\big{)}\big{)}$ where
•
A3.4.11.1 expresses “$U_{x}$ is a function from $V_{x}$ to $V$”.
•
A3.4.11.2 expresses “if $x^{\prime}$ is the first node in $(V_{x},E_{x})$ then $U_{x}(x^{\prime})=W_{x}(x^{\prime})$”.
•
A3.4.11.3 expresses “if $x^{\prime}$ is the last node in $(V_{x},E_{x})$ then $U_{x}(x^{\prime})=F(x)$”.
•
A3.4.11.4 expresses “$x^{\prime}$ is not the first node in $(V_{x},E_{x})$”.
•
A3.4.11.5 expresses “$\mathrm{sum}(U_{x}(x^{\prime\prime}),W_{x}(x^{\prime}),U_{x}(x^{\prime}))$”.
4.2 Second Strategy
The second strategy to define the class of hypercube graphs can be described in two steps.
i.
To identify every node $x$ in the input graph $\bf G$ with a different subset $S_{x}$ of a set $V^{\prime}\subset V$ of cardinality $\log_{2}|V|$, making sure that every subset of $V^{\prime}$ is assigned to some node of $\bf G$.
ii.
To check that for every pair of nodes $x$ and $y$ in $\bf G$, there is an edge between $x$ and $y$ iff $S_{x}$ can be obtained from $S_{y}$ by adding or removing a single element.
In second-order logic we can express this strategy as follows.
$$\displaystyle\varphi_{2}\equiv\exists R\big{(}$$
$$\displaystyle\exists V^{\prime}\big{(}\mathrm{A1}\wedge\forall S(\mathrm{A2}%
\rightarrow(\exists x(\mathrm{A3}\wedge\mathrm{A4})))\wedge\exists z(\mathrm{A%
5})\big{)}\wedge$$
$$\displaystyle\forall xy((E(x,y)\wedge E(y,x))\leftrightarrow\mathrm{A6})\big{)}$$
where
•
A1 expresses “$V^{\prime}\subset V\wedge V^{\prime}\neq\emptyset$”.
•
A2 expresses “$S\subseteq V^{\prime}\wedge S\neq\emptyset$”.
•
A3 expresses “$x$ is identified with $S$ via $R$”.
$\mathrm{A3}\equiv\forall v(R(x,v)\leftrightarrow S(v))$
•
A4 expresses “no other node $y\neq x$ can be identified with $S$ via $R$”.
$\mathrm{A4}\equiv\neg\exists y(x\neq y\wedge\forall v(R(y,v)\leftrightarrow S(%
v)))$
•
A5 expresses “all nodes, with the only exception of node $z$, are identified with some nonempty subset of $V^{\prime}$ via R”.
$\mathrm{A5}\equiv\neg\exists v\big{(}R(z,v)\big{)}\wedge\forall z^{\prime}\big%
{(}z\neq z^{\prime}\rightarrow\exists S(\mathrm{A5.1}\wedge\forall v(R(z^{%
\prime},v)\leftrightarrow S(v)))\big{)}$ where
–
A5.1 expresses “$S\neq\emptyset\wedge S\subseteq V^{\prime}$”.
•
A6 expresses “the set $S_{x}$ identified with $x$ can be obtained from the set $S_{y}$ identified with $y$ by adding or removing a single element”.
$\mathrm{A6}\equiv\exists v\big{(}\big{(}(R(x,v)\wedge\neg R(y,v))\vee(R(y,v)%
\wedge\neg R(x,v))\big{)}\wedge$
$\forall v^{\prime}\big{(}v^{\prime}\neq v\rightarrow(R(x,v^{\prime})%
\leftrightarrow R(y,v^{\prime}))\big{)}$
Remark 1
The formula $\varphi_{2}$ that expresses the second strategy has a prefix of second-order quantifiers of the form $\exists R\exists V^{\prime}\forall S$. Thus, it is in the class $\Sigma^{1}_{2}$. The existence of a formula in $\Sigma^{1}_{1}$ that expresses this second strategy is unlikely, since we must express that every subset $S$ is identified with some node in the graph. On the other hand, the formula $\varphi_{1}$ that expresses the first strategy, while considerably more cumbersome than $\varphi_{2}$, only uses existential second-order quantifiers and can be translated in a rather straightforward way into an equivalent $\Sigma^{1}_{1}$ formula. That is, we could transform the current quantification schema of the form
$$\forall xy\big{(}\exists V^{1}_{x}E^{2}_{x}B^{2}_{x}V^{1}_{y}E^{2}_{y}B^{2}_{y%
}\ldots\exists W^{2}_{x}U^{2}_{x}W^{2}_{y}U^{2}_{y}\ldots\big{)},$$
where the superindices added to the relation variables denote their arity, into an schema of the form
$$\big{(}\exists V^{2}_{x}E^{3}_{x}B^{3}_{x}V^{2}_{y}E^{3}_{y}B^{3}_{y}\ldots%
\exists W^{3}_{x}U^{3}_{x}W^{3}_{y}U^{3}_{y}\ldots\big{)},$$
where the prefix “$\forall xy$” is eliminated and the arity of every relation variable is increased in $1$, so that we can incorporate all nodes. Thus, for instance, every set $V^{1}_{x}$ corresponding to some node $x$ in a graph ${\bf G}$ is now encoded in the binary relation $V^{2}_{x}$ in such a way that $V^{1}_{x}=\{y|(x,y)\in V^{2}_{x}\}$. Then, we can simply express that the set $\{x|(x,y)\in V^{2}_{x}\}$ contains every node in the graph $\bf G$. Moreover, we can now omit $V_{y},E_{y},B_{y},W_{y}$ and $U_{y}$, since for every pair of nodes $x$ and $y$, their corresponding sets $V^{1}_{x}$ and $V^{1}_{y}$ will be both encoded into the binary relation $V^{2}_{x}$, and something similar will happen for the relations $E$, $B$, $W$ and $U$.
This is an important consideration since by Fagin-Stockmeyer characterization of the polynomial-time hierarchy [13] $\Sigma^{1}_{1}$ captures NP while $\Sigma^{1}_{2}$ captures $\mathrm{NP}^{\mathrm{NP}}$.
5 Quantified Boolean Formulae
A Boolean variable is any symbol to which we
can associate the truth values 0 and 1. Let $V$ be a countable set of Boolean
variables. The class of Boolean formulae over $V$ is the smallest class which
is defined by:
•
The Boolean constants 0 and 1 are Boolean formulae.
•
Every Boolean variable $x$ in $V$ is a Boolean formula.
•
If $\varphi$ and $\psi$ are Boolean formulae then $(\varphi\wedge\psi)$, $(\varphi\vee\psi)$ and $\neg(\varphi)$ are Boolean formulae.
The semantics of the Boolean formulae is given by the well-known semantics of the propositional logic.
A quantified Boolean formula over $V$, as defined by the influential Garey and
Johnson book on the theory of NP-Completeness [6], is a formula of the form
$$Q_{1}x_{1}Q_{2}x_{2}\ldots Q_{n}x_{n}(\varphi),$$
where $\varphi$ is a Boolean formula over $V$, $n\geq 0$, $x_{1},\ldots,x_{n}\in V$ and, for $1\leq i\leq n$, $Q_{i}$ is either “$\exists$” or “$\forall$”.
A variable that occurs in the Boolean formula but does not occur in the
prefix of quantifiers is called a free variable. We call QBF the set of
quantified Boolean formulae without free variables. As usual, for $k\geq 1$, $\mathrm{QBF}_{k}$ denotes
the fragment of QBF which consists of those formulae which start with an
existential block and have $k$ alternating blocks of quantifiers. Let $X\subset V$ be a
finite set of Boolean variables, we assume w.l.o.g. that a formula in
$\mathrm{QBF}_{k}$ over $X$ is of the form
$$\exists\bar{x}_{1}\forall\bar{x}_{2}\ldots Q\bar{x}_{k}(\varphi),$$
where for $1\leq i\leq k$, $\bar{x}_{i}=(x_{i1},\ldots,x_{il_{i}})$ is a
vector of $l_{i}$ different variables from $X$, $\exists\bar{x}_{i}$ denotes a
block of $l_{i}$ quantifiers of the form $\exists x_{i1},\ldots,\exists x_{il_{i}}$, $\forall\bar{x}_{i}$ denotes a
block of $l_{i}$ quantifiers of the form $\forall x_{i1},\ldots,\forall x_{il_{i}}$, $\varphi$ is a (quantifier free) Boolean formula
over $X$, $Q$ is “$\exists$” if $k$ is odd and “$\forall$” if $k$ is even, and the sets $X_{1},\ldots,X_{k}$ of variables
in $\bar{x}_{1},\ldots,\bar{x}_{k}$, respectively, form a partition of $X$.
We define next the notion of satisfiability of quantified Boolean
formulae. But first we introduce the concept of alternating valuations which
uses rooted binary trees to represent all possible valuations for a given formula, and paths
from the root to the leaves of such trees to represent individual valuations.
This unusual way of representing valuations is motivated by the way in
which we express in second-order logic the satisfiability problem for the
classes $\mathrm{QBF}_{k}$.
Let $\mathbf{T}_{v}$ be a rooted binary tree of
vocabulary $\sigma_{\mathbf{T}_{v}}=\{E,B,0,1\}$. That is, $\mathbf{T}_{v}$
is a maximally connected acyclic digraph in which every vertex has at most two child
vertices and, except for the root, has a unique parent. Here, $0$ and $1$ are
constant symbols which are interpreted as truth values and $B^{\mathbf{T}_{v}}$
is a total function which assigns a truth value $0^{\mathbf{T}_{v}}$ or
$1^{\mathbf{T}_{v}}$ to each vertex in $V$.
We say that $\mathbf{T}_{v}$ is an alternating valuation if the
following holds:
•
Every leaf of $\mathbf{T}_{v}$ is at the same depth $d$.
•
All vertices at a given depth, i.e., in the same level, have the same
out-degree.
•
If two vertices $a,b\in V$ are siblings, then $B^{\mathbf{T}_{v}}(a)\neq B^{\mathbf{T}_{v}}(b)$.
Let $\varphi\equiv\exists\bar{x}_{1}\forall\bar{x}_{2}\ldots Q\bar{x}_{k}(\psi)$ be a formula in $\mathrm{QBF}_{k}$, where $Q$ is “$\exists$” if $k$ is odd and
“$\forall$” if $k$ is even, and let $l_{j}$ for $1\leq j\leq k$ be the
length of the $j$-th alternating block of quantifiers. We say that an
alternating valuation $\mathrm{T}_{v}$ is applicable to $\varphi$, if the
depth of $\mathrm{T}_{v}$ is $l_{1}+\cdots+l_{k}-1$ and for every $1\leq i\leq l_{1}+\cdots+l_{k}$, it holds that:
•
All vertices at depth $i-1$ have no siblings if $1\leq i\leq l_{1}$ or
$l_{1}+l_{2}+1\leq i\leq l_{1}+l_{2}+l_{3}$ or $\cdots$ or $l_{1}+l_{2}+\cdots+l_{k^{\prime}-1}+1\leq i\leq l_{1}+l_{2}+\cdots+l_{k^{%
\prime}}$, where $k^{\prime}=k$ if the $k$-th
block of quantifiers is existential and $k^{\prime}=k-1$ otherwise.
•
All vertices at depth $i-1$ have exactly one sibling if $l_{1}+1\leq i\leq l_{1}+l_{2}$ or $l_{1}+l_{2}+l_{3}+1\leq i\leq l_{1}+l_{2}+l_{3}+l_{4}$ or
$\cdots$ or $l_{1}+l_{2}+\cdots+l_{k^{\prime\prime}-1}+1\leq i\leq l_{1}+l_{2}+\cdots+l_{k^%
{\prime\prime}}$, where $k^{\prime\prime}=k$ if the $k$-th
block of quantifiers is universal and $k^{\prime\prime}=k-1$ otherwise.
Let $\gamma=\exists\bar{x}_{1}\forall\bar{x}_{2}\ldots Q\bar{x}_{k}(\varphi)$ be a formula in $\mathrm{QBF}_{k}$ over $X$, and let $\mathbf{T}_{v}$ be an alternating
valuation applicable to $\gamma$. A leaf valuation $\mathbf{L}_{v}$ is a linear
subgraph of $\mathbf{T}_{v}$ of vocabulary $\sigma_{\mathbf{T}_{v}}$ which corresponds
to a path from the root to a leaf in $\mathbf{T}_{v}$. Let $v$ be a mapping from
the set of variables $X$ to $\{0,1\}$, i.e., a Boolean assignment, such that
for $x_{i}\in X$ the $i$-th variable in the prefix of quantifiers
of $\gamma$, it holds that $v(x_{i})=1$ iff $B^{\mathbf{L}_{v}}(n_{i})=1^{\mathbf{L}_{v}}$ for $n_{i}$ the $i$-th node in the linear
order induced by $E^{\mathbf{L}_{v}}$. We say that $\mathbf{L}_{v}$ satisfies
$\gamma$,
written $\mathbf{L}_{v}\models\gamma$, if the Boolean assignment $v$
satisfies $\varphi$. That is, if $\varphi$ is a Boolean variable $x_{i}$ in $X$, then $\mathbf{L}_{v}\models\varphi$ if $v(x_{i})=1$; if $\varphi=\neg(\psi)$, then $\mathbf{L}_{v}\models\varphi$ if $\mathbf{L}_{v}\not\models\psi$ (i.e., if it is not the
case that $\mathbf{L}_{v}\models\psi$); if $\varphi=(\psi\vee\alpha)$, then
$\mathbf{L}_{v}\models\varphi$ if either $\mathbf{L}_{v}\models\psi$ or
$\mathbf{L}_{v}\models\alpha$; and if $\varphi=(\psi\wedge\alpha)$, then
$\mathbf{L}_{v}\models\varphi$ if both $\mathbf{L}_{v}\models\psi$ and
$\mathbf{L}_{v}\models\alpha$.
Finally, we say that the alternating valuation $\mathbf{T}_{v}$ satisfies
$\gamma$ if every leaf valuation $\mathbf{L}_{v}$ of $\mathbf{T}_{v}$ satisfies
$\gamma$.
A Boolean formulae $\varphi$ in $\mathrm{QBF}_{k}$ is satisfiable if and only if
there is an alternating valuation $\mathbf{T}_{v}$ which satisfies $\varphi$;
otherwise $\varphi$ is unsatisfiable. $\mathrm{SATQBF}_{k}$ is the set of
$\mathrm{QBF}_{k}$ formulae that are satisfiable. $\mathrm{SATQBF}=\bigcup_{k\geq 1}\mathrm{SATQBF}_{k}$.
It is well known that $\mathrm{SATQBF}_{k}$ is complete for the level
$\Sigma^{p}_{k}$ of the polynomial-time hierarchy (see [6, 1] among
others sources). It is also well known that second-order logic captures the
polynomial-time hierarchy. In fact, there is an exact correspondence between
the prenex fragments of second-order logic with up to $k$ alternations of
quantifiers $\Sigma^{1}_{k}$ and the levels $\Sigma^{p}_{k}$ of the polynomial time
hierarchy [13]. Thus, for every $k$, $\mathrm{SATQBF}_{k}$ can be
defined in second-order logic, in fact, it can even be defined in
$\Sigma^{1}_{k}$. Regarding $\mathrm{SATQBF}$, we note that it is
$\mathrm{PSPACE}$-complete [13]. Since existential third-order logic
captures $\mathrm{NTIME}(2^{n^{{\cal O}(1)}})$ (see [7]) and
$\mathrm{PSPACE}\subseteq\mathrm{DTIME}(2^{n^{{\cal O}(1)}})\subseteq\mathrm{%
NTIME}(2^{n^{{\cal O}(1)}})$, we know that $\mathrm{SATQBF}$ can be
defined in existential third-order logic. In the following sections we present
a second-order formula that defines $\mathrm{SATQBF_{k}}$ and a
third-order formula that defines $\mathrm{SATQBF}$, respectively.
6 $\mathrm{SATQBF_{k}}$ in Second-Order Logic
Following a top-down approach, we present a detailed construction of a second-order formula
that defines $\mathrm{SATQBF}_{k}$. But first, we need to fix an encoding of
quantified Boolean formulae as relational structures.
There is a well-known correspondence between words and finite
structures. Let $A$ be a finite alphabet and let $\pi(A)$ be
the vocabulary $\{\leq\}\cup\{R_{a}:a\in A\}$, where $\leq$ is a binary
relation symbol and the $R_{a}$ are unary relation symbols. We can identify any word $v=a_{1}\ldots a_{n}$ in $A^{*}$ with a $\pi(A)$-structure
${\bf B}$, where the cardinality of $B$ equals the length of $v$,
$\leq^{\bf B}$ is a total order on ${\bf B}$, and, for each $R_{a}\in\pi(A)$,
$R_{a}^{\bf B}$ contains the positions in $v$ carrying an $a$,
$R_{a}^{\bf B}=\{b\in B:\textrm{for some $j$ $(1\leq j\leq n)$,}$
$\textrm{$b$ is the $j$-th element in the order $\leq^{\bf B}$ and $a_{j}=a$}\}$
Such structures are usually known as word models for $v$ ([3]).
As any two word models for $v$ are isomorphic, we can speak of the
word model for $v$.
Note that we can represent Boolean variables of the form $x_{n}$ by using a
symbol “$X$” followed
by a sequence of $n$ symbols “$|$”. For instance, we can write $X|||$ for
$x_{3}$. Thus using word models,
every quantified Boolean formula $\varphi$ can be viewed as a finite
relational structure $G_{\varphi}$ of the
following vocabulary.
$$\pi=\{\leq,P_{\neg},P_{\vee},P_{\wedge},P_{\exists},P_{\forall},P_{(},P_{)},P_%
{X},P_{|}\}$$
Example 3
If $\varphi$ is the quantified Boolean formula $\exists x_{1}\forall x_{2}((\neg x_{1})\vee x_{2})$, which using our notation for the variables
corresponds to $\exists X|\forall X||((\neg X|)\vee X||)$,
then the following $\pi$-structure ${\bf G}_{\varphi}$ (note that ${\bf G}_{\varphi}$ is a linear graph)
where $G_{\varphi}=\{1,\ldots,18\}$, $\leq^{{\bf G}_{\varphi}}$ is a total order on ${\bf G}_{\varphi}$,
$P^{{\bf G}_{\varphi}}_{\neg}=\{10\}$,
$P^{{\bf G}_{\varphi}}_{\vee}=\{14\}$, $P^{{\bf G}_{\varphi}}_{\wedge}=\emptyset$,
$P^{{\bf G}_{\varphi}}_{\exists}=\{1\}$, $P^{{\bf G}_{\varphi}}_{\forall}=\{4\}$, $P^{{\bf G}_{\varphi}}_{(}=\{8,9\}$, $P^{{\bf G}_{\varphi}}_{)}=\{13,18\}$, $P^{{\bf G}_{\varphi}}_{X}=\{2,5,11,15\}$, $P^{{\bf G}_{\varphi}}_{|}=\{3,6,7,12,16,17\}$, encodes $\varphi$.
We show next how to build a second-order logic formula $\varphi_{\mathrm{SATQBF}_{k}}$
such that, given a relational structure ${\bf G}_{\varphi}$ of vocabulary $\pi$, it
holds that ${\bf G}_{\varphi}\models\varphi_{\mathrm{SATQBF}_{k}}$ iff the
quantified Boolean formula $\varphi$ represented by ${\bf G}_{\varphi}$, is
satisfiable. That is, we show next how to build a second-order formula
$\varphi_{\mathrm{SATQBF}_{k}}$ of vocabulary $\pi$ that defines
$\mathrm{SATQBF}_{k}$. As mentioned earlier, we follow a top-down approach for the
construction of this formula. At the highest level of abstraction,
we can think of $\varphi_{\mathrm{SATQBF}_{k}}$ as a second-order formula that
expresses the following:
“There is an alternating valuation $${\bf T}_{v}$$ applicable to $$\varphi$$ that satisfies
$$\varphi$$”.
(A)
Recall that an alternating valuation ${\bf T}_{v}$ satisfies $\varphi$ iff every
leaf valuation ${\bf L}_{v}$ of ${\bf T}_{v}$ satisfies the quantifier-free part
$\varphi^{\prime}$ of $\varphi$. Also recall that every leaf
valuation ${\bf T}_{v}$ corresponds to a Boolean assignment $v$. Thus, if
$\varphi=\exists\bar{x}_{1}\forall\bar{x}_{2}\ldots Q\bar{x}_{k}(\varphi^{%
\prime})$, where for $1\leq i\leq k$, $\bar{x}_{i}=(x_{i1},\ldots,x_{il_{i}})$, $Q$ is “$\exists$” if $k$ is odd and “$\forall$” if $k$ is
even, $X_{1},\ldots,X_{k}$ are the set of variables
in $\bar{x}_{1},\ldots,\bar{x}_{k}$, respectively, and $\varphi^{\prime}$ is a
(quantifier free) Boolean formulae
over $X=X_{1}\cup\cdots\cup X_{k}$, then the expression in (A) can be divided in two parts:
AVS1
(Alternating Valuation that Satisfies $\varphi$, Part 1) which expresses
“There is a partial Boolean assignment $v_{1}$ on $X_{1}$,
such that for all partial Boolean assignments $v_{2}$ on $X_{2}$,
…,
there is (or “for all” if k is even) a partial Boolean assignment $v_{k}$ on $X_{k}$”.
AVS2
(Alternating Valuation that Satisfies $\varphi$, Part 2) which expresses
“The Boolean assignment $v=v_{1}\cup v_{2}\cup\cdots\cup v_{k}$ satisfies the (quantifier free) Boolean formula $\varphi^{\prime}$”.
For each partial Boolean assignment $v_{i}$ ($1\leq i\leq k$), we use
a second-order variable $V_{i}$ of arity one and two second-order variables
$E_{i}$ and $B_{i}$ of arity two, to store the encoding of each $v_{i}$ as a linear graph
$G_{i}=(V_{i},E_{i})$ with an associated function $B_{i}:V_{i}\rightarrow\{0,1\}$ (see Figure 6.1).
Correspondingly, we use a second-order variable $V_{t}$ of arity one and two
second-order variables $E_{t}$ and $B_{t}$ of arity two, to store the encoding of
each Boolean assignment $v$ (leaf valuation ${\bf T}_{v}$) as a linear graph
$G_{t}=(V_{t},E_{t})$ with an associated function $B_{t}:V_{t}\rightarrow\{0,1\}$. Figure 6.2 illustrates an alternating
valuation applicable to $\varphi$ and its corresponding encoding.
In the next subsection we describe the process followed to build
a second-order formula to express Statement AVS1. Then we describe in
Subsection 6.2, the corresponding process for Statement AVS2.
6.1 Expressing Statement AVS1
Let $k_{\exists}$ and $k_{\forall}$ be the index of the last existential
quantifier block and
the last universal quantifier block, respectively, in the prefix of $k$ blocks
of quantifiers of $\varphi$. We can express Statement AVS1 as follows:
$\exists V_{1}E_{1}B_{1}\forall V_{2}E_{2}B_{2}\cdots Q_{k}V_{k}E_{k}B_{k}%
\exists V_{t}E_{t}B_{t}U_{1},U_{2},\ldots,U_{k}\Big{(}\mathrm{A1}\wedge\mathrm%
{A2}\wedge\mathrm{A3}\wedge\mathrm{A4}\wedge\mathrm{A5}\wedge$
$\big{(}\big{(}\mathrm{A6}\wedge\mathrm{A7}\wedge\mathrm{A8}\wedge\mathrm{A9}%
\wedge\mathrm{A10}\wedge\mathrm{A11}\big{)}\rightarrow\mathrm{AVS2}\big{)}\Big%
{)}$ where
•
A1 expresses “$G_{t}=(V_{t},E_{t})$ is a linear graph”.
•
A2 expresses “The length of $G_{t}$ equals the number of variables that appear in the prefix of quantifiers of $\varphi$”.
•
A3 expresses “$G_{1}=(V_{1},E_{1}),G_{3}=(V_{2},E_{2}),\ldots,G_{k_{\exists}}=(V_{k_{\exists}%
},L_{k_{\exists}})$ are linear graphs”.
•
A4 expresses “$B_{1}:V_{1}\rightarrow\{0,1\}$, $B_{3}:V_{3}\rightarrow\{0,1\}$, …, $B_{k_{\exists}}:V_{k_{\exists}}\rightarrow\{0,1\}$ are total functions”.
•
A5 expresses “The lengths of the linear graphs $G_{1},G_{3},\ldots,G_{k_{\exists}}$ equal the lengths of their corresponding blocks of quantifiers in $\varphi$”.
•
A6 expresses “$V_{1},V_{2},\ldots,V_{k}$ are pairwise disjoint sets”.
•
A7 expresses “$G_{2}=(V_{2},E_{2}),G_{4}=(V_{4},E_{4}),\ldots,G_{k_{\forall}}=(V_{k_{\forall}%
},L_{k_{\forall}})$ are linear graphs”
•
A8 expresses “$B_{2}:V_{2}\rightarrow\{0,1\}$, $B_{4}:V_{4}\rightarrow\{0,1\}$, …, $B_{k_{\forall}}:V_{k_{\forall}}\rightarrow\{0,1\}$ are total functions”
•
A9 expresses “The lengths of the linear graphs $G_{2},G_{4},\ldots,G_{k_{\forall}}$ equal the lengths of their corresponding blocks of quantifiers in $\varphi$”.
•
A10 expresses “$U_{1}$ is a total injection from $G_{1}$ to the first part of $G_{t}$ and $U_{2}$ is a total injection from $G_{2}$ to the second part of $G_{t}$ …and $U_{k}$ is a total injection from $G_{k}$ to the $k$-th part of $G_{t}$”.
•
A11 expresses “$B_{t}:V_{t}\rightarrow\{0,1\}$ is a total function that coincides with $B_{1}$, $B_{2}$, $\ldots$, $B_{k}$”.
•
AVS2 expresses Statement AVS2 as described in Subsection 6.2.
Next, we discuss how to write the sub-formulae A1–A11 in second-order logic.
A1.
This is expressed by the auxiliary formula $\mathrm{LINEAR}(V_{t},E_{t})$, which is defined
in Subsection 6.3 below.
A2.
This is implied by the following statement which is expressed in further
detail in Subsection 6.2.1 (A).
“There is a partial surjective injection $V_{p}$ from the quantifier prefix of
$\varphi$ to $G_{t}$, which maps every $X$ in the prefix to its
corresponding node in $G_{t}$, and which preserves $\leq^{{\bf G}_{\varphi}}$ and
$E_{t}$”.
A3.
$\mathrm{LINEAR}(V_{1},E_{1})\wedge\mathrm{LINEAR}(V_{3},E_{3})\wedge\cdots%
\wedge\mathrm{LINEAR}(V_{k_{\exists}},E_{k_{\exists}})$, where the sub-formulae $\mathrm{LINEAR}(V_{i},E_{i})$
are as defined in Subsection 6.3.
A4.
$\forall t,p,p^{\prime}\Big{(}\bigwedge_{i=1,3,\ldots,k_{\exists}}\big{(}%
\mathrm{A4.1}\wedge\mathrm{A4.2}\wedge\mathrm{A4.3}\big{)}\Big{)}$
•
A4.1 expresses “$B_{i}$ is a function”.
$\mathrm{A4.1}\equiv((B_{i}(t,p)\wedge B_{i}(t,p^{\prime}))\rightarrow p=p^{%
\prime})$
•
A4.2 expresses “$B_{i}$ is total”.
$\mathrm{A4.2}\equiv(V_{i}(t)\rightarrow\exists p(B_{i}(t,p)))$
•
A4.3 expresses “the range of $B_{i}$ is $\{0,1\}$”.
$\mathrm{A4.3}\equiv(B_{i}(t,p)\rightarrow(\text{$p=1$ $\vee$ $p=0$}))$
where $p=0$ and $p=1$ have the obvious meaning and are defined in
Subsection 6.3.
A5.
If $k_{\exists}\neq k$, then
$\bigwedge_{1,3,\ldots,k_{\exists}}\big{(}\exists L^{\prime}v_{1}v_{2}\ldots v_%
{k_{\exists}}v_{k_{\exists}+1}(\alpha_{k_{\exists}}\wedge\zeta_{i})\big{)}$
where $\alpha_{k_{\exists}}$ is the formula template $\alpha_{i}$ instantiated
with $i=k_{\exists}$.
If $k_{\exists}=k$, then
$\big{(}\bigwedge_{1,3,\ldots,k_{\exists}-2}\big{(}\exists L^{\prime}v_{1}v_{2}%
\ldots v_{k_{\exists}-1}(\alpha_{k_{\exists-2}}\wedge\zeta_{i})\big{)}\big{)}%
\wedge\exists L^{\prime}v_{1}v_{2}\ldots v_{k}v_{e}(\beta_{1}\wedge\beta_{2}%
\wedge\beta_{3})$
where $\alpha_{k_{\exists}-2}$ is the formula template $\alpha_{i}$ instantiated
with $i=k_{\exists}-2$ (Note that $k_{\exists}-2$ is the previous to the last
existential block, and the subformulae $\beta_{1},\beta_{2}$ and $\beta_{3}$ take
care of the last block of quantifiers).
Next, we define the subformulae $\alpha_{i}$, $\beta_{1}$ $\zeta_{i}$, $\beta_{2}$ and $\beta_{3}$ in the listed order. For their definitions we use an auxiliary formula $\mathrm{PATH}_{\leq}(x,y)$ which is in turn defined in Subsection 6.3 below, and which expresses “the pair $(x,y)$ is in the transitive closure of the relation $\leq$”.
The subformula $\alpha_{i}$ is satisfied if, for $1\leq j\leq i$, $v_{j}$ is the position of the first quantifier of
the $j$-th block (when $i$ is not the last block of quantifiers).
$\alpha_{i}\equiv\big{(}P_{\exists}(v_{1})\wedge P_{\forall}(v_{2})\wedge\ldots%
\wedge P_{Q}(v_{i+1})\wedge\neg\exists x(x\neq v_{1}\wedge x\leq v_{1})\wedge$
$\mathrm{PATH}_{\leq}(v_{1},v_{2})\wedge\mathrm{PATH}_{\leq}(v_{2},v_{3})\wedge%
\cdots\wedge\mathrm{PATH}_{\leq}(v_{i},v_{i+1})\wedge$
$\neg\exists x(\mathrm{PATH}_{\leq}(v_{1},x)\wedge\mathrm{PATH}_{\leq}(x,v_{2})%
\wedge x\neq v_{1}\wedge x\neq v_{2}\wedge P_{\forall}(x))\wedge$
$\neg\exists x(\mathrm{PATH}_{\leq}(v_{2},x)\wedge\mathrm{PATH}_{\leq}(x,v_{3})%
\wedge x\neq v_{2}\wedge x\neq v_{3}\wedge P_{\exists}(x))\wedge$
$\ldots\wedge$
$\neg\exists x(\mathrm{PATH}_{\leq}(v_{i},x)\wedge\mathrm{PATH}_{\leq}(x,v_{i+1%
})\wedge x\neq v_{i}\wedge x\neq v_{i+1}\wedge P_{Q}(x))\big{)}$
where $P_{Q}$ is $P_{\forall}$ if $i$ is odd or $P_{\exists}$ if $i$ is even.
The subformula $\beta_{1}$ is satisfied if, for $1\leq j\leq i$, $v_{j}$ is the position of the first quantifier of
the $j$-th block.
$\beta_{1}\equiv\big{(}P_{\exists}(v_{1})\wedge P_{\forall}(v_{2})\wedge\ldots%
\wedge\mathrm{P1}(v_{k})\wedge P_{|}(v_{e})\wedge\neg\exists x(x\leq v_{1})\wedge$
$\mathrm{PATH}_{\leq}(v_{1},v_{2})\wedge\mathrm{PATH}_{\leq}(v_{2},v_{3})\wedge%
\cdots\wedge\mathrm{PATH}_{\leq}(v_{k},v_{e})\wedge$
$\neg\exists x(\mathrm{PATH}_{\leq}(v_{1},x)\wedge\mathrm{PATH}_{\leq}(x,v_{2})%
\wedge x\neq v_{1}\wedge x\neq v_{2}\wedge P_{\forall}(x))\wedge$
$\neg\exists x(\mathrm{PATH}_{\leq}(v_{2},x)\wedge\mathrm{PATH}_{\leq}(x,v_{3})%
\wedge x\neq v_{2}\wedge x\neq v_{3}\wedge P_{\exists}(x))\wedge$
$\ldots\wedge$
$\neg\exists x(\mathrm{PATH}_{\leq}(v_{k},x)\wedge\mathrm{PATH}_{\leq}(x,v_{e})%
\wedge x\neq v_{k}\wedge x\neq v_{e}\wedge\mathrm{P2}(x))\big{)}$
where $\mathrm{P1}$ is $P_{\exists}$ if $k$ is odd or $P_{\forall}$ if $k$ is even, and $\mathrm{P2}$ is $P_{\forall}$ if $k$ is
odd or $P_{\exists}$ if $k$ is even.
When $i$ is not the index of the last block of quantifiers, the subformula $\zeta_{i}$ is satisfied if $L^{\prime}$ is a bijection from the
indices of the symbols $X$ in the $i$-th
alternating block of quantifiers to $V_{i}$, which preserves $E_{i}$ and
$\mathrm{Next}_{X}=\{(a,b)\in\leq^{{\bf G}_{\varphi}}\,|\,\text{$a$ and $b$
are indices of symbols in the $i$-th block}\wedge P_{X}(a)\wedge P_{X}(b)%
\wedge\forall c((a\leq c\wedge c\leq b)\rightarrow\neg P_{X}(c))\}$ (i.e., the
order of appearance of the $X$’s in the $i$-th block of quantifiers in the prefix of
$\varphi$). This is illustrated in Figure 6.3. Recall that we encode in $G_{i}=(V_{i},E_{i},B_{i})$ a partial truth
assignment for the variables in the $i$-th alternating block of
quantifiers.
$\zeta_{i}\equiv\big{(}\mathrm{A5.1}\wedge\mathrm{A5.2}\wedge\mathrm{A5.3}%
\wedge\mathrm{A5.4}\wedge\mathrm{A5.5}\big{)}$ where
•
A5.1 defines the “domain of $L^{\prime}$”.
$\mathrm{A5.1}\equiv\forall x\big{(}(\mathrm{PATH}_{\leq}(v_{i},x)\wedge\mathrm%
{PATH}_{\leq}(x,v_{i+1})\wedge x\neq v_{i+1}\wedge P_{X}(x))\leftrightarrow%
\exists y(L^{\prime}(x,y))\big{)}$
•
A5.2 expresses “$L^{\prime}$ is surjective”.
$\mathrm{A5.2}\equiv\forall y\big{(}V_{i}(y)\rightarrow\exists z(L^{\prime}(z,y%
))\big{)}$
•
A5.3 expresses “$L^{\prime}$ preserves $\mathrm{Next}_{X}$ and $E_{i}$” which implies injectivity.
$\mathrm{A5.3}\equiv\forall sts^{\prime}t^{\prime}\Big{(}\big{(}L^{\prime}(s,t)%
\wedge L^{\prime}(s^{\prime},t^{\prime})\wedge s\neq s^{\prime}\wedge\mathrm{%
PATH}_{\leq}(v_{i},s)\wedge\mathrm{PATH}_{\leq}(s^{\prime},v_{i+1})$
$\wedge\mathrm{PATH}_{\leq}(s,s^{\prime})\wedge\neg\exists z(\mathrm{PATH}_{%
\leq}(s,z)\wedge\mathrm{PATH}_{\leq}(z,s^{\prime})\wedge$
$z\neq s\wedge z\neq s^{\prime}\wedge P_{X}(z))\big{)}\rightarrow E_{i}(t,t^{%
\prime})\Big{)}$
•
A5.4 defines the “range of $L^{\prime}$”.
$\mathrm{A5.4}\equiv\forall xy(L^{\prime}(x,y)\rightarrow V_{i}(y))$
•
A5.5 expresses “$L^{\prime}$ is a function”.
$\mathrm{A5.5}\equiv\forall xyz\big{(}(L^{\prime}(x,y)\wedge L^{\prime}(x,z))%
\rightarrow y=z\big{)}$
The subformula $\beta_{2}$ is satisfied if $L^{\prime}$ is a bijection from the
indices of the symbols $X$ in the $k$-th
alternating block of quantifiers to $V_{k}$, which preserves $E_{k}$ and
$\mathrm{Next}_{X}$ (i.e., the
order of appearance of the $X$’s in the $k$-th block of quantifiers in the prefix of
$\varphi$).
$\beta_{2}\equiv\big{(}\mathrm{A5.1^{\prime}}\wedge\mathrm{A5.2^{\prime}}\wedge%
\mathrm{A5.3^{\prime}}\wedge\mathrm{A5.4^{\prime}}\wedge\mathrm{A5.5^{\prime}}%
\big{)}$ where
•
$\mathrm{A5.1^{\prime}}$ defines the “domain of $L^{\prime}$”.
$\mathrm{A5.1^{\prime}}\equiv\forall x\big{(}(\mathrm{PATH}_{\leq}(v_{k},x)%
\wedge\mathrm{PATH}_{\leq}(x,v_{e})\wedge P_{X}(x))\leftrightarrow\exists y(L^%
{\prime}(x,y))\big{)}$
•
$\mathrm{A5.2^{\prime}}$ expresses “$L^{\prime}$ is surjective”.
$\mathrm{A5.2^{\prime}}\equiv\forall y\big{(}V_{k}(y)\rightarrow\exists z(L^{%
\prime}(z,y))\big{)}$
•
$\mathrm{A5.3^{\prime}}$ expresses “$L^{\prime}$ preserves $\mathrm{Next}_{X}$ and $E_{k}$” which implies injectivity.
$\mathrm{A5.3^{\prime}}\equiv\forall sts^{\prime}t^{\prime}\Big{(}\big{(}L^{%
\prime}(s,t)\wedge L^{\prime}(s^{\prime},t^{\prime})\wedge s\neq s^{\prime}%
\wedge\mathrm{PATH}_{\leq}(v_{k},s)\wedge\mathrm{PATH}_{\leq}(s^{\prime},v_{e})$
$\wedge\mathrm{PATH}_{\leq}(s,s^{\prime})\wedge\neg\exists z(\mathrm{PATH}_{%
\leq}(s,z)\wedge\mathrm{PATH}_{\leq}(z,s^{\prime})\wedge$
$z\neq s\wedge z\neq s^{\prime}\wedge P_{X}(z))\big{)}\rightarrow E_{k}(t,t^{%
\prime})\Big{)}$
•
$\mathrm{A5.4^{\prime}}$ defines the “range of $L^{\prime}$”.
$\mathrm{A5.4^{\prime}}\equiv\forall xy(L^{\prime}(x,y)\rightarrow V_{k}(y))$
•
$\mathrm{A5.5^{\prime}}$ expresses “$L^{\prime}$ is a function”.
$\mathrm{A5.5^{\prime}}\equiv\forall xyz\big{(}(L^{\prime}(x,y)\wedge L^{\prime%
}(x,z))\rightarrow y=z\big{)}$
The last subformula $\beta_{3}$ is satisfied if $v_{e}$ is the last symbol “$|$”
in the prefix of quantifiers of $\varphi$. We use $\mathrm{SUC}_{\leq}(x,y)$ to
denote that $x$ is the immediate successor of $y$ in the total order $\leq^{{\bf G}_{\varphi}}$. The formula that expresses $\mathrm{SUC}_{\leq}(x,y)$ is defined in Subsection 6.3.
$\beta_{3}\equiv\Big{(}\forall v^{\prime}\big{(}\mathrm{SUC}_{\leq}(v_{e},v^{%
\prime})\rightarrow\neg P_{|}(v^{\prime})\big{)}\wedge P_{|}(v_{e})\wedge$
$\forall v^{\prime}\big{(}\mathrm{PATH}_{\leq}(v_{e},v^{\prime})\rightarrow(%
\neg P_{\exists}(v^{\prime})\wedge\neg P_{\forall}(v^{\prime}))\big{)}\wedge$
$\exists xyw\forall v^{\prime}\big{(}P_{X}(x)\wedge P_{Q}(w)\wedge\mathrm{SUC}_%
{\leq}(x,y)\wedge\mathrm{SUC}_{\leq}(w,x)\wedge\mathrm{PATH}_{\leq}(y,v_{e})\wedge$
$((\mathrm{PATH}_{\leq}(v^{\prime},v_{e})\wedge\mathrm{PATH}_{\leq}(y,v^{\prime%
}))\rightarrow P_{|}(v^{\prime}))\big{)}\Big{)}$
where $P_{Q}$ is $P_{\exists}$ if $k$ is odd, or $P_{\forall}$ if $k$ is even.
A6.
Let $V_{i}\cap V_{j}=\emptyset$ denote $\forall x\big{(}(V_{i}(x)\rightarrow\neg V_{j}(x))\wedge(V_{j}(x)\rightarrow%
\neg V_{i}(x))\big{)}$, we can express that $V_{1},V_{2},\ldots,V_{k}$ are pairwise disjoint sets as follows.
$(V_{1}\cap V_{2}=\emptyset)\wedge(V_{1}\cap V_{3}=\emptyset)\wedge(V_{1}\cap V%
_{4}=\emptyset)\wedge\cdots\wedge(V_{1}\cap V_{k}=\emptyset)\wedge$
$(V_{2}\cap V_{3}=\emptyset)\wedge(V_{2}\cap V_{4}=\emptyset)\wedge\cdots\wedge%
(V_{2}\cap V_{k}=\emptyset)\wedge$
$\ldots\wedge(V_{k-1}\cap V_{k}=\emptyset)$
A7.
$\mathrm{LINEAR}(V_{2},E_{2})\wedge\mathrm{LINEAR}(V_{4},E_{4})\wedge\cdots%
\wedge\mathrm{LINEAR}(V_{k_{\forall}},E_{k_{\forall}})$,
where $\mathrm{LINEAR}(V_{i},E_{i})$
is as defined in Subsection 6.3.
A8.
$\forall t,p,p^{\prime}\Big{(}\bigwedge_{i=2,4,\ldots,k_{\forall}}\big{(}%
\mathrm{A8.1}\wedge\mathrm{A8.2}\wedge\mathrm{A8.3}\big{)}\Big{)}$ where
•
A8.1 expresses “$B_{i}$ is a function”.
$\mathrm{A8.1}\equiv((B_{i}(t,p)\wedge B_{i}(t,p^{\prime}))\rightarrow p=p^{%
\prime})$
•
A8.2 expresses “$B_{i}$ is total”.
$\mathrm{A8.2}\equiv(V_{i}(t)\rightarrow\exists p(B_{i}(t,p)))$
•
A8.3 expresses “the range of $B_{i}$ is $\{0,1\}$”.
$\mathrm{A8.3}\equiv(B_{i}(t,p)\rightarrow(p=1\vee p=0))$
where $p=0$ and $p=1$ have the obvious meaning and are defined in
Subsection 6.3.
A9.
If $k_{\forall}\neq k$, then
$\bigwedge_{2,4,\ldots,k_{\forall}}\big{(}\exists L^{\prime}v_{1}v_{2}\ldots v_%
{k_{\forall}}v_{k_{\forall}+1}(\alpha_{k_{\forall}}\wedge\zeta_{i})\big{)}$
where $\alpha_{k_{\forall}}$ is the formula template $\alpha_{i}$ instantiated
with $i=k_{\forall}$.
If $k_{\forall}=k$, then
$\big{(}\bigwedge_{2,4,\ldots,k_{\forall}-2}\big{(}\exists L^{\prime}v_{1}v_{2}%
\ldots v_{k_{\forall}-1}(\alpha_{k_{\forall-2}}\wedge\zeta_{i})\big{)}\big{)}%
\wedge\exists L^{\prime}v_{1}v_{2}\ldots v_{k}v_{e}(\beta_{1}\wedge\beta_{2}%
\wedge\beta_{3})$
where $\alpha_{k_{\forall}-2}$ is the formula template $\alpha_{i}$ instantiated
with $i=k_{\forall}-2$ (Note that $k_{\forall}-2$ is the previous to the last
universal block, and the subformulae $\beta_{1},\beta_{2}$ and $\beta_{3}$ take
care of the last block of quantifiers).
The subformulae $\alpha_{i}$, $\zeta_{i}$, $\beta_{1}$, $\beta_{2}$ and $\beta_{3}$ are
the same as in (A5).
A10.
$\Big{(}\mathrm{A10.1}\wedge\bigwedge_{2\leq i\leq k-1}\big{(}\mathrm{A10.2.i}%
\big{)}\wedge\mathrm{A10.3}\Big{)}$ where
•
A10.1 expresses “$U_{1}$ is a total injection from $V_{1}$ to $V_{t}$ such that: (a) preserves $E_{1}$ and $E_{t}$ and (b) $U_{1}($“first node in
the order $E_{1}$”$)=$“first node in the order $E_{t}$””.
•
A10.2.i expresses “$U_{i}$ is a total injection from $V_{i}$ to $V_{t}$ such that: (a) preserves $E_{i}$ and $E_{t}$ and (b) $U_{i}($“first node in the order $E_{i}$”$)=\mathrm{SUC}_{E_{t}}(U_{i-1}($“last node in the order $E_{i-1}$”$))$”.
•
A10.3 expresses “$U_{k}$ is a total injection from $V_{k}$ to $V_{t}$ such that: (a) preserves $E_{k}$ and $E_{t}$ and (b) $U_{k}($“first node in order $E_{k}$”$)=\mathrm{SUC}_{E_{t}}(U_{k-1}($“last node in order $E_{k-1}$”$))$”.
We describe next the second-order formula for A10.3 which is in turn illustrated in
Figure 6.4. Note that the node labeled $x$ in Figure 6.4 corresponds to the last node in the linear graph $G_{k-1}$ and that $x$ is mapped by the function $U_{k-1}$ to the node labelled $y$ in the linear graph $G_{t}$. Accordingly, $U_{k}$ maps the first node in the linear graph $G_{k}$ (i.e. the node labeled $u$), to the successor of node $y$ in $G_{t}$ (i.e. to the node labelled $t$).
$\mathrm{A10.3}\equiv\forall xytu\big{(}\mathrm{A10.3.1}\wedge\mathrm{A10.3.2}%
\wedge\mathrm{A10.3.3}\wedge\mathrm{A10.3.4}\big{)}$ where
•
A10.3.1 expresses “$U_{k}$ is a total injection from $V_{k}$ to $V_{t}$”.
$\mathrm{A10.3.1}\equiv((U_{k}(x,y)\wedge U_{k}(x,t))\rightarrow y=t)\wedge$
$((U_{k}(x,y)\wedge U_{k}(u,y))\rightarrow x=u)\wedge$
$(V_{k}(x)\rightarrow\exists y(U_{k}(x,y)))\wedge$
$(U_{k}(x,y)\rightarrow(V_{k}(x)\wedge V_{t}(y)))$
•
A10.3.2 expresses “preserves $E_{t}$”.
$\mathrm{A10.3.2}\equiv((U_{k}(x,y)\wedge U_{k}(u,t)\wedge E_{t}(y,t))%
\rightarrow E_{k}(x,u))$
•
A10.3.3 expresses “preserves $E_{k}$”.
$\mathrm{A10.3.3}\equiv((U_{k}(x,y)\wedge U_{k}(u,t)\wedge E_{k}(x,u))%
\rightarrow E_{t}(y,t))$
•
A10.3.4 expresses “$U_{k}($“first node in order $E_{k}$”$)=\mathrm{SUC}_{E_{t}}(U_{k-1}($“last node in order $E_{k-1}$”$))$”.
$\mathrm{A10.3.4}\equiv\big{(}\big{(}U_{k-1}(x,y)\wedge\neg\exists v(E_{k-1}(x,%
v))\wedge E_{t}(y,t)\wedge\neg\exists v(E_{k}(v,u)\wedge V_{k}(u))\big{)}$
$\rightarrow U_{k}(u,t)\big{)}$
A11.
$\forall xytpp^{\prime}\big{(}(B_{1}(t,p)\wedge U_{1}(t,y)\wedge B_{t}(y,p^{%
\prime}))\rightarrow p=p^{\prime}\big{)}\wedge$
$\forall xytpp^{\prime}\big{(}(B_{2}(t,p)\wedge U_{2}(t,y)\wedge B_{t}(y,p^{%
\prime}))\rightarrow p=p^{\prime}\big{)}\wedge$
$\ldots\wedge$
$\forall xytpp^{\prime}\big{(}(B_{k}(t,p)\wedge U_{k}(t,y)\wedge B_{t}(y,p^{%
\prime}))\rightarrow p=p^{\prime}\big{)}$
6.2 Expressing Statement AVS2
Statement AVS2 can be rephrased as follows:
$\exists\,V_{p}\,C\,E_{C}\,\mathit{ST}\,E_{\mathit{ST}}\,M\,C_{\wedge}\,C_{\vee%
}\,C_{\neg}\,C_{(}\,C_{)}\,C_{1}\,C_{0}\,H_{\phi}\,\big{(}\mathrm{AVS2.1}%
\wedge\mathrm{AVS2.2}\big{)}$
where
•
AVS2.1 expresses “There is a Boolean expression $\phi$ which is obtained from the quantifier-free part of $\varphi$ by replacing each occurrence of a variable by the corresponding truth value in $\{0,1\}$ assigned by the leaf valuation represented by $(G_{t},B_{t})$”.
•
AVS2.2 expresses “The Boolean expression $\phi$ evaluates to true”.
We describe next how to express AVS2.1 and AVS2.2 in
second-order logic.
6.2.1 Expressing AVS2.1
The idea is to define mappings to represent the relationships among the input
graph ${\bf G}_{\varphi}$, the graph $G_{t}$ and the quantifier-free part of the
input formulae. This is illustrated in Figure 6.5.
We can express AVS2.1 as follows:
$\mathrm{AVS2.1}\equiv A\wedge B\wedge C$ where
•
A expresses “$V_{p}$ is a partial bijection from the prefix of quantifiers of
$\varphi$ (restricted to the $X$’s that appear in the prefix) to $V_{t}$, which
maps every $X$ to its corresponding node in $G_{t}$, and which preserves $\leq^{{\bf G}_{\varphi}}$ and $E_{t}$”.
$\mathrm{A}\equiv\forall xyz\big{(}\mathrm{A1}\wedge\mathrm{A2}\wedge\mathrm{A3%
}\big{)}\wedge\forall sts^{\prime}t^{\prime}\big{(}\mathrm{A4}\big{)}$ where
–
A1 expresses “$V_{p}$ is a function”.
$\mathrm{A1}\equiv((V_{p}(x,y)\wedge V_{p}(x,z))\rightarrow y=z)$
–
A2 expresses “$V_{p}$ is injective”
$\mathrm{A2}\equiv((V_{p}(x,y)\wedge V_{p}(z,y))\rightarrow x=z)$
–
A3 defines the “domain and range of $V_{p}$”
$\mathrm{A3}\equiv\big{(}(P_{X}(x)\wedge\mathrm{PRED}_{\leq}(x,z)\wedge(P_{%
\exists}(z)\vee P_{\forall}(z)))\leftrightarrow\exists y(V_{t}(y)\wedge V_{p}(%
x,y))\big{)}$
where $\mathrm{PRED}_{\leq}(x,z)$ denotes the subformula that expresses that $z$ is the strict predecessor of
$x$ in the order $\leq^{{\bf G}_{\varphi}}$ (see Subsection 6.3).
–
A4 expresses “$V_{p}$ preserves $\leq^{{\bf G}_{\varphi}}$ and $E_{t}$”.
$(V_{p}(s,s^{\prime})\wedge V_{p}(t,t^{\prime})\wedge E_{t}(s^{\prime},t^{%
\prime}))\rightarrow$
$\big{(}\mathrm{PATH}_{\leq}(s,t)\wedge\forall z^{\prime}\big{(}(z^{\prime}\neq
s%
\wedge z^{\prime}\neq t\wedge\mathrm{PATH}_{\leq}(s,z^{\prime})\wedge\mathrm{%
PATH}_{\leq}(z^{\prime},t))\rightarrow$
$\neg P_{X}(z^{\prime})\big{)}\big{)}$
•
B expresses “$H_{\phi}$ is a partial surjective injection from the quantifier
free part of $\varphi$ to the formula $\phi$, encoded as the
first formula in $(C,E_{C})$ (see Figures 6.7 and 6.8),
which maps every $X$ in the quantifier-free part of $\varphi$ to the
corresponding position in the first formula in $(C,E_{C})$ (i.e. $\phi$),
which preserves $\wedge$, $\vee$, $\neg$, $($, $)$, $\leq^{{\bf G}_{\varphi}}$ and
$E_{C}$, and which ignores $|$”.
$\mathrm{B}\equiv\forall xy_{1}y_{2}z_{1}z_{2}\big{(}\mathrm{B1}\wedge\mathrm{B%
2}\wedge\mathrm{B3}\wedge\mathrm{B4}\big{)}\wedge\forall xx^{\prime}zy_{1}y_{2%
}z_{1}z_{2}\big{(}\mathrm{B5}\big{)}$ where
–
B1 expresses “$H_{\phi}$ is a function”.
$\mathrm{B1}\equiv\big{(}(H_{\phi}(x,y_{1},y_{2})\wedge H_{\phi}(x,z_{1},z_{2}))\rightarrow$
$(y_{1}=z_{1}\wedge y_{2}=z_{2}\wedge\exists x^{\prime}(P_{(}(x^{\prime})\wedge%
\mathrm{PATH}_{\leq}(x^{\prime},x))\wedge C(y_{1},y_{2}))\big{)}$
–
B2 expresses “$H_{\phi}$ is injective”.
$\mathrm{B2}\equiv\big{(}H_{\phi}(x,y_{1},y_{2})\wedge H_{\phi}(z,y_{1},y_{2})%
\rightarrow x=z\big{)}$
–
B3 expresses “the range of $H_{\phi}$ is the first formula in $(C,E_{C})$”.
$\mathrm{B3}\equiv\forall y_{1}^{\prime}y_{2}^{\prime}z_{1}^{\prime}z_{2}^{%
\prime}t_{1}^{\prime}t_{2}^{\prime}v^{\prime}v_{2}\big{(}\big{(}\mathit{ST}(v^%
{\prime})\wedge\neg\exists y(E_{\mathit{ST}}(y,v^{\prime}))\wedge$
$E_{\mathit{ST}}(v^{\prime},v_{2})\wedge M(v^{\prime},y_{1}^{\prime},y_{2}^{%
\prime})\wedge M(v_{2},z_{1}^{\prime},z_{2}^{\prime})\wedge$
$E_{C}(t_{1}^{\prime},t_{2}^{\prime},z_{1}^{\prime},z_{2}^{\prime})\wedge%
\mathrm{PATH}_{E_{C}}(y_{1}^{\prime},y_{2}^{\prime},y_{1},y_{2})\wedge$
$\mathrm{PATH}_{E_{C}}(y_{1},y_{2},t_{1}^{\prime},t_{2}^{\prime})\big{)}%
\rightarrow\exists x^{\prime}(H_{\phi}(x^{\prime},y_{1},y_{2}))\big{)}\wedge$
–
B4 expresses “the domain of $H_{\phi}$ corresponds to the quantifier free part of $\varphi$”.
$\mathrm{B4}\equiv\big{(}\exists x^{\prime}(P_{(}(x^{\prime})\wedge\mathrm{PATH%
}_{\leq}(x^{\prime},x))\rightarrow\exists y_{1}^{\prime}y_{2}^{\prime}(H_{\phi%
}(x,y_{1}^{\prime},y_{2}^{\prime}))\big{)}$
–
B5 expresses “$H_{\phi}$ preserves $\leq^{{\bf G}_{\varphi}}$ (ignoring “$|$”), $E_{C}$, $\wedge,\vee$, $(,)$ and $\neg$, and maps X to
$0/1$”.
$\mathrm{B5}\equiv\big{(}(H_{\phi}(x,y_{1},y_{2})\wedge H_{\phi}(z,z_{1},z_{2})%
\wedge E_{C}(y_{1},y_{2},z_{1},z_{2}))$
$\rightarrow(\mathrm{SUC_{\leq}}(x,z)\vee(\mathrm{PATH}_{\leq}(x,z)\wedge%
\forall x^{\prime}(\mathrm{PATH}_{\leq}(x,x^{\prime})\wedge$
$\mathrm{PATH}_{\leq}(x^{\prime},z)\wedge x^{\prime}\neq x\wedge x^{\prime}\neq
z%
)\rightarrow P_{|}(x^{\prime})))\big{)}\wedge$
$\big{(}H_{\phi}(x,y_{1},y_{2})\rightarrow((P_{(}(x)\wedge C_{(}(y_{1},y_{2}))\vee$
$(P_{)}(x)\wedge C_{)}(y_{1},y_{2}))\vee$
$(P_{\wedge}(x)\wedge C_{\wedge}(y_{1},y_{2}))\vee$
$(P_{\vee}(x)\wedge C_{\vee}(y_{1},y_{2}))\vee$
$(P_{\neg}(x)\wedge C_{\neg}(y_{1},y_{2}))\vee$
$(P_{X}(x)\wedge(C_{0}(y_{1},y_{2})\vee C_{1}(y_{1},y_{2}))))\big{)}$
•
C expresses “for every bijection $V_{0}$ from “$|\cdots|$” in “$QX|\cdots|$” (where Q is “$\exists$” or “$\forall$”)
to “$|\cdots|$” in “$(\ldots X|\cdots|\ldots)$”
that links a variable in the quantifier prefix of $\varphi$ with an occurrence of that variable in the quantifier-free part of it,
the variable in the quantifier free part of $\varphi$ which corresponds to the function $V_{0}$ is replaced in $\phi$ by the
value assigned to that variable by the leaf valuation $(G_{t},B_{t})$ (see Figures 6.5)”.
Note that in the formula below, $z_{0}$ represents the root in
$\mathrm{dom}(V_{0})$, $z_{f}$ represents the leaf in $\mathrm{dom}(V_{0})$, $y_{0}$ represents the root in
$\mathrm{ran}(V_{0})$, and $y_{f}$ represents the leaf in $\mathrm{ran}(V_{0})$ (see Figure 6.6).
Also note that $\phi$ is encoded in $(C,E_{C})$ starting in the node $M$(“first node in $(\mathit{ST},E_{\mathit{ST}})$”) and ending in the node $E^{-1}_{C}(M(\text{``second node in $(\mathit{ST},E_{\mathit{ST}})$''}))$, and that it is equivalent to the quantifier-free part of $\varphi$
with the variables replaced by $0$ or $1$ according to the leaf valuation
$(G_{t},B_{t})$ (this is further clarified in Subsection 6.2.2, also note Figures 6.7 and 6.8).
$\mathrm{C}\equiv\forall\,V_{0}\,\exists\,z_{0}\,y_{0}\,z_{f}\,y_{f}\,z_{0}^{%
\prime}\,y_{0}^{\prime}\,z_{f}^{\prime}\,y_{f}^{\prime}\,\big{(}(\mathrm{C1}%
\wedge\mathrm{C2}\wedge\mathrm{C3}\wedge\mathrm{C4}\wedge\mathrm{C5}\wedge%
\mathrm{C6})\rightarrow\mathrm{C7}\big{)}$
where
–
C1 expresses “$z_{0}$ is the root in $\mathrm{dom}(V_{0})$, $z_{f}$ is the leaf in $\mathrm{dom}(V_{0})$, $y_{0}$ is the root in $\mathrm{ran}(V_{0})$ and $y_{f}$ is the leaf in $\mathrm{ran}(V_{0})$”.
$\mathrm{C1}\equiv V_{0}(z_{0},y_{0})\wedge\neg\exists z^{\prime}y^{\prime}(%
\mathrm{PRED}_{\leq}(z_{0},z^{\prime})\wedge V_{0}(z^{\prime},y^{\prime}))\wedge$
$V_{0}(z_{f},y_{f})\wedge\neg\exists z^{\prime}y^{\prime}(\mathrm{SUC}_{\leq}(z%
_{f},z^{\prime})\wedge V_{0}(z^{\prime},y^{\prime}))\wedge$
$\forall z^{\prime}\big{(}(\mathrm{PATH}_{\leq}(z_{0},z^{\prime})\wedge\mathrm{%
PATH}_{\leq}(z^{\prime},z_{f}))\rightarrow\exists y^{\prime}(V_{0}(z^{\prime},%
y^{\prime}))\big{)}\wedge$
$\forall y^{\prime}\big{(}(\mathrm{PATH}_{\leq}(y_{0},y^{\prime})\wedge\mathrm{%
PATH}_{\leq}(y^{\prime},y_{f}))\rightarrow\exists z^{\prime}(V_{0}(z^{\prime},%
y^{\prime}))\big{)}$
–
C2 expresses “$V_{0}$ is a bijection from “$|\cdots|$” in “$QX|\cdots|$”
to “$|\cdots|$” in
“$(\ldots X|\cdots|\ldots)$” which preserves $\leq^{{\bf G}_{\varphi}}$”.
$\mathrm{C2}\equiv\forall xyvw\big{(}(V_{0}(x,y)\rightarrow(P_{|}(x)\wedge P_{|%
}(y)))\wedge$
$((V_{0}(x,y)\wedge V_{0}(x,v))\rightarrow y=v)\wedge$
$((V_{0}(x,y)\wedge V_{0}(w,y))\rightarrow x=w)\wedge$
$((V_{0}(x,y)\wedge V_{0}(v,w)\wedge\mathrm{SUC}_{\leq}(x,v))\rightarrow\mathrm%
{SUC}_{\leq}(y,w))\big{)}$
–
C3 expresses “$z_{0}^{\prime}$ is the predecessor of the root in $\mathrm{dom}(V_{0})$, i.e., it is the $X$ in the prefix of quantifiers”.
$\mathrm{C3}\equiv\mathrm{PRED}_{\leq}(z_{0},z_{0}^{\prime})\wedge P_{X}(z_{0}^%
{\prime})$
–
C4 expresses “$y_{0}^{\prime}$ is the predecessor of the root in $\mathrm{ran}(V_{0})$, i.e., it is the $X$ in the quantifier-free part”.
$\mathrm{C4}\equiv\mathrm{PRED}_{\leq}(y_{0},y_{0}^{\prime})\wedge P_{X}(y_{0}^%
{\prime})$
–
C5 expresses “$z_{f}^{\prime}$ is the successor of the leaf in $\mathrm{dom}(V_{0})$”.
$\mathrm{C5}\equiv\mathrm{SUC}_{\leq}(z_{f},z_{f}^{\prime})\wedge\neg P_{|}(z_{%
f}^{\prime})$
–
C6 expresses “$y_{f}^{\prime}$ is the successor of the leaf in $\mathrm{ran}(V_{0})$”.
$\mathrm{C6}\equiv\mathrm{SUC}_{\leq}(y_{f},y_{f}^{\prime})\wedge\neg P_{|}(y_{%
f}^{\prime})$
–
C7 expresses “$B_{t}(V_{p}(z_{0}^{\prime}))=H_{\phi}(y_{0}^{\prime})$”.
$\mathrm{C7}\equiv\forall\,x\,x^{\prime}\,\big{(}(V_{p}(z_{0}^{\prime},x)\wedge
B%
_{t}(x,x^{\prime}))\rightarrow$
$\exists z_{1}z_{2}(H_{\phi}(y_{0}^{\prime},z_{1},z_{2})\wedge$
$((\text{``$x^{\prime}=0$''}\wedge C_{0}(z_{1},z_{2}))\vee(\text{``$x^{\prime}=%
1$''}\wedge C_{1}(z_{1},z_{2}))))\big{)}$
6.2.2 Expressing AVS2.2
Now we need to check whether the formula $\phi$ built in the previous step,
evaluates to true. The idea is to evaluate one connective at a time, and one pair of
matching parenthesis at a time, until the final result becomes $1$.
Let us look at the example in Figure 6.7. Note that there are ten
evaluation steps, which correspond to ten “operators” (i.e., either
connectives or pairs of parenthesis). If there are at most $n$ symbols in
$\phi$, that means that the whole evaluation process needs at most $n$
evaluation steps. This is the reason for using pairs of elements to represent
the nodes of the graph $(C,E_{C})$, and quadruples to represent the edges. This
allows the whole evaluation process to take up to $n$ steps (where $n$ is the
length of the input formula). In each step, we have a Boolean sentence on
$\{0,1\}$ with up to $n$ symbols. Each node in the graph $(\mathit{ST},E_{\mathit{ST}})$ represents one such formula, and the function $M$ (for Marker)
is a pointer which tells us in which node in $(C,E_{C})$ that formula begins.
Note that in each evaluation step, either one or
two symbols are removed from the formula at the previous step.
Figure 6.8 further illustrates the graphs (A) and (B) of
Figure 6.7 with a horizontal orientation. Each evaluation step is called a stage. And the first
symbol in each stage is given by the marker function $M$.
Based on this description, we can express AVS2.2 in Section 6.2 as follows:
$\mathrm{A1}\wedge\mathrm{A2}\wedge\mathrm{A3}\wedge\mathrm{A4}\wedge\mathrm{A5}$ where
•
A1 expresses “$(C,E_{C})$ is a linear graph”.
•
A2 expresses “$(\mathit{ST},E_{\mathit{ST}})$ is a linear graph”.
•
A3 expresses “$M:\mathit{ST}\rightarrow C$ is an injective and total function that preserves $\mathrm{PATH}$ in $E_{\mathit{ST}}$ and $E_{C}$”.
•
A4 expresses “$C_{\wedge},C_{\vee},C_{\neg},C_{(},C_{)},C_{0},C_{1}$ are pairwise disjoint, and $C_{\wedge}\cup C_{\vee}\cup C_{\neg}\cup C_{(}\cup C_{)}\cup C_{0}\cup C_{1}=C$”.
•
A5 expresses “For every stage $x$, from stage $x$ to stage $x+1$, we need to follow the rules of evaluation (see Figure 6.7 part A). The formula in $(C,E_{C})$ at stage $x+1$ is the same as the formula at stage $x$, except for one of three possible sorts of changes, which correspond to the cases (a), (b) and (c) of Figure 6.9”.
We describe next how to express A1–A5 above in second-order
logic. See Section 6.3 for the auxiliary formulae used below.
$\mathrm{A1}\equiv\mathrm{LINEAR}(C,E_{C})$
$\mathrm{A2}\equiv\mathrm{LINEAR}_{2}(\mathit{ST},E_{\mathit{ST}})$
$\mathrm{A3}\equiv\forall s\,s^{\prime}\,t_{1}\,t_{2}\,k_{1}\,k_{2}\big{(}%
\mathrm{A3.1}\wedge\mathrm{A3.3}\wedge\mathrm{A3.3}\wedge\mathrm{A3.4}\big{)}$ where
•
A3.1 expresses “$M$ is a function, $M:\mathit{ST}\rightarrow C$”.
$\mathrm{A3.1}\equiv((M(s,t_{1},t_{2})\wedge M(s,k_{1},k_{2}))\rightarrow((t_{1%
}=k_{1}\wedge t_{2}=k_{2})\wedge\mathit{ST}(s)\wedge C(t_{1},t_{2})))$
•
A3.2 expresses “$M$ is injective”.
$\mathrm{A3.2}\equiv((M(s,k_{1},k_{2})\wedge M(t_{1},k_{1},k_{2}))\rightarrow s%
=t_{1})$
•
A3.3 expresses “$M$ is total”.
$\mathrm{A3.3}\equiv(\mathit{ST}(s)\rightarrow\exists t_{1}^{\prime}t_{2}^{%
\prime}(M(s,t_{1}^{\prime},t_{2}^{\prime})))$
•
A3.4 expresses “$M$ preserves PATH in $E_{\mathit{ST}}$ and $E_{C}$”.
$\mathrm{A3.4}\equiv((M(s,t_{1},t_{2})\wedge M(s^{\prime},k_{1},k_{2})\wedge%
\mathrm{PATH}_{\mathit{ST}}(s,s^{\prime}))\rightarrow\mathrm{PATH}_{E_{C}}(t_{%
1},t_{2},k_{1},k_{2}))$
$\mathrm{A4}\equiv\forall s_{1}s_{2}\big{(}(C_{\wedge}(s_{1},s_{2})\rightarrow%
\neg C_{\vee}(s_{1},s_{2}))\wedge(C_{\wedge}(s_{1},s_{2})\rightarrow\neg C_{%
\neg}(s_{1},s_{2}))\wedge$
$(C_{\wedge}(s_{1},s_{2})\rightarrow\neg C_{(}(s_{1},s_{2}))\wedge(C_{\wedge}(s%
_{1},s_{2})\rightarrow\neg C_{)}(s_{1},s_{2}))\wedge$
$(C_{\wedge}(s_{1},s_{2})\rightarrow\neg C_{0}(s_{1},s_{2}))\wedge(C_{\wedge}(s%
_{1},s_{2})\rightarrow\neg C_{1}(s_{1},s_{2}))\wedge\cdots\big{)}\wedge$
$\forall s_{1}s_{2}\big{(}C(s_{1},s_{2})\rightarrow(C_{\wedge}(s_{1},s_{2})\vee
C%
_{\vee}(s_{1},s_{2})\vee C_{\neg}(s_{1},s_{2})\vee C_{(}(s_{1},s_{2})\vee C_{)%
}(s_{1},s_{2})\vee$
$C_{0}(s_{1},s_{2})\vee C_{1}(s_{1},s_{2}))\big{)}\wedge$
$\forall s_{1}s_{2}\big{(}(C_{\wedge}(s_{1},s_{2})\rightarrow C(s_{1},s_{2}))%
\wedge(C_{\vee}(s_{1},s_{2})\rightarrow C(s_{1},s_{2}))\wedge$
$(C_{\neg}(s_{1},s_{2})\rightarrow C(s_{1},s_{2}))\wedge(C_{(}(s_{1},s_{2})%
\rightarrow C(s_{1},s_{2}))\wedge$
$(C_{)}(s_{1},s_{2})\rightarrow C(s_{1},s_{2}))\wedge(C_{0}(s_{1},s_{2})%
\rightarrow C(s_{1},s_{2}))\wedge$
$(C_{1}(s_{1},s_{2})\rightarrow C(s_{1},s_{2}))\big{)}$
$\mathrm{A5}\equiv\forall x\big{(}\mathit{ST}(x)\rightarrow\exists E_{v}\,f_{1}%
\,f_{2}\,l_{1}\,l_{2}\,f_{1}^{\prime}\,f_{2}^{\prime}\,l_{1}^{\prime}\,l_{2}^{%
\prime}\big{(}\alpha_{d}\vee\alpha_{e}\vee(\alpha_{0}\wedge(\alpha_{a}\vee%
\alpha_{b}\vee\alpha_{c}))\big{)}\big{)}$ where
The function $E_{v}$ maps the formula at stage $x$ to the formula at stage
$x+1$. The subformula $\alpha_{d}$ corresponds to the last transition, i.e., the
transition to the last formula in $(C,E_{C})$ (“$0$” or “$1$”). The
subformula $\alpha_{e}$ corresponds to the last formula in $(C,E_{C})$. The
subformulae $\alpha_{a}$, $\alpha_{b}$ and $\alpha_{c}$ correspond to the three
possible cases ($a$), ($b$) and ($c$) as in Figure 6.9, according
to which sort of operation is the one involved in the transition from the
formula in stage $x$ to the next formula in $(C,E_{C})$. Note that the
transition to the last formula $\alpha_{d}$ is necessarily an instance of case
($c$) in Figure 6.9. For case ($c$) in Figure 6.9, $E_{v}$
is not total in its domain, since $(v_{1},v_{2})\big{(}(\big{)}$ and $(w_{1},w_{2})\big{(})\big{)}$ are not mapped. For the last formula, $E_{v}$ is not injective,
since $(f_{1}^{\prime},f_{2}^{\prime})=(l_{1}^{\prime},l_{2}^{\prime})$ (i.e., $f_{1}^{\prime}=l_{1}^{\prime}$ and $f_{2}^{\prime}=l_{2}^{\prime}$) (see Figure 6.11).
$\alpha_{0}\equiv\mathrm{A5.1}\wedge\mathrm{A5.2}\wedge\mathrm{A5.3}\wedge%
\mathrm{A5.4}$ where
•
A5.1 expresses “$x$ is not the leaf in $E_{\mathit{ST}}$, and it is not the predecessor of the leaf”.
$\mathrm{A5.1}\equiv\exists yy_{1}(E_{\mathit{ST}}(x,y)\wedge E_{\mathit{ST}}(y%
,y_{1}))$
•
A5.2 expresses “$E_{v}:C\rightarrow C$ is a partial injection mapping the formula in $(C,E_{C})$ in stage $x$ to the formula in $(C,E_{C})$ in stage $E_{\mathit{ST}}(x)$”
$\mathrm{A5.2}\equiv\forall s_{1}s_{2}t_{1}t_{2}k_{1}k_{2}\big{(}((E_{v}(s_{1},%
s_{2},t_{1},t_{2})\wedge E_{v}(s_{1},s_{2},k_{1},k_{2}))\rightarrow$
$((t_{1}=k_{1}\wedge t_{2}=k_{2})\wedge C(s_{1},s_{2})\wedge C(t_{1},t_{2})))\wedge$
$((E_{v}(s_{1},s_{2},k_{1},k_{2})\wedge E_{v}(t_{1},t_{2},k_{1},k_{2}))\rightarrow$
$(s_{1}=t_{1}\wedge s_{2}=t_{2}))\big{)}$
•
A5.3 expresses “$((f_{1},f_{2}),(l_{1},l_{2}))$ and $((f_{1}^{\prime},f_{2}^{\prime}),(l_{1}^{\prime},l_{2}^{\prime}))$ are the delimiters of the two formulae as in Figure 6.10”.
$\mathrm{A5.3}\equiv M(x,f_{1},f_{2})\wedge\mathrm{A5.3.1}\wedge\mathrm{A5.3.2}%
\wedge\mathrm{5.3.3}$ where
–
A5.3.1 expresses “$M(E_{\mathit{ST}}(x),E_{C}(l_{1},l_{2}))$”.
–
A5.3.2 expresses “$E_{C}(l_{1},l_{2})=(f^{\prime}_{1},f^{\prime}_{2})$”.
–
A5.3.3 expresses “$E^{-1}_{C}(M(E_{\mathit{ST}}(E_{\mathit{ST}}(x))),l_{1}^{\prime},l_{2}^{\prime})$”.
•
A5.4 expresses “$E_{v}$ maps nodes from the subgraph induced by $((f_{1},f_{2}),(l_{1},l_{2}))$ to the subgraph induced by $((f_{1}^{\prime},f_{2}^{\prime}),(l_{1}^{\prime},l_{2}^{\prime}))$”.
$\mathrm{A5.4}\equiv\forall y_{1}y_{2}z_{1}z_{2}\big{(}E_{v}(y_{1},y_{2},z_{1},%
z_{2})\rightarrow$
$\big{(}\mathrm{PATH}_{E_{C}}(f_{1},f_{2},y_{1},y_{2})\wedge\mathrm{PATH}_{E_{C%
}}(y_{1},y_{2},l_{1},l_{2})\wedge$
$\mathrm{PATH}_{E_{C}}(f_{1}^{\prime},f_{2}^{\prime},z_{1},z_{2})\wedge\mathrm{%
PATH}_{E_{C}}(z_{1},z_{2},l_{1}^{\prime},l_{2}^{\prime})\big{)}\big{)}\wedge$
$E_{v}(f_{1},f_{2},f_{1}^{\prime},f_{2}^{\prime})\wedge E_{v}(l_{1},l_{2},l_{1}%
^{\prime},l_{2}^{\prime})$
$\alpha_{a}\equiv\exists v_{1}v_{2}w_{1}w_{2}v_{1}^{\prime}v_{2}^{\prime}w_{1}^%
{\prime}w_{2}^{\prime}p_{11}p_{12}p_{21}p_{22}p_{31}p_{32}p_{11}^{\prime}p_{12%
}^{\prime}\big{(}$
$\mathrm{A5.5}\wedge\mathrm{A5.6}\wedge\mathrm{A5.7}\wedge\mathrm{A5.8}\wedge%
\mathrm{A5.9}\big{)}$ where
•
A5.5 expresses “$((v_{1},v_{2}),(w_{1},w_{2}))$ and $((v_{1}^{\prime},v_{2}^{\prime}),(w_{1}^{\prime},w_{2}^{\prime}))$
define the window of change, that is the segment of the formula that is affected (changed) in the transition from stage $x$ to stage $x+1$ of the evaluation (see Cases (a) and (b) in Figure 6.10)”.
$\mathrm{A5.5}\equiv\mathrm{PATH}_{E_{C}}(f_{1},f_{2},v_{1},v_{2})\wedge\mathrm%
{PATH}_{E_{C}}(w_{1},w_{2},l_{1},l_{2})\wedge E_{C}(p_{11},p_{12},p_{21},p_{22%
})\wedge$
$E_{C}(p_{21},p_{22},p_{31},p_{32})\wedge E_{C}(v_{1},v_{2},p_{11},p_{12})%
\wedge E_{C}(p_{31},p_{32},w_{1},w_{2})$
$C_{(}(v_{1},v_{2})\wedge C_{)}(w_{1},w_{2})\wedge\mathrm{PATH}_{E_{C}}(f_{1}^{%
\prime},f_{2}^{\prime},v_{1}^{\prime},v_{2}^{\prime})\wedge\mathrm{PATH}_{E_{C%
}}(w_{1}^{\prime},w_{2}^{\prime},l_{1}^{\prime},l_{2}^{\prime})\wedge$
$E_{C}(v_{1}^{\prime},v_{2}^{\prime},p_{11}^{\prime},p_{12}^{\prime})\wedge E_{%
C}(p_{11}^{\prime},p_{12}^{\prime},w_{1}^{\prime},w_{2}^{\prime})\wedge E_{v}(%
p_{11},p_{12},p_{11}^{\prime},p_{12}^{\prime})\wedge$
$E_{v}(v_{1},v_{2},v_{1}^{\prime},v_{2}^{\prime})\wedge E_{v}(w_{1},w_{2},w_{1}%
^{\prime},w_{2}^{\prime})\wedge C_{(}(v_{1}^{\prime},v_{2}^{\prime})\wedge C_{%
)}(w_{1}^{\prime},w_{2}^{\prime})$
•
A5.6 expresses “$E_{v}$ preserves $E_{C}$ outside of the window of change, and preserves
left and right side of the window of change (see Figure 6.10)”.
$\mathrm{A5.6}\equiv\forall z_{11}z_{12}z_{21}z_{22}z_{11}^{\prime}z_{12}^{%
\prime}z_{21}^{\prime}z_{22}^{\prime}\big{(}$
$\big{(}(\mathrm{PATH}_{E_{C}}(f_{1},f_{2},z_{11},z_{12})\wedge\mathrm{PATH}_{E%
_{C}}(z_{21},z_{22},v_{1},v_{2})\wedge$
$E_{C}(z_{11},z_{12},z_{21},z_{22})\wedge E_{v}(z_{11},z_{12},z_{11}^{\prime},z%
_{12}^{\prime})\wedge E_{v}(z_{21},z_{22},z_{21}^{\prime},z_{22}^{\prime}))\rightarrow$
$(\mathrm{PATH}_{E_{C}}(f_{1}^{\prime},f_{2}^{\prime},z_{11}^{\prime},z_{12}^{%
\prime})\wedge\mathrm{PATH}_{E_{C}}(z_{21}^{\prime},z_{22}^{\prime},v_{1}^{%
\prime},v_{2}^{\prime})\wedge$
$E_{C}(z_{11}^{\prime},z_{12}^{\prime},z_{21}^{\prime},z_{22}^{\prime}))\big{)}\wedge$
$\big{(}(\mathrm{PATH}_{E_{C}}(w_{1},w_{2},z_{11},z_{12})\wedge\mathrm{PATH}_{E%
_{C}}(z_{21},z_{22},l_{1},l_{2})\wedge$
$E_{C}(z_{11},z_{12},z_{21},z_{22})\wedge E_{v}(z_{11},z_{12},z_{11}^{\prime},z%
_{12}^{\prime})\wedge E_{v}(z_{21},z_{22},z_{21}^{\prime},z_{22}^{\prime}))\rightarrow$
$(\mathrm{PATH}_{E_{C}}(w_{1}^{\prime},w_{2}^{\prime},z_{11}^{\prime},z_{12}^{%
\prime})\wedge\mathrm{PATH}_{E_{C}}(z_{21}^{\prime},z_{22}^{\prime},l_{1}^{%
\prime},l_{2}^{\prime})\wedge$
$E_{C}(z_{11}^{\prime},z_{12}^{\prime},z_{21}^{\prime},z_{22}^{\prime}))\big{)}%
\big{)}$
•
A5.7 expresses “$E_{v}$ preserves symbols in left side of the window of change”.
$\mathrm{A5.7}\equiv\forall z_{11}z_{12}z_{11}^{\prime}z_{12}^{\prime}\big{(}$
$\big{(}\mathrm{PATH}_{E_{C}}(f_{1},f_{2},z_{11},z_{12})\wedge\mathrm{PATH}_{E_%
{C}}(z_{11},z_{12},v_{1},v_{2})\wedge E_{v}(z_{11},z_{12},z_{11}^{\prime},z_{1%
2}^{\prime})\big{)}$
$\rightarrow\big{(}\mathrm{PATH}_{E_{C}}(f_{1}^{\prime},f_{2}^{\prime},z_{11}^{%
\prime},z_{12}^{\prime})\wedge\mathrm{PATH}_{E_{C}}(z_{11}^{\prime},z_{12}^{%
\prime},v_{1}^{\prime},v_{2}^{\prime})\wedge$
$\big{(}(C_{(}(z_{11},z_{12})\wedge C_{(}(z_{11}^{\prime},z_{12}^{\prime}))\vee%
(C_{)}(z_{11},z_{12})\wedge C_{)}(z_{11}^{\prime},z_{12}^{\prime}))\vee$
$(C_{\wedge}(z_{11},z_{12})\wedge C_{\wedge}(z_{11}^{\prime},z_{12}^{\prime}))%
\vee(C_{\vee}(z_{11},z_{12})\wedge C_{\vee}(z_{11}^{\prime},z_{12}^{\prime}))\vee$
$(C_{0}(z_{11},z_{12})\wedge C_{0}(z_{11}^{\prime},z_{12}^{\prime}))\vee(C_{1}(%
z_{11},z_{12})\wedge C_{1}(z_{11}^{\prime},z_{12}^{\prime}))\vee$
$(C_{\neg}(z_{11},z_{12})\wedge C_{\neg}(z_{11}^{\prime},z_{12}^{\prime}))\big{%
)}\big{)}\big{)}$
•
A5.8 expresses “$E_{v}$ preserves symbols in right side of the window of change”.
$\mathrm{A5.8}\equiv\forall z_{11}z_{12}z_{11}^{\prime}z_{12}^{\prime}\big{(}$
$\big{(}\mathrm{PATH}_{E_{C}}(w_{1},w_{2},z_{11},z_{12})\wedge\mathrm{PATH}_{E_%
{C}}(z_{11},z_{12},l_{1},l_{2})\wedge E_{v}(z_{11},z_{12},z_{11}^{\prime},z_{1%
2}^{\prime})\big{)}$
$\rightarrow\big{(}\mathrm{PATH}_{E_{C}}(w_{1}^{\prime},w_{2}^{\prime},z_{11}^{%
\prime},z_{12}^{\prime})\wedge\mathrm{PATH}_{E_{C}}(z_{11}^{\prime},z_{12}^{%
\prime},l_{1}^{\prime},l_{2}^{\prime})\wedge$
$\big{(}(C_{(}(z_{11},z_{12})\wedge C_{(}(z_{11}^{\prime},z_{12}^{\prime}))\vee%
(C_{)}(z_{11},z_{12})\wedge C_{)}(z_{11}^{\prime},z_{12}^{\prime}))\vee$
$(C_{\wedge}(z_{11},z_{12})\wedge C_{\wedge}(z_{11}^{\prime},z_{12}^{\prime}))%
\vee(C_{\vee}(z_{11},z_{12})\wedge C_{\vee}(z_{11}^{\prime},z_{12}^{\prime}))\vee$
$(C_{0}(z_{11},z_{12})\wedge C_{0}(z_{11}^{\prime},z_{12}^{\prime}))\vee(C_{1}(%
z_{11},z_{12})\wedge C_{1}(z_{11}^{\prime},z_{12}^{\prime}))\vee$
$(C_{\neg}(z_{11},z_{12})\wedge C_{\neg}(z_{11}^{\prime},z_{12}^{\prime}))\big{%
)}\big{)}\big{)}$
•
A5.9 expresses “In $(p_{11}^{\prime},p_{12}^{\prime})$ we get the result of applying the operator $\theta$ in $(p_{21},p_{22})$ to the Boolean values $b_{1}$, in $(p_{11},p_{12})$, and $b_{2}$ in $(p_{31},p_{32})$ (see (a) in Figure 6.9)”.
$\mathrm{A5.9}\equiv\big{(}(C_{0}(p_{11},p_{12})\wedge C_{0}(p_{31},p_{32})%
\wedge C_{\wedge}(p_{21},p_{22})\wedge C_{0}(p_{11}^{\prime},p_{12}^{\prime}))\vee$
$(C_{0}(p_{11},p_{12})\wedge C_{0}(p_{31},p_{32})\wedge C_{\vee}(p_{21},p_{22})%
\wedge C_{0}(p_{11}^{\prime},p_{12}^{\prime}))\vee$
$(C_{0}(p_{11},p_{12})\wedge C_{1}(p_{31},p_{32})\wedge C_{\wedge}(p_{21},p_{22%
})\wedge C_{0}(p_{11}^{\prime},p_{12}^{\prime}))\vee$
$(C_{0}(p_{11},p_{12})\wedge C_{1}(p_{31},p_{32})\wedge C_{\vee}(p_{21},p_{22})%
\wedge C_{1}(p_{11}^{\prime},p_{12}^{\prime}))\vee$
$(C_{1}(p_{11},p_{12})\wedge C_{0}(p_{31},p_{32})\wedge C_{\wedge}(p_{21},p_{22%
})\wedge C_{0}(p_{11}^{\prime},p_{12}^{\prime}))\vee$
$(C_{1}(p_{11},p_{12})\wedge C_{0}(p_{31},p_{32})\wedge C_{\vee}(p_{21},p_{22})%
\wedge C_{1}(p_{11}^{\prime},p_{12}^{\prime}))\vee$
$(C_{1}(p_{11},p_{12})\wedge C_{1}(p_{31},p_{32})\wedge C_{\wedge}(p_{21},p_{22%
})\wedge C_{1}(p_{11}^{\prime},p_{12}^{\prime}))\vee$
$(C_{1}(p_{11},p_{12})\wedge C_{1}(p_{31},p_{32})\wedge C_{\vee}(p_{21},p_{22})%
\wedge C_{1}(p_{11}^{\prime},p_{12}^{\prime}))\big{)}$
The subformulae $\alpha_{b}$ and $\alpha_{c}$ that correspond to the cases ($b$)
and ($c$) in Figure 6.9, are similar to $\alpha_{a}$.
For the clarity of presentation, we omit those formulae. Furthermore, it should be clear how to build
them using $\alpha_{a}$ as template. Moreover, the complete formulae can be found
in [12]. We present next the remaining two subformulae,
namely $\alpha_{d}$ and $\alpha_{e}$.
$\alpha_{d}\equiv\exists y\big{(}E_{\mathit{ST}}(x,y)\wedge\neg\exists z(E_{%
\mathit{ST}}(y,z))\wedge$
$\exists p_{11}p_{12}p_{11}^{\prime}p_{12}^{\prime}\big{(}M(x,f_{1},f_{2})%
\wedge M(y,p_{11}^{\prime},p_{12}^{\prime})\wedge$
$E_{C}(f_{1},f_{2},p_{11},p_{12})\wedge E_{C}(p_{11},p_{12},l_{1},l_{2})\wedge E%
_{C}(l_{1},l_{2},p_{11}^{\prime},p_{12}^{\prime})\wedge$
$\neg\exists p_{21}^{\prime}p_{22}^{\prime}(E_{C}(p_{11}^{\prime},p_{12}^{%
\prime},p_{21}^{\prime},p_{22}^{\prime}))\wedge$
$C_{(}(f_{1},f_{2})\wedge C_{)}(l_{1},l_{2})\wedge$
$((C_{1}(p_{11},p_{12})\wedge C_{1}(p_{11}^{\prime},p_{12}^{\prime}))\vee(C_{0}%
(p_{11},p_{12})\wedge C_{0}(p_{11}^{\prime},p_{12}^{\prime})))\big{)}\big{)}$
Note that the first line in $\alpha_{d}$ expresses “$x$ is the predecessor of the leaf in $E_{\mathit{ST}}$”, so that this case corresponds to the last transition (see Figure 6.11). Also note that the last transition is necessarily an instance of case (c) in Figure 6.9”
$\alpha_{e}\equiv\mathrm{A5.10}\wedge\exists p_{1}^{\prime}p_{2}^{\prime}\big{(%
}M(x,p_{1}^{\prime},p_{2}^{\prime})\wedge\mathrm{A5.11}\wedge\mathrm{A5.12}%
\big{)}$ where
•
A5.10 expresses “$x$ is the leaf in $E_{\mathit{ST}}$”.
$\mathrm{A5.10}\equiv\neg\exists y(E_{\mathit{ST}}(x,y))$
•
A5.11 expresses “$(p_{1}^{\prime},p_{2}^{\prime})$ is the leaf in $E_{C}$”.
$\mathrm{A5.11}\equiv\neg\exists y_{1}^{\prime}y_{2}^{\prime}(E_{C}(p_{1}^{%
\prime},p_{2}^{\prime},y_{1}^{\prime},y_{2}^{\prime}))$
•
A5.12 expresses “the last formula in $(C,E_{C})$ is $1$”.
$\mathrm{A5.12}\equiv C_{1}(p_{1}^{\prime},p_{2}^{\prime})$
6.3 Auxiliary Formulae
For the sake of completeness, we define next the remaining auxiliary formulae used through the previous
subsections. We assume an edge relation $E$ and a total order $\leq$.
“$x=0$” $\equiv\neg\exists y(y\neq x\wedge y\leq x)$
“$x=1$” $\equiv\exists y(y\neq x\wedge y\leq x\wedge\neg\exists z(z\neq x\wedge z\neq y%
\wedge y\leq z\wedge z\leq x)\wedge\neg\exists z(z\neq y\wedge z\leq y))$
$\mathrm{SUC}_{\leq}(x,y)\equiv x\leq y\wedge y\neq x\wedge\neg\exists z(z\neq x%
\wedge z\neq y\wedge x\leq z\wedge z\leq y)$
$\mathrm{PRED}_{\leq}(y,x)\equiv\mathrm{SUC}_{\leq}(x,y)$
$\mathrm{PATH_{E}}(v,w)$ is used to denote the following formula which is
satisfied by a given graph ${\bf G}$ iff $(v,w)$ is in the transitive closure
of the relation $E^{\bf G}$.
$\mathrm{PATH}_{E}(v,w)\equiv v=w\,\vee\exists V^{\prime}E^{\prime}\big{(}V^{%
\prime}(v)\wedge V^{\prime}(w)\wedge\mathrm{A1}\wedge\mathrm{A2}\wedge\mathrm{%
A3}\wedge\mathrm{A4}\wedge\mathrm{A5}\big{)}$
•
A1 expresses “$(V^{\prime},E^{\prime})$ is a subgraph of $(V,E)$ with no loops”.
$\mathrm{A1}\equiv\forall xy(E^{\prime}(x,y)\rightarrow(V^{\prime}(x)\wedge V^{%
\prime}(y)\wedge E(x,y)))\wedge\forall x(V^{\prime}(x)\rightarrow V(x))\wedge%
\forall x(\neg E^{\prime}(x,x))$
•
A2 expresses “$v$ is the only minimal node”.
$\mathrm{A2}\equiv\neg\exists x(E^{\prime}(x,v))\wedge\forall y((V^{\prime}(y)%
\wedge y\neq v)\rightarrow\exists x(E^{\prime}(x,y)))$
•
A3 expresses “$w$ is the only maximal node”.
$\mathrm{A3}\equiv\neg\exists x(E^{\prime}(w,x))\wedge\forall y((V^{\prime}(y)%
\wedge y\neq w)\rightarrow\exists x(E^{\prime}(y,x)))$
•
A4 expresses “all nodes except $v$ have input degree $1$”.
$\mathrm{A4}\equiv\forall z((V^{\prime}(z)\wedge z\neq v)\rightarrow\exists x(E%
^{\prime}(x,z)\wedge\forall y((V^{\prime}(y)\wedge E^{\prime}(y,z))\rightarrow
y%
=x)))$
•
A5 expresses “all nodes except $w$ have output degree $1$”.
$\mathrm{A5}\equiv\forall z((V^{\prime}(z)\wedge z\neq w)\rightarrow\exists x(E%
^{\prime}(z,x)\wedge\forall y((V^{\prime}(y)\wedge E^{\prime}(z,y))\rightarrow
y%
=x)))$
That is, $\mathrm{PATH}_{E}(v,w)$ expresses “$(V^{\prime},E^{\prime})$ is a linear subgraph of $(V,E)$, with minimal node
$v$ and maximal node $w$”. We use a similar
strategy to define the next auxiliary formula $\mathrm{LINEAR}(V,E)$ which expresses “$(V,E)$ is a linear graph”.
$\mathrm{LINEAR}(V,E)\equiv\forall xy(\mathrm{PATH}_{E}(x,y)\vee\mathrm{PATH}_{%
E}(y,x))\wedge$
$(\exists xy(x\neq y)\rightarrow\forall x(\neg E(x,x)))\wedge$
$\exists vw\big{(}V(v)\wedge V(w)\wedge$
$\neg\exists x(E(x,v))\wedge\forall y((V(y)\wedge y\neq v)\rightarrow\exists x(%
E(x,y)))\wedge$
$\neg\exists x(E(w,x))\wedge\forall y((V(y)\wedge y\neq w)\rightarrow\exists x(%
E(y,x)))\wedge$
$\forall z((V(z)\wedge z\neq v)\rightarrow$
$\exists x(E(x,z)\wedge\forall y((V(y)\wedge E(y,z))\rightarrow y=x)))$
$\forall z((V(z)\wedge z\neq w)\rightarrow$
$\exists x(E(z,x)\wedge\forall y((V(y)\wedge E(z,y))\rightarrow y=x)))\big{)}$
Note that we only allow loops in a linear graph when it has only one node.
In a similar way we can define the second-order formula $\mathrm{LINEAR}_{2}(Vâ,Eâ)$ where the free second-order variables have arity $2$ and $4$ respectively.
We also use the formula $\mathrm{PATH}_{\mathrm{E_{C}}}(x_{1},x_{2},y_{1},y_{2})$ with free first-order variables $x_{1}$, $x_{2}$, $y_{1}$, $y_{2}$, where the set of vertices is a binary relation, and the set of edges is a $4$-ary relation (see Figures 6.7 and 6.8).
7 $\mathrm{SATQBF}$ in Third-Order Logic
In this section we show how to build a formula in third-order logic that
expresses $\mathrm{SATQBF}$. We omit the tedious details of the subformulae
which can be built following the same patterns than in the detailed exposition of the
second-order formula for $\mathrm{SATQBF}_{k}$.
Roughly, we first express the existence of a third-order alternating valuation ${\bf T}_{v}$ applicable to a given $\mathrm{QBF}$ formula $\varphi$. Then we proceed
to evaluate the quantifier-free part $\varphi^{\prime}$ of $\varphi$ on each leaf
valuation ${\bf L}_{v}$ of ${\bf T}_{v}$. For this part we use the same
second-order subformulae than for $\mathrm{SATQBF}_{k}$. That is, from
$\varphi^{\prime}$ and ${\bf L}_{v}$, we build a Boolean sentence $\phi$ on $\{0,1\}$ by
replacing each occurrence of a Boolean variable $x$ in $\varphi^{\prime}$ by a constant
$0$ or $1$ according to the Boolean value assigned by ${\bf L}_{v}$ to $x$, and
then we evaluate $\phi$.
Unlike the case with $\mathrm{SATQBF}_{k}$ in which the input formulae all have a same fixed number $k$ of alternating blocks of quantifiers, in the case of $\mathrm{SATQBF}$ the number of alternating blocks $k\geq 1$ of quantifiers in the input formulae is not fixed. That is,
we need to take into account that the input formula can have any arbitrary number $k\geq 1$ of alternating blocks of quantifiers. We assume w.l.o.g. that the quantification in the input formula $\varphi$ has the form
$\exists x_{11}\cdots\exists x_{1l_{1}}\forall x_{21}\cdots\forall x_{2l_{2}}%
\exists x_{31}\cdots\exists x_{3l_{3}}\cdots Qx_{k1}\cdots Qx_{kl_{k}}($
$\varphi^{\prime}(x_{11},\ldots,x_{1l_{1}},x_{21}\ldots,x_{2l_{2}},x_{31},%
\ldots,x_{3l_{3}},\ldots,x_{k1},\ldots,x_{kl_{k}}))$
where $k\geq 1$, the formula
$\varphi^{\prime}$ is a quantifier-free Boolean formula and $Q$ is $\exists$ if $k$ is
odd, or $\forall$ if $k$ is even. To represent the formulae as relational structures, we use the same encoding based in word models as in Section 6.
We present a sketch of the third-order formula $\varphi_{\mathrm{SATQBF}}$ that expresses $\mathrm{SATQBF}$. We follow a top-down approach,
leaving most of the fine details of the formulae in the lowest level of abstraction as an exercise for the reader.
At the highest level of abstraction, we can think of $\varphi_{\mathrm{SATQBF}}$ as a third-order formula that expresses.
“There is a third-order alternating valuation $${\bf T}_{v}$$ applicable to
$$\varphi$$, which satisfies $$\varphi$$”.
At the next level of abstraction we can express $\varphi_{\mathrm{SATQBF}}$ in third-order logic as follows.
$\exists{\cal V}_{t}\,{\cal E}_{t}\,{\cal B}_{t}\,V_{t}\,E_{t}\,\Big{(}\mathrm{%
A1}\wedge\mathrm{A2}\wedge\mathrm{A3}\wedge\mathrm{A4}\wedge\mathrm{A5}\big{)}$ where
•
A1 expresses “${\cal B}_{t}:{\cal V}_{t}\rightarrow\{0,1\}$”.
•
A2 expresses “$G_{t}=(V_{t},E_{t})$ is a linear graph which represents the sequence of
quantified variables in $\varphi$”.
•
A3 expresses “$({\cal V}_{t},{\cal E}_{t})$ is a third-order binary tree with all its leaves at the
same depth, which is in turn equal to the length of $(E_{t},V_{t})$”.
•
A4 expresses “$({\cal V}_{t},{\cal E}_{t},{\cal B}_{t})$ is a third-order alternating
valuation ${\bf T}_{v}$ applicable to $\varphi$, i.e., all the nodes in $({\cal V}_{t},{\cal E}_{t})$ whose depth correspond to a universally quantified
variable in the prefix of quantifiers of $\varphi$, have exactly one sibling, and its
value under ${\cal B}_{t}$ is different than that of the given node, and
all the nodes whose depth correspond to an existentially quantified variable
in the prefix of quantifiers of $\varphi$, are either the root or have no
siblings”.
•
A5 expresses “Every leaf valuation in $({\cal V}_{t},{\cal E}_{t},{\cal B}_{t})$
satisfies $\varphi^{\prime}$”.
Recall that we use uppercase calligraphic letters for third-order variables and plain uppercase letters for second-order variables. In particular, ${\cal V}_{t}$, ${\cal E}_{t}$ and ${\cal B}_{t}$ are third-order variables while $V_{t}$ and $E_{t}$ are second-order variables.
Finally, we describe the strategies to express A2–A5 in third-order logic.
A2.
$\mathrm{LINEAR}(V_{t},E_{t})\wedge\mathrm{A2.1}$ where
–
A2.1 expresses “The length of $G_{t}$ is equal to the number of variables in the prefix of
quantifiers of $\varphi$. That is, there is a relation $V_{p}$ which is a partial bijection from the quantifier prefix of $\varphi$
(restricted to the $X$’s in the quantifier prefix) to $V_{t}$, which maps every
$X$ in the quantifier prefix to its corresponding node in $G_{t}$, and which preserves $E_{t}$
and $\leq^{{\bf G}_{\varphi}}$ in $G_{t}$ and $\varphi$ (restricted to the $X$’s in the
quantifier prefix), respectively”.
See (A) in Subsection 6.2.1 for more details.
A3.
Let ${\cal E}_{t}\restriction_{{\cal S}_{d}}$ denote the restriction of the third-order relation ${\cal E}_{t}$ to the nodes in the third-order set ${\cal S}_{d}$. We can express A3 as follows:
$\mathrm{A3.1}\wedge\mathrm{A3.2}\wedge\mathrm{A3.3}\wedge\mathrm{A3.4}$ where
–
A3.1 expresses “$({\cal V}_{t},{\cal E}_{t})$ is a third-order connected graph that has one root and one or
more leaves”.
–
A3.2 expresses “Except for the root node, all nodes in $({\cal V}_{t},{\cal E}_{t})$ have input
degree $1$”.
–
A3.3 expresses “Except for the leaf nodes, all nodes in $({\cal V}_{t},{\cal E}_{t})$ have
output degree $1$ or $2$”.
–
A3.4 expresses “All leaf nodes in $({\cal V}_{t},{\cal E}_{t})$ have the same
depth, which is in turn equal to the length of $(V_{t},E_{t})$”
A3.1, A3.2 and A3.3 can be expressed in third-order logic as follows:
$\exists R\big{(}\forall Z({\cal V}_{t}(Z)\rightarrow\mathrm{PATH}_{{\cal E}_{t%
}}(R,Z))\wedge$
$\neg\exists S_{1}({\cal E}_{t}(S_{1},R))\wedge$
$\exists S_{1}(\neg\exists S_{2}({\cal E}_{t}(S_{1},S_{2})))\wedge$
$\forall Z(({\cal V}_{t}(Z)\wedge Z\neq R)\rightarrow\exists S_{1}({\cal E}_{t}%
(S_{1},Z)\wedge\forall S_{2}({\cal E}_{t}(S_{2},Z)\rightarrow S_{1}=S_{2})))%
\big{)}\wedge$
$\forall Z\big{(}{\cal V}_{t}(Z)\rightarrow\neg\exists S_{1}S_{2}S_{3}(S_{1}%
\neq S_{2}\wedge S_{2}\neq S_{3}\wedge S_{1}\neq S_{3}\wedge{\cal E}_{t}(Z,S_{%
1})\wedge{\cal E}_{t}(Z,S_{2})\wedge$
${\cal E}_{t}(Z,S_{3}))\big{)}$
Regarding A3.4, we can express it as follows:
$\forall X\big{(}\mathrm{A3.4.1}\rightarrow\big{(}\exists{\cal S}_{d}\,{\cal D}%
(\mathrm{A3.4.2}\wedge\mathrm{A3.4.3}\wedge{\cal S}_{d}(X)\wedge\mathrm{A3.4.4%
}\wedge\mathrm{A3.4.5})\big{)}\big{)}$ where
–
A3.4.1 expresses “$X$ is a leaf node in $({\cal V}_{t},{\cal E}_{t})$”.
–
A3.4.2 expresses “${\cal S}_{d}\subseteq{\cal V}_{t}$”.
–
A3.4.3 expresses “${\cal D}:V_{t}\rightarrow{\cal S}_{d}$ is a bijection that
preserves $E_{t}$ and ${\cal E}_{t}\restriction_{{\cal S}_{d}}$”.
–
A3.4.4 expresses “${\cal D}^{-1}(X)$ is the leaf node in $G_{t}=(V_{t},E_{t})$”.
–
A3.4.5 expresses “${\cal S}_{d}$ includes the root of $({\cal V}_{t},{\cal E}_{t})$”.
A4.
We can express A4 as follows (refer to Figures 6.2 and 6.5):
$\mathrm{A4.1}\wedge\forall x\forall{\cal S}_{d}\big{(}(V_{t}(x)\wedge\mathrm{A%
4.2}\wedge\mathrm{A4.3})\rightarrow$
$\forall{\cal D}((\mathrm{A4.4}\wedge\mathrm{A4.5})\rightarrow((\mathrm{A4.6}%
\rightarrow\mathrm{A4.7})\wedge(\mathrm{A4.8}\rightarrow\mathrm{A4.9})))\big{)}$ where
–
A4.1 expresses “${\cal B}_{t}$ is a total function from ${\cal V}_{t}$ to $\{0,1\}$”.
–
A4.2 expresses “${\cal S}_{d}\subseteq{\cal V}_{t}$”.
–
A4.3 expresses “$({\cal S}_{d},{\cal E}_{t}\restriction_{{\cal S}_{d}})$ is a linear graph which includes the root of $({\cal V}_{t},{\cal E}_{t})$”.
–
A4.4 expresses “${\cal D}$ is a bijection from the initial subgraph of $G_{t}$ up to $x$, to ${\cal S}_{d}$”.
–
A4.5 expresses “${\cal D}$ preserves $E_{t}$ and ${\cal E}_{t}\restriction_{{\cal S}_{d}}$”.
–
A4.6 expresses “the predecessor of $V^{-1}_{p}(x)$ in $\leq^{{\bf G}_{\varphi}}$ is $\forall$”.
–
A4.7 expresses “${\cal D}(x)$ has exactly one sibling in $({\cal V}_{t},{\cal E}_{t})$ and ${\cal B}_{t}$ of that sibling is not equal to ${\cal B}_{t}({\cal D}(x))$”.
–
A4.8 expresses “the predecessor of $V^{-1}_{p}(x)$ in $\leq^{{\bf G}_{\varphi}}$ is $\exists$”.
–
A4.9 expresses “${\cal D}(x)$ has no siblings in $({\cal V}_{t},{\cal E}_{t})$, or ${\cal D}(x)$ is the root in $({\cal V}_{t},{\cal E}_{t})$”.
A5.
$\forall{\cal S}_{v}\big{(}(\mathrm{A5.1}\wedge\mathrm{A5.2})\rightarrow\exists%
{\cal D}\,B_{t}\big{(}\mathrm{A5.3}\wedge\mathrm{A5.4}\wedge\mathrm{A5.5}\big{%
)}\big{)}$ where
–
A5.1 expresses “${\cal S}_{v}\subseteq{\cal V}_{t}$”.
–
A5.2 expresses “$({\cal S}_{v},{\cal E}_{t}\restriction_{{\cal S}_{v}})$ is a linear graph which includes the root and a leaf of $({\cal V}_{t},{\cal E}_{t})$”.
–
A5.3 expresses “${\cal D}$ is a bijection from $V_{t}$ to ${\cal S}_{v}$ which preserves $E_{t}$ and ${\cal E}_{t}\restriction_{{\cal S}_{v}}$”.
–
A5.4 expresses “$B_{t}$ is a total function from $V_{t}$ to $\{0,1\}$ which coincides with ${\cal B}_{t}({\cal S}_{v})$ w.r.t. ${\cal D}$”.
–
A5.5 expresses “the leaf valuation represented by $(V_{t},E_{t},B_{t})$ satisfies the quantifier-free subformula $\varphi^{\prime}$ of $\varphi$”.
Note that, A5.5 can be expressed as in Subsection 6.2.2.
Remark 2
Note that while in the third-order formulae in A4 and A5 we have used universal third-order quantification (for ${\cal S}_{d}$ and ${\cal D}$ in A4, and for ${\cal S}_{v}$ in A5), it is not actually needed, and existential third-order quantification is enough. These are the only sub-formulae where we have used universal third-order quantification. Hence, we strongly believe that our third-order formula can be translated in a rather technical way into an existential third-order formula.
Let us consider the sketch for an existential third-order formula equivalent to the formula in A4 (the existential formula for A5 is easier). We can say that for every node $x$ in the graph $(V_{t},E_{t})$, and for every set $Z$ that is a node in the third-order graph $({\cal V}_{t},{\cal E}_{t})$, and such that there is a third-order set ${\cal S}_{d}$ of nodes in the third-order graph $({\cal V}_{t},{\cal E}_{t})$, such that the restriction of the edge relation ${\cal E}_{t}$ to the third-order set ${\cal S}_{d}$, together with ${\cal S}_{d}$, form a (third-order) subgraph that is a linear graph whose root is the root of the third-order graph $({\cal V}_{t},{\cal E}_{t})$, and whose leaf is the set $Z$, and such that its length is the length of the initial subgraph of the graph $(V_{t},E_{t})$, up to the node $x$, if the variable represented by $x$ in the input formula $\varphi$ is universally quantified, then the node $Z$ in the third-order graph $({\cal V}_{t},{\cal E}_{t})$ has exactly one sibling in that graph, and that sibling has a different value assigned by ${\cal B}_{t}$ than the value assigned by ${\cal B}_{t}$ to $Z$. On the other hand, if the variable represented by $x$ in the input formula $\varphi$ is existentially quantified, then the node $Z$ in the third-order graph $({\cal V}_{t},{\cal E}_{t})$ has no sibling in that graph.
To say that “the third order graph induced by the set ${\cal S}_{d}$ in the graph $({\cal V}_{t},{\cal E}_{t})$, whose leaf is the set $Z$, has the same length as the initial subgraph of the graph $(V_{t},E_{t})$, up to the node $x$”, we say that there is a binary third-order relation $\cal D$ which is a bijection between the set of nodes in the initial subgraph of the graph $(V_{t},E_{t})$, up to the node $x$, and the third-order set ${\cal S}_{d}$, and which preserves $E_{t}$ and the restriction of the edge relation ${\cal E}_{t}$ to the third-order set ${\cal S}_{d}$.
8 Final Considerations
Let $\exists\mathrm{SO}^{\leq 2}$ denote the restriction of $\exists$SO to formulae with second-order
variables of arity $\leq 2$.
As pointed out in [2], it is open whether on graphs full $\exists$SO is strictly more expressive than $\exists\mathrm{SO}^{\leq 2}$. Also as pointed out in [2], no concrete example of a graph property in PSPACE that is not in binary NP has been found yet, even though it is known that such properties exist.
Hence, it would be worthwhile to find an example of a
PSPACE query on graphs that cannot be expressed in
$\exists\mathrm{SO}^{\leq 2}$. The gained experience
on writing non-trivial queries in second-order logic, can prove to be
a valuable platform to make progress on these kind of open problems. In
particular, we used a second-order variable of arity 4 in Section 6.
We used it to represent (together with other variables) a linear digraph which, for each of the
leaf valuations, encodes a
sequence of word models corresponding to the different stages of
evaluation of the quantifier free part
of the input QBF${}_{k}$ formula. Since
the size of the Boolean formula in each stage is linear in the size of the input
QBF${}_{k}$ formula, and the number of connectives in the formula is also
linear, the length of the complete sequence of Boolean formulae is
quadratic. Therefore, we conjecture that arity $4$ is actually a lower bound,
though we have not attempted to prove it yet.
In general, the exploration of properties which force us to work
with intermediate structures of size greater than linear w.r.t. the input,
seems a reasonable way of approaching these kind of open problems.
As noted earlier, there are second-order queries that are difficult to express in the language of second-order logic,
but which have an elegant and simple characterization in third-order logic. Therefore it would be interesting to explore possible characterizations of fragments of third-order logic that admit translations of their formulae to equivalent formulae in second-order logic.
This way, those fragments of third-order logic could be assimilated to high-level programming languages, while second-order logic would be the corresponding low-level programming language. In turn, this would allow us to express complex second-order queries with greater abstraction of the low-level details, thus minimizing the probability of error.
Bibliography
[1]
José Luis Balcázar, Joseph Díaz, and Joaquim Gabarró.
Structural Complexity I.
Texts in Theoretical Computer Science, EATCS. Springer, Berlin
Heidelberg New York, 2 edition, 1995.
[2]
Arnaud Durand, Clemens Lautemann, and Thomas Schwentick.
Subclasses of binary NP.
J. Log. Comput., 8(2):189–207, 1998.
[3]
Heinz-Dieter Ebbinghaus and Jörg Flum.
Finite Model Theory.
Perspectives in Mathematical Logic. Springer, Berlin Heidelberg New
York, 2nd edition, 1999.
[4]
Ronald Fagin.
Generalized first-order spectra and polynomial-time recognizable
sets.
In R. Karp, editor, Complexity of Computations, volume 7 of
SIAM-AMS Proc., pages 27–41, Providence, RI, 1974. American
Mathematical Society.
[5]
Flavio Antonio Ferrarotti.
Expressibility of Higher-Order Logics on Relational Databases:
Proper Hierarchies.
PhD thesis, Department of Information Systems, Massey University,
Wellington, New Zealand, June 2008.
[6]
M. R. Garey and David S. Johnson.
Computers and Intractability: A Guide to the Theory of
NP-Completeness.
W. H. Freeman, 1979.
[7]
Lauri Hella and José María Turull Torres.
Computing queries with higher-order logics.
Theor. Comput. Sci., 355(2):197–214, 2006.
[8]
Neil Immerman.
Descriptive Complexity.
Graduate Texts in Computer Science. Springer, Berlin Heidelberg New
York, 1999.
[9]
Michal Krynicki and Jose Maria Turull Torres.
Games on trees and syntactical complexity of formulas.
Logic Journal of the IGPL, 15(5-6):653–687, 2007.
[10]
Leonid Libkin.
Elements Of Finite Model Theory.
Texts in Theoretical Computer Science, EATCS. Springer, Berlin
Heidelberg New York, 2004.
[11]
Wei Ren.
Logic languages: Cubic graphs in second-order logic, 2008.
Research Report, Massey University, Directed by José María
Turull Torres.
[12]
Wei Ren.
On the descriptive complexity of satisfability on quantified boolean
formulas.
http://mro.massey.ac.nz/handle/10179/2629, 2011.
Master of Information Sciences Thesis, Massey University, Directed by
José María Turull Torres.
[13]
Larry J. Stockmeyer.
The polynomial-time hierarchy.
Theor. Comput. Sci., 3(1):1–22, 1976.
[14]
Celia Wrathall.
Complete sets and the polynomial-time hierarchy.
Theor. Comput. Sci., 3(1):23–33, 1976.
Received 21 December 2012. |
The Use of Generalized Information Dimension
in Measuring Fractal Dimension of Time Series
Y. Ashkenazy
Department of Physics, Bar-Ilan University, Ramat-Gan 52900, Israel
(December 2, 2020)
Abstract
An algorithm for calculating generalized fractal dimension of a
time series using the
general information function is presented. The algorithm is based on a strings
sort technique and requires $O(N\log_{2}N)$ computations. A rough estimate
for the number of points needed for the fractal dimension calculation is given.
The algorithm was tested on analytic example as well as well-known examples,
such as, the Lorenz attractor, the Rossler attractor, the van der Pol
oscillator, and the
Mackey-Glass equation, and compared, successfully, with previous results
published in the literature. The computation time for the algorithm
suggested in this paper is much less then the computation time according to
other methods.
pacs: PACS numbers:05.45.+b, 47.52.+j, 47.53.+n
††preprint:
[
]
I Background
In the recent decades the study of chaos theory has gathered momentum.
The complexity that can be found in many physical and biological systems
has been analyzed by the tools of chaos theory.
Characteristic properties, such as the Lyapunov
exponent, Kolmogorov entropy and the fractal dimension (FD), have been
measured in experimental systems. It is fairly easy to calculate signs of
chaos if the system can be represented by a set of non-linear ordinary
differential equations. In many cases it is very difficult to built a
mathematical model that can represent sharply the experimental system.
It is essential, for this purpose, to reconstruct a new phase space based on
the information that one can produce from the system. A global value that is
relatively simple to compute is the FD. The FD can give
an indication of the dimensionality and complexity of the system.
Since actual living biological systems
are not stable and the system complexity varies with time,
one can distinguish between different states of the system by the FD.
The FD can also determine whether a particular system is more complex than
other systems. However, since biological systems are very complex, it is better
to use all the information and details the system can provide. In this paper
we will present
an algorithm for calculating FD based on the geometrical structure
of the system. The method can provide important information, in addition, on
the geometrical of the system (as reconstructed from a time series).
The most common way of calculating FD is through the correlation function,
$C_{q}(r)$ (eq. 6). There is also other method of FD calculation based
on Lyapunov exponents and Kaplan-Yorke conjecture [2]
(eq. 27). However, the computation of the Lyapunov exponent spectrum
from a time series is very difficult and people usually try to avoid this
method111The basic difficulty is that for high FD there are some
exponents which are very close to zero; one can easily add an extra
unnecessary exponent
that can increase the dimensionality by one; this difficulty is most
dominant in Lyapunov exponents which have been calculated from a time series..
The algorithm which is presented in this paper is important, since it gives a
comparative method for calculating FD according to the correlation function.
The need for a additional method of FD calculation is critical in some types
of time series,
such as EEG series (which are produced by brain activity), since several
different
estimations for FD were published in the literature
[3, 4, 5, 6, 7].
A comparative algorithm
can help to reach final conclusions about FD estimate for the signal.
A very simple way to reconstruct a phase space from single time series
was suggested by Takens [8]. Giving a time series ${x_{i}},i=1\ldots N_{p}$, we build a new $n$ dimensional phase space in the
following way :
$$\left.\begin{array}[]{l}\vec{y}_{0}=\{x(t_{0}),x(t_{0}+\tau),x(t_{0}+2\tau),%
\ldots,x(t_{0}+(n-1)\tau)\}\\
\vec{y}_{1}=\{x(t_{1}),x(t_{1}+\tau),x(t_{1}+2\tau),\ldots,x(t_{1}+(n-1)\tau)%
\}\\
\vdots\\
\vec{y}_{N}=\{x(t_{N}),x(t_{N}+\tau),x(t_{N}+2\tau),\ldots,x(t_{N}+(n-1)\tau)%
\}\\
\\
t_{i}=t_{0}+i\Delta t\qquad\tau=m\Delta t\\
N=N_{p}-(n-1)m\qquad m=integer,\end{array}\right.$$
(1)
where $\Delta t$ is the sampling rate, $\tau$ corresponds to the interval
on the time series
that creates the reconstructed phase space (it is usually chosen to be the
first zero of the autocorrelation function [9], or the first minimum
of mutual information [10];
in this work we will use $m$ instead of $\tau$ as an index), and $N$ is
number of reconstructed vectors.
For ideal systems (an infinite number of points without external noise)
any $\tau$ can
be chosen. Takens proved that such systems converge to the real dimension
if $n\geq 2D+1$, where $D$ is the FD of the real system. According
to Ding et al [11] if $n\geq D$ then the FD of the
reconstructed phase space is equal to the actual FD.
In this paper we will first (sec. II) present regular analytic
methods for calculating the FD of a system (generalized correlation method and
generalized information method). An efficient
algorithm for calculating the FD from a time series based on string sorting
will be described in the next section (sec. III). The following
step is to check and compare the general correlation
methods with the general information method (see eq. (2))
using known examples (sec. IV).
Finally, we summarize in sec. V.
II Generalized information dimension and
generalized correlation dimension
II.1 Generalized information dimension
The basic way to calculate an FD is with the Shannon entropy,
$I_{1}(\varepsilon)$222Generally, one has to add a superscript, $n$,
the embedding
dimension in which the general information is calculated. At this stage we
assume that the embedding dimension is known. [12], of
the system. The entropy is just a particular case of the general information
which is defined in the following way [13] :
$$I_{q}(\varepsilon)=\frac{1}{1-q}\ln(\sum_{i=1}^{M(\varepsilon)}p_{i}^{q}),$$
(2)
where we divide the phase space to $M(\varepsilon)$ hypercubes of edge
$\varepsilon$. The probability to find a point in the $i^{th}$
hyper cube is denoted by $p_{i}$. The generalization, $q$, can be any
real number; we usually
use just an integer $q$. When we increase $q$, we give more
weight to more populated
boxes, and when we decrease $q$ the less occupied boxes become dominant.
For $q=1$, by applying the l’Hospital rule, eq. (2) becomes :
$$I_{1}(\varepsilon)=-\sum_{i=1}^{M(\varepsilon)}p_{i}\ln p_{i},$$
(3)
where $I_{1}(\varepsilon)$ is referred to also as the Shannon entropy
[12] of the system.
The definition of the general information dimension (GID) is :
$$D_{q}=-\lim_{\varepsilon\to 0}\frac{I_{q}(\varepsilon)}{\ln\varepsilon},$$
(4)
for $N\to\infty$ ($N$ is the number of points).
However, in practice, the requirement ${\varepsilon\to 0}$ is not
achieved, and an average over a range of $\varepsilon$ is required.
In some cases this average is not sufficient
because several values of $I_{q}(\varepsilon)$
can be computed for the same $\varepsilon$. To illustrate this
we use a $2D$ Cantor
set, presented in Fig. 1. We start from a square. From 9 sub-squares
we erase the 5 internal squares. We continue with this evolution for each
remainder square. This procedure is continued, in principle,
to infinity. The FD, $D_{0}$, of the
$2D$ Cantor set is $\ln 4/\ln 3$ (when $q=0$, $I_{0}$ is just the number of
nonempty squares and $D_{0}$ is the logarithmic ratio between nonempty
squares and $\varepsilon$, where the square edge is normalized to one) as shown
in Fig 1. However, it is possible to locate the $2D$ grid in such
a way that there are 16 nonempty squares, giving rise to
$D_{0}=2\ln 4/\ln 3$, twice the time of real FD of the system. This
illustration shows that when $\varepsilon$ is not small, different
positions of the grid can lead to different FD’s. In this case it is clear
that we have to locate the grid in such a way
that minimum squares will be nonempty
(it is easy to show that this claim for the minimum is true for every $q$).
In general, we can say that one must locate the hyper-grid so that
the general information is minimum :
$$\tilde{I_{q}}(\varepsilon)\equiv\min\frac{1}{1-q}\ln(\sum_{i=1}^{M(\varepsilon%
)}p_{i}^{q}),$$
(5)
This proper location of the hyper-grid reduces the influence of surface
boxes that are partly contained in the attractor. The requirement
of a minimum
gives a good estimate for the GID when $\varepsilon$ is not small.
II.2 Generalized correlation dimension
In 1983 Grassberger and Procaccia presented a new method for
calculating $D_{2}$ [14]. According to this method, it is possible to
calculate dimension just from the correlation between different
points, without direct connection to the phase space, and therefore
easy to use. Some years later, a more general correlation function
was suggested by Pawelzik and Schuster [15] :
$$C_{q}(r)=\Biggl{[}\frac{1}{N}\sum_{i=1}^{N}\biggl{[}\frac{1}{N}\sum_{j=1}^{N}%
\Theta\bigl{(}r-\left|\vec{x}_{i}-\vec{x}_{j}\right|\bigr{)}\biggr{]}^{q-1}%
\Biggr{]}^{1\over q-1},$$
(6)
where $N$ is number of points, $\vec{x}_{i}$ is a point of the system
and $q$ is the generalization. $\Theta(x)$ is the Heaviside step function
$$\Theta(x)=\left\{\begin{array}[]{ll}0&{\rm when~{}}x\leq 0\\
1&{\rm when~{}}x>0.\end{array}\right.$$
(7)
According to this method we have to calculate the generalized
probability to find any two points within a distance $r$. This method is
some kind of integration on eq. (2). It is not necessary to compute
the real distance (e.g. $\left|\left|\vec{x}\right|\right|=\sqrt{x_{1}^{2}+\cdots+x_{n}^{2}}$
where n is the phase space dimension); it is equivalent to
calculating the probability to find any two points in a hyper-box where one
of them is located in the middle (e.g. $\left|\left|\vec{x}\right|\right|=\max_{1\leq i\leq n}\left|x_{i}\right|$ [16]).
It is easier to compute the last possibility. For the special case
of $q=1$, eq. (6) can be written (by applying the l’Hospital’s rule
[10]) as :
$$\ln C_{1}(r)=\frac{1}{N}\sum_{i=1}^{N}\ln\biggl{[}\frac{1}{N}\sum_{j=1}^{N}%
\Theta\bigl{(}r-\left|\vec{x}_{i}-\vec{x}_{j}\right|\bigr{)}\biggr{]}.$$
(8)
The generalized correlation dimension (GCD) has a similar definition to
the GID (eq. (4)) :
$$D_{q}=\lim_{N\to\infty;r\to 0}\frac{\ln C_{q}(r)}{\ln r}.$$
(9)
Both GID and GCD methods should give identical results.
The GCD method is easy use and gives smooth curves. On the other hand,
the method requires $O(N^{2})$ computations333If one looks just at small
$r$ values, the method requires just $O(N)$ computations [16]..
Also, the smooth curves due to
averaging over all distances are associated with a loss in information based
on the attractor structure.
As we pointed out earlier, we usually have a limited number of points,
forcing us to calculate dimension at large $r$. Thus an error enters the
calculation of FD. The minimum number of points needed for the FD
calculation has been discussed in several papers and different constraints
suggested (for example [18] and [19]). In this paper
we will use the $N_{min}$ of the Eckmann and Ruelle [19] constraint :
$$D_{2}<\frac{2\log_{10}N}{\log_{10}\Bigl{(}\frac{1}{\rho}\Bigr{)}},$$
(10)
under the following conditions :
$$\rho=\frac{r}{r_{0}}\ll 1\qquad\frac{1}{2}N^{2}\rho^{D}\gg 1,$$
(11)
with reference to Grassberger and Procaccia method. Here, $r_{0}$ is the
attractor diameter, and
$r$ is the maximum distance that gives reliable results in eq. (6) when
$q=2$. The normalized distance, $\rho$, must be small because
of the misleading influence of the hyper-surface.
A $\rho$ too large (close to 1) can cause incorrect results since we
take into account also hyper-boxes that are not
well occupied because the major volume is outside the attractor.
However, if one has a long time series, then it is not necessary to compute
all $N^{2}$ relations in eq. (6). One can compute eq. (6) for
certain reference points such that the conditions in eq. (11) will
still hold. Then, eq. (6) can be written as follows :
$$\displaystyle C_{q}(r)$$
$$\displaystyle=$$
$$\displaystyle\Biggl{[}\frac{1}{N_{ref}}\sum_{i=1}^{N_{ref}}\biggl{[}\frac{1}{N%
_{data}-2w-1}\times$$
(12)
$$\displaystyle\sum_{j=1}^{N_{data}}\Theta\bigl{(}r-\left|\vec{x}_{i\cdot s}-%
\vec{x}_{j}\right|\bigr{)}\biggr{]}^{q-1}\Biggr{]}^{1\over q-1},$$
where
$$\left|i\cdot s-j\right|>w\qquad s={\lfloor{\frac{N_{data}}{N_{ref}}}\rfloor}.$$
For each reference point we calculate the correlation function over all data
points. The step for the time series is $s$. To neglect short
time correlation one must also introduce a cutoff, $w$. Usually
$w\approx m$ ($m$ corresponds to the $\tau$ from (1))
[20]. The number of distances $\left|\vec{x}_{i}-\vec{x}_{j}\right|$ will be $P=N_{ref}(N_{data}-2w-1)$ instead of $N^{2}$. The new form
of eq. (10) is:
$$D<\frac{\log_{10}P}{\log_{10}\Bigl{(}\frac{1}{\rho}\Bigr{)}},$$
(13)
and the conditions (11) become
$$\rho=\frac{r}{r_{0}}\ll 1\qquad\frac{1}{2}P\rho^{D}\gg 1.$$
(14)
Although we discussed here a minimum number
of data points needed for the dimension
calculation, one must choose $\rho$ such that the influence from
the surface is negligible (the growth of the surface is exponential
to the embedding dimension). That gives us an upper limit for $\rho$.
On the other hand, in order to find the FD we must average over a range of
$r$, giving us a lower limit for $\rho$; therefore $N_{min}$
is determined according to the lower limit, giving rise to larger amount of
points.
III Algorithm
III.1 Background
In this section we describe an algorithm for GID method444Computer
programs are available at
http://faculty.biu.ac.il/$\sim$ashkenaz/FDprog/.
The algorithm is based on a string sort and can be useful
both for GID and GCD methods.
In previous works there have been several suggestions to compute DIG
[21, 22, 23, 24, 25, 26]. The key
idea of some of those methods [23, 24, 25] is to rescale
the coordinates of each point and to express them in binary form. Then, the
edge-box size can be initialized by the lowest value possible and then
double it each time. Those methods require $O(N\log N)$ operations. Another
method [22] uses a recursive algorithm to find the FD. The
algorithm
starts from a $d$ dimensional box which contains all data points. Then, this
box is divided into $2^{d}$ sub boxes, and so on. The empty boxes are not
considered, and those which contain only one point are marked. This procedure
requires only $O(N)$ operations (for a reasonable series length this fact does
not make a significant difference [22]).
As pointed out in ref. [22], in spite of the efficiency of the
above algorithm (speed and resolution), it is quite difficult to converge to
the FD of a high dimensional system. We were aware to this difficulty, and
suggested a smoothing term to solve it (as will be explained in this
section). Basically, we allow any choice of edge length (in contrast to power
of 2 edge size of the above methods) and optimal location of the grid is
searched. This procedure leads to convergence to the FD.
One of the works that calculates GID was done by Caswell and Yorke
[21]. They calculate FD of $2D$ maps.
They have proved that it is very efficient to
divide the phase space into circles instead of squares in spite of the
neglected area between the circles. However, this
approach is not suitable for higher phase space dimensions. The reason might
be that
the volume of the hyper-balls compared to the entire volume of the attractor
decreases according to the attractor dimension,
and in this way we lose most of the information that is included in the
attractor, since the space between the hyper-balls is not taken into account.
It is easy to show
that the ratio between the volume of the hyper-sphere with radius
$R$, $V_{S}$, and the volume of the hyper-box with edge size of
$2R$, $V_{B}$ (the sphere is in the
box), is
$$\frac{V_{S}(2n)}{V_{B}(2n)}=\frac{1}{2\cdot 4\cdot 6\cdots 2n}\Bigl{(}\frac{%
\pi}{2}\Bigr{)}^{n},$$
(15)
for even phase space dimension, and,
$$\frac{V_{S}(2n-1)}{V_{B}(2n-1)}=\frac{1}{1\cdot 3\cdot 5\cdots(2n-1)}\Bigl{(}%
\frac{\pi}{2}\Bigr{)}^{n-1},$$
(16)
for odd phase space dimension. It is clear that the ratios (15) and
(16) tend to zero for large $n$.
III.2 Algorithm
Let us call the time series $x(i)$. The form of
the reconstructed vector, (1),
depends on the jumping, $m$, that creates the phase
space. We order the time series in the following way :
$$\left.\begin{array}[]{l}x(0),x(m),x(2m),\ldots,x((l-1)m),\\
x(1),x(1+m),x(1+2m),\ldots,x(1+(l-1)m),\\
\vdots\\
x(m-1),x(m-1+m),\ldots,x(m-1+(l-1)m),\\
\end{array}\right.$$
(17)
where $l=\lfloor\frac{N_{p}}{m}\rfloor$. Let us denote the new series
as $\tilde{x}_{i}$
($i=0\ldots(lm-1)$; we lost $N_{p}-lm$ data points). If we take
$n$ consecutive numbers in each row, we create a reconstructed vector
in the $n$ dimensional embedding dimension.
At this stage, we fit a string (for each $\varepsilon$) to the
number series (17). One character is actually a number between
$0\ldots 255$,
and thus, it is possible to divide one edge of the hyper-box to 255 parts
(we need one of the 256 characters for sign). The correspondence is
applied in the following way. We search for the minimum value in $\tilde{x}_{i}$
series, and denote it as $\tilde{x}_{min}$ (similarly, we denote the maximum
values as $\tilde{x}_{max}$).
The corresponding character to $\tilde{x}_{i}$ is :
$$y_{i}=\lfloor\frac{\tilde{x}_{i}-\tilde{x}_{min}}{\varepsilon}\rfloor.$$
(18)
The value of $\varepsilon$ is in the range
$$\frac{(\tilde{x}_{max}-\tilde{x}_{min})}{255}\leq\varepsilon\leq(\tilde{x}_{%
max}-\tilde{x}_{min}).$$
(19)
If we take now $n$ consecutive characters (string), we have the “address”
of the hypercube in which the corresponding vector is found.
Let us represent the string as follows :
$$\left.\begin{array}[]{l}s_{0}=y_{0}y_{1}\ldots y_{n-1},s_{1}=y_{1}\ldots y_{n}%
,\ldots,s_{l-n}=y_{l-n}\ldots y_{l-1},\\
s_{l-n+1}=y_{l}\ldots y_{l+n-1},\ldots,s_{2(l-n)+1}=y_{2l-n}\ldots y_{2l-1},\\
\vdots\\
s_{(m-1)(l-n+1)}=y_{(m-1)l}\ldots y_{(m-1)l+n-1},\ldots,\\
\qquad\qquad\qquad\qquad\qquad s_{m(l-n+1)-1}=y_{ml-n}\ldots y_{ml-1}.\\
\end{array}\right.$$
(20)
Obviously, we do not have to keep each string $s_{i}$ in the memory; we can keep
just the pointers to the beginning of the strings in the characters series,
$y_{i}$. The number of vectors / strings is $m(l-n+1)$.
As mentioned, each string is actually the address of a hyper-box (which the
vector is in it) in a hyper-grid that cover the attractor. The first
character is the position of the first coordinate of the hyper-box on the
first edge of the hyper-grid. The second character
denotes the location on the second edge, and so on. We actually grid the
attractor with a hyper-grid so that any edge in this grid can be divided into
255 parts. Thus the maximal number of boxes is $255^{n}$, where $n$ is the
embedding dimension (there is no limit for $n$). Most of those boxes are
empty, and we keep just the occupied boxes in the hyper-grid.
The next step is to check how many vectors fall in the same box, or, in other
words, how many identical “addresses” there are. For this we have to sort
the vector that points to strings, $s_{i}$, in increasing order (one string is
less then other when the first character that is not equal is less then the
parallel character; e.g., ’abcce’$<$’abcda’ since the fourth character in the
first string, c, is less then the fourth character in the second string, d).
The above process is illustrated in Fig. 2.
The most efficient way to sort $N$ elements is the “quick sort”
[27, 28]. It requires just $O(N\log_{2}N)$ computations.
However, this sort algorithm is not suitable for our propose, since after the
vector is sorted, and we slightly increase the size of the edge of the
hyper-box, $\varepsilon$,
there are just a few changes in the vector, and most of it remains
sorted. Thus, one has to use another method which requires less computations
for this kind of situation, because the quick sort requires
$O(N\log_{2}N)$ computations independently of the initial state. The sort
that we used was a “shell sort” [27], which requires $O(N^{3/2})$
computations in worst case, around $O(N^{5/4})$ computations for random
series, and $O(N)$ operations for an almost sorted vector.
Thus, if we compute the information function for $m$ different
$\varepsilon$ values $O(N\log_{2}N+mN)$ operations will be needed.
After sorting we count how many identical vectors there are in each box.
Suppose that we want to calculate GID for embedding dimensions
$n_{min}\ldots n_{max}$, and generalizations $q_{min}\ldots q_{max}$.
We count in
the sorted pointers vector identical strings containing $n_{max}$ characters.
Once we detect a difference we know that the coordinate of the box
has changed. We detect the first different location. First
difference in the last character of the string means that it is possible
to observe the difference just in the $n_{max}^{th}$ embedding dimension, and
in lower
embedding dimension it is impossible to observe the difference. If the
first mismatch is in the $n^{th}$ character in the string, then the box
changes for embedding
dimensions higher than or equal to $n$. We hold a counter vector for the
different embedding dimensions, and we will assign zero for those embedding
dimensions in which we observed a change.
In this stage, we have to add to the results matrix the probability to fall in
a box according to eqs. (2), (3). We continue with this method
to the end of the pointer vector and then print the results matrix for
the current $\varepsilon$ and continue to the next $\varepsilon$.
It is easy to show that one does not have to worry about the range of the
generalization, $q$, because for large time series with $10^{6}$ data
points the range of computational $q$’s is -100..100.
The method of string sorting can be used also for calculating GCD. Our
algorithm is actually a generalization of the method that was suggested by
Grassberger [16]. The algorithm is specially efficient for small
$r$ ($r<\frac{1}{10}D$ where $D$ is the attractor diameter) and it requires
$O(N\log_{2}N)$ computations instead of $O(N^{2})$ computations (according to the
Grassbeger algorithm it requires $O(N)$ computations instead $O(N^{2})$
computations).
The basic idea of Grassberger was that for small $r_{max}$ values, where
$r_{max}$ is the maximum distance for which the correlation function is
computed, one does not have
to consider distances larger then $r_{max}$, distances which require most of
the computation time. Thus, it is enough to consider just the neighboring
boxes (we use the definition distance $\left|\left|\vec{x}\right|\right|=\max_{1\leq i\leq n}\left|x_{i}\right|$ in eq. (6)),
for which their edge is equal to $r_{max}$, since distances greater then
$r_{max}$ are not
considered in eq. (6). If, for example, one wants to calculate the
correlation of $r\leq r_{max}=\frac{1}{10}D$ (in accordance with
conditions
(10) (11) (13) (14)), then (in $2D$ projection)
it is enough to calculate distances in 9 squares instead 100 squares
(actually, it is enough to consider just 5 squares since eq. (6) is
symmetric). According to Grassberger, one has to build a matrix for which any
cell in it points to the list of data points located in it. It is possible to
generalize this method to a $3D$ projection matrix.
The generalization to higher dimensions can be done very easily according to
the string sort method. As a first step one has to prepare a string $y_{i}$
(18), which is the same operation as gridding the attractor by a
hyper-grid of size $r$. The next step is to sort strings $s_{i}$ (20).
For each reference point in eq. (12) the boxes adjoining the
box with the reference point in it should be found. Now, it is left to find
the distances between
the reference point to other points in neighboring boxes (we keep the pointer
vector from string $y_{i}$ to the initial series and vice versa), and calculate
the correlation according to eq. (12).
The algorithm described above is especially efficient for very complex systems
with high FD. For example, for EEG series produce by brain activity, it well
known (except for very special cases such as epilepsy [17])
that the FD is, at least, four. In this case $n_{min}=4$, and instead
of calculating distances in $l^{4}$ boxes ($l$ is the ratio between the attractor
diameter and $r$) it is sufficient to calculate distances in $3^{4}=81$ boxes.
If, for example, $l=9$, just $3^{4}/9^{4}=\frac{1}{81}$ distances must be computed
(or even less if one take into account also the symmetry of eq. (12)).
In this way, despite the $O(N\log_{2}N)$ operations needed for the initial
sort, one can reduce significantly the computation time.
III.3 Testing of GID algorithm and number of data points needed for
calculating GID
Let us test the algorithm of GID on random numbers. We create a random series
between 0 to 1 (uniform distribution). We then reconstruct the phase space
according to Takens theory (1). For any embedding dimension that we
would like to reconstruct, the phase space we will get a dimension that
is the same as the embedding dimension, since the new reconstructed vectors
are random in the new phase space, and hence fill the space
(in our case, a hyper-cube of edge 1).
For embedding dimension 1 (a simple line), if the edge length is $\varepsilon$,
then the probability to fall in any edge is also $\varepsilon$. The number of
edges $\varepsilon$ that covers the series range $[0,1)$ is $\lfloor\frac{1}{\varepsilon}\rfloor$. The probability to fall in the last edge
is $1\%\varepsilon$ (the residue of the division). For
generalization $q$ one gets :
$$\sum_{i=1}^{M(\varepsilon)}p_{i}^{q}=\lfloor 1/\varepsilon\rfloor\varepsilon^{%
q}+(1\%\varepsilon)^{q}.$$
(21)
For embedding dimension 2, one has to square (21) since the probability
distribution on different sides is equal.
For embedding dimension $n$ eq. (2) becomes :
$$I_{q}(\varepsilon)=\frac{1}{1-q}\ln[\lfloor 1/\varepsilon\rfloor\varepsilon^{q%
}+(1\%\varepsilon)^{q}]^{n}.$$
(22)
Opening eq. (22) according to the binomial formula will give all
different
combinations for the partly contained boxes. Notice that this grid location
fulfills requirement (5) for the minimum information function.
In Fig. 3 we present $-I_{2}(\varepsilon)$, both,
according to the GID algorithm, and according to eq. (22). There is full
correspondence between the curves, and, as we find,
it is possible to see that the slope
(dimension) in the embedding dimension $n$ is equal to the dimension itself.
When $-I_{2}(\varepsilon)\approx-10$, the curves separate. There is lower
convergence (when $-I_{2}(\varepsilon)\approx-12.5$) since the edge is too
small, and the number of reconstructed vectors is equal to the number of
nonempty boxes.
The “saw tooth” that seen in Fig. 3 caused by partial boxes that
fall on the edge of the hyper-box. The “saw tooth” appears when
there is an integer number of edges that covers the big edge. The slope in
the beginning of every saw tooth is equal to zero (that comes from the extremum
condition of (22)). The “saw teeth” become smaller when
$\varepsilon$ decrease, since the relative number of boxes that fall on the
surface in small compared to the entire number of boxes.
It is possible to calculate roughly the number of points needed for the
dimension calculation in a homogeneous system. The separation point of
Fig. 3 is
approximately around $-I_{2}(\varepsilon)\approx-10$, for every embedding
dimension. Thus, the number of hyper-boxes at this point is $e^{10}$
(since $-10=\ln\left(M(\varepsilon)p^{2}\right)$ where $p\sim\frac{1}{M(\varepsilon)}$ in our example). The
graph come to a saturation around $-I_{2}(\varepsilon)\approx-12.5$, giving
rise to $e^{12.5}$ data points. The number of points per box is just,
$e^{2.5}\approx 12$. Thus, generally, when there are less then 12 points per
box, the value of $I_{q}(\varepsilon)$ is unreliable. Let us denote the minimum
edges needed for the calculation as $m_{min}$ ($m_{min}$ is just
$\lfloor 1/\varepsilon\rfloor$, and thus one can define a lower value for
$\varepsilon$). The number of hyper-boxes in embedding dimension $n$ is
$m_{min}^{n}$. Thus, the minimum number of points needed for computing the
FD of an attractor with an FD $D$, is:
$$N=12m_{min}^{D}.$$
(23)
This estimate is less then what was required by Smith [18] ($42^{D}$)
and larger then the requirement of Eckmann and Ruelle [19] (if we
take, for example, $m_{min}={\frac{1}{\rho}}=10$
then according to (10)
$10^{D/2}$ points is needed and according to (23) $12\times 10^{D}$
points
is needed). However, as we pointed out earlier, (23) is a
rough estimation, and one can converge to a desired FD even with less points.
IV Examples
In this section we will present the results of both GID and GCD on well known
systems, such as the Lorenz attractor and the Rossler attractor.
The FD of those
results are well known by various methods, and we will present almost
identical results achieved by our methods. In the following examples, we
choose one of the coordinates to be the time series; we normalize the time
series to be in the range of $0\ldots 1$.
IV.1 Lorenz Attractor
The Lorenz attractor [29] is represented by a set of three
first order non-linear differential equations :
$$\displaystyle{dx\over dt}$$
$$\displaystyle=$$
$$\displaystyle\sigma(y-x)$$
$$\displaystyle{dy\over dt}$$
$$\displaystyle=$$
$$\displaystyle-xz+rx-y$$
(24)
$$\displaystyle{dz\over dt}$$
$$\displaystyle=$$
$$\displaystyle xy-bz.$$
Although it is much simpler to find the FD from the real phase-space $(x,y,z)$,
we prefer the difficult case of finding the FD just from one coordinate to
show the efficiency of Takens theorem. We choose the $z^{th}$ component to be
the time series from which we reconstruct the phase space (the parameter values
are: $\sigma=16$, $r=45.92$, $b=4$). We took $32768$ data points and $m=6$ (the
jumping on the time series which creates the reconstructed phase space; the
time step is $\Delta t=0.02$). The generalization that we used was
$q=2$. The embedding dimensions were $n=1..7$. The FD of Lorenz attractor
is $2.07$ ([30] and others).
In Fig. 4 we show the results of GCD and GID methods. As
expected from the GCD curves, in Fig. 4a we see very smooth curves,
for which the slopes
(in the center parts of the curves) converge approximately to dimension
$2.08$. In Fig. 4b we see some jumping in the GID curves. It can be
clearly seen that for small $\varepsilon$, one can not see the jumping while
for large $\varepsilon$ the non-monotonicity is much stronger. This is typical
behavior for GID curves; large $\varepsilon$ (or larger box edge) reflects more
variability of $-I_{2}(\varepsilon)$ to the hype- grid location. To smooth the
curves in order to get reliable results, it is necessary to find a proper
location of the hyper-grid that gives minimum general information
$\tilde{I}_{2}(\varepsilon)$ (or maximum $-\tilde{I}_{2}(\varepsilon)$).
We have performed 7 comparisons
to find that proper location. As for GCD, we also fitted approximate
linear curves (dashed lines) which lead to dimension $2.09$ (Fig. 4c),
and thus, agree well with previous results. To make sure of the validity of the
dimension results we add here (Fig. 4d) graphs of successive slopes of
curves in Fig. 4c, and conclude with dimension results versus
embedding dimension according to several generalizations ($q=2..5$, Fig.
4e). In both graphs, the approximate dimension is $2.03$.
IV.2 Rossler Attractor
Let us examine briefly our second example, the Rossler attractor
[31]. The
differential equations which generate the trajectory are :
$$\displaystyle\frac{dx}{dt}$$
$$\displaystyle=$$
$$\displaystyle-z-y$$
$$\displaystyle\frac{dy}{dt}$$
$$\displaystyle=$$
$$\displaystyle x+ay$$
(25)
$$\displaystyle\frac{dz}{dt}$$
$$\displaystyle=$$
$$\displaystyle b+z(x-c).$$
The parameter values are : $a=0.15$, $b=0.2$, $c=10$. The time step is $t=0.2$
and $m=8$. We use the $y$ direction to be the time series. The number of data
points is $N=16384$. In Fig. 5 we show the results of GID method
of embedding dimension $n=1..7$. There are many more discontinuities in the
unsmoothed curves (Fig. 5a) compared to the Lorenz attractor,
reflecting a sensitivity to the grid location.
The smooth curves in Fig. 5b are produced after 20 comparisons. Again,
we approximate the slopes of central linear part of the curves in two different
ways and find dimension $2.05$, which is in agreement with dimension $2.03$
calculated by the use of the GCD method, and through the Lyapunov
exponent [30].
IV.3 Van der Pol Oscillator
Another example that will be tested is the van der Pol oscillator
[32]. This system
was investigated in the beginning of the century in order to find a model for
the heart beat variability (among other uses of this equation). The behavior
of the system (with parameter values that will be used) is chaotic although it
looks quite periodic. Thus, we expect a smooth behavior of the geometrical
structure of the attractor. The equation of motion is :
$${\ddot{x}}-\alpha(1-x^{2})\dot{x}+kx=f\cos\Omega t.$$
(26)
The parameter values are : $\Omega=2.446$, $\alpha=5$, $k=1$ and $f=1$. Habib
and Ryne [33] and others [34] have found that the
Lyapunov exponents of the system are $\lambda_{1}\approx 0.098$, $\lambda_{2}=0$
and $\lambda_{3}\approx-6.84$,
and thus, according to the Kaplan and Yorke formula [2],
$$D_{L}=j+{\sum_{i=1}^{j}\lambda_{i}\over|\lambda_{j+1}|}\approx 2+{0.098\over 6%
.84}\approx 2.014$$
(27)
where $j$ is defined by the condition that $\sum_{i=1}^{j}\lambda_{i}>0$
and $\sum_{i=1}^{j+1}\lambda_{i}<0$. The fact that the system has a
“periodic” nature is reflected in this very low
FD (as known, the minimal FD of a chaotic system is 2).
In Fig. 6 we present the calculation of $D_{1}$ which we suppose to be
close to $D_{L}$. The approximate slope lead to an FD of $D_{1}\sim 2.02$, which
is in good agreement to $D_{L}$. Notice that minimization of $I_{1}(\varepsilon)$
did not succeed for large $\varepsilon$ values although we used 7 grid
comparisons; that fact can cause a small error in the slope estimation, since
usually the slope is determined from the central part of the curves.
IV.4 Mackey-Glass Equation
Up to now we examined the relation between GID and GCD on systems with finite
dimension, for which their FD can be calculated by using methods based on
the true integrated vector of the system. A class of systems which is more
close to reality are systems with infinite dimensional dynamics. A delay
differential equation of the type
$$\frac{dx(t)}{dt}=F(x(t),x(t-\tau))$$
(28)
belongs to this class. In order to solve this kind of equation, one needs to
initiate $x(t)$ in the range $0\leq t\leq\tau$, and then, by assuming that the
system can be represented by a finite set of discrete values of $x(t)$, it is
possible to evaluate the system dynamics step by step. The solution is
considered
to be accurate if one converges to the same global properties, such as,
Lyapunov exponents and FD, by different methods [35],
and without dependence on the number of intervals to which the function is
divided [14]555Note that the dynamics which is calculated
by the use of different methods, or, equivalently, by a different number of
discrete integration intervals, does not have to behave identically. In fact,
one can expect similar dynamics if the system is not chaotic, but, if the
system is chaotic, it is sensitive to initial condition and to the integration
method, and even moreover, it is sensitive to integration accuracy
[36, 37]. However, the global properties of a system
converge to the same values, by using different integration methods and using
different integration accuracy [37].. Delay equations, such as
eq. (28), describe systems in which a stimulus has a delay response.
There are many practical examples from control theory, economics, population
biology, and other fields.
One of the known examples is the model of blood cell production in patients
with leukemia, formulated by Mackey and Glass [38] :
$$\dot{x}(t)=\frac{ax(t-\tau)}{1+[x(t-\tau)]^{c}}-bx(t).$$
(29)
Following refs. [35, 14], the parameter values are :
$a=0.1$, $b=0.2$, and $c=10$. We confine ourselves to $\tau=30.0$.
As in ref. [14], we choose the time series to be
$\{x(t),x(t+\tau),x(t+2\tau),\ldots\}$666In fact, the common
procedure of reconstructing a phase space from a time series is described in
eq. (1). According to this method, one has to build the reconstructed
vectors by the use of a jumping choice $m$ which is determined by the first
zero
of the autocorrelation function, or the first minimum of the mutual information
function [9]. However, we used the same series as in
ref. [14] in order to compare results., as well as the integration
method which is described in this reference. The length of the time series is
$N=131072$.
The FD ($D_{2}$) calculation of the Mackey-Glass equation is presented in
Fig. 7. The embedding dimensions are, $n=1\ldots 9$.
In Fig. 7a, the correlation function, $C_{2}(r)$ is
shown. One notices that there are two regions in each one of which
there is convergence to a certain slope. These regions are separate by a
dashed line. In Fig.
7b, the average local slope of Fig. 7a is shown. One can
identify easily two convergences, the first in the neighborhood of
$\sim 3.0$, and the second around $\sim 2.4$. Thus, there are two
approximations for the FD, $D_{2}$, pointing to two different scales. The first
approximation is similar to the FD that was calculated in ref. [14].
However, the GID graphs which are presented in Fig. 7c and d lead to
an FD, $D_{2}\sim 2.45$, which seems to be very close to the second convergence
of Fig. 7b. Notice that the convergence to $D_{2}\sim 2.4$ appears,
in both methods, in the neighborhood of the same box size ($\sim 2.1$).
V Summary
In this work we develop a new algorithm for calculation of a general
information, $I_{q}(n)$, which is based on strings sorting (the method of string
sort can be used
also to calculate the conventional GCD method). According to our algorithm,
one can divide the phase space into
255 parts in each hyper-box edge. The algorithm requires $O(N\log_{2}N)$
computations, where $N$ is the number of reconstructed vectors. A rough
estimate for the number of points needed for the FD calculation was given.
The algorithm, which can
be used in a regular system with known equations of motion, was tested
on a reconstructed phase space (which was built according to Takens theorem).
The general information graphs have non monotonic curves, which can be smoothed
by the requirement for minimum general information. We examine our algorithm
on some well known examples, such as, the Lorenz attractor, the Rossler
attractor, the
van der Pol oscillator and others, and show that the FD that was computed by
the GID method is almost identical to the well known FD’s of those systems.
In practice, the computation time of an FD using the GID method, was much less
than for the GCD method. For a typical time series with $32768$ data points,
the computation time needed for the GCD method was about nine times greater
than the
computation time of GID method (when we do not restrict ourselves to small
$r$ values and we compute all $N^{2}$ relations). Thus, in addition to the fact
that the algorithm developed in
this paper enables the use of comparative methods (which is crucial in some
cases), the algorithm is generally faster.
The author wishes to thank J. Levitan, M. Lewkowicz and L.P. Horwitz for very
useful discussions.
References
[1]
[2]
J.L. Kaplan and J.A. Yorke, in: Functional
Differential Equations and Approximations of Fixed Points, Lecture Notes
in Mathematics 730, eds. H.-O. Peitgen and H.-O. Walther
(Springer Verlag, Berlin, 1979).
[3]
A. Babloyantz, C. Nicolis, and M. Salazar, Phys. Rev. A
111, 152 (1985); A. Babloyantz and A. Destexhe, Proc. Natl. Acad. Sci.
USA 83 3513 (1986).
[4]
E. Basar, ed., Chaos in the Brain
(Springer, Berlin, 1990).
[5]
G. Mayer-Kress, F.E. Yates, L. Benton, M. Keidel, W.S.
Tirsch and S.J. Poepple, Math. Biosci. 90, 155 (1988).
[6]
K. Saermark, J. Lebech, C.K. Bak, and A. Sabers, in:
Springer Series in Brain Dynamics, 2 eds. E Basar and T.H. Bullock (Springer,
Berlin, 1989) p. 149.
[7]
K. Saermark, Y. Ashkenazy, J. Levitan, and M. Lewkowicz,
Physica A 236, 363 (1997).
[8]
F. Takens, in : Dynamical Systems and Turbulence, eds.
D. Rand and L.S. Young, (Springer Verlag, Berlin, 1981).
[9]
Discussions about the choose of $\tau$ see for example :
H.G. Schuster, Deterministic Chaos, (Physik Verlag, Weinheim, 1989);
A.M. Fraser and H.L. Swinney, Phys. Rev. A 33, 1134 (1986);
J.-C. Roux, R.H. Simoyi and H.L. Swinney, Physica D 8, 257 (1983).
[10]
H.D.I. Abarbanel, R. Brown, J.J. Sidorowich,
and L.S. Tsimring, Rev. Mod. Phys. 65, 1331 (1993).
[11]
M. Ding, C. Grebobi, E. Ott, T. Sauer and J.A. Yorke, Physica
D 8, 257 (1993).
[12]
C.E. Shannon and W. Weaver, The Mathematical Theory
of Information, (University of Ill. Press, Urbana, 1949).
[13]
J. Balatoni and A. Renyi, in : Selected Papers of A.
Renyi, ed. P. Turan, (Akademiai K. Budapest, 1976), Vol. 1, p. 588.
[14]
P. Grassberger and I. Procaccia, Physica D 9,
189 (1983).
[15]
K. Pawelzik and H.G. Schuster, Phys. Rev. A 35,
481 (1987).
[16]
P. Grassberger, Phys. Lett. A 148, 63 (1990).
[17]
A. Babloyantz and A. Destexhe, in: From Chemical
to Biological Organization, eds. M. Markus, S. Muller, and G. Nicolis
(Springer Verlag, Berlin, 1988).
[18]
L.A. Smith, Phys. Lett. A 133, 283 (1988).
[19]
J.-P. Eckmann and D. Ruelle, Physica D 56,
185 (1992).
[20]
J. Theiler, Phys. Rev. A 34, 2427 (1986).
[21]
W.E. Caswell and J.A. York, in: Dimensions and
Entropies in Chaotic Systems, ed. G. Mayer-Kress,
(Springer Verlag, Berlin, 1986).
[22]
T.C.A. Molteno, Phys. Rev. E 48, 3263 (1993).
[23]
X.-J. Hou, R. Gilmore, G. Mindlin, and H. Solari, Phys. Lett.
A 151, 43 (1990).
[24]
A. Block, W. von Bloh, and H. Schnellnhuber, Phys. Rev. A
42, 1869 (1990).
[25]
L. Liebovitch and T. Toth, Phys. Lett. A 141, 386
(1989).
[26]
A. Kruger, Comp. Phys. Comm. 98, 224 (1996).
[27]
W.H. Press, S.A. Teukolsky, W.T. Vetterling, and B.P. Flannery,
Numerical Recipes in C, 2nd edn. (Cambridge University, Cambridge, 1995).
[28]
D.E. Knooth, The Art of Computer Programming vol. 3
- Sorting and Searching, (Addison-Wesley, Reading, Mass., 1975);
There is other sort algorithms that require even less computation
($O(N)$) such as radix sorting, A.V. Aho, J.E. Hopcroft, and J.D. Ullman,
Data Structures and Algorithms, (Addison-Wesley, Mass., 1983).
[29]
E.N. Lorenz, J. Atmos. Sci. 20, 130 (1963).
[30]
A. Wolf, J.B. Swift, H.L. Swinney and J.A. Vastano,
Physica D 16, 285 (1985).
[31]
O.E. Rossler, Phys. Lett. 57A, 397 (1976).
[32]
B. van der Pol, Phil. Mag. (7) 2, 978 (1926);
B. van der Pol and van der Mark, Phil. Mag. (7) 6, 763 (1928).
[33]
S. Habib and R.D. Ryne, Phys. Rev. Lett. 74,
70 (1995).
[34]
K. Geist, U. Parlitz, and W.Lauterborn, Prog. Theor. Phys.
83, 875 (1990).
[35]
J.D. Farmer, Physica D 4, 366 (1981).
[36]
J.M. Greene, J. Math. Phys. 20, 1183 (1979).
[37]
Y. Ashkenazy, C. Goren, and L.P. Horwitz, Phys. Lett.
A, 243, 195 (1998).
[38]
M.C. Mackey and L. Glass, Science 197, 287 (1977). |
Preservation of complete Baireness
Sergey Medvedev
South Ural State University, Chelyabinsk, Russia
[email protected]
Abstract.
The main result is the following.
Let $f\colon X\rightarrow Y$ be a continuous mapping of a completely Baire space $X$ onto a hereditary weakly Preiss-Simon regular space $Y$ such that the image of every open subset of $X$ is a resolvable set in $Y$. Then $Y$ is completely Baire.
The classical Hurewicz theorem about closed embedding of the space of rational numbers into metrizable spaces is generalized to weakly Preiss-Simon regular spaces.
Key words and phrases:
resolvable set, completely Baire space, weakly Preiss-Simon space, rational numbers
2010 Mathematics Subject Classification:
54H05, 03E15
One of the main tasks considered in topology is to find the properties of topological spaces that are preserved when moving to the image of the space with this property. For example, Sierpiński [11] and Vainštain [12] have established that if $Y$ is an image of a completely metrizable space $X$ under a closed mapping or an open mapping, then $Y$ is completely metrizable.
A natural question to ask is which maps preserve complete metrizability.
Ostrovsky [10] proved that if a continuous mapping from a zero-dimensional Polish (separable, completely metrizable) space $X$ onto a metrizable space $Y$ takes any clopen set $W\subset X$ to the union of an open set and a closed set, then $Y$ is completely metrizable. He raised the question whether the same is true when the images are the intersection of an open set and a closed set.
Gao and Kieftenbeld [4] showed that if the image under $f\colon X\rightarrow Y$ of every open set or every closed set in a Polish space $X$ is a resolvable set in $Y$, then $Y$ is Polish. Holický and Pol [6] established similar results for the nonseparable case. The latter result solves Ostrovsky’s problem for metrizable spaces.
It is notorious [3, Theorem 4.3.26] that a topological space is completely metrizable if and only if it is a Čech-complete metrizable space.
Holický [5, Corollary 3.3] proved among other things that if $f\colon X\rightarrow Y$ is a continuous mapping from a Čech-complete space $X$ onto a metrizable space $Y$ such that $f$ takes open sets in $X$ to resolvable sets in $Y$, then $Y$ is completely metrizable.
Let us pass now from a Čech-complete metrizable space to a completely Baire space.
Suppose we are given a continuous mapping $f\colon X\rightarrow Y$ from a zero-dimensional metrizable space $X$ onto a metrizable space $Y$ such that $f$ takes clopen sets in $X$ to resolvable sets in $Y$. Holický and Pol [6, Corollary 4] showed that if $X$ is completely Baire, then so is $Y$.
Our main Theorem 3 generalizes this statement to regular spaces.
In order to obtain it, we strengthen the Hurewicz theorem [7] about closed embedding of the space of rational numbers into metrizable spaces, see Theorem 1.
Notation. For all undefined terms see [3].
A set $E$ in a space $X$ is said to be resolvable [8, § 12] if it can be represented as
$$E=(F_{1}\setminus F_{2})\cup(F_{3}\setminus F_{4})\cup\ldots\cup(F_{\xi}\setminus F_{\xi+1})\cup\ldots,$$
where $\langle F_{\xi}\rangle$ forms a decreasing transfinite sequence of closed sets in $X$. Equivalently, $E$ is resolvable if, for every closed set $F\subset X$, the boundary of $F\bigcap E$ relative to $F$ is nowhere dense in $F$. Resolvable sets form a field of subsets of $X$ which contains closed sets.
A topological space $X$ is called a Baire space if the intersection of countably many dense open sets in $X$ is dense. We call a space completely Baire (or hereditarily Baire cf. [6]) if every closed subspace of it is Baire. By Hurewicz’s theorem [7], a metrizable space is completely Baire if and only if it contains no closed homeomorphic copy of rational numbers.
A topological space $X$ is called Preiss-Simon at a point $x\in X$ if for any subset $A\subset X$ with $x\in\overline{A}$ there is a sequence $\langle U_{n}\colon n\in\omega\rangle$ of nonempty open subsets of $A$ that converges to $x$ in the sense that each neighborhood of $x$ contains all but finitely many sets $U_{n}$. A space $X$ is called a Preiss-Simon space if $X$ is Preiss-Simon at each point $x\in X$. It is clear that every first countable space is Preiss-Simon and each Preiss–Simon space is Frèchet–Urysohn.
We shall say that a space $X$ is weakly Preiss-Simon at a point $x\in X$ if there is a sequence $\langle U_{n}\colon n\in\omega\rangle$ of nonempty open subsets of $X$ that converges to $x$. Denote by $wPS(X)$ the set of points $x\in X$ at which $X$ is weakly Preiss-Simon. A space $X$ is called weakly Preiss-Simon space if the set $wPS(X)$ is dense in $X$. A space $X$ is called hereditary weakly Preiss-Simon space if every nonempty closed subset of $X$ is a weakly Preiss-Simon space, i. e., the set $wPS(F)$ is dense in $F$ for every nonempty closed set $F\subset X$. Clearly, every Preiss-Simon space is hereditary weakly Preiss-Simon.
Lemma 1.
Let $F$ be a nowhere dense closed subset of a regular space $X$. Suppose $X$ is weakly Preiss-Simon at a non-isolated point $x$. Then for every neighborhood $O$ of $x$ there is a sequence $\langle U_{n}\colon n\in\omega\rangle$ of nonempty open sets such that:
(a)
$\langle U_{n}\colon n\in\omega\rangle$ converges to $x$,
(b)
$\overline{U_{n}}\subset O\setminus\{x\}$ and $\overline{U_{n}}\bigcap F=\emptyset$ for every $n\in\omega$,
(c)
$\overline{U_{n}}\bigcap\overline{U_{k}}=\emptyset$ provided $n\neq k$,
(d)
$\overline{\bigcup\{U_{n}\colon n\in\omega\}}=\{x\}\cup\bigcup\{\overline{U_{n}}\colon n\in\omega\}$.
Proof.
Take a sequence $\langle V_{k}\colon k\in\omega\rangle$ of nonempty open sets in $X$ that converges to $x$. The neighborhood $O$ of $x$ contains all but finitely many sets $V_{k}$. Choose the least number $k_{0}$ with $V_{k_{0}}\subset O$. The point $x$ is non-isolated in $X$. Then there is a point $x_{0}\in V_{k_{0}}\setminus(\{x\}\bigcup F)$ and its neighborhood $U_{0}$ satisfying $\overline{U_{0}}\subset V_{k_{0}}\setminus(\{x\}\bigcup F)$. Similarly, choose the least $k_{1}$ with $V_{k_{1}}\subset O\setminus\overline{U_{0}}$. Find a point $x_{1}\in V_{k_{1}}\setminus\{x\}$ and its neighborhood $U_{1}$ satisfying $\overline{U_{1}}\subset V_{k_{1}}\setminus(\{x\}\bigcup F)$. Clearly, $\overline{U_{0}}\bigcap\overline{U_{1}}=\emptyset$. Next, choose the least $k_{2}$ with $V_{k_{2}}\subset O\setminus(\overline{U_{0}}\bigcup\overline{U_{1}})$. Find a point $x_{2}\in V_{k_{2}}\setminus\{x\}$ and its neighborhood $U_{2}$ satisfying $\overline{U_{2}}\subset V_{k_{2}}\setminus(\{x\}\bigcup F)$, and so on. One can check that all conditions (a)–(d) hold.
∎
The following statement generalizes the Hurewicz theorem [7] about closed embedding of the space of rational numbers $\mathbb{Q}$ into metrizable spaces. Note that Theorem 1 was proved in [9] for a regular space $X$ of the first category with a dense first-countable subset.
Theorem 1.
Let $X$ be a regular space of the first category such that the set $wPS(X)$ is dense in $X$. Then $X$ contains a closed copy of the space of rational numbers.
Proof.
Let $X=\bigcup\{X_{n}\colon n\in\omega\}$, where each $X_{n}$ is a closed nowhere dense subset of $X$. Without loss of generality, $X_{0}\subset X_{1}\subset\ldots$.
By induction on the tree $\omega^{<\omega}$, we shall construct points $x_{t}\in wPS(X)$, open sets $U_{t}$, and sequences $\langle U_{t\hat{\,}k}\colon k\in\omega\rangle$ of open sets such that the following conditions hold for each $t\in\omega^{<\omega}$:
(a)
$U_{t}$ is a neighborhood of the point $x_{t}$,
(b)
$\overline{U_{t}}\bigcap X_{n}=\emptyset$ whenever $t\in\omega^{n+1}$,
(c)
$\overline{U_{t\hat{\,}k}}\subset U_{t}\setminus\{x_{t}\}$ for each $k\in\omega$,
(d)
the sequence $\langle U_{t\hat{\,}k}\colon k\in\omega\rangle$ converges to $x_{t}$,
(e)
$\overline{U_{t\hat{\,}i}}\bigcap\overline{U_{t\hat{\,}k}}=\emptyset$ provided $i\neq k$,
(f)
$\overline{\bigcup\{U_{t\hat{\,}k}\colon k\in\omega\}}=\{x_{t}\}\cup\bigcup\{\overline{U_{t\hat{\,}k}}\colon k\in\omega\}$.
The space $X$ has no isolated points as a set of the first category.
For the base step of the induction, take a point $x_{\emptyset}\in wPS(X)$ and $U_{\emptyset}=X$. By Lemma 1 with $F=X_{0}$ and $O=X$, there is a sequence $\langle U_{k}\colon k\in\omega\rangle$ of nonempty open sets satisfying conditions (a)–(f) for $n=0$.
Suppose the points $x_{s}$ and the sequences $\langle U_{s\hat{\,}k}\colon k\in\omega\rangle$ of open sets have been constructed for every $s\in\omega^{j}$ with $j\leq n$. Fix $s\in\omega^{n}$. The set $X_{n+1}$ is nowhere dense in the open set $U_{s\hat{\,}i}$ for every $i\in\omega$. Then we can find a point $x_{s\hat{\,}i}\in wPS(X)\bigcap(U_{s\hat{\,}i}\setminus X_{n+1})$. By Lemma 1 with $F=X_{n+1}$ and $O=U_{s\hat{\,}i}$, there is a sequence $\langle U_{s\hat{\,}i\hat{\,}k}\colon k\in\omega\rangle$ of nonempty open sets satisfying conditions (a)–(f) for $t=s\hat{\,}i$. This completes the induction step.
Clearly, the set $Z=\bigcup\{x_{t}\colon t\in\omega^{<\omega}\}$ is countable and has no isolated points. For every $t\in\omega^{<\omega}$ from conditions (d) and (e) it follows that the sequence $\langle V_{t\hat{\,}n}\colon n\in\omega\rangle$ forms a countable base at the point $x_{t}$ with respect to $Z$, where
$$V_{t\hat{\,}n}=Z\bigcap\left(U_{t}\setminus\cup\{\overline{U_{t\hat{\,}k}}\colon k<n\}\right).$$
Hence, $Z$ is a second-countable space. The Urysohn theorem [3, Theorem 4.2.9] implies that $Z$ is metrizable. According to the Sierpiński theorem [11] (see also [3, Exercise 6.2.A]), $Z$ is homeomorphic to the space of rational numbers.
It remains to verify that $Z$ is closed in $X$. For each $n\in\omega$ define the sets
$$\displaystyle Z_{n}$$
$$\displaystyle=\bigcup\left\{x_{s}\colon s\in\omega^{j},j\leq n\right\},$$
$$\displaystyle P_{n}$$
$$\displaystyle=\bigcup\left\{\overline{U_{t\hat{\,}k}}\colon t\in\omega^{n},k\in\omega\right\}.$$
Then $Z=\bigcup\{Z_{n}\colon n\in\omega\}$, $Z_{n}\cap P_{n}=\emptyset$, and $Z\subset Z_{n}\bigcup P_{n}$ for each $n$.
By induction on $n\in\omega$, let us check that the union $Z_{n}\cup P_{n}$ is closed in $X$ for each $n$.
For $n=0$, the set $Z_{0}\cup P_{0}$ is closed in $X$ by Lemma 1. We now show that $Z_{n+1}\cup P_{n+1}$ is closed under the induction assumption on closeness of $Z_{n}\cup P_{n}$. Suppose that, on the contrary, there is a point
$$x\in\overline{(Z_{n+1}\cup P_{n+1})}\setminus(Z_{n+1}\cup P_{n+1}).$$
Since $(Z_{n+1}\setminus Z_{n})\cup P_{n+1}\subset P_{n}$, we have $x\in\overline{Z_{n}\cup P_{n}}=Z_{n}\cup P_{n}$. Then $x\in P_{n}\setminus Z_{n+1}$.
From (c) and (e) it follows that there is a unique index $t\hat{\,}k\in\omega^{n+1}$ with $x\in\overline{U_{t\hat{\,}k}}\subset P_{n}$. According to (f), we have
$$\overline{P_{n+1}}\cap\overline{U_{t\hat{\,}k}}=\{x_{t\hat{\,}k}\}\cup\bigcup\left\{\overline{U_{t\hat{\,}k\hat{\,}i}}\colon i\in\omega\right\}.$$
By (e), $x\in\overline{U_{t\hat{\,}k\hat{\,}j}}$ for some $j$. Then $x\in P_{n+1}$, a contradiction.
Thus, the set $Z_{n}\cup P_{n}$ is closed in $X$ for each $n$.
By construction, $Z\subset Z_{n}\cup P_{n}$ for each $n$. Then $\overline{Z}\subset\bigcap\{Z_{n}\cup P_{n}\colon n\in\omega\}$.
Striving for a contradiction, suppose there is a point $x\in\overline{Z}\setminus Z$. Fix the least $i$ with $x\in X_{i}$. Obviously, $x\notin Z_{i}$. Condition (b) implies that $P_{i}\cap X_{i}=\emptyset$. Then $x\notin Z_{i}\cup P_{i}$, a contradiction.
∎
Let us recall the well-known theorem of van Douwen [2]. For a first countable regular space $X$, he proved that every closed subspace of $X$ is a Baire space if and only if $X$ has no countable closed crowded subspace. Here a space is said to be crowded if it has no isolated points. Note that up to homeomorphism the space of rational numbers is the only countable first-countable crowded regular space [11]. In other words, for a first countable regular space $X$, van Douwen proved that $X$ is completely Baire if and only if the space of rational numbers does not embed as a closed subspace into $X$.
Now we obtain a generalization of this result.
Theorem 2.
A hereditary weakly Preiss-Simon regular space $X$ is completely Baire if and only if the space of rational numbers does not embed as a closed subspace into $X$.
Proof.
Clearly, if $X$ contains a closed copy of the space of rational numbers, then it cannot be completely Baire. For the other direction, aiming for a contradiction, suppose that $X$ is not completely Baire. Then $X$ contains a closed subset $F$ of the first category. It remains to apply Theorem 1 to $F$.
∎
Lemma 2.
Let $f\colon X\rightarrow\mathbb{Q}$ be a continuous mapping of a Baire space $X$ onto the space of rational numbers $\mathbb{Q}$.
Then there is an open set $U\subset X$ such that $f(U)$ is not resolvable.
Proof.
Denote by $X^{\circ}_{q}$ the interior of $f^{-1}(q)$ for every point $q\in\mathbb{Q}$. Put $\mathbb{Q}^{\circ}=\{q\in\mathbb{Q}\colon X^{\circ}_{q}\neq\emptyset\}$.
Let us verify that the open set $A=\bigcup\{X^{\circ}_{q}\colon q\in\mathbb{Q}^{\circ}\}$ is dense in $X$. Suppose the contrary and choose a nonempty open set $W\subset X$ missing $A$. For each point $q\in\mathbb{Q}\setminus\mathbb{Q}^{\circ}$ the intersection $f^{-1}(q)\cap W$ is a closed nowhere dense subset of $W$. For each point $q\in\mathbb{Q}^{\circ}$ the intersection $f^{-1}(q)\cap W=(f^{-1}(q)\setminus A)\cap W$ is the same. This implies that $W$ is a set of the first category on itself. On the other hand, $W$ is a Baire space as an open subset of the Baire space $X$, a contradiction.
Then $\mathbb{Q}^{\circ}=f(A)$ is dense in $\mathbb{Q}$ because $f$ is continuous.
Take $B\subset\mathbb{Q}^{\circ}$ such that $B$ is dense, co-dense in $\mathbb{Q}$. Clearly, $B$ is not a resolvable subset of $\mathbb{Q}$. The open set $U=\bigcup\{X^{\circ}_{q}\colon q\in B\}$ is as required.
∎
Now we are ready to give the main result of the paper.
Theorem 3.
Let $f\colon X\rightarrow Y$ be a continuous mapping of a completely Baire space $X$ onto a hereditary weakly Preiss-Simon regular space $Y$ such that the image of every open subset of $X$ is a resolvable set in $Y$. Then $Y$ is completely Baire.
Proof.
Striving for a contradiction, suppose that $Y$ is not completely Baire. By Theorem 2, $Y$ contains a closed copy $Q$ of the space of rational numbers.
The inverse image $f^{-1}(Q)$ is a Baire space as a closed subset of $X$. By Lemma 2, there is an open in $f^{-1}(Q)$ set $U\subset f^{-1}(Q)$ such that $f(U)$ is not resolvable. Take an open set $U^{*}\subset X$ with $U^{*}\bigcap f^{-1}(Q)=U$. Then $f(U^{*})$ is not resolvable because the closed subset $Q\bigcap f(U^{*})=f(U)$ of a resolvable set must be resolvable. A contradiction.
∎
References
[1]
[2]
E.K. van Douwen, Closed copies of rational numbers, Comm. Math. Univ. Carol., 28 (1987) 137–139.
[3]
R. Engelking, General topology, PWN, Warszawa, 1977.
[4]
S. Gao and V. Kieftenbeld,
Resolvable maps preserve complete metrizability, Proc. Amer. Math. Soc., 138:6 (2010) 2245–2252.
[5]
Petr Holický,
Preservation of completeness by some continuous maps,
Topology Appl., 157 (2010) 1926–1930.
[6]
P. Holický and R. Pol,
On a question by Alexey Ostrovsky concerning preservation of completeness, Topology Appl., 157 (2010) 594–596.
[7]
W. Hurewicz, Relativ perfekte Teile von Punktmengen und Mengen (A), Fund. Math. 12 (1928) 78–109.
[8]
K. Kuratowski, Topology, Vol. 1. PWN, Warszawa, 1966.
[9]
S.V. Medvedev,
On a problem for spaces of the first category, Vestn. Mosk. Univ., Ser. I, Mat. Mekh., 41:2 (1986) 84–86 (in Russian).
English transl.: Mosc. Univ. Math. Bull., 41:2 (1986) 62–65.
[10]
A. Ostrovsky,
New basic result in classical descriptive set theory: Preservation of completeness, Topology Appl., 156 (2009) 1749–1751.
[11]
W. Sierpiński, Sur une propriété topologique des ensembles dénombrables denses en soi, Fund. Math., 1 (1920) 11–16.
[12]
I. A. Vainštain,
On closed mappings, Moskov. Gos. Univ. Uč. Zap. 155:5 (1952) 3–53 (in Russian). |
[
[
${}^{1}$Princeton University, Princeton, USA,
${}^{2}$Kapteyn Institute, Groningen, The Netherlands,
${}^{3}$University of Washington, Seattle, USA,
${}^{4}$Max-Planck-Inst. für Astroph., Garching, Germany
(#1; ?? and in revised form ??)
Abstract
We summarize the detection rates at wavelengths other than optical for $\sim$99,000
galaxies from the Sloan Digital Sky Survey (SDSS) Data Release 1 “main”
spectroscopic sample. The analysis is based on positional cross-correlation
with source catalogs from ROSAT, 2MASS, IRAS, GB6, FIRST, NVSS and WENSS
surveys. We find that the rest-frame UV-IR broad-band galaxy SEDs form a remarkably
uniform, nearly one parameter, family. As an example, the SDSS $u$ and $r$
band data, supplemented with redshift, can be used to predict $K$ band magnitudes
measured by 2MASS with an rms scatter of only 0.2 mag; when measurement uncertainties
are taken into account, the astrophysical scatter appears not larger than $\sim$0.1 mag.
††volume: #1\checkfont
eurm10
\checkfontmsam10
The Interplay among Black Holes, Stars and ISM in Galactic
Nuclei] Multiwavelength View of SDSS Galaxies
M. Obrić et al.]M. Obrić${}^{1,2}$
Ž. Ivezić${}^{1,3}$, R.H. Lupton${}^{1}$, G. Kauffmann${}^{4}$, G.R. Knapp${}^{1}$, J.E. Gunn${}^{1}$,
D. Schlegel${}^{1}$, M.A. Strauss${}^{1}$, S. Anderson${}^{3}$, et al.
2004
222
\pagerange1–8
\jnameThe Interplay among Black Holes, Stars and ISM
in Galactic Nuclei
\editorsTh. Storchi Bergmann, L.C. Ho & H.R. Schmitt, eds.
1. The fractions of SDSS galaxies detected at other wavelengths
We report initial results from a program aimed at the characterization of
multiwavelength properties of galaxies detected by SDSS. The details about the
positional cross-correlation method, and analysis of matched samples, will be
described elsewhere (Obrić et al. 2004); here we summarize the detection
rates and demonstrate a tight correlation between the UV and IR colors
of galaxies.
Table 1 lists the detection rates for galaxies from SDSS “main” spectroscopic
sample (a flux-limited sample, $r_{Pet}<17.77$, for details see Strauss et al.
2002) by ROSAT, 2MASS, IRAS, GB6, FIRST, NVSS and WENSS surveys. Galaxies
are separated into those with and without emission lines (using line measurements
and classification criteria from Kauffmann et al. 2003), and emission line
galaxies are further separated, using Baldwin-Phillips-Terlevich diagram,
into “star-forming”, “AGNs”, and “unclassified”. The detection rate of
SDSS galaxies is below 10% in all surveys except 2MASS.
2. Predicting 2MASS K flux from UV/optical SDSS flux
Galaxies form a nearly one-dimensional sequence in various optical
color-color diagrams (the correlation among colors is especially tight
for rest-frame colors, with a scatter of only $\sim$0.03 mag perpendicular
to the locus, Smolčić et al. 2004). We report here that this one-dimensionality
of galaxy spectral energy distributions (SEDs) extends to 2MASS wavelengths:
it is possible to predict 2MASS $K$ band flux with an rms of 0.2 mag
using SDSS $u$ and $r$ band measurements111We use UV/visual fluxes
($u$ and $r$ bands) to predict near-IR flux because this is the “hardest”
wavelength combination with most astrophysical implications (according to
“common wisdom”, such a relationship should not be very accurate due to
the effects of starbursts and dust extinction); predicting, for example,
2MASS $J$ band flux from SDSS $z$ band flux is trivial because these two
bands are adjacent in wavelength space..
The SDSS-based $K$ magnitude prediction is determined from $K_{SDSS}=r_{Pet}-(r-K)$, where $r_{Pet}$ is the SDSS $r$ band Petrosian
magnitude (Strauss et al. 2002), and $(r-K)$ is a best fit to the
observed $r-K$ colors for SDSS/2MASS galaxies.
It depends on $u-r$ color
(essentially a position along the Hubble sequence), and on redshift,
to account for
K-correction effects (we use default 2MASS magnitudes and SDSS model colors):
only2pages…
$\phantom{mala}(r-K)=A+B\,(u-r)+C\,(u-r)^{2}+D\,(u-r)^{3}+E\,z+F\,z^{2}\phantom%
{mirelajezlocesta}(2.1)\\
$
where (A, B, C, D, E, F)=(1.115, 0.940, -0.165, 0.00851, 4.92, -9.10),
and $z$ is redshift. This fit predicts 2MASS $K$ band magnitudes with an
rms scatter of 0.20 mag, which depends on neither color nor redshift, and
is nearly Gaussian (see Figure 1).
To correct for aperture and resolution effects
between SDSS and 2MASS, that presumably depend on galaxy profile, or nearly
equivalently on galaxy color (Strateva et al. 2000), we add to $(r-K)$
($0.496-0.154\,R_{50}^{z}$) for galaxies with $u-r<2.22$ and
($0.107-0.045\,R_{50}^{z}$) for redder galaxies, where $R_{50}^{z}$ is the
radius containing 50% of Petrosian flux in the $z$ band.
This correction has a negligible effect on rms scatter, and only removes a correlation
of $K_{SDSS}-K_{2MASS}$ residuals with galaxy size.
The median residuals, as a function of $u-r$ and $z$, do not exceed 0.03 mag.
The rms scatter decreases to 0.15 mag at the bright end
($K<12$). Given typical measurement errors in $u$, $r$, $R_{50}^{z}$ and
$K$, we conservatively conclude that the true astrophysical scatter of
$K$ band magnitudes predicted from the blue part of SED is less than $\sim$0.1
mag. Similarly, the relation $(J-K)=2.172\,z+0.966$, where $z$ is
redshift, predicts $J-K$ measured by 2MASS with an rms scatter of 0.11
mag (0.07 mag at the bright end), and no significant residuals with
respect to $K$, $u-r$ and redshift. These tight correlations demonstrate
remarkable one-dimensionality of galaxy spectral energy distributions from
UV to IR wavelengths.
References
[Kauffmann et al.(2003)]
Kauffmann, G., Heckman, T.M., White, S.D.M.,
and 19 co-authors 2003, MNRAS 341, 33
[Obrić et al.(2004)]
Obrić, M., Ivezić, Ž., Kauffmann, G.
and 15 co-authors 2004, MNRAS, in prep.
[Smolčić et al.(2004)]
Smolčić, V., Ivezić, Ž.,
Gaćeša, M., and 14 co-authors 2004, MNRAS, in prep.
[Strateva et al.(2001)]
Strateva, I., Ivezić, Ž., Knapp, G.R.,
and 26 co-authors 2001, AJ 122, 1861
[Strauss et al.(2002)]
Strauss, M.A., Weinberg, D., Lupton, R.H.,
and 36 co-authors 2002, AJ 124, 1810 |
\articleNo
301
\volnumber1
\hauthor
M.Yu. Skulskyy
11institutetext:
Lviv Polytechnic National University, Department of Physics, 79013, Lviv, Ukraine 11email: [email protected]
Formation of magnetized spatial structures in the Beta Lyrae system
III. Reflection of magnetically controlled matter in circumbinary structures in helium lines, in particular arising from metastable levels
M.Yu. Skulskyy
(2021; October 7, 2020; November 3, 2020; October 7, 2020)
Abstract
Spatial gaseous structures in the Beta Lyrae system have been studied with the fact of change in the longitudinal component of the donor’s magnetic field during the orbital period in mind. The investigation was based primarily on the study of the dynamics of the circumstellar structures surrounding the binary system as a whole. The special emphasis was placed on the study of complex helium lines, in particular those arising from metastable levels. A number of different observable facts from the ultraviolet to the red spectral region were analyzed. The configuration of the donor magnetic field is a factor that not only enhances mass transfer and influences the formation of spatial gas structures between stellar components but, to some extent, also affects the outflow of matter and the formation of external gas structures around this interacting binary system. Together with previous articles (Skulskyy, 2020a, b), the pieces of evidence of this work, confirming the reflection of magnetically controlled matter in circumbinary structures, define the basis for a coherent picture of the mass exchange between components and outflows of matter outwards.
keywords: binaries: individual: Beta Lyrae – emission-line: magnetic field: mass-loss
††volume: 51
1 Introduction
In previous papers, the focus was on the research of the relationship between the structure of the donor magnetic field and its reflection in the characteristic physical parameters of the visible and infrared Beta Lyrae spectra (Skulskyy, 2020a, b). Taking into account the spatial configuration of the donor magnetic field, the emphasis was put on certain facts of its reflection in the moving magnetized accretions structures between the stellar components of this binary system. The data of absolute spectrophotometry and spectral analysis, changes along with the orbital phases of the radial velocities and intensities of complex emission-absorption lines were examined, primarily based on the own spectral and spectrophotometric observations in the visual range spectrum. This was due to a series of conscious and consistent spectral observations and relevant studies conducted mainly in 1980-1995, after the discovery of the donor magnetic field. Such a method of approach has already made it possible to make certain generalizations and reflect moving magnetized structures as the original phenomenon of mass transfer, which is inherent to this binary system.
It is clear that by the time when nothing was known about the donor’s magnetic field and its possible effect on the physical conditions in the near and far gas structures of this binary system, a number of significant scientific studies had been completed. Some physical characteristics and certain parameters of radiating gas structures near both stellar components or in the common shell were often described by the known paradigms, which were later rejected or modified. This is especially true of researchers who have studied the spectral lines of the shell and the characteristics of the spectrum in the far ultraviolet region. A large number of scientific materials are available for their coverage and further rethinking. In this article, special attention is paid to the scientific works in order to investigate the influence of the magnetic field on the dynamics of the circumstellar structure surrounding this binary system as a whole. Clearly, to create a coherent picture, such studies should be conducted in comparison with the already obtained results of the investigation of the magnetized gas structures both near the donor and the gainer and between them.
The previous researches of Skulskyy (2020a, b) demonstrate an efficient collision of magnetized plasma in the phases of observation of magnetic poles on the donor surface. These shock collisions with the accretion disk in the phases of the secondary quadrature, in which the high-temperature environment and the system of formed accretion flows are observed, are especially noted. There are obvious correlations between the phase variability of the donor magnetic field and the corresponding variability of dynamic and energy characteristics of different complex lines, but above all the strongest emission lines H${}_{\alpha}$ and He I $\lambda$ 7065. Moreover, the phase boundaries of the location of the magnetic pole on the surface of the donor, above which the matter outflows are formed, were discovered. The additional loss of matter from the donor surface is observed mainly from the donor surface in the phase region round 0.855 P, i.e., when the magnetic field pole is facing the gainer. This allowed us to predict certain correlations between the behavior of the above strong lines, as well as certain helium lines formed in shell gas structures extending outward from the binary system. Therefore, in this article, closer attention is paid to such lines of helium, primarily those arising from metastable levels.
It should be recalled that the donor’s magnetic field changes significantly during the orbital period (which is close to 12.94 d); phase changes in the range from zero to one are tied to the main eclipse of this binary system in the visible spectral region when the more massive accretor obscures of the bright donor; the schematic model of the Beta Lyrae system and the picture of mass transfer are shown in Fig.1 of Skulskyy (2020a), which should be kept in mind for a better understanding of the geometry and physical properties of this interacting binary system in further analysis; as this article is a continuation of a number of ongoing comprehensive studies of the proposed topic, references to previous results are quite natural.
2 Magnetic field and gaseous structures surrounding the binary system
2.1 Magnetic field and the He I lines arising from metastable levels around the visual spectrum
The gas shell surrounding the Beta Lyrae system was discovered long ago and its structure has always been studied on the basis of helium lines originating from metastable levels. First of all, it concerned the strongest emission-absorption line of the He I $\lambda$ 3888 in the violet spectrum. Well-known traditional studies of the broadening shell in the complex line of He I $\lambda$ 3888 (e.g. Struve, 1941; Sahade et al., 1959) were subsequently extended through the study of this shell in the infrared line of the He I $\lambda$ 10830 of this triplet series (e.g. Girnyak et al., 1978). It should draw attention that the contours of these lines contain narrow and deep absorption components with negative radial velocities in all orbital phases. If they exceed the parabolic velocities of moving microparticles in the binary system, it has traditionally been believed that the structural components of this shell, in which such lines are formed, move outward from this binary system. The He I $\lambda\lambda$ 3888, 10830 lines have a metastable level of $2^{3}$S. This level is highly populated with electrons and the effects of collision are crucial for the manifestation of these lines in the spectrum. Indeed, absorption components of these triplet lines of helium are well visible in the diluted shell structure around this binary system. At the same time, these helium lines have a very strong emission component in their contours, in many respects similar to those in lines H${}_{\alpha}$ and He I $\lambda$ 7065 (see Skulskyy, 2020b), where clear intercorrelations were discovered between the phase variability of the donor’s magnetic field and dynamic and energy characteristics of these strong emission lines. This may indicate the general nature of the formation and spatial localization of all these emissions. It can be supposed that the absorption in the He I $\lambda\lambda$ 3888, 10830 lines, which cuts through their emission component, arises as self-absorption and should reflect the specificity of the matter loss from the donor surface, which, showing certain structural features of the magnetic field, reaches the Roche cavity. The advanced attempt to re-study the characteristic spectral features of these and other He I lines, which arise from metastable levels, in combination with the corresponding manifestation in the spectrum of the binary system of the donor magnetic field configuration presents new opportunities in this study.
2.2 Magnetic field and He I lines arising from metastable levels in the violet spectrum
2.2.1 The above question and known article by Sahade et al. (1959)
Based on the concept of the existing configuration of the donor’s magnetic field, the careful work of Sahade et al. (1959) can be considered as one of the most consistent and informative studies of the He I $\lambda$ 3888 line. The results of this investigation are based on studies of 195 plates with a dispersion of 10 Å/mm obtained from observations of Beta Lyrae on the Mount Wilson 100-inch reflector over 12 orbital cycles in the 1955 season (the article presents an atlas of photographic spectrograms in the region $\lambda\lambda$ 3680-4580). However, only in 4-7 cycles in June-July they were performed most fully to be suitable for a detailed analysis. In Fig. 1, which is a copy of Fig. 13 from the article of Sahade et al. (1959), “radial velocity measurements from the shell line at He I $\lambda$ 3888” are presented. As it is seen at the bottom of Fig. 1, where the two absorption components “with very definite” velocities of about -170 km/s and -130 km/s are shown, they converge to one absorption component near the 0.85 P phase of the visibility of the magnetic pole on the donor surface. These absorption components demonstrate very clear sequences over the orbital phases, and, according to the understanding of Sahade et al. (1959), their formation “must occur in a shell that surrounds the whole system”. Their radial velocities are the largest and differ in curves shape from such curves of radial velocities of the absorption components that are at the top of Fig. 1. The latter more resemble the absorption components superimposed on emissions in lines H${}_{\alpha}$ and He I $\lambda$ 7065 (reflecting the gas flows between the donor and the gainer; see Section 2.3 in Skulskyy (2020b)).
It is well known that the He I $\lambda$ 3888 line also has an emission component, which consists only of the very high red peak in contrast to the two-peak emission, with superimposed absorption, at strong lines H${}_{\alpha}$ and He I $\lambda$ 7065 (the latter are also above the continuous spectrum in all orbital phases). The violet peak in the He I $\lambda$ 3888 line is virtually absented owing to deep violet absorption, showing appreciable changes in observations of different investigators (see, for example, Skulsky & Kos, 2011). Fig. 2, which is a copy of Fig. 18 from the article of Sahade et al. (1959), illustrates the radial velocities curve from the red emission peak at the He I $\lambda$ 3888 line. One sees two maxima at the 0.35 P and 0.85 P phases, i.e., in the phases of visibility of the magnetic field poles on the donor surface (the small local maximum in the 0.6 P phase may reflect the collision of the gas flow with the accretion disk). The similar two maxima are observed on the radial velocity curve of the red peak at the He I $\lambda$ 4472 line (see Fig. 19 in Sahade et al. (1959)). The red peaks of these helium lines of the violet spectrum exhibit the same behavior as red peaks at the H${}_{\alpha}$ and He I $\lambda$ 7065 lines. The spectral features of all components at the H${}_{\alpha}$ and He I $\lambda$ 7065 lines were considered in detail on the basis of Figures 5 and 6 in Skulskyy (2020b). These maxima at the 0.35 P and 0.85 P phases match clearly the phases of the two maxima on the curve of the effective magnetic field strength of the donor (see also Fig. 1 in Skulskyy, 2020b). Being synchronized in orbital phases, all these emission components have clear similarities, their physical nature and spatial interconnection seem indisputable. It can be assumed that the formation of red peaks in these lines, including such for the line He I $\lambda$ 3888, should be associated with the moving radiation near the donor surface, close to the poles of the magnetic field on this surface. This corresponds to the conclusions of Skulskyy (2020b) on the study of dynamics of emission components in the H${}_{\alpha}$ and He I $\lambda$ 7065 lines.
The dynamic behavior of both the above absorptions of the He I $\lambda$ 3888 line in Sahade et al. (1959), which are shown at the bottom in Fig. 1, is even more interesting. Their double nature has not been so clearly defined or repeated by investigators in other studies (see, for example Harmanec & Scholz (1993); Skulsky & Kos (2011), which analyzed such studies from the early 20th century). It should be noted that the radial velocity of the center of mass of this binary system, as determined from the interstellar Ca II lines by Sahade et al. (1959), is -15.2 km/s. The $\gamma$ - velocity determined in Table 9, for the set of orbital elements of Beta Lyrae computed on the basis of all 192 spectral plates in the summer of 1955, is -16.0 km/s. The set of orbital elements, calculated from observations taken between 0.85 P and 0.15 P phases, i.e., with the data discarded during the main (or primary) eclipse, showed -15.0 km/s for the -velocity. That is, both absorption components with radial velocities of approximately -170 and -130 km/s, should be considered as components with average velocities of about -155 and -115 km/s. Modern stellar masses of this binary system and the inclination of the orbit (see in Skulskyy, 2020a) allow affirming that the parabolic velocity of the moving gas particles reaches a value slightly smaller, but close to -115 km/s, which is sufficient to reach the Roche cavity with the Lagrange point L2. A simple calculation also shows (Burnashev & Skulskij, 1991)
that the velocity of the particles, which is sufficient to reach the Roche cavity with the Lagrange point L3, is near to -160 km/s. That is, the absorption component with an average radial velocity of approximately -155 km/s, or its possible greater values, may indicate a mass loss in a wider range of directions, in particular in the direction of the Lagrange points L4 and L5.
Here it is important to emphasize that Sahade et al. (1959) themselves noted: “there are striking changes with phase and cycle in the spectrum” and “the great complexity of the spectrum of Beta Lyrae makes it impossible to give a description which would do justice to the wealth of information contained in the material”. Indeed, at some points, this spectral material is presented as a general description. Now it is possible to consider the data of the observation in the light of modern ideas and concepts, in a more detailed manner. The behavior of its identified two absorption components of the He I $\lambda$ 3888 line should be investigated as carefully as possible, supplementing the analysis of radial velocities, given primarily in Table 3 from Sahade et al. (1959). But here there are also several other lines of helium, in particular those arising from metastable levels, which on the basis of tabular data and graphs will help to significantly expand the scope of research.
Based on Table 3, it is possible to estimate the behavior of the absorption of the He I $\lambda$ 3888 line with a smaller averaged radial velocity. In most orbital phases, Table 3 shows measurements of its radial velocities in the range of 125-135 km/s. This corresponds to a matter outflow velocity of about -115 km/s relative to the center of gravity of the binary system, i.e., with the first parabolic velocity. The most “striking changes with phase and cycle in the spectrum” are visible in the phases (0.95-0.13) P within a wide primary eclipse (see Fig. 1), when the donor is in the process of the eclipse by the gainer hidden in the accretion disk. At the time, according to the data of Table 3, the radial velocity of this absorption, during several days of observations at the boundary from 6 to 7 and from 11 to 12 orbital cycles, is observed in the range from -95 to -105 km/s, i.e., smaller than the parabolic velocity. The significant decrease in the radial velocity of this absorption component at the boundary of these orbital cycles indicates that the gas, flowing in the direction of the gainer in its Roche cavity, can only reach nearer structures of the accretion disk. But this is not observed in the same phases of the zero cycle and at the boundary of 4 to 5 orbital cycles, where the averaged radial velocity is approximately -115 km/s relative to the center of gravity. This indicates that the flows of matter during mass transfer are not laminar and have a variable nature within several cycles. Because in both events the motion of matter is visible above the accretion disk, i.e., the observer behind the gainer records a significant elevation of the moving matter above the plane of the orbit, this determines a considerable height of the rim of this disk from the side of the donor. As shown in Skulskyy (2020b), the vertical component of the gaseous flow in the direction of the gainer can be formed due to the configuration of the donor’s magnetic field, in which the magnetic axis is directed at a significant angle to the plane of the orbit of this binary system.
It should be noted that in phases (0.95-0.13) P of the main eclipse, the more negative absorption component in the He I $\lambda$ 3888 line also shows the various radial velocities in different cycles. In phases (0.95-0.99) P on June 8 of the 4th cycle and on July 4 of the 6th cycle this absorption component shows the average radial velocity of -173 km/s (or -158 km/s relative to the center of gravity), reflecting the motion of matter over the accretion disk, which approaches the velocity of its exit outside the Roche cavity with the Lagrange point L3. The dynamics of gaseous flows recorded in phases (0.03-0.13) P of the 5th cycle when measuring the radial velocity of this absorption component on 8 spectrograms on June 9 and 16 spectrograms on June 10? is most intriguing. In phases (0.03-0.05) P of the first date, the average radial velocity was -170 km/s with its growth within 5 km/s, however, in phases (0.11-0.13) P on June 10, its average the velocity was -180 km/s (on the first 4 plates it was -175 km/s, on the next 4 plates in phase near 0.12 P has reached the average velocity of -185 km/s, then it was -181 km/s and -180 km/s, respectively). Confident achievement of the radial velocity of -185 km (-170 km/s relative to the center of gravity) indicates that in this direction, the velocity of the particles was sufficient for the free escaping of the matter from the binary system with the intersection here of the boundary of the Roche cavity with the Lagrange point L3. But in the phases (0.03-0.05) P on July 5 and (0.12-0.14) P on July 6 of the 7th cycle, the average radial velocity was -171 km/s, which is much smaller than the velocity of -185 km/s in the 0.12 P phase on June 10. Moreover, for example, in the phases (0.987-0.002) P at the boundary of cycles 11 and 12, this absorption component of the He I $\lambda$ 3888 line revealed a significantly lower average radial velocity of -163 km/s (or only -148 km/s relative to the center of gravity). That is, passing over the center of the accretion disk, the moving matter with such velocity, most likely, must be captured by the outer side of the revolving accretion disk. Therefore, the more negative absorption component of the He I $\lambda$ 3888 line also indicates that the matter flows during mass transfer are not laminar and have a variable character during one or several cycles.
For comparison of the behavior of the He I $\lambda$ 3888 line during phases of the main eclipse, it makes sense to evaluate the behavior of other helium absorption lines arising from metastable levels. Sahade et al. (1959) measured the radial velocities from He I $\lambda\lambda$ 3614, 3965 lines of the singlet series. The radial velocities of their absorption components, as well as for the He I $\lambda$ 3888 line, are listed in Table 3 and shown in Fig. 3 (as a copy of Fig. 14 from the article Sahade et al. (1959)). In almost all phases, the more negative component of these singlet lines forms a successive curve with an average radial velocity close to -115 km/s relative to the center of gravity (i.e., close to the first parabolic velocity) with complex deviations from it in phases (0.95-0.13) P. This also applies to phases (0.95-0.05) P around the middle of the main eclipse both for the zero’s cycle (April 5 as the first observation date in 1955) and at the boundary of the 11th and 12th cycles (September 7 as the last observation date). But an important difference in the variability of these lines around the middle of the main eclipse concerns the observation periods of June 8-10 at the boundary of the 4th and 5th cycles and July 4-5 at the boundary of the 6th and 7th cycles. It is in these June and July observations that two narrower absorption features in these singlet helium lines are clearly distinguished. The development of events should be considered from the end of the 4th cycle, noting that on June 7 in the phases (0.88-0.90) P of direct visibility of the magnetic pole on the donor surface there was the average radial velocity of this absorption of about -115 km/s relative to the center of gravity. A clear rise of the radial velocity of the above absorption at its increment from -130 km/s to -135 km/s for 8 spectrograms occurred on June 8 within the phases (0.96-0.98) P. On June 9, within the phases (0.03-0.05) P, this absorption component during 8 spectrograms reached the radial velocity from -146 km/s for the first 4 spectrograms to -156 km/s for the next 4. Such velocity of moving matter is close to the second parabolic velocity. However, on June 10, within the phases (0.11-0.13) P, this average radial velocity on all 16 spectrograms fell slightly to -148 km/s with respect to the center of gravity. For the first time, these spectrograms clearly manifested the second component of absorption that had an average radial velocity of -97 km/s, somewhat smaller than the parabolic velocity (near -115 km/s relative to the center of gravity). On July 4 of the 6th cycle, 8 spectrograms within the phases (0.96-0.99) P recorded an average radial velocity of -155 km/s for the first absorption component and -98 km/s for the secondary one. On July 5, 7 spectrograms within phases (0.04-0.06) P of the 7th cycle recorded an average radial velocity of -151 km/s for the first absorption component and -98 km/s for the secondary one. However, the next night, July 6, unexpected changes were recorded: all 12 spectrograms in phases of (0.12-0.14) P showed that the radial velocity of the first absorption component dropped to -105 km/s, and the second absorption component disappeared regarding the measurement of radial velocities.
Thus, in the June and July observations, these singlet helium lines, becoming double in certain phases of the main eclipse, reveal asynchronous differences. Between the two orbital rotations from June 10 to July 6, in the same phases (0.11-0.14) P, the spectrum of the binary system represents very different expressed sets of both absorption components of these lines. If on June 10 both components of these lines became clearly double at radial velocities of -148 km/s and -98 km/s, then after two orbital cycles on July 6 the first of the two absorption components diminished its velocity to -105 km/s, i.e., slightly less than the parabolic velocity, but the second absorption component was absent. Moreover, if in the 5th cycle in June both components of these lines with velocities of -148 km/s and -98 km/s became visible only in the phases (0.11-0.13) P, then in the 6th cycle similar radial velocities -155 km/s and -98 km/s of both these components were measured in phases (0.98-0.99) P on July 4 and close to such velocities -151 km/s and -98 km/s on July 5 (in the 7th cycle) in phases (0.04-0.06) P (the next night on July 6, in phases (0.11-0.14) P the first absorption component had the radial velocity only -105 km/s, and the second absorption component unexpectedly disappeared and was invisible on the following nights of this 7th cycle). Hence, it should be noted that within the main eclipse the He I $\lambda\lambda$ 3614, 3964 lines show two phase segments of significant changes in their intensity and dynamics, which may indicate disordered changes in the density of the accretion disk. They are grouped close to the 0.95 P and 0.05 P phases, which, reflecting the formed dynamically changing structures of the disk, correspond simultaneously to the phases of the maximum manifestation of satellite lines (see Skulskyy (2020a) and more fully Skulskij (1992)). The third region of such changes is formed by the denser matter flow in the phases (0.10-013) P, which opens after the center of the eclipse.
Comparing the changes in the radial velocities in the absorption components of singlet and triplet helium lines arising from metastable levels, it should be especially noted the observation on June 10 in the phase near 0.12 P. In the He I $\lambda$ 3888 line, there was recorded the radial velocity -170 km/s relative to the center of gravity, which is sufficient for the motion matter from the binary system in any direction, including its exit outside the Roche cavity with the Lagrange point L3. At the same time, for the first time, a clear bifurcation of He I $\lambda\lambda$ 3614, 3964 lines (of the singlet series) into two absorption components with radial velocities comparable to two parabolic velocities was recorded. It is also important that Table 4 and Fig. 15 in Sahade et al. (1959) show that these observations on June 10 recorded on all 16 spectrograms the strongly shifted radial velocities for absorptions, which cut through emissions of H${}_{\gamma}$ and He I $\lambda\lambda$ 4472, 4026 lines. Shifts in the position of these absorptions make up -151 km/s (as the average on the first eight spectrograms) and -141 km/s (on the last eight of them) relative to the center of gravity of the binary system. A similar phenomenon was recorded only during the zero’s cycle on April 6 in phase 0.10 P, when measurements of these lines showed on the obtained spectrograms the radial velocity of -157 km/s relative to the center of gravity of the binary system, at clear excess of the second parabolic velocity, sufficient to an outflow matter via stellar wind from the Lagrange point L3.
The events described above, recorded in the behavior of the radial velocities of the absorption components of helium and hydrogen lines in the violet spectrum, show, within the main eclipse, not laminar but to some extent an unpredictable sharply variable mass transfer between the components of the binary system. In close phases of different dates, the observer behind the mass gainer registers the approach of the matter from the donor to the gainer at different radial velocities above the surface of the accretion disk. It is clear that when such velocities of a matter are slightly smaller than the first parabolic velocity, this matter settles down on the surface of the accretion disk, from the donor side. When the radial velocities of the moving substance exceed the first parabolic velocity or are a significant fraction of the second parabolic velocity, such gas flows can be captured directly by the gainer or the outer parts of the accretion disk. Repeatedly recorded velocities of matter surpassing the second parabolic velocity confirm the direct exit of matter beyond the close circular structures of the binary system and the possibility of forming a distant shell. This agrees with the observations of the radio-nebula, which surrounds the Beta Lyrae system, indicating the substantial non-conservative mass-loss during the evolution of this interacting system (Umana et al., 2000). The obtained generalized picture does not yet have data that would allow us to make specific predictions of the above events. They can be caused by a holistic system of secondary periods and resonances that change synchronously with the growth of the orbital period, variable substructures of the outer layers of the accretion disk, the peculiar spatial configuration of the donor’s magnetic field, or other reasons considered in Skulskyy (2020a, b). In addition, the findings on the variable dynamic behavior of the absorption components of the above helium lines may to some extent confirm the results of disk-forming motions of matter in gas-dynamic modeling of stellar structures and accretion disk, given by Bisikalo et al. (2000).
Analyzing the behavior of the radial velocities of the absorption components in complex helium and hydrogen lines, the previous points in this section focused on the details of the transfer of matter from the donor to the gainer within the main eclipse, i.e., in the direction of the gravitational axis of the binary system. It is equally important to evaluate their behavior in the phases associated with the transfer of moving matter in the direction of the axis of the donor’s magnetic field. It should be noted (Skulskyy, 2020a) that a dipole magnetic field with the axis in the direction of the orbital phases (0.355-0.855) P has a maximum magnetic field on the donor surface at the 0.855P phase, which reflects the location of this magnetic pole close to the gainer. In addition, the magnetic dipole axis inclined to the orbit plane of the binary system by 28\degr, and therefore, this magnetic pole on the surface of the donor in phase 0.855P is also above the plane of the orbit. This means that the ionized gas directed by the donor’s magnetic field moves in the direction of the dipole axis from the donor’s surface and, deviating along the magnetic field lines towards the accretion disk, can rise above the front edge of the disk without much energy loss.
The tables and graphs in Sahade et al. (1959) allow of stand out certain aspects related to the moving matter in the direction of the axis of the magnetic field of the donor. It is reasonable to estimate first the behavior of radial velocities of absorption components of the above lines in orbital phases near 0.85 P, i.e. in the phases of direct visibility on the donor surface of the magnetic pole facing the gainer. From Fig. 1 it is seen that in phases around 0.8 P both absorption components in the He I $\lambda$ 3888 line form the single absorption with the radial velocity near -135 km/s. This is sufficient for a direct entering of matter in to an external medium in this direction through the Roche cavity with the Lagrange point L2 (the first parabolic velocity relative to the center of gravity of the binary system is near -115 km/s). In these phases, the more negative absorption component of the He I $\lambda\lambda$ 3614, 3964 singlet lines shows a single average radial velocity close to -115 km/s (not showing any bifurcation, see Fig. 3). Several absorption groups with much smaller radial velocities, which are present in other phases, are completely absent in phases before 0.855 P, i.e., the flow of magnetized plasma is manifested as a single whole. In the phases (0.88-0.90) P after 0.855 P, the physical conditions changed. The more negative absorption component in the He I $\lambda$ 3888 line reached the average radial velocity of -156 km/s, and the absorption of the He I $\lambda$ 3888 line with a lower average radial velocity approached to a radial velocity near -115 km/s, i.e., to the second and first parabolic velocities, respectively. Evidencing the certain stratification of their formation, the He I lines of both series arising from metastable levels emphasize the average direction of plasma flows in the direction of the donor’s magnetic field axis and localization in phases around 0.855 P on the donor surface of the magnetic pole close to the gainer. This indicates that in the phases of direct visibility on the donor surface of this magnetic pole there is an outflow of magnetized matter outwards from the donor surface, as one of the important directions of non-conservative loss of matter during the evolution of this interacting system.
Based on Sahade et al. (1959) observations in 1955 it can be stated that the radial velocities in the phases near both poles of the magnetic field in the He I $\lambda$ 3888 absorption line showed a radial velocity much higher than the first parabolic velocity (see Fig. 1). However, the stellar wind behaves somewhat differently in the phases of the first quadrature, in particular in the vicinity of the 0.355P phase of the magnetic field pole on the donor surface. This applies, for example, to both the phase interval (0.25-0.5) P in the He I $\lambda$ 3888 line (see Fig. 1) and phases of (0.25-0.45) P in the H${}_{\alpha}$ and He I $\lambda\lambda$ 4472, 4026 lines (see Figures 13, 15, and 16 and Tables 3 and 4 in Sahade et al. (1959)). Several groups of radial velocities inherent to the corresponding absorption components are visible. The group of absorptions with positive radial velocities around +20 km/s, i.e., +35 km/s relative to the center of gravity of the binary system is important. It indicates that within the (0.35$\pm$ 0.1) P phases of the visibility of the magnetic pole on the donor surface, a runoff of magnetized matter from the observer is formed in the direction of the axis of the donor’s magnetic field (0.35-0.85) P. Reflecting the thermal velocity of matter from the surface of the donor, which reaches its Roche cavity, the radial velocity in these phases indicates the direction of the stellar wind from the surface of the donor. It is directed from the observer along the magnetic axis of the donor to the visibility phases of 0.855P on the donor surface of the magnetic pole facing the gainer, where the magnetic field of the donor reaches its maximum value. As it can be seen from Fig. 3, such a group of radial velocities is not clearly detected in the lines He I $\lambda\lambda$ 3614, 3964, indicating some stratification of physical conditions in the moving matter. But these lines, like the lines He I $\lambda$ 3888, H${}_{\alpha}$ and He I $\lambda\lambda$ 4472, 4026 have absorption components, which clearly are changed in radial velocities from -60 km/s to -40 km/s in phases of (0.30-0.45) P. This exhibits certain variable events near this pole of the donor’s magnetic field that are related, for example, to variations in these phase ranges of the total flux of polarization recorded in Hoffman et al. (1998) and Lomax et al. (2012). To some extent, these changes were studied on the basis of helium lines arising from metastable levels by Skulsky & Kos (2011). Now, this requires detailed coverage and comparison with the evidence of Sahade et al. (1959).
2.2.2 Helium lines arising from metastable levels in more modern studies
As it was previously pointed out, the issues of the dynamics of circumstellar gaseous structures using lines He I $\lambda\lambda$ 3888, 3964, 5015, which arise from metastable levels, were studied by Skulsky & Kos (2011). During 2008-2010, they obtained almost 100 echelle spectrograms with a resolution of about 45 000 on the 2-m reflector of the Peak Terskol Observatory. The narrow absorption components of the He I $\lambda\lambda$ 3964, 5015 showed that the variability of radial velocities and intensities with the phase of the orbital period is practically correlated: both lines formed at approximately the same distance from the center of the system and participate in the same processes. The intensities of these lines reach a minimum in the phases near (0.45-0.60) P, when the donor closes the gap between it and the gainer, simultaneously highlighting the phase direction (0.6-0.1) P perpendicular to the axis of the donor’s magnetic field. The intensity of these lines also decreases sharply within less than 0.1 P in the phases of 0.35 P and 0.85 P of polar regions on the donor’s surface (more obviously in the He I $\lambda$ 3964 line, see Fig. 1 in Skulsky & Kos (2011)). But particularly sharp changes are clearly seen in the negative radial velocities of the He I $\lambda\lambda$ 3964, 5015 lines, which indicates complex processes in the orbital structures of this binary system (see Fig. 4, which is a copy of Fig. 2 from Skulsky & Kos (2011)). It can be seen that within 0.1 P in the 0.35 P and 0.85 P phases, which correspond to the location of the magnetic poles on the donor surface, there is a sharp drop in the radial velocity from -55 km/s to -125 km/s, i.e., by 70 km/s relative to the center of gravity of the binary system ($\gamma$ = -18 km/s according to measurements of the radial velocities of the sharp interstellar line Ca II $\lambda$ 3933).
The behavior of the radial velocities curve of these helium lines is actually clearly associated with the magnetic field curve obtained in the course of measurements of the Zeeman splitting at Si II $\lambda\lambda$ 6347, 6371 lines by Skulskij & Plachinda (1993). The effective intensity of the longitudinal component of the magnetic field varies during the orbital period within $\pm$ 200 G, but as it can be seen from Figures 5 and 6 of Skulskyy (2020a), rapid changes in the magnetic field curve are clearly visible only in the phases near the poles of the magnetic field. In both cases, the polarity of the magnetic field varies within 0.1 P in the 0.355 P phase of the first quadrature and in the 0.855 P phase of the second quadrature. The behavior of the photographic curve of the magnetic field is also characterized by rapid changes in the magnetic field curve around the 0.855 P phase (i.e. the phase of observation of the magnetic pole facing the gainer) and in the phases around the main eclipse.
Note that the sharp peak of radial velocities with a rapid change in their values is observed in Fig. 4 in the phase of about 0.67 P (one can assume that the opening gas flow is in collision with the accretion disk). Similarly formed two narrow peaks of radial velocities are observed in the same phases of the main eclipse, in which the radial velocities of the opposite sign are observed in the satellite lines, characterizing the outer edges of the accretion disk. It should be concluded that due to different physical parameters at the edges of the accretion disk a complex structure of plasma flows is formed, which may participate primarily in mass transfer between components of the binary system, as well as in the outflow of matter outside it. In addition, note that the rapid changes of radial velocities near the main eclipse in the He I $\lambda\lambda$ 3965, 5015 lines (see Fig. 4) are to some extent related to such events recorded in the similar phases (0.88-0.90) P and (0.96-0.98) P of radial velocities from the He I $\lambda\lambda$ 3614, 3965 lines according to their measurements in Sahade et al. (1959) (their behavior was discussed above since the end of the 4th cycle on June 8, 1955). Such a coincidence is not accidental.
Based on observations of the Peak Terskol Observatory in 2008-2010, Skulsky & Kos (2011) also investigated the behavior of radial velocities of the He I $\lambda$ 3888 line for 1900-2010. Some dynamic characteristics are shown in Fig. 5 as a copy of Fig. 3 from their paper. Without commenting on the nature of the peculiar understatement of radial velocities on the Potsdam 1900-1909 spectra (one can agree with the thorough analysis of this problem conducted by Harmanec & Scholz (1993)), one should note certain points (given the average radial velocity of the center of gravity of this binary system $\gamma$ = -18 km/s, which is determined within 2 km/s on the spectrographs of the Crimean, Ondrejov and Peak Terskol Observatories). The range of measurements of radial velocities according to our Crimean observations in 1966-68 and 1972, our more recent observations in 2008-2010 at Peak Terskol, as well as observations in 1991 in Ondrejov, lies in a range from -118 km/s to -142 km/s at the average radial velocity near -130 km/s. There are no such sharp jumps in radial velocities inherent in the curve of radial velocities of the He I $\lambda\lambda$ 3964, 5015 sharp lines shown above in Fig. 4. Well-provided observations of the 1991 season at the Ondrejov Observatory show the average radial velocity in the He I $\lambda$ 3888 line, which is only about -127 km/s, but the radial velocity curve based on the data in Table 4 and Fig. 3 by Harmanec & Scholz (1993) shows a hump with a value of -122 km/s in the phase range (0.70-0.85) P of the visibility of the magnetic pole on the donor surface (and -132 km/s in the phase of 0.05 P). This is consistent with the convergence of both absorption dependences on He I $\lambda$ 3888 (see Fig. 1) to a single absorption component, which yields a reduced radial velocity to -(130-135) km/s in the phase range (0.75-0.85) P of the visibility of this magnetic pole on the donor surface as shown in Sahade et al. (1959). In both papers, the flow of matter in the direction of the axis of the donor’s magnetic field is significantly greater than the first parabolic velocity. If the difference of 10 km/s in radial velocities between two distant seasons of active observation is real, then this could be a confirmation of long-term changes in the non-laminar loss of moving matter out of the circumstellar structures of this binary system. It should be noted that none of the researchers of the He I $\lambda$ 3888 line observed this type of bifurcation of its absorption into two different components during the almost entire orbital period. In this aspect, the article by Sahade et al. (1959) remains original and its consideration required special research. This is one of the reasons why this issue was not considered in the report of Skulsky & Kos (2011). Note that the He I $\lambda$ 3888 line behaves somewhat differently in the known article by Flora & Hack (1975). This thorough article also deserves a detailed review, which for some reason has not been conducted for a long time.
2.2.3 Magnetic field and behavior of shell lines in the known article by Flora & Hack (1975)
In this study, it is necessary to consider the radial velocity curve of the He I $\lambda$ 3888 sharp absorption line in the paper of Flora & Hack (1975), the variability of which seems somewhat unusual. In most orbital phases these radial velocities should now be considered much smaller than the first parabolic velocity for this binary system (perhaps this is one of the reasons for the inactive use of this work). Flora & Hack (1975) gave the results of the radial velocity (RV) measurements and the profiles and intensities of the most significant lines, which were taken during the international campaign in the period of July 18 - August 1, 1971. The rich spectrographic material (lacking only one day at the orbital period that is about 13 days) consisted of 70 spectrograms of the enough high-dispersion (7 and 12 Å/mm) obtained at the 1.52-m telescope of the Haute Provance Observatory. Flora & Hack (1975) indicate “an asymmetrical distribution of matter in the envelope surrounding the whole system”. They conclude that the behavior of the RV curves for the Balmer and He I shell absorption cores must be real because these lines are virtually independent of blending with absorption lines of the B9 star (or now the donor). It is also noted that “none various absorption or emission features show any correlation with the invisible companion” (or now the gainer). This motivates to consider their rich observational material, drawing on modern concepts in the approach to its interpretation.
Fig. 6, along with other spectral lines, shows the data of measurements of the radial velocities of the He I $\lambda$ 3888 line, taken directly from the tables of the article of Flora & Hack (1975). If the average velocity of the center of gravity of the binary system is -17 km/s (according to Table 3 in interstellar components line Ca II $\lambda$ 3933 it is -15 km/s, and in lines of Na I $\lambda$ 5899 and 5895 it is -19 km/s), all applied radial velocities should in Fig. 6 shift up 17 km/s, in the direction of reducing negative velocities. Then, in phases (0.65-0.85) P, the He I $\lambda$ 3888 line will show the radial velocities of about -80 km/s relative to the center of gravity of the binary system (which is actually much smaller than the current first parabolic velocity for this binary system). This result immediately revealed several important points in relation to further research. First, the radial velocity -80 km/s is comparable to the average radial velocity obtained on Potsdam spectrograms in the early 20th century, the reality of which has been repeatedly discussed and questioned (see, for example, Harmanec & Scholz (1993)). Second, the stability in profiles, intensity, and radial velocities of the He I $\lambda$ 3888 line in orbital cycles during the 1971 season can be questioned, based, for example, on significant changes in these spectral characteristics in the observations of Sahade et al. (1959). The question arises of changing the directions of active loss of matter both between stellar components and in relation to circumstellar medium, as well as the reasons for the variable density of such outflows.
Regarding the direction of interpretation of Fig. 6, it is necessary to pay attention to its three parts, which, according to the classification of Flora & Hack (1975), present the radial velocity curves of two of the non-metastable He I $\lambda\lambda$ 5875, 6678 lines of the highest intensities, the metastable lines of He I $\lambda\lambda$ 3888, 5015, 3447, and the shell components of the Na I $\lambda\lambda$ 5889, 5895 lines. One can consider a central part of Fig. 6, which shows three of the five lines measured in Flora & Hack (1975) arising from metastable levels. The absorption components of these lines form dome-shaped curves of radial velocities in the phases of the second quadrature. The maxima of these dome-shaped curves were recorded in consecutive days of observations in the phases (0.78 - 0.77) P and (0.84 - 0.86) P. Similar curves show two more such lines He I $\lambda$ 3964 and He I $\lambda$ 6313, the radial velocities of which in these phases are located mainly between the curves of radial velocities of the He I $\lambda$ 5015 and He I $\lambda$ 3447 lines (in Fig. 6 there were not introduced so as not to overload the general picture). All five lines of these helium lines with certain potential difference between the upper and lower levels demonstrate clear stratification in the radial velocities, but the matter outflows have velocity much smaller than the parabolic velocity.
Since the observer in phases (0.75 - 0.85) P looks perpendicularly to the line of centers of stellar components and along the axis of the dipole magnetic field of the donor (the motion of matter to the observer), the most adequate explanation of this picture can be considered as the consequences of loss of matter directly from the donor surface near its magnetic pole in phase 0.855 P and its subsequent transfer to the gainer. It should be noted that the He I $\lambda$ 3888 line shows the single absorption, the radial velocity of which confidently reaches the current value of the first parabolic velocity only in the vicinity of phases (0.0 - 0.1) P, i.e. in the directions of motion of the matter to the gainer, as well as in phases near 0.45 P. It is known that in phases (0.42 - 0.49) P a hot spot is registered on the accretion disk, according to Burnashev & Skulskij (1991) and Lomax et al. (2012). Here, the interpretation of such velocities can be related to the outflow of matter from the donor surface or, conversely, to this surface, reflecting the specifics of the location of the pole of the magnetic field on this surface. Indeed, Burnashev & Skulskij (1991) showed that the observer begins to see excess radiation (hot surface on the donor) only from the 0.355P phase, i.e., along the axis of the magnetic field when its pole opens on the donor surface (see Fig. 1 in Skulskyy (2020b)). This is more likely when the pole of the magnetic field is significantly shifted to the upper hemisphere on the surface of the donor (the magnetic dipole axis is inclined to the orbit plane of the binary system by 28\degr (Skulskij, 1985). Indeed, the hot surface on the donor continues to be observed almost to the 0.50P phase, demonstrating during the orbital phases (0.38-0.49) P the clear variability of the absolute radiation flux in the H${}_{\alpha}$ emission line and the rapid variability of the spectrum in the H${}_{\alpha}$ emission region (see also Alexeev & Skulskij, 1989). This may be a reflection of the collisions of the hot plasma with the donor surface, directed (to the observer with the radial velocity close to the first parabolic velocity) along the lines of force of the magnetic field. Note that in these phases, there is a bifurcation of the absorption component so that both shallow absorption components are superimposed on the strong emission component of lines H${}_{\alpha}$, He I $\lambda$ 6678, and He I $\lambda$5875, respectively, in Fig. 3d, 6a, and 7a in Flora & Hack (1975). The more violet-shifted component of this absorption shows that the tabular radial velocity approaches in the phase of 0.45 P to current values of parabolic velocity (see Fig. 6). Their occurrence can also be explained by the motion of the magnetic plasma along the lines of force of the magnetic field. In the He I $\lambda\lambda$ 5875, 6678 lines the red-shifted component of this absorption shows much smaller radial velocities close to the velocities in the lines of the donor atmosphere. Interestingly, Flora & Hack (1975) suggested that the violet-shifted absorption in phase 0.45 P may reflect the escape of gas matter through the Lagrange point in front of the donor. Such a hypothesis requires more observational evidence.
The radial velocity curve of the sodium lines resonates with the previous remark. The bottom of Fig. 6 does not so clearly show that the observation phase of 0.375 P (near the phase of 0.355 P of direct vision of the pole of the magnetic field on the surface of the donor) was missed. In the phase range (0.30 - 0.45) P one can assume both a certain minimum of this curve near the 0.355 P phase and a small local hump. In each case, the curve of radial velocities of the Na I $\lambda\lambda$ 5889, 5895 absorption lines clearly manifests itself as the wide maximum of negative velocities near the 0.355 P phase and their clear narrow minimum in the phase 0.855 P, i.e., in the phases of visibility of the magnetic field poles according to the photographic curve of the effective magnetic field strength of the donor (see Fig. 1 in Skulskyy (2020b)). The presence of narrow interstellar components of the Na I lines allowed us to check the tabular data and to clarify the radial velocities of the shell components of these sodium lines directly from Figs. 7a and 7b of Flora & Hack (1975). Fig. 6 shows that the observation in phases (0.8 - 0.9) P is especially interesting. Since the interstellar components in Na I $\lambda\lambda$ 5889, 5895 lines are -19 km/s, the radial velocity in the direction of the observer in these phases is the smallest one and is only of about -70 km/s, which is much smaller than the first parabolic velocity. Here the observer looks perpendicular to the donor surface along the axis of the magnetic field. Important is the fact of rapid dynamic changes in the moving plasma in the ranges of 0.1 P in phases near 0.855 P, i.e., of the center of the magnetic pole on the donor surface. Within 0.1 P, the movement of the plasma clearly slows down, stops at a certain level, and accelerates again. Hence, the observations of Flora & Hack (1975) clearly state the fact of actual variations in the sign of the RV progression curves of both helium and sodium lines reflecting the direction of motion of matter around the 0.855 P phase. It can be argued that the observations of Flora & Hack (1975) during the 1971 international campaign clearly identified the region of the donor surface that is responsible for the location of the pole of the magnetic field facing the gainer. This is consistent with studies of Zeilik et al. (1982) and Skulsky & Kos (2011), confirming CCD research of the magnetic field by Skulskij & Plachinda (1993).
Radial velocities in the Na I $\lambda\lambda$ 5889, 5895 lines in Fig. 6 do not show a detailed picture of their possible variability in phases (0.95 - 0.05) P of eclipsing of the donor by the gainer wrapped in the accretion disk. The observer sees in phases (0.60 - 0.85) P a gradual decrease in the radial velocities by 30 km/s, which in phase 0.855 P immediately changed to their such gradual increase in phases (0.85 - 0.10) P. One should think that the spatial configuration of the donor magnetic field reflects the direction (0.6 - 0.1) P, perpendicular to the direction (0.35 - 0.85) P of the donor magnetic axis. This is the direction of movement of the plasma perpendicular to the lines of force of the magnetic field in the space between the stellar components under the action of gravity of a massive gainer.
It should also be noted that the radial velocity curves of the He I $\lambda\lambda$ 5875, 6678 absorption lines show in certain respects similar behavior.
Their average curve at the top in Fig. 6 is made using measurements of the radial velocities of the line He I $\lambda\lambda$ 5875 as a line, the theoretical intensity of which is seven times greater than that of the He I $\lambda$ 6678 line.
The radial velocities of the He I $\lambda$ 6678 line are located slightly higher than the velocities of the He I $\lambda$ 5875 lines, but in general, the behavior of these lines with the orbital phase is similar.
The extreme values of their radial velocity curves correspond to phases of approximately 0.6 P and 0.1 P, reflecting the direction (0.6 - 0.1) P, perpendicular to the magnetic axis of the donor with the direction (0.35 - 0.85) P.
These curves are shifted relative to the 0.855 P phase by about a quarter of the orbital period, as it is seen in comparison, for example, with the RV-curves in Fig. 6 for the Na I $\lambda\lambda$ 5889, 5895 lines.
The absorption component with radial velocities, which changes rapidly in the phases near (0.60 - 0.65) P, is well seen in the higher members of the singlet lines of helium, in particular, as shown in Fig. 6 in the He I $\lambda$ 3447 line.
Starting from the phases (0.60-0.65) P, this may indicate the occurrence of gas flows that are directed perpendicularly to the axis of the donor’s magnetic field in the direction of the gainer.
Similar behavior in the phases of (0.1 - 0.6) P is demonstrated in the central depth variations of the He I $\lambda\lambda$ 3888, 3964 shell lines at the maximum value in the phase of 0.1 P, and its minimum in the phase of 0.6 P (see Fig. 14 of Flora & Hack (1975)).
It should be noted that all this is consistent with the picture of the directions of the plasma motion of magnetically controlled circumstellar matter in the helium stars according to the study of Shore & Brown (1990) (see Figure 11 in the paper of Shore & Brown (1990), and its representation as Fig. 8 in Skulskyy (2020b)).
Their model can be illustrative for the donor as an oblique magnetic rotator of intermediate obliquity because the dipole axis of its magnetic field deviates relative to the orbital plane of the binary system, and the center of the dipole of the donor’s magnetic field is significantly shifted in the direction of the gainer. The outflows of matter from the donor surface in the mass transfer picture are possible both in the direction of the donor’s magnetic axis and in the direction perpendicular to this axis (see Skulskyy, 2020b). Some proof of this type of phenomenon may be the behavior of a complex line H${}_{\alpha}$ along the surface of the donor. In particular, Fig. 15 of Flora & Hack (1975) shows that the equivalent width of the absorption, measured with respect to the total area of the emission in the H${}_{\alpha}$ line, is halved in the phase direction (0.1 - 0.6) P, and the total equivalent width of emission plus the absorption of the H${}_{\alpha}$ line shows two deep minima in the phases (0.25-0.45) P and about 0.855 P, i.e. near the phases of visibility of both magnetic poles on the donor surface. Of course, to some extent, this is the result of reflecting the characteristics of the donor that fills its Roche cavity: the loss of matter along the axis of the magnetic field at its pole and the loss of matter at the Lagrange point L1, amplified by the massive gainer and deflected by Coriolis forces to the 0.1 P phase, are significantly different. At the same time, the variable motions of the plasma are reflected more clearly in the phases around 0.855 P of the visibility of the magnetic pole, facing the gainer, in accordance with the changes in the polarity of the magnetic field in the observations of Skulskij & Plachinda (1993).
2.3 Magnetic field and the He I $\lambda$ 10830 line arising from metastable level
If the structure and behavior of the He I $\lambda$ 3888 line as the second member of the principal series of triplet helium were studied during the first half of the XX century, then to some extent successful observations and studies of the He I $\lambda$ 10830 line, as the first member of this series, was carried out by Alduseva & Esipov (1969) only in 1961-1962. They used a contact image-converter as an image receiver mounted on a 12-inch reflector. The spectral resolution was 4.8 Å(11 days of observations in 1961) and 2.8 Å at 9 days of observations in 1962. This highlighted the complex structure of the line contour and gave some initial data on the change with the orbital period of the shell radial velocities and equivalent widths of the emission in the He I $\lambda$ 10380 line during the orbital period. They indicated that “the line appears in the very external parts of the shell, surrounding Beta Lyrae as a whole”. Morgan et al. (1974) made the next attempt in the study of the He I $\lambda$ 10380 line. They observed on the rapid-scan interferometer at the coudé focus of the 2.7-m telescope at McDonald Observatory in April 1973, but received only two spectra in the 0.90 P and 0.13 P phases, respectively. The best resolution of 4.7 Å in the 0.90 P phase allowed them to record the value of the absorption component -140 km/s with an error of 60 km/s.
Our more complete observations of the He I $\lambda$ 10830 line by Girnyak et al. (1978) were conducted in April-November 1974. There are no data on the following observations, and these studies need more attention. The observations were performed using an electron-optical image-converter of type FKT-1A (S1) mounted on the 50-inch telescope of the Crimean Astrophysical Observatory. During 26 nights of this season, 59 spectrograms with a dispersion of 48 Å/mm were obtained, which allowed measuring radial velocities with an error of 10 km/s. As it can be seen from Fig. 7, which is a copy of Fig. 1 from Girnyak et al. (1978), this line in the Beta Lyrae spectrum represents the strong emission that is several times higher with respect to the continuum than such emission in the He I $\lambda$ 3888 line. The considerable width of the He I $\lambda$ 10830 line indicates the movements in the gas plasma structures with velocities up to 800 km/s. The deep absorption component, which cuts through the emission, is shifted to the short-wave side. Girnyak et al. (1978) noted that the total intensity and ratio of its short-wavelength and long-wavelength emission components change significantly with the orbital phase. This is a new important factor that requires additional attention and further research.
Girnyak et al. (1978) proposed processed physical parameters of the He I $\lambda$ 10830 line, introduced in a general table and five figures. Fig. 8 reproduces, on the basis of tabular values, two of these five figures,
which are presented below as the radial velocity curve of the absorption component of the He I $\lambda$ 10830 line and the sum of its absolute fluxes from violet and red emissions components given in erg s${}^{-1}$ A${}^{-1}$ km/s, multiplied by $10^{33}$ (contours in the He I $\lambda$ 10830 line were built in absolute energy units based on six-color colorimetry (Burnashev & Skulskij, 1978)). Fig. 8 indicates some new interesting points. They concern the upper part of Fig. 8, which illustrates the variability in the orbital phase of the ratio of the intensity of the short-wave and long-wave components of emission, i.e. the dependence of $I_{v}/I_{r}=f(P)$, which we built here on the basis of the tabular data. In Fig. 8, bottom, it is seen that the $V_{r}=f(P)$ curve of the absorption component, which separates these emission components, remains approximately the same in the orbital phases at an average value of about -130 km/s (taking into account the radial velocity of -18 km/s for the center of gravity of this binary system). The largest scattering of data and rapid changes in radial velocities are observed in the range of phases (0.7-0.9) P, and a particularly sharp velocity change is shown in the phases around 0.855 P of the visibility of the magnetic pole, facing the gainer. Similar behavior in these phases there is in changes of contours in the He I $\lambda$ 10830 line (see the middle part of Fig.8). Despite some sharp changes in the total emission, which can occur at times from tens of seconds to tens of minutes (see, e.g., Alexeev & Skulskij (1989) and Skulskyy (2020b)), some trend is observed to a sinusoidal curve of the absolute flux of the total emission with extrema in the phases of (0.35-0.45) P and 0.85 P. It is known that these phases correspond to the phases of both extrema of the donor’s magnetic field (Burnashev & Skulskij, 1991; Skulskij & Plachinda, 1993). This picture is more clearly manifested in the dependence of $I_{v}/I_{r}=f(P)$ (the upper part of Fig. 8), characterizing the expected variability of shell structures in this binary system. The dependence of $I_{v}/I_{r}=f(P)$ showed a clear decrease in the $I_{v}/I_{r}$ ratio after the 0.855 P phase (of the direct visibility of the magnetic field pole on the donor surface) to the minimum in the main eclipse phase and its rapid rise to the phase 0.15 P, when the main flow of matter from the donor to the gainer, deflected by the Coriolis forces, begins after the main eclipse. It should be noted that these observations do not provide enough data for more definite conclusions in the phases (0.95-0.05) P of the main eclipse.
Nevertheless, the radial velocity curve of the He I $\lambda$ 10830 line looks similar to that of the He I $\lambda$ 3888 line defined by Skulsky & Kos (2011) over five different seasons (see Fig. 5 above). Figs. 5 and 8 show that the average values of the radial velocity of the absorption components of the first two triplet helium lines are close to -130 km/s relative to the center of gravity of the binary system, which significantly exceeds the first parabolic velocity. It could be assumed that the sharp absorption components of both lines are formed in a dynamically identical medium moving in all directions relative to the center of gravity of the binary system. Moreover, the strong emission of the He I $\lambda$ 10830 line does not explicitly distort the position of the absorption component, and the radial velocity of the emission center as a whole turned out to be close to the orbital velocity of the gainer, based on the previously established mass ratio of the stellar components (Skulskij, 1975). However, Figs. 1 and 6, which respectively relate to the studies of Sahade et al. (1959) and Flora & Hack (1975), showed that the behavior of the He I $\lambda$ 3888 line at different seasons and epochs can significantly differ. Such facts are real, they required a detailed consideration of some, unnoticed by Girnyak et al. (1978), dynamical characteristics of the He I $\lambda$ 10830 line at certain parallels in the study of the lines He I $\lambda$ 3888 and He I $\lambda$ 10830.
First of all, this applies to the phase interval (0.75-0.9) P around the 0.855 P phase of the direct visibility of the magnetic field pole on the donor surface close to the gainer. Note that the observations of 1971 by Flora & Hack (1975) and our observations of 1974 by Girnyak et al. (1978) can be attributed to fairly close seasons and the dynamic behavior of the shell lines He I $\lambda$ 3888 and He I $\lambda$ 10830 may have some similarity. As it can be seen from the bottom of Fig. 8, our observations were most frequent in the (0.7-0.9) P phases, showing rapid changes in the radial velocity at large amplitudes of their variability. The observation data at these phases were rebuilt on a larger scale in Fig. 9. It can be seen that the radial velocity curve of the He I $\lambda$ 10830 line in these phases, like the domed-shaped curve in Fig. 6 for the He I $\lambda$ 3888 line, also has a domed-shaped peak at a clear rise in negative radial velocity immediately after the phase 0.855 P of direct visibility of the magnetic pole on the donor surface. This peak of the radial velocities of the He I $\lambda$ 10830 line that changes in the ranges of 0.1 P within 20 km/s is twice smaller than that for the He I $\lambda$ 3888 in Fig. 6. At least, this points to some analogies in the dynamic parameters of gas structures in which these triplet helium lines were formed in the first half of the 1970s. The decrease of negative radial velocities at approaching the 0.855 P phase and their subsequent rapid increase after this phase in the He I $\lambda$ 10830 line (as well as in metastable lines of He I $\lambda\lambda$ 3888, 5015, 3447, and the shell components of the Na I $\lambda\lambda$ 5889, 5895 lines - see Fig. 6 built on the data of Flora & Hack (1975)) indicates the similarity of physical processes along the donor surface near this magnetic field pole.
At the same time, the He I $\lambda$ 3888 and He I $\lambda$ 10830 lines show a physically significant difference in radial velocities. It is clearly seen that if in the He I $\lambda$ 10830 line in Fig. 9 in all phases, including phases (0.6-0.8) P, the radial velocities are greater than the parabolic velocity, then in the He I $\lambda$ 3888 line according to Fig.6 in phases (0.6-0.8) P, they are much smaller than the parabolic velocity. This indicates the nonstationarity of the outflow of matter and the variable flow density in the direction of the dipole axis of the magnetic field. The clear changes, which were noticed in the absolute radiation flux (the central part of Fig. 8) within (0.8-0.9) P, also reflect the complex structure of the emission formation near 0.855 P in this short phase interval with a width of approximately 0.1 P. This emphasizes that rapid changes are inherent in these phases of both energy and dynamic parameters of the magnetized plasma. The rapid change in radial velocities near 0.855 P may indicate that the vector components of the magnetized plasma motion at the edges of the magnetic field pole are superimposed on the total plasma motion in the direction of the donor magnetic field axis. This one can also see from Fig.4, which demonstrates similar variations in the radial velocities of the He I $\lambda\lambda$ 3964, 5015 lines in the short phase interval with a width of approximately 0.1 P in phases of the visibility of both magnetic field poles on the donor surface. The observations of Girnyak et al. (1978) in phases (0.25-0.45) P are not dense enough to conduct a detailed analysis of the behavior of the He I $\lambda$ 10830 line in the phases of visibility of the magnetic field pole on the donor surface in the first quadrature, although, in these phases, the radial velocities of this line are also subject to abrupt changes similar to Fig.4. However, as a whole, the above research facts confirm the general picture of the behavior of energetics and dynamic parameters of magnetized plasma caused by a certain configuration of the donor’s magnetic field. The study of the spectral characteristics of the He I $\lambda$ 10830 line has led to certain clarifications in the formation of the picture of the transfer matter between stellar components and its movement into the medium around the Beta Lyrae system.
3 Conclusions and discussion
The study of shell structures in the Beta Lyrae system seemed natural on the basis of previous studies by Skulskyy (2020a, b), which involved a continuous interaction of the outer shell with a complex system of plasma flows between the donor and the gainer. It is clear that the spectral lines formed in external structures, in particular the helium lines arising from metastable levels, were studied. Scientific articles were used for the study, which, on the basis of qualitative observations, covered the orbital period well. There are few similar works, in particular Sahade et al. (1959); Flora & Hack (1975); Girnyak et al. (1978), which are half a century old. It is to some extent well that the authors of these works presented their results without the influence of modern interpretations, say, without information about the magnetic field of the donor or ideas about bipolar jet structures associated with the accretion disk. However, these articles have a lot of important spectral material in the form of published tables and graphs. As a whole, this material, together with data from more recent publications of Harmanec & Scholz (1993) and Skulsky & Kos (2011), have not been analyzed in more detail. Here an attempt was made to overcome this shortcoming.
All researched articles are distinctive. They differ significantly in the energy and dynamic parameters of the moving plasma in the medium near each component and between stellar components of the binary system. First, significant long-term differences were found within half a century in the movements of matter to the gainer and in external structures. Second, there are significant differences in the physical parameters of this plasma within one observation season and their variations from cycle to cycle (especially in Sahade et al. (1959)). It was established (see Fig. 1) only in Sahade et al. (1959) that: “He I $\lambda$ 3888 has two components with very definite and fairly constant velocities” of about -155 km/s and -115 km/s (in the 1955 season), i.e., close to the second and first parabolic velocities. In Flora & Hack (1975), the absorption component of the He I $\lambda$ 3888 line in certain phases of the 1971 season is approximately -80 km/s, which is much less than the parabolic velocity (see Fig. 6). The averaged radial velocity of the absorption component in the He I $\lambda$ 3888 line during the 1991 season is -127 km/s (see also Fig.3 in Harmanec & Scholz (1993)), which is very different from that in the previous two articles. The He I $\lambda\lambda$ 3964, 5015 lines (Fig. 4) show rapid changes in radial velocities in the range from -55 km/s to -125 km/s in short phase intervals up to 0.1 P, in particular close to the magnetic field poles, which were recorded in the seasons 2008-2010 in Skulsky & Kos (2011). It seems that all these differences are not periodically ordered. It should be emphasized that all researched articles show the averaged radial velocity in the He I $\lambda\lambda$ 3888, 10830, i.e., the velocity of the outflowed plasma in external structures, to be of about -130 km/s, except for one orbital cycle in 1971, in which observations by Flora & Hack (1975) were made.
The article gradually reveals the analyzed facts from the available spectral material, and since there is quite a substantial amount of them, they are further presented as touches of the obtained research picture in a possibly concise form. It turned out that this picture is based on fragments of a generalized investigation during the orbital period. They can be grouped into certain phase intervals as a reflection of slightly different physical processes occurring in them. Formally, they can be divided into five interconnected intervals. The phase range (0.95-0.05) P reflects nonstationary processes that occur during the transfer of mass to the gainer at the donor eclipse by the accretion disk. The phase range (0.07-0.14) P reflects the motion of the main gas flow deflected by Coriolis forces outside the accretion disk. It is necessary to distinguish the behavior of plasma structures near the phases of visibility of both poles of the magnetic field (0.30-0.45) P and (0.75-0.90) P, as well as within the range of phases (0.4-0.6) P eclipse of the accretion disk by the donor.
There are a number of effects and events associated with the eclipsing of the donor surface by the accretion disk in the phases (0.9-0.1) P of the main eclipse of the binary system. It can be noted that within phases (0.95-0.05) P of the deeper eclipse, the He I $\lambda\lambda$ 3614, 3964 and He I $\lambda$ 3888 absorption lines show, according to Sahade et al. (1959), two main regions of changes in their intensities and radial velocities. They are grouped near the 0.95 P and 0.05 P phases, which corresponds to the phases of maximum accretion disk density. This follows from many studies, in particular from phase changes of the equivalent widths of satellite lines in 1991-1992 (see Fig. 2 in Skulskij (1993c), or from the light curve based on observations by the BRITE Satellites in May-October 2016 (see Fig. 2 in Rucinski et al. (2018)). In these phases, there is the greatest variability of the radial velocities of both absorption components of the He I $\lambda\lambda$ 3614, 3965 and He I $\lambda$ 3888 lines. Their more negative shifted component varies in the range from -140 km/s to -160 km/s, sometimes reaching the second parabolic velocity (near -160 km/s) without a certain regularity in orbital cycles. This also applies to the center of the eclipse in different orbital cycles in phases (0.98-0.02) P, where the radial velocities also show variation within the same limits (for example, in phases (0.98-0.99) P at the end of the 6th cycle on July 4, 1955, their radial velocity was -(155-160) km/s, that is sufficient for an outflow matter from the Lagrange point L3). It is clear that in the range (0.95-0.05) P the dynamic behavior of moving matter reflected in the more negative absorption component in the He I $\lambda\lambda$ 3614, 3964 and He I $\lambda$ 3888 lines can be the product of instability processes of the matter transfer between the stellar components. The less negative absorption component of the He I $\lambda$ 3888 line in these phases reaches only the first parabolic velocity, and in the He I $\lambda\lambda$ 3614, 3964 lines they only approach it or are much smaller, also demonstrating the fact of non-laminar, but clearly variable events during the mass transfer. It should be noted that the observations of Sahade et al. (1959) were carried out in the 1955 season with significant phase gaps during the day (its time is 0.077 P), which does not allow to clarify the structure of the accretion disk. The important fact of the detection of disk-shaped structures in phases (0.95-0.05) P are the observations of Skulsky & Kos (2011) during 2008-2010, which clearly show rapid changes of radial velocities from -55 km/s to -125 km/s in He I $\lambda\lambda$ 3965, 5015 singlet lines (see Fig. 4) in two short phase intervals.
Thus, the variable nature of the flows of matter during mass transfer in phases of the main eclipse indicates eruption processes on the donor, the appearance of matter clots over the rims of the disk from the side of the donor and their movement in the direction to the observer over the surface of the accretion disk accelerated by the massive gainer. Incidentally, Skulskyy (2020b) showed that the vertical component of the plasma flows in the direction of the gainer can be formed due to the configuration of the donor’s magnetic field, in which the magnetic axis is inclined at a significant angle to the orbit of the binary system. The structure of gas flows above the surface of the accretion disk significantly depends on the location of the magnetic pole on the donor surface relative to its distance to the gainer and relative to the plane of the orbit. These remarks also apply to the third region of active events in the phases (0.9-0.1) P reflected in the changeable and denser flows of matter in the phases (0.07-014) P, which start after the center of the main eclipse of the binary system. The dynamic behavior of the moving matter in the He I $\lambda\lambda$ 3614, 3964 and He I $\lambda$ 3888 lines is similar to that in the phase range (0.95-0.05) P. Interestingly, the highest radial velocity for the observation season in Sahade et al. (1959) was recorded in phases (0.11-0.13) P of the 5th cycle on June 10, 1955. The more negative absorption component of the He I $\lambda$ 3888 line showed the velocity of -(160-170) km/s, which is more than enough for the outflow of the matter from the Lagrange point L3. These more dense gas flows, when deflected by Coriolis forces outside the accretion disk in the Roche cavity of the gainer, push the outer part of the disk to the gainer surface (see Skulskij, 1992; Skulskyy, 2020a).
The range of phases (0.4-0.6) P eclipse of the accretion disk by the donor, as a specific part of the orbital period, deserves a separate study (by analogy with the studies in phases (0.9-0.1) P of the primary eclipse). The preliminary study of existing observations in this phase range leads to certain conclusions and ideas about the behavior of physical parameters in the vicinity of both stellar components. Although they require additional observations, they were worth paying attention to. Analysis of Tables 3 and 4 together with their illustrations 13-16 in the article Sahade et al. (1959) (see also Fig. 1-3 above), as well as the study of Skulskij (1992, 1993b, 1993a, 1993c); Ak et al. (2007) and others, leads to the conclusion that in these phases, in addition to complex shell lines formed near the accretion disk, clear changes of the curve of radial velocities of the gainer lines are visible. Their behavior can be explained by the Rossiter effect in the passage of the donor in front of the accretion disk. In phases (0.38-0.50) P the right part of the disk is observed (its left part being closed by the donor); in phases (0.45-0.55) P the central parts of the disk, closed by the donor, pass in front of the observer; in phases (0.5-0.6) P dominates in these lines the left part of the accretion disk (the donor obscures the right part of the disk). Approximately at 0.62 P, this effect disappears (further there are opened phases of the collision of the main flow into the accretion disk and the region of the magnetic field pole on the donor surface). In particular, this effect is well observed in phases (0.38-0.62) P in the behavior of the gainer radial velocities of the Si II $\lambda\lambda$ 6347, 6371 lines, for example, in Fig. 3 in Skulskij (1992), where it was revealed, and, especially in Fig. 5 in Ak et al. (2007). It is more likely that these lines are formed near the central part of the disk.
The visibility phases of both poles of the magnetic field (0.30-0.45) P and (0.75-0.90) P as specific parts of the orbital period are important for obtaining a general picture of the influence of the donor’s magnetic field on moving magnetized gas structures in near and farther medium regarding the donor. Some of the key facts identified in this study are, first of all, illustrated by the behavior of the radial velocities of helium lines arising from metastable levels. Fig. 1 shows that only within phases of about 0.1 P near the phases 0.855 P, both absorption components of the He I $\lambda$ 3888 line in Sahade et al. (1959) converge into a single absorption with radial velocities in the range -(125-135) km/s, i.e., with a radial velocity much greater than the first parabolic velocity. This indicates that in the phase of direct visibility of the pole of the magnetic field on the surface of the donor facing the gainer, the matter outflows in the direction of the axis of the donor magnetic field, perpendicular to the surface of the donor. According to Fig. 3, the radial velocities from the shells of the He I $\lambda\lambda$ 3614, 3964 singlet lines, although showing some stratification of their formation relative to the line He I $\lambda$ 3888, indicate that the He I lines of both series in phases close to 0.855 P behave in these phases in a similar way. The particularly sharp peaks of radial velocities with a rapid change in their values are observed in Fig. 4 in the He I $\lambda\lambda$ 3964, 5015 lines according to Skulsky & Kos (2011). Within 0.1 P in the 0.355 P and 0.855 P phases, there are sharp changes in the radial velocity from -55 km/s to -125 km/s, or 70 km/s relative to the center of gravity of the binary system. This indicates complex processes in the plasma structures of this binary system, which are close to the location of the magnetic poles on the donor surface. It should be recognized that Fig. 9 also confirms that the rapid changes in the radial velocity of the He I $\lambda$ 10830 line in phases immediately close to 0.855 P (at the mean velocities over the orbital period near -130 km/s - see Fig. 8) indicates the reality of rapid changes in plasma movement and events in this phase range.
Evidencing certain stratification of their formation, the He I lines of both series show in theses phases a certain relationship between their dynamic and energy parameters. Figure 9 also shows the parallel changes in the profile of the He I $\lambda$ 10830 line expressed through the dependence of $I_{v}/I_{r}=f(P)$, i.e., the ratio of the intensity of the short-wave and long-wave components of these emission profiles. This dependence is similar to the radiation increasing near the 0.855 P phase on light curves in the range of (3.5-4.6) m in Zeilik et al. (1982), which allowed to identify (see Skulskyy, 2020b) on the donor surface the region of this magnetic field pole. Figure 2 illustrates also the radial velocity curve from the red emission peak at the He I $\lambda$ 3888 line. Such two maxima are observed on the radial velocity curve of the red peak at the H${}_{\alpha}$ line (see Figures 5 in Skulskyy (2020b). The maxima at the 0.355 P and 0.855 P phases coincide with the phases of the two maxima on the curve of the effective magnetic field strength of the donor (see Figure 1 in Skulskyy (2020b) or Figure 2 in Burnashev & Skulskij (1991)). Moreover, the radial velocity curve for the Gaussian center of emission in H${}_{\alpha}$, as well as for such centers on radial velocity curves of the He I $\lambda\lambda$ 6678, 7065, and Si II $\lambda\lambda$ 6347, 6371 lines, see Figure 6 in Skulskyy (2020b), demonstrate that both their maxima match the extrema of the effective magnetic field strength of the donor (a similar shift on the $V_{r}$-curve of the center of the He I $\lambda$ 10830 emission as a whole is discussed by Girnyak et al. (1978)). The generation of the radiation flux and the formation of the emission-absorption profile of these lines can be largely carried out under the influence of the existing configuration of the donor magnetosphere in the interconnected spatial structures.
A similar conclusion is supported by the observations of Flora & Hack (1975). Conducted within one orbital cycle, they showed unusual variability of physical parameters within the phases (0.70-0.90) P close to the 0.855 P phase of the magnetic field pole close to the gainer. Fig. 6 shows a much lower velocity of the moving plasma than the first parabolic velocity in all five helium lines arising from metastable levels. This clearly indicates nonstationarity of the outflow of matter and the variable flow density in the direction of the dipole axis of the magnetic field. The behavior of radial velocities in phases (0.8-0.9) P in the shell lines Na I $\lambda\lambda$ 5889, 5895 is also particularly important. During 0.1 P, the moving plasma slows down, stops at a certain point, and accelerates again, stating the fact of actual variations in the sign of this movement. Thus, the study of observations of Flora & Hack (1975) also confirmed that the observer, looking perpendicular to the surface of the donor along the axis of the magnetic field, reveals the region of the donor surface that is responsible for the location of the pole of the magnetic field facing the gainer and the loss of matter directly from the donor surface in phase 0.855 P with its subsequent transfer to the gainer.
However, the fact of localization of the poles of the magnetic field on the surface of the donor, reflected in many studies, needs some clarification, based on the spatial location of its magnetic axis. Let us turn to the model of the donor as a magnetic rotator, by analogy to the model of a helium star as an oblique magnetic rotator presented by Shore & Brown (1990). This model (see Fig. 8 in Skulskyy (2020b)) may be quite illustrative since the dipole axis of its magnetic field is deviated by 28\degr relative to the orbital plane of the binary system (Skulskij, 1985). The center of the donor magnetic dipole is displaced by 0.08 of the distance between the centers of gravity of both components toward the gainer’s center. It is also clear that the magnetic pole, located on the surface of the donor and observed in phases of about 0.855 P, is more effective in terms of the amount of transferred matter to the gainer, which is confirmed by all studies. The ionized gas, directed by the magnetic field of the donor in the direction of its dipole axis from the surface of the donor, is deflected along the lines of the magnetic field primarily to the accretion disk. However, a certain amount of charged particles will move along the lines of force of the magnetic field in the direction to the second pole of the magnetic field on the surface of the donor, heating its surface. And, if in the first quadrature in the phases around 0.355 P the magnetic pole on the surface of the donor is located above the plane of the orbit, then in the phases around 0.855 P the magnetic pole on the surface of the donor is below the plane of the orbit, or vice versa. The detection of these poles also depends on the inclination of the plane of the orbit. Not everything is clear in this part because the inclination of the orbit in current research is taken from i = 81\degr in Mennickent & Djurašević (2013) to i = 93.5\degr in Mourard et al. (2018). This is close to the binary orbital inclination i = 90\degr, and regardless of whether the pole of the magnetic field is in the upper or lower hemisphere on the donor surface, the observer, due to the projection of the rounded shape of the surface on the line of sight, must register certain deviations of the surface heating maximum from the 0.355 P phase. This is confirmed by the absolute spectrophotometry of Burnashev & Skulskij (1991) of 1974-1985. They showed that the observer starts to register the excess radiation (a hot spot on the donor surface) only from the 0.355 P phase directed along the axis of the magnetic field (see Figure 1 in Skulskyy (2020b)). The excess of this radiation disappears near the 0.50 P phase, demonstrating during the orbital phases (0.37-0.49) P the rapid variability of the absolute radiation flux in the H${}_{\alpha}$ emission line and the continuum around this line (see Skulskyy, 2020b; Alexeev & Skulskij, 1989). The maximum of the excess of this radiation is really shifted from the 0.355 P phase and corresponds to the phases (0.43-0.47) P. Thus, in these phases, some surface heating on the donor surface may be formed due to nonstationary shock collisions of ionized gas, directed along the magnetic field lines to the magnetic field pole, the location of which is given by the spatial configuration of the magnetic field dipole. The phase range (0.43-0.47) P corresponds to the known minimum on the polarization curves for Beta Lyrae studied by Appenzeller & Hiltner (1967) and Lomax et al. (2012), which is interpreted as formed by collisions of gas flows with the accretion disk during the scattering of radiation by free electrons. This encourages the research of the reflection of the spatial configuration of the donor’s magnetic field in the polarization observations.
Acknowledgements.The author is thankful to V.I. Kudak for consultations.
References
Ak et al. (2007)
Ak, H., Chadima, P., Harmanec, P., et al., New findings supporting the
presence of a thick disc and bipolar jets in the $\beta$ Lyrae
system. 2007, A&A, 463, 233, DOI: 10.1051/0004-6361:20065536
Alduseva & Esipov (1969)
Alduseva, V. Y. & Esipov, V. F., The line He I $\lambda$
10830 in $\beta$ Lyr shell. 1969, AZh, 46, 113
Alexeev & Skulskij (1989)
Alexeev, G. N. & Skulskij, M. Y., Rapid variability of the spectrum of
$\beta$ Lyrae in the $H_{\alpha}$ region. 1989, Bull. Spec. Astroph. Obs., 28, 21
Appenzeller & Hiltner (1967)
Appenzeller, I. & Hiltner, W. A., True polarization curves for Beta
Lyrae. 1967, ApJ, 149, 353, DOI: 10.1086/149258
Bisikalo et al. (2000)
Bisikalo, D. V., Harmanec, P., Boyarchuk, A. A., Kuznetsov, O. A., &
Hadrava, P., Circumstellar structures in the eclipsing binary
$\beta$ Lyr A. Gasdynamical modelling confronted with
observations. 2000, A&A, 353, 1009
Burnashev & Skulskij (1978)
Burnashev, V. I. & Skulskij, M. Y., Absolute spectrophotometry of
$\beta$ Lyr. 1978, \krym, 58, 64
Burnashev & Skulskij (1991)
Burnashev, V. I. & Skulskij, M. Y., H${}_{{\alpha}}$
photometry and magnetic field of $\beta$ lyrae. 1991, \krym, 83, 108
Flora & Hack (1975)
Flora, U. & Hack, M., Spectrographic observations of $\beta$
Lyr during the international campaign of 1971. 1975, A&AS, 19,
57
Girnyak et al. (1978)
Girnyak, M. B., Skulskij, M. Y., Shanin, G. I., & Shcherbakov, A. G.,
The investigation of the emission line of He I $\lambda$ 10830
Å in the spectrum of Beta Lyrae. 1978, \krym, 58, 75
Harmanec & Scholz (1993)
Harmanec, P. & Scholz, G., Orbital elements of $\beta$ Lyrae
after the first 100 years of investigation. 1993, A&A, 279, 131
Hoffman et al. (1998)
Hoffman, J. L., Nordsieck, K. H., & Fox, G. K., Spectropolarimetric
evidence for a bipolar flow in Beta Lyrae. 1998, AJ, 115, 1576,
DOI: 10.1086/300274
Lomax et al. (2012)
Lomax, J. R., Hoffman, J. L., Elias, Nicholas M., I., Bastien, F. A.,
& Holenstein, B. D., Geometrical constraints on the hot spot in Beta
Lyrae. 2012, ApJ, 750, 59, DOI: 10.1088/0004-637X/750/1/59
Mennickent & Djurašević (2013)
Mennickent, R. E. & Djurašević, G., On the accretion disc and
evolutionary stage of $\beta$ Lyrae. 2013, MNRAS, 432, 799, DOI: 10.1093/mnras/stt515
Morgan et al. (1974)
Morgan, T. H., Potter, A. E., & Kondo, Y., Complex infrared emission
features in the spectrum of Beta Lyrae. 1974, ApJ, 190, 349,
DOI: 10.1086/152883
Mourard et al. (2018)
Mourard, D., Brož, M., Nemravová, J. A., et al., Physical
properties of $\beta$ Lyrae A and its opaque accretion disk.
2018, A&A, 618, A112, DOI: 10.1051/0004-6361/201832952
Rucinski et al. (2018)
Rucinski, S. M., Pigulski, A., Popowicz, A., et al., Light-curve
instabilities of $\beta$ Lyrae observed by the BRITE
satellites. 2018, AJ, 156, 12, DOI: 10.3847/1538-3881/aac38b
Sahade et al. (1959)
Sahade, J., Huang, S. S., Struve, O., & Zebergs, V., The spectrum of
Beta Lyrae. 1959, Transactions of the American Philosophical Society,
49, 1
Shore & Brown (1990)
Shore, S. N. & Brown, D. N., Magnetically controlled circumstellar matter
in the helium-strong stars. 1990, ApJ, 365, 665, DOI:
10.1086/169520
Skulskij (1975)
Skulskij, M. Y., Quantitative analysis of the spectrum of Beta Lyrae IV.
Line identifications for the faint component and the mass of both stars.
1975, AZh, 52, 710
Skulskij (1985)
Skulskij, M. Y., The magnetic field of the Beta-Lyrae system. 1985, \sal, 11, 21
Skulskij (1992)
Skulskij, M. Y., Study of $\beta$ Lyrae CCD spectra. Absorbtion
lines, orbital elements and disk structure of the gainer. 1992, \sal,
18, 287
Skulskij (1993a)
Skulskij, M. Y., Spectra of $\beta$ Lyr. Matter transfer and
circumstellar structures in presence of the donor’s magnetic field.
1993a, Astron. Lett., 19, 45
Skulskij (1993b)
Skulskij, M. Y., Study of $\beta$ Lyrae spectra - the Si II
$\lambda\lambda$ 6347, 6371 doublet and the discovery of the
cyclic variability of equivalent widths of lines of the loser’s ”magnetized”
atmosphere. 1993b, Astron. Lett., 19, 19
Skulskij (1993c)
Skulskij, M. Y., The spectrum of $\beta$ Lyrae: the SiII
$\lambda\lambda$ 6347, 6371 doublet in 1992 and its variation
from season to season. 1993c, Astron. Lett., 19,
160
Skulskij & Plachinda (1993)
Skulskij, M. Y. & Plachinda, S. I., A study of the magnetic field of the
bright component of $\beta$ Lyr in the SiII $\lambda\lambda$ 6347, 6371 lines. 1993, Pisma Astron. Zh., 19, 517
Skulsky & Kos (2011)
Skulsky, M. Y. & Kos, E. S., On the dynamics of circumstellar gaseous
structures and magnetic field of $\beta$ Lyrae. 2011, in Magnetic Stars. Proceedings of the International Conference, held in the
Special Astrophysical Observatory of the Russian AS, August 27- September 1,
2010, ed. I. Romanyuk & D. Kudryavtsev, 259–263
Skulskyy (2020a)
Skulskyy, M. Y., Formation of magnetized spatial structures in the Beta
Lyrae system. I. Observation as a research background of this phenomenon.
2020a, \caos, 50, 681, DOI:
10.31577/caosp.2020.50.3.681
Skulskyy (2020b)
Skulskyy, M. Y., Formation of magnetized spatial structures in the Beta
Lyrae system. II. Observation as a research background of this phenomenon.
2020b, \caos, 50, 717, DOI:
10.31577/caosp.2020.50.4.717
Struve (1941)
Struve, O., The Spectrum of $\beta$ Lyrae. 1941, ApJ,
93, 104, DOI: 10.1086/144249
Umana et al. (2000)
Umana, G., Maxted, P. F. L., Trigilio, C., et al., Resolving the radio
nebula around Beta Lyrae. 2000, A&A, 358, 229
Zeilik et al. (1982)
Zeilik, M., Heckert, P., Henson, G., & Smith, P., Infrared photometry
of Beta Lyrae: 1977-1982. 1982, AJ, 87, 1304, DOI:
10.1086/113217 |
Efficient ground state preparation in variational quantum eigensolver
with symmetry breaking layers
Chae-Yeun Park
Institute for Theoretical Physics, University of Cologne, 50937 Köln, Germany
Abstract
Variational quantum eigensolver (VQE) that solves the ground state of a Hamiltonian is a promising application of noisy intermediate-scale quantum devices.
The quantum alternating operator Ansatz (QAOA) is one of the most widely studied Ansätze for this purpose and solves many Hamiltonians reliably.
However, because of inherited symmetries from the Hamiltonian, the QAOA is not necessarily good at solving problems with symmetry broken nature.
In this paper, we propose a variational Ansatz with symmetry breaking layers for solving those systems.
Notably, our Ansatz finds a constant-depth circuit for solving a symmetry broken ground state which is impossible for the QAOA.
We also propose a simple learning technique that can choose a particular symmetry broken state among degenerate ground states.
I Introduction
Experimental progress in controlling quantum systems has allowed the first demonstration of quantum computational advantage [1] in the recent years, and a noisy quantum device with hundreds of qubits is getting visible.
Those devices, which are often dubbed as noisy intermediate-scale quantum (NISQ) [2] computers, are expected to solve a practical computational problem beyond the reach of classical computers.
Among possible applications, variational quantum eigensolver (VQE) [3, 4] that solves the ground state problem of quantum many-body Hamiltonian has gained lots of attention recently (see Ref. [5] for a recent review).
The VQE works by combining a parameterized quantum circuit and a classical optimization algorithm: A quantum circuit evaluates the expectation value of the Hamiltonian and its derivatives for the output quantum state, whereas a classical optimizer finds better parameters that minimize the energy.
As solving the ground state of quantum Hamiltonians are difficult for classical computers, one may easily get an advantage of VQEs.
Still, it is unclear which Ansatz and classical optimizer should be used to efficiently solve a given Hamiltonian.
One of the most widely studied Ansätze for solving translational invariant many-body Hamiltonians is the quantum alternating operator Ansatz (QAOA) [6] .
Inspired by a short-depth quantum algorithm for solving combinatory optimization problems [7], this Ansatz is composed of rotating gates using the terms of the Hamiltonian.
Even though the QAOA solves many different Hamiltonians reliably [8, 9, 10], however,
it is not necessarily good at solving problems with symmetry broken nature which prevail in many-body systems.
Indeed, there are local Hamiltonians whose ground states cannot be generated by this Ansatz in constant-depth albeit such a circuit exists, as the circuit obeys the same symmetry as the Hamiltonian [11].
In this paper, we devise a symmetry breaking Ansatz and explore its power for solving many-body Hamiltonians.
We first show that our Ansatz can generate a ground state within a different symmetry protected sector than the input state in constant-depth, which is impossible for the bare QAOA.
We next use our symmetry breaking Ansatz to prepare a particular symmetry broken state among degenerate ground states. This is done by simply adding a symmetry penalizing term to the loss function.
II Symmetry breaking Ansatz
The first model we study is the transverse field Ising (TFI) model for $N$ qubits defined by the Hamiltonian
$$\displaystyle H_{\rm TFI}=-\sum_{i=1}^{N}Z_{i}Z_{i+1}-h\sum_{i=1}^{N}X_{i}$$
(1)
where we impose the periodic boundary condition $Z_{N+1}=Z_{1}$. This model has two distinct (ferromagnetic and paramagnetic) phases depending on the strength of $h$ that are protected by the spin flip symmetry $P=\prod_{i}X_{i}$ [12]. The critical point of this model $h=1$ is well known.
This implies that if there is a circuit that commutes with $P$ and connects two ground states in different phases, then the circuit depth must be larger than a constant [13].
On the other hand, there is a finite-depth circuit that connects two different ground states if we do not restrict to such a symmetry, as the system is gapped unless $h=1$ [13, 14].
As a common Ansatz for preparing the ground state only utilizes terms within the Hamiltonian, it fails to represent such a circuit.
For example, the QAOA Ansatz for this Hamiltonian is given by $\prod_{k=D}^{1}L_{x}(\phi_{k})L_{zz}^{\rm odd}(\kappa_{k})L_{zz}^{\rm even}(\vartheta_{k})\ket{+}^{\otimes N}$ where $L_{x}(\phi)=\exp[-i\phi\sum_{i}X_{i}]$, $L_{zz}^{\rm odd}(\kappa)=\exp[-i\kappa\sum_{i=1}^{N/2}Z_{2i-1}Z_{2i}]$, and $L_{zz}^{\rm even}(\vartheta)=\exp[-i\vartheta\sum_{i=1}^{N/2}Z_{2i}Z_{2i+1}]$. As all gates commute with $P$, i.e. $[P,L_{zz}^{\rm odd}]=[P,L_{zz}^{\rm even}]=[P,L_{x}]=0$, and the input state $\ket{+}^{\otimes N}$ is the ground state of the Hamiltonian when $h\rightarrow\infty$, we know that preparing the ground state for $h<1$ requires circuit depth larger than a constant.
Indeed, theoretical and numerical studies have found that this type of Ansatz needs depth $D\geq N/2$ to prepare the ground state faithfully [15, 8].
We now add symmetry breaking layers in VQE Ansatz and see whether it can achieve lower circuit depth for preparing the ground state.
Our Ansatz for the TFI is given as
$$\displaystyle\ket{\psi(\boldsymbol{\theta})}=\prod_{j=D}^{1}L_{z}(\phi_{j})L_{x}(\kappa_{j})L_{zz}(\vartheta_{j})\ket{+}^{\otimes N}$$
(2)
where $L_{zz}(\theta_{j})=\exp[-i\theta_{j}\sum_{i}Z_{i}Z_{i+1}]$, $L_{x}(\kappa_{j})=\exp[-i\kappa_{j}\sum_{i}X_{i}]$, $L_{z}(\phi_{j})=\exp[-i\phi_{j}\sum_{i}Z_{i}]$, and $\boldsymbol{\theta}=\{\vartheta_{j},\kappa_{j},\phi_{j}\}_{j=1}^{D}$ is a set of all parameters. All layers in the Ansatz preserve the translational symmetry but the $L_{z}(\phi_{j})$ layers break the symmetry $P$ of the Hamiltonian.
To observe the effect of symmetry breaking layers, we first present the relative errors for different system sizes, depths, with and without the symmetry breaking $L_{z}(\phi_{j})$ layers.
We optimize parameters using the quantum natural gradient [16, 17, 18]: For each epoch $t$, we update parameters as $\boldsymbol{\theta}_{t+1}=\boldsymbol{\theta}_{t}-\eta(\mathcal{F}+\lambda_{t}\mathbb{1})^{-1}\nabla_{\theta}\braket{\psi(\boldsymbol{\theta}_{t})}{H}{\psi(\boldsymbol{\theta}_{t})}$ where $\eta$ is the learning rate, $\mathcal{F}=(\mathcal{F}_{ij})$ is the quantum Fisher matrix, and $\lambda_{t}$ is a (step dependent) regularization constant.
We choose this optimizer as it works more reliably in solving the ground problem both for classical neural networks [19] and VQEs [9].
For the quantum Fisher matrix, we use the centered one $\mathcal{F}=\mathcal{F}_{ij}^{\rm c}=\mathfrak{Re}\{\braket{\partial_{\theta_{i}}\psi(\boldsymbol{\theta}_{t})}{\partial_{\theta_{j}}\psi(\boldsymbol{\theta}_{t})}-\braket{\partial_{\theta_{i}}\psi(\boldsymbol{\theta}_{t})}{\psi(\boldsymbol{\theta}_{t})}\braket{\psi(\boldsymbol{\theta}_{t})}{\partial_{\theta_{j}}\psi(\boldsymbol{\theta}_{t})}\}$ [17] mostly (unless otherwise stated), but the uncentered one $\mathcal{F}_{ij}^{\rm nc}=\mathfrak{Re}\{\braket{\partial_{\theta_{i}}\psi(\boldsymbol{\theta}_{t})}{\partial_{\theta_{j}}\psi(\boldsymbol{\theta}_{t})}\}$ [16] is also considered when it improves the performance.
The difference between two can be understood using the notion of the projected Hilbert space [18].
For the hyperparameters, we typically use $\eta=0.01$ and $\lambda_{t}=\max(100.0\times 0.9^{t},10^{-3})$.
We show the optimized normalized energies $\widetilde{E}=(E_{\rm VQE}-E_{\rm GS})/E_{\rm GS}$ for different $N$ and $D$ in Fig. 1 when $h=0.5$.
We have used the Fisher matrix with the centering term $\mathcal{F}^{\rm c}$ and initial values of parameters $\{\vartheta_{j},\kappa_{j}\,\phi_{j}\}$ are sampled from the normal distribution $\mathcal{N}(0,\sigma^{2})$ with $\sigma=0.001$ besides $D=3,5$ at Fig. 1(b).
When $D=3$ (all $N$) and $5$ (for $N\geq 12$), this initialization does not give any better energy than the Ansatz without symmetry breaking layers.
We instead found that initializing parameters for the symmetry breaking layers $\{\phi_{j}\}$ with samples from $\mathcal{N}(2\pi/D,\sigma^{2})$ finds better optima in these cases.
However, this initialization does not change the results for $D=4$, and performs even worse when $D\geq 6$ (see Appendix A for detail comparisons).
Fig. 1(a) shows that the ground state is only found for $D\geq N/2$ when symmetry breaking layers are absent, which is consistent with Refs. [8, 9].
On the other hand, results with symmetry breaking layers [Fig. 1(b)] clearly demonstrate that converged energies are significantly improved for $5\leq D<N/2$.
Most importantly, converged normalized energies are $\leq 10^{-7}$ for all $N$ when $D\geq 9$, which implies that our Ansatz finds a constant-depth circuit for solving the ground state.
In addition, the results show that there is a finite-size effect up to $N=18$ where the accurate ground state is only obtained when $D\geq N/2$.
However, there is a huge gap between the converged energies and the true ground state when $D=4$.
In fact, the results for $D=4$ are even worse than those of $D=3$, which signals that the optimizer gets stuck in local minima. This type of convergence problem is already observed in Ref. [9].
To obtain a better convergence for $D=4$, we employ a transfer learning technique.
Instead of starting from a randomly initialized circuit, we insert a block in the middle of the converged circuit from $D=3$ and perturb all parameters by adding small numbers sampled from $\mathcal{N}(0,\sigma^{\prime 2})$ (where we typically use $\sigma^{\prime}=0.01$). We then optimize the full circuit using the quantum natural gradient.
We show converged energies from the transfer learning in Fig. 2.
Fig. 2(a) illustrates that the transfer learning succeeds to find a better optima for $D=4$. However, the result for $(N,D)=(10,5)$ is worse than that of the random initialization, which implies that the transfer learning does not necessarily find the global optimia.
We additionally plot a learning curve for $(N,D)=(18,4)$ in Fig. 2(b) where the initial energy is much higher due to an added perturbation but it eventually finds a lower energy state.
We next consider the transverse field cluster (TFC) model, the Hamiltonian of which is given as
$$\displaystyle H_{\rm TFC}=-\sum_{i=1}^{N}Z_{i}X_{i+1}Z_{i+2}-h\sum_{i=1}^{N}X_{i}.$$
(3)
It is known that the phase transition of this model also takes place at $h=1$ [20].
As the terms $Z_{i}X_{i+1}Z_{i+2}$ are mutually commuting, the ground state at $h=0$ is the common eigenvector of those operators which is also known as the cluster state.
Two relevant symmetries that determine the ground state of this model are $P_{1}=\prod_{i=1}^{N/2}X_{2i}$ and $P_{2}=\prod_{i=1}^{N/2}X_{2i-1}$ [21, 22].
We thus expect that a circuit cannot bring the product state $\ket{+}^{\otimes N}$ to the ground state of the Hamiltonian when $h<1$ in a constant depth as long as it commutes with $P_{1}$ and $P_{2}$.
We study whether a symmetry breaking layer can improve it. To see this, we construct our Ansatz
$$\displaystyle\ket{\psi(\boldsymbol{\theta})}$$
$$\displaystyle=\prod_{j=D}^{1}L_{z}^{\rm even}(\phi_{j})L_{z}^{\rm odd}(\chi_{j})$$
$$\displaystyle\qquad\qquad\times L_{x}(\kappa_{j})L_{zxz}(\theta_{j})\ket{+}^{\otimes N}$$
(4)
where $L_{zxz}(\theta_{j})=\exp[-i\theta_{j}\sum_{i=1}^{N}Z_{i}X_{i+1}Z_{i+2}]$, $L_{z}^{\rm even/odd}(\phi_{j})=\exp[-i\phi_{j}\sum_{i=1}^{N/2}Z_{2i/2i+1}]$.
Converged normalized energies with and without symmetry breaking layers for $h=0.5$ are shown in Fig. 3.
The results without symmetry breaking layers show that circuits with $D\geq\lfloor N/4\rfloor$ solve the ground state accurately.
However, the symmetry breaking layers only improve the results for $D=3$ when $N\geq 16$ for this Hamiltonian.
In addition, such an improvement is observed only when the uncentered Fisher matrix $\mathcal{F}^{\rm nc}$ and the initial value $2\pi/D$ for symmetry breaking layers are used.
We moreover see that our Ansatz does not find a constant-depth circuit for solving the ground state based upon the results up to $N=20$.
We attribute this to the finite-size effect and expect that much larger system size $N$ (which is beyond our computational capacity) is required to find the constant-depth circuit.
III Preparing a symmetry broken ground state
We next consider the cluster Hamiltonian [Eq. (3) with $h=0$] with the open boundary condition, which is given as
$$\displaystyle H_{\rm cluster}=-\sum_{i=1}^{N-2}Z_{i}X_{i+1}Z_{i+2}.$$
(5)
Ground states of this model are stabilized by $N-2$ terms ($Z_{i}X_{i+1}Z_{i+2}\ket{\psi_{\rm GS}}=\ket{\psi_{\rm GS}}$ for $i\in[1,\cdots,N-2]$), thus $4$-fold degenerate. As the operators $P_{1}$ and $P_{2}$ commute with all stabilizers, they further define the ground state manifold.
For the QAOA, the ground state must be $+1$ eigenstate of $P_{1}$ and $P_{2}$ as they commute with the circuit and $P_{1,2}\ket{+}^{\otimes N}=+1\ket{+}^{\otimes N}$.
In contrast, we show our Ansatz [Eq. (4)] (with $L_{zxz}(\theta_{j})$ in the open boundary condition) can be used to prepare a particular state within the manifold.
The main idea is minimizing the expectation value of $O=H_{\rm cluster}+\alpha_{1}P_{1}+\alpha_{2}P_{2}$ for suitable $\alpha_{1}$ and $\alpha_{2}$, instead of the Hamiltonian itself. We expect the obtained state is an eigenstate with the corresponding eigenvalue $\pm 1$ of $P_{i}$ when $\mathrm{sign}(\alpha_{i})=\mp 1$.
As an example, we study the VQE for preparing the ground state with $P_{1}=P_{2}=-1$ using $(N,D)=(14,8)$.
We have tested VQEs with the varying learning rates $\eta\in[0.01,0.025,0.05,0.1]$ and $\alpha_{1}=\alpha_{2}=\alpha\in[0.1,0.3,0.5,1.0,2.0,4.0]$, and found that the desired state can be prepared by the best for $\eta=0.025$, $\alpha=2.0$. Especially, all VQE runs have converged to $\braket{P_{1}}=\braket{P_{2}}=-1$ (accuracy within $10^{-8}$) with these values of $\eta$ and $\alpha$.
The resulting learning curve is shown in Fig. 4(a).
However, the converged energy $\widetilde{E}\approx 1.502\times 10^{-3}$ for $D=8$ may not be satisfying compared to the other models we have studied above.
Thus, we further study whether increasing $D$ helps the convergence in Fig. 4(b).
The results show that the converged energies are getting accurate as we increase $D$. Precisely, $47$ instances out of $48$ independent VQEs runs have converged to $\widetilde{E}\leq 10^{-8}$ when $D=20$.
Acknowledgement
The author thanks David Wierichs for helpful discussions.
This project was funded by the Deutsche Forschungsgemeinschaft under Germany’s Excellence Strategy - Cluster of Excellence Matter and Light for Quantum Computing (ML4Q) EXC 2004/1-390534769 and within the CRC network TR 183 (project grant 277101999) as part of project B01.
The numerical simulations were performed on the JUWELS cluster at the Forschungszentrum Juelich.
Source codes for the current paper can be found in Github repository [23].
Appendix A Optimization of the Ansatz with symmetry breaking layers
When the Ansatz contains symmetry breaking layers, an optimization algorithm easily stuck in local minima. We here compare several different set-ups for the transverse field Ising model we have studied in Sec. II.
We show results from the transverse Ising model with different initial values and using the centered and uncentered Fisher matrix in Fig. 5.
We can see that the result with $\{\phi_{j}\}$ initialized from $\mathcal{N}(0,\sigma^{2})$ and using the centered Fisher matrix is the most natural.
However, when $D=3,5$, the results with $\{\psi_{j}\}$ initialized from $\mathcal{N}(2\pi/D,\sigma^{2})$ show better convergence.
We have also found that using the uncentered Fisher matrix improves convergence for $D=4$ where the centered Fisher matrix failed to find an appropriate optima. Our results suggest that the learning landscape of the VQEs with symmetry breaking layers is rugged especially when the parameters are not sufficient to describe the ground state accurately.
References
Arute et al. [2019]
F. Arute, K. Arya,
R. Babbush, D. Bacon, J. C. Bardin, R. Barends, R. Biswas, S. Boixo, F. G. S. L. Brandao, D. A. Buell, and et al., Quantum supremacy using a programmable
superconducting processor, Nature 574, 505–510 (2019).
Preskill [2018]
J. Preskill, Quantum Computing in
the NISQ era and beyond, Quantum 2, 79 (2018).
Peruzzo et al. [2014]
A. Peruzzo, J. McClean,
P. Shadbolt, M.-H. Yung, X.-Q. Zhou, P. J. Love, A. Aspuru-Guzik, and J. L. O’brien, A variational eigenvalue solver on a photonic quantum
processor, Nature communications 5, 1 (2014).
McClean et al. [2016]
J. R. McClean, J. Romero,
R. Babbush, and A. Aspuru-Guzik, The theory of variational hybrid quantum-classical
algorithms, New Journal of Physics 18, 023023 (2016).
Cerezo et al. [2020]
M. Cerezo, A. Arrasmith,
R. Babbush, S. C. Benjamin, S. Endo, K. Fujii, J. R. McClean, K. Mitarai, X. Yuan,
L. Cincio, et al., Variational quantum
algorithms, arXiv preprint arXiv:2012.09265 (2020).
Hadfield et al. [2019]
S. Hadfield, Z. Wang,
B. O’Gorman, E. G. Rieffel, D. Venturelli, and R. Biswas, From the quantum approximate optimization algorithm to a quantum
alternating operator ansatz, Algorithms 12, 34 (2019).
Farhi et al. [2014]
E. Farhi, J. Goldstone, and S. Gutmann, A quantum approximate optimization
algorithm, arXiv
preprint arXiv:1411.4028 (2014).
Ho and Hsieh [2019]
W. W. Ho and T. H. Hsieh, Efficient variational simulation of
non-trivial quantum states, SciPost Phys 6, 29 (2019).
Wierichs et al. [2020]
D. Wierichs, C. Gogolin, and M. Kastoryano, Avoiding local minima in variational
quantum eigensolvers with the natural gradient optimizer, Physical Review Research 2, 043246 (2020).
Wiersema et al. [2020]
R. Wiersema, C. Zhou,
Y. de Sereville, J. F. Carrasquilla, Y. B. Kim, and H. Yuen, Exploring entanglement and optimization within the hamiltonian
variational ansatz, PRX Quantum 1, 020319 (2020).
Bravyi et al. [2020]
S. Bravyi, A. Kliesch,
R. Koenig, and E. Tang, Obstacles to variational quantum optimization from
symmetry protection, Physical Review Letters 125, 260505 (2020).
Chen et al. [2011a]
X. Chen, Z.-C. Gu, and X.-G. Wen, Complete classification of one-dimensional gapped
quantum phases in interacting spin systems, Physical Review B 84, 235128 (2011a).
Chen et al. [2011b]
X. Chen, Z.-C. Gu, and X.-G. Wen, Classification of gapped symmetric phases in
one-dimensional spin systems, Physical Review B 83, 035107 (2011b).
Bachmann et al. [2012]
S. Bachmann, S. Michalakis, B. Nachtergaele, and R. Sims, Automorphic equivalence
within gapped phases of quantum lattice systems, Communications in Mathematical Physics 309, 835 (2012).
Mbeng et al. [2019]
G. B. Mbeng, R. Fazio, and G. Santoro, Quantum annealing: A journey through
digitalization, control, and hybrid quantum variational schemes, arXiv preprint
arXiv:1906.08948 (2019).
McArdle et al. [2019]
S. McArdle, T. Jones,
S. Endo, Y. Li, S. C. Benjamin, and X. Yuan, Variational ansatz-based quantum simulation of imaginary time evolution, npj Quantum Information 5, 1 (2019).
Stokes et al. [2020]
J. Stokes, J. Izaac,
N. Killoran, and G. Carleo, Quantum natural gradient, Quantum 4, 269 (2020).
Hackl et al. [2020]
L. Hackl, T. Guaita,
T. Shi, J. Haegeman, E. Demler, and J. I. Cirac, Geometry of variational methods: dynamics of closed quantum
systems, SciPost Phys. 9, 48 (2020).
Park and Kastoryano [2020]
C.-Y. Park and M. J. Kastoryano, Geometry of learning
neural quantum states, Physical
Review Research 2, 10.1103/physrevresearch.2.023232 (2020).
Zonzo and Giampaolo [2018]
G. Zonzo and S. M. Giampaolo, n-cluster models in a
transverse magnetic field, Journal of Statistical Mechanics: Theory and Experiment 2018, 063103 (2018).
Son et al. [2011]
W. Son, L. Amico, and V. Vedral, Topological order in 1d cluster state protected by
symmetry, Quantum Information Processing 11, 1961–1968 (2011).
Zeng et al. [2019]
B. Zeng, X. Chen, D.-L. Zhou, and X.-G. Wen, Quantum
information meets quantum matter (Springer-Verlag
New York, 2019).
Park [2021]
C.-Y. Park, https://github.com/chaeyeunpark/efficient-vqe-symmetry-breaking
(2021). |
Canonical metric on the space of symplectic invariant tensors and its applications
Shigeyuki Morita
[email protected]
Graduate School of Mathematical Sciences,
University of Tokyo,
3-8-1 Komaba,
Meguro-ku, Tokyo, 153-8914, Japan
Dedicated to the memory of Professor Akio Hattori
Abstract.
Let $\Sigma_{g}$ be a closed oriented surface of genus $g$
and let $H_{\mathbb{Q}}$ denote $H_{1}(\Sigma_{g};\mathbb{Q})$ which we understand
to be the standard symplectic vector space over $\mathbb{Q}$ of dimension $2g$.
We introduce a canonical metric on the space $(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}$
of symplectic invariant tensors by analyzing the structure
of the vector space $\mathbb{Q}\mathcal{D}^{\ell}(2k)$ generated by linear chord diagrams
with $2k$ vertices. This space, equipped with a certain
inner product, serves as a universal model for $(H^{\otimes 2k})^{\mathrm{Sp}}$
for any $g$. We decompose
$\mathbb{Q}\mathcal{D}^{\ell}(2k)$ as an orthogonal direct sum of eigenspaces $E_{\lambda}$ where
$\lambda$ is indexed by
the set of all the Young diagrams with $k$ boxes.
We give a formula for the eigenvalue
$\mu_{\lambda}$ of $E_{\lambda}$ and thereby
we obtain a complete
description of how the spaces $(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}$ degenerate according
as the genus decreases
from the stable range $g\geq k$
to the last case $g=1$
with the largest eigenvalue
$2g(2g+1)\cdots(2g+k-1)$.
As an application of our canonical metric, we obtain certain relations among the
Mumford-Morita-Miller tautological classes, in a systematic way, which hold in the tautological
algebra in cohomology of the moduli space of curves. We also indicate other
possible applications
such as characteristic classes of
transversely symplectic foliations and a project with T. Sakasai and M. Suzuki [30]
where we study the structure of
the symplectic derivation Lie algebra.
Key words and phrases:symplectic tensor, symmetric group, Young diagram, tautological algebra, moduli space of curves, transversely symplectic foliation
2000 Mathematics Subject Classification: Primary 20C30;32G15 , Secondary 20J06; 55R40
1. Introduction and statements of the main results
Let $\Sigma_{g}$ be a closed oriented surface of genus $g$.
We denote $H_{1}(\Sigma_{g};\mathbb{Q})$ simply by $H_{\mathbb{Q}}$ and let
$$\mu:H_{\mathbb{Q}}\otimes H_{\mathbb{Q}}\rightarrow\mathbb{Q}$$
be the intersection pairing which is a non-degenerate skew symmetric
bilinear form. If we choose a symplectic basis of $H_{\mathbb{Q}}$, then
the automorphism group of $(H_{\mathbb{Q}},\mu)$ can be identified with the symplectic
group $\mathrm{Sp}(2g,\mathbb{Q})$ and $H_{\mathbb{Q}}$ serves as the fundamental
representation of $\mathrm{Sp}(2g,\mathbb{Q})$. Let $(H_{\mathbb{Q}}^{\otimes{2k}})^{\mathrm{Sp}}$
denote the $\mathrm{Sp}(2g,\mathbb{Q})$-invariant subspace of the tensor product
$H_{\mathbb{Q}}^{\otimes{2k}}$. Now consider the bilinear mapping
$$\mu^{\otimes 2k}:H_{\mathbb{Q}}^{\otimes{2k}}\otimes H_{\mathbb{Q}}^{\otimes{2%
k}}\rightarrow\mathbb{Q}$$
defined by
$$(u_{1}\otimes\cdots\otimes u_{2k})\otimes(v_{1}\otimes\cdots\otimes v_{2k})%
\mapsto\prod_{i=1}^{2k}\,\mu(u_{i},v_{i})\quad(u_{i},v_{i}\in H_{\mathbb{Q}}).$$
which is clearly symmetric. Hence the restriction of
$\mu^{\otimes 2k}$ to
the subspace $(H_{\mathbb{Q}}^{\otimes{2k}})^{\mathrm{Sp}}\otimes(H_{\mathbb{Q}}^{\otimes{2k%
}})^{\mathrm{Sp}}$
induces a symmetric bilinear form on $(H_{\mathbb{Q}}^{\otimes{2k}})^{\mathrm{Sp}}$.
Recall that irreducible representations of the symmetric group
$\mathfrak{S}_{k}$ is indexed by Young diagrams
$\lambda=[\lambda_{1}\lambda_{2}\cdots\lambda_{h}]$
whose number of boxes, denoted by $|\lambda|=\lambda_{1}+\cdots\lambda_{h}$,
is $k$. We denote the corresponding representation by $\lambda_{\mathfrak{S}_{k}}$.
For each Young diagram $\lambda$ as above, we define two types
of Young diagrams, denoted by $2\lambda$ and $\lambda^{\delta}$, with $2k$ boxes. One is defined as
$2\lambda=[2\lambda_{1}2\lambda_{2}\cdots 2\lambda_{h}]$ and the other is
defined to be $\lambda^{\delta}=[\lambda_{1}\lambda_{1}\lambda_{2}\lambda_{2}\cdots\lambda_{h%
}\lambda_{h}]$.
For example, if $\lambda=[431]$, then
$2\lambda=[862]$ and $\lambda^{\delta}=[443311]$. We call Young
diagrams with the former type (resp. the latter type)
of even type (resp. with multiple double floors).
Theorem 1.1.
For any $g$, the symmetric bilinear form $\mu^{\otimes 2k}$ on $(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}$
is positive definite so that it defines a metric on this space. Furthermore, there exists an orthogonal
direct sum decomposition
$$(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}\cong\bigoplus_{|\lambda|=k,\ h(%
\lambda)\leq g}U_{\lambda}$$
in terms of certain subspaces $U_{\lambda}$.
With respect to the natural action of $\mathfrak{S}_{2k}$
on $(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}$, each subspace $U_{\lambda}$
is an irreducible $\mathfrak{S}_{2k}$-submodule and we have an isomorphism
$$U_{\lambda}\cong(\lambda^{\delta})_{\mathfrak{S}_{2k}}.$$
In other words, the above orthogonal
direct sum decomposition gives also the irreducible decomposition of the
$\mathfrak{S}_{2k}$-module $(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}$.
Corollary 1.2.
The dimension of the space of invariant tensors $(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}$
is given by
$$\dim\,(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}=\sum_{|\lambda|=k,\ h(%
\lambda)\leq g}\mathrm{dim}\,(\lambda^{\delta})_{\mathfrak{S}_{2k}}.$$
Remark 1.3.
Recall here that there are a number of explicit formulas for the dimension of irreducible representations of
symmetric groups, e.g. the hook length formula, the number of standard tableaux on the
Young diagram and the formula of Frobenius (cf [4]). In particular
$$\dim\,U_{[k]}=\dim\,[k^{2}]_{\mathfrak{S}_{2k}}=\frac{1}{k+1}\binom{2k}{k}%
\quad(\text{the $k$-th Catalan number})$$
is equal to $\dim\,(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}$ for the case $g=1$.
We mention that, in this case, a basis of this space was given
in a classical paper [32] by Rumer, Teller and Weyl.
Furthermore Mihailovs [21] extended this work to obtain a nice
basis of $(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}$
for any $g$.
However, their basis is different from ours because it
depends essentially on the choice of ordering of tensors so that it is not canonical.
A linear chord diagram with $2k$ vertices is a partition of the set
$\{1,2,\cdots,2k\}$ into $k$-tuple
$$C=\{\{i_{1},j_{1}\},\cdots,\{i_{k},j_{k}\}\}\quad$$
of pairs $\{i_{\ell},j_{\ell}\}\ (\ell=1,\cdots,k)$. Here we assume
$$i_{1}<\cdots<i_{k},\ i_{\ell}<j_{\ell}\ (\ell=1,\cdots,k)$$
so that the above expression is uniquely determined.
We consider $C$ to be a graph with $2k$ vertices and $k$ edges (chords)
each of which connects $i_{\ell}$ with $j_{\ell}$ for $\ell=1,\cdots,k$.
Let
$\mathcal{D}^{\ell}(2k)$
denote the set of all the linear chord diagrams with $2k$ vertices.
It has $(2k-1)!!$ elements and let $\mathbb{Q}\mathcal{D}^{\ell}(2k)$ be the vector
space over $\mathbb{Q}$ spanned by $\mathcal{D}^{\ell}(2k)$.
The symmetric group $\mathfrak{S}_{2k}$ acts on $\mathcal{D}^{\ell}(2k)$ naturally
so that $\mathbb{Q}\mathcal{D}^{\ell}(2k)$ has a structure of an $\mathfrak{S}_{2k}$-module.
We define a symmetric bilinear mapping
$$\langle\ \,,\ \rangle:\mathbb{Q}\mathcal{D}^{\ell}(2k)\times\mathbb{Q}\mathcal%
{D}^{\ell}(2k)\rightarrow\mathbb{Q}[g]$$
by setting
$$\langle C,C^{\prime}\rangle=(-1)^{k-r}(2g)^{r}$$
for any two elements $C,C^{\prime}\in\mathcal{D}^{\ell}(2k)$, where $r$ denotes the number
of connected components of the graph $C\cup C^{\prime}$
(here we consider interiors of $2k$ edges to be mutually disjoint), and extending this bilinearly.
If we let $g$ to take a specific value, then the above can be understood
to be a symmetric bilinear pairing. It is clear that this pairing
is $\mathfrak{S}_{2k}$-invariant. Namely the equality
$$\langle\gamma(C),\gamma(C^{\prime})\rangle=\langle C,C^{\prime}\rangle\quad(C,%
C^{\prime}\in\mathcal{D}^{\ell}(2k))$$
holds for any $\gamma\in\mathfrak{S}_{2k}$. Let $M_{2k}$ denote
the corresponding intersection matrix. In other words, it is the matrix
which represents the linear mapping
$$\mathcal{M}:\mathbb{Q}\mathcal{D}^{\ell}(2k)\rightarrow\mathbb{Q}\mathcal{D}^{%
\ell}(2k)$$
defined as
$$\mathcal{M}(C)=\sum_{D\in\mathcal{D}^{\ell}(2k)}\langle C,D\rangle\,D.$$
Theorem 1.4.
The intersection matrix $M_{2k}$ is positive semi-definite for any $g$,
so that it defines a metric on
$\mathbb{Q}\mathcal{D}^{\ell}(2k)$ depending on $g$.
Furthermore, there exists an orthogonal
direct sum decomposition
$$\mathbb{Q}\mathcal{D}^{\ell}(2k)\cong\bigoplus_{|\lambda|=k}E_{\lambda}$$
in terms of eigenspaces $E_{\lambda}$ of $\mathcal{M}$ which
are defined independently of the genus $g$.
The eigenvalue of $E_{\lambda}$ , denoted by $\mu_{\lambda}$, is given by
the following formula
$$\mu_{\lambda}=\prod_{\text{b: box of $\lambda$}}(2g-2s_{b}+t_{b})$$
where $s_{b}$ denotes the number of columns of $\lambda$
which are on the left of the column containing $b$ and
$t_{b}$ denotes the number of rows which are above the
row containing $b$.
With respect to the natural action of $\mathfrak{S}_{2k}$
on $\mathbb{Q}\mathcal{D}^{\ell}(2k)$, $E_{\lambda}$
is an irreducible $\mathfrak{S}_{2k}$-submodule and we have an isomorphism
$$E_{\lambda}\cong(2\lambda)_{\mathfrak{S}_{2k}}.$$
In other words, the above orthogonal
direct sum decomposition gives also the irreducible decomposition of the
$\mathfrak{S}_{2k}$-module $\mathbb{Q}\mathcal{D}^{\ell}(2k)$.
Theorem 1.5.
The correspondence
$$\lambda\mapsto\mu_{\lambda}$$
defines an injective mapping from the set of all the Young diagrams
$\lambda$ with $k$ boxes into the space of polynomials in $g$ of degree $k$.
Example 1.6.
The largest (resp. the smallest) eigenvalue occurs for $\lambda=[1^{k}]$
(resp. $\lambda=[k]$) and the corresponding eigenvalues are
$$\mu_{[1^{k}]}=2g(2g+1)\cdots(2g+k-1),\quad\mu_{[k]}=2g(2g-2)\cdots(2g-2k+2).$$
The space $\mathbb{Q}\mathcal{D}^{\ell}(2k)$ is a universal model for
the symplectic invariant tensors because we can make the following construction.
First there exists a natural homomorphism
$$\Phi:\mathbb{Q}\mathcal{D}^{\ell}(2k)\rightarrow(H_{\mathbb{Q}}^{\otimes 2k})^%
{\mathrm{Sp}}$$
which is surjective for any $g$ and “anti” $\mathfrak{S}_{2k}$-equivariant
(see Proposition 2.1 for details).
In the dual setting, there exists also a natural homomorphism
$$\mathcal{K}:H_{\mathbb{Q}}^{\otimes 2k}\rightarrow\mathbb{Q}\mathcal{D}^{\ell}%
(2k)$$
which enumerate all the contractions and can detect $\mathrm{Sp}$-invariant
component of any given tensor
(see section $\lx@sectionsign 3$, in particular Proposition 3.2, for details).
We use the same symbol $\mathcal{K}$ for its restriction
to the subspace $(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}$.
Theorem 1.7.
The following diagram is commutative
$$\begin{CD}\mathbb{Q}\mathcal{D}^{\ell}(2k)@>{\mathcal{M}}>{}>\mathbb{Q}%
\mathcal{D}^{\ell}(2k)\\
@V{\Phi}V{}V\Big{\|}\\
(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}@>{\mathcal{K}}>{}>\mathbb{Q}%
\mathcal{D}^{\ell}(2k)\end{CD}$$
and all the homomorphisms $\Phi,\mathcal{M},\mathcal{K}$ are isomorphisms for $g\geq k$.
For any $\lambda$ with $|\lambda|=k$, $U_{\lambda}=\Phi(E_{\lambda^{\prime}})$
where $\lambda^{\prime}$ denotes the conjugate
Young diagram of $\lambda$.
Furthermore the subspaces $\mathcal{K}(U_{\lambda})=E_{\lambda^{\prime}}\subset\mathbb{Q}\mathcal{D}^{%
\ell}(2k)$ are mutually
orthogonal to each other with respect to
the ordinary Euclidean metric on $\mathbb{Q}\mathcal{D}^{\ell}(2k)$ induced by taking
$\mathcal{D}^{\ell}(2k)$ as an orthonormal basis.
This work was originally motivated by an attempt, begun some years ago, to extend
our former work [28] on the structure of
the tautological algebra of the moduli space of curves.
Although it turned out to require further works in this direction, we present in $\lx@sectionsign 6$
an explicit set of relations among the Mumford-Morita-Miller
tautological classes by making use of the main result of this
paper (see Theorem 6.3 and Conjecture 6.5). On the other hand,
in a joint work with T. Sakasai and M. Suzuki [30], where we study
the structure of the symplectic derivation Lie algebra, we
use the canonical metric introduced in this paper and found another role of it.
Because of these, we describe a detailed proof of the above results.
Acknowledgement The author would like to thank Carel Faber,
Richard Hain and Nariya Kawazumi for enlightening discussions
about the cohomology of the mapping class groups
on various occasions in the $1990$’s and early $2000$’s.
He also would like to thank
Takuya Sakasai and Masaaki Suzuki
for their helps in recent years
which greatly encouraged him to write up this paper.
2. Space of linear chord diagrams and $(H^{\otimes 2k})^{\mathrm{Sp}}$
For each linear chord diagram $C\in\mathcal{D}^{\ell}(2k)$, define
$$a_{C}\in(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}$$
by permuting the elements $(\omega_{0})^{\otimes 2k}$ in such a way that the $s$-th part
$(\omega_{0})_{s}$ of this tensor product goes to $(H_{\mathbb{Q}})_{i_{s}}\otimes(H_{\mathbb{Q}})_{j_{s}}$, where $(H_{\mathbb{Q}})_{i}$
denotes the $i$-th component of $H_{\mathbb{Q}}^{\otimes 2k}$, and multiplied by the factor
$$\mathrm{sgn}\,C=\mathrm{sgn}\begin{pmatrix}1&2&\cdots&2k-1&2k\\
i_{1}&j_{1}&\cdots&i_{k}&j_{k}\end{pmatrix}.$$
Here $\omega_{0}\in(H_{\mathbb{Q}}^{\otimes 2})^{\mathrm{Sp}}$ denotes the symplectic class.
Now define a linear mapping
$$\Phi:\mathbb{Q}\mathcal{D}^{\ell}(2k)\rightarrow(H_{\mathbb{Q}}^{\otimes 2k})^%
{\mathrm{Sp}}$$
by setting $\Phi(C)=a_{C}$.
Proposition 2.1.
The correspondence
$$\Phi:\mathbb{Q}\mathcal{D}^{\ell}(2k)\rightarrow(H_{\mathbb{Q}}^{\otimes 2k})^%
{\mathrm{Sp}}$$
is surjective for any $g$ and bijective for any $g\geq k$. Furthermore this correspondence
is “anti” $\mathfrak{S}_{2k}$-equivariant in the sense that
$$\Phi(\gamma(C))=\mathrm{sgn}\,\gamma\ \gamma(\Phi(C))$$
for any $C\in\mathbb{Q}\mathcal{D}^{\ell}(2k)$ and $\gamma\in\mathfrak{S}_{2k}$.
Proof.
The surjectivity follows from a classical result of Weyl on symplectic invariants.
The latter part can be shown by comparing how the signs of permutations act
on the symplectic invariant tensors and the linear chord diagrams.
∎
Proposition 2.2.
The following diagram is commutative
$$\begin{CD}\mathbb{Q}\mathcal{D}^{\ell}(2k)\otimes\mathbb{Q}\mathcal{D}^{\ell}(%
2k)@>{\langle\ ,\ \rangle}>{}>\mathbb{Q}[g]\\
@V{\Phi\otimes\Phi}V{}V@V{}V{\text{evaluation}}V\\
(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}\otimes(H_{\mathbb{Q}}^{\otimes 2k}%
)^{\mathrm{Sp}}@>{\mu^{\otimes 2k}}>{}>\mathbb{Q}.\end{CD}$$
To prove this, we prepare a few facts.
Lemma 2.3.
Let $x_{1},\cdots,x_{g},y_{1},\cdots,y_{g}$ be a symplectic basis of $H_{\mathbb{Q}}$ and let
$$\omega_{0}=\sum_{i=1}^{g}\{x_{i}\otimes y_{i}-y_{i}\otimes x_{i}\}\in(H_{%
\mathbb{Q}}^{\otimes 2})^{\mathrm{Sp}}$$
be the symplectic class.
$\mathrm{(i)}\ \text{For any two elements $u,v\in H_{\mathbb{Q}}$, we have the identity}$
$$\mu(u,v)=\sum_{i=1}^{g}\{\mu(u,x_{i})\mu(v,y_{i})-\mu(u,y_{i})\mu(v,x_{i})\}.$$
$\mathrm{(ii)}\ \text{For any two elements $u,v\in H_{\mathbb{Q}}$, we have the%
identity}$
$$\mu(u,v)=\mu^{\otimes 2}(u\otimes v,\omega_{0}).$$
Proof.
It is easy to verify (i) while (ii) is just a restatement of (i) in terms of the linear mapping
$$\mu^{\otimes 2}:H_{\mathbb{Q}}^{\otimes 2}\otimes H_{\mathbb{Q}}^{\otimes 2}%
\rightarrow\mathbb{Q}.$$
∎
Lemma 2.4 (Lemma 3.3 of [28]).
For any two indices $i,j$ with $1\leq i<j\leq 2k$, let
$$p_{ij}:H_{\mathbb{Q}}^{\otimes 2k}\rightarrow H_{\mathbb{Q}}^{\otimes 2k}$$
be the linear map defined by first taking the intersection number of the $i$-th entry with
the $j$-th entry and then putting the element $\omega_{0}$ there. Namely we set
$$\displaystyle p_{ij}(u_{1}\otimes\cdots\otimes u_{i}\otimes\cdots$$
$$\displaystyle\otimes u_{j}\otimes\cdots\otimes u_{2k})=$$
$$\displaystyle\mu(u_{i},u_{j})\sum_{s=1}^{g}$$
$$\displaystyle\{u_{1}\otimes\cdots\otimes x_{s}\otimes\cdots\otimes y_{s}%
\otimes\cdots\otimes u_{2k})$$
$$\displaystyle-u_{1}\otimes\cdots\otimes y_{s}\otimes\cdots\otimes x_{s}\otimes%
\cdots\otimes u_{2k}\}.$$
Then we have
$$p_{ij}(a_{C})=\begin{cases}2g\ a_{C}\quad\{i,j\}\in C\\
-a_{C^{\prime}}\quad\{i,j\}\notin C\end{cases}$$
where $C^{\prime}\in\mathcal{D}^{\ell}(2k)$ is the linear chord diagram defined as follows.
Let $i^{\prime},j^{\prime}$ be the
indices such that $\{i,j^{\prime}\},\{i^{\prime},j\}\in C$. Then
$$C^{\prime}=C\setminus\{\{i,j^{\prime}\},\{i^{\prime},j\}\}\cup\{\{i,j\},\{i^{%
\prime},j^{\prime}\}\}.$$
Proof of Proposition 2.2.
By the linearity, it is enough to prove the following. For any two
linear chord diagrams $C,C^{\prime}\in\mathcal{D}^{\ell}(2k)$, we have the identity
(1)
$$\mu^{\otimes 2k}(a_{C}\otimes a_{C^{\prime}})=\langle C,C^{\prime}\rangle=(-1)%
^{k-r}(2g)^{r}$$
where the right equality is the definition of the pairing
$\langle\ ,\ \rangle$ and $r$ denotes the number of connected components of the graph $C\cup C^{\prime}$.
To prove the above equality (1), we decompose $\mu^{\otimes 2k}$ as
$$\mu^{\otimes 2k}=\prod_{\ell=1}^{k}\mu_{i_{\ell},j_{\ell}}^{\otimes 2}$$
where $C=\{\{i_{1},j_{1}\},\cdots,\{i_{k},j_{k}\}\}$ as before and $\mu_{i_{\ell},j_{\ell}}^{\otimes 2}(u\otimes v)$
$(u,v\in H_{\mathbb{Q}}^{\otimes 2k})$ denotes
taking the product of $\mu(u_{i_{\ell}},v_{i_{\ell}})$ and $\mu(u_{j_{\ell}},v_{j_{\ell}})$.
Now we consider the first chord $\{i_{1},j_{1}\}$ of $C$
(actually $i_{1}=1$) and apply $\mu_{i_{1},j_{1}}^{\otimes 2}$ on $a_{C}\otimes a_{C^{\prime}}$. By Lemma 2.3
this can be calculated by applying the operator $p_{i_{1}j_{1}}$ on $a_{C^{\prime}}$ and by Lemma 2.4 the answer is
given as follows. Namely if $\{i_{1},j_{1}\}\in C^{\prime}$, then we obtain the number $\mu^{\otimes 2}(\omega_{0}\otimes\omega_{0})=2g$
and the linear chord diagram $C^{\prime}$ is unchanged so that $C^{\prime}_{1}=C^{\prime}$
while if $\{i_{1},j_{1}\}\notin C^{\prime}$, then we obtain the factor $-1$ and the new linear chord diagram $C^{\prime}_{1}$ is
the one obtained from $C^{\prime}$ by replacing $\{i_{1},j^{\prime}_{1}\},\{i^{\prime}_{1},j_{1}\}\in C^{\prime}$ by $\{i_{1},j_{1}\},\{i^{\prime}_{1},j^{\prime}_{1}\}$.
In the latter case, it is easy to observe that the number of connected components of the graph $C\cup C^{\prime}_{1}$
increases by one than that of the graph $C\cup C^{\prime}$ because one of the connected components of
the latter graph is split into two components one of which corresponds to the common edge $\{i_{1},j_{1}\}$
of the two graphs $C$ and $C^{\prime}_{1}$. Next we consider the second chord $\{i_{2},j_{2}\}$ of $C$
and apply $\mu_{i_{2},j_{2}}^{\otimes 2}$ on $a_{C}\otimes a_{C^{\prime}_{1}}$. By the same argument as above,
if $\{i_{2},j_{2}\}\in C^{\prime}_{1}$, then we obtain the number $2g$
and the linear chord diagram $C^{\prime}_{1}$ is unchanged so that $C^{\prime}_{2}=C^{\prime}_{1}$
while if $\{i_{2},j_{2}\}\notin C^{\prime}_{1}$, then we obtain the factor $-1$ and the new linear chord diagram $C^{\prime}_{2}$ is
the one obtained from $C^{\prime}_{1}$ by replacing $\{i_{2},j^{\prime}_{2}\},\{i^{\prime}_{2},j_{2}\}\in C^{\prime}_{1}$ by $\{i_{2},j_{2}\},\{i^{\prime}_{2},j^{\prime}_{2}\}$.
In the latter case, the number of connected components of the graph $C\cup C^{\prime}_{2}$
increases by one than that of the graph $C\cup C^{\prime}_{1}$ because one of the connected components of
the latter graph is split into two components one of which corresponds to the common edge $\{i_{2},j_{2}\}$
of the two graphs $C$ and $C^{\prime}_{2}$. Continuing in this way, we finally arrive at the last one $C^{\prime}_{k}$ which
coincides with $C$ and along the way we obtain the factor $(2g)$ $r$-times while the factor $(-1)$
$(k-r)$ times for the evaluation of $\mu^{\otimes 2k}$. This completes the proof.
∎
Proposition 2.5.
The irreducible decomposition of the $\mathfrak{S}_{2k}$-module $\mathbb{Q}\mathcal{D}^{\ell}(2k)$
is given by
$$\mathbb{Q}\mathcal{D}^{\ell}(2k)=\bigoplus_{|\lambda|=k}E_{\lambda}$$
where $E_{\lambda}$ is isomorphic to $(2\lambda)_{\mathfrak{S}_{2k}}$.
Proof.
The symmetric group $\mathfrak{S}_{2k}$ acts on $\mathcal{D}^{\ell}(2k)$ naturally and
this action is clearly transitive. If we denote by $C_{0}\in\mathcal{D}^{\ell}(2k)$ the
linear chord diagram
$$C_{0}=\{\{1,2\},\{3,4\},\cdots,\{2k-1,2k\}\}$$
then the isotropy group of it
is the subgroup $H_{k}\subset\mathfrak{S}_{2k}$
consisting of permutations which fix $C_{0}$ considered as an element
of the power set of $\{1,2,\cdots,2k\}$. $H_{k}$ is isomorphic to
$$(\mathfrak{S}_{2}\times\overset{\text{$k$-times}}{\cdots}\times\mathfrak{S}_{2%
})\rtimes\mathfrak{S}_{k}$$
and we have a bijection
(2)
$$\mathfrak{S}_{2k}/H_{k}\cong\mathcal{D}^{\ell}(2k)$$
from the set of left $H_{k}$-cosets of $\mathfrak{S}_{2k}$ to $\mathcal{D}^{\ell}(2k)$.
Note that the number of elements of both sets is equal to $(2k-1)!!$. It is easy to see that the
above bijection (2) is an isomorphism of $\mathfrak{S}_{2k}$-set.
As is well known, the $\mathfrak{S}_{2k}$-module $\mathfrak{S}_{2k}/H_{k}$
is isomorphic to the induced representation $\mathrm{Ind}^{\mathfrak{S}_{2k}}_{H_{k}}1_{H_{k}}$,
where $1_{H_{k}}$ denotes the one dimensional trivial representation of $H_{k}$.
Hence the proof of the present proposition is reduced to the following one.
∎
Proposition 2.6.
Let $\mathfrak{S}_{2k}$ be the symmetric group of order $2k$ and let $H_{k}$ denote the subgroup
defined above. Then the irreducible decomposition of the induced representation $\mathrm{Ind}^{\mathfrak{S}_{2k}}_{H_{k}}1_{H_{k}}$
is given by
$$\mathrm{Ind}^{\mathfrak{S}_{2k}}_{H_{k}}1_{H_{k}}\cong\bigoplus_{|\lambda|=k}(%
2\lambda)_{\mathfrak{S}_{2k}}.$$
This is an elementary fact about representations of symmetric groups
and can be proved by applying standard argument in the theory of representations of finite groups
(see e.g. [10]).
However, in order to motivate later arguments and also for completeness,
we give a proof along the line of the ingredient of this paper.
We first recall a few terminologies about representations of the symmetric groups from [4].
We will use them also in section $\lx@sectionsign 5$.
For a Young diagram $\lambda$ with $|\lambda|=d$, choose a Young tableau $T$ (say a standard one) on $\lambda$.
Then we set
$$\displaystyle P_{\lambda}$$
$$\displaystyle=\{\gamma\in\mathfrak{S}_{d};\text{$\gamma$ preserves each row}\}$$
$$\displaystyle Q_{\lambda}$$
$$\displaystyle=\{\gamma\in\mathfrak{S}_{d};\text{$\gamma$ preserves each column}\}$$
$$a_{\lambda}=\sum_{\gamma\in P_{\lambda}}\gamma,\quad b_{\lambda}=\sum_{\gamma%
\in Q_{\lambda}}\mathrm{sgn}\,\gamma\,\gamma$$
The element
$$c_{\lambda}=a_{\lambda}b_{\lambda}\in\mathbb{Q}[\mathfrak{S}_{d}]$$
is called a Young symmetrizer. We can also use the following element
$$c^{\prime}_{\lambda}=b_{\lambda}a_{\lambda}\in\mathbb{Q}[\mathfrak{S}_{d}]$$
which is another form of Young symmetrizer and can play the same role as above.
In this paper, we use both ones.
Next we recall the following well known result.
Theorem 2.7 (Mackey’s restriction formula, see e.g. [34]).
Let $G$ be a finite group and let $H,K$ be subgroups of $G$. Let
$\rho$ be a representation of $H$ and for each $(H,K)$-double coset
representative $g$, let $\rho_{g}$ be the representation of $H_{g}=gHg^{-1}\cap K$
defined by $\rho_{g}(x)=\rho(g^{-1}xg)\ (x\in gHg^{-1})$.
Then we have the equality
$$\mathrm{Res}^{G}_{K}\,\mathrm{Ind}^{G}_{H}\,\rho=\bigoplus_{g\in K\backslash G%
/H}\mathrm{Ind}^{K}_{H_{g}}\,\rho_{g}.$$
Proof of Proposition 2.6.
First, we show that
(3)
$$\left|H_{k}\backslash\mathfrak{S}_{2k}/H_{k}\right|=p(k).$$
Namely the number of elements of the double cosets
$H_{k}\backslash\mathfrak{S}_{2k}/H_{k}$ is equal to $p(k)$ which is the
number of partitions of $k$. Recall that the set of left $H_{k}$-cosets of $\mathfrak{S}_{2k}$
can be canonically identified with $\mathcal{D}^{\ell}(2k)$. $H_{k}$ acts on this set
from the left and it is easy to see that the space of orbits of this action can be
identified with the set of $H_{k}$-double cosets of $\mathfrak{S}_{2k}$.
Now define the $C_{0}$-relative type of an element $C\in\mathcal{D}^{\ell}(2k)$,
denoted by $\tau_{C_{0}}(C)$,
as follows. Consider the graph $\Gamma_{C}=C_{0}\cup C$. Each connected component
of $\Gamma_{C}$ contains even number of vertices from the set of all vertices $\{1,2,\cdots,2k\}$.
If we enumerate all these even numbers in the decreasing order, then we obtain
an even type Young diagram $2\lambda$ where $\lambda$ is
a partition of $k$ (equivalently a Young diagram with $k$ boxes).
We put $\tau_{C_{0}}(C)=\lambda$. For example $\tau_{C_{0}}(C_{0})=[1^{k}]$.
Then we can show the following. For any two linear chord diagrams $C,D\in\mathcal{D}^{\ell}(2k)$,
the corresponding left $H_{k}$-cosets belong to the same right $H_{k}$-orbits if and only if
$\tau_{C_{0}}(C)=\tau_{C_{0}}(D)$. Also, for any Young diagram $\lambda$ with $|\lambda|=k$,
there exists $C\in\mathcal{D}^{\ell}(2k)$ such that $\tau_{C_{0}}(C)=\lambda$.
Since there are exactly $p(k)$ Young diagrams $\lambda$ with $|\lambda|=k$, the claim
(3) is proved.
Next, by the Frobenius reciprocity theorem, for any Young diagram $\mu$ with $|\mu|=2k$ we have
(4)
$$\text{multiplicity of $\mu_{\mathfrak{S}_{2k}}$ in $\mathrm{Ind}^{\mathfrak{S}%
_{2k}}_{H_{k}}1_{H_{k}}$}=\mathrm{dim}\,(\mathrm{Res}^{\mathfrak{S}_{2k}}_{H_{%
k}}\,\mu_{\mathfrak{S}_{2k}})^{H_{k}}.$$
We show that for any Young diagram $\lambda$ with $|\lambda|=k$
(5)
$$\mathrm{dim}\,(\mathrm{Res}^{\mathfrak{S}_{2k}}_{H_{k}}\,(2\lambda)_{\mathfrak%
{S}_{2k}})^{H_{k}}\geq 1.$$
To prove this, it suffices to show that
$$g_{\lambda}=\left(\sum_{\gamma\in H_{k}}\gamma\right)c_{(2\lambda)}\not=0\ \in%
\mathbb{Q}[\mathfrak{S}_{2k}]$$
because this is clearly an $H_{k}$-invariant element.
Here we choose a standard tableau $T$ on $2\lambda$ to be the one defined as follows.
We give the numbers $1,2,\cdots,2\lambda_{1}$ to fill the first row from the left to the right
and then the numbers $2\lambda_{1}+1,\cdots,2\lambda_{1}+2\lambda_{2}$ to the second row
from the left to the right
and so on, where $\lambda=[\lambda_{1}\cdots\lambda_{h}]$. Consider
$k$ ordered pairs
$$\langle 1,2\rangle,\langle 2,3\rangle,\cdots,\langle 2k-1,2k\rangle$$
of numbers each of which is placed on some row of $2\lambda$. Now the subgroup
$H_{k}$ is isomorphic to the semi-direct product $(\mathfrak{S}_{2})^{k}\rtimes\mathfrak{S}_{k}$
and the $i$-th $\mathfrak{S}_{2}$ acts on $\langle 2i-1,2i\rangle$ naturally while $\mathfrak{S}_{k}$
acts on the above pairs by permutations (preserving each order). It follows that we can write
$$\sum_{\gamma\in H_{k}}=\sum_{\gamma\in\mathfrak{S}_{k}}\gamma\sum_{\delta\in(%
\mathfrak{S}_{2})^{k}}\delta$$
and also that $(\mathfrak{S}_{2})^{k}$ is contained in $P_{2\lambda}$ which is the subgroup of
$\mathfrak{S}_{2k}$ consisting of elements which permute the numberings
but within each row of $2\lambda$.
Therefore
$$g_{\lambda}=\left(\sum_{\gamma\in\mathfrak{S}_{k}}\gamma\right)\left(\sum_{%
\delta\in(\mathfrak{S}_{2})^{k}}\delta\right)a_{2\lambda}b_{2\lambda}=2^{k}%
\left(\sum_{\gamma\in\mathfrak{S}_{k}}\gamma\right)c_{2\lambda}$$
because $\delta\,a_{2\lambda}=a_{2\lambda}$ for any $\delta\in P_{2\lambda}$.
Recall that the coefficient of the unit element $1\in\mathfrak{S}_{2k}$ in the expression $c_{2\lambda}$ is $1$.
To prove the assertion $g_{\lambda}\not=0$, it suffices to show that the coefficient of $1$ is different from $0$.
Suppose for some elements $\gamma\in\mathfrak{S}_{k},g\in P_{2\lambda},h\in Q_{2\lambda}$, the
equality
(6)
$$\gamma gh=1$$
holds. It implies that $gh\in\mathfrak{S}_{k}$. Let $\langle 2i-1,2i\rangle$ be any one of the above ordered
pair. Then we can write
$$gh\langle 2i-1,2i\rangle=\langle gh(2i-1),gh(2i)\rangle=\langle 2j-1,2j\rangle$$
for some $j$, because the action of $gh=\gamma^{-1}\in\mathfrak{S}_{2k}$ reduces to permutations of the ordered pairs preserving
the order of each pair. Recall that $h(2i-1)$ (resp. $h(2i))$ lies in the same column as $2j-1$ (resp. $2j$).
Since $g$ preserves each row and still $gh(2i-1)=2j-1$ and $gh(2i)=2j$ belong to the same row, we can conclude that
$h(2j-1)$ and $h(2j)$ belong to the same row. It follows that the image of $\langle 2i-1,2i\rangle$ under
$h$ is again one of the ordered pairs. Since $\langle 2i-1,2i\rangle$ was an arbitrary ordered pair,
we can conclude that $h$ is contained in the subgroup $\mathfrak{S}_{k}\subset\mathfrak{S}_{2k}$.
Now observe that the $sgn$ of any element of this subgroup $\mathfrak{S}_{k}$, but considered as
an element of $\mathfrak{S}_{2k}$, is $+1$ because it acts on $\{1,2,\cdots,2k\}$
by an even permutation. Thus, whenever the equality (6) occurs, we have $\mathrm{sgn}\,h=1$.
We can now conclude that the coefficient of $1$ in $g_{\lambda}$ is different from $0$ as claimed.
Finally, we prove
(7)
$$\mathrm{dim}\,\left(\mathrm{Res}^{\mathfrak{S}_{2k}}_{H_{k}}\,\mathrm{Ind}^{%
\mathfrak{S}_{2k}}_{H_{k}}1_{H_{k}}\right)^{H_{k}}=p(k).$$
It is easy to see that the representation
$\mathrm{Res}^{\mathfrak{S}_{2k}}_{H_{k}}\,\mathrm{Ind}^{\mathfrak{S}_{2k}}_{H_%
{k}}1_{H_{k}}$
is nothing other than the one associated with the left $H_{k}$-action on the
set $\mathcal{D}^{\ell}(2k)\cong\mathfrak{S}_{2k}/H_{k}$. Therefore, by a well-known
basic fact, the multiplicity of the trivial representation occurring in it
is the same as the number of orbits of this action which is $p(k)$ by
(3).
More precisely, we can apply Theorem 2.7
to the most simple case where $G=\mathfrak{S}_{2k},H=K=H_{k}$ and $\rho$ is the
one dimensional trivial representation. Then we obtain
$$\mathrm{Res}^{\mathfrak{S}_{2k}}_{H_{k}}\,\mathrm{Ind}^{\mathfrak{S}_{2k}}_{H_%
{k}}1_{H_{k}}=\bigoplus_{g\in H_{k}\backslash{\mathfrak{S}_{2k}}/H_{k}}\mathrm%
{Ind}^{H_{k}}_{(H_{k})_{g}}\,1_{(H_{k})_{g}}.$$
Taking the $H_{k}$-trivial part, this implies
$$\left(\mathrm{Res}^{\mathfrak{S}_{2k}}_{H_{k}}\,\mathrm{Ind}^{\mathfrak{S}_{2k%
}}_{H_{k}}1_{H_{k}}\right)^{H_{k}}=\bigoplus_{g\in H_{k}\backslash{\mathfrak{S%
}_{2k}}/H_{k}}\left(\mathrm{Ind}^{H_{k}}_{(H_{k})_{g}}\,1_{(H_{k})_{g}}\right)%
^{H_{k}}.$$
Again by the Frobenius reciprocity theorem, we have
$$\mathrm{dim}\,\left(\mathrm{Ind}^{H_{k}}_{(H_{k})_{g}}\,1_{(H_{k})_{g}}\right)%
^{H_{k}}=1\ $$ for any $$g\in H_{k}\backslash{\mathfrak{S}_{2k}}/H_{k}$$
and the claim (7) follows.
Since there are exactly $p(k)$ Young diagrams $\lambda$ with $|\lambda|=k$, (5) implies
$$\mathrm{dim}\,(\mathrm{Res}^{\mathfrak{S}_{2k}}_{H_{k}}\,\mu_{\mathfrak{S}_{2k%
}})^{H_{k}}=\begin{cases}1\quad\text{if $\mu=2\lambda$ for some $\lambda$}\\
0\quad\text{otherwise}.\end{cases}$$
In view of (4), we can now conclude
$$\mathrm{Ind}^{\mathfrak{S}_{2k}}_{H_{k}}1_{H_{k}}\cong\bigoplus_{|\lambda|=k}(%
2\lambda)_{\mathfrak{S}_{2k}}$$
as required.
∎
Proposition 2.8.
A basis for the subspace $E_{\lambda}$ is given by
$$\{c_{\tau}(C_{0});\tau\in ST(\lambda)\}$$
where $ST(\lambda)$ denotes the set of all the standard tableaux on the Young
diagram $2\lambda$ and $c_{\tau}$ denotes the Young symmetrizer corresponding to
$\tau\in ST(\lambda)$.
Proposition 2.9.
The correspondence
$$\lambda\mapsto\mu_{\lambda}$$
defines an injective mapping from the set of all the Young diagrams
$\lambda$ with $k$ boxes into the space of polynomials in $g$ of degree $k$.
Proof.
It is enough to recover the shape of any Young diagram $\lambda=[\lambda_{1}\cdots\lambda_{h}]$ from the
corresponding polynomial $\mu_{\lambda}$.
Recall that $\mu_{\lambda}$ is the product of $\mu_{b}=2g-2s_{b}+t_{b}$ where $b$ runs through all the boxes of
$\lambda$. It is clear that the largest (reap. the smallest) one among $\mu_{b}$ is $(2g+h-1)$ (resp. $(2g-2\lambda_{1}+2)$).
It follows that $\mu_{\lambda}$ determines $h$, which is the number of rows of $\lambda$, and $\lambda_{1}$.
Now consider the quotient
$$\mu_{\lambda}^{(1)}=\frac{\mu_{\lambda}}{2g(2g+1)\cdots(2g+h-1)(2g-2)\cdots(2g%
-2\lambda_{1}+2)}$$
which is again a polynomial in $g$ of degree $(k-h-\lambda_{1}+1)$. Let $\lambda^{(1)}$ be the
Young diagram obtained from $\lambda$ by deleting the first row and the first column. It has
$(k-h-\lambda_{1}+1)$ boxes. For example, if $\lambda=[432]$, then $\lambda^{(1)}=[21]$.
Now $\mu_{\lambda}^{(1)}$ is expressed as the product of $\mu_{b}$ where $b$ runs through
all the boxes of $\lambda^{(1)}$ which we consider as a subset of the original $\lambda$.
Now it is easy to see that the largest (reap. the smallest) one among
$\mu_{b}\ (b\in\lambda^{(1)}$ is $(2g+h_{2}-3)$ (resp. $(2g-2\lambda_{2}+3)$)
where $h_{2}$ denotes the number of boxes of the second column of $\lambda$.
It follows that we can recover $h_{2}$ and $\lambda_{2}$. Continuing this kind of argument,
we see that we can eventually recover the shape of $\lambda$. This completes the proof.
∎
3. Orthogonal decomposition of $(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}$
In this section, we deduce Theorem 1.1 from Theorem 1.4 whose proof
will be given in the next section. See Table 1 for an example of the orthogonal direct
sum decomposition.
Proof of Theorem 1.1.
In view of Proposition 2.1 and Proposition 2.2, we have only to determine
the kernel of the surjective linear mapping
$$\Phi:\mathbb{Q}\mathcal{D}^{\ell}(2k)\rightarrow(H_{\mathbb{Q}}^{\otimes 2k})^%
{\mathrm{Sp}}$$
when we fix a specific genus $g$.
Let us denote the eigenvalue of $E_{\lambda}$ by $\mu_{\lambda}(g)$ to indicate its dependence
on the genus. The smallest factor in the expression of $\mu_{\lambda}(g)$,
considered as a polynomial in $g$ of degree $k$, is
$(2g-2\lambda_{1}+2).$
Therefore we have
$$\mu_{\lambda}(g)\not=0\ \Leftrightarrow\ \lambda_{1}\leq g$$
It follows that
$$\mathrm{Ker}\,\Phi=\bigoplus_{|\lambda|=k,\lambda_{1}>g}E_{\lambda}$$
and $\Phi$ induces an isometric linear isomorphism
$$\bar{\Phi}:\bigoplus_{|\lambda|=k,\lambda_{1}\leq g}E_{\lambda}\cong(H_{%
\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}$$
which is also an “anti” -isomorphism as $\mathfrak{S}_{2k}$-modules.
Now for each Young diagram $\lambda$ with $|\lambda|=k$ and
$h(\lambda)\leq g$, we set
$$U_{\lambda}=\bar{\Phi}(E_{\lambda^{\prime}})$$
where $\lambda^{\prime}$ denotes the conjugate Young diagram of $\lambda$ so that
$\lambda^{\prime}_{1}=h(\lambda)\leq g$. Also as an $\mathfrak{S}_{2k}$-module,
we have isomorphisms
$$U_{\lambda}\cong E_{\lambda^{\prime}}\otimes[1^{2k}]_{\mathfrak{S}_{2k}}\cong(%
2\lambda^{\prime})_{\mathfrak{S}_{2k}}\otimes[1^{2k}]_{\mathfrak{S}_{2k}}\cong%
\lambda^{\delta}_{\mathfrak{S}_{2k}}$$
where the last isomorphism follows because it is easy to see that the identity
$$(2\lambda^{\prime})^{\prime}=\lambda^{\delta}$$
holds for any Young diagram $\lambda$. Thus we obtain an orthogonal direct sum decomposition
$$(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}=\bigoplus_{|\lambda|=k,h(\lambda)%
\leq g}U_{\lambda}$$
completing the proof.
∎
If we are concerned only with the easier part of Theorem 1.1, namely
the irreducible decomposition of $(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}$ as an $\mathfrak{S}_{2k}$-module,
then there exists an alternative proof based on a classical result together with an argument
in the symplectic representation theory. We briefly sketch it here. First recall the following classical result.
Theorem 3.1 ([4],Theorem 6.3. (2)).
The $\mathrm{GL}(2g,\mathbb{Q})$-irreducible decomposition of $H_{\mathbb{Q}}^{\otimes 2k}$ is give by
$$H_{\mathbb{Q}}^{\otimes 2k}\cong\bigoplus_{|\lambda|=2k,h(\lambda)\leq 2g}\,%
\lambda_{\mathrm{GL}}^{\oplus\dim\,\lambda_{\mathfrak{S}_{2k}}}$$
where $\lambda_{\mathrm{GL}}$ denotes the irreducible representation of $\mathrm{GL}(2g,\mathbb{Q})$
corresponding to $\lambda$.
If we combine this with the following fact
$$\mathrm{dim}\,(\lambda_{\mathrm{GL}})^{\mathrm{Sp}}=\begin{cases}&1\quad(\text%
{$\lambda$: multiple double floors})\\
&0\quad(\text{otherwise})\end{cases}$$
which follows from the restriction law corresponding to
the pair $\mathrm{Sp}(2g,\mathbb{Q})\subset\mathrm{GL}(2g,\mathbb{Q})$
(see Proposition 4.1 of [29]), then we obtain a proof of
the easier part of Theorem 1.1. We mention here that,
in a recent paper [1],
Enomoto and Satoh describes the $\mathrm{Sp}$-irreducible
decomposition of $H_{\mathbb{Q}}^{\otimes 2k}$.
In a dual setting to the homomorphism
$\Phi:\mathbb{Q}\mathcal{D}^{\ell}(2k)\rightarrow(H_{\mathbb{Q}}^{\otimes 2k})^%
{\mathrm{Sp}}$
given in $\lx@sectionsign$2, we define a linear mapping
$$\mathcal{K}:H_{\mathbb{Q}}^{\otimes 2k}\rightarrow\mathbb{Q}\mathcal{D}^{\ell}%
(2k)$$
by
$$\mathcal{K}(\xi)=\sum_{C\in\mathcal{D}^{\ell}(2k)}\alpha_{C}(\xi)\,C\quad(\xi%
\in H_{\mathbb{Q}}^{\otimes 2k})$$
where $\alpha_{C}\in\mathrm{Hom}(H_{\mathbb{Q}}^{\otimes 2k},\mathbb{Q})^{\mathrm{Sp}}$ is defined as
$$\alpha_{C}(u_{1}\otimes\cdots\otimes u_{2k})=\mathrm{sgn}\,C\prod_{s=1}^{k}\mu%
(u_{i_{s}},u_{j_{s}})\ (u_{i}\in H_{\mathbb{Q}}).$$
We use the same symbol $\mathcal{K}$ to denote its
restriction to the subspace $(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}$.
Proposition 3.2.
$\mathrm{(i)}\ \text{
The following diagram is commutative}$
$$\begin{CD}\mathbb{Q}\mathcal{D}^{\ell}(2k)@>{\mathcal{M}}>{}>\mathbb{Q}%
\mathcal{D}^{\ell}(2k)\\
@V{\Phi}V{}V\Big{\|}\\
(H_{\mathbb{Q}}^{\otimes 2k})^{\mathrm{Sp}}@>{\mathcal{K}}>{}>\mathbb{Q}%
\mathcal{D}^{\ell}(2k)\end{CD}$$
and the lower homomorphism $\mathcal{K}$ is
injective for any $g$ and bijective for any $g\geq k$.
$\mathrm{(ii)\ }$
For any element $\xi\in E_{\lambda}\subset\mathbb{Q}\mathcal{D}^{\ell}(2k)$, we have the following equality
$$\mathcal{K}(\Phi(\xi))=\mu_{\lambda\,}\xi.$$
Proof.
First we prove $\mathrm{(i)}$. The commutativity of the diagram follows by comparing the definitions
of the homomorphisms $\mathcal{M},\Phi$ and $\mathcal{K}$. The latter claim follows from the basic
result of Weyl that any $\mathrm{Sp}$-invariant can be described in terms of various contractions.
Next we prove $\mathrm{(ii)}$. Since $\xi$ is an eigenvector of $\mathcal{M}$
corresponding to the eigenvalue $\mu_{\lambda}$, we have $\mathcal{M}(\xi)=\mu_{\lambda}\,\xi$.
If we combine this with the commutativity just proved above, then the claim follows.
∎
Proof of Theorem 1.7.
The fact that $U_{\lambda}=\Phi(E_{\lambda^{\prime}})$ follows from Proposition 2.1.
The last claim follows from $\mathrm{(ii)}$ of the above Proposition 3.2.
∎
4. Proof of Theorem 1.4
In this section, we prove Theorem 1.4.
For that, we use a few terminologies about representations of the symmetric groups from [4]
which we recalled in $\lx@sectionsign 2$.
Let
$$\hat{\ \ }:\mathbb{Q}\mathfrak{S}_{2k}\rightarrow\mathbb{Q}\mathfrak{S}_{2k}$$
denote the anti-involution induced by the anti-automorphism
$\mathfrak{S}_{2k}\ni\gamma\mapsto\gamma^{-1}\in\mathfrak{S}_{2k}$.
Lemma 4.1.
$\mathrm{(i)\ }$ For any Young diagram $\lambda$, the following equalities hold
$$\hat{a}_{\lambda}=a_{\lambda},\quad\hat{b}_{\lambda}=b_{\lambda},\quad\hat{c}_%
{\lambda}=c^{\prime}_{\lambda},\quad\hat{c^{\prime}}_{\lambda}=c_{\lambda}.$$
$\mathrm{(ii)\ }$ For any element $x\in\mathbb{Q}\mathfrak{S}_{2k}$ and $\xi,\eta\in\mathbb{Q}\mathcal{D}^{\ell}(2k)$, the following equality holds
$$\langle\xi,x\eta\rangle=\langle\hat{x}\,\xi,\eta\rangle.$$
Proof.
$\mathrm{(i)}$ follows from the definitions and $\mathrm{(ii)}$
follows from the fact that the pairing $\langle\ ,\ \rangle$ is $\mathfrak{S}_{2k}$-equivariant.
∎
Now, for each Young diagram $\lambda$ with $k$ boxes, we have a map
$$\left(\mathbb{Q}[\mathfrak{S}_{2k}]\,c^{\prime}_{2\lambda}\right)\otimes%
\mathcal{D}^{\ell}(2k)\rightarrow\mathbb{Q}\mathcal{D}^{\ell}(2k).$$
By Proposition 2.5 and the property of the Young symmetrizer,
the image of this map is precisely the subspace $E_{\lambda}\subset\mathbb{Q}\mathcal{D}^{\ell}(2k)$
which is the unique summand isomorphic to $(2\lambda)_{\mathfrak{S}_{2k}}$.
Proposition 4.2.
Two subspaces $E_{\lambda}$ and $E_{\mu}$ of $\mathbb{Q}\mathcal{D}^{\ell}(2k)$ are mutually orthogonal
to each other whenever $\lambda\not=\mu$. Namely we have
$$\langle E_{\lambda},E_{\mu}\rangle=\{0\}.$$
Proof.
It suffices to show that, for any $C,D\in\mathcal{D}^{\ell}(2k)$
$$\langle c_{2\lambda}C,c^{\prime}_{2\mu}D\rangle=0$$
because the elements of the form $c_{2\lambda}C$ generate the subspace $E_{\lambda}$ for any $\lambda$
and the same is true if we replace $c_{2\lambda}C$ by $c^{\prime}_{2\lambda}C$.
On the other hand, by Lemma 4.1 we have
$$\displaystyle\langle c_{2\lambda}C,c^{\prime}_{2\mu}D\rangle$$
$$\displaystyle=\langle\hat{c^{\prime}}_{2\mu}c_{2\lambda}C,D\rangle$$
$$\displaystyle=\langle{c}_{2\mu}c_{2\lambda}C,D\rangle$$
$$\displaystyle=0.$$
The last equality follows because, as is well known, ${c}_{\mu}c_{\lambda}=0$
whenever $\mu\not=\lambda$.
This completes the proof.
∎
Proposition 4.3.
The subspace $E_{\lambda}\subset\mathbb{Q}\mathcal{D}^{\ell}(2k)$ is an eigenspace for the linear mapping
$\mathcal{M}$ for any $\lambda$.
Proof.
First we prove that $E_{\lambda}$ is an $\mathcal{M}$-invariant subspace,
namely $\mathcal{M}(E_{\lambda})\subset E_{\lambda}$. For this, it suffices to show that
$\mathcal{M}(E_{\lambda})$ is orthogonal to $E_{\mu}$ for any $\mu\not=\lambda$.
This is because we already know that the whole space
$\mathbb{Q}\mathcal{D}^{\ell}(2k)$ can be written as
$$\mathbb{Q}\mathcal{D}^{\ell}(2k)=\bigoplus_{|\lambda|=k}E_{\lambda}\quad\text{%
(orthogonal direct sum)}$$
(see Proposition 2.5 and Proposition 4.2).
Recall that $\mathcal{M}(E_{\lambda})$ is generated by the elements of the form
$\sum_{D}\langle c_{2\lambda}C_{1},D\rangle D$
while $E_{\mu}$ is generated by the elements of the form
$c^{\prime}_{2\mu}\,C_{2}$
where $C_{1},C_{2}\in\mathcal{D}^{\ell}(2k)$.
Now we fix $C_{1},C_{2}\in\mathcal{D}^{\ell}(2k)$ and compute
$$\displaystyle\left\langle\sum_{D}\langle c_{2\lambda}C_{1},D\rangle D,c^{%
\prime}_{2\mu}\,C_{2}\right\rangle$$
$$\displaystyle=\sum_{D}\langle c_{2\lambda}C_{1},D\rangle\langle D,c^{\prime}_{%
2\mu}\,C_{2}\rangle$$
$$\displaystyle=\sum_{D}\langle c_{2\lambda}C_{1},D\rangle\langle c_{2\mu}D,C_{2}\rangle$$
$$\displaystyle=\sum_{D}\langle c_{2\lambda}C_{1},c^{\prime}_{2\mu}D\rangle%
\langle D,C_{2}\rangle$$
$$\displaystyle=\sum_{D}\langle c_{2\mu}c_{2\lambda}C_{1},D\rangle\langle D,C_{2}\rangle$$
$$\displaystyle=0$$
because $c_{2\mu}c_{2\lambda}=0$. Therefore $\mathcal{M}(E_{\lambda})\bot E_{\mu}$ as claimed.
Next observe that $\mathcal{M}$ is an $\mathfrak{S}_{2k}$-equivariant linear mapping because,
for any $\gamma\in\mathfrak{S}_{2k}$ and $C\in\mathcal{D}^{\ell}(2k)$, we have
$$\displaystyle\mathcal{M}(\gamma\,C)$$
$$\displaystyle=\sum_{D\in\mathcal{D}^{\ell}(2k)}\langle\gamma\,C,D\rangle D$$
$$\displaystyle=\sum_{D\in\mathcal{D}^{\ell}(2k)}\langle C,\gamma^{-1}\,D\rangle
D$$
$$\displaystyle=\sum_{D\in\mathcal{D}^{\ell}(2k)}\langle C,D\rangle\gamma\,D$$
$$\displaystyle=\gamma(\mathcal{M}(C)).$$
It follows that any eigenspace of $\mathcal{M}$ is an $\mathfrak{S}_{2k}$-submodule.
Thus $E_{\lambda}$ is an $\mathcal{M}$-invariant and
$\mathfrak{S}_{2k}$-irreducible subspace of $\mathbb{Q}\mathcal{D}^{\ell}(2k)$.
We can now conclude that it is an eigenspace as claimed.
∎
Remark 4.4.
It follows from Theorem 1.7 that, besides our inner product, the
subspaces $E_{\lambda}\subset\mathbb{Q}\mathcal{D}^{\ell}(2k)$ are mutually orthogonal to each other also with respect to
the Euclidean metric on $\mathbb{Q}\mathcal{D}^{\ell}(2k)$ induced by taking $\mathcal{D}^{\ell}(2k)$ as
an orthonormal basis. Indeed, this fact can be proved
by a standard argument in the theory of representations of symmetric groups
without mentioning our inner product on $\mathbb{Q}\mathcal{D}^{\ell}(2k)$.
The above argument is an adaptation of this to the case of our symmetric pairing
$\langle\ ,\ \rangle$ which depends on $g$. It follows that the answer should be given by rescaling
eigenspaces by certain factors. Thus, in addition to the positivity, the essential point of our work is the
explicit determination of these factors as below.
Proof of Theorem 1.4.
In view of Proposition 2.5 and the arguments following it, the proof for the latter part
of this theorem is finished so that it remains to prove the former
part.
We know already that each space $E_{\lambda}$ is an eigenspace (see Proposition 4.3).
Therefore for any choice of a tableau on $2\lambda$ and for any $C\in\mathcal{D}^{\ell}(2k)$,
we should have an equality
$$\mathcal{M}(c^{\prime}_{2\lambda}C)=\sum_{D\in\mathcal{D}^{\ell}(2k)}\langle c%
^{\prime}_{2\lambda}C,D\rangle\,D=\mu_{\lambda}c^{\prime}_{2\lambda}C$$
where $\mu_{\lambda}$ denotes the eigenvalue of $E_{\lambda}$. It follows that, for any $C,D$, we have
$$\langle c^{\prime}_{2\lambda}C,D\rangle=\text{$\mu_{\lambda}\cdot$ (%
coefficient of D in $c^{\prime}_{2\lambda}C$)}.$$
Our task is to determine $\mu_{\lambda}$ by the above equality and for that we may choose
$C,D$ arbitrarily. We choose $C=D=C_{0}(2k)$ where $C_{0}(2k)$ denotes the following linear chord diagram
$$C_{0}(2k)=\{\{1,2\},\cdots,\{2k-1,2k\}\}\in\mathcal{D}^{\ell}(2k).$$
We set
$$v_{2\lambda}(2k)=c^{\prime}_{2\lambda}C_{0}(2k),\quad m_{\lambda}=\text{%
coefficient of $C_{0}(2k)$ in $v_{2\lambda}(2k)$}$$
and we determine $\mu_{\lambda}$ from the identity
$$\langle v_{2\lambda}(2k),C_{0}(2k)\rangle=m_{\lambda}\mu_{\lambda}.$$
The rest of the proof is divided into three cases.
$\mathrm{(I)}\ $ The case of $\lambda=[k]$.
In this case, the group $P_{2\lambda}$ is the whole group while $Q_{2\lambda}$ is the trivial
group. Hence, for any tableau $\tau$, we have
$$v_{[2k]}=c_{[2k]}\,C_{0}(2k)=2^{k}k!\sum_{C\in\mathcal{D}^{\ell}(2k)}C$$
so that $m_{[k]}=2^{k}k!$. Furthermore it was already proved in [28] (proof of
Proposition 4.1) that
$$\left\langle\sum_{C\in\mathcal{D}^{\ell}(2k)}C,C_{0}(2k)\right\rangle=2g(2g-2)%
\cdots(2g-2k+2).$$
Therefore
$$\mu_{[k]}=2g(2g-2)\cdots(2g-2k+2)$$
as claimed.
$\mathrm{(II)}\ $ The case of $\lambda=[1^{k}]$.
The case $k=1$ was already treated in $\mathrm{(I)}\ $ so that we assume $k\geq 2$ and compute
the value of
$$\langle c^{\prime}_{[2^{k}]}C_{0},C_{0}\rangle$$
step by step as follows. We have
$$c^{\prime}_{[2^{k}]}=b_{[2^{k}]}a_{[2^{k}]}$$
and we choose a standard tableau $\tau$ on $[2^{k}]$ as below
$$\begin{bmatrix}1&2\\
3&4\\
\cdots&\cdots\\
2k-1&2k\end{bmatrix}.$$
Then the element $b_{[2^{k}]}$ can be described as
$$b_{[2^{k}]}=t_{k}t_{k-1}\cdots t_{3}t_{2}$$
where
$$\displaystyle t_{2}$$
$$\displaystyle=\{1-(13)\}\{1-(24)\},\ t_{3}=\{1-(15)-(35)\}\{1-(26)-(46)\},\cdots,$$
$$\displaystyle t_{k}$$
$$\displaystyle=\{1-(1,2k-1)-(3,2k-1)-\cdots-(2k-3,2k-1)\}\cdot$$
$$\displaystyle \{1-(2,2k)-(4,2k)-\cdots-(2k-2,2k)\}.$$
We define
$$c_{1}=a_{[2^{k}]},\quad c_{s}=t_{s}\cdots t_{2}a_{[2^{k}]}\quad(s=2,\cdots,k)$$
and compute
$$\langle c_{s}C_{0},C_{0}\rangle,\quad m_{s}=\text{coefficient of $C_{0}$ in $c%
_{s}C_{0}$}$$
inductively. Here the essential point in our computation is Lemma 2.4.
In fact, we use this lemma repeatedly in the whole of later argument and
we find a “perfect canceling” in sgns of various operations on linear chord diagrams.
In the first case $s=1$,
the group $P_{[2^{k}]}\cong\mathfrak{S}_{2}^{k}$ acts on $C_{0}$ trivially. Therefore
$$c_{1}C_{0}=a_{[2^{k}]}C_{0}=2^{k}C_{0},\quad m_{1}=2^{k}$$
and we have
$$\langle c_{1}C_{0},C_{0}\rangle=m_{1}(2g)^{k}.$$
Next we compute the case $k=2$
$$\langle c_{2}C_{0},C_{0}\rangle=2^{k}\langle t_{1}C_{0},C_{0}\rangle,\quad%
\text{$m_{2}$= coefficient of $C_{0}$ in $c_{2}C_{0}$}.$$
Among the $4$ elements in
$$t_{2}=\{1-(13)\}\{1-(24)\}=1+(13)(24)-(13)-(24)$$
exactly $2$ elements, namely $1,(13)(24)$ preserve $C_{0}$ unchanged.
Therefore $m_{2}=2m_{1}$ and if we apply Lemma 2.4, we can conclude that
$$\langle t_{2}C_{0},C_{0}\rangle=(2g)\{2\cdot 2g-2\cdot(-1)\}(2g)^{k-2}=(2g)2(2%
g+1)(2g)^{k-2}.$$
Therefore we have
$$\langle t_{2}C_{0},C_{0}\rangle=m_{2}(2g)(2g+1)(2g)^{k-2}.$$
Next we compute
$$\langle c_{3}C_{0},C_{0}\rangle=2^{k}\langle t_{3}t_{2}C_{0},C_{0}\rangle,%
\quad\text{$m_{3}$= coefficient of $C_{0}$ in $c_{3}C_{0}$}.$$
Among the $9$ elements in
$$t_{3}=\{1-(15)-(35)\}\{1-(26)-(46)\}$$
exactly $3$ elements, namely $1,(15)(26),(35)(46)$ preserve $C_{0}$ unchanged.
Therefore $m_{3}=3m_{2}$ and if we apply Lemma 2.4, we can conclude that
$$\langle t_{3}t_{2}C_{0},C_{0}\rangle=(2g)2(2g+1)\{3\cdot 2g-(9-3)\cdot(-1)\}(2%
g)^{k-3}=(2g)2(2g+1)3(2g+2)(2g)^{k-3}.$$
Therefore we have
$$\langle t_{3}t_{2}C_{0},C_{0}\rangle=m_{3}(2g)(2g+1)(2g+2)(2g)^{k-3}.$$
Continuing the computation like this, we can prove that
the equalities
$$\displaystyle\langle t_{s}\cdots t_{1}C_{0},C_{0}\rangle$$
$$\displaystyle=m_{s}(2g)(2g+1)(2g+2)\cdots(2g+s)(2g)^{k-s-1}$$
$$\displaystyle m_{s}$$
$$\displaystyle=2^{k}s!$$
hold for all $s=1,\cdots,k$.
We check the inductive argument from the $s$-th step to the next $(s+1)$-th step.
Among the $(s+1)^{2}$ elements in the expansion of
$$\displaystyle\{1-(1,2s+1)-(3,2s+1)-\cdots-(2s-1,2s+1)\}$$
$$\displaystyle\{1-(2,2s+2)-(4,2s+2)-\cdots-(2s,2s+2)\}$$
exactly the following $(s+1)$ elements preserve $C_{0}$ unchanged, namely
$$1,(1,2s+1)(2,2s+2),\cdots,(2s-1,2s+1)(2s,2s+2).$$
Therefore $m_{s+1}=(s+1)m_{s}$ and if we apply Lemma 2.4, we can conclude that
$$\displaystyle\langle t_{s+1}t_{s}\cdots t_{2}C_{0},C_{0}\rangle$$
$$\displaystyle=m_{s}(2g)(2g+1)\cdots(2g+s-1)\{(s+1)\cdot 2g-((s+1)^{2}-(s+1))%
\cdot(-1)\}(2g)^{k-s}$$
$$\displaystyle=(s+1)m_{s}2g(2g+1)\cdots(2g+s)(2g)^{k-s}.$$
Therefore we have
$$\langle t_{s+1}\cdots t_{2}C_{0},C_{0}\rangle=m_{s+1}(2g)(2g+1)\cdots(2g+s)(2g%
)^{k-s-1}$$
as required.
$\mathrm{(III)}\ $ The general case.
Finally we consider the general case. Roughly speaking, we combine the above two arguments, one is
the case $\mathrm{(I)}$ and the other is the case $\mathrm{(II)}$, to treat the general case.
Let $\lambda=[\lambda_{1}\lambda_{2}\cdots\lambda_{h}]$ be a general Young diagram with $|\lambda|=k$.
Choose a standard tableau on $2\lambda$, say $1,2,\cdots,2\lambda_{1}-1,2\lambda_{1}$ on the
first row, $2\lambda_{1}+1,2\lambda_{1}+2,\cdots,2\lambda_{1}+2\lambda_{2}-1,2\lambda_{1}+%
2\lambda_{2}$ on the
second row, and so on. Then we can write
$$P_{2\lambda}=P_{2\lambda_{1}}\times P_{2\lambda_{2}}\times\cdots\times P_{2%
\lambda_{h}}$$
where $P_{2\lambda_{i}}\ (i=1,\cdots,h)$ denotes the symmetric group consisting of permutations of
labels attached to the boxes in the $i$-th row of $2\lambda$.
Thus we have
$$a_{2\lambda}=a_{2\lambda_{1}}a_{2\lambda_{2}}\cdots a_{2\lambda_{h}}.$$
If we apply the result of case $\mathrm{(I)}\ $ to each row of $2\lambda$, then we obtain
$$\displaystyle\langle a_{2\lambda}C_{0},C_{0}\rangle=$$
$$\displaystyle 2^{\lambda_{1}}\lambda_{1}!\cdot 2g(2g-2)\cdots(2g-2\lambda_{1}+%
2)\cdot$$
$$\displaystyle 2^{\lambda_{2}}\lambda_{2}!\cdot 2g(2g-2)\cdots(2g-2\lambda_{2}+%
2)\cdot$$
$$\displaystyle\cdots$$
$$\displaystyle 2^{\lambda_{h}}\lambda_{h}!\cdot 2g(2g-2)\cdots(2g-2\lambda_{h}+2)$$
and
$$\text{coefficient of $C_{0}$ in $a_{2\lambda}C_{0}\ =\prod_{i=1}^{h}2^{\lambda%
_{i}}\lambda_{i}!=2^{k}\prod_{i=1}^{h}\lambda_{i}!$}.$$
Next we consider the element $b_{2\lambda}$. The group $Q_{2\lambda}$ can be described as
$$Q_{2\lambda}=Q_{2\lambda_{1}}\times Q_{2\lambda_{1}-1}\times\cdots\times Q_{2}%
\times Q_{1}$$
where $Q_{j}\ (j=1,2,\cdots,2\lambda_{1})$ denotes the symmetric group consisting of permutations of the
labels attached to the $j$-th column of $2\lambda$.
Here we mention that $2\lambda_{1}$ is the number of columns of $2\lambda$ and
also both of $Q_{1}$ and $Q_{2}$ are isomorphic to $\mathfrak{S}_{h}$.
Now we decompose the element $b_{2\lambda}$ as
$$b_{2\lambda}=b_{\lambda_{1}}\cdots b_{2}b_{1}$$
where
$$b_{j}=\prod_{\gamma\in Q_{2j-1}}(1+\mathrm{sgn}\,\gamma)\prod_{\delta\in Q_{2j%
}}(1+\mathrm{sgn}\,\delta).$$
First we consider the element $b_{1}a_{2\lambda}C_{0}$.
It can be seen that
we can apply the result of case $\mathrm{(II)}\ $ to the first two columns of $2\lambda$ from the left to obtain
$$\displaystyle\langle b_{1}a_{2\lambda}C_{0},C_{0}\rangle=$$
$$\displaystyle\,h!\,2^{\lambda_{1}}\lambda_{1}!\cdot 2g(2g-2)\cdots(2g-2\lambda%
_{1}+2)\cdot$$
$$\displaystyle 2^{\lambda_{2}}\lambda_{2}!\cdot[(2g+1)(2g-1)]\cdot(2g-4)\cdots(%
2g-2\lambda_{2}+2)\cdot$$
$$\displaystyle\cdots$$
$$\displaystyle 2^{\lambda_{h}}\lambda_{h}!\cdot[(2g+h-1)(2g+h-3)]\cdot(2g-4)%
\cdots(2g-2\lambda_{h}+2)$$
$$\displaystyle=$$
$$\displaystyle\,h!\,2^{k}\prod_{i=1}^{h}\lambda_{i}!\cdot 2g(2g-2)\cdots(2g-2%
\lambda_{1}+2)\cdot$$
$$\displaystyle[(2g+1)(2g-1)]\cdot(2g-4)\cdots(2g-2\lambda_{2}+2)\cdot$$
$$\displaystyle\cdots$$
$$\displaystyle[(2g+h-1)(2g+h-3)]\cdot(2g-4)\cdots(2g-2\lambda_{h}+2)$$
$$\displaystyle=$$
$$\displaystyle\,h!\,2^{k}\prod_{i=1}^{h}\lambda_{i}!\cdot[2g(2g+1)\cdots(2g+h-1%
)]\cdot[(2g-2)(2g-1)\cdots(2g+h-3)]\cdot$$
$$\displaystyle[(2g-4)(2g-6)]^{\lambda^{\prime}_{2}}\cdots[(2g-2\lambda_{1}+4)(2%
g-2\lambda_{1}+2)]^{\lambda^{\prime}_{\lambda_{1}}}$$
and
coefficient of $$C_{0}$$ in $$b_{1}a_{2\lambda}C_{0}\ =h!\,2^{k}\prod_{i=1}^{h}\lambda_{i}!$$
where $\lambda^{\prime}=[h\lambda^{\prime}_{2}\cdots\lambda^{\prime}_{\lambda_{1}}]$ denotes the conjugate Young diagram of $\lambda$.
Next we can apply the result of case $\mathrm{(II)}\ $ to the two columns,
the third and the fourth from the left of $2\lambda$, to obtain
$$\displaystyle\langle b_{2}b_{1}a_{2\lambda}C_{0},C_{0}\rangle=$$
$$\displaystyle h!\lambda^{\prime}_{2}!\,2^{k}\prod_{i=1}^{h}\lambda_{i}!\cdot 2%
g(2g-2)\cdots(2g-2\lambda_{1}+2)\cdot$$
$$\displaystyle(2g+1)(2g-1)\cdot[(2g-3)(2g-5)]\cdot(2g-8)\cdots(2g-2\lambda_{2}+%
2)\cdot$$
$$\displaystyle\cdots$$
$$\displaystyle(2g+h-1)(2g+h-3)\cdot[(2g+h-5)(2g+h-7)]\cdot(2g-8)\cdots(2g-2%
\lambda_{h}+2)$$
$$\displaystyle=$$
$$\displaystyle\,h!\lambda^{\prime}_{2}!\,2^{k}\prod_{i=1}^{h}\lambda_{i}!\cdot[%
2g(2g+1)\cdots(2g+h-1)]\cdot[(2g-2)(2g-1)\cdots(2g+h-3)]\cdot$$
$$\displaystyle\cdot[(2g-4)(2g-3)\cdots(2g+\lambda^{\prime}_{2}-1)]\cdot[(2g-6)(%
2g-5)\cdots(2g+\lambda^{\prime}_{2}-3)]\cdot$$
$$\displaystyle[(2g-8)(2g-10)]^{\lambda^{\prime}_{3}}\cdots[(2g-2\lambda_{1}+4)(%
2g-2\lambda_{1}+2)]^{\lambda^{\prime}_{\lambda_{1}}}$$
and
$$\text{coefficient of $C_{0}$ in $b_{2}b_{1}a_{2\lambda}C_{0}\ =h!\lambda^{%
\prime}_{2}!\,2^{k}\prod_{i=1}^{h}\lambda_{i}!$}.$$
Continuing in this way until the last two columns from the right of $\lambda$,
we finally obtain
$$\displaystyle\langle c^{\prime}_{2\lambda}C_{0},C_{0}\rangle=$$
$$\displaystyle m_{2\lambda}\cdot 2g(2g-2)\cdots(2g-2\lambda_{1}+2)\cdot$$
$$\displaystyle(2g+1)(2g-1)\cdots(2g-2\lambda_{2}+3)\cdot$$
$$\displaystyle\cdots$$
$$\displaystyle(2g+h-1)(2g+h-3)\cdots(2g+h-2\lambda_{h}+1)$$
$$\displaystyle m_{2\lambda}=$$
$$\displaystyle\prod_{i=1}^{h}2^{\lambda_{i}}\lambda_{i}!\prod_{j=1}^{\lambda_{1%
}}\lambda^{\prime}_{j}!=2^{k}\prod_{i=1}^{h}\lambda_{i}!\prod_{j=1}^{\lambda_{%
1}}\lambda^{\prime}_{j}!.$$
Here we would like to emphasize again that an essential use of Lemma 2.4 is made in the above computation.
This completes the proof.
∎
5. Topological study of the tautological algebra of the moduli space of curves
As already mentioned in the introduction, this work was originally motivated by
a topological study of the tautological algebra of the moduli space of curves.
Let $\mathcal{M}_{g}$ be the mapping class group of $\Sigma_{g}$
and let $\mathcal{R}^{*}(\mathcal{M}_{g})$ denote the tautological algebra in cohomology of
$\mathcal{M}_{g}$ . Namely it is the subalgebra of $H^{*}(\mathcal{M}_{g};\mathbb{Q})$
generated by the Mumford-Morita-Miller tautological classes $e_{i}\in H^{2i}(\mathcal{M}_{g};\mathbb{Q})$.
We also consider the mapping class group $\mathcal{M}_{g,*}$ of
$\Sigma_{g}$ relative to the base point $*\in\Sigma_{g}$ and the corresponding
tautological algebra $\mathcal{R}^{*}(\mathcal{M}_{g,*})$. In this case,
it is the subalgebra of $H^{*}(\mathcal{M}_{g,*};\mathbb{Q})$ generated by
$e_{i}$ and the Euler class $e\in H^{2}(\mathcal{M}_{g,*};\mathbb{Q})$.
In the context of algebraic geometry, there is a more refined form of the tautological algebra.
Namely, if we denote by $\mathbf{M}_{g}$ the moduli space of curves of genus $g$, then
its tautological algebra, denoted by $\mathcal{R}^{*}(\mathbf{M}_{g})$, is defined to be the
subalgebra of the Chow algebra $\mathcal{A}^{*}(\mathbf{M}_{g})$ of $\mathbf{M}_{g}$ generated
by the tautological classes $\kappa_{i}\in\mathcal{A}^{i}(\mathbf{M}_{g})$ introduced by Mumford.
There is a natural surjection $\mathcal{R}^{*}(\mathbf{M}_{g})\rightarrow\mathcal{R}^{*}(\mathcal{M}_{g})$
which sends $\kappa_{i}$ to $(-1)^{i+1}e_{i}$.
There have been obtained many interesting results about the tautological algebras,
Faber’s beautiful conjecture given in [2] together with his related works
being a strong motivation for them.
In particular, Faber and Zagier proposed a set of conjectural relations in
$\mathcal{R}^{*}(\mathbf{M}_{g})$ and Faber in fact proved that it
gives an affirmative answer to his conjecture including the
most difficult part, namely the perfect pairing property, up to genus $g\leq 23$.
Very recently, Pandharipande and Pixton [31] proved a remarkable
result that the Faber-Zagier relations are actual relations in $\mathcal{R}^{*}(\mathbf{M}_{g})$ for all $g$.
However the structure of $\mathcal{R}^{*}(\mathbf{M}_{g})$ for general $g$ remains mysterious.
For more details of these developments, we refer to Faber’s recent survey paper
[3] as well as the references therein.
The main tools in these works belong to algebraic geometry which are quite powerful.
A theorem of Looijenga given in [18], which is one of the earliest
results in this area, is a typical example.
On the other hand, there are a few topological approaches to this subject.
The first one was given in our papers [23][24].
Recently Randal-Williams [33] greatly generalized this
to obtain very interesting results. See also
Yin [35] for the case of $\mathcal{R}^{*}(\mathbf{M}_{g,1})$
and Grigoriev [7] for the case of higher dimensional manifold bundles.
In [28] we adopted a closely related but a bit different
approach, namely a combination of topology and representation theory.
In the next section we extend this approach and
give a systematic way of producing relations among
the $\mathrm{MMM}$ tautological classes in $\mathcal{R}^{*}(\mathcal{M}_{g})$ and $\mathcal{R}^{*}(\mathcal{M}_{g,*})$.
For that, in the remaining part of this section, we summarize former results.
By analyzing the natural action of $\mathcal{M}_{g}$ on the third nilpotent quotient of
$\pi_{1}\Sigma_{g}$,
in [27] we constructed the following commutative diagram
(8)
$$\begin{CD}\pi_{1}\Sigma_{g}@>{}>{}>[1^{2}]_{\mathrm{Sp}}\tilde{\times}H_{%
\mathbb{Q}}\\
@V{}V{}V@V{}V{}V\\
\mathcal{M}_{g,*}@>{\rho_{2}}>{}>(([1^{2}]^{\text{torelli}}_{\mathrm{Sp}}%
\oplus[2^{2}]_{\mathrm{Sp}})\tilde{\times}\wedge^{3}H_{\mathbb{Q}})\rtimes%
\mathrm{Sp}(2g,\mathbb{Q})\\
@V{}V{p}V@V{}V{}V\\
\mathcal{M}_{g}@>{\bar{\rho}_{2}}>{}>([2^{2}]_{\mathrm{Sp}}\tilde{\times}U_{%
\mathbb{Q}})\rtimes\mathrm{Sp}(2g,\mathbb{Q}).\end{CD}$$
This extends our earlier work in [25] in which we treated the case of the action on the second nilpotent quotient.
Here $U_{\mathbb{Q}}=\wedge^{3}H_{\mathbb{Q}}/H_{\mathbb{Q}}$ and $[2^{2}]_{\mathrm{Sp}}$ is the unique $\mathrm{Sp}$-representation
corresponding to the Young diagram $[2^{2}]$ which appears in both of $\wedge^{2}U^{*}_{\mathbb{Q}}$
and $\wedge^{2}(\wedge^{3}H^{*}_{\mathbb{Q}})$.
Also $([2^{2}]_{\mathrm{Sp}})$ denotes the ideal generated by $[2^{2}]_{\mathrm{Sp}}$.
$(([1^{2}]^{\text{torelli}}_{\mathrm{Sp}}\oplus[2^{2}]_{\mathrm{Sp}})$ is defined similarly
(see [27] for details).
In our joint works with Kawazumi [12][13], we determined the homomorphisms
$\rho^{*}_{2},\bar{\rho}^{*}_{2}$ in cohomology which are induced by the representations
$\rho_{2},\bar{\rho}_{2}$ in the above diagram (8). More precisely, we proved that the images of them
are precisely the tautological algebras and also that these representations are isomorphisms in the
stable range of Harer [8] (so that the images are in fact isomorphic to
the whole rational cohomology algebras of the mapping class groups
by the definitive result of Madsen and Weiss [19]).
In the proof of these results, the work of Kawazumi in [11]
played a crucial role.
Thus we have the following commutative diagram
(9)
$$\begin{CD}\left(\wedge^{*}(\wedge^{3}H^{*}_{\mathbb{Q}})/([1^{2}]^{\text{%
torelli}}_{\mathrm{Sp}}\oplus[2^{2}]_{\mathrm{Sp}})\right)^{\mathrm{Sp}}@>{%
\rho_{2}^{*}}>{\text{stably $\cong$}}>\mathcal{R}^{*}(\mathcal{M}_{g,*})\\
@A{}A{}A@A{}A{p^{*}}A\\
\left(\wedge^{*}U^{*}_{\mathbb{Q}}/([2^{2}]_{\mathrm{Sp}})\right)^{\mathrm{Sp}%
}@>{\bar{\rho}_{2}^{*}}>{\text{stably $\cong$}}>\mathcal{R}^{*}(\mathcal{M}_{g%
}).\end{CD}$$
Here $([2^{2}]_{\mathrm{Sp}})$ denotes the ideal generated by $[2^{2}]_{\mathrm{Sp}}$.
$(([1^{2}]^{\text{torelli}}_{\mathrm{Sp}}\oplus[2^{2}]_{\mathrm{Sp}})$ is defined similarly.
Furthermore we have given explicit formulas for the homomorphisms $\rho_{2}^{*},\bar{\rho}_{2}^{*}$
in terms of graphs which we now recall.
Let $\mathcal{G}$ be the set of isomorphism classes of connected trivalent graphs
and let $\mathbb{Q}[\Gamma;\Gamma\in\mathcal{G}]$ denote the polynomial algebra generated by
$\mathcal{G}$. We define the degree of a graph $\Gamma\in\mathcal{G}$ to be the number
of vertices of it. Thus the subspace $\mathbb{Q}[\Gamma;\Gamma\in\mathcal{G}]^{(2k)}$ consisting
of homogeneous polynomials of degree $2k$ can be naturally identified with the linear space
spanned by trivalent graphs with $2k$ vertices. Also let $\mathcal{G}^{0}$
be the subset of $\mathcal{G}$ consisting of connected trivalent graphs
without loops.
In [25], the following commutative diagram
was constructed
(10)
$$\begin{CD}\mathbb{Q}[\Gamma;\Gamma\in\mathcal{G}]@>{}>{}>(\wedge^{*}(\wedge^{3%
}H^{*}_{\mathbb{Q}}))^{\mathrm{Sp}}\\
@A{\iota_{g}}A{}A@A{}A{}A\\
\mathbb{Q}[\Gamma;\Gamma\in\mathcal{G}^{0}]@>{}>{}>(\wedge^{*}U^{*}_{\mathbb{Q%
}})^{\mathrm{Sp}}\end{CD}$$
where both of the horizontal homomorphisms are surjective
and bijective in a certain stable range and $\iota_{g}$ denotes some
twisted inclusion.
Then in [13], the following commutative diagram
$$\begin{CD}\mathbb{Q}[\Gamma;\Gamma\in\mathcal{G}]@>{\alpha}>{}>\mathbb{Q}[g][e%
,e_{1},e_{2},\cdots]\\
@A{\iota_{g}}A{}A@A{}A{\cup}A\\
\mathbb{Q}[\Gamma;\Gamma\in\mathcal{G}^{0}]@>{\beta}>{}>\mathbb{Q}[g][e_{1},e_%
{2},\cdots]\end{CD}$$
was constructed which realizes the composition of the commutative diagrams (10) followed by (9)
by means of explicit formulas $\alpha$ and $\beta$. In particular, for each trivalent graph
$\Gamma$ with $2k$ vertices, we have the associated characteristic class
$$\alpha_{\Gamma}\in\mathbb{Q}[g][e,e_{1},e_{2},\ldots]^{(2k)}.$$
If we fix a genus $g$, then
the commutative diagram (9) is isomorphic to the following
$$\begin{CD}\mathbb{Q}[\Gamma;\Gamma\in\mathcal{G}]/\widetilde{W}@>{\alpha_{g}}>%
{\text{stably $\cong$}}>\mathbb{Q}[e,e_{1},e_{2},\cdots]\\
@A{\iota_{g}}A{}A@A{}A{\cup}A\\
\mathbb{Q}[\Gamma;\Gamma\in\mathcal{G}^{0}]/W@>{\beta_{g}}>{\text{stably $%
\cong$}}>\mathbb{Q}[e_{1},e_{2},\cdots]\end{CD}$$
where $\widetilde{W}$ denotes certain equivalence relation in $\mathbb{Q}[\Gamma;\Gamma\in\mathcal{G}]$
induced by graphical operations on trivalent graphs including the Whitehead move
(or the $\mathrm{IH}$ move) and $W$ is a similar one.
We mention that the relation between the ideal $([2^{2}]_{\mathrm{Sp}})$ and
the $\mathrm{IH}$ move for the lower case
was first studied by Garoufalidis and Nakamura in [5].
6. A set of relations in $\mathcal{R}^{*}(\mathcal{M}_{g})$ and $\mathcal{R}^{*}(\mathcal{M}_{g,*})$
In this section, we apply Theorem 1.1 to the description of the tautological algebras
$\mathcal{R}^{*}(\mathcal{M}_{g})$ and $\mathcal{R}^{*}(\mathcal{M}_{g,*})$ in terms of certain symplectic invariant tensors
as well as trivalent graphs,
which we recall in the previous section, and obtain relations among the $\mathrm{MMM}$ tautological classes
in a systematic way. This extends our former work in [28].
The symplectic invariant tensors we are concerned is $\left(\wedge^{2k}(\wedge^{3}H_{\mathbb{Q}})\right)^{\mathrm{Sp}}$
which can be considered as both a natural quotient space as well as the
subspace of $(H_{\mathbb{Q}}^{\otimes 6k})^{\mathrm{Sp}}$. Furthermore this subspace
can be described by means of the natural action of $\mathfrak{S}_{6k}$ on the
whole space $(H_{\mathbb{Q}}^{\otimes 6k})^{\mathrm{Sp}}$. It follows that the orthogonal
direct sum decomposition given by Theorem 1.1 induces that
of the space $\left(\wedge^{2k}(\wedge^{3}H_{\mathbb{Q}})\right)^{\mathrm{Sp}}$
so that we obtain the following commutative diagram
(11)
$$\begin{CD}(H_{\mathbb{Q}}^{\otimes 6k})^{\mathrm{Sp}}@>{\text{quotient, %
contain}}>{\supset}>\left(\wedge^{2k}(\wedge^{3}H_{\mathbb{Q}})\right)^{%
\mathrm{Sp}}\\
\Big{\|}\Big{\|}\\
\bigoplus_{|\lambda|=3k}U_{\lambda}@>{\text{quotient, contain}}>{\supset}>%
\bigoplus_{|\lambda|=3k}L_{\lambda}\end{CD}$$
where $L_{\lambda}$ denotes certain subspace of $U_{\lambda}$.
The totality of the upper quotient homomorphism
in this diagram for all $k$ can be realized graphically by the following collapsing operation.
Definition 6.1 ([28]).
For each $C\in\mathcal{D}^{\ell}(6k)$, let $\Gamma_{C}$ denote the trivalent graph defined as follows.
We understand $C$ as a disjoint union of $3k$ chords each of which connects
the two points $i_{s},j_{s}\ (s=1,2,\cdots,3k)$ in the set of vertices $\{1,2,\cdots,6k\}$.
Then $\Gamma_{C}$ is the graph obtained from $C$ by joining the three vertices
in each of the sets $\{1,2,3\},$ $\{4,5,6\},$ $\cdots,$ $\{6k-2,6k-1,6k\}$ to a single point.
Thus $\Gamma_{C}$ is a (possibly disconnected) trivalent graph with $2k$ vertices.
By extending this operation linearly, we obtain a linear mapping
$$\mathbb{Q}\mathcal{D}^{\ell}(6k)\rightarrow\mathbb{Q}[\Gamma;\Gamma\in\mathcal%
{G}]^{(2k)}.$$
For each element $\xi\in\mathbb{Q}\mathcal{D}^{\ell}(6k)$, we denote its image
under the above mapping by $\Gamma_{\xi}$.
Then we have the following commutative diagram
(12)
$$\begin{CD}\mathbb{Q}\mathcal{D}^{\ell}(6k)@>{\text{collapse}}>{\supset}>%
\mathbb{Q}[\Gamma;\Gamma\in\mathcal{G}]^{(2k)}\\
\Big{\|}\Big{\|}\\
\bigoplus_{|\lambda|=3k}E_{\lambda}@>{\text{collapse}}>{\supset}>\bigoplus_{|%
\lambda|=3k}L^{\prime}_{\lambda}\end{CD}$$
where $L^{\prime}_{\lambda}$ denotes the subspace corresponding to $L_{\lambda}$
(and in fact isomorphic to $L_{\lambda^{\prime}}$).
In view of (10), this is equivalent to the diagram (11).
We need one more terminology to formulate our result and that concerns the
following direct sum decomposition
(13)
$$(\wedge^{2k}(\wedge^{3}H_{\mathbb{Q}}))^{\mathrm{Sp}}\cong\bigoplus_{p=0}^{2k}%
(\wedge^{2k-p}U_{\mathbb{Q}}\otimes\wedge^{p}H_{\mathbb{Q}})^{\mathrm{Sp}}.$$
This is based on the fact that $\wedge^{3}H_{\mathbb{Q}}$ is not irreducible as an $\mathrm{Sp}$-module
but is a sum of two irreducible summands $\wedge^{3}H_{\mathbb{Q}}\cong U_{\mathbb{Q}}\oplus H_{\mathbb{Q}}$.
Notice that the above decomposition is an unstable operation in the sense that it depends on the
genus $g$ and also that it cannot be described in terms of the action of the symmetric group.
To realize the above decomposition (13) graphically,
we recall a few terminologies from our paper [28] in suitably modified forms.
Consider the mapping
(14)
$$\begin{split}\displaystyle\wedge^{2k}&\displaystyle(\wedge^{3}H_{\mathbb{Q}})%
\ni\xi=\xi_{1}\wedge\cdots\wedge\xi_{2k}\\
&\displaystyle\mapsto\xi^{(p)}=\sum_{i_{1}<\cdots<i_{p}}\xi_{1}\wedge\cdots%
\wedge\widetilde{C}\xi_{i_{1}}\wedge\cdots\wedge\widetilde{C}\xi_{i_{p}}\wedge%
\cdots\wedge\xi_{2k}\in\wedge^{2k}(\wedge^{3}H_{\mathbb{Q}})\end{split}$$
where $\xi_{i}\in\wedge^{3}H_{\mathbb{Q}}$, $0\leq p\leq 2k$ and $\widetilde{C}:\wedge^{3}H_{\mathbb{Q}}\rightarrow\wedge^{3}H_{\mathbb{Q}}$
is the mapping defined by
$$\widetilde{C}(u\wedge v\wedge w)=\frac{1}{g-1}\,(\mu(u,v)w+\mu(v,w)u+\mu(w,u)v%
)\wedge\omega_{0}.$$
Now let $\Gamma$ be a trivalent graph with $2k$ vertices and let $V_{\Gamma}$ denote the
set of vertices. Fix a subset $F\subset V_{\Gamma}$ consisting of $p$ vertices where $0\leq p\leq 2k$.
We denote by $\mathcal{P}(F)$ the set of all ways of choices $\varpi$ where, for each vertex
$v\in F$, $\varpi(v)$ specifies $2$ edges emerging from $v$ out of the $3$ such edges.
Since there are $3$ edges emerging from each vertex, $\mathcal{P}(F)$
has $3^{p}\binom{2k}{p}$ elements.
For each $\varpi\in\mathcal{P}(F)$, we define a trivalent graph $\Gamma_{\varpi}$ and a number
$\ell(\varpi)$ as follows. At each vertex $v\in F$, cut the $2$ edges emerging from it which correspond to $\varpi(v)$
at the one third point along each edge from $v$. Then we connect the two endpoints of the shortened edges emerging
from $v$ to make a loop. At the same time, we also connect the two endpoints of the remaining
$2$ edges. After performing this operation at every vertex $v\in F$, we are left with a trivalent graph
together with certain number of disjoint circles and/or intervals. We define $\Gamma_{\varpi}$ to be this
trivalent graph throwing away the other parts. The number $\ell(\varpi)$ is defined to be that of the
disjoint circles. Namely it is the sum of the number of loops and that of double edges “occurring”
in $\varpi$.
Definition 6.2 (graphical contraction).
Let $\Gamma$ be a trivalent graph with $2k$ vertices. For each $0\leq p\leq 2k$, we define
$$\Gamma^{(p)}=\frac{1}{(2g-2)^{p}}\sum_{F\subset V_{\Gamma}}\sum_{\varpi\in%
\mathcal{P}(F)}(-1)^{p}(-2g)^{\ell(\varpi)}{\Gamma}_{\varpi}\ \in\mathbb{Q}[%
\Gamma;\Gamma\in\mathcal{G}]^{(2k)}$$
where $F$ runs through all subsets of $V_{\Gamma}$ consisting of $p$ elements.
Furthermore we set
$$\bar{\Gamma}=\Gamma-\Gamma^{(1)}+\Gamma^{(2)}-\cdots+\Gamma^{(2k)}.$$
Finally we extend these operations on the whole space $\mathbb{Q}[\Gamma;\Gamma\in\mathcal{G}]^{(2k)}$ by
linearity.
In summary, each element $\xi\in\mathbb{Q}\mathcal{D}^{\ell}(6k)$ gives rise to the associated element
$\Gamma_{\xi}\in\mathbb{Q}[\Gamma;\Gamma\in\mathcal{G}]^{(2k)}$ by Definition 6.1
and then we have the following series of elements
$$\Gamma_{\xi}=\Gamma_{\xi}^{(0)},\Gamma_{\xi}^{(1)},\cdots,\Gamma_{\xi}^{(2k)},%
\bar{\Gamma}_{\xi}$$
belonging to $\mathbb{Q}[\Gamma;\Gamma\in\mathcal{G}]^{(2k)}$ by Definition 6.2.
Then it was shown in [28] that the above graphical operations realize the decomposition
(13).
Now we are ready to state our main result in this section.
Theorem 6.3.
$\mathrm{(i)\ }$ In the tautological algebra $\mathcal{R}^{*}(\mathcal{M}_{g})$, the following relations hold.
$$\text{For all $\xi\in E_{\lambda}\subset\mathbb{Q}\mathcal{D}^{\ell}(6k)$ with%
$\lambda_{1}>g$,
we have $\alpha_{\bar{\Gamma}_{\xi}}=0\in\mathcal{R}^{2k}(\mathcal{M}_{g})$}.$$
$\mathrm{(ii)\ }$ In the tautological algebra $\mathcal{R}^{*}(\mathcal{M}_{g,*})$, the following relations hold.
$$\text{For all $\xi\in E_{\lambda}\subset\mathbb{Q}\mathcal{D}^{\ell}(6k)$ with%
$\lambda_{1}>g$,
we have $\alpha_{\Gamma^{(p)}_{\xi}}=0\in\mathcal{R}^{2k}(\mathcal{M}_{g,*})$ %
for any $0\leq p\leq 2k$}.$$
Proof.
We first prove $\mathrm{(ii)}$.
Let $\xi\in E_{\lambda}\subset\mathbb{Q}\mathcal{D}^{\ell}(6k)$ be any element. Then we have $\Phi(\xi)\in U_{\lambda^{\prime}}$
by Theorem 1.7. If we further assume that $\lambda_{1}>g$, then $h(\lambda^{\prime})=\lambda_{1}>g$
so that $\Phi(\xi)=0\in(H_{\mathbb{Q}}^{\otimes 6k})^{\mathrm{Sp}}$ by Theorem 1.1.
It follows that the projected image of $\Phi(\xi)$ in $(\wedge^{2k}(\wedge^{3}H_{\mathbb{Q}}))^{\mathrm{Sp}}$
is also trivial. Furthermore all the components of this element with respect to the direct sum
decomposition (13) are trivial because they are obtained by applying the
operations (14). Since these tensorial operations are realized by the graphical
operations given in Definition 6.1 and Definition 6.2, the claim holds.
Next we prove $\mathrm{(i)}$. The graphical operation
$$\Gamma\mapsto\bar{\Gamma}=\Gamma-\Gamma^{(1)}+\Gamma^{(2)}-\cdots+\Gamma^{(2k)}$$
given in Definition 6.2 realizes the tensorial operation of taking the
$(\wedge^{2k}U_{\mathbb{Q}})^{\mathrm{Sp}}$ part out of the whole $(\wedge^{2k}(\wedge^{3}H_{\mathbb{Q}}))^{\mathrm{Sp}}$
in the direct sum decomposition (13).
On the other hand, the
diagram (9) shows that this operation
extracts exactly the
“basic” part with respect to the projection $\mathcal{M}_{g,*}\rightarrow\mathcal{M}_{g}$,
namely the part which comes from the ”base” $\mathcal{M}_{g}$
(observe here that we have a natural
injection $\mathcal{R}^{*}(\mathcal{M}_{g})\subset\mathcal{R}^{*}(\mathcal{M}_{g,*})$).
The claim follows.
∎
Remark 6.4.
The first case $\lambda=[3k]$ in the above theorem was treated in our former paper [28] completely.
In this case $2\lambda=[6k]$ is the one dimensional trivial representation of $\mathfrak{S}_{6k}$
so that $\mathrm{dim}\,E_{\lambda}=1$. We have determined the relations
$\alpha_{\bar{\Gamma}_{\xi}}=0\in\mathcal{R}^{2k}(\mathcal{M}_{g})$
explicitly for all $k$ and $g=3k-1,3k-2,\ldots,2$, where $\xi$ denotes a non-trivial
unique (up to scalars) element of $E_{\lambda}$. In this computation, the eigenvalue
$$\mu_{[3k]}=2g(2g-2)\cdots(2g-6k+2)$$
of $E_{[3k]}$, which is the smallest one, played an important role.
As an application of this first case, we showed that $\mathcal{R}^{*}(\mathcal{M}_{g})$
is generated by the first $[g/3]$ $\mathrm{MMM}$ classes. Later Ionel [9] proved a stronger result that this fact holds
at the Chow algebra level, namely in $\mathcal{R}^{*}(\mathbf{M}_{g})$.
The second case corresponds to $\lambda=[3k-1,1]$ and $\mathrm{dim}\,E_{\lambda}=9k(2k-1)$,
the third case corresponds to $\lambda=[3k-2,2]\ (k\geq 2)$ and $\mathrm{dim}\,E_{\lambda}=\frac{1}{2}k(3k-1)(6k-1)(6k-7)$,
and so on. Schematically, we may write
$$\displaystyle[3k]\mapsto 0\quad(g\leq 3k-1)\ \Rightarrow\ \text{$\mathcal{R}^{%
*}(\mathcal{M}_{g})$ is generated by the first $[g/3]$ $\mathrm{MMM}$ classes}$$
$$\displaystyle[3k-1,1]\mapsto 0\quad(g\leq 3k-2)$$
$$\displaystyle[3k-2,2],[3k-2,1^{2}]\mapsto 0\quad(g\leq 3k-3)$$
$$\displaystyle[3k-3,3],[3k-3,21],[3k-3,1^{3}]\mapsto 0\quad(g\leq 3k-4)$$
$$\displaystyle[3k-4,4],[3k-4,31],[3k-4,2^{2}],[3k-4,21^{2}],[3k-3,1^{4}]\mapsto
0%
\quad(g\leq 3k-5)$$
$$\displaystyle\cdots$$
Although we have to consider the collapsing $E_{\lambda}\rightarrow L_{\lambda^{\prime}}$ described
in diagram (12), it seems reasonable to expect that many relations can be obtained in this way.
We would like to pursue this in future.
In view of the fact that Theorem 1.1 gives a complete description how the
$\mathrm{Sp}$-invariant tensors degenerate according as the genus decreases from the stable range to the
last case $g=1$, it would be reasonable to make the following.
Conjecture 6.5.
The relations given in Theorem 6.3 make a complete set of relations for the tautological algebras
$\mathcal{R}^{*}(\mathcal{M}_{g})$ and $\mathcal{R}^{*}(\mathcal{M}_{g,*})$.
7. Concluding remarks
As was already mentioned in the introduction, in a joint work with T. Sakasai and M. Suzuki, we found another
role of the canonical metric introduced in this paper. In this work, we are studying the structure of the
symplectic derivation Lie algebra $\mathfrak{h}_{g,1}$ which is the Lie algebra consisting of symplectic
derivations of the free Lie algebra generated by $H_{\mathbb{Q}}$. This Lie algebra is a very important subject in topology
and also in other related fields of mathematics including number theory.
In particular, the symplectic invariant part $\mathfrak{h}_{g,1}^{\mathrm{Sp}}$
of this Lie algebra concerns many deep questions in various contexts, e.g. homology cobordism invariants for homology cylinders,
the rational cohomology group of the outer automorphism group of free groups and the Galois obstructions
in the theory of arithmetic mapping class groups.
We obtain several results about $\mathfrak{h}_{g,1}^{\mathrm{Sp}}$ and
we refer to our paper [30] for details.
Finally we would like to mention another possible application of our metric.
That is concerned with “harmonic representatives” of cohomology classes
which arise in various symplectic Lie algebras, by which we mean graded Lie algebras
such that each graded piece is an $\mathrm{Sp}$-module and the bracket operation
is an $\mathrm{Sp}$-morphism.
We are keeping in mind the Lie algebra $\mathfrak{h}_{g,1}$ mentioned above
and more generally Kontsevich’s graph (co)homology theory (see [14][15]).
It might be worthwhile to investigate whether harmonic representatives of
cohomology classes of these Lie algebras would uncover some unknown geometric property of them.
In particular, the case of the Gel’fand-Fuks cohomology of formal Hamiltonian vector fields
on $H_{\mathbb{R}}=H_{\mathbb{Q}}\otimes\mathbb{R}$, both in the stable as well as the unstable contexts, seems
to be interesting. In the simplest case of formal Hamiltonian vector fields on the plane,
Gel’fand, Kalinin and Fuks found in [6] an exotic class and Metoki
found another one in [20]. The problem of determining whether these classes are non-trivial
as characteristic classes of transversely symplectic foliations remains unanswered up to the present.
In the framework due to Kontsevich [16], it was shown in our paper with Kotschick [17]
that the Gel’fand-Kalinin-Fuks class can be decomposed into a product of certain leaf cohomology class
and the transverse symplectic form and in [22] Mikami proved a similar statement for the
Metoki class. It would be worthwhile to study whether the harmonic representatives of these leaf cohomology
classes would shed light on the above difficult problem.
References
[1]
N. Enomoto, T. Satoh,
New series in the Johnson cokernels of the mapping class groups of surfaces,
Alg. Geom. Top. 14 (2014) 627–669.
[2]
C. Faber,
A conjectural description of of the tautological ring of the moduli space of curves,
in “Moduli of curves and abelian varieties”,
109–129, Aspects Math., E33, Vieweg, Braunschweig, 1999.
[3]
C. Faber,
Tautological algebras of the moduli space of curves,
in “Moduli Spaces of Riemann Surfaces”,
IAS/Park City Mathematical Series, Vol. 20,
197–219, Amer. Math. Soc., 2013.
[4]
W. Fulton, J. Harris,
“Representation Theory”,
Graduate Texts in Mathematics 129,
Springer Verlag, 1991.
[5]
S. Garoufalidis, H. Nakamura,
Some IHX-type relations on trivalent graphs and symplectic representation theory,
Math. Res. Lett. 5 (1998) 391–402.
Corrigendum: Math. Res. Lett. 14 (2007) 689–690.
[6]
I.M. Gelfand, D.I. Kalinin, D.B. Fuks,
The cohomology of the Lie algebra of Hamiltonian formal vector fields,
(Russian), Funkcional Anal. i Pril. 6 (1972), 25–29;
Engl. transl. in Funk. Anal. Appl. 6 (1973), 193–196.
[7]
I. Grigoriev,
Relations among characteristic classes of manifold bundles,
preprint, arXiv:1310.6804[math.AT].
[8]
J. Harer,
Stability of the homology of the mapping class groups of orientable surfaces,
Ann. of Math. (2) 121 (1985), 215–249.
[9]
E. Ionel,
Relations in the tautological ring of $\mathcal{M}_{g}$,
Duke Math. J. 129 (2005), 157–186.
[10]
N. Iwahori,
“Representation Theory of Symmetric Groups and General Linear Groups”,
in Japanese,
Iwanami Shoten, 1978.
[11]
N. Kawazumi,
A generalization of the Morita-Mumford classes to extended mapping class groups
for surfaces,
Invent. Math. 131 (1998), 137–149.
[12]
N. Kawazumi, S. Morita,
The primary approximation to the cohomology of the moduli space of curves and stable characteristic classes,
Math. Res. Lett. 3 (1996) 629–642.
[13]
N. Kawazumi, S. Morita,
The primary approximation to the cohomology of the moduli space of curves and cocycles for the Mumford-Morita-Miller classes,
preprint, Univ. of Tokyo, 2001.
[14]
M. Kontsevich,
Formal (non)commutative symplectic geometry,
in: “The Gel’fand Mathematical Seminars, 1990–1992”,
Birkhäuser, Boston (1993) 173–187.
[15]
M. Kontsevich,
Feynman diagrams and low-dimensional topology,
in: “First European Congress of Mathematics, Vol. II (Paris, 1992)”,
Progr. Math. 120, Birkhäuser, Basel (1994), 97–121.
[16]
M. Kontsevich,
Rozansky-Witten invariants via formal geometry, Compos. Math. 115
(1999), 115–127.
[17]
D. Kotschick, S. Morita,
The Gelfand-Kalinin-Fuks class and characteristic classes of transversely symplectic foliations,
preprint, arXiv:0910.3414[math.SG].
[18]
E. Looijenga,
On the tautological ring of $\mathcal{M}_{g}$,
Invent. Math. 121 (1995), 411–419.
[19]
I. Madsen, M. Weiss,
The stable moduli space of Riemann surfaces: Mumfordfs conjecture,
Ann. of Math. (2) 165 (2007), 843–94.
[20]
S. Metoki,
Non-trivial cohomology classes of Lie algebras of volume preserving formal vector fields,
Ph.D. thesis, University of Tokyo, 2000.
[21]
A. Mihailovs,
Symplectic tensor invariants, wave graphs and S-tris,
preprint, arXiv:9803102[math.RT].
[22]
K. Mikami,
An affirmative answer to a conjecture for Metoki class,
preprint, 2013.
[23]
S. Morita,
Families of Jacobian manifolds and characteristic classes of surface bundles, $I$,
Ann. Inst. Fourier (Grenoble) 39 (3) (1989), 777–810.
[24]
S. Morita,
Families of Jacobian manifolds and characteristic classes of surface bundles, $II$,
Math. Proc. Cambridge Philos. Soc. 105 (1989), 79–101.
[25]
S. Morita,
A linear representation of the mapping class group of orientable surfaces and characteristic classes of surface bundles,
Topology of Teichmüller spaces, World Scientific,
1996, 159–186.
[26]
S. Morita,
Structure of the mapping class groups
of surfaces: a survey and a prospect,
Geometry and Topology monographs 2,
Proceedings of the Kirbyfest (1999), 349–406.
[27]
S. Morita,
Structure of the mapping class group and symplectic representation theory,
l’Enseignement Math. Monographs 38 (2001), 577–596.
[28]
S. Morita,
Generators for the tautological algebra of the moduli space of curves,
Topology 42 (2003), 787–819.
[29]
S. Morita, T. Sakasai, M. Suzuki,
Computations in formal symplectic geometry and characteristic classes of
moduli spaces, preprint, arXiv:1207.4350[math.AT],
to appear in Quantum Topology.
[30]
S. Morita, T. Sakasai, M. Suzuki,
Structure of symplectic invariant Lie subalgebras
of symplectic derivation Lie algebras, preprint.
[31]
R. Pandharipande, A. Pixton,
Relations in the tautological ring of the moduli space of curves,
preprint, arXiv:1301.4561[math.AG].
[32]
G. Rumer, E. Teller, H. Weyl,
Eine für die Valenztheorie geeignete Basis der binaren Vectorinvarianten,
Nachr. Ges. Wiss. Göttingen Math. -Phys. Kl. (1932), 499–504.
[33]
O. Randal-Williams,
Relations among tautological classes revisited,
Adv. Math. 231 (2012) 1773–1785.
[34]
J.-P. Serre,
“Linear Representations of Finite Groups”,
Translated from the second French edition by Leonard L. Scott. Graduate Texts in
Mathematics, 42. Springer Verlag, New York, Heidelberg, 1977.
[35]
Q. Yin,
On the tautological rings of $\mathcal{M}_{g,1}$ and its universal Jacobian,
preprint, arXiv:1206.3783[math.AG]. |
AML-SVM: Adaptive Multilevel Learning with Support Vector Machines
1st Ehsan Sadrfaridpour
School of Computing
Clemson University
Clemson, USA
[email protected]
2nd Korey Palmer
School of Computing
Clemson University
Clemson, USA
[email protected]
3rd Ilya Safro
School of Computing
Clemson University
Clemson, USA
[email protected]
Abstract
The support vector machines (SVM) is one of the most widely used and practical optimization based classification models in machine learning because of its interpretability and flexibility to produce high quality results. However, the big data imposes a certain difficulty to the most sophisticated but relatively slow versions of SVM, namely, the nonlinear SVM.
The complexity of nonlinear SVM solvers and the number of elements in the kernel matrix quadratically increases with the number of samples in training data.
Therefore, both runtime and memory requirements are negatively affected.
Moreover, the parameter fitting has extra kernel parameters to tune, which exacerbate the runtime even further.
This paper proposes an adaptive multilevel learning framework for the nonlinear SVM, which addresses these challenges, improves the classification quality across the refinement process, and leverages multi-threaded parallel processing for better performance.
The integration of parameter fitting in the hierarchical learning framework and adaptive process to stop unnecessary computation significantly reduce the running time while increase the overall performance.
The experimental results demonstrate reduced variance on prediction over validation and test data across levels in the hierarchy, and significant speedup compared to state-of-the-art nonlinear SVM libraries without a decrease in the classification quality. The code is accessible at https://github.com/esadr/amlsvm.
Classification, Multilevel Computation, Large-scale learning, Support Vector Machine
\definechangesauthor
[name=”Ehsan”, color=blue]es
\definechangesauthor[name=”Korey”, color=green]kp
\definechangesauthor[name=”Ilya”, color=red]is
\bstctlcite
IEEEexample:BSTcontrol
I Introduction
Support vector machine (SVM) is a widely used family of classification methods that leverage the principle of separating hyperplane. Technically, this is achieved by solving underlying regularized optimization model adjusting which can provide highly accurate and interpretable classification.
Training linear SVM is very fast and can scale to millions of data points and features without using significant high-performance computing (HPC) resources.
For problems that are not linearly separable, the nonlinear SVM uses the kernel trick by implicitly projecting the data into the higher-dimensional space to separate it by a hyperplane.
Nonlinear SVM usually reaches higher prediction quality on complex datasets. However, it comes with a price tag of being not scalable in comparison to its linear version.
Solving the Lagrangian dual problem is typically the way to cope with regularized nonlinear SVM models with the underlying convex quadratic programming (QP) problem. In a number of libraries (such as LibSVM [1]) multiple methods have been implemented for solving both primal and dual problems.
The complexity of the convex QP solvers for nonlinear SVM often scales between $O(n^{2}f)$ to $O(n^{3}f)$ [2] , where $n$ is the number of data points, and $f$ is the number of features.
Therefore, as $n$ increases, the running time of the solver also increases, which hinders the usage of the nonlinear SVM for massive data sets.
As the size of many datasets continues to grow due to the advancements in technologies such as high-throughput sequencing, e-commerce [3] and the Internet of Things [4], more scalable machine learning algorithms are required.
Therefore, while a nonlinear SVM is fast on small datasets and can provide highly accurate prediction, more research is required to develop scalable nonlinear SVM solvers for massive datasets.
Our previous framework, Multilevel SVM (MLSVM), is a library that scales to millions of data points and exhibits up to two orders of magnitude faster running time compared to LIBSVM on reported benchmark datasets [5].
The MLSVM leverages multilevel algorithms that are inspired by the algebraic multigrid and some of its restricted versions [6]. These algorithms are known to be successful in accelerating computational optimization and modeling tasks without a loss in the quality. Examples include hypergraph partitioning [7], graph generation [8], and clustering [9]. Such multilevel methods have two essential phases, namely, coarsening and uncoarsening [6].
The coarsening phase gradually reduces the original problem size and generates a chain of smaller problems that approximate the original one. This is done by constructing
a (possibly fuzzy) hierarchy of aggregated data points.
The uncoarsening phase starts from the smallest generated problem and gradually uses generated approximations to refine the solution of the original problem. During the uncoarsening phase, a refinement leverages a solution inherited from the previous coarser problem and is performed using a local processing solver to avoid any heavy computation.
The multilevel SVM [10, 11, 5] is an approach to accelerate traditional SVM model solvers which also often improves the model quality as it exploits the geometry of data before starting the training. The key principles of multilevel SVM are as follows:
1) Learning the separating hyperplane occurs within a small proximity to support vectors. 2) Perform computationally expensive parts of the training only at the coarse levels. 3) The best final model is not necessarily the finest one.
We discussed several other principles in [5]. Among them is the one that discusses the way multilevel SVM copes with imbalanced data.
While the MLSVM library [5] provided fast runtime and highly accurate predictions on massive datasets, we observed several unexpected results on benchmark datasets that motivated us to extend that work and introduce the Adaptive MLSVM in this paper.
In multilevel algorithms [12, 6], it is expected to observe that the quality of optimized objective is improved at each next finer level during the uncoarsening. As such, we expected to see a non-decreasing quality improvement across the uncoarsening. However, on some most difficult benchmarks, we observed that while the trained coarse models have exhibited reasonable to optimal prediction quality, some of the middle levels experienced a significant drop in the quality. In some cases, these middle level drops were continuing to the fine levels and in some were gradually improved back to the high quality results.
Although the observed quality drops were rare, they often led to accumulating and increasing errors during the uncoarsening. Indeed, low quality support vectors may potentially be disaggregated into the data points that are even more distant to the optimal hyperplane which gives too much freedom to the optimizer. Other stages in the MLSVM pipeline such as the filtering of data points after too aggressive disaggregation were also affected.
Our Contribution In this paper, we propose the Adaptive Multilevel SVM, a successful approach to detect the problem of inconsistent learning quality during the uncoarsening and efficiently mitigate it. At each level of the multilevel hierarchy, we detect the problem of quality decrease by validating the model using the finest level data.
In the adaptive multilevel SVM learning framework, we adjust the training data by filling the training data gap with the misclassified validation data points, retraining the model and improving the quality with a new set of support vectors.
Our exhaustive experimental results on the benchmark datasets demonstrate the proposed method recovers the multilevel framework from such quality drops and achieves higher quality in comparison to the non-adaptive multilevel SVM as well as other state of the art solvers.
In addition, we speed up even more the runtime compared to non-adaptive multilevel SVM and reduce the prediction quality variance.
Moreover, in the new version of the MLSVM library we implement the multi-thread support for parameter fitting to speed up the model training at each level.
Our implementation is open-source and available at http://github.com/esadr/amlsvm.
II Related Works
Improving the performance of SVM has been widely studied with a common goal of improving the training time without a significant decrease in prediction accuracy such as recent researches
[13, 14, 15, 16].
The complexity of nonlinear SVM on massive datasets is a potential target to develop scalable SVMs.
A challenge with solving the quadratic programming problem for the kernel SVM is the kernel matrix of size $n^{2}$, where $n$ is the number of training data points. The matrix requires large memory, which is not feasible for massive datasets.
The main categories of accelerated solvers follow one of the following approaches: a) target the solver performance directly, b) partition the original data, or c) considering alternative representation for the training data (including the multilevel approach), and their combinations. Some of the improvements also rely on the advancement of software infrastructure, hardware, and distributed frameworks.
Solver performance:
A Sequential Minimal Optimization (SMO)[17] has been used to solve the underlying quadratic programming problem.
The ThunderSVM [15] directly works on improving the performance of the solver using multiple cores and GPUs.
Partition original training data:
Instead of solving the QP with a large number of data points, a set of smaller problems can be generated and solved independently. This reduces the running time of training and memory utilization required to store the kernel matrix. A drawback to this approach is the quality of partitioning the original data and the quality of the approach to combine all the trained models to drive a final model.
DC-SVM[18] used adaptive clustering for dividing the training data
and relied on coordinate descent method to converge to a solution from multiple models.
A disadvantage for these approaches is relying on the partitioning or clustering
that is sensitive to the order of the training data, number of clusters. Assignment of a point to a partition or cluster is strict, which limits the points to move between clusters or participate in multiple clusters.
Graph representation of training data:
There are two schemas for partitioning the data. The earlier approaches rely on the original data in the feature space, while newer approaches such as[11, 5, 13] rely on a graph representation of data. The graph representation of the training data is constructed as a preprocessing step for partitioning which is explained in detail in section III-B.
The graph representation provides the opportunity to leverage the multilevel paradigm that has been used successfully in a wide range of problems. We briefly mention some of the multilevel research related to graphs such as graph partitioning [19], graph clustering [20], and image segmentation [21].
The advantage of using a graph and multilevel paradigm such as Algebraic Multigrid (AMG) is to exploit more relaxed and less strict assignments of the data points to smaller aggregates compare to clustering, which has strict assignments and larger cluster sizes.
The position of AML-SVM:
Our proposed framework (AML-SVM) has a robust coarsening, which allows partial participation of points in multiple aggregates through gradual assignment and relaxation steps. Our results demonstrate that a small training data at the coarsest level can be used to train a model with high accuracy.
AML-SVM can directly leverage the performance improvements introduced by advanced solvers with multi-core or GPU support to achieve even faster runtime without any change in the coarsening or refinement process.
The uncoarsening (refinement) phase as a general step in multilevel methods has the potential for improvement of initial solutions and carrying essential information such as support vectors, and parameters to other levels.
The improvements proposed in this paper reduce the variance between the quality of models at various levels and reduce the sensitivity of the framework to configuration parameters.
Furthermore, these improvements, along with multiple core support for faster parameter fitting, can be used for ensemble models that we have not explored.
III Preliminaries
Consider a set $\mathcal{J}$ of input samples that contains $n$ data points denoted by $x_{i}$, where $x_{i}\in\mathbb{R}^{d}$, $1\leq i\leq n$. Each data point $x_{i}$ has a corresponding label $y_{i}\in\{-1,1\}$.
The SVM as a binary classification finds the largest margin hyperplane that separates the two classes of labeled data. The solution to the following optimization problem produces the largest margin hyperplane, which is defined by $w$, and $b$.
minimize
$$\displaystyle\frac{1}{2}\lVert w\rVert^{2}+{C}\sum_{i=1}^{n}\xi_{i}$$
(1)
subject to
$$\displaystyle y_{i}(w^{T}\phi(x_{i})+b)\geq 1-\xi_{i},$$
$$\displaystyle\quad i=1,\dots,n$$
$$\displaystyle\xi_{i}\geq 0,$$
$$\displaystyle\quad i=1,\dots,n.$$
The misclassification is penalized using slack variables $\{\xi_{i}\}_{i=1}^{n}$.
The parameter $C>0$ controls the magnitude of penalty for miscalssified data points.
The primal formulation in (1) is known as the soft margin SVM [22].
The SVM takes advantage of kernel $\phi:\mathbb{R}^{d}\rightarrow\mathbb{R}^{p}~{}(d\leq p)$ to map data points to higher-dimensional space. The kernel measures the similarity for pairs of points $x_{i}$ and $x_{j}$.
The Gaussian kernel (RBF), $\exp(-\gamma||x_{i}-x_{j}||^{2})$, which is known to be generally reliable [23], is the default kernel in this paper.
For imbalanced datasets, different weights can be assigned to classes using the weighted SVM (WSVM) to controllably penalize points in the minority class $C^{+}$ (e.g., rare events).
The set of slack variables is split into two disjoint sets $\{\xi_{i}^{+}\}_{i=1}^{n^{+}}$, and $\{\xi_{i}^{-}\}_{i=1}^{n^{-}}$, respectively.
In WSVM, the objective of (1) is changed into
minimize
$$\displaystyle\frac{1}{2}\lVert w\rVert^{2}+C\big{(}{W^{+}}\sum_{i=1}^{n^{+}}%
\xi_{i}^{+}+{W^{-}}\sum_{j=1}^{n^{-}}\xi_{j}^{-}\big{)}.$$
(2)
The imbalanced data sets contain significanlty fewer positively labeled data points. Hence, the data points in the minority class is denoted by $\mathcal{J}^{+}$, where size of minority class is $n^{+}=|\mathcal{J}^{+}|$.
The rest of the points belongs to the majority class which is denoted by $\mathcal{J}^{-}$, where $n^{-}=|\mathcal{J}^{-}|$, and $\mathcal{J}=\mathcal{J}^{+}\cup\mathcal{J}^{-}$.
We assign the class weight as the inverse of the total number of points in that class. For instance, the weight for minority class is $W^{+}=\frac{1}{n^{+}}$
In the our multilevel framework, one of the basic steps is aggregating data points. The aggregation can be performed either on full points or their fractions (i.e., a data point can be split into fractions, and different fractions can contribute themselves to different aggregates). Therefore, a coarse level data point normally contains several finer points or their fractions. Because of this, a data point can be more or less representative in comparison to other data points. We define and use the volume property for each point and calculate it as number of partial or fully participated original points in that aggregate.
We denote $v_{i}$ as volume for point $x_{i}$ which represent the internal importance of point $i$ among points in the same class.
Moreover, we consider the class importance for the final weight of each data point. The calculation for both classes are the same, hence, we only present the positive class. We calculate the class weights based on sum of the volumes in the class, i.e.,
$W^{+}=\frac{1}{\sum\limits_{i\in\mathcal{J}^{+}}v_{i}}.$ We calculate the final weight for each point as product of its volume weight times class weight
$\forall i\in\mathcal{J}^{+(-)}\quad W_{i}=W^{+(-)}\frac{v_{i}}{\sum\limits_{j%
\in\mathcal{J}^{+(-)}}v_{j}}.$
The instance weight based on volume and class, helps to improve prediction quality for the smaller class.
III-A Two-level problem
To describe the coarsening, uncoarsening and refinement algorithms, we introduce the two-level problem notation that can be extended into a full multilevel hierarchy (see Figure LABEL:fig:aml_svm). In Algorithm LABEL:alg:aml_general_refinement, we will use subscript $(\cdot)_{f}$ and $(\cdot)_{c}$ to represent fine and coarse variables, respectively. For example, the data points of two consecutive levels, namely, fine and coarse, will be denoted by $\mathcal{J}_{f}$, and $\mathcal{J}_{c}$, respectively. The sets of fine and coarse support vectors are denoted by $\textsf{sv}_{f}$, and $\textsf{sv}_{c}$, respectively. We will also use a subscript in the parentheses to indicate the level number in the hierarchy where appropriate. For example, $\mathcal{J}_{(i)}$ will denote the set of data points at level $i$.
III-B Proximity graphs
The original MLSVM framework [5] is based on the algebraic multigrid coarsening scheme for graphs (see coarsening in [24, 25]) that creates a hierarchy of data proximity graphs.
Initially, at the finest level, $\mathcal{J}$ is represented as two approximate $k$-nearest neighbor ($k$NN) graphs $G^{+}_{(0)}=({\textbf{V}}^{+},E^{+})$, and $G^{-}_{(0)}=({\textbf{V}}^{-},E^{-})$ for minority and majority classes, respectively, where each $x_{i}\in{\textbf{V}}^{+(-)}$ corresponds to a node in $G^{+(-)}_{(0)}$.
A pair of nodes in $G_{(0)}^{+(-)}$ is connected with an edge that belongs to $E^{+(-)}$ if one of these nodes belongs to a set of $k$-nearest neighbors of another.
Throughout the multilevel hierarchies, in two-level representation, the fine and coarse level graphs will be denoted by $G^{+(-)}_{f}=({\textbf{V}}^{+(-)}_{f},E^{+(-)}_{f})$, and $G^{+(-)}_{c}=({\textbf{V}}^{+(-)}_{c},E^{+(-)}_{c})$, respectively. For $k$NN, our experiments show $k=10$, for all tested datasets, provides a good trade-off between computational cost of calculating $k$NN and average node degree for $G^{+(-)}$.
III-C Learning Challenge
In the hierarchical framework, it is expected that the first solution is only a reasonable (but not necessarily the best) approximation of the optimal solution. Therefore, extra steps are required to improve coarse solutions gradually.
In the context of the classification task, the initial solution should provide a reasonable decision boundary. We can measure the quality of the initial solution with performance measures on validation data, which we cover in section IV. We can express it more accurately by
$Q_{(1)}\leq Q_{(2)}...\leq Q_{(k)},$
when $Q_{(i)}$ is the classification quality at level $i$. We denote this expected trend with a continuous increase in quality as the natural trend and Figure 1(a) demonstrates it for the Ringnorm data set.
However, this is not what we observed over exhaustive experiments with various data sets. In our experimental results, there were some cases in which the classification quality was dropped significantly during the refinement. The consequences of unexpected phenomena are as follows.
The training process for the models at the following levels of refinement will be based on sub-optimal training data. As explained earlier, the training data at each level of refinement is filtered to contain only the neighbor points of SVs of a model at the coarser level. When the SVs are sub-optimal, the filtering adds the points which are not representative of the data for the SVM to learn an optimal decision boundary.
Moreover, the number of SVs may increase dramatically without significant improvement in the classification quality. The increase in the number of SVs drastically increases the number of SV’s neighbors, which forms a larger training data at the next level. Therefore, it reduces the performance of the rest of levels in the refinement.
We recognized a pattern that the quality improves during refinement but at some levels, declines significantly. We denote it as unnatural trend and Figure 1(b) demonstrates an example of the unnatural trend for the Cod-RNA dataset.
Benefits of detecting and addressing the decrease in quality during the refinement are:
•
Improving the classification quality of the finer levels and in general
•
Improving the computational performance of the refinement process by improving the quality of training data set for the finer levels.
•
Reducing the computational cost of refinement through the Early Stopping process when there is not a significant improvement in quality by training larger models
•
Reducing the variance for the classification quality across levels
•
Preserve smaller training set during the refinement
Intuitively, the coarsening procedure gradually creates approximated (or summarized) representations of the finest level (original) data. An initial model we train at the coarsest level, provides a large margin hyperplane which can be used as the final solution.
On the one hand, at the finest level, rich data can easily lead to over-fitted models, a phenomenon frequently observed in practice [31]. On the other hand, over-compressed data representation may lead to an under-fitted model due to lack of detail information.
Our experiments confirm that more than half of the best models are obtained from the coarse (but not coarsest) and middle levels which typically prevents over- and under-fitting.
In general, if the best models were produced by the finest and middle levels, we recommend to use the model obtained at the middle levels to avoid potential over-fitting. This recommendation is based on the observation that same quality models can be generated by different hyperplanes but finest models may contain a large number of support vectors that can lead to over-fitting and slower prediction. However, this is a hypothesis that requires further exploration. In our experiments, no additional parameters or conditions are introduced to choose the final model. We simply choose the best model among those generated at different levels.
III-D Proposed Adaptive Method to Recover the Quality Drop
The drop in classification quality is equivalent to an increase in the number of points that are misclassified. Therefore, a classifier quality can be improved by training on more similar data to misclassified data points. The training data at each level is a sub-sample of data around the decision boundary (hyperplane) at the coarser level. The data is sampled from the neighborhood of SVs from the trained model at the coarser level. The points which are misclassified in the training data are not crucial for improvement since the SVM did not find a better solution for the decision boundary. However, the validation data points which are misclassified are the crucial points that need to be classified correctly in this step. It is worth to notice, this step does not rely on any information from the test data and only rely on the validation and training data.
Increasing the number of data points usually help the classifier to learn a better decision boundary. A $p$,$n$ number of neighbors of misclassified validation data points are found in the data from positive and negative class, respectively. The neighbor points are added to the training data, and the model is retrained. The $p$, $n$ can start from 1 and increase to larger values. However, as these parameters increase, the number of data points that are added to training data increase with respect to the validation data set size and affect the computational performance in the current level and overall framework.
We mention important considerations of earlier work in the MLSVM framework.
It is essential to mention that for the validation data, we used $FF$ method in MLSVM that is based on the sampling of the original training data at the finest level.
Therefore, the size of the validation data set for large data sets is significantly larger than the training data at the coarser levels. For instance, a data set with $5M$ data points such as SUSY with k-fold cross-validation using $k=5$ has $4M$ training data at the finest level. Using the validation sample rate of $10\%$, the validation size at all levels has $400k$ data points. At the coarsest level, the training size is less than $2\times M$ (600 samples with default $M=300$)
, and it continues to grow slowly over the next few levels of refinement. Suppose the accuracy is $80\%$ at a level. The maximum number of points that can be added to training data would be $400k*20\%*(p+n)$ or $80k*(p+n)$. This would be an impractical training size. Therefore, we limit the number of $p$ and $n$ to one. In our experiments, many miclassified points have common neighbors which were not in the training. We only add distinct points into training set, hence, the increase is training size is much smaller.
The focus of MLSVM is on extremely imbalanced data sets.
The majority of data points belong to the larger class in highly imbalanced data sets. Therefore, a more restrict sampling ratio helps to reduce the number of data points from the majority or larger class in the validation data. However, the minority or smaller class typically has ten times fewer data points, and its size would not cause performance problems. Therefore, a more significant sample rate, such as $50\%$ or more for minority class compares to $10\%$ or less for majority class, both reduce the number of points in the validation data and make the validation data better representative sample to evaluate the quality of the classifier. A validation set is significant for the framework, which allows us to find the best model from many models at various levels. If the validation data is not representative, the quality of the model will severely affect the test data, which is the ultimate goal of training a classifier.
The validation data size for a large data set with a sample ratio of $10\%$ is still huge. Using smaller values such as $2\%$, the validation data was not representative of test data anymore in the past. With proposed new unbalanced sampling ratios for validation, the $2\%$ validation sample ratio of majority class, and $50\%$ sample ratio of the minority class, reduce the computational challenge.
The notation used in Algorithm 1 is explained in Table I.
IV Experimental Results
First, we report the comparison between the ML-SVM algorithm and the state-of-the-art SVM algorithms such as LIBSVM and DC-SVM in terms of classification quality and computational performance. Then, we compare the proposed AML-SVM method and ML-SVM method using new set of experiments with more details.
IV-A Performance measures
Performance measures are metrics which are used to evaluate the classification quality of the model. We report the
accuracy (ACC), recall or sensitivity (SN), specificity (SP), and geometric mean (G-mean). They defined as
${\textrm{ACC}=\frac{TP+TN}{FP+TN+TP+FN}},{\textrm{SN}=\frac{TP}{TP+FN}},{%
\textrm{SP}=\frac{TN}{TN+FP}},$
${\textrm{G-mean}=\sqrt{\textrm{SP}\cdot\textrm{SN}}},$
Where $TN$, $TP$, $FP$, and $FN$ correspond to the numbers of real negative, true positive, false positive, and false negative points.
Our primary metric for comparison is G-mean, which measures the balance between classification quality on both the majority and minority classes. This metric is illuminating for imbalanced classification as a low G-mean is an indication of low-quality classification of the positive data points even if the negative points classification is of high quality. This measure indicates the over-fitting of the negative class and under-fitting of the positive class, a critical problem in imbalanced data sets.
In both ML and AML -SVM frameworks, many models are trained. We need to provide one value for prediction as to the final model performance measure. Therefore, we evaluate all the models using the validation set and select the best model and report the performance measures of the selected model over the test (hold-out) set in the results.
The detail of imbalanced data sets used in our experiment is presented in Table II. The imbalance ratio of data sets is denoted by $\epsilon$.
IV-B LIBSVM, DC-SVM and ML-SVM with AML-SVM Comparison
For self completeness of this paper and convenience of the readers, we present all the relevant results from the ML-SVM paper. The following results provides the clear comparison between the state-of-the-art methods which set the base for next section which we provide more detailed results for comparing the ML-SVM and AML-SVM methods.
IV-C Implementation
The ML-SVM was developed using C++ and PETSc library [32].
The following libraries are used for various parts of the framework. The FLANN library for finding approximate k-nearest neighbors before coarsening, and METIS library
for graph partitioning during the refinement.
The LIBSVM [1] as the underlying solver for SVM. The implementation is based on a single CPU core, and all the speedups are through algorithmic improvement without leveraging parallel processing.
The proposed framework is developed based on the ML-SVM library.
Moreover, the OpenMP library is speed up the parameter fitting at each level using multi-threading.
Since the number of parameter fitting used for experiments is 9 in the first stage and 4 in the second stage of nested uniform design, we do not evaluate the speedups by increasing the number of threads. The coarsening implementation is sequential and based on a single CPU core.
All experiments for data sets with less than 1M data points have executed on a single machine with CPU Intel Xeon E5-2665 2.4 GHz and 64 GB RAM. For larger data sets, we used one single machine with CPU Intel Xeon E5-2680v3 2.5 GHz and 128 GB RAM.
IV-D ML-SVM with AML-SVM Comparison
The same experiment setup and hardware configuration is used for the results in this section and next section. Tables III and
IV demonstrate a comparison across various libraries [5]. The C-Time and W-Time are used to report CPU time and wall time respectively.
However, the purpose of experiments and the version of ML-SVM are different. The experiments are designed to have a better understanding of the multilevel framework and refinement process on each level of the hierarchical framework, while the earlier experiments are designed to compare the overall frameworks regardless of internal details.
The CPU time is accurate metric for computational time of sequential programs such as LIBSVM, DC-SVM, and ML-SVM. Therefore, the CPU time is presented for them. However, The AML-SVM leverages multicore parallel processing which can reduce the wall time, hence, both CPU and wall times are reported.
In all experiments the data is normalized using z-score.
The computational time reported in all experiments contains generating the $k$-NN graph. The computational time is reported in seconds unless it is explicitly mentioned otherwise.
In each class, a part of the data is assigned to be the test data using $k$-fold cross validation. We experimented with $k=5$ and $10$ (no significant difference was observed). The experiments are repeated $k$ times to cover all the data as test data. The data randomly shuffled for each $k$-fold cross validation.The presented results are the averages of performance measures for all $k$ folds.
Data points which are not in the test data are used as the training data in $\mathcal{J}^{+(-)}$.
The test data is never used for any training or validation purposes.
We run 18 experiments per dataset using nested cross-validation with k-fold where $k\in\{5,10\}$. with various configurations to evaluate the performance of the proposed method (AML). For each data set, we evaluated the combination of important parameters such as interpolation order (r), validation sample ratio (v), and where we stop to partition the data in ML method, which we define as stopping criteria for the AML method. We controlled all the other variables only to consider the effect of new method for an exactly similar configuration. Each combination of configuration provide slightly different results per dataset, and therefore, we evaluate 18 combinations of parameters and present the statistical information per dataset. The goal of the experiment is to have a well-designed experiment for comparing the current and new methods. We provide the related results from the MLSVM paper for convenient comparison of AML method with other states of the arts SVM solvers such as LIBSVM and DC-SVM.
For each configuration, we present three sets of results over k-fold cross-validation: (1) Classification quality: we plot the statistics of G-mean at each level of uncoarsening; (2) Computational Performance: the average running time for both coarsening and uncoarsening framework is reported; and (3) An overall comparison only for the final results of the whole framework without details for levels are provided in Tables III to V. We only plot the statistics for Clean, Cod, and SUSY data set in figure 2. The rest of plots are available online at the GitHub repo.
The number of levels in coarsening and uncoarsening is the same, and we use them interchangeably for the rest of this section.
The Advertisement dataset has 3,279 data points and 1,558 features. It has a large number of features and a small number of data points. The number of levels in the coarsening phase, only depends on the number of data points. Therefore, there are only 3 levels in (un)coarsening phase for the Advertisement dataset.
Both ML-SVM and AML-SVM frameworks have identical results for the first and second levels. However, the AML-SVM achieved higher quality on the final level. All the qualities are reported based on G-mean performance measures.
The Buzz dataset has 140,707 data points, 77 features, and 7 levels of coarsening. The quality of both methods is comparable except for a significant drop on the 2nd level for the ML-SVM quality.
The Clean dataset has 6,598 data points, 166 features, and 5 levels of coarsening. The ML-SVM’s quality is descending from the 3rd level to the 5th level. However, while the AML-SVM has not decreased, in the 4th and 5th levels, it achieves the highest quality across all the levels.
The Cod-RNA dataset is related to Breast Cancer. It has 59,535 data points and 8 features. It has a similar decreasing trend as the Clean dataset.
The Forest (Cover Type) dataset is one of the larger data sets with 581,012 data points and 54 features. Contrary to the earlier data sets, the quality (G-mean) at the first (coarsest) level is lower than most of the levels. The ML-SVM quality is increased in the 2nd level but continues to descend til level six. The quality at 7th level is increased with a slight decrease in the following levels.
The AML-SVM recovers the descend at the third level and achieves the highest G-mean in third and fourth levels.
We designed early stopping for the AML-SVM to reduce the computational complexity. Therefore, as the size of training data reaches a threshold $\theta$ in Algorithm LABEL:alg:aml_general_refinement, the refinement process stops.
The ML-SVM has a descending trend for quality on the Letter data set similar to Clean and Cod-RNA data sets. However, AML-SVM has no decrease in quality.
The Ringnorm dataset has 7,400 data points and 20 features. The ascending quality is observed for both methods without any drop in qualities.
The Twonorm dataset has the same size and number of features as the Ringnorm.The ML-SVM quality drops in the 3rd and 4th levels. The adaptive methods achieve high-quality results on all levels.
The quality of ML-SVM for both HIGGS and SUSY data set has a similar trend of high quality at the coarsest levels following with lower quality in the middle levels. At the finer levels, there is a slight improvement, but the variance of qualities is larger than the coarser levels.
The adaptive method for the SUSY dataset achieves high quality up to level 6. In levels 7 and 8 the quality of adaptive method drops as well, and at level 8, the adaptive method stops. For the HIGGS dataset, the adaptive method has better quality compared to the normal method. At level 9, the adaptive method stops to reduce the computational complexity.
V Conclusion
In this paper, we introduced a novel adaptive recovery technique for under/over-fitting challenges in hierarchical frameworks which performs well on nonlinear support vector machines. Our results based on 18 distinct configurations of essential parameters for both coarsening and uncoarsening phases show that the adaptive approach is less sensitive to changes in hyper-parameters. A significant advantage is improved computational time.The adaptive method has reduced the prediction quality variance across various levels in the multilevel framework, which provides a more robust solution.
On massive data sets, the training size may increase significantly at the fine levels. Our proposed early stopping functionality improves the computational performance as it prevents computationally expensive fine level training. In addition, we developed multi-threading support for parameter fitting, which utilizes more processing power on single computing node and increases the overall performance.
The proposed approach can be extended to a hierarchical learning framework with a clear objective for quality.
While our exhaustive results demonstrate the ability of classifiers to learn a high-quality model at the coarse and middle levels, the model selection over many models, or training an ensemble model can be studied further.
For large-scale problems, sampling is a traditional approach to reduce the computational time for intermediate tasks such as feature selection or dimensionality reduction. The proposed framework is a useful alternative for blind sampling of data.
References
[1]
C.-C. Chang and C.-J. Lin, “Libsvm: A library for support vector machines,”
ACM transactions on intelligent systems and technology (TIST), vol. 2,
no. 3, pp. 1–27, 2011.
[2]
H. P. Graf, E. Cosatto, L. Bottou, I. Dourdanovic, and V. Vapnik, “Parallel
support vector machines: The cascade SVM,” in Advances in neural
information processing systems, 2004, pp. 521–528.
[3]
M. Heidari and S. Rafatirad, “Using transfer learning approach to implement
convolutional neural network to recommend airline tickets by using online
reviews,” in IEEE 2020 15th International Workshop on Semantic and
Social Media Adaptation and Personalization, SMAP 2020, 2020.
[4]
H. V. Jagadish, “Big data and science: Myths and reality,” Big Data
Research, vol. 2, no. 2, pp. 49–52, 2015.
[5]
E. Sadrfaridpour, T. Razzaghi, and I. Safro, “Engineering fast multilevel
support vector machines,” Machine Learning, pp. 1–39, 2019.
[6]
A. Brandt and D. Ron, “Chapter 1 : Multigrid solvers and multilevel
optimization strategies,” in Multilevel Optimization and VLSICAD,
J. Cong and J. R. Shinnerl, Eds. Kluwer, 2003.
[7]
R. Shaydulin, J. Chen, and I. Safro, “Relaxation-based coarsening for
multilevel hypergraph partitioning,” Multiscale Modeling &
Simulation, vol. 17, no. 1, pp. 482–506, 2019.
[8]
A. Gutfraind, I. Safro, and L. A. Meyers, “Multiscale Network Generation,”
in FUSION ’15: Proceedings of the 18th IEEE International Conference on
Information Fusion, 2015, pp. 158–165.
[9]
P. D’Ambra, L. Cutillo, and P. S. Vassilevski, “Bootstrap amg for spectral
clustering,” Computational and Mathematical Methods, vol. 1, no. 2,
p. e1020, 2019.
[10]
E. Sadrfaridpour, S. Jeereddy, K. Kennedy, A. Luckow, T. Razzaghi, and
I. Safro, “Algebraic multigrid support vector machines,” European
Symposium on Artificial Neural Networks, Computational Intelligence and
Machine Learning (ESANN), 2017.
[11]
T. Razzaghi and I. Safro, “Scalable multilevel support vector machines,” in
International Conference on Computational Science (ICCS), Procedia
Computer Science, vol. 51. Elsevier,
2015, pp. 2683–2687.
[12]
D. Ron, I. Safro, and A. Brandt, “Relaxation-based coarsening and multiscale
graph organization,” Multiscale Modeling & Simulation, vol. 9,
no. 1, pp. 407–423, 2011.
[13]
S. Schlag, M. Schmitt, and C. Schulz, “Faster support vector machines,” in
ALENEX. SIAM, 2019, pp.
199–210.
[14]
S. Shalev-Shwartz, Y. Singer, N. Srebro, and A. Cotter, “Pegasos: Primal
estimated sub-gradient solver for svm,” Mathematical programming,
vol. 127, no. 1, pp. 3–30, 2011.
[15]
Z. Wen, J. Shi, Q. Li, B. He, and J. Chen, “Thundersvm: A fast svm library on
gpus and cpus,” The Journal of Machine Learning Research, vol. 19,
no. 1, pp. 797–801, 2018.
[16]
Z. Zhang and T. W. Chow, “Maximum margin multisurface support tensor machines
with application to image classification and segmentation,” Expert
Systems with Applications, vol. 39, no. 1, pp. 849–860, 2012.
[17]
L. J. Cao, S. S. Keerthi, C.-J. Ong, J. Q. Zhang, and H. P. Lee, “Parallel
sequential minimal optimization for the training of support vector
machines,” Neural Networks, IEEE Transactions on, vol. 17, no. 4, pp.
1039–1049, 2006.
[18]
C.-J. Hsieh, S. Si, and I. Dhillon, “A divide-and-conquer solver for kernel
support vector machines,” in International conference on machine
learning, 2014, pp. 566–574.
[19]
P. Sanders and C. Schulz, “Engineering multilevel graph partitioning
algorithms,” in European Symposium on Algorithms. Springer, 2011, pp. 469–480.
[20]
I. Dhillon, Y. Guan, and B. Kulis, “A fast kernel-based multilevel algorithm
for graph clustering,” in Proceedings of the 11th ACM SIGKDD
International Conference on Knowledge Discovery and Data Mining
(KDD’05). ACM Press, 2005, pp.
629–634.
[21]
L. Grady and E. L. Schwartz, “Isoperimetric graph partitioning for image
segmentation,” IEEE Trans. on Pat. Anal. and Mach. Int, vol. 28, pp.
469–475, 2006.
[22]
Q. Wu and D.-X. Zhou, “Svm soft margin classifiers: linear programming versus
quadratic programming,” Neural computation, vol. 17, no. 5, pp.
1160–1187, 2005.
[23]
F. E. Tay and L. Cao, “Application of support vector machines in financial
time series forecasting,” Omega: The International Journal of
Management Science, vol. 29, no. 4, pp. 309–317, 2001.
[24]
I. Safro, D. Ron, and A. Brandt, “Graph minimum linear arrangement by
multilevel weighted edge contractions,” Journal of Algorithms,
vol. 60, no. 1, pp. 24–41, 2006.
[25]
I. Safro, D. Ron, and A. Brandt, “Multilevel algorithms for linear ordering problems,” ACM
Journal of Experimental Algorithmics, vol. 13, 2008.
[26]
I. Safro, P. Sanders, and C. Schulz, “Advanced coarsening schemes for graph
partitioning,” in Experimental Algorithms, R. Klasing, Ed. Berlin, Heidelberg: Springer Berlin Heidelberg,
2012, pp. 369–380.
[27]
S. Leyffer and I. Safro, “Fast response to infection spread and cyber attacks
on large-scale networks,” Journal of Complex Networks, vol. 1, no. 2,
pp. 183–199, 2013.
[28]
I. Safro and B. Temkin, “Multiscale approach for the network
compression-friendly ordering,” J. Discrete Algorithms, vol. 9,
no. 2, pp. 190–202, 2011.
[29]
U. Trottenberg and A. Schuller, Multigrid. Orlando, FL: Academic Press, 2001.
[30]
C. Huang, Y. Lee, D. Lin, and S. Huang, “Model selection for support vector
machines via uniform design,” Computational Statistics & Data
Analysis, vol. 52, no. 1, pp. 335–346, 2007.
[31]
T. Dietterich, “Overfitting and undercomputing in machine learning,”
ACM computing surveys (CSUR), vol. 27, no. 3, pp. 326–327, 1995.
[32]
S. Balay, S. Abhyankar, and et al., “PETSc users manual,” Argonne National
Laboratory, Tech. Rep. ANL-95/11 - Revision 3.7, 2016. [Online]. Available:
http://www.mcs.anl.gov/petsc |
\preprinttext
AUTHORS’ PREPRINT
\onlineid1105
\vgtccategoryResearch
\vgtcpapertypeVIS Full paper
\authorfooter
Jingyuan Liu and Chiew-Lan Tai are with Hong Kong University of Science and Technology. E-mail: [email protected]. Nazmus Saquib is with Tero Labs, California, United States.
Zhutian Chen is with Harvard University.
Rubaiat Habib Kazi and Li-Yi Wei are with Adobe Research.
Hongbo Fu is with City University of Hong Kong.
VCoach: A Customizable Visualization and Analysis System for Video-based Running Coaching
Jingyuan Liu
Nazmus Saquib
Zhutian Chen
Rubaiat Habib Kazi
Li-Yi Wei
Hongbo Fu
and Chiew-Lan Tai
Abstract
Videos are accessible media for analyzing sports postures and providing feedback to athletes.
Existing video-based coaching systems often present feedback on the correctness of poses by augmenting videos with visual markers either manually by a coach or automatically by computing key parameters from poses.
However, previewing and augmenting videos limit the analysis and visualization of human poses due to the fixed viewpoints, which confine the observation of captured human movements and cause ambiguity in the augmented feedback.
Besides, existing sport-specific systems with embedded bespoke pose attributes can hardly generalize to new attributes; directly overlaying two poses might not clearly visualize the key differences that viewers would like to pursue.
To address these issues,
we analyze and visualize human pose data with customizable viewpoints and attributes in the context of common biomechanics of running poses, such as joint angles and step distances.
Based on existing literature and a formative study, we have designed and implemented a system, VCoach, to provide feedback on running poses for amateurs.
VCoach provides automatic low-level comparisons of the running poses between a novice and an expert, and visualizes the pose differences as part-based 3D animations on a human model.
Meanwhile, it retains the users’ controllability and customizability in high-level functionalities, such as navigating the viewpoint for previewing feedback and defining their own pose attributes through our interface.
We conduct a user study to verify our design components and conduct expert interviews to evaluate the usefulness of the system.
keywords: Human Pose, Video Processing, Sports Data Analysis
\CCScatlist\CCScat
K.6.1Management of Computing and Information SystemsProject and People ManagementLife Cycle;
\CCScatK.7.mThe Computing ProfessionMiscellaneousEthics
\teaser
We present VCoach, a novel analysis and visualization system for providing feedback on running pose correction for amateur runners.
Our system compares two running poses (a) with respect to both common running pose attributes and specific attributes defined by users (e), and visualizes the differences with 3D animations (d).
The profile of running poses in the sample video (b) and differences in running poses (c) are previewed with glyphs.
\vgtcinsertpkg
1 Introduction
Running is a globally popular exercise, and many runners want to avoid injuries and improve their performance.
Not everyone can have access to human coaches, and thus various online materials and mobile apps have emerged to provide guidance on achieving correct running forms.
As with general sports training, an accessible means for novice sports players is to learn from pre-recorded performances of coaches or professional players by performing and comparing the same actions.
Despite the previous video-based systems for providing posture feedback [6, 11], analyzing and visualizing the differences in posture data in videos remain challenging, as discussed below.
According to the taxonomy of comparison-based visualization [15], existing visualizations for human pose comparison include displaying related poses in two videos side-by-side (juxtaposition) [48, 47, 37], overlaying one pose onto another (superposition) [11], and augmenting video with visual markers (explicit encoding) [46].
However, the main limitation of these video-based pose comparison techniques is that the appearances of observational biomechanical measurements, such as angles and distances, are often subject to changing viewpoints (see the toy example in Figure 1).
For sports coaching systems, such an ambiguity problem affects both the observation and the feedback.
When observing the actions in videos, the 3D human pose attributes might be distorted due to perspective shortening and thus fail to reflect the actual biomechanical measurements.
In visualization, the shapes of graphical annotation markers overlaid on videos are also subject to changing viewpoints, and are thus ambiguous in providing accurate corrective feedback to be perceived by amateur runners.
To promote spatial awareness, prior studies have attempted to analyze reconstructed 3D poses [13], fuse videos in multiviews [46], and use situated AR [28] and immersive visualization [10, 24].
Thanks to the emerging methods in monocular human reconstruction in computer vision [9, 16], reconstructing 3D poses has become an effective and accessible solution for videos.
Besides the ambiguity problem, another consideration is the data attributes for comparison, which can be classified as parametric and non-parametric.
Parametric pose features (e.g., knee angle) are sport-specific and pre-defined by domain experts [6].
The embedded bespoke knowledge makes sport-specific systems hard to scale and support users’ needs for individual customization.
Alternatively, non-parametric comparison avoids embedding bespoke knowledge by comparing the transferred and overlaid human poses [11].
Novices would need to infer the corrective feedback based on their perceptions.
To address the above-mentioned issues, we aim to develop an interactive system to analyze and visualize differences in human biomechanical data.
Our system, VCoach, provides intuitive and customizable corrective feedback for amateur runners.
To achieve this goal, we worked closely with experts in Sports Science to identify its designs based on the coaching process in practice.
As shown in VCoach: A Customizable Visualization and Analysis System for Video-based Running Coaching, our system takes as input a sample user video from an amateur runner and an exemplar video from an expert runner, and automatically performs pose analysis tasks, such as reconstructing 3D poses from videos and computing pose differences.
The differences are then visualized as short animations on a 3D human body model (VCoach: A Customizable Visualization and Analysis System for Video-based Running Coaching(d)) to resemble the dynamic demonstration of human coaches in practice.
To reduce the ambiguity of visualization, we propose to augment 3D visual markers onto the 3D body model instead of the video, such that users can either preview under our suggested viewpoints or manually navigate through viewpoints for better perception.
VCoach embeds pre-defined biomechanical attributes that are commonly used for analyzing running poses (e.g., leaning angle and foot landing position).
To support the analysis of attributes users are interested in but not embedded in the system (e.g., vertical bend angle of knees and height of feet),
we also provide an interface (VCoach: A Customizable Visualization and Analysis System for Video-based Running Coaching(e)) that allows users (advanced amateur runners or coaches) to manually label biomechanics.
The user-customized attribute will then be retrieved from both the sample and exemplar videos for comparison in the same way as those pre-defined attributes.
This attribute generalization is facilitated by a design of mappings for biomechanical data that unifies the representations of attributes, their differences and users’ interactions to label the attributes.
Specifically, we make use of the semantic model definition of SMPL 3D human mesh model [31].
Users annotate and define measurements on a 3D SMPL body model in T-pose, such that the defined attributes can be retrieved across multiple videos using model correspondence.
We design a user study and expert interviews to evaluate the design components and the overall effectiveness of our system.
For the scope of the user study we focus on adults in moderate-speed running (jogging), since this is the most common type and demography for running exercises.
The human pose analysis model in VCoach can generalize to other user groups than adult amateur runners, such as children and professional runners, with adapted visualizations of pose differences (e.g., cartoons for children and infographics with detailed figures for professional runners).
By replacing the current pre-defined attributes with key attributes of other sports, VCoach can also be generalized to support the posture analysis of new techniques, such as in skating [12] and high jump [38].
2 Related Work
Pose Coaching Systems
Previous research work on video-based running pose analysis is limited, partly because in-the-wild running poses contain larger variations in appearance than other sports with more confined locomotion ranges, such as yoga [6] and golf [35].
Running dynamics, such as ground contact time and vertical oscillation, require specific combinations of hardware to capture (e.g., [52]).
In the following, we review posture coaching systems in general sports.
According to how the bespoke knowledge of a specific sport is introduced into the system, existing coaching tools span the spectrum from fully-manual to fully-automatic, as illustrated in Figure 2.
The other dimension is whether the poses are captured in 2D (videos) or in 3D (MoCap or Kinect).
The fully-manual coaching tools require human coaches to either manually annotate on video playbacks to suggest improvements [48, 37], or analyze data of running gaits captured by MoCap [45].
MotionPro [35] supports manual selection of keypoints on each of the video frames such that some quantities, such as ball trajectory and 2D angles, can be obtained to facilitate analysis.
Kinovea [22] and OnForm [37] further simplify the manual tracking by providing basic processing of videos (e.g., automatically track objects and estimate 2D human poses).
On the automatic (right) side of the spectrum, a few video-based coaching tools assess the movements based on the reconstructed 2D poses from videos using embedded rules for a specific type of sports, such as skiing (AI Coach) [50] and yoga [6].
Such systems would require extensive domain knowledge to design.
To avoid bespoke knowledge, some systems compute suggestions based on the comparisons between novices’ actions with experts’ reference actions.
For example, MotionMA [49] and ReactiveVideo [11] align the experts’ poses captured by Kinect onto the novices’ poses in videos to visualize the difference in postures.
AIFit [13] mines and highlights the most significantly different features from the comparisons of reconstructed 3D poses from videos.
Even though AIFit is fully automatic, the dominant differences might not reflect informative feedback to the sport.
VCoach closes the gap in both dimensions in this spectrum:
the input is monocular videos such that it removes the constraint of indoor controlled environments, but it analyzes and visualizes in 3D to ensure spatial awareness.
It automatically performs low-level tasks but allows users the controllability to introduce high-level bespoke knowledge to the system.
Video-based Sports Data Analysis
A previous work [8] has classified general video-based sports data analysis into four levels: image level, object level, event level, and tactic level.
We adopt the same taxonomy as that in [8] and review w.r.t. video-based human pose data in sports.
Image-level analysis mainly includes video effects, such as slow-motion playback and displaying frames side-by-side [48].
Image-level analysis does not involve further image understanding from video frames, and thus the image contents would need to be analyzed manually (e.g., by a human coach).
Object-level mainly includes obtaining parameters of a single human instance, such as human pose estimation [50] and motion tracking [29, 22].
In sports videos object-level analysis is often more challenging than that in ordinary videos due to motion blurs, large subject displacements and complex sports poses (e.g., high diving).
Prior studies addressing these challenges include adopting sports motion priors [7], collecting sports motion datasets [40], and capturing human motions with multi-modal references [18].
Event-level analysis mainly includes recognition tasks from video streams, such as action recognition [40], action quality assessment [27], and key frame detection [54].
Tactic-level is mainly involved in ball games, such as soccer [43], table tennis [8], and basketball [3], by parsing the movements of athletes and objects from videos.
VCoach performs object-level analysis, but it focuses on local pose attributes rather than whole-body poses.
User-Customizability in UI
The goal of promoting user-customizability is to generalize to new instances other than those embedded in the systems, without requiring end-users’ explicit programming.
For example, in gesture recognition, a few systems, such as KinectScript [36] and Visual Gesture Builder [33] allow users to interactively define gestures by recording a few repetitions.
MotionMA [49] and YouMove [1] allow users to define movements via Programming by Demonstration (PbD).
Besides gestures and movements instances, other finer analysis tasks involve users’ specification of which body part(s) to analyze.
A medical research analysis tool, DeepLabCut [32], allows manual labeling of body parts across animal species for training data-driven models.
Kinovea [22] and RealitySketch [44] allow users to manually select points to track on top of videos, and customized joint angles can be further computed from the tracked points.
While such keypoint definitions apply to a specific video, in this work we develop a systematic set of mappings for users to customize reusable human pose biomechanics across videos.
3 Formative Study
At the beginning of this project we set out to decide the directions and the scope of a sports coaching system suitable for amateurs, which include but are not limited to runners.
We conducted a survey on potential target users to understand their usual ways of obtaining feedback on posture correctness in practising sports (subsection 3.1).
We also interviewed three experts on human locomotion to inform our design (subsection 3.2).
The results of this formative study form a set of design requirements for our system (subsection 3.3).
3.1 Target User Survey
To investigate the demands of potential target users (amateur sports players), we conducted a survey via the Amazon Mechanical Turk (MTurk).
We designed a questionnaire with three questions: (1) “What sport(s) do you frequently practise?” (2) “Have you paid attention to the correctness of your body postures while practising the sport(s)?” (3) “If yes, please describe how you get feedback on the correctness of your postures; if not, please explain why not.”
We distributed 120 questionnaires in total, and filtered out obvious spam responses according to the quality of the short answers to question (3).
Eventually 70 effective answers were collected.
Figure 3 shows the summaries of responses.
Among the responses, jogging/running accounts for the most, followed by football.
Other mentioned sports include those involving posture correctness, such as yoga and swimming.
24.3% of the subjects said they only depended on learned instructions of the actions but obtained no feedback; 21.4% of respondents stated that they got feedback from a coach or peers.
Other main feedback includes:
5.7% used outcome (e.g., score) as an indicator of posture correctness,
15.7% used feeling (e.g., tense on lower back) as an indicator,
and 8.6% adopted extra training on postures.
One respondent said he/she video-recorded the actions when practising gymnastics, and two responses explicitly said that they did not get any feedback since no one was watching.
Through this survey we learned that the public has the awareness of the importance of maintaining good postures, and there is a need for accessible posture analysis tools.
Based on the survey results, we set the focus of our system to jogging, due to its popularity and the requirement on correct postures to avoid injuries, without needing to consider ball/racket trajectories for instrument sports or tactics for team sports.
3.2 Expert Interviews
In order to understand the process and the key factors of human movement analysis, we conducted semi-structured interviews with three experts, two were medical doctors in Sports Medicine working in a hospital (E1, E2), and the other one (E3) was a researcher in Sports Science in a startup company studying performance analysis in sport.
During the interviews we first invited the participants to describe a representative case in which human movement analysis is involved in their daily practice.
During the description, they were asked to identify what is the routine they analyze human movements, what are the key factors they focus on, and what is the decision process based on their observations.
Then we raised open questions such as difficulties in human movement analysis, and the role of video-based analysis in practice.
All of the three experts mentioned that human movement analysis is based on gold standards, i.e., comparisons with the normal values in rehabilitation exercises or with top athletes’ postures and performances in sports.
Even for a full-body movement only a few key factors are concerned in evaluation (deterministic models [17]).
For example, E1 described a case of imbalance testing, where the key factors were movement accuracy and time required for completion.
E3 emphasized the advantage of externally-focused training over internally-focused training [53].
He pointed out that even though real-time feedback provides direct guidance, it would distract a subject during the action by interfering the subject’s intention of movements.
He also mentioned that since a coach’s attention is limited, he/she often can only focus on a specific body part during instruction, and that it would be ideal to analyze other parts during playback.
Since our system is focused on running, throughout the project we closely worked with E3 and another expert (E4), a third-year postgraduate student in Sports Science, who was involved after this formative study.
We initiated discussions with them as needed via remote chats.
3.3 Design Requirements
From the expert interviews on the human movement analysis, as well as the limitations of existing systems, we identify the following design requirements:
R1 - The tool should be accessible to users without an expert.
The potential users of our system might have no domain knowledge to determine the posture correctness directly from their videos.
This can be mitigated by comparing their videos with another video involving standard running poses from a professional runner and learning from the differences.
Our system should not only include key factors to running, but should also allow users to easily introduce other key factor(s) in case needs arise, instead of embedding redundant bespoke knowledge of running in the system.
Our system should be as easy to use as possible for novice users.
R2 - The comparison should adapt to variations.
The videos input by users may contain large variations on the running poses, due to viewpoints and subjects’ physical characteristics.
The comparison should be able to factor out these interferences and focus on only factors that indicate running posture incorrectness.
R3 - The visualization should highlight the part-based differences.
As pointed out by E3, the attention of both coaches and athletes is limited, they are often advised to correct one part at a time.
Thus instead of showing all the mistakes at the same time, our system should show the differences in each body part separately.
E3 also mentioned that for both coaches and athletes the quantitative figures do not make sense; they desire a direct corrective suggestion.
Thus instead of presenting analysis results as infographics, we need to design an intuitive way to demonstrate the differences.
R4 - The system should enable user interactivity.
As suggested by E4 in a later discussion, when a coach corrects an action, he/she usually first points out the mistakes, and then shows the correct action.
Our system should also follow this routine.
Following the design requirement R1, since there is no remote coach explaining the results, our system should allow users to explore the feedback to make the most sense out of it.
4 System Overview
We design our system VCoach based on the aforementioned requirements.
Since we target novice users, the overall system workflow follows the “overview first, details-on-demand” principle [41].
Users input videos and preview suggestions through the user interface (VCoach: A Customizable Visualization and Analysis System for Video-based Running Coaching).
The input to our system contains two videos (VCoach: A Customizable Visualization and Analysis System for Video-based Running Coaching(a)): a sample running video to be analyzed, and an exemplar running video for comparison (R1).
Upon loading the two videos, our system automatically processes the videos to reconstruct 3D human poses, normalizes the motions (R2), and segments the videos into running cycles.
Our system then performs the pose analysis by aligning the sample and exemplar running pose sequences based on 3D pose similarity, and retrieves the pre-defined key attributes to conduct comparisons (R1).
The suggestions for correction are generated based on the part-based differences from the comparison (R3), and directly reflect on a timeline tailored for running pose sequences (VCoach: A Customizable Visualization and Analysis System for Video-based Running Coaching(c)).
Those attributes that require improvement are represented with glyphs.
By clicking on each glyph on the timeline (R4), a detailed instruction for improving the corresponding attribute is shown as a short 3D animation of a body part on a human model in the suggestion preview window (VCoach: A Customizable Visualization and Analysis System for Video-based Running Coaching(d)).
Users can rotate the body model to navigate through viewpoints for better perception (R4).
For other pose attributes that are not embedded in our system as pre-defined attributes, the users can interactively label (R4) on a 3D body model via the query editor (VCoach: A Customizable Visualization and Analysis System for Video-based Running Coaching(e)).
The labeled attributes will then be retrieved and analyzed from the videos in the same way as the pre-defined attributes.
Our system contains five modules, as shown in Figure 4.
5 Visual Design
In this section we first describe the formulation of the data (attributes of running poses) we study to design our system.
Then we propose three mappings based on the data, i.e., the representation of data, the visualization of their differences, and the user operations to interactively define the attribute of each type.
5.1 Data Attributes
The data attributes in our system include both pre-defined attributes that are commonly used for evaluating running poses, and user-defined attributes for their customized analysis.
To determine the common data attributes for running pose correction, we collected a corpus of running pose tutorials by searching with key words “running pose tutorials”, “running pose corrections”, “running techniques”, “running form”, etc., from Google and YouTube.
The current corpus contains 55 items (including 37 videos and 18 articles).
The data attributes are summarized from the corpus into four types, as shown in Figure 5.
We conducted another interview with E4 to verify the coverage of these attributes in running pose evaluation in practice.
The fourth type “categorical data” is different from the previous three in that they are not computed from comparison with exemplar poses, but computed directly based on the other three classes (i.e., first compute a value and then discretize it into a category by a certain threshold).
Thus we focus on the design for the first three types, but support the visualization of the categorical data for commonly evaluated attributes in running.
5.2 Representing attributes
In this section we summarize the visual encoding of the positional, angular, and temporal attributes.
Positional attributes (Figure 6(a)) are defined as the relative distance between two points (classified as type P1), or the position of a point from a specific axis (P2).
For example, the trajectory of the wrist is its relative distance to the body center (P1).
Another example is the knee lift, which is a vertical distance from the knee joint to the body center (P2).
Angular attributes (Figure 6(b)) are defined as either the angle formed by three endpoints (classified as type A1), or the orientation of a vector formed by two joints with respect to an axis (A2).
For example, the elbow angle (A1) is an angle formed by the shoulder, the elbow and the wrist joint.
The leaning of the upper body (A2) is the orientation of the vector pointing from the root joint to the neck joint w.r.t. the z-axis.
Temporal attributes are defined as either a single moment (T1) or a time range within a running cycle (T2).
We use a temporal axis to show the temporal context.
The temporal axis (Figure 6(c)) is a fixed full running cycle, with the three dots from left to right respectively corresponding to the states of right foot landing ($RL$), left foot landing ($LL$), and right foot landing for the next cycle.
The positioning of the human center on the temporal axis reflects the state of the current pose within the running cycle.
5.3 Visualizing attribute differences
This section introduces the representation of the differences in data attributes.
Such differences are mainly used for presenting feedback, i.e., from an incorrect configuration to a correct one.
We define a set of visuals for attribute differences (Figure 6(d)), which are unified with the attribute representation.
Positional difference is shown by two points and an arrow pointing from the wrong position to the correct position.
Angular difference is shown by two vectors forming a wedge to show an angular difference.
Temporal difference is represented by a red marker segment on the temporal axis showing a temporal offset.
For example, the red segment along the forward temporal axis direction indicates the current event should appear later.
5.4 User operations
In this section we introduce the user operations (Figure 6(e)) for defining their own data attributes under the three data attribute classes.
Specifically, the query editor in our user interface (VCoach: A Customizable Visualization and Analysis System for Video-based Running Coaching(e)) contains a 3D viewer presenting the 3D human body model in T-pose, radio buttons for specifying properties and two draggable cursors (red lines) on top of a running cycle diagram for specifying timings.
A user may either refer to the mesh or skeleton of the body model and directly mouse-click on the body model to select joints; our system will snap the mouse click to the nearest joint.
A user first selects the attribute type by selecting either the angle button or distance button for angular and positional attributes, respectively, or directly dragging the temporal cursors for a temporal attribute.
To edit a positional attribute, a user first specifies the joint to track, and then specifies the base point (P1).
When the user further selects an axis, only the component of the selected dimension will be recorded (P2).
To edit an angular attribute, a user either selects three endpoints in order on the body model (A1), or two points and one axis (A2).
To edit a temporal attribute, the user either moves one cursor to specify a moment (T1), or both cursors to specify a time range (T2).
Our system will record a phase or a phase range accordingly.
When the positional and angular attributes are associated with an event, the user also moves the temporal cursor to specify the timing.
Please refer to the demo video for the authoring process of “left foot landing position” example.
5.5 Design of Attributes Overview
In this section we discuss the design of the overview for the problems reflected from the comparison.
The overview should show which attributes appear in question in the sample video and their timings.
We thus propose to use glyphs for representing attributes and a timeline tailored for running to organize them temporally.
Glyphs
We designed two types of glyphs for the four classes of the attributes, namely suggestion glyphs and profile glyphs.
Suggestion glyphs are icons for each of the three classes of attributes in Figure 6, i.e., positional, angular and temporal attributes in the collected corpus, whose values are continuous variables and are compared with those in the exemplars.
As shown in Figure 7(a-c), the suggestion glyphs are designed based on the idea of traffic signs that augment markers to symbols, such that users do not need to memorize the encoding, but can easily get familiar with the meaning of the icons and can interpret the meaning by intuition.
The profile glyphs are used to represent categorical attributes which do not need comparison with the exemplar.
We adopt the idea from the dance notations [34] to discretize complex human movements into reference planes (sagittal,
frontal and horizontal).
As shown in Figure 7(d), we use three transverse planes that capture the joints with a large degree of freedom, i.e., foot, knee, and shoulder.
Then the motions of these joints in relation to the body center are reflected by their projections into the three planes.
For example, by referring to the projection of wrists, users gain an intuitive profile of whether the wrists cross the body’s middle line in front of the chest.
In the transverse plane for feet, beyond showing the relative landing position to the body center, the triplet stacked squares further show the strike mode (fore-foot, mid-foot or rear-foot strike) of each foot by highlighting one of the blocks at the corresponding position.
Timeline
A characteristic of a running pose attribute sequence is that it is temporally periodical, and each period can be divided into a right-phase and a left-phase.
Based on this characteristic, we propose to design a timeline that transforms the temporal space into a running event space.
As shown in VCoach: A Customizable Visualization and Analysis System for Video-based Running Coaching(c), the horizontal axis is a complete running cycle, and the vertical axes correspond to the attributes of the left side of the body, right side of the body, and middle, respectively.
All the data attributes are summarized among cycles to be shown on the timeline.
Our system will automatically select significant errors, with the sizes of the glyphs proportional to the significance of the errors of a particular type.
We have conducted a pilot study to verify the above design against their alternatives.
For glyph design, the alternatives include a set of simplified icons highlighting the body parts in question, and color and shape encoding.
For timeline design, the alternatives are an ordinary linear timeline of a video that is not segmented into running cycles, and a spiral timeline displaying all running cycles without summarizations.
We invited two users, both of them are novices to running, and one of them with design background.
We introduced the overall function of our system along with the two sets of designs, and then let them vote on which representation they prefer.
Both of them chose the semantic glyph and the aggregated timeline, because they thought the semantic icons are intuitive and can be easily remembered.
As novice users they do not desire all the occurrences of the problems, but rather what kinds of problems appear in their running; thus the aggregated timeline is more preferable.
6 Data Analysis Model
In this section we introduce the methods of the backend modules in VCoach (Figure 4): video processing, pose analysis, and feedback.
6.1 Video Processing
3D Pose Reconstruction and Normalization
When the sample and the exemplar videos are loaded into the system, the pose at each frame is retargeted onto the SMPL models, denoted as $M_{s}$ for the sample video and $M_{e}$ for the exemplar video.
The retargeting (reconstruction) is implemented with TCMR [9], which is a monocular pose reconstruction method achieving state-of-the-art accuracy on challenging outdoor video datasets.
$M_{s}$ and $M_{e}$ are then rotated to a unified global orientation to facilitate comparison (VCoach: A Customizable Visualization and Analysis System for Video-based Running Coaching(a)).
The video frames are cropped to maximize their preview in windows.
Temporal Segmentation
The running pose sequences in both the sample and exemplar videos are segmented by the key frames of foot landing and foot extension.
Since the action of running is periodical, we adopt the phase variable of human locomotion, as in [20].
A full running cycle thus contains four key phases, in “right foot landing” ($phase=0$), “right foot extension” ($phase=0.25$), “left foot landing” ($phase=0.5$), and “left foot extension” ($phase=0.75$) order.
These four key phases are detected from the local extrema of the foot trajectories.
6.2 Pose Analysis
Sequence Alignment
Given the detected key phases, the running pose sequences in the sample and exemplar videos are first temporarily aligned at key phases, and then aligned at a finer level between each two key phases using the dynamic time warping technique [2].
We use joint rotations to measure human pose similarity [30].
Attributes Retrieval
Each data attribute is represented as a meta data tuple: $\left[name,type,J_{A},J_{o},J_{B},axis,side,phase\right]$,
where $J_{A},J_{o},J_{B}$ are the joint IDs of the attribute endpoints in the body model (as shown in Figure 6);
$side$ is one of the “left”, “neutral” and “right”;
$axis$ and $phase$ are the related axis and timing of the attribute; they are left empty if not applicable.
For the attributes embedded in VCoach (Figure 5) the meta tuples are pre-defined.
For customized attributes, the meta is formed from users’ input from the query editor.
Our attribute retrieval program parses the meta tuple and outputs retrieved values from the videos.
The retrieved values are then used for comparison.
Comparison
Since different attributes have different scales and units, we normalize the attribute values to the range $\left[0,1\right]$.
Then the differences in the attribute values are computed as the relative errors between the attributes from the sample video and those from the exemplar video.
We set a threshold of $25\%$ to select the significantly different attributes and scale the sizes of the suggestion icons according to the relative errors.
6.3 Feedback
Animation-based Demonstration
The corrective suggestion from pose comparison is conveyed by animating a 3D human model.
To make the demo easily understandable, the animation follows the design guideline as data-GIF [42].
The animation contains two key frames corresponding to the wrong pose and the same pose with a specific body part in the position as the exemplar pose, respectively.
Specifically, we use the joint rotations to drive the model: for angular attributes, the intermediate frames are interpolated with the joint rotations of $J_{o}$; while for positional attributes, the animation is interpolated with the joint rotations of the parent joint of $J_{o}$ along the kinematics tree.
The 3D animations are augmented with visual markers to highlight differences, as in Figure 6(b).
Suggestive Viewpoint
Since the animation of corrective suggestion is in 3D, we would like to demonstrate it at the most informative viewpoint.
While there are prior studies on the automatic selection of viewpoints for previewing a 3D mesh, the definition and criteria of the optimal viewpoints are often dependent on the purpose, such as to demonstrate region visual saliency [26], to set man-made models in upright orientation [14], and to incorporate modelers’ creation processes [5].
Previous studies on optimal viewpoints for human poses mainly include reducing prediction uncertainty in estimating 3D pose [21] and metrics defined over body part visibility [25].
In VCoach, since we would like to provide suggestions w.r.t. specific 3D local pose attributes, we develop a set of schemes to suggest viewpoints according to the geometry of the attributes.
The main idea is to minimize the ambiguity in the attributes due to camera projection, while preserving the human model as the spatial context.
Based on this goal, we make use of the normal vector formed by the 3D attributes to decide the orientation of the viewpoint (see Figure 9).
We further use the side of the body to determine whether to revert a normal to its opposite direction.
For example, to present an attribute on the right side of the body, the camera should also be placed to the right facing the body model.
The up direction of the viewpoint is along the average of the two vectors.
We also determine whether to revert the up direction according to whether it keeps the human model heading upwards.
Even though we present the 3D animation in the suggested viewpoint, users can still manually change the viewpoint to explore the corrective suggestion.
7 Results
In this section, we show the results of a user study evaluating the visualizations of posture correction feedback in VCoach and the baseline methods (subsection 7.1) for novices, and expert interviews (subsection 7.2) to evaluate the overall effectiveness of the system in pose correction.
7.1 User Evaluations
The main purpose of the user study is to evaluate the improvement of VCoach in promoting novices’ perception of running pose differences over existing methods (see Baselines).
It also evaluates the effectiveness of other components (e.g., viewpoint navigation and summarization of feedback) in assisting novices’ perceptions of running pose improvements.
Apparatus
VCoach was implemented with PyQt5 on a PC running Win10 (Intel x64 i5 [email protected], 8.00GB RAM).
Due to the current local COVID-19 regulation, the user study was conducted via Zoom with remote screen control.
Baselines
The baseline methods are visualizations of pose differences via juxtaposition and superposition, as shown in Figure 10.
We implement the baselines as follows.
For juxtaposition, we used the setup in [48] and put two running poses side-by-side.
To facilitate the preview, the two poses are cropped with subjects’ bounding boxes in videos, and the two videos are temporally synchronized using joint rotations.
For superposition, we adopted the method in [11].
Since [11] is based on Kinect, we transformed the 3D pose in a temporally correspondent exemplar frame and aligned it to the pose in the sample video frame at the body center, such that the temporally synchronized exemplar pose is overlaid on the sample video frame for comparison.
Participants
12 members from a local university were invited to participate in the user study (a1$\sim$a12, aged 23$\sim$32, 3 female).
Except for a1 and a7, all the other participants practise running more than once a week, but do not have access to professional coaches.
a12 stated that he was once curious about the correctness of his running poses and searched for mobile apps providing running pose checking functions but could not find a suitable one.
a2 focused on foot landing during running to avoid injuries; a6 used body senses after running as feedback.
a3, a10 and a11 said that they used mirrors during fitness workout, but obtained no feedback on pose correctness during running.
Task
We prepared 9 sample videos (V1$\sim$V9) covering all of the ten pre-defined attributes.
They were collected from running tutorial videos such that the ground-truth of the mistakes in running poses was known from the coaches’ comments in the videos, such as foot landing in front of the body (the braking position) and insufficient knee lift.
The difficulty level of the videos was controlled by containing only one main problem.
The general task for the participants was to explore the corrective feedback from videos using either VCoach or the baseline methods in a think-aloud manner, and complete a questionnaire afterwards.
The user study contained three sessions: two sessions using our system with and without the suggestive viewpoints, and one session using the baseline methods.
The order of the three sessions was counter-balanced, and the order of the nine videos was randomized among the three sessions (three videos for each session).
During training, we first gave a detailed tutorial on the operations of VCoach as well as the baseline system.
The participants then tried freely to get familiar with both systems.
In the session using VCoach without suggestive viewpoints (denoted as “VCoach-w/o”), we disabled the suggestive viewpoint function, and the participants would need to manually navigate the viewpoints to preview the 3D animations.
The system recorded the participants’ navigation activities in the suggestion preview window, parameterized by viewpoint azimuth and elevation, and the duration of each viewpoint.
In another session using VCoach (denoted as “VCoach”), the suggestive viewpoint function was enabled; the participants could also manually navigate, and their navigation activities were also recorded.
In the session using the baseline methods (denoted as “Baseline”), the participants explored the corrective feedback by comparing running poses in videos in either juxtaposition or superposition visualization.
After the sessions, the participants completed a designed questionnaire (Table 1) in a 7-point Likert Scale (1 is Strongly Disagree and 7 is Strongly Agree), and a standard System Usability Scale (SUS) [4].
The user study with each participant took about 90 minutes.
Effectiveness of Pose Difference Visualization
We first investigate the effectiveness of VCoach in presenting feedback compared with the baseline system.
Q10 explicitly asked the comparison between VCoach and the baseline methods, where 10 out of 12 participants strongly agreed that VCoach was more effective in conveying feedback than the baselines.
We recorded the time required to explore the running pose problem(s) in each video, as shown in Figure 12(a).
Paired t-tests on exploration time required for each video between sessions “VCoach” and “Baseline” showed that using VCoach with the suggestive viewpoint significantly requires less time to obtain the desired feedback ($p=0.019$).
However, there is no significance on exploration time between sessions “VCoach-w/o” and “Baseline” ($p=0.519$).
We evaluated the accuracy via the successful rate of the participants’ discovered mistakes matched the ground-truth mistakes as commented by the coaches in videos.
In sessions “VCoach-w/o” and “VCoach” the successful rate was $100\%$.
In other words, all the participants could figure out the problem(s) in the running poses with the visualization provided by VCoach.
In contrast, the successful rate was $77.8\%$ in session “Baseline”.
From the participants’ think-aloud in session “Baseline”, they often referred to the superposition visualization more than the juxtaposition visualization, especially when the subjects in the sample and exemplar videos are running in different directions.
For superposition in the baseline system, a6 and a8 said that they would refer to the lower limbs more often than upper limbs, since upper limbs were often occluded and misaligned due to differences in limb lengths.
Effectiveness of System Components
We then investigate the influence of specific design components on users’ perception of feedback on running pose correction.
Q6 asked the participants to rate the key component in VCoach, which visualizes pose differences via animations of local body parts on a human model.
8 out of 12 participants strongly agreed that such visualization was helpful for understanding, and the other four chose agreed.
The component that received the most disagreement is the preview of normalized poses from the sample and exemplar videos shown in juxtaposition (VCoach: A Customizable Visualization and Analysis System for Video-based Running Coaching(a) middle).
Since their orientations are often different from those in the original videos, the participants stated that referring to them increased the cognitive load by having to imagine the transformation to understand.
Thus even though normalized poses are crucial to computing pose differences, they do not necessarily contribute to users’ visual comparison.
During the participants’ think-aloud in sessions “VCoach-w/o” and “VCoach”, they often directly moved on to check the glyphs on the timeline after loading both videos.
After watching the animation, they sometimes checked the sample video frame to verify the problem.
At first they sometimes also referred to the exemplar frame to verify the animation, but many of them skipped the exemplar frame later because they found the corrective feedback illustrated by the animation was trust-worthy.
We also evaluated the usefulness of the design component of suggestive viewpoint.
We would like to figure out the following two questions: (1) do users find previewing the animations of pose correction under a certain viewpoint yields better perception?
(2) If yes, do our suggestive viewpoints match the preferred viewpoints selected by users?
We thus analyze the usage of viewpoint selection during the user study.
In session “VCoach-w/o”, the average number of times the participants manually changed the viewpoint was 7.36 times per video, compared with 2.05 times per video in session “VCoach”.
A paired t-test on the numbers of manual navigation between sessions “VCoach-w/o” and “VCoach” shows that enabling the suggestive viewpoint function significantly reduces users’ manual navigation ($p=0.00059$).
To answer question (2), we further analyze the relevance of the participants’ manually-selected viewpoints with the suggested viewpoints computed by our system in session “VCoach-w/o”.
We analyzed previewing viewpoints that lasted more than one second and considered those with a duration less than one second as the navigation process.
The average errors of azimuth and elevation relative to $360^{\circ}$ were $3.19\%$ and $4.52\%$, respectively, indicating a good match between our suggestive viewpoints and preferred viewpoints by the participants.
In the rating of the usefulness of suggestive viewpoint, seven participants chose “strongly agree”, and four of them explicitly stated during exploration that this function was very convenient.
a2 in session S1 asked whether the suggestive viewpoint function could be enabled, because she found this function especially useful when she was comparing the magnitudes of corrections on foot landing position.
a4 found the suggestive viewpoint more useful in observing upper limbs because they often suffer from heavier occlusions by the body torso than lower limbs.
Interestingly, a12 rated “Neutral” in Q9.
He explained that since he studied exoskeleton robotics, he was more used to imagining the attributes using the sagittal, coronal and transverse planes as reference, rather than using the human body as a spatial context.
Since VCoach targets at novice users without human movement analysis background, and most participants found the suggestive viewpoint function convenient, it can serve as a helpful option in VCoach.
System Usability
In the training session, all the participants could get familiar with VCoach within 5 minutes by completing a pipeline of operations, including loading videos, previewing frames and poses, and navigating on the timeline to preview animations of suggestions.
The SUS score for all the ten questions in the SUS questionnaire was 83.125 on average (SD: 10.56), out of a scale of
100, indicating the good usability of VCoach.
In post-study interviews with the participants, they commented favorably towards VCoach.
For example, a3: “Besides clarity, the summarization in VCoach helps me form a better impression of frequent mistakes. With VCoach I don’t even have to browse the entire video, but only need to refer to the frames the system has highlighted for me.”
The participants also commented on the potential generalization of VCoach in other scenarios.
Specifically, a11: “This tool is solving a very practical problem. I can see how it is useful in running and can imagine it generalizes to many other sports.”
a12 (from exoskeleton robotics background): “… current rehabilitation training often relies on wearable sensors to detect patients’ biomechanics, such as joint angular velocities and accelerations.
Such a video-based tool is promising in providing a non-invasive means to analyze patients’ movements.”
Evaluation of Query Editor
From the user study we also evaluate the easiness of use of the query editor, specifically, how efficiently and accurately users can edit a pose data attribute.
There is no baseline method for this task.
We chose three frequently used data attributes from each of the classes in the pre-defined attributes, and asked the participants to edit the attributes using the query editor in our interface.
The three attributes were: “foot landing position” (P2), “elbow angle” (A1) and “foot contact time” (T2).
They covered all the operations on the query editor.
The participants were given sample running video clips as references.
As shown in Figure 12(b), the average editing time for the three attributes were 95.36s ($SD=37.71$), 39.91s ($SD=10.11$) and 38.64s ($SD=14.03$).
On average the editing of the foot landing position took the longest time, since it required the most operations covering all the components on the query editor.
The successful rates that the participants can implement the same attribute as our pre-defined was $83.3\%$, $100\%$, and $91.7\%$, respectively.
In the failure cases, a3 failed the temporal attribute, because he misunderstood the question and labeled the time between two consecutive foot landings instead.
a4 and a10 both correctly annotated the positional attribute on the human model, but forgot to associate with the timing for foot landing by dragging the timeline cursor.
Through this experiment we verified that novice users could easily understand and implement the representative attributes with minimal training.
Even though for most amateur runners the pre-defined attributes would suffice, they can annotate their interested attributes via the query editor with reasonable efforts.
7.2 Expert Interviews
We conducted expert interviews to evaluate the overall usefulness of our system in helping amateur runners correct running poses.
Two experts with running backgrounds were invited: one was a licensed running coach (E5); the other was a professional marathon runner (E6).
The two interview sessions were conducted separately, and each session lasted 50 minutes.
During the interviews we provided a detailed introduction of functions in VCoach with three demonstrations of usage scenarios, and then invited them to try the system freely.
Both experts strongly agreed that VCoach would benefit a lot of runners.
E5: “Not only beginners, but experienced runners are also often bothered by the problems of running pose correctness. I can expect this tool serves a lot of runners.”
They also appreciated that the design rationale of VCoach is very reasonable for practical usage.
E5 said that coaching is a highly personalized process; and thus there is no absolute “correct” running pose regulated by numbers, such as the legal range of elbow angle in degree.
A significant advantage of the design of VCoach is that it does not directly classify a runner as right or wrong, but retains the flexibility to compare with various running poses to show the differences.
E5 thus finds VCoach especially useful for novices to iteratively adjust to different exemplars to find their most suitable poses.
E6 commented that the design of VCoach is similar to the idea of the “champion model” for elite athletes, such as
Su Bingtian, who was trained by shortening the gaps (on both poses and capabilities) with elite exemplars.
This comment is consistent with E3’s advice in the formative study.
We also invited experts to comment on the positioning of VCoach in training in real life.
E5: “It is suitable for the majority of ordinary runners.
But for severely over-weight people, asking them to resemble the running of ordinary people might cause injury instead of reducing it; they should seek for professional advice instead.”
E6 suggested that if the athletes’ parameters (mainly including height, leg lengths and training years) in the videos are accessible, it would be helpful to also suggest exemplars to users according to the similarity in these parameters, since runners with similar body configurations are more likely to have similar suitable running poses.
8 Conclusion and Future Work
We have presented a novel system, VCoach, for assisting amateur runners in improving their running poses.
We designed the system based on the design requirements formed from the literature research and expert interviews.
VCoach embeds common running pose attributes based on a collected corpus, and also provides an interface for users to customize attributes.
VCoach analyzes the poses from a sample video and an exemplar video in 3D, and visualizes the pose differences via 3D animations on a human body model.
Our user study showed that demonstrating pose corrective feedback via 3D animations is more effective than displaying frames side-by-side or overlaying the correct poses onto the sample frames.
There are several limitations and possible future work directions for VCoach.
In the current setting the running pose attributes are analyzed and visualized independently.
But there are certain correlations among the attributes, e.g., a higher knee lift might yield a larger stride.
A potential improvement is to incorporate human body harmonics [23, 19] to further summarize the problematic attributes.
Besides, in our user study we mainly evaluated the effectiveness of the visualization in VCoach in providing intuitive pose correction feedback.
It would be meaningful to conduct a long-term user study with participants from running backgrounds to further evaluate the effectiveness of VCoach in promoting running forms in practice.
Finally, currently VCoach focuses on the kinematics measurements (e.g., angles and positions).
However, more professional analysis [51] would require kinetics measurements, such as ground reaction force (braking force) [55] and muscle elastic energy [39].
Since the measure of kinetics parameters is currently limited to biomechanics laboratories, developing methods that recover the kinetics from videos would increase accessibility to many fields, including but not limited to sports posture analysis.
References
[1]
F. Anderson, T. Grossman, J. Matejka, and G. Fitzmaurice.
Youmove: enhancing movement training with an augmented reality
mirror.
In Proceedings of the 26th annual ACM symposium on User
interface software and technology, pp. 311–320, 2013.
[2]
D. J. Berndt and J. Clifford.
Using dynamic time warping to find patterns in time series.
In KDD workshop, vol. 10, pp. 359–370. Seattle, WA, USA:,
1994.
[3]
V. Bettadapura, C. Pantofaru, and I. Essa.
Leveraging contextual cues for generating basketball highlights.
In Proceedings of the 24th ACM international conference on
Multimedia, pp. 908–917, 2016.
[4]
J. Brooke et al.
Sus-a quick and dirty usability scale.
Usability evaluation in industry, 189(194):4–7, 1996.
[5]
H.-T. Chen, T. Grossman, L.-Y. Wei, R. M. Schmidt, B. Hartmann, G. Fitzmaurice,
and M. Agrawala.
History assisted view authoring for 3d models.
In Proceedings of the SIGCHI Conference on Human Factors in
Computing Systems, pp. 2027–2036, 2014.
[6]
H.-T. Chen, Y.-Z. He, and C.-C. Hsu.
Computer-assisted yoga training system.
Multimedia Tools and Applications, 77(18):23969–23991, 2018.
[7]
X. Chen, A. Pang, W. Yang, Y. Ma, L. Xu, and J. Yu.
Sportscap: Monocular 3d human motion capture and fine-grained
understanding in challenging sports videos.
International Journal of Computer Vision, 129(10):2846–2864,
2021.
[8]
Z. Chen, S. Ye, X. Chu, H. Xia, H. Zhang, H. Qu, and Y. Wu.
Augmenting sports videos with viscommentator.
IEEE Transactions on Visualization and Computer Graphics,
28(1):824–834, 2022. doi: 10 . 1109/TVCG . 2021 . 3114806
[9]
H. Choi, G. Moon, J. Y. Chang, and K. M. Lee.
Beyond static features for temporally consistent 3d human pose and
shape from a video.
In Proceedings of the IEEE/CVF Conference on Computer Vision and
Pattern Recognition, pp. 1964–1973, 2021.
[10]
X. Chu, X. Xie, S. Ye, H. Lu, H. Xiao, Z. Yuan, Z. Chen, H. Zhang, and Y. Wu.
Tivee: Visual exploration and explanation of badminton tactics in
immersive visualizations.
IEEE Transactions on Visualization and Computer Graphics,
28(1):118–128, 2021.
[11]
C. Clarke, D. Cavdir, P. Chiu, L. Denoue, and D. Kimber.
Reactive video: Adaptive video playback based on user motion for
supporting physical activity.
In Proceedings of the 33rd Annual ACM Symposium on User
Interface Software and Technology, pp. 196–208, 2020.
[12]
V. Colman and U. Persyn.
Kinematics analysis of a new straight line skating technique.
In ISBS-Conference Proceedings Archive, 2000.
[13]
M. Fieraru, M. Zanfir, S. C. Pirlea, V. Olaru, and C. Sminchisescu.
Aifit: Automatic 3d human-interpretable feedback models for fitness
training.
In Proceedings of the IEEE/CVF Conference on Computer Vision and
Pattern Recognition, pp. 9919–9928, 2021.
[14]
H. Fu, D. Cohen-Or, G. Dror, and A. Sheffer.
Upright orientation of man-made objects.
In ACM SIGGRAPH 2008 papers, pp. 1–7. 2008.
[15]
M. Gleicher, D. Albers, R. Walker, I. Jusufi, C. D. Hansen, and J. C. Roberts.
Visual comparison for information visualization.
Information Visualization, 10(4):289–309, 2011.
[16]
M. Habermann, W. Xu, M. Zollhöfer, G. Pons-Moll, and C. Theobalt.
Livecap: Real-time human performance capture from monocular video.
ACM Trans. Graph., 38(2), Mar. 2019. doi: 10 . 1145/3311970
[17]
J. Hay.
The biomechanics of sports techniques.
Prentice-Hall, 1978.
[18]
Y. He, A. Pang, X. Chen, H. Liang, M. Wu, Y. Ma, and L. Xu.
Challencap: Monocular 3d capture of challenging human performances
using multi-modal references.
In Proceedings of the IEEE/CVF Conference on Computer Vision and
Pattern Recognition, pp. 11400–11411, 2021.
[19]
R. N. Hinrichs.
Whole body movement: coordination of arms and legs in walking and
running.
Multiple muscle systems, pp. 694–705, 1990.
[20]
D. Holden, T. Komura, and J. Saito.
Phase-functioned neural networks for character control.
ACM Transactions on Graphics (TOG), 36(4):1–13, 2017.
[21]
S. Kiciroglu, H. Rhodin, S. N. Sinha, M. Salzmann, and P. Fua.
Activemocap: Optimized viewpoint selection for active human motion
capture.
In Proceedings of the IEEE/CVF Conference on Computer Vision and
Pattern Recognition, pp. 103–112, 2020.
[22]
Kinovea.
A microscope for your videos.
https://www.kinovea.org/, 2012.
[23]
J. P. Kuhtz-Buschbeck and B. Jing.
Activity of upper limb muscles during human walking.
Journal of Electromyography and Kinesiology, 22(2):199–206,
2012.
[24]
G. Kurillo, R. Bajcsy, K. Nahrsted, and O. Kreylos.
Immersive 3d environment for remote collaboration and training of
physical activities.
In 2008 IEEE Virtual Reality Conference, pp. 269–270. IEEE,
2008.
[25]
B. Kwon, J. Huh, K. Lee, and S. Lee.
Optimal camera point selection toward the most preferable view of 3-d
human pose.
IEEE Transactions on Systems, Man, and Cybernetics: Systems,
52(1):533–553, 2022. doi: 10 . 1109/TSMC . 2020 . 3004338
[26]
C. H. Lee, A. Varshney, and D. W. Jacobs.
Mesh saliency.
In ACM SIGGRAPH 2005 Papers, pp. 659–666. 2005.
[27]
Y. Li, X. Chai, and X. Chen.
Scoringnet: Learning key fragment for action quality assessment with
ranking loss in skilled sports.
In Asian Conference on Computer Vision, pp. 149–164. Springer,
2018.
[28]
T. Lin, R. Singh, Y. Yang, C. Nobre, J. Beyer, M. A. Smith, and H. Pfister.
Towards an Understanding of Situated AR Visualization for
Basketball Free-Throw Training.
2021.
[29]
J. Liu, P. Carr, R. T. Collins, and Y. Liu.
Tracking sports players with context-conditioned motion models.
In Proceedings of the IEEE Conference on Computer Vision and
Pattern Recognition, pp. 1830–1837, 2013.
[30]
J. Liu, M. Shi, Q. Chen, H. Fu, and C.-L. Tai.
Normalized human pose features for human action video alignment.
In Proceedings of the IEEE/CVF International Conference on
Computer Vision, pp. 11521–11531, 2021.
[31]
M. Loper, N. Mahmood, J. Romero, G. Pons-Moll, and M. J. Black.
SMPL: A skinned multi-person linear model.
ACM Trans. Graphics (Proc. SIGGRAPH Asia), 34(6):248:1–248:16,
Oct. 2015.
[32]
A. Mathis, P. Mamidanna, K. M. Cury, T. Abe, V. N. Murthy, M. W. Mathis, and
M. Bethge.
Deeplabcut: markerless pose estimation of user-defined body parts
with deep learning.
Nature neuroscience, 21(9):1281–1289, 2018.
[33]
Microsoft.
Visual gesture builder (vgb), 2017.
[34]
E. Mirzabekiantz.
Benesh movement notation for humanoid robots?
In Dance Notations and Robot Motion, pp. 299–317. Springer,
2016.
[35]
MotionPro.
Motion analysis software for all sports, 2018.
[36]
M. Nebeling, D. Ott, and M. C. Norrie.
Kinect analysis: A system for recording, analysing and sharing
multimodal interaction elicitation studies.
In Proceedings of the 7th ACM SIGCHI Symposium on Engineering
Interactive Computing Systems, EICS ’15, p. 142–151. Association for
Computing Machinery, New York, NY, USA, 2015. doi: 10 . 1145/2774225 . 2774846
[37]
OnForm.
Video analysis for skill development in any sport, 2021.
[38]
V. Panoutsakopoulos and I. Kollias.
3-d biomechanical analysis of women’s high jump.
New Studies in Athletics, 27(3):31–48, 2012.
[39]
A. Seth, J. L. Hicks, T. K. Uchida, A. Habib, C. L. Dembia, J. J. Dunne, C. F.
Ong, M. S. DeMers, A. Rajagopal, M. Millard, et al.
Opensim: Simulating musculoskeletal dynamics and neuromuscular
control to study human and animal movement.
PLoS computational biology, 14(7):e1006223, 2018.
[40]
D. Shao, Y. Zhao, B. Dai, and D. Lin.
Finegym: A hierarchical video dataset for fine-grained action
understanding.
In Proceedings of the IEEE/CVF conference on computer vision and
pattern recognition, pp. 2616–2625, 2020.
[41]
B. Shneiderman.
The eyes have it: A task by data type taxonomy for information
visualizations.
In The craft of information visualization, pp. 364–371.
Elsevier, 2003.
[42]
X. Shu, A. Wu, J. Tang, B. Bach, Y. Wu, and H. Qu.
What makes a data-gif understandable?
IEEE Transactions on Visualization and Computer Graphics,
27(2):1492–1502, 2020.
[43]
M. Stein, H. Janetzko, A. Lamprecht, T. Breitkreutz, P. Zimmermann,
B. Goldlücke, T. Schreck, G. Andrienko, M. Grossniklaus, and D. A. Keim.
Bring it to the pitch: Combining video and movement data to enhance
team sport analysis.
IEEE transactions on visualization and computer graphics,
24(1):13–22, 2017.
[44]
R. Suzuki, R. H. Kazi, L.-y. Wei, S. DiVerdi, W. Li, and D. Leithinger.
Realitysketch: Embedding responsive graphics and visualizations in ar
through dynamic sketching.
In Proceedings of the 33rd Annual ACM Symposium on User
Interface Software and Technology, UIST ’20, pp. 166–181, 2020. doi: 10 . 1145/3379337 . 3415892
[45]
Q. System.
A leading provider of precision motion capture and 3d positioning
tracking system, 2015.
[46]
R. Tang, X.-D. Yang, S. Bateman, J. Jorge, and A. Tang.
Physio@home: Exploring visual guidance and feedback techniques for
physiotherapy exercises.
In Proceedings of the 33rd Annual ACM Conference on Human
Factors in Computing Systems, pp. 4123–4132, 2015.
[47]
H. Techiques.
Hudl: Performance analysis tools for sports teams and athletes at
every level, 2007.
[48]
TechSmith.
Coach’s eye.
https://www.coachseye.com, 2011.
[49]
E. Velloso, A. Bulling, and H. Gellersen.
Motionma: motion modelling and analysis by demonstration.
In Proceedings of the SIGCHI Conference on Human Factors in
Computing Systems, pp. 1309–1318, 2013.
[50]
J. Wang, K. Qiu, H. Peng, J. Fu, and J. Zhu.
Ai coach: Deep human pose estimation and analysis for personalized
athletic training assistance.
In Proceedings of the 27th ACM International Conference on
Multimedia, pp. 374–382, 2019.
[51]
R. X. Wei, Z. Y. Chan, J. H. Zhang, G. L. Shum, C.-Y. Chen, and R. T. Cheung.
Difference in the running biomechanics between preschoolers and
adults.
Brazilian Journal of Physical Therapy, 25(2):162–167, 2021.
[52]
P. W. Woundefinedniak, M. Zbytniewska, F. Kiss, and J. Niess.
Making Sense of Complex Running Metrics Using a Modified Running
Shoe.
2021.
[53]
G. Wulf, C. Shea, and R. Lewthwaite.
Motor skill learning and performance: a review of influential
factors.
Medical education, 44(1):75–84, 2010.
[54]
G. Yan and M. Woźniak.
Accurate key frame extraction algorithm of video action for aerobics
online teaching.
Mobile Networks and Applications, pp. 1–10, 2022.
[55]
D. Zecha, C. Eggert, M. Einfalt, S. Brehm, and R. Lienhart.
A convolutional sequence to sequence model for multimodal dynamics
prediction in ski jumps.
In Proceedings of the 1st International Workshop on Multimedia
Content Analysis in Sports, pp. 11–19, 2018. |
Pathways for virus assembly around nucleic acids
Jason D Perlmutter
Matthew R Perkett
Michael F Hagan
[email protected]
Martin Fisher School of Physics, Brandeis University,
Waltham, MA, USA.
\PrependGraphicsExtensions
*.pdf,.PDF
I Abstract
Understanding the pathways by which viral capsid proteins assemble around their genomes could identify key intermediates as potential drug targets. In this work we use computer simulations to characterize assembly over a wide range of capsid protein-protein interaction strengths and solution ionic strengths. We find that assembly pathways can be categorized into two classes, in which intermediates are either predominantly ordered or disordered. Our results suggest that estimating the protein-protein and the protein-genome binding affinities may be sufficient to predict which pathway occurs. Furthermore, the calculated phase diagrams suggest that knowledge of the dominant assembly pathway and its relationship to control parameters could identify optimal strategies to thwart or redirect assembly to block infection. Finally, analysis of simulation trajectories suggests that the two classes of assembly pathways can be distinguished in single molecule fluorescence correlation spectroscopy or bulk time resolved small angle x-ray scattering experiments.
II Introduction
In many virus families, the spontaneous assembly of a protein shell (capsid) around the viral genome is an essential step in the viral life cycle Hagan2014 . These families include most viruses with single stranded RNA (ssRNA) genomes, as well as the Hepadnaviridae (e.g. hepatitis B virus, HBV). Understanding the mechanisms which underlie this cooperative assembly process could facilitate efforts to develop anti-viral drugs that block or derail the formation of infectious particles (for reviews see Zlotnick2011 ; Prevelige2011 ) and promote efforts to reengineer them for biomedical delivery. In this article, we explore how the interactions between the molecular components determine the mechanism of assembly, and how these interactions can be altered by changing solution conditions or mutagenesis to modulate assembly pathways.
The most detailed knowledge of capsid-RNA interactions comes from structural analysis of assembled viral particles. Atomic resolution structures of capsids assembled around ssRNA have been obtained by x-ray crystallography and/or cryo-electron microscopy (cryo-EM) (e.g. Fox1994 ; Valegard1997 ; Johnson2004 ; Tihova2004 ; Krol1999 ; Stockley2007 ; Toropova2008 ; Lucas2002 ; Valegard1994 ; Worm1998 ; Grahn2001 ; Valegard1997 ; Helgstrand2002 ; Schneemann2006 ).
The packaged NAs are less ordered than their protein containers and thus have been more difficult to characterize. However cryo-EM experiments have identified that the nucleotide densities are nonuniform, with a peak near the inner capsid surface and relatively low densities in the interiorTihova2004 ; Zlotnick1997 ; Conway1999 . While atomistic detail has not been possible in these experiments, all-atom models have been derived from equilibrium simulations Freddolino2006 ; Devkota2009 ; Zeng2012 . In some cases, striking image reconstructions reveal that the packaged RNA adopts the symmetry of the overlying capsid (e.g. Tihova2004 ; Toropova2008 ; Schneemann2006 ; Bakker2012 ; Ford2013 ). While it has been proposed that this order arises as a function of the assembly mechanism for several viruses Larson2001 ; Dykeman2011 ; Dykeman2013 , computational analysis of polyelectrolyte configurations inside capsids also indicate that capsid-polymer interactions can generically drive spatial organization of the packaged polymer Siber2008 ; Hu2008c ; Devkota2009 ; Forrey2009 ; Harvey2009 ; Belyi2006 ; Angelescu2006 ; Zhang2004a ; Lee2008 ; Angelescu2008 ; Jiang2009 ; Elrad2010 ; Perlmutter2013 . Theoretical works have also characterized the relationship between the NA charge and structure and the length which is optimal for packaging Perlmutter2013 ; Schoot2013 ; Hu2008a ; Belyi2006 ; Schoot2005 ; Angelescu2006 ; Siber2008 ; Ting2011 ; Ni2012 ; Siber2010 ; Siber2012 .
In addition to this structural data on assembled capsids, an extensive combination of mass spectrometry, assembly kinetics experiments, constraints from assembled capsid structures, and mathematical modeling has delineated assembly pathways for several viruses, with a particular focus on the role of interactions between CPs and specific RNA sequences called packaging signals. Recent single molecule fluorescence correlation spectroscopy (smFCS) experiments indicate that, for these viruses, assembly around the viral genome is more robust and proceeds by a different mechanism as compared to around heterologous RNA Borodavka2012 . Yet, in other cases capsid proteins show no preference for genomic RNA over heterologous RNA (e.g. HBV Porterfield2010 ), and cowpea chlorotic mottle virus (CCMV) proteins preferentially encapsidate heterologous RNA (from BMV) over the genomic CCMV RNA with equivalent length Comas-Garcia2012 . Furthermore, experimental model systems in which capsid proteins assemble into icosahedral capsids around synthetic polyelectrolytes or other polyanions Hiebert1968 ; Bancroft1969 ; Dixit2006 ; Loo2006 ; Goicochea2007 ; Huang2007 ; Loo2007 ; Sikkema2007 ; Sun2007 ; Hu2008 ; Comellas-Aragones2009 ; Crist2009 ; Chang2008 ) demonstrate that specific RNA sequences are not required for capsid formation or cargo packaging.
Thus, a complete picture of capsid assembly mechanisms requires understanding how assembly pathways depend on those features which are generic to polyelectrolytes, as well as those which are specific to viral RNAs.
In previous work on assembly around a simple model for a polymer, Elrad and Hagan Elrad2010 proposed that mechanisms for assembly around a cargo (i.e. RNA, polymer, or nanoparticle) can be classified on the basis of two extreme limits. In the first (originally proposed by McPherson McPherson2005 and then Refs Hagan2008 ; Devkota2009 ), strong protein-cargo interactions drive proteins to adsorb ‘en masse’ onto the cargo in a disordered manner, meaning there are few protein-protein interactions. Once enough subunits are bound, subunits undergo cooperative rearrangements (potentially including dissociation of excess subunits) to form an ordered capsid. This mechanism has been observed in recent simulations Mahalik2012 ; Perlmutter2013 ; Elrad2010 ; Hagan2008 ; Perkett2014 . In the second limit, where protein-protein interactions dominate, a small partial capsid nucleates on the cargo, followed by a growth phase in which individual proteins or small oligomers sequentially add to the growing capsid. This class of pathways resembles the nucleation-and-growth mechanism by which empty capsids assemble Endres2002 , except that the polymer plays an active role by stabilizing protein-protein interactions and by enhancing the flux of proteins to the assembling capsid Hu2007 ; Kivenson2010 ; Elrad2010 .
It is difficult to determine assembly mechanisms directly from experiments, due to the small size ($\lesssim 10$ of nm) and transience ($\sim$ms) of most intermediates. Observations in vitro suggest that both mechanisms may be viable. Kler et al. Kler2012 used time-resolved X-ray scattering (trSAXS) to monitor SV40 capsid proteins assembling around ssRNA. The scattering profiles at all time points during assembly could be decomposed into unassembled components (RNA + protein subunits) and complete capsid; the absence of any signal corresponding to a large disordered intermediate suggests this assembly follows the nucleation-and-growth (ordered) assembly mechanism Kler2012 .
Other observations suggest that viruses can assemble through the en masse mechanism. Refs. Garmann2013 ; Cadena-Nava2012 found that in vitro assembly CCMV assembly was most productive when performed in two steps: (1) at low salt (strong protein-RNA interactions) and neutral pH (weak protein-protein interactions) the proteins undergo extensive adsorption onto RNA, then (2) pH is reduced to activate binding of protein-protein binding Garmann2013 . Similarly, a recent observation of capsid protein assembly around charge-functionalized nanoparticles found that assembly initially proceeded through nonspecific aggregation of proteins and nanoparticles, followed by the gradual extrusion of nanoparticles within completed capsids Malyutin2013 . These experiments used viral proteins with relatively weak protein-protein interactions (CCMV and BMV) Zlotnick2013 and moderate salt concentrations ($100-150$mM). The Kler et al. Kler2012 ; Kler2013 experiments considered SV40 proteins, which have strong protein-protein interactions Zlotnick2013 , and high salt ($250$mM) Together, these in vitro experiments suggest that productive assembly could proceed by either the en masse or nucleation-and-growth mechanism.
In this work, we use dynamical simulations to investigate the extent to which the assembly mechanism can be controlled by tuning solution ionic strength and protein-protein attractions. We extend a model that was recently used to calculate the thermostability and assembly yields of viral particles as a function of protein charge and nucleic acid (NA) length and structure. Those previous simulations found quantitative agreement between predicted NA lengths that optimize capsid thermostability and viral genome length for seven viruses Perlmutter2013 . Here, we perform extensive new simulations of assembly, in which protein-protein interactions, the sequence of charges in capsid protein-NA binding domains, and the solution ionic strength are varied. We find that, by varying these control parameters, the assembly mechanism can be systematically varied between the two extreme limits described above. Our results suggest that knowledge of protein-protein and protein-NA binding affinities may be sufficient to predict which assembly mechanism will occur, and we estimate relative protein-NA binding interactions for three viruses (based on non-specific electrostatic interactions). These findings suggest that assembly mechanisms can be rationally designed through choice of solution conditions and mutagenesis of capsid protein-protein interfaces and protein-NA binding domains.
Finally, by calculating hydrodynamic radii and SAXS profiles associated with assembly intermediates, we show that assembly mechanisms can be distinguished by experimental techniques recently applied to virus assembly, respectively smFCS Borodavka2012 ) and trSAXS Kler2012 . While the NA is represented by a linear polyelectrolyte in most of the simulations, we obtain qualitatively similar results when considering a model for base-paired NAs developed in Ref. Perlmutter2013 .
III Results
To study how capsid assembly around a polyelectrolyte depends on the strength of protein subunit-subunit and subunit-polyelectrolyte interactions, we performed Brownian dynamics simulations with a recently developed model Perlmutter2013 (Figure 1).
This model was motivated by the observation Kler2012 ; Kler2013 that purified simian virus 40 (SV40) capsid proteins assemble around ssRNA molecules to form capsids composed of 12 homopentamer subunits. The capsid is modeled as a dodecahedron composed of 12 pentagonal subunits (each of which represents a rapidly forming and stable pentameric intermediate, which then more slowly assembles into the complete capsid, as is the case for SV40 in vitro around ssRNA Li2002 ; Kler2012 ; Kler2013 ). Subunits are attracted to each other via attractive pseudoatoms at the vertices (type ‘A’) and driven toward a preferred subunit-subunit angle by repulsive ‘Top’ pseudoatoms (type ‘T’) and ‘Bottom’ pseudoatoms (type ‘B’) (see Fig. 1 and the Methods section). These attractions represent the interactions between capsid protein subunits that arise from hydrophobic, van der Waals, and electrostatic interactions as well as hydrogen bonding Hagan2014 ). The magnitude of these interactions varies between viral species Zlotnick2013 and can be experimentally tuned by pH and salt concentration Ceres2002 ; Kegel2004 ; Hagan2014 . Here, the attraction strength is controlled by the model parameter $\varepsilon_{\text{ss}}$. In order to relate this potential energy to the free energy of dimerization, we have run a separate series of calculations, where in the absence of cargo, we find that the free energy of subunit of dimerization is $g_{\text{ss}}{=}5.0-1.5*\varepsilon_{\text{ss}}$ (see SI section 12). Throughout this article energies are given in units of the thermal energy, $k_{\mathrm{B}}T$.
Capsid assembly around nucleic acids and other polyelectrolytes is primarily driven by electrostatic interactions between negative charges on the encapsulated polyelectrolyte and positive charges on the inner surfaces of capsid proteins Hagan2014 . We consider a linear bead-spring polyelectrolyte, with a charge of $âe$ per bead and a persistence length comparable to that of ssRNA in the absence of base pairing. Positive charges on capsid proteins are located in flexible polymers affixed to the inner surface of the model subunit, which represent the highly charged, flexible terminal tails known as arginine rich motifs that are typical of positive-sense ssRNA virus capsid proteins (e.g., Schneemann2006 ).
Electrostatics are modeled using Debye-Huckel interactions, where the Debye screening length ($\lambda_{\text{D}}$) is determined by the ionic strength $I$, equivalent to the concentration $C_{\text{salt}}$ of a monovalent salt, as $\lambda_{\text{D}}\approx 0.3/I^{1/2}$ with $\lambda_{\text{D}}$ in nm and $I$ in molar units. Perlmutter et al. Perlmutter2013 showed that Debye-Huckel interactions compare well to simulations with explicit counterions for the parameter values under consideration; furthermore, inclusion of multivalent ions at experimentally relevant concentrations had only a small effect on properties such as the optimal genome length. The effect of base-pairing on the optimal length was also considered in that reference Perlmutter2013 .
III.1 Kinetic phase diagram
We first consider the predominant assembly products (Fig. 2) and the packaging efficiency (Fig. 3) as a function of Debye length $\lambda_{\text{D}}$ and subunit-subunit interaction strength $\varepsilon_{\text{ss}}$. The packaging efficiency is defined as the fraction of trajectories in which the polyelectrolyte is completely encapsulated by a well-formed capsid, which contains 12 subunits each of which interact with 5 neighbors. We refer to this as a kinetic phase diagram Hagan2006 ; Hagan2014 because we characterize products at a finite observation time of $t_{\text{f}}{=}2\times 10^{8}$ time steps, which is long enough that assemblages do not vary significantly with time (except for under weak interactions, see below), but is not sufficient to equilibrate kinetic traps if there are large activation barriers Elrad2010 ; Hagan2014 . We see that for the range of simulated ionic strengths ($1\leq C_{\text{salt}}\leq 500$ mM or $10\geq\lambda_{\text{D}}\geq 0.4$ nm) assembly yields are highest for $\varepsilon_{\text{ss}}{=}5k_{\mathrm{B}}T$ and $C_{\text{salt}}{=}100$ mM (the parameter values focused on in Perlmutter2013 ), and that for moderate subunit-subunit interaction strengths ($4\leq\varepsilon_{\text{ss}}\leq 6,k_{\mathrm{B}}T$) yield to remain high as the ionic strength is increased to about 300 mM ($\lambda_{\text{D}}\approx 0.6$ nm). For higher salt concentrations, yields are depressed by the appearance of long-lived on-pathway intermediates. As will be discussed further below, weakening the electrostatic interaction between the polymer and protein limits the ability of the polymer to promote assembly. Although we expect that these simulations would eventually result in complete capsids, the low yield at our finite measurement time reflects the fact that assembly is less efficient than for lower salt concentrations. At the highest ionic strength considered ($C_{\text{salt}}{=}500$ mM), the most prevalent outcome is that no nucleation occurs. At lower salt concentrations ($C_{\text{salt}}\geq 10$ mM), rapid adsorption of a super-stroichiometric quantity of subunits results in malformed capsids.
At larger-than-optimal protein-protein interaction strengths ($\varepsilon_{\text{ss}}>6$) assembly yields are lower for two reasons. The first, and more frequent outcome, is long-lived malformed structures with strained interactions. This kinetic trap arises in a wide variety of assembly systems when interactions become strong in comparison to the thermal energy, because strained interactions are unable to anneal before additional subunits lock them in place Hagan2014 ; Hagan2011 ; Grant2011 . In our simulations, we found that these structures frequently result from an incorrect merger between two partial capsid intermediates; even when each individual interaction is too weak to lock-in the non-optimal structure, multiple erroneous interactions formed by two partial capsids are effectively irreversible on our timescale. The tendency for oligomer binding to lead to malformed structures was seen previously in the context of empty capsid assembly Hagan2006 ; Whitelam2009 . Here, the polymer helps to bring oligomers together, and thus this trap arises when nucleation on the polymer is faster than growth of a nucleus into a complete capsid. This trap resembles Geminivirus particles, which are composed of a union of two nearly complete capsids Bottcher2004 .
The second obstacle to polyelectrolyte encapsulation arises at the highest protein-protein interaction strengths studied ($\varepsilon_{\text{ss}}\geq 8k_{\mathrm{B}}T$), for which subunits not associated with the polyelectrolyte undergo spontaneous assembly. The resulting off-polyelectrolyte assembly depletes the pool of available monomers and small oligomers available for assembly on the polyelectrolyte, leading to a form of the monomer-starvation kinetic trap previously discussed for empty capsid assembly Hagan2014 ; Endres2002 . Triggering formation of empty capsids and thus preventing nucleic acid encapsidation by strengthening subunit-subunit interactions has been suggested as a mode of action for a putative antiviral drug for hepatitis B virus Katen2010 ; Katen2013 .
At smaller-than-optimal protein-protein interaction strengths ($\varepsilon_{\text{ss}}<4$) assembly is unsuccessful for two reasons, depending upon the ionic strength. At smaller values ($C_{\text{salt}}\leq 300$ mM), electrostatic interactions are relatively strong, and many proteins adsorb onto to the polymer. However, because of the weak protein-protein interaction strength, these proteins do not form stable capsids, predominantly because nucleation is slow in comparison to $t_{\text{f}}$ (the final observation time, $t_{\text{f}}{=}2\times 10^{8}$ time steps). In a minority of cases, a nucleus will form, but completion is prevented by the excess number of subunits adsorbed to the polyelectrolyte.
We refer to the resulting configurations as disordered, due to the lack of ordered binding between protein subunits. At larger ionic strengths ($C_{\text{salt}}>300$ mM) electrostatic interactions are relatively weak, and individual subunits rapidly desorb from the polyelectrolyte. In this regime, assembly requires a fluctuation in the number of adsorbed subunits which leads to nucleation of a partial capsid intermediate which has enough subunit-polyelectrolyte interactions to be stable against rapid desorption. The nucleation rate decreases exponentially with subunit-subunit interaction strength Hagan2014 ; Kivenson2010 , and thus most simulations at high salt and weak subunit-subunit interactions never undergo nucleation. (We categorize simulations with fewer than 3 subunits adsorbed to the polyelectrolyte and no subunit-subunit interactions (i.e. no progress towards assembly) as ‘unnucleated’.)
Importantly, we expect that trajectories in this region of parameter space will eventually undergo nucleation. Thus, as the finite assembly time $t_{\text{f}}$ is increased the region of successful assembly will expand to lower values of $\varepsilon_{\text{ss}}$ and higher ionic strength, until eventually reaching values below which capsid assembly is thermodynamically unstable (see Hagan2006 , Fig. 7). To confirm this possibility, we used the Markov state model (MSM) approach described in Ref. Perkett2014 to characterize assembly with $\varepsilon_{\text{ss}}{=}5,6$ and $C_{\text{salt}}{=}500$ mM (see Figs. 4C and S4 (in the SI)). In contrast, the malformed capsids encountered under large parameter values typically will not resolve on any relevant timescales, since numerous strong interactions would need to be broken Hagan2014 . Thus, the boundaries of successful assembly at large $\varepsilon_{\text{ss}}$ values are insensitive to $t_{\text{f}}$.
Several additional conclusions can be drawn from the variation of packaging efficiency (Fig. 3). Firstly, the yield of complete virus-like particles is relatively high for moderate subunit-subunit interaction strengths $\varepsilon_{\text{ss}}\in[4,6]$ across a range of ionic strengths ($C_{\text{salt}}\in[100,300]$ mM). Even above ($C_{\text{salt}}>300$ mM) and below ($C_{\text{salt}}{=}10$ mM) this range we observe moderate yields of complete particles. Secondly, as electrostatic interactions are weakened (moving to the left), the subunit-subunit interaction strength which optimizes the yield increases (i.e. from $\varepsilon_{\text{ss}}=5k_{\mathrm{B}}T$ at $C_{\text{salt}}=100mM$ to $\varepsilon_{\text{ss}}=6k_{\mathrm{B}}T$ at $C_{\text{salt}}:150-400mM$ to $\varepsilon_{\text{ss}}=7k_{\mathrm{B}}T$ at $C_{\text{salt}}=500mM$). This result suggests that one interaction type can compensate for the other within a limited range. However, though all successful capsids contain the same subunit geometry, the mechanism by which they form depends on the relative interaction strengths, as discussed next.
III.2 Assembly mechanisms
As noted previously Hagan2008 ; Elrad2010 ; Hagan2014 , pathways for assembly around a central core such as a polyelectrolyte can be roughly separated into two classes. In the first class (Fig. 4a), which we refer to as the ‘en masse’ mechanism, subunits first adsorb onto the polyelectrolyte in a disordered manner, followed by cooperative rearrangements to form an ordered capsid. In the second class (Fig. 4b), referred to as the nucleation-and-growth mechanism, a small partial capsid nucleates on the polyelectrolyte followed by the sequential, reversible addition of subunits or small oligomers until assembly completes. In contrast to the earlier models which considered a qualitative subunit-polyelectrolyte interaction, we study here how assembly pathways depend on the ionic strength.
To quantify the degree of order along assembly pathways, we record the total number of subunits adsorbed to the polyelectrolyte $n_{\text{ad}}$ and the number of subunits in the largest cluster $n$. Trajectories that pass through configurations with a large value of $n_{\text{free}}=n_{\text{ad}}-n$ are disordered, with many adsorbed subunits not participating in ordered assemblies. In Fig. 4, these quantities are shown as a function of time averaged over all simulation trajectories (leading to successful assembly or not), for parameter sets that respectively lead to the en masse mechanism (Fig. 4A) and nucleation-and-growth mechanism (Fig. 4B). In the first case, there are strong subunit-polyelectrolyte interactions (low ionic strength, $C_{\text{salt}}{=}100$ mM) and weak subunit-subunit interactions ($\varepsilon_{\text{ss}}{=}3k_{\mathrm{B}}T$). Subunits therefore initially adsorb nonspecifically and form only transient subunit-subunit interactions, leading to a rapid rise in $n_{\text{ad}}$ with $n\approx 0$. Once enough subunits are adsorbed ($\sim 12$ around an optimal-length polyelectrolyte for this model with a 12-subunit capsid), a cooperative fluctuation in subunit configurations eventually leads to a stable nucleus and then rapid completion of the ordered capsid geometry. Since this nucleation process is stochastic, there is a distribution of waiting times and thus a more gradual increase in the average cluster size $n$ (see Fig. S2). In the nucleation-and-growth case, on the other hand, the subunit-polyelectrolyte interactions are weak ($C_{\text{salt}}{=}300$ mM) and the subunit-subunit interactions are strong ($\varepsilon_{\text{ss}}{=}6k_{\mathrm{B}}T$). There is limited nonspecific subunit adsorption onto the polyelectrolyte, adsorbed subunits form relatively strong associations, and thus $n_{\text{ad}}$ and $n$ increase nearly in lockstep. Snapshots from typical trajectories for each of these parameter sets are shown in Fig. 4A.
To visualize the degree of order as a function of parameter values, we define a trajectory-averaged order parameter $\bar{n}_{\text{free}}$, which is $n_{\text{free}}$ averaged over all configurations with $4\leq n\leq 6$ and over all trajectories at a given parameter setElrad2008 . Large values of this parameter ($\bar{n}_{\text{free}}\gtrsim 5$) indicate the en masse mechanism, while small values ($\bar{n}_{\text{free}}\lesssim 2$) indicate the nucleation-and-growth mechanism. As shown in Fig. 5, the degree of order generally increases with ionic strength and subunit-subunit interaction strength, with the most ordered assembly occurring at $C_{\text{salt}}=500$ mM (where on average fewer than one subunit is adsorbed nonspecifically) and $\varepsilon_{\text{ss}}\geq 6k_{\mathrm{B}}T$. However, notice that for $\varepsilon_{\text{ss}}=3$ assembly is always disordered; for such weak subunit-subunit interactions the critical nucleus size is large and a high density of adsorbed subunits is required to achieve nucleation. On the other hand, for moderate subunit-subunit interactions we do not observe the extreme limit of the en masse mechanism even for low ionic strength. Though a low ionic strength drives strong nonspecific subunit adsorption, absorbed subunits collide frequently due to cooperative polymer motions and subunit sliding along the polymer Hu2007 ; Elrad2010 . For $\varepsilon_{\text{ss}}>3$ absorbed subunits therefore achieve nucleation before nonspecific absorption has time to saturate.
The nucleation-and-growth trajectories can be further classified based on the relative timescales of nucleation and growth. When nucleation is slow compared to growth (Figs. 4C and S4), the reaction is effectively two-state — each critical nucleus rapidly proceeds to completion, leading to low concentrations of intermediates which are essentially undetectable in a bulk experiment. When nucleation and growth timescales are comparable, multiple capsids within a spatial region can be assembling simultaneously, and thus potentially could be detected in bulk experiments. Below, we consider whether it is possible to experimentally distinguish between the latter case, ordered assembly with rapid nucleation, and en masse assembly pathways characterized by disordered intermediates.
Biological ARM sequences. The assembly simulations described above consider a simplified peptide ARM, which contains only 5 positive ARM charges (Fig. 1). However, our previous work demonstrated that the ARM amino acid sequence (specifically the charges) can significantly affect binding to a polyelectrolyte Perlmutter2013 . Furthermore, results from a more simplified model suggested that the polyelectrolyte binding affinity is a determining factor for the assembly mechanism Elrad2010 . To test this hypothesis and to place our results in the context of specific viruses, we calculated the polyelectrolyte binding affinity for the simplified ARM and three ARM sequences from biological viruses (SV40, BMV, PC2), with each amino acid classified as neutral, cationic, or anionic. These calculations were performed by setting the subunit-subunit interaction strength to zero and recording the average number of adsorbed subunits at varying $C_{\text{salt}}$ (or Debye length $\lambda_{\text{D}}$).
The measured equilibrium linear densities of adsorbed subunits, $c_{\text{eq}}$, are shown as a function of $C_{\text{salt}}$ in Fig. 6A-C. In all cases the linear density increases monotonically with decreasing $C_{\text{salt}}$; saturating at a maximum density.
The simplified ARM has the largest binding affinity, despite having the lowest net positive charge ($+5$, 0 neutral segments) of the four species considered. Comparison with SV40 ($+6$, 14 neutral segments) and BMV ($+9$, 33 neutral segments) illustrates the expected result that that neutral segments decrease the binding affinity, particularly at high salt. The PC2 subunits, with net charge $+22$ demonstrate markedly different behavior, with significant subunit absorption at the highest salt concentration simulated (500 mM), but saturating at about 300 mM ionic strength due to subunit-subunit charge repulsions. Variations in adsorption density with subunit concentration are shown in Fig. S1.
In Fig. 6D we connect these measures of adsorption with assembly mechanism by plotting the assembly order parameter $\bar{n}_{\text{free}}$ as a function of $c_{\text{eq}}$ (controlled by varying $C_{\text{salt}}$), for several values of the subunit dimerization free energy (determined by $\varepsilon_{\text{ss}}$ and $C_{\text{salt}}$) for the 5-ARM and SV40 models. Plotting against $c_{\text{eq}}$ rather than $C_{\text{salt}}$ ($\lambda_{\text{D}}$) allows us to overlay data from these two models while accounting for the differences in affinity due to ARM sequence described above. In support of the proposed link between binding affinity and assembly mechanism, we find rough agreement in the measured assembly order parameters between the two models.
The results indicate that ARM sequence can significantly influence the assembly mechanism. For example, if we define $\bar{n}_{\text{free}}\leq 2$ as the nucleation-growth mechanism, Fig. 6D indicates that nucleation-growth occurs for $c_{\text{eq}}\leq c_{\text{eq}}^{*}$ with the threshold value $c_{\text{eq}}^{*}=0.0375$ for $\varepsilon_{\text{ss}}\geq 2.5k_{\mathrm{B}}T$. From Fig. 6A-C, we can then identify the threshold values of ionic strength $C_{\text{salt}}^{*}$, above which the nucleation-growth mechanism will occur: $C_{\text{salt}}^{*}\approx 300$ mM for the 5-ARM and $C_{\text{salt}}^{*}\approx 175$ mM for the BMV and SV40 models; while PC2 is below the threshold value for all salt concentrations considered. This allows us to predict, for example, that recent experiments on SV40 assembly (at $C_{\text{salt}}{=}250$mM and observed strong subunit-subunit attraction) would have a very low $\bar{n}_{\text{free}}$ ($\sim 1$), which is consistent with SAXS observations Kler2012 .
III.3 Experimental Observables
We now seek to provide experimental signatures through which the capsid assembly pathways discussed above can be distinguished. We focus on two experimental techniques which have recently been applied to respectively probe the formation of individual capsids and bulk assembly kinetics.
smFCS measurements on individual capsids can distinguish assembly mechanisms. Borodavka et al. Borodavka2012 used single molecule fluorescence correlation spectroscopy (smFCS) to monitor the hydrodynamic radii $R_{\text{H}}$ of nucleocapsid complexes during assembly of MS2 and STNV capsid proteins around cognate RNA or non-cognate RNA. Assembly around cognate RNA was characterized by either constant $R_{\text{H}}$ or, in some trajectories, a collapsed complex followed by gradual increase until reaching the size of an assembled capsid. In contrast, assembly around non-cognate RNA led to an increase in $R_{\text{H}}$ before eventually decreasing to the size of the capsid. The difference between these two behaviors can be attributed to sequence-specific ‘packaging signals’ on cognate RNA that interact with the capsid proteins. In this article we do not consider the effect of packaging signals (these will be considered in a subsequent article); instead, we consider whether the pathways described in the previous section can be distinguished by this experimental technique.
We estimated the hydrodynamic radii $R_{\text{H}}$ for polymer-subunit intermediates using the program HYDROPRO, which has been shown to accurately predict $R_{\text{H}}$ for large protein complexes Ortega2011 . The resulting $R_{\text{H}}$ are shown during representative en masse and ordered assembly trajectories in Fig. 7E. We see that the complex $R_{\text{H}}$ first increases as subunits adsorb onto the complex and then decreases as subunits assemble (Fig. 7E). However, the en masse mechanism leads to a much larger and longer duration increase in $R_{\text{H}}$, due to the extensive and long-lived disordered adsorption of unassembled subunits. The difference in $R_{\text{H}}$ between en masse and ordered trajectories is conserved across their respective parameter ranges (see SI Fig. S3 for other trajectories), and also occurs for assembly trajectories around the model nucleic acid with intramolecular base- pairing developed in Ref. Perlmutter2013 (see SI Fig. S3C,F). These results suggest that smFCS can distinguish among the classes of assembly pathways observed in our simulations. They are consistent with an interpretation of the Borodavka et al. Borodavka2012 results in which assembly around the non-cognate RNA proceeds via the disordered mechanism while packaging signals lead to an ordered mechanism.
To further characterize the differences in polymer conformations between disordered and ordered assembly pathways, we show the polymer radius of gyration, $R_{\text{g}}$, during assembly trajectories in Fig. 7. In contrast to $R_{\text{H}}$, contributions of the capsid proteins are not included in $R_{\text{g}}$. While the results in Fig. 7 are averaged over multiple trajectories, example individual trajectories are shown in SI Fig. S2. In all cases of successful assembly, the polymer is gradually compacted from its free size into its encapsidated size. However, at fixed $C_{\text{salt}}{=}100$mM, the average rate of compaction increases with $\varepsilon_{\text{ss}}$, with a dramatic increase in rate for $\varepsilon_{\text{ss}}>3k_{\mathrm{B}}T$ (Fig. 7A). Similarly, decreasing $C_{\text{salt}}$ increases the rate of compaction (Fig. 7B). Notice that the rate of polymer compaction is not determined by the assembly mechanism — increased order correlates with faster compaction in Fig. 7A but with slower compaction in Figure 7B. When the $R_{\text{g}}$ is plotted as a function of number of adsorbed subunits ($n_{\text{ad}}$), the en masse and ordered pathways clearly split into two groups (Fig. 7C). However, this distinction disappears when $R_{\text{g}}$ is plotted as a function of the number of subunits in the largest cluster ($n$, Fig. 7D). Taken together, these data demonstrate that polymer compaction is driven by adsorbed subunits forming ordered intermediates, with the rate of compaction consequently mirroring the rate of capsid assembly.
trSAXS measurments of bulk assembly kinetics can distinguish assembly mechanisms. While smFCS can detect individual assembly intermediates, Kler et al. Kler2012 ; Kler2013 recently used time resolved small angle x-ray scattering (trSAXS) to elucidate structures of assembling capsids in bulk. They found that the SAXS profiles at all time points could be decomposed into scattering contributions from the unassembled and complete components, suggesting that assembly proceeded by an effectively two-state reaction with undetectable levels of intermediates (i.e. the nucleation-and-growth pathway with relatively slow nucleation). While it is evident that profiles from profiles from a two-state reaction can be decomposed in this way, we investigated the extent to which SAXS profiles from the other pathway classes (en masse or nucleatation-and-growth with rapid nucleation) can be distinguished from the two-state case.
First, Fig. 8A,B shows SAXS profiles calculated (using CRYSOL Svergun1995 ) from simulation snapshots along ensembles of en masse and ordered assembly trajectories. For each parameter set, SAXS profiles are averaged over 6 time windows. In both cases, the first profile is dominated by scattering from free subunits (Fig. 8C) and the final time segment shows clear minima and maxima corresponding to the complete capsid (Fig. 8C). For comparison, Fig. 8C presents the SAXS profiles for ordered subunit clusters ranging in size from a single subunit (black) to a complete capsid (yellow). As the capsid grows distinct minima and maxima appear and become more pronounced. We note that the positions of the minima and maxima in the complete model capsid are similar to those observed experimentally for SV40 Kler2012 .
To test the extent to which these trajectories can be distinguished from the two-state case, we attempted to fit SAXS profiles using a linear combination of the profiles for unassembled polymer and subunits and the complete capsid. The resulting fits for the second fifth of the trajectory (where intermediates are most plentiful) are shown in Figs. 8D–F. At this stage, the ordered systems contain mostly incomplete subunit clusters (6-11 subunits), while the disordered simulations contain mostly adsorbed but unassembled subunits. For the ordered simulations, we find that the fit reproduces all of the features of the intermediates, except at low q. In contrast, the intermediates in the disordered trajectory display a shoulder at $q\sim 0.3nm^{-1}$ that is not observed in any of the unassembled or assembled components. This shoulder is a distinct feature of the disordered intermediate, and thus could be used to identify this class of pathways. Finally, as expected, SAXS profiles from trajectories at parameter sets (Fig. 4C) which lead to two-state kinetics are very well fit by a linear combination of polymer/subunit and complete capsid (Fig. 8F).
A significant distinction between the SAXS profiles is that the ordered pathways lead to an isosbestic point at $q\sim 0.3nm^{-1}$ (as seen in SV40 experimentsKler2012 ), whereas the disordered pathways do not. The presence of an isosbestic point is frequently assumed to indicate two-state behaviour; however, it occurs in our data for ordered trajectories even when the reaction is far from two-state due to rapid nucleation. In these cases the isosbestic point appears due to due to the similarity in scattering from the ordered intermediates and the complete capsid. This suggests that an isosbestic point may distinguish ordered from disordered assemblies, but is less sensitive to the extent to which the reaction kinetics can be approximated as two-state (i.e., how undetectable the intermediate concentrations are).
IV Discussion
Our simulations demonstrate that capsid assembly around a NA or other cargo can proceed through two mechanisms, and that which mechanism occurs depends on the relative strengths of protein-protein and protein-cargo interactions. The assembly mechanism can be rationally designed by tuning these interactions through solution conditions or mutagenesis of capsid protein-NA binding domains. However, because productive assembly requires weak interactions, the parameters must be tuned within relatively narrow ranges, and alterations which increase (decrease) the strength of one type of interaction must be compensated by a decrease (increase) in the other. Our results suggest that the subunit-cargo dissociation constant is an important parameter, whose value might be used to map specific viruses onto our phase diagrams (Figs. 2, 3 ,5), although experimental tests of this capability are required. Finally, we have shown that the nature of assembly pathways can be inferred from the results of recently developed techniques to monitor the assembly of individual capsids or bulk assembly kinetics.
Our simulations predict that a single viral species can be driven to assemble via different mechanisms in vitro by changing solution conditions. In particular, under a constant subunit-subunit binding energy ($\varepsilon_{\text{ss}}\sim 5-6k_{\mathrm{B}}T$) robust assembly occurs for a range of solvent conditions, with highly ordered or disordered assemblies occurring depending on salt concentration ($C_{\text{salt}}\in[50,400]$ mM). To our knowledge, this prediction has not yet been realized experimentally, although the signatures of the two classes of assembly pathways have been seen in experiments on different viruses and/or different cargoes Kler2012 ; Garmann2013 ; Cadena-Nava2012 ; Malyutin2013 ; Borodavka2012 .
One recent experimental study sought to test the role of of $C_{\text{salt}}$ and subunit-subunit attractions (controlled by solution $p$H) on in vitro assembly of CCMV around RNA Garmann2013 . In some regards, these experiments mirror our observations, with malformed capsids observed for strong subunit-subunit attractions, disordered products for weak subunit-subunit attractions, and well-formed capsids at intermediate conditions. However, robust assembly was only observed for a two-step process: first $C_{\text{salt}}$ was reduced (from $1M$) to drive RNA-subunit interactions, and then secondly $p$H was reduced to activate subunit-subunit attractions. The resulting assembly pathways resemble the en masse mechanism described here. On the other hand, one-step assembly led to malformed particles even at moderate $p$H, suggesting an inability to assemble through a nucleation-and-growth mechanism in these experiments. In our simulations, this outcome would only occur at high salt concentrations (e.g. $C_{\text{salt}}\sim 400-500$mM, see Figs. 2, 3), where the narrow range of $\varepsilon_{\text{ss}}$ leading to successful assembly indicates that parameters must be finely tuned. Reproducing such a lack of successful assembly at moderate salt concentrations would require a reduction in the orientation-dependence of the subunit-subunit interaction potential (see Methods), or introduction of additional factors as discussed below. Experiments in which solution conditions are changed for other viruses which do undergo one step assembly (e.g. SV40 Kler2012 ; Kler2013 ), may elucidate which of these possibilities to pursue and would directly test our prediction that the assembly mechanism can be controlled by solution conditions.
Understanding capsid assembly mechanisms and their location within the assembly phase diagram has important implications for the design of antiviral agents. As one example, we consider the recently developed class of HBV inhibitors based on phenylpropenamides Katen2010 ; Katen2013 , which act by increasing the strength of subunit-subunit interactions, driving subunits to assemble in the absence of their genome and thus increasing the generation of empty, non-infective capsids Katen2010 ; Katen2013 . Comparing Figs. 3 and 5 shows that a virus that undergoes ordered assembly (e.g. $\varepsilon_{\text{ss}}\sim 6-7$, $C_{\text{salt}}\sim 300$ mM), sits close to parameters that support empty capsid assembly, which are demarcated by dashed lines in Fig. 3. Thus only a small increase in subunit-subunit interactions is required to trigger unproductive RNA-free assembly. In contrast, a much larger perturbation would be required to achieve empty capsid assembly for a virus that assembles via the en masse mechanism.
Outlook.
We have described two classes of assembly pathways and several distinct failure modes (Fig. 2) that arise when assembly is driven by nonspecific electrostatic subunit-cargo interactions. Our phase diagrams can serve as a starting point to understand how virus-specific features, such as packaging signals Stockley2013 , allosteric NA-induced Kler2013 ; Stockley2007 or ‘autosteric’ protein-induced conformational changes Caspar1980 , base pairing-induced NA-structure Perlmutter2013 ; Yoffe2008 ; Borodavka2012 , or subcellular localization Bamunusinghe2011 can robustly avoid failure modes amidst the crowded intracellular milieu while enabling selective assembly around the viral genome Routh2012 ; Ford2013 ; Rao2006 . For example, allosteric or autosteric conformational changes may allow for strong subunit-subunit interactions on the NA while avoiding the off-cargo assembly we observe at large $\varepsilon_{\text{ss}}$. Systematically studying how these additional factors expand (or contract) regions of parameter space that lead to successful assembly will ultimately reveal how viruses have optimized their structures and interactions for robust assembly in vivo and how their assembly in vivo or in vitro can be manipulated for biomedical or materials science applications.
V Methods
Model. We have recently a presented a complete description of our model system, which we summarize here briefly Perlmutter2013 and in the SI. Our model subunits are based upon that previous used to simulate assembly of empty capsids Wales2005 ; Fejer2009 ; Johnston2010 which we extended previously to model assembly around cargo Perlmutter2013 . The pseudoatoms in the capsid subunit model are illustrated in Fig. 1. Subunit assembly is mediated through an attractive Morse potential between Attractor (‘A’) pseudoatoms located at each subunit vertex. The Top (‘T’) pseudoatoms interact with other ‘T’ psuedoatoms through a potential consisting of the repulsive term of the Lennard-Jones (LJ) potential, the radius of which is chosen to favor a subunit-subunit angle consistent with a dodecahedron (116 degrees). The Bottom (‘B’) pseudoatom has a repulsive LJ interaction with ‘T’ pseudoatoms, intended to prevent ‘upside-down’ assembly. The ‘T’, ‘B’, and ‘A’ pseudoatoms form a rigid body Wales2005 ; Fejer2009 ; Johnston2010 . See Refs. Schwartz1998 ; Hagan2006 ; Hicks2006 ; Nguyen2007 ; Wilber2007 ; Nguyen2008 ; Nguyen2009 ; Johnston2010 ; Wilber2009a ; Wilber2009 ; Rapaport1999 ; Rapaport2004 ; Rapaport2008 ; Hagan2008 ; Elrad2010 ; Hagan2011 ; Mahalik2012 ; Levandovsky2009 for related models.
To model electrostatic interaction with a negatively charged NA or polyelectrolyte we extend the model as follows. Firstly, to better represent the capsid shell we add a layer of ‘Excluder’ pseudoatoms which have a repulsive LJ interaction with the polyelectrolyte and the ARMs. Each ARM is modeled as a bead-spring polymer, with one bead per amino acid. The ‘Excluders’ and first ARM segment are part of the subunit rigid body. ARM beads interact through repulsive LJ interactions and, if charged, electrostatic interactions modelled by a Debye-Huckel potential. We note that repulsion between subunits due to the positive charges does affect the magnitude of the subunit-subunit interaction. Previously, we estimated this repulsion contributes $1k_{\mathrm{B}}T$ to the dimerization free energy Perlmutter2013 . This contribution (and entropic terms) are not included in $\varepsilon_{\text{ss}}$, which is the magnitude of the Morse potential depth. See SI section VII.2) for a discussion of binding free energies.
Simulations and units. Simulations were performed with the Brownian Dynamics algorithm of HOOMD, which uses the Langevin equation to evolve positions and rigid body orientations in time Anderson2008 ; nguyen2011 ; LeBard2012 . Simulations were run using a set of fundamental units. The fundamental energy unit is selected to be $E_{\text{u}}\equiv 1k_{\mathrm{B}}T$. The unit of length $D_{\text{u}}$ is set to the circumradius of a pentagonal subunit, which is taken to be $1D_{\text{u}}\equiv 5$ nm so that the dodecahedron inradius of $1.46D_{\text{u}}=7.3$ nm gives an interior volume consistent with that of the smallest $T{=}1$ capsids. Assembly simulations were run at least 10 times for each set of parameters, each of which were concluded at either completion, persistent malformation, or $2\times 10^{8}$ time steps. For all dynamics simulations there were 60 subunits with $\mbox{box size}{=}200\times 200\times 200$ nm, resulting in a concentration of $12\mu$M.
SAXS profile and hydrodynamic radius estimations. Small Angle X-ray Scattering (SAXS) analysis was performed using CRYSOL Svergun1995 . For this analysis the all-atom structure of an SV40 pentameric subunit Stehle1996 was aligned into the position of each coarse-grained subunit and the polymer was replaced with a carbon chain. We note that this entails significant simplification: segments of the protein which were not resolved in the crystal structure were not reconstructed and there is no optimization of structure at the all-atom resolution. We believe this approximation is suitable, given that our analysis is limited to the X-ray scattering profile in the small angle regime, which reflects $\sim$nm scale structural features. Fitting of the scattering profiles was performed using least squares fitting. Hydrodynamic radius analysis was performed using HYDROPRO Ortega2011 . This program is capable of calculating the hydrodynamic radius of large protein complexes. In order to perform this analysis, we treated the synthetic polymer as an amino acid chain. Though this is a gross approximation, it has a negligible effect: removing the polymer entirely does not change the $R_{\text{H}}$ trend and only alters the magnitude by $\sim 1\%$. Simulations were visualized using VMD Humphrey1996 .
VI Acknowledgments
We gratefully acknowledge Adam Zlotnick for insightful discussion and critical reading of the manuscript. This work was supported by Award Number R01GM108021 from the National Institute Of General Medical Sciences. Computational resources were provided by the NSF through XSEDE computing resources (Longhorn, Maverick, Condor, and Keeneland) and the Brandeis HPCC which is partially supported by NSF-MRSEC-0820492.
References
[1]
Joshua A. Anderson, Chris D. Lorenz, and A. Travesset.
General purpose molecular dynamics simulations fully implemented on
graphics processing units.
J Comput Phys, 227(10):5342–5359, 2008.
[2]
D. G. Angelescu, R. Bruinsma, and P. Linse.
Monte Carlo simulations of polyelectrolytes inside viral
capsids.
Phys Rev E, 73(4):041921, 2006.
[3]
Daniel George Angelescu and Per Linse.
Modelling of icosahedral viruses.
Curr Opin Colloid Interface Sci, 13(6):389–394, 2008.
[4]
Saskia E Bakker, Robert J Ford, Amy M Barker, Janice Robottom, Keith Saunders,
Arwen R Pearson, Neil A Ranson, and Peter G Stockley.
Isolation of an asymmetric RNA uncoating intermediate for a
single-stranded RNA plant virus.
Journal of Molecular Biology, 417(1):65–78, 2012.
[5]
Devinka Bamunusinghe, Jang-Kyun Seo, and A. L. N. Rao.
Subcellular Localization and Rearrangement of Endoplasmic Reticulum
by Brome Mosaic Virus Capsid Protein.
J Virol, 85(6):2953–2963, 2011.
[6]
Jb Bancroft, Hiebert E, and Bracker Ce.
Effects Of Various Polyanions On Shell Formation Of Some Spherical
Viruses.
Virology, 39(4):924–930, 1969.
[7]
V. A. Belyi and M. Muthukumar.
Electrostatic origin of the genome packing in viruses.
Proc Natl Acad Sci U S A, 103(46):17174–17178, 2006.
[8]
Alexander Borodavka, Roman Tuma, and Peter G. Stockley.
Evidence that viral RNAs have evolved for efficient, two-stage
packaging.
Proc Natl Acad Sci U S A, 109(39):15769–15774, 2012.
[9]
B. Bottcher, S. Unseld, H. Ceulemans, R. B. Russell, and H. Jeske.
Geminate structures of African cassava mosaic virus.
J Virol, 78(13):6758–6765, 2004.
[10]
G. R. Bowman, K. A. Beauchamp, G. Boxer, and V. S. Pande.
Progress and challenges in the automated construction of Markov
state models for full protein systems.
J Chem Phys, 131(12), 2009.
[11]
G. R. Bowman, D. L. Ensign, and V. S. Pande.
Enhanced Modeling via Network Theory: Adaptive Sampling of Markov
State Models.
J Chem Theory Comput, 6(3):787–794, 2010.
[12]
G. R. Bowman, V. A. Voelz, and V. S. Pande.
Atomistic Folding Simulations of the Five-Helix Bundle Protein
lambda(6-85).
J Am Chem Soc, 133(4):664–667, 2011.
[13]
Ruben D. Cadena-Nava, Mauricio Comas-Garcia, Rees F. Garmann, A. L. N. Rao,
Charles M. Knobler, and William M. Gelbart.
Self-Assembly of Viral Capsid Protein and RNA Molecules of
Different Sizes: Requirement for a Specific High Protein/RNA Mass Ratio.
J Virol, 86(6):3318–3326, 2012.
[14]
Donald L Caspar.
Movement and self-control in protein assemblies. Quasi-equivalence
revisited.
Biophysical journal, 32(1):103–138, 1980.
[15]
P. Ceres and A. Zlotnick.
Weak protein-protein interactions are sufficient to drive assembly
of hepatitis B virus capsids.
Biochemistry, 41(39):11525–11531, 2002.
[16]
Connie B. Chang, Charles M. Knobler, William M. Gelbart, and Thomas G. Mason.
Curvature dependence of viral protein structures on encapsidated
nanoemulsion droplets.
ACS Nano, 2(2):281–286, FEB 2008.
[17]
J. D. Chodera, N. Singhal, V. S. Pande, K. A. Dill, and W. C. Swope.
Automatic discovery of metastable states for the construction of
Markov models of macromolecular conformational dynamics.
J Chem Phys, 126(15), 2007.
[18]
John D. Chodera, William C. Swope, Frank Noe, Jan-Hendrik Prinz, Michael R.
Shirts, and Vijay S. Pande.
Dynamical reweighting: improved estimates of dynamical properties
from simulations at multiple temperatures.
J Chem Phys, 134(24):244107, 2011.
[19]
Mauricio Comas-Garcia, Ruben D. Cadena-Nava, A. L. N. Rao, Charles M. Knobler,
and William M. Gelbart.
In vitro quantification of the relative packaging efficiencies of
single-stranded RNA molecules by viral capsid protein.
J Virol, 2012.
[20]
Marta Comellas-Aragones, Andres de la Escosura, A. J. Dirks, Anne van der Ham,
Anna Fuste-Cune, Jeroen J. L. M. Cornelissen, and Roeland J. M. Nolte.
Controlled Integration of Polymers into Viral Capsids.
Biomacromolecules, 10(11):3141–3147, NOV 2009.
[21]
J. F. Conway and A. C. Steven.
Methods for reconstructing density maps of ”single” particles from
cryoelectron micrographs to subnanometer resolution.
J of Struct Biol, 128(1):106–118, 1999.
[22]
R. M. Crist, S. A. K. Datta, A. G. Stephen, F. Soheilian, J. Mirro, R. J.
Fisher, K. Nagashima, and A. Rein.
Assembly Properties of Human Immunodeficiency Virus Type 1
Gag-Leucine Zipper Chimeras: Implications for Retrovirus Assembly.
J Virol, 83(5):2216–2225, 2009.
[23]
P. Deuflhard and M. Weber.
Robust Perron cluster analysis in conformation dynamics.
Linear Algebra and Its Applications, 398:161–184, 2005.
[24]
B. Devkota, A. S. Petrov, S. Lemieux, M. B. Boz, L. Tang, A. Schneemann, J. E.
Johnson, and S. C. Harvey.
Structural and Electrostatic Characterization of Pariacoto Virus:
Implications for Viral Assembly.
Biopolymers, 91(7):530–538, 2009.
[25]
S. K. Dixit, N. L. Goicochea, M. C. Daniel, A. Murali, L. Bronstein, M. De,
B. Stein, V. M. Rotello, C. C. Kao, and B. Dragnea.
Quantum dot encapsulation in viral capsids.
Nano Lett, 6(9):1993–1999, 2006.
[26]
E. C. Dykeman, N. E. Grayson, K. Toropova, N. A. Ranson, P. G. Stockley, and
R. Twarock.
Simple Rules for Efficient Assembly Predict the Layout of a Packaged
Viral RNA.
J Mol Biol, 408(3):399–407, 2011.
[27]
Eric C Dykeman, Peter G Stockley, and Reidun Twarock.
Packaging signals in two single-stranded RNA viruses imply a
conserved assembly mechanism and geometry of the packaged genome.
Journal of molecular biology, 425(17):3235–3249, 2013.
[28]
O. M. Elrad and M. F. Hagan.
Nano Lett, 8(11):3850–3857, 2008.
[29]
O.M. Elrad and M. F. Hagan.
Encapsulation of a polymer by an icosahedral virus.
Phys Biol, 7:045003, 2010.
[30]
D. Endres and A. Zlotnick.
Model-based analysis of assembly kinetics for virus capsids or other
spherical polymers.
Biophys J, 83(2):1217–1230, 2002.
[31]
Szilard N. Fejer, Tim R. James, Javier Hernandez-Rojas, and David J. Wales.
Energy landscapes for shells assembled from pentagonal and hexagonal
pyramids.
Phys Chem Chem Phys, 11(12):2098–2104, 2009.
[32]
Robert J Ford, Amy M Barker, Saskia E Bakker, Robert H Coutts, Neil A Ranson,
Simon EV Phillips, Arwen R Pearson, and Peter G Stockley.
Sequence-specific, RNA–protein interactions overcome electrostatic
barriers preventing assembly of satellite tobacco necrosis virus coat
protein.
Journal of molecular biology, 425(6):1050–1064, 2013.
[33]
C. Forrey and M. Muthukumar.
Electrostatics of capsid-induced viral RNA organization.
J Chem Phys, 131(10), 2009.
[34]
J. M. Fox, J. E. Johnson, and M. J. Young.
RNA/Protein Interactions in Icosahedral Virus Assembly.
Seminars in Virology, 5(1):51–60, 1994.
[35]
Peter L. Freddolino, Anton S. Arkhipov, Steven B. Larson, Alexander McPherson,
and Klaus Schulten.
Molecular Dynamics Simulations of the Complete Satellite Tobacco
Mosaic Virus.
Structure, 14(3):437–449, 2006.
[36]
Rees F. Garmann, Mauricio Comas-Garcia, Ajaykumar Gopal, Charles M. Knobler,
and William M. Gelbart.
The Assembly Pathway of an Icosahedral Single-Stranded {RNA} Virus
Depends on the Strength of Inter-Subunit Attractions.
Journal of Molecular Biology, (0):–, 2013.
[37]
Nancy L. Goicochea, Mrinmoy De, Vincent M. Rotello, Suchetana Mukhopadhyay, and
Bogdan Dragnea.
Core-like particles of an enveloped animal virus can self-assemble
efficiently on artificial templates.
Nano Lett, 7(8):2281–2290, 2007.
[38]
E. Grahn, T. Moss, C. Helgstrand, K. Fridborg, M. Sundaram, K. Tars, H. Lago,
N. J. Stonehouse, D. R. Davis, P. G. Stockley, and L. Liljas.
Structural basis of pyrimidine specificity in the MS2 RNA
hairpin-coat-protein complex.
RNA-a Publication of the RNA Society, 7(11):1616–1627, 2001.
[39]
James Grant, Robert L. Jack, and Stephen Whitelam.
Analyzing mechanisms and microscopic reversibility of
self-assembly.
J Chem Phys, 135(21), 2011.
[40]
M. F. Hagan.
Controlling viral capsid assembly with templating.
Phys Rev E, 77:051904, 2008.
[41]
M. F. Hagan and D. Chandler.
Dynamic pathways for viral capsid assembly.
Biophys J, 91(1):42–54, 2006.
[42]
M. F. Hagan, O. M. Elrad, and R. L. Jack.
Mechanisms of kinetic trapping in self-assembly and phase
transformation.
J Chem Phys, 135:104115, 2011.
[43]
M.F. Hagan.
Modeling Viral Capsid Assembly.
Adv Chem Phys, 155, 2014.
http://arxiv.org/abs/1301.1657.
[44]
S. C. Harvey, A. S. Petrov, B. Devkota, and M. B. Boz.
Viral assembly: a molecular modeling perspective.
Phys Chem Chem Phys, 11(45):10553–10564, 2009.
[45]
C. Helgstrand, E. Grahn, T. Moss, N. J. Stonehouse, K. Tars, P. G. Stockley,
and L. Liljas.
Investigating the structural basis of purine specificity in the
structures of MS2 coat protein RNA translational operator hairpins.
Nucleic Acids Res, 30(12):2678–2685, 2002.
[46]
S. D. Hicks and C. L. Henley.
Irreversible growth model for virus capsid assembly.
Phys Rev E, 74(3):031912, 2006.
[47]
E Hiebert, JE Bancroft, and CE Bracker.
Assembly In Vitro Of Some Small Spherical Viruses, Hybrid Viruses,
And Other Nucleoproteins.
Virology, Vol 34 Issue 3(3):492–509, 1968.
[48]
N. S. Hinrichs and V. S. Pande.
Calculation of the distribution of eigenvalues and eigenvectors in
Markovian state models for molecular dynamics.
J Chem Phys, 126(24), 2007.
[49]
L. H. Hu, A. Y. Grosberg, and R. Bruinsma.
Are DNA transcription factor proteins Maxwellian Demons?
Biophys J, 95(3):1151–1156, 2008.
[50]
T. Hu, R. Zhang, and B. I. Shklovskii.
Electrostatic theory of viral self-assembly.
Physica a-Statistical Mechanics and Its Applications,
387(12):3059–3064, 2008.
[51]
Tao Hu and B. I. Shklovskii.
Kinetics of viral self-assembly: Role of the single-stranded RNA
antenna.
Phys Rev E, 75:051901, 2007.
[52]
Tao Hu, Rui Zhang, and B. I. Shklovskii.
Electrostatic theory of viral self-assembly: a toy model.
Physica A, 387:3059, 2008.
[53]
Xinlei Huang, Lyudmila M. Bronstein, John Retrum, Chris Dufort, Irina
Tsvetkova, Stella Aniagyei, Barry Stein, Galen Stucky, Brandon McKenna,
Nicholas Remmes, David Baxter, C. Cheng Kao, and Bogdan Dragnea.
Self-assembled virus-like particles with magnetic cores.
Nano Lett, 7(8):2407–2416, AUG 2007.
[54]
W. Humphrey, A. Dalke, and K. Schulten.
VMD: Visual molecular dynamics.
J Mol Graph, 14(1):33–38, 1996.
[55]
G. Jayachandran, V. Vishal, and V. S. Pande.
Using massively parallel simulation and Markovian models to study
protein folding: Examining the dynamics of the villin headpiece.
J Chem Phys, 124(16), 2006.
[56]
T. Jiang, Z. G. Wang, and J. Z. Wu.
Electrostatic Regulation of Genome Packaging in Human Hepatitis B
Virus.
Biophys J, 96(8):3065–3073, 2009.
[57]
J. M. Johnson, D. A. Willits, M. J. Young, and A. Zlotnick.
Interaction with capsid protein alters RNA structure and the
pathway for in vitro assembly of Cowpea chlorotic mottle virus.
J Mol Biol, 335(2):455–464, 2004.
[58]
I. G. Johnston, A. A. Louis, and J. P. K. Doye.
Modelling the self-assembly of virus capsids.
J Phys : Condens Matter, 22(10), 2010.
[59]
S. P. Katen, S. R. Chirapu, M. G. Finn, and A. Zlotnick.
Trapping of Hepatitis B Virus Capsid Assembly Intermediates by
Phenylpropenamide Assembly Accelerators.
ACS Chemical Biology, 5(12):1125–1136, 2010.
[60]
Sarah P Katen, Zhenning Tan, Srinivas Reddy Chirapu, M G Finn, and Adam
Zlotnick.
Assembly-directed antivirals differentially bind quasiequivalent
pockets to modify hepatitis B virus capsid tertiary and quaternary
structure.
Structure, 21(8):1406–16, August 2013.
[61]
W. K. Kegel and P. van der Schoot.
Competing hydrophobic and screened-Coulomb interactions in hepatitis
B virus capsid assembly.
Biophys J, 86(6):3905–3913, 2004.
[62]
A Kivenson and MF Hagan.
Mechanisms of viral capsid assembly around a polymer.
Biophys J, Vol 99 Issue 2(2):619–628, 2010.
[63]
Stanislav Kler, Roi Asor, Chenglei Li, Avi Ginsburg, Daniel Harries, Ariella
Oppenheim, Adam Zlotnick, and Uri Raviv.
RNA Encapsidation by SV40-Derived Nanoparticles Follows a Rapid
Two-State Mechanism.
J Am Chem Soc, 134(21):8823–8830, 2012.
[64]
Stanislav Kler, Joseph Che-Yen Wang, Mary Dhason, Ariella Oppenheim, and Adam
Zlotnick.
Scaffold properties are a key determinant of the size and shape of
self-assembled virus-derived particles.
ACS Chem. Biol., 8(12):2753–61, December 2013.
[65]
MA Krol, NH Olson, J Tate, JE Johnson, TS Baker, and P Ahlquist.
RNA-controlled polymorphism in the in vivo assembly of 180-subunit
and 120-subunit virions from a single capsid protein.
Proc Natl Acad Sci U S A, 96(24):13650–13655, NOV 23 1999.
[66]
Thomas J. Lane, Gregory R. Bowman, Kyle Beauchamp, Vincent A. Voelz, and
Vijay S. Pande.
Markov State Model Reveals Folding and Functional Dynamics in
Ultra-Long MD Trajectories.
J Am Chem Soc, 133(45):18413–18419, 2011.
[67]
Steven B Larson and Alexander McPherson.
Satellite tobacco mosaic virus RNA: structure and implications for
assembly.
Current opinion in structural biology, 11(1):59–65, 2001.
[68]
David N. LeBard, Benjamin G. Levine, Philipp Mertmann, Stephen A. Barr, Arben
Jusufi, Samantha Sanders, Michael L. Klein, and Athanassios Z.
Panagiotopoulos.
Self-assembly of coarse-grained ionic surfactants accelerated by
graphics processing units.
Soft Matter, 8(8):2385–2397, 2012.
[69]
S. Lee and T. T. Nguyen.
Radial distribution of RNA genomes packaged inside spherical
viruses.
Phys Rev Lett, 100(19), 2008.
[70]
Artem Levandovsky and Roya Zandi.
Nonequilibirum assembly, retroviruses, and conical structures.
Phys Rev Lett, 102(19):198102–198102, 2009.
[71]
P. P. Li, A. Nakanishi, S. W. Clark, and H. Kasamatsu.
Formation of transitory intrachain and interchain disulfide bonds
accompanies the folding and oligomerization of simian virus 40 Vp1 in the
cytoplasm.
Proc Natl Acad Sci U S A, 99(3):1353–1358, 2002.
[72]
L Loo, RH Guenther, VR Basnayake, SA Lommel, and S Franzen.
Controlled encapsidation of gold nanoparticles by a viral protein
shell.
J Am Chem Soc, 128(14):4502–4503, APR 12 2006.
[73]
LiNa Loo, Richard H. Guenther, Steven A. Lommel, and Stefan Franzen.
Encapsidation of nanoparticles by Red Clover Necrotic Mosaic Virus.
J Am Chem Soc, 129(36):11111–11117, SEP 12 2007.
[74]
R. W. Lucas, S. B. Larson, and A. McPherson.
The crystallographic structure of brome mosaic virus.
J Mol Biol, 317(1):95–108, 2002.
[75]
J. P. Mahalik and M. Muthukumar.
Langevin dynamics simulation of polymer-assisted virus-like
assembly.
J Chem Phys, 136(13):135101, 2012.
[76]
Andrey G Malyutin and Bogdan Dragnea.
Budding Pathway in the Templated Assembly of Viruslike Particles.
J Phys Chem B, September 2013.
[77]
JD Mcghee and PHV Hippel.
Theoretical Aspects Of Dna-protein Interactions - Cooperative And
Non-cooperative Binding Of Large Ligands To A One-dimensional Homogeneous
Lattice.
J. Mol. BioEssays, 86(2):469–489, 1974.
[78]
A. McPherson.
Micelle formation and crystallization as paradigms for virus
assembly.
Bioessays, 27(4):447–458, 2005.
[79]
H. D. Nguyen, V. S. Reddy, and C. L. Brooks.
Deciphering the kinetic mechanism of spontaneous self-assembly of
icosahedral capsids.
Nano Lett, 7(2):338–344, 2007.
[80]
HD Nguyen and CL Brooks.
Generalized structural polymorphism in self-assembled viral
particles.
Nano Lett, 8:4574, 2008.
[81]
Hung D. Nguyen, Vijay S. Reddy, and Charles L. Brooks.
Invariant polymorphism in virus capsid assembly.
J Am Chem Soc, 131(7):2606–14, 2009.
[82]
Trung Dac Nguyen, Carolyn L. Phillips, Joshua A. Anderson, and Sharon C.
Glotzer.
Rigid body constraints realized in massively-parallel molecular
dynamics on graphics processing units.
Comput Phys Commun, 182(11):2307–2313, 2011.
[83]
P. Ni, Z. Wang, X. Ma, N. C. Das, P. Sokol, W. Chiu, B. Dragnea, M. F. Hagan,
and C. C Kao.
An Examination of the Electrostatic Interactions between the
N-Terminal Tail of the Coat Protein and RNA in Brome Mosaic Virus.
J Mol Biol, 419:284–300, 2012.
[84]
F. Noe, C. Schutte, E. Vanden-Eijnden, L. Reich, and T. R. Weikl.
Constructing the equilibrium ensemble of folding pathways from short
off-equilibrium simulations.
Proc Natl Acad Sci U S A, 106(45):19011–19016, 2009.
[85]
A Ortega, D Amorós, and J García de La Torre.
Prediction of hydrodynamic and other solution properties of rigid
proteins from atomic-and residue-level models.
Biophysical journal, 101(4):892–898, 2011.
[86]
V. S. Pande, K. Beauchamp, and G. R. Bowman.
Everything you wanted to know about Markov State Models but were
afraid to ask.
Methods, 52(1):99–105, 2010.
[87]
S. Park, D. L. Ensign, and V. S. Pande.
Bayesian update method for adaptive weighted sampling.
Phys Rev E, 74(6), 2006.
[88]
Matthew R Perkett and Michael F. Hagan.
Using Markov State Models to Study Self-Assembly.
Submitted for publication.
[89]
Jason D. Perlmutter, Cong Qiao, and Michael F. Hagan.
Viral genome structures are optimal for capsid assembly.
eLife, 2:e00632, 2013.
[90]
J. Zachary Porterfield, Mary Savari Dhason, Daniel D. Loeb, Michael Nassal,
Stephen J. Stray, and Adam Zlotnick.
Full-Length Hepatitis B Virus Core Protein Packages Viral and
Heterologous RNA with Similarly High Levels of Cooperativity.
J Virol, 84(14):7174–7184, 2010.
[91]
Peter E Prevelige.
New approaches for antiviral targeting of HIV assembly.
J. Mol. Biol., 410(4):634–40, July 2011.
[92]
Jan-Hendrik Prinz, John D. Chodera, Vijay S. Pande, William C. Swope, Jeremy C.
Smith, and Frank Noe.
Optimal use of data in parallel tempering simulations for the
construction of discrete-state Markov models of biomolecular dynamics.
J Chem Phys, 134(24), 2011.
[93]
A. L. N. Rao.
Genome packaging by spherical plant RNA viruses.
In Annual Review of Phytopathology, volume 44 of Annual
Review of Phytopathology, pages 61–87. 2006.
[94]
D. C. Rapaport, J. E. Johnson, and J. Skolnick.
Supramolecular self-assembly: molecular dynamics modeling of
polyhedral shell formation.
Comput Phys Commun, 122:231–235, 1999.
[95]
D.C. Rapaport.
Self-assembly of polyhedral shells: a molecular dynamics study.
Phys. Rev. E., 70:051905, 2004.
[96]
D.C. Rapaport.
The role of reversibility in viral capsid growth: A paradigm for
self-assembly.
Phys Rev Lett, 101:186101, 2008.
[97]
Andrew Routh, Tatiana Domitrovic, and John E Johnson.
Host RNAs, including transposons, are encapsidated by a eukaryotic
single-stranded RNA virus.
Proceedings of the National Academy of Sciences,
109(6):1907–1912, 2012.
[98]
A. Schneemann.
The structural and functional role of RNA in icosahedral virus
assembly.
Annu Rev Microbiol, 60:51–67, 2006.
[99]
Paul Schoot and Roya Zandi.
Impact of the topology of viral RNAs on their encapsulation by virus
coat proteins.
Journal of Biological Physics, 39(2):289–299, 2013.
[100]
R. Schwartz, P. W. Shor, P. E. Prevelige, and B. Berger.
Local rules simulation of the kinetics of virus capsid
self-assembly.
Biophys J, 75(6):2626–2636, 1998.
[101]
A. Siber and R. Podgornik.
Nonspecific interactions in spontaneous assembly of empty versus
functional single-stranded RNA viruses.
Phys Rev E, 78:051915, 2008.
[102]
Antonio Siber, Anze Losdorfer Bozic, and Rudolf Podgornik.
Energies and pressures in viruses: contribution of nonspecific
electrostatic interactions.
Phys Chem Chem Phys, 14(11):3746–3765, 2012.
[103]
Antonio Siber, Roya Zandi, and Rudolf Podgornik.
Thermodynamics of nanospheres encapsulated in virus capsids.
Phys Rev E, 81(5):051919, 2010.
[104]
F. D. Sikkema, M. Comellas-Aragones, R. G. Fokkink, B. J. M. Verduin, Jjlm
Cornelissen, and R. J. M. Nolte.
Monodisperse polymer-virus hybrid nanoparticles.
Org Biomol Chem, 5(1):54–57, 2007.
[105]
T. Stehle, S. J. Gamblin, Y. W. Yan, and S. C. Harrison.
The structure of simian virus 40 refined at 3.1 angstrom
resolution.
Structure, 4(2):165–182, 1996.
[106]
Peter G. Stockley, Ottar Rolfsson, Gary S. Thompson, Gabriella Basnak, Simona
Francese, Nicola J. Stonehouse, Steven W. Homans, and Alison E. Ashcroft.
A simple, RNA-mediated allosteric switch controls the pathway to
formation of a T=3 viral capsid.
J. Mol. BIo., 369:541–552, 2007.
[107]
Peter G Stockley, Reidun Twarock, Saskia E Bakker, Amy M Barker, Alexander
Borodavka, Eric Dykeman, Robert J Ford, Arwen R Pearson, Simon EV Phillips,
Neil A Ranson, et al.
Packaging signals in single-stranded RNA viruses: nature’s
alternative to a purely electrostatic assembly mechanism.
Journal of biological physics, 39(2):277–287, 2013.
[108]
J. Sun, C. DuFort, M. C. Daniel, A. Murali, C. Chen, K. Gopinath, B. Stein,
M. De, V. M. Rotello, A. Holzenburg, C. C. Kao, and B. Dragnea.
Core-controlled polymorphism in virus-like particles.
Proc Natl Acad Sci U S A, 104(4):1354–1359, 2007.
[109]
D Svergun, C Barberato, and MHJ Koch.
CRYSOL-a program to evaluate X-ray solution scattering of biological
macromolecules from atomic coordinates.
Journal of Applied Crystallography, 28(6):768–773, 1995.
[110]
W. C. Swope, J. W. Pitera, and F. Suits.
Describing protein folding kinetics by molecular dynamics
simulations. 1. Theory.
J Phys Chem B, 108(21):6571–6581, 2004.
[111]
W. C. Swope, J. W. Pitera, F. Suits, M. Pitman, M. Eleftheriou, B. G. Fitch,
R. S. Germain, A. Rayshubski, T. J. C. Ward, Y. Zhestkov, and R. Zhou.
Describing protein folding kinetics by molecular dynamics
simulations. 2. Example applications to alanine dipeptide and beta-hairpin
peptide.
J Phys Chem B, 108(21):6582–6594, 2004.
[112]
M. Tihova, K. A. Dryden, T. V. L. Le, S. C. Harvey, J. E. Johnson, M. Yeager,
and A. Schneemann.
Nodavirus coat protein imposes dodecahedral RNA structure
independent of nucleotide sequence and length.
J Virol, 78(6):2897–2905, 2004.
[113]
Christina L. Ting, Jianzhong Wu, and Zhen-Gang Wang.
Thermodynamic basis for the genome to capsid charge relationship in
viral encapsidation.
Proc Natl Acad Sci U S A, 108(41):16986–16991, 2011.
[114]
K. Toropova, G. Basnak, R. Twarock, P. G. Stockley, and N. A. Ranson.
The three-dimensional structure of genomic RNA in bacteriophage
MS2: Implications for assembly.
J Mol Biol, 375:824–836, 2008.
[115]
K. Valegard, J. B. Murray, P. G. Stockley, N. J. Stonehouse, and L. Liljas.
Crystal structure of an bacteriophage RNA coat protein operator
complex.
Nature, 371(6498):623–626, 1994.
[116]
K. Valegard, J. B. Murray, N. J. Stonehouse, S. vandenWorm, P. G. Stockley, and
L. Liljas.
The three-dimensional structures of two complexes between
recombinant MS2 capsids and RNA operator fragments reveal
sequence-specific protein-RNA interactions.
J Mol Biol, 270(5):724–738, 1997.
[117]
S. H. E. van den Worm, N. J. Stonehouse, K. Valegard, J. B. Murray, C. Walton,
K. Fridborg, P. G. Stockley, and L. Liljas.
Crystal structures of MS2 coat protein mutants in complex with
wild-type RNA operator fragments.
Nucleic Acids Res, 26(5):1345–1351, 1998.
[118]
P. van der Schoot and R. Bruinsma.
Electrostatics and the assembly of an RNA virus.
Phys Rev E, 71(6):061928, 2005.
[119]
D. J. Wales.
The energy landscape as a unifying theme in molecular science.
Phil Trans R Soc A, 363(1827):357–375, 2005.
[120]
Stephen Whitelam, Carl Rogers, Andrea Pasqua, Chad Paavola, Jonathan Trent, and
Phillip L. Geissler.
The Impact of Conformational Fluctuations on Self-Assembly:
Cooperative Aggregation of Archaeal Chaperonin Proteins.
Nano Lett, 9:292–297, 2009.
[121]
A. W. Wilber, J. P. K. Doye, and A. A. Louis.
Self-assembly of monodisperse clusters: Dependence on target
geometry.
J Chem Phys, 131(17), 2009.
[122]
A. W. Wilber, J. P. K. Doye, A. A. Louis, E. G. Noya, M. A. Miller, and
P. Wong.
Reversible self-assembly of patchy particles into monodisperse
icosahedral clusters.
J Chem Phys, 127(8), 2007.
[123]
Alex W. Wilber, Jonathan P. K. Doye, Ard A. Louis, and Anna C. F. Lewis.
Monodisperse self-assembly in a model with protein-like
interactions.
J Chem Phys, 131(17):175102, 2009.
[124]
A. M. Yoffe, P. Prinsen, A. Gopal, C. M. Knobler, W. M. Gelbart, and
A. Ben-Shaul.
Predicting the sizes of large RNA molecules.
Proc Natl Acad Sci U S A, 105(42):16153–16158, 2008.
[125]
Y. Zeng, S.B. Larson, C.E. Heitsch, A. McPherson, and S.C. Harvey.
A Model for the Structure of Satellite Tobacco Mosaic Virus.
J Struct Biol, 2012.
[126]
Z. L. Zhang and S. C. Glotzer.
Self-assembly of patchy particles.
Nano Lett, 4(8):1407–1413, 2004.
[127]
A. Zlotnick, N. Cheng, S. J. Stahl, J. F. Conway, A. C. Steven, and P. T.
Wingfield.
Localization of the C terminus of the assembly domain of hepatitis B
virus capsid protein: Implications for morphogenesis and organization of
encapsidated RNA.
Proc Natl Acad Sci U S A, 94(18):9556–9561, 1997.
[128]
A. Zlotnick and S. Mukhopadhyay.
Virus assembly, allostery and antivirals.
Trends Microbiol, 19(1):14–23, 2011.
[129]
Adam Zlotnick, J. Zachary Porterfield, and Joseph Che-Yen Wang.
To Build a Virus on a Nucleic Acid Substrate.
Biophys J, 104(7):1595–1604, 2013.
VII Supporting Information
VII.1 Model potentials and parameters
The model details are described in Ref. [89]; we summarize them here for the convenience of the reader. In our model, all potentials can be decomposed into pairwise interactions. Potentials involving capsomer subunits further decompose into pairwise interactions between their constituent building blocks – the excluders, attractors, ‘Top’ and ‘Bottom’, and ARM pseudoatoms. It is convenient to write the total energy of the system as the sum of 6 terms: a capsomer-capsomer $U_{\mathrm{cc}}$ part (which does not include interactions between ARM pseudoatoms), capsomer-polymer $U_{\mathrm{cp}}$, capsomer-ARM $U_{\mathrm{ca}}$, polymer-polymer $U_{\mathrm{pp}}$, polymer-ARM $U_{\mathrm{pa}}$, and ARM-ARM $U_{\mathrm{aa}}$ parts. Each is summed over all pairs of the appropriate type:
$$\displaystyle U=$$
$$\displaystyle\sum_{\mathrm{cap\ }{i}}\sum_{\mathrm{cap\ }{j<i}}U_{\mathrm{cc}}%
+\sum_{\mathrm{cap\ }{i}}\sum_{\mathrm{poly\ }{j}}U_{\mathrm{cp}}+\sum_{%
\mathrm{cap\ }{i}}\sum_{\mathrm{ARM\ }{j}}U_{\mathrm{ca}}+$$
$$\displaystyle\sum_{\mathrm{poly\ }{i}}\sum_{\mathrm{poly\ }{j<i}}U_{\mathrm{pp%
}}+\sum_{\mathrm{poly\ }{i}}\sum_{\mathrm{ARM\ }{j}}U_{\mathrm{pa}}+\sum_{%
\mathrm{tail\ }{i}}\sum_{\mathrm{ARM\ }{j<i}}U_{\mathrm{aa}}$$
(1)
where $\sum_{\mathrm{cap\ }{i}}\sum_{\mathrm{cap\ }{j<i}}$ is the sum over all distinct pairs of capsomers in the system, $\sum_{\mathrm{cap\ }{i}}\sum_{\mathrm{poly\ }{j}}$ is the sum over all capsomer-polymer pairs, etc.
The capsomer-capsomer potential $U_{\mathrm{cc}}$ is the sum of the attractive interactions between complementary attractors, and geometry guiding repulsive interactions between ‘Top’ - ‘Top’ pairs and ‘Top’ - ‘Bottom’ pairs. There are no interactions between members of the same rigid body, but ARMs are not rigid and thus there are intra-subunit ARM-ARM interactions. Thus, for notational clarity, we index rigid bodies and non-rigid pseudoatoms in Roman, while the pseudoatoms comprising a particular rigid body are indexed in Greek. E.g., for capsomer $i$ we denote its attractor positions as $\{\bm{a}_{i\alpha}\}$ with the set comprising all attractors $\alpha$, its ‘Top’ positions $\{\bm{t}_{i\alpha}\}$, and its ‘Bottom’ positions $\{\bm{b}_{i\alpha}\}$. The capsomer-capsomer interaction potential between two capsomers $i$ and $j$ is then defined as:
$$\displaystyle U_{\mathrm{cc}}$$
$$\displaystyle(\{\bm{a}_{i\alpha}\},\{\bm{t}_{i\alpha}\},\{\bm{b}_{i\alpha}\},%
\{\bm{a}_{j\beta},\{\bm{t}_{j\beta}\},\{\bm{b}_{j\beta}\})=$$
$$\displaystyle\sum_{\alpha,\beta}^{N_{\mathrm{t}}}\varepsilon\mathcal{L}\left(%
\left|\bm{t}_{i\alpha}-\bm{t}_{j\beta}\right|,\ \sigma_{\mathrm{t}}\right)+$$
$$\displaystyle\sum_{\alpha,\beta}^{N_{\mathrm{b}},N_{\mathrm{t}}}\varepsilon%
\mathcal{L}\left(\left|\bm{b}_{i\alpha}-\bm{t}_{j\beta}\right|,\ \sigma_{%
\mathrm{b}}\right)+$$
$$\displaystyle\sum_{\alpha,\beta}^{N_{\mathrm{a}}}\varepsilon\mathcal{M}\left(%
\left|\bm{a}_{i\alpha}-\bm{a}_{j\beta}\right|,\ r_{\mathrm{0}},\varrho,r_{%
\text{cut}}\right)$$
(2)
where $\varepsilon$ is an adjustable parameter which both sets the strength of the capsomer-capsomer attraction at each attractor site and scales the repulsive interactions which enforce the dodecahedral geometry. $N_{\mathrm{t}}$, $N_{\mathrm{b}}$, and $N_{\mathrm{a}}$ are the number of ‘Top’, ‘Bottom’, and attractor pseudoatoms respectively in one capsomer, $\sigma_{\mathrm{t}}$ and $\sigma_{\mathrm{b}}$ are the effective diameters of the ‘Top’ – ‘Top’ interaction and ‘Bottom’ – ‘Top’ interactions, which are set to 10.5 nm and 9.0 nm respectively throughout this work, $r_{\mathrm{0}}$ is the minimum energy attractor distance, set to 1 nm, $\varrho$ is a parameter determining the width of the attractive interaction, set to 2.5, and $r_{\text{cut}}$ is the cutoff distance for the attractor potential, set to 10.0 nm.
The function $\mathcal{L}$ is defined as the repulsive component of the Lennard-Jones potential shifted to zero at the interaction diameter:
$$\mathcal{L}(x,\sigma)\equiv\left\{\begin{array}[]{ll}\left(\frac{\sigma}{x}%
\right)^{12}-1&:x<\sigma\\
0&:\mathrm{otherwise}\end{array}\right.$$
(3)
The function $\mathcal{M}$ is a Morse potential:
$$\mathcal{M}(x,r_{\mathrm{0}},\varrho)\equiv\left\{\begin{array}[]{ll}\left(e^{%
\varrho\left(1-\frac{x}{r_{\mathrm{0}}}\right)}-2\right)e^{\varrho\left(1-%
\frac{x}{r_{\mathrm{0}}}\right)}&:x<r_{\text{cut}}\\
0&:\mathrm{otherwise}\end{array}\right.$$
(4)
The capsomer-polymer interaction is a short-range repulsion that accounts for excluded-volume. For capsomer $i$ with excluder positions $\{\bm{x}_{i\alpha}\}$ and polymer subunit $j$ with position $\bm{R}_{j}$, the potential is:
$$\displaystyle U_{\mathrm{cp}}(\{\bm{x}_{i\alpha}\},\bm{R}_{j})$$
$$\displaystyle=$$
$$\displaystyle\sum_{\alpha}^{N_{\mathrm{x}}}\mathcal{L}\left(|\bm{x}_{i\alpha}-%
\bm{R}_{j}|,\sigma_{\mathrm{xp}}\right)$$
(5)
where $N_{\text{x}}$ is the number of excluders on a capsomer and $\sigma_{\mathrm{xp}}=0.5(\sigma_{\text{x}}+\sigma_{\text{p}})$ is the effective diameter of the excluder – polymer repulsion. The diameter of the polymer bead is $\sigma_{\text{p}}=0.5$ nm and the diameter for the excluder beads is $\sigma_{\text{x}}=3.0$ nm for the $T{=}1$ model and $\sigma_{\text{x}}=5.25$ nm for the $T{=}3$ model.
The capsomer-ARM interaction is a short-range repulsion that accounts for excluded-volume. For capsomer $i$ with excluder positions $\{\bm{x}_{i\alpha}\}$ and ARM subunit $j$ with position $\bm{R}_{j}$, the potential is:
$$\displaystyle U_{\mathrm{cA}}(\{\bm{x}_{i\alpha}\},\bm{R}_{j})$$
$$\displaystyle=$$
$$\displaystyle\sum_{\alpha}^{N_{\mathrm{x}}}\mathcal{L}\left(|\bm{x}_{i\alpha}-%
\bm{R}_{j}|,\sigma_{\mathrm{xA}}\right)$$
(6)
with $\sigma_{\mathrm{xA}}=0.5(\sigma_{\text{x}}+\sigma_{\text{A}})$ as the effective diameter of the excluder - ARM repulsion with $\sigma_{\text{A}}=0.5$ nm the diameter of an ARM bead.
The polymer-polymer non-bonded interaction is composed of electrostatic repulsions and short-ranged excluded-volume interactions. These polymers also contain bonded interactions which are only evaluated for segments occupying adjacent positions along the polymer chain and angular interactions which are only evaluated for three sequential polymer segments. As noted in the main text, electrostatics are represented by Debye Huckel interactions.
$$\displaystyle U_{\mathrm{pp}}(\bm{R}_{i},\bm{R}_{j},\bm{R}_{k})$$
$$\displaystyle=$$
$$\displaystyle\left\{\begin{array}[]{l}\mathcal{K}_{\mathrm{bond}}(R_{ij},%
\sigma_{\mathrm{p}},k_{\mathrm{bond}})\\
\quad:\{i,j\}\ \mathrm{bonded}\\
\mathcal{K}_{\mathrm{angle}}(R_{ijk},k_{\mathrm{angle}})\\
\quad:\{i,j,k\}\ \mathrm{angle}\\
\mathcal{L}(R_{ij},\sigma_{\mathrm{p}})+\mathcal{U}_{\mathrm{DH}}(R_{ij},q_{%
\text{p}},q_{\text{p}},\sigma_{\mathrm{p}})\\
\quad:\{i,j\}\ \mathrm{nonbonded}\\
\end{array}\right.$$
(7)
where $R_{ij}\equiv|\bm{R}_{i}-\bm{R}_{j}|$ is the center-to-center distance between the polymer subunits, $q_{\text{p}}=-1$ is the valence of charge on each polymer segment, and $\mathcal{U_{\mathrm{DH}}}$ is a Debye-Huckel potential smoothly shifted to zero at the cutoff:
$$\displaystyle\mathcal{U}_{\mathrm{DH}}$$
$$\displaystyle(r,q_{1},q_{2},\sigma{})\equiv$$
(8)
$$\displaystyle\left\{\begin{array}[]{l}\frac{q_{1}q_{2}l_{\mathrm{b}}\ e^{%
\sigma{}/\lambda_{\mathrm{D}}}}{\lambda_{\mathrm{D}}+\sigma{}}\left(\frac{e^{-%
r/\lambda_{\mathrm{D}}}}{r}\right)\\
\quad:x<2\lambda_{\mathrm{D}}\\
\frac{(r_{cut}^{2}-r^{2})^{2}(r_{cut}^{2}+2r^{2}-3r_{on}^{2}))}{(r_{cut}^{2}-2%
r_{on}^{2})^{3}}\frac{q_{1}q_{2}l_{\mathrm{b}}\ e^{\sigma{}/\lambda_{\mathrm{D%
}}}}{\lambda_{\mathrm{D}}+\sigma{}}\left(\frac{e^{-r/\lambda_{\mathrm{D}}}}{r}%
\right)\\
\quad:2\lambda_{\mathrm{D}}<x<3\lambda_{\mathrm{D}}\\
0\\
\quad:\mathrm{otherwise}\end{array}\right.$$
(9)
$\lambda_{\mathrm{D}}$ is the Debye length, $l_{\mathrm{b}}$ is the Bjerrum length, and $q_{1}$ and $q_{2}$ are the valences of the interacting charges.
Bonds are represented by a harmonic potential:
$$\mathcal{K}_{\mathrm{bond}}(R_{ij},\sigma,k_{\mathrm{bond}})\equiv\frac{k_{%
\mathrm{bond}}}{2}(R_{ij}-\sigma)^{2}.$$
(10)
Angles are also represented by a harmonic potential:
$$\mathcal{K}_{\mathrm{angle}}(R_{ijk},k_{\mathrm{angle}})\equiv\frac{k_{\mathrm%
{angle}}}{2}(\vartheta_{ijk})^{2}$$
(11)
where $\vartheta_{ijk}$ is the angle formed by the sequential polymer units $i,j,k$.
The ARM-ARM interaction is similar to the polymer-polymer interaction, consisting of non-bonded interactions composed of electrostatic repulsions and short-ranged excluded-volume interactions. These ARMs also contain bonded interactions which are only evaluated for segments occupying adjacent positions along the polymer chain:
$$\displaystyle U_{\mathrm{aa}}(\bm{R}_{i},\bm{R}_{j})$$
$$\displaystyle=$$
$$\displaystyle\left\{\begin{array}[]{l}\mathcal{K}_{\mathrm{bond}}(R_{ij},%
\sigma_{\mathrm{a}},k_{\mathrm{bond}})\\
\quad:\{i,j\}\ \mathrm{bonded}\\
\mathcal{L}(R_{ij},\sigma_{\mathrm{a}})+\mathcal{U}_{\mathrm{DH}}(R_{ij},q_{i}%
,q_{j},\sigma_{\mathrm{a}})\\
\quad:\{i,j\}\ \mathrm{nonbonded}\\
\end{array}\right.$$
(12)
where $R_{ij}\equiv|\bm{R}_{i}-\bm{R}_{j}|$ is the center-to-center distance between the ARM subunits and $q_{i}$ is the valence of charge on ARM segment $i$.
Finally, the ARM-Polymer interaction is the sum of short-ranged excluded-volume interactions and electrostatic interactions:
$$U_{\mathrm{pa}}(\bm{R}_{i},\bm{R}_{j})=\mathcal{L}(R_{ij},\sigma_{\mathrm{ap}}%
)+\mathcal{U}_{\mathrm{DH}}(R_{ij},q_{i},q_{j},\sigma_{\mathrm{ap}})$$
(13)
VII.2 Calculation of binding free energy estimates
Subunit-Subunit binding free energy. Our method of calculating the subunit-subunit binding free energy was described previously [89] and is similar to that presented in our previous work [29, 42]. Briefly, subunits were modified such that only one edge formed attractive bonds, limiting complex formation to dimers. We measured the relative concentration of dimers and monomers for a range of attraction strengths ($\varepsilon_{\text{ss}}$). The free energy of binding along that interface is then $g_{\text{ss}}/k_{\mathrm{B}}T=-\ln(c_{\text{ss}}/K_{\mathrm{d}})$ with standard state concentration $c_{\text{ss}}=1$ M and $K_{\mathrm{d}}$ in molar units. We can then correct for the multiplicity of dimer conformations, by adding in the additional term $-T\Delta s_{\mathrm{c}}=\ln(25/2)k_{\mathrm{B}}T$, where the five pentagonal edges are assumed to be distinguishable, but complex orientations which differ only through global rotation are not. For subunits which do not contain ARMs, the free energy is well fit by the linear expression $g_{\text{ss}}/k_{\mathrm{B}}T=-1.5\varepsilon_{\text{ss}}-Ts_{\mathrm{b}}$ where $Ts_{\mathrm{b}}=-5.0k_{\mathrm{B}}T$, and these values are used in Figures 3 and 5. Calculations were also performed with ARMs at varying $C_{\text{salt}}$. For $C_{\text{salt}}{=}100$ mM, $g_{\text{ss}}$ increases by $\sim 0.5k_{\mathrm{B}}T$ due to ARM-ARM repulsion for the simple 5-ARM model. For the SV40 ARM, $g_{\text{ss}}$ increases by $\sim 2k_{\mathrm{B}}T$ at $C_{\text{salt}}{=}100$ mM.
Subunit-Polymer interaction. Our method of estimating the subunit-polymer binding free energy is also similar to that presented in our previous work [29, 42]. In these simulations, the subunit-subunit attraction is eliminated (i.e. $\varepsilon_{\text{ss}}{=}0k_{\mathrm{B}}T$), and the average number of subunits adsorbed to the polymer is measured as a function of the subunit concentration and salt concentration (Fig. S1A). With sufficient data, the dissociation constant and free energy as a function of salt concentration could be determined using the McGhee-von Hippel formulation, wherein a bound ligand occupies multiple binding sites on the polymer [77]. In our previous simulations the number of binding sites occupied by a subunit was determined by the model [29, 42], whereas here the number of polymer sites occupied per subunit emerges from collective interactions. As our estimate, we use the number of subunits per polymer length for the maximal saturation observed, which for our subunits containing 25 ARM charges (5 ARMs of length 5 each) results in $p=22$. This leads to estimates in the free energy of subunit-polymer binding of $\sim 3-6k_{\mathrm{B}}T$ for $C_{\text{salt}}=500-100$ mM. However, we found systematic variations of our estimate for changing total subunit concentration, suggesting that our sampling was incomplete.
VII.3 Constructing the Markov state model
Markov state models (MSMs) have been used extensively to study protein folding [12, 11, 10, 111, 110, 92, 87, 86, 84, 66, 55, 18, 48, 17, 23], and are one of the few methods that can describe out-of-equilibrium dynamical processes that include rare events, such as a nucleation barrier crossing. Our strategy for using MSMs to study capsid self-assembly is described in detail in Ref. [88]; we summarize the key points here. We construct an MSM by running many short (relative to the assembly timescale) simulations using a ratcheting procedure, which starts simulations from configurations based on the most poorly sampled states. System configurations are partitioned into states such that configurations which interconvert rapidly are collected in the same state while those which interconvert slowly are in separate states. The separation of timescales provided by such decomposition ensures that the system behaves Markovian beyond the relatively short intra-state relaxation timescales. Inter-state transition probabilities are then measured after a ‘lag time’ which is longer than intra-state relaxation times.
We follow the simplest approach described in Ref. [88], in which states are defined by the largest cluster size $n$ and the total number of subunits adsorbed to the polyelectrolyte $n_{\text{ad}}$. Using this state decomposition produced 74 states for the slow nucleation parameters ($\varepsilon_{\text{ss}}=5k_{\mathrm{B}}T$, $C_{\text{salt}}=500$mM) with a lag time of $4\times 10^{5}$ simulation steps sufficient to maintain the Markov property.
MSM Calculations. The transition probability matrix $\mathbf{T}(\tau)$ is calculated by column-normalizing the count matrix $\mathbf{C}(\tau)$, in which each element $C_{ji}$ gives the total number of transitions from state $i$ to state $j$ measured at a lag time $\tau$. The count matrix is calculated from the many, relatively short, trajectories run in parallel using the ratcheting procedure described below. After diagonalizing $\mathbf{T}(\tau)$, the time-dependent state probabilities can be written as
$$\displaystyle\vec{P}(t;\tau)$$
$$\displaystyle=\sum_{i=1}^{N}\ket{i}\bra{i}\ket{\vec{P}(0)}e^{-\lambda_{i}t}$$
$$\displaystyle\lambda_{i}$$
$$\displaystyle=-\log(\omega_{i})/\tau$$
(14)
where $\omega_{i}$ is the $i^{\text{th}}$ eigenvalue of $\mathbf{T}(\tau)$ and $\bra{i}$ and $\ket{i}$ are the corresponding left/right eigenvectors, which are assumed to be normalized. Since $\mathbf{T}(\tau)$ is generally not Hermitian the left and right eigenvectors are not equivalent. There is only one unit eigenvalue, whose associated right eigenvector corresponds to the equilibrium distribution, while all other eigenvalues are positive and real [110]. The average value of an order parameter, $Q$, can then be calculated as a function of time from
$$\bar{Q}(t)=\sum_{i=1}^{N}P_{i}(t;\tau)Q_{i}$$
(15)
with $Q_{i}$ as the order parameter value for state $i$ and $P_{i}(t;\tau)$ as the time dependent probability of state $i$ (from Eq. 14). Eq. 15 was used to calculate the mean cluster size $n$ and mean number adsorbed $n_{\text{ad}}$ in SI Fig. 4C. In order to calculate the probability of a given cluster size as a function of time in SI Fig. S4, the same procedure was followed except that all states for a given $n$ were lumped together to give a single curve for each $n$.
Ratcheting Procedure. Because of the Markov property, starting coordinates for simulations can be chosen to efficiently generate good statistics for all of the relevant transition elements. Many simulations are run in parallel for a time $t_{\text{s}}$ , which must be longer than the lag time $\tau$, but can be much shorter than the longest relaxation timescale. For the slow nucleation calculation in this paper, we continuously ran $100$ simulations in parallel starting from an unassembled bath of subunits. New simulations were seeded from the states with the fewest number of simulation starts, biasing sampling to unexplored regions of state space. Our ratcheting procedure initially grouped states only by the number of subunits adsorbed to polymer $n_{\text{ad}}$. However, after discovering states for each cluster size on the pathway to assembly, we ran a large number of simulations in parallel seeded from a random selection of existing states. This crude, but effective approach could be improved by using adaptive sampling [11], which explicitly considers the error in the MSM when seeding new simulations. A total of $6.3\times 10^{9}$ simulation steps were used to build the MSM. |
AEI-2000-078 OHSTPY-HEP-T-00-032 hep-th/0012080
Tachyon condensation and universality of DBI action
G. Arutyunov${}^{*,**}$ 111emails: agleb, theisen @aei-potsdam.mpg.de; frolov,
tseytlin @mps.ohio-state.edu
${}^{**}$On leave of absence from Steklov Mathematical Institute,
Gubkin str. 8,
117966, Moscow, Russia,
S. Frolov${}^{\dagger,**}$,
S. Theisen${}^{*}$
and A.A. Tseytlin${}^{\dagger}$
222Also at Imperial College, London and Lebedev Institute, Moscow
${}^{*}$
Max-Planck-Institut für Gravitationsphysik,
Albert-Einstein-Institut,
Am Mühlenberg 1, D-14476 Golm, Germany
${}^{\dagger}$ Department of Physics,
The Ohio State University,
Columbus, OH 43210-1106, USA
Abstract
We show that a low-energy action for
massless fluctuations around a tachyonic soliton background
representing a codimension one D-brane coincides with the
Dirac-Born-Infeld action. The scalar modes which
describe transverse oscillations of the D-brane
are translational collective coordinates of the soliton.
The appearance of the DBI action is a universal
feature independent of details of a tachyon effective action,
provided it has the structure implied by the open
string sigma model partition function.
1 Introduction
In the original perturbative
string-theory description
D-branes [1] are specified by boundary conditions
on the open strings ending on them. Their
collective coordinates
are identified with massless modes of the
open strings, and the “acceleration-independent”
part of their action – the DBI action [2] –
can be derived
(using T-duality)
as a reduction of Born-Infeld action
(directly from the open string sigma model
partition function [3]
or from conformal invariance condition [4],
see [2, 5, 6]).
Supersymmetric
D-branes appear also in another guise
as black-hole type solitons of $D=10$ type II
supergravity
[7, 8] which carry Ramond-Ramond charges [9].
Their collective coordinates,
determined by the
massless fluctuation modes in these
backgrounds,
can be related to
the parameters of spontaneously broken gauge
symmetries [10]. With
an appropriate non-linear parametrization
of the supergravity fields in terms of the collective
coordinates, one should be able to derive the
corresponding DBI actions directly from the
type II supergravity action.111To a large extent
this should follow
essentially from reparametrization
invariance,
implying that the action of a boosted 11-d Schwarzschild
black hole should be $M\int dt\sqrt{1-v^{2}}$, and
T- and S- dualities of type II supergravities.
Recently, a new, third,
description of D-branes as
solitons was suggested [11, 12].
Arguments supporting the proposal
that D-branes can be interpreted as tachyonic solitons
were given
using a variety of approaches:
boundary conformal field theory (see, e.g., [13]),
Witten’s open string field theory [14]
(see, e.g., [15]), non-commutative
field theory obtained in large $B_{\mu\nu}$ limit
(see, e.g., [16]) and simple
low-energy effective Lagrangian models
with specific tachyon potentials [17, 18, 19].
Remarkably, the
latter models were shown to follow
from the boundary string field theory (BSFT)
[20, 21] in [22, 23, 24].
The BSFT approach, consistently restricted to the lowest-level
(renormalizable) tachyon
and vector couplings, is essentially the same
as the off-shell sigma model approach [25, 3, 26]
being based on the disc partition function
of the open string sigma model.
The partition function encodes the information not only about
the beta-functions but also about the field space
metric which relates them
to field equations of motion, and thus
allows to smoothly interpolate between the standard perturbative
tachyon
vacuum and a new non-trivial vacuum at the minimum
of the tachyon potential.222A simple
consequence of the sigma model
approach to the tachyon condensation
is the background independence of
the tachyon potentials discussed in the
framework of Witten’s open string field
theory in [27]: the zero mode
of the tachyon field which is the only argument
of the potential
does not feel any closed string background.
One finds that
there are solitonic solutions corresponding to D-branes of
lower dimensions
[17, 18],
and that the descent relations between D-brane tensions
hold exactly [23, 24].
These results
have been generalized to include a background
gauge field
[28, 29, 30, 31, 32, 26].333Since a constant
field strength $F_{\mu\nu}$ enters the
actions in
the same way as a constant $B_{\mu\nu}$-field,
the sigma model approach
should provide a
natural explanation for some
of the results for noncommutative tachyon
condensation
in [16].
From the boundary sigma model point of view, one starts with a
conformal theory with $d$ Neumann boundary conditions in the UV
and adds relevant (tachyon) perturbations driving the theory
to an IR fixed point that corresponds to a new (stable or unstable)
vacuum with $(d-1-p)$ Dirichlet boundary conditions. The IR fixed
point is then interpreted as a closed string vacuum with a Dp-brane.
Given that the tension of a Dp-brane is correctly
reproduced [23, 24],
a further crucial test
is to find the spectrum and the effective action
for light modes on
a Dp-brane obtained as a result of tachyon condensation.
The aim of the present paper is to show that
(the “acceleration-independent” part of) the action
for the massless scalar modes
$\Phi$ on the tachyonic soliton and the massless
$U(1)$ vector field $A$ is indeed the standard DBI action.
In the context of the Sen’s proposal to describe D-branes
as open string field theory solitons
the massless modes representing transverse D-brane fluctuations
appear as collective coordinates.
The presence of these massless scalars is
a general phenomenon independent of details of an
effective field theory action, and is a consequence
of spontaneous breaking of the
translational symmetry [33].
The existence of a massless vector mode on the soliton
is related to
the fact that the tachyon is coupled to the (abelian)
open string
vector field in a non-minimal
way, i.e. only through the field strength $F_{\mu\nu}$.444Note that
one does not find massless vector modes and DBI actions
for the collective coordinates for brane
solitons in familiar scalar-vector field-theory systems
where
scalars are coupled to vector fields in a
minimal way.
The general structure
of the dependence of the effective action on $F_{\mu\nu}$
and its coupling to the tachyon $T$ is dictated
by the open string sigma model [3, 4, 26]:
the action is an integral of
a product of the Born-Infeld Lagrangian
for the vector field and some function of $T$ and its derivatives
contracted with functions of $F_{\mu\nu}$.
This structure is
the only essential assumption one needs
to show that the scalar and vector zero modes always
combine into the standard DBI action,
irrespective of all other details of the effective action.555The precise
value of the tension
of the resulting
Dp-brane does depend of course on details of an
effective string field theory action.
To reproduce the expected tension
one should compute the disc partition
at the proper conformal point, or minimize
the “trial actions” depending on a
finite number of parameters
obtained in the BSFT framework
in [23, 24].
While the universality of the DBI action holds irrespective of the
presence of a tachyonic mode in the spectrum of fluctuations around
the soliton, the form of an action for the tachyonic
mode666The
tachyonic mode is always present in open bosonic string theory
context, and is
absent on the kink solution describing condensation of an
unstable D9-brane tachyon to produce a D8-brane, which
is a stable BPS-object in type IIA theory. does depend on
details of an effective string field theory action (the
action for the tachyonic mode
on D24-brane was shown [34] to have the same form as the
one on the D25-brane).
We should add that while finding the DBI action
for the D-branes described by tachyonic solitons was
expected – after all, the required information is
contained in the open string sigma model which gives
both, the effective action for the original brane
and the action for massless fluctuations of the solitonic brane –
to see how this derivation works in detail seems
quite instructive. In particular, this clarifies
how the brane collective coordinates should appear
in the string world sheet action which may be relevant
for off-shell aspects of D-branes
(cf. [35]).
In Section 2 we shall
present a derivation of the DBI action from an
effective field theory action with the structure implied by
the open string sigma model. In Section 3 we shall
discuss how one can see the same
directly in the world-sheet
approach, identifying the boundary couplings of the D-brane
collective coordinates.
2 Effective field theory picture
2.1 Tachyon–vector actions
The string sigma model approach to
field theory
describing dynamics of an unstable D25-brane
in bosonic string theory or
a non-BPS D9-brane in type IIA string theory
leads to an effective action for the tachyon
and vector gauge fields
of the following general form (for a review, see [26])
777In the units we shall use $\alpha^{\prime}=2$, and our
normalizations for $T$ and $A_{\mu}$ are given in Section 3.
$$\displaystyle S=\int d^{d}x\sqrt{\det(\delta_{\mu\nu}+F_{\mu\nu})}\ {\cal L}(T%
,F,\partial T,\partial F,\partial\partial T,\ldots).$$
(1)
The tachyon field couples to the vector
field through its strength, i.e. through the combinations
(and their derivatives)888The case when the starting point is
a Dp-brane with a non-zero number of transverse dimensions
represented by massless scalars is obtained by “dimensionally reducing”
$F_{\mu\nu}$ matrix.
the
$$\displaystyle G^{\mu\nu}=\left(\frac{1}{1+F}\right)^{(\mu\nu)}=\left(\frac{1}{%
1-F^{2}}\right)^{\mu\nu}\ ,\ \ \ \ \ \ \ H^{\mu\nu}=\left(\frac{1}{1+F}\right)%
^{[\mu\nu]}=-\left(\frac{F}{1-F^{2}}\right)^{\mu\nu}\ ,$$
(2)
where $G^{\mu\nu}$ plays the role of an effective metric
as seen by open string excitations in
the presence of a constant gauge field background [4].
We will be interested in
deriving an effective action for massless modes on
a soliton solution up to “acceleration-dependent” (i.e.
second and higher derivative) terms.
For that reason we may restrict our consideration to
the terms in $S$ which do not
depend on derivatives of $F_{\mu\nu}$.
We will study only codimension one D-branes, i.e.
soliton solutions represented by the tachyon field
depending only on one
coordinate $x_{1}$ and the vector field having
constant field strength $F_{\mu\nu}$.
As will be shown below, such backgrounds
satisfy the equations of motion for
the vector field only if
$F_{1i}=0$ ($i=2,...10$), i.e., if the matrices in (2)
obey
$G^{1i}=H^{1i}=0$.
Since in deriving the effective action for massless modes we drop
their
second and higher derivatives, none of the
possible $H^{\mu\nu}$
dependent terms in (1)
contribute to the effective action,
i.e. we are allowed to use
an action which depends only on $G^{\mu\nu}$ and $T$ and its
derivatives of
any order, i.e.
$$\displaystyle S=\int d^{d}x\sqrt{\det(\delta_{\mu\nu}+F_{\mu\nu})}\ {\cal L}(G%
^{\mu\nu}(F),T,\partial T,\partial\partial T,\ldots)\ .$$
(3)
Though any
action of the form (3) will
lead to the DBI action for the massless soliton modes,
it is useful to recall several explicit
actions of that type
which were previously discussed in the
context of tachyon condensation.
The low-energy two-derivative effective
action for a D25-brane in open bosonic string theory is (3)
with ${\cal L}$ given by [32, 26]
$$\displaystyle{\cal L}_{25}=T_{25}\ {\rm e}^{-T}\ (1+T)\bigl{(}1+G^{\mu\nu}%
\partial_{\mu}T\partial_{\nu}T\bigr{)}.$$
(4)
The corresponding expression for the
non-BPS D9-brane in type IIA theory is [24, 26]
$$\displaystyle{\cal L}_{9}=T_{9}\ {\rm e}^{-\frac{1}{4}T^{2}}\left(1+G^{\mu\nu}%
\partial_{\mu}T\partial_{\nu}T+{1\over 2}\log(\frac{4}{{\rm e}})\ G^{\mu\nu}T^%
{2}\partial_{\mu}T\partial_{\nu}T\right).$$
(5)
The actions (4) and (5)
are obtained in special schemes chosen to
reproduce the standard
values of the tachyon masses around the perturbative $T=0$ vacua.
The actions (4) and (5)
(or similar model actions in [17, 18])
have soliton solutions representing codimension one D-branes, but they
do not reproduce the expected D-brane tensions.
However, one can find
“interpolating” actions which
coincide with (4) and (5) in the
two-derivative approximation and lead to
the correct D-brane tensions.
In particular, for a non-BPS D9-brane such action
has [24, 31, 26]
$$\displaystyle{\cal L}_{9}=T_{9}\ {\rm e}^{-\frac{1}{4}T^{2}}\left(\sqrt{\pi}%
\frac{\Gamma(1+G^{\mu\nu}\partial_{\mu}T\partial_{\nu}T)}{\Gamma(\frac{1}{2}+G%
^{\mu\nu}\partial_{\mu}T\partial_{\nu}T)}+\log(\frac{4}{{\rm e}})\ G^{\mu\nu}T%
\partial_{\mu}\partial_{\nu}T\right).$$
(6)
Here the first term can be shown to coincide with the action
found in [24, 31], and the scheme-dependent coefficient
of the last
term is chosen to reproduce the
correct value of the tachyon mass at
$T=0$ [26].
The action (3),(6)
is given by the open superstring partition function
on a disc.
In the sigma model approach
the tachyon coupling at the boundary of the disc
can be represented as
$$\displaystyle T(x(\tau))=T(x)+\xi^{\mu}(\tau)\partial_{\mu}T(x)+\frac{1}{2}\xi%
^{\mu}(\tau)\xi^{\nu}(\tau)\partial_{\mu}\partial_{\nu}T(x)+\cdots,$$
(7)
where $\tau$ is the angular coordinate on the boundary of the disc,
and $x(\tau)=x+\xi(\tau)$, $\int d\tau\xi(\tau)=0$.
The derivatives of the tachyon field are then treated as
independent, and the
partition function is computed exactly in $\partial_{\mu}T$, and up to
the first order in $\partial_{\mu}\partial_{\nu}T$.
2.2 Collective coordinates and DBI action
We shall assume that the equations for $T$ and $A_{\mu}$
following from (3)
have a soliton solution with
$$\displaystyle T=T(x_{1})\ ,\ \ \ \ \ \ F_{\mu\nu}={\rm const}\ .$$
(8)
Following Sen [11],
we shall identify the solution with a D(d-2)-brane.
For example, for model two-derivative actions
one finds the kink
$T(x)=ux_{1}$ [18, 24] in the case of the type IIA
D9-brane, and the lump $T(x)=ux_{1}^{2}$
[17, 23] in the case of the bosonic D25-brane.
These solutions represent exact conformal field theories in the limit
$u\rightarrow\infty$ [13, 23, 24].
It is straightforward to see
that for any solution for $T$ found for $F_{\mu\nu}=0$
there is a corresponding solution for $F_{\mu\nu}=const$.
Since the vector potential $A_{\mu}$ is a dynamical field,
one is still to check that its equation
is satisfied for the above background.
To do that
we shall restrict ourselves to
the simplest case when the action can be put in the form
$$\displaystyle S=\int d^{d}x\sqrt{\det(\delta_{\mu\nu}+F_{\mu\nu})}\ {\cal L}(G%
^{\mu\nu}(F)\partial_{\mu}T\partial_{\nu}T,T)\ .$$
(9)
The explicit examples of the
actions considered above are all of that type.
Then varying (9) with respect to
$A_{\mu}$ one gets (assuming (8))
$$\displaystyle\partial_{\mu}\Biggl{[}(I+F)^{-1}_{[\mu\nu]}\ {\cal L}+2(I+F)^{-1%
}_{1[\mu}(I+F)^{-1}_{\nu]1}(\partial_{1}T)^{2}{\cal L}^{\prime}\Biggr{]}=0.$$
(10)
Here ${\cal L}={\cal L}(K,T),\ K\equiv G^{\mu\nu}\partial_{\mu}T\partial_{\nu}T$
and the derivative ${\cal L}^{\prime}$
is with respect to the first argument $K$.
Since ${\cal L}$ depends only on $x_{1}$
this equation reduces to
$$\displaystyle(I+F)^{-1}_{[\nu 1]}\ \partial_{1}\bigl{[}{\cal L}-2G^{11}(%
\partial_{1}T)^{2}{\cal L}^{\prime}\bigr{]}=0\ .$$
(11)
This equation is satisfied if
$$\displaystyle(I+F)^{-1}_{[\nu 1]}=\bigl{(}\frac{F}{1-F^{2}}\bigr{)}_{1\nu}=0\ ,$$
(12)
i.e. if $F_{1\nu}=0$, or if
$$\displaystyle{\cal L}-2G^{11}(\partial_{1}T)^{2}{\cal L}^{\prime}={\rm const}\,.$$
(13)
The latter condition does not hold
for a generic Lagrangian,
so we will always require that $F_{1\nu}=0$.
Then
the effective metric $G^{\mu\nu}$ simplifies, so that
$$\displaystyle G^{11}=1,\ \ \ \quad G^{1i}=0.$$
(14)
Note also that in the axial gauge, $A_{1}(x_{1},x_{i})=0$,
the condition
$F_{1\nu}=0$ implies that $A_{i}(x_{i})$ does
not depend on $x_{1}$, as one would
expect for a massless
vector field on a D(d-2)-brane.
It follows from the translational invariance
of the action (3)
that fluctuations around
the soliton solution contain a zero mode $\Phi(x_{i})$, which
satisfies the equation of motion for a massless scalar
in $d-1$ dimensions.
As usual (see, e.g., [33]),
it is natural to identify this zero mode
with a collective coordinate
describing fluctuations of the D(d-2)-brane
in the transverse $x_{1}$ direction.
For small fluctuations and in the semiclassical approximation,
the effective action
for the zero mode is obtained by substituting
$T\rightarrow T(x_{1})-\partial_{1}T(x_{1})\Phi(x_{i})$
into (3),
and integrating over $x_{1}$.
To describe arbitrary fluctuations in the semiclassical
approximation one may then consider the standard ansatz
$T\rightarrow T(x_{1}-\Phi(x_{i}))$.
However, in general, one should take into account
the global $SO(d)_{G}$
invariance of the action
(3) implied by the open string sigma model.
The subscript
$G$ means that the general linear
transformations of the $d$
coordinates $x\rightarrow\Lambda x$ should
preserve the constant metric $G^{\mu\nu}$, i.e.
$$\Lambda G\Lambda^{T}=G.$$
Assuming that $\Lambda$ has the form
$$\displaystyle\Lambda_{\mu\nu}=\frac{1}{\beta}\tilde{\Lambda}_{\mu\nu},\quad\ %
\ \tilde{\Lambda}_{11}=1,\ \ \quad\tilde{\Lambda}_{1i}=-V_{i}\ ,$$
(15)
we get
$$\displaystyle\beta=\sqrt{1+G^{ij}V_{i}V_{j}}\ .$$
(16)
Consider now the following ansatz for the
tachyon in terms of the zero mode $\Phi(x_{i})$
$$\displaystyle T=T(y_{1})\ ,\ \ \ \ \ \ \ \ y_{1}\equiv\frac{x_{1}-\Phi(x_{i})}%
{\beta(\partial\Phi,F)}\ ,$$
(17)
$$\displaystyle\beta(\partial\Phi,F)=\sqrt{1+G^{ij}(F)\partial_{i}\Phi\partial_{%
j}\Phi}\ .$$
(18)
Note that if $\Phi(x_{i})$ is
of the form $\Phi(x_{i})=V_{i}x^{i}$
(i.e. brane has “constant velocity”),
then $y_{1}=(\Lambda x)_{1}$ and $\Lambda$ belongs to the
invariance group $SO(d)_{G}$.
Taking into account
that the full tachyon field in (17)
depends only on one combination $y_{1}$
of coordinates, we can rewrite (3) as follows
(ignoring higher-derivative $\partial^{n}\Phi,\,n>1$ terms)
$$\displaystyle S=T_{d-2}\int d^{d-1}x\ \beta(\partial\Phi,F)\sqrt{\det(\delta_{%
\mu\nu}+F_{\mu\nu})}\ ,$$
(19)
where the integral over $y_{1}$
$$\displaystyle T_{d-2}\equiv\int dy_{1}{\cal L}(T(y_{1}),\partial_{1}T(y_{1}),%
\partial_{1}\partial_{1}T(y_{1}),\ldots)$$
(20)
does not depend on the field strength999As follows from
the sigma model considerations,
$G^{\mu\nu}$ can be
contracted only with the tachyon derivative terms.
Thus only $G^{11}$ could
appear in (19), but it is equal to 1 according to
(14).
and thus
defines the D(d-2)-brane tension.
Taking into account that (for $F_{i1}=0$)
$$\displaystyle\beta(\partial\Phi,F)\sqrt{\det(\delta_{\mu\nu}+F_{\mu\nu})}=%
\sqrt{\det(\delta_{ij}+F_{ij}+\partial_{i}\Phi\partial_{j}\Phi)}\ ,$$
(21)
we arrive at the standard DBI action for a D(d-2)-brane
$$\displaystyle S=T_{d-2}\int d^{d-1}x\sqrt{\det(\delta_{ij}+F_{ij}+\partial_{i}%
\Phi\partial_{j}\Phi)}\ .$$
(22)
To get the DBI action it was essential to
introduce the collective coordinate dependence as
in (17).
Indeed, the above derivation is a generalization of the argument
used to show that a point-like soliton of a relativistically
invariant (e.g. 2-d) scalar field
theory is described by the standard particle
action $M\int dt\sqrt{1-v^{2}}$: one starts with the static
solution and applies a boost $x\rightarrow{x-vt\over\sqrt{1-v^{2}}}$.
The ansatz (17) of course reduces to
$T(x_{1})-\partial_{1}T(x_{1})\Phi(x_{i})$ for small fluctuations,
but the real reason for the specific non-linear structure
of (17) is to ensure the decoupling
of massive solitonic modes from the massless mode,
so that there is no term linear in the massive modes in the action
(in the approximation where
we neglect second and higher derivatives of $\Phi$).
Thus one can consistently put all massive modes to zero, and consider only
the
$\Phi$-dependent terms.101010This is somewhat
similar to what happens
in consistent Kaluza-Klein reductions of compactified gravity models.
If one would use the standard form of the
collective coordinate dependence $T=T(x_{1}-\Phi(x_{i}))$
then to recover
the DBI action one would have to take into account
the massive mode contributions
to the low-energy effective action and at the end
to make a proper non-linear
redefinition of the collective coordinate $\Phi$.
3 World-sheet picture
Here we shall relate the discussion of the
DBI action from the effective field theory point of view
in the previous section to the
world-sheet sigma model considerations,
with the aim to identify the boundary couplings
corresponding to the tachyonic soliton collective coordinate.
3.1 Standard boundary couplings
The starting point in the usual perturbative description
of D-branes in flat space
is an exact free 2-d conformal field
theory in the bulk with
$p+1$ Neumann and $d-p-1$ Dirichlet boundary conditions.
The marginal boundary interactions are then represented by
the boundary sigma model action [2, 1]111111For the sake
of clarity here we shall
consider only the bosonic string theory
but will ignore tachyonic coupling on the resulting brane.
$$\displaystyle I_{\partial\Sigma}=\frac{1}{4\pi}\int_{\partial\Sigma}d\tau%
\biggl{[}\sum_{i=1}^{p+1}A_{i}(x^{1},...,x^{p+1})\dot{x}^{i}+\sum_{a=p+2}^{d}%
\Phi_{a}(x^{1},...,x^{p+1})\partial_{n}{x}^{a}\biggr{]}\,,$$
(23)
where the couplings
$A_{i}$ and $\Phi_{a}$ are the gauge field and the scalars on
the world-volume of the $Dp$-brane, respectively ($\partial_{n}$
is the normal derivative to the boundary $\partial\Sigma$).
The open string sigma model
partition function on the disc is then the DBI
action for $A_{i}$ and $\Phi_{a}$ (ignoring terms
higher than first derivative in the fields).
On the other hand, to describe tachyon condensation
from a space-filling brane to a
lower-dimensional brane one should start
with the following boundary theory
$$\displaystyle I_{\partial\Sigma}=\frac{1}{4\pi}\int_{\partial\Sigma}d\tau\bigl%
{[}T(x)+A_{\mu}(x)\dot{x}^{\mu}\bigr{]}\ .$$
(24)
In the presence of this interaction the standard Neumann boundary
conditions for the open string coordinate
fields $x^{\mu}$ are modified
to
$$\displaystyle\partial\Sigma:\ \ \ \ \ \ \ \ \partial_{n}x_{\mu}+F_{\mu\nu}\dot%
{x}^{\nu}+\partial_{\mu}T=0\ .$$
(25)
A particularly simple case for which
the open bosonic world-sheet theory remains solvable is obtained by taking
$A_{\mu}$ to have a constant field strength and switching on
the tachyon with a
quadratic profile in one direction, i.e.
$T(x)=a+\frac{1}{2}u\,x^{2}_{1}$. In this case the boundary conditions
are
$$\displaystyle\partial\Sigma:\ \ \ \ \ \partial_{n}x_{1}+F_{1i}\dot{x}_{i}+u\,x%
_{1}=0,~{}~{}~{}~{}~{}~{}~{}~{}~{}\partial_{n}x_{i}+F_{i1}\dot{x}_{1}+F_{ij}%
\dot{x}_{j}=0\,.$$
(26)
For $u=0$ these
are the usual Neumann conditions modified by the
presence of the constant strength $F_{\mu\nu}$ (and thus the
corresponding
partition function is the BI action for $F_{\mu\nu}$ with
$\mu=1,\ldots,d$).
On the other hand, when the tachyon condenses
into the vacuum $a,u\rightarrow\infty$, the field $x_{1}$ becomes
constrained to vanish at the boundary, i.e.
is subject to
the Dirichlet boundary condition ($x_{1}|_{\partial\Sigma}=0$). This
leads to a complete decoupling of $F_{1i}$ components
in (24),(26).
Thus, in the IR fixed point one gets the BI action for $F_{ij}$
components only.
Following this approach, it appears
that we are missing the dependence
on the massless scalar fields describing
the transverse motion of the brane.
To account for them, one should
understand how the normal derivative coupling of the type
introduced from the beginning in the standard D-brane description
(23) is effectively induced
in the process of tachyon condensation.
This is what we are going to explain below.
3.2 Generalized boundary sigma model and collective coordinate
coupling
From the sigma model viewpoint,
the role of the tachyonic field is to control
a number of “free” (Neumann)
space-time dimensions, and in this respect
the condensation may be compared with
dimensional reduction. However,
in the process of reduction
the BI action in $p+1$ dimensions
does lead to
the DBI action in $p$ dimensions with $A_{1}\rightarrow\Phi$.
The procedure of dimensional
reduction is of course formally related
to T-duality transformation
along one world-volume dimension $x_{1}$ (implying that $A_{1}$
becomes a scalar field describing the transverse fluctuations).
In the world-sheet description,
the T-duality transformation simply exchanges the
normal and tangential derivatives in the boundary vertex
operators [1].
This suggests that
in order to make contact with the D-brane description
based on (23)
one may try to generalize the boundary action
(24) by adding from the beginning an
additional
normal derivative coupling (with a new coefficient
function $V_{\mu}$)121212Such sigma model
and its renormalization was discussed in detail in [36]
and references there.
$$\displaystyle I_{\partial\Sigma}=\frac{1}{4\pi}\int_{\partial\Sigma}d{\tau}%
\biggl{[}T(x)+A_{\mu}(x)\dot{x}^{\mu}+V_{\mu}(x)\partial_{n}x^{\mu}\biggr{]}\,.$$
(27)
Adding $V_{\mu}$ term may seem
irrelevant – in
the standard perturbative vacuum where all $x^{\mu}$
are subject to
the Neumann boundary conditions this term
decouples: computing correlation functions or
the disc partition function in perturbation theory
in powers of couplings one finds that they do not
depend on $V_{\mu}$.131313There exists a (point-splitting)
regularization
preserving the Neumann boundary conditions,
and, therefore, any correlation
function involving a normal-derivative term vanishes.
Note also that lifted to the bulk of the world sheet,
this coupling becomes $\int_{\Sigma}\partial^{a}(V_{\mu}(x)\partial_{a}x^{\mu})$,
and thus (modulo a term proportional to the $x^{\mu}$ equation of motion)
redefines the target space metric by a diffeomorphism-type
term.
More precisely, one
should determine the boundary conditions
dynamically,
by minimizing the total string action $I=I_{\Sigma}+I_{\partial\Sigma}$.
Then (depending on a calculation procedure)
the normal-derivative term may contribute
to the partition
function, but this still does not imply
the appearance
of a new physical degree of freedom associated with $V_{\mu}$.
Indeed, one may interpret such normal derivative couplings
as additional pure gauge modes
of open string theory, which can be removed
by a gauge choice.
Indeed, the general form of the
boundary conditions should be
$$\displaystyle\partial\Sigma:\ \ \ \ \ \ \partial_{n}x_{\mu}+N_{\mu\nu}\dot{x}^%
{\nu}+N_{\mu}=0,$$
(28)
where $N_{\mu\nu}$ and $N_{\mu}$ should be determined
from the requirement that
the variation of the total string action ${1\over 8\pi}\int_{\Sigma}(\partial x)^{2}+I_{\partial\Sigma}$ vanishes.
One finds then the following equations141414These
boundary conditions differ from the ones used in [36]
and references there.
$$\displaystyle N_{\mu}=\partial_{\mu}\tilde{T}\ ,\ \ \ \ \ \tilde{T}\equiv T-N_%
{\nu}V^{\nu}\ ,$$
(29)
$$\displaystyle N_{\mu\nu}=\tilde{F}_{\mu\nu}\ ,\ \ \ \ \ \ \tilde{F}_{\mu\nu}%
\equiv F_{\mu\nu}-\partial_{\mu}(N_{\rho\nu}V^{\rho})+\partial_{\nu}(N_{\rho%
\mu}V^{\rho})\,.$$
(30)
It is then easy to see (cf. (28),(25)) that
one can replace (27) by the usual
boundary term (24) without $V_{\mu}$ but with transformed
tachyon
$T\rightarrow\tilde{T}$ and vector $A_{\mu}\rightarrow\tilde{A}_{\mu}$ fields:
$$\displaystyle T=(1+V^{\mu}\partial_{\mu})\tilde{T}\ ,\ \ \ \ \ \ \ \ \ \ A_{%
\mu}=\tilde{A}_{\mu}-\tilde{F}_{\mu\nu}V^{\nu}\,.$$
(31)
Thus adding the normal-derivative term
amounts simply to a redefinition of the
original open string sigma model couplings in (24).
The central point, however, is that
in the presence of a nontrivial
tachyon condensate the normal derivative coupling
becomes relevant – it provides an adequate
description of dynamics of the soliton translational zero modes,
allowing one to recover the usual
D-brane description in the IR.
Assuming that there is
a tachyonic condensate breaking
translational invariance in $x_{1}$-direction,
and that $V_{1}$ depends only on
$x_{i}$, one sees that
$V_{1}$ can be identified with
the translational collective coordinate $\Phi$.
Indeed, for small $V_{1}$
$$\displaystyle\tilde{T}\approx T-V_{1}\partial_{1}T\approx T(x_{1}-V_{1}),\quad%
\ \ \ \ \tilde{A_{i}}\approx A_{i}-V_{1}\partial_{1}A_{i}\approx A_{i}(x_{1}-V%
_{1}),$$
(32)
which is the standard collective coordinate dependence.
Another argument for this identification can be given by
analyzing the boundary conditions (28) in the IR fixed point
$a,u\rightarrow\infty$ for the quadratic tachyon coupling
$T(x_{1})=a+\frac{1}{2}ux_{1}^{2}$.
At this point the boundary condition (28) reduces to the
following Dirichlet condition for $x_{1}$
$$\displaystyle\partial\Sigma:\ \ \ \ \ \ \ \ \ x_{1}=\Phi(x_{i})\,,\ \ \ \ \ \ %
\Phi\equiv V_{1}(x_{i})\ ,$$
(33)
and the modified Neumann condition for $x_{i}$ (cf. (26))
$$\displaystyle\partial\Sigma:\ \ \ \ \ \ \ \ \partial_{n}x_{i}+\partial_{i}\Phi%
\partial_{n}x_{1}+F_{ij}\dot{x}_{j}=0\,.$$
(34)
The boundary conditions (33) and (34) define the mixed
Dirichlet-Neumann sigma-model.
The vanishing of the one-loop $\beta$-functions
for this model is known [2]
to be equivalent to the equations of motion
for space-time fields $A_{i}$ and $\Phi$ following from DBI.
One may ask then why in order
to derive the DBI action in Section 2 we needed to
use the non-linear ansatz for the tachyon (17)
instead of the
transformed tachyon and gauge fields in (32).
The redefined tachyon coupling $\tilde{T}$
with the $V_{1}=\Phi$ -dependence as in (17)
can be obtained by starting with the boundary action
(27)
with
normal-derivative couplings of an appropriate
non-linear form $f(V,\partial_{n}x,\dot{x})$.
This would modify the boundary conditions (28)
and the $V_{\mu}$-dependence of the transformed tachyon and vector fields.
From the sigma model point of view the
difference between the linearized
collective coordinate dependence
in (32) and the non-linear
one in (17) should be in contact terms only.
Such terms are crucial for implementing the symmetries of
string theory in a manifest way (see, e.g., [37]),
but they do not
contribute to the (scheme-independent)
one-loop $\beta$-functions.
This explains why one can
reproduce correct $\beta$-functions corresponding to the
DBI action in the approach of [2]
using the linear expressions
(32).
To derive the DBI action as the open string sigma model
partition function (which is sensitive to a choice of a scheme
preserving underlying symmetries of the theory)
one needs, however, to use
the non-linear ansatz (17)
for the collective coordinate dependence,
substituting it into the tachyon expansion
(7) and integrating over the fluctuations
$\xi^{\mu}$.
Acknowledgements
G.A. was supported by the DFG and by the European Commission
RTN programme HPRN-CT-2000-00131 in which G.A. and S.T.
are associated to U. Bonn,
and in part by RFBI grant N99-01-00166 and by INTAS-99-1782.
S.T. also acknowledges support from GIF,
the German-Israeli foundation
for Scientific Research.
S.F. and A.A.T are supported
by the U.S. Department of Energy under grant
No. DE-FG02-91ER-40690. S.F. is also supported
in part by RFBI grant N99-01-00190, and
A.A.T. – by the
INTAS project 991590
and PPARC SPG grant PPA/G/S/1998/00613.
References
[1]
J. Dai, R. Leigh and J. Polchinski,
New connections between string theories,
Mod. Phys. Lett. A4, 2073 (1989).
[2]
R. Leigh, Dirac-Born-Infeld action from Dirichlet sigma
model,
Mod. Phys. Lett. A4, 2767 (1989).
[3]
E.S. Fradkin and A.A. Tseytlin,
“Nonlinear Electrodynamics From Quantized Strings,”
Phys. Lett. B163, 123 (1985).
[4]
A. Abouelsaood, C. G. Callan, C. R. Nappi and S. A. Yost,
“Open Strings In Background Gauge Fields,”
Nucl. Phys. B280, 599 (1987).
[5]
C. Bachas,
“D-brane dynamics,”
Phys. Lett. B374, 37 (1996)
[hep-th/9511043].
[6]
A.A. Tseytlin,
“Self-duality of Born-Infeld action and Dirichlet 3-brane of type IIB
superstring theory,”
Nucl. Phys. B469, 51 (1996)
[hep-th/9602064].
H. Dorn and H. J. Otto,
“Remarks on T-duality for open strings,”
Nucl. Phys. Proc. Suppl. 56B, 30 (1997)
[hep-th/9702018].
[7]
G. T. Horowitz and A. Strominger,
“Black strings and P-branes,”
Nucl. Phys. B360, 197 (1991).
[8]
M. J. Duff and J. X. Lu,
“Elementary five-brane solutions of D = 10 supergravity,”
Nucl. Phys. B354, 141 (1991).
[9]
J. Polchinski,
“Dirichlet-Branes and Ramond-Ramond Charges,”
Phys. Rev. Lett. 75, 4724 (1995)
[hep-th/9510017].
[10]
C. G. Callan, J. A. Harvey and A. Strominger,
“Worldbrane actions for string solitons,”
Nucl. Phys. B367, 60 (1991);
M. J. Duff, R. R. Khuri and J. X. Lu,
“String solitons,”
Phys. Rept. 259, 213 (1995)
[hep-th/9412184];
D. M. Kaplan and J. Michelson,
“Zero Modes for the D=11 Membrane and Five-Brane,”
Phys. Rev. D53, 3474 (1996)
[hep-th/9510053].
[11]
A. Sen,
“Descent relations among bosonic D-branes,”
Int. J. Mod. Phys. A14, 4061 (1999)
[hep-th/9902105];
“Stable non-BPS bound states of BPS D-branes,”
JHEP 9808, 010 (1998)
[hep-th/9805019];
“Tachyon condensation on the brane antibrane system,”
JHEP 9808, 012 (1998)
[hep-th/9805170];
[12]
A. Sen,
“Non-BPS states and branes in string theory,”
hep-th/9904207;
“Supersymmetric world-volume action for non-BPS D-branes,”
JHEP 9910, 008 (1999)
[hep-th/9909062].
[13]
J. A. Harvey, D. Kutasov and E. J. Martinec,
“On the relevance of tachyons,”
hep-th/0003101.
C. G. Callan, I. R. Klebanov, A. W. Ludwig and J. M. Maldacena,
“Exact solution of a boundary conformal field theory,”
Nucl. Phys. B422, 417 (1994)
[hep-th/9402113].
[14]
E. Witten, “Noncommutative geometry and string
field theory”, Nucl. Phys. B268, 253 (1986).
[15]
V. A. Kostelecky and S. Samuel,
“On A Nonperturbative Vacuum For The Open Bosonic String,”
Nucl. Phys. B336, 263 (1990);
A. Sen and B. Zwiebach,
“Tachyon condensation in string field theory,”
JHEP 0003, 002 (2000)
[hep-th/9912249].
J. A. Harvey and P. Kraus,
“D-branes as unstable lumps in bosonic open string field theory,”
JHEP 0004, 012 (2000)
[hep-th/0002117].
N. Moeller and W. Taylor,
“Level truncation and the tachyon in open bosonic string field
theory,”
Nucl. Phys. B583, 105 (2000)
[hep-th/0002237];
R. de Mello Koch, A. Jevicki, M. Mihailescu and R. Tatar,
“Lumps and p-branes in open string field theory,”
Phys. Lett. B482, 249 (2000)
[hep-th/0003031].
N. Moeller, A. Sen and B. Zwiebach,
“D-branes as tachyon lumps in string field theory,”
JHEP 0008, 039 (2000)
[hep-th/0005036].
[16]
K. Dasgupta, S. Mukhi and G. Rajesh,
“Noncommutative tachyons,”
JHEP 0006, 022 (2000)
[hep-th/0005006];
J. A. Harvey, P. Kraus, F. Larsen and E. J. Martinec,
“D-branes and strings as non-commutative solitons,”
JHEP 0007, 042 (2000)
[hep-th/0005031];
J. A. Harvey, P. Kraus and F. Larsen,
“Exact noncommutative solitons,”
hep-th/0010060.
[17]
J. A. Minahan and B. Zwiebach,
“Field theory models for tachyon and gauge field string dynamics,”
JHEP 0009, 029 (2000)
[hep-th/0008231].
[18]
J. A. Minahan and B. Zwiebach,
“Effective tachyon dynamics in superstring theory,”
hep-th/0009246.
[19]
J. A. Minahan and B. Zwiebach,
“Gauge fields and fermions in tachyon effective field theories,”
hep-th/0011226.
[20]
E. Witten, “On background-independent
open-string field theory,” Phys. Rev. D46, 5467
(1992), hep-th/9208027.
“Some computations in background-independent
off-shell string theory,” Phys. Rev. D47, 3405
(1993), hep-th/9210065.
[21]
S. Shatashvili, “Comment on the background
independent
open string theory,” Phys. Lett. B311, 83
(1993),
hep-th/9303143;
“On the problems with background independence in string theory,”
hep-th/9311177.
[22]
A. A. Gerasimov and S. L. Shatashvili, “On exact tachyon
potential in open string field theory”, hep-th/0009103.
[23]
D. Kutasov, M. Marino and G. Moore,
“Some exact results on tachyon condensation in string field theory”,
JHEP 0010, 045 (2000) hep-th/0009148.
[24]
D. Kutasov, M. Marino and G. Moore,
“Remarks on tachyon condensation in superstring field theory”,
hep-th/0010108.
[25]
E.S. Fradkin and A.A. Tseytlin,
“Quantum String Theory Effective Action,”
Nucl. Phys. B261, 1 (1985).
[26]
A.A. Tseytlin,
“Sigma model approach to string theory effective actions with tachyons,”
hep-th/0011033.
[27]
A. Sen,
“Universality of the tachyon potential,”
JHEP 9912, 027 (1999)
[hep-th/9911116].
[28]
K. Li and E. Witten,
“Role of short distance behavior
in off-shell open string field
theory,”
Phys. Rev. D48, 853 (1993)
[hep-th/9303067].
[29]
L. Cornalba,
“Tachyon condensation in large magnetic fields with background independent
string field theory,”
hep-th/0010021.
[30]
K. Okuyama,
“Noncommutative tachyon from background independent open string field
theory,”
hep-th/0010028.
[31]
O. Andreev,
“Some computations of partition functions and
tachyon potentials in background independent
off-shell string theory,”
hep-th/0010218.
[32]
A. A. Gerasimov and S. L. Shatashvili,
“Stringy Higgs mechanism and the fate of open strings,”
hep-th/0011009.
[33]
N. H. Christ and T. D. Lee,
“Quantum Expansion Of Soliton Solutions,”
Phys. Rev. D12, 1606 (1975);
J.-L. Gervais, A. Jevicki and B. Sakita,
“Perturbation expansion around extended-particle states in quantum
field
theory,”
Phys. Rev. D12, 1038 (1975);
E. Tomboulis,
“Canonical Quantization Of Nonlinear Waves,”
Phys. Rev. D12, 1678 (1975).
[34]
S. Moriyama and S. Nakamura,
“Descent relation of tachyon condensation from boundary string
field theory,”
hep-th/0011002.
[35]
W. Fischler, S. Paban and M. Rozali,
“Collective Coordinates for D-branes,”
Phys. Lett. B381, 62 (1996)
[hep-th/9604014].
S. Rey,
“Collective coordinate quantization of Dirichlet branes,”
hep-th/9604037.
Y. Kazama,
“Scattering of quantized Dirichlet particles,”
Nucl. Phys. B504, 285 (1997)
[hep-th/9705111].
[36]
W. Kummer and D. V. Vassilevich,
“Renormalizability of the open string sigma model and emergence of
D-branes,”
JHEP 0007, 012 (2000)
[hep-th/0006108].
[37]
M.B. Green and N. Seiberg,
“Contact Interactions In Superstring Theory,”
Nucl. Phys. B299, 559 (1988).
O.D. Andreev and A.A. Tseytlin,
“Partition Function Representation For The Open
Superstring Effective Action: Cancellation Of Mobius
Infinities And Derivative Corrections To
Born-Infeld Lagrangian,”
Nucl. Phys. B311, 205 (1988).
[38] |
Coherent hard x-rays from attosecond
pulse train-assisted harmonic generation
Michael Klaiber${}^{1,2}$, Karen Z. Hatsagortsyan${}^{1}$, Carsten Müller${}^{1}$ and Christoph H. Keitel${}^{1,*}$
Abstract
High-order harmonic generation from atomic systems is considered in the crossed fields of a relativistically strong infrared laser and a weak attosecond-pulse train of soft x-rays. Due to one-photon ionization by the x-ray pulse, the ionized electron obtains a starting momentum that compensates the relativistic drift which is induced by the laser magnetic field, and allows the electron to efficiently emit harmonic radiation upon recombination with the atomic core in the relativistic regime. In this way, short pulses of coherent hard x-rays of up to 40 keV energy and 10 as duration can be brought about.
\address
${}^{1}$Max-Planck Institut für Kernphysik, Saupfercheckweg 1, D-69117 Heidelberg, Germany
${}^{2}$Theoretische Quantendynamik, Physikalisches Institut der Universität, H.-Herder-Str. 3, D-79104 Freiburg, Germany
${}^{*}$Corresponding author: [email protected]
\ocis
020.2649 Strong field laser physics; 020.4180 Multiphoton processes; 320.7120 Ultrafast phenomena
Coherent, short x-ray pulses are auspicious both for fundamental physics and applied science [1]. One of the successful ways to produce coherent short-wavelength radiation is based on high-order harmonic generation (HHG) [2].
With this state-of-the-art technique, coherent x-ray photons of about 1 keV [3] and short XUV pulses of about $100$ as [4] are generated via HHG in atomic gas jets. A further increase of the emitted photon energies cannot be achieved straightforwardly by an enhancement of the driving laser intensity as the interaction regime becomes relativistic and the returning electron misses the ionic core due to the relativistic drift that suppresses HHG [5]. Various methods to counteract the relativistic drift have been proposed, which have either limited scope of applicability, or are challenging to realize [6, 7]. On a different front, the attosecond laser technique is developing [2]. In particular, attosecond pulses superimposed onto an infrared laser field have recently been used to realize an attosecond streak camera [8] and to control HHG spectra in the nonrelativistic regime [9].
In this letter we investigate HHG in the relativistic regime which is driven by an infrared laser field crossed with an attosecond pulse train (APT) of soft x-ray radiation. In this case, the APT not only fixes the ionization time [9], but also induces an initial electron momentum that can compensate the subsequent relativistic drift in the infrared laser field. This permits electron recollisions with the atomic core and efficient HHG in the relativistic domain. We employ a crossed-beam setup where the polarization direction of the APT is parallel to the propagation direction of the laser field. Then the electron emission probability in the direction along or opposite the laser propagation is the largest and, in the second case, the compensation of the relativistic drift is most efficient (see Fig. 1). We calculate the HHG yield in this setup in the moderately relativistic regime and show a large increase in the HHG yield as compared to the conventional case of a driving laser field only.
Let us see how large the required energy $\Omega$ of the x-ray photons (equivalent to the carrier frequency of the APT)
has to be in order to compensate the relativistic drift. The kinetic energy of the electron directly
after one-photon ionization by the x-ray pulse is
$$\displaystyle\varepsilon_{0}=\Omega-I_{p},$$
(1)
with the ionization potential $I_{p}$ (atomic units are used throughout). For the drift compensation, the initial energy has to be of the same order of magnitude as the drift energy: $\varepsilon_{0}\approx p_{d}^{2}/2\approx c^{2}\xi^{4}/32$, with the drift momentum $p_{d}\approx c\xi^{2}/4$ [5], the relativistic field strength parameter $\xi=E_{0}/c\omega$, the electric field of the laser $E_{0}$, the laser frequency $\omega$, and the speed of light $c$. The regime is relativistic when the electron drift distance is larger than the wave-packet spreading: $\sqrt{2I_{p}}(\xi^{3}/16)(c/\omega)>1$ [5]. The intensity should be restricted, though, to prevent over-the-barrier ionization: $\xi<I_{p}^{3/2}/(4c\omega)$ [5]. These conditions are satisfied in the relativistic tunneling regime with $\xi>0.14$ and $I_{p}>2.4$ a.u. We employ a moderately relativistic infrared laser field of $\xi\approx 0.33$ (ponderomotive energy $U_{p}=500$ a.u.) and $I_{p}=5.5$ a.u. (e.g., Be${}^{2+}$ ions). Then the drift compensation demands soft x-rays of $\Omega\approx 9$ a.u. ($245$ eV) [10]. Under these conditions, a HHG cutoff $\omega_{c}\approx 3.17U_{p}$ of $43$ keV is attained.
We consider the interaction of an atomic system with the following superposition of a laser field and an APT.
The laser field is propagating in $z$-direction and defined by the vector potential $\mathbf{A}(\eta)=\mathbf{e}_{x}(cE_{0}/\omega)\cos\eta$, with $\eta=\omega(t-z/c)$
and the unit vector $\mathbf{e}_{x}$ in polarization direction.
The pulses in the APT are Gaussian-shaped, propagate in $y$-direction, and have the electric field
$\mathbf{E}^{a}(y,t)=\mathbf{e}_{z}E^{a}_{0}g(y,t)\sin\Omega(t-y/c)$, with the unit vector $\mathbf{e}_{z}$ in polarization direction,
the envelope $g(y,t)=\exp(-2(t-y/c-t_{0})^{2}/\tau^{2})$, the time delay $t_{0}$ between the laser field and the APT, and the width $\tau$ of the pulses. The repetition rate of the attosecond pulses is twice the laser frequency.
The electron dynamics can be described by the Klein-Gordon equation in the single-active-electron approximation and the radiation gauge:
$$\displaystyle\left(\partial^{\mu}\partial_{\mu}+c^{2}\right)\Psi({\rm x})=(V_{%
L}+V_{X}+V_{AI}+V_{H})\Psi({\rm x}),$$
(2)
with the electron-ion interaction operator $V_{AI}=2iV/c^{2}\partial_{t}+V^{2}/c^{2}$,
the atomic potential $V$, the time-space coordinate ${\rm x}=(ct,\mathbf{x})$, the electron-laser interaction operator $V_{L}=2i\mathbf{A}(\eta)\cdot\boldsymbol{\nabla}/c-\mathbf{A}(\eta)^{2}/c^{2}$,
the electron-APT interaction operator $V_{X}=2i\mathbf{A}^{a}(\eta)\cdot\boldsymbol{\nabla}/c-\mathbf{A}^{a}(\eta)^{2%
}/c^{2}$,
the APT vector potential $\mathbf{A}^{a}$, and the interaction operator between the electron and the quantized harmonic field $V_{H}({\rm x})=2\mathbf{A}_{H}({\rm x})/c\cdot(i\boldsymbol{\nabla}/c-\mathbf{%
A}_{tot}({\rm x})/c)$ (the $\mathbf{A}_{H}^{2}$ term is omitted as the harmonic field is a perturbation);
$\mathbf{A}_{H}({\rm x})=c\sqrt{2\pi/\omega_{H}}{\mathbf{e}}_{H}^{*}b^{\dagger}%
\exp(i\omega_{H}t-i{\bf k}_{H}\cdot{\bf x})$ is the vector potential of the harmonic field in second quantization,
with the unit vector ${\mathbf{e}}_{H}$ in polarization direction and the harmonic-photon creation operator $b^{\dagger}$, $\mathbf{A}_{tot}=\mathbf{A}+\mathbf{A}^{a}$,
$\omega_{H}$ is the harmonic frequency, and $\mathbf{k}_{H}$ the wave vector.
The differential HHG rate for the $n^{th}$ harmonic is given by $dw_{n}/d\Omega=n(\omega/c)^{3}|M_{n}|^{2}$ [7], where the HHG amplitude in the strong-field approximation reads
$$\displaystyle M_{n}$$
$$\displaystyle=$$
$$\displaystyle-i\int d^{4}{\rm x}^{\prime}\int d^{4}{\rm x}^{\prime\prime}\left%
\{\Phi({\rm x}^{\prime})^{*}V_{H}({\rm x}^{\prime})\right.$$
(3)
$$\displaystyle\times$$
$$\displaystyle\!\!\!\!\!\left.G^{V}_{L}({\rm x}^{\prime},{\rm x}^{\prime\prime}%
)\,\mathbf{x}^{\prime\prime}\cdot\left[\mathbf{E}({\rm x}^{\prime\prime})+%
\mathbf{E}^{a}({\rm x}^{\prime\prime})\right]\Phi({\rm x}^{\prime\prime})%
\right\},$$
with $\mathbf{E}=-\partial_{t}\mathbf{A}/c$. Here the APT can be treated in the dipole approximation since $E^{a}_{0}/\Omega c\ll 1$, and $c/\Omega\gg a_{B}$, with the Bohr radius $a_{B}$.
Taking into account the shortness of the pulses in the APT ($\tau=4$ a.u.),
we further assume that the dynamics of the electron wave-packet after the ionization is governed by the laser field only. Therefore, the Klein-Gordon Green function in the total field can be replaced by the Volkov-Klein-Gordon Green function in the laser field $G^{V}_{L}({\rm x}^{\prime},{\rm x}^{\prime\prime})$ [11]. Note that the wave function of the ground state $\Phi({\rm x})$ is an eigenstate of the energy operator in the radiation gauge [12].
The integral in Eq. (3) is evaluated with the saddle-point method.
Two ionization scenarios are described by Eq. (3). In the first scenario, the electron is both tunnel-ionized and propagated in the continuum by the laser field. In the other scenario, the process is initiated by the APT via single-photon ionization and the laser field takes over the electron in the continuum. In the relativistic regime, HHG due to the first scenario is heavily damped because of the relativistic drift. Therefore, we will consider the second scenario of the process. In this case, the saddle point equation reflects the energy conservation of Eq.(1) at the moment of ionization: $\varepsilon_{\mathbf{p}}(\eta^{\prime\prime})-c^{2}=\Omega-I_{p}$.
In Fig. 2 the HHG spectrum is displayed via Eq. (3).
A moderately relativistic infrared laser field of $2\times 10^{17}$ W/cm${}^{2}$ ($\xi\approx 0.33$) and an APT of soft x-rays where each pulse has a duration of $100$ as and an energy of $300$ pJ is employed.
Two cases are considered: 1) one-photon ionization by the APT
with ensuing propagation by the laser field and HHG in the APT propagation direction, and 2) tunneling ionization and propagation in a single laser field with HHG in the laser propagation direction. In the first case,
we use an APT with Gaussian pulses, or with delta pulses as in Ref. [9]. For the second case, two calculations are presented: the first one is a fully relativistic calculation via the Klein-Gordon equation, whereas the second one treats the laser field in dipole approximation (DA). The latter serves as an indicator for ideal HHG via the tunneling-recombination mechanism without the relativistic drift. Due to the drift, in the sole laser wave the HHG signal is strongly suppressed by more than 6 orders of magnitude compared to the DA results.
By superimposing a weak APT onto the infrared field, the HHG yield is enhanced by nearly 5 orders of magnitude and almost reaches the DA level. Moreover, the HHG rate can be increased proportionally to the APT intensity. Thus, the APT assistance enables recollisions with energies of 40 keV at significant rates. The description of the APT via delta-pulses is not justified for the parameters applied, since $\Omega\tau\ll 1$ [13], but gives a qualitatively correct picture. Note that the different HHG emission directions considered in 1) and 2) ensure phase-matching between the radiation of different atoms.
We now investigate the stability of the HHG enhancement against variation of the APT parameters.
Firstly, we vary the phase delay between the two fields in Fig. 3a. Phase delay changes of about 0.15 rad have only a minor influence on the near-cutoff HHG rates, whereas changes of 0.3 rad reduce them significantly. This is due to the attosecond pulse duration of $\omega\tau=0.2$ rad: a phase delay variation
of this order or larger reduces the electric field in the pulse whenever trajectories leading to high final energies are launched.
This condition also determines the limited spatial area, where
the APT assistance can strongly enhance the HHG yield.
Secondly, we vary the APT carrier frequency $\Omega$ at the optimal phase delay of the fields.
Fig. 3b shows that for small carrier frequencies the rate for high-energy harmonics is strongly reduced, since the energy supplied by the APT is too small to compensate the relativistic drift in the laser field.
For larger values of the carrier frequency starting from $\Omega\approx 9$ a.u., efficient recombination with the atomic core becomes possible.
From the spectra shown, a slow decrease in the rates for higher carrier frequencies can be deduced. This is due to the scaling of the dipole-matrix element on the electron energy.
The optimal carrier frequency thus amounts to about 9 a.u.; higher values up to 11 a.u. slightly reduce the process efficiency.
A rough estimation, taking into account the spatial restriction of the phase matched emitters, shows that $100$ harmonic photons of $42$ keV photon energy can be emitted within a $1$ keV spectral window at an ion density of $10^{16}$ cm${}^{-3}$ (the electron excursion length is less than the interatomic distance), a laser pulse duration of $100$ fs, and an interaction volume of $10$ mm${}^{3}$.
The HHG in the relativistic regime considered in this paper can also be used for generating ultra-short attosecond pulses via an appropriate filtering of harmonics.
An analysis of the harmonic chirp shows that the optimal spectral window is about 150 eV. Here, two trajectories are responsible for the HHG, which does not allow for a broad window.
As a result, short pulses of 10 as duration can be produced.
In conclusion, we have shown that the newly emerging field of attosecond physics can
be applied for the benefit of modifying relativistic electron recollision dynamics in laser fields.
With the proposed setup HHG with energies of 40 keV and significant rates as well as the generation of ultra-short pulses with a duration of 10 as is possible. Limiting factors for the respective macroscopic yields are the restricted areas which allow for phase-matching among the different ions and a high ion density.
This research was supported in part by the National Science Foundation under Grant No. PHY99-07949. We acknowledge the hospitality of the Kavli Institute for Theoretical Physics.
References
[1]
P. M. Paul, et al., Science 292, 1689 (2001); Y. Mairesse et al., ibid. 302, 1540 (2003); R. Kienberger et al., Nature 427, 817 (2004);
J. Itatani, et al., ibid. 432, 867 (2004);
M. Uiberacker, ibid. 446, 627 (2007);
P. B. Corkum and F. Krausz, Nature Phys. 3, 381 (2007).
[2]
P. Agostini and L. F. DiMauro, Rep. Prog. Phys. 67, 813 (2004).
[3]
J. Seres et al., Nature 433, 596 (2005).
[4]
G. Sansone et al., Science 314, 443 (2006).
[5]
W. Becker et al., Adv. At. Mol. Opt. Phys. 48, 35 (2001);
Y. I. Salamin, et al., Phys. Rep. 427, 42 (2006).
[6]
V. D. Taranukhin, Laser Phys. 10, 330 (2000);
C. H. Keitel and S. X. Hu, Appl. Phys. Lett. 80, 541 (2002);
C. C. Chirilǎ et al., Phys. Rev. A 66, 063411 (2002); Phys. Rev. Lett. 93, 243603 (2004);
N. Milosevic, P. B. Corkum, and T. Brabec, ibid. 92, 013002 (2004);
G. Mocken, and C. H. Keitel, J. Phys. B 37, L275 (2004);
B. Henrich, K. Z. Hatsagortsyan and C. H. Keitel, Phys. Rev. Lett. 93, 013601 (2004);
Q. Lin, S. Li, and W. Becker, Opt. Lett. 31, 2163 (2006);
R. Fischer, M. Lein, and C. H. Keitel, Phys. Rev. Lett. 97, 143901 (2006), J. Phys. B 40, F113 (2007);
M. Verschl, and C. H. Keitel, ibid. 40, F69 (2007); Europhys. Lett. 77, 64004 (2007).
[7]
M. Klaiber, K. Z. Hatsagortsyan, and C. H. Keitel, Phys. Rev. A 73 051803(R) (2006); 75 063413 (2007).
[8]
M. Hentschel et al., Nature 414, 509 (2001);
M. Drescher et al., Science 291, 1923 (2001);
J. Itatani et al., Phys. Rev. Lett. 88, 173903 (2002);
M. Kitzler et al., ibid. 88, 173904 (2002).
[9]
A. Bandrauk and N. H. Shon, Phys. Rev. A 66, 031401(R) (2002);
K. J. Schafer et al., Phys. Rev. Lett. 92, 023003 (2004);
A. Bandrauk, S. Chelkowski and S. Goudreau, J. Mod. Phys. 52, 411 (2005);
C. Figueira de Morisson Faria et al., Phys. Rev. A 74, 053416 (2006);
A. Henrich et al., J. Phys. B 39, S275 (2006).
[10]
Strong attosecond pulses of up to keV photon energy can be produced, e.g., via laser pulse interaction with overdense plasmas, see G. Tsakiris et al. New J. Phys. 8, 19 (2006).
[11]
D. B. Milošević, S. Hu, and W. Becker, Phys. Rev. A 63, 011403(R) (2001); Laser Phys. 12, 389 (2002).
[12]
M. Klaiber, K. Z. Hatsagortsyan, C. H. Keitel, Phys. Rev. A 73, 053411 (2006).
[13]
D. Dimitrovski, E. Solov’ev and J. S. Briggs, Phys. Rev. A 72, 043411 (2005). |
Anharmonic oscillators in the complex plane, $\mathcal{PT}$-symmetry, and real eigenvalues.
Kwang C. Shin
Department of Mathematics, University of West Georgia, Carrollton, GA 30118 USA
(Date:: December 5, 2020)
Abstract.
For integers $m\geq 3$ and $1\leq\ell\leq m-1$, we study the eigenvalue problems
$-u^{\prime\prime}(z)+[(-1)^{\ell}(iz)^{m}-P(iz)]u(z)=\lambda u(z)$ with the boundary conditions that $u(z)$ decays to zero as
$z$ tends to infinity along the rays $\arg z=-\frac{\pi}{2}\pm\frac{(\ell+1)\pi}{m+2}$ in the complex plane,
where $P$ is a polynomial of degree at most $m-1$. We provide asymptotic expansions of the eigenvalues
$\lambda_{n}$. Then we show that if the eigenvalue problem is $\mathcal{PT}$-symmetric, then the eigenvalues are
all real and positive with at most finitely many exceptions. Moreover, we show that when $\gcd(m,\ell)=1$, the
eigenvalue problem has infinitely many real eigenvalues if and only if its translation or itself is
$\mathcal{PT}$-symmetric. Also, we will prove some other interesting direct and inverse spectral results.
2010 Mathematics Subject Classification: 34L40, 34L20, 81Q12
Key words: Anharmonic oscillators, asymptotics of the eigenvalues, $\mathcal{PT}$-symmetry
1. Introduction
In this paper, we study Schrödinger eigenvalue problems with real and complex polynomial potentials in the complex plane under various decaying boundary conditions. We provide explicit asymptotic formulas relating the index $n$ to a series of fractional powers of the eigenvalue $\lambda_{n}$ (see Theorem 1.1). Also, we recover the polynomial potentials from asymptotic formula of the eigenvalues (see Theorem 1.4 and Corollary 2.3) as well as applications to the so-called $\mathcal{PT}$-symmetric Hamiltonians (see Theorems 1.2 and 1.3).
For integers $m\geq 3$ and $1\leq\ell\leq m-1$, we consider the Schrödinger eigenvalue
problem
(1.1)
$$\left(H_{\ell}u\right)(z):=\left[-\frac{d^{2}}{dz^{2}}+(-1)^{\ell}(iz)^{m}-P(%
iz)\right]u(z)=\lambda u(z),\quad\text{for
some $\lambda\in\mathbb{C}$},$$
with the boundary condition that
(1.2)
$$\text{$u(z)\rightarrow 0$ as $z\rightarrow\infty$ along the two rays}\quad\arg
z%
=-\frac{\pi}{2}\pm\frac{(\ell+1)\pi}{m+2},$$
where $P$ is a polynomial of degree at most $m-1$ of the form
$$P(z)=a_{1}z^{m-1}+a_{2}z^{m-2}+\cdots+a_{m-1}z+a_{m},\quad a_{j}\in\mathbb{C}%
\,\,\text{\,for $1\leq j\leq m$}.$$
If a nonconstant function $u$ satisfies (1.1) with some $\lambda\in\mathbb{C}$ and the boundary condition
(1.2), then we call $\lambda$ an eigenvalue of $H_{\ell}$ and $u$ an eigenfunction of
$H_{\ell}$ associated with the eigenvalue $\lambda$. Sibuya [27] showed that the eigenvalues of $H_{\ell}$ are the zeros of an entire function of order
$\rho:=\frac{1}{2}+\frac{1}{m}\in(0,1)$ and hence, by the Hadamard factorization theorem (see, e.g., [7, p. 291]),
there are infinitely many eigenvalues. We call the entire function the Stokes multiplier (or the spectral
determinant), and the algebraic multiplicity of an eigenvalue $\lambda$ is the order of the zero $\lambda$ of the
Stokes multiplier. Also, the geometric multiplicity of an eigenvalue $\lambda$ is the number of linearly
independent eigenfunctions associated with the eigenvalue $\lambda$, that is $1$ for every eigenvalue $\lambda$ [16, §7.4].
We number the eigenvalues $\{\lambda_{n}\}_{n\geq N_{0}}$ in the
order of nondecreasing magnitudes, counting their algebraic
multiplicities. We will show that the magnitude of large
eigenvalues is strictly increasing (see, Lemma 2.2) and
hence, there is a unique way of ordering large eigenvalues, but
this is not guaranteed for small eigenvalues. However, how we
order these small eigenvalues will not affect results in this
paper. Throughout this paper, we will use $\lambda_{n}$ to denote
the eigenvalues of $H_{\ell}$ without explicitly indicating their
dependence on the potential and the boundary condition. Also, we
let
$$a:=(a_{1},a_{2},\ldots,a_{m})\in\mathbb{C}^{m}$$
be the coefficient vector
of $P(z)$.
The anharmonic oscillators $H_{\ell}$ with the various boundary conditions (1.2) are considered in
[4, 24]. When $m$ is even and $\ell=\frac{m}{2}$, $H_{\frac{m}{2}}$ is a Schrödinger operator in
$L^{2}(\mathbb{R})$ (see, e.g.,
[1, 2, 5, 11, 12, 15, 18, 19]). This is self-adjoint if the potential $V(z)=(-1)^{\ell}(iz)^{m}-P(iz)$ is real on the real line, and
non-self-adjoint if the potential is non-real.
Some particular classes of $H_{1}$ have been studied extensively in recent years in the context of theory of
$\mathcal{PT}$-symmetry [3, 6, 10, 14, 20, 26, 28, 29]. The $H_{\ell}$ is
$\mathcal{PT}$-symmetric if the potential $V$ satisfies $\overline{V(-\overline{z})}=V(z)$, $z\in\mathbb{C}$, that is equivalent to $a\in\mathbb{R}^{m}$. In this paper, we will generalize results in [26] (where $H_{1}$ is studied) to $1\leq\ell\leq m-1$ and introduce some new results.
These results are consequences of the following asymptotic expansion of the eigenvalues.
Theorem 1.1.
For each integer $m\geq 3$ and $1\leq\ell\leq m-1$, there exists an integer $N_{0}=N_{0}(m,\ell)$ such that the
eigenvalues $\{\lambda_{n}\}_{n\geq N_{0}}$ of $H_{\ell}$ satisfy
(1.3)
$$n+\frac{1}{2}=\sum_{j=0}^{m+1}c_{\ell,j}(a)\lambda_{n}^{\frac{1}{2}+\frac{1-j}%
{m}}+\eta_{\ell}(a)+O\left(\lambda_{n}^{-\rho}\right)\quad\text{as $n\to\infty%
$},$$
where $c_{\ell,j}(a)$ and $\eta_{\ell}(a)$ are defined in (2.4) and (2.5), respectively.
Also, we obtain the partial reality of the eigenvalues for $\mathcal{PT}$-symmetric $H_{\ell}$.
Theorem 1.2.
If $H_{\ell}$ is $\mathcal{PT}$-symmetric, then eigenvalues are all real with at most finitely many exceptions.
Proof.
If $u(z)$ is an eigenfunction associated with the eigenvalue $\lambda$ of a $\mathcal{PT}$-symmetric $H_{\ell}$,
then $\overline{u(-\overline{z})}$ is also an eigenfunction associated with $\overline{\lambda}$. In Corollary
2.2, we will show that $|\lambda_{n}|<|\lambda_{n+1}|$ for all large $n$ and $\arg(\lambda_{n})\to 0$.
Thus, $\lambda_{n}$ are real and positive for all large $n$ since $|\lambda|=|\overline{\lambda}|$.
∎
Many $\mathcal{PT}$-symmetric operators have real eigenvalues only [10, 22, 23]. However, there are some $\mathcal{PT}$-symmetric $H_{\ell}$ that produce a finite number of non-real eigenvalues [2, 8, 9, 14]. When $H_{\ell}$ is self-adjoint, the spectrum is real. Conversely, when the spectrum is real, what can we conclude about $H_{\ell}$? The next theorem provides a necessary and sufficient condition for $H_{\ell}$ to have infinitely many real eigenvalues.
Theorem 1.3.
Suppose that $\gcd(m,\ell)=1$. Then $H_{\ell}$ with the potential $V(z)=-(iz)^{m}+P(iz)$ has infinitely many real
eigenvalues if and only if $H_{\ell}$ with the potential $V(z-z_{0})$ for some $z_{0}\in\mathbb{C}$ is
$\mathcal{PT}$-symmetric.
The next theorem reveals an interesting feature of the
eigenvalues as a sequence:
Theorem 1.4.
Suppose that $\gcd(m,\ell)=1$. Let $\{\lambda_{n}\}_{n\geq N_{0}}$ and
$\{\widetilde{\lambda}_{n}\}_{n\geq N_{0}}$ be the eigenvalues of
$H_{\ell}$ with the potentials $V$ and $\widetilde{V}$,
respectively. Suppose that
$\lambda_{n}-\widetilde{\lambda}_{n}=o\left(1\right)$ as $n\to\infty$.
Then $\widetilde{V}(z)=V(z-z_{0})$ for some $z_{0}\in\mathbb{C}$ and $\lambda_{n}=\widetilde{\lambda}_{n}$ for all $n\geq N_{0}$ after, if needed, small eigenvalues are reordered.
The asymptotic expansions of the eigenvalues of $H_{\ell}$ with $\ell=\lfloor\frac{m}{2}\rfloor$ have been studied
in, for example, [13, 15, 18].
Maslov [18] computed the first three terms of asymptotic expansions of $\lambda_{n}^{\frac{3}{4}}$,
where $\lambda_{n}$ are the eigenvalues of $-\frac{d^{2}}{dx^{2}}u+x^{4}u=\lambda u,\,\,u\in L^{2}(\mathbb{R}).$
Helffer and Robert [15] considered
$$-\frac{d^{2k}}{dx^{2k}}u+(x^{2m}+p(x))u=\lambda u,\quad u\in L^{2}(\mathbb{R}),$$
where $k,\,m$ are positive integers and where $p(\cdot)$ is a real polynomial of degree at most $2m-1$.
They obtained the existence of asymptotic expansions of the eigenvalues to all orders, and suggested an explicit
way of computing the coefficients of the asymptotic expansion. In particular, for the case when the potential is
$\varepsilon x^{4}+x^{2}$, $\varepsilon>0$, Helffer and Robert [15] computed the first nine terms of the
asymptotic expansion of $\lambda_{n}^{\frac{3}{4}}$. Also, Fedoryuk [13, §3.3] considered (1.1)
with complex polynomial potentials and with (1.2) for $\ell=\lfloor\frac{m}{2}\rfloor$ and computed the
first term in the asymptotic expansion. Also, Sibuya [27] computed the first term in the asymptotic
expansion for $\ell=1$.
This paper is organized as follows. In Section 2, we define $c_{\ell,j}(a)$ and some other notations. Also, we invert (1.3), expressing $\lambda_{n}$ as a series of fractional powers of the index $n$ and prove Theorems 1.3 and 1.4, and other interesting direct and inverse spectral results. In Section 3, we introduce some properties of the solutions of the differential equation in (1.1), due to Hille [16] and Sibuya [27]. We study the asymptotic of the Stokes multiplier associated with $H_{1}$ in Section 4 and treat the general case $H_{\ell}$ in Section 5. In Section 6, we relate the eigenvalues of $H_{\ell}$ with the zeros of the Stokes multiplier. We prove Theorem 1.1 for $1\leq\ell<\frac{m}{2}$ in Section 7 and for $\ell\geq\frac{m}{2}$ in Section 8.
2. Notations and corollaries
In this section, we will define $c_{\ell,j}(a)$ and some other notations and introduce some corollaries of Theorem 1.1.
We define, for nonnegative integers $k,j$,
(2.1)
$$b_{j,k}({a})$$ is the coefficient of $$\frac{1}{z^{j}}$$ in
$${\frac{1}{2}\choose{k}}\left(\frac{a_{1}}{z}+\frac{a_{2}}{z^{2}}+\dots+\frac{a%
_{m}}{z^{m}}\right)^{k}$$ and
(2.2)
$$b_{j}({a})=\sum_{k=0}^{j}b_{j,k}({a}),\quad j\in\mathbb{N}.$$
Notice that $b_{0,0}(a)=b_{0}(a)=1$, $b_{j,0}(a)=0$ if $j\geq 1$, and $b_{j,k}(a)=0$ if $j<k$ or $k<\frac{j}{m}$.
We will also use
(2.3)
$$K_{m,j}({a})=\sum_{k=0}^{j}K_{m,j,k}\,b_{j,k}({a})\,\,\,\text{for $j\geq 0$,}$$
where for $k<\frac{j}{m}$, $K_{m,j,k}=0$ and for $\frac{j}{m}\leq k\leq j$,
$$\displaystyle K_{m,j,k}$$
$$\displaystyle:=\left\{\begin{array}[]{cl}\frac{B\left(\frac{1}{2},1+\frac{1}{m%
}\right)}{2\cos\left(\frac{\pi}{m}\right)}&\text{if $j=k=0$},\\
&\\
-\frac{2}{m}&\text{if $j=k=1$},\\
&\\
\frac{2}{m}\left(\ln 2-\sum_{s=1}^{k-1}\frac{1}{2s-1}\right)&\text{if $j=\frac%
{m}{2}+1$, $m$ even,}\\
&\\
\frac{1}{m}B\left(k-\frac{j-1}{m},\,\frac{j-1}{m}-\frac{1}{2}\right)&\text{if %
$j\not=1$ or $j\not=\frac{m}{2}+1$.}\end{array}\right.$$
Now we are ready to define $c_{\ell,j}(a)$ as follows: for $1\leq j\leq m+1$,
(2.4)
$$c_{\ell,j}(a)=-\frac{2}{\pi}\sum_{k=0}^{j}(-1)^{(\ell+1)k}K_{m,j,k}b_{j,k}(a)%
\sin\left(\frac{(j-1)\ell\pi}{m}\right)\cos\left(\frac{(j-1)\pi}{m}\right),$$
and we also define
(2.5)
$$\eta_{\ell}(a)=\left\{\begin{array}[]{rl}(-1)^{\frac{\ell-1}{2}}\frac{2\nu(a)}%
{m}&\text{if $\ell$ is odd,}\\
&\\
0&\text{if otherwise}\end{array}\right.\,\,\text{and }\,\,\mu({a})=\frac{m}{4}%
-\nu({a}),$$
where
$$\displaystyle\nu({a})=\left\{\begin{array}[]{rl}b_{\frac{m}{2}+1}({a})&\text{%
if $m$ is even,}\\
&\\
0&\text{if $m$ is odd.}\end{array}\right.$$
2.1. Further direct spectral results
One can invert the asymptotic formulas (1.3) to obtain formulas for $\lambda_{n}$ in terms of $n$.
Corollary 2.1.
One can compute $d_{\ell,j}({a})$ explicitly such that
(2.6)
$$\lambda_{n}=\sum_{j=0}^{m+1}d_{\ell,j}({a})\left(n+\frac{1}{2}\right)^{\frac{2%
m}{m+2}\left(1-\frac{j}{m}\right)}+O\left(n^{-\frac{4}{m+2}}\right),$$
where $d_{\ell,0}(a)=\left(\pi^{-1}B\left(\frac{1}{2},1+\frac{1}{m}\right)\sin\left(%
\frac{\ell\pi}{m}\right)\right)^{-\frac{2m}{m+2}}>0$.
Proof.
Equations (1.3) is an asymptotic equation and it can be solved for $\lambda_{n}$, resulting in
(2.6). For details, see, for example, [24, §5].
∎
In the next corollary, we provide an asymptotic formula for the nearest neighbor spacing of the eigenvalues. And
the large eigenvalues increase monotonically in magnitude and have their argument approaching zero.
Corollary 2.2.
The space between successive eigenvalues is
(2.7)
$$\lambda_{n+1}-\lambda_{n}\underset{n\to+\infty}{=}\frac{2m}{m+2}\,d_{\ell,0}%
\cdot\left(n+\frac{1}{2}\right)^{\frac{m-2}{m+2}}+o\left(n^{\frac{m-2}{m+2}}%
\right).$$
In particular, $\lim_{n\to\infty}\left|\lambda_{n+1}-\lambda_{n}\right|=\infty$ and
$\lim_{n\to+\infty}\arg(\lambda_{n})=0.$ Hence:
$$\left|\lambda_{n}\right|<\left|\lambda_{n+1}\right|\,\,\,\text{for all large $%
n$}.$$
Proof.
These claims are consequences of (2.6) and the generalized binomial expansion. For details, see, for example, [24, §6].
∎
Suppose that $\lambda_{n}$ are eigenvalues of $H_{\ell}$ for some $P$. Then the degree $m$ of the polynomial
potential can be recovered by
$$\frac{2m}{m+2}=\lim_{n\to\infty}\frac{\ln(|\lambda_{n}|)}{\ln n}<2.$$
The next corollary shows that one can recover the polynomial potential from the asymptotic formula for the
eigenvalues.
Corollary 2.3.
Suppose that $\gcd(m,\ell)=1$ and that $\{\lambda_{n}\}_{n\geq N_{0}}$ are the eigenvalues of $H_{\ell}$ with the
potential $V$. If
(2.8)
$$\sum_{j=0}^{m+1}c_{j}^{*}\lambda_{n}^{\frac{1}{2}+\frac{1-j}{m}}+O\left(%
\lambda_{n}^{-\frac{1}{2}-\frac{1}{m}}\right){=}\left(n+\frac{1}{2}\right)$$
as $n\to\infty$ for some $c_{j}^{*}\in\mathbb{C}$, or if
(2.9)
$$\lambda_{n}=\sum_{j=0}^{m+1}d_{j}^{*}\left(n+\frac{1}{2}\right)^{\frac{2m}{m+2%
}\left(1-\frac{j}{m}\right)}+O\left(n^{-\frac{4}{m+2}}\right),$$
for some $d_{j}^{*}\in\mathbb{C}$, then there exists $V_{0}$ such that for each polynomial $V$ with which $H_{\ell}$ generates
$\{\lambda_{n}\}_{n\geq N_{0}}$, $V(z)=V_{0}(z-z_{0})$ for some $z_{0}\in\mathbb{C}$.
Proof.
The coefficients $c_{\ell,j}({a})$ and $d_{\ell,j}({a})$ have the following properties.
(i) The $c_{\ell,j}({a})$ and $d_{\ell,j}({a})$ are all real polynomials in terms of the coefficients ${a}$ of $P(x)$.
(ii) The coefficients $c_{\ell,0}({a})$ and $d_{\ell,0}({a})$ do not depend on ${a}$ (they are constants).
(iii) For $1\leq j\leq m$, the polynomials $c_{\ell,j}({a})$ and $d_{\ell,j}({a})$ depend only on $a_{1},\,a_{2},\dots,a_{j}$.
Furthermore, if $\gcd(m,\ell)=1$ and if $j\not=1,\,\frac{m}{2}+1$, then $c_{\ell,j}({a})$ and $d_{\ell,j}({a})$ are
non-constant linear functions of $a_{j}$.
Here we will sketch the proof when (2.8) is assumed. First, notice that $c_{1}^{*}=0$. Otherwise, $\lambda_{n}$
are not the eigenvalues of $H_{\ell}$, according to (2.4). Choose $a_{1}=0$. Then since $c_{\ell,2}({a})$ is a non-constant linear function of $a_{2}$ and depends on $a_{1}$ and $a_{2}$, $c_{\ell,2}({a})=c_{2}^{*}$
determines $a_{2}$. If $m$ is even, $c_{\ell,\frac{m}{2}+1}(a)=0$ and we use $\eta_{\ell}(a)$ instead of
$c_{\ell,\frac{m}{2}+1}(a)$. Suppose that for $2\leq j\leq m-1$, the $c_{2}^{*},\dots,c_{j}^{*}$ uniquely determine
$a_{2},\,a_{3},\dots,a_{j}$. Then by (iii), $c_{\ell,j+1}({a})=c_{j+1}^{*}$ determines $a_{j+1}$ uniquely. So by
induction, one can determine all $a_{j}$ for $2\leq j\leq m$ and $V_{0}$ is the potential with these $a_{j}$’s. Then
$V(z)=V_{0}(z-z_{0})$ for some $z_{0}\in\mathbb{C}$. Otherwise, the eigenvalues $\{\lambda_{n}\}$ do not satisfy (2.8).
The case when (2.9) holds can be handled similarly.
∎
Next, we will prove Theorems 1.3 and 1.4.
Proof of Theorems 1.3.
If $V(z-z_{0})$ for some $z_{0}\in\mathbb{C}$ is $\mathcal{PT}$-symmetric, then by Theorem 1.2, all but finitely many eigenvalues are real and hence, there are infinitely many real eigenvalues.
Suppose that $H_{\ell}$ with the potential $V(z)$ has infinitely many real eigenvalues. Then we can always find $z_{0}\in\mathbb{C}$ so that $V(z-z_{0})$ has no $z^{m-1}$-term, that is, $a_{1}=0$. We will show that $V(z-z_{0})$ is $\mathcal{PT}$-symmetric. Since there are infinitely many real eigenvalues, $c_{\ell,j}(a)$, $0\leq j\leq m+1$, and $\eta_{\ell}$ in (1.3) are real. Also, the sine term in (2.4) does not vanish except for $j=1$. And the cosine term does not vanish except for $j=\frac{m}{2}+1$ when $m$ is even in which $\eta_{\ell}$ replaces the role of $c_{\ell,\frac{m}{2}}$.
Like we did for proof of Corollary 2.3, since $a_{1}=0$, from the above properties of $c_{\ell,j}$ and $\eta_{\ell}$, by the induction, we can show that $a_{j}$ for $2\leq j\leq m$ are all real and hence, $V(z-z_{0})$ is $\mathcal{PT}$-symmetric.
∎
Proof of Theorems 1.4.
Suppose that
$\lambda_{n}-\widetilde{\lambda}_{n}=o\left(1\right)$ as $n\to\infty$. Then $d_{\ell,j}(a)=d_{\ell,j}(\widetilde{a})$ for all $1\leq j\leq m+1$. Then Corollary 2.3 completes proof.
∎
3. Properties of the solutions
In this section, we introduce work of Hille [16] and Sibuya [27] about properties of the solutions of (1.1).
First, we scale equation (1.1) because many facts that we
need later are stated for the scaled equation. Let $u$ be a
solution of (1.1) and let $v(z)=u(-iz)$.
Then $v$ solves
(3.1)
$$-v^{\prime\prime}(z)+[(-1)^{\ell+1}z^{m}+P(z)+\lambda]v(z)=0.$$
When $\ell$ is odd, (3.1) becomes
(3.2)
$$-v^{\prime\prime}(z)+[z^{m}+P(z)+\lambda]v(z)=0.$$
Later we will handle the case when $\ell$ is even.
Since we scaled the argument of $u$, we must rotate the boundary conditions. We state them in a more general
context by using the following definition.
Definition .
The Stokes sectors $S_{k}$ of the equation (3.2)
are
$$S_{k}=\left\{z\in\mathbb{C}:\left|\arg(z)-\frac{2k\pi}{m+2}\right|<\frac{\pi}{%
m+2}\right\}\quad\text{for}\quad k\in\mathbb{Z}.$$
See Figure 1.
It is known from Hille [16, §7.4] that every nonconstant solution of (3.2) either decays to
zero or blows up exponentially, in each Stokes sector $S_{k}$.
Lemma 3.1 ([16, §7.4]).
(i)
For each $k\in\mathbb{Z}$, every solution $v$ of (3.2) is asymptotic to
(3.3)
$$(const.)z^{-\frac{m}{4}}\exp\left[\pm\int^{z}\left[\xi^{m}+P(\xi)+\lambda%
\right]^{\frac{1}{2}}\,d\xi\right]$$
as $z\rightarrow\infty$ in every closed subsector of $S_{k}$.
(ii)
If a nonconstant solution $v$ of (3.2) decays in $S_{k}$, it must blow up
in $S_{k-1}\cup S_{k+1}$. However, when $v$ blows up in $S_{k}$, $v$ need not be decaying in $S_{k-1}$ or in $S_{k+1}$.
Lemma 3.1 (i) implies that if $v$ decays along one ray in $S_{k}$, then it decays along all rays in $S_{k}$.
Also, if $v$ blows up along one ray in $S_{k}$, then it blows up along all rays in $S_{k}$. Thus, the boundary
conditions (1.2) with $1\leq\ell\leq m-1$ represent all decaying boundary conditions.
Still with $\ell$ odd, the two rays in (1.2) map, by $z\mapsto-iz$, to the rays
$\arg(z)=\pm\frac{(\ell+1)\pi}{m+2}$ which are the center rays of the Stokes sectors $S_{\frac{\ell+1}{2}}$ and
$S_{-\frac{\ell+1}{2}}$ and the boundary conditions (1.2) on $u$ become
$$v$$ decays to zero in the Stokes sector $$S_{\frac{\ell+1}{2}}$$ and $$S_{-\frac{\ell+1}{2}}$$.
When $\ell$ is even, we let
$y(z)=v(\omega^{-\frac{1}{2}}z)$ so that (3.1) becomes
(3.4)
$$-y^{\prime\prime}(z)+[z^{m}+\omega^{-1}P(\omega^{-\frac{1}{2}}z)+\omega^{-1}%
\lambda]y(z)=0,$$
where
$$\omega=\exp\left[\frac{2\pi i}{m+2}\right]$$
and hence, $\omega^{-{\frac{m}{2}+1}}=-1$. For these cases, the boundary conditions (1.2) become
$$y$$ decays to zero in the Stokes sector $$S_{\frac{\ell+2}{2}}$$ and $$S_{-\frac{\ell}{2}}$$.
The following theorem is a special case of Theorems 6.1, 7.2, 19.1 and 20.1 of Sibuya [27] that is the
main ingredient of the proofs of the main results in this paper. For this we will use $r_{m}=-\frac{m}{4}$ if $m$ is
odd, and $r_{m}=-\frac{m}{4}-b_{\frac{m}{2}+1}(a)$ if $m$ is even.
Theorem 3.2.
Equation (3.2), with $a\in\mathbb{C}^{m}$, admits a solution $f(z,a,\lambda)$ with the following properties.
(i)
$f(z,a,\lambda)$ is an entire function of $z,a$ and $\lambda$.
(ii)
$f(z,a,\lambda)$ and $f^{\prime}(z,a,\lambda)=\frac{\partial}{\partial z}f(z,a,\lambda)$ admit
the following asymptotic expansions. Let $\varepsilon>0$. Then
$$\displaystyle f(z,a,\lambda)=$$
$$\displaystyle\qquad z^{r_{m}}(1+O(z^{-1/2}))\exp\left[-F(z,a,\lambda)\right],$$
$$\displaystyle f^{\prime}(z,a,\lambda)=$$
$$\displaystyle-z^{r_{m}+\frac{m}{2}}(1+O(z^{-1/2}))\exp\left[-F(z,a,\lambda)%
\right],$$
as $z$ tends to infinity in the sector $|\arg z|\leq\frac{3\pi}{m+2}-\varepsilon$, uniformly on each compact set of $(a,\lambda)$-values .
Here
$$F(z,a,\lambda)=\frac{2}{m+2}z^{\frac{m}{2}+1}+\sum_{1\leq j<\frac{m}{2}+1}%
\frac{2}{m+2-2j}b_{j}(a)z^{\frac{1}{2}(m+2-2j)}.$$
(iii)
Properties (i) and (ii) uniquely determine the solution $f(z,a,\lambda)$ of (3.2).
(iv)
For each fixed $a\in\mathbb{C}^{m}$ and $\delta>0$, $f$ and $f^{\prime}$ also admit the asymptotic expansions,
(3.5)
$$\displaystyle f(0,a,\lambda)=$$
$$\displaystyle[1+O\left(\lambda^{-\rho}\right)]\lambda^{-1/4}\exp\left[L(a,%
\lambda)\right],$$
(3.6)
$$\displaystyle f^{\prime}(0,a,\lambda)=$$
$$\displaystyle-[1+O\left(\lambda^{-\rho}\right)]\lambda^{1/4}\exp\left[L(a,%
\lambda)\right],$$
as $\lambda\to\infty$ in the sector $|\arg(\lambda)|\leq\pi-\delta$, uniformly on each compact set of
$a\in\mathbb{C}^{m}$, where
$$\displaystyle L(a,\lambda)=\left\{\begin{array}[]{rl}&\int_{0}^{+\infty}\left(%
\sqrt{t^{m}+P(t)+\lambda}-t^{\frac{m}{2}}-\sum_{j=1}^{\frac{m+1}{2}}b_{j}(a)t^%
{\frac{m}{2}-j}\right)\,dt\quad\text{if $m$ is odd,}\\
&\int_{0}^{+\infty}\left(\sqrt{t^{m}+P(t)+\lambda}-t^{\frac{m}{2}}-\sum_{j=1}^%
{\frac{m}{2}}b_{j}(a)t^{\frac{m}{2}-j}-\frac{b_{\frac{m}{2}+1}}{t+1}\right)\,%
dt\quad\text{if $m$ is even.}\end{array}\right.$$
(v)
The entire functions $\lambda\mapsto f(0,a,\lambda)$ and $\lambda\mapsto f^{\prime}(0,a,\lambda)$ have orders $\frac{1}{2}+\frac{1}{m}$.
Proof.
In Sibuya’s book [27], see Theorem 6.1 for a proof of (i) and (ii); Theorem 7.2 for a proof of (iii) with the error terms $o(1)$;
and Theorem 19.1 for a proof of (iv). Moreover, (v) is a consequence of (iv) along with Theorem 20.1 in
[27]. The error terms in (iii) are improved from $o(1)$ to $O\left(\lambda^{-\rho}\right)$ in [25]. Note that properties (i), (ii) and (iv) are summarized on pages 112–113 of Sibuya [27].
∎
Remarks.
Throughout this paper, we will deal with numbers like
$\left(\omega^{\alpha}\lambda\right)^{s}$ for some $s\in\mathbb{R}$ and $\alpha\in\mathbb{C}$. As usual, we will use
$$\omega^{\alpha}=\exp\left[\alpha\frac{2\pi i}{m+2}\right]$$
and if $\arg(\lambda)$ is specified, then
$$\arg\left(\left(\omega^{\alpha}\lambda\right)^{s}\right)=s\left[\arg(\omega^{%
\alpha})+\arg(\lambda)\right]=s\left[{\rm Re}\,(\alpha)\frac{2\pi}{m+2}+\arg(%
\lambda)\right],\quad s\in\mathbb{R}.$$
Lemma 3.3.
Let $m\geq 3$ and $a\in\mathbb{C}^{m}$ be fixed. Then
(3.7)
$$L(a,\lambda)=\sum_{j=0}^{\infty}K_{m,j}(a)\lambda^{\frac{1}{2}+\frac{1-j}{m}}-%
\frac{\nu(a)}{m}\ln(\lambda)$$
as $\lambda\to\infty$ in the sector $|\arg(\lambda)|\leq\pi-\delta$, uniformly on each compact set of
$a\in\mathbb{C}^{m}$.
Proof.
See [24] for a proof.
∎
Sibuya [27] proved the following corollary, directly from Theorem 3.2, that will be used later in
Sections 4 and 5.
Corollary 3.4.
Let $a\in\mathbb{C}^{m}$ be fixed. Then $L(a,\lambda)=K_{m}\lambda^{\frac{1}{2}+\frac{1}{m}}(1+o(1))$ as $\lambda$ tends
to infinity in the sector $|\arg\lambda|\leq\pi-\delta$, and hence
(3.8)
$${\rm Re}\,\left(L(a,\lambda)\right)=K_{m}\cos\left(\frac{m+2}{2m}\arg(\lambda)%
\right)|\lambda|^{\frac{1}{2}+\frac{1}{m}}(1+o(1))$$
as $\lambda\to\infty$ in the sector $|\arg(\lambda)|\leq\pi-\delta$.
In particular, ${\rm Re}\,\left(L(a,\lambda)\right)\to+\infty$ as $\lambda\to\infty$ in any closed subsector of the
sector $|\arg(\lambda)|<\frac{m\pi}{m+2}$. In addition, ${\rm Re}\,\left(L(a,\lambda)\right)\to-\infty$ as
$\lambda\to\infty$ in any closed subsector of the sectors $\frac{m\pi}{m+2}<|\arg(\lambda)|<\pi-\delta$.
Based on Corollary 3.4, Sibuya [27, Thm. 29.1] also computed the leading term in
(1.3) for $\ell=1$. Also, Sibuya [27] constructed solutions of (3.2) that decays
in $S_{k}$, $k\in\mathbb{Z}$. Before we introduce this, we let
(3.9)
$$G^{\ell}(a):=(\omega^{(m+1)\ell}a_{1},\omega^{m\ell}a_{2},\ldots,\omega^{2\ell%
}a_{m})\quad\text{for}\quad\ell\in\frac{1}{2}\mathbb{Z}.$$
Then we have the following lemma, regarding some properties of $G^{\ell}(\cdot)$.
Lemma 3.5.
For $a\in\mathbb{C}^{m}$ fixed, and $\ell_{1},\ell_{2},\ell\in\frac{1}{2}\mathbb{Z}$,
$G^{\ell_{1}}(G^{\ell_{2}}(a))=G^{\ell_{1}+\ell_{2}}(a)$, and
$$b_{j,k}(G^{\ell}(a))=\omega^{((m+2)k-j)\ell}b_{j,k}(a),\quad\ell\in\frac{1}{2}%
\mathbb{Z}.$$
In particular,
$$b_{j}(G^{\ell}(a))=\omega^{-j\ell}b_{j}(a),\,\,\ell\in\mathbb{Z}.$$
Next, recall that the function
$f(z,a,\lambda)$ in Theorem 3.2 solves (3.2) and decays to zero exponentially as $z\rightarrow\infty$ in
$S_{0}$, and blows up in $S_{-1}\cup S_{1}$. One can check that the function
$$f_{k}(z,a,\lambda):=f(\omega^{-k}z,G^{k}(a),\omega^{2k}\lambda),\quad k\in%
\mathbb{Z},$$
which is obtained by scaling $f(z,G^{k}(a),\omega^{2k}\lambda)$ in the $z$-variable, also solves (3.2). It is clear
that $f_{0}(z,a,\lambda)=f(z,a,\lambda)$, and that $f_{k}(z,a,\lambda)$ decays in $S_{k}$ and blows up in $S_{k-1}\cup S_{k+1}$ since
$f(z,G^{k}(a),\omega^{2k}\lambda)$ decays in $S_{0}$. Since no nonconstant solution decays in two consecutive Stokes sectors (see Lemma 3.1
(ii)), $f_{k}$ and $f_{k+1}$ are linearly independent and hence any solution of (3.2) can be expressed as a linear combination of these two.
Especially, for each $k\in\mathbb{Z}$ there exist some coefficients $C_{k}(a,\lambda)$ and $\widetilde{C}_{k}(a,\lambda)$ such that
(3.10)
$$f_{k}(z,a,\lambda)=C_{k}(a,\lambda)f_{0}(z,a,\lambda)+\widetilde{C}_{k}(a,%
\lambda)f_{-1}(z,a,\lambda).$$
We then see that
(3.11)
$$C_{k}(a,\lambda)=-\frac{\mathcal{W}_{k,-1}(a,\lambda)}{\mathcal{W}_{-1,0}(a,%
\lambda)}\quad\text{and}\quad\widetilde{C}_{k}(a,\lambda)=\frac{\mathcal{W}_{k%
,0}(a,\lambda)}{\mathcal{W}_{-1,0}(a,\lambda)},$$
where $\mathcal{W}_{j,\ell}=f_{j}f_{\ell}^{\prime}-f_{j}^{\prime}f_{\ell}$ is the Wronskian of $f_{j}$ and $f_{\ell}$. Since both
$f_{j},\,f_{\ell}$ are solutions of the same linear equation (3.2), we know that the Wronskians are
constant functions of $z$. Also, $f_{k}$ and $f_{k+1}$ are linearly independent, and hence
$\mathcal{W}_{k,k+1}\not=0$ for all $k\in\mathbb{Z}$.
Also, the following is an easy consequence of (3.10) and (3.11). For each $k,\ell\in\mathbb{Z}$ we have
$$\displaystyle\mathcal{W}_{\ell,k}(a,\lambda)$$
$$\displaystyle=C_{k}(a,\lambda)\mathcal{W}_{\ell,0}(a,\lambda)+\widetilde{C}_{k%
}(a,\lambda)\mathcal{W}_{\ell,-1}(a,\lambda)$$
(3.12)
$$\displaystyle=-\frac{\mathcal{W}_{k,-1}(a,\lambda)\mathcal{W}_{\ell,0}(a,%
\lambda)}{\mathcal{W}_{-1,0}(a,\lambda)}+\frac{\mathcal{W}_{k,0}(a,\lambda)%
\mathcal{W}_{\ell,-1}(a,\lambda)}{\mathcal{W}_{-1,0}(a,\lambda)}.$$
Moreover, we have the following lemma that is useful later on.
Lemma 3.6.
Suppose $k,\,j\in\mathbb{Z}$. Then
(3.13)
$$\mathcal{W}_{k+1,j+1}(a,\lambda)=\omega^{-1}\mathcal{W}_{k,j}(G(a),\omega^{2}%
\lambda),$$
and $\mathcal{W}_{0,1}(a,\lambda)=2\omega^{\mu(a)}$, where $\mu(a)=\frac{m}{4}-\nu(a)$.
Proof.
See Sibuya [27, pp. 116-118].
∎
4. Asymptotics of $\mathcal{W}_{-1,1}(a,\lambda)$
In this section, we introduce asymptotic expansions of $\mathcal{W}_{-1,1}(a,\lambda)$ as $\lambda\to\infty$ along
the rays in the complex plane [24].
First, we provide an asymptotic expansion of the Wronskian $\mathcal{W}_{0,j}(a,\lambda)$ of $f_{0}$ and $f_{j}$
that will be frequently used later.
Lemma 4.1.
Suppose that $1\leq j\leq\frac{m}{2}+1$. Then for each $a\in\mathbb{C}^{m}$,
(4.1)
$$\mathcal{W}_{0,j}(a,\lambda)=[2i\omega^{-\frac{j}{2}}+O\left(\lambda^{-\rho}%
\right)]\exp\left[L(G^{j}(a),\omega^{2j-m-2}\lambda)+L(a,\lambda)\right],$$
as $\lambda\to\infty$ in the sector
(4.2)
$$-\pi+\delta\leq\pi-\frac{4j\pi}{m+2}+\delta\leq\arg(\lambda)\leq\pi-\delta.$$
Next, we provide an asymptotic expansion of $\mathcal{W}_{-1,1}(a,\lambda)$ as $\lambda\to\infty$ in the sector
near the negative real axis.
Theorem 4.2.
Let $m\geq 3$, $a\in\mathbb{C}^{m}$ and $0<\delta<\frac{\pi}{m+2}$ be fixed. Then
(4.3)
$$\mathcal{W}_{-1,1}(a,\lambda)=[2i+O\left(\lambda^{-\rho}\right)]\exp\left[L(G^%
{-1}(a),\omega^{-2}\lambda)+L(G(a),\omega^{-m}\lambda)\right],$$
as $\lambda\to\infty$ along the rays in the sector
(4.4)
$$\pi-\frac{4\pi}{m+2}+\delta\leq\arg(\lambda)\leq\pi+\frac{4\pi}{m+2}-\delta.$$
Proof.
This is an easy consequence of Lemma 4.1 with $j=2$ and (3.13).
∎
Also, for integers $m\geq 4$ we provide an asymptotic expansion of $\mathcal{W}_{-1,1}(a,\lambda)$ as
$\lambda\to\infty$ in the sector $|\arg(\lambda)|\leq\pi-\delta$.
Theorem 4.3.
Let $a\in\mathbb{C}^{m}$ and $0<\delta<\frac{\pi}{2(m+2)}$ be fixed. If $m\geq 4$ then
$$\displaystyle\mathcal{W}_{-1,1}(a,\lambda)=$$
$$\displaystyle[2\omega^{\frac{1}{2}+\mu(a)}+O\left(\lambda^{-\rho}\right)]\exp%
\left[L(G^{-1}(a),\omega^{-2}\lambda)-L(a,\lambda)\right]$$
(4.5)
$$\displaystyle+[2\omega^{\frac{1}{2}+\mu(a)+2\nu(a)}+O\left(\lambda^{-\rho}%
\right)]\exp\left[L(G(a),\omega^{2}\lambda)-L(a,\lambda)\right],$$
as $\lambda\to\infty$ in the sector
(4.6)
$$-\pi+\delta\leq\arg(\lambda)\leq\pi-\delta.$$
Next, we provide an asymptotic expansion of $\mathcal{W}_{-1,1}(a,\lambda)$ as $\lambda\to\infty$ along the rays
in the upper- and lower- half planes.
Corollary 4.4.
Let $m\geq 4$, $a\in\mathbb{C}^{m}$ and $0<\delta<\frac{\pi}{m+2}$ be fixed. Then
$$\mathcal{W}_{-1,1}(a,\lambda)=[2\omega^{\frac{1}{2}+\mu(a)}+O\left(\lambda^{-%
\rho}\right)]\exp\left[L(G^{-1}(a),\omega^{-2}\lambda)-L(a,\lambda)\right],$$
as $\lambda\to\infty$ in the sector $\delta\leq\arg(\lambda)\leq\pi-\delta$.
Also,
$$\mathcal{W}_{-1,1}(a,\lambda)=[2\omega^{\frac{1}{2}+\mu(a)+2\nu(a)}+O\left(%
\lambda^{-\rho}\right)]\exp\left[L(G(a),\omega^{2}\lambda)-L(a,\lambda)\right],$$
as $\lambda\to\infty$ in the sector $-\pi+\delta\leq\arg(\lambda)\leq-\delta$.
Proof.
We will determine which term in (4.5) dominates in the upper and lower half planes.
Since, by (3.8),
$${\rm Re}\,(L(a,\lambda))=K_{m}\cos\left(\frac{m+2}{2m}\arg(\lambda)\right)|%
\lambda|^{\frac{1}{2}+\frac{1}{m}}(1+o(1)),$$
we have
$$\displaystyle\left[{\rm Re}\,(L(G^{-1}(a),\omega^{-2}\lambda))-{\rm Re}\,(L(a,%
\lambda))\right]-\left[{\rm Re}\,(L(G(a),\omega^{2}\lambda))-{\rm Re}\,(L(a,%
\lambda))\right]$$
$$\displaystyle=K_{m}\left[\cos\left(-\frac{2\pi}{m}+\frac{m+2}{2m}\arg(\lambda)%
\right)-\cos\left(\frac{2\pi}{m}+\frac{m+2}{2m}\arg(\lambda)\right)\right]|%
\lambda|^{\frac{1}{2}+\frac{1}{m}}(1+o(1))$$
$$\displaystyle=2K_{m}\sin\left(\frac{2\pi}{m}\right)\sin\left(\frac{m+2}{2m}%
\arg(\lambda)\right)|\lambda|^{\frac{1}{2}+\frac{1}{m}}(1+o(1)).$$
Thus, the first term in (4.5) dominates as $\lambda\to\infty$ along the rays in the upper half plane, and the second term dominates in the lower half plane. This completes the proof.
∎
Proof of Theorem 4.3.
In [24], $C(a,\lambda)$ is used for $\frac{\mathcal{W}_{-1,1}(a,\lambda)}{\mathcal{W}_{0,1}(a,\lambda)}$
and asymptotics of $C(a,\lambda)$ are provided. Notice that
$\mathcal{W}_{-1,1}(a,\lambda)=2\omega^{\mu(a)}C(a,\lambda)$.
Theorem 13 in [24] implies
(4.5) for the sector
(4.7)
$$\pi-\frac{4\lfloor\frac{m}{2}\rfloor\pi}{m+2}+\delta\leq\arg(\lambda)\leq\pi-%
\frac{4\pi}{m+2}-\delta.$$
Theorem 14 in [24] implies that
$$\displaystyle\mathcal{W}_{-1,1}(a,\lambda)=$$
$$\displaystyle[2\omega^{\frac{1}{2}+\mu(a)}+O\left(\lambda^{-\rho}\right)]\exp%
\left[L(G^{-1}(a),\omega^{-2}\lambda)-L(a,\lambda)\right]$$
$$\displaystyle+[2\omega^{1+2\mu(a)+4\nu(a)}+O\left(\lambda^{-\rho}\right)]\exp%
\left[-L(G^{2}(a),\omega^{2-m}\lambda)-L(a,\lambda)\right],$$
as $\lambda\to\infty$ in the sector $\pi-\frac{8\pi}{m+2}+\delta\leq\arg(\lambda)\leq\pi-\delta$. One can check that the first term dominates in this sector, by using an argument similar to that in the proof of Corollary 4.4.
Also, Theorem 15 in [24] implies that
$$\displaystyle\mathcal{W}_{-1,1}(a,\lambda)=$$
$$\displaystyle[2\omega^{1+2\mu(a)}+O\left(\lambda^{-\rho}\right)]\exp\left[-L(a%
,\omega^{-m-2}\lambda)-L(^{-2}(a),\omega^{-4}\lambda)\right]$$
$$\displaystyle+[2\omega^{\frac{1}{2}+\mu(a)+2\nu(a)}+O\left(\lambda^{-\rho}%
\right)]\exp\left[L(G(a),\omega^{-m}\lambda)-L(a,\omega^{-m-2}\lambda)\right],$$
as $\lambda\to\infty$ in the sector $\pi+\delta\leq\arg(\lambda)\leq\pi+\frac{8\pi}{m+2}-\delta$. One can check that the second term dominates in this sector. Then we replace $\lambda$ by $\omega^{m+2}\lambda$ to convert the sector here to $-\pi+\delta\leq\arg(\lambda)\leq-\pi+\frac{8\pi}{m+2}-\delta$. This completes the proof.
∎
Theorem 4.5.
Let $m=3$ and let $a\in\mathbb{C}^{m}$ and $0<\delta<\frac{\pi}{m+2}$ be fixed. Then
$$\displaystyle\mathcal{W}_{-1,1}(a,\lambda)=$$
$$\displaystyle[-2\omega^{-\frac{5}{4}}+O\left(\lambda^{-\rho}\right)]\exp\left[%
L(G^{4}(a),\omega^{-2}\lambda)-L(a,\lambda)\right]$$
$$\displaystyle-[2i\omega^{\frac{5}{2}}+O\left(\lambda^{-\rho}\right)]\exp\left[%
-L(G^{2}(a),\omega^{-1}\lambda)-L(a,\lambda)\right],$$
as $\lambda\to\infty$ in the sector
$-\delta\leq\arg(\lambda)\leq\pi-\delta$.
Also,
$$\displaystyle\mathcal{W}_{-1,1}(a,\lambda)=$$
$$\displaystyle[-2i\omega^{\frac{5}{2}}+O\left(\lambda^{-\rho}\right)]\exp\left[%
-L(a,\omega^{-5}\lambda)-L(G^{-2}(a),\omega^{-4}\lambda)\right]$$
$$\displaystyle+[2\omega^{\frac{15}{4}}+O\left(\lambda^{-\rho}\right)]\exp\left[%
L(G(a),\omega^{-3}\lambda)-L(a,\omega^{-5}\lambda)\right],$$
as $\lambda\to\infty$ in the sector $-\pi+\delta\leq\arg(\lambda)\leq\delta$.
Proof.
See Theorems 14 and 15 in [24] for a proof.
∎
5. Asymptotics of $\mathcal{W}_{-1,n}(a,\lambda)$
In this section, we will provide asymptotic expansions of $\mathcal{W}_{-1,n}(a,\cdot)$, zeros of which will be
closely related with the eigenvalues of $H_{n}$.
First, we treat the cases when $1\leq n<\lfloor\frac{m}{2}\rfloor$.
Theorem 5.1.
Let $1\leq n<\lfloor\frac{m}{2}\rfloor$ be an integer. Then $\mathcal{W}_{-1,n}(a,\cdot)$ admits the following
asymptotic expansion
(5.1)
$$\mathcal{W}_{-1,n}(a,\lambda)=[2\omega^{\frac{2-n}{2}+\mu(G^{n-1}(a))}+O\left(%
\lambda^{-\rho}\right)]\exp\left[L(G^{-1}(a),\omega^{-2}\lambda)-L(G^{n-1}(a),%
\omega^{2(n-1)}\lambda)\right],$$
as $\lambda\to\infty$ in the sector
(5.2)
$$-\frac{2(n-1)\pi}{m+2}+\delta\leq\arg(\lambda)\leq\pi-\frac{4n\pi}{m+2}+\delta.$$
Also,
$$\displaystyle\mathcal{W}_{-1,n}(a,\lambda)$$
$$\displaystyle=[2\omega^{\frac{2-n}{2}+\mu(G^{n-1}(a))}+O\left(\lambda^{-\rho}%
\right)]\exp\left[L(G^{-1}(a),\omega^{-2}\lambda)-L(G^{n-1}(a),\omega^{2(n-1)}%
\lambda)\right]$$
(5.3)
$$\displaystyle+[2\omega^{\frac{2-n}{2}+\mu(G^{-1}(a))}+O\left(\lambda^{-\rho}%
\right)]\exp\left[L(G^{n}(a),\omega^{2n}\lambda)-L(a,\lambda)\right],$$
as $\lambda\to\infty$ in the sector
(5.4)
$$-\frac{2(n-1)\pi}{m+2}-\delta\leq\arg(\lambda)\leq-\frac{2(n-1)\pi}{m+2}+\delta.$$
Proof.
First we will prove (5.1) for the sector
(5.5)
$$-\frac{2(n-1)\pi}{m+2}+\delta\leq\arg(\lambda)\leq\pi-\frac{4n\pi}{m+2}-\delta$$
and the second part of the theorem by induction on $n$.
The case when $n=1$ is trivially satisfied by Theorem 4.3 and Corollary 4.4 since $\mu(a)+2\nu(a)=\mu(G^{-1}(a))$.
Suppose that (5.1) holds in the sector (5.5) for $n-1$. From this induction hypothesis we have
$$\displaystyle\mathcal{W}_{0,n}(a,\lambda)$$
$$\displaystyle=\omega^{-1}\mathcal{W}_{-1,n-1}(G(a),\omega^{2}\lambda)$$
(5.6)
$$\displaystyle=[2\omega^{-\frac{n-1}{2}+\mu(G^{n-1}(a))}+O\left(\lambda^{-\rho}%
\right)]\exp\left[L(a,\lambda)-L(G^{n-1}(a),\omega^{2(n-1)}\lambda)\right],$$
as $\lambda\to\infty$ in the sector
$$-\frac{2(n-2)\pi}{m+2}+\delta\leq\arg(\omega^{2}\lambda)\leq\pi-\frac{4(n-1)%
\pi}{m+2}-\delta,$$
that is,
(5.7)
$$-\frac{2n\pi}{m+2}+\delta\leq\arg(\lambda)\leq\pi-\frac{4n\pi}{m+2}-\delta.$$
Also, from Lemma 4.1 if $1\leq j\leq\frac{m}{2}+1$, then we have
(5.8)
$$\mathcal{W}_{0,j}(a,\lambda)=[2i\omega^{-\frac{j}{2}}+O\left(\lambda^{-\rho}%
\right)]\exp\left[L(G^{j}(a),\omega^{2j-m-2}\lambda)+L(a,\lambda)\right],$$
as $\lambda\to\infty$ in the sector
$$\pi-\frac{4j\pi}{m+2}+\delta\leq\arg(\lambda)\leq\pi-\delta.$$
We solve (3.12) for $\mathcal{W}_{\ell,-1}(a,\lambda)$ and set $\ell=n$ to get
(5.9)
$$\mathcal{W}_{-1,n}(a,\lambda)=\frac{\mathcal{W}_{-1,0}(a,\lambda)\mathcal{W}_{%
n,k}(a,\lambda)}{\mathcal{W}_{0,k}(a,\lambda)}+\frac{\mathcal{W}_{0,n}(a,%
\lambda)\mathcal{W}_{-1,k}(a,\lambda)}{\mathcal{W}_{0,k}(a,\lambda)}$$
Set $k=\lfloor\frac{m}{2}\rfloor$. Then since $1\leq k-n<k=\lfloor\frac{m}{2}\rfloor$, using (3.13),
$$\displaystyle\mathcal{W}_{-1,n}(a,\lambda)$$
$$\displaystyle=\frac{2\omega^{\mu(G^{-1}(a))}\mathcal{W}_{0,k-n}(G^{n}(a),%
\omega^{2n}\lambda)}{\omega^{n-1}\mathcal{W}_{0,k}(a,\lambda)}+\frac{\mathcal{%
W}_{0,n}(a,\lambda)\mathcal{W}_{0,k+1}(G^{-1}(a),\omega^{-2}\lambda)}{\omega^{%
-1}\mathcal{W}_{0,k}(a,\lambda)}$$
$$\displaystyle=\frac{2\omega^{\mu(G^{-1}(a))}[2i\omega^{-\frac{k-n}{2}}+O\left(%
\lambda^{-\rho}\right)]\exp\left[L(G^{n}(a),\omega^{2k-m-2}\lambda)+L(G^{n}(a)%
,\omega^{2n}\lambda)\right]}{\omega^{n-1}[2i\omega^{-\frac{k}{2}}+O\left(%
\lambda^{-\rho}\right)]\exp\left[L(G^{k}(a),\omega^{2k-m-2}\lambda)+L(a,%
\lambda)\right]}$$
$$\displaystyle+\frac{[2\omega^{-\frac{n-1}{2}+\mu(G^{n-1}(a))}+O\left(\lambda^{%
-\rho}\right)]\exp\left[L(a,\lambda)-L(G^{n-1}(a),\omega^{2(n-1)}\lambda)%
\right]}{\omega^{-1}[2i\omega^{-\frac{k}{2}}+O\left(\lambda^{-\rho}\right)]%
\exp\left[L(G^{k}(a),\omega^{2k-m-2}\lambda)+L(a,\lambda)\right]}$$
$$\displaystyle\times[2i\omega^{-\frac{k+1}{2}}+O\left(\lambda^{-\rho}\right)]%
\exp\left[L(G^{k}(a),\omega^{2k-m-2}\lambda)+L(G^{-1}(a),\omega^{-2}\lambda)\right]$$
$$\displaystyle=[2\omega^{\frac{2-n}{2}+\mu(G^{-1}(a))}+O\left(\lambda^{-\rho}%
\right)]\exp\left[L(G^{n}(a),\omega^{2n}\lambda)-L(a,\lambda)\right]$$
(5.10)
$$\displaystyle+[2\omega^{\frac{2-n}{2}+\mu(G^{n-1}(a))}+O\left(\lambda^{-\rho}%
\right)]\exp\left[L(G^{-1}(a),\omega^{-2}\lambda)-L(G^{n-1}(a),\omega^{2(n-1)}%
\lambda)\right],$$
where we used (5.6) for $\mathcal{W}_{0,n}(a,\lambda)$ and (5.8) for everything else, provided
that $\lambda$ lies in (5.7) and that
$$\displaystyle\pi-\frac{4\left(\lfloor\frac{m}{2}\rfloor-n\right)\pi}{m+2}+\delta\leq$$
$$\displaystyle\arg(\omega^{2n}\lambda)\leq\pi-\delta$$
$$\displaystyle\pi-\frac{4\lfloor\frac{m}{2}\rfloor\pi}{m+2}+\delta\leq$$
$$\displaystyle\arg(\lambda)\leq\pi-\delta$$
$$\displaystyle\pi-\frac{4\left(\lfloor\frac{m}{2}\rfloor+1\right)\pi}{m+2}+\delta\leq$$
$$\displaystyle\arg(\omega^{-2}\lambda)\leq\pi-\delta,$$
that is,
$$-\frac{2n\pi}{m+2}+\delta\leq\arg(\lambda)\leq\pi-\frac{4n\pi}{m+2}-\delta.$$
Thus, the second part of the theorem is proved by induction.
Next in order to prove the first part of the theorem for the sector (5.5), we will determine which term in (5.10) dominates as $\lambda\to\infty$. To do that, we look at
$$\displaystyle{\rm Re}\,\left(L(G^{-1}(a),\omega^{-2}\lambda)-L(G^{n-1}(a),%
\omega^{2(n-1)}\lambda)\right)-{\rm Re}\,\left(L(G^{n}(a),\omega^{2n}\lambda)-%
L(a,\lambda)\right)$$
$$\displaystyle=K_{m}\left[\cos\left(-\frac{2\pi}{m}+\frac{m+2}{2m}\arg(\lambda)%
\right)-\cos\left(\frac{2(n-1)\pi}{m}+\frac{m+2}{2m}\arg(\lambda)\right)\right.$$
$$\displaystyle-\left.\left(\cos\left(-\frac{2n\pi}{m}+\frac{m+2}{2m}\arg(%
\lambda)\right)-\cos\left(\frac{m+2}{2m}\arg(\lambda)\right)\right)\right]|%
\lambda|^{\frac{1}{2}+\frac{1}{m}}(1+o(1))$$
$$\displaystyle=2K_{m}\sin\left(\frac{n\pi}{m}\right)\left[\sin\left(\frac{(n-2)%
\pi}{m}+\frac{m+2}{2m}\arg(\lambda)\right)\right.$$
$$\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad\left.+\sin\left(\frac{n\pi}{%
m}+\frac{m+2}{2m}\arg(\lambda)\right)\right]|\lambda|^{\frac{1}{2}+\frac{1}{m}%
}(1+o(1))$$
(5.11)
$$\displaystyle=4K_{m}\sin\left(\frac{n\pi}{m}\right)\cos\left(\frac{\pi}{m}%
\right)\sin\left(\frac{(n-1)\pi}{m}+\frac{m+2}{2m}\arg(\lambda)\right)|\lambda%
|^{\frac{1}{2}+\frac{1}{m}}(1+o(1)),$$
that tends to positive infinity as $\lambda\to\infty$ (and hence the second term in (5.10) dominates) if $-\frac{2(n-1)\pi}{m+2}+\delta\leq\arg(\lambda)\leq\pi-\frac{4n\pi}{m+2}-\delta$.
We still need to prove (5.1) for the sector
(5.12)
$$\pi-\frac{4n\pi}{m+2}-\delta\leq\arg(\lambda)\leq\pi-\frac{4n\pi}{m+2}+\delta,$$
for which we use induction on $n$ again.
When $n=1$, (5.1) holds by Lemma 4.4.
Suppose that (5.1) in the sector (5.12) for $n-1$ with $2\leq n<\lfloor\frac{m}{2}\rfloor$.
Then (3.13) and (5.9) with $k=n+1$ yield
$$\mathcal{W}_{-1,n}(a,\lambda)=\frac{2\omega^{\mu(G^{-1}(a))}\mathcal{W}_{0,1}(%
G^{n}(a),\omega^{2n}\lambda)}{\omega^{n-1}\mathcal{W}_{0,n+1}(a,\lambda)}+%
\frac{\mathcal{W}_{-1,n-1}(G(a),\omega^{2}\lambda)\mathcal{W}_{0,n+2}(G^{-1}(a%
),\omega^{-2}\lambda)}{\mathcal{W}_{0,n+1}(a,\lambda)}.$$
If $\pi-\frac{4n\pi}{m+2}-\delta\leq\arg(\lambda)\leq\pi-\frac{4n\pi}{m+2}+\delta$, then $\pi-\frac{4(n-1)\pi}{m+2}-\delta\leq\arg(\omega^{2}\lambda)\leq\pi-\frac{4(n-1%
)\pi}{m+2}+\delta$. So
(5.13)
$$\displaystyle\mathcal{W}_{-1,n}(a,\lambda)$$
$$\displaystyle=\frac{4\omega^{\mu(G^{-1}(a))+\mu(G^{n}(a))}}{\omega^{n-1}%
\mathcal{W}_{0,n+1}(a,\lambda)}+\frac{\mathcal{W}_{-1,n-1}(G(a),\omega^{2}%
\lambda)\mathcal{W}_{0,n+2}(G^{-1}(a),\omega^{-2}\lambda)}{\mathcal{W}_{0,n+1}%
(a,\lambda)}$$
$$\displaystyle=\frac{4\omega^{\mu(G^{-1}(a))+\mu(G^{n}(a))}}{\omega^{n-1}[2i%
\omega^{-\frac{n+1}{2}}+O\left(\lambda^{-\rho}\right)]\exp\left[L(G^{n+1}(a),%
\omega^{2n-m}\lambda)+L(a,\lambda)\right]}$$
$$\displaystyle+\frac{[2\omega^{\frac{3-n}{2}+\mu(G^{n-1}(a))}+O\left(\lambda^{-%
\rho}\right)]\exp\left[L(a,\lambda)-L(G^{n-1}(a),\omega^{2(n-1)}\lambda)\right%
]}{[2i\omega^{-\frac{n+1}{2}}+O\left(\lambda^{-\rho}\right)]\exp\left[L(G^{n+1%
}(a),\omega^{2n-m}\lambda)+L(a,\lambda)\right]}$$
$$\displaystyle\times[2i\omega^{-\frac{n+2}{2}}+O\left(\lambda^{-\rho}\right)]%
\exp\left[L(G^{n+1}(a),\omega^{2n-m}\lambda)+L(G^{-1}(a),\omega^{-2}\lambda)\right]$$
$$\displaystyle=[-2i\omega^{\frac{3-n}{2}+\mu(G^{-1}(a))+\mu(G^{n}(a))}+O\left(%
\lambda^{-\rho}\right)]\exp\left[-L(G^{n+1}(a),\omega^{2n-m}\lambda)-L(a,%
\lambda)\right]$$
$$\displaystyle+[2\omega^{\frac{2-n}{2}+\mu(G^{n-1}(a))}+O\left(\lambda^{-\rho}%
\right)]\exp\left[L(G^{-1}(a),\omega^{-2}\lambda)-L(G^{n-1}(a),\omega^{2(n-1)}%
\lambda)\right],$$
where we use the induction hypothesis for $\mathcal{W}_{-1,n-1}(G(a),\omega^{2}\lambda)$ and use (4.1) for
$\mathcal{W}_{0,n+1}(a,\lambda)$ and $\mathcal{W}_{0,n+2}(G^{-1}(a),\omega^{-2}\lambda)$. Next, we use an argument
similar (5.11) to complete the induction step. Thus, the theorem is proved.
∎
Next we investigate $\mathcal{W}_{0,\lfloor\frac{m}{2}\rfloor}(a,\lambda)$.
Theorem 5.2.
If $m\geq 4$ is an even integer, then
$$\displaystyle\mathcal{W}_{-1,\lfloor\frac{m}{2}\rfloor}(a,\lambda)$$
(5.14)
$$\displaystyle=-[2\omega^{2+\mu(G^{-1}(a))+\mu(G^{\frac{m}{2}}(a))}+O\left(%
\lambda^{-\rho}\right)]\exp\left[-L(G^{{\frac{m}{2}+1}}(a),\lambda)-L(a,%
\lambda)\right]$$
$$\displaystyle-[2\omega^{2+\mu(a)+\mu(G^{\frac{m-2}{2}}(a))}+O\left(\lambda^{-%
\rho}\right)]\exp\left[-L(G^{\frac{m-2}{2}}(a),\omega^{m-2}\lambda)-L(G^{m}(a)%
,\omega^{m-2}\lambda)\right].$$
as $\lambda\to\infty$ in the sector
(5.15)
$$-\pi+\frac{4\pi}{m+2}-\delta\leq\arg(\lambda)\leq-\pi+\frac{4\pi}{m+2}+\delta.$$
If $m\geq 4$ is an odd integer, then
$$\displaystyle\mathcal{W}_{-1,\lfloor\frac{m}{2}\rfloor}(a,\lambda)$$
$$\displaystyle=[2\omega^{\frac{5}{4}}+O\left(\lambda^{-\rho}\right)]\exp\left[-%
L(G^{\frac{m+1}{2}}(a),\omega^{-1}\lambda)-L(a,\lambda)\right]$$
(5.16)
$$\displaystyle+[2\omega^{\frac{5}{4}}+O\left(\lambda^{-\rho}\right)]\exp\left[L%
(G^{m+1}(a),\omega^{-2}\lambda)-L(G^{\frac{m-3}{2}}(a),\omega^{m-3}\lambda)%
\right].$$
as $\lambda\to\infty$ in the sector
$$-\pi+\frac{4\pi}{m+2}+\delta\leq\arg(\lambda)\leq-\pi+\frac{6\pi}{m+2}+\delta.$$
Proof.
We will use (5.13) with $n=\lfloor\frac{m}{2}\rfloor$, that is,
$$\mathcal{W}_{-1,\lfloor\frac{m}{2}\rfloor}(a,\lambda)=\frac{4\omega^{\mu(G^{-1%
}(a))+\mu(G^{\lfloor\frac{m}{2}\rfloor}(a))}}{\omega^{\lfloor\frac{m}{2}%
\rfloor-1}\mathcal{W}_{0,\lfloor\frac{m}{2}\rfloor+1}(a,\lambda)}+\frac{%
\mathcal{W}_{-1,\lfloor\frac{m}{2}\rfloor-1}(G(a),\omega^{2}\lambda)\mathcal{W%
}_{0,\lfloor\frac{m}{2}\rfloor+2}(G^{-1}(a),\omega^{-2}\lambda)}{\mathcal{W}_{%
0,\lfloor\frac{m}{2}\rfloor+1}(a,\lambda)}.$$
When $m$ is even, say $m=2k$,
$$\displaystyle\mathcal{W}_{0,k+2}(G^{-1}(a),\omega^{-2}\lambda)$$
$$\displaystyle=\mathcal{W}_{m+2,k+2}(G^{-1}(a),\omega^{-2}\lambda)$$
$$\displaystyle=-\omega^{-k-3}\mathcal{W}_{-1,k-1}(G^{k+2}(a),\omega^{2k+4}\lambda)$$
$$\displaystyle=\omega^{-2}\mathcal{W}_{-1,k-1}(G^{k+2}(a),\omega^{2}\lambda).$$
So
$$\displaystyle\mathcal{W}_{-1,k}(a,\lambda)$$
$$\displaystyle=\frac{4\omega^{\mu(G^{-1}(a))+\mu(G^{k}(a))}}{\omega^{k-1}%
\mathcal{W}_{0,k+1}(a,\lambda)}+\frac{\mathcal{W}_{-1,k-1}(G(a),\omega^{2}%
\lambda)\mathcal{W}_{0,k+2}(G^{-1}(a),\omega^{-2}\lambda)}{\mathcal{W}_{0,k+1}%
(a,\lambda)}$$
$$\displaystyle=\frac{4\omega^{\mu(G^{-1}(a))+\mu(G^{k}(a))}}{\omega^{k-1}%
\mathcal{W}_{0,k+1}(a,\lambda)}+\frac{\mathcal{W}_{-1,k-1}(G(a),\omega^{2}%
\lambda)\mathcal{W}_{-1,k-1}(G^{k+2}(a),\omega^{2}\lambda)}{\omega^{2}\mathcal%
{W}_{0,k+1}(a,\lambda)}$$
Since $\lambda$ lies in (5.15),
$$-\frac{2\left(\lfloor\frac{m}{2}\rfloor-2\right)\pi}{m+2}+\delta\leq-\pi+\frac%
{8\pi}{m+2}-\delta\leq\arg(\omega^{2}\lambda)\leq\pi-\frac{4\left(\lfloor\frac%
{m}{2}\rfloor-1\right)\pi}{m+2}+\delta.$$
$$\displaystyle\mathcal{W}_{-1,k}(a,\lambda)$$
$$\displaystyle=\frac{4\omega^{\mu(G^{-1}(a))+\mu(G^{k}(a))}}{\omega^{k-1}%
\mathcal{W}_{0,k+1}(a,\lambda)}+\frac{\mathcal{W}_{-1,k-1}(G(a),\omega^{2}%
\lambda)\mathcal{W}_{0,k+2}(G^{-1}(a),\omega^{-2}\lambda)}{\mathcal{W}_{0,k+1}%
(a,\lambda)}$$
$$\displaystyle=\frac{4\omega^{\mu(G^{-1}(a))+\mu(G^{k}(a))}}{\omega^{k-1}%
\mathcal{W}_{0,k+1}(a,\lambda)}+\frac{\mathcal{W}_{-1,k-1}(G(a),\omega^{2}%
\lambda)\mathcal{W}_{-1,k-1}(G^{k+2}(a),\omega^{2}\lambda)}{\omega^{2}\mathcal%
{W}_{0,k+1}(a,\lambda)}$$
$$\displaystyle=\frac{4\omega^{\mu(G^{-1}(a))+\mu(G^{k}(a))}}{\omega^{k-1}[2i%
\omega^{-\frac{k+1}{2}}+O\left(\lambda^{-\rho}\right)]\exp\left[L(G^{k+1}(a),%
\omega^{2k-m}\lambda)+L(a,\lambda)\right]}$$
$$\displaystyle+\frac{[2\omega^{\frac{3-k}{2}+\mu(G^{k-1}(a))}+O\left(\lambda^{-%
\rho}\right)]\exp\left[L(a,\lambda)-L(G^{k-1}(a),\omega^{2(k-1)}\lambda)\right%
]}{\omega^{2}[2i\omega^{-\frac{k+1}{2}}+O\left(\lambda^{-\rho}\right)]\exp%
\left[L(G^{k+1}(a),\omega^{2k-m}\lambda)+L(a,\lambda)\right]}$$
$$\displaystyle\times[2\omega^{\frac{3-k}{2}+\mu(G^{2k}(a))}+O\left(\lambda^{-%
\rho}\right)]\exp\left[L(G^{k+1}(a),\lambda)-L(G^{2k}(a),\omega^{2(k-1)}%
\lambda)\right]$$
$$\displaystyle=[-2i\omega^{\frac{3-k}{2}+\mu(G^{-1}(a))+\mu(G^{k}(a))}+O\left(%
\lambda^{-\rho}\right)]\exp\left[-L(G^{k+1}(a),\lambda)-L(a,\lambda)\right]$$
$$\displaystyle-[2i\omega^{\frac{3-k}{2}+\mu(G^{k-1}(a))+\mu(G^{2k}(a))}+O\left(%
\lambda^{-\rho}\right)]\exp\left[-L(G^{k-1}(a),\omega^{2(k-1)}\lambda)-L(G^{2k%
}(a),\omega^{2(k-1)}\lambda)\right]$$
$$\displaystyle=-[2i\omega^{\frac{3-k}{2}+\mu(G^{-1}(a))+\mu(G^{k}(a))}+O\left(%
\lambda^{-\rho}\right)]\exp\left[-L(G^{k+1}(a),\lambda)-L(a,\lambda)\right]$$
$$\displaystyle-[2i\omega^{\frac{3-k}{2}+\mu(a)+\mu(G^{k-1}(a))}+O\left(\lambda^%
{-\rho}\right)]\exp\left[-L(G^{k-1}(a),\omega^{2(k-1)}\lambda)-L(G^{2k}(a),%
\omega^{2(k-1)}\lambda)\right],$$
where we used (5.1) for $\mathcal{W}_{-1,k-1}(G(a),\cdot)$ and $\mathcal{W}_{-1,k-1}(G^{k+2}(a),\cdot)$,
and (4.1) for $\mathcal{W}_{0,k+1}(a,\cdot)$. Finally, we use $\omega^{-\frac{m+2}{4}}=-i$, to get the
desired asymptotic expansion of $\mathcal{W}_{-1,\lfloor\frac{m}{2}\rfloor}(a,\lambda)$.
Next we investigate the case when $m$ is odd, say $m=2k+1$ (so $\lfloor\frac{m}{2}\rfloor=k$).
$$\mathcal{W}_{0,\lfloor\frac{m}{2}\rfloor+1}(a,\lambda)=\omega^{-1}\mathcal{W}_%
{-1,\lfloor\frac{m}{2}\rfloor}(G(a),\omega^{2}\lambda)$$
and
$$\displaystyle\mathcal{W}_{0,k+2}(G^{-1}(a),\omega^{-2}\lambda)$$
$$\displaystyle=\mathcal{W}_{m+2,k+2}(G^{-1}(a),\omega^{-2}\lambda)$$
$$\displaystyle=-\omega^{-k-2}\mathcal{W}_{0,k+1}(G^{k+1}(a),\omega^{2k+2}\lambda)$$
$$\displaystyle=\omega^{-\frac{1}{2}}\mathcal{W}_{0,k+1}(G^{k+1}(a),\omega^{-1}%
\lambda).$$
Similarly to the proof of the theorem for $m$ even,
$$\displaystyle\mathcal{W}_{-1,k}(a,\lambda)$$
$$\displaystyle=\frac{4\omega^{\mu(G^{-1}(a))+\mu(G^{k}(a))}}{\omega^{k-1}%
\mathcal{W}_{0,k+1}(a,\lambda)}+\frac{\mathcal{W}_{-1,k-1}(G(a),\omega^{2}%
\lambda)\mathcal{W}_{0,k+2}(G^{-1}(a),\omega^{-2}\lambda)}{\mathcal{W}_{0,k+1}%
(a,\lambda)}$$
$$\displaystyle=\frac{4\omega^{\mu(G^{-1}(a))+\mu(G^{k}(a))}}{\omega^{k-1}%
\mathcal{W}_{0,k+1}(a,\lambda)}+\frac{\mathcal{W}_{-1,k-1}(G(a),\omega^{2}%
\lambda)\mathcal{W}_{0,k+1}(G^{k+1}(a),\omega^{-1}\lambda)}{\omega^{\frac{1}{2%
}}\mathcal{W}_{0,k+1}(a,\lambda)}$$
$$\displaystyle=\frac{4\omega^{\mu(G^{-1}(a))+\mu(G^{k}(a))}}{\omega^{k-1}[2i%
\omega^{-\frac{k+1}{2}}+O\left(\lambda^{-\rho}\right)]\exp\left[L(G^{k+1}(a),%
\omega^{2k-m}\lambda)+L(a,\lambda)\right]}$$
$$\displaystyle+\frac{[2\omega^{\frac{3-k}{2}+\mu(G^{k-1}(a))}+O\left(\lambda^{-%
\rho}\right)]\exp\left[L(a,\lambda)-L(G^{k-1}(a),\omega^{2(k-1)}\lambda)\right%
]}{\omega^{\frac{1}{2}}[2i\omega^{-\frac{k+1}{2}}+O\left(\lambda^{-\rho}\right%
)]\exp\left[L(G^{k+1}(a),\omega^{2k-m}\lambda)+L(a,\lambda)\right]}$$
$$\displaystyle\times[2i\omega^{-\frac{k+1}{2}}+O\left(\lambda^{-\rho}\right)]%
\exp\left[L(G^{2k+2}(a),\omega^{2k-m-1}\lambda)+L(G^{k+1}(a),\omega^{-1}%
\lambda)\right]$$
$$\displaystyle=[-2i\omega^{\frac{3-k}{2}+\mu(G^{-1}(a))+\mu(G^{k}(a))}+O\left(%
\lambda^{-\rho}\right)]\exp\left[-L(G^{k+1}(a),\omega^{-1}\lambda)-L(a,\lambda%
)\right]$$
$$\displaystyle+[2\omega^{\frac{2-k}{2}+\mu(G^{k-1}(a))}+O\left(\lambda^{-\rho}%
\right)]\exp\left[L(G^{2k+2}(a),\omega^{-2}\lambda)-L(G^{k-1}(a),\omega^{2(k-1%
)}\lambda)\right]$$
$$\displaystyle=-[2i\omega^{\frac{3-k}{2}+\frac{m}{2}}+O\left(\lambda^{-\rho}%
\right)]\exp\left[-L(G^{k+1}(a),\omega^{-1}\lambda)-L(a,\lambda)\right]$$
$$\displaystyle+[2\omega^{\frac{2-k}{2}+\frac{m}{4}}+O\left(\lambda^{-\rho}%
\right)]\exp\left[L(G^{2k+2}(a),\omega^{-2}\lambda)-L(G^{k-1}(a),\omega^{2(k-1%
)}\lambda)\right],$$
where we use (5.1) for $\mathcal{W}_{-1,k-1}(G(a),\cdot)$, and use (4.1) for
$\mathcal{W}_{0,k+1}(a,\cdot)$ and $\mathcal{W}_{0,k+1}(G^{k+1}(a),\cdot)$. Finally, we use
$\omega^{\frac{m+2}{4}}=i$, to get the asymptotic expansion of
$\mathcal{W}_{-1,\lfloor\frac{m}{2}\rfloor}(a,\lambda)$. This completes the proof.
∎
The order of an entire function $g$ is defined by
$$\limsup_{r\rightarrow\infty}\frac{\log\log M(r,g)}{\log r},$$
where $M(r,g)=\max\{|g(re^{i\theta})|:0\leq\theta\leq 2\pi\}$ for $r>0$.
If for some positive real numbers $\sigma,\,c_{1},\,c_{2}$, we have $\exp[c_{1}r^{\sigma}]\leq M(r,g)\leq\exp[c_{2}r^{\sigma}]$ for all large $r$, then the order of $g$ is $\sigma$.
Corollary 5.3.
Let $1\leq n\leq\frac{m}{2}$. Then
the entire functions $\mathcal{W}_{-1,n}(a,\cdot)$ are of order $\frac{1}{2}+\frac{1}{m}$, and hence they have infinitely many zeros in the complex plane. Moreover, $\mathcal{W}_{-1,n}(a,\cdot)$ have at most finitely many zeros in the sector
$$-\frac{2(n-1)\pi}{m+2}+\delta\leq\arg(\lambda)\leq\pi+\frac{4\pi}{m+2}-\delta.$$
6. Relation between eigenvalues of $H_{\ell}$ and zeros of $\mathcal{W}_{-1,n}(a,\cdot)$
In this section, we will relate the eigenvalues of $H_{\ell}$ with zeros of some entire function
$\mathcal{W}_{-1,n}(a,\cdot)$.
Suppose that $\ell=2s-1$ is odd with $1\leq\ell\leq m-1$. Then (1.1) becomes (3.2) by the
scaling $v(z)=u(-iz)$, and $v$ decays in the Stokes sectors $S_{-s}$ and $S_{s}$. Since $f_{s-1}$ and $f_{s}$ are
linearly independent, for some $D_{s}$ and $\widetilde{D}_{s}$ one can write
$$f_{-s}(z,a,\lambda)=D_{s}(a,\lambda)f_{s-1}(z,a,\lambda)+\widetilde{D}_{s}(a,%
\lambda)f_{s}(z,a,\lambda).$$
Then one finds
$$D_{s}(a,\lambda)=\frac{\mathcal{W}_{-s,s}(a,\lambda)}{\mathcal{W}_{s-1,s}(a,%
\lambda)}\quad\text{and}\quad\widetilde{D}_{s}(a,\lambda)=\frac{\mathcal{W}_{-%
s,s-1}(a,\lambda)}{\mathcal{W}_{s,s-1}(a,\lambda)}.$$
Also it is easy to see that $\lambda$ is an eigenvalue of $H_{\ell}$ if and only if $D_{s}(a,\lambda)=0$ if and only
if $\mathcal{W}_{-s,s}(a,\lambda)=0$. Since
$\mathcal{W}_{-s,s}(a,\lambda)=\omega^{s-1}\mathcal{W}_{-1,2s-1}(G^{-s+1}(a),%
\omega^{-2s+2}\lambda)$, by Corollary
5.3 $\mathcal{W}_{-s,s}(a,\lambda)$ has at most finitely many zeros in the sector $-\frac{2(2s-2)\pi}{m+2}+\delta\leq\arg(\omega^{-2s+2}\lambda)\leq\pi+\frac{4%
\pi}{m+2}-\delta,$ that is,
$$\delta\leq\arg(\lambda)\leq\pi+\frac{2(2s+1)\pi}{m+2}-\delta.$$
Next, by symmetry one can show that $\mathcal{W}_{-s,s}(a,\lambda)$ has at most finitely many zeros in the sector
$\pi\leq\arg(\lambda)\leq 2\pi-\delta$. For that, we examine $H_{\ell}$ with $P(z)$ replaced by
$\overline{P(\overline{z})}$ whose coefficient vector is
$\overline{a}:=(\overline{a}_{1},\,\overline{a}_{2},\dots,\overline{a}_{m})$. Then one sees that
$\mathcal{W}_{-s,s}(a,\lambda)=0$ if and only if $\mathcal{W}_{-s,s}(\overline{a},\overline{\lambda})=0$.
$\mathcal{W}_{-s,s}(\overline{a},\overline{\lambda})$ has at most finitely many zeros in the sector
$\delta\leq\arg(\overline{\lambda})\leq\pi$ by the arguments above. Thus, $\mathcal{W}_{-s,s}(a,\lambda)$ has at
most finitely many zeros in the sector $\pi\leq\arg(\lambda)\leq 2\pi-\delta$, and has infinitely many zeros in
the sector $|\arg(\lambda)|\leq\delta$ since it is an entire function of order $\frac{1}{2}+\frac{1}{m}\in(0,1)$.
Suppose that $\ell=2s$ is even with $1\leq\ell\leq m-1$. Then (1.1) becomes (3.4) by the
scaling $y(z)=u(-i\omega^{-\frac{1}{2}}z)$, and $y$ decays in the Stokes sectors $S_{-s}$ and $S_{s+1}$. We then
see that the coefficient vector $\widetilde{a}$ of the polynomial $\omega^{-1}P(\omega^{-\frac{1}{2}}z)$ becomes
$$\widetilde{a}=G^{-\frac{1}{2}}(a).$$
Now one can express $f_{-s}$ as a linear combination of
$f_{s}$ and $f_{s+1}$ as follows.
$$f_{-s}(z,\widetilde{a},\omega^{-1}\lambda)=\frac{\mathcal{W}_{-s,s+1}(%
\widetilde{a},\omega^{-1}\lambda)}{\mathcal{W}_{s,s+1}(\widetilde{a},\omega^{-%
1}\lambda)}f_{s}(z,\widetilde{a},\omega^{-1}\lambda)+\frac{\mathcal{W}_{-s,s}(%
\widetilde{a},\omega^{-1}\lambda)}{\mathcal{W}_{s+1,s}(\widetilde{a},\omega^{-%
1}\lambda)}f_{s+1}(z,\widetilde{a},\omega^{-1}\lambda).$$
Thus, $\lambda$ is an eigenvalue of $H_{\ell}$ if and only if
$\mathcal{W}_{-s,s+1}(\widetilde{a},\omega^{-1}\lambda)=0$. Since
$$\mathcal{W}_{-s,s+1}(\widetilde{a},\omega^{-1}\lambda)=\omega^{s-1}\mathcal{W}%
_{-1,2s}(G^{-s+1}(\widetilde{a}),\omega^{-2s+1}\lambda),$$
by Corollary 5.3,
$\mathcal{W}_{-s,s+1}(\widetilde{a},\omega^{-1}\lambda)$ has at most finitely many zeros in the sector
$-\frac{2(2s-1)\pi}{m+2}+\delta\leq\arg(\omega^{-2s+1}\lambda)\leq\pi+\frac{4%
\pi}{m+2}-\delta,$ that is,
$$\delta\leq\arg(\lambda)\leq\pi+\frac{4s\pi}{m+2}-\delta.$$
This is true for each $a\in\mathbb{C}^{m}$. So one can show that $\mathcal{W}_{-s,s+1}(\widetilde{a},\omega^{-1}\lambda)$
has at most finitely many zeros in the sector $\pi\leq\arg(\lambda)\leq 2\pi-\delta$ by symmetry, similar to the
case when $\ell$ is odd. Thus, $\mathcal{W}_{-s,s+1}(a,\lambda)$ has infinitely many zeros in the sector
$|\arg(\lambda)|\leq\delta$.
7. Proof of Theorem 1.1 when $1\leq\ell<\lfloor\frac{m}{2}\rfloor$
In this section, we prove Theorem 1.1 for
$1\leq\ell<\lfloor\frac{m}{2}\rfloor$ and in doing so, we will use
the following proposition on univalent functions.
Proposition 7.1 ([17, p. 216]).
Let $A(\mu)$ be analytic in the region $S=\{\mu\in\mathbb{C}:\alpha\leq\arg(\mu)\leq\beta,|\mu|\geq M\}$ for some
$\alpha,\beta\in\mathbb{R}$ with $\beta-\alpha<\pi$ and for some $M>0$.
Suppose that $A(\mu)\to 0$ as $\mu\to\infty$ in $S$. Then there
exist $\widetilde{\alpha},\widetilde{\beta},\widetilde{M}$ such
that $\mu(1+A(\mu))$ is univalent in $\{\mu\in\mathbb{C}:\widetilde{\alpha}\leq\arg(\mu)\leq\widetilde{\beta},|\mu|%
\geq\widetilde{M}\}\subset S$.
Proof.
See [17, p. 216] or [27, Thm. 3.4] for a proof.
∎
We will consider two cases; when $\ell$ is odd and when $\ell$ is
even.
Proof of Theorem 1.1 when $1\leq\ell<\lfloor\frac{m}{2}\rfloor$ is odd.
We will closely follow his proof of Theorem 29.1 in [27]
where Sibuya computed the leading term in the asymptotics
(1.3) for $\ell=1$.
Suppose that $1\leq\ell=2s-1<\lfloor\frac{m}{2}\rfloor$ is odd.
Recall that when $\ell$ is odd, $\lambda$ is an eigenvalue of
$H_{\ell}$ if and only if
$\mathcal{W}_{-s,s}(a,\lambda)=0$. Also, in Section 6 we showed that $\mathcal{W}_{-s,s}(a,\lambda)$ has all zeros except finitely many in the sector
$|\arg(\lambda)|\leq\delta$ and hence,
all the eigenvalues $\lambda$ of $H_{\ell}$ lie in the sector $|\arg(\lambda)|\leq\delta$ if $|\lambda|$ is large
enough.
Since
$$\mathcal{W}_{-s,s}(a,\lambda)=\omega^{s-1}\mathcal{W}_{-1,2s-1}(G^{-s+1}(a),%
\omega^{-2s+2}\lambda),$$
we will use (5.3) to investigate asymptotics of large eigenvalues. Suppose that $\mathcal{W}_{-s,s}(a,\lambda)=0$ and $|\lambda|$ is large enough.
Then from (5.3) with $n=2s-1$, and with $a$ and $\lambda$ replaced by $G^{-s+1}(a)$ and $\omega^{-2s+2}\lambda$, respectively, we have
$$\displaystyle\left[1+O\left(\lambda^{-\rho}\right)\right]$$
$$\displaystyle\exp\left[L(G^{s}(a),\omega^{2s}\lambda)-L(G^{-s}(a),\omega^{-2s}%
\lambda)\right]$$
$$\displaystyle\times\exp\left[L(G^{s-1}(a),\omega^{2s-2}\lambda)-L(G^{-s+1}(a),%
\omega^{-2s+2}\lambda)\right]=-\omega^{2\nu(G^{s}(a))}.$$
Also, since $\left[1+O\left(\lambda^{-\rho}\right)\right]=\exp\left[O\left(\lambda^{-\rho}%
\right)\right]$ and
$\omega^{2\nu(G^{s}(a))}=\exp\left[\frac{4\pi\nu(G^{s}(a))}{m+2}\right]$,
$$\displaystyle\exp\left[L(G^{s}(a),\omega^{2s}\lambda)-L(G^{-s}(a),\omega^{-2s}%
\lambda)\right]$$
(7.1)
$$\displaystyle\times\exp\left[L(G^{s-1}(a),\omega^{2s-2}\lambda)-L(G^{-s+1}(a),%
\omega^{-2s+2}\lambda)-\frac{4\pi\nu(G^{s}(a))}{m+2}+O\left(\lambda^{-\rho}%
\right)\right]=-1.$$
For each odd integer $\ell=2s-1$ in $1\leq\ell<\lfloor\frac{m}{2}\rfloor$, we define
$$\displaystyle h_{m,\ell}(\lambda)=$$
$$\displaystyle L(G^{s}(a),\omega^{2s}\lambda)-L(G^{-s}(a),\omega^{-2s}\lambda)$$
$$\displaystyle+L(G^{s-1}(a),\omega^{2s-2}\lambda)-L(G^{-s+1}(a),\omega^{-2s+2}%
\lambda)-\frac{4\pi\nu(G^{s}(a))}{m+2}+O\left(\lambda^{-\rho}\right).$$
Then by Corollary 3.4,
$$\displaystyle h_{m,\ell}(\lambda)$$
$$\displaystyle=K_{m}\left(e^{\frac{2s\pi}{m}i}-e^{-\frac{2s\pi}{m}i}+e^{\frac{2%
(s-1)\pi}{m}i}-e^{-\frac{2(s-1)\pi}{m}i}\right)\lambda^{\frac{m+2}{2m}}(1+o(1))$$
$$\displaystyle=2iK_{m}\left(\sin\left(\frac{2s\pi}{m}\right)+\sin\left(\frac{2s%
\pi}{m}-\frac{2\pi}{m}\right)\right)\lambda^{\frac{m+2}{2m}}(1+o(1))$$
(7.2)
$$\displaystyle=4iK_{m}\cos\left(\frac{\pi}{m}\right)\sin\left(\frac{(2s-1)\pi}{%
m}\right)\lambda^{\frac{m+2}{2m}}(1+o(1))\quad\text{as}\quad\lambda\to\infty,$$
in the sector $|\arg(\lambda)|\leq\delta$. Since $K_{m}>0$ and $0<\frac{(2s-1)\pi}{m}<\pi$, the function
$h_{m,\ell}(\cdot)$ maps the region $|\lambda|\geq M_{1}$ for some large $M_{1}$ and $|\arg(\lambda)|\leq\delta$ into
a region containing $|\lambda|\geq M_{2}$ for some large $M_{2}$ and $|\arg(\lambda)-\frac{\pi}{2}|\leq\varepsilon_{1}$
for some $\varepsilon_{1}>0$. Also, $h_{m,\ell}(\cdot)$ is analytic in this region due to the analyticity of
$f(0,a,\lambda)$ in Theorem 3.2. Following Sibuya, we will show that for every large positive integer $n$
there exists $\lambda_{n}$ such that
(7.3)
$$h_{m,\ell}(\lambda_{n})=\left(2n+1\right)\pi i.$$
Next, Proposition 7.1 with $\mu=\lambda^{\frac{m+2}{2m}}$ implies that there exist $M_{1}^{\prime}>0$ and
$0<\delta^{\prime}<\frac{\pi}{6}$ such that $\mu(1+A(\mu))$ is univalent in the sector $|\arg(\mu)|\leq\delta^{\prime}$ and
$|\mu|\geq M_{1}^{\prime}.$ Thus, for each $n\in\mathbb{Z}$, there is at most one $\mu_{n}$ in the sector $|\arg(\mu)|\leq\delta^{\prime}$
and $|\mu|\geq M_{1}^{\prime}$ such that
(7.4)
$$\mu_{n}(1+A(\mu_{n}))=B_{n},$$
where $A$ is the error term in (7.2) and
$$B_{n}=\frac{\left(2n+1\right)\pi}{4K_{m}\cos\left(\frac{\pi}{m}\right)\sin%
\left(\frac{(2s-1)\pi}{m}\right)}.$$
Then there exists $M_{1}^{\prime\prime}\geq M_{1}^{\prime}$ such that in the sector $|\arg(\mu)|\leq\delta^{\prime}$ and $|\mu|\geq M_{1}^{\prime\prime}$,
$$|A(\mu)|<\frac{1}{5}.$$
Also we can take $n$ large enough so that $\frac{1}{2}B_{n}\geq M_{1}^{\prime\prime}$. Next, we define
$$A_{n}:=\{|A(\mu)|:|\arg(\mu)|\leq\delta^{\prime},|\mu|\geq B_{n}\}.$$
Then $\lim_{n\to\infty}A_{n}=0$
Suppose that $n$ is so large that $A_{n}\leq\frac{1}{2}\sin(\delta^{\prime})<\frac{1}{4}$. Then the disk defined by
$|\mu-B_{n}|\leq\frac{A_{n}}{1-2A_{n}}B_{n}$ is contained in the sector $|\arg(\mu)|\leq\delta^{\prime}$ and
$|\mu|\geq\frac{1}{2}B_{n}$. Moreover, on the circle $|\mu-B_{n}|=\frac{A_{n}}{1-2A_{n}}B_{n}$,
$$\displaystyle|\mu-B_{n}|-|\mu||A(\mu)|$$
$$\displaystyle\geq|\mu-B_{n}|-|\mu|A_{n}$$
$$\displaystyle\geq|\mu-B_{n}|(1-A_{n})-A_{n}B_{n}$$
$$\displaystyle\geq\frac{1-A_{n}}{1-2A_{n}}A_{n}B_{n}-A_{n}B_{n}>0.$$
Thus, by the Rouché’s theorem in complex analysis (see, e. g., [7, p. 125]), $\mu-B_{n}$ and
$\mu(1+A(\mu))-B_{n}$ have the same number of zeros in the disk and hence, (7.4) has exactly one $\mu_{n}$ in
the sector. Therefore, there exists $N=N(m,a)>0$ such that (7.3) has exactly one solution $\lambda_{n}$ for
all integers $n\geq N$.
Next, since
$$\displaystyle-\frac{\nu(G^{s}(a))}{m}\ln(\omega^{2s}\lambda)+\frac{\nu(G^{-s}(%
a))}{m}\ln(\omega^{-2s}\lambda)$$
$$\displaystyle-\frac{\nu(G^{s-1}(a))}{m}\ln(\omega^{2s-2}\lambda)+\frac{\nu(G^{%
-s+1}(a))}{m}\ln(\omega^{-2s+2}\lambda)-\frac{4\nu(G^{s}(a))}{m+2}\pi i$$
$$\displaystyle=$$
$$\displaystyle\frac{\nu(G^{s}(a))}{m}\left(-\ln(\omega^{2s}\lambda)+\ln(\omega^%
{-2s}\lambda)+\ln(\omega^{2s-2}\lambda)-\ln(\omega^{-2s+2}\lambda)\right)-%
\frac{4\nu(G^{s}(a))}{m+2}\pi i$$
$$\displaystyle=$$
$$\displaystyle-\frac{4}{m}\nu(G^{s}(a))\pi i,$$
using Corollary 3.3, (7.3) becomes
(7.5)
$$\left(2n+1\right)\pi i=\sum_{j=0}^{m+1}e_{\ell,j}(a)\lambda_{n}^{\frac{1}{2}+%
\frac{1-j}{m}}-\frac{4}{m}\nu(G^{s}(a))\pi i+O\left(\lambda^{-\rho}\right),$$
where
$$\displaystyle e_{\ell,j}(a)=$$
$$\displaystyle K_{m,j}(G^{s}(a))\omega^{2s\left(\frac{1}{2}+\frac{1-j}{m}\right%
)}-K_{m,j}(G^{-s}(a))\omega^{-2s\left(\frac{1}{2}+\frac{1-j}{m}\right)}$$
$$\displaystyle+K_{m,j}(G^{s-1}(a))\omega^{2(s-1)\left(\frac{1}{2}+\frac{1-j}{m}%
\right)}-K_{m,j}(G^{-s+1}(a))\omega^{-2(s-1)\left(\frac{1}{2}+\frac{1-j}{m}%
\right)}$$
$$\displaystyle=$$
$$\displaystyle K_{m,j}(a)\left(\omega^{-sj+2s\left(\frac{1}{2}+\frac{1-j}{m}%
\right)}-\omega^{sj-2s\left(\frac{1}{2}+\frac{1-j}{m}\right)}\right)$$
$$\displaystyle+K_{m,j}(a)\left(\omega^{-(s-1)j+2(s-1)\left(\frac{1}{2}+\frac{1-%
j}{m}\right)}-\omega^{(s-1)j-2(s-1)\left(\frac{1}{2}+\frac{1-j}{m}\right)}\right)$$
$$\displaystyle=$$
$$\displaystyle 4iK_{m,j}(a)\sin\left(\frac{(1-j)\ell\pi}{m}\right)\cos\left(%
\frac{(1-j)\pi}{m}\right),$$
where we used Lemma 3.5.
In summary, we have showed that for each $a\in\mathbb{C}^{m}$, there exists
$N_{0}=N_{0}(m,a)\leq N$ such that $\{\lambda_{n}\}_{n\geq N_{0}}$ is the
set of all eigenvalues that satisfy (7.5), since
there are only at most finitely many eigenvalues outside
$|\arg(\lambda)|\leq\varepsilon$ or inside $|\lambda|\leq M$ for
any $\varepsilon>0,\,M>0$. In order to complete proof, we need to
show that $N_{0}$ does not depend on $a\in\mathbb{C}^{m}$.
We will prove that for each $R>0$, if $|a|\leq R$, then $N_{0}(m,a)=N_{0}(m,0)$.
The error terms in (3.5) and (3.6) are uniform on the closed ball $|a|\leq R$ and hence, so is $A(\mu)$ as $\mu\to\infty$
in the sector $|\arg(\mu)|\leq\delta^{\prime}$. Thus, we can choose $N$ independent of $|a|\leq R$. Since $\mu(1+A(\mu))$ is univalent
in the sector $|\arg(\mu)|\leq\delta^{\prime}$ and $|\mu|\geq M_{1}^{\prime\prime}\geq M_{1}^{\prime}$, by the implicit function theorem and (7.4) the function $a\mapsto\mu_{n}(a)$ is
continuous on $|a|\leq R$ for each $n\geq N$ and hence, so is $a\mapsto\lambda_{n}(a)$ for $n\geq N$.
Next, we claim that for all $|a|\leq R$, there are exactly the same number of
eigenvalues that are not in $\{\lambda_{n}(a)\}_{n\geq N}$. To prove this, we will use the Hurwitz’s theorem (see, e. g., [7, p. 152]) in complex analysis. That is,
if a sequence of analytic functions converges uniformly to an
analytic function on any compact sets, then eventually functions in the sequence and the limit function have the
same number of zeros in any open set whose boundary does not contain any zeros of the limit function. The
Hurwitz’s theorem implies that since the eigenvalues are the zeros of the entire function
$\mathcal{W}_{-s,s}(a,\lambda)$, they vary continuously as $a$ and hence, there is no sudden appearance or
disappearance of eigenvalues. Also, none of the eigenvalues that are not in $\{\lambda_{n}(a)\}_{n\geq N}$ can be
continuously mapped to $\lambda_{n}(a)$ for some $n\geq N$ as $a$ varies and hence, the claim is proved. This
completes the proof.
∎
Next we prove Theorem 1.1 for $1\leq\ell<\lfloor\frac{m}{2}\rfloor$ is even.
Proof of Theorem 1.1 when $1\leq\ell<\lfloor\frac{m}{2}\rfloor$ is even.
Proof is very similar to the case when $1\leq\ell<\lfloor\frac{m}{2}\rfloor$ is odd. Let $\ell=2s$ for some
$s\in\mathbb{N}$.
Recall that $\lambda$ is an eigenvalue of $H_{\ell}$ if and only if
$\mathcal{W}_{-1,2s}(G^{-s+1}(\widetilde{a}),\omega^{-2s+1}\lambda)=0$. Then from (5.3), we have
$$\displaystyle\exp\left[L(G^{s+1}(\widetilde{a}),\omega^{2s+1}\lambda)-L(G^{-s}%
(\widetilde{a}),\omega^{-2s-1}\lambda)\right]$$
(7.6)
$$\displaystyle\times\exp\left[L(G^{s}(\widetilde{a}),\omega^{2s-1}\lambda)-L(G^%
{-s+1}(\widetilde{a}),\omega^{-2s+1}\lambda)+O\left(\lambda^{-\rho}\right)%
\right]=-1,$$
where $\widetilde{a}=G^{-\frac{1}{2}}(a)$ and where we used
$\left[1+O\left(\lambda^{-\rho}\right)\right]=\exp[O\left(\lambda^{-\rho}\right)]$
again.
Like in the case when $1\leq\ell<\lfloor\frac{m}{2}\rfloor$ is
odd, from Lemma 3.3, we have
(7.7)
$$\left(2n+1\right)\pi i=\sum_{j=0}^{m+1}e_{\ell,j}(a)\lambda_{n}^{\frac{1}{2}+%
\frac{1-j}{m}}+O\left(\lambda^{-\rho}\right),$$
where for $0\leq j\leq m+1$,
$$\displaystyle e_{\ell,j}(a)=$$
$$\displaystyle K_{m,j}(G^{s+\frac{1}{2}}(a))\omega^{(2s+1)\left(\frac{1}{2}+%
\frac{1-j}{m}\right)}-K_{m,j}(G^{-s-\frac{1}{2}}(a))\omega^{-(2s+1)\left(\frac%
{1}{2}+\frac{1-j}{m}\right)}$$
$$\displaystyle+K_{m,j}(G^{s-\frac{1}{2}}(a))\omega^{(2s-1)\left(\frac{1}{2}+%
\frac{1-j}{m}\right)}-K_{m,j}(G^{-s+\frac{1}{2}}(a))\omega^{-(2s-1)\left(\frac%
{1}{2}+\frac{1-j}{m}\right)}$$
$$\displaystyle=$$
$$\displaystyle\sum_{k=0}^{j}(-1)^{k}K_{m,j,k}b_{j,k}(a)\left(\omega^{-j\frac{%
\ell+1}{2}+(\ell+1)\left(\frac{1}{2}+\frac{1-j}{m}\right)}-\omega^{j\frac{\ell%
+1}{2}-(\ell+1)\left(\frac{1}{2}+\frac{1-j}{m}\right)}\right.$$
$$\displaystyle\qquad\qquad\qquad\qquad\qquad\left.+\omega^{-j\frac{\ell-1}{2}+(%
\ell-1)\left(\frac{1}{2}+\frac{1-j}{m}\right)}-\omega^{j\frac{\ell-1}{2}-(\ell%
-1)\left(\frac{1}{2}+\frac{1-j}{m}\right)}\right)$$
$$\displaystyle=$$
$$\displaystyle 4i\sum_{k=0}^{j}(-1)^{k}K_{m,j,k}b_{j,k}(a)\sin\left(\frac{(1-j)%
\ell\pi}{m}\right)\cos\left(\frac{(1-j)\pi}{m}\right),$$
where we used Lemma 3.5 and $\ell=2s$. Then we use the
arguments in the case when $\ell$ is odd to show that there exists
$N_{0}=N_{0}(m)$ such that the set $\{\lambda_{n}\}_{n\geq N_{0}}$ of all
eigenvalues satisfy (7.7) and hence, the proof is
completed.
∎
8. Proof of Theorem 1.1 when $\ell=\lfloor\frac{m}{2}\rfloor$ and when $\frac{m}{2}<\ell\leq m-1$.
In this section, we prove Theorem 1.1 for
$\ell=\lfloor\frac{m}{2}\rfloor$. We first prove the theorem when
$m$ is even, and later, we will treat the cases when $m$ is odd.
Then at the end of the section, we will prove the theorem when
$\frac{m}{2}<\ell\leq m-1$, by scaling. Proof of existence of
$N_{0}=N_{0}(m)$ is the same as in Section 7, so below we
will omit this part of proof.
8.1. When $m$ is even
We further divide the case into when $\ell$ is odd and when $\ell$ is even.
Proof of Theorem 1.1 when $m$ is even and $\ell=\lfloor\frac{m}{2}\rfloor$ is odd.
Suppose that $m$ is even and $\ell=\frac{m}{2}=2s-1$ for some $s\in\mathbb{N}$. Recall that $\lambda$ is an eigenvalue of
$H_{\ell}$ if and only if $\mathcal{W}_{-s,s}(a,\lambda)=0$ if and only if
$\mathcal{W}_{-1,2s-1}(G^{-s+1}(a),\omega^{-2s+2}\lambda)=0$.
So if $\mathcal{W}_{-1,2s-1}(G^{-s+1}(a),\omega^{-2s+2}\lambda)=0$ and if $|\lambda|$ is large enough, then from
(5.14),
$$\displaystyle\exp\left[L(G^{3s-1}(a),\omega^{2s-2}\lambda)-L(G^{s+1}(a),\omega%
^{-2s+2}\lambda)\right]$$
$$\displaystyle\times\exp\left[L(G^{s-1}(a),\omega^{2s-2}\lambda)-L(G^{-s+1}(a),%
\omega^{-2s+2}\lambda)+O\left(\lambda^{-\rho}\right)\right]=-\omega^{4\nu(G^{s%
}(a))}.$$
Then by Lemma 3.3,
$$\left(\frac{8\nu(G^{s}(a))}{m+2}+2n+1\right)\pi i=\sum_{j=0}^{m+1}e_{\ell,j}(a%
)\lambda_{n}^{\frac{1}{2}+\frac{1-j}{m}}+\frac{16(s-1)\nu(G^{s}(a))}{m(m+2)}%
\pi i+O\left(\lambda^{-\rho}\right),$$
where for $0\leq j\leq m+1$, the coefficients $e_{\ell,j}(a)$ are given by
$$\displaystyle e_{\ell,j}(a)=$$
$$\displaystyle K_{m,j}(G^{3s-1}(a))\omega^{2(s-1)\left(\frac{1}{2}+\frac{1-j}{m%
}\right)}-K_{m,j}(G^{s+1}(a))\omega^{-2(s-1)\left(\frac{1}{2}+\frac{1-j}{m}%
\right)}$$
$$\displaystyle+K_{m,j}(G^{s-1}(a))\omega^{2(s-1)\left(\frac{1}{2}+\frac{1-j}{m}%
\right)}-K_{m,j}(G^{-s+1}(a))\omega^{-2(s-1)\left(\frac{1}{2}+\frac{1-j}{m}%
\right)}$$
$$\displaystyle=$$
$$\displaystyle K_{m,j}(a)\left(\omega^{(s+1)j+2(s-1)\left(\frac{1}{2}+\frac{1-j%
}{m}\right)}-\omega^{-(s+1)j-2(s-1)\left(\frac{1}{2}+\frac{1-j}{m}\right)}\right.$$
$$\displaystyle\qquad\qquad\qquad\qquad\left.+\omega^{-(s-1)j+2(s-1)\left(\frac{%
1}{2}+\frac{1-j}{m}\right)}-\omega^{(s-1)j-2(s-1)\left(\frac{1}{2}+\frac{1-j}{%
m}\right)}\right)$$
$$\displaystyle=$$
$$\displaystyle 2iK_{m,j}(a)\left(1+(-1)^{j}\right)\sin\left(\frac{(1-j)(2s-2)%
\pi}{m}\right)$$
$$\displaystyle=$$
$$\displaystyle 4iK_{m,j}(a)\sin\left(\frac{(1-j)\pi}{2}\right)\cos\left(\frac{(%
1-j)\pi}{m}\right),$$
where we used Lemma 3.5.
∎
Proof of Theorem 1.1 when $m$ is even and $\ell=\lfloor\frac{m}{2}\rfloor$ is even.
Let $\ell=\frac{m}{2}=2s$ for some $s\in\mathbb{N}$. Recall that $\lambda$ is an eigenvalue of $H_{\ell}$ if and only if
$\mathcal{W}_{-1,2s}(G^{-s+1}(\widetilde{a}),\omega^{-2s+1}\lambda)=0$.
Suppose that
$\mathcal{W}_{-1,2s}(G^{-s+1}(\widetilde{a}),\omega^{-2s+1}\lambda)=\mathcal{W}%
_{-1,2s}(G^{-s+\frac{1}{2}}(a),\omega^{-2s+1}\lambda)=0$.
Then by (5.14),
$$\displaystyle\exp\left[L(G^{3s+\frac{1}{2}}(a),\omega^{2s-1}\lambda)-L((G^{s+%
\frac{3}{2}}(a),\omega^{-2s+1}\lambda)\right]$$
$$\displaystyle\times\exp\left[L(G^{s-\frac{1}{2}}(a),\omega^{2s-1}\lambda)-L(G^%
{-s+\frac{1}{2}}(a),\omega^{-2s+1}\lambda)+O\left(\lambda^{-\rho}\right)\right%
]=-1.$$
Then like before,
$$\left(2n+1\right)\pi i=\sum_{j=0}^{m+1}e_{\ell,j}(a)\lambda_{n}^{\frac{1}{2}+%
\frac{1-j}{m}}+O\left(\lambda^{-\rho}\right),$$
where for $0\leq j\leq m+1$,
$$\displaystyle e_{\ell,j}(a)=$$
$$\displaystyle K_{m,j}(G^{3s+\frac{1}{2}}(a))\omega^{(2s-1)\left(\frac{1}{2}+%
\frac{1-j}{m}\right)}-K_{m,j}((G^{s+\frac{3}{2}}(a))\omega^{-(2s-1)\left(\frac%
{1}{2}+\frac{1-j}{m}\right)}$$
$$\displaystyle+K_{m,j}(G^{s-\frac{1}{2}}(a))\omega^{(2s-1)\left(\frac{1}{2}+%
\frac{1-j}{m}\right)}-K_{m,j}(G^{-s+\frac{1}{2}}(a))\omega^{-(2s-1)\left(\frac%
{1}{2}+\frac{1-j}{m}\right)}$$
$$\displaystyle=$$
$$\displaystyle 2i\sum_{k=0}^{j}(-1)^{k}K_{m,j,k}b_{j,k}(a)\left(1+(-1)^{j}%
\right)\sin\left(\frac{(1-j)(2s-1)\pi}{m}\right)$$
$$\displaystyle=$$
$$\displaystyle 4i\sum_{k=0}^{j}(-1)^{k}K_{m,j,k}b_{j,k}(a)\sin\left(\frac{(1-j)%
\pi}{2}\right)\cos\left(\frac{(1-j)\pi}{m}\right).$$
∎
8.2. When $m$ is odd
We divide the case into when $\ell$ is odd and when $\ell$ is even.
Proof of Theorem 1.1 when $m$ and $\ell=\lfloor\frac{m}{2}\rfloor$ are odd.
Let $m$ and $\ell=\frac{m-1}{2}=2s-1$ be odd. Suppose that $\lambda$ is an eigenvalue of $H_{\ell}$. Since
$\lambda$ is an eigenvalue of $H_{\ell}$ if and only if
$\mathcal{W}_{-1,2s-1}(G^{-s+1}(a),\omega^{-2s+2}\lambda)=0$,
if $|\lambda|$ is large enough, then from (5.16),
$$\displaystyle\exp\left[L(G^{s-1}(a),\omega^{2s-2}\lambda)-L(G^{-s+1}(a),\omega%
^{-2s+2}\lambda)\right]$$
$$\displaystyle\times\exp\left[-L(G^{3s+1}(a),\omega^{-2s}\lambda)-L(G^{s+1}(a),%
\omega^{-2s+1}\lambda)+O\left(\lambda^{-\rho}\right)\right]=-1.$$
Then, from Lemma 3.3,
$$\left(2n+1\right)\pi i=\sum_{j=0}^{m+1}e_{\ell,j}(a)\lambda_{n}^{\frac{1}{2}+%
\frac{1-j}{m}}+O\left(\lambda^{-\rho}\right),$$
where for $0\leq j\leq m+1$,
$$\displaystyle e_{\ell,j}(a)=$$
$$\displaystyle K_{m,j}(G^{s-1}(a))\omega^{(2s-2)\left(\frac{1}{2}+\frac{1-j}{m}%
\right)}-K_{m,j}(G^{-s+1}(a))\omega^{(-2s+2)\left(\frac{1}{2}+\frac{1-j}{m}%
\right)}$$
$$\displaystyle-K_{m,j}(G^{3s+1}(a))\omega^{-2s\left(\frac{1}{2}+\frac{1-j}{m}%
\right)}-K_{m,j}(G^{s+1}(a))\omega^{(-2s+1)\left(\frac{1}{2}+\frac{1-j}{m}%
\right)}$$
$$\displaystyle=$$
$$\displaystyle K_{m,j}(a)\left(\omega^{-j(s-1)+(2s-2)\left(\frac{1}{2}+\frac{1-%
j}{m}\right)}-\omega^{j(s-1)-(2s-2)\left(\frac{1}{2}+\frac{1-j}{m}\right)}\right.$$
$$\displaystyle\qquad\qquad\qquad\left.-\omega^{-j(3s+1)-2s\left(\frac{1}{2}+%
\frac{1-j}{m}\right)}-\omega^{-j(s+1)-(2s-1)\left(\frac{1}{2}+\frac{1-j}{m}%
\right)}\right)$$
$$\displaystyle=$$
$$\displaystyle K_{m,j}(a)\left(\omega^{(1-j)(2s-2)\left(\frac{1}{2}+\frac{1}{m}%
\right)}-\omega^{-(1-j)(2s-2)\left(\frac{1}{2}+\frac{1}{m}\right)}\right.$$
$$\displaystyle\qquad\qquad\qquad\left.-\omega^{js-2s\left(\frac{1}{2}+\frac{1-j%
}{m}\right)}+\omega^{-js+2s\left(\frac{1}{2}+\frac{1-j}{m}\right)}\right)$$
$$\displaystyle=$$
$$\displaystyle 2iK_{m,j}(a)\left(\sin\left(\frac{(m-3)(1-j)\pi}{2m}\right)+\sin%
\left(\frac{(m+1)(1-j)\pi}{2m}\right)\right)$$
$$\displaystyle=$$
$$\displaystyle 4iK_{m,j}(a)\sin\left(\frac{(1-j)\ell\pi}{m}\right)\cos\left(%
\frac{(1-j)\pi}{m}\right).$$
∎
Proof of Theorem 1.1 when $m$ is odd and $\ell=\lfloor\frac{m}{2}\rfloor$ is even.
Let $\ell=\frac{m-1}{2}=2s$ for some $s\in\mathbb{N}$. Then $\lambda$ is an eigenvalue of $H_{\ell}$ if and only if
$\mathcal{W}_{-1,2s}(G^{-s+1}(\widetilde{a}),\omega^{-2s+1}\lambda)=0$.
If $\mathcal{W}_{-1,2s}(G^{-s+1}(\widetilde{a}),\omega^{-2s+1}\lambda)=0$ and if $|\lambda|$ is large enough, then
from (5.16),
$$\displaystyle\exp\left[L(G^{s}(\widetilde{a}),\omega^{2s-1}\lambda)-L(G^{-s+1}%
(\widetilde{a}),\omega^{-2s+1}\lambda)\right]$$
$$\displaystyle\times\exp\left[-L(G^{3s+3}(\widetilde{a}),\omega^{-2s-1}\lambda)%
-L(G^{s+2}(\widetilde{a}),\omega^{-2s}\lambda)+O\left(\lambda^{-\rho}\right)%
\right]=-1.$$
Then,
from Lemma 3.3,
$$\left(2n+1\right)\pi i=\sum_{j=0}^{m+1}e_{\ell,j}(a)\lambda_{n}^{\frac{1}{2}+%
\frac{1-j}{m}}+O\left(\lambda^{-\rho}\right),$$
where for $0\leq j\leq m+1$,
$$\displaystyle e_{\ell,j}(a)=$$
$$\displaystyle K_{m,j}(G^{s-\frac{1}{2}}(a))\omega^{(2s-1)\left(\frac{1}{2}+%
\frac{1-j}{m}\right)}-K_{m,j}((G^{-s+\frac{1}{2}}(a))\omega^{-(2s-1)\left(%
\frac{1}{2}+\frac{1-j}{m}\right)}$$
$$\displaystyle-K_{m,j}(G^{3s+\frac{5}{2}}(a))\omega^{-(2s+1)\left(\frac{1}{2}+%
\frac{1-j}{m}\right)}-K_{m,j}(G^{s+\frac{3}{2}}(a))\omega^{-2s\left(\frac{1}{2%
}+\frac{1-j}{m}\right)}$$
$$\displaystyle=$$
$$\displaystyle 2i\sum_{k=0}^{j}(-1)^{k}K_{m,j,k}b_{j,k}(a)\left[\sin\left(\frac%
{(1-j)(2s-1)\pi}{m}\right)+\sin\left(\frac{(1-j)(2s+1)\pi}{m}\right)\right]$$
$$\displaystyle=$$
$$\displaystyle 4i\sum_{k=0}^{j}(-1)^{k}K_{m,j,k}b_{j,k}(a)\sin\left(\frac{(1-j)%
\ell\pi}{m}\right)\cos\left(\frac{(1-j)\pi}{m}\right).$$
∎
Theorem 1.1 for $1\leq\ell\leq\frac{m}{2}$ has been proved. Next we prove Theorem 1.1 for
$\frac{m}{2}<\ell\leq m-1$, by the change of the variables $z\mapsto-z$.
8.3. When $\frac{m}{2}<\ell\leq m-1$
Proof of Theorem 1.1 when $\frac{m}{2}<\ell\leq m-1$.
Suppose that $\ell>\frac{m}{2}$. If $u$ is an eigenfunction of $H_{\ell}$, then $v(z)=u(-z,\lambda)$ solves
$$-v^{\prime\prime}(z)+\left[(-1)^{-\ell}(-iz)^{m}-P(-iz)\right]v(z)=\lambda v(z),$$
and
$$\text{$v(z)\rightarrow 0$ exponentially, as $z\rightarrow\infty$ along the two%
rays}\quad\arg z=-\frac{\pi}{2}\pm\frac{((m-\ell)+1)\pi}{m+2}.$$
The coefficient vector of $P(-z)$ is $((-1)^{m-1}a_{1},(-1)^{m-2}a_{2},\dots,(-1)^{1}a_{m-1},a_{m}).$ Certainly,
$$\sin\left(\frac{(1-j)(m-\ell)\pi}{m}\right)=(-1)^{j}\sin\left(\frac{(1-j)\ell%
\pi}{m}\right).$$
Also, one can find from Lemma 3.5 that for $0\leq k\leq j$,
$$b_{j,k}((-1)^{m-1}a_{1},(-1)^{m-2}a_{2},\dots,-a_{m-1},a_{m})=(-1)^{mk-j}b_{j,%
k}(a_{1},a_{2},\dots,a_{m-1},a_{m}).$$
Moreover, $c_{m-\ell,j}((-1)^{m-1}a_{1},(-1)^{m-2}a_{2},\dots,-a_{m-1},a_{m})=c_{\ell,j}(%
a_{1},a_{2},\dots,a_{m-1},a_{m}).$
This completes proof of Theorem 1.1.
∎
Acknowledgments
The author thanks Mark Ashbaugh, Fritz Gesztesy, Richard Laugesen, Boris Mityagin, Grigori Rozenblioum, and Alexander Turbiner for helpful discussions and references.
References
[1]
A. Avila.
Convergence of an exact quantization scheme.
Comm. Math. Phys., 249 (2004), 305–318.
[2]
C. M. Bender and P. D. Mannheim.
$\mathcal{PT}$ symmetry and necessary and sufficient conditions for the reality of energy eigenvalues,
Phys. Lett. A, 374 (2010), 1616–1620.
[3]
C. M. Bender and S. Boettcher.
Real spectra in non-Hermitian Hamiltonians having $\mathcal{PT}$-symmetry.
Phys. Rev. Lett., 80 (1998), 5243–5246.
[4]
C. M. Bender and A. Turbiner.
Analytic continuation of eigenvalue problems.
Phys. Lett. A, 173, (1993), 442–446.
[5]
E. Caliceti, F. Cannata, and S. Graffi.
$\mathcal{PT}$ symmetric Schrödinger operators: reality of the perturbed eigenvalues.
SIGMA, 6 (2010), 009, 8 pp.
[6]
E. Caliceti, S. Graffi, and M. Maioli.
Perturbation theory of odd anharmonic oscillators.
Comm. Math. Phys., 75 (1980), 51–66.
[7]
J. B. Conway.
Functions of one complex variable.
Springer-Verlag, New York, 1978.
[8]
E. Delabaere and D. T. Trinh.
Spectral analysis of the complex cubic oscillator.
J. Phys. A: Math. Gen., 33 (2000), 8771–8796.
[9]
P. Dorey, C. Dunning, A. Lishman, and R. Tateo.
$\mathcal{PT}$-symmetry breaking and exceptional points for a class of inhomogeneous complex potentials.
J. Phys. A: Math. Theor., 42 (2009), 465302, 41pp.
[10]
P. Dorey, C. Dunning, and R. Tateo.
Spectral equivalences, Bethe ansatz equations, and reality properties in $\mathcal{PT}$-symmetric quantum mechanics.
J. Phys. A: Math. Gen, 34 (2001), 5679–5704.
[11]
A. Eremenko and A. Gabrielov.
Analytic continuation of eigenvalues of a quartic oscillator.
Comm. Math. Phys., 287 (2009), no. 2, 431–457.
[12]
A. Eremenko, A. Gabrielov, and B. Shapiro.
High energy eigenfunctions of one-dimensional Schrödinger operators with polynomial potentials.
Comput. Methods Funct. Theory, 8 (2008), no. 1-2, 513–529.
[13]
M. V. Fedoryuk.
Asymptotic Analysis.
Springer-Verlag, New York, 1993.
[14]
C. R. Handy.
Generating converging bounds to the (complex) discrete states of the $P^{2}+iX^{3}+i\alpha X$ Hamiltonian.
J. Phys. A: Math. Gen., 34 (2001), 5065–5081.
[15]
B. Helffer and D. Robert.
Asymptotique des niveaux d’énergie pour des hamiltoniens a un degre de liberté.
Duke Math. J., 49 (1982), 853–868.
[16]
E. Hille.
Lectures on Ordinary Differential Equations.
Addison-Wesley, Reading, Massachusetts, 1969.
[17]
T. Kimura.
On the iteration of analytic functions.
Funkcial. Ekvac., 14 (1971), 197–238.
[18]
V. P. Maslov.
Thórie des perturbations et méthodes asymptotiques.
Dunod, Paris, 1972.
[19]
J. B. McLeod and E. C. Titchmarsh.
On the asymptotic distribution of eigenvalues.
Quart. J. Math. Oxford (2), 10 (1959), 313–320.
[20]
A. Mostafazadeh.
Pseudo-Hermiticity versus PT Symmetry: The necessary condition for the reality of the spectrum of a non-Hermitian Hamiltonian.
J. Math. Phys., 43 (2002), 205–214.
[21]
A. Mostafazadeh.
Pseudo-Hermiticity and generalized PT- and CPT-symmetries.
J. Math. Phys., 44 (2003), 974–989.
[22]
K. C. Shin.
On the reality of the eigenvalues for a class of $\mathcal{PT}$-symmetric oscillators,
Comm. Math. Phys., 229 (2002), 543–564.
[23]
K. C. Shin.
The potential $(iz)^{m}$ generates real eigenvalues only,
under symmetric rapid decay conditions,
J. Math. Phys., 46 (2005), 082110, 17pp.
[24]
K. C. Shin
Eigenvalues of $\mathcal{PT}$-symmetric oscillators with polynomial potentials,
J. Phys. A: Math. Gen., 38 (2005), 6147-6166.
[25]
K. C. Shin.
Asymptotics of eigenvalues of non-self adjoint Schrödinger operators on a
half-line,
Comput. Methods and Func. Theor., 10 (2010), 111 – 133.
[26]
K. C. Shin.
Anharmonic Oscillators with Infinitely Many Real
Eigenvalues and $\mathcal{PT}$-Symmetry,
SIGMA, 6 (2010), 015, 9 pp.
[27]
Y. Sibuya.
Global theory of a second order linear ordinary differential equation with a polynomial coefficient.
North-Holland Publishing Company, Amsterdam-Oxford, 1975.
[28]
M. Znojil.
$\mathcal{PT}$-symmetric harmonic oscillators.
Phys. Lett. A, 259 (1999), 220–223.
[29]
M. Znojil.
Complete set of inner products for a discrete $\mathcal{PT}$-symmetric square-well Hamiltonian.
J. Math. Phys., 50 (2009), 122105, 19 pp.
email contact: [email protected] |
Quasiparticle velocities in 2D electron/hole liquids with spin-orbit coupling
D. Aasen
Stefano Chesi
W. A. Coish
Department of Physics, McGill University, Montréal, Québec H3A 2T8, Canada
(December 5, 2020)
Abstract
We study the influence of spin-orbit interactions on quasiparticle dispersions in two-dimensional electron and heavy-hole liquids in III-V semiconductors. To obtain closed-form analytical results, we restrict ourselves to spin-orbit interactions with isotropic spectrum and work within the screened Hartree-Fock approximation, valid in the high-density limit. For electrons having a linear-in-momentum Rashba (or, equivalently, Dresselhaus) spin-orbit interaction, we show that the screened Hartree-Fock approximation recovers known results based on the random-phase approximation and we extend those results to higher order in the spin-orbit coupling. While the well-studied case of electrons leads only to a weak modification of quasiparticle properties in the presence of the linear-in-momentum spin-orbit interaction, we find two important distinctions for hole systems (with a leading nonlinear-in-momentum spin-orbit interaction). First, the group velocities associated with the two hole-spin branches acquire a significant difference in the presence of spin-orbit interactions, allowing for the creation of spin-polarized wavepackets in zero magnetic field. Second, we find that the interplay of Coulomb and spin-orbit interactions is significantly more important for holes than for electrons and can be probed through the quasiparticle group velocities. These effects should be directly observable in magnetotransport, Raman scattering, and femtosecond-resolved Faraday rotation measurements. Our results are in agreement with a general argument on the velocities, which we formulate for an arbitrary choice of the spin-orbit coupling.
pacs: 71.10.-w, 71.70.Ej, 71.45.Gm, 73.61.Ey
I Introduction
Semiconductor heterostructures offer the possibility of forming two-dimensional liquids with tunable sheet density $n_{s}$. In an idealized model where the carriers have parabolic dispersion with band mass $m$ and interact with Coulomb forces,TheBook the only relevant quantity is the dimensionless Wigner-Seitz radius $r_{s}=1/\sqrt{\pi a_{B}^{2}n_{s}}$, where $a_{B}=\hbar^{2}\epsilon_{r}/me^{2}$ is the effective Bohr radius ($\epsilon_{r}$ is the dielectric constant). Since $r_{s}$ serves as the interaction parameter, changing $n_{s}$ allows for a systematic study of the effects of the Coulomb interaction. In particular, properties of quasiparticle excitations such as their dispersion and lifetime are significantly modified due to electron-electron interactions.TheBook
Great attention has been paid in recent years to band-structure effects involving the spin degree of freedom.WinklerBook The strength and form of spin-orbit interaction (SOI) can be controlled in two-dimensional liquids through the choice of materials, the type of carriers (electrons/holes), and details of the confinement potential. For example, it is possible to change the coupling constant with external gates.Datta1989 ; Nitta1997 ; Johnson2009 In addition to detailed studies of single-particle properties, the problem of understanding the effects of SOI in the presence of Coulomb interactions is a topic of ongoing investigations.
Quasiparticle properties in the presence of SOIChen1999 ; Saraga2005 ; Nechaev2009 ; Nechaev2010 ; Agarwal2011 have been examined primarily accounting for RashbaBychov1984a ; Bychov1984b and/or Dresselhaus SOI,Dresselhaus1955 ; Dyakonov1986 which are dominant in electronic systems. The SOI results in two distinct spin subbands, with two associated Fermi surfaces. The effects on quasiparticles are usually very small; at each of the two Fermi surfaces the quasiparticle dispersionChen1999 ; Saraga2005 ; Agarwal2011 and lifetimeSaraga2005 ; Nechaev2009 ; Nechaev2010 are almost unaffected by SOI, except in the case of very large SOI coupling.Nechaev2010 ; Berman2010 In fact, it was found with Rashba SOI that the corrections to these quantities linear in the SOI coupling are absent.Saraga2005 Although explicit calculations are performed within perturbative approximation schemes, notably the random-phase approximation (RPA),Saraga2005 ; Nechaev2009 ; Nechaev2010 the SOI leading-order cancellation is valid non-perturbatively (to all orders in $r_{s}$).ChesiExact Similar arguments hold for other physical quantities.ChesiExact ; Abedinpour2010 For example, values of the ground-state energy obtained with Monte Carlo simulationsAmbrosetti2009 for up to $r_{s}=20$ could be reproduced with excellent accuracy by simply neglecting SOI corrections to the exchange-correlation energy.ChesiHighDens Noticeable exceptions exhibiting larger SOI effects are spin-textured broken symmetry phases,ChesiHFstates ; Juri2008 non-analytic corrections to the spin susceptibility,Zak2010 ; Zak2011 and the plasmon dispersion.Agarwal2011 All these examples involve the presence of spin polarization (either directlyChesiHFstates ; Juri2008 ; Zak2010 or indirectlyAgarwal2011 ), in which case the arguments of Ref. ChesiExact, do not apply.
Another interesting situation occurs when the SOI has a nonlinear dependence on momentum, thus cannot be written as a spin-dependent gauge potential.Aleiner2001 ; Coish2006 ; Abedinpour2010 Then, the approximate cancellations mentioned above are not expected. Winkler has shown that the dominant SOI induced by heterostructure asymmetry is cubic in momentum for heavy holes in III-V semiconductors.Winkler2000 This theoretical analysis was later shown to be in good agreement with magnetoresistance experiments.Winkler2002 Recently, the relevance of this cubic-in-momentum model was supported by the anomalous sign and magnetic-field dependence of spin polarization in quantum point contacts.Rokhinson2004 ; ChesiQPC SOI quadratic in momentum can also be induced for heavy holes by an in-plane magnetic field.Bulaev2007 ; Chesi07April
That these band-structure effects can substantially modify standard many-body results is confirmed by recent Shubnikov-de Haas oscillation measurements in low-density hole systems.Winkler2005 ; Chiu2011 For example, a surprisingly small Coulomb enhancement of the $g$-factor has been reported Winkler2005 and puzzling results have also been obtained for the effective masses $m_{\pm}$ of the two hole-spin sub-bands ($\sigma=\pm$).Chiu2011 A mechanism for the small $g$-factor enhancement was suggested in Ref. Chesi07April, : if SOI strongly distorts the ground state spin structure, the exchange energy becomes ineffective in promoting full polarization of the hole system.
In this paper, we focus on the effective masses $m_{\pm}$. We note that $m_{\pm}$ are directly related to the quasiparticle group velocities $v_{\pm}$ at the Fermi surfaces. With this in mind, we find it more transparent to discuss the effects of SOI and electron-electron interactions on $m_{\pm}$ in terms of wavepacket motion, as illustrated in Fig. 1. If an unpolarized wavepacket is injected at the Fermi energy (with average momentum along a given direction), the subsequent motion is very different depending on whether the SOI is linear ($n=1$) or non-linear ($n=2,3$). In the former case ($n=1$), corresponding to electrons, we have $v_{+}\simeq v_{-}$ and the motion is essentially equivalent to the case without SOI. In contrast, for holes with strong SOI, $v_{+}\neq v_{-}$ and the two spin components become spatially separated. The separation between the two spin components can become quite sizable, and it should be possible to observe such an effect with, e.g., Faraday-rotation imaging techniques.Crooker1996 ; Kikkawa1997 Electron-electron interactions, in addition to modifying the average velocity $v=(v_{+}+v_{-})/2$ (an effect which is well-known without SOITheBook ; Tan2005 ; Drummond2009 ), are also reflected on the velocity difference $(v_{+}-v_{-})$ between the two spin branches.
With these motivations in mind, we pursue a study of the quasiparticle group velocity in the presence of linear Rashba and non-linear SOI. An accurate analytical treatment of the electron-electron interactions can be carried out at high density, and we restrict ourselves to this interesting limit. The regime of strong electron-electron interactions is much more difficult to treat (see Ref. Drummond2009, for a Monte Carlo study without SOI) but it represents a relevant topic for future investigations ($r_{s}\sim 6-12$ in Refs. Winkler2005, ; Chiu2011, ).
This paper is organized as follows: In Sec. II we introduce a model Hamiltonian including a generalized SOIChesi07April and we briefly review its non-interacting properties, demonstrating that injected holes will separate into spin-polarized wavepackets. The Coulomb interaction is treated in Sec. III by extending the classic treatment of Ref. Janak1969, . We describe several results of this screened Hartree-Fock approach in detail, focusing on the quasiparticle group velocities and the interplay of spin-orbit and Coulomb interaction effects. A discussion of these results is given in Sec. IV. In Appendix A, we give a general argument showing that corrections to the velocity from any linear-in-momentum SOI can always be neglected to lowest order. Finally, a number of technical details are provided in Appendices B and C.
II Noninteracting problem
In the high-density limit, $r_{s}$ is small and the system is well-described by a non-interacting single-particle Hamiltonian. We consider here the following model,Chesi07April including a generalized SOI with a linear-, quadratic-, and cubic-in-momentum dependence for $n=1,2,3$:
$$H_{0}=\frac{p^{2}}{2m}+i\gamma\frac{p_{-}^{n}\sigma_{+}-p_{+}^{n}\sigma_{-}}{2},$$
(1)
where ${\bf p}$ is the momentum operator, $m$ the band mass, $p_{\pm}=p_{x}\pm ip_{y}$, $\sigma_{\pm}=\sigma_{x}\pm i\sigma_{y}$, and $\gamma$ the generalized spin-orbit coupling, with $\boldsymbol{\sigma}$ the vector of Pauli matrices. The physical justification of this model has been given in Ref. Chesi07April, : the $n=1$ Hamiltonian contains the Rashba SOI present in electronic systems,Bychov1984a ; Bychov1984b while $n=3$ includes the analogous term generated by an asymmetric confinement potential for holes.Winkler2000 ; Winkler2002 ; ChesiQPC Finally, the $n=2$ case is also relevant for holes, in the presence of an in-plane magnetic field.Chesi07April ; Bulaev2007 While SOI terms of different form generally coexist (see Appendix A), we assume here that one $n$ value is dominant. This greatly simplifies the problem by preserving the isotropy of the electron liquid in the x-y plane and is a good approximation for several relevant situations. For example, it was found for holesWinkler2000 ; Winkler2002 ; ChesiQPC that the $n=3$ term can be much larger than corrections to the SOI due to bulk inversion asymmetry.WinklerBook ; Rashba1988 Diagonalizing $H_{0}$ yields the energy spectrum
$$E_{\sigma}^{0}(k)=\frac{\hbar^{2}k^{2}}{2m}+\sigma\gamma\hbar^{n}k^{n},$$
(2)
with $\sigma=\pm$ labeling the two chiral spin branches. The corresponding eigenfunctions areChesi07April
$$\varphi_{\mathbf{k}\sigma}(\mathbf{r})=\frac{e^{i\mathbf{k}\cdot\mathbf{r}}}{%
\sqrt{2L^{2}}}\begin{pmatrix}1\\
i\sigma e^{in\theta_{\mathbf{k}}}\end{pmatrix},$$
(3)
where $\mathbf{k}$ is a wavevector in the x-y plane, $\theta_{\bf k}$ is the angle $\mathbf{k}$ makes with the x-axis, and $L$ is the linear size of the system.
It is useful at this point to introduce a dimensionless quantity $g$ characterizing the strength of the spin-orbit coupling:Chesi07April
$$g=\frac{\gamma\hbar^{n}k_{F}^{n}}{E_{F}},$$
(4)
where $E_{F}=\hbar^{2}k_{F}^{2}/2m$ is the Fermi energy without SOI, written in terms of the Fermi wavevector $k_{F}=\sqrt{2\pi n_{s}}=\sqrt{2}/(a_{\mathrm{B}}r_{s})$. While $\gamma$ has different physical dimensions for each value, $n=1,2,3$, the dimensionless coupling $g$ always gives the ratio of the spin-orbit energy to the kinetic energy. The coupling $g$ thus plays a role analogous to that of $r_{s}$ for the Coulomb interaction. Taking $\gamma$ to be independent of the density (this is not always the caseWinklerBook ), then from Eq. (4) we have $g\propto k_{F}^{n-2}\propto r_{s}^{2-n}$ (since $E_{F}\propto k_{F}^{2}$). This suggests that in the high-density limit ($r_{s}\to 0$), the effects of SOI are suppressed for electrons ($n=1$), but remain constant ($n=2$) or are enhanced ($n=3$) for holes. This simple estimate already indicates a qualitative difference for holes relative to electrons. We will see that this difference is indeed significant in the following sections.
In the presence of SOI the two spin bands $(\sigma=\pm)$ have different densities $n_{\pm}$, giving the total 2D sheet density $n_{s}=n_{+}+n_{-}$. Keeping $n_{s}$ (thus $k_{F}$) fixed gives a constraint on the Fermi wavevectors $k_{\pm}$ for the $\sigma=\pm$ bands, $k_{+}^{2}+k_{-}^{2}=2k_{F}^{2}$. We can then characterize the solution to this equation with a single parameter $\chi$:
$$k_{\pm}=k_{F}\sqrt{1\mp\chi}.$$
(5)
The parameter $\chi=(n_{-}-n_{+})/n_{s}$ gives the chirality and is determined by both the SOI and electron-electron interaction.comment_gen_chi We will assume for definiteness that $\gamma\geq 0$ such that $\chi\geq 0$ and $k_{+}\leq k_{-}$. For a fixed generalized spin-orbit coupling $\gamma$, the non-interacting value of $\chi$ can be determined by setting the Fermi energies of the two bands equal, i.e., $E^{0}_{+}(k_{+})=E^{0}_{-}(k_{-})$. This equation immediately gives a relationship between $g$ and $\chi$,
$$g=\frac{2\chi}{(1+\chi)^{n/2}+(1-\chi)^{n/2}}.$$
(6)
We denote the solution of Eq. (6) by $\chi_{0}(g)$, which gives the non-interacting Fermi wavevectors $k_{\pm}^{0}=k_{F}\sqrt{1\mp\chi_{0}}$. Explicit expressions for $\chi_{0}(g)$ are given in Ref. Chesi07April, . We only cite here the small-$g$ behavior, which is easily found from Eq. (6):
$$\chi_{0}(g)\simeq g.$$
(7)
We are mainly interested here in the properties of the quasiparticles and, in particular, their group velocities $v_{\pm}$. The fact that the dispersion relation (2) is a function only of the magnitude $k$ is a consequence of the model being isotropic with respect to $\mathbf{k}$, which allows us to discuss the magnitude of the group velocity on the Fermi surfaces,
$$v_{\pm}^{0}=\frac{1}{\hbar}\frac{\partial E_{\pm}^{0}(k)}{\partial k}\Big{|}_{%
k=k^{0}_{\pm}}=\frac{\hbar k^{0}_{\pm}}{m}\pm n\gamma(\hbar k^{0}_{\pm})^{n-1}.$$
(8)
The above expression can be evaluated explicitly in terms of the non-interacting chirality $\chi_{0}(g)$:
$$\displaystyle\frac{v_{\pm}^{0}}{v_{F}}=$$
$$\displaystyle\sqrt{1\mp\chi_{0}(g)}\pm\frac{n}{2}g[1\mp\chi_{0}(g)]^{(n-1)/2}$$
(9)
$$\displaystyle\simeq 1\pm\frac{g}{2}(n-1),$$
where $v_{F}=\hbar k_{F}/m$ is the Fermi velocity in the absence of SOI and in the second line we have expanded the result to lowest order in $g$, by making use of Eq. (7). Equation (9) shows that there is no relative difference in group velocity for the $\sigma=\pm$ bands when $n=1$. In contrast, when $n=2,3$ a sizable correction linear in $g$ is present. This difference reflects itself on the evolution of an initially unpolarized wavepacket injected at the Fermi surface, schematically illustrated in Fig. 1. While for electrons the wavepacket remains unpolarized, for holes the two spin components spatially separate with time. Electron-electron interactions modify the non-interacting velocities $v^{0}_{\pm}$, but the qualitative difference introduced by SOI between electrons ($n=1$) and holes ($n=2,3$) remains essentially unchanged. A similar behavior holds for other properties of the electron liquid as well:ChesiExact vanishing corrections to lowest order in $g$ were found for the quasiparticle lifetime,Saraga2005 ; Nechaev2010 the occupation,Chesi07April and the exchange-correlation energy,ChesiHighDens if only Rashba SOI ($n=1$) is included.
Finally, another relevant quasiparticle observable is the effective mass. This has recently been measured in hole systems through experiments on quantum oscillations.Chiu2011 This physical quantity is simply given by $m_{\pm}=\hbar k_{\pm}/v_{\pm}$ and is thus essentially equivalent to $v_{\pm}$.
III Screened Hartree-Fock approximation
Realistically, charged particles interact through the Coulomb potential so it is interesting to understand how the presence of SOI modifies the behavior of the quasiparticles. The fully interacting Hamiltonian is given by:
$$H=\sum_{i}H_{0}^{(i)}+\frac{1}{2}\sum_{i\neq j}\frac{e^{2}}{\epsilon_{r}|{\bf r%
}_{i}-{\bf r}_{j}|},$$
(10)
where $H_{0}^{(i)}$ (for electron $i$) is as in Eq. (1) and the presence in (10) of a uniform neutralizing background is understood. Although many sophisticated techniques exist to approach this problem,TheBook ; Drummond2009 the simplest approximation to the quasiparticle self-energy is obtained by only including the exchange contribution:
$$E_{\sigma}(k)=E_{\sigma}^{0}(k)+\Sigma^{x}_{\sigma}(k),$$
(11)
where
$$\Sigma^{x}_{\sigma}(k)=-\sum_{\mathbf{k}^{\prime}\sigma^{\prime}}\frac{(1+%
\sigma\sigma^{\prime}\cos{n\theta^{\prime}})}{2L^{2}}n_{\mathbf{k}^{\prime}%
\sigma^{\prime}}V(|\mathbf{k}-\mathbf{k}^{\prime}|).$$
(12)
Here, $n_{\mathbf{k^{\prime}}\pm}=\Theta(k_{\pm}-k^{\prime})$ is the occupation at $T=0$ for the $\sigma=\pm$ band, respectively, with $\Theta(x)$ the Heaviside step function. The first factor in the summation of Eq. (12), involving the angle $\theta^{\prime}$ between $\mathbf{k^{\prime}}$ and $\mathbf{k}$, arises from the scalar product of the non-interacting spinors [Eq. (3)], and takes into account the specific nature of the spin-orbit interaction ($n=1,2,3$).
To lowest order, $V(q)$ is the Fourier transform of the bare Coulomb potential, $2\pi e^{2}/(\epsilon_{r}q)$. As is well known,TheBook this form of the Coulomb interaction leads to an unphysical divergence in the quasiparticle velocity. By considering an infinite resummation in perturbation theory, the screening of the Coulomb interaction removes the divergence, e.g., in the RPA approximation. Finally, by approximating the dielectric function in the effective interaction by its zero-frequency long-wavelength limit, the RPA self-energy gives Eq. (12) with
$$V(q)=\frac{2\pi e^{2}/\epsilon_{r}}{q+\sqrt{2}r_{s}k_{F}}.$$
(13)
This screened Hartree-Fock approximation with SOI,Chen1999 ; Agarwal2011 notwithstanding its simplicity, becomes accurate in the high-density limit.
III.1 Renormalized occupation
By including the SOI, we can verify that Eq. (12) gives the correct high-density behavior for the Fermi wavevectors $k_{\pm}$. These are modified by electron-electron interactions from their non-interacting values $k_{\pm}^{0}$.Schliemann2006 ; Chesi07April From Eq. (12), $k_{\pm}$ can be obtained by equating the chemical potentials in the two spin branches:
$$E_{+}(k_{+})=E_{-}(k_{-}).$$
(14)
After taking the continuum limit, this equation is rewritten in dimensionless form as follows:
$$\displaystyle(y_{+}^{n}+y_{-}^{n})g=$$
$$\displaystyle 2\chi+\frac{r_{s}}{\sqrt{2}}\sum_{\sigma\sigma^{\prime}}\int_{0}%
^{2\pi}\!\frac{d\theta}{2\pi}\!\int_{0}^{y_{\sigma^{\prime}}}\!dy$$
(15)
$$\displaystyle\times\frac{y(\sigma+\sigma^{\prime}\cos{n\theta})}{\sqrt{y^{2}+y%
_{\sigma}^{2}-2yy_{\sigma}\cos{\theta}}+\sqrt{2}r_{s}},$$
where we have rescaled the wavevectors $k=k_{F}y$ and defined $y_{\pm}=\sqrt{1\mp\chi}$. The integral on the right-hand side is the correction from the exchange term and we have verified that Eq. (15) gives Eq. (6) for $r_{s}=0$. We note that, for a given value of the SOI, $\chi(r_{s},g)$ enters in a rather complicated way in Eq. (15), being involved in the integration limits of the exchange term as well as the integrand. In practice, instead of solving Eq. (15) for $\chi$, it is convenient to evaluate $g$ for a given value of $\chi$ and numerically invert the function $g(r_{s},\chi)$.
The values of $k_{\pm}=k_{F}y_{\pm}$ were obtained in Ref. Chesi07April, through a different procedure, i.e., by minimizing the total energy (including the exchange contribution) of non-interacting states.comment_variational Although both methods are unreliable at $r_{s}>1$, they both become accurate in the high-density limit, $r_{s}<1$. In fact, the only difference in the two approaches is due to the presence of the Thomas-Fermi screening wavevector in Eq. (15) and, by neglecting $\sqrt{2}r_{s}$ in the denominator of the second line, the equation from the variational treatment is recovered. In particular, expanding Eq. (15) at small $r_{s}$ and $g$ gives the same result found in Ref. Chesi07April, :
$$\chi(r_{s},g)\simeq g\left(1-\frac{\sqrt{2}r_{s}}{\pi}\sum_{j=0}^{n}\frac{1}{2%
j-1}\right).$$
(16)
Details of the derivation of Eq. (16) are given in Appendix B. Two salient features of Eq. (16) are:Chesi07April (i) For $n=1$ there is no correction to the noninteracting result $\chi(r_{s},g)\simeq g$. On the other hand, the linear dependence on $g$ is actually modified by electron-electron interactions at $n=2,3$. (ii) The effect of the electron-electron interactions is a reduction of $\chi(r_{s},g)$ from the non-interacting value. This result could be rather surprising, having in mind the well-known enhancement of spin polarization caused by the exchange energyTheBook (when the spin-splitting is generated by a magnetic field). However, $\chi$ does not correspond here to a real spin-polarization, which is zero. Instead, $\chi$ is simply related to the population difference of the two chiral spin subbands.
III.2 Quasiparticle velocity
In the screened Hartree-Fock approximation, the group velocities at the Fermi surfaces are given by:
$$\displaystyle v_{\pm}=$$
$$\displaystyle\,\frac{1}{\hbar}\frac{\partial E_{k\pm}}{\partial k}\Big{|}_{k=k%
_{\pm}}$$
$$\displaystyle=$$
$$\displaystyle\,\frac{\hbar k_{\pm}}{m}\pm n\gamma(\hbar k_{\pm})^{n-1}$$
(17)
$$\displaystyle-\sum_{\mathbf{k}^{\prime}\sigma^{\prime}}\frac{(1\pm\sigma^{%
\prime}\cos{n\theta^{\prime}})}{2L^{2}}n_{\mathbf{k}^{\prime}\sigma^{\prime}}%
\left[\frac{\partial}{\partial k}V(|\mathbf{k}\!-\!\mathbf{k^{\prime}}|)\right%
]_{k=k_{\pm}}.$$
In general, we can discuss all corrections to $v_{\pm}$ by introducing the following notation:
$$\frac{v_{\pm}}{v_{F}}=1+\delta v(r_{s})+\delta v_{\pm}^{0}(g)+\delta v_{\pm}(r%
_{s},g),$$
(18)
where $\delta v(r_{s})$ is the (spin-independent) correction due to electron-electron interactions at $g=0$, which has been the subject of many theoretical and experimental studies (see, e.g., Ref. TheBook, , Chiu2011, ; Janak1969, ; Tan2005, ; Drummond2009, , and references therein). In the approximation (III.2), it is given byJanak1969
$$\delta v(r_{s})=-\frac{\sqrt{2}r_{s}}{\pi}+\frac{r_{s}^{2}}{2}+\frac{r_{s}(1-r%
_{s}^{2})}{\sqrt{2}\pi}\frac{\cosh^{-1}(\sqrt{2}/r_{s})}{\sqrt{1-r_{s}^{2}/2}}.$$
(19)
As is known,TheBook this approximation gives the correct leading behavior at small $r_{s}$: $\delta v\simeq-(r_{s}\ln r_{s})/(\sqrt{2}\pi)$. The second nontrivial term in Eq. (18) is the non-interacting correction purely due to SOI
$$\delta v^{0}_{\pm}(g)=\frac{v^{0}_{\pm}(g)}{v_{F}}-1,$$
(20)
which only depends on $g$ [see Eq. (9)]. Finally, $\delta v_{\pm}(r_{s},g)$ collects all remaining corrections.
A pictorial representation of the physical meaning of the three terms is shown in Fig. 1 for an unpolarized wavepacket injected at the Fermi energy. At high density, as seen in Eq. (19) and illustrated in Fig. 1(a), the group velocity is larger than without electron-electron interactions. In Fig. 1(b) we depict the generic situation with SOI. In the presence of SOI, the two spin branches have different group velocities. An initially unpolarized wavepacket then splits into its two spin components. Both the SOI and electron-electron interactions influence the relative velocity $(v_{+}-v_{-})$. The separation after a time $\tau$ is $\delta=|v^{0}_{+}-v_{-}^{0}|\tau$ for the non-interacting case and is modified by $\delta v_{\pm}$ with electron-electron interactions: $\Delta=|v_{+}-v_{-}|\tau$ [see Fig. 1(b)].
The presence of ‘interference’ terms, $\delta v_{\pm}(r_{s},g)$, becomes clear from Eq. (III.2). These terms are due to the interplay of many-body interactions with SOI. A first contribution to $\delta v_{\pm}(r_{s},g)$, which we refer to as the ‘self-energy contribution’, comes directly from the exchange integral [third line of Eq. (III.2)]: due to the presence of two distinct Fermi wavevectors $k_{\pm}$, the result obviously contains corrections to Eq. (19) which depend on $g$ (in addition to $r_{s}$). A second contribution to $\delta v_{\pm}(r_{s},g)$ comes indirectly from the non-interacting part and we refer to it as the ‘repopulation contribution’. Since the Fermi wavevectors $k_{\pm}$ are modified from the non-interacting values $k_{\pm}^{0}$, the second line of Eq. (III.2) gives a result distinct from $v^{0}_{\pm}$. The sign of this repopulation contribution is easily found by noting that, as discussed in Sec. III.1, the exchange energy reduces the value of $\chi$ (for $n=2,3$). This corresponds to an increase (decrease) of $k_{+}$ ($k_{-}$), i.e., a positive (negative) correction to $v_{+}$ ($v_{-}$). The effect would be to enhance the difference in velocity between the two branches, as illustrated in Fig. 1(b) ($\Delta>\delta$). However, to establish the ultimate form and sign of $\delta v_{\pm}(r_{s},g)$ requires a detailed calculation of both self-energy and repopulation contributions, which is presented below for some interesting cases. The total spin-dependent part of the velocity, $\delta v_{\pm}^{0}+\delta v_{\pm}$, can then be compared to the simple non-interacting effect, $\delta v_{\pm}^{0}$.
III.3 $n=1$: higher-order corrections in $g$
We begin the analysis of Eq. (III.2) by rewriting it in a more explicit way. To this end, we use
$$\frac{\partial}{\partial k}V(|\mathbf{k}-\!\mathbf{k}^{\prime}|)=-\hat{\mathbf%
{k}}\cdot\frac{\partial}{\partial{\mathbf{k}^{\prime}}}V(|\mathbf{k}-\mathbf{k%
}^{\prime}|),$$
(21)
where $\hat{\bf k}={\bf k}/k$. This allows us to integrate Eq. (III.2) by parts, which leads to:
$$\displaystyle\frac{v_{\pm}}{v_{F}}=y_{\pm}\pm\frac{g}{2}ny_{\pm}^{n-1}$$
$$\displaystyle+r_{s}\sum_{\sigma}\int_{0}^{2\pi}\frac{d\theta^{\prime}}{8\pi}%
\frac{\sqrt{2}\cos{\theta^{\prime}}(1\pm\sigma\cos{n\theta^{\prime}})y_{\sigma%
}}{\sqrt{y_{\pm}^{2}+y_{\sigma}^{2}-2y_{\pm}y_{\sigma}\cos{\theta^{\prime}}}+r%
_{s}\sqrt{2}}$$
$$\displaystyle\pm r_{s}\int_{0}^{2\pi}\frac{d\theta^{\prime}}{8\pi}\int_{y_{+}}%
^{y_{-}}\frac{\sqrt{2}n\sin{n\theta^{\prime}}\sin{\theta^{\prime}}\,dy^{\prime%
}}{\sqrt{y_{\pm}^{2}+{y^{\prime}}^{2}-2y^{\prime}y_{\pm}\cos{\theta^{\prime}}}%
+r_{s}\sqrt{2}}$$
$$\displaystyle=L_{1}+L_{2}+L_{3}.$$
(22)
Notice that, in the integration by parts of Eq. (III.2) in the continuum limit, two types of terms enter: those corresponding to the second line of Eq. (III.3) ($L_{2}$) involve the derivative of $n_{\mathbf{k^{\prime}}\pm}$. This results in a delta function in the $dk^{\prime}$ integral, which can then be easily evaluated. Thus, only the integral in $d\theta^{\prime}$ is left. The second type of term involves the derivative of $\cos{n\theta^{\prime}}$:
$$\hat{\mathbf{k}}\cdot\frac{\partial}{\partial\mathbf{k^{\prime}}}\cos{n\theta^%
{\prime}}=\frac{n}{k^{\prime}}\sin{n\theta^{\prime}}\sin{\theta^{\prime}},$$
(23)
and indeed this angular factor appears in the third line of Eq. (III.3) ($L_{3}$).
Eq. (III.3) can always be evaluated numerically, after obtaining the values of $\chi$ (thus $y_{\pm}=\sqrt{1\mp\chi}$) from Eq. (15). By specializing to the small $g,r_{s}$ limit for $n=1$ SOI, it is known that the linear-in-$g$ correction to $v_{\pm}$ vanishes.Saraga2005 ; ChesiExact Thus, an expansion to second order in $g$ has to be performed, which has been done in Ref. Saraga2005, in the context of the RPA treatment of the quasiparticle properties. To verify the validity of the simpler screened Hartree-Fock procedure, it is interesting to perform the same expansion for our Eq. (III.3).
In fact, the two calculations bear some similarities since $L_{2}$ is the same as the boundary term $B_{\rm boundary}^{(u\to 0_{+})}$ of the RPA treatment, see Eq. (69) of Ref. Saraga2005, . Thus, we can borrow the expansion in small $g,r_{s}$:
$$L_{2}\simeq-\frac{r_{s}}{\sqrt{2}\pi}\left(\ln\frac{r_{s}}{2\sqrt{2}}+2\pm%
\frac{2}{3}\,g-\frac{g^{2}}{8}\ln g\right);\quad(n=1),$$
(24)
where terms of order $O(r_{s}^{2},r_{s}g^{2})$ have been omitted. To the same order of approximation, we have for $L_{1}$:
$$L_{1}\simeq 1-\frac{g^{2}}{8};\quad(n=1),$$
(25)
which can also be obtained by expanding Eq. (9) (with $n=1$). Since $\chi$ only receives $O(r_{s}g^{3}\ln g)$ corrections from electron-electron interactions,ChesiHighDens it is sufficient to use the non-interacting value, $\chi_{0}(g)$, to this order of approximation. Thus, the repopulation induced by electron-electron interactions has a negligible effect on $v_{\pm}$ in this case. The situation will be different for $n=2,3$. Expanding $L_{3}$ for small $r_{s},g$ yields
$$L_{3}\simeq\pm\frac{\sqrt{2}r_{s}}{3\pi}g;\quad(n=1),$$
(26)
which cancels the linear-in-$g$ term of Eq. (24), as expected for $n=1$.
We note that in the screened Hartree-Fock approximation we are able to obtain an analytic result for the leading term of $L_{3}$, and some details of the derivation can be found in Appendix C. In contrast, in the RPA treatment of Ref. Saraga2005, it was not possible to expand the more involved corresponding term, $B_{\rm int}$, in a fully analytic fashion. The cancellation of the linear-in-$g$ contribution was indicated by a general argumentSaraga2005 ; ChesiExact and confirmed through numerical study. Additionally, the absence of higher-order terms which modify Eq. (24) was inferred numerically in Ref. Saraga2005, . Although the final results of both approaches (RPA and screened Hartree-Fock) agree, the situation is clearly more satisfactory within the framework of Eq. (III.3), since cancellation of the linear term in Eq. (24) can be checked exactly and the expansion can be carried out to higher order systematically. The final result, computed to higher order in $g$, reads (for $n=1$):
$$\displaystyle\frac{v_{\pm}}{v_{F}}$$
$$\displaystyle\simeq 1+\delta v(r_{s})+\delta v^{0}_{\pm}(g)$$
(27)
$$\displaystyle+\frac{\sqrt{2}r_{s}}{16\pi}\left[g^{2}\left(\ln\frac{g}{8}+\frac%
{3}{2}\right)\pm\frac{g^{3}}{6}\left(\ln\frac{g^{7}}{8^{5}}+\frac{319}{20}%
\right)\right].$$
The second line of Eq. (27) represents the expansion of $\delta v_{\pm}(r_{s},g)$ including all terms up to $O(r_{s}g^{3}$). For $n=1$, the non-interacting result gives the same velocity for both spin branches [$\delta v^{0}_{\pm}(g)$ is actually independent of $\pm$ for this particular model]. The $O(r_{s}g^{3})$ term in Eq. (27) is therefore quite interesting. It shows that a small difference in velocity exists. This is a genuine effect of electron-electron interactions.
The accuracy of Eq. (27) can be seen in Fig. 2: it becomes accurate at very small values of $r_{s}$ (as shown in the inset) while at larger more realistic values of $r_{s}$ it still gives the correct magnitude of the effect. The numerical example of Fig. 2 also shows that $\delta v_{\pm}(r_{s},g)$ is very small. In fact, it is generally much smaller than the non-interacting correction: since $\delta v^{0}_{\pm}(g)\simeq-g^{2}/8$ [see Eq. (25)] the $~{}r_{s}g^{2}\ln g$ term becomes larger only if $r_{s}\ln g\gg 1$. This condition is only satisfied at extremely small values of $g$ (if $r_{s}$ is small as well) at which SOI effects are hardly of any relevance. Thus, Eq. (27) implies that the effects of SOI and electron-electron interactions are essentially decoupled for $n=1$. This picture changes substantially for hole systems ($n=2,3$), as we show in the next section.
III.4 Corrections to the velocity for $n=2,3$
We now apply the discussion of the previous section to the SOI more appropriate for holes and point out some important differences. Expansions for small $r_{s}$ and $g$ are given in Appendix C. From Eqs. (56) and (57) and using $\chi=g+O(r_{s}g,g^{3})$ [Eq. (16)], we find the following expressions for the self-energy contribution. For $n=2$:
$$L_{2}+L_{3}\simeq\,\delta v(r_{s})+\frac{\sqrt{2}r_{s}}{4\pi}\left[\pm\frac{g}%
{3}+g^{2}\left(\ln\frac{g}{8}+2\right)\right],$$
(28)
and for $n=3$:
$$L_{2}+L_{3}\simeq\,\delta v(r_{s})+\frac{\sqrt{2}r_{s}}{4\pi}\left[\pm\frac{8}%
{15}g+\frac{g^{2}}{4}\left(9\ln\frac{g}{8}+\frac{613}{30}\right)\right].$$
(29)
At variance with the case of $n=1$, the linear-in-$g$ term does not vanish here. Thus, we find an appreciable correction to the velocity. This correction has opposite sign in the two branches and is positive for the $+$ (higher-energy) branch.
A second contribution to the $g$-linear correction comes from $L_{1}$. Expanding $L_{1}$ in terms of $\chi$ and using Eq. (16) gives, for $n=2$,
$$L_{1}\simeq 1+\delta v^{0}_{\pm}(g)+\frac{\sqrt{2}r_{s}}{4\pi}\left(\pm\frac{2%
}{3}g+g^{2}\right);\quad(n=2),$$
(30)
and for $n=3$:
$$L_{1}\simeq 1+\delta v^{0}_{\pm}(g)+\frac{\sqrt{2}r_{s}}{4\pi}\left(\pm\frac{1%
6}{15}g+\frac{32}{15}g^{2}\right);\quad(n=3).$$
(31)
Thus, the repopulation contribution is present in this case and has the sign discussed at the end of Sec. III.2 (it is positive for the $+$ branch).
As it turns out, the self-energy, repopulation, and non-interacting contributions to the velocity have the same sign. The three contributions therefore have a cooperative effect in enhancing the difference in velocity between the two spin branches. Of course, based on the high-density theory presented here, we cannot tell if this conclusion holds at all densities. We also note that the $g$-linear term of the self-energy correction [Eq. (28) or (29)] is always half of the corresponding repopulation correction [Eq. (30) or (31)]. Again, we have not investigated if this curious relation only occurs within this approximation scheme or if it is more general.
Finally, we give the complete result for $n=2$
$$\displaystyle\frac{v_{\pm}}{v_{F}}$$
$$\displaystyle\simeq 1+\delta v(r_{s})+\delta v^{0}_{\pm}(g)$$
(32)
$$\displaystyle+\frac{\sqrt{2}r_{s}}{4\pi}\left[\pm g+g^{2}\left(\ln\frac{g}{8}+%
3\right)\right];\quad(n=2),$$
and for $n=3$
$$\displaystyle\frac{v_{\pm}}{v_{F}}$$
$$\displaystyle\simeq 1+\delta v(r_{s})+\delta v^{0}_{\pm}(g)$$
(33)
$$\displaystyle+\frac{\sqrt{2}r_{s}}{4\pi}\left[\pm\frac{8}{5}g+g^{2}\left(9\ln%
\frac{g}{8}+\frac{869}{120}\right)\right];\quad(n=3),$$
and show two numerical examples in Figs. 3 and 4.
Fig. 3 is a comparison of $\delta v_{\pm}(r_{s},g)$ for $n=1$ and $n=3$: it is clear that the dependence on $g$ is very weak for the electron case ($n=1$) and the magnitude of $\delta v_{\pm}(r_{s},g)$ is much larger for holes ($n=3$) with SOI of comparable strength. Fig. 4 shows the separation $\Delta$ between the two spin components of an initially unpolarized wave packet for a fixed travel distance of $1~{}\mu$m:
$$\Delta=2\,\frac{|v_{+}-v_{-}|}{v_{+}+v_{-}}\times(1~{}\mu{\rm m}).$$
(34)
In addition to including both $n=1$ and $n=3$, in Fig. 4 we plot the non-interacting value $\Delta_{0}$, obtained by substituting $v_{\pm}\to v^{0}_{\pm}$ in Eq. (34). For $n=1$ the non-interacting velocities $v_{\pm}^{0}$ are the same ($\Delta_{0}=0$). The effect of electron-electron interactions is not visible. Thus, the wavepacket remains essentially unsplit and unpolarized ($\Delta\simeq 0$) and the only significant influence is on the average velocity $(v_{+}+v_{-})/2$, from Eq. (19) [see Fig. 1(a)]. In contrast, a large splitting is found for $n=3$ where $\Delta,\Delta_{0}$ can reach a large fraction of the traveling distance. Since spin-polarized ballistic transport is observed in two-dimensional hole systems on $\mu$m-scales (e.g., in spin-focusing experimentsRokhinson2004 ; ChesiQPC ) and the typical wavepacket traveling time is $1-50$ ps (depending on the density), Fig. 4 suggests that a direct optical imaging of the wavepacket separation should be within reach of femtosecond-resolved Faraday rotation measurements.Crooker1996 ; Kikkawa1997
As for electron-electron interaction effects, we see in Fig. 4 that a visible difference between $\Delta$ and $\Delta_{0}$ exists. The difference is quite small, due to the fact that we are considering here the weak-coupling limit, and all the interaction corrections are proportional to $r_{s}<1$. $\Delta_{0}$ is modified here by $\sim 2-3\%$ and we collect some representative values in Table 1. It can be seen in Table 1 that, while $\Delta_{0}$ does not change with $r_{s}$, the interaction correction $\Delta-\Delta_{0}$ grows at lower densities (see Table 1). Since experiments on hole systems can reach values as large as $r_{s}=6-12$,Winkler2005 ; Chiu2011 it is reasonable to expect significant effects from $\delta v_{\pm}$ in this low-density regime. As a reference, in electron systems, $\delta v$ changes from $\sim 5\%$ to $-30\%$ for $r_{s}\sim 1$ to $6$.Tan2005 ; Drummond2009
It might be surprising to see $\Delta<\Delta_{0}$ in Fig. 4. This is due to the interplay of two competing interaction effects. It is true that, as discussed already, the difference between $v_{\pm}$ is enhanced by $\delta v_{\pm}$. The two spin components therefore split faster in the interacting case. However, the situation shown in Fig. 4 is distinct from that shown in Fig. 1 since a constant traveling distance (and not time $\tau$) is assumed. At high density the mean velocity $(v_{+}+v_{-})/2$ is greater for the interacting gas ($\delta v>0$), which allows the wavepackets less time to separate. As it turns out, the latter effect is dominant in Fig. 4.
Interestingly, the sign of $\delta v$ changes at low density.TheBook This would imply a cooperative effect of $\delta v_{\pm}$ and $\delta v$ on $\Delta$ if $\delta v_{\pm}$ does not change sign. Unfortunately, to infer the behavior of $\delta v_{\pm}$ at low density requires a much more sophisticated approach.
IV Conclusion
We have presented a theory of the quasiparticle group velocity at high density, in the presence of SOI of different types. Contrasting the behavior of electron and hole systems, we find several intriguing differences. We have shown explicitly that the lowest-order cancellations of SOI effects occur only for the electronic case, when the SOI is approximately linear in momentum (e.g., a strong Rashba or Dresslhaus SOI is present). On the other hand, SOI terms non-linear in ${\bf p}$ are often dominant in hole systems.Winkler2000 ; Winkler2002 ; ChesiQPC Thus, larger effects of the SOI and a non-trivial interplay with electron-electron interactions are expected for holes on general grounds.
As an important motivation for future theoretical studies, hole liquids can be realized in the laboratory with strong SOI and large values of $r_{s}$. For example, the spin-subband population difference at zero field is of order $15-20\%$ in Ref. Winkler2002, , with $n_{s}\simeq 2-4\times 10^{10}~{}{\rm cm}^{-2}$. With a hole effective mass $m\simeq 0.2m_{0}$ in GaAs these densities correspond to $r_{s}\simeq 9-12$. For electrons, materials with strong SOI typically have small effective masses, which results in much lower values of $r_{s}$. A diluted electron liquid with $n_{s}\simeq 2\times 10^{10}~{}{\rm cm}^{-2}$ gives $r_{s}\simeq 1.2$, using InAs parameters ($m=0.023m_{0}$).
Discussing the large-$r_{s}$ regime of holes would require extending many-body perturbation theorySaraga2005 ; Nechaev2009 ; Nechaev2010 or Monte CarloAmbrosetti2009 approaches, so far only applied to linear SOI. The high-density regime studied here would represent a well-controlled limit of these theories for the quasiparticle dispersion. In addition to being relevant for transport measurements of the effective mass, the significant difference in group velocity at the Fermi surface of the two spin branches could also be addressed by Raman scattering experiments, demonstrated for electron systems in Ref. Jusserand1992, , or via time-resolved Faraday-rotation detection of the spin-polarization.Crooker1996 ; Kikkawa1997 A similar discussion should hold with $n=2,3$ for other physical observables and many-body effects. For example, studying the compressibilityTheBook ; Eisenstein1992 in the presence of SOI and extending the $n=1$ discussion of the quasiparticle lifetimeSaraga2005 ; Nechaev2009 ; Nechaev2010 would also be topics of interest. Finally, the problem of including in our framework a more general form of SOI than Eq. (1) is clearly of practical relevance. However, as discussed in Appendix A, we expect that our qualitative picture on the different role of linear and non-linear SOI remains valid also in this more general situation.
Acknowledgements.We thank J. P. Eisenstein and D. L. Maslov for useful discussions and acknowledge financial support from CIFAR, NSERC, and FQRNT.
Appendix A General spin-orbit coupling
A general SOI contains terms with linear and non-linear dependence on momentum, and does not necessarily have the isotropic form assumed in Eq. (1). For definiteness, we suppose that there is no magnetic field, so that quadratic terms are not present:
$$H=\sum_{i}\frac{p_{i}^{2}}{2m}+H^{\rm so}_{1}+H^{\rm so}_{3}+H_{\rm el-el},$$
(35)
where $H^{\rm so}_{1}$, $H^{\rm so}_{3}$, and $H_{\rm el-el}$ give, respectively, the linear-in-momentum spin-orbit, the cubic-in-momentum spin-orbit, and electron-electron interactions. To show that $H^{\rm so}_{1}$ always has a small effect, we consider the unitary transformation $H^{\prime}=e^{-S}He^{S}$ with $S$ defined by $[S,\sum_{i}p_{i}^{2}/2m]=H^{\rm so}_{1}$, giving
$$H^{\prime}\simeq\frac{p^{2}}{2m}+H^{\rm so}_{3}+H_{\rm el-el}-[S,H^{\rm so}_{1%
}+H^{\rm so}_{3}].$$
(36)
In the same transformed frame, the velocity operator of electron $i$ is given by ${\bf v}^{\prime}_{i}=e^{-S}(\partial H/\partial{\bf p}_{i})e^{S}$, which yields
$${\bf v}^{\prime}_{i}\simeq\frac{{\bf p}_{i}}{m}+\frac{\partial{H^{\rm so}_{3}}%
}{\partial{\bf p}_{i}}-[S,\frac{\partial{H^{\rm so}_{1}}}{\partial{\bf p}_{i}}%
+\frac{\partial{H^{\rm so}_{3}}}{\partial{\bf p}_{i}}].$$
(37)
In deriving Eq. (36) we have used the fact that $[S,H_{\rm el-el}]=0$. For example, for an isotropic SOI as in Eq. (1),
$$S=i\frac{m\gamma}{\hbar}\sum_{i}(x_{i}\sigma_{y,i}-y_{i}\sigma_{x,i}).$$
(38)
The property $[S,H_{\rm el-el}]=0$ is valid for a general linear-in-momentum SOI, including a combination of Rashba and Dresselhaus SOI.Aleiner2001 However, the same identity $[S,H_{\rm el-el}]=0$ does not hold for a transformation with $[S,\sum_{i}p_{i}^{2}/2m]=H^{\rm so}_{3}$, i.e., a transformation that is designed to remove the non-linear component from the non-interacting Hamiltonian. In writing Eq. (37), we have used $[S,{\bf p}_{i}/m]=\partial{H^{\rm so}_{1}}/\partial{\bf p}_{i}$, which implies a cancellation of the $H^{\rm so}_{1}$ contribution to ${\bf v}^{\prime}_{i}$ to lowest order. Again, this cancellation is only valid for linear-in-momentum SOI.
By introducing dimensionless couplings $g_{1,3}$ associated with $H^{\rm so}_{1,3}$, in direct analogy with Eq. (4), we see that the commutators in Eqs. (36) and (37) are of quadratic or bilinear order in the couplings ($\sim g_{1}^{2}$ and $\sim g_{1}g_{3}$). This indicates on general grounds that $H^{\rm so}_{3}$ has the largest effect on the quasiparticle velocity if $g_{1}\lesssim g_{3}\ll 1$. In this case, if we are interested in lowest-order effects, we can neglect both anticommutators in Eqs. (36) and (37), which is equivalent to neglecting $H^{\rm so}_{1}$ in the original Hamiltonian (35). Thus, to leading order all results we report for the quasiparticle velocities due to a pure cubic-in-momentum spin-orbit interaction also apply in the case of a mixed linear-plus-cubic spin-orbit interaction (with the caveat that we consider only the isotropic form of cubic SOI).
Appendix B Derivation of Eq. (16)
As discussed in the text, at high density we can neglect the $\sqrt{2}r_{s}$ in the integrand of Eq. (15) (second line). For small $g$, the value of $\chi$ is also small and we can perform an expansion of the exchange integral. First notice that the constant term at $\chi=0$ is missing, because the integration limits simply become $y_{\pm}=1$ and the integrand vanishes upon the summation on $\sigma,\sigma^{\prime}$. Therefore we only need to compute the linear term in $\chi$:
$$\displaystyle\left[\frac{\partial}{\partial\chi}\sum_{\sigma\sigma^{\prime}}%
\int_{0}^{2\pi}\frac{d\theta}{2\pi}\int_{0}^{y_{\sigma^{\prime}}}\frac{(\sigma%
+\sigma^{\prime}\cos{n\theta})ydy}{\sqrt{y^{2}+y_{\sigma}^{2}-2yy_{\sigma}\cos%
{\theta}}}\right]_{\chi=0}$$
$$\displaystyle=$$
$$\displaystyle\int_{0}^{2\pi}\frac{d\theta}{2\pi}\left[\int_{0}^{1}\frac{2(1-y%
\cos{\theta})ydy}{(1+y^{2}-2y\cos{\theta})^{\frac{3}{2}}}-\frac{\cos{n\theta}}%
{\sin{\frac{\theta}{2}}}\right],$$
(39)
and after evaluating the $dy$ integral in the square parenthesis, Eq. (B) gives
$$\displaystyle\int_{0}^{2\pi}$$
$$\displaystyle\frac{d\theta}{2\pi}\Bigg{[}2+\frac{2\cos{\theta}-\cos n\theta-1}%
{\sin{\frac{\theta}{2}}}$$
$$\displaystyle-2$$
$$\displaystyle\,\ln\left(1+\frac{1}{\sin{\frac{\theta}{2}}}\right)\cos{\theta}%
\Bigg{]}=\frac{4}{\pi}\sum_{j=0}^{n}\frac{1}{2j-1}.$$
(40)
Equation (16) is then easily obtained from Eq. (15) by neglecting all the cubic terms in the small parameters $g,\chi,r_{s}$ [e.g., the left side of Eq. (15) is $(y_{+}^{2}+y_{-}^{2})g=2g+O(g\chi^{2})$].
Appendix C Small $r_{s},g$ expansions
We give in this appendix some details on the expansions of Eq. (III.3):
$$\frac{v_{F\pm}}{v_{F}}=y_{\pm}\pm\frac{g}{2}ny_{\pm}^{n-1}+\frac{\sqrt{2}r_{s}%
}{16\pi}(I_{1}+I_{2}+I_{3}),$$
(41)
where we have split $L_{2}$ in its two contributions ($I_{1,2}$ refer to $\sigma=\pm$) and $I_{3}$ corresponds to $L_{3}$:
$$\displaystyle I_{1}$$
$$\displaystyle=$$
$$\displaystyle\int_{0}^{2\pi}d\theta\frac{\sqrt{2}y_{\pm}\cos{\theta}(1+\cos{n%
\theta})}{\sqrt{2}y_{\pm}\sin{\theta/2}+r_{s}},$$
(42)
$$\displaystyle I_{2}$$
$$\displaystyle=$$
$$\displaystyle\int_{0}^{2\pi}d\theta\frac{\sqrt{2}y_{\mp}\cos{\theta}(1-\cos{n%
\theta})}{\sqrt{1-y_{+}y_{-}\cos{\theta}}+r_{s}},$$
(43)
$$\displaystyle I_{3}$$
$$\displaystyle=$$
$$\displaystyle\pm\int_{0}^{2\pi}d\theta\int_{y_{+}}^{y_{-}}dy\frac{2n\sin{n%
\theta}\sin{\theta}}{\sqrt{y_{\pm}^{2}+y^{2}-2yy_{\pm}\cos{\theta}}+r_{s}\sqrt%
{2}},$$
(44)
Notice that these integrals have an explicit dependence on $r_{s}$ and $y_{\pm}=\sqrt{1\mp\chi}$. So, it is easier to perform first the expansion in the two small parameters $r_{s},\chi$. The final results in the main text are given in terms of the physical couplings of the hamiltonian: $r_{s}$ and $g$. Those final expression are easily obtained by substituting the value of $\chi$ in terms of $r_{s}$ and $g$ ($\chi\simeq g$ in first approximation).
The first integral, Eq. (42), can be evaluated exactly. In particular for $n=1$ we obtain
$$\displaystyle I_{1}=$$
$$\displaystyle-\frac{40}{3}+8\pi\delta(1-\delta^{2})+16\delta^{2}$$
(45)
$$\displaystyle+8(1-3\delta^{2}+2\delta^{4})\frac{\tanh^{-1}\sqrt{1-\delta^{2}}}%
{\sqrt{1-\delta^{2}}}$$
where $\delta=r_{s}/\sqrt{2}y_{\pm}$. This expression can then be easily expanded in $r_{s},\chi$ and an analogous procedure is followed for $n=2,3$. To lowest-order in $r_{s}$, we can set $r_{s}=0$ in $I_{2}$ and $I_{3}$. Similarly to the $I_{1}$ angular integral above, the $d\theta$ integrals of $I_{2}$ and $I_{3}$ at $r_{s}=0$ can be computed analytically for $n=1,2,3$. For $I_{2}$ this yields directly the desired function of $\chi$. For $I_{3}$ we still need to perform a last integration in $dy$. Since the integration region is of size $\sim\chi$ around $y=1$, we can expand the integrand in the small parameter $(y-1)$ and perform the integration in $dy$ order-by-order, which allows us to extract the leading terms of the expansion in $\chi$. For $n=1$ all this gives
$$\displaystyle\delta I_{1}$$
$$\displaystyle\simeq$$
$$\displaystyle\mp 4\chi-2\chi^{2}\mp\frac{4}{3}\chi^{3},$$
(46)
$$\displaystyle\delta I_{2}$$
$$\displaystyle\simeq$$
$$\displaystyle\mp\frac{4}{3}\chi+\chi^{2}\left(\ln\frac{\chi}{8}+\frac{13}{6}%
\right)\pm\frac{\chi^{3}}{2}\left(\ln\frac{\chi}{8}+\frac{3}{2}\right),$$
(47)
$$\displaystyle\delta I_{3}$$
$$\displaystyle\simeq$$
$$\displaystyle\pm\frac{16}{3}\chi+\frac{4}{3}\chi^{2}\pm\chi^{3}\left(\frac{2}{%
3}\ln\frac{\chi}{16}+\frac{389}{120}\right),$$
(48)
where only the corrections $\delta I_{\alpha}=I_{\alpha}(\chi)-I_{\alpha}(\chi=0)$ are listed, since terms independent on $\chi$ simply give the small $r_{s}$ expansion of the well known Eq. (19). Here, terms of order $O(r_{s}\chi,\chi^{4})$ are omitted, while it is interesting to keep the $O(\chi^{3})$ terms, since they give the leading spin splitting. Indeed, it is easily checked that the linear terms cancel
$$\sum_{\alpha=1}^{3}\delta I_{\alpha}\simeq\,\chi^{2}\left(\frac{3}{2}+\ln\frac%
{\chi}{8}\right)\pm\frac{\chi^{3}}{6}\left(\ln\frac{\chi^{7}}{8^{5}}+\frac{319%
}{20}\right),$$
(49)
which immediately leads to Eq. (27).
For $n=2,3$ we can proceed in a similar way. The spin splitting appears now already to linear order in $\chi$. By keeping the first subleading correction in $\chi$ we have for $n=2$:
$$\displaystyle\delta I_{1}$$
$$\displaystyle\simeq$$
$$\displaystyle\mp 4\chi-2\chi^{2},$$
(50)
$$\displaystyle\delta I_{2}$$
$$\displaystyle\simeq$$
$$\displaystyle\pm\frac{16}{15}\chi+\chi^{2}\left(4\ln{\frac{\chi}{8}}+\frac{134%
}{15}\right),$$
(51)
$$\displaystyle\delta I_{3}$$
$$\displaystyle\simeq$$
$$\displaystyle\pm\frac{64}{15}\chi+\frac{16}{15}\chi^{2},$$
(52)
and for $n=3$:
$$\displaystyle\delta I_{1}$$
$$\displaystyle\simeq$$
$$\displaystyle\mp 4\chi-2\chi^{2},$$
(53)
$$\displaystyle\delta I_{2}$$
$$\displaystyle\simeq$$
$$\displaystyle\pm\frac{212}{105}\chi+\chi^{2}\left(9\ln{\frac{\chi}{8}}+\frac{8%
99}{42}\right),$$
(54)
$$\displaystyle\delta I_{3}$$
$$\displaystyle\simeq$$
$$\displaystyle\pm\frac{144}{35}\chi+\frac{36}{35}\chi^{2}.$$
(55)
The final results are for $n=2$:
$$\sum_{\alpha=1}^{3}\delta I_{\alpha}\simeq\,\pm\frac{4}{3}\chi+4\chi^{2}\left(%
\ln\frac{\chi}{8}+2\right),$$
(56)
and for $n=3$:
$$\sum_{\alpha=1}^{3}\delta I_{\alpha}\simeq\,\pm\frac{32}{15}\chi+\chi^{2}\left%
(9\ln\frac{\chi}{8}+\frac{613}{30}\right).$$
(57)
From Eqs. (56) and (57) we immediately obtain Eqs. (28) and (29).
References
(1)
F. Giuliani and G. Vignale, Quantum Theory of the Electron
Liquid (Cambridge University Press, Cambridge, 2005).
(2)
R. Winkler, Spin-Orbit Coupling Effects in Two-Dimensional
Electron and Hole Systems (Springer, Berlin, 2003).
(3)
S. Datta and B. Das, Appl. Phys. Lett. 56, 665 (1989).
(4)
J. Nitta, T. Akazaki, H. Takayanagi, and T. Enoki, Phys. Rev. Lett. 78, 1335 (1997).
(5)
H. C. Koo, J. H. Kwon, J. Eom, J. Chang, S. H. Han, and M. Johnson, Science 325, 1515 (2009).
(6)
G.-H. Chen and M. E. Raikh, Phys. Rev. B 60, 4826 (1999).
(7)
D. S. Saraga and D. Loss,, Phys. Rev. B 72, 195319 (2005).
(8)
I. A. Nechaev, M. F. Jensen, E. D. L. Rienks, V. M. Silkin, P. M. Echenique, E. V. Chulkov, and P. Hofmann, Phys. Rev. B 80, 113402 (2009).
(9)
I. A. Nechaev, P. M. Echenique, and E. V. Chulkov, Phys. Rev. B 81, 195112 (2010).
(10)
A. Agarwal, S. Chesi, T. Jungwirth, J. Sinova, G. Vignale, and M. Polini, Phys. Rev. B 83, 115135 (2011).
(11)
Y. A. Bychkov and E. Rashba, JETP Lett. 39, 78 (1984).
(12)
Y. A. Bychkov and E. Rashba, J. Phys. C 17, 6039 (1984).
(13)
G. Dresselhaus, Phys. Rev. 100, 580 (1955).
(14)
M. I. D’yakonov and V.Yu. Kachorovskii, Sov. Phys. Semicond. 20, 110 (1986).
(15)
D. H. Berman and M. E. Flatté, Phys. Rev. Lett. 105, 157202 (2010).
(16)
S. Chesi and G. F. Giuliani, Phys. Rev. B 83, 235308 (2011).
(17)
S. H. Abedinpour, G. Vignale, and I. V. Tokatly, Phys. Rev. B 81, 125123 (2010).
(18)
A. Ambrosetti, F. Pederiva, E. Lipparini, and S. Gandolfi, Phys. Rev. B 80, 125306 (2009).
(19)
S. Chesi and G. F. Giuliani, Phys. Rev. B 83, 235309 (2011).
(20)
G. F. Giuliani and S. Chesi, Proceedings of Highlights in the
Quantum Theory of Condensed Matter (Edizioni della Normale, Pisa, 2005); S. Chesi, G. Simion, and G. F. Giuliani, e-print
arXiv:cond-mat/0702060; S. Chesi, Ph.D. thesis, Purdue University, 2007.
(21)
L. O. Juri and P. I. Tamborenea, Phys. Rev. B 77, 233310 (2008).
(22)
R. A. Żak, D. L. Maslov, and D. Loss, Phys. Rev. B 82, 115415 (2010).
(23)
R. A. Żak, D. L. Maslov, and D. Loss, arXiv:1112.4786.
(24)
I. L. Aleiner and V. I. Falko, Phys. Rev. Lett. 87, 256801 (2001).
(25)
W. A. Coish, V. N. Golovach, J. C. Egues, and D. Loss, phys. stat. sol (b) 243, 3658 (2006).
(26)
R. Winkler, Phys. Rev. B 62, 4245 (2000).
(27)
R. Winkler, H. Noh, E. Tutuc, and M. Shayegan, Phys. Rev. B 65, 155303 (2002).
(28)
L. P. Rokhinson, V. Larkina, Y. B. Lyanda-Geller, L. N. Pfeiffer,
and K. W. West, Phys. Rev. Lett. 93, 146601 (2004).
(29)
S. Chesi, G. F. Giuliani, L. P. Rokhinson, L. N. Pfeiffer, and K. W. West, Phys. Rev. Lett. 106, 236601 (2011).
(30)
D. V. Bulaev and D. Loss, Phys. Rev. Lett. 98, 097202 (2007).
(31)
S. Chesi and G. F. Giuliani, Phys. Rev. B 75, 155305 (2007).
(32)
E. I. Rashba and E. Ya. Sherman, Phys. Lett. A 129, 175 (1988).
(33)
R. Winkler, E. Tutuc, S. J. Papadakis, S. Melinte, M. Shayegan, D. Wasserman, and S. A. Lyon, Phys. Rev. B 72, 195321 (2005).
(34)
Y. T. Chiu, M. Padmanabhan, T. Gokmen, J. Shabani, E. Tutuc, M. Shayegan, and R. Winkler, e-print arXiv:1106.4608.
(35)
S. A. Crooker, J. J. Baumberg, F. Flack, N. Samarth, and D. D. Awschalom, Phys. Rev. Lett. 77, 2814 (1996).
(36)
J. M. Kikkawa and D. D. Awschalom, Phys. Rev. Lett. 80, 4313 (1998).
(37)
Y.-W. Tan, J. Zhu, H. L. Stormer, L. N. Pfeiffer, K. W. Baldwin, and K. W. West, Phys. Rev. Lett. 94, 016405 (2005).
(38)
N. D. Drummond and R. J. Needs, Phys. Rev. B 80, 245104 (2009).
(39)
J. F. Janak, Phys. Rev. 178, 1416 (1969).
(40)
To deal with extremely low densities at which only one spin band is occupied and the occupation is a ring in momentum space, it is also useful to generalize $\chi$ to values larger than one.ChesiHighDens However, we do not consider this low-density regime in this paper.
(41)
J. Schliemann, Phys. Rev. B 74, 045214 (2006).
(42)
The method of Ref. Chesi07April, is slightly preferable, in that it leads to variational wavefunctions with a lower total energy.
(43)
B. Jusserand, D. Richards, H. Peric, and B. Etienne, Phys. Rev. Lett. 69, 848 (1992).
(44)
J. P. Eisenstein, L. N. Pfeiffer, and K. W. West, Phys. Rev. Lett. 68, 674 (1992). |
Coupled-Channel Effects for the Bottomonium with Realistic Wave Functions
Yu Lu${}^{1,3}$
111Email address:[email protected] ,
Muhammad Naeem Anwar ${}^{2,3}$
222Email address:[email protected] ,
Bing-Song Zou${}^{2,3}$
333Email address:[email protected]
${}^{1}$Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049, China,
${}^{2}$CAS Key Laboratory of Theoretical Physics, Institute of Theoretical Physics,
Chinese Academy of Sciences, Beijing 100190, China
${}^{3}$University of Chinese Academy of Sciences, Beijing 100049, China
(November 19, 2020)
Abstract
With Gaussian expansion method (GEM),
realistic wave functions are used to calculate coupled-channel effects for the bottomonium under the framework of ${}^{3}P_{0}$ model.
The simplicity and accuracy of GEM are explained.
We calculate the mass shifts, probabilities of the $B$ meson continuum, $S-D$ mixing angles,
strong and dielectric decay widths.
Our calculation shows that both $S-D$ mixing and the $B$ meson continuum can contribute to
the suppression of the vector meson’s dielectric decay width.
We suggest more precise measurements on the radiative decays of $\Upsilon(10580)$ and $\Upsilon(11020)$ to distinguish these two effects.
The above quantities are also calculated with simple harmonic oscillator (SHO) wave function approximation for comparison.
The deviation between GEM and SHO indicates that it is essential to treat the wave functions accurately for near threshold states.
1 Introduction
Heavy quarkonium is a multi-scale system covering all regimes of quantum chromodynamics (QCD) which make it an ideal place to study strong interactions [1].
Despite the success of QCD in high energy region, due to asymptotic freedom,
non-perturbative effect dominates at low energies and brings problems to perturbative calculation.
One tool to study this non-perturbative effect is lattice QCD.
However, due to its huge calculation work, it is still unable to calculate all the physical quantities with the current computation power.
Another important approach is to develop various phenomenological models.
Among these phenomenological models, quark model is a prominent one.
Under quark model framework,
various types of interactions have been suggested by various groups,
and they have achieved many impressive successes (see e.g. Refs. [2, 3, 4, 5, 6]).
However, these potential models cannot be the whole story.
One important missing ingredient is the mechanism to generate quark-antiquark pairs,
which enlarge the Fock space of the initial state,
i.e. the initial state contains multi-quark components.
These multi-quark components will change the Hamiltonian of the potential model,
causing mass shift and mixing between states with same quantum numbers
or directly contributing to open channel strong decay if the initial state is above threshold.
These consequences can be summarized as unquenched effects or coupled-channel effects.
Coupled-channel effects have been considered at least 30 years ago by
Törnqvist et al. in Refs. [7, 8, 9, 10, 11];
they extended the quark model to be unquenched quark model.
Despite the fact that the underlying quark pair creation mechanism is not fully understood up to now,
still there are different phenomenological models to decode the mystery,
such as ${}^{3}P_{0}$ model [12, 13, 14],
flux-tube breaking model [15, 16],
microscopic decay models [4, 17, 18].
Among these, the most simple and successful one is the ${}^{3}P_{0}$ model,
where the generated light quark pair share the same quantum number as vacuum.
Even though ${}^{3}P_{0}$ model is extensively studied by many people,
however, almost all the calculations are using SHO wave function approximation to simplify the calculation
(see e.g. [19, 18, 20, 21, 22, 23, 24, 25, 26]).
A simple yet powerful method to handel the wave function precisely is still not widely known.
We propose using the Gaussian expansion method (GEM) [27] to accurately evaluate the wave function convolution.
There have been already some works related to GEM.
In Refs. [28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41],
GEM is adopted to calculate the wave functions under variational method approach.
Coupled-channel effects with GEM are only studied for some specific cases,
such as $X(3872)$ and $P$ wave $D_{s}$ mesons [42, 43].
In Refs. [44, 45, 46],
the authors also use GEM to calculate the spectrum and open channel strong decays of light mesons and some specific charmonia,
where the coupled channel induced mass shift is not considered.
Even though the mass shifts can be partly absorbed by redefining the potential,
the potential model cannot describe near-threshold effects [47].
We want to emphasize that the mass shifts and open channel strong decays are directly correlated by coupled-channel effects,
so it is essential to evaluate them under a same framework and calculate them precisely.
So far, a precise evaluation and a thorough discussion of the coupled-channel effects are still missing,
and the validity of SHO approximation is yet to be clarified.
In this paper, we fill these gaps by a thorough discussion of coupled-channel effects for the bottomonium
and we also predict some important results on the dielectric and radiative decays of vector mesons
which are going to be tested by experiment.
The paper is organized as follows.
In Sec. 2, we explain the details of Cornell potential and ${}^{3}P_{0}$ model,
where we deduce the formula of mass shift, open channel strong decay width and $S-D$ mixing.
In Sec. 3, we focus on the calculation details and GEM,
where the advantages of GEM are elucidated and the procedure to fit the wave function is explained.
Sec. 4 is devoted to discussing the possible impacts of coupled-channel effects for the bottomonium
on the spectrum, open channel strong decays, probabilities of the $B$ meson continuum,
the $S-D$ mixing and the vector meson’s dielectric and radiative decays.
We also explicitly show the deviation between GEM and SHO approximation.
Finally, we give a short summary of this work in Sec. 5.
2 Theoretical Framework
2.1 Cornell Potential Model
As the quenched limit,
the wave functions for the heavy quarkonium are obtained by solving the Schrödinger equation
with the well-known Cornell potential [4, 17]
$$V(r)=-\frac{4}{3}\frac{\alpha}{r}+\lambda r+c,$$
(1)
where $\alpha,\lambda$ and $c$ stand for the strength of color Coulomb potential,
the strength of linear confinement and mass renormalization, respectively.
To restore the hyperfine or fine structures of the bottomonium,
we use the following form of the spin dependent interactions
$$V_{s}(r)=\left(\frac{2\alpha}{m^{2}_{b}r^{3}}-\frac{\lambda}{2m^{2}_{b}r}%
\right)\vec{L}\cdot\vec{S}+\frac{32\pi\alpha}{9m_{b}^{2}}\tilde{\delta}(r)\vec%
{S}_{b}\cdot\vec{S}_{\bar{b}}+\frac{4\alpha}{m^{2}_{b}r^{3}}\left(\frac{\vec{S%
}_{b}\cdot\vec{S}_{\bar{b}}}{3}+\frac{(\vec{S}_{b}\cdot\vec{r})(\vec{S}_{\bar{%
b}}\cdot\vec{r})}{r^{2}}\right),$$
(2)
where $\vec{L}$ denotes relative orbital angular momentum,
$\vec{S}=\vec{S}_{b}+\vec{S}_{\bar{b}}$ is the total spin of the $b$ quark pairs
and $m_{b}$ is the $b$ quark mass.
Since the non-relativistic expansion will fail if two composite quarks are very close to each other,
instead of Dirac $\delta$ function in the second term,
we use the smeared delta function,
which can be written as
$\tilde{\delta}(r)=(\sigma/\sqrt{\pi})^{3}e^{-\sigma^{2}r^{2}}$ [48, 47].
The Hamiltonian of the Schrödinger equation in quenched limit is represented as
$$H_{0}=2m_{b}+\frac{p^{2}}{m_{b}}+V(r)+V_{s}(r).$$
(3)
We treat the spin dependent term as a perturbation and
the spatial wave functions are obtained by solving Schrödinger equation numerically using Numerov’s method [49].
2.2 ${}^{3}P_{0}$ Model and Coupled-Channel Effects
For the coupled channel calculation,
we adopt the widely used ${}^{3}P_{0}$ model or quark pair creation model,
which is first proposed by L. Micu [12] in 1969 and then extended by A. Le Yaouanc et al. in 1970s [13, 14].
In this model, the generated quark pairs have vacuum quantum number $J^{PC}=0^{++}$.
After simple arithmetic, one can conclude that the relative orbital angular momentum and total spin are both equal to 1
and total angular momentum is 0.
In the notation of ${}^{2S+1}L_{J}$, one should write it as ${}^{3}P_{0}$ which explains the model’s name.
The interaction Hamiltonian can be expressed as
$$H_{I}=2m_{q}\gamma\int d^{3}x\bar{\psi}_{q}\psi_{q},$$
(4)
where $m_{q}$ is the produced quark mass,
$\gamma$ is the dimensionless coupling constant.
Since the probability to generate heavier quarks is suppressed,
we use the effective strength $\gamma_{s}=\frac{m_{q}}{m_{s}}\gamma$ in the following calculation,
where $m_{q}=m_{u}=m_{d}$ is the constituent quark mass of up (or down) quark and $m_{s}$ is strange quark mass.
The ${}^{3}P_{0}$ Hamiltonian induces not only open-flavor strong decays of the heavy quarkonium above threshold,
but also coupled-channel effects.
As sketched by Fig. 1, the experimentally observed state should be a mixture of pure quarkonium state (bare state) and $B$ meson continuum.
Put it in to formula, the physical or experimentally observed state $|{A}\rangle$ should be expressed as
$$|{A}\rangle=c_{0}|{\psi_{0}}\rangle+\sum_{BC}\int d^{3}p\,c_{BC}(p)|{BC;p}\rangle,$$
(5)
where $c_{0}$ and $c_{BC}$ stand for the normalization constants of the bare state and $B$ meson continuum, respectively.
$|{\psi_{0}}\rangle$ is normalized to 1 and $|{A}\rangle$ is also normalized to 1 if it lies below $B\bar{B}$ threshold.
$|{BC;p}\rangle$ is normalized as $\langle{BC;p_{1}}|{B^{\prime}C^{\prime};p_{2}}\rangle=\delta^{3}(p_{1}-p_{2})%
\delta_{BB^{\prime}}\delta_{CC^{\prime}}$,
where $p$ is the momentum of $B$ meson in $|{A}\rangle$’s rest frame.
Combine the Cornell potential and the dynamics of quark pair generation,
we get the full Hamiltonian,
$$H=H_{0}+H_{BC}+H_{I},$$
(6)
with the following relations
$$\displaystyle H_{0}|{\psi_{0}}\rangle$$
$$\displaystyle=$$
$$\displaystyle M_{0}|{\psi_{0}}\rangle$$
(7)
$$\displaystyle H_{0}|{BC;p}\rangle$$
$$\displaystyle=$$
$$\displaystyle 0$$
(8)
$$\displaystyle H_{BC}|{\psi_{0}}\rangle$$
$$\displaystyle=$$
$$\displaystyle 0$$
(9)
$$\displaystyle H_{BC}|{BC;p}\rangle$$
$$\displaystyle=$$
$$\displaystyle E_{BC}|{BC;p}\rangle$$
(10)
$$\displaystyle H|{A}\rangle$$
$$\displaystyle=$$
$$\displaystyle M|{A}\rangle,$$
(11)
where $M_{0}$ is the bare mass of the charmonium state and can be solved directly from Schrödinger equation.
The interaction between $B$ mesons is neglected,
so the energy of meson continuum can be expressed as
$E_{BC}=\sqrt{m_{B}^{2}+p^{2}}+\sqrt{m_{C}^{2}+p^{2}}.$
When Eq. (11) is projected onto each component, we immediately get
$$\displaystyle\langle{\psi_{0}}|H|{\psi}\rangle=c_{0}M=c_{0}M_{0}+\int d^{3}p\,%
c_{BC}(p)\langle{\psi_{0}}|H_{I}|{BC;p}\rangle,$$
(12)
$$\displaystyle\langle{BC;p}|H|{\psi}\rangle=c_{BC}(p)M=c_{BC}(p)E_{BC}+c_{0}%
\langle{BC;p}|H_{I}|{\psi_{0}}\rangle.$$
(13)
Solve $c_{BC}$ from Eq. (13), substitute back to Eq. (12) and eliminate the $c_{0}$ on both sides,
we get a integral equation
$$M=M_{0}+\Delta M,$$
(14)
where
$$\Delta M=\sum_{BC}\int d^{3}p\,\frac{|\langle{BC;p}|H_{I}|{\psi_{0}}\rangle|^{%
2}}{M-E_{BC}-i\epsilon}.$$
(15)
The sum of $BC$ is restricted to the ground state $B_{(s)}$ mesons, i.e. $B\bar{B},B\bar{B}^{*}+h.c.,B^{*}\bar{B}^{*},B_{s}\bar{B}_{s},B_{s}\bar{B}^{*}_%
{s}+h.c.,B^{*}_{s}\bar{B}^{*}_{s}$.
Note that the $i\epsilon$ term is added to handle the situation when $m_{A}>m_{B}+m_{C}$.
In this case, $\Delta M$ will pick up an imaginary part
$$\mathrm{Im}(\Delta M)=\sum_{BC}\pi P_{B}\frac{E_{B}E_{C}}{m_{A}}|\langle{BC;P_%
{B}}|H_{I}|{\psi_{0}}\rangle|^{2},$$
(16)
which is equal to one half of the the decay width.
$P_{B}$ and $E_{B}$ respectively denote the momentum and energy of $B$ meson.
The wave function overlap integration lies in the term
$$\langle{BC;P_{B}}|H_{I}|{\psi_{0}}\rangle=\sum_{\text{polarization}}\int d^{3}%
k\phi_{0}(\vec{k}+\vec{P}_{B})\phi_{B}^{*}(\vec{k}+x_{B}\vec{P}_{B})\phi_{C}^{%
*}(\vec{k}+x_{C}\vec{P}_{B})|\vec{k}|Y_{1}^{m}(\theta_{\vec{k}},\phi_{\vec{k}}),$$
(17)
where $x_{B}=m_{4}/(m_{1}+m_{4}),x_{C}=m_{3}/(m_{2}+m_{3})$ and
$m_{1}=m_{2}=m_{Q},m_{3}=m_{4}$ respectively denote the $b$ quark and the light quark mass.
Once $M$ is solved, the coefficient of different components can be worked out either.
For states below threshold, the normalization condition $|{A}\rangle$ can be rewritten as
$$|c_{0}|^{2}+\int d^{3}p|c_{BC}|^{2}=1$$
(18)
after the substitution of $c_{BC}$, we get the probability of the $b\bar{b}$ component
$$P_{b\bar{b}}:=|c_{0}|^{2}=1/\left(1+\sum_{BCLS}\int_{0}^{\infty}dp\frac{p^{2}|%
\mathcal{M}^{LS}|^{2}}{(M-E_{BC})^{2}}\right),$$
(19)
where $|\mathcal{M}^{LS}|^{2}$ is represented as
$$|\mathcal{M}^{LS}|^{2}=\int d\Omega_{B}\,|\langle{BC;P_{B}}|H_{I}|{\psi_{0}}%
\rangle|^{2}.$$
(20)
2.3 Coupled Channel Induced S-D Mixing
From the quark model’s perspective,
the spatial wave functions of $J^{PC}=1^{--}$ family can be both $S$ and $D$ wave.
It is natural to expect that the experimentally observed vector states are mixing of $S$ and $D$ waves.
As in the case of conventional meson coupling with $B\bar{B}$ continuum,
we rewrite it into a matrix form
$$\left(\begin{array}[]{cc}M_{0}&\int d^{3}p\langle{\psi_{0}}|H_{I}|{BC}\rangle%
\\
\langle{BC}|H_{I}|{\psi_{0}}\rangle&E_{BC}\\
\end{array}\right)\left(\begin{array}[]{c}c_{0}\\
c_{BC}\\
\end{array}\right)=M\left(\begin{array}[]{c}c_{0}\\
c_{BC}\\
\end{array}\right),$$
(21)
where the integration part should be understood as a formal notation,
and one need to insert all the $p$ dependent part into the integral.
For example, in the above case,
one may naively get the following form after diagonalization
$$(M-M_{0})(M-E_{BC})=\int d^{3}p|\langle{\psi_{0}}|H_{I}|{BC}\rangle|^{2}.$$
(22)
However, the correct form should be understand as Eq.(15),
where the $(M-E_{BC})$ term have moved to the integration.
The advantage of the matrix form is that one can easily see its structure and
can be easily generalized to $S-D$ mixing case.
Under the assumption that $(n+1)S$ mix only with $nD$, we have
$$\left(\begin{array}[]{ccc}M_{S}^{0}&H_{T}&\int d^{3}p\langle{\psi_{S}}|H_{I}|{%
BC}\rangle\\
H_{T}&M_{D}^{0}&\int d^{3}p\langle{\psi_{D}}|H_{I}|{BC}\rangle\\
\langle{BC}|H_{I}|{\psi_{S}}\rangle&\langle{BC}|H_{I}|{\psi_{D}}\rangle&E_{BC}%
\\
\end{array}\right)\left(\begin{array}[]{c}c_{S}\\
c_{D}\\
c_{BC}\\
\end{array}\right)=M\left(\begin{array}[]{c}c_{S}\\
c_{D}\\
c_{BC}\\
\end{array}\right).$$
(23)
The $S-D$ mixing induced by tensor part of the potential is so small,
typically around $0.8^{\circ}$ in our calculation (see also the appendix A in Ref. [50]),
so its quite reasonable to set $H_{T}=0$.
After this approximation, one can reexpress $c_{BC}$ in terms of $c_{S},c_{D}$ and easily get
$$\left(\begin{array}[]{cc}M_{S}^{0}+\Delta M_{S}&\Delta M_{SD}\\
\Delta M_{DS}&M_{D}^{0}+\Delta M_{D}\\
\end{array}\right)\left(\begin{array}[]{c}c_{S}\\
c_{D}\\
\end{array}\right)=M\left(\begin{array}[]{c}c_{S}\\
c_{D}\\
\end{array}\right),$$
(24)
where
$$\displaystyle\Delta M_{f}$$
$$\displaystyle=$$
$$\displaystyle\int d^{3}p\frac{|\langle{\psi_{f}}|H_{I}|{BC}\rangle|^{2}}{M-E_{%
BC}-i\epsilon}\qquad(f=S,D),$$
(25)
$$\displaystyle\Delta M_{SD}=\Delta M_{DS}^{*}$$
$$\displaystyle=$$
$$\displaystyle\int d^{3}p\frac{\langle{\psi_{S}}|H_{I}|{BC}\rangle\langle{BC}|H%
_{I}|{\psi_{D}}\rangle}{M-E_{BC}-i\epsilon}.$$
(26)
From the above equation, both the mass and the relative ratio $c_{S}/c_{D}$ can be worked out.
For states below threshold, the probability can be solved once the mass is known,
which is a generalization of Eq. (19)
$$|c_{S}|^{2}+|c_{D}|^{2}+\sum_{BC}\int d^{3}p\frac{1}{(M-E_{BC})^{2}}(|c_{S}|^{%
2}H_{S,BC}^{2}+|c_{D}|^{2}H_{D,BC}^{2}+2\text{Re}[c_{S}c_{D}^{*}H_{S,BC}H_{BC,%
D}])=1,$$
(27)
where $H_{f,i}$ stands for$\langle{f}|H_{I}|{i}\rangle$.
One will get a complex solution of $M=M_{\text{BW}}+i\Gamma/2$ if $M_{\text{BW}}>m_{B}+m_{\bar{B}}$,
where $M_{\text{BW}}$ represents the Breit-Wigner mass of the resonance,
and $\Gamma$ is the decay width after considering $S-D$ mixing.
As a cross check, one can also calculate the decay width directly with the following formula,
$$\Gamma_{SD}=2\Big{(}|c_{S}|^{2}\mathrm{Im}(\Delta M_{S})+|c_{D}|^{2}\mathrm{Im%
}(\Delta M_{D})+2\mathrm{Re}\big{(}c_{S}^{*}c_{D}\mathrm{Im}(\Delta M_{SD})%
\big{)}\Big{)}.$$
(28)
Eq. (24) is much more difficult to solve than the Eq. (15).
The method we use to solve this equation will be discussed in the next section.
3 Parameter Selection and Gaussian Expansion Method
3.1 Parameter Selection
As a first step, we tune the wave functions to be consistent with the dielectric decay widths of $\Upsilon(nS)$ for $n\leq 3$.
Theoretically, dielectric decay widths can be expressed as [51, 52, 53, 54]
$$\Gamma_{ee}=\beta\frac{4\alpha^{2}e^{2}_{b}}{M^{2}_{nS}}|c_{S}\,R_{nS}(0)+c_{D%
}\frac{5}{2\sqrt{2}m_{b}^{2}}R_{nD}^{\prime\prime}(0)|^{2}\\
$$
(29)
where $\beta=(1-16\alpha_{s}/3\pi)$ is the QCD radiative correction,
and $e_{b}=-1/3$ is the $b$ quark charge in the unit of electron charge.
$R_{nS}(0)$ denotes the radical $S$ wave function at the origin,
and $R_{nD}^{\prime\prime}(0)$ is the second derivative of the radical $D$ wave function at the origin.
$c_{S}$ and $c_{D}$ respectively denote the normalization coefficients before $S$ and $D$ wave.
Note that from the perspective of coupled channels,
they are not restricted to be real-valued and $|c_{S}|^{2}+|c_{D}|^{2}\neq 1$.
For below threshold states, the correct normalization is given by Eq. (27).
Nevertheless, if the imaginary part of the $\Delta M$ are neglected in Eq. (24),
the corresponding solutions will be real,
and one can easily get the feel of how big the mixing is by defining
$\tan\theta:=|c_{S}/c_{D}|$ for $S$ wave dominate states and $\tan\theta:=|c_{D}/c_{S}|$ for $D$ wave dominate states.
There is also an argument for the above formula that
the QCD corrections of higher order may be important [50],
thus $\beta$ has to be treat as an effective constant.
So in order to reduce the parameter’s uncertainty,
we tune the wave functions to reproduce $\Gamma_{ee}(nS)/\Gamma_{ee}(1S),(n=2,3)$(see Fig. 6).
3.2 Gaussian Expansion Method
There are at least two ways to solve Eq. (15).
First one is recursion method,
which based on the observation that the mass shift is expected to be small compared with the bare mass.
i.e. set $M=m_{0}$ as the first step and do the integration in Eq. (15) to get the mass shift,
then set $M=m_{0}+\Delta m$ again and so on until the result converges.
One can even make a further approximation and only do the first step recursion.
However, this method only applies to the single channel mass shift formula (15).
In $S-D$ mixing cases, such as Eq. (24),
the mass difference between $M_{S}^{0}+\Delta M_{S},M_{D}^{0}+\Delta M_{D}$ is small,
even a small error in the off-diagonal term in Eq. (24) will ruin the prediction of the $S-D$ mixing angle.
The second way is to solve the equation by brute-force,
i.e. for the energy ranges we are interested in,
work out all of the integrations in Eq. (15) or Eq. (24) at specific energy point.
We use this method despite of its huge calculation work.
The benefit is that we can extract a lot of information about the wave function’s impact on the mass shift.
One can also change the ${}^{3}P_{0}$’s coupling constant $\gamma$ or the mass renormalization constant $c$ to see the possible consequences.
The high precision work will be not convincing if there is no way to precisely evaluate the integration,
which has a key ingredient — the wave function.
One can indeed evaluate the amplitude pure numerically as the authors do in Refs. [55, 47],
however, we still want analytic expressions which is more convenient
if we want to change the parameters and then repeat the calculations.
In order to achieve that,
various groups approximate the wave functions by simple harmonic oscillators (SHOs) approximation (see e.g. [19, 18, 20, 23, 22]).
The oscillator parameters $\beta$s are usually settled down by requiring that
the root mean square radii to be equal to the initial states [56, 57, 58]
or maximizing its overlap with the numerical wave function [18].
To improve the accuracy, people also expand the true wave function in terms of SHOs (see e.g. Refs. [59, 60]).
As a consequence, one will get fairly complicated analytic expression for highly excited states.
For example, expression (2.12) in Ref. [59].
Due to the highly oscillated behavior of the excited SHOs,
one would need a large number of SHOs to achieve an ideal precision.
We think that it is necessary to fully respect the wave function and make a precise calculation of the transition amplitude,
which is shown to be essential for the states near threshold.
In this work, we get both the analytic expression and the high precision
by using the Gaussian expansion method (GEM) proposed by Hiyama et al. [27].
This method has the observation that bound state’s wave function can be expanded in Gaussian bases as the following
$$\psi_{NLM}(r)=\left(\sum_{i=1}^{n}c_{i}\beta_{i}^{L+\frac{3}{2}}e^{-\frac{1}{2%
}\beta_{i}^{2}r^{2}}r^{L}\right)Y_{L}^{M}(\theta,\varphi),$$
(30)
where $\beta_{i}$s and $c_{i}$s denote oscillator parameters and corresponding coefficients, respectively.
$n$ is the number of Gaussian basis.
In this work, $n=5\sim 20$ for initial states and 5 for $B$ mesons,
and $\beta_{i}$s lie in the range of $0.1\sim 5$GeV.
Compared with the SHO basis, the Gaussian basis is no longer orthogonal.
So a little trick is used to speed up the fitting procedure.
As explained in Ref [27],
$\beta_{i}$s are set to be a geometric series.
Instead of increasing the number of SHO basis for fixed $\beta$,
GEM can both increase the bases number and change $\beta$ to improve the fit.
The quality of the fitting by GEM is quite impressive which is shown in Fig. 2.
Because the ${}^{3}P_{0}$ model calculation is easily done in momentum space,
we need to make a Fourier transformation of the position space wave function.
One benefit of the SHO wave function is that it is invariant after Fourier transform
apart from the substitution $\beta\to 1/\beta$.
Since Gaussian basis is ground state SHO wave function, it naturally keeps this property.
That means, after fitting position space wave functions,
we can rebuild the momentum space wave functions by
$$\psi_{NLM}(p)=\left(\sum_{i=1}^{n}c_{i}\beta_{i}^{-(L+\frac{3}{2})}e^{-\frac{p%
^{2}}{2\beta_{i}^{2}}}p^{L}\right)Y_{L}^{M}(\theta,\varphi).$$
(31)
What makes GEM simple is that there are a minimum number of polynomials in the integration
which simplifies the expression from the beginning.
GEM is quite universal and it is not limited to the wave functions obtained by solving the non-relativistic Schrödinger equation.
To work out the analytic expression,
one has to deal with the associated Laguerre polynomials if SHO wave functions are involved
and also the sophisticated angular integration.
Although in Ref. [61], Roberts and B. Silvestre-Brac show us the general method to do the integration,
and there are analytical expressions [59],
however, these expressions are quite lengthy and they only apply to ${}^{3}P_{0}$ model.
The complexity can be bypassed if we transfer the spherical harmonics into Cartesian form [62] from the beginning.
After this transformation, the general form of the integration can be compactly expressed as
$$\iiint dp_{x}dp_{y}dp_{z}\exp(-p^{\mathrm{T}}.A.p-B.p-C)f(p_{x},p_{y},p_{z}),$$
(32)
where $p,A,B$ and $C$ in the exponent denote $(p_{x},p_{y},p_{z})^{\mathrm{T}}$, $3\times 3$ real symmetric matrix,
three-vector and constant, respectively.
$f(p_{x},p_{y},p_{z})$ is nothing but polynomial, so Eq. (32) is a standard form of Gaussian integration
which can be easily done even manually.
After the integration is done, we can easily transform it back to spherical basis by the substitution
$P_{x}=P\sin\theta\cos\varphi,P_{y}=P\sin\theta\sin\varphi,P_{z}=P\cos\theta$.
Another benefit of this transformation is that it can easily handle much more complicated polynomials $f(p_{x},p_{y},p_{z})$
which may show up in other quark pair creation models.
4 Result and Discussion
4.1 Mass Shift and Open Channel Strong Decay
From Tab. 2 and 3,
one can find that the mass shifts are generally same between GEM and SHOs with a few exceptions for near threshold states.
The mass shifts in a same multiplet are also almost equal
simply because that their wave functions are identical and their bare masses are approximately equal.
This conclusion is consistent with the loop theorem in Ref. [63].
For the states below threshold, $\Delta M$s are all negative,
and closer to the threshold gets more deduction of the mass.
With GEM, this conclusion is true even for states slightly above threshold.
This conclusion differs with Refs. [22, 23], where SHO are used to calculate the mass shift.
Take $h_{b}$ family as an example, our mass shift grows with the mass going higher no matter whether we use GEM or SHO,
however, in Ref. [22] and Ref. [23],
the largest mass shift happens to $h_{b}(1P)$ and $h_{b}(2P)$, respectively.
For states above threshold,
the mass shift behavior becomes complicated (see Fig. 4)
and it is not appropriate to draw the conclusion that the mass shift of state above the $B\bar{B}$ threshold is positive.
This conclusion is only true for asymptotically large mass, and in this case, $\Delta M\propto 1/M$.
We should also point out that this mathematical fact does not mean it will definitely happen.
The reason is that when mass becomes bigger, more $B$ meson channels will contribute,
and one cannot tell the sign of $\Delta M$ before summing all possible channels’ contributions in Eq. (15).
In order to study this sensitivity,
we also plot the dependence on the initial state mass of $\Delta M$ and decay width
for the vector meson above threshold in Fig. 4 and Fig. 5, respectively.
As a concrete example, one can see this sensitivity by comparing $\Upsilon(4S)$ with $\Upsilon(6S)$.
Compared with $\Upsilon(4S)$, the wave function of $\Upsilon(6S)$ has more nodes,
however, basing on this fact one cannot conclude that the $\Delta M$’s behavior is more complicated.
The important reason is that the bare mass of $\Upsilon(6S)$ is also farther from threshold,
causing the average of the wave function overlap integration in Eq. (17).
Note also that absolute value of the mass shift of $\Upsilon(6S)$ calculated in SHO is larger than GEM,
however, in $\Upsilon(4S)$ case,
we have the opposite conclusion if we choose the lowest intersection point of $\Delta M$ and $M-M_{0}$.
From Tab. 2,
one can find that the masses predicted in Ref. [22] are generally closer to the experimental data.
However, we want to stress that the spectrum is an important but not the only criterion to judge whose parameters are better.
As shown in Fig. 6,
the dielectric decay ratios $\Gamma_{ee}/\Gamma_{ee}(1S)$ calculated with the parameters given in Ref. [22]
are generally smaller than experimental measurements before coupled-channel effects are taken into account.
As will be discussed in Sec. 4.2,
coupled-channel effects will suppress rather than enhance these ratios,
so their parameters are difficult to explain the dielectric decays of vector mesons despite of their success in the spectrum.
The complicated structure of the mass shift of $\Upsilon(4S)$ needs further discussion.
Even though the curve of GEM and SHO share some common features,
the small difference is sufficient to generate a large discrepancy of the mass shift.
Another interesting feature of this plot is that GEM has three solutions,
implying that more resonances may pop up compared with potential model prediction.
This sensitivity can also be seen by the decay width behavior in Fig. 5.
Over a large energy range $10.58\sim 10.73$ GeV,
the decay width of $\Upsilon(4S)$ calculated by GEM can be around two times as large as SHO.
Our decay width plot of $\Upsilon(4S)$ also share some resemblance with Fig.2 in Ref. [64],
where the prediction of SHO is not calculated.
The deviations of mass shift and decay width tell us that
it is necessary to adopt the realistic wave functions other than SHO approximation in the coupled channel calculation.
Of course, one may argue that, since the bare mass is 140MeV heavier
than the experimental measurement,
if the bare mass is tuned closer to the $B\bar{B}$ threshold,
the difference between GEM and SHO would be small and we will get one solution.
However, we want to stress that the bare mass is directly related to the wave function,
in the case where we have a smaller bare mass, the wave function will also be different,
thus cause different mass shift behavior.
This sensitivity also reminds us that taking only one step approximation
in the recursive method to solve Eq. (15) may cause a large error,
so an accurate treatment of wave function and a precise method to solve Eq. (15)
are essential for near threshold states.
If $\Upsilon(10580),\Upsilon(10860)$ and $\Upsilon(11020)$ are treated to be pure $S$ or $D$ wave,
we get the open channel decay width shown in Tab. 5.
It is worthy to note that this assumption is oversimplified, so the absolute value cannot be treated too seriously.
4.2 S-D Mixing and Dielectric Decay
As explained in Sec. 2 and sketched in Fig. 1,
coupled-channel effects will also induce mixing among states with same $J^{PC}$.
In this paper, we focus on the mixing between $\Upsilon(S)$ and $\Upsilon_{1}(D)$ family.
The Cornell potential model tells us that
the mass splitting between $\Upsilon((n+1)S)$ and $\Upsilon_{1}(nD)$ is smaller than other configurations,
such as $\Upsilon(nS)$ and $\Upsilon((n+1)S)$ or $\Upsilon_{1}(nD)$ and $\Upsilon_{1}((n+1)D)$ states,
so its quite reasonable to assume that the mixing only happens to $\Upsilon((n+1)S)$ and $\Upsilon_{1}(nD)$.
The masses and corresponding mixing angles after considering $S-D$ mixing is listed in Tab. 6.
In Eq. 24, the overall phase before $c_{S}$ and $c_{D}$ is nonphysical,
so we are free to set the phase of $c_{D}$ to be 0, i.e. $c_{D}\geq 0$.
Under this convention and the normalization condition Eq. (27),
the ratio $c_{S}/c_{D}$ is adequate to fix the value of $c_{S}$ and $c_{D}$.
If the imaginary part of $\Delta M_{f}$ and $\Delta M_{SD}$ in Eq. (24) are neglected,
one would get real solutions both for $M$ and $c_{S}/c_{D}$.
After this approximation, one can deduce the mixing angles.
However, the definition of $c_{S}$ and $c_{D}$ doesn’t exist for states above threshold [56, 65].
Despite of this difficulty, we follow Ref. [7],
assuming that these open channels’ contribution are neglected.
Under this assumption, quantities related to $S-D$ mixing are shown in Tab. 6.
From Tab. 6, we found that the masses barely change
after considering $S-D$ mixing for states below threshold,
indicating that the mixing angles are approximately 0.
So it is reasonable to treat $\Upsilon(1S),\Upsilon(2S)$ and $\Upsilon(3S)$ as pure $S$ wave states.
This conclusion also agrees with the loop theorem in Ref. [63].
From Eq. (29),
we also learn that the dielectric decay of $\Upsilon_{1}(D)$ is suppressed by the $b$ quark mass $m_{b}$,
so the small mixing also provide a natural explanation why these $D$ wave vector mesons are difficult to find at $e^{+}e^{-}$ collider.
From Tab. 6, one can also read off the open channel strong decay widths after considering $S-D$ mixing for states above threshold.
However, one cannot compare the imaginary part of $M_{\text{comp}}$ directly with experimental data
because its real part (which is the Breit-Wigner mass) does not equal to the experimental mass,
thus their phase space for $B\bar{B}$ are different from experiment.
A natural and direct consequence of non-negligible $S-D$ mixing
is the suppression of $\Gamma_{ee}(S)$ or enhancement of $\Gamma_{ee}(D)$.
As can be seen from Fig. 6,
the dielectric decay width of $\Upsilon(10580)$ and $\Upsilon(11020)$ is highly suppressed experimentally.
Under the assumption that $\Upsilon(10580)$ and $\Upsilon(11020)$ are $S$ wave dominate states,
one may be tempted to introduce a large $S-D$ mixing angle for these highly excited states (see e.g. Ref. [50]).
The unexpected large central value of $\Gamma_{ee}(\Upsilon(10860))$ seems to favor a small $S-D$ mixing angle,
however, due to its large errors,
a mixing angle as large as $27^{\circ}$ can also reproduce the data which gives the decay width lies at the lower bound [50].
Of course, more precise measurement of $\Upsilon(10860)$’s dielectric decay will tell us
whether the claim of large $S-D$ mixing is correct or not,
if $S-D$ mixing is fully responsible for this suppression.
As can be seen from Tab. 6,
except for $4S-3D$ case, which will be discussed shortly,
we get a rather small mixing angle not only for below threshold states,
but also for highly excited states.
It seems that the coupled channel formalism cannot explain the suppression of $\Gamma_{ee}$.
However, we want to point it out that $S-D$ mixing is not the only way to suppress $\Gamma_{ee}(S)$.
Even though the $S-D$ mixing angles are small,
$\Gamma_{ee}$ can still be suppressed by the $B$ meson continuum.
The suppression due to $B$ meson continuum is not difficult to understand.
If the conventional mesons have non-negligible components of $B$ meson pairs,
these meson pairs have to undergo one more ${}^{3}P_{0}$ vertex before annihilating into $e^{+}e^{-}$ pairs.
Since the Hamiltonian is small compared with Cornell potential,
it is reasonable to discard the contribution of these meson pairs, then $\Gamma_{ee}$ is suppressed.
This suppression is universal both for $S$ and $D$ wave vector m considering esons,
in contrast to $S-D$ mixing, which enhances the $\Gamma_{ee}(D)$.
We take into account both $S-D$ mixing (see Tab. 6) and $B$ meson pair suppression mechanism (see Tab. 4) in this work,
and the results of $\Gamma_{ee}/\Gamma_{ee}(1S)$ are shown in Fig. 6.
From Fig. 6, one can see that the coupled channel results agree well with experiment except for $\Upsilon(11020)$.
The large suppression of $\Gamma_{ee}(\Upsilon(10580))$ deserves more explanation.
From quark model’s perspective, $\Gamma_{ee}(\Upsilon(10580))$ is suggested to be $4S$ or $3D$ state.
If it is $S$ wave dominate, the mixing angle is about $9^{\circ}$,
which is still not big enough to reproduce $\Gamma_{ee}$.
In fact, according to our calculation, the major suppression comes from the $B$ meson pairs.
Even though $\Upsilon(10580)$ is above $B\bar{B}$ threshold,
it is still below and quite close to $B\bar{B}^{*}$ threshold.
As $h_{b}(3P),\chi_{b0}(3P)$ and $\chi_{b0}(3P)$ shown to us in Tab. 4,
closer to the threshold means bigger probabilities of $B$ meson continuum.
Like the authors did in Ref. [7],
we neglect $B\bar{B}$’s probability and work out the probabilities of other channels.
The probabilities of $B\bar{B}^{*}+h.c.,B^{*}\bar{B}^{*},B_{s}\bar{B}_{s},B_{s}\bar{B}_{s}^{*}+h.c.,%
B_{s}^{*}\bar{B}^{*}_{s}$ are
$20.06\%,11.7\%,0.125\%,0.37\%,0.51\%$, respectively, that means $P_{b\bar{b}}=67.2\%$.
So as an estimation, one will get only two thirds of the decay width predicted by the potential model.
For $3D$ dominant states with Breit-Wigner mass $10.731$GeV in Tab. 6,
its large mixing angle may grasp one’s attention.
Because that it is more difficult to generate at $e^{+}e^{-}$ collider compared with $4S$ dominant states,
and its mass is $80$ MeV heavier than $4S$ dominant states,
we do not consider it as $\Upsilon(10580)$.
In the $4S-3D$ mixing case,
because of the oscillation behavior of $\Delta M_{f}$ and $\langle{\psi_{f}}|H_{I}|{\psi_{i}}\rangle$,
there is one more pair of solutions of $M$ in Eq. (24) with GEM.
For $4S$ dominate state, $m_{\text{comp}}=10.673+0.0989i,m_{\text{real}}=10.675,\theta=18.18^{\circ}$,
and for $3D$ dominate state, $m_{\text{comp}}=10.718+0.0441i,m_{\text{real}}=10.7233,\theta=18.4^{\circ}$.
With the same reasons we also do not consider it as $\Upsilon(10580)$.
Another interesting detail of the $B$ meson continuum is
the slightly increased ratio of $\Gamma_{ee}/\Gamma_{ee}(1S)$.
The mixing angles of $5S-4D$ and $6S-5D$ are so small that $\Gamma_{ee}$ barely change,
nevertheless, due to the small $B$ meson continuum component of $\Upsilon(1S)$,
$\Gamma_{ee}(1S)$ will be suppressed about 0.013, as a consequence,
the ratio $\Gamma_{ee}/\Gamma_{ee}(1S)$ becomes slightly larger after considering coupled-channel effects.
For $\Upsilon(10876)$ and $\Upsilon(11020)$,
all the ground state $B$ meson channels are open.
we can no longer deduce the probabilities of these $B$ meson pairs with Eq. (24).
There is no $B$ meson continuum suppression in this work.
This can reproduce the dielectric decay width of $\Upsilon(10876)$ but not $\Upsilon(11020)$.
As shown in Fig. 6,
there is a notable discrepancy between our calculation and experiment on $\Upsilon(11020)$ dielectric decay width.
This issue may come from the two assumptions we use to simplify the calculation.
One is that we only consider the mixing between $6S$ and $5D$.
In fact, with the increase of radial quantum number,
the energy levels of $S$ or $D$ wave will become denser,
so the mixing may exist between several $S$ and $D$ wave states.
Another is the probabilities of excited $B$ meson pairs are neglected.
This may also cause problems.
For example, the $\Upsilon(11020)$ is only $26$ MeV lighter than $m_{B^{*}}+m_{B_{1}}$,
so a large suppression of $P_{b\bar{b}}$ is naturally expected,
causing a large suppression of the dielectric decay width.
It is possible to distinguish $S-D$ mixing and $B$ meson pairs suppression mechanism by the measurement of the radiative decay.
Theoretically, the $E1$ transition can be represented by [67, 54, 68]
$$\Gamma(n^{2S+1}L_{J}\to n^{\prime 2S^{\prime}+1}L^{\prime}_{J^{\prime}}+\gamma%
)=\frac{4}{3}C_{fi}\delta_{SS^{\prime}}e^{2}_{b}\alpha|\langle{f}|r|{i}\rangle%
|^{2}E^{3}_{\gamma},$$
(33)
where $e_{b}=-\frac{1}{3}$. $\alpha$ and $E_{\gamma}$ respectively denote the fine structure constant and the energy of the emitted photon.
$\langle{f}|r|{i}\rangle$ and $C_{fi}$ are represented by
$$\displaystyle\langle{f}|r|{i}\rangle$$
$$\displaystyle=$$
$$\displaystyle\int_{0}^{\infty}R_{f}(r)R_{i}(r)r^{3}dr,$$
(34)
$$\displaystyle C_{fi}$$
$$\displaystyle=$$
$$\displaystyle\max(L,L^{\prime})(2J^{\prime}+1)\left\{\begin{array}[]{ccc}L^{%
\prime}&J^{\prime}&S\\
J&L&1\end{array}\right\}^{2}.$$
(35)
From Eq. (33), we have
$$\displaystyle r_{\gamma}(S)$$
$$\displaystyle:=$$
$$\displaystyle\frac{\Gamma(\Upsilon(S)\to\chi_{b2}(1P)+\gamma)}{\Gamma(\Upsilon%
(S)\to\chi_{b0}(1P)+\gamma)}=5\left(\frac{E_{\gamma 2}}{E_{\gamma 0}}\right)^{%
3},$$
(36)
$$\displaystyle r_{\gamma}(D)$$
$$\displaystyle:=$$
$$\displaystyle\frac{\Gamma(\Upsilon_{1}(D)\to\chi_{b2}(1P)+\gamma)}{\Gamma(%
\Upsilon_{1}(D)\to\chi_{b0}(1P)+\gamma)}=\frac{1}{20}\left(\frac{E_{\gamma 2}}%
{E_{\gamma 0}}\right)^{3},$$
(37)
where $E_{\gamma 2}$ and $E_{\gamma 0}$ respectively represent the photon energy of
$V\to\chi_{b2}+\gamma$ and $V\to\chi_{b0}+\gamma$. ($V$ stands for initial vector state.)
From PDG data [66], we have $r_{\gamma}(2S)=1.91\pm 0.29$ and $r_{\gamma}(3S)=3.82\pm 1.05$,
and the theoretical predictions of Eq. (36) and Eq. (37) are $r_{\gamma}(2S)=1.57$, $r_{\gamma}(3S)=3.6$,
$r_{\gamma}(1D)=0.0157$ and $r_{\gamma}(2D)=0.036$.
So it is reasonable to treat $\Upsilon(2S)$ and $\Upsilon(3S)$ as pure $S$ wave,
and our conclusion of small $S-D$ mixing angle is consistent with experiment for vector bottomonia below threshold.
If the all vector bottomonia observed are $S$ wave dominant,
the small $\Gamma_{ee}$ of $\Upsilon(10580)$ and $\Upsilon(11020)$ naturally requires a large mixing angle
under $S-D$ mixing mechanism, causing a large suppression of $r_{\gamma}(S)$.
On the contrary, $B$ meson continuum suppresses $P_{b\bar{b}}$, leaving the ratio $r_{\gamma}$ unchanged.
Given that the deduced $S-D$ mixing angle is small, we expect a large $r_{\gamma}$.
Unfortunately, the data on the radiative decay widths of
$\Upsilon(10580),\Upsilon(10860)$ and $\Upsilon(11020)$ is still not available so far.
A precise measurement of radiative decay will definitely tell us more about their internal structures.
For states above threshold,
the predicted spectra and decay widths agree not very well with experimental data.
There are two reasons to cause this issue.
As with most work, the meson loops of excited $B$ mesons are ignored,
however, this assumption may be not appropriate for highly excited states.
For example, $\Upsilon(11020)$ is already $20$ MeV heavier than $B_{1}\bar{B}$ threshold.
Second reason comes from the non-relativistic approximation of our bare mass.
In principle, relativistic corrections will be more important with the binding energy goes high,
so both the wave functions and the bare masses will change accordingly.
However, including contributions of excited $B$ mesons
and re-fitting the spectrum and decay widths involves much more work, which lies beyond this work.
It still remains a challenge to reproduce the spectra and dielectric or hadronic decay patterns.
5 Summary
In this paper, we make a thorough and precise calculation of coupled-channel effects in the framework of ${}^{3}P_{0}$ model with GEM
for the bottomonium.
The results of the spectrum, open channel strong decays, probabilities of the $B$ meson continuum,
the $S-D$ mixing and the vector meson’s dielectric decays are explicitly shown.
In order to study the near threshold effects,
we also plot the mass dependence of the mass shift and open channel decay widths for pure $S$ and $D$ wave vector mesons.
For $\Upsilon(4S)$, the decay width of GEM can be two times as large as SHO over a wide energy range,
and the mass shift is around three times as large.
These big deviations indicate that SHO is not a good approximation for near threshold states,
even though the oscillator parameters are carefully selected to reproduce the root mean square radius of the corresponding mesons.
With the consideration of coupled-channel effects, we get small $S-D$ mixing angles except for $\Upsilon(4S)$.
We point it out that, for $S$ wave dominant vector states,
$S-D$ mixing is not the only mechanism to suppress their dielectric decay widths,
$B$ meson continuum can also lead to the suppression.
With this $B\bar{B}$ suppression mechanism at hand,
we still succeed to reproduce the dielectric decays of vector bottomonia except for $\Upsilon(11020)$.
The deviation of the spectrum and decays between our predictions and experimental measurements
may be due to the neglect of excited $B$ meson continuum in coupled-channel effects or
the non-relativistic approximation in the quenched limit.
$S-D$ mixing will cause the suppression of the ratio in Eq. (36) for $S$ wave dominant state,
as a contrary, the $B$ meson continuum does not change this ratio.
We suggest BaBar and Belle to make precise measurements on the radiative decays of
$\Upsilon(10580),\Upsilon(10860)$ and $\Upsilon(11020)$ to distinguish these two effects.
Acknowledgements
The authors are grateful to
Meng Ce, Gui-Jun Ding, David R. Entem, Feng-Kun Guo, Yu. S. Kalashnikova, Bilal Masud, Jia-Lun Ping and E. Santopinto
for useful discussions and suggestions.
This work is supported by the National Natural Science Foundation of China under Grants No. 11261130311 (CRC110 by DFG and NSFC).
M. Naeem Anwar is supported by CAS-TWAS President’s Fellowship for International Ph.D Students.
References
[1]
N. Brambilla et al.
Eur. Phys. J.
C71 (2011) 1534,
arXiv:1010.5827 [hep-ph].
[2]
A. Martin
Phys. Lett.
B100 (1981) 511.
[3]
R. A. Bertlmann and A. Martin
Nucl. Phys.
B168 (1980) 111–136.
[4]
E. Eichten, K. Gottfried, T. Kinoshita, K. D. Lane, and T.-M. Yan
Phys. Rev. D17 (1978) 3090.
[Erratum: Phys. Rev.D21,313(1980)].
[5]
W. Buchmuller and S. H. H. Tye
Phys. Rev. D24 (1981) 132.
[6]
S. Godfrey and N. Isgur
Phys. Rev. D32 (1985) 189–231.
[7]
K. Heikkila, S. Ono, and N. A. Tornqvist
Phys. Rev. D29 (1984) 110.
[Erratum: Phys. Rev.D29,2136(1984)].
[8]
S. Ono and N. A. Tornqvist
Z. Phys. C23
(1984) 59.
[9]
N. A. Tornqvist and P. Zenczykowski
Phys. Rev. D29 (1984) 2139.
[10]
S. Ono, A. I. Sanda, and N. A. Tornqvist
Phys. Rev. D34 (1986) 186.
[11]
S. Ono, A. I. Sanda, N. A. Tornqvist, and J. Lee-Franzini
Phys. Rev. Lett.
55 (1985) 2938.
[12]
L. Micu
Nucl. Phys.
B10 (1969) 521–526.
[13]
A. Le Yaouanc, L. Oliver, O. Pene, and J. C. Raynal
Phys. Rev. D8 (1973) 2223–2234.
[14]
A. Le Yaouanc, L. Oliver, O. Pene, and J. C. Raynal
Phys. Rev. D9 (1974) 1415–1419.
[15]
R. Kokoski and N. Isgur Phys. Rev. D 35 no. 3, (Feb., 1987) 907–933.
[16]
F. E. Close and P. R. Page
Nucl. Phys.
B443 (1995) 233–254,
arXiv:hep-ph/9411301
[hep-ph].
[17]
E. Eichten, K. Gottfried, T. Kinoshita, K. D. Lane, and T.-M. Yan
Phys. Rev. D21 (1980) 203.
[18]
E. S. Ackleh, T. Barnes, and E. S. Swanson
Phys. Rev. D54 (1996) 6811–6829,
arXiv:hep-ph/9604355
[hep-ph].
[19]
T. Barnes, N. Black, and P. R. Page
Phys. Rev.
D68 (2003) 054014,
arXiv:nucl-th/0208072
[nucl-th].
[20]
T. Barnes, F. E. Close, P. R. Page, and E. S. Swanson
Phys. Rev. D55 (1997) 4157–4188,
arXiv:hep-ph/9609339
[hep-ph].
[21]
Y.-c. Yang, Z. Xia, and J. Ping
Phys. Rev.
D81 (2010) 094003,
arXiv:0912.5061 [hep-ph].
[22]
J.-F. Liu and G.-J. Ding
Eur. Phys. J.
C72 (2012) 1981,
arXiv:1105.0855 [hep-ph].
[23]
J. Ferretti and E. Santopinto
Phys. Rev.
D90 no. 9, (2014) 094022,
arXiv:1306.2874 [hep-ph].
[24]
C. Chen, X.-L. Chen, X. Liu, W.-Z. Deng, and S.-L. Zhu
Phys. Rev.
D75 (2007) 094017,
arXiv:0704.0075 [hep-ph].
[25]
C. S. An and B. Saghai Phys. Rev. C85 (2012) 055203,
arXiv:1204.0300 [hep-ph].
[26]
C. S. An, B. C. Metsch, and B. S. Zou
Phys. Rev.
C87 no. 6, (2013) 065207,
arXiv:1304.6046 [hep-ph].
[27]
E. Hiyama, Y. Kino, and M. Kamimura
Prog. Part. Nucl.
Phys. 51 (2003) 223–307.
[28]
J. Segovia, A. M. Yasser, D. R. Entem, and F. Fernandez
Phys. Rev. D78 (2008) 114033.
[29]
J. Segovia, C. Albertus, D. R. Entem, F. Fernandez, E. Hernandez, and M. A.
Perez-Garcia Phys.
Rev. D84 (2011) 094029,
arXiv:1107.4248 [hep-ph].
[30]
J. Segovia, D. R. Entem, and F. Fern¨¢ndez
Phys. Lett.
B715 (2012) 322–327,
arXiv:1205.2215 [hep-ph].
[31]
J. Segovia, D. R. Entem, F. Fernandez, and E. Hernandez
Int. J. Mod. Phys.
E22 (2013) 1330026,
arXiv:1309.6926 [hep-ph].
[32]
J. Segovia, D. R. Entem, and F. Fernandez
Nucl. Phys.
A915 (2013) 125–141,
arXiv:1301.2592 [hep-ph].
[33]
Y. Yang and J. Ping Phys. Rev. D81 (2010) 114025,
arXiv:1004.2444 [hep-ph].
[34]
C. Deng, J. Ping, F. Wang, and T. Goldman
Phys. Rev. D82 (2010) 074001.
[35]
Y. Yang, J. Ping, C. Deng, and H.-S. Zong
J. Phys.
G39 (2012) 105001,
arXiv:1105.5935 [hep-ph].
[36]
C. Deng, J. Ping, H. Wang, P. Zhou, and F. Wang
Phys. Rev.
D86 (2012) 114035,
arXiv:1201.1768 [hep-ph].
[37]
Q.-X. Gao, Y.-C. Yang, and J. Ping
J. Phys.
G39 (2012) 045001.
[38]
C. Deng, J. Ping, Y. Yang, and F. Wang
Phys. Rev.
D86 (2012) 014008,
arXiv:1202.4167 [hep-ph].
[39]
C. Deng, J. Ping, Y. Yang, and F. Wang
Phys. Rev.
D88 no. 7, (2013) 074007,
arXiv:1306.6725 [hep-ph].
[40]
J. Segovia, D. R. Entem, and F. Fernandez
Phys. Rev.
D91 no. 9, (2015) 094020,
arXiv:1502.03827 [hep-ph].
[41]
J. Segovia, P. G. Ortega, D. R. Entem, and F. Fern¨¢ndez
Phys. Rev.
D93 no. 7, (2016) 074027,
arXiv:1601.05093 [hep-ph].
[42]
P. G. Ortega, J. Segovia, D. R. Entem, and F. Fernandez
Phys. Rev.
D81 (2010) 0054023,
arXiv:0907.3997 [hep-ph].
[43]
P. G. Ortega, J. Segovia, D. R. Entem, and F. Fernandez
arXiv:1603.07000 [hep-ph].
[44]
H. Wang, Z. Yan, and J. Ping
Eur. Phys. J.
C75 no. 5, (2015) 196,
arXiv:1412.7068 [hep-ph].
[45]
H. Wang, Y. Yang, and J. Ping
Eur. Phys. J.
A50 (2014) 76.
[46]
X. Chen, Z. Yan, and J. Ping
Int. J. Mod. Phys.
A29 no. 30, (2014) 1450175.
[47]
B.-Q. Li, C. Meng, and K.-T. Chao
Phys. Rev.
D80 (2009) 014012,
arXiv:0904.4068 [hep-ph].
[48]
T. Barnes, S. Godfrey, and E. S. Swanson
Phys. Rev.
D72 (2005) 054026,
arXiv:hep-ph/0505002
[hep-ph].
[49]
B. Numerov Astron.
Nachr. 230 (1927) 359¨C364.
[50]
A. M. Badalian, B. L. G. Bakker, and I. V. Danilkin
Phys. Atom. Nucl.
73 (2010) 138–149,
arXiv:0903.3643 [hep-ph].
[51]
R. Van Royen and V. F. Weisskopf
Nuovo Cim. A50
(1967) 617–645.
[Erratum: Nuovo Cim.A51,583(1967)].
[52]
P. Moxhay and J. L. Rosner
Phys. Rev. D28 (1983) 1132.
[53]
R. Barbieri, E. d’Emilio, G. Curci, and E. Remiddi
Nucl. Phys.
B154 (1979) 535–546.
[54]
J. L. Rosner Phys.
Rev. D64 (2001) 094002,
arXiv:hep-ph/0105327
[hep-ph].
[55]
S. Ono
Z. Phys. C8
(1981) 7–12.
[56]
Yu. S. Kalashnikova
Phys. Rev.
D72 (2005) 034010,
arXiv:hep-ph/0506270
[hep-ph].
[57]
S. Godfrey and K. Moats
Phys. Rev.
D92 no. 5, (2015) 054034,
arXiv:1507.00024 [hep-ph].
[58]
Z.-Y. Zhou and Z. Xiao Eur. Phys. J. A50 no. 10, (2014) 165,
arXiv:1309.1949 [hep-ph].
[59]
H. G. Blundell.
PhD thesis, Carleton U., 1996.
arXiv:hep-ph/9608473
[hep-ph].
[60]
I. V. Danilkin and Yu. A. Simonov
Phys. Rev.
D81 (2010) 074027,
arXiv:0907.1088 [hep-ph].
[61]
W. Roberts and B. Silvestre-Brac
Few-Body Syst. 11 no. 4, (Dec., 1992) 171–193.
[62]
H. B. Schlegel and M. J. Frisch
Int. J. Quantum Chem.
54 (1995) 83.
[63]
T. Barnes and E. S. Swanson
Phys. Rev.
C77 (2008) 055206,
arXiv:0711.2080 [hep-ph].
[64]
S. Ono
Phys. Rev. D23 (1981) 1118.
[65]
V. Baru, C. Hanhart, Yu. S. Kalashnikova, A. E. Kudryavtsev, and A. V.
Nefediev Eur. Phys.
J. A44 (2010) 93–103,
arXiv:1001.0369 [hep-ph].
[66]
Particle Data Group Collaboration, K. A. Olive et al.
Chin. Phys.
C38 (2014) 090001.
[67]
W. Kwong and J. L. Rosner
Phys. Rev. D38 (1988) 279.
[68]
B.-Q. Li and K.-T. Chao
Phys. Rev.
D79 (2009) 094004,
arXiv:0903.5506 [hep-ph]. |
What Makes Kevin Spacey Look Like Kevin Spacey
Supasorn Suwajanakorn Ira Kemelmacher-Shlizerman Steven M. Seitz
University of Washington
Abstract
We reconstruct a controllable model of a person from a large photo collection that captures his or her persona, i.e., physical appearance and behavior. The ability to operate on unstructured photo collections enables modeling a huge number of people, including celebrities and other well photographed people without requiring them to be scanned. Moreover, we show the ability to drive or puppeteer the captured person B using any other video of a different person A. In this scenario, B acts out the role of person A, but retains his/her own personality and character. Our system is based on a novel combination of 3D face reconstruction, tracking, alignment, and multi-texture modeling, applied to the puppeteering problem. We demonstrate convincing results on a large variety of celebrities derived from Internet imagery and video.
1 Introduction
Kevin Spacey has appeared in many acting roles over the years. He’s played characters with a wide variety of temperaments and personalities. Yet, we always recognize him as Kevin Spacey.
Why? Is it his shape? His appearance? The way he moves?
Inspired by Doersch et al’s “What Makes Paris Look Like Paris” [10], who sought to capture the essence of a city, we seek to capture an actor’s persona.
But what defines a persona, how can we capture it, and how will we know if we’ve succeeded?
Conceptually, we want to capture how a person appears in all possible scenarios. In the case of famous actors, there’s a wealth of such data available, in the form of photographic and video (film and interview) footage.
If, by using this data, we could somehow synthesize footage of Kevin Spacey in any number and variety of new roles, and they all look just like Kevin Spacey, then we have arguably succeeded in capturing his persona.
Rather than creating new roles from scratch, which presents all sorts of challenges unrelated to computer vision, we will assume that we have video footage of one person (actor A), and we wish to replace him with actor B, performing the same role.
More specifically, we define the following problem:
Input:
1) all available photos and/or videos of actor B, and 2) a photo collection and a single video V of actor A
Output:
a video V${}^{\prime}$ of actor B performing the same role as actor A in V, but with B’s personality and character.
Figure LABEL:fig:teaser presents example results with Kevin Spacey as actor B, and two other celebrities (Daniel Craig and George Bush) as actor A.
The problem of using one face to drive another is a form of puppetry, which has been explored in the graphics literature
e.g., [20, 25, 18]. The term avatar is also used sometimes to denote this concept of a puppet. What makes our work unique is that we derive the puppet (actor B) automatically from large photo collections.
Our answer to the question of what makes Kevin Spacey look like Kevin Spacey is in the form of a demonstration, i.e., a system that is capable of very convincing renderings of one actor believably mimicking the behavior of another. Making this work well is challenging, as we need to determine what aspects are preserved from actor A’s performance and actor B’s personality. For example, if actor A smiles, should actor B smile in the exact same manner? Or use actor B’s own particular brand of smile?
After a great deal of experimentation, we obtained surprisingly convincing results using the following simple recipe: use actor B’s shape, B’s texture, and A’s motion (adjusted for the geometry of B’s face). Both the shape and texture model are derived from large photo collections of B, and A’s motion is estimated using a 3D optical flow technique.
We emphasize that the novelty of our approach is in our application and system design, not in the individual technical ingredients.
Indeed, there is a large literature on face reconstruction [9, 2, 14, 21] and tracking and alignment techniques e.g., [24, 15]. Without a doubt, the quality of our results is due in large part to the strength of these underlying algorithms from prior work. Nevertheless, the system itself, though seemingly simple in retrospect, is the result of many design iterations, and shows results that no other prior art is capable of. Indeed, this is the first system capable of building puppets automatically from large photo collections and driving them from videos of other people.
2 Related Work
Creating a realistic controllable model of a person’s face is an extremely challenging problem due to the high degree of variability in the human face shape and appearance. Moreover, the shape and texture are highly coupled:
when a person smiles, the 3D mouth and eye shape changes, and wrinkles and creases appear and disappear which changes the texture of the face.
Most research on avatars focuses on animated, non-human faces [18, 24]. The canonical example is that a person drives an animated character, e.g., a dog, with his/her face. The driverâs face can be captured by a webcam or structured light device such as Kinect, the facial expressions are then transferred to a blend shape model that connects the driver and the puppet and then coefficients of the blend shape model are applied to the puppet to create a similar facial expression. Recent techniques can operate in real-time, with a number of commercial systems now available, e.g., faceshift.com (based on [24]), and Adobe Project Animal [1].
The blend shape model that is used for non-human avatars typically capture only large scale expression deformations. Capturing fine details remains an open challenge. Some authors have explored alternatives to blend shape models for non-human characters by learning shape transfer functions [29], and dividing the shape transfer to several layers of detail [27, 18].
Creating a model of a real person, however, is an even more challenging task due to the extreme detail that is required for creation of a realistic face. One way of capturing fine details is by having the person participate in a sequence of lab sessions and use multiple synchronized and calibrated lights and camera rigs [2]. For example, light stages were used for creation of the Benjamin Button movie–to create an avatar of Brad Pitt in an older age [9]. For this, Brad Pitt had to participate in numerous sessions of capturing every possible expression his face can make according to the Facial Action coding system [11]. These expressions were later used to create a personalized blend shape model and transferred to an artist created sculpture of an older version of him. This approach produces amazing results, however, requires active participation of the person and takes months to execute.
Automatic methods for expression transfer of people used multilinear models created from 3D scans [22] or structured light data [6], and transfered differences in expressions of the driver’s mesh to the puppet’s mesh through direct deformation transfer, e.g., [20, 25, 18] or through coefficients that represent different face shapes [24, 22], or driven by speech [7]. These approaches either account only for large scale deformations or do not handle texture changes on the puppet.
This paper is about creating expression transfer in 3D with high detail models and accounting for texture changes that occur due to expression change. Change in texture was previously considered by [19] via image based wrinkles transfer using ratio images, where editing of facial expression used only a single photo [28], face swapping [3, 8], reenactment [12], and age progression [17]. These approaches changed a person’s appearance by transferring changes in texture from another person, and typically focus on a small range of expressions. Finally, [13] showed that it is possible to create a puppetry effect by simply comparing two youtube videos (of the driver and puppet) and finding similarly looking (based on metrics of [16]) pairs of photos. However, the results simply recalled the best matching frame at each time instant, and did not synthesize continuous motion.
In this paper, we show that it is possible to leverage a completely unconstrained photo collection of the person (e.g., Internet photos) in a simple but highly effective way to create texture changes, applied in 3D.
3 Overview
Given a photo collection of the driver and the puppet, our system (illustrated in Figure 1) first reconstructs a rigid 3D average model of the driver and the puppet. Next, given a video of the driver, it estimates 3D flow from each video frame to the driver’s average model. This flow is then transfered onto the average model of the puppet creating a sequence of shapes that move like the driver (Section 4). In the next stage, high detail consistent texture is generated for each frame that accounts for changes in facial expressions (Section 5).
4 3D Dynamic Mesh Creation
By searching for “Kevin Spacey” on Google’s image search we get a large collection of photos that are captured under various poses, expressions, and lightings. In this section, we describe how we estimate an average 3D model of the driver and the puppet, and deform it according to a video or a sequence of photos of the driver. Figure 2 illustrates the shape creation process.
3D Average Model Estimation.
We begin by detection of face and fiducial points (corners of eyes, mouth, nose) in each photo using CMU’s face tracker IntraFace [26]. We next align all the faces to a canonical coordinate frame and reconstruct an average rigid 3D shape of Spacey’s face. For 3D average shape reconstruction we follow [14] with the modification of non-rigidly aligning photos prior to 3D reconstruction. We describe the non-rigid alignment step in Section 5. The same reconstruction pipeline is applied on the driver and the puppet photo collections, resulting in two average 3D rigid models.
Dynamic 3D Model.
Next, we create a dynamic model of the puppet that is deformed according to the driver’s non-rigid motions. For the driver, we are given a video or sequence of photos. The first step is to reconstruct the 3D flow that deforms the driver’s 3D average model to the expression of the driver in every single frame of the input video, using the method of [21]. The geometric transformation is given as a 3D translation field $T:\mathbb{R}^{3}\rightarrow\mathbb{R}^{3}$ applied on a driver’s average shape.
Given a reconstructed mesh at frame $i$ of a driver $M_{D}^{i}(u,v):\mathbb{R}^{2}\rightarrow\mathbb{R}^{3}$ parametrized on an image plane $(u,v)$ from a depth map, and the average mesh over the entire frame sequence $\overline{M}_{D}$, the goal is to transfer the translation field $M_{D}^{i}-\overline{M}_{D}$ to the puppet’s base mesh $M_{P}$ to produce $M_{P}^{i}$. To transfer the deformation, we first establish correspondence between $M_{D}$ and $M_{P}$ through a 2D optical flow algorithm between the puppet’s and driver’s 2D averages from their photo collections.
[15] has shown that we can obtain correspondence between two very different people by projecting one average onto the appearance subspace of the other by this matching illumination, and then run an optical flow algorithm between the resulting projections. With this flow, we can apply the deformation of the driver on the same facial features of the puppet. However, the direct deformation from the driver may not be suitable for the puppet, for example, if their eye sizes are different, the deformation needed to blink will be different. We solve this by scaling the magnitude of the deformation to fit each puppet as follows (Figure 2): Let the deformation vector from the driver at vertex $M_{D}(u,v)$, be $\Delta(u,v)$. We first find the nearest vertex to $M_{D}(u,v)+\Delta(u,v)$ in euclidean distance on the driver mesh, denoted by $M_{D}(s,t)$. Through the flow between $M_{D}$ and $M_{P}$ we computed earlier, we can establish a corresponding pair $(M_{P}(u^{\prime},v^{\prime}),M_{P}(s^{\prime},t^{\prime}))$ on the puppet mesh. The magnitude-adjusted deformation at $M_{P}(u^{\prime},v^{\prime})$ is then computed by $\hat{\Delta}(u,v)(\hat{\Delta}(u,v)\cdot\Delta^{\prime})$ where $\hat{\Delta}=\frac{\Delta}{\|\Delta\|}$ and $\Delta^{\prime}=M_{P}(s^{\prime},t^{\prime})-M_{P}(u^{\prime},v^{\prime})$. In addition, since the flow between the driver and puppet can be noisy around ambiguous, untextured regions, we perform the standard denoising on the term $f(u,v)=(\hat{\Delta}(u,v)\cdot\Delta^{\prime})$ to obtain a regularized field $f^{*}(u,v)$. The particular denoising algorithm we use is ROF denoising with the Huber norm and TV regularization. The final puppet’s mesh is constructed as $M_{P}^{i}(u,v)=M_{P}(u,v)+\hat{\Delta}(u,v)f^{*}(u,v)$.
5 High detail Dynamic Texture Map Creation
In the previous section, we have described how to create a dynamic mesh of the puppet. This section will focus on creation of a dynamic texture. The ultimate set of texture maps should be consistent over time (no flickering, or color change), have the facial details of the puppet, and change according to the driver’s expression, i.e., when the driver is laughing, creases around the mouth and eye wrinkles may appear on the face. For the latter it is particularly important to account for the puppet’s identity–some people may have wrinkles while others won’t. Thus, a naive solution of copying the expression detail from the driver’s face will generally not look realistic. Instead, we leverage a large unconstrained photo collection of the puppet’s face. The key intuition is that to create a texture map of a smile, we can find many more smiles of the person in the collection. While these smiles are captured under different pose, lighting, white balance, etc. they have a common high detail that can be transfered to a new texture map.
Our method works as follows. Given the target expression which is either the configuration of fiducials on the driver’s face (that represents e.g., a rough shape of a smile) or by a reference photo if the driver is the same person as the puppet, we first warp all the photos in the puppet’s collection to the given expression. We then create a multi-scale weighted average that perserves a uniform illumination represented by the lower frequency bands and enhances details in the higher frequency bands. Next we explain each of these in more detail.
Non-rigid warping of the photos.
Each photo in the puppet’s photo collection has 49 fiducial points that we detected. Next we frontalize the face by marking the same fiducials on a generic 3D face model and solve a Perspective-n-Point problem to estimate the 3D pose of the face in the photo. The model is then back-projected to produce a frontal-warp version of each photo. Let the rigid pose-corrected fiducials in each photo be $F^{i}\in\mathbb{R}^{2\times 49}$ and the target fiducials be $F^{T}$. Given two sets of fiducials we estimate a smooth mapping $r$ that transforms the $i$-th photo to the target expression using a smooth variant of thin-plate splines [23] which minimizes the following objective:
$$\min_{r}\sum_{i=1}^{n}\|F^{i}-r(F^{T})\|^{2}+\lambda\iint r_{xx}^{2}+2r_{xy}^{%
2}+r_{yy}^{2}\,\textrm{d}x\,\textrm{d}y$$
(1)
The optimal mapping $r$ satisfying this objective can be represented with a radial basis function $\phi(x)=x^{2}\log x$ and efficiently solved with a linear system of equations [23]. Given the optimal $r$, we can then warp each face photo to the target expression by backward warping. However, this warping relies only on a sparse set of fiducials and the resulting warp field can be too coarse to capture shape changes required to match the target expression which results in a blurry average around eyes and mouth (Figure 5 b). To refine the alignment, we perform an additional dense warping step by exploiting appearance subspaces based on [21, 15]. The idea is to warp all photos to their average, which now has the target expression, through optical flow between illumination-matched pairs. Specifically, let the $i^{\text{th}}$ face image after TPS warping be $I^{i}$, its projection onto the rank 4 appearance subspace of the TPS warped photos be $\hat{I}^{i}$. The refined warp field is then simply the flow from $I^{i}$ to $\hat{I}^{i}$. In the case where a reference photo is available, (Figure 4), we can further warp the entire collection to the reference photo by computing an optical flow that warps $\hat{I}^{T}$ to $I^{T}$, denoted by $F_{{\hat{I}^{T}}\rightarrow I^{T}}$, and compute the final warp field by composing $F_{{I^{i}}\rightarrow\hat{I}^{i}}\circ F_{\hat{I}^{i}\rightarrow{\hat{I}^{T}}}%
\circ F_{{\hat{I}^{T}}\rightarrow I^{T}}=F_{{I^{i}}\rightarrow\hat{I}^{i}}%
\circ F_{{\hat{I}^{T}}\rightarrow I^{T}}$ from the fact that $F_{\hat{I}^{i}\rightarrow{\hat{I}^{T}}}$ is an identity mapping.
Adding high-detail.
Given the set of aligned face photos, we compute a weighted average of the aligned photos where the weights measure the expression similarity to the target expression and the confidence of high-frequency details. We measure expression similarity by $L_{2}$ norm of the difference between the source and target fiducial points, and high-frequency details by the response of a Laplacian filter. A spatially-varying weight $W^{i}_{jk}$ for face $i$ at pixel $(j,k)$ is computed as:
$$W^{i}_{jk}=\exp\left(\frac{-\|F^{T}-F^{i}\|^{2}}{2\sigma^{2}}\right)\cdot(L^{i%
}_{jk})^{\alpha}$$
(2)
where $L^{i}_{jk}$ is the response of a Laplacian filter on face image $i$ at pixel $(j,k)$. An average produced with this weighting scheme produces blending artifacts, for example if high-frequency details from many photos with various illuminations are blended together (Figure 6). To avoid this problem, the blending is done in a multi-scale framework, which blends different image frequency separately. In particular, we construct a Laplacian pyramid for every face photo and compute the weighted average of each level from all the pyramids according to the normalized $W^{i}_{jk}$, then collapse the average pyramid to create a final texture.
With real photo collections, it is rarely practical to assume that the collection spans any expression under every illumination. One problem is that the final texture for different expressions may be averaged from a subset of photos that have different mean illuminations which results in an inconsistency in the overall color or illumination of the texture. This change in the color, however, is low-frequency and is mitigated in the multi-scale framework by preferring a uniform weight distribution in the lower frequency levels of the pyramid. We achieve this is by adding a uniform distribution term, which dominates the distribution in the coarser levels:
$$W^{i}_{jk}=\left(\exp\left(\frac{-\|F^{T}-F^{i}\|^{2}}{2\sigma^{2}}\right)+%
\tau l^{-\beta}\right)\cdot(L^{i}_{jk})^{\alpha}$$
(3)
where $l\in\{0,\ldots,p-1\}$ and $l=0$ represents the coarsest level of a pyramid with $p$ levels, and $\tau$ and $\beta$ are constants.
6 Experiments
In this section, we describe implementation details, runtime, and our results.
Implementation details
In Section 4, the 3D average models for both driver and puppet are reconstructed using [21] which outputs meshes as depth maps with a face size around 194 x 244 (width x height) pixels. To find correspondence between the driver and puppet for deformation transfer purpose, we project the 2D average of the puppet onto the rank-4 appearance subspace of the driver, then compute an optical flow using Ce Liu’s implementation based on Brox et al.[4] and Bruhn et al. [5] with parameters ($\alpha,$ ratio, minWidth, outer-,inner-,SOR-iterations) = $(0.02,0.85,20,4,1,40)$. The ROF denoising algorithm used for adjusting deformation magnitude has only two parameters: The weight constant for the TV regularization which is set to 1, and the Huber epsilon to 0.05. In Section 5, $\lambda$ in TPS warping objective is set to 10, and the dense warping step uses the same optical flow implementation but with $\alpha=0.3$. For Equation 3, $(\alpha,\beta,\tau)=(1,20,1)$, $\sigma$ is typically set to 10 but can vary around $6-12$ for different sizes and qualities of the photo collections (See 7).
Runtime We test our system on a single CPU core of a quad-core Intel [email protected]. Computing a deformed puppet mesh based on the driver sequence takes 0.2 second per frame with 0.13 second spent on denoising. Synthesizing a texture which includes TPS warping, dense warping, and multi-scale pyramid blending takes 0.34 second per frame on average.
Evaluating a puppetry system objectively are extremely hard, and there exists no accuracy metric or benchmark to evalute such system. Ground-truth shapes for evaluating deformation transfer across two people cannot be captured as this requires the puppet person whose shape will be captured, to perform exactly like a driver sequence, which is not possible unless the person is the driver themselves. However, such a setup of self puppetry to evalute the reconstructed geometry requires no deformation transfer and does not evaluate our system. Evaluating the synthesized textures is also qualitative in nature as the average texture we generate cannot be pixel-wise compared to the reference. We provide results and input references for qualitative comparisons and point out areas where further improvement can be done.
From Google Images, we gathered around 200 photos for celebrities and politicians in Figure 7. We generated output puppetry sequences of those people performing various facial expressions driven by YouTube Videos of Kevin Spacey and George W. Bush in the top row. These 3D models are generated by warping an average model of each person with 3D optical flow transfered from the driver (top). So, to render these texture-mapped models, we only synthesize textures in their neutral expressions for the average models but use the target expressions to calculate the blending weights. The identities of these puppets are well-preserved and remain recognizable even when driven by the same source, and the transformation provides plausible output for puppets with different genders, ethnicities, skin colors, or facial features. Facial details are enhanced and change dynamically according to the reference expressions, for example, in creases around the mouth in the last column. We strongly encourage the readers to watch our supplementary videos for these results.
In Figure 4, we show the capability to recreate consistent textures with similar expressions as reference photos in the photo collection. In other words, we are able to “regenerate” each photo in the entire collection so that they appear as if the person is performing different expressions within the same video or photograph captures. Note that each reference here is part of the photo collection used in the averaging process. Texture results for references outside the photo collection is in Figure 8. We compare our method with 3 baseline approaches: 1. A single static average is TPS warped to the reference. This approach produces textures that lack realistic changes such as wrinkles and creases, and shapes that only roughly match the reference (e.g. eyes in column (ii) second row which appear bigger than the reference) because the warping can only rely on sparse fiducial points. 2. A weighted average of the photo collection using identical weights as our method. With this approach, creases can be seen, but the overall texture colors appear inconsistent when there is a variation in the mean color of different high-weighted sets of photos. The overall textures look blurry as there is no alignment done for each photo, and the shapes (eyes in the third row) do not match the reference when the number of similar photos in the collection is small. 3. An improved weighted average with prewarping step which includes TPS and dense warping similar to our pipeline. The prewarping step improves the shapes and the sharpness of the faces, but the textures remain inconsistent. Our method in column (v) produces sharp, realistic, and consistent textures with expression-dependent details and is able to match references with strong illuminations or in black-and-white in Figure 4 (b). Since the references are part of the averaging process, some high-frequency details such as wrinkles are transfered to the output texture. However, the low-frequency details such as shading effects, soft shadow under the nose (in the last example, middle row), or highlights (in the second example, last row) are averaged out in the multi-scale blending and are not part of the final textures.
In Figure 8, we show self-puppetry results where we render output 3D models from [21] with our textures. Similarly to Figure 7, we only synthesize textures in neutral expressions for the average models with blending weights calculated based on the target expressions. The reference photos are excluded from the photo collection in the averaging process. Our textures remain consistent when the references have different lightings and look realistic from various angles. In the fourth reference in the last row, our textures have wrinkles but are less pronounced than the input reference, which is due partly to the fact that the number of photos with wrinkles in the collection is less than 5%.
7 Discussion
The quality of the synthesized textures highy depends on many aspects of the photo collection which include the number and resolutions of the photos, expression and light varations. Since the textures are synthesized based on the assumption that we can find photos with similar expressions, the results will degrade with smaller photo collection (less expression variation). In that situation, the method needs to take into account less-similar photos with a larger standard deviation in Equation 3 resulting in a less pronouced expression. If the standard deviation is kept small, high-frequency details can flicker when the rendered models from video input are played in sequence. Higher resolution photos directly contribute to a sharper average. Our method is less sensitive to having small light variations, in contrast to expression variations, because the shading differences are of low-frequency and can be shared across a wider range of photos in the coarser levels of pyramid.
When a photo collection contains in the order of thousands photos such as when we extract frames from all movies starring a particular actress, additional characteristics of photos can be used to fine-tune the similarity measure in the averaging process such as the directions of lights in the scene to enable a religthing capability or the age of the person (e.g. from a regressor) to synthesize textures at different ages. Only a small modification is needed to implement these changes in our framework. It is also useful to learn the association between the apperance of facial details and facial motions to help with unseen expressions that may share common facial details with already existing photos in the collection.
8 Conclusion
We presented the first system that allows reconstruction of a controllable 3D model of any person from a photo collection toward the goal of capturing persona. The reconstructed model has time-varying, expression-dependent textures and can be controlled by a video sequence of a different person. This capability opens up the ability to create puppets for any photo collection of a person, without requiring them to be scanned. Furthermore, we believe that the insights from this approach (i.e., using actor B’s shape and texture but A’s motion), will help drive future research in this area.
References
[1]
Adobe project animal.
http://blogs.adobe.com/aftereffects/2014/10/weve-created-an-animal-2.html.
[2]
O. Alexander, G. Fyffe, J. Busch, X. Yu, R. Ichikari, A. Jones, P. Debevec,
J. Jimenez, E. Danvoye, B. Antionazzi, et al.
Digital ira: Creating a real-time photoreal digital actor.
In ACM SIGGRAPH 2013 Posters, page 1. ACM, 2013.
[3]
D. Bitouk, N. Kumar, S. Dhillon, P. Belhumeur, and S. K. Nayar.
Face swapping: automatically replacing faces in photographs.
In ACM Transactions on Graphics (TOG), volume 27, page 39. ACM,
2008.
[4]
T. Brox, A. Bruhn, N. Papenberg, and J. Weickert.
High accuracy optical flow estimation based on a theory for warping.
In Computer Vision-ECCV 2004, pages 25–36. Springer, 2004.
[5]
A. Bruhn, J. Weickert, and C. Schnörr.
Lucas/kanade meets horn/schunck: Combining local and global optic
flow methods.
International Journal of Computer Vision, 61(3):211–231, 2005.
[6]
C. Cao, Y. Weng, S. Zhou, Y. Tong, and K. Zhou.
Facewarehouse: a 3d facial expression database for visual computing.
Visualization and Computer Graphics, IEEE Transactions on,
20(3):413–425, 2014.
[7]
Y. Cao, W. C. Tien, P. Faloutsos, and F. Pighin.
Expressive speech-driven facial animation.
ACM Transactions on Graphics (TOG), 24(4):1283–1302, 2005.
[8]
K. Dale, K. Sunkavalli, M. K. Johnson, D. Vlasic, W. Matusik, and H. Pfister.
Video face replacement.
ACM Transactions on Graphics (TOG), 30(6):130, 2011.
[9]
P. Debevec.
The light stages and their applications to photoreal digital actors.
SIGGRAPH Asia Technical Briefs, 2012.
[10]
C. Doersch, S. Singh, A. Gupta, J. Sivic, and A. A. Efros.
What makes paris look like paris?
ACM Trans. Graph., 31(4):101:1–101:9, July 2012.
[11]
P. Ekman and W. V. Friesen.
Facial action coding system.
1977.
[12]
P. Garrido, L. Valgaerts, O. Rehmsen, T. Thormaehlen, P. Perez, and
C. Theobalt.
Automatic face reenactment.
In Computer Vision and Pattern Recognition (CVPR), 2014 IEEE
Conference on, pages 4217–4224. IEEE, 2014.
[13]
I. Kemelmacher-Shlizerman, A. Sankar, E. Shechtman, and S. M. Seitz.
Being john malkovich.
In Computer Vision–ECCV 2010, pages 341–353. Springer, 2010.
[14]
I. Kemelmacher-Shlizerman and S. M. Seitz.
Face reconstruction in the wild.
In Computer Vision (ICCV), 2011 IEEE International Conference
on, pages 1746–1753. IEEE, 2011.
[15]
I. Kemelmacher-Shlizerman and S. M. Seitz.
Collection flow.
In Computer Vision and Pattern Recognition (CVPR), 2012 IEEE
Conference on, pages 1792–1799. IEEE, 2012.
[16]
I. Kemelmacher-Shlizerman, E. Shechtman, R. Garg, and S. M. Seitz.
Exploring photobios.
In ACM Transactions on Graphics (TOG), volume 30, page 61. ACM,
2011.
[17]
I. Kemelmacher-Shlizerman, S. Suwajanakorn, and S. M. Seitz.
Illumination-aware age progression.
In Computer Vision and Pattern Recognition (CVPR), 2014 IEEE
Conference on, pages 3334–3341. IEEE, 2014.
[18]
N. Kholgade, I. Matthews, and Y. Sheikh.
Content retargeting using parameter-parallel facial layers.
In Proceedings of the 2011 ACM SIGGRAPH/Eurographics Symposium
on Computer Animation, pages 195–204. ACM, 2011.
[19]
Z. Liu, Y. Shan, and Z. Zhang.
Expressive expression mapping with ratio images.
In Proceedings of the 28th annual conference on Computer
graphics and interactive techniques, pages 271–276. ACM, 2001.
[20]
R. W. Sumner and J. Popović.
Deformation transfer for triangle meshes.
ACM Transactions on Graphics (TOG), 23(3):399–405, 2004.
[21]
S. Suwajanakorn, I. Kemelmacher-Shlizerman, and S. M. Seitz.
Total moving face reconstruction.
In Computer Vision–ECCV 2014, pages 796–812. Springer, 2014.
[22]
D. Vlasic, M. Brand, H. Pfister, and J. Popović.
Face transfer with multilinear models.
In ACM Transactions on Graphics (TOG), volume 24, pages
426–433. ACM, 2005.
[23]
G. Wahba.
Spline models for observational data, volume 59.
Siam, 1990.
[24]
T. Weise, S. Bouaziz, H. Li, and M. Pauly.
Realtime performance-based facial animation.
In ACM Transactions on Graphics (TOG), volume 30, page 77. ACM,
2011.
[25]
T. Weise, H. Li, L. Van Gool, and M. Pauly.
Face/off: Live facial puppetry.
In Proceedings of the 2009 ACM SIGGRAPH/Eurographics Symposium
on Computer Animation, pages 7–16. ACM, 2009.
[26]
X. Xiong and F. De la Torre.
Supervised descent method and its applications to face alignment.
In Computer Vision and Pattern Recognition (CVPR), 2013 IEEE
Conference on, pages 532–539. IEEE, 2013.
[27]
F. Xu, J. Chai, Y. Liu, and X. Tong.
Controllable high-fidelity facial performance transfer.
ACM Transactions on Graphics (TOG), 33(4):42, 2014.
[28]
F. Yang, L. Bourdev, E. Shechtman, J. Wang, and D. Metaxas.
Facial expression editing in video using a temporally-smooth
factorization.
In Computer Vision and Pattern Recognition (CVPR), 2012 IEEE
Conference on, pages 861–868. IEEE, 2012.
[29]
M. D. Zeiler, G. W. Taylor, L. Sigal, I. Matthews, and R. Fergus.
Facial expression transfer with input-output temporal restricted
boltzmann machines.
In Advances in Neural Information Processing Systems, pages
1629–1637, 2011. |
Free homotopy algebras
Martin Markl
[email protected]
Mathematical Institute of the Academy
Žitná 25
115 67 Praha 1
The Czech Republic
Abstract
An explicit description of free strongly homotopy associative and
free strongly homotopy Lie algebras is given. A variant of the
Poincaré-Birkhoff-Witt theorem for the universal enveloping
A($m$)-algebra of a strongly homotopy Lie algebra is formulated.
keywords: Free algebra, A($m$)-algebra, L($m$)-algebra, PBW theorem
: 08B20
††thanks: This work was supported by the
grant GA ČR 201/96/0310 and MŠMT ME603.
Introduction
This note was originated many years ago as my reaction to questions of
several people how free strongly homotopy algebras can be described
and what can be said about the structure of the universal enveloping
A($m$)-algebra of an L($m$)-algebra constructed
in [3], and then circulated as a “personal
communication.” I must honestly admit that it contains no really deep
result and that everything I did was that I expanded definitions and
formulated a couple of statements with more or less obvious proofs.
A($m$)-algebras and their strict
homomorphisms [5, pages 147–148] form an equationally
given algebraic category ${\tt A}(m)$. It follows from general theory
that the forgetful functor to the category ${\tt gVect}$ of graded
vector spaces, $\Box:{\tt A}(m)\to{\tt gVect}$, has a left adjoint
${\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}:{\tt gVect}\to{\tt A}(m)$. Given a graded vector space $X\in{\tt gVect}$, the object ${\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}(X)\in{\tt A}(m)$ is the free
A($m$)-algebra on the graded vector space $X$. We will also, for
$n<m$, consider forgetful functors $\Box:{\tt A}(m)\to{\tt A}(n)$
and their left adjoints ${\raisebox{7.2pt}{\tiny$m\hskip-2.27622pt:\hskip-1.707165ptn$}\hskip-9.0pt{%
\mathfrak{A}}\hskip 2.0pt}:{\tt A}(n)\to{\tt A}(m)$; here
the case $n=1$ is particularly important, because ${\raisebox{7.2pt}{\tiny$m\hskip-2.27622pt:\hskip-1.707165pt1$}\hskip-9.0pt{%
\mathfrak{A}}\hskip 2.0pt}:{\tt dgVect}\to\mbox{{\tt A}($m$)}$ describes the free A($m$)-algebra
generated by a differential graded vector space.
We believe there is no need to emphasize the important rôle
of free objects in mathematics. Each A($m$)-algebra is a quotient of a
free one and free A($m$)-algebras were used in our definition of the
universal enveloping algebra of a strongly homotopy
Lie algebra [3, page 2154]. It
could be useful to have a concrete description of these free
algebras, as explicit as, for example,
the description of free associative algebras by tensor algebras.
This brief note gives such a description in terms of planar trees.
Replacing planar trees by
non-planar ones, one can equally easy represent also free
strongly homotopy Lie algebras.
Surprisingly, these free algebras are
simpler objects that their strict counterparts
and admit a nice linear basis
(Section 3), while
free (strict) Lie algebras are very complicated
objects
(see, for example, [9, Chapter IV]).
This might be explained by the fact that axioms of strongly
homotopy algebras are certain resolutions of
strict axioms, thus they are more amenable. They
manifest some properties of distributive
laws [6] – they are ‘directed,’ therefore one can
obtain unique representatives of elements of the corresponding
operads in terms of admissible trees, like
in Proposition 1.2, with obvious implications for the
structure of free algebras.
In the first section of this note we describe the operad ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-8.0pt\underline{\mathcal{A}}}$ for
A($m$)-algebras111underlining indicates that it is a
non-$\Sigma$ operad and then, in Section 2, free A($m$)-algebras. In
the third section we modify these constructions to free strongly
homotopy Lie algebras. In the last section we prove a
Poincaré-Birkhoff-Witt-type theorem for strongly homotopy Lie
algebras and formulate a problem related to the structure of the
category of modules. In the appendix we then recall the language of
trees used throughout the paper.
1 Operad for A($m$)-algebras
In this note we assume a certain familiarity with operads, a good
reference book for this subject is [8].
For simplicity we suppose the ground ring ${\bf k}$ to be a field of
characteristic zero, though the results of this and the following
section remain true over an arbitrary ring.
Let $m$ be a natural number or $\infty$. Recall
(see [10] for the original definition,
but we use the sign convention of [5])
that an A($m$)-algebra
is a graded ${\bf k}$-module $A$ together with a set $\{\mu_{k};1\leq k\leq m,\ k<\infty\}$
of degree $k-2$-linear maps, $\mu_{k}:\bigotimes^{k}A\to A$, such that
$$\sum_{\lambda=0}^{n-1}\sum_{k=1}^{n-\lambda}(-1)^{\omega}\cdot\mu_{n-k+1}(a_{1%
},\dots,a_{\lambda},\mu_{k}(a_{\lambda+1},\dots,a_{\lambda+k}),a_{\lambda+k+1}%
,\dots,a_{n})=0$$
for all homogeneous $a_{1},\ldots,a_{n}\in A$, $n\leq m$, where the sign
is given as
$$\omega:={k+\lambda+k\lambda+k(|a_{1}|+\dots+|a_{\lambda}|)}.$$
It easily follows from the above axiom that $\mu_{1}:A\to A$ is a
degree $-1$ differential. This operation will play a special rôle
and we denote it by $\partial:=\mu_{1}$. A($m$)-algebras are algebras
over a certain non-$\Sigma$ operad ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-8.0pt\underline{\mathcal{A}}}=\{%
{\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-8.0pt\underline{\mathcal{A}}}(n)%
\}_{n\geq 1}$ in the monoidal category of graded vector spaces. This operad
can be constructed as follows.
Let ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.2pt\underline{\mathcal{F}}}=\{%
{\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.2pt\underline{\mathcal{F}}}(n)%
\}_{n\geq 1}$ be the free graded non-$\Sigma$
operad $\underline{\Gamma}(\xi_{1},\xi_{2},\ldots,\xi_{m})$ generated by operations
$\xi_{i}$, $l\leq i\leq m$, where each $\xi_{i}$ has arity $i$ and degree
$i-2$. Alternatively, let ${\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}(n)$ be the set of all connected
directed planar trees with $n$ input edges, all of whose vertices have
at most $m$ input edges; we admit also ‘unary’ vertices with only one
input edge.
Let ${\rm Vert}(T)$ denote the set of vertices of
a tree $T\in{\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}(n)$. For $v\in{\rm Vert}(T)$, let ${\rm val}(v)$ be the number of
input edges of $v$. We assign to each $T\in{\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}(n)$ a degree
$\deg(T)$ by the formula
$$\deg(T):=\sum_{v\in{\rm Vert}(T)}({\rm val}(v)-2).$$
Then ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.2pt\underline{\mathcal{F}}}(n)$ can be naturally identified with the ${\bf k}$-linear
span of the set ${\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}(n)$,
$${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.2pt\underline{\mathcal{F}}}(n)%
={\rm Span}_{{\bf k}}({\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}%
(n)),$$
(1)
see [8, Section II.1.9]. The operadic
structure is, in this language, defined by grafting the corresponding
trees. This means that, for $T\in{\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}(a)$ and $S\in{\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}(b)$,
the operadic $i$-th circle
product [8, Definition II.1.16] $T\circ_{i}S$ is the tree obtained by grafting the tree $S$ at the $i$-th
input edge of $T$. Similarly, for $T\in{\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}(l)$ and $S_{j}\in{\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}(k_{j})$, $1\leq j\leq l$, the operadic
composition [8, Definition II.1.4]
$\gamma(T;S_{1},\ldots,S_{l})=T(S_{1},\ldots,S_{l})\in{\raisebox{7.2pt}{\tiny$m%
$}\hskip-5.8pt{\mbox{$\tt T$}}}(k_{1}+\cdots+k_{l})$ is the tree obtained by grafting $S_{j}$ at the
$j$-th input edge of $T$, for each $1\leq j\leq l$.
The operad ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-8.0pt\underline{\mathcal{A}}}$ is then defined as ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.2pt\underline{\mathcal{F}}}/{%
\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-6.0pt\underline{\mathcal{I}}}$, where ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-6.0pt\underline{\mathcal{I}}}$
is the operadic ideal generated by the elements
$$\Phi_{n}:=\sum_{\lambda=0}^{n-1}\sum_{k=1}^{n-\lambda}(-1)^{\lambda(k+1)+k}\xi%
_{n-k+1}\circ_{\lambda+1}\xi_{k}\in{\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}%
\hskip-7.2pt\underline{\mathcal{F}}}(n),\ n\leq m.$$
(2)
In the tree language of (1), the generators $\xi_{j}$
are represented by $j$-corollas (denoted by the same symbol)
$\xi_{j}\in{\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}(j)$; recall that the $j$-corolla is the unique
tree having exactly one vertex. Relations (2) then represent
certain linear combinations of trees, indicated
in Figure 1.
Elements of ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-8.0pt\underline{\mathcal{A}}}(n)$ are equivalence classes of linear
combinations of trees from ${\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}(n)$, modulo the relations of
Figure 1. Surprisingly, there exists, for any $n\geq 1$, a
very natural basis for the ${\bf k}$-vector space ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-8.0pt\underline{\mathcal{A}}}(n)$.
Definition 1.1.
We say that a tree $T\in{\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}(n)$ is admissible if all its unary
vertices (i.e. vertices with just one input edge) are input vertices
of the tree $T$.
Let us denote by ${\hskip 8.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-14.0pt{{\rm Adm}}}(n)\subset{%
\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}(n)$ the subset consisting of
admissible trees. Then the following proposition holds.
Proposition 1.2.
Admissible $n$-trees form, for $n\geq 1$ a linear basis of ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-8.0pt\underline{\mathcal{A}}}(n)$,
$${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-8.0pt\underline{\mathcal{A}}}(n)%
\cong{\rm Span}_{{\bf k}}({\hskip 8.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-14.0pt%
{{\rm Adm}}}(n)).$$
The proposition follows from the following lemma.
Lemma 1.3.
There exists a unique map
$\Omega:{\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}(n)\to{\rm Span%
}_{{\bf k}}({\hskip 8.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-14.0pt{{\rm Adm}}}(n))$
such that, for each $T\in{\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}(n)$,
$$\Omega(T)\equiv T\mbox{ modulo the relations of Figure~{}\ref{picture}}.$$
Proof.
If $T$ is admissible, we put
$\Omega(T):=T$. If $T$ is not admissible, it contains a subtree of
the form $\xi_{1}\circ_{1}\xi_{j}$, with some $1\leq j\leq m$, or, pictorially,
$\bullet$$\bullet$$\cdots$$\xi_{1}$$\xi_{j}$
We may also assume that $j\geq 2$, otherwise $T\equiv 0$ modulo ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-6.0pt\underline{\mathcal{I}}}$.
Using the relations of Figure 1, we can replace the above
subtree
by a linear combination of trees of the form
$\cdots$$\cdots$$\cdots$$\bullet$$\bullet$$\xi_{s}$$\xi_{t}$
with $s\geq 2$. This enables us to move,
by local replacements, the vertex $\xi_{1}$
towards the inputs of the tree. Repeating this process sufficiently
many times, we get the requisite presentation $\Omega(T)$. It is
easy to see that this presentation is unique, compare also similar
arguments in the proof of [6, Theorem 2.3].
∎
Proposition 1.2 immediately implies that
$$\dim_{{\bf k}}({\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-8.0pt\underline{%
\mathcal{A}}}(n))=2^{n}\cdot b^{m}_{n},\ \mbox{ for }m\leq 2,\ n\leq 1.$$
(3)
where $b^{m}_{n}$ is the number of all planar rooted reduced (i.e. without
unary vertices) $n$-trees, all of whose vertices have arities $\leq m$.
Clearly, $b^{m}_{n}$ equals the number of cells of the $m$-skeleton of
Stasheff’s associahedron $K_{n}$, compare also Example 2.2. Another
kind of information about the size of the non-$\Sigma$-operad ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-8.0pt\underline{\mathcal{A}}}$
is given in:
Proposition 1.4.
Let us consider the generating function
$${\hskip 3.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\varphi}}(t):=\sum_{n\geq 1%
}\dim_{{\bf k}}({\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-8.0pt\underline{%
\mathcal{A}}}(n))t^{n}.$$
Then ${\hskip 3.0pt\raisebox{7.2pt}{\tiny$1$}\hskip-7.0pt{\varphi}}(t)=2t$ while, for $m\geq 2$, $\varphi(t)={\hskip 3.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\varphi}}(t)$ solves
the equation
$$\varphi(t)-2t=\varphi^{2}(t)(1+\varphi(t)+\cdots+\varphi^{m-2}(t)).$$
(4)
Proof.
Formula (4) is probably known, though we were not
able to find a reference in the literature. We give a proof that can
be easily modified to obtain the analogous
formula (10) for L($m$)-algebras.
Let $d_{n}^{m}:=\mbox{card}({\hskip 8.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-14.0pt{{%
\rm Adm}}}(n))$. By Proposition 1.2,
$d_{n}^{m}:=\dim_{{\bf k}}({\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-8.0pt%
\underline{\mathcal{A}}}(n))$. Let us denote, just in this proof,
by ${\tt T}_{n}$ the set of all planar reduced $n$-trees all of whose
vertices have at most $m$ input edges. Let $r_{n}:=\mbox{card}({\tt T}_{n})$. We claim that, for any $n\geq 2$,
$$r_{n}=\sum_{i_{1}+i_{2}=n}\hskip-4.0ptr_{i_{1}}r_{i_{2}}+\hskip-4.0pt\sum_{i_{%
1}+i_{2}+i_{3}=n}\hskip-6.0ptr_{i_{1}}r_{i_{2}}r_{i_{3}}+\cdots+\hskip-7.0pt%
\sum_{i_{1}+\cdots+i_{m}=n}\hskip-7.0ptr_{i_{1}}\cdots r_{i_{m}}$$
(5)
and
$$d_{n}^{m}=2^{n}r_{n}.$$
(6)
The first equation follows from the decomposition
$${\tt T}_{n}={\tt T}_{n,2}\sqcup{\tt T}_{n,3}\sqcup\cdots\sqcup{\tt T}_{n,m},$$
where ${\tt T}_{n,i}\subset{\tt T}_{n}$ is the subset of trees whose root
has $i$ input edges. The second equation follows from the
observation that each admissible tree $T\in{\hskip 8.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-14.0pt{{\rm Adm}}}(n)$ can be obtained
from a tree $S\in{\tt T}_{n}$ by grafting terminal unary trees at the
input legs of $S$; there are $2^{n}$ ways to do this. If we denote
$\psi(t):=\sum_{n\geq 1}r_{n}t^{n}$, then (5) implies
$$\psi(t)-t=\psi^{2}(t)(1+\cdots+\psi^{m-2}(t)),$$
while (6) implies that $\varphi(t)=\psi(2t)$. This finishes
the proof.
∎
Example 1.5.
For $m=\infty$, equation (4) can be rewritten as
$$2\varphi^{2}(t)-\varphi(t)(1+2t)+2t=0.$$
For $m=2$, (4)
gives $\varphi(t)-2t=\varphi^{2}(t)$,
therefore
$$\varphi(t)=\frac{1-\sqrt{(1-8t)}}{2}.$$
A little exercise in Taylor expansions results in
$$d^{2}_{n}=\frac{2^{2n-1}(2n-3)!!}{n!}\hskip 10.0pt\mbox{ for }n\geq 2,$$
where as usual
$$(2n-3)!!=1\cdot 3\cdot 5\cdots(2n-3).$$
2 Free A($m$)-algebras
The following theorem relies on the notation introduced in the previous
section. Recall namely that
${\hskip 8.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-14.0pt{{\rm Adm}}}(n)$ denotes the set of admissible $n$-trees in
the sense of Definition 1.1, ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-8.0pt\underline{\mathcal{A}}}(n)%
\cong{\rm Span}_{{\bf k}}({\hskip 8.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-14.0pt%
{{\rm Adm}}}(n))$ and
$\Omega$ is the map of Lemma 1.3.
Recall also that we denoted by $\xi_{j}$ the $j$-corolla.
If $T_{1},\ldots,T_{j}$ are trees, then
$\xi_{j}(T_{1},\ldots,T_{j})$ is the tree obtained by
grafting the tree
$T_{i}$ at the $i$-th input edge of the tree $\xi_{j}$, $1\leq i\leq j$.
Theorem 2.1.
Let $X$ be a graded vector space. Then the free A($m$)-algebra $\ {\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}(X)$
on $X$ can be described as ${\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}(X)=\bigoplus_{n\geq 1}{%
\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}{}^{n}(X)$, with
$${\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}{}^{n}(X):={\hskip 2.0pt%
\raisebox{7.2pt}{\tiny$m$}\hskip-8.0pt\underline{\mathcal{A}}}(n)\otimes X^{%
\otimes n}.$$
The structure operations $\mu_{1},\ldots,\mu_{m}$ are defined as follows. Let
$\vec{v}_{j}\in X^{\otimes k_{j}}$ and $T_{j}\in{\hskip 8.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-14.0pt{{\rm Adm}}}(k_{j})$, $1\leq j\leq k$. Then
$$\mu_{k}(T_{1}\otimes\vec{v}_{1},\ldots,T_{k}\otimes\vec{v}_{k}):=\Omega(\xi_{k%
}(T_{1},\ldots,T_{k}))\otimes(\vec{v}_{1}\otimes\cdots\otimes\vec{v}_{k}).$$
Proof.
It follows from general
theory [8, Section II.1.4] that the free
$\underline{\mathcal{P}}$-algebra $F_{\underline{\mathcal{P}}}(X)$
on a graded vector space $X$, where $\underline{\mathcal{P}}$ is a
non-$\Sigma$ operad, is given by $F_{\underline{\mathcal{P}}}(X)=\bigoplus_{n\geq 1}F_{\underline{\mathcal{P}}}^%
{n}(X)$, with
$F_{\underline{\mathcal{P}}}^{n}(X):=\underline{\mathcal{P}}(n)\otimes X^{%
\otimes n}$, with the algebra structure induced
from the operad structure of $\underline{\mathcal{P}}$. Our theorem
is then obtained by taking $\underline{\mathcal{P}}:={\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-8.0pt%
\underline{\mathcal{A}}}$ and
applying the results on the structure of this operad proved in the
previous section.
∎
Observe that the free algebra ${\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}(X)$ is bigraded. The internal
grading is the grading of the underlying vector space, while
the external grading is given by the decomposition
$${\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}(X)=\bigoplus_{n\geq 1}{%
\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}{}^{n}(X).$$
Because all main structures considered in this note (vector
spaces, A($m$)-algebras, L($m$)-algebras, etc.) have an implicit internal
grading, we will usually mean by saying that an object is graded the
presence of an external grading. We believe that the actual meaning
will always be clear from the context.
Example 2.2.
There exists a simple way to encode elements of
${\hskip 8.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-14.0pt{{\rm Adm}}}(n)$. Let ${{\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-8.0ptB}}(n)$ be the set of all meaningful bracketings
of strings $\epsilon_{1}\epsilon_{2}\cdots\epsilon_{n}$, where
$\epsilon_{i}\in\{\circ,\bullet\}$ and no pair of brackets
encompasses more than $m$ terms. A moment’s reflection shows that
${\hskip 8.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-14.0pt{{\rm Adm}}}(n)\cong{{%
\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-8.0ptB}}(n)$, $n\geq 1$. Thus
$$\displaystyle{ \raisebox{7.2pt}{\tiny$m$}\hskip-14.0pt{{\rm Adm}}}(1)$$
$$\displaystyle=$$
$$\displaystyle\{\circ,\bullet\},\ m\geq 1,$$
$$\displaystyle{ \raisebox{7.2pt}{\tiny$1$}\hskip-14.0pt{{\rm Adm}}}(2)$$
$$\displaystyle=$$
$$\displaystyle\emptyset,\ {\hskip 8.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-14.0pt{%
{\rm Adm}}}(2)=\{(\circ\circ),(\circ\bullet),(\bullet\circ),(\bullet\bullet)\}%
,\ m\geq 2,$$
$$\displaystyle{ \raisebox{7.2pt}{\tiny$1$}\hskip-14.0pt{{\rm Adm}}}(3)$$
$$\displaystyle=$$
$$\displaystyle\emptyset,$$
$$\displaystyle{ \raisebox{7.2pt}{\tiny$2$}\hskip-14.0pt{{\rm Adm}}}(3)$$
$$\displaystyle=$$
$$\displaystyle\{((\circ\circ)\circ),((\bullet\circ)\circ),((\bullet\bullet)%
\circ),\ldots,(\circ(\circ\circ)),(\bullet(\circ\circ)),(\bullet(\bullet\circ)%
),\ldots\},$$
$$\displaystyle{ \raisebox{7.2pt}{\tiny$m$}\hskip-14.0pt{{\rm Adm}}}(3)$$
$$\displaystyle=$$
$$\displaystyle {\hskip 8.0pt\raisebox{7.2pt}{\tiny$2$}\hskip-14.0pt{{\rm Adm}}}%
(3)\cup\{(\circ\circ\circ),(\bullet\circ\circ),(\bullet\bullet\circ),\ldots\},%
\ m\geq 3,\mbox{ etc.}$$
We conclude that ${\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}(X)$ is spanned by all meaningful bracketings of
strings of “variables” $x$ and $\partial x$, $\deg(\partial x)=\deg(x)-1$,
$x\in X$, with no pair of brackets encompassing more than $m$ terms.
Thus ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$1$}\hskip-5.0pt{\mathfrak{A}}}(X)={\hskip 2%
.0pt\raisebox{7.2pt}{\tiny$1$}\hskip-5.0pt{\mathfrak{A}}}{}^{1}(X)$ is the free differential
complex generated by $X$, ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$1$}\hskip-5.0pt{\mathfrak{A}}}(X)\cong X%
\oplus\partial X$. Similarly,
$$\displaystyle{\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}{}^{2}(X)$$
$$\displaystyle\cong$$
$$\displaystyle(X\otimes X)\oplus(\partial X\otimes X)\oplus(X\otimes\partial X)%
\oplus(\partial X\otimes\partial X),\ \mbox{for $m\geq 2$,}$$
$$\displaystyle{ \raisebox{7.2pt}{\tiny$2$}\hskip-5.0pt{\mathfrak{A}}}{}^{3}(X)$$
$$\displaystyle\cong$$
$$\displaystyle((X\otimes X)\otimes X)\oplus((\partial X\otimes X)\otimes X)%
\oplus((X\otimes\partial X)\otimes X)$$
$$\displaystyle \oplus\ ((X\otimes X)\otimes\partial X)\oplus((\partial X%
\otimes\partial X)\otimes X)\oplus\ ((\partial X\otimes X)\otimes\partial X)$$
$$\displaystyle \oplus\ ((X\otimes\partial X)\otimes\partial X)\oplus((%
\partial X\otimes\partial X)\otimes\partial X)\oplus(X\otimes(X\otimes X))$$
$$\displaystyle \oplus\ (\partial X\otimes(X\otimes X))\oplus(X\otimes(%
\partial X\otimes X))\oplus(X\otimes(X\otimes\partial X))$$
$$\displaystyle \oplus\ (\partial X\otimes(\partial X\otimes X))\oplus(%
\partial X\otimes(X\otimes\partial X))\oplus(X\otimes(\partial X\otimes%
\partial X))$$
$$\displaystyle \oplus\ (\partial X\otimes(\partial X\otimes\partial X)),$$
$$\displaystyle{{\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}}{}^{3}(X)$$
$$\displaystyle\cong$$
$$\displaystyle{ \raisebox{7.2pt}{\tiny$2$}\hskip-5.0pt{\mathfrak{A}}}{}^{3}(X)%
\oplus(X{\otimes}X{\otimes}X)\oplus(\partial X{\otimes}X{\otimes}X)\oplus(X{%
\otimes}\partial X{\otimes}X)\oplus(X{\otimes}X{\otimes}\partial X)$$
$$\displaystyle \oplus\ (\partial X{\otimes}\partial X{\otimes}X)\oplus(%
\partial X{\otimes}X{\otimes}\partial X)\oplus(X{\otimes}\partial X{\otimes}%
\partial X)$$
$$\displaystyle \oplus\ (\partial X{\otimes}\partial X{\otimes}\partial X),%
\ \mbox{for $m\geq 3$, etc.}$$
We see that
$$\dim(\hskip 2.0pt{\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}{}^{n}(%
X))=(2\dim(X))^{n}\cdot b^{m}_{n},\ n\geq 1,\ m\geq 2,$$
where $b^{m}_{n}$ is as in (3).
Unital A($m$)-algebras.
Recall [5, page 148] that an A($m$)-algebra
$A=(A,\{\mu_{n}\})$ is unital if there exists an element $1=1_{A}\in A$ such that $\mu_{2}(a,1_{A})=\mu_{2}(1_{A},a)=a$, for all $a\in A$, and that $\mu_{n}(a_{1},\dots,a_{i-1},1_{A},a_{i+1},\dots,a_{n})=0$, for all $n\not=2$ and $1\leq i\leq n$. The free unital A($m$)-algebra
can be constructed as ${\rm Span}_{{\bf k}}(1)\oplus{\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak%
{A}}}(X)$, with the structure
operations defined in an obvious way.
Important modifications. One may as well consider, for any $n<m$, the forgetful functor $\Box:\mbox{{\tt A}($m$)}\to\mbox{{\tt A}($n$)}$.
This functor has a left adjoint ${\raisebox{7.2pt}{\tiny$m\hskip-2.27622pt:\hskip-1.707165ptn$}\hskip-9.0pt{%
\mathfrak{A}}\hskip 2.0pt}:\mbox{{\tt A}($n$)}\to\mbox{{\tt A}($m$)}$ which can be described as follows. Given an A($n$)-algebra
$(V,m_{1},m_{2},\ldots,m_{n})$, consider the free A($m$)-algebra ${\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}(V)$ on
the graded vector space $V$, and let $\mu_{1},\ldots,\mu_{m}$ be the
structure operations of ${\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}(V)$. Then
$${\raisebox{7.2pt}{\tiny$m\hskip-2.27622pt:\hskip-1.707165ptn$}\hskip-9.0pt{%
\mathfrak{A}}\hskip 2.0pt}(V,m_{1},m_{2},\ldots,m_{n})\cong{\raisebox{7.2pt}{%
\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}(V)/(\mu_{i}=m_{i},\ i\leq n).$$
A very important special case is $n=1$ when we obtain a left adjoint
${\raisebox{7.2pt}{\tiny$m\hskip-2.27622pt:\hskip-1.707165pt1$}\hskip-9.0pt{%
\mathfrak{A}}\hskip 2.0pt}:\mbox{{\tt dgVect}}\to\mbox{{\tt A}($m$)}$ to the forgetful
functor $\Box:\mbox{{\tt A}($m$)}\to\mbox{{\tt dgVect}}$ to the category
of differential graded vector spaces. One can easily verify
that
$${\raisebox{7.2pt}{\tiny$m\hskip-2.27622pt:\hskip-1.707165pt1$}\hskip-9.0pt{%
\mathfrak{A}}\hskip 2.0pt}(V,\partial)\cong\bigoplus_{n\geq 1}({\hskip 2.0pt%
\raisebox{7.2pt}{\tiny$m$}\hskip-8.0pt\underline{\mathcal{A}}}(n),\partial)%
\otimes(V,\partial)^{\otimes n},$$
(7)
where $({\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-8.0pt\underline{\mathcal{A}}},\partial)$ is now the non-$\Sigma$ operad describing
A($m$)-algebras as algebras in the monoidal category of differential vector spaces. Let us
recall [7, page 1493] that ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-8.0pt\underline{\mathcal{A}}}=%
\underline{\Gamma}(\xi_{2},\ldots,\xi_{m})$ with the differential
$$d(\xi_{n}):=\sum(-1)^{(b+1)(i+1)+b}\cdot\xi_{a}\circ_{i}\xi_{b},$$
where the summation runs over all $a,b\geq 2$ with $a+b=n+1$ and $1\leq i\leq a$.
Formula (7) is in a perfect harmony
with [8, Definition II.1.24] describing
free operad algebras in a general monoidal category.
Example 2.3.
Clearly ${\raisebox{7.2pt}{\tiny$m\hskip-2.27622pt:\hskip-1.707165pt1$}\hskip-9.0pt{%
\mathfrak{A}}\hskip 2.0pt}{}^{1}(V,\partial)\cong(V,\partial)$, and,
for $m\geq 2$, ${\raisebox{7.2pt}{\tiny$m\hskip-2.27622pt:\hskip-1.707165pt1$}\hskip-9.0pt{%
\mathfrak{A}}\hskip 2.0pt}{}^{2}(V,\partial)\cong(V\otimes V)$ with the
induced differential. As graded vector spaces
$$\displaystyle{\raisebox{7.2pt}{\tiny$2\hskip-2.27622pt:\hskip-1.707165pt1$}%
\hskip-9.0pt{\mathfrak{A}}\hskip 2.0pt}{}^{3}(V,\partial)$$
$$\displaystyle\cong$$
$$\displaystyle((V\otimes V)\otimes V)\oplus(V\otimes(V\otimes V)),$$
$$\displaystyle{\raisebox{7.2pt}{\tiny$m\hskip-2.27622pt:\hskip-1.707165pt1$}%
\hskip-9.0pt{\mathfrak{A}}\hskip 2.0pt}{}^{3}(V,\partial)$$
$$\displaystyle\cong$$
$$\displaystyle{\raisebox{7.2pt}{\tiny$2\hskip-2.27622pt:\hskip-1.707165pt1$}%
\hskip-9.0pt{\mathfrak{A}}\hskip 2.0pt}{}^{3}(V)\oplus(V{\otimes}V{\otimes}V),%
\mbox{ for $m\geq 3$,}$$
$$\displaystyle{\raisebox{7.2pt}{\tiny$2\hskip-2.27622pt:\hskip-1.707165pt1$}%
\hskip-9.0pt{\mathfrak{A}}\hskip 2.0pt}{}^{4}(V,\partial)$$
$$\displaystyle\cong$$
$$\displaystyle(((V{\otimes}V){\otimes}V){\otimes}V)\oplus(V{\otimes}(V{\otimes}%
(V{\otimes}V)))\oplus((V{\otimes}V){\otimes}(V{\otimes}V))$$
$$\displaystyle \oplus\ ((V{\otimes}(V{\otimes}V)){\otimes}V)\oplus(V{%
\otimes}((V{\otimes}V){\otimes}V)),$$
$$\displaystyle{\raisebox{7.2pt}{\tiny$3\hskip-2.27622pt:\hskip-1.707165pt1$}%
\hskip-9.0pt{\mathfrak{A}}\hskip 2.0pt}{}^{4}(V,\partial)$$
$$\displaystyle\cong$$
$$\displaystyle{\raisebox{7.2pt}{\tiny$2\hskip-2.27622pt:\hskip-1.707165pt1$}%
\hskip-9.0pt{\mathfrak{A}}\hskip 2.0pt}{}^{4}(V,\partial)\oplus((V{\otimes}V{%
\otimes}V){\otimes}V)\oplus(V{\otimes}(V{\otimes}V{\otimes}V))\oplus((V{%
\otimes}V){\otimes}V{\otimes}V)$$
$$\displaystyle \oplus\ (V{\otimes}(V{\otimes}V){\otimes}V)\oplus(V{\otimes%
}V{\otimes}(V{\otimes}V)),$$
$$\displaystyle{\raisebox{7.2pt}{\tiny$m\hskip-2.27622pt:\hskip-1.707165pt1$}%
\hskip-9.0pt{\mathfrak{A}}\hskip 2.0pt}{}^{4}(V,\partial)$$
$$\displaystyle\cong$$
$$\displaystyle{\raisebox{7.2pt}{\tiny$3\hskip-2.27622pt:\hskip-1.707165pt1$}%
\hskip-9.0pt{\mathfrak{A}}\hskip 2.0pt}{}^{4}(V,\partial)\oplus(V{\otimes}V{%
\otimes}V{\otimes}V{\otimes}V)\mbox{ for $m\geq 4$, etc.}$$
The action of the differential in ${\raisebox{7.2pt}{\tiny$m\hskip-2.27622pt:\hskip-1.707165pt1$}\hskip-9.0pt{%
\mathfrak{A}}\hskip 2.0pt}(V,\partial)$ can be
easily read off from the axioms of A($m$)-algebras, for example
$$\displaystyle\partial(x{\otimes}y{\otimes}z)=((x{\otimes}y){\otimes}z)-(x{%
\otimes}(y{\otimes}z))\hskip 100.0pt$$
$$\displaystyle -(\partial x{\otimes}y{\otimes}z)-(%
-1)^{|a|}(x{\otimes}\partial y{\otimes}z)-(-1)^{|a|+|b|}(x{\otimes}y{\otimes}%
\partial z),$$
where $(x{\otimes}y{\otimes}z)\in{\raisebox{7.2pt}{\tiny$m\hskip-2.27622pt:\hskip-1.7%
07165pt1$}\hskip-9.0pt{\mathfrak{A}}\hskip 2.0pt}{}^{3}(V)$, $m\geq 3$. The above
formula illustrates a surprising fact that the differential in ${\raisebox{7.2pt}{\tiny$m\hskip-2.27622pt:\hskip-1.707165pt1$}\hskip-9.0pt{%
\mathfrak{A}}\hskip 2.0pt}(V,\partial)$ is nontrivial even if the differential $\partial$ on $V$ is
zero, because in this case still
$$\partial(x{\otimes}y{\otimes}z)=((x{\otimes}y){\otimes}z)-(x{\otimes}(y{%
\otimes}z))\ !$$
The dimension of the homogeneous components of ${\raisebox{7.2pt}{\tiny$m\hskip-2.27622pt:\hskip-1.707165pt1$}\hskip-9.0pt{%
\mathfrak{A}}\hskip 2.0pt}(V,\partial V)$ clearly equals
$$\dim{{\raisebox{7.2pt}{\tiny$m\hskip-2.27622pt:\hskip-1.707165pt1$}\hskip-9.0%
pt{\mathfrak{A}}\hskip 2.0pt}{}^{n}(V,\partial)}=(\dim(V))^{n}\cdot b^{m}_{n},%
\mbox{ for }n\geq 1,\ m\geq 2,$$
where the integers $b^{m}_{n}$ are as in (3).
3 Free strongly homotopy Lie algebras
In this section we indicate how to modify previous results to strongly
homotopy Lie algebras. Recall ([4]
or [3, Definition 2.1]) that an L($m$)-structure on a graded vector space $L$ is a system $\{l_{k};\ 1\leq k\leq m,\ k<\infty\}$ of degree $k-2$ linear maps $l_{k}:\otimes^{k}L\to L$ which are antisymmetric in the sense that
$$l_{k}(x_{\sigma(1)},\dots,x_{\sigma(k)})=\chi(\sigma)l_{k}(x_{1},\dots,x_{n})$$
(8)
for all $\sigma\in\Sigma_{n}$ and $x_{1},\dots,x_{n}\in L$.
Here $\chi(\sigma):={\rm sgn}(\sigma)\cdot\epsilon(\sigma)$, where
$\epsilon(\sigma)$ is the Koszul sign of the permutation $\sigma$,
see [3, page 2148] for details.
Moreover, the
following generalized form of the Jacobi identity is supposed to be
satisfied for any $n\leq m$:
$$\sum_{i+j=n+1}\sum_{\sigma}\chi(\sigma)(-1)^{i(j-1)}l_{j}(l_{i}(x_{\sigma(1)},%
\dots,x_{\sigma(i)}),x_{\sigma(i+1)},\dots,x_{\sigma(n)})=0,$$
(9)
where the summation is taken over all $(i,n-i)$-unshuffles with
$i\geq 1$.
The operad ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathcal{L}}}=\{{\hskip 2.%
0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathcal{L}}}(n)\}_{n\geq 1}$ describing L($m$)-algebras as algebras in the category of graded vector spaces can be
constructed as follows. Let ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.2pt{\mathcal{F}}}=\{{\hskip 2.%
0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.2pt{\mathcal{F}}}(n)\}_{n\geq 1}$ be the
free graded $\Sigma$-operad $\Gamma(\zeta_{1},\zeta_{2},\ldots,\zeta_{m})$
generated by the operations $\zeta_{i}$, $l\leq i\leq m$, where each
$\zeta_{i}$ has $i$ inputs, degree $i-2$ and spans the signum
representation of the symmetric group $\Sigma_{i}$. Using trees, this
operad can be described as follows. Let ${\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}(n)$ be now the set of
all directed labeled (non-planar, abstract) trees with $n$-input
edges, all of whose vertices have at most $m$ input edges. We again
admit also ‘unary’ vertices with only one input edge and ‘labeled’
means that the input legs are labeled by natural numbers $1,\ldots,n$.
Then, as before,
$${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.2pt{\mathcal{F}}}(n)\cong{\rm
Span%
}_{{\bf k}}({\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}(n)).$$
We put ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathcal{L}}}:={\hskip 2.0%
pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.2pt{\mathcal{F}}}/{\hskip 2.0pt\raisebox{%
7.2pt}{\tiny$m$}\hskip-6.0pt{\mathcal{I}}}$, where the operadic ideal ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-6.0pt{\mathcal{I}}}$ is now
generated by relations whose pictorial presentation is indicated
in Figure 2.
By similar arguments as in the previous sections we easily infer that
there is a natural basis of the ${\bf k}$-vector space ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathcal{L}}}(n)$, formed
by the set ${\hskip 8.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-14.0pt{{\rm Adm}}}(n)$ of admissible $n$-trees, whose definition is
formally the same as in Definition 1.1, only removing everywhere
the adjective “planar.” The size of the operad ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathcal{L}}}$ is
described in the following proposition whose proof is similar to that
of Proposition 1.4.
Proposition 3.1.
The generating function
$\vartheta(t)={\buildrel\mbox{\scriptsize$m$}\over{\vartheta}}(t):=\sum_{n\geq 1%
}\frac{1}{n!}\dim_{{\bf k}}({\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.0%
pt{\mathcal{L}}}(n))t^{n}$
satisfies
the equation
$$\vartheta(t)-2t=\vartheta^{2}(t)\left(\frac{1}{2!}+\frac{1}{3!}\vartheta(t)+%
\cdots+\frac{1}{m!}\vartheta^{m-2}(t)\right).$$
(10)
Example 3.2.
For $m=\infty$, equation (10) becomes
$2\vartheta(t)-2t-1=\exp(\vartheta(t))$. For $m=2$,
equation (10) says that $\vartheta(t)=1-\sqrt{1-4t^{2}}$, therefore
$$\dim_{{\bf k}}({\buildrel\mbox{\scriptsize$2$}\over{\mathcal{L}}}(n))=2^{n}%
\cdot(2n-3)!!,\ n\geq 1.$$
As in Lemma 1.3, we have a map $\Omega:{\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}(n)\to{\rm Span%
}_{{\bf k}}({\hskip 8.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-14.0pt{{\rm Adm}}}(n))$ with the property that $\Omega(T)\equiv T$ modulo
relations of Figure 2. Free L($m$)-algebras are then
described as follows.
Theorem 3.3.
Let $Y$ be a graded vector space. The free L($m$)-algebra ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{L}}}(Y)$
generated by $Y$ can defined as ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{L}}}(Y)=%
\bigoplus_{n\geq 1}{\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{%
\mathfrak{L}}}{}^{n}(Y)$, where
$${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{L}}}{}^{n}(Y):={%
\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathcal{L}}}(n)\otimes_{%
\Sigma_{n}}Y^{\otimes n}.$$
The structure operations $l_{1},\ldots,l_{m}$ are given by the
following rule.
Let $\vec{v}_{j}\in Y^{\otimes k_{j}}$ and $T_{j}\in{\hskip 8.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-14.0pt{{\rm Adm}}}(k_{j})$, $1\leq j\leq k$. Then
$$l_{k}(T_{1}\otimes_{\Sigma_{k_{1}}}\vec{v}_{1},\ldots,T_{k}\otimes_{\Sigma_{k_%
{1}}}\vec{v}_{k}):=\Omega(\xi_{k}(T_{1},\ldots,T_{k}))\otimes_{\Sigma_{k_{1}+%
\cdots+k_{k}}}(\vec{v}_{1}\otimes\cdots\otimes\vec{v}_{k}).$$
Remark 3.4.
In exactly the same fashion as for A($m$)-algebras one may define
left adjoints ${\hskip 2.0pt\raisebox{7.2pt}{\scriptsize$m\hskip-2.27622pt:\hskip-1.707165ptn%
$}\hskip-9.0pt{\mathcal{L}}}:\mbox{{\tt L}($n$)}\to\mbox{{\tt L}($m$)}$ to
the forgetful functor $\Box:\mbox{{\tt L}($m$)}\to\mbox{{\tt L}($n$)}$, $n<m$, as
$${\hskip 2.0pt\raisebox{7.2pt}{\scriptsize$m\hskip-2.27622pt:\hskip-1.707165ptn%
$}\hskip-9.0pt{\mathcal{L}}}\ (L,\lambda_{1},\lambda_{2},\ldots,\lambda_{n})={%
\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{L}}}(L)/(\lambda_%
{i}=l_{i},\ i\leq n),$$
where $l_{i}$ are the structure operations of ${\hskip 2.0pt\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{L}}}(L)$.
4 Reflections on the PBW theorem
As in the case of ordinary Lie algebras, there exists the
symmetrization functor from the category A($m$) of A($m$)-algebras to the
category L($m$) of L($m$)-algebras. More precisely, we proved
in [3, Theorem 3.1]:
Theorem 4.1.
Any A($m$)-structure $\{\mu_{n}:\bigotimes^{n}V\longrightarrow V\}$
on a differential graded vector space $V$ induces an L($m$)-structure
$\{l_{n}:\bigotimes^{n}V\longrightarrow V\}$ on the same
differential graded vector space, with
$$l_{n}(v_{1}\otimes\dots\otimes v_{n}):=\sum_{\sigma\in{\mathcal{S}}_{n}}\chi(%
\sigma)\mu_{n}(v_{\sigma(1)}\otimes\dots\otimes v_{\sigma(n)}),\ 2\leq n\leq m.$$
This correspondence defines a functor (the symmetrization)
$(-)_{L}:{\tt A}(m)\longrightarrow{\tt L}(m)$.
In [3, Theorem 3.3], we also proved that:
Theorem 4.2.
There exists a functor $\mathcal{U}_{m}:{\tt L}(m)\longrightarrow{\tt A}(m)$ that is left adjoint to the functor $(-)_{L}:{\tt A}(m)\longrightarrow{\tt L}(m)$.
The algebra $\mathcal{U}_{m}(L)$ of the previous theorem is called
the universal enveloping A($m$)-algebra
of the L($m$)-algebra $L$.
We gave, in [3, page 2154],
the following explicit construction of
${\mathcal{U}}_{m}(L)$.
Start with the free unital A($m$)-algebra ${\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}(L)$ generated by the
underlying vector space $L$ of the L($m$)-algebra $L=(L,\{l_{n}\})$ and let
$\{\mu_{n}\}$ be the A($m$)-structure maps of ${\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}(L)$. Let
$I$ denote the A($m$)-ideal in ${\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}(L)$ generated by the relations
$$\sum_{\sigma\in S_{n}}\chi(\sigma)\mu_{n}(\xi_{\sigma(1)},\dots,\xi_{\sigma(n)%
})=l_{n}(\xi_{1},\dots,\xi_{n}),\mbox{ for }\xi_{1},\dots,\xi_{n}\in L,\ n\leq
m.$$
Then we put $\mathcal{U}_{m}(L):={\raisebox{7.2pt}{\tiny$m$}\hskip-7.0pt{\mathfrak{A}}}(L)/I$. This universal
enveloping A($m$)-algebra is equipped with the canonical map (in fact, an
inclusion) $\iota:L\to\mathcal{U}_{m}(L)$.
Remark 4.3.
Hinich and Schechtman introduced
in [2], for any L($m$)-algebra
$L$ (they did it for $m=\infty$, but the general case is an
obvious modification), an associative algebra (which they
also called the universal enveloping algebra), having the property
that the category of modules (in a suitable sense, see for
example [3, Definition 5.1]) over an
L($m$)-algebra $L$ is equivalent to the category of left modules over
the universal enveloping associative algebra of $L$.
It is unclear whether there exists a functor ${\mathcal{G}}:\mbox{{\tt L}($m$)}\to\mbox{{\tt A}($m$)}$ such that the category of
modules over an L($m$)-algebra $L$ would be naturally equivalent to
the category of left modules, in the sense
of [5, page 157], over the A($m$)-algebra $\mathcal{G}(L)$. Our universal enveloping A($m$)-algebra functor whose
definition we recalled above probably does not have this property.
While it can be easily shown that each left module over the
A($m$)-algebra $\mathcal{U}_{m}(L)$ naturally defines a module over
$L$, we doubt that each $L$-module uniquely extends into a left
module over $\mathcal{U}_{m}(L)$. The classical universal enveloping
associative algebra of a strict Lie algebra is constructed as a
left adjoint to the symmetrization functor. By a miracle, this
functor also happens to describe the category of $L$-modules, but this
mystic phenomenon does not take place in the category of
L($m$)-algebras.
Moreover, there is no reason to believe that such a functor
$\mathcal{G}$ exits. The category of modules over a given
L($m$)-algebra $L$ is a well-behaved abelian category, therefore it
should be the category of modules over an associative
algebra, which turns out to be the one constructed by Hinich and
Schechtman. On the other hand, we do not see why this abelian
category should be at the same time the category of left modules
over some A($m$)-algebra. Let us close this remark with
Problem 4.4.
Let $L$ be an L($m$)-algebra. Does there exist an A($m$)-algebra $A$ with the property that the category of modules
over $L$ is equivalent to the category of left modules over $A$?
The simplest form of the P(oincaré)-B(irkhoff)-W(itt) theorem for an
ordinary Lie algebra $L=(L,[-,-])$ says that the associated graded
algebra $G^{*}(L)$ of the universal enveloping algebra ${\mathcal{U}}(L)$
of $L$ is isomorphic to the free commutative associative algebra
$S^{*}(L)$ (the polynomial ring) on the vector space $L$. More
precisely, recall that
$${\mathcal{U}}(L)=T(L)/(x{\otimes}y-y{\otimes}x=[x,y],\ x,y\in L),$$
where $T(L)$ is the free associative algebra (tensor algebra) on the
vector space $L$. The algebra ${\mathcal{U}}(L)$ is filtered, the ascending
filtration being given by ${\mathcal{U}}_{p}(L):=$ vector space generated by
linear combinations of elements which can be presented as a product of
$\ \leq p$ elements in the augmentation ideal of $T(L)$. Clearly
${\mathcal{U}}_{0}(L)={\bf k}$ and ${\mathcal{U}}_{1}(L)={\bf k}\oplus L$. If $G^{*}(L)$ denotes
the associated graded algebra, then one proves that the natural map
$L=G^{1}(L)\to G^{*}(L)$ is a monomorphism and $G^{*}(L)$ is generated by
$L$. It is immediate to see that $xy=yx$, for arbitrary $x,y\in L\subset G^{*}(L)$. We formulate the following trivial observation.
Observation 4.5.
Suppose that an associative algebra $A$ admits a set of mutually
commutative generators. Then $A$ is commutative.
From this observation we infer that $G^{*}(L)$ is commutative,
therefore there exists, by the universal property, a commutative
algebra map $S^{*}(L)\to G^{*}(L)$, induced by the inclusion $L\subset G^{*}(L)$. The PBW222not to be mistaken with
Praha-Berlin-Warszawa cycling race theorem says that this map is an
isomorphism of graded commutative associative algebras.
Let us try to guess which form the PBW theorem for the universal
enveloping A($m$)-algebra ${\mathcal{U}}_{m}(L)$ of an L($m$)-algebra $L$ may possibly
have. The classical PBW theorem compares the associated graded of
${\mathcal{U}}(L)$ to a commutative associative algebra. Commutative
associative algebras occur in the symbolic exact sequence (which may
be found in [1, page 228])
$$\mbox{Commutative associative algebras}\subset\mbox{Associative algebras}%
\stackrel{{\scriptstyle\pi}}{{\longrightarrow}}\mbox{Lie algebras.}$$
Here the ‘projection’ $\pi$ is given by the symmetrization of the
associative product. Commutative associative
algebras are also the quadratic duals of Lie algebras
(and vice versa) [1, Theorem 2.1.11],
but this coincidence is misleading. Summing up,
the classical PBW theorem says that the associated graded $G^{*}(L)$
is the free algebra in the kernel of the symmetrization map $\pi$.
One would naturally expect something similar
also in the L($m$)-algebra case. The kernel
of the symmetrization map $(-)_{L}$ of
Theorem 4.1 clearly
consists of A($m$)-algebras
$A=(A,\{\mu_{n}\})$ such that
$$\sum_{\sigma\in S_{n}}\chi(\sigma)\mu_{n}(a_{\sigma(1)},\ldots,a_{\sigma(n)})=%
0,\ n\geq 2,\ a_{1},\dots,a_{n}\in A.$$
(11)
Therefore one tends to believe that A($m$)-algebras enjoying this form of
symmetry are analogs of the polynomial algebra $S^{*}(L)$ from the
classical PBW theorem. This is, however, not exactly so.
To see why, let $L=(L,\{l_{n}\})$ be an L($m$)-algebra and define
inductively an ascending filtration of the universal enveloping
A($m$)-algebra ${\mathcal{U}}_{m}(L)$ by ${\mathcal{U}}_{m,0}(L):={\bf k}$, ${\mathcal{U}}_{m,1}(L):={\mathcal{U}}_{m,0}\cup\iota(L)$ and ${\mathcal{U}}_{m,p}(L)$ being the vector space
generated by elements of the form $\mu_{k}(x_{1},\dots,x_{k})$, $k\geq 2$, $x_{i}\in{\mathcal{U}}_{m,p_{i}}(L)$ for $1\leq i\leq k$ and $\sum_{1}^{k}p_{i}\leq p$.
The associated graded A($m$)-algebra $G^{*}_{m}(L)$ is defined by
$$G^{*}_{m}(L)=\bigoplus_{q\geq 0}G^{q}_{m}(L),\ G^{q}_{m}(L):={\mathcal{U}}_{m,%
q}(L)/{\mathcal{U}}_{m,q-1}(L)$$
with the A($m$)-structure maps induced by that of ${\mathcal{U}}_{m}(L)$.
It is easily seen that $G^{*}_{m}(L)$
is generated by the image of the canonical
inclusion $\iota:L\hookrightarrow{\mathcal{U}}_{m,1}(L)\to G^{*}_{m}(L)$.
A moment’s reflection shows that the elements of $L\subset G_{m}^{*}(L)$
‘mutually commute’ in the sense that (11) is satisfied for
$a_{1},\ldots,a_{n}\in L$. Surprisingly,
this does not imply that (11)
is satisfied for all elements of $G^{*}_{m}(L)$ –
there is no analog of
Observation 4.5 for A($m$)-algebras!
One must instead consider A($m$)-algebras $S_{m}(V,\partial)$
defined as
$$S_{m}(V,\partial):={\raisebox{7.2pt}{\tiny$m\hskip-2.27622pt:\hskip-1.707165pt%
1$}\hskip-9.0pt{\mathfrak{A}}\hskip 2.0pt}(V,\partial)/{\mathfrak{I}}_{m},$$
where ${\mathfrak{I}}_{m}$ is the ideal generated by the relations
$$\sum_{\sigma\in S_{n}}\chi(\sigma)\mu_{n}(v_{\sigma(1)},\ldots,v_{\sigma(n)})=%
0\ \mbox{ for }\ 2\leq n\leq m\ \mbox{ and }\ v_{1},\dots,v_{n}\in V.$$
The algebra $S_{m}(V,\partial)$ is (externally) graded, with the grading induced from
the (external) grading of the free A($m$)-algebra ${\raisebox{7.2pt}{\tiny$m\hskip-2.27622pt:\hskip-1.707165pt1$}\hskip-9.0pt{%
\mathfrak{A}}\hskip 2.0pt}(V,\partial)$.
By a miracle, $S_{m}^{*}(V,\partial)$ has the
following very explicit description.
Choose a basis $\{f_{i}\}_{i\in I}$ of $V$ indexed by a linearly ordered
set $I$. Let ${\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}{}^{n}(V)$ be the set of all planar rooted reduced
$n$-trees $S$ whose vertices have arity $\leq m$. We moreover
suppose that the input legs of $S$ are labeled by elements of the
basis $\{f_{i}\}_{i\in I}$ and that the labels satisfy the following
condition:
Let $v$ be an input vertex of $S$ and let $e_{1},\dots,e_{k}$ be the
input legs of $v$, written in the order induced by the imbedding of $S$
into the plane. Let $f_{i_{1}},\dots,f_{i_{k}}$ be the corresponding labels.
Then it is not true that ${i_{1}}\geq i_{2}\geq\cdots\geq i_{k}$.
Proposition 4.6.
For each $n\geq 1$, $S^{n}_{m}(V)\cong{\rm Span}_{{\bf k}}({\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{%
\mbox{$\tt T$}}}{}^{n}(V))$.
Proof.
It follows from (7) that ${\raisebox{7.2pt}{\tiny$m\hskip-2.27622pt:\hskip-1.707165pt1$}\hskip-9.0pt{%
\mathfrak{A}}\hskip 2.0pt}(V,\partial)$ is spanned by the
set of planar directed reduced $n$-trees with all vertices of arity
$\leq m$ and the input legs labeled by the chosen basis
$\{f_{i}\}_{i\in I}$ of $V$. Let $v$ be an input vertex of $S$,
$e_{1},\ldots,e_{k}$ input legs of $v$ written in the order induced by
the plane and $f_{i_{1}},\dots,f_{i_{k}}$ their labels.
Suppose ${i_{1}}\geq i_{2}\geq\cdots\geq i_{k}$. If $i_{1}=i_{k}$, then
clearly $S\equiv 0$ modulo ${\mathfrak{I}}_{m}$. If $i_{1}\not=i_{k}$, we may replace $S$ modulo ${\mathfrak{I}}_{m}$ by a sum of trees
that differ from $S$ only by labels of $e_{1},\dots,e_{k}$ which are
changed to $i_{\sigma(1)},\dots,i_{\sigma(k)}$, for some $\sigma\in\Sigma_{k}$ such that it is not true that
${i_{\sigma(1)}}\geq i_{\sigma(2)}\geq\cdots\geq i_{\sigma(k)}$.
Repeating this process at any input vertex of $S$, we replace $S$
modulo ${\mathfrak{I}}_{m}$ by a sum of trees from ${\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}{}^{n}(V)$.
This replacement is obviously unique.
∎
Let us formulate our version of the PBW theorem for the universal
enveloping A($m$)-algebra of an L($m$)-algebra.
Theorem 4.7.
The canonical map $\iota:L\hookrightarrow{\mathcal{U}}_{m,1}(L)\to G^{*}_{m}(L)$
induces an isomorphism
$$\rho:S^{*}_{m}(L)\cong G^{*}_{m}(L)$$
of (externally) graded A($m$)-algebras.
Proof.
The proof relies on the fact that $\{\rho(S);\ S\in{\raisebox{7.2pt}{\tiny$m$}\hskip-5.8pt{\mbox{$\tt T$}}}{}^{n}%
(V)\}$
forms a basis of $G^{n}_{m}(V)$, which easily follows from the above
analysis. We leave the details to the reader.
∎
Example 4.8.
Let us analyze the case $m=2$ when
L($m$)-algebras are just differential graded
anti-commutative non-associative algebras. For an L(2)-algebra $L=(L,[-,-],\partial)$, ${\raisebox{7.2pt}{\tiny$2\hskip-2.27622pt:\hskip-1.707165pt1$}\hskip-9.0pt{%
\mathfrak{A}}\hskip 2.0pt}(L)$ is the free
non-associative dg algebra generated by $(L,\partial)$, the universal enveloping
A(2)-algebra of $L$ is given by
$$\mathcal{U}_{2}(L)={\raisebox{7.2pt}{\tiny$2\hskip-2.27622pt:\hskip-1.707165pt%
1$}\hskip-9.0pt{\mathfrak{A}}\hskip 2.0pt}/(xy-(-1)^{|x||y|}yx=[x,y],\ x,y\in L),$$
and the associated graded $G^{*}_{2}(L)$ is the quotient of ${\raisebox{7.2pt}{\tiny$2\hskip-2.27622pt:\hskip-1.707165pt1$}\hskip-9.0pt{%
\mathfrak{A}}\hskip 2.0pt}(L)$
by the ideal generated by $xy=(-1)^{|x||y|}yx$, for $x,y\in L$.
Remark 4.9.
An ordinary strict Lie algebra $L=(L,[-,-])$ can be considered,
for any $m\geq 2$, as an L($m$)-algebra with $l_{2}=[-,-]$ and $l_{n}=0$ if $n\not=2$. Its universal enveloping A($m$)-algebra
${\mathcal{U}}_{m}(L)$ is a fully-fledged A($m$)-algebra, therefore an
object completely different from the ordinary universal enveloping
associative algebra ${\mathcal{U}}(L)$. Thus our $\mathcal{U}_{m}(L)$ cannot be
considered as a generalization of the classical universal
enveloping algebra functor.
Appendix: the tree language
The following definitions can be found for example
in [8, Section II.1.5]. By a tree we mean a connected graph $T$ without loops. Let ${\rm Vert}(T)$
denote the set of vertices of $T$. A valence of a vertex $v\in{\rm Vert}(T)$ is the number of edges adjacent to $v$. A leg of
$T$ is an edge adjacent to a vertex of valence one, other edges of $T$
are interior. We in fact discard vertices of valence one at
the endpoints of the legs, therefore the legs are “half-edges” having
only one vertex. This can be formalized by introducing (generalized)
graphs as certain sets with involutions,
see [8, Definition II.5.37].
By a rooted or directed tree we mean a tree with a
distinguished output leg called the root. The
remaining legs are called the input legs of the tree. A tree
with $n$ input legs is called an $n$-tree. A rooted tree is
automatically oriented, each edge pointing towards the root.
The edges pointing towards a given vertex $v$ are called the input edges of $v$, the number of these input edges is then the
arity of $v$. Vertices of arity one are called unary, vertices
of arity two binary, vertices of arity three ternary, etc. A tree is
reduced if it has no unary vertices. A vertex $v\in{\rm Vert}(T)$ is an input vertex of $T$ if all its input edges are
input legs of $T$. With the above conventions accepted, we must admit
also the unique reduced 1-tree which consist of one edge which is both
its root and the input leg, and no vertices (soul without body).
Each tree can be imbedded into the plane. By a planar tree we
mean a tree with a specified (isotopy class of) imbedding. Reading
off counterclockwise the input legs of a planar tree, starting from
the leftmost one, gives a linear order of input legs. In the same
manner, input edges of each vertex of a planar tree are linearly
oriented. Vice versa, specifying linear orders of input edges of each
vertex of a rooted tree $T$ uniquely determines an isotopy class of
imbedding of $T$ into the plane.
Example 4.10.
In Figure 3, $f$ is the root, $a$ the root vertex,
$l,m,n,o,p,q,r,s,t,u$ the input legs and $g,j,k,i$ the interior edges
of $T$. Vertices $b,c,d,v$ are the input vertices of $T$. The vertex $d$
is unary, $v$ binary and the remaining vertices are ternary. The tree
in Figure 3 is not reduced because it contains an unary
vertex.
References
[1]
V. Ginzburg and M.M. Kapranov.
Koszul duality for operads.
Duke Math. Journal, 76(1):203–272, 1994.
[2]
V.A. Hinich and V.V. Schechtman.
Homotopy Lie algebras.
Advances in Soviet Mathem., 16(2):1–28, 1993.
[3]
T. Lada and M. Markl.
Strongly homotopy Lie algebras.
Communications in Algebra, 23(6):2147–2161, 1995.
[4]
T. Lada and J.D. Stasheff.
Introduction to sh Lie algebras for physicists.
International Journal of Theoretical Physics, 32(7):1087–1103,
1993.
[5]
M. Markl.
A cohomology theory for $A(m)$-algebras and applications.
Journ. Pure Appl. Algebra, 83:141–175, 1992.
[6]
M. Markl.
Distributive laws and Koszulness.
Annales de l’Institut Fourier, 46(4):307–323, 1996.
[7]
M. Markl.
Models for operads.
Communications in Algebra, 24(4):1471–1500, 1996.
[8]
M. Markl, S. Shnider, and J. D. Stasheff.
Operads in Algebra, Topology and Physics, volume 96 of Mathematical Surveys and Monographs.
American Mathematical Society, Providence, Rhode Island,
2002.
[9]
J.-P. Serre.
Lie Algebras and Lie Groups.
Benjamin, 1965.
[10]
J.D. Stasheff.
Homotopy associativity of H-spaces I,II.
Trans. Amer. Math. Soc., 108:275–312, 1963. |
Coherent Multiple Scattering Effect in DIS
Xiaofeng Guo and Jun Li
Department of Physics and Astronomy, Iowa State University
Ames, Iowa 50011, U.S.A.
Abstract
We present a QCD calculation of the coherent multiple scattering effect
on single hadron production in semi-inclusive lepton-nucleus
deep inelastic scattering (DIS). We show that the destructive
interference of coherent multiple scattering leads to suppression
in hadron productions. The suppression is similar
to the expectations of energy loss calculations.
We compare our calculation with experimental data.
keywords:
Multiple scattering, energy loss, deep inelastic scattering
PACS: 12.38.Bx 13.85.Ni 24.85.+p 25.75.-q
††thanks: Supported in part by the National Science Fundation,
grant PHY-0340729.
1. Introduction The strong suppression of high transverse momentum hadron in
relativistic heavy ion collisons was considered to be
an evidence for the QCD quark-gluon plasma [1].
The suppression was believed to be the result of
medium induced
radiative energy loss of high energy partons [2].
However, recent data indicate that heavy quarks would have to lose
the same amount of energy as that of a light quark
if the radiative energy loss is the only source of the suppression
[3]. On the other hand, we expect heavy quarks to lose
much less energy than a light quark because of its mass [4].
In this talk, we present a new source of suppresion from
the destructive multiple scattering in a nuclear medium [5].
We calculated the single hadron production
in DIS in perturbative QCD, and
found that the coherent interference results into a suppression
in the hadron production rate.
2. Single hard scattering We consider the production of a single hadron of momentum $p_{h}$ in
lepton-nucleus deep inelastic scattering (DIS),
$e(k_{1})+A(p)\longrightarrow e(k_{2})+h(p_{h})+X$, with the target
atomic weight $A$ and average nucleon momentum $p=P_{A}/A$.
The semi-inclusive DIS cross section,
$$\frac{d{\sigma}_{eA\rightarrow ehX}}{dx_{B}dQ^{2}dz_{h}}=\frac{1}{8\pi}\,\frac%
{e^{4}}{x_{B}^{2}s^{2}Q^{2}}\,L^{\mu\nu}(k_{1},k_{2})\,\frac{dW^{A}_{\mu\nu}}{%
dz_{h}}\ ,$$
(1)
where $s=(p+k_{1})^{2}$ is the total invariant mass of the lepton-nucleon
system, and
$z_{h}\equiv{p\cdot p_{h}}/{p\cdot q}={2x_{B}p\cdot p_{h}}/{Q^{2}}$
is the fraction of the photon momentum carried by the observed hadron.
The Bjorken $x_{B}=Q^{2}/(2p\cdot q)$ with $Q^{2}=-q^{2}$.
The leptonic tensor $L^{\mu\nu}(k_{1},k_{2})={\rm Tr}(\gamma\cdot k_{1}\gamma^{\mu}\gamma\cdot k_{2%
}\gamma^{\nu})/2$. For single scattering
at the lowest order, the hadronic tensor can be factorized as
$$\frac{dW^{A}_{\mu\nu}}{dz_{h}}=\frac{1}{2}\,e^{T}_{\mu\nu}\sum_{q}\,e_{q}^{2}%
\,A\,\phi^{A}_{q}(x_{B},Q^{2})\,D_{q}(z_{h},Q^{2})\,,$$
(2)
where $e^{T}_{\mu\nu}=[p_{\mu}q_{\nu}+q_{\mu}p_{\nu}]/{p\cdot q}+p_{\mu}p_{\nu}({2x_{%
B}}/{p\cdot q})-g_{\mu\nu}$,
$\phi^{A}_{q}(x_{B},Q^{2})$ is nuclear quark distribution
per nucleon of flavor $q$, and $D_{q}(z_{h},Q^{2})$ is quark-to-hadron
fragmentation function.
The nuclear dependence of Eq. (2)
and Eq. (1) is
limited to the nuclear dependence of $\phi^{A}_{q}(x_{B},Q^{2})$.
3. Coherent multiple scattering Multiple scattering
introduces the medium size enhanced, or the $A^{1/3}$-type,
nuclear effects. We adopt the leading pole approaximation
to extract the leading contributions in powers of
$A^{1/3}$[6].
For double scattering at lowest order in $\alpha_{s}$,
we only need to consider two diagrams in Fig. 1.
When $x_{B}\geq x_{c}\sim 0.1$ [7],
we consider coherent multiple scattering effects
between the scattered quark and the partons of the nuclear target
at the same impact parameter in the Breit frame
and obtain [5]
$$\displaystyle\frac{dW_{\mu\nu}^{(1)}}{dz_{h}}$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{2}e_{\mu\nu}^{T}\,\sum_{q}e_{q}^{2}\left[\frac{4\pi^{2}%
\alpha_{s}}{3}\right]\frac{z_{h}}{Q^{2}}\,T_{qg}^{A}(x_{B},Q^{2})\,\frac{dD(z_%
{h},Q^{2})}{dz_{h}}$$
(3)
$$\displaystyle\approx$$
$$\displaystyle\frac{1}{2}e_{\mu\nu}^{T}\,\sum_{q}e_{q}^{2}\left(A\phi^{A}_{q}(x%
_{B},Q^{2})\right)\left[\frac{z_{h}\xi^{2}(A^{1/3}-1)}{Q^{2}}\right]\frac{dD(z%
_{h},Q^{2})}{dz_{h}}$$
(4)
where $\sum_{q}$ runs over all quark and antiquark flavors.
The twist-4 quark-gluon correlation function, $T_{qg}^{A}(x_{B},Q^{2})$,
is defined in Ref. [6].
In deriving Eq. (4), we assumed that a large nucleus
is made of color neutral nucleons with a constant density.
The quantity $\xi^{2}$ defined in Ref. [7]
is proportional to a matrix element of a gluon density operator, and
represents a characteristic scale of quark interaction with
nuclear medium.
If $x_{B}<x_{c}$, the hadronic tensor in Eqs. (3) and
(4) has an additional term proportional to
$x_{B}\frac{d}{dx_{B}}\phi_{q}(x_{B},Q^{2})$ [5].
For $x_{B}\geq x_{c}$, we generalized our result in Eq. (4) to
$n$ scattering [5]
$$\displaystyle\frac{dW_{\mu\nu}^{(n)}}{dz_{h}}$$
$$\displaystyle\approx$$
$$\displaystyle\frac{1}{2}e_{\mu\nu}^{T}\,\sum_{q}e_{q}^{2}\left(A\phi^{A}_{q}(x%
_{B},Q^{2})\right)\frac{1}{n!}\left[\frac{z_{h}\xi^{2}(A^{1/3}-1)}{Q^{2}}%
\right]^{n}\frac{d^{n}D(z_{h},Q^{2})}{dz_{h}^{n}}\,.$$
(5)
We sum the $A^{1/3}$-type contributions in Eq. (5) to
$N$ scatterings, and take
$N\approx\infty$ because the effective value of $\xi^{2}$ is
relatively small. We obtain
$$\displaystyle\frac{dW^{A}_{\mu\nu}}{dz_{h}}$$
$$\displaystyle\approx$$
$$\displaystyle\frac{1}{2}e_{\mu\nu}^{T}\,\sum_{q}e_{q}^{2}\left(A\phi^{A}_{q}(x%
_{B},Q^{2})\right)\sum_{n=0}^{N}\frac{1}{n!}\left[\frac{z_{h}\xi^{2}(A^{1/3}-1%
)}{Q^{2}}\right]^{n}\frac{d^{n}D(z_{h},Q^{2})}{dz_{h}^{n}}$$
(6)
$$\displaystyle\approx$$
$$\displaystyle A\,\frac{1}{2}e_{\mu\nu}^{T}\,\sum_{q}e_{q}^{2}\,\phi^{A}_{q}(x_%
{B},Q^{2})\,D_{q}\left(z_{h}+\frac{z_{h}\xi^{2}(A^{1/3}-1)}{Q^{2}},Q^{2}\right%
)\,.$$
Eq. (6) is our main result.
The net effect of coherent multiple scattering of
a propagating quark in the medium,
is equivalent to a shift, $\Delta z$, in the variable $z_{h}$
of the quark fragmentation function $D_{q}(z_{h},Q^{2})$ with
$\Delta z=z_{h}\,\xi^{2}(A^{1/3}-1)/Q^{2},$
which leads to a suppression of the production rate
[5].
4. Comparison with HERMES data We compare our result with HERMES data [8] in Fig. 2.
$R_{M}$ is the ratio
of hadron production rate rate per DIS event for a nuclear target
$A$, $R^{A}$, to that of a deuterium target $D$, $R^{D}$ [5],
$$R_{M}\equiv\frac{R^{A}}{R^{D}}\approx\frac{\sum_{q}e_{q}^{2}\,\phi^{A}_{q}(x_{%
B},Q^{2})\,D_{q}(z_{h}+\Delta z,Q^{2})}{\sum_{q}e_{q}^{2}\,\phi^{D}_{q}(x_{B},%
Q^{2})\,D_{q}(z_{h},Q^{2})}\frac{\sum_{q}e_{q}^{2}\,\phi^{D}_{q}(x_{B},Q^{2})}%
{\sum_{q}e_{q}^{2}\,\phi^{A}_{q}(x_{B},Q^{2})}\ .$$
(7)
The dashed lines are from Eq. (7) and the dot-dashed lines
includes only the double scattering.
At large $z_{h}$, resummed theory curves are much
steeper than the data.
This is because the formation time for the hadron or
corresponding pre-hadron state is much shorter at large $z_{h}$,
and therefore, the number of scattering for the propagating parton
at large $z$, or the shift, should be reduced [5].
Since the hadron formation time is proportional to
$(1-z_{h})$ [9], we modify
$\Delta z$ by multiplying a factor $(1-z_{h})$ and
produce the solid lines in Fig. 2.
The suppression from the coherent multiple scattering, calculated here,
complements the suppression from medium-induced radiation
[10], which should give additional suppression and
bring down the curves in Fig. 2 [5].
5. Summary and outlook We showed that
the coherent multiple scattering of a propagating quark
in the medium, without the induced radiation,
can change the quark fragmentation or
hadron’s production rate. The net effect of
leading power contributions in medium length
is equivalent to a shift in the fragmentation
function’s $z_{h}\rightarrow z_{h}+\Delta z$, and
at the lowest order, $\Delta z$ is given by
an universal matrix element.
Our result is complementary to the
energy loss of induced radiation.
However, beyond the leading order, the separation
of the collisional energy loss and that of induced
radiation will depend on the factorization scheme and
is not unique, and needs further study.
Our approach can be generalized to hadron productions in $p+A$
and $A+A$ collisions.
References
[1]
S. S. Adler et al. [PHENIX Collaboration],
Phys. Rev. Lett. 91, 072301 (2003); and references therein.
[2]
M. Gyulassy and X. N. Wang,
Nucl. Phys. B 420, 583 (1994);
X. N. Wang, M. Gyulassy and M. Plumer,
Phys. Rev. D 51, 3436 (1995);
R. Baier, Y. L. Dokshitzer, A. H. Mueller, S. Peigne and D. Schiff,
Nucl. Phys. B 484, 265 (1997);
B. G. Zakharov,
JETP Lett. 63, 952 (1996); JETP Lett. 65, 615 (1997);
U. A. Wiedemann,
Nucl. Phys. B 588, 303 (2000)
[3]
S. S. Adler et al. [PHENIX Collaboration],
Phys. Rev. Lett. 96, 032301 (2006);
J. Bielcik [STAR Collaboration],
Nucl. Phys. A 774, 697 (2006)
[4]
Y. L. Dokshitzer and D. E. Kharzeev,
Phys. Lett. B 519, 199 (2001)
[5]
X. F. Guo, and J. Li, in preparation.
[6]
M. Luo, J.-W. Qiu and G. Sterman, Phys. Rev. D49, 4493 (1994).
[7]
J. W. Qiu and I. Vitev,
Phys. Rev. Lett. 93, 262301 (2004).
[8]
A. Airapetian et al. [HERMES Collaboration],
Eur. Phys. J. C 20, 479 (2001).
[9]
A. Accardi, in this proceedings.
[10]
X. F. Guo and X. N. Wang,
Phys. Rev. Lett. 85, 3591 (2000);
X. N. Wang and X. F. Guo,
Nucl. Phys. A 696, 788 (2001);
E. Wang and X. N. Wang,
Phys. Rev. Lett. 89, 162301 (2002). |
Theta Functions on Covers of Symplectic Groups
Solomon Friedberg
and
David Ginzburg
(corresponding author) [email protected]
Abstract.
We study the automorphic theta representation $\Theta_{2n}^{(r)}$ on the $r$-fold cover of the symplectic group $Sp_{2n}$. This representation is obtained from the residues of Eisenstein series on this group.
If $r$ is odd,
$n\leq r<2n$, then under a natural hypothesis on the theta representations, we show that
$\Theta_{2n}^{(r)}$ may be used to construct a generic representation
$\sigma_{2n-r+1}^{(2r)}$ on the $2r$-fold cover of $Sp_{2n-r+1}$. Moreover, when $r=n$ the
Whittaker functions of this representation attached to factorizable data
are factorizable, and the unramified local factors may be computed in terms of $n$-th order Gauss sums.
If $n=3$ we prove these results, which in that case pertain to the six-fold cover of $Sp_{4}$, unconditionally.
We expect that in fact the representation constructed here, $\sigma_{2n-r+1}^{(2r)}$, is precisely $\Theta_{2n-r+1}^{(2r)}$; that
is, we conjecture relations between theta representations on different covering groups.
Key words and phrases:Symplectic group, metaplectic cover, theta representation, descent integral, unipotent orbit, generic representation,
Whittaker function
2010 Mathematics Subject Classification: Primary 11F70; Secondary 11F27, 11F55
This work was supported by the BSF,
grant number 2012019, and by the NSF, grant number 1500977 (Friedberg).
1. Introduction
The classical metaplectic group is a double cover of a symplectic group, and may be defined over a local field or the ring of adeles of a number field. This group arises in the study of theta functions,
which may be constructed directly as sums over the global rational points of an isotropic subspace. This construction
of theta functions appears
to be special to the double cover case.
To generalize, let $r>1$ be any integer, and let
$F$ be a number field with a full set of $r$-th roots of unity. Then there is an $r$-fold cover of the symplectic
group $Sp_{2n}(\mathbb{A})$, where $\mathbb{A}$ is the adeles of $F$,
denoted $Sp_{2n}^{(r)}(\mathbb{A})$. When $r>2$ one may construct theta functions for such a group, but only indirectly – theta
functions are obtained
as the multi-residues of the minimal parabolic (Borel) Eisenstein series on $Sp_{2n}^{(r)}(\mathbb{A})$ at
the “right-most” (suitably normalized) pole. (This construction obtains when $r=2$ as well.)
As such, these theta functions are automorphic forms and give an
automorphic representation $\Theta_{2n}^{(r)}$, called the theta representation.
Since for a general cover the theta functions may only be defined via residues, it is considerably more challenging
to determine basic information such as their Whittaker or other Fourier
coefficients in this situation. Indeed, the Fourier coefficients of the theta function
have not been determined for covers of $SL_{2}$ of any degree higher than $4$. And when $n>1$ very little is known.
Moreover, except in the double cover case there are few examples of any automorphic functions on such a group, there
is no information about when their Whittaker coefficients might be factorizable (one expects rarely), and it is not clear
to what extent one might expect relations between theta functions or other automorphic
forms on different covers of different groups.
In a first example of such information, the authors [6] have established
relations between
the Whittaker coefficients of certain automorphic functions on different covering groups in two specific cases, related to the conjectures
of Patterson and Chinta-Friedberg-Hoffstein concerning the Fourier coefficients of theta functions on
covers of $GL_{2}$ . The key to doing so was to adapt
descent methods, originally used by Ginzburg-Rallis-Soudry [9] in the context of
algebraic groups or double covers, to higher degree covering groups. In particular, the treatment of
the adelic version of the conjecture
of Chinta-Friedberg-Hoffstein relied on the study of the theta function on the three-fold cover of $Sp_{4}$.
In this work we investigate theta functions and descent integrals on covers of symplectic groups of arbitrary rank.
As we shall explain, for $r$ odd a descent construction allows one to pass from
the theta representation $\Theta_{2n}^{(r)}$ to a representation $\sigma_{2n-r+1}^{(2r)}$ realized
inside the space of automorphic square-integrable functions on
$Sp^{(2r)}_{2n-r+1}(F)\backslash Sp^{(2r)}_{2n-r+1}(\mathbb{A})$. We expect that this representation is in fact the
theta representation $\Theta_{2n-r+1}^{(2r)}$ (Conjecture 2), and we show that $\sigma_{2n-r+1}^{(2r)}$ has non-zero
projection to this representation.
In Theorem 1 below we establish that if $r$ is odd,
$n\leq r<2n$, then the representation $\sigma_{2n-r+1}^{(2r)}$ is generic. Moreover, if $r=n$,
we show that the Whittaker coefficients of
$\sigma_{n+1}^{(2n)}$ arising from factorizable inputs are factorizable.
As a first new case, this is true of the Whittaker coefficients
of the descent representation on the six-fold cover of $Sp_{4}$.
Then in Theorem 2 we show that the unramified local
contributions to the Whittaker coefficients in the factorizable
case may be expressed as sums of Whittaker coefficients of the theta representation on the $n$-fold cover of $GL_{n}$.
These coefficients are $n$-th order Gauss sums by the work of Kazhdan and Patterson [12]. Thus we exhibit a new
class of automorphic forms on covering groups whose Whittaker coefficients are factorizable with algebraic factors at good places.
These results are proved on
two conditions. One concerns the specific unipotent orbit attached to $\Theta_{2n}^{(r)}$ (Conjecture 1).
This is an object of on-going study [4]; in the last Section we give a sketch of the proof of this Conjecture in the
first new case, $n=r=3$. A
second concerns the precise characters within a given unipotent orbit which support a nonzero Fourier coefficient
(Assumption 1). This is not needed in all cases, and in particular is not needed in the factorizable case
$\sigma_{n+1}^{(2n)}$.
Our results concerning Conjecture 1 are thus sufficient to establish unconditionally that
the Whittaker function of the descent to the six-fold cover of $Sp_{4}$ is Eulerian for factorizable data, and to compute its unramified
local factor in terms of cubic Gauss sums. We expect that this descent is in fact the theta function on the six-fold cover of $Sp_{4}$.
2. Definition and Properties of the Theta Representation
Let $n$ and $r$ be two natural numbers. Let $F$ be a number field containing a full set $\mu_{r}$ of $r$-th roots of unity,
and let $\mathbb{A}$ be the adeles of $F$. Let $Sp_{2n}$ denote the symplectic group consisting of the $2n\times 2n$ matrices leaving
invariant the standard symplectic form $<x,y>=\sum_{i=1}^{n}(x_{i}y_{2n-i+1}-x_{n+i}y_{n-i+1})$.
Denote by $Sp_{2n}^{(r)}(\mathbb{A})$ the metaplectic $r$-fold cover of the symplectic group $Sp_{2n}(\mathbb{A})$. This group
consists of pairs $(g,\zeta)$ with $g\in Sp_{2n}(\mathbb{A})$ and $\zeta\in\mu_{r}$; it
may be
obtained in the standard way from covers of the local groups $Sp_{2n}(F_{\nu})$ as $\nu$ ranges over the places of $F$,
identifying the copies of $\mu_{r}$. The multiplication in the local group is determined by a cocyle $\sigma$.
If $\sigma_{BLS}$ is the 2-cocycle of
Banks-Levy-Sepanski [1] for $SL_{2n}(F_{\nu})$ and $w$ is the permutation such that the conjugate of $Sp_{2n}$
by $w$ preserves the symplectic form $<x,y>^{\prime}=\sum_{i=1}^{n}x_{i}y_{n+i}-y_{i}x_{n+i}$, then we take
$\sigma(g_{1},g_{2})=\sigma_{BLS}(wg_{1}w^{-1},wg_{2}w^{-1})$.
We shall be concerned with the theta representation defined on the group $Sp_{2n}^{(r)}(\mathbb{A})$, denoted $\Theta_{2n}^{(r)}$. This representation is defined as
the space spanned by the residues of the Borel Eisenstein series on $Sp_{2n}^{(r)}({\mathbb{A}})$, similarly to the definition for the general linear
group in [12], p. 118. Our first task is to give a brief account of this construction.
Let $B_{2n}$ denote the standard Borel subgroup of $Sp_{2n}$, and let $T_{2n}$ denote the maximal torus of $Sp_{2n}$ which is a subgroup of $B_{2n}$. Working either locally at a completion of $F$ or globally over $\mathbb{A}$, if $H$ is any algebraic subgroup of $Sp_{2n}$, we let $H^{(r)}$ denote its inverse image in $Sp_{2n}^{(r)}$. We will call this inverse image parabolic or unipotent if $H$ has this property.
Let $Z(T_{2n}^{(r)})$ denote the center of $T_{2n}^{(r)}$.
Suppose first that $r$ is odd.
Given a character $\mu$ of $T_{2n}$, we may use it to define a genuine character,
again denoted $\mu$, of $Z(T_{2n}^{(r)})$.
(The notion of genuine depends on an embedding of the group of $r$-th roots of unity $\mu_{r}$ into $\mathbb{C}^{\times}$. We will fix this and omit it from the notation.)
Extending it trivially to any maximal abelian subgroup of $T_{2n}^{(r)}$ and then inducing up, we obtain a representation of $Sp_{2n}^{(r)}$ which we denote by $\text{Ind}_{B_{2n}^{(r)}}^{Sp_{2n}^{(r)}}\mu$. This representation is determined uniquely by the choice of $\mu$ and this procedure may be carried out both locally and globally. Here we consider
unnormalized induction, and shall include the modular character $\delta_{B_{2n}}^{1/2}$ when we require normalized induction.
Let $s_{i}$ be complex variables, and let $\mu$ be the character of $T_{2n}$ given by
$$\mu(\text{diag}(a_{1},\ldots,a_{n},a_{n}^{-1},\ldots,a_{1}^{-1}))=|a_{1}|^{s_{%
1}}\cdots|a_{n}|^{s_{n}}.$$
If this construction is carried out over a local field, then $|a|$ denotes the normalized local absolute value, while if it is
carried out over $\mathbb{A}$, then $|a|$ denotes the product of these over all places of $F$.
One may form the induced representation $\text{Ind}_{B_{2n}^{(r)}({\mathbb{A}})}^{Sp_{2n}^{(r)}({\mathbb{A}})}\mu\delta%
_{B_{2n}}^{1/2}$, and
for each vector in this space, one may construct the multi-variable Eisenstein series $E(h,s_{1},\ldots,s_{n})$ defined on the group $Sp_{2n}^{(r)}({\mathbb{A}})$. Computing the constant term as in [12], Prop. II.1.2, we deduce that the poles of the partial intertwining operator associated with the long Weyl element of $Sp_{2n}$ are determined by
(1)
$$\frac{\prod_{i<j}[\zeta_{S}(r(s_{i}-s_{j}))\zeta_{S}(r(s_{i}+s_{j}))]\prod_{i}%
\zeta_{S}(rs_{i})}{\prod_{i<j}[\zeta_{S}(r(s_{i}-s_{j})+1)\zeta_{S}(r(s_{i}+s_%
{j})+1)]\prod_{i}\zeta_{S}(rs_{i}+1)}.$$
Here $S$ is a finite set of all places including the archimedean places and all finite places $\mu$ such that $|r|_{\nu}\neq 1$
and $\zeta_{S}(s)$ is the partial global zeta function.
The expression (1) has a multi-residue at the point
$$s_{n}=\frac{1}{r},\qquad r(s_{i}-s_{i+1})=1.$$
From this we deduce that the Eisenstein series $E(h,s_{1},\ldots,s_{n})$ has a multi-residue at that point, and we denote the residue representation by $\Theta_{2n}^{(r)}$. If
$\mu_{0}$ denotes the character of $T_{2n}$ defined by
$$\mu_{0}(\text{diag}(a_{1},\ldots,a_{n},a_{n}^{-1},\ldots,a_{1}^{-1}))=|a_{1}|^%
{\frac{n}{r}}\cdots|a_{n-1}|^{\frac{2}{r}}|a_{n}|^{\frac{1}{r}},$$
then it follows that $\Theta_{2n}^{(r)}$ is a subquotient of the induced representation $\text{Ind}_{B_{2n}^{(r)}({\mathbb{A}})}^{Sp_{2n}^{(r)}({\mathbb{A}})}\mu\delta%
_{B_{2n}}^{1/2}$. From this we deduce that $\Theta_{2n}^{(r)}$ is also a subrepresentation of the induced representation
$\text{Ind}_{B_{2n}^{(r)}({\mathbb{A}})}^{Sp_{2n}^{(r)}({\mathbb{A}})}\chi_{Sp_%
{2n}^{(r)},\Theta}$ where
$$\chi_{Sp_{2n}^{(r)},\Theta}(\text{diag}(a_{1},\ldots,a_{n},a_{n}^{-1},\ldots,a%
_{1}^{-1}))=|a_{1}|^{\frac{n(r-1)}{r}}\cdots|a_{n-1}|^{\frac{2(r-1)}{r}}|a_{n}%
|^{\frac{r-1}{r}}.$$
We turn to even degree coverings. We shall only be concerned with the covers of degree twice an odd integer.
If $r$ is odd, the definition of the theta representation for $Sp_{2m}^{(2r)}$ is similar.
There is a small difference since the maximal parabolic subgroup of $Sp_{2m}$ whose Levi part is $GL_{m}$ splits under the double cover. Because of this, when computing the intertwining operator corresponding to the long Weyl element,
one finds that its poles are determined by
$$\frac{\prod_{i<j}[\zeta_{S}(r(s_{i}-s_{j}))\zeta_{S}(r(s_{i}+s_{j}))]\prod_{i}%
\zeta_{S}(2rs_{i})}{\prod_{i<j}[\zeta_{S}(r(s_{i}-s_{j})+1)\zeta_{S}(r(s_{i}+s%
_{j})+1)]\prod_{i}\zeta_{S}(2rs_{i}+1)}.$$
Accordingly we define $\Theta_{2m}^{(2r)}$ to be the multi-residue of the Eisenstein series at the point
$$s_{m}=\frac{1}{2r},\qquad r(s_{i}-s_{i+1})=1.$$
Then the representation $\Theta_{2m}^{(2r)}$ is a subquotient of the
induced representation $\text{Ind}_{B_{2m}^{(r)}({\mathbb{A}})}^{Sp_{2m}^{(2r)}({\mathbb{A}})}\mu_{e}%
\delta_{B_{2m}}^{1/2}$ where $\mu_{e}$ denotes the character of $T_{2m}$ defined by
$$\mu_{e}(\text{diag}(a_{1},\ldots,a_{m},a_{m}^{-1},\ldots,a_{1}^{-1}))=|a_{1}|^%
{\frac{2m-1}{2r}}\cdots|a_{m-1}|^{\frac{3}{2r}}|a_{m}|^{\frac{1}{2r}}.$$
It follows that $\Theta_{2m}^{(2r)}$ is also a subrepresentation of the induced representation $\text{Ind}_{B_{2m}^{(r)}({\mathbb{A}})}^{Sp_{2m}^{(2r)}({\mathbb{A}})}\chi_{Sp%
_{2m}^{(2r)},\Theta}$ where
$$\chi_{Sp_{2n}^{(2r)},\Theta}(\text{diag}(a_{1},\ldots,a_{m},a_{m}^{-1},\ldots,%
a_{1}^{-1}))=|a_{1}|^{\frac{2mr-(2m-1)}{2r}}\cdots|a_{m-2}|^{\frac{6r-5}{2r}}|%
a_{m-1}|^{\frac{4r-3}{2r}}|a_{m}|^{\frac{2r-1}{2r}}.$$
We now develop some properties of the theta representations that follow from induction in stages, or, equivalently,
by taking an $(n-1)$-fold residue of $E(h,s_{1},\dots,s_{n})$ to obtain a maximal parabolic Eisenstein series attached to
theta representations of lower rank groups.
Let $1\leq a\leq n$. Denote by $P_{2n,a}$ the maximal parabolic subgroup of $Sp_{2n}$ whose Levi part is $GL_{a}\times Sp_{2(n-a)}$, and let $L_{2n,a}$ denote its unipotent radical. We write $i$ for the inclusion of $GL_{a}$ in $P_{2n,a}$ in this paragraph
but we suppress $i$ afterwards.
From the block compatibility of the cocycle $\sigma_{BLS}$ (Banks-Levy-Sepanski [1]) and a short computation it follows
that if $g_{1},g_{2}\in GL_{a}$ over a local field then
$\sigma(i(g_{1}),i(g_{2}))=\sigma_{GL_{a}}(g_{1},g_{2})^{2}(\det g_{1},\det g_{%
2})^{-1}$, where $\sigma_{GL_{a}}$ is the metaplectic 2-cocycle
of [1] for $GL_{a}$.
Thus for both the $r$ and $2r$-fold covers, $i$ determines an $r$-fold cover of $GL_{a}$. We write this cover $GL_{a}^{(r)}$ (which case we are in will always be clear from context).
Suppose first that $r$ is odd.
Let $\Theta_{GL_{a}}^{(r)}$ denote the theta representation of the group $GL_{a}^{(r)}({\mathbb{A}})$, as constructed in [12]
(or the corresponding character if $a=1$).
Then it follows from induction in stages that
$\text{Ind}_{B_{2n}^{(r)}({\mathbb{A}})}^{Sp_{2n}^{(r)}({\mathbb{A}})}\mu_{0}%
\delta_{B_{2n}}^{1/2}$ is equal to $\text{Ind}_{P_{2n,a}^{(r)}({\mathbb{A}})}^{Sp_{2n}^{(r)}({\mathbb{A}})}(\Theta%
_{GL_{a}}^{(r)}\otimes\Theta_{2(n-a)}^{(r)})\delta_{P_{2n,a}}^{\frac{r+1}{2r}}$.
From this we deduce the following. Let $E_{a}(g,s)$ denote the Eisenstein series of $Sp_{2n}^{(r)}({\mathbb{A}})$ associated with the induced representation $\text{Ind}_{P_{2n,a}^{(r)}({\mathbb{A}})}^{Sp_{2n}^{(r)}({\mathbb{A}})}(\Theta%
_{GL_{a}}^{(r)}\otimes\Theta_{2(n-a)}^{(r)})\delta_{P_{2n,a}}^{s}$.
(For the construction of such Eisenstein series more generally, see Brubaker and Friedberg [2].)
Then the representation
$\Theta_{2n}^{(r)}$ is a residue of $E_{a}(g,s)$ at the point $s=(r+1)/2r$. Of course, this can also be verified directly by studying the corresponding intertwining operators.
In the case of a cover of degree $2r$, $r$ odd, the situation is roughly similar.
In this case, the representation $\text{Ind}_{B_{2m}^{(r)}({\mathbb{A}})}^{Sp_{2m}^{(2r)}({\mathbb{A}})}\mu_{e}%
\delta_{B}^{1/2}$ is equal to $\text{Ind}_{P_{2m,a}^{(r)}({\mathbb{A}})}^{Sp_{2m}^{(2r)}({\mathbb{A}})}(%
\Theta_{GL_{a}}^{(r)}\otimes\Theta_{2(m-a)}^{(2r)})\delta_{P_{2n,a}}^{\frac{(r%
+1)(2m-a)+r}{2r(2m-a+1)}}$. Hence if we let $E_{a}(g,s)$ denote the Eisenstein series associated with the induced representation $\text{Ind}_{P_{2m,a}^{(r)}({\mathbb{A}})}^{Sp_{2m}^{(2r)}({\mathbb{A}})}(%
\Theta_{GL_{a}}^{(r)}\otimes\Theta_{2(m-a)}^{(2r)})\delta_{P_{2n,a}}^{s}$, then we deduce that $\Theta_{2m}^{(2r)}$ is the residue of $E_{a}(g,s)$ at the point $s=\frac{(r+1)(2m-a)+r}{2r(2m-a+1)}$.
From these observations, we will deduce the following proposition. Here and below,
matrices are embedded in metaplectic groups by the trivial
section without additional notation, and we call metaplectic elements diagonal or unipotent when their projections to the
linear group are.
Proposition 1.
Suppose that $r$ is odd.
Let $\theta_{2n}^{(r)}$ be in the space of $\Theta_{2n}^{(r)}$. Then there exist functions $\theta_{GL_{a}}^{(r)}\in\Theta_{GL_{a}}^{(r)}$,
$\theta_{2(n-a)}^{(r)}\in\Theta_{2(n-a)}^{(r)}$
such that for all diagonal $g$ in $GL_{a}^{(r)}({\mathbb{A}})$ which lies in the center of the Levi part of the parabolic group $P_{2n,a}^{(r)}({\mathbb{A}})$ and for all unipotent $h\in Sp^{(r)}_{2(n-a)}({\mathbb{A}})$, $v\in GL_{a}^{(r)}({\mathbb{A}})$, one has
(2)
$$\int\limits_{L_{2n,a}(F)\backslash L_{2n,a}({\mathbb{A}})}\theta_{2n}^{(r)}(u(%
gv,h))du=\chi_{Sp_{2n}^{(r)},\Theta}(g)\,\theta_{GL_{a}}^{(r)}(v)\,\theta_{2(n%
-a)}^{(r)}(h).$$
A similar identity holds in the even-degree cover case.
The requirement that $h,v$ be unipotent could be dropped at the expense of a cocyle; being unipotent guarantees that it is $1$.
Proof.
Since $\Theta_{2n}^{(r)}$ is the residue of the Eisenstein series $E_{a}(\cdot,s)$, we first consider the constant term
(3)
$$\int\limits_{L_{2n,a}(F)\backslash L_{2n,a}({\mathbb{A}})}E_{a}(u(gv,h),s)\,du.$$
For $\text{Re}(s)$ large we can unfold the Eisenstein series and deduce that (3) is equal to
(4)
$$\sum_{w\in P_{2n,a}(F)\backslash Sp_{2n}(F)/P_{2n,a}(F)}\int\limits_{L_{2n,a}^%
{w}(F)\backslash L_{2n,a}({\mathbb{A}})}\sum_{\gamma\in(w^{-1}P_{2n,a}(F)w\cap
P%
_{2n,a}(F))\backslash P_{2n,a}(F)}f_{s}(w\gamma u(gv,h))\,du.$$
Here $L_{2n,a}^{w}=w^{-1}L_{2n,a}w\cap L_{2n,a}^{-}$ where $L_{2n,a}^{-}$ is the conjugate of $L_{2n,a}$ by the long Weyl element. Also, we have $f_{s}\in\text{Ind}_{P_{2n,a}^{(r)}({\mathbb{A}})}^{Sp_{2n}^{(r)}({\mathbb{A}})%
}(\Theta_{GL_{a}}^{(r)}\otimes\Theta_{2(n-a)}^{(r)})\delta_{P_{2n,a}}^{s}$.
Notice that all elements in $P_{2n,a}(F)\backslash Sp_{2n}(F)/P_{2n,a}(F)$ can be chosen to be Weyl elements.
Similarly to [13], Section 1.2, one can check that for all Weyl elements in $P_{2n,a}(F)\backslash Sp_{2n}(F)/P_{2n,a}(F)$
which are not the long Weyl element, the inner summation is a certain Eisenstein series or product
of such series. Moreover, one can also check that the Eisenstein series in
the corresponding summand in (4) is holomorphic at $s=(r+1)/2r$. Hence, taking the residue in (3) and (4) at $s=(r+1)/2r$, we are left only with the long
Weyl element, and for that element we obtain the identity
$$\int\limits_{L_{2n,a}(F)\backslash L_{2n,a}({\mathbb{A}})}\theta_{2n}^{(r)}(u(%
gv,h))du=\text{Res}_{s=(r+1)/2r}M_{w}f_{s}(gv,h)$$
where $M_{w}$ is the intertwining operator attached to $w$.
From this identity (2) follows.
∎
Similar statements for other groups are established in [3] Proposition 3.4
and [6] Proposition 1.
We end this section with a general conjecture about the unipotent orbit attached to the representation $\Theta_{2n}^{(r)}$,
and give a consequence of this conjecture.
Recall that if $\pi$ denotes an automorphic representation of a reductive group,
then the set ${\mathcal{O}}(\pi)$ was defined in [7]. It is the largest unipotent orbit that supports a nonzero coefficient
for this representation.
The extension of the definition of this set to the metaplectic groups is
clear.
In this paper we are interested in the set ${\mathcal{O}}(\Theta_{2n}^{(r)})$. The conjecture regarding this set is given as follows. Assume that $r<2n$. Write $2n=a(n,r)r+b(n,r)$ where $a(n,r)$ and $b(n,r)$ are both nonnegative numbers such that $0\leq b(n,r)\leq r-1$.
We recall that a partition is a symplectic partition if any odd number in the partition occurs with even multiplicity. As defined in [5], given a partition $\lambda$ of $2n$, we define the $Sp$ collapse of $\lambda$ to be the greatest symplectic partition which is smaller than $\lambda$. We have
Conjecture 1.
Let $r$ be an odd number.
If $r<2n$, then the set ${\mathcal{O}}(\Theta_{2n}^{(r)})$ consists of the partition which is the $Sp$ collapse of the partition $(r^{a(n,r)}b(n,r))$. If $r>2n$, then the representation $\Theta_{2n}^{(r)}$ is generic.
The conjecture is studied in [4]. We shall write Conjecture $1_{r,n}$ for this statement for a specific pair $(r,n)$.
We remark that an extension to even degree covers has not been formulated at this time.
To give a consequence of this conjecture, let $r$ be odd, let $a$ be an integer with $1\leq a\leq n-(r-1)/2$,
and let $b=n-(r-1)/2-a$.
Let $U_{2n,n-b}$ be the unipotent radical of the standard parabolic subgroup of $Sp_{2n}$ whose Levi part is $GL_{1}^{n-b}\times Sp_{2b}$. (In particular, $U_{2n,n}$ is the standard maximal unipotent subgroup of $Sp_{2n}$.)
Let $U_{2n,n-b,1}$ be the subgroup of $U_{2n,n-b}$ which consists of all matrices $u=(u_{i,j})\in U_{2n,n-b}$ such that $u_{n-b,i}=0$ for all $n-b+1\leq i\leq n$. Let $\underline{\alpha}=(\alpha_{1},\ldots,\alpha_{a-1})$ where for all $i$ we have $\alpha_{i}\in\{0,1\}$. Let $\psi$ be a nontrivial character of $F\backslash{\mathbb{A}}$,
which will be fixed throughout this paper.
Let
$\psi_{U_{2n,n-b,1},\underline{\alpha}}$ be the character of $U_{2n,n-b,1}$ given by
$$\psi_{U_{2n,n-b,1},\underline{\alpha}}(u)=\psi\left(\sum_{i=1}^{a-1}\alpha_{i}%
u_{i,i+1}+\sum_{j=a}^{n-b}u_{j,j+1}\right).$$
Then we have the following result, which will be used later.
Lemma 1.
Suppose $r$ is odd, $r<2n$, and Conjecture 1${}_{r,n-i_{0}}$ holds for all $0\leq i_{0}\leq a-1$. Then
the integral
(5)
$$\int\limits_{U_{2n,n-b,1}(F)\backslash U_{2n,n-b,1}({\mathbb{A}})}\theta_{2n}^%
{(r)}(ug)\,\psi_{U_{2n,n-b,1},\underline{\alpha}}(u)\,du$$
is zero for all choices of data, that is, for all $\theta_{2n}^{(r)}\in\Theta_{2n}^{(r)}$.
Proof.
Consider first the case where $\alpha_{i}=1$ for all $i$. Then the Fourier coefficient (5) is the Fourier coefficient which corresponds to the unipotent orbit $((2n-2b)1^{2b})$. From Conjecture 1${}_{r,n}$ we
have that ${\mathcal{O}}(\Theta_{2n}^{(r)})$ consists of the partition which is the $Sp$ collapse of the partition $(r^{a(n,r)}b(n,r))$. Hence $((2n-2b)1^{2b})$ is greater than or not related to ${\mathcal{O}}(\Theta_{2n}^{(r)})$. Indeed, this follows from the relation $2b=2n-2a-r+1$ which implies that $((2n-2b)1^{2b})=((r+2a-1)1^{2b})$. Thus, if $\alpha_{i}=1$ for all $i$, then the integral (5) is zero for all choices of data.
Next assume that at least one of the scalars $\alpha_{i}$ is zero. Let $i_{0}\leq a-1$ be the largest index such that $\alpha_{i_{0}}=0$. Then the integral (5) is equal to
(6)
$$\displaystyle\int\limits_{H(F)\backslash H({\mathbb{A}})}\int\limits_{L_{2n,i_%
{0}}(F)\backslash L_{2n,i_{0}}({\mathbb{A}})}\int\limits_{U_{2(n-i_{0}),n-i_{0%
}-b,1}(F)\backslash U_{2(n-i_{0}),n-i_{0}-b,1}({\mathbb{A}})}\\
\displaystyle\theta_{2n}^{(r)}(luhg)\,\psi_{U_{2(n-i_{0}),n-i_{0}-b,1}}(u)\,%
\psi_{H,\underline{\alpha}}(h)\,du\,dl\,dh.$$
Here $H$ is a certain unipotent subgroup of $GL_{i_{0}}$ which will not be important to us.
Notice that the integration over $L_{2n,i_{0}}(F)\backslash L_{2n,i_{0}}({\mathbb{A}})$ is the constant term along this unipotent group. Therefore, it follows from Proposition 1 above
that the integration along the quotient space $U_{2(n-i_{0}),n-i_{0}-b,1}(F)\backslash U_{2(n-i_{0}),n-i_{0}-b,1}({\mathbb{A}})$ is the coefficient of the representation $\Theta_{2(n-i_{0})}^{(r)}$ corresponding to the unipotent orbit $((2(n-i_{0}-b))1^{2b})$. Thus the vanishing of the integral (6) will follow if this unipotent orbit is greater than or not related to the
orbit ${\mathcal{O}}(\Theta_{2(n-i_{0})}^{(r)})$. From Conjecture 1${}_{r,n-i_{0}}$, this is the case if $2(n-i_{0}-b)>r$. Since this
inequality follows from the relations $2(n-a-b)=r-1$ and $a>i_{0}$, the result is proved.
∎
3. The Descent Construction
Recall that $U_{2m,k}$ denotes the unipotent radical of the standard parabolic subgroup of
$Sp_{2m}$ whose Levi part is $GL_{1}^{k}\times Sp_{2(m-k)}$. The quotient group $U_{2m,k-1}\backslash U_{2m,k}$ may be identified with the Heisenberg group in $2(m-k)+1$ variables, ${\mathcal{H}}_{2(m-k)+1}$. Indeed, there is a homomorphism $l:U_{2m,k}\mapsto{\mathcal{H}}_{2(m-k)+1}$ which is onto, and whose kernel is the group $U_{2m,k-1}$.
For the rest of this section we suppose that $r>1$ is odd. Set $r^{\prime}=(r-1)/2$.
For $\alpha\in F^{*}$ let $\psi_{U,\alpha}$ be the character
of $U_{r-1,r^{\prime}}(F)\backslash U_{r-1,r^{\prime}}({\mathbb{A}})$ given by
$$\psi_{U,\alpha}(u)=\psi(u_{1,2}+u_{2,3}+\cdots+u_{r^{\prime}-1,r^{\prime}}+%
\alpha u_{r^{\prime},r^{\prime}+1})\qquad u=(u_{i,j}).$$
From now on we shall suppose that Conjecture 1${}_{r,r^{\prime}}$ holds, that is, that
the representation $\Theta_{r-1}^{(r)}$ is generic. This is known if $r=3$ [12].
Thus there is an $\alpha\in F^{*}$ and a choice of data, i.e. an automorphic function $\theta_{r-1}^{(r)}$ in the space of $\Theta_{r-1}^{(r)}$, such that the integral
(7)
$$\int\limits_{U_{r-1,r^{\prime}}(F)\backslash U_{r-1,r^{\prime}}({\mathbb{A}})}%
\theta_{r-1}^{(r)}(ug)\,\psi_{U,\alpha}(u)\,du$$
is not zero.
Replacing $\psi$ by $\psi^{\alpha^{-1}}$ and changing the function, we may assume that $\alpha=1$. In other words,
there is a choice of data $\theta_{r-1}^{(r)}$ such that the integral
(8)
$$\int\limits_{U_{r-1,r^{\prime}}(F)\backslash U_{r-1,r^{\prime}}({\mathbb{A}})}%
\theta_{r-1}^{(r)}(ug)\,\psi_{U}(u)\,du$$
is not zero, where $\psi_{U}$ is defined as $\psi_{U,\alpha}$ with $\alpha=1$.
Remark: In fact we expect that for all $\alpha\in F^{*}$, the integral (7) will be nonzero for some choice of data.
Let $\Theta_{2m}^{\psi,\phi}$ denote the theta representation attached to the Weil representation and defined on the double cover of $Sp_{2m}({\mathbb{A}})$. Here $\phi$ is a Schwartz function on ${\mathbb{A}}^{m}$.
Since $r$ is odd, using the isomorphism
$\mu_{2r}\cong\mu_{2}\times\mu_{r}$ we may map $g\in Sp_{2n-r+1}^{(2r)}({\mathbb{A}})$ to its image in the double and $r$-fold covers;
we will not introduce separate notation for this.
To define the descent construction, consider the function of $g\in Sp_{2n-r+1}^{(2r)}({\mathbb{A}})$ given by
(9)
$$f(g)=\int\limits_{U_{2n,r^{\prime}}(F)\backslash U_{2n,r^{\prime}}({\mathbb{A}%
})}\theta_{2n-r+1}^{\psi,\phi}(l(u)g)\,\theta_{2n}^{(r)}(ug)\,\psi_{U_{2n,r^{%
\prime}}}(u)\,du.$$
Here $\theta_{2n-r+1}^{\psi,\phi}$ and $\theta_{2n}^{(r)}$ are vectors in the spaces of the representations
$\Theta_{2n-r+1}^{\psi,\phi}$ and $\Theta_{2n}^{(r)}$ (resp.), $\psi_{U_{2n,r^{\prime}}}$ is the character of $U_{2n,r^{\prime}}({\mathbb{A}})$ given by $\psi_{U_{2n,r^{\prime}}}(u_{i,j})=\psi(u_{1,2}+u_{2,3}+\cdots+u_{r^{\prime}-1,%
r^{\prime}})$, and $g\in Sp_{2n-r+1}$ is embedded in $Sp_{2n}$ as
$$g\mapsto\begin{pmatrix}I_{r^{\prime}}&&\\
&g&\\
&&I_{r^{\prime}}\end{pmatrix}.$$
(One could also consider descent integrals similar to (9) if $r$ is even but we shall not do so here.)
Then $f(g)$ is a genuine automorphic function defined on $Sp_{2n-r+1}^{(2r)}({\mathbb{A}})$.
Let $\sigma_{2n-r+1}^{(2r)}$ denote the representation of $Sp_{2n-r+1}^{(2r)}({\mathbb{A}})$ generated by all the functions $f(g)$.
4. Computation of the Constant Term of the Descent Integral
Let $V$ denote any standard unipotent subgroup of $Sp_{2n-r+1}$, and let $\psi_{V}$
be a character of $V(F)\backslash V({\mathbb{A}})$, possibly trivial.
In this paper we will compute integrals of the type
(10)
$$\int\limits_{V(F)\backslash V({\mathbb{A}})}f(vg)\,\psi_{V}(v)\,dv,$$
where $f(g)$ is given by (9).
Using (9), we arrive at an iterated integral. We first unfold the theta function $\theta_{2n-r+1}^{\psi,\phi}$. Collapsing summation and integration, and then using the formulas for the action of the Weil representation
$\omega_{\psi}$, the integral (10) is equal to
(11)
$$\int\limits_{{\mathbb{A}}^{n-r^{\prime}}}\int\limits_{V(F)\backslash V({%
\mathbb{A}})}\int\limits_{U_{2n,r^{\prime},1}(F)\backslash U_{2n,r^{\prime},1}%
({\mathbb{A}})}\omega_{\psi}(g)\phi(x)\,\theta_{2n}^{(r)}(uvj(x)g)\,\psi_{U_{2%
n,r^{\prime},1}}(u)\,\psi_{V}(v)\,du\,dv\,dx.$$
Here $x=(x_{1},\ldots,x_{n-r^{\prime}})\in\mathbb{A}^{n-r^{\prime}}$ is embedded in $Sp_{2n}$ via the map
$j(x)=I_{2n}+x_{1}e^{*}_{r^{\prime},r^{\prime}+1}+x_{2}e^{*}_{r^{\prime},r^{%
\prime}+2}+\cdots+x_{n-r^{\prime}}e^{*}_{r^{\prime},n}$,
where $e_{i,j}$ denotes the $(i,j)$-th elementary matrix and $e^{*}_{i,j}=e_{i,j}-e_{2n-j+1,2n-i}$.
Recall that the group $U_{2n,r^{\prime},1}$ is the subgroup of $U_{2n,r^{\prime}}$ consisting of all matrices $u=(u_{i,j})\in U_{2n,r^{\prime}}$ such that $u_{r^{\prime},k}=0$ for all $r^{\prime}+1\leq k\leq n$. The character $\psi_{U_{2n,r^{\prime},1}}(u)$ is defined as the product of $\psi_{U_{2n,r^{\prime}}}(u)$ and the character $\psi^{0}_{U_{2n,r^{\prime},1}}$ of $U_{2n,r^{\prime},1}$ given by $\psi^{0}_{U_{2n,r^{\prime},1}}(u)=\psi(u_{r^{\prime},2n-r^{\prime}+1})$.
At this point we consider the case when $V=L_{2n-r+1,a}$ where $1\leq a\leq n-r^{\prime}$ and $\psi_{V}$ is the trivial character. Thus,
$V$ is the unipotent radical of the standard maximal parabolic subgroup of $Sp_{2n-r+1}$ whose
Levi part is $GL_{a}\times Sp_{2n-2a-r+1}$. We recall that $2b=2n-2a-r+1$.
Let $w_{a}$ denote the Weyl group element of $Sp_{2n}$ defined by
$$w_{a}=\begin{pmatrix}&I_{a}&&&\\
I_{n-a-b}&&&&\\
&&I_{2b}&&\\
&&&&I_{n-a-b}\\
&&&I_{a}&\end{pmatrix}.$$
Since $\theta_{2n}^{(r)}$ is left-invariant under rational points, we may conjugate by $w_{a}$ from left to right.
Doing so, we see that the integral (11) is equal to
(12)
$$\displaystyle\int\limits_{{\mathbb{A}}^{n-r^{\prime}}}\int\limits_{Mat_{(n-a-b%
-1)\times a}(F)\backslash Mat_{(n-a-b-1)\times a}({\mathbb{A}})}\int\limits_{V%
_{0}(F)\backslash V_{0}({\mathbb{A}})}\int\limits_{U_{0}(F)\backslash U_{0}({%
\mathbb{A}})}\\
\displaystyle\omega_{\psi}(g)\phi(x)\,\theta_{2n}^{(r)}(u_{0}v_{0}k(y)w_{a}j(x%
)g)\,\psi_{V_{0}}(v_{0})\,du_{0}\,dv_{o}\,dy\,dx.$$
The notations are defined as follows. Recall that $L_{2n,a}$ denotes the unipotent radical of the standard maximal parabolic subgroup of $Sp_{2n}$ whose Levi part is $GL_{a}\times Sp_{2(n-a)}$. The group $U_{0}$ is defined to be the subgroup of $L_{2n,a}$ consisting of all matrices in $Sp_{2n}$ of the form
$$\begin{pmatrix}I_{a}&0&*&*&*\\
&I_{n-a-b}&0&0&*\\
&&I_{2b}&0&*\\
&&&I_{n-a-b}&0\\
&&&&I_{a}\end{pmatrix}.$$
The embedding of $Mat_{(n-a-b-1)\times a}$ into $Sp_{2n}$ is given by
$$k(y)=\begin{pmatrix}I_{a}&&&&\\
y&I_{n-a-b}&&&\\
&&I_{2b}&&\\
&&&I_{n-a-b}&\\
&&&y^{*}&I_{a}\end{pmatrix},$$
where $y^{*}$ is chosen to make the matrix symplectic. Here the embedding of $y$ is such that it consists of all matrices of size $(n-a-b)\times a$ such that its bottom row is zero. Finally $V_{0}$ is equal to the group $U_{2(n-a),n-a-b,1}$, defined similarly to the definition of $U_{2n,a,1}$ reviewed after (11) (in particular, the group $V_{0}$ is a subgroup of $Sp_{2(n-a)}$), and
the character $\psi_{V_{0}}:=\psi_{U_{2(n-a),n-a-b},1}$ as defined there. In coordinates,
$\psi_{V_{0}}(v)=\psi(v_{a+1,a+2}+v_{a+2,a+3}+\cdots+v_{n-b-1,n-b}+v_{n-b,n+b+1})$.
The next step is to carry out some root exchanges. The notion of root exchange was described in detail in the authors’
paper [6], Section 2.2; a more abstract formulation may be found in Ginzburg, Rallis and Soudry
[11], Section 7.1.
In our context, we perform the following root exchange. For $z\in Mat_{a\times(n-a-b)}$, let $m(z)$ in $Sp_{2n}$ be given by
$$m(z)=\begin{pmatrix}I_{a}&z&&&\\
&I_{n-a-b}&&&\\
&&I_{2b}&&\\
&&&I_{n-a-b}&z^{*}\\
&&&&I_{a}\end{pmatrix}.$$
We embed the group $Mat_{a\times(n-a-b-1)}(\mathbb{A})$ inside $Sp_{2n}(\mathbb{A})$ by considering all matrices $m(z)$ as above such that the first column of $z$ is zero. Expand the integral (12) along the group so obtained. Performing the root exchange with the group of all matrices $k(y)$ such that $y\in Mat_{(n-a-b-1)\times a}({\mathbb{A}})$, one sees that the integral (12) is equal to
(13)
$$\displaystyle\int\limits_{{\mathbb{A}}^{n-r^{\prime}}}\int\limits_{Mat_{(n-a-b%
-1)\times a}({\mathbb{A}})}\int\limits_{V_{0}(F)\backslash V_{0}({\mathbb{A}})%
}\int\limits_{L^{0}_{2n,a}(F)\backslash L^{0}_{2n,a}({\mathbb{A}})}\\
\displaystyle\omega_{\psi}(g)\phi(x)\,\theta_{2n}^{(r)}(uv_{0}k(y)w_{a}j(x)g)%
\,\psi_{V_{0}}(v_{0})\,du\,dv_{0}\,dy\,dx.$$
Here $L^{0}_{2n,a}$ is the subgroup of $L_{2n,a}$ which consists of all matrices $u=(u_{i,j})\in L_{2n,a}$ such that $u_{i,a+1}=0$ for all $1\leq i\leq a$.
Consider the quotient space of $L^{0}_{2n,a}\backslash L_{2n,a}$.
This quotient may be naturally identified with the column vectors of size $a$. This group embeds in $Sp_{2n}$ as the group of all matrices of the form $I_{2n}+z_{1}e^{*}_{1,a+1}+z_{2}e^{*}_{2,a+1}+\cdots+z_{a}e^{*}_{a,a+1}$. Expand the integral (13) along this group. The group $GL_{a}(F)$ acts on this expansion with two orbits.
The contribution of the nontrivial orbit to the integral (13) is the expression
(14)
$$\displaystyle\int\limits_{{\mathbb{A}}^{n-r^{\prime}}}\int\limits_{Mat_{(n-a-b%
-1)\times a}({\mathbb{A}})}\int\limits_{V_{0}(F)\backslash V_{0}({\mathbb{A}})%
}\int\limits_{L_{2n,a}(F)\backslash L_{2n,a}({\mathbb{A}})}\\
\displaystyle\omega_{\psi}(g)\phi(x)\,\theta_{2n}^{(r)}(uv_{0}k(y)w_{a}j(x)g)%
\,\psi_{V_{0}}(v_{0})\,\psi_{L_{2n,a}}(u)\,du\,dv_{0}\,dy\,dx,$$
where for $u=(u_{i,j})\in L_{2n,a}(F)\backslash L_{2n,a}({\mathbb{A}})$, $\psi_{L_{2n,a}}(u)=\psi(u_{a,a+1})$.
We claim that the integral (14) is zero for all choices of data.
To prove this, we carry out similar expansions repeatedly. To start, consider the quotient space $L_{2n,a-1}\backslash L_{2n,a}$. This space may be identified with the column vectors of size $a-1$, embedded in $Sp_{2n}$ as the group of all matrices of the form $I_{2n}+z_{1}e^{*}_{1,a}+z_{2}e^{*}_{2,a}+\cdots+z_{a-1}e^{*}_{a-1,a}$. Expand the integral (14) along this group. The group $GL_{a-1}(F)$ acts on this expansion with two orbits. If $a-1>1$ we continue in this way. Then the vanishing
of the integral (14) for all choices of data then follows from Lemma 1 above.
We conclude that the integral (10) with $V=U_{2n-r+1,a}$ and $\psi_{V}=1$ is equal to
(15)
$$\displaystyle\int\limits_{{\mathbb{A}}^{n-r^{\prime}}}\int\limits_{Mat_{(n-a-b%
-1)\times a}({\mathbb{A}})}\int\limits_{V_{0}(F)\backslash V_{0}({\mathbb{A}})%
}\int\limits_{L_{2n,a}(F)\backslash L_{2n,a}({\mathbb{A}})}\\
\displaystyle\omega_{\psi}(g)\phi(x)\,\theta_{2n}^{(r)}(uv_{0}k(y)w_{a}j(x)g)%
\,\psi_{V_{0}}(v_{0})\,du\,dv_{0}\,dy\,dx.$$
We shall prove momentarily that (15) is nonzero for some choice of data. Supposing this for the moment,
let us analyze this integral when $g=\text{diag}(tI_{a},I_{2n-r-2a+1},t^{-1}I_{a})$ is an element in the center of the Levi part of the maximal parabolic subgroup of $Sp_{2n-r+1}$ whose Levi part is $GL_{a}\times Sp_{2n-r-2a+1}$, embedded in the $r$-fold metaplectic group via the trivial section. Suppose also that $t$ is an $r$-th power. Inserting $g$ into (15), we
move it from right to left in the function $\theta_{2n}^{(r)}$. Moving it past $j(x)$ and making the corresponding variable change in $x$ gives a factor of $|t|^{-a}$,
as well as a factor of $\gamma(t^{a})|t|^{a/2}$ from the action of the Weil representation. Then moving it past $k(y)$ we obtain a factor of $|t|^{-a(n-a-b-1)}$ from the corresponding change of variables in $y$. Clearly, $g$ commutes with $V_{0}$, and finally, applying Proposition 1 above we also get a factor of
$$\delta_{P_{2n,a}}^{\frac{r-1}{2r}}(\text{diag}(tI_{a},I_{2(n-a)},t^{-1}I_{a}))%
=|t|^{a(r-1)(2n-a+1)/2r}$$
(here, recall that $P_{2n,a}$ is the maximal parabolic subgroup of $Sp_{2n}$ whose unipotent radical is $L_{2n,a}$).
We conclude that, for $g=\text{diag}(tI_{a},I_{2(n-a)},t^{-1}I_{a})$ as above and for $V=U_{2n-r+1,a}$, we have
$$\int\limits_{V(F)\backslash V({\mathbb{A}})}f(vg)\,dv=|t|^{\beta}\int\limits_{%
V(F)\backslash V({\mathbb{A}})}f(v)\,dv$$
where
$$\beta=\frac{a(r-1)(2n-a+1)}{2r}-\frac{a}{2}-a(n-a-b-1).$$
It follows from Section 2 that $\chi_{Sp_{2n-r+1}^{(2r)},\Theta}(g)=|t|^{\beta}$.
Finally, we prove that the integral (15) is not zero for some choice of data. This will
also imply that the representation $\sigma_{2n-r+1}^{(2r)}$ is not zero.
Suppose instead that (15) is zero for all choices of data. Arguing as in [6], Lemma 1, we may ignore the integration over the $x$ and $y$ variables, and deduce that the integral
$$\int\limits_{V_{0}(F)\backslash V_{0}({\mathbb{A}})}\int\limits_{L_{2n,a}(F)%
\backslash L_{2n,a}({\mathbb{A}})\,}\theta_{2n}^{(r)}(uv_{0})\,\psi_{V_{0}}(v_%
{0})\,du\,dv_{0}$$
is zero for all choices of data. From Proposition 1 we deduce that the integral
(16)
$$\int\limits_{V_{0}(F)\backslash V_{0}({\mathbb{A}})}\theta_{2(n-a)}^{(r)}(v_{0%
}h)\,\psi_{V_{0}}(v_{0})\,dv_{0}$$
is zero for all choice of data. Here $h\in Sp_{2(n-a)}^{(r)}({\mathbb{A}})$. From the definition of the group $V_{0}$ and the character $\psi_{V_{0}}$ it follows that the above Fourier coefficient corresponds to the unipotent orbit $((2(n-a-b))1^{2b})$. Since $2b=2n-2a-r+1$, this unipotent orbit is the same as $((r-1)1^{2n-2a-r+1})=((r-1)1^{2b})$.
At this point we would like to use instances of Conjecture 1, but we need to be careful. The Fourier coefficients attached to the orbit $((r-1)1^{2b})$ are given by the integrals
(17)
$$\int\limits_{V_{0}(F)\backslash V_{0}({\mathbb{A}})}\theta_{2b+r-1}^{(r)}(v_{0%
}h)\,\psi_{V_{0},\alpha}(v_{0})\,dv_{0}$$
with $\alpha\in F^{*}$ and $\psi_{V_{0},\alpha}$ defined on $V_{0}\subset Sp_{2(n-a)}$ by
$$\psi_{V_{0},\alpha}(v)=\psi(v_{1,2}+v_{2,3}+\cdots+v_{r^{\prime}-1,r^{\prime}}%
+\alpha v_{r^{\prime},2b+r^{\prime}+1}).$$
Conjecture 1 asserts that there is an $\alpha$ such that the integral (17) is not zero for some choice of data. However, we need to prove that the integral (16) is not zero for some choice of data, that is, the integral (17) with $\alpha=1$. This does not follow from Conjecture 1.
To prove that integral (16) is not zero for some choice of data, we will use the non-vanishing of the integral (8). Indeed, suppose instead that (16) is zero for all choices of data. Then the integral
(18)
$$\int\limits_{L_{2b,b}(F)\backslash L_{2b,b}({\mathbb{A}})}\int\limits_{V_{0}(F%
)\backslash V_{0}({\mathbb{A}})}\theta_{2b+r-1}^{(r)}(v_{0}l)\,\psi_{V_{0}}(v_%
{0})\,dv_{0}\,dl$$
is zero for all choices of data. We recall that $L_{2b,b}$ is the unipotent radical of
the maximal parabolic subgroup of $Sp_{2b}$ whose Levi part is $GL_{b}$. Arguing as in the computation of (11), we obtain that the vanishing of (18) implies the vanishing of the integral
(19)
$$\int\limits_{U_{r-1,r^{\prime}}(F)\backslash U_{r-1,r^{\prime}}({\mathbb{A}})}%
\int\limits_{L_{2b+r-1,b}(F)\backslash L_{2b+r-1,b}({\mathbb{A}})}\theta_{2b+r%
-1}^{(r)}(lv_{0})\,\psi_{V_{0}}(v_{0})\,dl\,dv_{0}.$$
But the vanishing of (19) for all choice of data implies that (8) is zero for all choice of data,
by Proposition 1. This is a contradiction.
Since the constant terms have been bounded and the descent is a compact integral of an automorphic form, the descent
and its derivatives have uniformly moderate growth. Then from truncation as in the Corollary in [14], I.2.12, together with
the above calculations, one
obtains the following result.
Proposition 2.
Let $r$ be odd, $r<2n$, and suppose that Conjecture 1${}_{r,i}$ holds for all $i$ with $r^{\prime}\leq i\leq n$.
Then the representation $\sigma_{2n-r+1}^{(2r)}$ is not zero and it is a sub-representation of $L^{2}(Sp_{2n-r+1}^{(2r)}(F)\backslash Sp_{2n-r+1}^{(2r)}({\mathbb{A}}))$. Moreover its projection to the residual spectrum is
nonzero.
One could also deduce this following [6], Propositions 4 and 5, by a sequence
of expansions along abelian unipotent radicals. We omit the details, as in fact
we expect that a stronger statement holds (compare [6], Conjecture 1).
Conjecture 2.
The representation $\sigma_{2n-r+1}^{(2r)}$ is equal to $\Theta_{2n-r+1}^{(2r)}$.
5. The Whittaker Coefficient of the Descent
In this section we study those cases in which the descent representation $\sigma_{2n-r+1}^{(2r)}$ is generic.
Let us explain how these cases are predicted by the
dimension equation of the second author [8]. It follows from this equation
together with (9) that the following equation should hold, where the dimension of a representation means
its Gelfand-Kirillov dimension:
(20)
$$\text{dim}\ \Theta_{2n-r+1}^{\psi,\phi}\ +\text{dim}\ \Theta_{2n}^{(r)}=\text{%
dim}\ U_{2n,r^{\prime}}+\text{dim}\ \sigma_{2n-r+1}^{(2r)}.$$
Since $\Theta_{2n-r+1}^{\psi,\phi}$ is the minimal representation of $Sp_{2n-r+1}^{(2)}({\mathbb{A}})$,
its dimension is $\frac{1}{2}(2n-r+1)$. An easy computation shows that $\text{dim}\ U_{2n,r^{\prime}}=\frac{1}{4}(r-1)(4n-r+1)$. If we want the representation $\sigma_{2n-r+1}^{(2r)}$ to be generic,
its dimension should equal $\frac{1}{4}(2n-r+1)^{2}$. Using these values, (20) gives
(21)
$$\text{dim}\ \Theta_{2n}^{(r)}=n^{2}-n+\frac{1}{2}(r-1).$$
It is not hard to check that for a representation $\Theta_{2n}^{(r)}$ which satisfies both
equation (21) and Conjecture 1${}_{r,n}$, we must have $n\leq r<2n$. If $n=2k$, then
for all $0\leq i\leq k-1$ the dimension of the unipotent orbit $((4k-2i-2)(2i+2))$ is equal to
$n^{2}-n+\frac{1}{2}(r-1)$ with $r=4k-2i-1$. If $n=2k+1$, then
for all $0\leq i\leq k-1$ the dimension of the unipotent orbit $((4k-2i)(2i+2))$ is equal to
$n^{2}-n+\frac{1}{2}(r-1)$ with $r=4k-2i+1$. In this case we also have the unipotent orbit
$((2k+1)^{2})$ whose dimension is $n^{2}-n+\frac{1}{2}(r-1)$ with $r=2k+1$. The above argument motivates the following
result.
Theorem 1.
Suppose that $r$ is odd, $n\leq r<2n$. Suppose that Conjecture 1${}_{r,i}$ holds for $r^{\prime}\leq i\leq n$. If
$r\neq n,n+1$, suppose that Assumption 1 below holds as well.
Then the representation $\sigma_{2n-r+1}^{(2r)}$ is generic. Moreover, in the special case where $r=n$, the Whittaker coefficient of
$\sigma_{n+1}^{(2n)}$ attached to factorizable data $\theta_{2n}^{(n)}\in\Theta_{2n}^{(n)}$ and $\phi$ is factorizable.
We shall see below that the hypotheses concerning Conjecture 1${}_{r,i}$ are satisfied when $r=n=3$.
Proof.
The proof is based on a computation of the integral (10)
with $V$ the maximal unipotent subgroup of $Sp_{2n-r+1}$ and $\psi_{V}$ the Whittaker character of $V$.
According to Conjecture 1${}_{r,n}$ the unipotent orbit attached to the
representation $\Theta_{2n}^{(r)}$ is computed as follows.
Let $r=2n-2l+1$ for convenience, and
write $2n=a(n,r)(2n-2l+1)+b(n,r)$. Then $a(n,r)=1$ and $b(n,2)=2l-1$. Thus, the unipotent orbit
attached to the representation $\Theta_{2n}^{(r)}$ is the $Sp$ collapse of
$((2n-2l+1)(2l-1))$. This is $((2n-2l)(2l))$ unless $n=2l-1$, in which case it is $((2l-1)^{2})$. This latter is the case $n=r$.
We consider first the case where the unipotent orbit attached to the representation $\Theta_{2n}^{(r)}$
is of the form $((2n-2l)(2l))$. (In the above notations we have $l=i+1$.)
We first describe the set of Fourier coefficients attached to this unipotent orbit.
Let $P^{\prime}_{l}$ denote the standard parabolic subgroup of $Sp_{2n}$ whose Levi part is $GL_{1}^{n-2l}\times GL_{2}^{l}$. We embed this group in $Sp_{2n}$ as the group of all matrices of the form $\text{diag}(a_{1},\ldots,a_{n-2l},g_{1},\ldots,g_{l},g_{l}^{*},\ldots,g_{1}^{*%
},a_{2n-l}^{-1},\ldots,a_{1}^{-1})$. Here $a_{i}\in GL_{1}$ and $g_{j}\in GL_{2}$ and the starred entries are determined by the requirement that the matrix be in $Sp_{2n}$. Let $U^{\prime}_{l}$ denote the
unipotent radical of $P^{\prime}_{l}$. Then the Fourier coefficients of an automorphic function $\varphi$ attached to the
unipotent orbit $((2n-2l)(2l))$ are given by
(22)
$$\int\limits_{U^{\prime}_{l}(F)\backslash U^{\prime}_{l}({\mathbb{A}})}\varphi(%
ug)\,\psi_{U^{\prime}_{l},a,b}(u)\,du$$
where $a,b\in F^{*}$ and the character $\psi_{U^{\prime}_{l},a,b}$ is defined by
(23)
$$\psi_{U^{\prime}_{l},a,b}(u)=\psi\left(\sum_{i=1}^{n-2l}u_{i,i+1}+\sum_{j=n-2l%
+1}^{n-2}u_{j,j+2}+au_{n-1,n+2}+bu_{n,n+1}\right).$$
Thus, to assert that the unipotent orbit $((2n-2l)(2l))$ is attached to a certain
representation, means first that there is a choice of $a$ and $b$ such that the Fourier
coefficient (22) is not zero for some choice of data, and second that for all orbits
which are greater than or not related to $((2n-2l)(2l))$ all corresponding
Fourier coefficients are zero. However, Conjecture 1${}_{r,n}$ does not specify for which values
of $a$ and $b$ the integral (22) is nonzero for some choice of data.
To prove the Theorem, we will need a compatibility assumption which is stated as follows.
Assumption 1.
Suppose that $r\neq n,n+1$. Then there is a choice of vector $\varphi$ in the space of $\Theta_{2n}^{(r)}$
such that the integral (22) with $a=1$ is not zero.
This assumption, which will not be needed when the orbit is of the form $(n^{2})$,
is used to guarantee the non-vanishing of the
Whittaker coefficient of the descent.
At this point we
begin the computation of the Whittaker coefficient when $n\neq r$; the case $n=r$ will be treated separately. We compute the integral (10) with $V=U_{2n-r+1,n-r^{\prime}}$, that is,
(24)
$$\int\limits_{U_{2n-r+1,n-r^{\prime}}(F)\backslash U_{2n-r+1,n-r^{\prime}}({%
\mathbb{A}})}f(ug)\,\psi_{U,b}(u)\,du.$$
As above $r^{\prime}=(r-1)/2$ and the character $\psi_{U,b}$, $b\in F^{*}$, is given by
$$\psi_{U,b}=\psi(u_{1,2}+u_{2,3}+\cdots+u_{n-r^{\prime}-1,n-r^{\prime}}+bu_{n-r%
^{\prime},n-r^{\prime}+1}).$$
Similarly to the analysis of (10) and (11), the integral (24)
is equal to
(25)
$$\displaystyle\int\limits_{{\mathbb{A}}^{n-r^{\prime}}}\int\limits_{U_{2n-r+1,n%
-r^{\prime}}(F)\backslash U_{2n-r+1,n-r^{\prime}}({\mathbb{A}})}\int\limits_{U%
_{2n,r^{\prime},1}(F)\backslash U_{2n,r^{\prime},1}({\mathbb{A}})}\\
\displaystyle\omega_{\psi}(g)\phi(x)\,\theta_{2n}^{(r)}(uvj(x)g)\,\psi_{U_{2n,%
r^{\prime},1}}(u)\,\psi_{U,b}(v)\,du\,dv\,dx.$$
Define the following Weyl element $w_{0}$ of $Sp_{2n}$. First, let $w$ denote the Weyl element of
$GL_{n}$ given by $w=\begin{pmatrix}I_{n-2l+1}&\\
&\nu\end{pmatrix}$ with $l=n-r^{\prime}$ and $\nu=(\nu_{i,j})\in GL_{2l-1}$ defined
as follows. Let $\nu_{i,j}=1$ for the pairs
$(i,j)=(1,l);(3,l+1);(5,l+2);\ldots;(2l-1,2l-1)$ and $(i,j)=(2,1);(4,2);(6,3)\ldots(2l-2,l-1)$ and $\nu_{i,j}=0$ otherwise.
Then let $w_{0}=\begin{pmatrix}w&\\
&w^{*}\end{pmatrix}$ where the star indicates that the matrix is in $Sp_{2n}$.
Since $\theta_{2n}^{(r)}$ is left-invariant under $w_{0}$, inserting $w_{0}$ into $\theta_{2n}^{(r)}$ and moving it rightward,
the integral (25) is equal to
(26)
$$\int\limits_{{\mathbb{A}}^{n-r^{\prime}}}\int\limits_{V_{1}(F)\backslash V_{1}%
({\mathbb{A}})}\int\limits_{U^{\prime}_{l,1}(F)\backslash U^{\prime}_{l,1}({%
\mathbb{A}})}\omega_{\psi}(g)\phi(x)\,\theta_{2n}^{(r)}(uvw_{0}j(x)g)\,\psi_{U%
^{\prime}_{l},1,b}(u)\,du\,dv\,dx.$$
Here $U^{\prime}_{l,1}$ is the subgroup consisting of $u=(u_{i,j})\in U^{\prime}_{l}$ such that
$u_{2m,2k+1}=0$ for all $1\leq m\leq l-1$ and $m\leq k\leq l-1$, and the character $\psi_{U^{\prime}_{l},1,b}$ is the
character $\psi_{U^{\prime}_{l},a,b}$ defined in (23) with $a=1$. Also, the group $V_{1}$ is the unipotent
subgroup of $Sp_{2n}$ consisting of all matrices of the form $I_{2n}+\sum_{i,j}r_{i,j}e_{i,j}^{*}$ where
the sum is over all pairs $(i,j)=(2k+1,2m)$ with $1\leq m\leq l-1$ and $m-1\leq k\leq l-2$.
Let $Z_{l}$ denote the unipotent subgroup of $U^{\prime}_{l}$ consisting of all matrices of the form
$I_{2n}+\sum_{i,j}r_{i,j}e_{i,j}^{*}$ where the sum is over all pairs $(i,j)=(2m,2k+1)$ with
$1\leq m\leq l-1$ and $m\leq k\leq l-1$.
Then $Z_{l}$ is a subgroup of $U^{\prime}_{l}$ which satisfies the condition
that $U^{\prime}_{l}=Z_{l}\cdot U^{\prime}_{l,1}$. We perform a root exchange in the integral (26).
Expanding the integral (26) along the quotient $Z_{l}(F)\backslash Z_{l}({\mathbb{A}})$
and exchanging roots with the group $V_{1}$, we deduce that this integral is equal to
(27)
$$\int\limits_{{\mathbb{A}}^{n-r^{\prime}}}\int\limits_{V_{1}({\mathbb{A}})}\int%
\limits_{U^{\prime}_{l}(F)\backslash U^{\prime}_{l}({\mathbb{A}})}\omega_{\psi%
}(g)\phi(x)\,\theta_{2n}^{(r)}(uvw_{0}j(x)g)\,\psi_{U^{\prime}_{l},1,b}(u)\,du%
\,dv\,dx.$$
Using [6], Lemma 1, we deduce that the integral (27) is not zero for some choice of data if and only if the integral (22) with $a=1$ is not zero for some choice of data.
When $r\neq n+1$, i.e. $n\neq 2l$, we use Assumption 1 to deduce the nonvanishing of integral (27), or equivalently,
the nonvanishing of a Whittaker coefficient of $\sigma_{2n-r+1}^{(r)}$.
When $r=n+1$, so $n=2l$, we may deduce the nonvanishing of the integral
(27) without requiring an analogue of Assumption 1.
Indeed, this situation is essentially the same as the cases studied in Ginzburg, Rallis and Soudry [9, 10].
In those works, the residue representation is defined on the linear group
$Sp_{2n}({\mathbb{A}})$ rather than a covering group. In our case, we are concerned with the $r=(n+1)$-fold cover, and the representation we use to define the descent is $\Theta_{2n}^{(n+1)}$. However, the crucial ingredient for the computations of the Whittaker coefficients is the unipotent orbit attached to the residue representations. In both cases it is the orbit $(n^{2})$. Since the computations for the proof of the Theorem involve only manipulations of unipotent groups, the cover does not enter.
Since the notations in those references are different, for the convenience of the reader we sketch the argument.
Starting with the integral (27), we choose $b=-1/4$. Let $\gamma=\begin{pmatrix}1&\\
-1&1\end{pmatrix}\begin{pmatrix}1&\tfrac{1}{2}\\
&1\end{pmatrix}$, and define
$\gamma_{0}=\text{diag}(\gamma,\ldots,\gamma,\gamma^{*},\ldots,\gamma^{*})\in Sp%
_{2n}(F)$. This is
analogous to the matrix $a$ defined in [9], (4.8).
Since $\theta_{2n}^{(n+1)}$ is invariant under $\gamma_{0}$, inserting it in the integral (27) and moving
it rightward we obtain
(28)
$$\int\limits_{{\mathbb{A}}^{n-r^{\prime}}}\int\limits_{V_{1}({\mathbb{A}})}\int%
\limits_{U^{\prime}_{l}(F)\backslash U^{\prime}_{l}({\mathbb{A}})}\omega_{\psi%
}(g)\phi(x)\,\theta_{2n}^{(n+1)}(u\gamma_{0}vw_{0}j(x)g)\,\psi_{U^{\prime}_{l}%
}(u)\,du\,dv\,dx$$
where $\psi_{U^{\prime}_{l}}$ is given by
$$\psi_{U^{\prime}_{l}}(u)=\psi(u_{1,3}+u_{2,4}+u_{3,5}+u_{4,6}+\cdots+u_{n-3,n-%
1}+u_{n-2,n}+u_{n-1,n+1}).$$
Define the following Weyl element $w_{0}^{\prime}$ of $Sp_{2n}$. If $w_{0}^{\prime}=(w^{\prime}_{0,i,j})$ then
$w^{\prime}_{0,i,2i-1}=1$ for all $1\leq i\leq n$. All other entries are $0,\pm 1$ and are determined uniquely such that $w_{0}^{\prime}\in Sp_{2n}$. This Weyl element was denoted by $\nu$ in [9], p. 881. Inserting $w_{0}^{\prime}$ in the integral (28), we obtain
(29)
$$\int\limits_{{\mathbb{A}}^{n-r^{\prime}}}\int\limits_{V_{1}({\mathbb{A}})}\int%
\limits_{Y(F)\backslash Y({\mathbb{A}})}\int\limits_{U_{2n,n,0}(F)\backslash U%
_{2n,n,0}({\mathbb{A}})}\omega_{\psi}(g)\phi(x)\,\theta_{2n}^{(n+1)}(uyw_{0}^{%
\prime}\gamma_{0}vw_{0}j(x)g)\,\psi_{U_{2n,n}}(u)\,du\,dy\,dv\,dx.$$
Here $Y$ is the unipotent group consisting of all matrices in $Sp_{2n}$ of the form $\begin{pmatrix}I_{n}&\\
y&I_{n}\end{pmatrix}$ such that $y_{i,j}=0$ if $i\geq j$. The group $U_{2n,n,0}$ is the
subgroup of $U_{2n,n}$ such that $u_{i,j}=0$ for all $1\leq i\leq n$ and $n+1\leq j\leq n+i$. Finally,
the character $\psi_{U_{2n,n}}$ is the character of $U_{2n,n}$ defined by $\psi_{U_{2n,n}}(u)=\psi(u_{1,2}+u_{2,3}+\cdots u_{n-1,n})$.
At this point we carry out a sequence of root exchanges. This process is described in detail in [9]
Section 5 so we will omit the details here. Carrying out this process, we deduce that the integral
(27) is equal to
(30)
$$\int\limits_{{\mathbb{A}}^{n-r^{\prime}}}\int\limits_{V_{1}({\mathbb{A}})}\int%
\limits_{Y({\mathbb{A}})}\int\limits_{U_{2n,n}(F)\backslash U_{2n,n}({\mathbb{%
A}})}\omega_{\psi}(g)\phi(x)\,\theta_{2n}^{(n+1)}(uyw_{0}^{\prime}\gamma_{0}vw%
_{0}j(x)g)\,\psi_{U_{2n,n}}(u)\,du\,dy\,dv\,dx.$$
Then once again [6], Lemma 1, implies that integral
(30) is not zero for some choice of data if and only if the inner integration along
the quotient $U_{2n,n}(F)\backslash U_{2n,n}({\mathbb{A}})$ is not zero for some choice of data. The
nonvanishing of this integral follows from Proposition 1
with $a=n$ and the genericity of the representation $\Theta_{GL_{n}}^{(n+1)}$ [12].
To complete the proof of the Theorem we need to consider the case when $n$ is odd and $r=n$. In
this case we have $l=(n+1)/2$. To use formulas from the previous case, we change the notation, and now
let $P^{\prime}_{l}$ be the standard parabolic subgroup of $Sp_{2n}$ whose Levi part is $GL_{1}\times GL_{2}^{{(n-1)}/{2}}$ and $U^{\prime}_{l}$ be its unipotent radical. (Above we would have written $P^{\prime}_{l-1}$, $U^{\prime}_{l-1}$.)
We again
start with the integral (24), define $w_{0}$ as above, and after conjugation we obtain the
integral (26), with the groups $U^{\prime}_{l,1}$ and $V_{1}$ changed as follows.
Now $U^{\prime}_{l,1}$ is the subgroup of $U^{\prime}_{l}$ consisting of $u=(u_{i,j})\in U^{\prime}_{l}$ such that $u_{2m-1,2k}=0$ for all $1\leq m\leq l-1$
and $m\leq k\leq l-1$, and $V_{1}$ is the unipotent subgroup of $Sp_{2n}$ consisting of all matrices of the form $I_{2n}+\sum_{i,j}r_{i,j}e_{i,j}^{*}$ where the sum ranges over all pairs $(i,j)=(2k,2m-1)$ where $1\leq m\leq l-1$ and $m-1\leq k\leq l-2$.
Let $Z_{l}$ now be the subgroup of $U^{\prime}_{l}$ consisting of all matrices of the form
$I_{2n}+\sum_{i,j}r_{i,j}e_{i,j}^{*}$ where the sum is over all pairs $(i,j)=(2m-1,2k)$ where
$1\leq m\leq l-1$ and $m\leq k\leq l-1$ together with the condition that $r_{1,2}=0$. Notice that
in this case it is not true that $U^{\prime}_{l}=Z_{l}\cdot U^{\prime}_{l,1}$. Performing root exchange similarly to our treatment of the
integral (26) we obtain
$$\int\limits_{{\mathbb{A}}^{n-r^{\prime}}}\int\limits_{V_{1}({\mathbb{A}})}\int%
\limits_{U^{\prime}_{l,0}(F)\backslash U^{\prime}_{l,0}({\mathbb{A}})}\omega_{%
\psi}(g)\phi(x)\,\theta_{2n}^{(n)}(uvw_{0}j(x)g)\,\psi_{U^{\prime}_{l},1,b}(u)%
\,du\,dv\,dx.$$
Here, $U^{\prime}_{l,0}$ is the subgroup of $U^{\prime}_{l}$ consisting of all matrices $u=(u_{i,j})\in U^{\prime}_{l}$
such that $u_{1,2}=0$, and the character $\psi_{U^{\prime}_{l},1,b}$ is given by
$$\psi_{U^{\prime}_{l},1,b}(u)=\psi\left(\sum_{i=1}^{n-2}u_{i,i+2}+u_{n-1,n+2}+%
bu_{n,n+1}\right).$$
Expand the above integral along the group $z(y)=I_{2n}+ye_{1,2}^{*}$. We obtain
$$\int\limits_{{\mathbb{A}}^{n-r^{\prime}}}\int\limits_{V_{1}({\mathbb{A}})}\sum%
_{\alpha\in F}\int\limits_{F\backslash{\mathbb{A}}}\int\limits_{U^{\prime}_{l,%
0}(F)\backslash U^{\prime}_{l,0}({\mathbb{A}})}\omega_{\psi}(g)\phi(x)\,\theta%
_{2n}^{(n)}(z(y)uvw_{0}j(x)g)\,\psi_{U^{\prime}_{l},1,b}(u)\,\psi(\alpha y)\,%
du\,dy\,dv\,dx.$$
Assume that $b=-1/4$. Then it is not hard to check that if $\alpha\neq-1/2$ then the two inner
integrations give a coefficient which is associated to the unipotent orbit $((n+1)(n-1))$.
Since $n$ is odd this is a symplectic partition. However, it follows from Conjecture 1${}_{n,n}$
that the unipotent orbit associated with $\Theta_{2n}^{(n)}$ is $(n^{2})$. Therefore, the only
nonzero contribution to the above integral is from $\alpha=1/2$. Let $\gamma=\begin{pmatrix}1&\\
1&1\end{pmatrix}\begin{pmatrix}1&-\tfrac{1}{2}\\
&1\end{pmatrix}$ and let
$\gamma_{0}=\text{diag}(\gamma,\ldots,\gamma,\gamma^{*},\ldots,\gamma^{*})\in Sp%
_{2n}(F)$. Using the invariance of
$\theta_{2n}^{(n)}$ by
this element and moving it rightward, we obtain the integral
(31)
$$\int\limits_{{\mathbb{A}}^{n-r^{\prime}}}\int\limits_{V_{1}({\mathbb{A}})}\int%
\limits_{U^{\prime}_{l}(F)\backslash U^{\prime}_{l}({\mathbb{A}})}\omega_{\psi%
}(g)\phi(x)\,\theta_{2n}^{(n)}(u\gamma vw_{0}j(x)g)\,\psi_{U^{\prime}_{l}}(u)%
\,du\,dv\,dx,$$
where $\psi_{U^{\prime}_{l}}$ is given by
$$\psi_{U^{\prime}_{l}}(u)=\psi(u_{1,2}+u_{2,4}+u_{3,5}+\cdots+u_{n-3,n-1}+u_{n-%
2,n}+u_{n-1,n+1}).$$
Next we introduce the Weyl element $w_{0}^{*}\in Sp_{2n}$ defined by
$$w_{0}^{*}=\begin{pmatrix}1&&\\
&w_{0}^{\prime}&\\
&&1\end{pmatrix}.$$
(Here $w_{0}^{\prime}$ was defined following (28).) Using invariance by $w_{0}^{*}$ and moving this element rightward,
(31)
is equal to
$$\int\limits_{{\mathbb{A}}^{n-r^{\prime}}}\int\limits_{V_{1}({\mathbb{A}})}\int%
\limits_{Y_{0}(F)\backslash Y_{0}({\mathbb{A}})}\int\limits_{U_{2n,n,0}(F)%
\backslash U_{2n,n,0}({\mathbb{A}})}\omega_{\psi}(g)\phi(x)\,\theta_{2n}^{(n)}%
(uy_{0}w_{0}^{*}\gamma_{0}vw_{0}j(x)g)\,\psi_{U_{2n,n}}(u)\,du\,dy_{0}\,dv\,dx.$$
Here $U_{2n,n,0}$ is now the subgroup of $U_{2n,n}$ of $u=(u_{i,j})\in U_{2n,n}$
such that $u_{i,j}=0$ for $2\leq i\leq n$ and $n+1\leq j\leq n+i-1$, and $Y_{0}$ is the group of all matrices of the form
$$y_{0}=\begin{pmatrix}1&&\\
&y^{\prime}&\\
&&1\end{pmatrix}$$
such that $y^{\prime}\in Y$ (where $Y$, as well as the character $\psi_{U_{2n,n}}$, was defined following (29)).
Carrying out root exchanges similarly to our treatment of (29), we obtain the integral
(32)
$$\int\limits_{{\mathbb{A}}^{n-r^{\prime}}}\int\limits_{V_{1}({\mathbb{A}})}\int%
\limits_{Y_{0}({\mathbb{A}})}\int\limits_{U_{2n,n}(F)\backslash U_{2n,n}({%
\mathbb{A}})}\omega_{\psi}(g)\phi(x)\,\theta_{2n}^{(n)}(uy_{0}w_{0}^{*}\gamma_%
{0}vw_{0}j(x)g)\,\psi_{U_{2n,n}}(u)\,du\,dy_{0}\,dv\,dx.$$
It is not hard to check that the integral (32) is not zero for some choice of data.
To complete the proof of the Theorem, observe that Proposition 1
allows us to express the integral of $\psi_{U_{2n,n}}$ times any right translate of $\theta_{2n}^{(n)}$ over
$U_{2n,n}(F)\backslash U_{2n,n}({\mathbb{A}})$ in terms of the Whittaker function of the representation $\Theta_{GL_{n}}^{(n)}$.
Since the Whittaker functional of this representation is unique up to scalars (see Kazhdan-Patterson [12], Theorem
II.2.1), we conclude that the integral (32) is
factorizable provided the data $\theta_{2n}^{(n)}\in\Theta_{2n}^{(n)}$ and $\phi$ are factorizable.
∎
6. Some Local Computations
We have just seen that for factorizable data the Whittaker coefficient of
$\sigma_{n+1}^{(2n)}$, $n$ odd, can be expressed as a product of local Whittaker functionals. We now analyze the corresponding
local integral in the unramified case. Consider a
place $\nu$ where all data is unramified. In this Section, $F$ denotes the corresponding local field, and in
all integrals we understand that we are taking the $F$-valued points of the algebraic groups that are shown.
We continue to use the convention that all matrices are embedded in the corresponding metaplectic groups via the trivial section, unless otherwise noted.
In this section, let $T_{2n}^{(n)}$ denote the inverse image in the local $n$-fold metaplectic group of the
standard maximal torus of $Sp_{2n}(F)$,
and let $Z(T_{2n}^{(n)})$ denote its center.
Let $p$ be a generator of the maximal ideal in the ring of integers ${\mathcal{O}}$ of $F$.
Since $n$ is odd, $-1$ is an $n$-th power
and hence $(p,p)=1$. This implies that $(p^{k_{1}},p^{k_{2}})=1$ for all integers $k_{1}$, $k_{2}$.
The subgroup of $T_{2n}^{(n)}$ generated by $Z(T_{2n}^{(n)})$ and by all matrices
$\text{diag}(p^{k_{1}},\ldots,p^{k_{n}},p^{-k_{n}},\ldots,p^{-k_{1}})$ is a maximal abelian subgroup of
$T_{2n}^{(n)}$. Denote this subgroup by $T_{2n,0}^{(n)}$. For a diagonal matrix
$t=\text{diag}(a,a^{*})\in Sp_{2n}(F)$ we shall denote $t_{0}=a\in GL_{n}(F)$. Also, recall that since we are in the unramified case the
maximal compact subgroup $Sp_{2n}({\mathcal{O}})$ of $Sp_{2n}(F)$ splits under the $n$-fold cover; fix an
injection $\iota$
of $Sp_{2n}({\mathcal{O}})$ into $Sp_{2n}^{(n)}(F)$.
Let $f_{W}^{(n)}$ be the function on $Sp_{2n}(F)$ defined as follows. First, it is right-invariant under
$\iota(Sp_{2n}({\mathcal{O}}))$.
Second, $f_{W}^{(n)}(ug)=\psi_{U_{2n,n}}(u)f_{W}^{(n)}(g)$ for all $u\in U_{2n,n}(F)$. (The character $\psi_{U_{2n,n}}$ was defined
before (30) above.) Finally, for all $(t,\zeta)\in T_{2n,0}^{(n)}$ ($t=\text{diag}(a,a^{*})\in Sp_{2n}(F)$, $\zeta\in\mu_{n})$ we have
$f_{W}^{(n)}((t,\zeta))=\zeta\,\delta_{Q}^{\frac{n-1}{2n}}(t)\,W_{GL_{n}}^{(n)}%
((t_{0},1))$. Here $W_{GL_{n}}^{(n)}$
is the normalized Whittaker function for the local theta representation of $GL_{n}^{(n)}$ that is unramified with respect to the
subgroup $GL_{n}(\mathcal{O})$ embedded in $GL_{n}^{(n)}(F)$ in a way that is compatible with $\iota$,
and $Q=P_{2n,n}$ is the maximal parabolic subgroup of $Sp_{2n}$ whose Levi part is $GL_{n}$.
(Recall from [12] that the space of
Whittaker
functionals for the theta function of $GL_{n}^{(n)}(F)$ is one dimensional.)
Consider the integral
(33)
$$W_{Sp_{n+1}}^{(2n)}(g)=\int\limits_{F^{n-r^{\prime}}}\int\limits_{V_{1}}\int%
\limits_{Y_{0}}\omega_{\psi}(g)\phi(x)\,f_{W}^{(n)}(y_{0}w_{0}^{*}\gamma_{0}vw%
_{0}j(x)g)\,dy_{0}\,dv\,dx,$$
where the function $\phi$ is the normalized unramified Schwartz function.
It may be desirable to compute $W_{Sp_{n+1}}^{(2n)}(g)$ for an arbitrary torus element; however,
a general formula would be quite complicated. We will concentrate on a special
case (though for $n=3$ it is general)
where the computations, though not trivial, are simpler. Let $n_{1},n_{2}$ be non-negative integers,
$a=p^{n_{1}}$ and $b=p^{n_{2}}$, and let
$$g=\text{diag}(ab,b,I_{n-3},b^{-1},a^{-1}b^{-1}).$$
Substituting this element in (33), changing variables
in $x$ and using the invariant properties of the functions involved, we obtain
$$\gamma(a)\,|ab^{2}|^{-1/2}\int\limits_{V_{1}}\int\limits_{Y_{0}}f_{W}^{(n)}(y_%
{0}w_{0}^{*}\gamma_{0}vw_{0}g)\,dy_{0}\,dv.$$
Here $\gamma(a)$ is the Weil factor and the factor $\gamma(a)\,|ab^{2}|^{-1/2}$ is obtained from the change
of variables in $x$ together with the properties of the Weil representation. We can ignore the
integration over $x$ since $\phi$ is the unramified Schwartz function.
The matrices $\gamma_{0}$ and $\gamma$ were defined before (5) above. We have
the factorization
$\gamma=\begin{pmatrix}1&1\\
&1\end{pmatrix}\begin{pmatrix}&-1\\
1&\end{pmatrix}\begin{pmatrix}1&\tfrac{1}{2}\\
&1\end{pmatrix},$ and this induces a factorization of $\gamma_{0}$:
$\gamma_{0}=\gamma_{0}^{\prime}w_{0}^{\prime}\gamma_{0}^{\prime\prime}$. Changing variables in $V_{1}$ we may ignore the matrix $\gamma_{0}^{\prime\prime}$.
Moving $\gamma_{0}^{\prime}$ to the left and using the left transformation of $f_{W}^{(n)}$ by matrices
in $U_{2n,n}$, we obtain the integral
$$\gamma(a)\,|ab^{2}|^{-1/2}\int\limits_{V_{1}}\int\limits_{Y_{0}}f_{W}^{(n)}(y_%
{0}w_{0}^{*}w_{0}^{\prime}vw_{0}g)\,\psi_{Y_{0}}(y_{0})\,dy_{0}\,dv.$$
Here we recall that $n$ is odd and that $Y_{0}$ consists of all matrices of
the form
$$y_{0}=\begin{pmatrix}1&&&\\
&I_{n-1}&&\\
&y&I_{n-1}&\\
&&&1\end{pmatrix}\quad\text{with~{}}y\in\text{Mat}_{(n-1)\times(n-1)},y_{i,j}=%
0~{}\text{for all~{}}i\geq j;$$
we have $\psi_{Y_{0}}(y_{0})=\psi(y_{\frac{n-1}{2},\frac{n+1}{2}})$.
Next we move the matrix $g$ to the left and the Weyl element $w_{0}^{*}w_{0}^{\prime}w_{0}$ to the right. Doing so, we
obtain the integral
(34)
$$\gamma(a)\,|ab^{2}|^{-\frac{n-2}{2}}\int\limits_{V_{2}}\int\limits_{Y_{0}}f_{W%
}^{(n)}(ty_{0}v_{2})\,\psi_{Y_{0}}(y_{0})\,dy_{0}\,dv_{2}$$
with $t=\text{diag}(ab,b,I_{2n-4},b^{-1},a^{-1}b^{-1})$. The group
$V_{2}$ consists of all matrices of the form $\left(\begin{smallmatrix}I_{n}&\\
v&I_{n}\end{smallmatrix}\right)$ where
$v=\sum_{i,j}r_{i,j}(e_{i,j}+e_{n-j+1,n-i+1})$ and the sum is over all pairs $(i,j)$ such that $1\leq i\leq\frac{n-3}{2}$ and $1\leq j\leq i+1$. The extra factor of
$|ab^{2}|^{-\frac{n-3}{2}}$ is obtained from the change of variables in $V_{1}$.
Integral (34) is the analogous to the integral that appears on the right hand side of
Ginzburg-Rallis-Soudry [10], p. 255, Eqn. (3.7). Indeed, the matrices $y_{0}$ and $v_{2}$ can be viewed
as matrices of $Sp_{2n-2}$ embedded in $Sp_{2n}$ in the obvious way. This comparison enables
us to proceed as in [10], pp. 256-260, and conclude that integral (34) is a sum
of two terms. First we have
$$\gamma(a)\,|ab^{2}|^{-\frac{n-2}{2}}f_{W}^{(n)}(t)=\gamma(a)|ab^{2}|^{\frac{2n%
-1}{2n}}W_{GL_{n}}^{(n)}(t_{0}).$$
This corresponds to the term which is denoted as Case 1 in [10], p. 260. The second term
we obtain is the one corresponding to Case 2 in that reference. It is equal to
(35)
$$\gamma(a)|ab^{2}|^{-\frac{n-2}{2}}\int\limits_{|h|>1}f_{W}^{(n)}(ty(h))\,\psi(%
h)\,|h|^{\frac{(n-1)(n-3)}{2}}\,dh,$$
with $y(h)=\text{diag}(I_{2},h^{-1}I_{n-2},hI_{n-2},I_{2})$. Write $h=p^{-m}\epsilon$ where $\epsilon$
is a unit. Then (35) is equal to
$$\gamma(a)\,|ab^{2}|^{-\frac{n-2}{2}}\sum_{m=1}^{\infty}q^{m\left(1+\frac{(n-1)%
(n-3)}{2}\right)}\int\limits_{|\epsilon|=1}f_{W}^{(n)}(ty(p^{-m}\epsilon))\,%
\psi(p^{-m}\epsilon)\,d\epsilon.$$
The factorization $y(p^{-m}\epsilon)=y(p^{-m})y(\epsilon)$ contributes, via the cocycle defining the metaplectic group, the factor
$(\epsilon,p)_{n}^{m(n-2)}$. Hence the above integral is equal to
(36)
$$\gamma(a)\,|ab^{2}|^{-\frac{n-2}{2}}\sum_{m=1}^{\infty}q^{m\left(1+\frac{(n-1)%
(n-3)}{2}\right)}f_{W}^{(n)}(ty(p^{-m}))\int\limits_{|\epsilon|=1}(\epsilon,p)%
_{n}^{m(n-2)}\,\psi(p^{-m}\epsilon)\,d\epsilon.$$
The inner integral is zero unless $m=1$, and when $m=1$ it is equal to $q^{-1/2}G_{n-2}^{(n)}(p)$
where the last term is a normalized $n$-th order Gauss sum, as in [6], Section 6. Substituting this into (36) we
obtain
$$\gamma(a)\,|ab^{2}|^{-\frac{n-2}{2}}q^{-\frac{(n-2)(2n-1)}{2n})}W_{GL_{n}}^{(n%
)}(t_{0}y(p)_{0}).$$
Thus we have established the following result, relating the Whittaker coefficients of the descent on a suitable
cover of the symplectic group to the Whittaker coefficients of the theta function on a cover of the general linear group.
Theorem 2.
Let $n\geq 3$ be an odd integer. Let $a=p^{n_{1}}$, $b=p^{n_{2}}$, with $n_{1},n_{2}\geq 0$. Let
$$t=\text{\rm diag}(ab,b,I_{n-3},b^{-1},a^{-1}b^{-1})\in Sp_{n+1}(F),$$
$$t_{0}=\text{\rm diag}(ab,b,I_{n-2})\in GL_{n}(F),\quad t_{1}=\text{\rm diag}(%
ab,b,pI_{n-2})\in GL_{n}(F).$$
Suppose that Conjecture 1${}_{n,i}$ holds for $(n-1)/2\leq i\leq n$. Then we have the identity
(37)
$$W_{Sp_{n+1}}^{(2n)}(t)=\gamma(a)\,|ab^{2}|^{\frac{2n-1}{2n}}\left(W_{GL_{n}}^{%
(n)}(t_{0})+q^{-\frac{(n-2)(2n-1)}{2n}}W_{GL_{n}}^{(n)}(t_{1})\right).$$
7. A Case of Conjecture 1
Suppose $n=3$. Then Conjecture 1${}_{3,i}$ is known for $i=1$ ([12], Corollary II.2.6) and $i=2$ ([6],
Proposition 3).
In this section we sketch the proof of Conjecture 1${}_{3,3}$.
Proposition 3.
Conjecture 1${}_{3,3}$ holds. That is, ${\mathcal{O}}(\Theta_{6}^{(3)})=(3^{2})$.
As a consequence, the hypotheses of Theorem 2 are satisfied for $n=3$, and we conclude unconditionally the
Whittaker function for the descent on the $6$-fold cover of $Sp_{4}$ is Eulerian, and
given at unramified places by formula (37).
Proof.
To prove the Proposition we need to prove that the representation $\Theta_{6}^{(3)}$ has no nonzero Fourier coefficients
corresponding to the unipotent orbits $(6)$ and $(42)$. Once we do so, an argument similar to the one given in
Ginzburg [7], Thm. 3.1, will imply that this representation also has no nonzero Fourier coefficients corresponding to the orbit $(41^{2})$. We also need to show that $\Theta_{6}^{(3)}$ has a nonzero Fourier coefficient
corresponding to the unipotent orbit $(3^{2})$. This last follows from Proposition 1. So the critical step is the
vanishing of Fourier coefficients attached to the two unipotent orbits $(6)$ and $(42)$.
For convenience, we drop the prior notation for unipotent subgroups and characters, and will introduce the ones needed anew.
To prove this claim for $\Theta_{6}^{(3)}$ we make use of the fact that this representation can be obtained as a residue of an Eisenstein series induced from a suitable theta function of a smaller rank group. This is explained in Section 2 above.
First we consider the unipotent orbit $(6)$.
Since $\Theta_{6}^{(3)}$ is a residue of $E_{\Theta_{4}}^{(3)}(g,s)$, in order to prove that
$\Theta_{6}^{(3)}$ is not generic it is enough to prove that $E_{\Theta_{4}}^{(3)}(g,s)$ is not generic. Moreover, let $U$ denote the standard maximal unipotent of $Sp_{6}$, and let $\psi_{U}$ denote a Whittaker character of $U$;
without loss we take $\psi_{U}(u)=\psi(u_{1,2}+u_{2,3}+au_{3,4})$ with
$a\in F^{*}$. Then a standard unfolding argument shows that the integral
$$\int\limits_{U(F)\backslash U({\mathbb{A}})}E_{\Theta_{4}}^{(3)}(ug,s)\,\psi_{%
U}(u)\,du$$
is not zero for some choice of data if and only if the representation
$\Theta_{4}^{(3)}$ is generic. However, it follows from Friedberg-Ginzburg
[6], Lemma 2, that this representation is not generic. Hence, for $\text{Re}(s)$ large, the above integral is zero for all choices of data, and by continuation it follows that $\Theta_{6}^{(3)}$ is not generic.
Next we consider the Fourier coefficients corresponding to the unipotent orbit $(42)$. Let $V\subseteq U$ denote the subgroup
$V=\{u=(u_{i,j})\in U\ :\ u_{2,3}=u_{4,5}=0\}$. This is the unipotent radical of the parabolic subgroup of $Sp_{6}$ whose Levi part is $L=GL_{1}\times GL_{2}$. The embedding of $L$ into $Sp_{6}$ is given by $L=\{\text{diag}(g_{1},g_{2},g_{2}^{*},g_{1}^{-1})\ :\ g_{1}\in GL_{1};\ g_{2}%
\in GL_{2}\}$. Let $\alpha_{1},\alpha_{2}\in F$ and
$\alpha_{3},\alpha_{4}\in F^{*}$. Consider the Fourier coefficient defined by
(38)
$$\int\limits_{V(F)\backslash V({\mathbb{A}})}\varphi(vg)\,\psi_{V,\alpha_{i}}(v%
)\,dv,$$
where $\psi_{V,\alpha_{i}}(v)=\psi(\alpha_{1}v_{1,2}+\alpha_{2}v_{1,3}+\alpha_{3}v_{3%
,4}+\alpha_{4}v_{2,5})$.
This Fourier coefficient is
associated with the unipotent orbit $(42)$ provided the connected component of the stabilizer of $\psi_{V,\alpha_{i}}$ inside
$L(F)$ is trivial (see for example Ginzburg [7]).
If $\alpha_{3}\alpha_{4}=-\epsilon^{2}$ for some $\epsilon\in F^{*}$, then one can find a suitable discrete element $\gamma_{L}\in L$
and $\alpha\in F^{*}$ such that the integral (38) is equal to
(39)
$$\int\limits_{V(F)\backslash V({\mathbb{A}})}\varphi(v\gamma_{L}g)\,\widetilde{%
\psi}_{V,\alpha}(v)\,dv,$$
where $\widetilde{\psi}_{V,\alpha}(v)=\psi(\alpha v_{1,2}+v_{1,3}+v_{2,4})$.
We will say that the Fourier coefficient (38) is of split type if $\alpha_{3}\alpha_{4}=-\epsilon^{2}$ for some $\epsilon\in F^{*}$,
and non-split otherwise.
Since at half of the places $\nu$ we have the condition $\alpha_{3}\alpha_{4}=-\epsilon^{2}$ for some $\epsilon\in F_{\nu}^{*}$,
to complete the proof it is enough to
prove that for every finite local unramified place, the Jacquet
module $J_{V,\widetilde{\psi}_{V,\alpha}}(\Theta_{6}^{(3)})$ is zero for all characters $\widetilde{\psi}_{V,\alpha}$
with $\alpha\in F^{*}$, where $\Theta_{6}^{(3)}$ now denotes the local theta representation at $\nu$.
To do so we consider the Jacquet module associated with the unipotent orbit $(3^{2})$.
Let $R$ denote the subgroup of $U$ defined by $R=\{u\in U\ :\ u_{1,2}=u_{3,4}=u_{5,6}=0\}$.
Then $R$ is the unipotent radical of the maximal parabolic subgroup
of $Sp_{6}$ whose Levi part is $GL_{2}\times SL_{2}$.
Let $\psi_{R}$ denote the character $\psi_{R}(r)=\psi(r_{1,3}+r_{2,4})$ of $R$ (that is, of $R(F_{\nu})$; from now on we take the $F_{\nu}$
points of the algebraic groups in question without changing the notation).
Next, we compute
$J_{RN,\psi_{R}\psi_{\alpha}}(\Theta_{6}^{(3)})$, where $N$ is the unipotent radical of $SL_{2}$ embedded in $Sp_{6}$ as the group of all matrices
$$n(x)=\begin{pmatrix}1&x&&&&\\
&1&&&&\\
&&1&x&&\\
&&&1&&\\
&&&&1&-x\\
&&&&&1\end{pmatrix},$$
and $\psi_{\alpha}$ is the
character of $N$ defined by $\psi_{\alpha}(n(x))=\psi(\alpha x)$, with
$\alpha\in F_{\nu}$.
Using the one-parameter subgroups $\{z(k)=I_{6}+ke_{3,4}\}$ and $\{y(m)=I_{6}+me_{2,3}-me_{4,5}\}$,
we perform root exchanges using the Lemma in Section 2 of [9].
This gives the isomorphism of Jacquet modules
(40)
$$J_{RN,\psi_{R}\psi_{\alpha}}(\Theta_{6}^{(3)})\cong J_{V,\widetilde{\psi}_{V,%
\alpha}}(\Theta_{6}^{(3)}).$$
Thus, to prove Prop. 3, it is enough to prove that for all
$\alpha\neq 0$, the Jacquet module $J_{RN,\psi_{R}\psi_{\alpha}}(\Theta_{6}^{(3)})$ is zero.
It is not hard to check that the stabilizer of $\psi_{R}$ inside $GL_{2}\times SL_{2}$ is $SL_{2}$ embedded diagonally. Thus,
$J_{R,\psi_{R}}(\Theta_{6}^{(3)})$, if not zero, defines an unramified
representation of $SL_{2}$.
Let $\sigma$ denote an irreducible unramified representation of
$SL_{2}$ such that the space
$$\text{Hom}_{SL_{2}}(\sigma,J_{R,\psi_{R}}(\Theta_{6}^{(3)}))$$
is not zero. Since $\sigma$ is unramified it is a sub-representation
of $\text{Ind}_{B}^{SL_{2}}\chi\delta_{B}^{1/2}$. Applying Frobenius reciprocity
we deduce that the space
$$\text{Hom}_{GL_{1}}(\chi,J_{RN,\psi_{R}}(\Theta_{6}^{(3)}))$$
is not zero. We shall show below that we must have $\chi=1$, which implies that $\sigma=1$.
Since the Jacquet module is always of finite length, we then conclude that all Jacquet modules
$J_{RN,\psi_{R}\psi_{\alpha}}(\Theta_{6}^{(3)})$ with $\alpha\neq 0$ are
zero. Thus once we have established that $\chi$ must be $1$ (and in particular,
that $J_{R,\psi_{R}}(\Theta_{6}^{(3)})$ is nonzero), the Proposition will be proved.
To do so, we use the isomorphism (40) above with $\alpha=0$.
If $\alpha=0$, then the Jacquet module $J_{RN,\psi_{R}\psi_{0}}(\Theta_{6}^{(3)})=J_{RN,\psi_{R}}(\Theta_{6}^{(3)})%
\cong J_{V,\psi_{V}}(\Theta_{6}^{(3)})$, if not zero, defines an unramified character of $GL_{1}$, which is embedded in $Sp_{6}$ as the
group of all matrices $\{t(a)\}$, with $t(a)={\text{diag}}(a,a^{-1},a,a^{-1},a,a^{-1})$.
Here $\psi_{V}=\widetilde{\psi}_{V,\alpha}$ with $\alpha=0$.
For $1\leq i\leq 3$, let $w_{i}$ denote the three simple reflections of the Weyl group of $Sp_{6}$. In matrices we have
$$w_{1}=\begin{pmatrix}&1&&&&\\
1&&&&&\\
&&1&&&\\
&&&1&&\\
&&&&&1\\
&&&&1&\end{pmatrix};\ \ w_{2}=\begin{pmatrix}1&&&&&\\
&&1&&&\\
&1&&&&\\
&&&&1&\\
&&&1&&\\
&&&&&1\end{pmatrix};\ \ w_{3}=\begin{pmatrix}1&&&&&\\
&1&&&&\\
&&&1&&\\
&&-1&&&\\
&&&&1&\\
&&&&&1\end{pmatrix}.$$
Let $Y=\{y_{1}(m_{1})=I_{6}+m_{1}e_{4,3}\}$, and let $U_{1}$ be the subgroup of the maximal unipotent subgroup $U$ of $Sp_{6}$ consisting
of all matrices $u=(u_{i,j})\in U$ such that $u_{2,4}=u_{3,5}=u_{3,4}=0$. Let $\psi_{U_{1}}(u_{1})=\psi(u_{1}(1,2)+u_{1}(2,3))$. Then
$U_{1}Y=(w_{3}w_{2})V(w_{3}w_{2})^{-1}$.
Hence, $J_{V,\psi_{V}}(\Theta_{6}^{(3)})$ is isomorphic to
$J_{U_{1},\psi_{U_{1}}}(\Theta_{6}^{(3)})$. Let $X=\{I_{6}+ke_{2,4}+ke_{3,5}\}$, and perform a root exchange of this unipotent subgroup with the
group $Y$. Then, since $\Theta_{6}^{(3)}$ is not generic,
we obtain that $J_{U_{1},\psi_{U_{1}}}(\Theta_{6}^{(3)})$ is isomorphic
to $J_{U,\psi_{U}}(\Theta_{6}^{(3)})$.
This Jacquet module is not zero. Indeed, using an argument similar to
[3] Theorem 2.3, the nonvanishing of this Jacquet module
follows from the genericity of the theta representation of $GL_{3}^{(3)}$.
From this we deduce that the Jacquet module
$J_{R,\psi_{R}}(\Theta_{6}^{(3)})$ is not zero.
Moreover, the Jacquet module $J_{U,\psi_{U}}(\Theta_{6}^{(3)})$ acts by the character $|a|^{4}$ under the torus
$\text{diag}(aI_{3},a^{-1}I_{3})\in Sp_{6}$. Taking into an account the
various root exchanges, we deduce that the group
$GL_{1}=\{t(a)\}$ acts trivially on the Jacquet module $J_{RN,\psi_{R}}(\Theta_{6}^{(3)})$.
Thus $\chi$ above must be $1$.
This completes the proof of the Proposition.
∎
References
[1]
Banks, William D.; Levy, Jason; Sepanski, Mark R.
Block-compatible metaplectic cocycles.
J. Reine Angew. Math. 507 (1999), 131-163.
[2]
Brubaker, Ben; Friedberg, Solomon. Whittaker coefficients of metaplectic Eisenstein series.
Geometric and Functional Analysis (GAFA) 25(4) (2015), 1180-1239.
[3]
Bump, Daniel; Friedberg, Solomon; Ginzburg, David.
Small representations for odd orthogonal groups. Int. Math. Res. Not. 2003,
no. 25 (2003), 1363-1393.
[4]
Cai, Yuanqing. Thesis, Boston College (in progress).
[5]
Collingwood, David; McGovern, William M.
Nilpotent orbits in semisimple Lie algebras.
Van Nostrand Reinhold Mathematics Series. Van Nostrand Reinhold Co., New York, 1993.
[6]
Friedberg, Solomon; Ginzburg, David. Descent and theta functions for metaplectic groups. To appear
in J. European Math. Soc.
arXiv:1403.3930.
[7]
Ginzburg, David.
Certain conjectures relating unipotent orbits to automorphic representations. Isr. J. Math. 151 (2006), 323-355.
[8]
Ginzburg, David. Towards a classification of global integral constructions and functorial liftings using the small representations method. Adv. Math. 254 (2014), 157-186.
[9]
Ginzburg, David; Rallis, Stephen; Soudry, David. On a correspondence between cuspidal representations of
$GL_{2n}$ and $\widetilde{Sp}_{2n}$. J. Amer. Math. Soc. 12 (1999), no. 3, 849-907.
[10]
Ginzburg, David; Rallis, Stephen; Soudry, David. Lifting cusp forms on $GL_{2n}$ to $\widetilde{Sp}_{2n}$: the unramified correspondence. Duke Math. J. 100 (1999), no. 2, 243-266.
[11]
Ginzburg, David; Rallis, Stephen; Soudry, David. The Descent Map from Automorphic Representations of $GL(n)$ to Classical Groups. World Scientific Publishing Co. Pte. Ltd., Singapore, 2011.
[12]
Kazhdan, David A.; Patterson, S. J. Metaplectic forms.
Publ. Math. Inst. Hautes Études Sci. 59 (1984), 35-142.
[13]
S. S. Kudla; S. Rallis. Poles of Eisenstein series and $L$-functions. Isr. Math. Conf. Proc., Vol. 3, Weizmann Science Press of Isr., Jerusalem (1990), pp. 81-110.
[14]
Mœglin, C.; Waldspurger, J.-L. Spectral decomposition and Eisenstein series.
Une paraphrase de l’Écriture [A paraphrase of Scripture].
Cambridge Tracts in Mathematics, 113. Cambridge University Press, Cambridge, 1995. |
Value-Decomposition Networks For Cooperative Multi-Agent Learning
Peter Sunehag
DeepMind
[email protected] &Guy Lever
DeepMind
[email protected] &Audrunas Gruslys
DeepMind
[email protected] &Wojciech Marian Czarnecki
DeepMind
[email protected] &Vinicius Zambaldi
DeepMind
[email protected]
&Max Jaderberg
DeepMind
[email protected] &Marc Lanctot
DeepMind
[email protected] &Nicolas Sonnerat
DeepMind
[email protected] &Joel Z. Leibo
DeepMind
[email protected] &Karl Tuyls
DeepMind & University of Liverpool
[email protected] &Thore Graepel
DeepMind
[email protected]
Abstract
We study the problem of cooperative multi-agent reinforcement learning with a single joint reward signal. This class of learning problems is difficult because of the often large combined action and observation spaces. In the fully centralized and decentralized approaches, we find the problem of spurious rewards and a phenomenon we call the “lazy agent” problem, which arises due to partial observability. We address these problems by training individual agents with a novel value decomposition network architecture, which learns to decompose the team value function into agent-wise value functions. We perform an experimental evaluation across a range of partially-observable multi-agent domains and show that learning such value-decompositions leads to superior results, in particular when combined with weight sharing, role information and information channels.
Value-Decomposition Networks For Cooperative Multi-Agent Learning
Peter Sunehag
DeepMind
[email protected]
Guy Lever
DeepMind
[email protected]
Audrunas Gruslys
DeepMind
[email protected]
Wojciech Marian Czarnecki
DeepMind
[email protected]
Vinicius Zambaldi
DeepMind
[email protected]
Max Jaderberg
DeepMind
[email protected]
Marc Lanctot
DeepMind
[email protected]
Nicolas Sonnerat
DeepMind
[email protected]
Joel Z. Leibo
DeepMind
[email protected]
Karl Tuyls
DeepMind & University of Liverpool
[email protected]
Thore Graepel
DeepMind
[email protected]
1 Introduction
We consider the cooperative multi-agent reinforcement learning (MARL) problem (Panait and Luke, 2005, Busoniu et al., 2008, Tuyls and Weiss, 2012), in which a system of several learning agents must jointly optimize a single reward signal – the team reward – accumulated over time. Each agent has access to its own (“local”) observations and is responsible for choosing actions from its own action set. Coordinated MARL problems emerge in applications such as coordinating self-driving vehicles and/or traffic signals in a transportation system, or optimizing the productivity of a factory comprised of many interacting components. More generally, with AI agents becoming more pervasive, they will have to learn to coordinate to achieve common goals.
Although in practice some applications may require local autonomy, in principle the cooperative MARL problem could be treated using a centralized approach, reducing the problem to single-agent reinforcement learning (RL) over the concatenated observations and combinatorial action space. We show that the centralized approach consistently fails on relatively simple cooperative MARL problems in practice. We present a simple experiment in which the centralised approach fails by learning inefficient policies with only one agent active and the other being ‘‘lazy’’. This happens when one agent learns a useful policy, but a second agent is discouraged from learning because its exploration would hinder the first agent and lead to worse team reward.111For example, imagine training a 2-player soccer team using RL with the number of goals serving as the team reward signal. Suppose one player has become a better scorer than the other. When the worse player takes a shot the outcome is on average much worse, and the weaker player learns to avoid taking shots (Hausknecht, 2016).
An alternative approach is to train independent learners to optimize for the team reward. In general each agent is then faced with a non-stationary learning problem because the dynamics of its environment effectively changes as teammates change their behaviours through learning (Laurent et al., 2011). Furthermore, since from a single agent’s perspective the environment is only partially observed, agents may receive spurious reward signals that originate from their teammates’ (unobserved) behaviour. Because of this inability to explain its own observed rewards naive independent RL is often unsuccessful: for example Claus and Boutilier (1998) show that independent $Q$-learners cannot distinguish teammates’ exploration from stochasticity in the environment, and fail to solve even an apparently trivial, 2-agent, stateless, $3\times 3$-action problem and the general Dec-POMDP problem is known to be intractable (Bernstein et al., 2000, Oliehoek and Amato, 2016). Though we here focus on 2 player coordination, we note that the problems with individual learners and centralized approaches just gets worse with more agents since then, most rewards do not relate to the individual agent and the action space grows exponentially for the fully centralized approach.
One approach to improving the performance of independent learners is to design individual reward functions, more directly related to individual agent observations. However, even in the single-agent case, reward shaping is difficult and only a small class of shaped reward functions are guaranteed to preserve optimality w.r.t. the true objective (Ng et al., 1999, Devlin et al., 2014, Eck et al., 2016). In this paper we aim for more general autonomous solutions, in which the decomposition of the team value function is learned.
We introduce a novel learned additive value-decomposition approach over individual agents. Implicitly, the value decomposition network aims to learn an optimal linear value decomposition from the team reward signal, by back-propagating the total $Q$ gradient through deep neural networks representing the individual component value functions. This additive value decomposition is specifically motivated by avoiding the spurious reward signals that emerge in purely independent learners.The implicit value function learned by each agent depends only on local observations, and so is more easily learned. Our solution also ameliorates the coordination problem of independent learning highlighted in Claus and Boutilier (1998) because it effectively learns in a centralised fashion at training time, while agents can be deployed individually.
Further, in the context of the introduced agent, we evaluate weight sharing, role information and information channels as additional enhancements that have recently been reported to improve sample complexity and memory requirements (Hausknecht, 2016, Foerster et al., 2016, Sukhbaatar et al., 2016). However, our main comparison is between three kinds of architecture; Value-Decomposition across individual agents, Independent Learners and Centralized approaches. We investigate and benchmark combinations of these techniques applied to a range of new interesting two-player coordination domains. We find that Value-Decomposition is a much better performing approach than centralization or fully independent learners, and that when combined with the additional techniques, results in an agent that consistently outperforms centralized and independent learners by a big margin.
1.1 Other Related Work
Schneider et al. (1999) consider the optimization of the sum of individual reward functions, by optimizing local compositions of individual value functions learnt from them. Russell and Zimdars (2003) sums the $Q$-functions of independent learning agents with individual rewards, before making the global action selection greedily to optimize for total reward. Our approach works with only a team reward, and learns the value-decomposition autonomously from experience, and it similarly differs from the approach with coordination graphs (Guestrin et al., 2002) and the max-plus algorithm (Kuyer et al., 2008, van der Pol and Oliehoek, 2016).
Other work addressing team rewards in cooperative settings is based on difference rewards (Tumer and Wolpert, 2004), measuring the impact of an agent’s action on the full system reward. This reward has nice properties (e.g. high learnability), but can be impractical as it requires knowledge about the system state (Colby et al., 2016, Agogino and Tumer, 2008, Proper and Tumer, 2012). Other approaches can be found in Devlin et al. (2014), HolmesParker et al. (2016), Babes et al. (2008).
2 Background
2.1 Reinforcement Learning
We recall some key concepts of the RL setting (Sutton and Barto, 1998), an agent-environment framework (Russell and Norvig, 2010) in which an agent sequentially interacts with the environment over a sequence of timesteps, $t=1,2,3,\ldots$, by executing actions and receiving observations and rewards, and aims to maximize cumulative reward. This is typically modelled as a Markov decision process (MDP) (e.g. Puterman, 1994) defined by a tuple $\langle\mathcal{S,A},\mathcal{T}_{1},\mathcal{T},R\rangle$ comprising the state space $\mathcal{S}$, action space $\mathcal{A}$, a (possibly stochastic) reward function $R:\mathcal{S\times A\times S}\to\mathbb{R}$ start state distribution $\mathcal{T}_{1}\in\mathcal{P(S)}$ and transition function $\mathcal{T}:\mathcal{S\times A}\rightarrow\mathcal{P(S)}$, where $\mathcal{P(X)}$ denotes the set of probability distributions over the set $\mathcal{X}$. We use $\bar{R}$ to denote the expected value of $R$. The agent’s interactions give rise to a trajectory $(S_{1},A_{1},R_{1},S_{2},...)$ where $S_{1}\sim\mathcal{T}_{1}$, $S_{t+1}\sim\mathcal{T}(\cdot|S_{t},A_{t})$ and $R_{t}=R(S_{t},A_{t},S_{t+1})$, and we denote random variables in upper-case, and their realizations in lower-case. At time $t$ the agent observes $o_{t}\in\mathcal{O}$ which is typically some function of the state $s_{t}$, and when the state is not fully observed the system is called a partially observed Markov decision process (POMDP).
The agent’s goal is to maximize expected cumulative discounted reward with a discount factor $\gamma$, ${\cal R}_{t}:=\sum_{t=1}^{\infty}{\gamma^{t-1}R_{t}}$. The agent chooses actions according to a policy: a (stationary) policy is a function $\pi:\mathcal{S}\to\mathcal{P(A)}$ from states to probability distributions over $\mathcal{A}$. An optimal policy is one which maximizes expected cumulative reward. In fully observed environments, stationary optimal policies exist. In partially observed environments, the policy usually incorporates past agent observations from the history $h_{t}=a_{1}o_{1}r_{1},...,a_{t-1}o_{t-1}r_{t-1}$ (replacing $s_{t}$). A practical approach utilized here, is to parameterize policies using recurrent neural networks.
$V^{\pi}(s):=\mathbb{E}[\sum_{t=1}^{\infty}\gamma^{t-1}R(S_{t},A_{t},S_{t+1})|S%
_{1}=s;A_{t}\sim\pi(\cdot|S_{t})]$ is the value function and the action-value function is $Q^{\pi}(s,a):=\mathbb{E}_{S^{\prime}\sim\mathcal{T}(\cdot|s,a)}[R(S,a,S^{%
\prime})+\gamma V(S^{\prime})]$ (generally, we denote the successor state of $s$ by $s^{\prime}$). The optimal value function is defined by $V^{*}(s)=\sup_{\pi}V^{\pi}(s)$ and similarly $Q^{*}(s,a)=\sup_{\pi}Q^{\pi}(s,a)$. For a given action-value function $Q:\mathcal{S}\times\mathcal{A}\to\mathbb{R}$ we define the (deterministic) greedy policy w.r.t. $Q$ by $\pi(s):=\mathop{\arg\max}_{a\in\mathcal{A}}Q(s,a)$ (ties broken arbitrarily). The greedy policy w.r.t. $Q^{*}$ is optimal (e.g. Szepesvári, 2010).
2.2 Deep $Q$-Learning
One method for obtaining $Q^{*}$ is $Q$-learning which is based on the update $Q_{i+1}(s_{t},a_{t})=(1-\eta_{t})Q_{i}(s_{t},a_{t})+\eta_{t}(r_{t}+\gamma\max_%
{a}Q_{i}(s_{t+1},a))$, where $\eta_{t}\in(0,1)$ is the learning rate. We employ the $\varepsilon$-greedy approach to action selection based on a value function, which means that with $1-\varepsilon$ probability we pick $\mathop{\arg\max}_{a}Q_{i}(s,a)$ and with probability $\varepsilon$ a random action. Our study focuses on deep architectures for the value function similar to those used by Mnih et al. (2015), and our approach incorporates the key techniques of target networks and experience replay employed there, making the update into a stochastic gradient step. Since we consider partially observed environments our $Q$-functions are defined over agent observation histories, $Q(h_{t},a_{t})$, and we incorporate a recurrent network similarly to Hausknecht and Stone (2015). To speed up learning we add the dueling architecture of Wang et al. (2016) that represent $Q$ using a value and an advantage function, including multi-step updates with a forward view eligibility trace (e.g. Harb and Precup, 2016) over a certain number of steps. When training agents the recurrent network is updated with truncated back-propagation through time (BPTT) for this amount of steps. Although we concentrate on DQN-based agent architectures, our techniques are also applicable to policy gradient methods such as A3C (Mnih et al., 2016).
2.3 Multi-Agent Reinforcement Learning
We consider problems where observations and actions are distributed across $d$ agents, and are represented as $d$-dimensional tuples of primitive observations in $\mathcal{O}$ and actions in $\mathcal{A}$. As is standard in MARL, the underlying environment is modeled as a Markov game where actions are chosen and executed simultaneously, and new observations are perceived simultaneously as a result of a transition to a new state (Littman, 1994, 2001, Hu and Wellman, 2003, Busoniu et al., 2008).
Although agents have individual observations and are responsible for individual actions, each agent only receives the joint reward, and we seek to optimize ${\cal R}_{t}$ as defined above. This is consistent with the Dec-POMDP framework (Oliehoek et al., 2008, Oliehoek and Amato, 2016).
If we denote $\bar{h}:=(h^{1},h^{2},...,h^{d})$ a tuple of agent histories, a joint policy is in general a map $\pi:\mathcal{H}^{d}\to\mathcal{P}(\mathcal{A}^{d})$; we in particular consider policies where for any history $\bar{h}$, the distribution $\pi(\bar{h})$ has independent components in $\mathcal{P}(\mathcal{A})$. Hence, we write $\pi:\mathcal{H}^{d}\to\mathcal{P}(\mathcal{A})^{d}$. The exception is when we use the most naive centralized agent with a combinatorial action space, aka joint action learners.
3 A Deep-RL Architecture for Coop-MARL
Building on purely independent DQN-style agents (see Figure 2), we add enhancements to overcome the identified issues with the MARL problem. Our main contribution of value-decomposition is illustrated by the network in Figure 2.
The main assumption we make and exploit is that the joint action-value function for the system can be additively decomposed into value functions across agents,
$$Q((h^{1},h^{2},...,h^{d}),(a^{1},a^{2},...,a^{d}))\approx\sum_{i=1}^{d}\tilde{%
Q}_{i}(h^{i},a^{i})$$
where the $\tilde{Q}_{i}$ depends only on each agent’s local observations. We learn $\tilde{Q}_{i}$ by backpropagating gradients from the $Q$-learning rule using the joint reward through the summation, i.e. $\tilde{Q}_{i}$ is learned implicitly rather than from any reward specific to agent $i$, and we do not impose constraints that the $\tilde{Q}_{i}$ are action-value functions for any specific reward. The value decomposition layer can be seen in the top-layer of Figure 2. One property of this approach is that, although learning requires some centralization, the learned agents can be deployed independently, since each agent acting greedily with respect to its local value $\tilde{Q}_{i}$ is equivalent to a central arbiter choosing joint actions by maximizing the sum $\sum_{i=1}^{d}\tilde{Q}_{i}$.
For illustration of the idea consider the case with 2 agents (for simplicity of exposition) and where rewards decompose additively across agent observations222Or, more generally, across agent histories., $r({\mathbf{s}},{\mathbf{a}})=r_{1}(o^{1},a^{1})+r_{2}(o^{2},a^{2})$, where $(o^{1},a^{1})$ and $(o^{2},a^{2})$ are (observations, actions) of agents 1 and 2 respectively. This could be the case in team games for instance, when agents observe their own goals, but not necessarily those of teammates. In this case we have that
$$Q^{\pi}({\mathbf{s}},{\mathbf{a}})={\mathbb{E}}[\sum_{t=1}^{\infty}\gamma^{t-1%
}r({\mathbf{s}}_{t},{\mathbf{a}}_{t})|{\mathbf{s}}_{1}={\mathbf{s}},{\mathbf{a%
}}_{1}={\mathbf{a}};\pi]$$
$$={\mathbb{E}}[\sum_{t=1}^{\infty}\gamma^{t-1}r_{1}(o^{1}_{t},a^{1}_{t})|{%
\mathbf{s}}_{1}={\mathbf{s}},{\mathbf{a}}_{1}={\mathbf{a}};\pi]+{\mathbb{E}}[%
\sum_{t=1}^{\infty}\gamma^{t-1}r_{2}(o^{2}_{t},a^{2}_{t})|{\mathbf{s}}_{1}={%
\mathbf{s}},{\mathbf{a}}_{1}={\mathbf{a}};\pi]$$
$$=:\bar{Q}^{\pi}_{1}({\mathbf{s}},{\mathbf{a}})+\bar{Q}^{\pi}_{2}({\mathbf{s}},%
{\mathbf{a}})$$
where $\bar{Q}^{\pi}_{i}({\mathbf{s}},{\mathbf{a}}):={\mathbb{E}}[\sum_{t=1}^{\infty}%
\gamma^{t-1}r_{1}(o^{i}_{t},a^{i}_{t})|{\mathbf{s}}_{1}={\mathbf{s}},{\mathbf{%
a}}_{1}={\mathbf{a}};\pi],i=1,2$.
The action-value function $\bar{Q}^{\pi}_{1}({\mathbf{s}},{\mathbf{a}})$ – agent 1’s expected future return – could be expected to depend more strongly on observations and actions $(o^{1},a^{1})$ due to agent 1 than those due to agent 2. If $(o^{1},a^{1})$ is not sufficient to fully model $\bar{Q}^{\pi}_{1}({\mathbf{s}},{\mathbf{a}})$ then agent 1 may store additional information from historical observations in its LSTM, or receive information from agent 2 in a communication channel, in which case we could expect the following approximation to be valid
$$Q^{\pi}({\mathbf{s}},{\mathbf{a}})=:\bar{Q}^{\pi}_{1}({\mathbf{s}},{\mathbf{a}%
})+\bar{Q}^{\pi}_{2}({\mathbf{s}},{\mathbf{a}})\approx\tilde{Q}^{\pi}_{1}(h^{1%
},a^{1})+\tilde{Q}^{\pi}_{2}(h^{2},a^{2})$$
Our architecture therefore encourages this decomposition into simpler functions, if possible. We see that natural decompositions of this type arise in practice (see Section 4.4).
One approach to reducing the number of learnable parameters, is to share certain network weights between agents.
Weight sharing also gives rise to the concept of agent invariance, which is useful for avoiding the lazy agent problem.
Definition 1 (Agent Invariance).
If for any permutation (bijection) $p:\{1,...,d\}\to\{1,...,d\}$,
$$\pi(p(\bar{h}))=p(\pi(\bar{h}))$$
we say that $\pi$ is agent invariant.
It is not always desirable to have agent invariance, when for example specialized roles are required to optimize a particular system. In such cases we provide each agent with role information, or an identifier. The role information is provided to the agent as a 1-hot encoding of their identity concatenated with every observation at the first layer. When agents share all network weights they are then only conditionally agent invariant, i.e. have identical policies only when conditioned on the same role. We also consider information channels between agent networks, i.e. differentiable connections between agent network modules. These architectures, with shared weights, satisfy agent invariance.
4 Experiments
We introduce a range of two-player domains, and experimentally evaluate the introduced value-decomposition agents with different levels of enhancements, evaluating each addition in a logical sequence. We use two centralized agents as baselines, one of which is introduced here again relying on learned value-decomposition, as well as an individual agent learning directly from the joint reward signal. We perform this set of experiments on the same form of two dimensional maze environments used by Leibo et al. (2017), but with different tasks featuring more challenging coordination needs. Agents have a small $3\times 5\times 5$ observation window, the first dimension being an RGB channel, the second and third are the maze dimensions, and each agent sees a box 2 squares either side and 4 squares forwards, see Figures 2 and 2. The simple graphics of our domains helps with running speed while, especially due to their multi-agent nature and severe partial observability and aliasing (very small observation window combined with map symmetries), they still pose a serious challenge and is comparable to the state-of-the-art in multi-agent reinforcement learning (Leibo et al., 2017), which exceeds what is common in this area (Tuyls and Weiss, 2012).
4.1 Agents
Our agent’s learning algorithm is based on DQN (Mnih et al., 2015) and includes its signature techniques of experience replay and target networks, enhanced with an LSTM value-network as in Hausknecht and Stone (2015) (to alleviate severe partial observability), learning with truncated back-propagation through time, multi-step updates with forward view eligibility traces (Harb and Precup, 2016) (which helps propagating learning back through longer sequences) and the dueling architecture (Wang et al., 2016) (which speeds up learning by generalizing across the action space). Since observations are from a local perspective, we do not benefit from convolutional networks, but use a fully connected linear layer to process the observations.
Our network architectures first process the input using a fully connected linear layer with $32$ hidden units followed by a ReLU layer, and then an LSTM, with $32$ hidden units followed by a ReLU layer, and finally a linear dueling layer, with $32$ units. This produces a value function $V(s)$ and advantage function $A(s,a)$, which are combined to compute a $Q$-function $Q(s,a)=V(s)+A(s,a)$ as described in Wang et al. (2016). Layers of $32$ units are sufficiently expressive for these tasks with limited observation windows.
The architectures (see Appendix B for detailed diagrams) differ between approaches by what is input into each layer. For architectures without centralization or information channels, one observation of size $3\times 5\times 5$ is fed to the first linear layer of $32$ units, followed by the ReLU layer and the LSTM (see Figure 2). For the other (information channels and centralized) agents, $d$ such observations are fed separately to identical such linear layers and then concatenated into $64$ dimensional vectors before passing though ReLUs to an LSTM.
For architectures with information channels we concatenate the outputs of certain layers with those of other agents. To preserve agent invariance, the agent’s own previous output is always included first. For low-level communication, the signal’s concatenation is after the first fully connected layer, while for high-level communication the concatenation takes place on the output of the LSTM layer. Note, that this has the implication that what starts as one agent’s gradients are back-propagated through much of the other agents network, optimizing them to serve the purposes of all agents. Hence, representing in that sense, a higher degree of centralization than the lower-level sharing.
We have found a trajectory length of $8$, determining both the length of the forward view and the length of the back propagation through time is sufficient for these domains. We use an eligibility trace parameter $\lambda=0.9$. In particular, the individual agents learning directly from the joint reward without decomposition or information channels, has worse performance with lower $\lambda$. The Adam (Kingma and Ba, 2014) learning rate scheme initialized with $0.0001$ is uniformly used, and further fine-tuning this per agent (not domain) does not dramatically change the total performance. The agents that we evaluate are listed in the table above.
4.2 Environments
We use 2D grid worlds with the same basic functioning as Leibo et al. (2017), but with different tasks we call Switch, Fetch and Checkers. We have observations of byte values of size $3\times 5\times 5$ (RGB), which represent a window depending on the player’s position and orientation by extending $4$ squares ahead
and $2$ squares on each side. Hence, agents are very short-sighted. The actions are: step forward, step backward,
step left, step right, rotate left, rotate right, use beam
and stand still. The beam has no effect in our games, except for lighting up a row or column of squares straight ahead with yellow. Each player appears as a blue square in its own observation, and the other player, when in the observation window, is shown in red for Switch and Escape, and lighter blue for Fetch. We use three different maps shown in Figure 3 for both Fetch and Switch and a different one for Checkers, also shown in Figure 3 (bottom right). The tasks repeat as the agents succeed (either by full reset of the environment in Switch and Checkers or just by pickup being available again in Fetch), in training for 5,000 steps and 2,000 in testing.
Switch: The task tests if two agents can effectively coordinate their use of available routes on two maps with narrow corridors. The task is challenging because of strong observation aliasing. The two agents appear on different ends of a map, and must reach a goal at the other end. If agents collide in a corridor, then one agent needs to leave the corridor to allow the other to pass. When both players have reached their goal the environment is reset. A point is scored whenever a player reaches a goal.
Fetch: The task tests if two agents can synchronize their behaviour, when picking up objects and returning them to a drop point. In the Fetch task both players start on the same side of the map and have pickup points on the opposite side. A player scores $3$ points for the team for pick-up, and another $5$ points for dropping off the item at the drop point near the starting position. Then the pickup is available to either player again. It is optimal for the agents to cycle such that when one player reaches the pickup point the other returns to base, to be ready to pick up again.
Checkers:
The map contains apples and lemons. The first player is very sensitive and scores $10$ for the team for an apple (green square) and $-10$ for a lemon (orange square). The second, less sensitive player scores $1$ for the team for an apple and $-1$ for a lemon. There is a wall of lemons between the players and the apples. Apples and lemons disappear when collected, and the environment resets when all apples are eaten. It is important that the sensitive agent eats the apples while the less sensitive agent should leave them to its team mate but clear the way by eating obstructing lemons.
4.3 Results
We compare the eight approaches listed in Table 4.1, on the seven tasks. Each is run ten times, with different random seeds determining spawn points in the environment, as well as initializations of the neural networks.
We calculated curves of the average performance over 50,000 episodes (plots in Appendix A) for each approach on each task and we display the normalized area under the curve in Figure 4. Figure 6 displays the normalized final performance averaged over runs and the last 1,000 episodes. Average performance across tasks is also shown for both ways of evaluation.
The very clear conclusion is that architectures based on value-decomposition perform much better, with any combination of other techniques or none, than the centralized approach and individual learners. The centralized agent with value-decomposition is better than the combinatorially centralized as well as individual learners while worse than the more individual agents with value-decomposition.
We particularly see the benefit of shared weights on the hard task of Fetch with one corridor. Without sharing, the individual value-decomposition agent suffers from the lazy agent problem. The agent with weight sharing and role information also perfectly learns the one corridor Fetch task. It performs better than the agent just sharing weights on Switch, where coordination, in particular with one corridor, is easier with non-identical agents. Further, shared weights are problematic for the Checkers task because the magnitude of rewards (and hence the value function) from one agent is ten times higher than for the other agent.
Adding information channels does increase learning complexity because the input comes from more than one agent. However, the checkers task, designed for the purpose, shows that it can be very useful. Overall, the low-level channels where the agent’s LSTM processes the combined observations of both agents turned out to learn faster in our experiments than the more centralized high level communication (after the LSTM).
4.4 The Learned $Q$-Decomposition
Figure 6 shows the learned $Q$-decomposition for the value-decomposition network, using shared weights, in the game of Fetch. A video of the corresponding game can be seen at Video (2017). Spikes correspond to pick-up events (short spikes, 3 reward points), and return events (large spikes, 5 reward points). These are separated into events due to agent 1 (blue spikes) and agent 2 (red spikes). This disambiguation is for illustration purposes only: the environment gives a reward to the whole team for all of these events. The total $Q$-function is seen in yellow, clearly anticipating the team reward events, and dropping shortly afterwards. The component $Q$-functions $\tilde{Q}_{1}$ and $\tilde{Q}_{2}$ for agents 1 and 2 are shown in green and purple. These have generally disambiguated the $Q$-function into rewarding events separately attributable to either player. The system has learned to autonomously decompose the joint $Q$-function into sensible components which, when combined, result in an effective $Q$-function. This would be difficult for independent learners since many rewards would not be observed by both players, see e.g. the situation at 15-16 seconds in the corresponding video available at Video (2017).
5 Conclusions
We study cooperative multi-agent reinforcement learning where only a single joint reward is provided to the agents. We found that the two naive approaches, individual agents learning directly from team reward, and fully centralized agents, provide unsatisfactory solutions as previous literature has found in simpler environments, while our value-decomposition networks do not suffer from the same problems and shows much better performance across a range of more complex tasks. Further, the approach can be nicely combined with weight sharing and information channels, leading to agents that consistently optimally solve our new benchmark challenges.
Value-decomposition networks are a step towards automatically decomposing complex learning problems into local, more readily learnable sub-problems. In future work we will investigate the scaling of value-decomposition with growing team sizes, which make individual learners with team reward even more confused (they mostly see rewards from other agents actions), and centralized learners even more impractical. We will also investigate decompositions based on non-linear value aggregation.
References
Agogino and Tumer (2008)
A. K. Agogino and K. Tumer.
Analyzing and visualizing multiagent rewards in dynamic and
stochastic environments.
Journal of Autonomous Agents and Multi-Agent Systems,
17(2):320–338, 2008.
Babes et al. (2008)
M. Babes, E. M. de Cote, and M. L. Littman.
Social reward shaping in the prisoner’s dilemma.
In 7th International Joint Conference on Autonomous Agents and
Multiagent Systems (AAMAS 2008), Estoril, Portugal, May 12-16, 2008, Volume
3, pages 1389–1392, 2008.
Bernstein et al. (2000)
D. S. Bernstein, S. Zilberstein, and N. Immerman.
The complexity of decentralized control of Markov Decision
Processes.
In UAI ’00: Proceedings of the 16th Conference in Uncertainty
in Artificial Intelligence, Stanford University, Stanford, California, USA,
June 30 - July 3, 2000, pages 32–37, 2000.
Busoniu et al. (2008)
L. Busoniu, R. Babuska, and B. D. Schutter.
A comprehensive survey of multiagent reinforcement learning.
IEEE Transactions of Systems, Man, and Cybernetics Part C:
Applications and Reviews, 38(2), 2008.
Claus and Boutilier (1998)
C. Claus and C. Boutilier.
The dynamics of reinforcement learning in cooperative multiagent
systems.
In Proceedings of the Fifteenth National Conference on
Artificial Intelligence and Tenth Innovative Applications of Artificial
Intelligence Conference, AAAI 98, IAAI 98, July 26-30, 1998, Madison,
Wisconsin, USA., pages 746–752, 1998.
Colby et al. (2016)
M. Colby, T. Duchow-Pressley, J. J. Chung, and K. Tumer.
Local approximation of difference evaluation functions.
In Proceedings of the Fifteenth International Joint Conference
on Autonomous Agents and Multiagent Systems, Singapore, May 2016.
Devlin et al. (2014)
S. Devlin, L. Yliniemi, D. Kudenko, and K. Tumer.
Potential-based difference rewards for multiagent reinforcement
learning.
In Proceedings of the Thirteenth International Joint Conference
on Autonomous Agents and Multiagent Systems, May 2014.
Eck et al. (2016)
A. Eck, L. Soh, S. Devlin, and D. Kudenko.
Potential-based reward shaping for finite horizon online POMDP
planning.
Autonomous Agents and Multi-Agent Systems, 30(3):403–445, 2016.
Foerster et al. (2016)
J. N. Foerster, Y. M. Assael, N. de Freitas, and S. Whiteson.
Learning to communicate with deep multi-agent reinforcement learning.
In Advances in Neural Information Processing Systems 29: Annual
Conference on Neural Information Processing Systems 2016, December 5-10,
2016, Barcelona, Spain, pages 2137–2145, 2016.
Guestrin et al. (2002)
C. Guestrin, M. G. Lagoudakis, and R. Parr.
Coordinated reinforcement learning.
In Proceedings of the Nineteenth International Conference on
Machine Learning, ICML ’02, pages 227–234, San Francisco, CA, USA, 2002.
Morgan Kaufmann Publishers Inc.
ISBN 1-55860-873-7.
URL http://dl.acm.org/citation.cfm?id=645531.757784.
Harb and Precup (2016)
J. Harb and D. Precup.
Investigating recurrence and eligibility traces in deep Q-networks.
In Deep Reinforcement Learning Workshop, NIPS 2016, Barcelona,
Spain, 2016.
Hausknecht (2016)
M. J. Hausknecht.
Cooperation and Communication in Multiagent Deep Reinforcement
Learning.
PhD thesis, The University of Texas at Austin, 2016.
Hausknecht and Stone (2015)
M. J. Hausknecht and P. Stone.
Deep recurrent Q-learning for partially observable MDPs.
CoRR, abs/1507.06527, 2015.
HolmesParker et al. (2016)
C. HolmesParker, A. Agogino, and K. Tumer.
Combining reward shaping and hierarchies for scaling to large
multiagent systems.
Knowledge Engineering Review, 2016.
to appear.
Hu and Wellman (2003)
J. Hu and M. P. Wellman.
Nash q-learning for general-sum stochastic games.
Journal of Machine Learning Research, 4:1039–1069,
2003.
Kingma and Ba (2014)
D. P. Kingma and J. Ba.
Adam: A method for stochastic optimization.
CoRR, abs/1412.6980, 2014.
URL http://arxiv.org/abs/1412.6980.
Kuyer et al. (2008)
L. Kuyer, S. Whiteson, B. Bakker, and N. A. Vlassis.
Multiagent reinforcement learning for urban traffic control using
coordination graphs.
In Machine Learning and Knowledge Discovery in Databases,
European Conference, ECML/PKDD 2008, Antwerp, Belgium, September 15-19,
2008, Proceedings, Part I, pages 656–671, 2008.
Laurent et al. (2011)
G. J. Laurent, L. Matignon, and N. L. Fort-Piat.
The world of independent learners is not Markovian.
Int. J. Know.-Based Intell. Eng. Syst., 15(1):55–64, 2011.
Leibo et al. (2017)
J. Z. Leibo, V. Zambaldi, M. Lanctot, J. Marecki, and T. Graepel.
Multi-agent Reinforcement Learning in Sequential Social Dilemmas.
In Proceedings of the 16th International Conference on
Autonomous Agents and Multiagent Systems (AAMAS 2017), Sao Paulo, Brazil,
2017.
Littman (1994)
M. L. Littman.
Markov games as a framework for multi-agent reinforcement learning.
In Machine Learning, Proceedings of the Eleventh International
Conference, Rutgers University, New Brunswick, NJ, USA, July 10-13, 1994,
pages 157–163, 1994.
Littman (2001)
M. L. Littman.
Friend-or-foe q-learning in general-sum games.
In Proceedings of the Eighteenth International Conference on
Machine Learning (ICML 2001), Williams College, Williamstown, MA, USA, June
28 - July 1, 2001, pages 322–328, 2001.
Mnih et al. (2015)
V. Mnih, K. Kavukcuoglu, D. Silver, A. Rusu, J. Veness, M. Bellemare,
A. Graves, M. Riedmiller, A. Fidjeland, G. Ostrovski, S. Petersen,
C. Beattie, A. Sadik, I. Antonoglou, H. King, D. Kumaran, D. Wierstra,
S. Legg, and D. Hassabis.
Human-level control through deep reinforcement learning.
Nature, 518(7540):529–533, 02 2015.
Mnih et al. (2016)
V. Mnih, A. P. Badia, M. Mirza, A. Graves, T. P. Lillicrap, T. Harley,
D. Silver, and K. Kavukcuoglu.
Asynchronous methods for deep reinforcement learning.
In Proceedings of the 33nd International Conference on Machine
Learning, ICML 2016, New York City, NY, USA, June 19-24, 2016, pages
1928–1937, 2016.
Ng et al. (1999)
A. Y. Ng, D. Harada, and S. J. Russell.
Policy invariance under reward transformations: Theory and
application to reward shaping.
In Proceedings of the Sixteenth International Conference on
Machine Learning (ICML 1999), Bled, Slovenia, June 27 - 30, 1999, pages
278–287, 1999.
Oliehoek and Amato (2016)
F. A. Oliehoek and C. Amato.
A Concise Introduction to Decentralized POMDPs.
SpringerBriefs in Intelligent Systems. Springer, 2016.
Oliehoek et al. (2008)
F. A. Oliehoek, M. T. J. Spaan, and N. A. Vlassis.
Optimal and approximate q-value functions for decentralized pomdps.
J. Artif. Intell. Res. (JAIR), 32:289–353, 2008.
Panait and Luke (2005)
L. Panait and S. Luke.
Cooperative multi-agent learning: The state of the art.
Autonomous Agents and Multi-Agent Systems, 11(3):387–434, 2005.
Proper and Tumer (2012)
S. Proper and K. Tumer.
Modeling difference rewards for multiagent learning (extended
abstract).
In Proceedings of the Eleventh International Joint Conference
on Autonomous Agents and Multiagent Systems, Valencia, Spain, June 2012.
Puterman (1994)
M. Puterman.
Markov Decision Processes: Discrete Stochastic Dynamic
Programming.
Wiley, New York, 1994.
Russell and Norvig (2010)
S. J. Russell and P. Norvig.
Artificial Intelligence: A Modern Approach.
Prentice Hall, Englewood Cliffs, NJ, $3^{nd}$ edition, 2010.
Russell and Zimdars (2003)
S. J. Russell and A. Zimdars.
Q-decomposition for reinforcement learning agents.
In Machine Learning, Proceedings of the Twentieth International
Conference (ICML 2003), August 21-24, 2003, Washington, DC, USA, pages
656–663, 2003.
Schneider et al. (1999)
J. G. Schneider, W. Wong, A. W. Moore, and M. A. Riedmiller.
Distributed value functions.
In Proceedings of the Sixteenth International Conference on
Machine Learning (ICML 1999), Bled, Slovenia, June 27 - 30, 1999, pages
371–378, 1999.
Sukhbaatar et al. (2016)
S. Sukhbaatar, A. Szlam, and R. Fergus.
Learning multiagent communication with backpropagation.
CoRR, abs/1605.07736, 2016.
URL http://arxiv.org/abs/1605.07736.
Sutton and Barto (1998)
R. Sutton and A. Barto.
Reinforcement Learning.
The MIT Press, 1998.
Szepesvári (2010)
C. Szepesvári.
Algorithms for Reinforcement Learning.
Synthesis Lectures on Artificial Intelligence and Machine Learning.
Morgan & Claypool Publishers, 2010.
Tumer and Wolpert (2004)
K. Tumer and D. Wolpert.
A survey of collectives.
In K. Tumer and D. Wolpert, editors, Collectives and the Design
of Complex Systems, pages 1–42. Springer, 2004.
Tuyls and Weiss (2012)
K. Tuyls and G. Weiss.
Multiagent learning: Basics, challenges, and prospects.
AI Magazine, 33(3):41–52, 2012.
van der Pol and Oliehoek (2016)
E. van der Pol and F. A. Oliehoek.
Coordinated deep reinforcement learners for traffic light control.
NIPS Workshop on Learning, Inference and Control of Multi-Agent
Systems, 2016.
Video (2017)
Video.
Video for the q-decomposition plot.
2017.
URL https://youtu.be/aAH1eyUQsRo.
Wang et al. (2016)
Z. Wang, T. Schaul, M. Hessel, H. van Hasselt, M. Lanctot, and N. de Freitas.
Dueling network architectures for deep reinforcement learning.
In Proceedings of the International Conference on Machine
Learning (ICML), pages 1995–2003, 2016.
Appendix A: Plots
Appendix B: Diagrams |
Density-matrix renormalization group study of the linear conductance
in quantum wires coupled to interacting leads or phonons
Jan-Moritz Bischoff
[email protected]
Eric Jeckelmann
[email protected]
Leibniz Universität Hannover, Institut für Theoretische Physik, Appelstraße 2, D-30167
Hannover, Germany
(Draft of December 1, 2020)
Abstract
In a previous paper [J.-M. Bischoff and E. Jeckelmann, Phys. Rev. B 96, 195111 (2017)]
we introduced a density-matrix renormalization group method for calculating the linear conductance of
one-dimensional correlated quantum systems
and demonstrated it on homogeneous spinless fermion chains with impurities.
Here we present extensions of this method to inhomogeneous systems, models with phonons,
and the spin conductance of electronic models. The method is applied to a spinless fermion wire-lead model,
the homogeneous spinless Holstein model, and the Hubbard model. Its capabilities are demonstrated
by comparison with the predictions of Luttinger liquid theory combined with
Bethe Ansatz solutions and other numerical methods.
We find a complex behavior for quantum wires coupled to interacting leads
when the sign of the interaction (repulsive/attractive) differs in wire and leads.
The renormalization of the conductance given by the Luttinger parameter in purely fermionic systems
is shown to remain valid in the Luttinger liquid phase of the Holstein model with phononic degrees of freedom.
I Introduction
Confining electrons to one dimension results in various surprising properties because of the fermionic nature of the
particles [1; 2; 3].
In practice, one-dimensional electron systems can be realized in semiconductor wires [4],
atomic wires deposited on a substrate [5; 6], carbon nanotubes [7],
or atomic and molecular wire junctions [8; 9; 10; 11].
In contrast to the behavior of a three dimensional metal, which is well described by the Fermi liquid paradigm,
theory predicts that one-dimensional conductors should exhibit the behavior of a Luttinger liquid [2].
In particular, the transport properties have been studied and controversially discussed
for more than three decades [2; 3].
In a previous paper [12]
we introduced a density-matrix renormalization group method (DMRG) for calculating the linear conductance of
one-dimensional correlated quantum systems with short-range interactions at zero-temperature based on the Kubo formalism [13; 14].
By taking advantage of the area law for the entanglement entropy [15]
the DMRG method provides us with the most efficient method for calculating the properties of
these systems [16; 17; 18; 19].
Our method combines DMRG with a finite-size scaling of dynamical correlation functions to
compute the conductance in the thermodynamic limit.
The method was demonstrated on the homogeneous spinless fermion chains with impurities [12],
for which well-established results were available [2; 20; 21; 22].
In this paper we develop the method further to treat a variety of complexer models.
We present extensions to wire-lead systems with different interactions in the wire and leads,
to electron-phonon models, and to spinfull fermions (electrons).
These extensions are necessary steps toward future studies of more realistic models that
are able to describe quantitatively
the experimental realizations of one-dimensional electronic
conductors [4; 5; 6; 7; 8; 9; 10; 11].
Here, we apply our method to a spinless fermion wire-lead model, the homogeneous spinless Holstein model, and the Hubbard model
to demonstrate its current capabilities and limitations using the predictions of Luttinger liquid theory combined with
Bethe Ansatz solutions and numerical results found in the literature.
The method is summarized in the next section. The extension and results for inhomogeneous systems such as
the spinless fermion wire-lead system are presented in Sec. III.
The extension to systems with phonon degrees of freedom and the results for the spinless Holstein model
are shown in Sec. IV.
Section V describes the generalization to spinfull fermion (electron) model and to the calculation of the spin
conductance as well as our results for the spin conductance of the Hubbard model.
Finally, Sec. VI contains our conclusion and outlook.
II Method
In this section we summarize our DMRG method for calculating the conductance of one-dimensional correlated quantum systems.
Full details can be found in our previous paper [12].
We consider a one-dimensional lattice of $M$ sites. It consists of left and right segments (called leads)
with approximately $(M-M_{\text{W}})/2$ sites
each and a central segment (called wire) of $M_{\text{W}}$ sites. A current can be generated by applying a potential bias. We assume that the potential
remains constant in the leads and drop linearly in the wire, i.e. it has the shape
$$C(j)=\begin{cases}\hfill\hphantom{-}\frac{1}{2}\hfill&\text{ for $j\leq j_{1}$%
}\\
\hfill-\frac{j-j_{1}}{j_{2}-j_{1}}+\frac{1}{2}\hfill&\text{ for $j_{1}\leq j%
\leq j_{2}$}\\
\hfill-\frac{1}{2}\hfill&\text{ for $j\geq j_{2}$}\end{cases}$$
(1)
where $j_{1}$ and $j_{2}$ are the first and last site in the wire, respectively.
This is illustrated in Fig. 1.
The resulting linear conductance can be calculated directly from dynamical current-current correlation functions
in the thermodynamic limit using Kubo’s linear response theory [13].
The dynamical DMRG method [23; 24] can be used to compute the imaginary part of the frequency-resolved current-current correlation functions
for a finite lattice size $M$
$$G_{J,\eta}(\omega)=\left\langle 0\left|J\frac{\eta}{(E_{0}-H+\hbar\omega)^{2}+%
\eta^{2}}J\right|0\right\rangle$$
(2)
where $H$ is the Hamilton operator, $|0\rangle$ is its ground state, $E_{0}$ is the ground-state energy,
$J$ is the current operator in the wire, and $\eta>0$ is a small number.
The Hamilton operator and the current operator depend on the system studied and thus will be given explicitly
in the next sections where we present our results for specific models.
It is important to realize, that the wire segment (and consequently both lead segments) is defined by the potential profile
(1) and thus, in practice, by the lattice region, where the current operator $J$ acts.
It is possible but not necessary to include any information about the wire-lead layout in the Hamilton operator $H$,
such as different interactions or different degrees of freedom for wire and leads.
In fact, in our first work we tested
our method with homogeneous systems only, i.e. the degrees of freedom and interactions were identical for all sites,
but for the presence of up to two impurities (on-site potentials) in the middle of the wire.
From the DMRG data for (2) we can calculate the “finite-size conductance”
$$G(M)=\left.\frac{q^{2}}{\omega}\left[G_{J,\eta(M)}(\omega)-G_{J,\eta(M)}(-%
\omega)\right]\right|_{\omega=0}$$
(3)
for a given system size $M$ using the scaling $\eta(M)=C/M$, where
$q$ is the charge carried by each fermion.
The constant $C$ depends on the properties of the system investigated.
For all results presented here we used $C=48t$ or $96t$.
Extrapolating $G(M)$ to infinite system length $M$ yields our estimation of the conductance $G$ in the thermodynamic limit.
Here, we used chain lengths up to $M=2000$.
For a noninteracting tight-binding chain it can be shown analytically that $G(M)$ converges to the exact result for $M\rightarrow\infty$,
i.e. the quantum of conductance
$$G_{0}=n\frac{q^{2}}{h}$$
(4)
where $n=1$ for spinless fermions and $n=2$ for electrons.
For correlated systems we can only compute (2) and (3)
numerically.
In practice, there are two main practical issues. First, DMRG must be able to calculate
$G(M)$ with enough accuracy for large system sizes $M$. Second,
$G(M)$ should converge neatly to the actual value of the conductance for $M\gg 1$.
In our previous work we showed, that this method yields correct results for the renormalized
conductance of the one-component Luttinger liquid in the homogeneous spinless fermion chain
as well as for one and two impurities added to the chain.
In the next sections we will show that this approach can also be applied to
inhomogeneous systems, electronic Hamiltonians and models with phonons.
In all our numerical results and figures we use $q=1$, and $\hbar=1$.
This yields $G_{0}=\frac{n}{2\pi}$. Therefore, we show $2\pi G(M)$ in our figures.
III Inhomogeneous wire-lead system
As mentioned above our method was demonstrated in Ref. [12]
for homogeneous systems only.
In real systems, however, wire and leads are of different nature.
For instance, it is often assumed that metallic leads are Fermi liquids and thus they are modelled
by noninteracting tight-binding chains attached to the interacting
wire [25; 26; 14; 27; 28; 29; 30; 31; 32].
More generally, one may have to consider
different interactions in the leads and in the wire as well as additional degrees of freedom (such as phonons)
in the wires.
Therefore, we have generalized our method to this situation and tested it on a spinless fermion model with different nearest-neighbor interactions
in the wire ($V_{\text{W}}$) and in the leads ($V_{\text{L}}$). Explicitly, this corresponds
to the Hamiltonian
$$\displaystyle H=$$
$$\displaystyle-t\sum\limits_{j=2}^{M}\left(c^{\dagger}_{j}c_{j-1}+c^{\dagger}_{%
j-1}c_{j}\right)$$
(5)
$$\displaystyle+V_{\text{W}}\sum\limits_{j=j_{1}+1}^{j_{2}}\left(n_{j}-\frac{1}{%
2}\right)\left(n_{j-1}-\frac{1}{2}\right)$$
$$\displaystyle+V_{\text{L}}\sum\limits_{j=2}^{j_{1}}\left(n_{j}-\frac{1}{2}%
\right)\left(n_{j-1}-\frac{1}{2}\right)$$
$$\displaystyle+V_{\text{L}}\sum\limits_{j=j_{2+1}}^{M}\left(n_{j}-\frac{1}{2}%
\right)\left(n_{j-1}-\frac{1}{2}\right).$$
The operator $c^{\dagger}_{j}$ ($c_{j}$) creates (annihilates) a spinless fermion at position $j$ ($=1,\dots M$)
while $n_{j}=c^{\dagger}_{j}c_{j}$ counts the number of spinless fermions on this site.
The hopping amplitude is identical for wire and leads and sets the energy unit $t=1$.
We assume that the system has an even number of sites $M$ and is occupied by $M/2$ spinless fermions (half filling).
From the potential profile (1) follows that the current operator is given by
$$J=\frac{1}{M_{W}-1}\frac{it}{\hbar}\sum\limits_{j=j_{1}+1}^{j_{2}}\left(c^{%
\dagger}_{j}c_{j-1}-c^{\dagger}_{j-1}c_{j}\right).$$
(6)
In principle, the wire edge sites $j_{1}$ and $j_{2}$ in the definition of the Hamiltonian (5)
could be different from those used for the potential profile (and thus the current operator).
Here we assume that the screening of the potential difference between source and drain occurs
entirely (and linearly) in the wire and thus we use the same values for $j_{1}$ and $j_{2}$ in both the Hamilton operator (5)
and the current operator (6).
Field-theoretical approaches predict that
the conductance of a homogeneous Luttinger liquid is renormalized as
$$G=KG_{0}$$
(7)
where $K$ is the Luttinger parameter [20; 21; 22].
We could reproduce this result in our previous work for the homogeneous spinless fermion model,
i.e., for $V_{\text{W}}=V_{\text{L}}$ in (5).
On the contrary, field-theoretical approaches predict that
the conductance through a Luttinger liquid wire connected
to leads is not renormalized by interactions in the wire and is
entirely determined by the lead properties [33; 34; 35; 36; 37].
In particular, if the leads are Luttinger liquids their Luttinger parameter $K_{\text{L}}$
determines the conductance through (7) [34; 36].
However, field-theoretical approaches assume perfect contacts, i.e. without any single-particle
backscattering [36].
Numerical calculations for lattice models of quantum wires connected
to noninteracting leads show that the conductance deviates from the lead conductance $G_{0}$
when backscattering is taken into account.
In the lattice model (5)
single-particle backscattering is due to the sharp change of the on-site potential
at the boundaries between wire and leads
[$-V_{\text{W}}$ in the wire, $-V_{\text{L}}$ in the leads, and
$-(V_{\text{W}}+V_{\text{L}})/2$ on the boundary sites $j_{1}$ and $j_{2}$],
which is required to maintain a constant density in equilibrium.
We have found that the calculation of the frequency-resolved correlation
function (2) with dynamical DMRG is not more difficult for an inhomogeneous system
($V_{\text{W}}\neq V_{\text{L}}$) than for a homogeneous one ($V_{\text{W}}=V_{\text{L}}$).
To investigate the conductance of a Luttinger liquid between leads, however, we
would have to compute the conductance $G(M)$ in the limit of long wires $M_{\text{W}}\gg 1$
while maintaining much longer leads ($M\gg M_{\text{W}})$.
In our study of homogeneous systems [12] we found that the conductance scales
with $M_{\text{W}}/M$ and thus we were able to determine its value in the thermodynamic limit
using a fixed (and small) value of $M_{\text{W}}$. For the wire–lead system (5)
with $V_{\text{W}}\neq V_{\text{L}}$ considered here, we have found that the scaling
with $M$ and $M_{\text{W}}$ (up to $M_{\text{W}}=82$) is much more complicated. As a consequence, it is not possible
to compute the conductance of an infinitely long wire in most cases. Thus we discuss
here the results obtained for short wires and present only results for $M_{\text{W}}=10$.
In Fig. 2 we show the scaling of $G(M)$ with the system length $M$ for noninteracting
leads ($V_{\text{L}}=0$) and various interaction strengths $V_{\text{W}}$ in the wire.
The extrapolated values $G=\lim_{M\rightarrow\infty}G(M)$ are shown in Fig. 3.
For repulsive interactions ($V_{\text{W}}>0$) we observe a progressive decrease
of $G$ with increasing coupling $V_{\text{W}}$.
The conductance of the spinless fermion model connected to noninteracting
leads was investigated previously for repulsive interactions
using DMRG and the functional renormalization group
method [25; 26].
Our results for $G$ agree quantitatively with the (DMRG) results presented in Ref. [25]
for a slightly longer wire ($M_{\text{W}}=12$), as shown in Fig. 3.
In contrast, we observe in Fig. 2
that $G(M)$ tends to $G_{0}$ (i.e., the conductance of the noninteracting leads)
for attractive interactions $0>V_{\text{W}}>-2t$.
The non-monotonic behavior of $G(M)$ for $V_{\text{W}}=-1.9t$ in Fig. 2
and the resulting large deviation of the extrapolated value from $G_{0}$ in Fig. 3 illustrate
the problems that one encounters when analyzing finite-size effects in the wire-lead system.
For an interaction $V_{\text{W}}>2t$ the homogeneous half-filled spinless fermion model is in an insulating
charge-density-wave (CDW) phase and thus we expect that $G$ diminishes exponentially
with increasing wire length $M_{\text{W}}$.
In the Luttinger liquid phase with repulsive interactions ($2t\geq V_{\text{W}}>0$)
a power-law suppression of $G$ with increasing wire length was observed previously [26].
This deviation from the field-theoretical predictions mentioned above is due to the
single-particle backscattering caused by the sharp change of the interaction
at the wire boundaries.
The scaling of $G$ with the wire length in the attractive Luttinger liquid phase
($-2t<V_{\text{W}}<0$) is not known.
As explained above, we can not determine the conductance for infinitely long wire
($M_{\text{W}}\rightarrow\infty$) with the present method.
Nevertheless, our data suggest that for attractive wire interactions within the Luttinger liquid regime the noninteracting leads
determine the conductance of the system in agreement with field theory. This is in stark contrast to repulsive wire interactions,
for which even a small $V_{\text{W}}>0$ results in a reduction of $G$ from the noninteracting lead conductance $G_{0}$,
as already reported previously [36; 25; 26].
While there are numerous studies of the conductance of wires connected to noninteracting
leads [33; 34; 35; 36; 37; 25; 26; 14; 27; 28; 29; 30; 31; 32],
there seem to be only a few field-theoretical results for interacting leads [34; 36; 31].
Figure 4 shows the scaling of $G(M)$ with the system size $M$ for leads with a repulsive
interaction $V_{\text{L}}=2t$. From the Bethe Ansatz solution of the homogeneous spinless fermion model
we know that this corresponds to a Luttinger liquid parameter $K_{\text{L}}=1/2$ [38]
and thus to a conductance $G=G_{0}/2$ for perfect contacts according to field theory.
Clearly, all our results for $G(M)$ with $V_{\text{W}}>0$ converge toward this value and thus agree
with the field-theoretical predictions.
Surprisingly, this also holds for the wire with $V_{\text{W}}=3t$, which
would correspond to a CDW insulator in the thermodynamic limit $M_{\text{W}}\rightarrow\infty$.
Therefore, we expect that $G$ will decrease with increasing $M_{\text{W}}$
for $V_{\text{W}}=3t$.
For a noninteracting wire or an attractive interaction in the wire ($0\geq V_{\text{W}}>-2t$), however,
we observe in Fig. 4
a different, non-monotonic behavior of $G(M)$, which hinders the extrapolation of $G(M)$ to the limit $M\rightarrow\infty$.
Apparently, $G$ is lower than the lead conductance $G_{0}/2$ for $V_{\text{W}}=0$ and $-t$.
Figure 5 shows the scaling of $G(M)$ with system length $M$
for leads with an attractive interaction $V_{\text{L}}=-t$.
This corresponds to a Luttinger liquid parameter $K_{\text{L}}=3/2$ [38]
and thus to a conductance $G=(3/2)G_{0}$ for perfect contacts according to field theory.
We see in Fig. 5 that $G(M)$ seems to converge to different values for different
repulsive couplings in the wire ($V_{\text{W}}>0$).
Unfortunately, we cannot determine the limit $G=\lim_{M\rightarrow\infty}G(M)$
accurately even with system lengths up to $M=2000$.
The values of $G(M)$ for $M=2000$ are above the conductance of an homogeneous
Luttinger liquid with the corresponding interaction $V_{\text{W}}$. They seem to increase further with $M$
but to converge to lower values than the lead conductance $G=(3/2)G_{0}$.
Therefore, it appears that the conductance of the wire-lead system is lower than the lead conductance in
that case, similarly to what is found for noninteracting leads due to single-particle backscattering.
For $V_{\text{W}}=3t$ the conductance seems again to converge to a finite value. As for noninteracting leads
it should vanish in the limit $M_{\text{W}}\rightarrow\infty$ as this interaction strength leads to
a CDW insulating state in the wire.
For a noninteracting wire and for attractive interactions in the wire ($0\geq V_{\text{W}}\geq-2t$), however, we find
that $G\approx(3/2)G_{0}$ as predicted by field theory but it is difficult to extrapolate $G(M)$ reliably
close to the Luttinger phase boundary $V_{\text{W}}=-2t$.
In summary, our results agree with the known results for the conductance of the spinless
fermion wire-lead system and thus confirm the validity of our method.
Moreover, they also reveal complex qualitative differences in the attractive and repulsive
interaction regimes.
The conductance seems to be given by (7) with the lead Luttinger parameter
(as predicted by field theory) only when the interactions have the same sign in wire and leads.
The properties of short wires with attractive interactions connected to leads with repulsive
interactions (or vice versa) seem to be more complex than anticipated from the results reported so far
(using field theory) and should be investigated further.
IV Holstein chain
The interaction between charge carriers and phonons plays an important role for
transport properties [39; 40; 41], in particular
for atomic or molecular wire junctions between leads [8; 9; 10; 11].
Therefore,
we have to generalize our method to models that include phonon degrees of freedom coupled
to the charge carriers.
It is relatively straightforward to extend a DMRG program to models with phonons if one uses
a truncated eigenbasis of the boson number operator with at most $N_{\text{b}}$ states
to represent each phonon mode [42]. The only real challenge is the computational cost that
increases as $N^{3}_{\text{b}}$ and thus limit most applications to small $N_{\text{b}}$.
For Einstein phonons (i.e. dispersionless) and purely local couplings with the fermion degrees
of freedom
it is possible to speed up DMRG calculations using a pseudo-site representation for bosons described in [43].
Therefore, we have tested our method on the simplest model with these properties,
the spinless fermion Holstein model [44],
which is defined by the Hamiltonian
$$\displaystyle H=$$
$$\displaystyle-t\sum\limits_{j=2}^{M}\left(c^{\dagger}_{j}c_{j-1}+c^{\dagger}_{%
j-1}c_{j}\right)+\omega_{\text{b}}\sum\limits_{j=1}^{M}b^{\dagger}_{j}b_{j}$$
$$\displaystyle-g\,\omega_{\text{b}}\sum\limits_{j=1}^{M}\left(b^{\dagger}_{j}+b%
_{j}\right)n_{j},$$
(8)
where $\omega_{\text{b}}$ is the frequency of the Einstein phonons created (annihilated) by the boson operators $b^{\dagger}_{j}$ ($b_{j}$)
and $g$ is the dimensionless coupling between spinless fermions and phonons.
The spinless fermion operators $c^{\dagger}_{j}$, $c_{j}$, and $n_{j}$ have the same meaning as in the previous section
and the hopping term again sets the energy unit $t=1$, while we consider a system with an even number of sites $M$ filled
with $M/2$ spinless fermions.
This Hamiltonian is homogeneous and thus the wire and lead sections are defined by the potential profile (1) only.
The current operator is again given by (6).
We have found that calculating the conductance with our method
is more difficult for the Holstein model than for purely fermionic models such as (5).
This is due not only to the increase of the computational cost with the phonon cutoff $N_{\text{b}}$
but also to the fact that the dynamical DMRG algorithm [23; 24] often fails to converge in a reasonable time
for large $M$. Thus we can compute $G(M)$ systematically for smaller system sizes $M$ than in fermionic systems
and we show here results for $M$ up to $800$ sites only.
(One calculation for this system size requires about 600 CPU hours and 4GB of memory. Thus
one could certainly treat larger systems using supercomputer facilities.)
In contrast, the scaling of $G(M)$
with the system size is regular in the homogeneous Holstein wire as found previously for homogeneous
fermionic systems [12]. Thus we can estimate the conductance $G$ in the thermodynamic limit despite the short system lengths
as illustrated in Fig. 6 for the adiabatic regime ($\omega_{\text{b}}=0.1t$)
and in Fig. 7
for the intermediate regime ($\omega_{\text{b}}=t$).
The phase diagram of the half-filled spinless fermion Holstein model
was determined previously with DMRG [45] and quantum Monte Carlo computations [46].
It exhibits a Luttinger liquid phase at weak coupling or high phonon frequency.
The Luttinger liquid parameter $K$ was calculated from the long-wavelength limit
of the static structure factor [45; 46]. This corresponds to the
parameter $K$ determining the exponents in the power-law correlation functions
of the Luttinger liquid theory for purely fermionic systems [2].
As seen in Fig. 6,
we have found that the conductance corresponds to $K=G/G_{0}\approx 1$ in the adiabatic regime $\omega_{\text{b}}=0.1t$
for a coupling up to at least $g=1.5$ in agreement with the values determined from the
structure factor [45; 46].
In the intermediate regime $\omega_{\text{b}}=t$ shown in Fig. 7, the conductance $G(M)$
clearly decreases with increasing coupling $g$. The extrapolated values of $G$ for $1/M\rightarrow 0$
yield parameters $K=G/G_{0}$ that agree well with the values determined from
the structure factor in Ref. [45], as shown in Fig. 8.
In particular, we do not observe any sign of superconducting fluctuations ($K>1$ or $G>G_{0}$).
Unfortunately, we are not able to determine the conductance close
to the metal-insulator transition because this requires too much computational resources
(the phonon cutoff $N_{\text{b}}$ should be larger than the value $N_{\text{b}}=4$ used here).
In summary, our results for the conductance in the spinless fermion Holstein model
agree with the Luttinger parameter $K$ defined from the power-law decay of correlation functions
(or equivalently from the static structure factor).
This confirms the validity of our method for models with charge carriers
coupled to phonon degrees of freedom. It also shows that the relation (7)
for the conductance,
which was derived for purely fermionic systems using field theory [2; 20; 21; 22],
remains valid for the Luttinger liquid phase of models including phonons.
V Hubbard chain
A necessary but simple extension of our method is the generalization to electrons (i.e., fermions with spin $\frac{1}{2}$).
To illustrate this extension we consider the one-dimensional homogeneous Hubbard chain
with the Hamiltonian
$$H=-t\sum\limits_{j=2}^{M}\sum_{\sigma}\left(c^{\dagger}_{j,\sigma}c_{j-1,%
\sigma}+\text{h.c.}\right)+U\sum\limits_{j=1}^{M}n_{j,\uparrow}n_{j,\downarrow},$$
(9)
where $c^{\dagger}_{j,\sigma}$ ($c_{j,\sigma}$) creates (annihilates) an electron with spin $\sigma(=\uparrow,\downarrow)$
on site $j$ and $n_{j,\sigma}=c^{\dagger}_{j,\sigma}c_{j,\sigma}$ is the corresponding particle number operator.
The strength of the on-site coupling between electrons is given by the Hubbard parameter $U$
and the hopping term again sets the energy unit $t=1$.
We assume that the system length $M$ is even and that the system contains $M/2$ electrons of each spin (half filling).
The current operator for the electrons with spin $\sigma$ in the wire is
a simple generalization of (6)
$$J_{\sigma}=\frac{1}{M_{\text{W}}-1}\frac{it}{\hbar}\sum\limits_{j=j_{1}+1}^{j_%
{2}}\left(c^{\dagger}_{j,\sigma}c_{j-1,\sigma}-c^{\dagger}_{j-1,\sigma}c_{j,%
\sigma}\right).$$
(10)
The current operator for the charge transport is then given by
$$J_{\text{C}}=J_{\uparrow}+J_{\downarrow}.$$
(11)
Using this definition in Eqs. (2) and (3)
we obtain the (charge) conductance $G(M)$ of the electronic system.
The current operator for the spin transport is similarly given by
$$J_{\text{S}}=J_{\uparrow}-J_{\downarrow}.$$
(12)
Using this definition in Eqs. (2) and (3)
we obtain the spin conductance $G_{\text{S}}(M)$ of the electronic system.
This describes the linear response of the system to an external magnetic field with
the profile (1).
The one-dimensional Hubbard model is exactly solvable using the Bethe Ansatz method [47; 48].
Its low-energy properties are characterized by a separation of charge and spin excitations
and its gapless excitation modes are described by the Luttinger liquid theory.
At half filling there is an exact symmetry between the charge (spin) properties for $U$ and
the spin (charge) properties for a coupling $-U$. In particular, the ground state is a Mott insulator
with gapped charge excitations and gapless spin excitations for $U>0$, while
it is a Luther-Emery liquid [2] with gapless charge excitations and gapped spin excitations for $U<0$.
We have found that calculating the (charge or spin) conductance with our method is only slightly more difficult
for the homogeneous electronic model (9) with $U\neq 0$ than for the homogeneous spinless fermion model.
This is understandable because in both cases there is only one gapless excitation mode in the system.
As expected the charge conductance $G(M)$ for any $U$ is equal within the numerical errors
to the spin conductance $G_{\text{S}}(M)$ for an interaction $-U$. Thus we discuss only the latter case.
In Fig. 9 the scaling of $G_{\text{S}}(M)$ with $M$ is plotted for several values of the interaction $U$.
For noninteracting electrons $G_{\text{S}}(M)$ approaches $G_{0}$ as expected. [Note that the quantum of conductance (4)
for electrons ($n=2$) equals $2\pi G_{0}=2$ in our units].
The scattering of the $G_{\text{S}}(M)$ data and the resulting poor extrapolation are due to the lesser
accuracy of DMRG for a chain with two gapless excitation modes (i.e., for $U=0$) than for a single one (i.e. for $U\neq 0$).
For $U<0$, $G_{\text{S}}(M)$ vanishes in the thermodynamic limit as required for a phase with gapped spin excitations.
The qualitatively different behaviors for weak and strong attractive interactions is probably due to
the different correlation length, which is smaller than $M_{\text{W}}=10$ for $U=-8t$
but larger for $U=-2t$.
The Luttinger liquid parameter for the spinless spin excitations is $K_{\text{S}}=1$ in the Mott insulating phase at half
filling for any $U>0$.
Concordantly, we observe in Fig. 9 that $G_{\text{S}}(M)$ approaches $G_{0}$ for a weak repulsive interaction ($U=2t$).
For stronger interactions the convergence of $G_{\text{S}}(M)$ toward $G_{0}$ is less clear as shown by the results
for $U=8t$ in Fig. 9.
This is due to the small band width of the gapless spin excitations (of the order of $4t^{2}/U$ for large $U$)
which requires a smaller broadening $\eta\sim 4t^{2}/U$ and thus a larger $M$ to reach the same resolution
for the dynamical correlation functions (2).
VI Conclusion and Outlook
We have extended the DMRG method for the linear conductance of one-dimensional correlated lattice models to
more complex systems including interacting leads, coupling to phonons, and the electron spin properties.
The tests conducted in this work reveal
intriguing differences
when the sign of the interaction (attractive/repulsive) differs in wire and leads
and confirm that the renormalization of the conductance by the Luttinger parameter (7)
remains valid for the Luttinger liquid phase of homogeneous models with coupling to phonons.
Unfortunately, we have found that we cannot determine the conductance of a
Luttinger liquid wire between leads with a different interaction but can only investigate short
wires because of the complicated finite-size scaling.
Moreover, for models with coupling to phonon degrees of freedom the high computational cost of
dynamical DMRG calculations limits the system sizes that can be used.
Nevertheless,
the extensions presented in this work will certainly allows us to compute the conductance
of short electron-phonon-coupled wires connected to interacting leads (without phonons).
Therefore, we will be able to study more realistic models for realizations of quasi-one-dimensional
electronic conductors such as
atomic or molecular wire junctions [8; 9; 10; 11].
Moreover, we point out that our approach is not limited to the charge and spin conductance but could be extended to other
transport properties that are described by a local current operator, such as the energy current [49; 50; 51; 52].
Acknowledgements.
Jan Bischoff would like to thank the Lower Saxony PhD-Programme Contacts in Nanosystems for financial support.
We also acknowledge support from the DFG (Deutsche Forschungsgemeinschaft) through Grant No. JE 261/2-2 in the
Research Unit Advanced Computational Methods for Strongly Correlated Quantum Systems (FOR 1807).
The cluster system at the Leibniz Universität Hannover was used for the computations.
References
[1]
D. Baeriswyl and L. Degiorgi, Strong Interactions in Low Dimensions
(Kluwer Academic Publishers,
Dordrecht, 2004).
[2]
T. Giamarchi, Quantum Physics in One
Dimension, International Series of Monographs on Physics
(Clarendon Press, 2003).
[3]
A. Kawabata, Electron conduction in
one-dimension,
Rep.
Progr. Phys. 70, 219
(2007).
[4]
S. Tarucha, T. Honda, and
T. Saku, Reduction of quantized conductance
at low temperatures observed in 2 to 10 $\mu$m-long quantum wires,
Solid
State Communications
94, 413
(1995).
[5]
C. Blumenstein, J. Schäfer,
S. Mietke, S. Meyer,
A. Dollinger, M. Lochner,
X. Y. Cui, L. Patthey,
R. Matzdorf, and R. Claessen,
Atomically controlled quantum chains hosting a
Tomonaga-Luttinger liquid,
Nat Phys
7,
776
(2011).
[6]
Y. Ohtsubo, J.-i. Kishi,
K. Hagiwara, P. Le Fèvre,
F. Bertran, A. Taleb-Ibrahimi,
H. Yamane, S.-i. Ideta,
M. Matsunami, K. Tanaka, and
S.-i. Kimura, Surface
Tomonaga-Luttinger-liquid state on $\mathrm{Bi}/\mathrm{InSb}(001)$,
Phys.
Rev. Lett. 115, 256404
(2015).
[7]
M. Bockrath, D. H. Cobden,
J. Lu, A. G. Rinzler,
R. E. Smalley, L. Balents, and
P. L. McEuen, Luttinger-liquid behaviour in
carbon nanotubes,
Nature
397,
598
(1999).
[8]
N. Agraït, C. Untiedt,
G. Rubio-Bollinger, and S. Vieira,
Onset of energy dissipation in ballistic atomic wires,
Phys.
Rev. Lett. 88, 216803
(2002).
[9]
N. Agraït, C. Untiedt,
G. Rubio-Bollinger, and S. Vieira,
Electron transport and phonons in atomic wires,
Chemical
Physics
281, 231
(2002).
[10]
H. Ness and A. Fisher,
Coherent electron injection and transport in molecular wires:
inelastic tunneling and electron–phonon interactions,
Chemical
Physics
281, 279
(2002).
[11]
J. Hihath and N. Tao,
Electron–phonon interactions in atomic and molecular
devices,
Progress
in Surface Science
87, 189
(2012).
[12]
J.-M. Bischoff and E. Jeckelmann,
Density-matrix renormalization group method for the
conductance of one-dimensional correlated systems using the Kubo formula,
Phys.
Rev. B 96, 195111
(2017).
[13]
R. Kubo, Statistical-Mechanical Theory of
Irreversible Processes. I. General Theory and Simple
Applications to Magnetic and Conduction Problems,
J. Phys. Soc.
Jpn. 12, 570
(1957).
[14]
D. Bohr, P. Schmitteckert, and
P. Wölfle, DMRG evaluation of the
Kubo formula — Conductance of strongly interacting quantum systems,
EPL
(Europhysics Letters)
73, 246
(2006).
[15]
J. Eisert, M. Cramer, and
M. B. Plenio, Colloquium: Area laws for the
entanglement entropy,
Rev.
Mod. Phys. 82, 277
(2010).
[16]
S. R. White, Density matrix formulation for
quantum renormalization groups,
Phys.
Rev. Lett. 69, 2863
(1992).
[17]
S. R. White, Density-matrix algorithms for
quantum renormalization groups,
Phys.
Rev. B 48, 10345
(1993).
[18]
U. Schollwöck, The density-matrix
renormalization group in the age of matrix product states,
Ann.
Phys. 326, 96
(2011).
[19]
E. Jeckelmann, Density-Matrix
Renormalization Group Algorithms, Computational
Many-Particle Physics, edited by H. Fehske,
R. Schneider, and A. Weiße, volume
739 of Lecture Notes in Physics,
chapter 21, pp. 597–619
(Springer Berlin Heidelberg, 2008).
[20]
W. Apel and T. M. Rice,
Combined effect of disorder and interaction on the
conductance of a one-dimensional fermion system,
Phys.
Rev. B 26, 7063
(1982).
[21]
C. L. Kane and M. P. A. Fisher,
Transport in a one-channel Luttinger liquid,
Phys.
Rev. Lett. 68, 1220
(1992).
[22]
C. L. Kane and M. P. A. Fisher,
Transmission through barriers and resonant tunneling in an
interacting one-dimensional electron gas,
Phys.
Rev. B 46, 15233
(1992).
[23]
E. Jeckelmann, Dynamical density-matrix
renormalization-group method,
Phys.
Rev. B 66, 045114
(2002).
[24]
E. Jeckelmann and H. Benthien, Dynamical density-matrix renormalization group, Computational Many-Particle Physics, edited by
H. Fehske, R. Schneider, and
A. Weiße, volume 739 of
Lecture Notes in Physics, chapter 22,
pp. 621–635 (Springer Berlin
Heidelberg, 2008).
[25]
V. Meden and U. Schollwöck,
Conductance of interacting nanowires,
Phys.
Rev. B 67, 193303
(2003).
[26]
V. Meden, S. Andergassen,
W. Metzner, U. Schollwöck, and
K. Schönhammer, Scaling of the
conductance in a quantum wire,
Europhys.
Lett. 64, 769
(2003).
[27]
D. Bohr and P. Schmitteckert,
Strong enhancement of transport by interaction on contact
links,
Phys.
Rev. B 75, 241103
(2007).
[28]
A. Branschädel, G. Schneider, and
P. Schmitteckert, Conductance of
inhomogeneous systems: Real-time dynamics,
Ann. Phys.
(Berlin) 522, 657
(2010).
[29]
F. Heidrich-Meisner, I. González,
K. A. Al-Hassanieh, A. E. Feiguin,
M. J. Rozenberg, and E. Dagotto,
Nonequilibrium electronic transport in a one-dimensional
Mott insulator,
Phys.
Rev. B 82, 205110
(2010).
[30]
L. G. G. V. Dias da Silva, K. A.
Al-Hassanieh, A. E. Feiguin, F. A.
Reboredo, and E. Dagotto, Real-time
dynamics of particle-hole excitations in Mott insulator-metal junctions,
Phys.
Rev. B 81, 125113
(2010).
[31]
D. Morath, N. Sedlmayr,
J. Sirker, and S. Eggert,
Conductance in inhomogeneous quantum wires: Luttinger liquid
predictions and quantum Monte Carlo results,
Phys.
Rev. B 94, 115162
(2016).
[32]
F. Lange, S. Ejima,
T. Shirakawa, S. Yunoki, and
H. Fehske, Spin transport through a
spin-$\frac{1}{2}$ XXZ chain contacted to fermionic leads,
Phys.
Rev. B 97, 245124
(2018).
[33]
V. V. Ponomarenko, Renormalization of the
one-dimensional conductance in the Luttinger-liquid model,
Phys.
Rev. B 52, R8666
(1995).
[34]
I. Safi and H. J. Schulz,
Transport in an inhomogeneous interacting one-dimensional
system,
Phys.
Rev. B 52, R17040
(1995).
[35]
D. L. Maslov and M. Stone,
Landauer conductance of Luttinger liquids with leads,
Phys.
Rev. B 52, R5539
(1995).
[36]
K. Janzen, V. Meden, and
K. Schönhammer, Influence of the contacts
on the conductance of interacting quantum wires,
Phys.
Rev. B 74, 085301
(2006).
[37]
R. Thomale and A. Seidel,
Minimal model of quantized conductance in interacting
ballistic quantum wires,
Phys.
Rev. B 83, 115330
(2011).
[38]
J. Sirker, The Luttinger liquid and
integrable models,
Int. J.
Mod. Phys. B 26, 1244009
(2012).
[39]
J. M. Ziman, Electrons and Phonons: The
Theory of Transport Phenomena in Solids (Clarendon
Press, Oxford, 1960).
[40]
G. D. Mahan, Many-Particle Physics
(Kluwer Academic Publishers, New
York, 2000).
[41]
J. Sólyom, Fundamentals of the
Physics of Solids, Volume 2 - Electronic Properties
(Springer, Berlin,
2009).
[42]
E. Jeckelmann and H. Fehske,
Exact numerical methods for electron-phonon problems,
Riv.
Nuovo Cimento 30, 259
(2007).
[43]
E. Jeckelmann and S. R. White,
Density-matrix renormalization-group study of the polaron
problem in the Holstein model,
Phys.
Rev. B 57, 6376
(1998).
[44]
T. Holstein, Studies of polaron motion: Part
I. The molecular-crystal model,
Ann.
Phys. 8, 325
(1959).
[45]
S. Ejima and H. Fehske,
Luttinger parameters and momentum distribution function for
the half-filled spinless fermion Holstein model: A DMRG approach,
Europhs.
Lett. 87, 27001
(2009).
[46]
J. Greitemann, S. Hesselmann,
S. Wessel, F. F. Assaad, and
M. Hohenadler, Finite-size effects in
Luther-Emery phases of Holstein and Hubbard models,
Phys.
Rev. B 92, 245132
(2015).
[47]
E. H. Lieb and F. Y. Wu,
Absence of Mott transition in an exact solution of the
short-range, one-band model in one dimension,
Phys.
Rev. Lett. 20, 1445
(1968).
[48]
F. Essler, H. Frahm,
F. Göhmann, A. Klümper, and
V. Korepin, The One-Dimensional
Hubbard Model (Cambridge University Press,
Cambridge, 2005).
[49]
X. Zotos, F. Naef, and
P. Prelovsek, Transport and conservation
laws,
Phys.
Rev. B 55, 11029
(1997).
[50]
F. Heidrich-Meisner, A. Honecker,
D. C. Cabra, and W. Brenig,
Thermal conductivity of anisotropic and frustrated
spin-$\frac{1}{2}$ chains,
Phys.
Rev. B 66, 140406
(2002).
[51]
R. Steinigeweg, J. Herbrych,
X. Zotos, and W. Brenig,
Heat conductivity of the Heisenberg spin-1/2 ladder: From
weak to strong breaking of integrability,
Phys.
Rev. Lett. 116, 017202
(2016).
[52]
C. Karrasch, D. M. Kennes, and
F. Heidrich-Meisner, Thermal Conductivity
of the One-Dimensional Fermi-Hubbard model,
Phys.
Rev. Lett. 117, 116401
(2016). |
\makeheadline
\summaryWe summarise our recent work on the faint galaxy contribution
to the cosmic X-ray background (XRB). At bright X-ray fluxes (in the
ROSAT pass band), broad line QSOs dominate the X-ray source
population, but at fainter fluxes there is evidence for a significant
contribution from emission-line galaxies. Here we present statistical
evidence that these galaxies can account for a large fraction
of the XRB. We also demonstrate that these galaxies have significantly
harder X-ray spectra than QSOs in the ROSAT band. Finally we present
preliminary findings from infra-red spectroscopy on the nature of this
X-ray emitting galaxy population. We conclude that a hybrid
explanation consisting of obscured/Type 2 AGN surrounded by starburst
activity can explain the properties of these galaxies and perhaps the
origin of the entire XRB. END
\keywActive galactic nuclei, X-ray background, deep surveys
END
\AAAclaEND
\kap
Introduction
Deep ROSAT surveys have revealed that broad-line QSOs account for at
least $30\%$ of the total $0.5-2$ keV XRB. Analyses of the number
count distribution and luminosity function of QSOs, however, suggests
that they are unlikely to form more than $\sim 50\%$ of the XRB at
these energies (Boyle et al 1994, Georgantopoulos et al 1996),
although we note that recent work by Schmidt et al (1997) and Hasinger
et al (1997) disputes this claim. Furthermore, it has been known for
many years that the X-ray spectra of QSOs are much steeper than the
spectrum of the XRB, leading to the so-called ‘spectral paradox’
(Gendreau et al 1995). A faint source population with a flatter X-ray
spectrum is required to account for the remainder of this background
radiation.
In recent years it has become clear that a population of X-ray
luminous narrow emission-line galaxies (NELGs) could provide a
possible explanation, perhaps dominating the X-ray source counts at
faint fluxes (Boyle et al 1995, McHardy et
al. 1997, Carballo et al 1995, Griffiths et al 1996). We refer
to these X-ray loud objects as ‘narrow-line X-ray galaxies’ (NLXGs)
to distinguish them from the field galaxy population.
The nature of the X-ray emitting mechanism in these galaxies will be
discussed further in Section 4, but we will will conclude
that many
of these objects contain low luminosity or obscured AGN (see
eg. Comastri et al 1995). In this sense the entire XRB could still be
due to ‘AGN’ rather than two distinct classes of X-ray source. The
most significant problem in assigning faint galaxies to X-ray sources
comes from the confusion associated with the relatively large X-ray
error box (see Hasinger et al 1997). To overcome this confusion, one
must adopt a statistical approach.
\kap
Probing the galaxy contribution by cross-correlation techniques
The clearest evidence that faint galaxies could be important
contributors to the XRB came from the study of Roche et al (1995), who
cross-correlated faint ($B<23$) galaxy catalogues with deep X-ray
observations, thus avoiding the problems associated with source
confusion. In Almaini et al (1997b) we performed an independent test
of the Roche et al results on new deep ($\sim$50ks) $\it ROSAT$
exposures and for the first time attempted to measure the evolution in
the X-ray emissivity of faint galaxies with redshift. The
cross-correlation of unidentified X-ray sources with faint galaxies
indicated that these could account for $20\pm 7\%$ of all X-ray
sources to a limiting flux of
$\sim 4\times 10^{-15}$erg$\,$s${}^{-1}$cm${}^{-2}$ in the $0.5-2.0$ keV
band. To probe deeper, cross-correlations were also performed with the
residual, unresolved XRB images. Significant signals were again
obtained on all 3 $\it ROSAT$ fields (see Figure 1), independently
confirming the results obtained by Roche et al (1995).
To constrain the evolution of X-ray emissivity with redshift, we used
the magnitudes of the thousands of catalogue galaxies as probes of
their redshift distribution, adapting the formalism of Treyer & Lahav
(1996). We found evidence for very strong evolution in the mean X-ray
luminosity of the form $\left\langle{L_{x}}\right\rangle\propto(1+z)^{3.22\pm 0.98}$. This represents the first evidence that the
X-ray emission from faint galaxies evolves as strongly as AGN. Similar
results were obtained by analysing small numbers of bright narrow
emission-line galaxies emerging from deep $\it ROSAT$ exposures
(Griffiths et al 1996, Boyle et al 1995). A simple extrapolation to
z=2 suggests that faint galaxies can account for $\sim 40\pm 10\%$
of the total XRB at 1keV.
\kap
The X-ray spectra of faint ROSAT sources
In Almaini et al (1996) we presented an analysis of the X-ray spectra
of all X-ray sources from our deep survey. The results confirmed the
trend found by Hasinger et al (1993) and Vikhlinin et al (1995) that
the average source spectra harden towards fainter fluxes. Hardness
ratios for the stacked spectra of all sources as a function of flux
are shown on Figure 2(a).
Using the optical identifications available for most of these sources,
it was found that ordinary, broad-line QSOs are not responsible for
this spectral hardening (Figure 2b). This implied that the change in
mean spectra was due to the emergence of another source
population. Our cross-correlation results suggest that $20\pm 7$% of
the remaining, unidentified sources are due to faint galaxies, but
confusion problems prevent us from determining exactly which X-ray
sources are responsible. We therefore selected a restricted sample of
the most likely galaxy candidates with brighter optical magnitudes
($B<21.5$) and lying within 20 arcsec of the X-ray source. The
hardness ratios for the 15 emission-line galaxies satisfying this
criteria are shown in Figure 2b. We expect approximately 3 of these
galaxies to be spurious identifications, but nevertheless it is clear
that their mean X-ray spectra are significantly harder than those of
QSOs (Almaini et al 1996, Romero-Colmenero et al 1996). Hence the
X-ray spectra provide further evidence that NLXGs may finally provide
the solution to the origin of the XRB. The key test is to examine
their mean spectral properties with a large sample at higher energies,
which will be possible with AXAF or XMM.
\kap
Understanding the nature of NLXGs
There is mounting evidence that a population of X-ray luminous
‘galaxies’ could finally provide a solution to the origin of the XRB.
The important question now is to understand the nature of this unusual
activity. Using standard optical emission line ratios, these objects
appear on the borderline between starburst and Seyfert 2
classification and in most cases the distinction is very ambiguous
(Boyle et al 1995, McHardy et al 1997). One possibility is that
these are a new type of evolved starburst galaxy in which the X-ray
emission comes from massive X-ray binaries (Griffiths and Padovani
1990). No local starburst galaxies have been found with such hard
X-ray spectra however. These galaxies are also significantly more
X-ray luminous than any known starburst galaxies, and their far
infra-red fluxes are generally too low for the X-ray emission to be
entirely due to starforming activity (Iwasawa et al 1997). The most
likely explanation is that the hard X-rays come from an obscured
active nucleus. Such models can readily reproduce the spectrum of the
X-ray background (Comastri et al 1995). Recent work by Hasinger et al
(1997; see also these proceedings) has suggested that many of
these galaxies would indeed be classified as AGN in higher signal to
noise optical spectra, albeit as low luminosity and/or type 2 AGN.
If these objects are obscured AGN, one might expect to detect broad
emission lines in the infra-red waveband (the dust extinction at $K$
is approximately 6 magnitudes lower than in the $V$ band). We have
recently obtained UK Infra-Red Telescope CGS4 spectroscopy for a small
number of NLXGs. Of the 5 objects studied so far, none show any
evidence for the expected broad Paschen emission lines. In most
cases, this implies an absorbing column of at least $5\times 10^{22}$
atom cm${}^{-2}$ (if an AGN is present). Such columns are inconsistent
with the large ROSAT luminosities below 1keV, unless there is some
contribution from an additional source of soft X-ray flux. In at
least one object, the detection of highly ionized Si[VI] clearly
indicates the presence of an AGN, despite the non-detection of broad
Paschen lines (Figure 3). As such, this is similar to the optical
narrow-line objects detected by Hasinger et al (1997). A plausible way
to account for these properties is to postulate a hybrid model,
consisting of an obscured AGN surrounded by starforming activity.
Such a model would certainly explain the ambiguous optical line ratios
seen in these objects. The hard ASCA NLXG detected by Iwasawa et al
(1997) shows clear evidence for this dual behaviour. As discussed in
Fabian et al (1997), obscuration by nuclear starburst activity may be
an inevitable consequence of triggering a low luminosity AGN, and
could perhaps account for the obscuration of the entire hard XRB.
\kap
Conclusions
There is now overwhelming evidence for the existence of a population
of X-ray luminous galaxies, which could explain the remainder of the
XRB. Our cross-correlation analysis suggests that these objects could
account for $40\pm 10\%$ of the XRB at 1keV. In addition, the
individually identified galaxies show
significantly harder X-ray spectra than QSOs, more consistent with
that of the XRB. Obscured and/or low luminosity AGN provide the most
natural explanation for this activity, although our infra-red
spectroscopy and the optical properties of these galaxies suggest
the presence of an additional soft X-ray component. A hybrid model
consisting of an obscured AGN surrounded by starburst activity
provides the most natural explanation for all observed properties.
\sect
References
\rf
Almaini O., Shanks T., Boyle B.J., Griffiths R.E., Roche N., Stewart G.C.
& Georgantopoulos I., 1996, MNRAS 282, 295
\rfAlmaini O. & Fabian A.C., 1997a, MNRAS 288, L19
\rfAlmaini O., Shanks T., Griffiths R.E., Boyle B.J., Roche N.,
Georgantopoulos I., & Stewart G.C. 1997b, MNRAS 291, 372
\rfBoyle B.J., Griffiths R.E., Shanks T., Stewart G.C.,
Georgantopoulos I., 1994, MNRAS, 271,639
\rfBoyle B.J., McMahon R.G., Wilkes B.J., & Elvis M.,
1995, MNRAS 276, 315
\rfCarballo R. et al 1995, MNRAS 277, 1312
\rfComastri A., Setti G., Zamorani G. & Hasinger G.,
1995, A&A, 296, 1
\rfFabian A.C., Barcons X., Almaini O., 1997, MNRAS submitted
\rfGendreau K.C. et al, 1995, Publ. Astron. Soc. Japan, 47, L5-L9
\rfGeorgantopoulos I., Stewart G.C., Shanks T., Griffiths R.E., & Boyle B.J.,
1996, MNRAS 280, 276
\rfGriffiths R.E. & Padovani P. 1990, ApJ 360, 483
\rfGriffiths R.E., Della Ceca R., Georgantopoulos I., Boyle B.J.,
Stewart G.C., & Shanks T., 1996, MNRAS 281, 71
\rfHasinger G., Burg R., Giacconi R., Hartner G., Schmidt M.,
Trümper J., Zamorani G., 1993, A&A, 275, 1
\rfHasinger G., Burg R., Giacconi R., Schmidt M., Trumper J. & Zamorani G.,
1997, A&A in press
\rfIwasawa K., Fabian A.C., Brandt W.N., Crawford C.S.,
Almaini O., 1997, MNRAS, 291, L17
\rfMcHardy I. et al, 1997, MNRAS, In press
\rfRoche N., Shanks T., Georgantopoulos I., Stewart G.C., Boyle B.J., &
Griffiths R.E., 1995, MNRAS 273, L15
\rfSchmidt M. et al, 1997, A&A in press
\rfRomero-Colmenero E. et al, 1996, MNRAS 282, 94
\rfShanks T., Georgantopoulos I., Stewart G.C., Pounds K.A., Boyle B.J.
& Griffiths R.E., 1991 Nat 353, 315
\rfTreyer M.A. & Lahav O. 1996 MNRAS 280, 469
\rfVikhlinin A., Forman W., Jones C. & Murray S.,
1995 ApJ 451, 564
Omar Almaini, Institute of Astronomy, Madingley Road, Cambridge, CB3 OHA, e-mail: [email protected]
END |
Soft Open Charm Production in Heavy-Ion Collisions.
V. Topor Pop
McGill, University, Montreal, Canada, H3A 2T8
J. Barrette
McGill, University, Montreal, Canada, H3A 2T8
M. Gyulassy
Physics Department,
Columbia University, New York, N.Y. 10027
(February 23, 2009)
Abstract
Effects of strong longitudinal color
electric fields (SCF) on the open charm production in nucleus-nucleus
(A + A) collisions at 200A GeV are investigated
within the framework of
the HIJING/BB̄ v2.0 model.
A three fold increase of the effective string tension
due to in medium effects in A + A collisions,
results in a sizeable ($\approx$ 60-70 %) enhancement
of the total charm production cross sections
($\sigma^{\rm NN}_{{\it c}\,\bar{{\it c}}}$).
The nuclear modification factors show a
suppression at moderate transverse momentum (${\rm p}_{\rm T}$)
consistent with RHIC data.
At Large Hadron Collider energies
the model predicts an increase of $\sigma^{\rm NN}_{{\it c}\,\bar{{\it c}}}$
by approximately an order of magnitude.
pacs: 25.75.Dw, -25.75.Cj, 25.75.+r, 24.85.+p
The phase transition from hadronic degrees of freedom to
partonic degrees of freedom in ultra-relativistic nuclear collisions
is a central focus of experiments at the Relativistic
Heavy Ion Collider (RHIC).
Heavy-flavor quarks are an ideal probe to study early dynamics
in these nuclear collisions.
Several theoretical studies predict
wang_prl92 ; shuryak_prl92 ; geiger_prc93
a substantial enhancement of open
charm production associated to plasma formation of the deconfined
parton matter relative to the case of a purely hadronic scenario
without plasma formation.
A recent analysis shows that the dynamics of heavy-quarks at
RHIC are dominated by partonic or “pre-hadronic” interactions
in the strongly coupled plasma (sQGP) stage and can neither
be modeled by “hadronic interactions” nor described appropriately
by color screening alone cassing_08 .
Therefore, these quarks are key observables in the study
of thermalization of the initially created hot nuclear matter
nuxu_07 .
A review of heavy-flavor production in heavy-ion collisions
has been recently published vogt08_1 .
Direct reconstructed $D^{0}$ mesons
via hadronic channel ($D^{0}\,\rightarrow K\pi$)
in d + Au star_05 , Cu + Cu star_08_1 and
Au + Au star_08_2 collisions have been measured.
Due to the difficulty to reconstruct D-mesons hadronic decay vertex,
both STAR and PHENIX have studied open charm indirectly
via semileptonic decay to non-photonic electrons (NPE) or muons
star_05 ; star_08_2 ; star_07 ; yzhang_08 ; phenix_05 ; phenix_06_1 ; phenix_06_2 ; phenix_07 . Theory predict that charm quarks are produced by initial
gluon fusion Lin95 and
their production rates are expected to be well described
by perturbative Quantum Cromodynamics (pQCD)
at Fixed Order plus Next to-leading Logarithms (FONLL) cacciari_05 .
Total charm cross sections reported by both experiments
are however only compatible with the upper limit
of the FONLL predictions.
In addition, the data indicate a suppression
as large as that of light quarks star_07 ,phenix_07
while, due to their large mass and to the dead cone effect
charm quarks are predicted to loose
less energy than light quarks by gluon radiation in the medium
kharzeev_01 .
Recent model calculations based on in-medium charm
resonances/diffusion or collisional dissociation
Rapp_06 ; Vitev_07 ; Moore_05 ,
radiative energy loss via few hard scatterings Magdalena_06
or radiative energy loss via multiple soft collisions
Armesto_06 , have been applied to describe the non-photonic
electrons (NPE) spectra. They all predict less suppression than
that observed in experiments. On the other hand,
a good description of the nuclear modification factor
(NMF), ${\rm R}_{\rm AA}^{\rm npe}({\rm p}_{\rm T})$, was obtained in
non-perturbative time-dependent heavy-quark diffusion in the
Quark-Gluon Plasma (QGP) rapp_prl08 .
In previous papers prc72_top05 ; prc75_top07 we have shown
that the dynamics of strangeness production
deviates considerably from calculations based on Schwinger-like
estimates for homogeneous and constant color fields schwinger
and point to the
contribution of fluctuations of transient strong
color fields (SCF).
These fields are similar with those which could appear in a
“glasma” mclerran_08 at initial stage of the collisions.
In a scenario with QGP phase
transitions the typical field strength of SCF at RHIC energies
was predicted to be about 5-12 GeV/fm csernai_npa02 .
Recently Schwinger mechanism has been revisited cohen_jul08 and
pair production in time-dependent
electric fields has been studied gies_jul08 .
It is concluded that particles with large momentum are likely to have
been created earlier than particle with small momentum and
for very short temporal widths
($\Delta\tau\approx 10{\rm t}_{\rm c}$, where the Compton time
${\rm t}_{\rm c}=1/{\rm m}_{\rm c}$) the Schwinger formula strongly
underestimates the reachable particle number density.
In this paper we extend our study in the framework of
HIJING/BB̄ v2.0 model prc75_top07
to open charm productions.
We explore dynamical effects associated with
long range coherent fields (i.e strong color fields, SCF),
including baryon junctions and loops prc72_top05 , with emphasis
on the novel open charm observables measured at RHIC
in p+p and heavy-ion collisions.
Using this model we analyze the enhancement of total charm production
at 200A GeV energy.
For a uniform chromoelectric flux tube with field (E)
the pair production rate Biro:1984cf ; Gyulassy:1986jq ; cohen_jul08
per unit volume for a heavy quark is given by:
$$\Gamma=\frac{\kappa^{2}}{4\pi^{3}}{\text{exp}}\left(-\frac{\pi\,m_{Q}^{2}}{%
\kappa}\right)$$
(1)
where for $Q=c$ or $b$ , $m_{\rm Q}=1.27,\;{\rm or}\;4.16$
GeV (with $\pm 1\%$
uncertainty Steinhauser:2008pm )
Note that $\kappa=|eE|_{eff}=\sqrt{C_{2}(A)/C_{2}(F)}\,\kappa_{0}$
is the effective string tension
in terms of the vacuum string tension $\kappa_{0}\approx 1$ GeV/fm
and $C_{2}(A)$, $C_{2}(F)$ are the second order Casimir operators
(see Ref. Gyulassy:1986jq ).
In a nuclear collisions, the local longitudinal
field strength increases with the square root of the
number of color exchanges proportional to
the number of binary collisions per unit area
($\kappa(x_{\perp},b)\propto\sqrt{T_{A}A(x_{\perp},b)}$),
where for a given impact parameter
$b$ and transverse coordinate $x_{\perp}$, $T_{A}A\propto A^{2/3}$
is the Glauber A+A
binary collision distribution. Therefore, the effective string tension
$\kappa\propto A^{1/3}$.
A measurable rate for spontaneous pair production requires
“strong chromo electric fields”, such
that $\kappa/m_{\rm Q}^{2}\,\,>$ 1 at least some of the time.
On the average, longitudinal electric field “string” models
predict for heavier flavor a very suppressed production rate per unit
volume $\gamma_{Q}$ via the well known Schwinger formula schwinger ,
since
$$\gamma_{Q\bar{Q}}=\frac{\Gamma_{Q\bar{Q}}}{\Gamma_{q\bar{q}}}={\text{exp}}%
\left(-\frac{\pi(m_{Q}^{2}-m_{q}^{2})}{\kappa_{0}}\right)\ll 1$$
(2)
for $Q=c$ and $q=u,d$.
For a color rope on the other hand,
if the average string tension value ($<\kappa>$)
increases from 1.0 GeV/fm to 3.0 GeV/fm,
the rate $\Gamma$ for charm pairs to tunnel through the longitudinal field
increases from $\approx 1.4\,\cdot 10^{-12}$ to
$\approx 3.5\,\cdot 10^{-4}$ fm${}^{-4}$,
and this can lead to a net “soft” tunneling production
comparable to the initial “hard” FONLL pQCD production.
The conventional hard pQCD mechanism, mainly
gluon fusion wang_prl92 , is calculated via the PYTHIA
subroutines in HIJING/BB̄ v2.0.
The advantage of HIJING over PYTHIA
is the ability to include novel SCF color rope effects
that arise from longitudinal fields amplified by the random walk
in color space of the high x valence partons in
A+A collisions.
This random walk could induce
a very broad fluctuation spectrum of the effective string tension.
Thus, if the average or mean $<\kappa>=n\,\kappa_{0}$, then the typical
fluctuation is of order $1/\sqrt{n}$ which is large because
$n\approx 6$ for Au nuclei.
A Poisson fluctuation of effective $\kappa$ about the mean, gives a
strong bias toward less probable but larger
than the average value $<\kappa>$. This is amplified for heavy
quarks.
Here we do not investigate in details such fluctuations, but we will
estimate the effects of a larger effective value $\kappa>$ 3 GeV/fm
on the enhancement of $\sigma^{\rm NN}_{{\it c}\,\bar{{\it c}}}$.
Both STAR and PHENIX experiments have measured charm production cross
sections in several collision systems.
Figure 1 shows the measured total charm production cross
sections at mid-rapidity,
$d\,\sigma^{\rm NN}_{{\it c}\,\bar{{\it c}}}/dy$ (left
panel) and in all phase space,
$\sigma^{\rm NN}_{{\it c}\,\bar{{\it c}}}$ (right panel).
The data from both experiments seems to indicate a scaling
with number of binary collisions (${\rm N}_{\rm bin}$),
as expected because of the high mass of charm pairs produced in initial
nucleon-nucleon collisions Lin95 .
However, there is still an unresolved discrepancy of the order of
a factor of two between STAR and PHENIX data.
The predictions of HIJING/BB̄ v2.0 model without SCF (open crosses)
and including SCF effects (filled squares) are shown in the figure.
For completeness the results with SCF but no gluon
shadowing effects (open triangles) are also included.
However, in this scenario multiplicities at
mid-rapidity are strongly overestimated topor_prc03 .
The main parameters used in the calculations are given in
Table II of reference prc75_top07 ,
and corresponds to
strengths of strong color (electric) field dependent on collision system
($\kappa$ = 1.5; 2.0; 3.0 GeV/fm for p + p, d + Au, and
A + A collisions respectively).
In our calculations we estimate the total open charm
production (c + $\bar{\it c}$) cross section considering
the 12 lightest D-mesons (${\rm D}^{0}$, ${\bar{\rm D}}^{0}$,
${\rm D}^{0*}$, ${\bar{\rm D}}^{0*}$,
${\rm D}^{+}$, ${\bar{\rm D}}^{+}$, ${\rm D}^{+*}$, ${\bar{\rm D}}^{+*}$,
${\rm D}_{\rm s}$, ${\bar{\rm D}}_{\rm s}$,
${\rm D}^{*}_{\rm s}$, ${\bar{\rm D}}^{*}_{\rm s}$), and the hyperons
$\Lambda_{\rm c}$ and $\bar{\Lambda_{\rm c}}$.
The contribution of higher mass charm hyperons is negligible.
For calculations which take into consideration SCF effects (filled
squares) we obtain an increase of $60-70\,\%$ in comparison
with a scenario without SCF effects (open crosses).
These results describe well the PHENIX data within statistical
and systematical errors and are close to the upper limit of
uncertainty band of the pQCD FNOLL predictions cacciari_05 .
Our calculations also show that the scaling with ${\rm N}_{\rm bin}$
is only approximately satisfied, the reason being
an interplay between the mass dependent SCF and
shadowing effects, which act in opposite directions.
In fact, we calculate that only $60\,\%$ of total open charm production
(c + $\bar{\it c}$) comes from partons embedded within the target
and projectile.
The study of open charm production in d + Au collisions
allow to separate “cold nuclear matter” effects.
The initial production
of c $\bar{\it c}$ pairs by gluon fusion might be suppressed
due to gluon shadowing.
We recall that shadowing is a depletion of the low-momentum parton
distribution in a nucleon embedded in a nucleus compared to a free
nucleon; this leads to a lowering in the (scaled)
c + $\bar{\it c}$ production relative to p + p
collisions.
The shadowing in the regular HIJING parameterization hij_94
implemented also in our model seems to be too strong wang_pl527_02 .
There is a considerable uncertainty
(up to a factor of 3) in the amount of shadowing
predicted at RHIC energies by the different models with HIJING
predicting the strongest effect cms_enteria_08 .
This could explain
why the results for scaled cross sections in d + Au collisions
are smaller than those obtained for p + p
collisions (see Fig. 1 left panel).
We study if we can find scenarios that would give larger enhancement
of total cross sections for open charm production, than
those reported in Fig. 1 (filled squares),
and that would be consistent with the STAR data.
The random walk in color space of heavy quark could induce a broad spectrum
of the effective string tension.
Therefore, we study the effects of
a further increase of mean value of the string tension from
3.0 GeV/fm to 5.0 GeV/fm on ${\it c}\,\bar{{\it c}}$ pair production.
This results in only a modest increase of scaled
cross sections, $\sigma^{\rm sNN}_{{\it c}\,\bar{{\it c}}}=\sigma^{\rm AA}_{{\it c}\,\bar{{\it c%
}}}/{\rm N}_{\rm bin}$,
by approximately $20\,\%$ for central collisions.
For values between 5 - 10 GeV/fm a saturation
sets in, as an effect of energy and momentum conservation constraints.
In our model the multiplicative factor that accounts for
next to leading order corrections eskola_03 in calculations of
hard or semihard parton
scattering processes via pQCD is set to K = 2.
Increasing this factor to K = 3.5, as suggested
in reference eskola_03 results in an increase of
$\sigma^{\rm sNN}_{{\it c}\,\bar{{\it c}}}$ by approximately $70\%$
in central Au + Au collisions, but also overpredicts by $40\,\%$
the total charged particle production at
mid-rapidity (N${}_{\rm ch}$ (y=0)).
Therefore, we conclude that the large charm cross sections obtained by
the STAR collaboration cannot be explained within our phenomenology.
The ${\it D}^{0}$-mesons spectra are sensitive to
the dynamics of produced charm particles.
In Fig. 2 we present the calculated
$D^{0}$-mesons spectra for systems where data are available star_05 ,star_08_1 ,star_08_2 .
In all cases, the calculated yield is much smaller than
the STAR data, consistent with the results shown in Fig. 1.
The calculated spectra show little shoulder at low ${\rm p}_{\rm T}$
indicating small radial flow of ${\it D}^{0}$-mesons consistent with
the results of STAR yzhang_08 .
Figure 3 shows our predictions for
the Nuclear Modification factor (NMF),
${\rm R}_{\rm AA}({\rm p}_{\rm T})$ for $D^{0}$ and $\pi^{0}$ mesons.
Data (filled symbols) are NMF for non photonic
electrons, ${\rm R}_{\rm AA}^{\rm npe}({\rm p}_{\rm T})$
star_07 ,phenix_07 .
The data for $\pi^{0}$ meson (open symbols) are from reference
phenix_03 .
Note, that non photonic electrons include also electrons
from bottom (b) production (B $\rightarrow$ lX)
and the yields of $D^{0}$ mesons could
be affected by the decay (B $\rightarrow$ D).
For central (0-10 %) Au + Au
collisions we calculate a scaled total cross section for
bottom production with (without) SCF of
$\sigma^{\rm sNN}_{{\it b}\,\bar{{\it b}}}$ = 17.8 $\mu$b
($\sigma^{\rm sNN}_{{\it b}\,\bar{{\it b}}}$ = 0.86 $\mu$b).
These values are
few orders of magnitude lower than $\sigma^{\rm sNN}_{{\it c}\,\bar{{\it c}}}$
and this contribution is estimated to be
negligible for ${\rm p}_{\rm T}\,<6.0$ GeV/c.
In our calculations for low
${\rm p}_{\rm T}$ ($0\,<\,{\rm p}_{\rm T}\,<2.5$ GeV/c),
non-perturbative production mechanism via SCF
results in a split between $D^{0}$ and $\pi^{0}$ mesons.
The charged and $\pi^{0}$ mesons are suppressed due to
conservation of energy prc72_top05 .
The yields of the ${\it D}^{0}$-mesons are enhanced due to an increase
of $c\,\bar{c}$ pair production rate (see Eq. 1).
In central ($0-10\,\%$) Au + Au collisions a suppression
at moderate $p_{\rm T}$ ($4\,<\,{\rm p}_{\rm T}\,<\,6$ GeV/c)
as large as that of light quarks is observed
in contrast to previous theoretical studies kharzeev_01 ,
Magdalena_06 ; Rapp_06 ; Armesto_06 ; wicks_npa07 .
Our model predicts a suppression consistent with the data.
We can interpret this results as experimental evidence for
“in-medium mass modification” of charm quark, due to
possible induced chiral symmetry restoration kharzeev_05 .
An in-medium mass modification has also been predicted
near the phase boundary (i.e. at lower energy) in tolos_06 .
In contrast statistical hadronization model andronic_08
predicts no medium effects at top RHIC energy.
We performed calculations at the much higher
Large Hadrons Collider (LHC) energy
using parameters from reference top_jpg08 ,
i.e $\kappa$ = 2.0; 5.0 GeV/fm for p + p
and central (0-10 %) Pb + Pb collisions respectively.
The predicted charm production cross section
is approximately an order of magnitude larger than at RHIC energy.
We obtain $\sigma^{\rm NN}_{{\it c}\,\bar{{\it c}}}$ = 6.4 mb
in p+p collisions and a (scaled) cross section
$\sigma^{\rm sNN}_{{\it c}\,\bar{{\it c}}}$ = 2.8 mb
for central Pb + Pb collisions (${\rm N}_{\rm bin}$ = 960
and N${}_{\rm ch}$ (y=0) = 2500).
This indicates a clear violation of scaling with ${\rm N}_{\rm bin}$
at the LHC.
These values increase by a factor of 2 to 3 if the effects
of shadowing are not included
( N${}_{\rm ch}$ (y=0) $\approx$ 5000 and
$\sigma^{\rm sNN}_{{\it c}\,\bar{{\it c}}}$ $\approx$ 8.4 mb).
In summary, we studied the influence of possible strong homogenous
constant color fields in open charm production
in heavy-ion collisions by varying the effective string tension that
control QQ̄ pair creation rates. This is equivalent with
assuming an in medium mass modification of charm quark.
We show that this approach is an important
dynamical mechanism that can explain the observed
D-mesons enhancement production observed by the PHENIX experiments.
Our model is based on the time-independent color field while in
reality the production of QQ̄ pairs is a far-from-equilibrium,
time-dependent phenomenon. Thus to achieve more quantitative
conclusions, such mechanisms gies_jul08
should be considered in future generation of Monte Carlo codes.
The large cross sections reported by the STAR collaboration
remain unexplained within our study.
Solving the discrepancy between the measurements is important, since
confirmation of the STAR results
may indicate the importance of other dynamical mechanisms such
as pre-equilibrium production
from secondary parton cascades wang_prl92 ,
or hot-glue scenario shuryak_prl92 .
Acknowledgments: We thank C. Gale and S. Jeon for useful discussions.
This work was supported by the Natural Sciences and Engineering
Research Council of Canada.
This work was supported also by the Division of Nuclear Science,
of the U. S. Department of Energy under Contract No. DE-AC03-76SF00098 and
DE-FG02-93ER-40764.
References
(1)
B. Muller and X. -N. Wang,
Phys. Rev. Lett. 68, 2437 (1992).
(2)
E. Shuryak, Phys. Rev. Lett. 68, 3270 (1992).
(3)
K. Geiger, Phys. Rev. D 48, 4129 (1993).
(4)
O. Linnyk, E. L. Bratkovskaya, W. Cassing,
Int. J. Mod. Phys. E 17, 1367 (2008).
(5)
X. Zhu, M. Bleicher, S. L. Huang, K. Schweda, H. Stocker,
N. Xu, P. Zhuang, Phys. Lett. B 647, 366 (2007).
(6)
A. D. Frawley, T. Ulrich, R. Vogt,
Phys. Rept. 462, 125 (2008).
(7)
STAR Collaboration, J. Adams et al., Phys.
Rev. Lett. 94, 062301 (2005).
(8)
STAR Collaboration, S. L. Baumgart
arXiv:0805.4228 [nucl-ex].
(9)
STAR Collaboration, B. I. Abelev et al.,
arXiv:0805.0364 [nucl-ex], submitted to . Phys. Rev. Lett.
(10)
STAR Collaboration, B. I. Abelev et al.,
Phys. Rev. Lett. 98, 192301 (2007).
(11)
Y. Zhang, J. Phys. G 35,104022 (2008).
(12)
PHENIX Collaboration, A. Adare et al.,
Phys. Rev. Lett. 94, 082301 (2005).
(13)
PHENIX Collaboration, S. S. Adler et al.,
Phys. Rev. Lett. 96, 032301 (2006).
(14)
PHENIX Collaboration, S. S. Adler et al.,
Phys. Rev. Lett. 97, 252002 (2006).
(15)
PHENIX Collaboration, S. S. Adler et al.,
Phys. Rev. Lett. 98, 172301 (2007).
(16)
Z. Lin and M. Gyulassy, Phys. Rev. C 51,
2177 (1995).
(17)
R. Vogt, Eur. Phys. J. ST 155, 213 (2008);
M. Cacciari, P. Nason, and R. Vogt,
Phys. Rev. Lett. 95, 122001 (2005).
(18)
Yu. L. Dokshitzer and D. E. Kharzeev,
Phys. Lett. B 519, 199 (2001).
(19)
H. Hess, V. Greco and R. Rapp,
Phys. Rev. C 73, 034913 (2006).
(20)
A. Adil and I. Vitev, Phys. Lett. B 649,
139 (2007).
(21)
G. D. Moore and D. Teaney,
Phys. Rev. C 71, 064904 (2005).
(22)
M. Djordjevic, M. Gyulassy, R. Vogt, S. Wicks,
Phys. Lett. B 632, 81 (2006).
(23)
N. Armesto, M. Cacciari, A. Dainese, C. A. Salgado,
U. A. Wiedemann, Phys. Lett. B 637, 362 (2006).
(24)
H. van Hess, M. Mannarelli, V. Greco, R. Rapp,
Phys. Rev. Lett. 100, 192301 (2008).
(25)
V. Topor Pop, M. Gyulassy, J. Barrette, C. Gale,
Phys. Rev. C 72, 054901 (2005).
(26)
V. Topor Pop, M. Gyulassy, J. Barrette, C. Gale,
S. Jeon and R. Bellwied,
Phys. Rev. C 75, 014904 (2007).
(27)
J. S. Schwinger, Phys. Rev. 82, 664 (1951).
(28)
L. McLerran, J. Phys. G 35, 104001 (2008).
(29)
V. K. Magas, L. P. Csernai, D. Strotman,
Nucl. Phys. A712, 167 (2002).
(30)
T. D. Cohen and D. A. McGady,
Phys. Rev. D 78, 036008 (2008).
(31)
F. Hebenstreit, R. Alkofer and H. Gies,
Phys. Rev. D 78, 061701 (2008).
(32)
T. S. Biro, H. B. Nielsen and J. Knoll,
Nucl. Phys. B245, 449 (1984).
(33)
M. Gyulassy and A. Iwazaki,
Phys. Lett. B. 165, 157 (1985).
(34)
M. Steinhauser,
arXiv:0809.1925 [hep-ph].
(35)
V. Topor Pop et al., Phys. Rev. C
68, 054902 (2003).
(36)
X. -N. Wang, M. Gyulassy, Comput. Phys. Commun.
83, 307 (1994).
(37)
Shi-yuan Li, Xin. -N. Wang ,
Phys. Lett. B 527, 85 (2002).
(38)
CMS Collaboration, D. d’Enteria et al.,
J. Phys. G 35, 104039 (2008).
(39)
K. J. Eskola and H. Honkanen, Nucl. Phys.
A713, 167 (2003).
(40)
PHENIX Collaboration, S. S. Adler et al.,
Phys. Rev. Lett. 91, 072301 (2003).
(41)
S. Wicks, W. Horowitz, M. Djordjevic, M. Gyulassy, Nucl. Phys. A783, 493 (2007).
(42)
D. Kharzeev and K. Tuchin, Nucl. Phys. A753,
316 (2005).
(43)
L. Tolos, J. Schaffner-Bielich, H. Stocker,
Phys. Lett. B 635, 85 (2006).
(44)
A. Andronic, P. Braun-Munzinger, K. Redlich,
J. Stachel, J. Phys. G 35, 104155 (2008).
(45)
V. Topor Pop, J. Barrette, C. Gale, S. Jeon
and M. Gyulassy, J. Phys. G 35, 054001 (2008). |
On the Geometry of Spaces of Oriented Geodesics
Dmitri V. Alekseevsky$\,$,
School of Mathematics and Maxwell Insitute for
Mathematical Sciences, The Kings Buildings, JCMB, University of
Edinburgh,
Mayfield Road, Edinburgh, EH9 3JZ,UK.
[email protected]
Brendan Guilfoyle$\,$,
Department of Computing and Mathematics,
IT Tralee, Clash, Tralee,
County Kerry, Ireland.
[email protected]
Wilhelm Klingenberg$\,$.
School of Mathematical Sciences, University of Durham, Durham DH1
3LE, UK.
[email protected]
Abstract:
Let $M$ be either a simply connected pseudo-Riemannian space of constant curvature or a rank one Riemannian symmetric space other than $\mathbb{O}H^{2}$, and consider the space
$L(M)$ of oriented geodesics of $M$. The space $L(M)$ is a smooth
homogeneous manifold and in this paper we describe all invariant
symplectic structures, (para)complex structures, pseudo-Riemannian metrics and
(para)Kähler structure on $L(M)$.
Keywords: space of geodesics, rank one symmetric spaces, homogeneous Riemannian manifolds, pseudo-Riemannian metrics, symplectic structures, complex structures, Kähler structures, flag manifolds,
Cayley projective plane.
MSC 2000: Primary 53A25, Secondary 53B35.
This
work was supported by Leverhulme Trust, EM/9/2005/0069.
Contents
1 Introduction
1.1 Background
1.2 Main results
1.3 Outline of paper
2 The Space of Oriented Geodesics on a Riemannian Manifold
2.1 Tangent and unit sphere bundle
2.2 Smooth structure in the space of geodesics $L(M)$
2.3 Canonical symplectic structure on $L(M)$
3 Pseudo-Riemannian Spaces of Constant Curvature
3.1 Spaces of zero curvature
3.2 Spaces of constant non-zero curvature
4 Rank One Symmetric Spaces of Non-Constant Curvature
4.1 Complex projective and hyperbolic spaces
4.2 Quaternionic projective and hyperbolic spaces
4.3 Cayley projective plane
1 Introduction
1.1 Background
The geometry of the set of straight lines of projective space $P^{3}$ and Euclidean space $E^{3}$ is a classical
subject of investigations of such 19th century geometers as Grassmann, Plücker, F. Klein and Study.
They studied natural correspondences between submanifolds (i.e. points, curves and surfaces) in $E^{3}$ and submanifolds of $L(E^{3})$, the
space of oriented lines of $E^{3}$.
For example, a point $p\in E^{3}$ defines a surface in $L(E^{3})$ which consists of the oriented lines through $p$, a curve $C\subset E^{3}$
defines three curves in $L(E^{3})$ associated with the Frenet frame
of $C$, and a surface $S\subset E^{3}$ defines a surface in $L(E^{3})$ by its oriented normal lines. Conversely, distinguished (from the point of view of
intrinsic geometry of $L(E^{3})$) submanifolds determine special families of oriented lines in $E^{3}$.
In [14], Study identified the space $L(E^{3})$ of oriented lines in $E^{3}$ with the 2-sphere (the Study sphere) over the dual numbers
and defined and studied a notion of distance between oriented lines.
For a nice modern exposition and generalization of these results with applications to computational geometry, computer graphics
and visualization, see [9].
A natural complex structure in the space $L(E^{3})$ of oriented lines has been considered by Hitchin, who used it for a description of monopoles [7].
In [5], two of the authors defined a neutral Kähler structure in the space $L(E^{3})$ and gave its geometric description.
Recently, it has been used in the solution of a long-standing conjecture of Carathéodory [6].
In general, however, the space $L(M)$ of oriented geodesics of a (complete) Riemannian manifold is not a smooth manifold and has very bad topology.
But $L(M)$ is a smooth manifold if $(M,g)$ is either an Hadamard manifold (i.e. complete simply connected Riemannian manifold of non-positive curvature)
or a manifold with closed geodesics of the same length (aufwiederseen manifolds).
Symplectic and Riemannian structures in the space of geodesics $L(M)$ of aufwiederseen manifolds are discussed by Besse
[1], see also Reznikov [10] [11].
The symplectic form on the space of geodesics of a Hadamard manifold is described in [2].
Geometric structures in the space of oriented geodesics of hyperbolic 3-space are studied in [3], while Salvai has addressed the existence
and uniqueness of pseudo-Riemannian metrics in the spaces $L(E^{n})$ and $L(H^{n})$ [12] [13]. Note that both spaces are homogeneous manifolds
of the corresponding isometry group. Salvai proved that $L(E^{n})$ admits a pseudo-Riemannian metric invariant under
a transitive subgroup $G$ of the isometry group $I(E^{n})=E(n)=SO(n)\cdot E^{n}$ only for $n=3,7$ and gave an explicit description of the
corresponding metric.
1.2 Main results
The aim of the paper is to describe the natural geometric structures
in the space $L(M)$ of oriented geodesics of $M$, where $M$ is either a simply connected pseudo-Riemannian
space of constant curvature, or a rank one Riemannian symmetric space other than $\mathbb{O}H^{2}$. In these cases the space of geodesics $L(M)$ is a smooth
homogeneous manifold and we use Lie groups and Lie algebras to describe all invariant
symplectic structures, (para)complex structures, pseudo-Riemannian metrics and
(para)Kähler structure on $L(M)$.
More specifically, let $S^{p,q}=\{x\in E^{p+1,q},x^{2}=1\}$, where $E^{p+1,q}$ is ${\mathbb{R}}^{p+q+1}$ endowed with the flat metric of signature
$(p+1,q)$. Here, for $p=0$ we
assume in addition that $x^{0}>0$ so that $S^{p,q}$ is connected. The induced metric on $S^{p,q}$
has signature $(p,q)$ and is of constant curvature $1$.
Let $L^{+}(S^{p,q})$ (respectively $L^{-}(S^{p,q})$)
be the set of spacelike (respectively, timelike) geodesics in $S^{p,q}$ and similarly for $L^{\pm}(E^{p+1,q})$. We prove:
Main Theorem 1
For the flat pseudo-Euclidean spaces $E^{p+1,q}$
i)
The space $L^{-}(E^{p+1,q})=E(p+1,q)/SO(p,q)\cdot{\mathbb{R}^{+}}$ is a symplectic symmetric space with an invariant
Grassmann structure defined by a decomposition $\mbox{\goth m}=W\otimes{\mathbb{R}}^{2}$. Moreover, if $n=p+1+q=3$,
it has an invariant Kähler structure $(g,J)$ of neutral signature $(2,2)$. In addition, $L^{+}(E^{p+1,q})=L^{-}(E^{q,p+1})$.
while for the non-flat constant curvature manifolds $S^{p,q}$
ii)
Suppose that $p+q>3$. Then there exists a unique (up to scaling)
invariant symplectic structure $\omega$ and a unique (up to a sign)
invariant complex structure $I^{+}=J$ on $L^{+}(S^{p,q})$
and a unique (up to sign) invariant para-complex structure $K=I^{-}$ on $L^{-}(S^{p,q})$.
There exists unique (up to scaling) invariant
pseudo-Riemannian metric $g^{\varepsilon}=\omega\circ I^{\varepsilon}$ on $L^{\varepsilon}(S^{p,q})$ which is Kähler of signature
$(2(p-1),2q)$ for
$\varepsilon=+$ and para-Kähler (of neutral signature) for $\varepsilon=-$.
iii)
Suppose that $p+q=3$. Then there are 2-linearly independent invariant (parallel and closed)
2-forms $\omega,\omega^{\prime}$ on $L^{\varepsilon}(S^{p,q})$ with values $\omega_{\mbox{\goth m}}=\omega_{H}\otimes g_{V},\,\omega^{\prime}_{\mbox{\goth
m%
}}=g_{H}\otimes\omega_{V}$.
Any invariant metric has the form $h=\lambda g+\mu g^{\prime}$ where $g^{\prime}$ is a neutral metric with
value $g^{\prime}_{\mbox{\goth m}}=\omega_{H}\otimes\omega_{V}$. Any metric $h$ is Kähler for $\varepsilon=1$ (respectively,
para-Kähler for $\varepsilon=-$) with respect to the complex (respectively, para-complex) structure
$I^{\varepsilon}=I^{\varepsilon}_{H}\otimes 1$ with the Kähler form $h\circ I^{\pm}=\lambda\omega+\mu\omega^{\prime}$.
Moreover, the endomorphism $I^{\prime}=1\otimes I_{V}$ of m defines an invariant parallel
$h$-skew-symmetric complex structure of $L^{\varepsilon}(S^{p,q})$ if $\varepsilon=1,(p-1,q)=(2,0)$ or $(0,2)$ or
$\varepsilon=-$ and $(p,q-1)=(2,0)$ or $(0,2)$ and skew-symmetric parallel para-complex structure
if $\varepsilon=+,(p-1,q)=(1,1)$ or $\varepsilon=-,(p,q-1)=(1,1)$. The Kähler or para-Kähler structure
$(h,I^{\prime})$ has the Kähler form $h\circ I^{\prime}=\lambda\omega+\mu\omega^{\prime}$.
Consider now the rank one Riemannian symmetric spaces of non-constant
curvature. That is, $M=G/K$ is one of the projective spaces
$${\mathbb{C}}P^{n}=SU_{n+1}/U_{n},\qquad{\mathbb{H}}P^{n}=Sp_{n+1}/Sp_{1}\cdot
Sp%
_{n},\qquad{\mathbb{O}}P^{2}=F_{4}/Spin_{9}$$
or one of the dual hyperbolic spaces
$${\mathbb{C}}H^{n}=SU_{1,n}/U_{n},\qquad{\mathbb{H}}H^{n}=Sp_{1,n}/Sp_{1}\cdot
Sp%
_{n},\qquad{\mathbb{O}}H^{2}=F^{non-comp}_{4}/Spin_{9}.$$
For the projective spaces we prove:
Main Theorem 2
i)
The space $L({\mathbb{C}}P^{n})=SU_{n+1}/T^{2}\cdot SU_{n-1}$ has a one–parameter
family $\omega^{t}=\omega_{1}+t\omega_{0}$ of invariant symplectic forms (up to scaling) and four invariant
almost complex structures, up to sign, two of them being integrable.
All (almost) complex structures $J$ are compatible with $\omega^{t}$ i.e. they define an (almost)
Kähler or pseudo-Kähler metric $g=\omega^{t}\circ J$.
ii)
The spaces $L({\mathbb{H}}P^{n})=Sp_{n+1}/T^{1}\cdot Sp_{1}\cdot Sp_{n-1}$ and $L({\mathbb{O}}P^{2})=F_{4}/T^{1}\cdot Spin_{7}$
has a unique (up to scaling) invariant symplectic forms $\omega$ and
unique (up to sign) invariant complex structure $J$.
The pair $(\omega,J)$ defines a unique (up to scaling) invariant Kähler metric
$g=\omega\circ J$.
iii)
None of the above spaces have any invariant almost para-complex structures.
iv)
The canonical
symplectic structure on $L({\mathbb{C}}P^{n})$ is identified with $\omega_{1}$.
while for the hyperbolic spaces (other than $\mathbb{O}H^{2}$) we show
Main Theorem 3
i)
The space $L({\mathbb{C}}H^{n})=SU_{1,n}/T^{2}\cdot SU_{n-1}$ has a one–parameter family of
invariant symplectic structures $\omega^{t}=\omega_{1}+t\omega_{0}$,
and two (up to sign) invariant almost para-complex structures $K^{\pm}$, one of them
being integrable; and both are consistent
with $\omega^{t}$
i.e. $(K^{\pm},\omega^{t})$ defines a para-Kähler metric $g=\omega\circ K^{\pm}$.
ii)
The space $L({\mathbb{H}}H^{n})=Sp_{1,n}/T^{1}\cdot Sp_{1}\cdot Sp_{n-1}$ admits a unique (up to scaling)
invariant symplectic form $\omega$ and two (up to sign) invariant almost para-complex
structures, one of which is integrable. Both are compatible with $\omega$.
iii)
None of the above spaces have any invariant almost complex structures.
iv)
The canonical
symplectic structure on the space geodesics $L({\mathbb{C}}H^{n})$ is $\omega_{1}$.
In Table 1 we summarize the results of Main Theorems 1, 2 and 3.
1.3 Outline of paper
This paper is organised as follows. In the following section we consider the set of oriented geodesics $L=L(M)$ of a general complete
Riemannian manifold $M$.
Though $L$ may not be a manifold, we can still define
differential geometric objects on $L$ in terms of $G_{1}$-invariant objects on $SM$, where $SM$ is the unit sphere bundle and $G_{1}=\{expt\Gamma\}$ is the
geodesic flow. For example, the algebra of smooth functions ${\mathcal{F}}(L)$ on $L$ is defined as the algebra
of $G_{1}$-invariant functions on $SM$ (that is first integrals of the geodesic flow).
We sketch this approach in section 2 and define the canonical symplectic form on $L$. In the case when
$L$ is a manifold this coincides with
the standard symplectic form defined in [1].
Section 3 considers the oriented geodesics of pseudo-Riemannian manifolds of constant curvature, taking the flat and non-flat cases separately.
Main Theorem 1 follows from Theorems 1 and 2 of sections 3.1 and 3.2, respectively.
In section 4 we turn to rank one Riemannian symmetric spaces with non-constant curvature, dealing separately with the complex projective
and hyperbolic spaces,
the quaternionic projective and hyperbolic spaces and the Cayley plane. In particular, Main Theorem 2
follows from Theorem 3 of section 4.1, Theorem 5 of section 4.2 and Theorem 7 of section 4.3, while
Main Theorem 3
follows from Theorem 4 of section 4.1 and Theorem 6 of section 4.2.
2 The Space of Oriented Geodesics on a Riemannian Manifold
2.1 Tangent and unit sphere bundle
Let $M$ be an $n$-dimensional manifold and $\pi:TM\to M$ be the tangent bundle of $M$. Local coordinates $(x^{i})$ on $M$ give rise
to local coordinates
$(x^{i},y^{i})$ on $TM$ via $v_{x}=y^{i}\partial/\partial x^{i}$. Then $T_{(x,y)}TM\ni(x,y,\dot{x},\dot{y})$.
The vertical subspace $V_{(x,y)}TM$ is given by $(x,y,0,\dot{y})$ and we denote by $N$
the canonical endomorphism
$$N:T_{(x,y)}TM\to V_{(x,y)}TM\to 0:\;\;(x,y,\dot{x},\dot{y})\to(x,y,0,\dot{x})%
\to 0,$$
with $N^{2}=0$. Note that
$$\mathrm{Ker}\;N=\mathrm{Im}\;N=T^{v}(TM)=\{(x,y,0,\dot{y})\},$$
is the vertical subbundle of $T(TM)$. The restriction of the projection $\pi_{*}:T(TM)\to TM$
onto the vertical subspace is an isomorphism
$$\pi_{*}:T^{v}_{(x,y)}(TM)\to T_{x}M,\,\,(x,y,\dot{x},\dot{y})\mapsto(x,\dot{x}).$$
Assume now that $(M,g)$ is
Riemannian and $SM$ is the unit sphere bundle:
$$SM=\{(x,y)\in TM,\,g_{ij}(x)y^{i}y^{j}=1\}.$$
The metric $g$ induces an isomorphism of the tangent bundle
onto cotangent bundle $T^{*}M=\{\alpha=(x,p),\,\alpha=p_{i}dx^{i}\}$ given by
$$(x^{i},y^{i})\to(x^{i},p_{i}=g_{ij}y^{j}).$$
The pull-back of the canonical
one-form $\alpha=p_{i}dx^{i}$ and symplectic form $\omega=d\alpha$ of $T^{*}M$ are given by
$$\alpha^{g}=g^{*}\alpha=g_{ij}y^{i}dx^{j},\,\,\omega^{g}=g\circ\omega=g\circ d%
\alpha=d(g_{ij}y^{i})\wedge dx^{j}.$$
The tangent bundle of the unit sphere bundle has the canonical decomposition
$T_{(x,y)}SM=V_{(x,y)}SM+H^{g}_{(x,y)}SM$
into vertical space and horizontal subspaces.
Lemma 1
The sphere bundle of the Riemannian manifold $(M,g)$
admits the canonical contact structure
$$\theta=\alpha^{g}|SM,\;\;\;\alpha^{g}=g^{*}\alpha=g_{ij}y^{i}dx^{j},$$
where the associated Reeb vector field is the geodesic vector field given by
$$\Gamma=y^{i}\partial/\partial x^{i}-\Gamma^{i}_{jk}y^{j}y^{k}\partial/\partial
y%
^{i}.$$
The horizontal lift of $\partial/\partial x^{i}=\partial_{i}$ into $SM$ is given by
$$\nabla_{i}=\partial_{i}-\Gamma^{k}_{ij}(x)y^{j}\partial/\partial y^{k},(x,y)%
\in SM.$$
In particular, $\Gamma$ preserves $\theta$ and $d\theta$.
Proof. The last claims are verified as follows. A vector field along $x(t)$ given by $X(t)=y^{i}(x(t))\partial_{i}$ is parallel if
$$\dot{X}^{i}+\Gamma^{i}_{jk}(x(t))\dot{x}^{i}X^{j}=0.$$
Then $(x(t),X(t))$ is the horizontal lift of the curve $x(t)$. The horizontal space is spanned by
the tangent vectors of such lifts, namely
$$H_{(x,y)}={\rm{span}}\{(\dot{x}^{i},-\Gamma^{i}_{jk}\dot{x}^{j}y^{k})\}.$$
The vector field $\Gamma=y^{i}\partial/\partial x^{i}-\Gamma^{i}_{jk}y^{j}y^{k}\partial/\partial
y%
^{i}$ on $TM$ is tangent to $SM$ since
$$\nabla_{k}(g_{ij}y^{i}y^{j})=g_{ij,k}-g_{im}\Gamma^{m}_{kl}-g_{jm}\Gamma^{m}_{%
ki}=g_{ij,k}-\Gamma_{i,kj}-\Gamma_{j,ki}=0.$$
Moreover, $\Gamma$ is a geodesic vector field since its integral curves satisfy the geodesic equation. Therefore we have
the decomposition into vertical and horizontal parts
$$T_{(x,y)}SM=V_{(x,y)}SM+H_{(x,y)}SM=\{v^{i}\partial/\partial y^{i}:g_{ij}y^{i}%
v^{j}=0\}+\{u^{k}\nabla_{k}\}.$$
We compute
$$d\theta=g_{ij,k}y^{j}dx^{i}\wedge dx^{k}+g_{ij}dy^{j}\wedge dx^{i}.$$
One easily checks that $\theta(\Gamma)=1$ and $d\theta(\Gamma)=0$.
Clearly the form $d(g^{*}\alpha)=g^{*}d\alpha=g^{*}\omega$ is non-degenerate on $TM$. Therefore its restriction
to $SM$ has one-dimensional kernel spanned by $\Gamma$ and so $\Gamma$ is the Reeb vector field of the
contact form $\theta$ and it preserves $\theta$ and $d\theta$.
$\square$
We associate with the vector field $X=X^{i}\partial_{i}$ the function $f_{X}$ on $SM$ given by
$f_{X}(x,y)=g_{ij}X^{i}y^{j}$.
Lemma 2
The covariant derivative $\nabla_{i}X$ corresponds to the Lie derivative of $f_{X}$ in direction of the vector field
$\nabla_{i}$ :
$$f_{\nabla_{i}X}=\nabla_{i}f_{X}.$$
Proof. Applying the vector field $\nabla_{i}$ to the function $f_{X}$ on $SM$ we get
$$\begin{array}[]{rl}\nabla_{i}f_{X}&=(g_{kl}X^{k})_{,i}y^{l}-g_{kl}X^{k}\Gamma^%
{l}_{ij}y^{j},\\
&=g_{kl}X^{k}_{,i}y^{l}+(g_{kj,i}-\Gamma_{k,ij})X^{k}y^{j},\\
&=g_{kl}X^{k}_{,i}y^{l}-\frac{1}{2}(g_{kj,i}+g_{ki,j}-g_{ij,k})X^{k}y^{j},\\
&=g_{ka}(X^{a}_{,i}+\Gamma^{a}_{,li}X^{l})y^{k},\\
&=f_{\nabla_{i}X}.\end{array}$$
$\square$
Denote the tangent bundle without the zero section by $T^{\prime}M$. Then $T^{\prime}M=SM\times{\mathbb{R}}^{+}$ with coordinate
$r$ on the second factor. We denote $E=r\partial/r=y^{i}\partial/\partial y^{i}$ the Euler vector field.
Note that the symplectic form $\omega^{g}$ is homogeneous of degree one: $E\cdot\omega^{g}=\omega^{g}$ and $E\cdot(\omega_{g})^{-1}=-\omega_{g}$.
Denote the homogeneous functions of degree $k$ by ${\cal{F}}_{k}(T^{\prime}M)$ and the extension of
the function $f\in{\cal F}(SM)$ to ${\cal F}_{k}$ by $f_{(k)}$. Furthermore, denote the Poisson
structure by $\{f,g\}=\omega_{g}^{-1}(df,dg)$. Since $\omega_{g}^{-1}$ has degree $-1$, we have
$$\{{\cal F}_{k},{\cal F}_{l}\}\subset{\cal F}_{k+l-1}.$$
We identify ${\cal F}(SM)$ with ${\cal{F}}_{1}(T^{\prime}M)$, $f\mapsto\tilde{f}=f_{1}=f\otimes{\mathbb{R}}$
and define the Legendrian bracket in the space ${\cal F}(SM)$ by
$$\{f,h\}:=\{\tilde{f},\tilde{h}\}_{SM}.$$
$\square$
2.2 Smooth structure in the space of geodesics $L(M)$
We first consider the topology of $L(M)$.
Let $(M,g)$ be a smooth complete Riemannian $n$-dimensional
manifold. By geodesic we mean an oriented maximally extended
geodesic on $M$ and such a geodesic $\gamma$ has a natural
parameterization by arc-length $\gamma=\gamma(s)$ defined up to a
shift $s\rightarrow s+C$. We denote by $\Gamma$ the canonical geodesic vector
field on the unit sphere bundle $SM$.
The maximal integral curves of $\Gamma$ through $(x,v)\in SM$ have the form
$(\gamma(s),\gamma^{\prime}(s))=(\exp_{x}(sv),d/ds\;\exp_{x}(sv))$, where $\gamma(s)=\exp_{x}(sv)$ is the maximal geodesic defined by $(x,v)\in SM$.
The set $L(M)$ is identified with the set of orbits of the flow
generated by $\Gamma$, i.e. the maximal integral curves of $\Gamma$. Denote by $\pi:SM\rightarrow L(M)=SM/\Gamma$ the natural projection,
equip $L(M)$ with the weakest topology such that $\pi$ is
continuous, and call the resulting topological space $L(M)$ the
space of geodesics of M. In general, $L(M)$ is not Hausdorff, but if $M$ is complete and compact, then
$L(M)$ is compact.
Lemma 3
The projection $\pi:SM\rightarrow L(M)$ is an open map if $(M,g)$ is
a complete Riemannian manifold.
Proof. We have to prove that if $U\subset SM$ is an open set then $\pi(U)$ is open, i.e.
$V:=\pi^{-1}(\pi(U))\subset SM$ is open.
It is clear since $V=\cup_{t\in{\mathbb{R}}}\varphi_{t}U$ where $\varphi_{t}=\exp(t\Gamma)$ is
the 1-parameter group of transformations which
is generated by $\Gamma$.
$\square$
To define a smooth structure in $L(M)$ we consider:
Definition 1
i) A function $f$ on an open subset $U\subset L(M)$ is called smooth if its pull back
$\pi^{*}f$ is a smooth function on $\pi^{-1}U$. We identify the algebra
${\cal F}(U)$ of smooth functions on $U$ with the algebra
${\cal F}(U)^{\Gamma}$ of $\Gamma$-invariant functions on $\pi^{-1}U$. In particular,
${\cal F}(L(M))={\cal F}(SM)^{\Gamma}$. We denote by ${\cal F}_{\gamma}(L(M))$
the germ of smooth functions at $\gamma$.
ii) A tangent vector $v$ of $L(M)$ at $\gamma$ is a derivation $v:{\cal F}_{\gamma}(L(M))\rightarrow{\mathbb{R}}$,
i.e. $v:f\mapsto v\cdot f$ such that $v\cdot(fh)=h(\gamma)v\cdot f+f(\gamma)v\cdot(h).$
They give rise to a vector space denoted by $T_{\gamma}L(M).$
iii) A vector field on $U\subset L$ is a derivation of the algebra ${\cal F}(U)$.
We identify the Lie algebra $\mbox{\mcal X}(U)$ of vector fields with
the Lie algebra $\mbox{\mcal X}(\pi^{-1}U)^{\Gamma}$ of $\Gamma$-invariant vector fields on $\partial^{-1}U\subset SM$.
iv) A k-form $\omega\in\Omega^{k}(U)$ is a ${\cal F}(U)$- polylinear skew-symmetric map
$$\omega:\mbox{\mcal X}(U)\times.....\times\mbox{\mcal X}(U)\rightarrow\mbox{%
\mcal X}(U).$$
The standard definition of the exterior differential holds on $\Omega^{k}.$
Note in particular that if the manifold $M$ has a
dense geodesic, then $C^{\infty}(L)={\mathbb{R}}$ and there are no
non-trivial vectors and vector fields on $L$. But if we restrict $M$ to a sufficiently small neighborhoods
$M^{\prime}$ of a point, the algebra $\mathcal{F}(M^{\prime})$ will be non-trivial and we get a non-trivial Lie algebra of vector fields.
2.3 Canonical symplectic structure on $L(M)$
We now give another, more general, definition of tangent vectors in $T_{\gamma}L(M)$ in terms of Jacobi fields along
$\gamma$.
Definition 2
A Jacobi tangent vector $v\in T_{\gamma}L(M)$ at a point $\gamma\in L(M)$
is a Jacobi vector field $Y$ along $\gamma$ which is normal to $\gamma$. The space $T^{J}_{\gamma}(L(M)):=Jac_{\gamma}^{\perp}$ of such
vector fields is called the Jacobi tangent vector space of $L(M)$ at $\gamma$.
Note that $\dim T^{J}_{\gamma}(L(M))=2n-1$. It is useful to give a relation between the tangent vector space $T_{\gamma}L(M)$
and the Jacobi tangent vector space.
For a tangent vector $Y\in T_{m}M$ we denote by $Y_{x,y}^{v}\in V_{x,y}TM$
the vertical lift and by $Y_{x,y}^{h}\in H_{x,y}TM$ the horizontal lift. We denote by
$\gamma^{S}=(\gamma,\dot{\gamma})$ the natural lift of a geodesic $\gamma$ to $SM$.
Lemma 4
The horizontal lift $Y^{h}$ of a Jacobi field $Y\in Jac^{\perp}_{\gamma}=T^{J}_{\gamma}L(M)$ is a horizontal $\Gamma$-invariant vector field along
$\gamma^{S}\subset SM$. It defines a tangent vector
$$\hat{Y}:\mathcal{F}(L(M))\to\mathbb{R},\,\,f\mapsto Y^{h}(f).$$
The map $Y\mapsto\hat{Y}$ is a homomorphism of $T^{J}_{\gamma}(L(M))$ into $T_{\gamma}(L(M))$.
Note that Proposition 1.90 in [1] is incorrect: not every Jacobi field is the transverse field
of a geodesic variation, a counterexample being the vertical constant vector field along the
minimal geodesic of $x_{1}+x_{2}^{2}-x_{3}^{2}=1$.
We now define the canonical symplectic 2-form $\omega$ on $L(M)$.
Lemma 5
The 2-form $\omega=d\theta$ on $SM$ is $\Gamma$-horizontal $(\iota_{\Gamma}\omega=0)$ and $\Gamma$-invariant.
Proof. We have $\theta(\Gamma)=1$ and $\iota_{\Gamma}\omega=0$, therefore
$\Gamma\cdot\omega=(d\iota_{\Gamma}+\iota_{\Gamma}d)\omega=0$ and $\Gamma\in ker\;\omega$.
Therefore $d\theta$ pushes down to a closed 2-form $\omega=\omega_{L}$
on $L(M)=SM/\Gamma.$
$\square$
Now we describe a Poisson structure on $L(M)$, namely
Lemma 6
${\cal F}(L(M))={\cal F}(SM)^{\Gamma}$ is a subalgebra of the Lie algebra $({\mathcal{F}}(SM),\{,\}).$
Proof. It is known that the Hamiltonian field preserves the symplectic form. If $\Gamma\cdot f=0$, then $\Gamma\cdot\tilde{f}=0$
since $\Gamma\cdot r=0$. We then have $\Gamma\cdot\{f,h\}=\Gamma\cdot\{\tilde{f},\tilde{h}\}=\Gamma\cdot\omega_{g}^{-%
1}(d\tilde{f},d\tilde{h})|SM=0.$
We conclude that $L(M)$ has a canonical Poisson structure.
$\square$
One can check that if the form $\omega_{L}$ is non-degenerate, then the Poisson structure on $L(M)$ is associated with the
symplectic structure $\omega_{L}$, i.e.
$\{f,h\}=\omega_{L}^{-1}(df,dh),f,h\in{\cal F}(L)={\mathcal{F}}_{1}(SM)^{\Gamma}$ if $\Gamma\cdot f=\Gamma\cdot h=0$,
where
$\omega_{L}^{-1}(df,dh):=\omega^{-1}(d\tilde{f},d\tilde{h})$.
3 Pseudo-Riemannian Spaces of Constant Curvature
3.1 Spaces of zero curvature
Let $E=E^{p+1,q}$ be a pseudo-Euclidean vector space of signature
$(p+1,q)$ with basis $(e_{0}^{+},...,e_{p}^{+},e_{1}^{-},....,e_{q}^{-})$. The
scalar product is given by
$$g(X,Y)=<x,y>=\sum_{0}^{p}x_{+}^{i}y_{+}^{i}-\sum_{1}^{q}x_{-}^{j}y_{-}j.$$
Definition 3
A vector $v\in E^{p+1,q}$ is said to be timelike (respectively spacelike, null) if
its norm is negative (positive, vanishes). A straight line is said to be timelike
(respectively spacelike, null) if its tangent vector has that type.
The space of oriented timelike (respectively spacelike) geodesics of $E^{p+1,q}$ is denoted by $L^{-}(E^{p+1,q})$ (respectively $L^{+}(E^{p+1,q})$).
Note that by changing the sign of the metric $L^{+}(E^{p+1,q})=L^{-}(E^{q,p+1})$.
We denote the unit pseudosphere by
$S^{p,q}=\{x\in E^{p+1,q},\,\,x^{2}=1\}$. Here, for $p=0$ we
assume in addition that $x^{0}>0$ so that $S^{p,q}$ is connected. The induced metric
has signature $(p,q)$ and constant curvature $1$.
We denote by
$SO(E)=SO^{0}(p+1,q)$ the connected pseudo-orthogonal group which
preserves the scalar product and by $SO(E)_{e}=SO^{0}(p,q)$ the connected
subgroup which preserves the vector $e=e_{0}^{+}$. The group $SO^{0}(p,q)$
acts transitively on $S^{p,q}$ and we can identify $S^{p,q}$ with the quotient
$SO^{0}(p+1,q)/SO^{0}(p,q)$. Note that the
tangent space $T_{e}S^{p,q}$ has the orthonormal basis $(e_{1}^{+},...,e_{p}^{+},e_{1}^{-},....,e_{q}^{-})$
Any non-null oriented straight line in $E$ can be canonically written
in the form
$$\ell_{e,v}(t)=\{v+te\},$$
where $e$ is the unit tangent vector (s.t. $e^{2}=\pm 1$) and $v$ is a vector orthogonal to $e$.
So we can identify the space $L^{-}(E)$ of timelike lines in $E$ with the tangent bundle
$TS^{p,q}$. The group $E(p+1,q)=SO^{0}(p+1,q)\cdot E^{p+1,q}$ of pseudo-Euclidean motions acts in the space
$L^{-}(E)=TS^{p,q}$ of timelike lines $(e,v):=l_{e,v}$ by
$$T_{a}(e,v)=(e,v+a_{e^{\perp}}),\,\,A(e,v)=(Ae,Av),\,\,a\in E,\,\,A\in SO^{0}(p%
+1,q),$$
(3.1)
where $a_{e^{\perp}}=a-<a,e>e$.
The group $E(p+1,q)$ also naturally acts on $S^{p+1,q}$ with the kernel of effectivity $E^{p+1,q}$.
Proposition 1
The isometry group $E(p+1,q)$ acts transitively on the space $L^{-}(E)$ of straight lines with stabilizer
$SO(n-1)\cdot R^{+}$ and this action commutes with the projection $\pi:L^{-}(E)\to S^{p.q},(e,v)\to e.$
The proof follows from equation (3.1).
Proposition 2
A necessary condition that a subgroup $G\subset E(p+1,q)$ acts transitively on the space
$L^{-}(E^{p+1,q})$ of timelike geodesics is that its linear part $LG=G/{G\cap E}$ acts transitively on $S^{p,q}$.
If the group $G$ contains the group $T_{E}$ of parallel translations this condition is also sufficient.
Proof. The first claim follows from the previous Proposition. Assume now that $G$ contains $E$ and $LG$ acts transitively on $S^{p,q}$. Let $\ell=\ell_{e,v},\,\ell^{\prime}=\ell_{e^{\prime},v^{\prime}}$ be two lines. Using a transformation from $G$ we transform
$\ell_{e^{\prime},v^{\prime}}$ into a line $\ell^{\prime\prime}=\ell_{e,v^{\prime\prime}}$ with the tangent vector $e$ and then using parallel translation we
transforms $\ell^{\prime\prime}$ into $\ell$.
Corollary 1
Let $E=E^{n}$ be the Euclidean vector space. Then any connected subgroup $G$ of the group $E(n)=SO(E)\cdot E$ of Euclidean motions has the form $G=LG\cdot E$ where $LG\subset SO(E)$ is a connected orthogonal group which acts transitively on $S^{n-1}$, that is $LG$ is one of the groups
$$SO(n),\,U(n/2),\,SU(n/2),\,Sp(1)\cdot Sp(n/4),\,Sp(n/4),$$
$$G_{2},(n=7),\,Spin(7),\,(n=8),\,Spin(9),\,(n=16).$$
We identify the space $L^{-}(E^{p+1,q})$ with the homogeneous space
$$G/H=E(p+1,q)/SO(p,q)\cdot R^{*}$$
where $H=SO(p,q)\cdot R^{+}$ is the stabilizer of the line $\ell_{0}=\ell_{e_{0},0}$. Let $E=\mathbb{R}e_{0}+W$ be the orthogonal decomposition. We write
the corresponding reductive decomposition of the homogeneous space $G/H$ as
$$\mbox{\goth g}=\mathfrak{e}(p+1,q)={\mbox{\goth h}}+\mbox{\goth m}=(\mbox{%
\goth so}(W))+\mathbb{R}e_{0})+(U+W),$$
where $\mbox{\goth so}(E)=\mbox{\goth so}(W)+U$ is the reductive decomposition of $\mbox{\goth so}(E)$. In matrix notation,
elements from the algebra of Euclidean isometries $\mathfrak{e}(p+1,q)\subset\mbox{\goth gl}(n+1)$ can be written as
$$(A,\lambda e_{0},u,w)=\left(\begin{array}[]{ccc}A&u&w\\
-u^{t}&0&\lambda\\
0&0&0\end{array}\right),\,\lambda\in\mathbb{R},\,u,w\in\mathbb{R}^{p+q}.$$
The adjoint action of $(A,\lambda e_{0})$ on $\mbox{\goth m}=U\oplus W=\{(u,w)\}$ is given by
$$\mathrm{ad\;}_{(A,\lambda)}(u,w)=(Au,Aw+\lambda u),$$
and the bracket of two elements of $\rm=U\oplus W$ is given by
$$[(u,w),(u^{\prime},w^{\prime})]=-(u\wedge u^{\prime},(u\cdot w^{\prime}-u^{%
\prime}\cdot w)e_{0})\in{\mbox{\goth h}},$$
where $u\wedge u^{\prime}$ is an element from $\mbox{\goth so}(W)$ and dot means the standard scalar product of vectors from $\mathbb{R}^{n-1}=U=W$.
Note that $L^{-}(E^{p+1,q})=G/H$ is a symmetric manifold since $\mbox{\goth g}={\mbox{\goth h}}+\mbox{\goth m}$ is a symmetric decomposition.
The isomorphism $\mathrm{ad\;}_{e_{0}}:U\to W$ allows one to identify $U$ with $W$ and the tangent space $\mbox{\goth m}=U+W$ with a tensor product
$\mbox{\goth m}=W\otimes{\mathbb{R}}^{2}$, where $U=W\otimes f_{1}$, $W=W\otimes f_{2}$ and $f_{1},f_{2}$ is the standard basis of $\mathbb{R}^{2}$.
The isotropy representation $\mathrm{ad\;}_{{\mbox{\goth h}}}$ preserves the Grassmann structures $\mbox{\goth m}=W\otimes\mathbb{R}^{2}$
and $\mbox{\goth so}(W)$ act on the first factor $W$ and $\lambda e_{0}$ acts on the second factor $\mathbb{R}^{2}$ by the matrix
$$\lambda\;\mathrm{ad\;}_{e_{0}}=\left(\begin{array}[]{cc}0&0\\
-\lambda&0\end{array}\right).$$
Moreover, the isotropy action $\mathrm{ad\;}_{{\mbox{\goth h}}}|_{\mbox{\goth m}}$ preserves the metric $g^{W}=g|_{W}$
of signature $(p,q)$ in $W$ and the symplectic structure $\omega_{0}=f_{1}\wedge f_{2}$ in $\mathbb{R}^{2}$.
The tensor product $\omega^{\mbox{\goth m}}=g^{W}\otimes\omega_{0}$
defines a non-degenerate $\mathrm{ad\;}_{{\mbox{\goth h}}}$-invariant 2-form in
m which is extended to an invariant symplectic form $\omega$ in $L^{-}(E^{p+1,q})$. The form $\omega$ is closed since it is invariant and
the manifold $L^{-}(E^{p+1,q})$ is a symmetric space.
In the case of dimension $n=3$ the action $\mathrm{ad\;}_{\mbox{\goth so}(W)}$ on $W$ also preserves a 2-form $\omega^{W}$ (which is the volume form of $W$).
Hence we get an invariant metric $g^{\mbox{\goth m}}=\omega^{W}\otimes\omega_{0}$ on m which extends to an invariant pseudo-Riemannian metric
$g$ of signature $(2,2)$ on $L^{-}(E^{p+1,q})$. The quotient $J=g^{-1}\circ\omega$ is an invariant (hence, integrable)
complex structure and the pair $(g,J)$ is an invariant Kähler structure.
Summarizing, we get (cf. [12]):
Theorem 1
The space $L^{-}(E^{p+1,q})=E(p+1,q)/SO(p,q)\cdot{\mathbb{R}^{+}}$ is a symplectic symmetric space with an invariant Grassmann structure defined by a decomposition $\mbox{\goth m}=W\otimes{\mathbb{R}}^{2}$. Moreover, if $n=p+1+q=3$,
it has an invariant Kähler structure $(g,J)$ of neutral signature $(2,2)$.
3.2 Spaces of constant non-zero curvature
We now describe the space of oriented timelike and spacelike geodesics of the pseudo-Riemannian space $S^{p,q}$
of constant curvature 1. Any such geodesic
through $e\in S^{p,q}$ in direction of a unit vector $e_{1}^{\pm}$ with $(e_{1}^{\pm})^{2}=\pm 1$ is given by
$$\gamma^{+}=\gamma_{e_{1}^{+}}^{+}=cos(s)e+sin(s)e_{1}^{+}\;,\;\;\;\gamma^{-}=%
\gamma_{e_{1}^{-}}^{-}=ch(s)e+sh(s)e_{1}^{-}.$$
The subgroup of the stability group $SO^{0}(p,q)$
preserving the spacelike geodesic $\gamma^{+}$ is $SO(p-1,q)$ and the timelike
geodesic $\gamma^{-}$ is $SO(p,q-1)$. The one-parameter subgroup $SO(2)$ generated by the
element $e\wedge e_{1}^{+}$ preserves $\gamma^{+}$ and the one-parameter subgroup $SO(1,1)$ generated by $e\wedge e_{1}^{-}$ preserves $\gamma^{-}$. Since the group $SO^{0}(p+1,q)$ acts
transitively on the space $L^{+}(S^{p,q})$ of spacelike geodesics
and on the space $L^{-}(S^{p,q})$ of timelike geodesics, we can
represent these spaces as
$$L^{+}(S^{p,q})=SO^{0}(p+1,q)/SO(2)\cdot SO(p-1,q),$$
$$L^{-}(S^{p,q})=SO^{0}(p+1,q)/SO^{0}(1,1)\cdot SO(p,q-1).$$
To get the reductive decomposition associated with
these spaces, fix the orthogonal decomposition
$$E^{p+1,q}={\mathbb{R}}e\oplus{\mathbb{R}}e^{\pm}_{1}\oplus V^{\pm},$$
where $V^{+}$ is the vector space of signature $(p-1,q)$ with basis
$(e_{2}^{+},...,e_{p}^{+},e_{1}^{-},...,e_{q}^{-})$ and $V^{-}$ is the vector space
of signature $(p,q-1)$ with basis
$(e_{1}^{+},...,e_{p}^{+},e_{2}^{-},...,e_{q}^{-})$.
Using the metric, we identify the Lie
algebra $\mbox{\goth so}(p+1,q)$ of $SO^{0}(p+1,q)$ with the space of
bivectors $\Lambda^{2}(E^{p+1,q})$. Then the reductive decomposition
associated with the unit sphere bundles
$$S^{\pm}(S^{p,q})=\{e_{1}^{\pm}\in T_{e}S^{\partial,q},\,<e_{1}^{\pm},e_{1}^{%
\pm}>=\pm 1\}$$
is given by
$$\mbox{\goth so}(p+1,q)=\Lambda^{2}(V^{\pm})\oplus(e\wedge V^{\pm}\oplus e_{1}^%
{\pm}\wedge V^{\pm})\oplus{\mathbb{R}}(e\wedge e_{1}^{\pm}).$$
The bivector $e\wedge e_{1}^{+}$ (resp.,$e\wedge e_{1}^{+}$ ) is invariant under the stability subgroup $SO(p-1,q)$
(resp., $SO(p,q-1)$ )and defines
an invariant vector field $\Gamma$ on $S^{\pm}(S^{p,q})$, which is the geodesic field. It
is the velocity field of the right action of the subgroup $SO^{\pm}(2)=SO(2),\,SO(1,1)$ of
$SO(p+1,q)$. The space of geodesics is the quotient
$$L^{+}(S^{p,q})=SO(p+1,q)/SO(p-1,q)\cdot SO^{\pm}(2),\,\,L^{-}(S^{p,q})=SO(p+1,%
q)/SO(p,q-1)\cdot SO(1,1).$$
The corresponding reductive decomposition may be
written as
$$\mbox{\goth so}(p+1,q)={\mbox{\goth h}}^{\pm}\oplus\mbox{\goth m}^{\pm}={%
\mathbb{R}}(e\wedge e_{1}^{\pm})\oplus\Lambda^{2}(V^{\pm})\oplus(e\wedge V^{%
\pm}\oplus e_{1}^{\pm}\wedge V^{\pm}).$$
We identify $\mbox{\goth m}^{\pm}$ with the tangent space
$T_{\gamma^{\pm}}(L^{\pm}S^{p,q})$. There is also a natural identification with the tensor product
$\mbox{\goth m}^{\pm}=H\otimes V^{\pm}$, where $H={\rm span}(e,e_{1}^{\pm})\simeq{\mathbb{R}}^{2}$
is the 2-dimensional oriented pseudo-Euclidean vector space.
Then the action of the isotropy subalgebra ${\mbox{\goth h}}^{\pm}$ takes the form:
$$ad_{e\wedge e_{1}^{\pm}}:e\otimes x\mapsto(e\wedge e_{1}^{\pm})e\otimes x=-e_{%
1}^{\pm}\otimes x,$$
$$ad_{e\wedge e_{1}^{\pm}}:e_{1}^{\pm}\otimes x\mapsto(e\wedge e_{1}^{\pm})e_{1}%
^{\pm})\otimes x=\pm e\otimes x,$$
$$ad_{a\wedge b}:e^{\prime}\otimes x\mapsto(e^{\prime}\otimes(a\wedge b)x=e^{%
\prime}\otimes<b,x>a-<a,x>b,$$
for all $a,b,x\in V^{\pm}$ and $e^{\prime}\in{\mathbb{R}}^{2}$.
Note that $L^{\pm}(S^{p,q})$ is identified with the Grassmanian
$Gr_{2}^{\pm}({\mathbb{R}}^{p+1,q})$ of two-planes of signature $(2,0)$ or $(1,1)$ and the decomposition
$$T_{\gamma^{\pm}}(L^{\pm}S^{p,q})=H\otimes V^{\pm},$$
defines an invariant Grassmann
structure in $L^{\pm}(S^{p,q})$.
Denote by $H^{\varepsilon}$ two-dimensional vector space with a scalar product $g_{H}=g_{H}^{\varepsilon}$ of signature
$(2,0)$ for $\varepsilon=+$ and $(1,1)$ for $\varepsilon=-$, and by $I_{H}=I_{H}^{\varepsilon}$ the $SO(H^{\varepsilon})$-invariant
endomorphism
of $H^{\varepsilon}$ with $I_{H}^{2}=-\varepsilon 1$ and by $\omega_{H}=\omega_{H^{\varepsilon}}=g_{\varepsilon}\circ I_{H}$ the invariant volume form.
Let $(V,g_{V})$ be a pseudo-Euclidean vector space of dimension $m$. If $m=2$, we denote by $I_{V}$ the
$SO(V)$-invariant endomorphism with $I_{V}^{2}=-1$ for signature $(2,0)$ or $(0,2)$ and with $I_{V}^{2}=1$
for the signature $(1,1)$. Denote also by $\omega_{V}=g_{V}\circ I_{V}$ the volume form of $V$.
Lemma 7
i)
Any $SO(H^{\varepsilon})\times SO(V)$-invariant endomorphism of the space $W^{\varepsilon}=H^{e}\otimes V$ has the form $A=1\otimes A+I_{H}^{\varepsilon}\otimes B$ where $A,B\in\mbox{\goth gl}(V)^{SO(V)}$ are invariant endomorphisms of $V$.
ii)
Any invariant endomorphism $I$ of $W^{\varepsilon}$ different from $1$ with $I^{2}=\pm 1$ is given
(up to a sigh) by
$I_{\varepsilon}=I_{H}\otimes 1$ if $m>2$ and by
$I_{\varepsilon},I^{\prime}_{\varepsilon}:=1\otimes I_{V},I^{\prime\prime}_{%
\varepsilon}=I_{H}\otimes I_{V}$ if $m=2.$
iii)
Any invariant metric on $W^{\varepsilon}$ is proportional to $g:=g_{H}\otimes g_{V}$ if $m>2$ and is a
linear combination of the metric $g$ and the neutral metric $g^{\prime}:=\omega_{H}\otimes\omega_{V}$ otherwise.
iv)
The space of invariant 2-forms has the basis $\omega=\omega_{H}\otimes g_{V}$ if $m>2$ and
$\omega,\omega^{\prime}:=g_{H}\otimes\omega_{V}$ if $m=2$.
v)
The endomorphisms $I,I^{\prime}$ are skew-symmetric with respect to any invariant metric $h$ on $W^{\varepsilon}$,
hence define a Hermitian or para-Hermitian structure, and the endomorphism $I^{\prime\prime}$ is symmetric
with respect to any invariant metric $h$.
Note that the tensor product of two complex or two para-complex structures is a para-complex structure
and the tensor product of a complex and a para-complex structures is a complex structure.
Proof. To prove part i), it is sufficient to write the endomorphism $A$ in block matrix form with respect to
the decomposition $W=h_{1}\otimes V+h_{2}\otimes V$, where $h_{1},h_{2}$ is an orthonormal basis of
$H^{+}$ or isotropic basis of $H^{-}$ and write the conditions that it is
$SO(H)\times SO(V)$-invariant. Since the only invariant endomorphism of $V$ is a scalar if $m>2$
and is a linear combination of $1,I_{V}$ if $m=2$, part ii) follows from part i). Parts iii) and iv) follow
from the fact that the space of symmetric bilinear forms
$$S^{2}(H\otimes V)=S^{2}(H)\otimes S^{2}(V)+\Lambda^{2}H\otimes\Lambda^{2}(V),$$
and that the space of 2-forms
$\Lambda^{2}(H\otimes V)=\Lambda^{2}(H)\otimes S^{2}(V)+S^{2}(H)\otimes\Lambda^%
{2}V$.
Now part v) follows from parts iii) and iv).
$\square$
Since the spaces
$$L^{+}(S^{p,q})=SO^{0}(p+1,q)/SO(2)\cdot SO^{0}(p-1,q)$$
$$L^{-}(S^{p,q})=SO^{0}(p+1,q)/SO^{0}(1,1)\cdot SO^{0}(p,q-1)$$
of spacelike and timelike geodesics
are symmetric spaces, any Hermitian pair $(h,I)$ which consists of invariant pseudo-Euclidean
metric on $\mbox{\goth m}^{\varepsilon}$ and skew-symmetric invariant complex or para-complex structure $I$ (such that
$I^{2}=-1$ or $I^{2}=1$) defines an invariant Kähler or para-Kähler structure on $L^{\pm}S^{p,q}$.
We get the following theorem.
Theorem 2
i)
Let $L^{+}(S^{p,q})$, (respectively $L^{-}(S^{p,q})$)
be the space of spacelike (respectively timelike) geodesics in $S^{p,q}$ and
$p+q>3$. Then there exists a unique (up to scaling)
invariant symplectic structure $\omega$ and a unique (up to sign)
invariant complex structure $I^{+}=J$ on $L^{+}(S^{p,q})$
(respectively para-complex structure $K=I^{-}$ on $L^{-}(S^{p,q})$).
There exists unique (up to a scaling ) invariant
pseudo-Riemannian metric $g^{\varepsilon}=\omega\circ I^{\varepsilon}$ on $L^{\varepsilon}(S^{p,q})$ which is Kähler of signature
$(2(p-1),2q)$ for
$\varepsilon=+$ and para-Kähler (of neutral signature) for $\varepsilon=-$.
ii)
If $p+q=3$, then there are two linearly independent invariant (parallel and closed)
2-forms $\omega,\omega^{\prime}$ on $L^{\varepsilon}(S^{p,q})$ with values $\omega_{\mbox{\goth m}}=\omega_{H}\otimes g_{V},\,\omega^{\prime}_{\mbox{\goth
m%
}}=g_{H}\otimes\omega_{V}$.
Any invariant metric has the form $h=\lambda g+\mu g^{\prime}$ where $g^{\prime}$ is a neutral metric with
value $g^{\prime}_{\mbox{\goth m}}=\omega_{H}\otimes\omega_{V}$. Any metric $h$ is Kähler for $\varepsilon=1$ (respectively,
para-Kähler for $\varepsilon=-$) with respect to the complex (respectively, para-complex) structure
$I^{\varepsilon}=I^{\varepsilon}_{H}\otimes 1$ with the Kähler form $h\circ I^{\pm}=\lambda\omega+\mu\omega^{\prime}$.
Moreover, the endomorphism $I^{\prime}=1\otimes I_{V}$ of m defines an invariant parallel
$h$-skew-symmetric complex structure of $L^{\varepsilon}(S^{p,q})$ if $\varepsilon=1,(p-1,q)=(2,0)$ or $(0,2)$ or
$\varepsilon=-$ and $(p,q-1)=(2,0)$ or $(0,2)$ and skew-symmetric parallel para-complex structure
if $\varepsilon=+,(p-1,q)=(1,1)$ or $\varepsilon=-,(p,q-1)=(1,1)$. The Kähler or para-Kähler structure
$(h,I^{\prime})$ has the Kähler form $h\circ I^{\prime}=\lambda\omega+\mu\omega^{\prime}$.
One can easily check that the form $\omega$ is the canonical symplectic form of the space of
geodesics $L^{\varepsilon}(S^{p,q})$.
4 Rank One Symmetric Spaces of Non-Constant Curvature
In this section we discuss the invariant geometric structures on
the space $L(M)$ of oriented geodesics of a rank one Riemannian symmetric space of non-constant
curvature $M=G/K$, that is for the projective spaces
$${\mathbb{C}}P^{n}=SU_{n+1}/U_{n},\qquad{\mathbb{H}}P^{n}=Sp_{n+1}/Sp_{1}\cdot
Sp%
_{n},\qquad{\mathbb{O}}P^{2}=F_{4}/Spin_{9},$$
and the dual hyperbolic spaces
$${\mathbb{C}}H^{n}=SU_{1,n}/U_{n},\qquad{\mathbb{H}}H^{n}=Sp_{1,n}/Sp_{1}\cdot
Sp%
_{n},\qquad{\mathbb{O}}H^{2}=F^{non-comp}_{4}/Spin_{9}.$$
In all of these cases, the space of geodesics is a homogeneous manifold
$L(M)=G/H$ where the stability subgroup $H$ is the same for the compact and dual non-compact
case, and is given by
$$H=T^{2}\cdot SU_{n-1},\,\,T^{1}\cdot Sp_{1}\cdot Sp_{n-1},T^{1}\cdot Spin_{7}.$$
Moreover, in the case of a classical Lie group $G$, the space $L(M)$ is the adjoint orbit
$L(M)=\mathrm{Ad\;}_{G}I^{\varepsilon}$ of the element
$I^{\varepsilon}=h_{1}^{\varepsilon}={\rm{diag}}(I^{\varepsilon}_{2},0,0,...,0),$ where
$$I_{2}^{\varepsilon}=\left(\begin{array}[]{cc}0&-\varepsilon\\
1&0\end{array}\right),$$
and $\varepsilon=1$ in the compact case and $\varepsilon=-1$ otherwise.
Main Theorems 2 and 3 describe all invariant structures
(symplectic structures, complex and para-complex, Kähler and para-Kähler
structures) on the space of geodesics $L(M)$.
We prove these in three stages: first for the complex and
quaternionic projective spaces, then for their hyperbolic
counterparts and finally for the Cayley projective
plane.
4.1 Complex projective and hyperbolic spaces
We now describe the space of real geodesics in complex projective
space
$$M^{1}={\mathbb{C}}P^{n}=SU_{n+1}/U_{n}$$
and in complex hyperbolic space
$$M^{-1}={\mathbb{C}}H^{n}=SU_{1,n}/U_{n}.$$
We set $\mathfrak{g}^{1}=\mathfrak{su}_{n+1}$ and
$\mathfrak{g}^{-1}=\mathfrak{su}_{1,n}$. We then choose the
associated reductive decompositions $\mathfrak{g}^{\epsilon}=j(\mathfrak{u}_{n})+\mathfrak{p}^{\epsilon}$, $\epsilon=\pm 1$, where
$j(\mbox{\goth u}_{n})=\{\left(\begin{array}[]{cc}-trA&0\\
0&A\end{array}\right)\left|\quad A\in\mbox{\goth u}_{n}\right.\}$, $\mbox{\goth p}_{\epsilon}=\{\left(\begin{array}[]{cc}0&-\epsilon X^{*}\\
X&0\end{array}\right)\left|\quad\quad X\in\mathbb{C}^{n}\right.\}$,
and where $X$ is a column vector and $X^{*}$ denotes the
Hermitian conjugate. We identify $\mathfrak{p}^{\epsilon}$ with the
tangent space $T_{o}M^{\epsilon},o=eU_{n}$.
We next describe the stability subalgebra ${\mbox{\goth h}}^{\epsilon}$ of the
geodesic $\gamma=\rm{exp}(th_{1}^{\epsilon})(o)$, where the
element $h_{1}^{\varepsilon}$ is represented by the matrix
$$h_{1}^{\varepsilon}=I^{\epsilon}=\left(\begin{array}[]{cc}I^{\epsilon}_{2}&0\\
0&0\end{array}\right),\qquad\qquad\rm{and}\qquad\qquad I^{\epsilon}_{2}=\left(%
\begin{array}[]{cc}0&-\epsilon\\
1&0\end{array}\right).$$
We have that
${\mbox{\goth h}}^{\varepsilon}=Z_{\mbox{\goth g}}(h^{\epsilon}_{1})={\mathbb{R%
}}h^{\epsilon}_{1}+Z_{\mbox{\goth u}_{n}}(h^{\epsilon}_{1})$,
where $Z_{\mbox{\goth u}_{n}}(h^{\epsilon}_{1})$ is the centraliser of $h^{\epsilon}_{1}$
in $\mbox{\goth u}_{n}$. We now describe the reductive decomposition
$\mbox{\goth g}^{\epsilon}={\mbox{\goth h}}^{\epsilon}+\mbox{\goth l}^{\epsilon}$.
We have
$${\mbox{\goth h}}^{\epsilon}=\{A=\left(\begin{array}[]{ccc}i\alpha&-\epsilon%
\beta&0\\
\beta&i\alpha&0\\
0&0&A_{n-1}\end{array}\right)|\quad\alpha,\beta\in{\mathbb{R}},\quad A_{n-1}%
\in\mbox{\goth u}_{n-1},trA_{n-1}+2i\alpha=0\},$$
and the complimentary subspace is
$$\mbox{\goth l}^{\varepsilon}=\{X=(x_{1},x_{2},X_{1},X_{2})=\left(\begin{array}%
[]{ccc}ix_{1}&\varepsilon ix_{2}&-\varepsilon X_{1}^{*}\\
ix_{2}&-ix_{1}&-X_{2}^{*}\\
X_{1}&X_{2}&0\end{array}\right)|\quad x_{1},x_{2}\in{\mathbb{R}},\quad X_{1},X%
_{2}\in{\mathbb{C}}^{n-1}\}.$$
We may write
$${\mbox{\goth h}}^{\varepsilon}={\mathbb{R}}h_{0}+{\mathbb{R}}h^{\varepsilon}_{%
1}+\mathfrak{su}_{n-1},$$
where
$$h_{0}=\left(\begin{array}[]{ccc}i&0&0\\
0&i&0\\
0&0&\frac{-2i}{n-1}{\mathrm{I}d}_{n-1}\end{array}\right),\quad h^{\varepsilon}%
_{1}=I^{\varepsilon}=\left(\begin{array}[]{ccc}0&-\varepsilon&0\\
1&0&0\\
0&0&0\end{array}\right).$$
Similarly,
$$\mbox{\goth l}^{\varepsilon}=V_{0}^{\varepsilon}+V_{+}^{\varepsilon}+V_{-}^{%
\varepsilon},$$
where
$$V_{0}^{\varepsilon}=\{(x_{1},x_{2},0,0)\}={\mathbb{R}}E_{1}+{\mathbb{R}}E_{2}^%
{\varepsilon},\;\;V_{\pm}^{\varepsilon}=\{X_{\pm}=(0,0,X,\pm X),X\in{\mathbb{C%
}}^{n-1}\},$$
$$E_{1}=(1,0,0,0)=\left(\begin{array}[]{ccc}i&0&0\\
0&-i&0\\
0&0&0\end{array}\right),\quad E_{2}^{\varepsilon}=(0,1,0,0)=\left(\begin{array%
}[]{ccc}0&\varepsilon i&0\\
i&0&0\\
0&0&0\end{array}\right).$$
We denote the canonical Hermitian form in the
space of vector columns ${\mathbb{C}}^{n-1}$ by $\eta(X,Y)=X^{*}Y$ . Then
$$g(X,Y)=Re\;\eta(X,Y)=\frac{1}{2}(X^{*}Y+Y^{*}X),$$
$$\rho(X,Y)=Im\;\eta(X,Y)=\frac{1}{2i}(X^{*}Y-Y^{*}X).$$
For any $X\in{\mathbb{C}}^{n-1}$ we set $X_{\pm}=(0,0,X,\pm X).$
Lemma 8
We have the following commutator relations:
$$[E_{1},E_{2}^{\varepsilon}]=2h_{1}^{\varepsilon},$$
(4.1)
$$[E_{1},(0,0,X_{1},X_{2})]=(0,0,-iX_{1},iX_{2}),$$
(4.2)
$$[E_{2}^{\varepsilon},(0,0,X_{1},X_{2})]=(0,0,-iX_{2},-\varepsilon iX_{1}).$$
(4.3)
The isotropy action of ${\mbox{\goth h}}^{\varepsilon}$ on $\mbox{\goth l}^{\varepsilon}$ is given by
$${\rm ad}_{h_{0}}(x_{1},x_{2},X_{1},X_{2})=(0,0,-iX_{1},-iX_{2}),$$
$${\rm ad}_{h_{1}^{\varepsilon}}(x_{1},x_{2},X_{1},X_{2})=(-2\varepsilon x_{2},2%
x_{1},-X_{2},\varepsilon X_{1}),$$
$${\rm ad}_{A_{n-1}}(x_{1},x_{2},X_{1},X_{2})=(0,0,A_{n-1}X_{1},A_{n-1}X_{2}).$$
Moreover
$$[X_{\pm}^{1},Y_{\pm}^{1}]=2\rho(X,Y)(-h_{0}\mp E_{2}^{1})\quad\rm{mod}\quad%
\mbox{\goth su}_{n-1},$$
(4.4)
$$[X_{\pm}^{-1},Y_{\pm}^{-1}]=2\rho(X,Y)(E_{1}\pm E_{2}^{-1})\quad\rm{mod}\quad%
\mbox{\goth su}_{n-1},$$
(4.5)
$$[X_{+}^{1},Y_{-}^{-1}]=-2\rho(X,Y)E_{1}+2g(X,Y)h_{1}\quad\rm{mod}\quad\mbox{%
\goth su}_{n-1},$$
(4.6)
$$[X_{+}^{-1},Y_{-}^{-1}]=2\rho(X,Y)h_{0}+2g(X,Y)h_{-1}\quad\rm{mod}\quad\mbox{%
\goth su}_{n-1},$$
(4.7)
for all
$X_{\pm}^{\varepsilon},Y_{\pm}^{\varepsilon}\in V_{\pm}^{\varepsilon}.$
Proposition 3
The $\mathfrak{h}$-module $\mathfrak{l^{\varepsilon}}$ has the
following decomposition into irreducible components.
a) For $\varepsilon=1$
$$\begin{array}[]{cccc}\mathfrak{l}_{+}^{1}=&\quad V_{0}^{1}&\quad+\quad V_{+}^{%
1}&\quad+\quad V_{-}^{1}\\
\mbox{ad}_{h_{0}}:&0&-i\mathrm{Id\;}&\quad-i\mathrm{Id\;}\\
\mbox{ad}_{h_{1}^{\varepsilon}}:&2J_{0}&i\mathrm{Id\;}&\quad-i\mathrm{Id\;}\\
A_{n-1}:&0&A_{n-1}&\quad A_{n-1}.\end{array}$$
b) For $\varepsilon=-1$
$$\begin{array}[]{ccccc}\mathfrak{l}_{-}^{-1}=&\quad\mathbb{R}E_{+}&\quad+\quad%
\mathbb{R}E_{-}&\quad+\quad V_{+}^{-1}\quad+&\quad V_{-}^{-1}\\
\mbox{ad}_{h_{0}}:&0&0&-i\mathrm{Id\;}&-i\mathrm{Id\;}\\
\mbox{ad}_{h_{1}^{\varepsilon}}:&2&-2&\mathrm{Id\;}&-\mathrm{Id\;}\\
A_{n-1}:&0&0&A_{n-1}&A_{n-1},\end{array}$$
where
$$E_{\pm}=(1,\pm 1,0,0)=E_{1}\pm E_{2}^{-1}.$$
With this notation in case b), the commutation relations read as follows :
$$[E_{\pm},V_{\pm}]=0,\;[E_{+},E_{-}]=-4h_{1}^{-1},$$
(4.8)
$$[E_{+},X_{-}]=-2iX_{+},\;[E_{-},X_{+}]=-2iX_{-}\;$$
(4.9)
$$[X_{\pm},Y_{\pm}]=2\rho(X,Y)E_{\pm},\;\;\;\;[X_{+},Y_{-}]=2\rho(X,Y)h_{0}+2g(X%
,Y)h_{1}\quad\rm{mod}\;\mathfrak{su}_{n-1},$$
(4.10)
$\mathrm{for}\;\;X_{\pm}\in V_{\pm}^{-1}$.
Recall that any invariant 2-form on $L(M)$ is generated by an
$\mathrm{ad\;}_{\mbox{\goth h}}$-invariant two form $\omega$ on the tangent space
$\mbox{\goth l}=T_{0}L(M)$. Any such form may be represented as $\omega=d(B\circ h)$,
where $h\in Z({\mbox{\goth h}})$ is a central element and $B$ is the Killing form.
Theorem 3
i)
The only invariant almost complex structures on
$L({\mathbb{C}}P^{n})$ are defined by
$$J_{\varepsilon_{0}\varepsilon_{1}\varepsilon_{2}}=\varepsilon_{0}\frac{1}{2}{%
\rm ad}_{h_{1}^{1}}|_{V^{0}}\oplus\varepsilon_{1}{\rm ad}_{h_{1}^{1}}|_{V^{+}}%
\oplus\varepsilon_{2}{\rm ad}_{h_{1}^{1}}|_{V^{-}}=\varepsilon_{0}J^{V_{0}}%
\oplus\varepsilon_{1}J^{V_{+}}\oplus\varepsilon_{2}J^{V_{-}},$$
where $\varepsilon_{k}=\pm 1$. The integrable ones among these are (up to a sign) $J_{\pm}=(\pm J^{V_{0}})\oplus J^{V_{-}}\oplus J^{V_{+}}.$
ii)
Any $SU(n+1)$-invariant closed 2-form on
$L({\mathbb{C}}P^{n})$ is a linear combination $\omega=\lambda_{0}\omega_{0}+\lambda_{1}\omega_{1}$ of the invariant differential forms
defined by
$\omega_{0}=d(B\circ h_{0})|\mbox{\goth l}^{1}$ and $\omega_{1}=d(B\circ h_{1}^{1})|\mbox{\goth l}^{1}$ where $B$ is the Killing form.
Moreover
$$ker\;\omega_{0}=V_{0},\quad\omega_{0}(X_{\pm},Y_{\pm})=2\rho(X,Y),\quad\omega_%
{0}(V_{+}^{1},V_{-}^{1})=0,$$
$$\omega_{1}(E_{1},E_{2})=-2,\quad\omega_{1}(V_{\pm},V_{\pm})=0,\quad\omega_{1}(%
X_{+},Y_{-})=-\omega_{1}(Y_{-},X_{+})=2g(X,Y).$$
Here $X_{\pm}=(0,0,X,\pm X)$, $Y_{\pm}=(0,0,Y,\pm Y)\in\mathfrak{l^{1}}$
and
$$g(X,Y)=\mathrm{Re}\;\eta(X,Y),\omega(X,Y)=\mathrm{Im}\;\eta(X,Y)$$
are the real and imaginary parts of the standard Hermitian form
$\eta(X,Y)=X^{*}Y$ on $\mathbb{C}^{n-1}.$
iii)
The canonical
symplectic structure on the space oriented geodesics is $\omega_{1}$.
Corollary 2
Up to scaling, any invariant symplectic form on
$L({\mathbb{C}}P^{n})$ may written as
$$\omega^{t}=\omega_{1}+t\omega_{0},\;\;t\in\mathbb{R}$$
It is compatible with any invariant complex structure $J_{\pm}$.
That is, the pair $(\omega^{t},\pm J_{\pm})$ is a Kähler structure on
$L$.
Proof of the Theorem: The description in part i) follows directly from the previous
Proposition. For integrability, we calculate the Niejenhuis bracket
$$N_{J}(X,Y)=[JX,JY]-J[X,JY]-J[JX,Y]-[X,Y],$$
for
$X,Y\in\mbox{\goth l}=\mathbb{R}E_{1}+\mathbb{R}E_{2}^{\varepsilon}+V_{+}+V_{-}$. For
example, for $X_{+}\in V_{+},Y_{-}\in V_{-}$, $J=J_{\varepsilon_{0}1\epsilon}$ we
calculate :
$$\begin{array}[]{cc}N_{J}(X_{+},Y_{-})&=[JX_{+},JY_{-}]-J[X_{+},JY_{-}]-J[JX_{+%
},Y_{-}]-[X_{+},Y_{-}]\\
&=[iX_{+},\varepsilon iY_{-}]-J[X_{+},\varepsilon iY_{-}]-J[iX_{+},Y_{-}]-[X_{%
+},Y_{-}]\\
&=(-\varepsilon-1)[X_{+},Y_{-}]+(-\varepsilon-1)iJ[X_{+},Y_{-}].\\
\end{array}$$
Hence $N_{J}(X_{+},Y_{-})=0$ iff $J=J_{\varepsilon_{0}11}$ i.e. $\varepsilon=1$.
Similarly we calculate $N_{J}(X_{+},E_{1})$ :
$$\begin{array}[]{cc}N_{J}(X_{+},E_{1})&=[JX_{+},JE_{1}]-J[X_{+},JE_{1}]-J[JX_{+%
},E_{1}]-[X_{+},E_{1}]\\
&=[iX_{+},E_{2}]-J[X_{+},E_{2}]-J[-iX_{+},E_{1}]-[X_{+},E_{1}]\\
&=(i-J)[X_{+},E_{2}]+(iJ-1)[X_{+},E_{1}]\\
&=(-i+J)iX_{+}(iJ-1)X_{+}\end{array}$$
This always vanishes. The proof of part ii) follows from the fact that the center
$Z({\mbox{\goth h}}^{1})$ has basis $h_{0},h_{1}^{1}.$ To verify part iii) we consider
$h_{1}^{1}\in\mbox{\goth p}^{1}=T_{0}{\mathbb{C}}P^{n}.$ The stability subalgebra of this element is
${\mbox{\goth h}}_{h_{1}^{1}}={\mathbb{R}}h_{0}+\mathfrak{su}_{n-1}\subset{%
\mbox{\goth h}}^{1}$.
The sphere bundle $S{\mathbb{C}}P^{n}$ is identified with $SU_{n+1}/T^{1}\cdot SU_{n}$ with the
reductive decomposition $\mathfrak{su}_{n+1}=({\mathbb{R}}h_{0}+\mathfrak{su}_{n-1})+({\mathbb{R}}h_{1}%
^{1}+\mbox{\goth l}^{1}).$
The geodesic vector field $\Gamma$ on $S{\mathbb{C}}P^{n}$ is the invariant vector field
generated by the element $h_{1}^{1}$ and the contact form $\theta$ is the invariant form
associated the one-form $\theta_{0}=cB\circ h_{1}^{1}$. This shows that the canonical
form coincides with $\omega^{1}$ (up to scaling).
$\square$
Theorem 4
i)
There is no invariant almost
complex structure on the space $L(\mathbb{C}H^{n})$. There exist two
(up to sign) almost para-complex structures $K^{\pm}$ with
$(\pm 1)$-eigenspace decompositions given by
$$K^{+}:\mbox{\goth l}^{-1}=\mbox{\goth l}_{+}+\mbox{\goth l}_{-}=(\mathbb{R}E_{%
+}+V_{+})+(\mathbb{R}E_{-}+V_{-}),$$
$$K^{-}:\mbox{\goth l}^{-1}=\mbox{\goth l}_{+}+\mbox{\goth l}_{-}=(\mathbb{R}E_{%
+}+V_{-})+(\mathbb{R}E_{-}+V_{+}).$$
Only $K^{+}$ is integrable.
ii)
Any closed invariant two-form is a linear combination
of the form defined by
$$\omega_{0}=d(B\circ h_{0})|\mbox{\goth l}^{-1},\;\;\omega_{1}=d(B\circ h_{1}^{%
-1})|\mbox{\goth l}^{-1},$$
where $B$ is the Killing form.
Moreover, we have
$$\rm{ker}\,\omega_{0}=V_{0},\,\,\omega_{0}(V_{\pm},V_{\pm})=0,\;\;\omega_{0}(X_%
{+},Y_{-})=2\rho(X,Y),$$
$$\omega_{1}(E_{+},E_{-})=4,\;\omega_{1}(X_{+},Y_{-})=-\omega_{1}(Y_{-},X_{+})=2%
g(X,Y),$$
where $X_{+}=(0,0,X,X),\;Y_{-}=(0,0,Y,-Y).$
Proof: An invariant almost complex structure on $L(\mathbb{C}H^{n})$
preserves the one-dimensional $ad_{h_{1}^{-1}}$ - eigenspaces
$\mathbb{R}E_{\pm}$, which is impossible. Since $\mbox{\goth l}_{\pm}=\mathbb{R}E_{\pm}+V_{\pm}$
are subalgebras, the endomorphisms $K$ of $\mbox{\goth l}^{-1}$ with $K|\mbox{\goth l}_{\pm}=\pm\mathrm{Id}$ define an invariant para-complex structure. It is unique up
to sign since $[V_{\pm},V_{\pm}]=\mathbb{R}E_{\pm}$. This proves i).
The first claim of ii) follows from the remark that
$h_{0},h_{1}^{-1}$ form a basis of the center $Z({\mbox{\goth h}}^{-1})$. The explicit
formulas for $\omega_{0},\omega_{1}$ follow from equations (4.8), (4.9) and (4.10).
$\square$
Corollary 3
Any invariant symplectic form on $L\mathbb{C}H^{n}$ may
be written as
$$\omega^{t}=\omega_{1}+t\omega_{0}.$$
They are compatible with the para-complex structures $K^{\pm}$, i.e.
$(\omega^{t},K^{\pm})$ is a para-Kähler structure and, in particular, $g^{t}=\omega^{t}\circ K^{\pm}$ is a para-Kähler metric.
4.2 Quaternionic projective and hyperbolic spaces
Consider now the spaces $M^{+}=\mathbb{H}P^{n}=Sp_{n+1}/Sp_{1}\cdot Sp_{n}$ and $M^{-}=\mathbb{H}H^{n}=Sp_{1,n}/Sp_{1}\cdot Sp_{n}$. The
reductive decomposition $\mbox{\goth g}^{\varepsilon}=(sp_{1}+sp_{n})+\mbox{\goth p}^{\varepsilon}$ associated
to the homogeneous space $M^{\varepsilon}$ may be written as
$$sp_{1}+sp_{n}=\{\left(\begin{array}[]{cc}a&0\\
0&A_{n}\end{array}\right)\left|\;a\in{\rm{Im}}{\mathbb{H}}=\mbox{\goth sp}_{1}%
,\;\;A_{n}\in\mbox{\goth sp}_{n}\right.\},\;\mbox{\goth p}^{\varepsilon}=\{%
\left(\begin{array}[]{cc}0&-\varepsilon X^{*}\\
X&0\end{array}\right)\left|\quad X\in{\mathbb{H}}^{n}\right\}.$$
We next describe the stability subalgebra ${\mbox{\goth h}}^{\epsilon}$ of the
geodesic $\gamma=\rm{exp}(th_{1}^{\epsilon})(o)$, $o=e(Sp_{1}\cdot Sp_{n})\in M^{\varepsilon}$, which is the orbit of the one-parameter group
$\rm{exp}(th_{1}^{\varepsilon})$, where $h_{1}^{\varepsilon}=\mathrm{diag}(I_{2}^{\varepsilon},0)$.
We have ${\mbox{\goth h}}=Z_{\mbox{\goth g}}(h_{1}^{\epsilon})={\mathbb{R}}h_{1}^{%
\epsilon}+Z_{(sp_{1}+sp_{n})}(h_{1}^{\epsilon})$,
where $Z_{(sp_{1}+sp_{n})}(h_{1}^{\epsilon})$ is the
centraliser of $h_{1}^{\epsilon}$.
The reductive decomposition associated to $L(M)^{\varepsilon}$ may be written
as
$${\mbox{\goth h}}^{\epsilon}=\{A=\left(\begin{array}[]{ccc}a&-\epsilon\alpha&0%
\\
\alpha&a&0\\
0&0&A_{n-1}\end{array}\right)=ah_{0}+\alpha h_{1}^{\varepsilon}+A_{n-1},a\in sp%
_{1},\;\alpha\in{\mathbb{R}},A_{n-1}\in sp_{n-1}\},$$
where
$h_{0}=\rm{diag}(1,1,0),h_{1}^{\varepsilon}=\rm{diag}(I^{\varepsilon}_{2},0)$,
and the complimentary subspace is
$$\mbox{\goth l}^{\varepsilon}=\{X=(x_{1},x_{2},X_{1},X_{2})=\left(\begin{array}%
[]{ccc}x_{1}&\varepsilon x_{2}&-\varepsilon X_{1}^{*}\\
x_{2}&-x_{1}&-X_{2}^{*}\\
X_{1}&X_{2}&0\end{array}\right)|\quad x_{1},x_{2}\in sp_{1},\quad X_{1},X_{2}%
\in{\mathbb{H}}^{n-1}\}.$$
We set
$$E_{1}=\rm{diag}(1,-1,0)\;\;,\;\;E_{2}^{\varepsilon}=h_{1}^{\varepsilon}.$$
Then
$$\mbox{\goth l}^{\varepsilon}=\mathfrak{sp}_{1}E_{1}+\mathfrak{sp}_{1}E_{2}^{%
\varepsilon}+\{(0,0,X_{1},X_{2}),X_{i}\in{\mathbb{H}}^{n-1}\}.$$
The isotropy action of ${\mbox{\goth h}}^{\varepsilon}$ on $\mbox{\goth l}^{\varepsilon}$ is given by
$${\rm ad}_{ah_{0}}(x_{1},x_{2},X_{1},X_{2})=([a,x_{1}],[a,x_{2}],-X_{1}a,-X_{2}%
a),$$
$${\rm ad}_{h_{1}^{\varepsilon}}(x_{1},x_{2},X_{1},X_{2})=(-2\varepsilon x_{2},2%
x_{1},-X_{2},\varepsilon X_{1}),$$
$${\rm ad}_{A_{n-1}}(x_{1},x_{2},X_{1},X_{2})=(0,0,A_{n-1}X_{1},A_{n-1}X_{2}).$$
As in the complex case, we introduce the canonical Hermitian form $\eta(X,Y)=X^{*}Y$ on
${\mathbb{H}}^{n-1}$. We have the following commutator relations:
$$[x_{1}E_{1},x_{2}E_{2}^{\varepsilon}]=-(x_{1}x_{2}+x_{2}x_{1})h_{1}^{%
\varepsilon}=-2Re(x_{1}x_{2})h_{1}^{\varepsilon},$$
$$[x_{1}E_{1},y_{1}E_{1}]=[x_{1},y_{1}]E_{1},\;[x_{2}E_{2}^{\varepsilon},y_{2}E_%
{2}^{\varepsilon}]=\varepsilon[x_{2},y_{2}]h_{0},$$
$\begin{array}[]{cc}&[(0,0,X_{1},X_{2}),(0,0,Y_{1},Y_{2})]=-(g(X_{2},Y_{1})+g(Y%
_{2},X_{1}))h_{1}^{\varepsilon}\\
&-\big{(}\varepsilon\rho(X_{1},Y_{1})+\rho(X_{2},Y_{2})\big{)}h_{0}+\big{(}-%
\varepsilon\rho(X_{1},Y_{1})+\rho(X_{2},Y_{2})\big{)}E_{1}\;\rm{(mod}\;\;sp_{n%
-1}),\end{array}$
$$[x_{1}E_{1},(0,0,X_{1},X_{2})]=(0,0,-X_{1}x_{1},X_{2}x_{1}),$$
$$[x_{2}E_{2},(0,0,X_{1},X_{2})]=(0,0,-X_{2}x_{2},-\varepsilon X_{1}x_{2}).$$
For $\varepsilon=1$ we have the following decomposition of the ${\mbox{\goth h}}^{1}$-module
$\mbox{\goth l}^{1}$ into two irreducible submodules: $\mbox{\goth l}^{1}=V_{0}+V_{1},V_{0}=Im\mathbb{H}+Im\mathbb{H}=\mathbb{R}^{6},%
V_{1}=\mathbb{H}^{n-1}+\mathbb{H}^{n-1}$. Define an ${\mbox{\goth h}}^{1}$-invariant complex structure
$J^{V_{0}},J^{V_{1}}$ by
$$J^{V_{0}}=1/2ad_{h_{1}}|V_{0}:(x_{1},x_{2})\to(-x_{2},x_{1})\quad J^{V_{1}}=ad%
_{h_{1}}|V_{1}:(X_{1},X_{2})\to(-X_{2},X_{1}).$$
The commutator relations imply the following Theorem in a similar
way as in the complex case.
Theorem 5
i)
There exist (modulo sign) two invariant almost complex structures on
$L(\mathbb{H}P^{n})$ defined by $J_{\pm}=J^{V_{0}}\oplus\pm J^{V_{1}}$.
Only one of them, namely $J_{+}$, is integrable.
ii)
Up to scaling, there exists a unique invariant symplectic form $\omega$
on $L(\mathbb{H}P^{n})$ defined by $\omega=d(h_{1}^{\varepsilon})^{*}$. More precisely
we have
$$\omega((x_{1},x_{2},X_{1},X_{2}),(y_{1},y_{2},Y_{1},Y_{2}))=$$
$$+2Re(y_{1}x_{2})+2Re(y_{1}x_{2})+Reh(X_{1},Y_{2})-Reh(Y_{1},X_{2}).$$
iii)
The symplectic structure $\omega$ consistent with the complex structure
$J_{+}$ and gives rise to a Kähler structure.
For $\varepsilon=-1$ we have the decomposition into irreducible $(\mathrm{ad\;}_{{\mbox{\goth h}}^{-1}})$ -submodules
$$\mbox{\goth l}^{-1}=V_{2}+V_{-2}+V_{1}+V_{-1}$$
such that $\mathrm{ad\;}_{{\mbox{\goth h}}^{-1}}|V_{k}=k\mathrm{Id}$,
$V_{2}\simeq V_{-2}\simeq\mathrm{Im}{\mathbb{H}}\simeq\mathfrak{sp}_{1}$ and
$V_{1}\simeq V_{-1}\simeq\mathrm{Im}{\mathbb{H}}^{n-1}$ with the standard action of
${\mbox{\goth h}}^{-1}=\mathfrak{sp}_{1}+\mathfrak{sp}_{n-1}$.
We finally define two $\mathrm{ad}_{{\mbox{\goth h}}}$- invariant para-complex structures
$K_{\pm}$ on ${\mbox{\goth h}}_{-1}$ by
$$K_{+}|_{V_{1}+V_{2}}=1,\;\;K_{+}|_{V_{-1}+V_{-2}}=-1,\;\;K_{-}|_{V_{1}}=1,\;$$
$$K_{-}|_{V_{2}}=-1,\;\;K_{-}|_{V_{2}}=-1,\;\;K_{-}|_{V_{-1}}=-1,\;\;K_{-}|_{V_{%
-2}}=1.$$
We then have the following
Theorem 6
On $L(\mathbb{H}H^{n})$ there exist no invariant almost complex
structures and two (up to sign) unique invariant almost para-complex
structures $K_{\pm}$, with $K_{+}$ being integrable and $K_{-}$ being non-integrable.
4.3 Cayley projective plane
Let $M=\mathbb{O}P^{2}=F_{4}/Spin_{9}$ be the octonian projective
plane and $\mbox{\goth g}=\mbox{\goth f}_{4}=spin_{9}+\mbox{\goth p}$ the associated reductive
decomposition.
The isotropy group $Spin_{9}$ acts on the $16$-dimensional tangent space
$\mbox{\goth p}=T_{o}M$ by the spinor
representation with $15$-dimensional spheres as orbits. Let $\gamma=\rm{exp(t\,h_{1}))o},h_{1}\in\mbox{\goth p}$ be the geodesic through the point
$o=e(Spin_{16})\in M$.
The stability subgroup of $\gamma$ is $H=SO_{2}\cdot Spin_{7}$ and the stability subalgebra
${\mbox{\goth h}}=\mathbb{R}h_{1}+Z_{\mathfrak{spin_{9}}}(h_{1})=\mathbb{R}h_{1%
}+{\mathfrak{spin}}_{7}$.
We
identify the space of
geodesics in $M=\mathbb{O}P^{2}$ with $L(M)=F_{4}/SO_{2}\cdot Spin_{7}$.
Following [4] we choose the root system
$R=\{\pm\varepsilon_{i},\pm\varepsilon_{i}\pm\varepsilon_{j},1/2(\pm\varepsilon%
_{1}\pm\varepsilon_{2}\pm\varepsilon_{3}\pm\varepsilon_{4})\}$
of the complex Lie algebra $\mbox{\goth f}_{4}$ with
respect to a Cartan subalgebra $\mathfrak{a}$ and a system of
simple roots as follows
$$\alpha_{1}=1/2(\varepsilon_{1}-\varepsilon_{2}-\varepsilon_{3}-\varepsilon_{4}%
),\,\alpha_{2}=\varepsilon_{4},\,\alpha_{3}=\varepsilon_{3}-\varepsilon_{4},\,%
\alpha_{4}=\varepsilon_{4}-\varepsilon_{3}.$$
Here, $\varepsilon_{i},\;\;i=1,...,4$ is an orthonormal basis of the real space
$\mbox{\goth a}_{{\mathbb{R}}}=B^{-1}\circ\mathrm{span}_{\mathbb{R}}R$.
We may assume that $\frac{1}{2}d=-ih_{1}\in\mathfrak{a}^{{\mathbb{R}}}$ is the vector dual to the fundamental weight $\pi_{1}=\varepsilon_{1}$. Then the adjoint operator $ad_{d}$ defines a gradation
$$\mathfrak{f}_{4}=\mathfrak{g}_{-2}+\mathfrak{g}_{-1}+\mathfrak{g}_{0}+%
\mathfrak{g}_{1}+\mathfrak{g}_{2},$$
where $\mathfrak{g}_{0}=Z_{\mathfrak{f}_{4}}(h_{1})=\mathbb{C}h_{1}+\mathfrak{spin}_{%
7}^{\mathbb{C}},$ and
$\mathfrak{spin}_{7}^{\mathbb{C}}$ has the root system given by
$$\{\pm\varepsilon_{i},\pm\varepsilon_{i}\pm\varepsilon_{j},i,j=2,3,4\}.$$
The space $\mathfrak{g}_{\pm 1}$
is spanned by the root vectors with roots $1/2(\varepsilon_{1}\pm\varepsilon_{2}\pm\varepsilon_{3}\pm\varepsilon_{4})$ and
$\mathfrak{g}_{\pm 2}$ is spanned by the root vectors
with roots $\pm\varepsilon_{1},\pm(\varepsilon_{1}\pm\varepsilon_{i}),\;i=2,3,4.$
Let $\tau$ be the standard
compact involution of $\mathfrak{f}_{4}$ such that
$\mathfrak{f}_{4}^{\tau}$ is the compact real form of $\mathfrak{f}_{4}$.
Then the reductive decomposition associated with the space of
geodesics can be
written as
$\mathfrak{f}_{4}^{\tau}={\mbox{\goth h}}+\mbox{\goth l}=\mathfrak{g}_{0}^{\tau%
}+\big{(}(\mathfrak{g}_{-1}+\mathfrak{g}_{1})^{\tau}+(\mathfrak{g}_{-2}+%
\mathfrak{g}_{2})^{\tau}$ ).
The decomposition
$$\mbox{\goth l}^{\mathbb{C}}=\mbox{\goth l}^{10}+\mbox{\goth l}^{01}=(\mathfrak%
{g}_{1}+\mathfrak{g}_{2})+(\mathfrak{g}_{-1}+\mathfrak{g}_{-2})$$
defines a unique (up to sign) invariant complex structure $J$ on the space
of geodesics defined by $J|_{\mbox{\goth l}^{10}}=i\mathrm{Id\;},\,\,J|_{\mbox{\goth l}^{01}}=-i\mathrm%
{Id\;}$.
The 2-form $\omega=d(B\circ h_{1})$ associated with the central
element $h_{1}\in Z({\mbox{\goth h}})=\mathbb{R}h$ defines a unique
(up to scaling) symplectic form compatible with $J$, where $B$ is the Killing form. We get
Theorem 7
The space $L(\mathbb{O}P^{2})=F_{4}/SO_{2}\cdot Spin_{7}$ admits a
unique (up to a sign) invariant complex structure $J$, unique (up to a
scaling) invariant symplectic structure $\omega=dB\circ h_{1},h_{1}=\in Z({\mbox{\goth h}})$ and a unique
invariant Kähler structure $(\omega,J)$.
References
[1]
A.L. Besse, Manifolds all of whose geodesics are closed, Ergebnisse der Mathematik und ihrer Grenzgebiete 93,
Springer, Berlin, 1978.
[2]
E. Ferrand, Sur la structure symplectique de l éspace des g éod ésiques d ún espace de Hadamard, Geom. Dedicata 68 (1997) 79–89.
[3]
N. Georgiou and B. Guilfoyle, On the space of oriented geodesics of hyperbolic 3-space Rocky Mountain J. Math. (to appear).
[4]
V. Gorbatcevich, A.L. Onishchik, and E.B. Vinberg, Foundations of Lie theory and Lie transformation groups, Springer-Verlag, Berlin,
1997.
[5]
B. Guilfoyle and W. Klingenberg, A Neutral Kähler structure on the space of oriented lines, J. London Math. Soc. 72 (2005) 497–509.
[6]
B. Guilfoyle and W. Klingenberg, Proof of the Carathéodory conjecture by mean curvature flow in the space of oriented affine lines, (2008) math.DG/0808.0851
[7]
N.J. Hitchin, Monopoles and geodesics, Comm. Math. Phys. 83 (1982) 579-602.
[8]
S. Kobayashi and K. Nomizu, Foundations of differential geometry Vol. II, Wiley and Sons, New York, 1996.
[9]
H. Potmann and J. Wallner, Computational line geometry, Springer Verlag, Berlin, 2001.
[10]
A. Reznikov, The weak Blaschke conjecture for CP(n), Invent. Math. 117 (1994) 447–454.
[11]
A. Reznikov, Symplectic twistor spaces, Ann. Global Anal. Geom. 11 (1993) 109–118.
[12]
M. Salvai, On the geometry of the space of oriented lines in Euclidean space, Manuscripta Math. 118 (2005) 181–189.
[13]
M. Salvai, Geometry of the space of oriented lines in hyperbolic space, Glasgow Math. J. 49 (2007) 357-366.
[14]
E. Study, Von den Bewegungen und Umlegungen I, II, Math. Ann. 34 (1891) 441–566. |
Abstract
We study the local geometry of testing a mean vector within a
high-dimensional ellipse against a compound alternative. Given samples
of a Gaussian random vector, the goal is to distinguish whether the
mean is equal to a known vector within an ellipse, or equal to some
other unknown vector in the ellipse. Such ellipse testing problems
lie at the heart of several applications, including non-parametric
goodness-of-fit testing, signal detection in cognitive radio, and
regression function testing in reproducing kernel Hilbert spaces.
While past work on such problems has focused on the
difficulty in a global sense, we study difficulty in a way that is
localized to each vector within the ellipse. Our main result is to
give sharp upper and lower bounds on the localized minimax testing
radius in terms of an explicit formula involving the Kolmogorov width
of the ellipse intersected with a Euclidean ball. When applied to
particular examples, our general theorems yield interesting rates that
were not known before: as a particular case, for testing in Sobolev
ellipses of smoothness $\alpha$, we demonstrate rates that vary from
$(\sigma^{2})^{\frac{4\alpha}{4\alpha+1}}$, corresponding to the
classical global rate, to the faster rate $(\sigma^{2})^{\frac{8\alpha}{8\alpha+1}}$, achievable for vectors at favorable
locations within the ellipse. We also show that the optimal test for
this problem is achieved by a linear projection test that is based on
an explicit lower-dimensional projection of the observation vector.
The local geometry of testing in ellipses:
Tight control
via localized Kolmogorov widths
Yuting Wei$${}^{\dagger}$$
Martin J. Wainwright$${}^{\dagger,\star}$$
January 13, 2021
Department of Statistics$${}^{\dagger}$$, and
Department of Electrical
Engineering and Computer Sciences$${}^{\star}$$
UC Berkeley, Berkeley,
CA 94720
1 Introduction
Testing and estimation are two fundamental classes of problems in
statistics [24, 25]. In a classical
decision-theoretic framework, different methods are compared via their
risk functions, and under the minimax formalism, methods are compared
by their worst-case behavior over the entire parameter space. Such
global minimax risk calculation are reflective of typical behavior
when the risk function is close to constant over the entire parameter
space. On the other hand, when the risk function varies
substantially, then it is possible that a global minimax calculation
will be unduly conservative. In such settings, one is motivated to
study the notion of an adaptive or localized minimax risk (e.g., see
references [10, 4] for early work
in this vein).
Recent years have witnessed a rapidly evolving line of work on
studying notions of local minimax risk for both estimation and
testing. In the context of shape-constrained regression, Meyer and
Woodruffe [28] introduced a notion of degrees of
freedom that adapts to the geometry of the function being estimated.
Focusing on the problem of isotonic regression, Chatterjee et
al. [7] proved a range of convergence rates,
depending on the structure of the true regression function. For the
task of estimating a convex function at a point, Cai and
Low [6] proposed a local minimax criterion to
evaluate the performance of different estimators. This criterion was
adapted to establish a form of instance-dependent optimality for
convex set estimation by Cai et al. [5]. In the
context of hypothesis testing, Valiant and
Valiant [32] studied a class of compound
testing problems for discrete distributions, and characterized the
local minimax testing radius in the TV norm. Balakrishnan and
Wasserman [2] studied an analogous
problem for testing density functions, and also characterized the
local testing radius in TV norm.
In this paper, we study the local geometry of testing a mean vector
inside an ellipse against a compound alternative. More precisely,
consider an ellipse of the form $\mathcal{E}=\{\theta\in{}^{d}\mid\sum_{j=1}^{d}\theta_{j}^{2}/\mu_{j}\leq 1\}$,
where $\{\mu_{j}\}_{j=1}^{d}$ is a non-negative sequence that
defines the aspect ratios of the ellipse. Although we focus on
finite-dimensional ellipses, all of our theory is non-asymptotic and
explicit, so that we can obtain results for infinite-dimensional
ellipses by taking suitable limits. Given an observation of a
Gaussian random vector $y\in{}^{d}$, our goal is to test
whether its mean is equal to some known vector $\theta^{*}\in\mathcal{E}$, or
equal to some unknown vector $\theta\in\mathcal{E}$ that is suitably
separated from $\theta^{*}$ in Euclidean norm. As we discuss in more detail
below, such ellipse testing problems lie at the heart of a number of
applications, including non-parametric goodness-of-fit testing, signal
detection in cognitive radio, and testing regression functions in
kernel spaces.
Study of the ellipse testing problems date back to the seminal work of
Ingster [19, 20, 21] and
Ermakov [12], who focused their attention on the
special case when the null is zero ($\theta^{*}=0$), and the ellipse
is induced by a Sobolev function class. For this particular class of
testing problems, they provided a sharp characterization of the
minimax testing radius. A more general question is whether it is
possible to provide a geometric characterization of the minimax
testing radius, both as a function of the ellipse parameters
$\{\mu_{j}\}_{j=1}^{d}$ as well as of the location of null vector
$\theta^{*}$ within the ellipse. The main contribution of this paper
is to answer this question in the affirmative. In particular, we show
how for any ellipse—including the Sobolev ellipses as special
cases—the localized minimax testing radius is characterized by a
formula that involves the Kolmogorov width of a particular set. The
Kolmogorov width is a classical notion from approximation
theory [29], which measures the “size” of a set in
terms of how well it can be approximated by a lower-dimensional linear
subspace of
fixed dimension. Our formula involves the Kolmogorov width of the
ellipse intersected with an Euclidean ball around the null vector
$\theta^{*}$, which leads to an equation involving both $\mathcal{E}$ and
$\theta^{*}$ that can be solved to determine the localized minimax
risk. We show that the zero case ($\theta^{*}=0$) is the most
difficult, in that the localized minimax radius is largest at
this instance. Conversely, we exhibit other more favorable locations
within ellipses for which the local minimax testing radius can be
substantially smaller.
1.1 Some motivating examples
Before proceeding, let us consider some concrete examples so as to
motivate our study.
Example 1 (Non-parametric goodness-of-fit testing).
Consider a set of samples $\{x_{i}\}_{i=1}^{n}$ drawn from an
unknown distribution supported on a compact set $\mathcal{X}$. Assuming
that the unknown distribution has a density, the goodness-of-fit
problem is to test whether the samples have been drawn from some fixed
density $p^{*}$, in which case the null hypothesis holds, or according
to some other density $p\neq p^{*}$, referred to as the alternative
hypothesis. In the non-parametric version of this problem, the
alternative density $p$ is allowed to vary over a broad function
class; one example might be the class of all twice continuously
differentiable densities with second derivative $p^{\prime\prime}$ belonging to the
unit ball in $L^{2}(\mathcal{X})$. There is a very broad literature on this
topic, with the book of Ingster and
Suslina [22] covering many different
classes of alternative densities.
One way to approach non-parametric goodness-of-fit is via orthogonal
series expansions. In particular, if we let $\{\phi_{j}\}_{j=1}^{\infty}$
be an orthonormal basis for $L^{2}(\mathcal{X})$, then the null density
$p^{*}$ and alternative $p$ can be described in terms of their basis
expansion coefficients
$$\displaystyle\theta^{*}_{j}=\int_{\mathcal{X}}\phi_{j}(x)p^{*}(x)dx\;=\;%
\mathbb{E}_{p^{*}}[\phi_{j}(X)],\quad\mbox{and}\quad\theta_{j}=\int_{\mathcal{%
X}}\phi_{j}(x)p(x)dx\;=\;\mathbb{E}_{p}[\phi_{j}(X)],$$
respectively, for $j=1,2,\ldots$. Note that the sample averages
$y_{j}:\,=\frac{1}{n}\sum_{i=1}^{n}\phi_{j}(x_{i})$ are
unbiased estimates of coefficients of the true underlying density, so
the testing problem can be written as
$$\displaystyle\mathcal{H}_{0}:\quad y=\theta^{*}+w\quad\mbox{versus}\quad%
\mathcal{H}_{1}:\quad y=\theta+w,\quad\mbox{for some $\theta\neq\theta^{*}$}$$
where $w=\{w_{j}\}_{j=1}^{\infty}$ is a sequence of noise variables. A
typical smoothness constraint on the alternative—for instance, the
condition $\int_{\mathcal{X}}(p^{\prime\prime}(x))^{2}dx\leq 1$—amounts to requiring
that, in a suitably chosen sinusoidal basis, the vector $\theta$
belongs an ellipse with parameters of the form $\mu_{j}=cj^{-4}$.
Example 2 (Detection of unknown signals in noise).
In cognitive radio and other wireless applications
(e.g., [1, 8, 30]),
it is frequently of interest to test for the presence of unknown
signals that might potentially interfere with transmission. More
concretely, given an observation period $[0,T]$, one observes a
continuous-time waveform $\{y(t)\mid t\in[0,T]\}$, and the goal is to
determine whether it was generated by pure noise, or a by combination
of some band-limited signal $\theta(t)$ plus noise. See
Figure 1 for an illustration.
After applying the Fourier transform, the waveform and signal can be
represented by sequences $y\in\ell^{2}(\mathbb{N})$ and $\theta\in\ell^{2}(\mathbb{N})$, respectively. Assuming that the Fourier frequencies
are ordered from smallest to largest, a hard band-width constraint on
the signal corresponds to the restriction that $\theta_{j}=0$ for all
$j>B$, where the integer $B$ corresponds to the bandlimit cut-off.
Note that this bandwidth constraint defines a (degenerate) ellipse in
$\ell^{2}(\mathbb{N})$. More generally, a soft bandwidth constraint can be
imposed by requiring that the weighted sum $\sum_{j=1}^{\infty}\frac{\theta_{j}^{2}}{\mu_{j}}$ for some summable sequence of positive
coefficients $\{\mu_{j}\}_{j=1}^{\infty}$. As before, this constraint
means that the unknown signal vector $\theta$ must belong to an
ellipse $\mathcal{E}$.
Given this set-up, the problem of detecting the presence of an unknown
signal versus noise corresponds to testing the null that $y\sim N(0,\sigma^{2}I_{d})$ versus the alternative $y\sim N(\theta,\sigma^{2}I_{d})$ for some $\theta\in\mathcal{E}$. More generally, one
might be interested in testing between the null hypothesis $y\sim N(\theta^{*},\sigma^{2}I_{d})$, for some fixed signal vector
$\theta^{*}\in\mathcal{E}$, versus the same alternative. We refer the
reader to the
papers [1, 8, 30]
for further details on this application.
Example 3 (Regression function testing in kernel spaces).
In the problem of fixed design regression, we observe pairs $(x_{i},z_{i})\in\mathcal{X}\times\real$ of the form
$$\displaystyle z_{i}$$
$$\displaystyle=f(x_{i})+\xi_{i}\quad\mbox{for $i=1,\ldots,n$}.$$
(1)
Here $f:\mathcal{X}\rightarrow\real$ is an unknown function, and
$\{\xi_{i}\}_{i=1}^{n}$ are a sequence of zero-mean noise variables
with variance $\sigma^{2}$. Based on the observation vector $z\in{}^{n}$, we might be interested in testing various hypotheses
related to the unknown function $f$. For instance, is $f$ a non-zero
function? More generally, does $f$ deviate significantly from
some fixed class of parametric functions?
These questions are meaningful only if we impose some structures on the
underlying function space, and given the
observations (1) based on pointwise evaluation, one
of the most natural is based on reproducing kernel Hilbert spaces
(e.g., [34, 14]). These function spaces are defined by a
symmetric positive semidefinite kernel function $\mathcal{K}:\mathcal{X}\times\mathcal{X}\rightarrow\real$; as some simple examples, when
$\mathcal{X}=[0,1]$, the kernel function $\mathcal{K}(x,x^{\prime})=1+\min\{x,x^{\prime}\}$ defines a class of Lipschitz functions, whereas the Gaussian
kernel $\mathcal{K}(x,x^{\prime})=\exp(-\frac{1}{2t}(x-x^{\prime})^{2})$ with bandwidth
$t>0$ defines a class of infinitely differentiable functions.
For any choice of kernel, the pointwise observation
model (1) can be transformed to a sequence model
form. In particular, define the kernel matrix $K\in{}^{n\times n}$ with entries $K_{ij}=\mathcal{K}(x_{i},x_{j})/n$, and
let $\mu_{1}\geq\mu_{2}\geq\cdots\geq\mu_{n}\geq 0$ represent
its ordered eigenvalues. In terms of this notation, the pointwise
model with any function $f$ in the unit norm Hilbert ball can be
transformed to the model $y=\theta+w$, where the vector $\theta\in{}^{n}$ satisfies the ellipse constraint
$\sum_{j=1}^{n}\frac{\theta_{j}^{2}}{\mu_{j}}\leq 1$, and each entry
of the noise vector is zero-mean with variance $\frac{\sigma^{2}}{n}$.
We have thus arrived at another instance of our general ellipse
problem.
1.2 Problem formulation
Having understood the range of motivations for our problem,
let us set up the problem more precisely. Given a sequence of
positive scalars $\mu_{1}\geq\ldots\geq\mu_{d}>0$, consider the
associated ellipse
$$\displaystyle\mathcal{E}:\,=\Big{\{}\theta\in{}^{d}\mid\sum_{i=1}^{d}\frac{%
\theta_{i}^{2}}{\mu_{i}}\leq 1\Big{\}}.$$
(2)
Suppose that we make observations of the form $y=\theta+\sigma g$.
Here $\theta\in\mathcal{E}$ is an unknown vector, whereas $g\sim N(0,I_{d})$ is a noise vector, and the variance level $\sigma^{2}$ is known. For
a given vector $\theta^{*}\in\mathcal{E}$, our goal is to test the null
hypothesis $\theta=\theta^{*}$ versus the alternative $\theta\in\mathcal{E}\backslash\{\theta^{*}\}$.
Under this formulation, it is not possible to make any non-trivial
assertion about the power of any test, since the alternative allows
for vectors $\theta$ that are arbitrarily close to $\theta^{*}$. In
order to make quantitative statements, we need to exclude a certain
$\epsilon$-ball around $\theta^{*}$ from the alternative. Doing so
leads to the notion of the minimax testing radius associated
this composite decision problem. This minimax formulation was
introduced in the seminal work of Ingster and
co-authors [18, 22]; since
then, it has been studied by many other researchers
(e.g., [12, 31, 26, 27, 3]).
More precisely, for a given radius $\epsilon>0$, we consider the
compound testing problem specified by the observation model $y\sim N(\theta,\sigma^{2}I_{d})$, and the null and alternative
$$\displaystyle\mathcal{H}_{0}:\theta=\theta^{*}\in\mathcal{E}~{}~{}~{}\text{%
versus}~{}~{}~{}\mathcal{H}_{1}:\|\theta-\theta^{*}\|_{2}\geq\epsilon,\theta%
\in\mathcal{E}.$$
(3)
We thus have a sequence of testing problems indexed by $\epsilon>0$,
and our goal is to calibrate their difficulty in terms of the noise
level $\sigma^{2}$, the null vector $\theta^{*}$, and the local geometry
of the ellipse $\mathcal{E}$ around $\theta^{*}$. To be clear, all the
tests that we analyze in this paper are not given knowledge of
$\epsilon$; rather, it is a unknown quantity that is used to titrate
the difficulty of the problem. See Figure 2 for an
illustration of the testing problem (3).
Letting $\psi:{}^{d}\rightarrow\{0,1\}$ be any measurable
test function, we characterize its performance on the testing
problem (3) in terms of the sum of its (type I)
error under the null, and its (type II) error uniformly over the
alternative. In particular, we define the uniform error
$$\displaystyle\operatorname{Err}(\psi;\theta^{*},\epsilon):\,=\mathbb{E}_{%
\theta^{*}}[\psi(y)]+\sup_{\theta\in\mathcal{E},\|\theta-\theta^{*}\|_{2}\geq%
\epsilon}\mathbb{E}_{\theta}[1-\psi(y)],$$
(4)
where $\mathbb{E}_{\theta}$ denotes expectation over $y$ under the $N(\theta,\sigma^{2}I_{d})$ distribution.
For a given error level $\rho\in(0,1)$, we are interested in the
smallest setting of $\epsilon$ for which there exists some test with
uniform error at most $\rho$. More precisely, we define
$$\displaystyle\epsilon_{\text{{\tiny OPT}}}(\theta^{*};\rho,\mathcal{E}):\,=%
\inf\Big{\{}\epsilon\,\mid\,\inf_{\psi}\;\operatorname{Err}(\psi;\theta^{*},%
\epsilon)\leq\rho\Big{\}},$$
(5)
a quantity which we call the $\theta^{*}$-local minimax testing
radius. By definition, the local minimax testing radius
$\epsilon_{\text{{\tiny OPT}}}(\theta^{*};\rho,\mathcal{E})$ corresponds to the smallest separation
$\epsilon$ at which there exists some test that distinguishes
between the hypotheses $\mathcal{H}_{0}$ and $\mathcal{H}_{1}$ in
equation (3) with uniform error at most $\rho$.
Thus, it provides a fundamental characterization of the statistical
difficulty of the hypothesis testing as a function of $\theta^{*}.$
Our results also have consequences for the more classical global
minimax testing radius, given by
$$\displaystyle\epsilon_{\text{{\tiny OPT}}}(\rho):\,=\sup_{\theta^{*}\in%
\mathcal{E}}\epsilon_{\text{{\tiny OPT}}}(\theta^{*};\rho,\mathcal{E}).$$
As we discuss in the sequel, this global quantity has been
characterized for various Sobolev ellipses in past
work [22].
1.3 Overview of our results
Having set up the problem, let us now provide a high-level overview of
the main results of this paper. We will show that the testing radius is
fully described by a purely geometric quantity, namely the
Kolmogorov $k$-width of a set that localized at
$\theta^{*}$. Given a set $A$, the Kolmogorov width characterizes how
well the set is approximated by a $k$-dimensional linear subspace. It
is known that the Kolmogorov width provides risk upper bounds for the
truncated series estimator
(e.g. [23, 11]) and it also turns
out to be a key quantity in problems such as density estimation, as
well as compressed sensing (see the
papers [15, 9]). We discuss
the definition of the Kolmogorov width and its property at length in
Section 2.2.
We will also show that the optimal test, as will be discussed in sequel, is given by a linear projection test that is based on
the projection of observation vector to a particular lower dimensional
subspace. More discussions on its formulation and background can be
found in our Section 2.1. We state our main results and
some of their important consequences in Section 3.
Section 4 is devoted to the proofs of our main
theorems and corollaries. We defer the auxiliary lemmas for our main
proofs to the appendices.
2 Background
Before proceeding to the statements of our main results, we introduce some
background on linear projection tests, as well as the notion of Kolmogorov
width.
2.1 Linear projection tests
In this paper, we prove some upper bounds on $\epsilon_{\text{{\tiny OPT}}}(\theta^{*};\rho,\mathcal{E})$ via a concretely defined class of tests, in particular
those based on linear projections. Given an integer $k\in[d]:\,=\{1,\ldots,d\}$, let $\Pi_{k}$ denote an orthogonal
projection operator, mapping any vector $v\in{}^{d}$ to a
particular $k$-dimensional subspace. Any projection $\Pi_{k}$
defines a family of linear projection tests, indexed by the threshold
$\beta\geq 0$, of the following form
$$\displaystyle\psi_{\beta,\Pi_{k}}(y):\,=\mathbb{I}\big{[}\|\Pi_{k}(y-\theta^{*%
})\|_{2}\geq\beta\big{]}\;=\;\begin{cases}1&\mbox{if
$\|\Pi_{k}(y-\theta^{*})\|_{2}\geq\beta$}\\
0&\mbox{otherwise.}\end{cases}$$
(6)
We use $\Psi_{\text{{\tiny LPT}}}$ to denote the collection of all linear projection
tests
$$\displaystyle\Psi_{\text{{\tiny LPT}}}$$
$$\displaystyle:\,=\Big{\{}\psi_{\beta,\Pi_{k}}\;\mid k\in[d],\;\Pi_{k}\in%
\mathcal{P}_{k},\;\beta\in{}_{+}\Big{\}},$$
obtained by varying the dimension $k\in[d]$, the projection
$\Pi_{k}$ over the space $\mathcal{P}_{k}$ of all $k$-dimensional
orthogonal projections, and the threshold $\beta\geq 0$.
Given the family $\Psi_{\text{{\tiny LPT}}}$, we then define
$$\displaystyle\epsilon_{\text{{\tiny LPT}}}(\theta^{*};\rho,\mathcal{E}):\,=%
\inf\Big{\{}\epsilon\,\mid\,\inf_{\psi\in\Psi_{\text{{\tiny LPT}}}}\;%
\operatorname{Err}(\psi;\theta^{*},\epsilon)\leq\rho\Big{\}}.$$
(7)
a quantity which we call
$\theta^{*}$-local LPT testing radius. When the error level
$\rho$ is clear from the context, we adopt the shorthand notation
$\epsilon_{\text{{\tiny LPT}}}(\theta^{*};\mathcal{E})$ and $\epsilon_{\text{{\tiny OPT}}}(\theta^{*};\mathcal{E})$, with the implicit
understanding that all of our analysis depends on the choice of
$\rho$.
2.2 Kolmogorov width
Our characterization of testing difficulty involves a classical
geometric notion known as the Kolmogorov width, which we introduce
here; we refer the reader to the book by Pinkus [29] for
more details.
For a given compact set $C\subset{}^{d}$ and integer $k\in[d]$, the Kolmogorov $k$-width is a measure of how well the set
can be approximated by a $k$-dimensional subspace. More precisely,
recalling that $\mathcal{P}_{k}$ denotes the set of all $k$-dimensional
orthogonal linear projections, the Kolmogorov $k$-width,
denoted by $\omega_{k}$, is given by
$$\displaystyle\omega_{k}(C)=\min_{\Pi_{k}\in\mathcal{P}_{k}}\max_{\theta\in C}%
\|\theta-\Pi_{k}\theta\|_{2}.$$
(8)
Any $\Pi_{k}$ achieving the minimum in the saddle point
problem (8) is said to be an optimal
projection for $\omega_{k}(C)$. By definition, the Kolmogorov
widths are non-increasing as a function of the integer $k$—in
particular, we have
$$\displaystyle r(C)=\omega_{0}(C)\geq\omega_{1}(C)\geq\ldots\geq\omega_{d}(C)=0,$$
where $r(C):\,=\max_{\theta\in C}\|\theta\|_{2}$ is the diameter
of the compact set $C$.
3 Main results and their consequences
We now turn to the statement of our main results, along with a
discussion of some of their consequences.
3.1 Upper bound on the local minimax testing radius
In this section, we establish an upper bound on the minimax testing
radius for problem (3) localized to a given vector
$\theta^{*}$. In order to do so, we study the behavior of the linear
projection tests previously described in Section 2.1, It
turns out that LPT testing radius $\epsilon_{\text{{\tiny LPT}}}(\theta^{*};\mathcal{E})$ can be
characterized by the Kolmogorov $k$-width—as defined in
expression (8)—of a particular set.
Before stating the main result, let us first introduce some notation.
For each $\epsilon>0$, we define the upper critical dimension
$$\displaystyle k_{u}(\epsilon,\theta^{*},\mathcal{E}):\,=\arg\min_{1\leq k\leq d%
}\Big{\{}\omega_{k}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(\epsilon))\leq\frac%
{\epsilon}{\sqrt{2}}\Big{\}}.$$
(9)
Here set $\mathcal{E}_{\theta^{*}}:\,=\{\theta-\theta^{*}\mid\theta\in\mathcal{E}\}$
for the ellipse defined in expression (2) and
$\mathbb{B}(\epsilon):\,=\{v\in{}^{d}\mid\|v\|_{2}\leq\epsilon\}.$
For any fixed $\epsilon>0$, the Kolmogorov width
$\omega_{k}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(\epsilon))$ is non-increasing
with index $k$ and it goes to zero as $k$ approaches $d$ (where
$d$ can grow as well, then we have an infinite dimensional
ellipse) and $\omega_{d}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(\epsilon))=0$.
Therefore, the critical dimension is always well defined.
This dimension is used to define the optimal linear projection test at
$\theta^{*}$. We also use it to define for each error level $\rho\in(0,1/2]$, the upper critical radius
$$\displaystyle\epsilon_{u}(\theta^{*};\rho,\mathcal{E})$$
$$\displaystyle:\,=\inf\{\epsilon\mid\epsilon\geq\frac{8}{\sqrt{\rho}}\sigma^{2}%
\frac{\sqrt{k_{u}(\epsilon,\theta^{*},\mathcal{E})}}{\epsilon}\Big{\}}.$$
(10)
For each choice of error level $\rho$, this radius defines the
separation between null and alternative that can be distinguished with
error controlled at level $\rho$. From the fact that the function
$\epsilon\mapsto k_{u}(\epsilon,\theta^{*},\mathcal{E})$ is
non-increasing, it follows that $\epsilon_{u}(\theta^{*};\rho,\mathcal{E})$
always exists, and is guaranteed to be unique and strictly positive.
More details can be found in Section D.1.
With these two ingredients, we are ready to state our upper
bound for the LPT testing radius localized to $\theta^{*}$:
Theorem 1.
For any vector $\theta^{*}\in\mathcal{E}$ and any error level $\rho\in(0,1/2]$, we have
$$\displaystyle\inf_{\psi\in\Psi_{\text{{\tiny LPT}}}}\;\operatorname{Err}(\psi;%
\theta^{*},\delta)\leq\rho\qquad\text{ for all }\delta\geq\epsilon_{u}(\theta^%
{*};\rho,\mathcal{E}).$$
(11)
We prove this theorem in Section 4.1.
Recall the definition of the $\theta^{*}$-localized minimax radius
$\epsilon_{\text{{\tiny OPT}}}$ and $\theta^{*}$-localized LPT radius $\epsilon_{\text{{\tiny LPT}}}$ from
equations (5)
and (7), respectively. In terms of this
notation, Theorem 1 guarantees that
$$\displaystyle\epsilon_{\text{{\tiny OPT}}}(\theta^{*};\rho,\mathcal{E})\;%
\stackrel{{\scriptstyle(i)}}{{\leq}}\;\epsilon_{\text{{\tiny LPT}}}(\theta^{*}%
;\rho,\mathcal{E})\;\stackrel{{\scriptstyle(ii)}}{{\leq}}\;\epsilon_{u}(\theta%
^{*};\rho,\mathcal{E}).$$
To be clear, the first inequality (i) is a trivial consequence of the
definitions, whereas inequality (ii) is the substance of
Theorem 1.
Structure of optimal linear test:
Although we have stated
the theorem in an existential way, our analysis also gives an explicit
form of the test that achieves the error
guarantee (11). The construction of this test
consists of three steps:
•
For a given error level $\rho$, first compute the critical
radius $\epsilon_{u}(\theta^{*};\rho,\mathcal{E})$ using
expression (10).
•
Second, using equation (9), compute the induced
critical upper dimension
$$\displaystyle k^{*}_{u}:\,=k_{u}(\epsilon_{u}(\theta^{*};\rho,\mathcal{E}),%
\theta^{*},\mathcal{E}),$$
as well as the associated projection matrix $\Pi_{k^{*}_{u}}$ that
defines the Kolmogorov width indexed by $k^{*}_{u}$.
•
Third, apply the linear projection test
$$\displaystyle y\mapsto\mathbb{I}\Big{[}\|\Pi_{k^{*}_{u}}(y-\theta^{*})\|_{2}^{%
2}\geq\sigma^{2}(k^{*}_{u}+\sqrt{4k^{*}_{u}/\rho})\Big{]}.$$
(12)
The proof of the theorem is based on analyzing this linear
test (12) directly and explicitly controlling its type I
and type II error.
In order to illustrate the use of Theorem 1,
let us consider a very simple example.
Example 4 (Circular constraints).
Suppose that the ellipse $\mathcal{E}$ is simply a circle in
${}^{d}$, say with $\mu_{j}=d$ for all $j=1,\ldots,d$. Let us consider the case of testing at zero (so
that $\theta^{*}=0$) with a noise level $\sigma\in(0,1)$. For
any $\epsilon\in(0,\sqrt{d})$, we claim that the upper
critical dimension (9) is equal to $d$.
Indeed, for any projection matrix of dimension $k<d$, there
is at least one dimension that is missing, and this dimension
contains a vector of Euclidean length $\sqrt{d}$ inside the
sphere. Consequently, solving for the upper critical radius from
equation (10) yields that $\epsilon_{u}^{2}(0;\mathcal{E})=\frac{8}{\sqrt{\rho}}\sigma^{2}\sqrt{d}$.
Note that this rate matches the minimax testing radius for a subspace
of dimension $d$. This calculation makes sense, because for
this particular example—that is, testing at the zero vector for a
circle with the chosen scalings of $\mu_{j}$ and $\sigma$—the geometry
and boundary of the circle actually plays no role.
In Section 3.3, we consider some more substantive
applications of Theorem 1, and in particular,
demonstrate that the local minimax testing radius varies substantially
as a function of the location of $\theta^{*}$ within the ellipse.
3.2 Lower bound on the local minimax testing radius
Thus far, we have derived an upper bound for the localized testing
radius for a particular procedure—namely, the LPT. Of course, it is
of interest to understand when the LPT is actually an optimal test,
meaning that there is no other test that can discriminate between the
null and alternative for smaller separations. In this section, we use
information-theoretic methods to prove a lower bound on the localized
minimax testing radius $\epsilon_{\text{{\tiny OPT}}}(\theta^{*};\mathcal{E})$. Whenever the
radius $\epsilon_{\text{{\tiny LPT}}}(\theta^{*};\mathcal{E})$ achievable by a linear projection
test matches the lower bound for every $\theta^{*}$, we can claim the
linear projection tests are locally minimax optimal.
For the testing problem (3), it turns out that the
local minimax testing radius $\epsilon_{\text{{\tiny OPT}}}(\theta^{*};\mathcal{E})$ can also be
lower bounded via the Kolmogorov $k$-width, as previously defined in
equation (8), of a particular set. In order to
state this lower bound, we fix a triplet of positive constants $a,b,c$
linked by the relations
$$\displaystyle a>3b,\mbox{ and }c=c(a,b):\,=\frac{b}{8\sqrt{2}}-\frac{\sqrt{a^{%
2}-9b^{2}}}{12\sqrt{2}}>0.$$
(13)
Recall the recentered ellipse $\mathcal{E}_{\theta^{*}}:\,=\{\theta-\theta^{*}\mid\theta\in\mathcal{E}\}$, and Euclidean ball of radius $\epsilon$ centered at zero $\mathbb{B}(\epsilon):\,=\{v\in{}^{d}\mid\|v\|_{2}\leq\epsilon\}$.
For our upper bound, the upper critical dimension $k_{u}$ from
equation (9) played a central role. As the lower
analogue to this quantity, let us define the lower critical
dimension
$$\displaystyle k_{\ell}(\epsilon,\theta^{*},\mathcal{E}):\,=\arg\min_{1\leq k%
\leq d}\left\{\omega_{k}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(a\epsilon))%
\leq 3b\epsilon\right\},$$
(14)
where the fixed choice of constants $(a,b)$ should be understood
implicitly. We then define the lower critical radius
$$\displaystyle\epsilon_{\ell}(\theta^{*};\mathcal{E})$$
$$\displaystyle:\,=\sup\Big{\{}\epsilon\mid\epsilon\leq\frac{1}{4}\sigma^{2}%
\frac{\sqrt{k_{\ell}(\epsilon,\theta^{*},\mathcal{E})}}{\epsilon}\Big{\}},$$
(15)
From the fact that the function $\epsilon\mapsto k_{\ell}(\epsilon,\theta^{*},\mathcal{E})$ is non-increasing, it follows that $\epsilon_{\ell}$ is
well-defined and unique. More details can be found in
Appendix D.1.
We prove a lower bound as a function of the quantity
$\epsilon_{\ell}(\theta^{*};\mathcal{E})$ and also a second quantity, one which
measures its proximity to the boundary of the ellipse. More
precisely, for a given vector $\theta^{*}$ and constant $a>0$, we
define the mapping $\Phi:{}_{+}\rightarrow{}_{+}$ via
$$\displaystyle\Phi(\delta)$$
$$\displaystyle=\begin{cases}1&\mbox{if $\delta>\|\theta^{*}\|_{2}/a$}\\
1\wedge\min\Big{\{}r\geq 0\,\mid a^{2}\delta^{2}\leq\sum_{i=1}^{d}\frac{r^{2}}%
{(r+\mu_{i})^{2}}(\theta^{*}_{i})^{2}\Big{\}}&\mbox{otherwise.}\end{cases}$$
(16)
As shown in Appendix D.2, this mapping is well-defined, and
has the limiting behavior $\Phi(\delta)\rightarrow 0$ as $\delta\rightarrow 0^{+}$. Let us denote $\Phi^{-1}(x)$ as the largest
positive number of $\delta$ such that $\Phi(\delta)\leq x$. Note
that by this definition, we have $\Phi^{-1}(1)=\infty$.
We are now ready to state our lower bound for the optimal
testing radius localized to $\theta^{*}$:
Theorem 2.
Consider constants $a,b,c\in(0,1)$ satisfying the
conditions (13). Then for any $\theta^{*}\in\mathcal{E}$,
we have
$$\displaystyle\inf_{\psi}\;\operatorname{Err}(\psi;\theta^{*},\delta)\geq\frac{%
1}{2}\qquad\text{ whenever }\quad\delta\,\leq\,c\min\Big{\{}\epsilon_{\ell}(%
\theta^{*};\mathcal{E}),~{}~{}\Phi^{-1}\big{(}(\|\theta^{*}\|_{\mathcal{E}}^{-%
1}-1)^{2}\big{)}\Big{\}}.$$
(17)
See Section 4.2 for the proof of this theorem.
Remarks:
Regarding concrete choices of the constants $(a,b,c)$
that underlie this theorem, if we set $a=\frac{\sqrt{43}}{12}$ and
$b=\frac{1}{4}$, then $c=\frac{1}{288\sqrt{2}}$. Since our main interest
is to understand the scaling of the testing radius with respect to
$\sigma$ and the geometric parameters of the problem, we have made no
efforts to obtain the sharpest constants in the theorem statement.
Second, to clarify the role of the term $\Phi^{-1}\big{(}(\|\theta^{*}\|_{\mathcal{E}}^{-1}-1)^{2}\big{)}$, it is not of primary interest
and serves to measure whether or not $\theta^{*}$ is sufficiently far
inside the ellipse. Concretely, if we assume that $\|\theta^{*}\|_{\mathcal{E}}\leq 1/2$, then $(\|\theta^{*}\|_{\mathcal{E}}^{-1}-1)^{2}\geq 1$ therefore
$\Phi^{-1}\big{(}(\|\theta^{*}\|_{\mathcal{E}}^{-1}-1)^{2}\big{)}=\infty$ which
means that
$$\displaystyle\min\Big{\{}\epsilon_{\ell}(\theta^{*};\mathcal{E}),~{}~{}\Phi^{-%
1}\big{(}(\|\theta^{*}\|_{\mathcal{E}}^{-1}-1)^{2}\big{)}\Big{\}}=\epsilon_{%
\ell}(\theta^{*};\mathcal{E}).$$
A final remark on the proof of Theorem 2: it is
based on constructing a prior distribution on the alternative $\mathcal{H}_{1}$
that is independent of observation $y$. This distribution is
supported on precisely those points in $\mathcal{H}_{1}$ that are relatively
hard to distinguish from $\mathcal{H}_{0}$. Then the testing error can be
lower bounded through controlling the total variation (TV) distance between
two marginal likelihood functions.
3.3 Some consequences of our results
One useful consequence of Theorems 1
and 2 is in providing a sufficient condition for the
optimality of the LPT. In particular, suppose that the upper and
lower dimensions defined in equations (9)
and (14) differ only by constant pre-factors, which we
write as
$$\displaystyle k_{u}(\epsilon,\theta^{*},\mathcal{E})\asymp k_{\ell}(\epsilon;%
\theta^{*},\mathcal{E}).$$
(18)
Under this condition, it follows that the critical upper and lower
radii match—namely, that $\epsilon_{u}(\theta^{*};\mathcal{E})\asymp\epsilon_{\ell}(\theta^{*};\mathcal{%
E})$. Theorem 1 then
guarantees that the linear projection test at $\theta^{*}$ is locally
minimax optimal.
As previously discussed, the local optimality studied here provides a
finer-grained distinction than the usual notion of global minimax
optimality. All of the rates depend on the vector $\theta^{*}$ via
the set $\mathcal{E}_{\theta}^{*}\cap\mathbb{B}(\epsilon)$, whose shape changes as
$\theta^{*}$ moves across the ellipse. The resulting changes in the local
minimax radii can be quite substantial, as we now illustrate with some
concrete examples.
3.3.1 Testing at zero
We begin our exploration by considering the testing
problem (3) with $\theta^{*}=0$. In order to
characterize the $\theta^{*}$-localized minimax testing
radius (5), we define the functions
$$\displaystyle m_{u}(\delta;\mathcal{E}):\,=\arg\max_{1\leq k\leq d}\{\mu_{k}%
\geq\frac{1}{2}\delta^{2}\},~{}\text{ and }~{}m_{\ell}(\delta;\mathcal{E}):\,=%
\arg\max_{1\leq k\leq d}\{\mu_{k+1}\geq\frac{9}{16}\delta^{2}\}.$$
(19)
In order to ensure that both functions are well-defined, we restrict
our attention to $\delta$ in the interval $\big{(}0,\min\{\sqrt{2\mu_{1}},\frac{4}{3}\sqrt{\mu_{2}}\}\big{)}$. We now state
a consequence of Theorems 1 and 2 in
terms of these quantities:
Corollary 1.
Given an error level $\rho\in(0,1)$, the local minimax testing
radius $\epsilon_{\text{{\tiny OPT}}}(\theta^{*};\mathcal{E})$ at $\theta^{*}=0$ is bounded as
$$\displaystyle\sup\Big{\{}\delta\,\mid\,\delta\leq\frac{1}{4}\sigma^{2}\frac{%
\sqrt{m_{\ell}(\delta;\mathcal{E})}}{\delta}\Big{\}}\;\leq\;\epsilon_{\text{{%
\tiny OPT}}}(0;\mathcal{E})\;\leq\;\inf\Big{\{}\delta\,\mid\,\delta\geq\frac{8%
}{\sqrt{\rho}}\sigma^{2}\frac{\sqrt{m_{u}(\delta;\mathcal{E})}}{\delta}\Big{\}},$$
(20)
where we only consider $\delta\leq\min\{\sqrt{2\mu_{1}},\frac{4}{3}\sqrt{\mu_{2}}\}$.
We prove Corollary 1 in
Section 4.3.
The quantities defined on the left and right hand sides of
expression (20) exist due to the fact that
$m_{u}(\delta;\mathcal{E})$ and $m_{\ell}(\delta;\mathcal{E})$ are
non-increasing functions of $\delta.$ Putting inequalities in
(20) together, whenever $m_{u}$ and $m_{\ell}$ differ
only by constant factors, we have a tight characterization of the
testing radius $\epsilon_{\text{{\tiny OPT}}}(0;\mathcal{E})$ localized to $\theta^{*}=0$.
In order to illustrate, we now apply Corollary 1 to some
special cases of the ellipse $\mathcal{E}$, and see how the minimax testing
radius localized to zero varies as a function of the ellipse
parameters. In particular, let us consider two broad classes of decay
rates for the ellipse parameters $\{\mu_{j}\}_{j}$.
$\gamma$-exponential decay:
For some $\gamma>0$, suppose the scalars $\{\mu_{j}\}_{j}$ satisfies a decay
condition of the form $\mu_{j}=c_{1}\exp(-c_{2}j^{\gamma})$, where
$c_{1},c_{2}$ are universal constants. Such exponential decay rates arise
the problem of non-parametric regression testing
(Example 3), in particular for reproducing kernel
Hilbert spaces based on Gaussian kernels. From
expression (19), we know $k_{u}=C\log^{1/\gamma}(1/\epsilon)$ and $k_{B}=c\log^{1/\gamma}(1/\epsilon)$
and solving the two inequalities from equation (20)
yields
$$\displaystyle\epsilon_{\text{{\tiny OPT}}}^{2}(0)\asymp\sigma^{2}(\log(\frac{1%
}{\sigma^{2}}))^{1/2\gamma}.$$
(21)
Here $\asymp$ denotes equality up to a factor involving constants and
terms of the order $\log\log(1/\sigma^{2})$.
$\alpha$-polynomial decay:
Now suppose that the ellipse parameters $\{\mu_{j}\}_{j}$ satisfy a
decay condition of the form $\mu_{j}=c_{1}j^{-2\alpha}$ for some
$\alpha>1/2$. As previously discussed, polynomial decay of this
type arises in ellipses that are induced by Sobolev function spaces of
smoothness $\alpha$. From expression (19), we know
$k_{u}=C\epsilon^{-1/\alpha}$ and $k_{B}=c\epsilon^{-1/\alpha}$ and
solving the two inequalities in equation (20) yields
$$\displaystyle\epsilon_{\text{{\tiny OPT}}}^{2}(0)\asymp(\sigma^{2})^{\frac{4%
\alpha}{1+4\alpha}}.$$
(22)
Here our notation $\asymp$ denotes equality up to constants
independent of $(\sigma,d)$, so that
inequality (22) provides a characterization of the
testing radius localized at zero that is tight up to constant factors.
It is worth noticing that that rate (22) matches the
(non-localized) globally minimax optimal testing radius established by
Ingster [19, 20, 21]. Thus, testing at
zero is as hard as the hardest testing problem in the ellipse.
Intuitively, this correspondence makes sense because the ellipse
contains a relatively large volume around zero, meaning that the
compound alternative contains a larger number of possible vectors that
need to be distinguished from the null.
Ingster [19] proves the lower bound by the direct
analysis of a $\chi^{2}$-type test, which is then shown to be optimal.
For the estimation analogue of this problem, Ibragimov and
Khasminskii [16, 17] showed that
the global minimax rate for estimation scales as
$\epsilon_{\text{est}}\asymp(\sigma^{2})^{2\alpha/(1+2\alpha)}$. This slower rate confirms
the intuition that testing is an easier problem than estimation:
we can detect a vector of length much smaller than the accuracy
at which we can estimate.
Remark:
As a final comment,
Corollary 1 can be generalized to the $\ell_{p}$-ellipse
with $p\geq 2$, given by $\mathcal{E}_{p}:\,=\{\theta\in{}^{d}\mid\sum_{i=1}^{d}\frac{\theta_{i}^{p}}{%
\mu_{i}}\leq 1\}$. Let us again consider testing at $\theta^{*}=0.$ Using the
same technique to prove Corollary 1, it is easy to check
that inequalities in equation (20) still hold with the
definitions of $m_{u}$ and $m_{\ell}$ in
expression (19) replaced by
$$\displaystyle m_{u}(\delta;\mathcal{E}):\,=\arg\max_{1\leq k\leq d}\big{\{}\mu%
_{k+1}^{2/p}\geq\frac{1}{2}\delta^{2}\big{\}},\quad\mbox{and}\quad m_{\ell}(%
\delta;\mathcal{E}):\,=\arg\max_{1\leq k\leq d}\big{\{}\mu_{k}^{2/p}\geq\delta%
^{2}\big{\}}.$$
3.3.2 Testing at extremal vectors
In the previous section, we studied the behavior of the minimax
testing radius for $\theta^{*}=0$, and recovered some known results from
the classical literature. In this section, we study some non-zero
cases of the vector $\theta^{*}$. For concreteness, we restrict our
attention to vectors that are non-zero some coordinate $s\in[d]=\{1,\ldots,d\}$, and zero in all other
coordinates. Even for such simple vectors, our analysis reveals some
interesting scaling of the minimax testing radius that is novel.
Given an integer $s\in[d]$ and a positive scalar $\delta$, we
define
$$\displaystyle m_{u}(\delta;s)$$
$$\displaystyle:\,=\arg\max_{1\leq k\leq d}\Big{\{}\mu_{k}^{2}\geq\frac{1}{64}%
\delta^{2}\mu_{s}\Big{\}},~{}\text{ and }~{}m_{\ell}(\delta;s):\,=\arg\max_{1%
\leq k\leq d}\Big{\{}\mu^{2}_{k}\geq\delta^{2}\mu_{s}\Big{\}}.$$
(23)
So as to ensure that these functions are well-defined, we restrict our
attention to $\delta$ in the interval $\big{(}0,\mu_{1}/\sqrt{\mu_{s}}\big{)}$. By definition, we have $m_{u}(\delta;s)\geq m_{\ell}(\delta;s)$.
Before stating our corollary, let us define the upper and lower
critical radii as follows
$$\displaystyle t^{*}_{u}(s,\mathcal{E})$$
$$\displaystyle=\inf\Big{\{}\delta>0\mid\delta\geq\frac{8}{\sqrt{\rho}}\sigma^{2%
}\frac{\sqrt{m_{u}(\delta;s)}}{\delta}\Big{\}}$$
(24a)
$$\displaystyle\text{ and
}~{}~{}t^{*}_{\ell}(s,\mathcal{E})$$
$$\displaystyle=\sup\Big{\{}\delta>0\mid\delta\leq\frac{1}{4}\sigma^{2}\frac{%
\sqrt{m_{\ell}(\delta;s)}}{\delta}\Big{\}}.$$
(24b)
Using the fact that $\rho\in(0,1/2]$ and non-increasing functions
satisfy $m_{u}(\delta;s)\geq m_{\ell}(\delta;s)$, we have $t^{*}_{\ell}(s,\mathcal{E})\leq t^{*}_{u}(s,\mathcal{E})$. Again these two quantities do not depend
on the $\theta^{*}$ that we are testing at. We then define the set
$$\displaystyle C(s;\mathcal{E})$$
$$\displaystyle:\,=\Big{\{}\theta\in{}^{d}\mid\theta_{j}=0\quad\mbox{for all $j%
\neq s$, and}\quad\theta_{s}\in\big{[}\sqrt{\mu_{s}}-t^{*}_{u}(s,\mathcal{E}),%
\sqrt{\mu_{s}}-t^{*}_{\ell}(s,\mathcal{E})\big{]}\Big{\}}.$$
We are ready to state our corollary:
Corollary 2.
Consider any positive integer $s$ such that $t^{*}_{u}(s,\mathcal{E})\leq\sqrt{\mu_{s}}$. Then for any testing error level $\rho>0$, we
have
$$\displaystyle t^{*}_{\ell}(s,\mathcal{E})\,\leq\,\epsilon_{\text{{\tiny OPT}}}%
(\theta^{*};\mathcal{E})\,\leq\,t^{*}_{u}(s,\mathcal{E})\qquad\mbox{for all $%
\theta^{*}\in C(s;\mathcal{E})$.}$$
(25)
See Section 4.4 for the proof of
Corollary 2.
Corollary 2 characterizes the optimal testing radius for
vectors that are sufficiently close to the boundary of the ellipse, In
order to compare with the results of Section 3.3.1 for
testing at zero, let us apply Corollary 2 to some
special examples of the ellipse $\mathcal{E}$, and see how the testing
radius varies as we vary the index $s\in[d]$.
For these comparisons, we restrict our attention to ellipses
for which the parameters $\{\mu_{j}\}$ decay at a polynomial rate.
$\alpha$-Polynomial decay:
Suppose that the scalars $\{\mu_{j}\}_{j}$ satisfy a decay condition of the form
$\mu_{j}=c_{1}j^{-2\alpha}$ for some $\alpha>1/2$. Recalling
the definition (23), some calculations yield
$$\displaystyle\frac{\mu_{m_{u}}^{2}}{\mu_{s}}$$
$$\displaystyle=\frac{m_{u}^{-4\alpha}}{s^{-2\alpha}}~{}\geq~{}\frac{\delta^{2}}%
{64}~{}\Longrightarrow~{}m_{u}=\lfloor(64s^{2\alpha})^{\frac{1}{4\alpha}}%
\delta^{-\frac{1}{2\alpha}}\rfloor,\qquad\mbox{and}$$
$$\displaystyle\frac{\mu_{m_{\ell}}^{2}}{\mu_{s}}$$
$$\displaystyle=\frac{m_{\ell}^{-4\alpha}}{s^{-2\alpha}}~{}\geq~{}\delta^{2}~{}%
\Longrightarrow~{}m_{\ell}=\lfloor s^{\frac{1}{2}}\delta^{-\frac{1}{2\alpha}}\rfloor.$$
Solving the fixed point equations yields
$$\displaystyle t^{*}_{u}(s,\mathcal{E})=\Big{(}8^{\frac{4\alpha+1}{4\alpha}}%
\rho^{-\frac{1}{2}}s^{\frac{1}{4}}\sigma^{2}\Big{)}^{\frac{4\alpha}{8\alpha+1}%
}\qquad\text{and }~{}~{}t^{*}_{\ell}(s,\mathcal{E})=\Big{(}\frac{1}{4}s^{\frac%
{1}{4}}\sigma^{2}\Big{)}^{\frac{4\alpha}{8\alpha+1}}.$$
If we omit the dependence on the level $\rho$ and constants, the
optimal testing radius is given by
$$\displaystyle\epsilon_{\text{{\tiny OPT}}}^{2}(\theta^{*};\mathcal{E})\asymp(%
\sigma^{2}s^{1/4})^{\frac{8\alpha}{1+8\alpha}}\qquad\text{and }~{}~{}m_{u},m_{%
\ell}\asymp(\sigma^{-2}s^{2\alpha})^{\frac{2}{8\alpha+1}}$$
(26)
whenever111The constraint on $s$ arises from the fact that
$t^{*}_{u}(s,\mathcal{E})\leq\sqrt{\mu_{s}}$. $s\lesssim(\sigma^{2})^{-2/(4\alpha+1)}$. As the choice of $s$ is varied,
inequality (26) actually provides us with a range
of different results.
(a)
First considering $s=1$, suppose that $\theta^{*}$ is zero at
every coordinate except the first: we then have
$$\displaystyle\epsilon_{\text{{\tiny OPT}}}^{2}(\theta^{*};\mathcal{E})\asymp(%
\sigma^{2})^{\frac{8\alpha}{1+8\alpha}},$$
(27)
where $\asymp$ means equal up to a constant that is independent of
problem parameters $d$ and $\sigma$. To the best of our
knowledge, this is a novel result on ellipse testing. It shows that
that the local minimax rate at this extremal vector is very different
from the corresponding rate at zero, which as we have established in
our Section 3.3.1 is equal $(\sigma^{2})^{\frac{4\alpha}{1+4\alpha}}$.
(b)
On the other hand, let us consider a sequence (indexed by
$\sigma$) of the form $s=(\sigma^{2})^{-\beta}$ for some $\beta\in\Big{(}0,\frac{2}{(4\alpha+1)}\Big{)}$. Over such a sequence,
inequality (26) shows that the minimax
testing radius scales as
$$\displaystyle\epsilon_{\text{{\tiny OPT}}}^{2}(\theta^{*};\mathcal{E})\asymp(%
\sigma^{2})^{\frac{2\alpha(4-\beta)}{1+8\alpha}}.$$
(28)
As $\beta$ ranges from $0$ to $\frac{2}{(4\alpha+1)}$, the testing
radius $\epsilon_{\text{{\tiny OPT}}}^{2}(\theta^{*};\mathcal{E})$ ranges from
$(\sigma^{2})^{\frac{4\alpha}{1+4\alpha}}$ to
$(\sigma^{2})^{\frac{8\alpha}{1+8\alpha}}$.
4 Proofs of main results
We now turn to the proofs of our main results, with the
proof of Theorems 1 and 2
given in Sections 4.1 and 4.2
respectively, followed by the proofs of
Corollary 1 and 2 in
Sections 4.3 and 4.4, respectively.
In all cases, we defer the proofs of certain more technical lemmas to the appendices.
4.1 Proof of Theorem 1
From our discussion following the statement of
Theorem 1, recall that the optimal linear
test (12) is based on the statistic
$$\displaystyle\psi(y):\,=\mathbb{I}\big{[}\|\Pi_{k^{*}_{u}}(y-\theta^{*})\|_{2}%
^{2}\geq\sigma^{2}(k^{*}_{u}+\sqrt{4k^{*}_{u}/\rho})\big{]},$$
(29)
where the critical upper dimension $k^{*}_{u}:\,=k_{u}(\epsilon_{u}(\theta^{*};\rho,\mathcal{E}),\theta^{*},%
\mathcal{E})$ is
computed from equation (9).
From here onwards, we adopt $\epsilon$ and $k$ as convenient
shorthands for the quantities $\epsilon_{u}$ and $k^{*}_{u}$ respectively.
We let $\Pi_{k}$ denote the projection matrix that defines the
Kolmogorov $k$-width of $\mathcal{E}_{\theta}^{*}\cap\mathbb{B}(\epsilon)$—that is
$$\displaystyle\Pi_{k}:\,=\arg\min_{\Pi_{k}\in\mathcal{P}_{k}}\max_{\Delta\in%
\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(\epsilon)}\|\Delta-\Pi_{k}\Delta\|_{2}^%
{2}.$$
(30)
Controlling the type I error:
We first show that the
test statistic (29) has type I error at most
$\rho/2$. Under the null hypothesis $\mathcal{H}_{0}$, we have $y\sim N(\theta^{*},\sigma^{2}I_{d})$. Since $\Pi_{k}$ is an orthogonal
projection matrix, the rotation invariance of the Gaussian
distribution guarantees that
$$\displaystyle T_{k}$$
$$\displaystyle:\,=\|\Pi_{k}(y-\theta^{*})\|_{2}^{2}\stackrel{{\scriptstyle d}}{%
{=}}\|\tilde{w}\|_{2}^{2},\qquad\mbox{where $\tilde{w}\sim N(0,\sigma^{2}%
\mathbb{I}_{k})$.}$$
Consequently, the statistic $T_{k}/\sigma^{2}$ follows a
$\chi_{k}^{2}$-distribution, and hence
$$\displaystyle\mathbb{P}_{0}\big{[}\psi(y)=1\big{]}$$
$$\displaystyle=\mathbb{P}_{\theta^{*}}\big{[}T_{k}\geq\sigma^{2}(k+\sqrt{4k/%
\rho})\big{]}\leq\frac{\rho}{2},$$
(31)
where the last step follows from Chebyshev’s inequality.
Controlling the type II error:
Under the alternative
hypothesis $\mathcal{H}_{1}$, we have $y\sim N(\theta,\sigma^{2}I_{d})$
for some vector $\theta\in\mathcal{E}$ with $\|\theta-\theta^{*}\|_{2}\geq\epsilon$. Since $\Pi_{k}$ is a linear operator, we have
$$\displaystyle T_{k}=\|\Pi_{k}(y-\theta^{*})\|_{2}^{2}=\|\Pi_{k}w+\Pi_{k}(%
\theta-\theta^{*})\|_{2}^{2},\qquad\mbox{where $w\sim N(0,\sigma^{2}I_{d})$.}$$
Using the fact that Gaussian distribution is rotation invariant, the
statistic $T_{k}/\sigma^{2}$ follows a non-central $\chi^{2}_{k}(c_{0})$
distribution, defined by the recentering $c_{0}=\|\Pi_{k}(\theta-\theta^{*})\|_{2}^{2}$.
For each vector $\theta\in\mathcal{E}$ such that $\|\theta-\theta^{*}\|_{2}\geq\epsilon$, we claim that
$$\displaystyle c_{0}=\|\Pi_{k}(\theta-\theta^{*})\|^{2}_{2}\geq\min_{u\in%
\mathcal{E},\|u-\theta^{*}\|_{2}=\epsilon}\{\|\Pi_{k}(u-\theta^{*})\|^{2}_{2}\}.$$
(32)
We take inequality (32) as given for the moment,
returning to prove it at the end of this section. Since $\Pi_{k}$
is an orthogonal projection onto a subspace, the Pythagorean theorem
guarantees that
$$\displaystyle\|\Pi_{k}(u-\theta^{*})\|^{2}_{2}=\|(u-\theta^{*})\|_{2}^{2}-\|(u%
-\theta^{*})-\Pi_{k}(u-\theta^{*})\|^{2}_{2}.$$
Consequently, we have
$$\displaystyle c_{0}=\|\Pi_{k}(\theta-\theta^{*})\|^{2}_{2}$$
$$\displaystyle\stackrel{{\scriptstyle(i)}}{{\geq}}\min_{u\in\mathcal{E},\|u-%
\theta^{*}\|_{2}=\epsilon}\,\left\{\|(u-\theta^{*})\|_{2}^{2}-\|(u-\theta^{*})%
-\Pi_{k}(u-\theta^{*})\|^{2}_{2}\right\}$$
$$\displaystyle=\epsilon^{2}-\max_{u\in\mathcal{E},\|u-\theta^{*}\|_{2}=\epsilon%
}\|(u-\theta^{*})-\Pi_{k}(u-\theta^{*})\|_{2}^{2}$$
$$\displaystyle\geq\epsilon^{2}-\max_{u\in\mathcal{E},\|u-\theta^{*}\|_{2}\leq%
\epsilon}\|(u-\theta^{*})-\Pi_{k}(u-\theta^{*})\|_{2}^{2},$$
where inequality (i) makes use of inequality (32).
Recalling the definition of $\Pi_{k}$ from
equation (30), we find that
$$\displaystyle c_{0}$$
$$\displaystyle~{}\geq\epsilon^{2}-\min_{\Pi_{k}\in\mathcal{P}_{k}}\max_{\theta%
\in\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(\epsilon)}\|\Delta-\Pi_{k}\Delta\|_{%
2}^{2}\;=\;\epsilon^{2}-\omega_{k}^{2}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(%
\epsilon))\stackrel{{\scriptstyle(ii)}}{{\geq}}\frac{\epsilon^{2}}{2},$$
where inequality (ii) follows since the critical dimension $k$ was
chosen so that $\omega_{k}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(\epsilon))\leq\frac{\epsilon%
}{\sqrt{2}}$.
Now we are ready to bound the type II error. For any vector $\theta$
in the alternative, we have
$$\displaystyle\mathbb{P}_{\theta}[\psi(y)=0]$$
$$\displaystyle=\mathbb{P}_{\theta}\big{[}T_{k}\leq\sigma^{2}(k+\sqrt{4k/\rho})%
\big{]}=\mathbb{P}_{\theta}\Biggr{[}\frac{T_{k}-\sigma^{2}k-c_{0}}{\sigma^{2}%
\sqrt{k}}\leq\frac{2}{\sqrt{\rho}}-\frac{c_{0}}{\sigma^{2}\sqrt{k}}\Biggr{]}.$$
Together, the two inequalities $c_{0}\geq\epsilon^{2}/2$ and $\epsilon^{2}\geq 8\rho^{-1/2}\sigma^{2}\sqrt{k}$ imply that $c_{0}/\sigma^{2}\sqrt{k}\geq 4\rho^{-1/2}$, whence
$$\displaystyle\mathbb{P}_{\theta}[\psi(y)=0]$$
$$\displaystyle\leq\mathbb{P}_{\theta}\left[\frac{T_{k}-\sigma^{2}k-c_{0}}{%
\sigma^{2}\sqrt{k}}\leq\frac{2}{\sqrt{\rho}}-\frac{4}{\sqrt{\rho}}\right]\leq%
\frac{\rho}{2},$$
where the last inequality is again due to Chebyshev’s inequality.
Therefore whenever $\epsilon^{2}\geq 8\rho^{-1/2}\sigma^{2}\sqrt{k}$,
then type II error is upper bounded by
$$\displaystyle\sup_{\theta\in\mathcal{H}_{1}}\mathbb{P}_{\theta}\big{[}\psi(y)=%
0\big{]}\leq\frac{\rho}{2}.$$
(33)
Combining inequalities (31)
and (33) yields the claim.
Proof of inequality (32):
The only remaining detail in the proof of Theorem 1 is
to establish inequality (32). It suffices to show that
the minimum
$$\displaystyle\min_{u\in\mathcal{E},\;\|u-\theta^{*}\|_{2}\geq\epsilon}\|\Pi_{k%
}(\theta-\theta^{*})\|^{2}_{2}$$
(34)
is always achieved by a vector $u^{*}\in\mathcal{E}$ with $\|u^{*}-\theta^{*}\|_{2}=\epsilon$.
Noting the continuity of the objective and the compactness of the
constraint set, Weierstrass’ theorem ensures that the minimum is
always achieved. Leting $\widehat{u}$ be any vector that achieves the
minimum, we define the new vector
$$\displaystyle u^{*}$$
$$\displaystyle:\,=(1-\alpha)\theta^{*}+\alpha\widehat{u}\qquad\mbox{where $%
\alpha:\,=\frac{\epsilon}{\|\widehat{u}-\theta^{*}\|_{2}}\in(0,1]$.}$$
The convexity of $\mathcal{E}$ ensures that $u^{*}\in\mathcal{E}$, and
moreover, by construction, we have $\|u^{*}-\theta^{*}\|_{2}=\epsilon$. Thus, the vector $u^{*}$ is feasible for the original
problem (34), and by linearity of projection, we have
$$\displaystyle\|\Pi_{k}(u^{*}-\theta^{*}\|_{2}$$
$$\displaystyle\leq(1-\alpha)\|\Pi_{k}(0)\|_{2}+\alpha\|\Pi_{k}(\widehat{u}-%
\theta^{*})\|_{2}\;\leq\;\|\Pi_{k}(\widehat{u}-\theta^{*})\|_{2},$$
showing that $u^{*}$ is also a minimizer, as claimed.
4.2 Proof of Theorem 2
We now turn to the proof of the lower bound stated in
Theorem 2. This proof, as well as subsequent ones,
makes use of the Bernstein width as a lower bound on the Kolmogorov
width. Accordingly, we begin by introducing some background on it, as
well stating an auxiliary lemma that shows how to obtain testing error
lower bounds via Bernstein width.
4.2.1 Lower bounds via the Bernstein width
The Bernstein $k$-width of a compact set $C$ is the radius of the
largest $k+1$-dimensional ball that can be inscribed into $C$.
Typically, we use the Euclidean ($\ell_{2}$) norm to define the ball,
but more generally, we can define widths respect to the $\ell_{p}$-norms
with $p\in[1,\infty]$. More precisely, for a given integer $k\geq 1$, let $\mathcal{S}_{k+1}$ denote the set of all $(k+1)$-dimensional
subspaces. With this notation, the Bernstein $k$-width in
$\ell_{p}$-norm is given by
$$\displaystyle b_{k,p}(C)$$
$$\displaystyle:\,=\arg\max_{r\geq 0}\Big{\{}\mathbb{B}^{k+1}_{p}(r)\subseteq C%
\cap S\quad\mbox{for some subspace $S\in\mathcal{S}_{k+1}$}\Big{\}},$$
(35)
where $\mathbb{B}^{k+1}_{p}(r):\,=\{u\in{}^{k+1}\mid\|u\|_{p}\leq r\}$ denotes a $(k+1)$-dimensional $\ell_{p}$-ball of radius $r$
centered at zero.
An important fact is that the Kolmogorov $k$-width from
equation (8) can always be lower bounded by the
Bernstein $k$-width in $\ell_{2}$—that is,
$$\displaystyle\omega_{k}(C)$$
$$\displaystyle\geq b_{k,2}(C)\quad\mbox{for any $k=1,2,\ldots$.}$$
(36)
See Pinkus [29] for more details of this property.
Moreover, comparing the Bernstein widths in $\ell_{2}$ and $\ell_{\infty}$
norm, we have the sandwich relation
$$\displaystyle b_{k,\infty}(C)\;\leq\;b_{k,2}(C)\;\leq\;\sqrt{k+1}\;b_{k,\infty%
}(C),$$
(37)
which is an elementary consequence of the fact that $1\leq\frac{\|u\|_{2}\;\,}{\|u\|_{\infty}}\leq\sqrt{k+1}$ for any non-zero vector
$u\in{}^{k+1}$.
We now use these definitions to define a lower bound on the testing
error in terms of Bernstein width in $\ell_{\infty}$. Define the
$\ell_{\infty}$-Bernstein lower critical dimension as
$$\displaystyle k_{B}(\epsilon,\theta^{*},\mathcal{E}):\,=\arg\max_{1\leq k\leq d%
}\Big{\{}kb^{2}_{k-1,\infty}(\mathcal{E}_{\theta^{*}})\geq\epsilon^{2}\Big{\}},$$
(38a)
and the corresponding lower critical radius
$$\displaystyle\epsilon_{B}(\theta^{*};\mathcal{E})$$
$$\displaystyle:\,=\sup\Big{\{}\epsilon\mid\epsilon\leq\frac{1}{4}\sigma^{2}%
\frac{\sqrt{k_{B}(\epsilon,\theta^{*},\mathcal{E})}}{\epsilon}\Big{\}}.$$
(38b)
In terms of these quantities, we have the following auxiliary result:
Lemma 1.
For any vector $\theta^{*}\in\mathcal{E}$, we have
$$\displaystyle\inf_{\psi}\operatorname{Err}(\psi;\theta^{*},\delta)\geq\frac{1}%
{2}\qquad\text{for all }\delta\leq\epsilon_{B}(\theta^{*};\mathcal{E}).$$
(39)
See Appendix A for the proof of this lemma.
Lemma 1 is useful because evaluating the Bernstein
width is often easier than evaluating the Kolmogorov width. Note that
we also have a (possibly weaker) lower bound in term of
$\ell_{2}$-Bernstein widths. In particular, suppose that we define the
$\ell_{2}$-Bernstein critical dimension
$$\displaystyle\widetilde{k}_{B}(\epsilon,\theta^{*},\mathcal{E}):\,=\arg\max_{1%
\leq k\leq d}\Big{\{}b^{2}_{k-1,2}(\mathcal{E}_{\theta^{*}})\geq\epsilon^{2}%
\Big{\}}$$
(40)
and then define the critical radius $\epsilon_{B}$ using $\widetilde{k}_{B}$. By
the sandwich relation (37), the lower
bound (39) still holds.
4.2.2 Main portion of proof
Let $D:\,=\max_{\theta\in\mathcal{E}}\|\theta-\theta^{*}\|_{2}$
denote the diameter of the ellipse with respect to $\theta^{*}$. We
need only consider testing radii $\epsilon\leq D$, since otherwise,
the alternative is empty. Moreover, we need only consider noise
standard deviations $\sigma\leq D$, since otherwise even the simple
(non-compound) testing problem is hard.
Next, we claim that $\epsilon_{\text{{\tiny OPT}}}^{2}(\theta^{*};\mathcal{E})\geq\sigma^{2}$.
In order to establish this claim, we consider a simple hypothesis test
between a pair $(\theta^{*},\theta)$ both in the ellipse, and such
that $\|\theta^{*}-\theta\|_{2}=\delta$ for some $\delta\in[\sigma,D]$. This single-to-single test requires separation
$\delta^{2}\geq\sigma^{2}$, a classical fact that can be proved by
various arguments, for example, by analyzing the likelihood ratio and
using Neyman-Pearson lemma. Since our composite testing
problem (3) is at least as hard, it follows that
$\epsilon_{\text{{\tiny OPT}}}^{2}(\theta^{*};\mathcal{E})\geq\sigma^{2}$.
For the remainder of our analysis, we assume that $\sqrt{k_{\ell}}\geq 18$. (Otherwise, the squared critical radius $\epsilon_{u}$ from in
equation (15)) is upper bounded by $9\sigma^{2}/2$, so
that the claim of Theorem 2 follows by adjusting the
constant $c$, combined with the lower bound $\epsilon_{\text{{\tiny OPT}}}^{2}(\theta^{*};\mathcal{E})\geq\sigma^{2}$.) We now divide our analysis into two cases,
depending on whether or not $\|\theta^{*}\|_{\mathcal{E}}\leq 1/2$.
Case 1:
First, suppose that $\|\theta^{*}\|_{\mathcal{E}}\leq 1/2$, which
implies that $\Phi(\delta)\leq(\|\theta^{*}\|_{\mathcal{E}}-1)^{2}\leq 1$. In this
case, we apply the $\ell_{2}$-Bernstein-width based lower bound from
Lemma 1 to establish our theorem. Define the
quantities $\widetilde{k}_{B}(\epsilon,\theta^{*},\mathcal{E}):\,=\arg\max\limits_{1\leq k%
\leq d}\Big{\{}b_{k-1,2}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(a\epsilon))%
\geq b\epsilon\Big{\}}$, and
$$\displaystyle\epsilon_{B}(\theta^{*};\mathcal{E})$$
$$\displaystyle:\,=\sup\Big{\{}\epsilon\mid b^{2}\epsilon\leq\frac{1}{4}\sigma^{%
2}\frac{\sqrt{\widetilde{k}_{B}(\epsilon,\theta^{*},\mathcal{E})}}{\epsilon}%
\Big{\}}.$$
With this notation, an equivalent statement of 1
is that $\inf_{\psi}\operatorname{Err}(\psi;\theta^{*},\delta)\geq\frac{1}{2}$ for all $\delta\leq b\epsilon_{B}(\theta^{*};\mathcal{E})$.
(In this statement, we have replaced $b_{k-1,2}(\mathcal{E}_{\theta^{*}})$ by $b_{k-1,2}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(a\epsilon))$ for later theoretical convenience; doing so does
not change the definition as $a>b$.)
In order to prove Theorem 2, we need to show that
$c\epsilon_{\ell}(\theta^{*};\mathcal{E})\leq b\epsilon_{B}(\theta^{*};\mathcal%
{E})$, and we claim that it suffices to show that $k_{\ell}(\epsilon,\theta^{*},\mathcal{E})\leq\widetilde{k}_{B}(\epsilon,\theta%
^{*},\mathcal{E})$. Indeed, supposing that the latter inequality holds, the
definition of the lower critical radius in
equation (15) implies that
$$\displaystyle\epsilon_{\ell}(\theta^{*};\mathcal{E})$$
$$\displaystyle:\,=\sup\Big{\{}\epsilon\mid\epsilon\leq\frac{1}{4}\sigma^{2}%
\frac{\sqrt{k_{\ell}(\epsilon,\theta^{*},\mathcal{E})}}{\epsilon}\Big{\}}.$$
Comparing with the above definition of $\epsilon_{B}(\theta^{*};\mathcal{E})$ and using222 It should also be noted that the right
hand sides of both fixed point equations are non-increasing function
of $\epsilon$ while the left hand side is non-decreasing. the fact
that $b\in(0,1)$, we see that $\epsilon_{\ell}(\theta^{*};\mathcal{E})\leq\epsilon_{B}(\theta^{*};\mathcal{E})$. Finally, since $c<b$, it follows
that $c\epsilon_{\ell}(\theta^{*};\mathcal{E})\leq b\epsilon_{B}(\theta^{*};\mathcal%
{E})$.
Now it is only left for us to show that $k_{\ell}(\epsilon,\theta^{*},\mathcal{E})\leq\widetilde{k}_{B}(\epsilon,\theta%
^{*},\mathcal{E})$. By definition of
the critical lower dimension $k_{\ell}(\epsilon,\theta^{*},\mathcal{E})$ in
equation (14), we have
$\omega_{k_{\ell}-1}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(a\epsilon))>3b\epsilon$. So our proof for Case 1 will be complete once we establish
that $b_{k_{\ell},2}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(a\epsilon))\geq b\epsilon$. In order to do so, we use the following lemma.
Lemma 2.
The following chain of inequalities hold
$$\displaystyle\min\{a\epsilon,\frac{1}{2}\sqrt{\mu_{k+1}}\}\,\stackrel{{%
\scriptstyle(i)}}{{\leq}}\,b_{k,2}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(a%
\epsilon))\leq\omega_{k}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(a\epsilon))\,%
\stackrel{{\scriptstyle(ii)}}{{\leq}}\,\frac{3}{2}\omega_{k}(\mathcal{E}\cap%
\mathbb{B}(a\epsilon))\,\stackrel{{\scriptstyle(iii)}}{{=}}\,\frac{3}{2}\min\{%
a\epsilon,\sqrt{\mu_{k+1}}\}.$$
(41)
See Appendix B.1 for the proof of this claim.
From Lemma 2, we have the string of inequalities
$$\displaystyle 3b\epsilon<\omega_{k_{\ell}-1}(\mathcal{E}_{\theta^{*}}\cap%
\mathbb{B}(a\epsilon))\leq\frac{3}{2}\min\{a\epsilon,\sqrt{\mu_{k_{\ell}}}\},$$
and hence $b_{k_{\ell},2}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(a\epsilon))\geq\min\{a%
\epsilon,\frac{1}{2}\sqrt{\mu_{k_{\ell}}}\}\geq b\epsilon$,
which completes the proof in Case 1.
Case 2:
Otherwise, we may assume that $\|\theta^{*}\|_{\mathcal{E}}>1/2$,
in which case $\Phi(\delta/c)\leq(\|\theta^{*}\|_{\mathcal{E}}-1)^{2}<1$, and
hence by definition of the function $\Phi$, we have $\delta<c\|\theta^{*}\|_{2}/a$.
Our analysis in this case makes use of the following auxiliary result
(stated as Lemma 6 in Wei et al. [35]). Suppose
that we observe a Gaussian random vector $y\sim N(\Delta,\sigma^{2}I_{d})$. For a set $C$, we use $\operatorname{Err}(\psi;\{0\},C,\delta)$ to denote the uniform testing error defined by the compound
testing problem (in the Gaussian sequence model) of distinguishing the
null $\Delta=0$ from the compound alternative $\Delta\in C\cap\mathbb{B}^{c}(\delta)$, where $\mathbb{B}^{c}(\delta):\,=\{v\mid\|v\|_{2}\geq\delta\}$.
Lemma 3.
For every set $C$ and probability measure $\mathbb{Q}$ supported on $C\cap\mathbb{B}^{c}(\delta)$, we have
$$\displaystyle\inf_{\psi}\operatorname{Err}(\psi;\{0\},C,\delta)\geq 1-\frac{1}%
{2}\sqrt{\mathbb{E}_{\eta,\eta^{\prime}}\,\exp\Big{(}\frac{\langle\eta,\,\eta^%
{\prime}\rangle}{\sigma^{2}}\Big{)}-1},$$
(42)
where $\mathbb{E}_{\eta,\eta^{\prime}}$ denotes expectation with respect to an
i.i.d pair $\eta,\eta^{\prime}\sim\mathbb{Q}$.
In order to apply Lemma 3, we need to specify a
suitable set $C$ and probability distribution $\mathbb{Q}$. For our
application, the relevant set is $C=\mathcal{E}_{\theta}^{*}$. In
addition, we need to construct an appropriate distribution $\mathbb{Q}$
with which to apply Lemma 3. To this end, we
introduce two more auxiliary results. In the following statement, we
fix a constant $\epsilon>0$, as well as pair $a,b\in(0,1)$ such
that $a>3b$, and recall that $k_{\ell}(\epsilon)$ denotes the lower
critical dimension from equation (14). Also recall that
the ellipse norm $\|\theta\|_{\mathcal{E}}^{2}=\theta^{T}M\theta$ is
defined in terms of the diagonal matrix $M$ with diagonal
entries $\{\frac{1}{\mu_{i}}\}_{i=1}^{d}$.
Lemma 4.
For any vector $\theta^{*}\in\mathcal{E}$ such that
$\|\theta^{*}\|_{2}>a\epsilon$, there exists a vector $\theta^{\dagger}\in\mathcal{E}$, a collection of $d$-dimensional orthonormal
vectors $\{u_{i}\}_{i=1}^{k_{\ell}}$ and an upper triangular matrix of the
form
$$\displaystyle H:\,=\begin{bmatrix}1&h_{3,2}&h_{4,2}&\cdots&h_{k_{\ell},2}\\
&1&h_{4,3}&\cdots&h_{k_{\ell},3}\\
&&1&\cdots&h_{k_{\ell},4}\\
&&&\ddots&\vdots\\
&&&&1\end{bmatrix}\in{}^{(k_{\ell}-1)\times(k_{\ell}-1)},$$
(43)
with ordered singular values $\nu_{1}(H)\geq\nu_{2}(H)\geq\cdots\geq\nu_{k_{\ell}-1}(H)\geq 0$ such that:
(a)
The vectors $u_{1}$, $M\theta^{\dagger}$, and $\theta^{\dagger}-\theta^{*}$ are all scalar multiples of one another.
(b)
We have $\|\theta^{\dagger}-\theta^{*}\|_{2}=a\epsilon$.
(c)
For each $i\in[k_{\ell}-1]$, the $i^{th}$ column $H_{i,\cdot}$ of the matrix $H$ has its squared Euclidean norm upper
bounded as $\|H_{\cdot,i}\|_{2}^{2}\leq a^{2}/9b^{2}$.
(d)
For each $i\in[k_{\ell}-1]$, each of the vectors $\theta^{\dagger}\pm b\epsilon\underbrace{\begin{bmatrix}u_{2}&\cdots u_{k_{%
\ell}}\end{bmatrix}}_{:\,=U}H_{\cdot,i}$ belongs to the ellipse $\mathcal{E}$.
(e)
For each pair of integers $(s,t)\in[k_{\ell}-1]\times[k_{\ell}-2]$, we have
$$\displaystyle\nu_{s}\stackrel{{\scriptstyle(i)}}{{\leq}}\frac{a}{3b}\sqrt{%
\frac{k_{\ell}-1}{s}},~{}~{}\text{ and }~{}~{}\nu_{t+1}\stackrel{{\scriptstyle%
(ii)}}{{\geq}}1-\frac{t}{k_{\ell}-1}-\sqrt{\frac{a^{2}-9b^{2}}{9b^{2}}}.$$
(44)
See Appendix B.2 for the proof of this lemma.
Let us write the SVD of the matrix $H$ from
equation (43) in the form $Q\Sigma V^{T}$. Since $H$
is square and full rank, the matrices $Q$ and $V$ are square and
orthogonal, and $\Sigma$ is a square matrix of singular values. By
construction, we have $H^{T}H=V\Sigma^{2}V^{T}$.
Define matrix $A:\,=V^{T}(UH)^{T}M(UH)V$ and write its diagonal
as $(A_{1},\ldots,A_{k_{\ell}-1}).$
We introduce index set $\mathcal{M}(r):\,=\{i\mid A_{i}\leq A_{(r)}\}$
which contains those elements that lie below the $r$-th quantile of $A$’s diagonals elements.
We assume that $(k_{\ell}-1)/8$ is an integer; otherwise, the same argument can be applied by taking its floor function.
Let $m_{1}:\,=(k_{\ell}-1)/4$ and $m_{2}:\,=(k_{\ell}-1)/2$.
Fixing a sparsity level $s\leq m_{2}-m_{1}+1$, let $\mathcal{S}$ be the set
of all subsets of
set $F:\,=\{m_{1},m_{1}+1,\ldots,m_{2}\}\cap\mathcal{M}(7(k_{\ell}-1)/8)$ of size $s$.
Note that from definitions, the cardinality of $F$ is lower bounded by $(k_{\ell}-1)/8.$
Given a subset $S\in\mathcal{S}$
and a sign vector $z^{S}=(z^{S}_{1},\ldots,z^{S}_{k_{\ell}-1})\in\{-1,0,1\}^{k_{\ell}-1}$, define the vector
$$\displaystyle\theta^{S}:\,=\theta^{*}+b\epsilon\frac{1}{4\sqrt{2s}}XVz^{S},%
\qquad\Delta^{S}:\,=\theta^{S}-\theta^{*},$$
(45)
where the matrix $X=UH$ was previously defined in
4. We then make the following claim:
Lemma 5.
Under the assumption of Lemma 4 and $\Phi(\epsilon)\leq(\|\theta^{*}\|_{\mathcal{E}}-1)^{2}$, for any subset $S\in\mathcal{S}$, there is
a sign vector $z^{S}$ supported on $S$ such that:
$$\displaystyle\theta^{S}(z^{S})$$
$$\displaystyle\;\stackrel{{\scriptstyle(i)}}{{\in}}\;\mathcal{E},\qquad\text{%
and }~{}~{}\|\Delta^{S}\|_{2}\;\stackrel{{\scriptstyle(ii)}}{{\geq}}\;c\epsilon.$$
(46)
See Appendix B.3 for the proof of this lemma.
We now apply Lemma 3 with $\mathbb{Q}$ as the
uniform distribution on those perturbation vectors $\Delta^{S}=\theta^{S}-\theta^{*}$ that are defined via
expression (45). Note that the sign vector $z^{S}$ is
taken to be the one whose existence is guaranteed by
Lemma 5. For any $S\in\mathcal{S}$, due to
inequality (46), we have $\|\Delta^{S}\|_{2}\geq\epsilon$,
and thus, with the choice $\delta:\,=c\epsilon$, the
distribution $\mathbb{Q}$ is supported on set $\mathcal{E}_{\theta^{*}}\cap\mathbb{B}^{c}(\delta)$
For simplicity of the notation, let us introduce the shorthand
notation $t:\,=m_{2}-m_{1}+1$; notat that this integer satisfies
the relations $t=k_{\ell}/4\geq 81$ and $\lambda:\,=\frac{1}{32\sigma^{2}}b^{2}\epsilon^{2}\nu^{2}_{m_{1}}$, where $\nu_{m_{1}}$
denotes the $m_{1}$-th largest eigenvalue of matrix $H$.
Lemma 6.
Given the distribution $\mathbb{Q}$ defined as above, we have
$$\displaystyle\mathbb{E}_{\eta,\eta^{\prime}\sim\mathbb{Q}}\Big{[}\exp\big{(}%
\frac{\langle\eta,\,\eta^{\prime}\rangle}{\sigma^{2}}\big{)}\Big{]}$$
$$\displaystyle\leq\exp\left(-\left(1-\frac{1}{\sqrt{t}}\right)^{2}+\exp\left(%
\frac{2+\lambda}{\sqrt{t}-1}\right)\right).$$
(47)
See Appendix B.4 for the proof of this lemma.
We can now complete the proof of Case 2. If the radius satisfies
$\epsilon^{2}\leq\frac{\sqrt{k_{\ell}}}{4}\sigma^{2}$, as a consequence of
the eigenvalue bound (44), we find that
$$\displaystyle\frac{\lambda}{\sqrt{t}}=\frac{\epsilon^{2}b^{2}\nu^{2}_{m_{1}}}{%
32\sigma^{2}\sqrt{m_{2}-m_{1}+1}}\leq\frac{\sqrt{k_{\ell}}\cdot b^{2}\cdot%
\frac{4a^{2}}{9b^{2}}}{144\sqrt{m_{2}-m_{1}+1}}\leq\frac{1}{18},$$
where the last inequality uses $a<1$ and $m_{2}-m_{1}+1=k_{\ell}/4.$
With control of these quantities in hand, we can bound the testing
error. The right hand side in expression (47) can be
upper bounded as
$$\displaystyle\exp\left(-\left(1-\frac{1}{\sqrt{t}}\right)^{2}+\exp\left(\frac{%
2}{\sqrt{t}-1}+\frac{\sqrt{t}}{\sqrt{t}-1}\frac{\lambda}{\sqrt{t}}\right)%
\right)\leq\exp\left(-\left(1-\frac{1}{9}\right)^{2}+\exp\left(\frac{1}{4}+%
\frac{9}{8}\cdot\frac{1}{18}\right)\right)<2,$$
where we use the fact that $t=k_{\ell}/4\geq 81$. As a
consequence of Lemma 3, we have
$$\displaystyle\inf_{\psi}\operatorname{Err}(\psi;\{0\},\mathcal{E}_{\theta^{*}}%
,c\epsilon)\geq 1-\frac{1}{2}\Big{[}\mathbb{E}_{\eta,\eta^{\prime}}\,\exp(%
\frac{\langle\eta,\,\eta^{\prime}\rangle}{\sigma^{2}})-1\Big{]}^{1/2}\;>\;%
\frac{1}{2}\geq\rho,$$
which completes the proof of Case 2, and hence the theorem.
4.3 Proof of Corollary 1
Our proof of Corollary 1 is based on a particular way of sandwiching the upper and
lower critical radii, which we first introduce.
4.3.1 Sandwiching the upper and lower critical radii
Recall the upper and lower critical dimensions from
equations (9) and (14), respectively.
Suppose that we can find pair a pair of functions $(\hat{k}_{\ell},\hat{k}_{u})$
such that the sandwich relation
$$\displaystyle\hat{k}_{\ell}(\epsilon)\,\leq\,k_{\ell}(\epsilon)\,\leq\,k_{u}(%
\epsilon)\leq\hat{k}_{u}(\epsilon),~{}~{}\text{ for }\epsilon>0.$$
(48)
holds. Under an additional monotonicity property, any such pair can
be used to bound the interval $[\epsilon_{\ell},\epsilon_{u}]$.
Lemma 7.
Consider a pair of functions $\hat{k}_{u}$ and $\hat{k}_{\ell}$ that are each
non-increasing in $\epsilon$, and satisfy the sandwich
relation (48). Then we have $\big{[}\epsilon_{\ell},\;\epsilon_{u}\big{]}\subseteq\big{[}\widehat{\epsilon}%
_{\ell},\;\widehat{\epsilon}_{u}\big{]}$, where
$$\displaystyle\widehat{\epsilon}_{u}:\,=\inf\left\{\epsilon\mid\epsilon\geq%
\frac{8}{\sqrt{\rho}}\sigma^{2}\frac{\sqrt{\hat{k}_{u}(\epsilon)}}{\epsilon}%
\right\}\quad\mbox{and}\quad\widehat{\epsilon}_{\ell}:\,=\sup\left\{\epsilon%
\mid\epsilon\leq\frac{1}{4}\sigma^{2}\frac{\sqrt{\hat{k}_{\ell}(\epsilon)}}{%
\epsilon}\right\}.$$
Proof.
These bounds follow directly by replacing $k_{u}$ by $\hat{k}_{u}$
(respectively $k_{\ell}$ by $\hat{k}_{\ell}$) in the definition of $\widehat{\epsilon}_{u}$
(respectively $\widehat{\epsilon}_{\ell}$). The non-increasing nature of the functions
$(\hat{k}_{\ell},\hat{k}_{u})$ ensure that the quantities $(\widehat{\epsilon}_{\ell},\widehat{\epsilon}_{u})$ are
well-defined.
∎
4.3.2 Proof of upper bound
We prove the upper bound via an application of Lemma 7
combined with Theorem 1. In particular, we construct
functions333Strictly speaking, these functions also depend on
$\theta^{*}$ and the ellipse $\mathcal{E}$, but we omit this dependence
so as to simplify notation. $(f_{k}^{\ell},f_{k}^{u})$ that sandwich
local Kolmogorov width above and below by
$$\displaystyle f_{k}^{\ell}(\epsilon)~{}\leq~{}\omega_{k}(\mathcal{E}_{\theta^{%
*}}\cap\mathbb{B}(a\epsilon))\leq\omega_{k}(\mathcal{E}_{\theta^{*}}\cap%
\mathbb{B}(\epsilon))~{}\leq f_{k}^{u}(\epsilon)\qquad\mbox{for every integer %
$1\leq k\leq d$.}$$
(49)
Here $a\in(0,1)$ is the constant involved in the statement of
Theorem 2. It is then straightforward to verify
that the functions
$$\displaystyle\hat{k}_{u}(\epsilon):\,=\arg\min_{1\leq k\leq d}\big{\{}f_{k}^{u%
}(\epsilon)\leq\frac{\epsilon}{\sqrt{2}}\big{\}}~{}~{}\text{ and
}~{}~{}\hat{k}_{\ell}(\epsilon):\,=\arg\min_{1\leq k\leq d}\big{\{}f_{k}^{\ell%
}(\epsilon)\leq 3b\epsilon\big{\}}.$$
(50)
satisfy the conditions of Lemma 7, so that it can be
applied.
We begin by upper bounding the localized Kolmogorov
width (8) of $\mathcal{E}\cap\mathbb{B}(\epsilon)$. Let
$\Pi_{k}$ denote the orthogonal projection onto the first $k$
coordinates of the ellipse. We then have
$$\displaystyle\omega^{2}_{k}(\mathcal{E}_{\theta}^{*}\cap\mathbb{B}(\epsilon))%
\leq\max_{\theta\in\mathcal{E},\|\theta-\theta^{*}\|_{2}\leq\epsilon}\|(\theta%
-\theta^{*})-\Pi_{k}(\theta-\theta^{*})\|^{2}_{2}=\max_{\theta\in\mathcal{E},%
\|\theta\|_{2}\leq\epsilon}\sum_{i=k+1}^{d}\theta_{i}^{2}.$$
Since the scalars $\{\mu_{i}\}_{i=1}^{d}$ are ranked in
non-increasing order, for any vector $\theta\in\mathcal{E}$, we have
$$\displaystyle\sum_{i=k+1}^{d}\theta_{i}^{2}=\sum_{i=k+1}^{d}\mu_{i}\frac{%
\theta_{i}^{2}}{\mu_{i}}$$
$$\displaystyle\;\leq\;\mu_{k+1}\sum_{i=k+1}^{d}\frac{\theta_{i}^{2}}{\mu_{i}}\;%
\leq\;\mu_{k+1},$$
where the final inequality follows from the inclusion $\theta\in\mathcal{E}$. Consequently, we have proved that
$$\displaystyle\omega_{k}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(\epsilon))$$
$$\displaystyle\leq f_{k}^{u}(\epsilon)\;:\,=\;\underbrace{\min\{\epsilon,\sqrt{%
\mu_{k+1}}}_{=:f_{k}^{u}(\epsilon)}\}.$$
It is easy to check that the function $\epsilon\mapsto\hat{k}_{u}(\epsilon)$ from equation (50) is non-increasing, so
that an application of Lemma 7 yields the claimed upper
bound (20).
4.3.3 Proof of lower bound
We prove the lower bound by a combination of Lemma 7 and
Theorem 2, and in order to do so, we need to lower
bound the localized Kolmogorov width. As previously noted, the
Bernstein width (35) is always a lower bound on the
Kolmogorov width, so that
$$\displaystyle\omega_{k}(\mathcal{E}\cap\mathbb{B}(a\epsilon))\geq b_{k,2}(%
\mathcal{E}\cap\mathbb{B}(a\epsilon))=\underbrace{\min\{a\epsilon,\sqrt{\mu_{k%
+1}}\}}_{=:f_{k}^{\ell}(\epsilon)}$$
where the last inequality follows from the fact that we can always
inscribe a Euclidean ball of radius $\sqrt{\mu_{k+1}}$ centered at
zero inside the ellipse $\mathcal{E}$ truncated to its first
$(k+1)$-coordinates. It is straightforward to verify that the
corresponding $\hat{k}_{\ell}$ that is defined in (50) is
non-increasing in $\epsilon$, so that the the claimed lower
bound (20) follows by applying Lemma 7 with
Theorem 2.
4.4 Proof of Corollary 2
We again separate our proof into two parts, corresponding to the upper
bound and lower bounds. The upper bound is proved via an application
of Lemma 7 and Theorem 1. On the other
hand, we prove the lower bound by using Theorem 2
and Lemma 1 in conjunction with Lemma 7.
4.4.1 Proof of upper bound
Let us introduce the convenient shorthand $\theta^{*}_{-s}:\,=(\theta^{*}_{1},\ldots,\theta^{*}_{s-1},\theta^{*}_{s+1},%
\ldots,\theta^{*}_{d})$. In terms of the
the local minimax testing radius $\epsilon_{\text{{\tiny OPT}}}(\theta^{*};\mathcal{E})$
previously defined in equation (5),
Theorem 1 guarantees that $\epsilon_{\text{{\tiny OPT}}}(\theta^{*})\leq\epsilon_{u}(\theta^{*};\mathcal{%
E})$. Consequently, in order to establish
the upper bound stated in Corollary 2, it suffices to
show that for any vector $\theta^{*}$ with $\theta^{*}_{s}\geq\sqrt{\mu_{s}}-t^{*}_{u}(s,\mathcal{E})$ and $\theta^{*}_{-s}=0$, we have
$\epsilon_{u}(\theta^{*};\mathcal{E})\leq t^{*}_{u}(s,\mathcal{E})$.
In order to do so, it suffices to show that
$$\displaystyle\omega_{m_{u}(\delta,s)}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(%
\delta))\leq\frac{\delta}{\sqrt{2}}\qquad\mbox{for each $\delta>0$.}$$
(51)
It then follows that $m_{u}(\delta;s)\geq k_{u}(\delta)$, from
which an application of Lemma 7 implies that
$\epsilon_{u}(s,\mathcal{E})\leq t^{*}_{u}(s,\mathcal{E})$, as
desired.
Accordingly, the remainder of our argument is devoted to establishing
inequality (51). Let $\Pi_{m}$ denote the
orthogonal projection onto the span of the first $m$ standard
basis vectors $\{e_{i}\}_{i=1}^{m}$. By
definition (8), the localized Kolmogorov width is
upper bounded by
$$\displaystyle\omega_{m}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(\delta))\leq%
\underbrace{\max_{\theta\in C}\sqrt{\sum_{i=m+1}^{d}\theta_{i}^{2}}}_{=\,:T_{m%
}^{u}}\qquad\mbox{where}\quad C:\,=\big{\{}\theta\in{}^{d}\mid\sum_{i=1}^{d}%
\frac{\theta_{i}^{2}}{\mu_{i}}\leq 1,~{}~{}(\theta_{s}-\theta^{*}_{s})^{2}+%
\sum_{i\neq s}\theta_{i}^{2}\leq\delta^{2}\big{\}},$$
where we have used the fact that $\theta^{*}_{-s}=0.$
Let us introduce the shorthand $m\backslash s:\,=\{1,2,\ldots,s-1,s+1,\ldots,m\}$. Observe that setting $\theta_{j}\neq 0$ for any $j\in m\backslash s$ only makes the constraints defining
$C$ more difficult to satisfy, and does not improve the objective
function defining $T_{m}^{u}$. Therefore, we can reduce the problem to
$$\displaystyle\max_{\theta_{s},\{\theta_{i}\}_{i=m+1}^{d}}\sum_{i=m+1}^{d}%
\theta_{i}^{2},\quad\mbox{such that}\quad\sum_{i=m+1}^{d}\frac{\theta^{2}_{i}}%
{\mu_{i}}\leq 1\text{ and
}(\theta_{s}-\theta^{*}_{s})^{2}+\sum_{i=m+1}^{d}\theta^{2}_{i}\leq\delta^{2}.$$
Since the scalars $\{\mu_{i}\}_{i=m+1}^{d}$ are ranked in
non-increasing order, any optimal solution must satisfy $\theta_{i}=0$
for all $i\in\{m+2,\ldots,d\}$. Putting pieces together,
the optimal solution is defined by a pair $(\theta_{s},\theta_{m+1})$ that satisfy the relations
$$\displaystyle(\theta_{s}-\theta^{*}_{s})^{2}+\theta_{m+1}^{2}$$
$$\displaystyle=\delta^{2}~{}~{}\text{ and }~{}~{}\frac{\theta^{2}_{s}}{\mu_{s}}%
+\frac{\theta^{2}_{m+1}}{\mu_{m+1}}=1.$$
Solving these equations yields
$$\displaystyle\theta_{s}=\frac{\theta^{*}_{s}}{1-t}-\sqrt{\frac{\delta^{2}-\mu_%
{m+1}}{1-t}+\frac{t(\theta^{*}_{s})^{2}}{(1-t)^{2}}},\quad\mbox{and}\quad%
\theta_{m+1}^{2}=\delta^{2}-\left(\sqrt{\frac{\delta^{2}-\mu_{m+1}}{1-t}+\frac%
{t(\theta^{*}_{s})^{2}}{(1-t)^{2}}}-\frac{t\theta^{*}_{s}}{1-t}\right)^{2},$$
(52)
where we define $t:\,=\mu_{m+1}/\mu_{s}.$ We introduce the
following lemma.
Lemma 8.
For any $\delta\geq t^{*}_{u}(s,\mathcal{E})$ and $\sqrt{\mu_{s}}-t^{*}_{u}(s,\mathcal{E})\leq\theta^{*}_{s}\leq\sqrt{\mu_{s}}$, we have
$\theta_{m+1}^{2}\leq\frac{\delta^{2}}{2}$.
See Appendix C for the proof of this result.
Therefore, the localized Kolmogorov width is at most
$\omega^{2}_{m}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(\delta))\leq(T_{m}^{u})^%
{2}=\theta_{m+1}^{2}\leq\frac{\delta^{2}}{2}$, which
completes the proof of the upper bound stated in
Corollary 2.
4.4.2 Proof of lower bound
In terms of the the local minimax testing radius $\epsilon_{\text{{\tiny OPT}}}(\theta^{*};\mathcal{E})$ previously defined in
equation (5), Lemma 1
guarantees that $\epsilon_{\text{{\tiny OPT}}}(\theta^{*})\geq\epsilon_{B}(\theta^{*};\mathcal{%
E})$. Therefore, in order to establish the lower bound stated in
Corollary 2, it is sufficient for us to show that for
any vector $\theta^{*}$ with $\theta^{*}_{s}\leq\sqrt{\mu_{s}}-t^{*}_{\ell}(s,\mathcal{E})$ and $\theta^{*}_{-s}=0$, we have
$\epsilon_{B}(\theta^{*};\mathcal{E})\geq t^{*}_{\ell}(s,\mathcal{E})$.
Suppose that we can show that
$$\displaystyle mb^{2}_{m-1,\infty}(\mathcal{E}_{\theta^{*}})\geq\delta^{2},$$
(53)
where we have introduced the shorthand $m:\,=m_{\ell}(\delta;s)$. It then follows that then $k_{B}(\delta)\geq m$, from which
an application of Lemma 7 guarantees that
$\epsilon_{B}(\theta^{*};\mathcal{E})\geq t^{*}_{\ell}(s,\mathcal{E})$.
Here we apply Lemma 7 with respect to $k_{B}$ and $\epsilon_{B}$
replacing quantities $k_{\ell}$ and $\epsilon_{\ell}$ in the statement
which holds by the same argument of the original proof.
The remainder of our argument is to devoted to proving
inequality (53). Consider the set
$$\displaystyle\mathcal{H}$$
$$\displaystyle:\,=\left\{\theta\mid\theta_{s}=\theta^{*}_{s},~{}\theta_{m+1},%
\ldots,\theta_{d}=0,~{}\theta_{i}\in[-\frac{\delta}{\sqrt{|\mathcal{M}|}},%
\frac{\delta}{\sqrt{|\mathcal{M}|}}],~{}i\in\mathcal{M}\right\},~{}~{}\mathcal%
{M}:\,=\{1\leq i\leq m,i\neq s\}.$$
We claim that the set $\mathcal{H}$ is contained within the ellipse
$\mathcal{E}$. In order to see this, first note that $\mu_{i}$ are ranked in
a non-increasing order, so we have $\frac{\delta^{2}}{|\mathcal{M}|}\sum_{i\neq s}^{m}\frac{1}{\mu_{i}}\leq\delta^%
{2}/\mu_{m}$. It implies
$$\displaystyle\|\theta\|_{\mathcal{E}}^{2}=\frac{(\theta^{*}_{s})^{2}}{\mu_{s}}%
+\sum_{i\neq s}^{m}\frac{\delta^{2}}{\mu_{i}|\mathcal{M}|}\leq\frac{(\theta^{*%
}_{s})^{2}}{\mu_{s}}+\frac{\delta^{2}}{\mu_{m}}=1-\frac{2w}{\sqrt{\mu_{s}}}+%
\frac{w^{2}}{\mu_{s}}+\frac{\delta^{2}}{\mu_{m}},$$
(54)
where we plugged in $\theta^{*}_{s}=\sqrt{\mu_{s}}-w$ with $w\geq t^{*}_{\ell}(s,\mathcal{E}).$ On one hand, inequality $w\leq\sqrt{\mu_{s}}$
guarantees that $\frac{w^{2}}{\mu_{s}}-\frac{w}{\sqrt{\mu_{s}}}<0.$ On the
other hand, inequalities $\mu_{k}^{2}\geq\delta^{2}\mu_{s}$ and $\delta\leq t^{*}_{\ell}(s,\mathcal{E})$ yield that $\frac{\delta^{2}}{\mu_{k}}\leq\frac{\delta}{\sqrt{\mu_{s}}}\leq\frac{w}{\sqrt{%
\mu_{s}}}$. Combining
these two facts, we have $\|\theta\|_{\mathcal{E}}^{2}\leq 1$ for every $\theta\in\mathcal{H}$, namely set $\mathcal{H}\subset\mathcal{E}$.
It means that we are able to inscribe a $|\mathcal{M}|-1$-dimensional
$\ell_{\infty}$ ball with radius $\delta/\sqrt{|\mathcal{M}|}$ into
the ellipse $\mathcal{E}$, namely $mb^{2}_{m-1,\infty}(\mathcal{E}_{\theta^{*}})\geq\delta^{2}$. Putting pieces
together, we finish the proof of the lower bound of
Corollary 2.
5 Discussion
In this paper, we have studied the local geometry of compound testing
problems in ellipses. Our main contribution was to show that it is
possible to obtain a sharp characterization of the localized minimax
testing radius at a given vector $\theta^{*}$ in terms of an equation
involving the Kolmogorov width of a particular set. This set involves
the geometry of the ellipse, but is localized to the vector
$\theta^{*}$ via intersection with a Euclidean ball. This form of
localization and the resulting fixed point equation is similar to
localization techniques that are used in proving optimal rates for
estimation (e.g., [33]), and our work demonstrates that
similar concepts are also useful in establishing optimal testing
rates. We anticipate that this perspective will be useful in studying
adaptivity and local geometry in other testing problems.
Acknowledgments
This work was partially supported by Office of Naval Research grant
DOD ONR-N00014, and National Science Foundation grant NSF-DMS-1612948
to MJW.
Appendix A Proof of Lemma 1
In order to simplify notation, we use $k$ as a shorthand for $k_{B}$
throughout this proof. Our goal is to prove that if $\epsilon^{2}\leq\frac{\sqrt{k}\sigma^{2}}{4}$, then any test has testing error at least
$1/2$. Let $e_{i}\in{}^{d}$ denote the standard basis vector
with the $i$-th coordinate equal to one. The definition of $k$
guarantees that we can construct a $2^{k}$-sized collection of
perturbation vectors of the form
$$\displaystyle\theta_{b}=\theta^{*}+\frac{\epsilon}{\sqrt{k}}\sum_{i=1}^{k}b_{i%
}e_{i},~{}~{}~{}b\in\{-1,+1\}^{k}$$
(55)
such that each $\theta_{b}$ lies on the boundary of the $\ell_{\infty}$
ball around $\theta^{*}$ with radius $\frac{\epsilon}{\sqrt{k}}$.
This last property ensures that each $\theta_{b}$ belongs to the ellipse
$\mathcal{E}$. At the same time, we have $\|\theta_{b}-\theta^{*}\|^{2}=k(\frac{\epsilon}{\sqrt{k}})^{2}=\epsilon^{2}$,
valid for all Boolean vectors $b\in\{-1,+1\}^{k}$, so that each
perturbation vector $\theta_{b}$ indexes a distribution in the
alternative hypothesis class $\mathcal{H}_{1}$.
By Lemma 3, we can lower bound the testing error
as
$$\displaystyle\inf_{\psi}\operatorname{Err}(\psi;\{\theta^{*}\},\mathcal{E},%
\epsilon)=\inf_{\psi}\operatorname{Err}(\psi;\{0\},\mathcal{E}_{\theta^{*}},%
\epsilon)\geq 1-\frac{1}{2}\sqrt{\mathbb{E}_{\eta,\eta^{\prime}\sim\mathbb{Q}}%
\,\exp(\frac{\langle\eta,\,\eta^{\prime}\rangle}{\sigma^{2}})-1}.$$
Let $\mathbb{Q}$ denote the uniform distribution over the set $\{\theta_{b}-\theta^{*},b\in\{-1,+1\}^{k}\}$. Introducing the shorthand
$N:\,=2^{k}$, we have
$$\displaystyle\mathbb{E}_{\eta,\eta^{\prime}\sim\mathbb{Q}}\,\exp(\frac{\langle%
\eta,\,\eta^{\prime}\rangle}{\sigma^{2}})\;=\;\frac{1}{N^{2}}\sum_{b,b^{\prime%
}}\exp(\frac{(\theta_{b}-\theta^{*})^{T}(\theta_{b^{\prime}}-\theta^{*})}{%
\sigma^{2}})$$
$$\displaystyle=\frac{1}{N}\sum_{b}\exp(\frac{(\theta_{b}-\theta^{*})^{T}(%
\epsilon\text{1}_{d}/\sqrt{k})}{\sigma^{2}})$$
(56)
$$\displaystyle=2^{-k}\sum_{i=0}^{k}\binom{k}{i}\exp(\frac{\epsilon^{2}(k-2i)}{k%
\sigma^{2}})$$
$$\displaystyle=\Big{(}\frac{\exp(\epsilon^{2}/k\sigma^{2})+\exp(-\epsilon^{2}/k%
\sigma^{2})}{2}\Big{)}^{k},$$
where we have used the symmetry of the problem. It can be verified
via elementary calculation that $\frac{\exp(x)+\exp(-x)}{2}\leq 1+x^{2}$ for all $|x|\leq 1/2$. Setting $x=\frac{\epsilon^{2}}{k\sigma^{2}}$, we have $x\leq\frac{1}{4\sqrt{k}}<1/2$, and hence
$$\displaystyle\frac{1}{N^{2}}\sum_{b,b^{\prime}}\exp(\frac{(\theta_{b}-\theta^{%
*})^{T}(\theta_{b^{\prime}}-\theta^{*})}{\sigma^{2}})\leq(1+\frac{\epsilon^{4}%
}{k^{2}\sigma^{4}})^{k}\leq\exp(\frac{\epsilon^{4}}{k\sigma^{4}}),$$
where in the last step we used the standard bound $1+x\leq e^{x}$.
Therefore, the testing error is lower bounded by $1-\frac{1}{2}\sqrt{e^{\epsilon^{4}/k\sigma^{4}}-1}$. Thus, whenever $\epsilon^{2}\leq\sqrt{k}\sigma^{2}/4$, the testing error is lower bounded as
$\inf_{\psi}\operatorname{Err}(\psi;\{\theta^{*}\},\mathcal{E},\delta)\geq 1-%
\frac{1}{2}\sqrt{e^{1/16}-1}\geq\frac{1}{2}$, which completes the
proof of Lemma 1.
Appendix B Proofs for Theorem 2
In this appendix, we collect the proofs of various auxiliary results involved
in the proof of Theorem 2.
B.1 Proof of Lemma 2
Note that the inequality $b_{k,2}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(a\epsilon))\leq\omega_{k}(%
\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(a\epsilon))$ follows as an immediate consequence of the relation
between widths. It remains to prove inequalities (i)-(iii).
Proof of inequalities (i) and (iii):
Recalling the definition of the Bernstein width, we claim that
$\omega_{k}((\mathcal{E}\cap\mathbb{B}(a\epsilon)))=b_{k,2}(\mathcal{E}\cap%
\mathbb{B}(a\epsilon))=\min\{a\epsilon,\sqrt{\mu_{k+1}}\}$. In order
to prove this claim, it suffices to show that
$$\displaystyle\omega_{k}(\mathcal{E}\cap\mathbb{B}(a\epsilon))\stackrel{{%
\scriptstyle(i)}}{{\leq}}\min\{a\epsilon,\sqrt{\mu_{k+1}}\},\quad\mbox{and}%
\quad b_{k,2}(\mathcal{E}\cap\mathbb{B}(a\epsilon))\stackrel{{\scriptstyle(ii)%
}}{{\geq}}\min\{a\epsilon,\sqrt{\mu_{k+1}}\}.$$
Inequality (i) follows easily by direct calculation after taking the
$k$-dimensional projection in the definition of Kolmogorov width to be
projecting to the span of $\{e_{1},\ldots,e_{k}\}$. To show the second
part, note that any vector $v\in{}^{d}$ with $\|v\|_{2}\leq\sqrt{\mu_{k+1}}$ and $v_{i}=0$ for all $i=k+2,\ldots,d$
satisfies
$$\displaystyle\|v\|_{\mathcal{E}}^{2}=\sum_{i=1}^{d}\frac{v_{i}^{2}}{\mu_{i}}=%
\sum_{i=1}^{k+1}\frac{v_{i}^{2}}{\mu_{i}}\,\stackrel{{\scriptstyle(a)}}{{\leq}%
}\,\frac{1}{\mu_{k+1}}\sum_{i=1}^{k+1}v_{i}^{2}\stackrel{{\scriptstyle(b)}}{{%
\leq}}1,$$
where inequality (a) follows from the non-increasing ordering of the
sequence $\{\mu_{i}\}_{i=1}^{d}$ and inequality (b) follows from
the structure of vector $v.$ Consequently, the ellipse $\mathcal{E}$
contains a $k+1$-dimensional $\ell_{2}$ ball centered at $0$ of radius
$\sqrt{\mu_{k+1}}$. Therefore we have $b_{k,2}(\mathcal{E}\cap\mathbb{B}(a\epsilon))\geq\min\{a\epsilon,\sqrt{\mu_{k+%
1}}\}$. We
complete the proof of inequality (iii).
Consider every vector $v\in\mathcal{E}$ that lies in the
$k+1$-dimensional $\ell_{2}$ ball that specified above. Since
$\|\theta^{*}\|_{\mathcal{E}}\leq 1/2$, we have $\|2\theta^{*}\|_{\mathcal{E}}\leq 1$ which implies
that $2\theta^{*}\in\mathcal{E}.$ By the convexity of $\mathcal{E}$, we have $\theta^{*}+v/2\in\mathcal{E}$. It means that $b_{k,2}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(a\epsilon))\geq\min\{a\epsilon,%
\frac{1}{2}\sqrt{\mu_{k+1}}\}$, which completes the proof of
inequality (i).
Proof of inequality (ii):
Again consider projecting to the span of $\{e_{1},\ldots,e_{k}\}$,
therefore the Kolmogorov width can be controlled as
$$\displaystyle\omega_{k}^{2}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(a\epsilon))%
\leq(a\epsilon)^{2}\wedge\max\Big{\{}\sum_{i=k+1}^{d}\Delta_{i}^{2}~{}\mid~{}%
\|\theta^{*}+\Delta\|_{\mathcal{E}}\leq 1,~{}\|\Delta\|_{2}\leq a\epsilon\Big{%
\}}.$$
By the triangle inequality, we obtain $\|\Delta\|_{\mathcal{E}}\leq 1+\|\theta^{*}\|_{\mathcal{E}}\leq 3/2$ where the last inequality uses $\|\theta^{*}\|_{\mathcal{E}}\leq 1/2.$ Letting $\Delta_{k+1}=\sqrt{\mu_{k+1}}$ and $\Delta_{i}=0$ for
$i\neq k+1$, we have $\omega_{k}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(a\epsilon))\leq\min\{a%
\epsilon,3/2\sqrt{\mu_{k+1}}\}$ which
thus proves the inequality (ii).
B.2 Proof of 4
We break the proof of Lemma 4 into two parts. In the
first part, we construct the vector $\theta^{\dagger}$ and the collection
$\{u_{i}\}_{i=1}^{k_{\ell}}$, and show that the properties (a)–(d) hold.
In the second part, we show that matrix $H$ satisfies the
eigenvalue property.
B.2.1 Part I
Recall that the ellipse norm is defined as $\|\theta\|_{\mathcal{E}}^{2}=\sum_{i=1}^{d}\frac{\theta_{i}^{2}}{\mu_{i}}$, so that $\theta\in\mathcal{E}$ is equivalent to $\|\theta\|_{\mathcal{E}}\leq 1$. Recall that
$M$ denotes a diagonal matrix with diagonal entries
$(1/\mu_{1},\ldots,1/\mu_{d})$.
Constructing $\theta^{\dagger}$:
Let us first define a vector $\theta^{\dagger}\in\mathcal{E}$ that satisfies the
assumptions in Lemma 4. Define the function $\psi:(0,\infty)\rightarrow(0,\infty)$ via
$$\displaystyle\psi(r)$$
$$\displaystyle:\,=r^{2}\|M(\mathbb{I}_{d}+rM)^{-1}\theta^{*}\|_{2}^{2}=r^{2}%
\sum_{i=1}^{d}\frac{1}{(1+r/\mu_{i})^{2}}\frac{(\theta^{*}_{i})^{2}}{\mu_{i}^{%
2}}=\sum_{i=1}^{d}\frac{r^{2}}{(\mu_{i}+r)^{2}}(\theta^{*}_{i})^{2}.$$
Note that $\psi$ is a continuous and non-decreasing function on
$[0,\infty)$ such that $\psi(0)=0$ and $\lim_{r\rightarrow+\infty}\psi(r)=\|\theta^{*}\|$. Since $a\in(0,1)$, there must
exists some $r>0$ such that $\psi(r)=a\epsilon$. Given this
choice of $r$, we then define $\theta^{\dagger}:\,=(\mathbb{I}_{d}+rM)^{-1}\theta^{*}$. Since $r>0$, our definition ensures
that $\theta^{\dagger}\in\mathcal{E}$: more precisely, we have
$$\displaystyle\|\theta^{\dagger}\|_{\mathcal{E}}^{2}=\sum_{i=1}^{d}\frac{1}{(1+%
r/\mu_{i})^{2}}\frac{(\theta^{*}_{i})^{2}}{\mu_{i}}<\|\theta^{*}\|_{\mathcal{E%
}}^{2}\leq 1.$$
Now we are ready to construct orthogonal unit vectors
$u_{1},\ldots,u_{k_{\ell}}$ such that proper perturbations of $\theta^{\dagger}$
towards linear combinations of those directions still lie in the set
$\mathcal{E}.$ We do this in a sequential way.
Constructing the vector $u_{1}$.
Defining the vector $v_{1}:\,=\theta^{*}-\theta^{\dagger}$, the definition
of $\theta^{\dagger}$ guarantees that $v_{1}$ is parallel to both $\theta^{\dagger}-\theta^{*}$ and $M\theta^{\dagger}$, so that condition 1 of the lemma
holds. Note also that $\|\theta^{\dagger}-v_{1}\|_{\mathcal{E}}^{2}=\|\theta^{\dagger}+v_{1}\|_{%
\mathcal{E}}^{2}-4(\theta^{\dagger})^{T}Mv_{1}\leq 1$ because $\theta^{\dagger}+v_{1}=\theta^{*}\in\mathcal{E}$ and $(\theta^{\dagger})^{T}Mv_{1}=r\|M\theta^{\dagger}\|_{2}^{2}\geq 0$. These properties guarantee the
inclusion $\theta^{\dagger}\pm v_{1}\in\mathcal{E}$, whence by convexity, the
ellipse $\mathcal{E}$ contains the line segment connecting the two points
$\theta^{\dagger}\pm v_{1}$. We let $u_{1}$ be the normalized version of
$v_{1}$—namely, $u_{1}:\,=v_{1}/\|v_{1}\|_{2}$.
Constructing $u_{2}$.
Without loss of generality, we can assume that $k_{\ell}\geq 2$. By
definition of $k_{\ell}$, there exists a vector $\Delta\in\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(a\epsilon)$ satisfying $\|\Delta-\Pi_{v_{1}}(\Delta)\|_{2}\geq 3b\epsilon$. Accordingly, we may define
$$\displaystyle\Delta_{2}$$
$$\displaystyle:\,=\arg\max_{\begin{subarray}{c}\Delta\in\mathcal{E}_{\theta^{*}%
}\cap\mathbb{B}(a\epsilon):\\
\|\Delta-\Pi_{v_{1}}(\Delta)\|_{2}\geq 3b\epsilon\end{subarray}}\|\Delta-\Pi_{%
v_{1}}(\Delta)\|_{\mathcal{E}}.$$
(57)
With this choice, we set $v_{2}:\,=\Delta_{2}-\Pi_{v_{1}}(\Delta_{2})$
and $u_{2}:\,=v_{2}/\|v_{2}\|_{2}$. Note that these choices ensure that
$\|v_{2}\|_{2}\geq 3b\epsilon$ and $v_{2}\perp v_{1}$.
In order to complete the proof, it suffices to show that $\theta^{\dagger}\pm\frac{1}{3}v_{2}\in\mathcal{E}$. Indeed, when this inclusion holds,
then we have $\theta^{\dagger}\pm b\epsilon\begin{bmatrix}u_{2}&\cdots u_{k_{\ell}}\end{%
bmatrix}H_{\cdot,1}=\theta^{\dagger}\pm b\epsilon u_{2}\in\mathcal{E}.$ In order to show that the required
inclusion holds, we begin by noting that since $\mathcal{E}$ is a convex
set containing $\theta^{\dagger}-v_{1}$ and $\theta^{*}+\Delta_{2}$, it
contains the segment connecting these two points. In particular, we
have
$$\displaystyle\theta^{\dagger}+\frac{\|(\theta^{\dagger}-v_{1})-\theta^{\dagger%
}\|_{2}}{\|(\theta^{\dagger}-v_{1})-(\theta^{*}+\Pi_{v_{1}}(\Delta_{2}))\|_{2}%
}v_{2}=\theta^{\dagger}+\frac{\|v_{1}\|_{2}}{\|-2v_{1}-\Pi_{v_{1}}(\Delta_{2})%
\|_{2}}v_{2}\in\mathcal{E},$$
a result that follows from the proportional segments
theorem [13], when we consider a line
passing through $\theta^{\dagger}$ that is parallel to $v_{2}$. See also
Figure 3 for an illustration.
Since $\|\Delta_{2}\|_{2}\leq a\epsilon\leq\|v_{1}\|_{2}$ and $\|\Pi_{v_{1}}(\Delta_{2})\|_{2}\leq\|\Delta_{2}\|_{2}$, we thus have $\|-2v_{1}-\Pi_{v_{1}}(\Delta_{2})\|_{2}\leq 3\|v_{1}\|_{2}$ and it implies that $\theta^{\dagger}+\frac{1}{3}v_{2}\in\mathcal{E}$.
Finally, using the fact that $\theta^{\dagger}+\frac{1}{3}v_{2}\in\mathcal{E}$
and $v_{2}^{T}M\theta^{\dagger}=0$, we have
$$\displaystyle\|\theta^{\dagger}-\frac{1}{3}v_{2}\|_{\mathcal{E}}^{2}=\|\theta^%
{\dagger}+\frac{1}{3}v_{2}\|_{\mathcal{E}}^{2}-\frac{2}{3}v_{2}^{T}M\theta^{%
\dagger}\leq 1,$$
which completes the proof.
Constructing the remaining $u_{i}$.
We sequentially construct pairs $(v_{i},v^{\prime}_{i})$ for $i=3,\ldots,k_{\ell}$ as
follows and $u_{i}$ is a just a scaled version of $v_{i}$.
Given $v_{1},\ldots,v_{i-1}$ for $i\leq k_{\ell}$, by definition of $k_{\ell}$,
there still exists
some $\Delta_{i}\in\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(a\epsilon)$ such
that $\|\Delta_{i}-\Pi_{v_{1},\ldots,v_{i-1}}(\Delta_{i})\|_{2}\geq 3b\epsilon$. We then define
$$\displaystyle v_{i}:\,=\Delta_{i}-\Pi_{v_{1},\ldots,v_{i-1}}(\Delta_{i}),\quad%
\mbox{and}\quad v^{\prime}_{i}:\,=\Delta_{i}-\Pi_{v_{1}}(\Delta_{i}),$$
which ensures that $\|v_{i}\|_{2}\geq 3b\epsilon$ and
both $v_{i}$ and $v_{i}^{\prime}$ are perpendicular to $v_{1}.$
Finally, we set $u_{i}:\,=v_{i}/\|v_{i}\|_{2}$.
We now claim that
$$\displaystyle\theta^{\dagger}\pm\frac{1}{3}v_{i}^{\prime}\in\mathcal{E}.$$
(58)
Taking this claim as given for the moment, let us first
establish property (c) and (d).
Observe that the vectors
vectors $v_{1},\ldots,v_{i}$ are orthogonal, so we can write $v^{\prime}_{i}$ as the
linear combination of $v_{2},\ldots,v_{i}$. For $i\geq 3$, there exist
scalars $r_{i,j}$ such that
$$\displaystyle v^{\prime}_{i}=h_{i,2}v_{2}+\cdots+h_{i,i-1}v_{i-1}+v_{i},$$
and by orthogonality of those $v_{i}$, we have
$$\displaystyle\|v^{\prime}_{i}-v_{i}\|_{2}^{2}=h_{i,1}^{2}\|v_{2}\|_{2}^{2}+%
\cdots+h_{i,i-1}^{2}\|v_{i-1}\|_{2}^{2}.$$
For $i\geq 3$, since $(v^{\prime}_{i}-v_{i})\perp v_{i}$, it is further guaranteed that
$$\displaystyle\|v^{\prime}_{i}-v_{i}\|_{2}^{2}$$
$$\displaystyle=\|v^{\prime}_{i}\|_{2}^{2}-\|v_{i}\|_{2}^{2}\leq(a^{2}-9b^{2})%
\epsilon^{2}$$
where the last inequality follows from fact $v^{\prime}_{i}\in\mathbb{B}(a\epsilon)$ and the definition of fact that $\|v_{i}\|_{2}\geq 3b\epsilon$. Putting the pieces together, we have
$$\displaystyle h_{i,1}^{2}+\cdots+h_{i,i-1}^{2}\leq\frac{(a^{2}-9b^{2})\epsilon%
^{2}}{9b^{2}\epsilon^{2}}=\frac{a^{2}-9b^{2}}{9b^{2}}.$$
This proves property property (c) where $\|H_{i,\cdot}\|_{2}^{2}\leq 1+(a^{2}-9b^{2})/9b^{2}$. Thus by relation (58), for $i\geq 3$
we have shown
$$\displaystyle\theta^{\dagger}\pm b\epsilon x_{i-1}=\theta^{\dagger}\pm b%
\epsilon(h_{i,2}u_{2}+\cdots+h_{i,i-1}u_{i-1}+u_{i})\in\mathcal{E},$$
which completes the proof of property (d).
It is only left to prove relation (58). To see this
fact, first note $\|v_{i}\|_{2}\leq\|v^{\prime}_{i}\|_{2}$ due to Pythagoras
theorem. Also, by the maximality of $v_{2}$ in inequality
(57), note that $\|v^{\prime}_{i}\|_{\mathcal{E}}\leq\|v_{2}\|_{\mathcal{E}}$. Thus,
$$\displaystyle\|\theta^{\dagger}+\frac{1}{3}v^{\prime}_{i}\|_{\mathcal{E}}^{2}=%
\|\theta^{\dagger}\|_{\mathcal{E}}^{2}+\frac{1}{9}\|v^{\prime}_{i}\|_{\mathcal%
{E}}^{2}\leq\|\theta^{\dagger}\|_{\mathcal{E}}^{2}+\frac{1}{9}\|v_{2}\|_{%
\mathcal{E}}^{2}=\|\theta^{\dagger}+\frac{1}{3}v_{2}\|_{\mathcal{E}}^{2}\leq 1,$$
where we have used the fact that $v^{\prime}_{i}$ and $v_{2}$ are perpendicular to
$v_{1}=rM\theta^{\dagger}$ to ignore the cross terms, as well as the
fact that $\theta^{\dagger}+\frac{1}{3}v_{2}\in\mathcal{E}$. Thus $\theta^{\dagger}+\frac{1}{3}v^{\prime}_{i}\in\mathcal{E}$. We can also show $\theta^{\dagger}-\frac{1}{3}v^{\prime}_{i}\in\mathcal{E}$ similarly as before
since $\theta^{\dagger}Mv_{i}^{\prime}=0.$
B.2.2 Part II
We deal with inequalities (i) and (ii) separately.
For any $s\in[k_{\ell}-1]$, we have
$$\displaystyle s\nu_{s}^{2}\leq\sum_{i=1}^{k_{\ell}-1}\nu_{i}^{2}=\mbox{trace}(%
H^{T}H)\leq\sum_{i=1}^{k_{\ell}-1}\|H_{i,\cdot}\|_{2}^{2}\leq(k_{\ell}-1)\frac%
{a^{2}}{9b^{2}},$$
where the last inequality uses property (c).
It implies that $s\nu_{s}^{2}\leq(k_{\ell}-1)\frac{a^{2}}{9b^{2}}$ which proves inequality (i).
For any $t\in[k_{\ell}-2]$,
we claim that for any $t$-dimensional subspace $W\subseteq{}^{m}$, there exists some $\ell$, such that standard basis vector $e_{\ell}$ satisfies that
$\|e_{\ell}-\Pi_{W}(e_{\ell})\|_{2}^{2}\geq 1-\frac{t}{m}$. In order
to prove this claim, we take an orthonormal basis $z_{1},\ldots,z_{t}$ of
$W$ and extend it to an orthonormal basis $z_{1},\ldots,z_{m}$ for
${}^{m}$. We then have
$$\displaystyle\sum_{i=1}^{m}\|e_{i}-\Pi_{W}(e_{i})\|_{2}^{2}=\sum_{i=1}^{m}\sum%
_{j=t+1}^{m}\langle e_{i},z_{j}\rangle^{2}=\sum_{j=t+1}^{m}\sum_{i=1}^{m}%
\langle e_{i},z_{j}\rangle^{2}=m-t,$$
where the last equality is due to the fact that $e_{i},1\leq i\leq m$
forms a standard basis of ${}^{m}$, and $z_{j}$ is a unit vector. So the claim
holds by the pigeonhole principle.
Now let $H^{T}=\sum_{i=1}^{k_{\ell}-1}\sigma_{i}u_{i}v_{i}^{T}$ with
$\sigma_{1}\geq\cdots\geq\sigma_{k_{\ell}-1}$ be the SVD of $H^{T}$.
Take $W$ be the span of the first $t$ singular vectors
$v_{1},\ldots,v_{t}$. By taking dimension $m=k_{\ell}-1$, the above claim implies there
exists some $e_{\ell}$ such that $\|e_{\ell}-\Pi_{W}(e_{\ell})\|_{2}^{2}\geq 1-\frac{t}{k_{\ell}-1}$. Let $v:\,=(h_{\ell,2},\ldots,h_{\ell,\ell-1},0,\ldots,0)$ so that the $\ell-1$
row of $H^{T}$ can be written as $H^{T}_{\cdot,\ell-1}=e_{\ell}^{T}+v$. Note by definition of matrix $H$, we have $\|v\|_{2}^{2}\leq a^{2}-9b^{2}$; thus
$$\displaystyle\|v-\Pi_{W}(v)\|_{2}\leq\|v\|_{2}\leq\sqrt{a^{2}-9b^{2}}/3b.$$
Applying the triangle inequality yields
$$\displaystyle\|H_{\cdot,\ell-1}-\Pi_{W}(H_{\cdot,\ell-1})\|_{2}\geq\|e_{\ell}-%
\Pi_{W}(e_{\ell})\|_{2}-\|v-\Pi_{W}(v)\|_{2}\geq 1-\frac{t}{k_{\ell}-1}-\frac{%
\sqrt{a^{2}-9b^{2}}}{3b}.$$
(59)
Define the matrix $\tilde{H}:\,=\sum_{i=1}^{t}\sigma_{i}u_{i}v_{i}^{T}$, note that the rows of $\tilde{H}$ are the projections of
the rows of $H^{T}$ onto $W$, and therefore, the definition
of the operator norm implies that
$$\displaystyle\|H^{T}-\tilde{H}\|_{op}\geq\|H_{\cdot,\ell-1}-\Pi_{W}(H_{\cdot,%
\ell-1})\|_{2}.$$
It is easy to check that $\|H^{T}-\tilde{H}\|_{op}=\sigma_{t+1}$. Combining with inequality (59), we have
$$\displaystyle\sigma_{t+1}\geq 1-\frac{t}{k_{\ell}-1}-\frac{\sqrt{a^{2}-9b^{2}}%
}{3b},$$
which completes the proof of inequality (ii).
Combining our results from Parts I and II concludes the
proof of Lemma 4.
B.3 Proof of Lemma 5
We break our proof into two parts, corresponding to the two claims in
the lemma.
Proof of inequality (i):
This inequality is proved via a probabilistic argument. Recall that
the ellipse norm is defined as $\|x\|_{\mathcal{E}}^{2}=\sum_{i=1}^{d}\frac{x_{i}^{2}}{\mu_{i}}$, and that we write $x\in\mathcal{E}$ to mean to
$\|x\|_{\mathcal{E}}\leq 1$. We use $M$ to denote the diagonal matrix
with entries $(1/\mu_{1},\ldots,1/\mu_{d})$.
Let us define auxiliary vector
$$\displaystyle\tilde{\theta}^{S}:\,=\theta^{\dagger}+\frac{b\epsilon}{2\sqrt{2s%
}}XVz^{S},$$
(60)
where matrix $X:\,=UH$ is defined in Lemma 4
and matrix $H$ has eigen-decomposition $Q\Sigma V^{T}$.
We claim that the following two facts:
(a)
There exists a sign vector $z^{S}\in\{-1,0,1\}^{k_{\ell}-1}$
such that $\tilde{\theta}^{S}\in\mathcal{E}$.
(b)
Defining $v_{1}:\,=\theta^{*}-\theta^{\dagger}$, we then have
$\theta^{*}+v_{1}\in\mathcal{E}$.
Taking these two facts as given for now, we can prove inequality (i).
Noticing that $\theta^{\dagger}+v_{1}=\theta^{*}\in\mathcal{E}$ and $\theta^{\dagger}+2v_{1}=\theta^{*}+v_{1}\in\mathcal{E}$, by convexity of set $\mathcal{E}$,
if $\theta^{\dagger}+u\in\mathcal{E}$ for some vector $u$ then
$$\displaystyle\theta^{*}+\frac{1}{2}u=\theta^{\dagger}+v_{1}+\frac{1}{2}u=\frac%
{1}{2}(\theta^{\dagger}+u)+\frac{1}{2}(\theta^{\dagger}+2v_{1})\in\mathcal{E}.$$
Letting $u:\,=b\epsilon\frac{1}{\sqrt{s}}XVz^{S}$, then
$\theta^{S}:\,=\theta^{*}+\frac{1}{2}b\epsilon\frac{1}{\sqrt{s}}XVz^{S}=\theta^%
{*}+\frac{1}{2}u$. Then we have $\theta^{S}\in\mathcal{E}$.
It remains to establish claims (a) and (b) stated above.
Proof of claim (a):
Let $z^{S}\in\{-1,+1\}^{|S|}$ be a vector of i.i.d. Rademacher random
variables. Expanding the square yields
$$\displaystyle\|\tilde{\theta}^{S}\|_{\mathcal{E}}^{2}$$
$$\displaystyle=(\theta^{\dagger})^{T}M\theta^{\dagger}+\frac{b\epsilon}{\sqrt{2%
s}}(\theta^{\dagger})^{T}MXVz^{S}+\frac{b^{2}\epsilon^{2}}{8s}(XVz^{S})^{T}%
MXVz^{S}$$
$$\displaystyle=(\theta^{\dagger})^{T}M\theta^{\dagger}+b^{2}\epsilon^{2}\frac{1%
}{8s}(XVz^{S})^{T}MXVz^{S},$$
where the last inequality is due to the fact that $(\theta^{\dagger})^{T}MX=0.$ Let us consider the expectation $\mathbb{E}\|\tilde{\theta}^{S}\|_{\mathcal{E}}^{2}$. Note
that
$$\displaystyle\mathbb{E}(XVz^{S})^{T}MXVz^{S}$$
$$\displaystyle=\mathbb{E}~{}\mbox{trace}[(XVz^{S})^{T}MXVz^{S}]\;=\;\mathbb{E}~%
{}\mbox{trace}[V^{T}X^{T}MXVz^{S}(z^{S})^{T}],$$
where the last inequality uses the property of the trace function. By
linearity, we can switch the order of the expectation and trace
function, and doing so yields
$$\displaystyle\mathbb{E}~{}\mbox{trace}[V^{T}X^{T}MXVz^{S}(z^{S})^{T}]=\mbox{%
trace}[\underbrace{V^{T}X^{T}MXV}_{:\,=A}\mathbb{E}[z^{S}(z^{S})^{T}]]=\sum_{i%
\in S}A_{i}.$$
We claim $\sum_{i\in S}A_{i}\leq 8s\max_{i}\{x_{i}^{T}Mx_{i}\}$.
In order to show this, first note that
$$\displaystyle\mbox{trace}[A]=\mbox{trace}[V^{T}X^{T}MXV]=\mbox{trace}[X^{T}MX]%
=\sum_{i=1}^{k_{\ell}-1}x_{i}^{T}Mx_{i},$$
using the fact that $VV^{T}=\mathbb{I}$.
Since $S$ is the subset of $F:\,=\{m_{1},m_{1}+1,\ldots,m_{2}\}\cap\mathcal{M}(7(k_{\ell}-1)/8)$, so for $t:\,=\frac{7(k_{\ell}-1)}{8}$, we have
$$\displaystyle((k_{\ell}-1)-t)A_{(t)}\leq\sum_{i=t}^{k_{\ell}-1}A_{(i)}\leq%
\mbox{trace}[A]\leq(k_{\ell}-1)\max_{i}\{x_{i}^{T}Mx_{i}\}.$$
It implies that $A_{(t)}\leq 8\max_{i}\{x_{i}^{T}Mx_{i}\}$ which further verifies our claim.
Putting the pieces together, we have
$$\displaystyle\mathbb{E}\|\tilde{\theta}^{S}\|_{\mathcal{E}}^{2}$$
$$\displaystyle=(\theta^{\dagger})^{T}M\theta^{\dagger}+b^{2}\epsilon^{2}\frac{1%
}{8s}\mathbb{E}(XVz^{S})^{T}MXVz^{S}$$
$$\displaystyle\stackrel{{\scriptstyle(i)}}{{\leq}}(\theta^{\dagger})^{T}M\theta%
^{\dagger}+b^{2}\epsilon^{2}\frac{1}{8s}8s\max_{i}\{x_{i}^{T}Mx_{i}\}\stackrel%
{{\scriptstyle(ii)}}{{=}}\|\theta^{\dagger}+b\epsilon x_{j}\|_{\mathcal{E}}^{2%
}\stackrel{{\scriptstyle(iii)}}{{\leq}}1,$$
where step (i) uses claim on the $A_{i}$s, step (ii) uses fact that
$\theta^{\dagger}Mx_{j}=0$ and
step (iii) uses the fact that $\theta^{\dagger}+b\epsilon x_{j}\in\mathcal{E}$. Since the average squared ellipse norm is at most one,
there must exist at least one $\tilde{\theta}^{S}$ with
$\|\tilde{\theta}^{S}\|_{\mathcal{E}}^{2}\leq 1$, as claimed.
Proof of claim (b):
Since $v_{1}=\theta^{*}-\theta^{\dagger}$, the results of
Lemma 4 guarantee that
$$\displaystyle\|v_{1}\|_{\mathcal{E}}^{2}=\sum_{i=1}^{d}\left(\frac{r(\epsilon)%
}{r(\epsilon)+\mu_{i}}\right)^{2}(\theta^{*}_{i})^{2}$$
$$\displaystyle=\sum_{i=1}^{d}\frac{r(\epsilon)^{2}\mu_{i}}{(r(\epsilon)+\mu_{i}%
)^{2}}\frac{(\theta^{*}_{i})^{2}}{\mu_{i}}$$
$$\displaystyle\leq\|\theta^{*}\|_{\mathcal{E}}^{2}\;\max_{i=1,\ldots,d}\left\{%
\frac{r(\epsilon)^{2}\mu_{i}}{(r(\epsilon)+\mu_{i})^{2}}\right\}.$$
Since $(r(\epsilon)+\mu_{i})^{2}\geq 4r(\epsilon)\mu_{i}$, we have
$\|v_{1}\|_{\mathcal{E}}\leq\frac{\sqrt{r(\epsilon)}}{2}\|\theta^{*}\|_{%
\mathcal{E}}$. Triangle
inequality guarantees that
$$\displaystyle\|\theta^{*}+v_{1}\|_{\mathcal{E}}\leq\|\theta^{*}\|_{\mathcal{E}%
}+\|v_{1}\|_{\mathcal{E}}\leq\left(1+\frac{\sqrt{r(\epsilon)}}{2}\right)\|%
\theta^{*}\|_{\mathcal{E}}\leq 1,$$
where the last inequality uses assumption that
$r(\epsilon)=\Phi(\epsilon)<(\|\theta^{*}\|_{\mathcal{E}}^{-1}-1)^{2}$
and $\|\theta^{*}\|_{\mathcal{E}}\leq 1.$
Proof of inequality (ii):
Letting $\tilde{\Delta}^{S}:\,=\tilde{\theta}^{S}-\theta^{\dagger}$ and
by construction $\Delta^{S}=\frac{1}{2}\tilde{\Delta}^{S}.$
By definition, $\|\tilde{\Delta}\|_{2}^{2}=\frac{1}{8s}b^{2}\epsilon^{2}\|U\Sigma Vz^{S}\|_{2}%
^{2}.$ Substituting matrix $H$ as its
decomposition $Q\Sigma V^{T}$ yields
$$\displaystyle\|\tilde{\theta}^{S}-\theta^{\dagger}\|_{2}^{2}$$
$$\displaystyle=\frac{1}{8s}b^{2}\epsilon^{2}\|UQ\Sigma z^{S}\|_{2}^{2}=\frac{1}%
{8s}b^{2}\epsilon^{2}\|\Sigma z^{S}\|_{2}^{2},$$
(61)
where the last inequality uses the orthogonality of matrices $U,Q$.
Recall that set $S$ is a subset of $\{m_{1},\ldots,m_{2}\}$. Using
eigenvalue bound (44)(ii) from part (e) of
Lemma 4, we have $\|\Sigma z^{S}\|_{2}^{2}\geq(1-\frac{m_{2}}{t}-\frac{\sqrt{a^{2}-9b^{2}}}{3b})%
^{2}\|z^{S}\|_{2}^{2}$. Combining with equality (61) yields
inequality (ii).
Finally, putting the two parts together completes the proof
of Lemma 5.
B.4 Proof of Lemma 6
Let us denote the cardinality of $\mathcal{S}$ as $N:\,=\binom{m_{2}-m_{1}+1}{s}.$ Then the expectation can be written as
$$\displaystyle\mathbb{E}\exp\big{(}\frac{\langle\eta,\,\eta^{\prime}\rangle}{%
\sigma^{2}}\big{)}=\frac{1}{N^{2}}\sum_{S,S^{\prime}\in\mathcal{S}}\exp\big{(}%
\frac{1}{\sigma^{2}}\langle\Delta^{S},\,\Delta^{S^{\prime}}\rangle\big{)}.$$
From the definition (45) of $\theta^{S}$, we have
$$\displaystyle\langle\Delta^{S},\,\Delta^{S^{\prime}}\rangle=\frac{1}{32}b^{2}%
\epsilon^{2}\frac{1}{s}(z^{S})^{T}V^{T}H^{T}U^{T}UHVz^{S^{\prime}}=\frac{1}{32%
}b^{2}\epsilon^{2}\frac{1}{s}(z^{S})^{T}\Sigma^{2}z^{S^{\prime}},$$
where we used the relation $H^{T}H=V\Sigma^{2}V^{T}$, and the
orthogonality of the matrices $V$ and $U$. In order to further
control the right hand side, note that
$$\displaystyle(z^{S})^{T}\Sigma^{2}z^{S^{\prime}}=\sum_{i\in S\cap S^{\prime}}%
\nu_{i}^{2}z^{S}_{i}z^{S^{\prime}}_{i}\leq\nu^{2}_{m_{1}}\sum_{i\in S\cap S^{%
\prime}}|z^{S}_{i}z^{S^{\prime}}_{i}|=\nu^{2}_{m_{1}}|S\cap S^{\prime}|.$$
Combining these three inequalities above yields
$$\displaystyle\mathbb{E}\exp\big{(}\frac{\langle\eta,\,\eta^{\prime}\rangle}{%
\sigma^{2}}\big{)}\leq\frac{1}{N^{2}}\sum_{S,S^{\prime}\in\mathcal{S}}\exp\big%
{(}\frac{1}{32\sigma^{2}}b^{2}\epsilon^{2}\nu^{2}_{m_{1}}\frac{|S\cap S^{%
\prime}|}{s}\big{)}.$$
Let us now further compute the right hand side above. Recall that we
denote $t:\,=m_{2}-m_{1}+1$. Note that intersection cardinality
$|S\cap S^{\prime}|$ takes values in $\{0,1,\ldots,s\}$. Given every set
$S$ and integer $i\in\{0,1,\ldots,s\}$, the number of $S^{\prime}$ such
that $|S\cap S^{\prime}|=i$ equals to ${s\choose i}{t-s\choose s-i}$.
Consequently, if we let $\lambda:\,=\frac{1}{32\sigma^{2}}b^{2}\epsilon^{2}\nu^{2}_{m_{1}}$, we obtain
$$\displaystyle\mathbb{E}\exp\big{(}\frac{\langle\eta,\,\eta^{\prime}\rangle}{%
\sigma^{2}}\big{)}={t\choose s}^{-1}\sum_{i=0}^{s}{s\choose i}{t-s\choose s-i}%
e^{\lambda i/s}=\sum_{i=0}^{s}\frac{G_{i}z^{i}}{i!},$$
(62)
where $z:\,=e^{\lambda/s}$ and $G_{i}:\,=\frac{(s!(t-s)!)^{2}}{((s-i)!)^{2}t!(t-2s+i)!}$.
Let us set integer $s:\,=\lfloor\sqrt{t}\rfloor$. We claim
quantity $G_{i}$ satisfies the following bound
$$\displaystyle G_{i}\leq\exp\Big{(}-(1-\frac{1}{\sqrt{t}})^{2}+\frac{2i}{\sqrt{%
t}-1}\Big{)}\qquad\mbox{for all $i\in\{0,1,\ldots,s\}$.}$$
(63)
Taking expression (63) as given for now and plugging
into inequality (62), we have
$$\displaystyle\mathbb{E}\exp(\frac{\langle\eta,\,\eta^{\prime}\rangle}{\sigma^{%
2}})\leq\exp\Big{(}-(1-\frac{1}{\sqrt{t}})^{2}\Big{)}\sum_{i=0}^{s}\frac{(z%
\exp(\frac{2}{\sqrt{t}-1}))^{i}}{i!}$$
$$\displaystyle\,\stackrel{{\scriptstyle(a)}}{{\leq}}\,\exp\Big{(}-(1-\frac{1}{%
\sqrt{t}})^{2}\Big{)}\exp\left(z\exp(\frac{2}{\sqrt{t}-1})\right)$$
$$\displaystyle\,\stackrel{{\scriptstyle(b)}}{{\leq}}\,\exp\left(-\left(1-\frac{%
1}{\sqrt{t}}\right)^{2}+\exp\left(\frac{2+\lambda}{\sqrt{t}-1}\right)\right),$$
where step (a) follows from the standard power series expansion $e^{x}=\sum_{i=0}^{\infty}\frac{x^{i}}{i!}$ and step (b) follows by $z=e^{\lambda/s}$ and $s=\lfloor\sqrt{t}\rfloor>\sqrt{t}-1$. We have thus established
inequality (47).
It only remains to check inequality (63) for $G_{i}$.
Using the fact that $1-x\leq e^{-x}$, we have
$$\displaystyle G_{0}=\frac{((t-s)!)^{2}}{t!(t-2s)!}$$
$$\displaystyle=(1-\frac{s}{t})(1-\frac{s}{t-1})\cdots(1-\frac{s}{t-s+1})\leq%
\exp(-s\sum_{i=1}^{s}\frac{1}{t-s+i}).$$
(64)
Recalling that $s:\,=\lfloor\sqrt{t}\rfloor$, then we can
bound the sum in expression (64) as
$$\displaystyle s\sum_{i=1}^{s}\frac{1}{t-s+i}\geq s\sum_{i=1}^{s}\frac{1}{t}=%
\frac{s^{2}}{t}\geq(1-\frac{1}{\sqrt{t}})^{2},$$
which, when combined with inequality (64), implies that
$G_{0}\leq\exp(-(1-\frac{1}{\sqrt{t}})^{2})$.
Moreover, direct calculations yield that we have
$$\displaystyle\frac{G_{i}}{G_{i-1}}=\frac{(s-i+1)^{2}}{t-2s+i}\;\leq\;\frac{G_{%
1}}{G_{0}}\quad\mbox{for all $i=1,\ldots,s$,}$$
(65)
where the last inequality follows from the fact that
$\frac{(s-i+1)^{2}}{t-2s+i}$ is non-increasing with index $i$.
Therefore, recalling that $s=\lfloor\sqrt{t}\rfloor$, we
have
$$\displaystyle\frac{G_{i}}{G_{i-1}}\leq\frac{G_{1}}{G_{0}}\;\leq\;\frac{t}{t-2%
\sqrt{t}+1}=(1+\frac{1}{\sqrt{t}-1})^{2}\leq\exp(\frac{2}{\sqrt{t}-1}),$$
where the last inequality follows from $1+x\leq e^{x}.$ Putting pieces
together validates bound (63) thus finishing the proof
of inequality (47).
Appendix C Proof of Lemma 8
Letting $\Delta:\,=\sqrt{\frac{\delta^{2}-\mu_{m+1}}{1-t}+\frac{t(\theta^{*}_{s})^{2}}{%
(1-t)^{2}}}-\frac{t\theta^{*}_{s}}{1-t}$, we then have
$\theta_{m+1}^{2}=\delta^{2}-\Delta^{2}$. Direct calculations yield
$$\displaystyle\Delta=\frac{\frac{\delta^{2}-\mu_{m+1}}{1-t}+\frac{t(\theta^{*}_%
{s})^{2}}{(1-t)^{2}}-\frac{t^{2}(\theta^{*}_{s})^{2}}{(1-t)^{2}}}{\sqrt{\frac{%
\delta^{2}-\mu_{m+1}}{1-t}+\frac{t(\theta^{*}_{s})^{2}}{(1-t)^{2}}}+\frac{t%
\theta^{*}_{s}}{1-t}}=\frac{\frac{\delta^{2}-\mu_{m+1}}{1-t}+\frac{t(\theta^{*%
}_{s})^{2}}{1-t}}{\sqrt{\frac{\delta^{2}-\mu_{m+1}}{1-t}+\frac{t(\theta^{*}_{s%
})^{2}}{(1-t)^{2}}}+\frac{t\theta^{*}_{s}}{1-t}}.$$
Re-organizing the terms yields
$$\displaystyle\Delta$$
$$\displaystyle=\frac{\delta^{2}-\mu_{m+1}+t(\theta^{*}_{s})^{2}}{t\theta^{*}_{s%
}+\sqrt{(\delta^{2}-\mu_{m+1})(1-t)+t(\theta^{*}_{s})^{2}}}=\delta\Big{(}\frac%
{1+\frac{t(\theta^{*}_{s})^{2}-\mu_{m+1}}{\delta^{2}}}{\frac{t\theta^{*}_{s}}{%
\delta}+\sqrt{1-t+\frac{t\mu_{m+1}}{\delta^{2}}+\frac{t(\theta^{*}_{s})^{2}-%
\mu_{m+1}}{\delta^{2}}}}\Big{)}.$$
Now let us first analyze the denominator. Since $t$ is defined as
$\frac{\mu_{m+1}}{\mu_{s}}$, we obtain
$\frac{t\mu_{m+1}}{\delta^{2}}=\frac{\mu^{2}_{m+1}}{\delta^{2}\mu_{s}}$; by $\theta^{*}_{s}\leq\sqrt{\mu_{s}}$,
we obtain that $\frac{t\theta^{*}_{s}}{\delta}\leq\frac{t\sqrt{\mu_{s}}}{\delta}\leq\frac{\mu_%
{m+1}}{\delta\sqrt{\mu_{s}}}$. Plugging into the expression of $\Delta$ gives
$$\displaystyle\frac{\Delta}{\delta}\geq\frac{1+\frac{t(\theta^{*}_{s})^{2}-\mu_%
{m+1}}{\delta^{2}}}{\frac{\mu_{m+1}}{\delta\sqrt{\mu_{s}}}+\sqrt{1+\frac{\mu^{%
2}_{m+1}}{\delta^{2}\mu_{s}}+\frac{t(\theta^{*}_{s})^{2}-\mu_{m+1}}{\delta^{2}%
}-t}}.$$
Moreover, since $m$ is chosen as the maximum index that satisfies
$\mu_{k}^{2}\geq\frac{1}{64}\delta^{2}\mu_{s}$, it is guaranteed that
$\mu_{m+1}^{2}\leq\frac{1}{64}\delta^{2}\mu_{s}$ which further implies
that
$$\displaystyle\frac{\Delta}{\delta}\geq\frac{1+\frac{t(\theta^{*}_{s})^{2}-\mu_%
{m+1}}{\delta^{2}}}{\frac{1}{8}+\sqrt{1+\frac{1}{64}+\frac{t(\theta^{*}_{s})^{%
2}-\mu_{m+1}}{\delta^{2}}-t}}.$$
(66)
Now, in order to control $\frac{\Delta}{\delta}$, we only need to control
quantity $\frac{t(\theta^{*}_{s})^{2}-\mu_{m+1}}{\delta^{2}}$. Recall that
for any $\delta>t^{*}_{u}(s,\mathcal{E})$, we only consider those
$\theta^{*}$ satisfying $\theta^{*}_{s}\geq\sqrt{\mu_{s}}-t^{*}_{u}(s,\mathcal{E})$.
Let us write $\theta^{*}_{s}=\sqrt{\mu_{s}}-c\delta$ with $c\in[0,1]$, then
$$\displaystyle\frac{t(\theta^{*}_{s})^{2}-\mu_{m+1}}{\delta^{2}}=t\frac{(\sqrt{%
\mu_{s}}-c\delta)^{2}-\mu_{s}}{\delta^{2}}=\underbrace{-2c\frac{\mu_{m+1}}{%
\delta\sqrt{\mu_{s}}}}_{:\,=x}+c^{2}t.$$
Here $x:\,=-2c\frac{\mu_{m+1}}{\delta\sqrt{\mu_{s}}}\geq-2\frac{\mu_{m+1}}{\delta%
\sqrt{\mu_{s}}}\geq-\frac{1}{4}$ since
$c\leq 1$ and $\mu_{m+1}^{2}\leq\frac{1}{64}\delta^{2}\mu_{s}$. Now
we are ready to substitute this expression into
inequality (66). Therefore it is guaranteed that
$$\displaystyle\frac{\Delta}{\delta}\geq\frac{1+x+c^{2}t}{\frac{1}{8}+\sqrt{%
\frac{65}{64}+x+c^{2}t-t}}\geq\frac{1+x}{\frac{1}{8}+\sqrt{\frac{65}{64}+x}}%
\geq\min_{x\geq-1/4}\Big{\{}\frac{1+x}{\frac{1}{8}+\sqrt{\frac{65}{64}+x}}\Big%
{\}}=\frac{3}{4}.$$
Consequently, we have
$$\displaystyle\theta^{2}_{m+1}=\delta^{2}-\Delta^{2}\leq\delta^{2}-(\frac{3%
\delta}{4})^{2}\leq\frac{1}{2}\delta^{2}.$$
Appendix D Auxiliary results on the critical radii
In this appendix, we collect a number of auxiliary results concerning
the upper and low critical inequalities, as defined in
equations (10) and (15), respectively.
D.1 Existence and uniqueness
First, let us show that there exists a unique and strictly positive
solution to inequalities (10) and (15).
Let us prove this claim for the solution of
inequality (10), since the argument for
inequality (15) is entirely analogous. It suffices to
show that the function $g(t):\,=k_{u}(t,\theta^{*},\mathcal{E})$ is
non-increasing on the positive real line. Since the function $t\mapsto 1/t$ is strictly decreasing on the positive real line, we are
then guaranteed that the function $t\mapsto\sqrt{g(t)}/t=\sqrt{k_{u}(t,\theta^{*},\mathcal{E})}/t$ is strictly decreasing on the
positive real line. This property ensures that
inequality (10) has a unique and strictly positive
solution: the left-side function $t$ is strictly increasing, whereas
the right-side function is strictly decreasing.
It remains to show that $g(t)=k_{u}(t,\theta^{*},\mathcal{E})$ is
non-increasing. Equivalently, we need to show that for any $t>0$,
we have
$$\displaystyle g(ct)\;=\;k_{u}(ct,\theta^{*},\mathcal{E})\geq k_{u}(t,\theta^{*%
},\mathcal{E})\;=\;g(t)\qquad\mbox{for all $c\in(0,1]$.}$$
By linearity of the Kolmogorov width, we have
$$\displaystyle k_{u}(t,\theta^{*},\mathcal{E}):\,=\arg\min_{1\leq k\leq d}\left%
\{\omega_{k}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(t))\leq\frac{1}{2}t\right%
\}=\arg\min_{1\leq k\leq d}\left\{\omega_{k}(c(\mathcal{E}_{\theta^{*}}\cap%
\mathbb{B}(t)))\leq\frac{1}{2}c\,t\right\}.$$
(67)
Given any vector $v\in\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(t)$, convexity
ensures that $cv\in\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(ct)$, which is
equivalent to the containment $c\,\big{(}\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(t)\big{)}\subset\mathcal{E}_{%
\theta^{*}}\cap\mathbb{B}(ct)$. This
containment implies that
$$\displaystyle\omega_{k}(c(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(t)))\leq%
\omega_{k}(\mathcal{E}_{\theta^{*}}\cap\mathbb{B}(ct)).$$
Combined with our earlier inequality (67), we
conclude that $k_{u}(t,\theta^{*},\mathcal{E})\leq k_{u}(ct,\theta^{*},\mathcal{E})$, as desired.
D.2 Well-definedness of the function $\Phi$
In this appendix, we verify that the function $\Phi$ from
equation (16) is well-defined. In order to provide
intuition, Figure 4 provides an illustration of $\Phi$.
We begin with the case when $\delta<\|\theta^{*}\|_{2}/a$. For notation
simplicity, let $r(\delta):\,=\min\Big{\{}r\geq 0\,\mid a^{2}\delta^{2}\leq\sum_{i=1}^{d}\frac{%
r^{2}}{(r+\mu_{i})^{2}}(\theta^{*}_{i})^{2}\Big{\}}.$ Since each $\mu_{i}\geq 0$, function $f:r\to\sum_{i=1}^{d}\frac{r^{2}}{(r+\mu_{i})^{2}}(\theta^{*}_{i})^{2}$ is
non-decreasing in $r$. It is easy to check that
$$\displaystyle\lim_{r\to 0^{+}}f(r)=0,\qquad\text{ and }\lim_{r\to\infty}f(r)=%
\|\theta^{*}\|_{2}^{2}.$$
Then quantity $r(\delta)$ is uniquely defined and positive whenever
$\delta<\|\theta^{*}\|_{2}/a$. Note that as $\delta\to\frac{\|\theta^{*}\|_{2}}{a}$, $a^{2}\delta^{2}\to\|\theta^{*}\|_{2}^{2}$ therefore
$r(\delta)\to\infty.$
It is worth noticing that given any $\theta^{*}$ where $\|\theta^{*}\|_{2}$ does
not depend on $\delta$, $r$ goes to zero when $\delta\to 0$, namely
$\lim_{\delta\to 0^{+}}\Phi(\delta)=0.$
References
[1]
S. Atapattu, C. Tellambura, and H. Jiang.
Energy detection for spectrum sensing in cognitive radio.
Springer, 2014.
[2]
S. Balakrishnan and L. Wasserman.
Hypothesis testing for densities and high-dimensional multinomials:
Sharp local minimax rates.
arXiv preprint arXiv:1706.10003, 2017.
[3]
Y. Baraud.
Non-asymptotic minimax rates of testing in signal detection.
Bernoulli, 8(5):577–606, 2002.
[4]
P. J. Bickel, C. A. Klaassen, P. J. Bickel, Y. Ritov, J. Klaassen, J. A.
Wellner, and Y. Ritov.
Efficient and adaptive estimation for semiparametric models.
Johns Hopkins University Press Baltimore, 1993.
[5]
T. T. Cai, A. Guntuboyina, and Y. Wei.
Adaptive estimation of planar convex sets.
To appear in The Annals of Statistics, arXiv:1508.03744, 2017+.
[6]
T. T. Cai and M. G. Low.
A framework for estimation of convex functions.
Statistica Sinica, pages 423–456, 2015.
[7]
S. Chatterjee, A. Guntuboyina, B. Sen, et al.
On risk bounds in isotonic and other shape restricted regression
problems.
The Annals of Statistics, 43(4):1774–1800, 2015.
[8]
F. F. Digham, M.-S. Alouini, and M. K. Simon.
On the energy detection of unknown signals over fading channels.
IEEE Transactions on Communications, 55(1):21–24, 2007.
[9]
D. L. Donoho.
Compressed sensing.
IEEE Transactions on Information Theory, 52(4):1289–1306,
2006.
[10]
D. L. Donoho and J. M. Johnstone.
Ideal spatial adaptation by wavelet shrinkage.
Biometrika, 81(3):425–455, 1994.
[11]
D. L. Donoho, R. C. Liu, and B. MacGibbon.
Minimax risk over hyperrectangles, and implications.
The Annals of Statistics, pages 1416–1437, 1990.
[12]
M. S. Ermakov.
Minimax detection of a signal in a Gaussian white noise.
Theory of Probability & Its Applications, 35(4):667–679,
1991.
[13]
T. Friedrich et al.
Elementary geometry, volume 43.
American Mathematical Soc., 2008.
[14]
C. Gu.
Smoothing spline ANOVA models.
Springer Series in Statistics. Springer, New York, NY, 2002.
[15]
R. Hasminskii, I. Ibragimov, et al.
On density estimation in the view of Kolmogorov’s ideas in
approximation theory.
The Annals of Statistics, 18(3):999–1010, 1990.
[16]
I. A. Ibragimov and R. Z. Khasminskii.
Asymptotic properties of some nonparametrie estimaters in a
Gaussian white noise.
In: Proc 3rd Summer School on Probab, Theory and
Math. Stat., pages 31–64, 1978.
[17]
I. A. Ibragimov and R. Z. Khasminskii.
Statistical Estimation: Asymptotic Theory, volume 16.
Springer Science & Business Media, 2013.
[18]
Y. I. Ingster.
Minimax testing of nonparametric hypotheses on a distribution density
in the ${L}_{p}$-metrics.
Theory of Probability and Its Applications, 31(2):333–337,
1987.
[19]
Y. I. Ingster.
Asymptotically minimax hypothesis testing for nonparametric
alternatives. I.
Math. Methods Statist, 2:85–114, 1993a.
[20]
Y. I. Ingster.
Asymptotically minimax hypothesis testing for nonparametric
alternatives. II.
Math. Methods Statist, 3:171–189, 1993b.
[21]
Y. I. Ingster.
Asymptotically minimax hypothesis testing for nonparametric
alternatives. III.
Math. Methods Statist, 4:249–268, 1993c.
[22]
Y. I. Ingster and I. A. Suslina.
Nonparametric goodness-of-fit testing under Gaussian models,
volume 169.
Springer Science and Business Media, 2012.
[23]
A. Javanmard and L. Zhang.
The minimax risk of truncated series estimators for symmetric convex
polytopes.
In Information Theory Proceedings (ISIT), 2012 IEEE
International Symposium on, pages 1633–1637. IEEE, 2012.
[24]
E. L. Lehmann and G. Casella.
Theory of point estimation.
Springer Science & Business Media, 2006.
[25]
E. L. Lehmann and J. P. Romano.
Testing statistical hypotheses.
Springer Science & Business Media, 2006.
[26]
O. V. Lepski and V. G. Spokoiny.
Minimax nonparametric hypothesis testing: the case of an
inhomogeneous alternative.
Bernoulli, 5(2):333–358, 1999.
[27]
O. V. Lepski and A. B. Tsybakov.
Asymptotically exact nonparametric hypothesis testing in sup-norm and
at a fixed point.
Probability Theory and Related Fields, 117(1):17–48, 2000.
[28]
M. Meyer and M. Woodroofe.
On the degrees of freedom in shape-restricted regression.
The Annals of Statistics, pages 1083–1104, 2000.
[29]
A. Pinkus.
N-Widths in Approximation Theory.
Springer-Verlag, New York, 2012.
[30]
F. Shayegh and F. Labeau.
On signal detection in the presence of weakly correlated noise over
fading channels.
IEEE Transactions on Communications, 62(3):797–809, 2014.
[31]
V. G. Spokoiny.
Adaptive and spatially adaptive testing a nonparametric hypothesis.
Math. Methods Statist, 7:245–273, 1998.
[32]
G. Valiant and P. Valiant.
An automatic inequality prover and instance optimal identity testing.
In IEEE 55th Annual Symposium on Foundations of Computer Science
(FOCS), pages 51–60. IEEE, 2014.
[33]
S. van de Geer.
Empirical Processes in M-Estimation.
Cambridge University Press, 2000.
[34]
G. Wahba.
Spline models for observational data.
CBMS-NSF Regional Conference Series in Applied Mathematics. SIAM,
Philadelphia, PN, 1990.
[35]
Y. Wei, M. J. Wainwright, and A. Guntuboyina.
The geometry of hypothesis testing over convex cones: Generalized
likelihood tests and minimax radii.
arXiv preprint arXiv:1703.06810, 2017. |
Possible Coexistence of Antiferromagnetic and Ferromagnetic Spin Fluctuations in the Spin-triplet Superconductor UTe${}_{2}$ Revealed by ${}^{125}$Te NMR under Pressure
Devi V. Ambika
Ames Laboratory, U.S. DOE, Ames, Iowa 50011, USA
Department of Physics and Astronomy, Iowa State University, Ames, Iowa 50011, USA
Qing-Ping Ding
Ames Laboratory, U.S. DOE, Ames, Iowa 50011, USA
Department of Physics and Astronomy, Iowa State University, Ames, Iowa 50011, USA
Khusboo Rana
Ames Laboratory, U.S. DOE, Ames, Iowa 50011, USA
Department of Physics and Astronomy, Iowa State University, Ames, Iowa 50011, USA
Corey E. Frank
NIST Center for Neutron Research, National Institute of Standards and Technology, Gaithersburg, MD 20899, USA
Center for Nanophysics and Advanced Materials, Department of Physics, University of Maryland, College Park, MD 20742, USA
Elizabeth L. Green
National High Magnetic Field Laboratory, Florida State University, Tallahassee, Florida 32310, USA
Sheng Ran111Present address: Department of Physics, Washington University, St. Louis, MO
NIST Center for Neutron Research, National Institute of Standards and Technology, Gaithersburg, MD 20899, USA
Center for Nanophysics and Advanced Materials, Department of Physics, University of Maryland, College Park, MD 20742, USA
Nicholas P. Butch
NIST Center for Neutron Research, National Institute of Standards and Technology, Gaithersburg, MD 20899, USA
Center for Nanophysics and Advanced Materials, Department of Physics, University of Maryland, College Park, MD 20742, USA
Yuji Furukawa
Ames Laboratory, U.S. DOE, Ames, Iowa 50011, USA
Department of Physics and Astronomy, Iowa State University, Ames, Iowa 50011, USA
Abstract
A spin-triplet superconducting state mediated by ferromagnetic (FM) spin fluctuations has been suggested to occur in the newly discovered heavy-fermion superconductor UTe${}_{2}$. However, the recent neutron scattering measurements revealed the presence of antiferromagnetic (AFM) spin fluctuations in UTe${}_{2}$. Here, we report the ${}^{125}$Te nuclear magnetic resonance (NMR) studies of a single-crystal UTe${}_{2}$, suggesting the coexistence of FM and AFM spin fluctuations in UTe${}_{2}$. Owing to the two different Te sites in the compound, we conclude that the FM spin fluctuations are dominant within ladders and the AFM spin fluctuations originate from the inter-ladder magnetic coupling. Although AFM spin fluctuations exist in the system, the FM spin fluctuations in the ladders may play an important role in the appearance of the spin-triplet superconducting state of UTe${}_{2}$.
The interplay between magnetic fluctuations and superconductivity (SC) is one of the central issues in unconventional superconductors such as iron-based superconductors, high $T_{\rm c}$ cuprates, and heavy fermions.
UTe${}_{2}$ is a newly discovered heavy-fermion superconductor with a superconducting transition temperature $T_{\rm c}$ $\sim$ 1.6–2.0 K Ran2019 ; Aoki2019 ; Rosa2021 and was proposed to be located at the end member of U-based ferromagnetic (FM) superconductors.
Distinct from the previously discovered FM superconductors such as UGe${}_{2}$, UCoGe, and URhGe which exhibit long-range FM order Aoki20192 , UTe${}_{2}$ does not show any magnetic order down to 0.25 K Ran2019 ; Aoki2019 ; Ran2019_Nat ; Sundar2019 , making it unique in the family of U-based FM superconductors.
The observation of the nature of unconventional spin-triplet SC in UTe${}_{2}$, such as very anisotropic SC upper critical field $H_{\rm c2}$ exceeding the Pauli limit Ran2019 ; Aoki2019 ; Knebel2019 , multiple SC phases Braithwaite2019 ; Hayes2021 ; Thomas2020 , and time-reversal symmetry breaking Hayes2021 ; Wei2022 , has also sparked a large volume of research activity on the compound Aoki20192 .
Initially, FM spin fluctuations have been understandably considered to play an important role for the triplet paring, as suggested by muon spin relaxation Sundar2019 and nuclear magnetic resonance (NMR) measurements Tokunaga2019 .
On the other hand, recently, antiferromagnetic (AFM) spin fluctuations with the incommensurate wave-vector of $q$ = (0, 0.57, 0) have been detected by neutron scattering (NS) measurements Knafo2021 ; Duan2020 .
Those results indicate that the nature of the magnetic fluctuations in UTe${}_{2}$ are complicated and still under debate.
The nature of the magnetically ordered state as well as the SC phases under pressure is also still an open question.
With the application of pressure, two SC phases appear around 0.25 GPa Thomas2020 .
While the SC phase with lower $T_{\rm c}$ (SC1) is suppressed continuously with pressure, the SC phase with higher $T_{\rm c}$ (SC2) is enhanced and takes a maximum $T_{\rm c}$ $\sim$ 3 K at around 1.2 GPa, and is suppressed rapidly at higher pressures.
Above the critical pressure $p_{\rm c}$ $\sim$ 1.5 GPa, a magnetic phase appears Ran_PRB2020 ; Braithwaite2019 ; Thomas2020 ; Li2021 .
Initially, a FM ordered state was suggested for the pressure-induced magnetically ordered state Ran_PRB2020 ; Braithwaite2019 , however, recent studies proposed an AFM state Thomas2020 ; Li2021 .
NMR is a powerful technique to investigate low-energy spin fluctuations and superconducting properties from a microscopic point of view.
The temperature ($T$) dependence of the nuclear spin-lattice relaxation rate (1/$T_{1}$) reflects the wave vector $q$-summed dynamical susceptibility at nuclear sites.
On the other hand, NMR spectrum measurements, in particular, the Knight shift $K$, give us information on local static magnetic properties.
Furthermore, in the SC state, the temperature dependences of $K$ and 1/$T_{1}$ provide important information about the SC gap structure.
In fact, recent NMR measurements on single crystals revealed strong and slow spin fluctuations in the normal state Tokunaga2019 ; Tokunaga2022 and also provided key experimental results supporting the spin-triplet superconducting state in UTe${}_{2}$ Nagamine2019 ; Nagamine2021 ; Fujibayashi2022 .
In this paper, we have carried out ${}^{125}$Te NMR measurements using a ${}^{125}$Te-enriched single crystal of UTe${}_{2}$ to investigate the evolution of magnetic fluctuations in UTe${}_{2}$ under pressure.
The most striking result obtained here is the observation of the possible coexistence of AFM and FM spin fluctuations in UTe${}_{2}$.
Owing to the two different Te sites in the compound, the FM spin fluctuations are considered to be dominated within the ladders while the AFM spin fluctuations originate from the inter-ladder magnetic couplings.
The observed results are consistent with the recent NS data where intra-ladder and inter-ladder magnetic couplings are ferromagnetic and antiferromagnetic, respectively Knafo2021 ; Duan2020 .
The ${}^{125}$Te-enriched single crystal (3 $\times$ 1 $\times$ 0.2 mm${}^{3}$) of UTe${}_{2}$ with $T_{\rm c}$ = 1.6 K at zero magnetic field ($H$) at ambient pressure was synthesized by chemical vapor transport method using iodine as the transport agent Ran2019 .
NMR measurements of ${}^{125}$Te ($I$ = $\frac{1}{2}$, $\frac{\gamma_{\rm N}}{2\pi}$ = 13.454 MHz/T) nuclei were conducted using a laboratory-built phase-coherent spin-echo pulse spectrometer up to a pressure of 1.57 GPa with a NiCrAl/CuBe piston-cylinder cell using Daphne 7373 as the pressure transmitting medium.
Pressure calibration was accomplished by ${}^{63}$Cu nuclear quadrupole resonance in Cu${}_{2}$O Fukazawa2007 ; Reyes1992 at 77 K.
$T_{\rm c}$ = 2.3 K for 0.52 GPa and 2.8 K for 1.0 GPa at $H$ = 0 were determined by in-situ AC susceptibility measurements using an NMR tank circuit.
The values of $T_{\rm c}$ are consistent with previous papers Ran_PRB2020 ; Thomas2020 .
The ${}^{125}$Te-NMR spectra were obtained by sweeping $H$ at fixed NMR frequencies ($f$) where $H$ was applied parallel to the $b$ axis.
The 1/$T_{\rm 1}$ was measured with a saturation recovery method T1 .
UTe${}_{2}$ crystallizes in a body-centered orthorhombic structure with the $Immm$ space group Hutanu2020 where the U atoms form a two-leg ladder structure with legs along $a$ axis and rung along the $c$ axis as shown in Fig. 1(a) VESTA .
There are two crystallographic inequivalent Te sites occupying $4j$ and $4h$ sites with point symmetries $mm2$ and $m2m$, respectively.
Following the previous paper Tokunaga2019 , these sites are denoted by Te1 and Te2.
Te1 is located inside the distorted tetrahedron formed by the first and second nearest-neighbors four U atoms which are belonging to three ladders.
On the other hand, Te2 is surrounded by the four nearest neighbor U atoms forming a square-like structure within a ladder.
Thus NMR measurements for Te2 mainly pick up the local magnetic properties of each ladder while Te1 NMR provides the local information of magnetic properties related to inter-ladder coupling.
Those differences in the environments for the Te sites are important as discussed below.
Figure 1(b) shows the $H$-swept ${}^{125}$Te-NMR spectra at 30 K under various pressures ($p$ = 0–1.57 GPa) with $H$ parallel to the $b$ axis, where two lines corresponding to the two Te sites are observed.
The units of the horizontal axis are Knight shift $K$ defined by $K$ = (2$\pi$$f$/$\gamma_{\rm N}$ – $H$)/$H$.
The two lines with lower and higher $K$ values have been assigned to Te1 and Te2, respectively Tokunaga2019 .
The temperature and pressure dependences of full width at half maximum (FWHM) of lines for Te1 and Te2 are shown in Fig. 1(c).
At high $T$ above 200 K, the FWHMs are relatively small $\sim$ 0.8 mT for both Te1 and Te2 at ambient pressure, showing a good sample quality.
With decreasing $T$, FWHM increases and shows a broad maximum around 35 K, similar to the $T$ dependence of magnetic susceptibility $\chi$ and also $K$ as shown in Fig. 2(a).
This indicates that the $T$ dependence of FWHM reflects the $T$ dependence of $\chi$.
Similar $T$ dependence of FWHM can be observed with the lowered peak temperatures down to $\sim$20 K for $p$ = 0.52 and 1.0 GPa and $\sim$10 K at 1.57 GPa.
The large enhancements of FWHM at higher $p$ at low $T$ suggest the increase of $\chi$, as actually observed in the $K$ data.
The $T$ dependences of ${}^{125}$Te Knight shift for Te1 [$K$(Te1)] and Te2 [$K$(Te2)] sites determined by the peak positions of the NMR lines are shown in Fig. 2(a) for the measured pressures under $H$ $\parallel$ $b$.
All $K$ values for Te2 are greater than those for Te1 due to the different hyperfine coupling constants ($A_{\rm hf}$).
The $A_{\rm hf}$ values have been reported to be 34.1 and 51.8 kOe/$\mu_{\rm B}$ for Te1 and Te2, respectively, at ambient pressure Tokunaga2019 .
At ambient pressure, $K$s for both Te sites show a similar temperature dependence with a broad maximum around $T_{\rm max}$ $\sim$ 35–40 K, similar to the magnetic susceptibility data Ran2019 ; Aoki2019 ; Li2021 .
With increasing $p$, $T_{\rm max}$ shifts to lower temperatures to $\sim$30 K at $p$ = 0.52 GPa and to $\sim$25 K at $p$ = 1.0 GPa.
These behaviors are consistent with the previously reported $p$-dependence of magnetic susceptibility Li2021 and the recent NMR data Kinjo2022 .
At $p$ = 1.57 GPa, both $K$s keep increasing with decreasing $T$ and level off below $\sim$15 K without showing a clear maximum.
We were able to measure the spectrum down to 5.5 K although the signal intensity becomes weak below 10 K due to the shortening of nuclear spin-spin relaxation time $T_{2}$.
However, we could not observe any signals at 4.2 K.
This could be due to the pressure-induced short-range magnetically ordered state whose onset temperature has been reported to be $\sim$5 K at $p$ = 1.57 GPa Ran_PRB2020 ; Thomas2020 ; Li2021 .
We notice that the $T$ dependences of $K$(Te1) and $K$(Te2) are slightly different below $\sim$30 K, a little bit lower than $T_{\rm max}$ under $p$ $<$ 1.0 GPa.
This can be seen in Fig. 2(b) where the ratios of Knight shifts for the two Te sites, $K$(Te2)/$K$(Te1), are plotted as a function of $T$.
The ratio is nearly 1.48 at higher $T$ above $T_{\rm max}$ at ambient pressure and increases to 1.62 at 1.6 K.
This indicates that the $T$ dependences of $K$ for Te1 and Te2 are scaled above $\sim$30 K, but do not scale below 30 K.
Since the $T$ dependent part of $K$ is proportional to hyperfine coupling constant as $K(T)$ = $A_{\rm hf}$$\chi(T)$/$N_{\rm A}$$\mu_{\rm B}$
where $N_{\rm A}$ is Avogadro’s number and $\mu_{\rm B}$ is Bohr magneton, this indicates that the ratio of the $A_{\rm hf}$ for Te1 and Te2 changes slightly below $\sim$30 K.
To understand how the $A_{\rm hf}$ changes at low temperatures, we plotted $K$ as a function of the magnetic susceptibility.
As can be seen in Fig. 2(c), all the data points for Te2 are on the same straight line above and below 30 K, indicating no change in $A_{\rm hf}$ for this site.
From the slope of the line, $A_{\rm hf}$ is estimated to be 52.0 kOe/$\mu_{\rm B}$ which is very close to 51.8 kOe/$\mu_{\rm B}$ reported previously Tokunaga2019 .
On the other hand, although a clear linear relationship between $K$(Te1) and $\chi(T)$ can be seen above 30 K (shown by the solid red circles), the data points start deviating below the temperature as shown by the open red circles.
From the change in the slopes shown by solid and broken red lines in Fig. 2(c), we found the $A_{\rm hf}$ changes from 34.8 kOe/$\mu_{\rm B}$ above 30 K to 32.2 kOe/$\mu_{\rm B}$ at 1.6 K.
Thus we attribute the small increase in $K$(Te2)/$K$(Te1) below 30 K to the small reduction of $A_{\rm hf}$ for the Te1 site.
Since 30 K is close to the crossover temperature below which Fermi-liquid (FL) behavior 1/$T_{1}T$ = constant is observed as shown in Fig. 3, the results suggest that $A_{\rm hf}$ changes slightly in the FL state in UTe${}_{2}$.
The similar $T$ dependence of the ratios can be seen at 0.52 and 1.0 GPa, but the deviation from the constant values at high temperatures starts at slightly lower temperatures of $\sim$ 20 K for 0.52 GPa and $\sim$15 K for 1.0 GPa, respectively.
This is consistent with $T_{1}$ data where the crossover temperature to a low-temperature FL state decrease with increasing $p$.
It is interesting to point out that the ratios decrease with increasing $p$, indicating that the ratio of $A_{\rm hf}$ changes.
To determine the origin of the changes in the ratios under pressure, one needs $\chi$ data at the same pressure which are not available at present.
Further studies for the $\chi$ measurements are required to elucidate the origin.
At 1.57 GPa, the values of the ratio decrease slightly with $p$, but a clear upturn cannot be observed down to 5.5 K.
This may suggest no crossover to the FL state at 1.57 GPa down to $\sim$5.5 K.
It is also worth mentioning that, since UTe${}_{2}$ does not exhibit SC but has a magnetically ordered ground state under $p$ $>$ 1.5 GPa Thomas2020 , the unconventional SC state only appears in the FL state, as has been pointed out in Ref. Kinjo2022 .
Now we discuss magnetic fluctuations in UTe${}_{2}$ based on the results of $T_{1}$ measurements for both Te sites.
Figures 3(a)-3(d) show the $T$ dependence of 1/$T_{1}T$ at various pressures.
At ambient pressure, 1/$T_{1}T$ for both the Te sites exhibits Curie-Weiss (CW) behavior above $\sim$40 K.
Below $\sim$40 K, due to the hybridization between the localized 5$f$-electron and conduction electron bands, 1/$T_{1}T$ exhibits nearly constant behavior, a characteristics of heavy-fermion states.
The solid lines represents the FL behaviors of 1/$T_{1}T$ = 325 and 250 1/sK for Te1 and Te2, respectively.
These results are consistent with the data reported previously Tokunaga2019 .
The FL-crossover temperature defined as $T_{\rm FL}$ is also usually considered as Kondo temperature.
Under pressure, the crossover temperature decreases down to $T_{\rm FL}$ $\sim$ 20 and $\sim$ 10 K for $p$ = 0.52 and 1.0 GPa, respectively.
At the same time, the values of 1/$T_{1}T$ = constant increase to 260 (390) 1/sK for Te2(Te1) at 0.52 GPa and to 510 (765) 1/sK for Te2(Te1) at 1.0 GPa, suggesting the increase of the density of states at the Fermi energy and/or the enhancement of electron correlations in the FL state.
The CW behavior of 1/$T_{1}T$ above $\sim$40 K is nearly independent of $p$, however, we observed a different $T$ dependence of 1/$T_{1}T$ for the two Te sites at low temperatures.
With decreasing $T$ below $\sim$40 K, 1/$T_{1}T$ for Te1 is enhanced more than 1/$T_{1}T$ for Te2, which is clearly recognized under $p$ = 1.57 GPa.
Even for the case of $p$ = 0.52 and 1.0 GPa, one can see difference in the $T$ dependence of 1/$T_{1}T$ below $\sim$40 K.
The different $T$ dependence between Te1 and Te2 is more clearly seen in the 1/$T_{1}$ vs. $T$ plots in Figs. 3(e)-3(h).
1/$T_{1}$ for Te2 is nearly $T$ independent at high temperatures and shows a FL behavior below $T_{\rm FL}$ $\sim$ 20 K and $\sim$ 10 K for $p$ = 0.52 and 1.0 GPa, respectively.
1/$T_{1}$ for Te1 shows an enhancement where a clear difference in the $T$ dependence can be detected below $\sim$40 K down to $T_{\rm FL}$.
A similar difference in the $T$ dependence of 1/$T_{1}$ between Te1 and Te2 is also observed at $p$ = 1.57 GPa where the ground state is magnetic.
Those results indicate that Te1 and Te2 pick up different magnetic fluctuations.
The CW behavior in 1/$T_{1}T$ has been pointed out to be scaled with the $T$ dependence of the magnetic susceptibility under $H$ parallel to the $a$ axis (the magnetic easy axis), expected for FM spin fluctuations with $q$ = 0 Tokunaga2019 .
The FM spin fluctuations were also suggested from nuclear spin-spin relaxation time $T_{2}$ measurements Tokunaga2019 ; Tokunaga2022 .
Therefore, the further enhancements of 1/$T_{1}T$ at the Te1 site in comparison with those at the Te2 site indicate an additional contribution of magnetic fluctuations with $q$ $\neq$ 0, AFM spin fluctuations, at the Te1 site AFM_Te1 .
These results suggest the existence of AFM and FM spin fluctuations in the paramagnetic state of UTe${}_{2}$.
It is noted that the AFM spin fluctuations develop below around 40 K under pressure, although the fluctuations were not clearly observed at ambient pressure.
As mentioned above, Te2 is surrounded by four U atoms forming a ladder and picks up magnetic fluctuations originating from each ladder.
Thus our results indicate that FM spin fluctuations are dominant within the ladders.
On the other hand, since Te1 is surrounded by four U atoms belonging to three different ladders, it is possible to pick up magnetic fluctuations originating from not only intra-ladder but also inter-ladders magnetic couplings as schematically shown in Fig. 4.
Therefore, the AFM spin fluctuations detected at Te1 are considered to be due to AFM coupling between the ladders in UTe${}_{2}$.
These results are consistent with the NS measurements where the intra-ladder magnetic interactions are ferromagnetic for the leg and the rung directions and the magnetic interactions between the ladders is antiferromagnetic Duan2020 ; Knafo2021 .
It is noted that the NS measurements detect the AFM spin fluctuations at ambient pressure while our NMR data do not show clear AFM spin fluctuation at $p$ = 0.
Although the reason for the difference is not clear at present, it may be possible to explain by taking into consideration the different energy scale between the two experimental techniques because NMR may not detect the magnetic fluctuations if those energy (frequency) are much higher than NMR frequency.
If this were the case, our NMR results might suggest that the fluctuation frequency of the AFM spin fluctuations gets lower with increasing pressure.
Further NS experiments under pressure are required to clarify this.
Finally it is interesting to point out that the energy scale of $\sim$40 K (below which the AFM spin fluctuations develop under $p$) has been detected in other experiments even at ambient $p$ such as the observation of peak in the magnetic excitation spectrum by inelastic NS measurements Butch2022 , the peak $T$ in the magnetic susceptibility, hybridization gap observed in scanning tunneling spectroscopy Jiao2020 and so on.
Therefore, the energy of $\sim$40 K could be considered as one of the characteristic energy scales in determining the physical properties of the heavy-fermion UTe${}_{2}$.
In summary, we performed ${}^{125}$Te NMR measurements on UTe${}_{2}$ under pressure up to 1.57 GPa.
From the different temperature dependence of 1/$T_{1}T$ between the two different Te sites, Te1 and Te2, we suggest the coexistence of AFM and FM spin fluctuations in UTe${}_{2}$.
The FM spin fluctuations are considered to exist inside each ladder while the AFM spin fluctuations are assigned to originate from the inter-ladder magnetic interactions.
We point out that the FM spin fluctuations inside the ladders may play an important role in the appearance of the spin-triplet superconducting state in UTe${}_{2}$, although AFM spin fluctuations start to develop below $\sim$40 K which are clearly observed under pressure.
Further detailed investigations of the relationship between the magnitude of AFM/FM fluctuations and $T_{\rm c}$, the magnetic ordered states, and also about the superconducting properties under pressures are highly called for.
The authors thank A. Sapkota, B. Li, M. Hirata, and R. Yamamoto for helpful discussions. The research was supported by the U.S. Department of Energy (DOE), Office of Basic Energy Sciences, Division of Materials Sciences and Engineering. Ames Laboratory is operated for the U.S. DOE by Iowa State University under Contract No. DE-AC02-07CH11358. The National High Magnetic Field Laboratory is supported by the National Science Foundation through NSF/DMR-1644779 and the State of Florida.
References
(1)
S. Ran, C. Eckberg, Q.-P. Ding, Y. Furukawa, T. Metz, S. R. Saha, I.-L. Liu, M. Zic, H. Kim, J. Paglione, and N. P. Butch, Science 365, 684 (2019).
(2)
D. Aoki, A. Nakamura, F. Honda, D. Li, Y. Homma, Y. Shimizu, Y. J. Sato, G. Knebel, J.-P. Brison, A. Pourret, D. Braithwaite, G. Lapertot, Q. Niu, M. Vališka, H. Harima, and J. Flouquet, J. Phys. Soc. Jpn. 88, 043702 (2019).
(3)
P. F. S. Rosa, A. Weiland, S. S. Fender, B. L. Scott, F. Ronning, J. D. Thompson, E. D. Bauer, S. M. Thomas, arXiv:2110.06200v1
(4)
D. Aoki, K. Ishida, and J. Flouquet, J. Phys. Soc. Jpn. 88, 022001 (2019).
(5)
S. Sundar, S. Gheidi, K. Akintola, A. M. Cote, S. R. Dunsiger, S. Ran, N. P. Butch, S. R. Saha, J. Paglione, and J. E. Sonier, Phys. Rev. B 100, 140502(R) (2019).
(6)
S. Ran, I.-L. Liu, Y. Suk Eo, D. J. Campbell, P. M. Neves, W. T. Fuhrman, S. R. Saha, C. Eckberg, H. Kim, D. Graf, F. Balakirev, J. Singleton, J. Paglione, and N. P. Butch, Nat. Phys. 15, 1250 (2019).
(7)
G. Knebel, W. Knafo, A. Pourret, Q. Niu, M. Vališka, D. Braithwaite, G. Lapertot, M. Nardone, A. Zitouni, S. Mishra, I. Sheikin, G. Seyfarth, J.-P. Brison, D. Aoki, and J. Flouquet, J. Phys. Soc. Jpn. 88, 063707 (2019).
(8)
S. M. Thomas, F. B. Santos, M. H. Christensen, T. Asaba, F. Ronning, J. D. Thompson, E. D. Bauer, R. M. Fernandes, G. Fabbris, and P. F. S. Rosa, Sci. Adv. 6, eabc8709 (2020).
(9)
D. Braithwaite, M. Vališka, G. Knebel, G. Lapertot, J.-P. Brison, A. Pourret, M. E. Zhitomirsky, J. Flouquet, F. Honda, D. Aoki, Commun. Phys. 2, 147 (2019).
(10)
I. M. Hayes, D. S. Wei, T. Metz, J. Zhang, Y. S. Eo, S. Ran, S. R. Saha, J. Collini, N. P. Butch, D. F. Agterberg, A. Kapitulnik, and J. Paglione, Science 373, 797 (2021).
(11)
D. S. Wei, D. Saykin, O. Y. Miller, S. Ran, S. R. Saha, D. F. Agterberg, J. Schmalian, N. P. Butch, J. Paglione, and A. Kapitulnik,
Phys. Rev. B 105, 024521 (2022).
(12)
Y. Tokunaga, H. Sakai, S. Kambe, T. Hattori, N. Higa, G. Nakamine, S. Kitagawa, K. Ishida, A. Nakamura, Y. Shimizu, Y. Homma, D. Li, F. Honda, and D. Aoki, J. Phys. Soc. Jpn. 88, 073701 (2019).
(13)
C. Duan, K. Sasmal, M. B. Maple, A. Podlesnyak, J.-X. Zhu, Q. Si, and P. Dai, Phys. Rev. Lett. 125, 237003 (2020).
(14)
W. Knafo, G. Knebel, P. Steffens, K. Kaneko, A. Rosuel, J.-P. Brison, J. Flouquet, D. Aoki, G. Lapertot, and S. Raymond, Phys. Rev. B 104, L100409 (2021).
(15)
D. Li, A. Nakamura, F. Honda, Y. J. Sato, Y. Homma, Y. Shimizu, J. Ishizuka, Y. Yanase, G. Knebel, J. Flouquet, and D. Aoki, J. Phys. Soc. Jpn. 90, 073703 (2021).
(16)
S. Ran, H. Kim, I.-L. Liu, S. R. Saha, I. Hayes, T. Metz, Y. S. Eo, J. Paglione, and N. P. Butch, Phys. Rev. B 101, 140503(R) (2020).
(17)
Y. Tokunaga, H. Sakai, S. Kambe, Y. Haga, Y. Tokiwa, P. Opletal, H. Fujibayashi, K. Kinjo, S. Kitagawa, K.i Ishida, A. Nakamura, Y. Shimizu, Y. Homma, D. Li, F. Honda, and D. Aoki, J. Phys. Soc. Jpn. 91, 0237070 (2022)
(18)
G. Nakamine, S. Kitagawa, K. Ishida, Y. Tokunaga, H. Sakai, S. Kambe, A. Nakamura, Y. Shimizu, Y. Homma, D. Li, F. Honda, and D. Aoki, J. Phys. Soc. Jpn. 88, 113703 (2019).
(19)
G. Nakamine, K. Kinjo, S. Kitagawa, K. Ishida, Y. Tokunaga, H. Sakai, S. Kambe, A. Nakamura, Y. Shimizu, Y. Homma, D. Li, F. Honda, and D. Aoki, Phys. Rev. B 103, L100503 (2021).
(20)
H. Fujibayashi, G. Nakamine, K. Kinjo, S. Kitagawa, K. Ishida, Y Tokunaga, H. Sakai, S. Kambe, A. Nakamura, Y. Shimizu, Y. Homma, D. Li, F.i Honda, and D. Aoki, J. Phys. Soc. Jpn. 91, 043705 (2022).
(21)
H. Fukazawa, N. Yamatoji, Y. Kohori, C. Terakura, N. Takeshita, Y. Tokura, and H. Takagi, Rev. Sci. Instrum. 78, 015106 (2007).
(22)
A. P. Reyes, E. T. Ahrens, R. H. Heffner, P. C. Hammel, and J. D. Thompson, Rev. Sci. Instrum. 63, 3120 (1992).
(23)
$1/T_{1}$ at each $T$ was determined by fitting the nuclear magnetization $M$ versus time $t$ using the exponential function $1-M(t)/M(\infty)=e^{-(t/T_{1})}$ for ${}^{125}$Te NMR, where $M(t)$ and $M(\infty)$ are the nuclear magnetization at $t$ after the saturation and the equilibrium nuclear magnetization at $t$ $\rightarrow$ $\infty$, respectively.
(24)
V. Hutanu, H. Deng, S. Ran, W. T. Fuhrman, H. Thoma and N. P. Butch, Acta Cryst. B76, 137 (2020).
(25)
The crystal structure was drawn by using VESTA [K. Momma and F. Izumi, J. Appl. Crystallogr., 44, 1272-1276 (2011)].
(26)
K. Kinjo, H. Fujibayashi, G. Nakamine, S. Kitagawa, K. Ishida, Y. Tokunaga, H. Sakai, S. Kambe, A. Nakamura, Y. Shimizu, Y. Homma, D. Li, F. Honda, and D. Aoki, Phys. Rev. B 105, L140502 (2022).
(27)
It is noted that AFM spin fluctuations will not be canceled out at Te1 due to the local configuration with respect to the four U atoms forming the distorted tetrahedron as shown in Fig. 1(a), which is not a high symmetry point in the crystal.
In addition, if the AFM fluctuations are characterized by the incommensurate wave vector determined by the neutron scattering measurements, the AFM spin fluctuations will not be canceled out at the Te1 site.
(28)
N. P. Butch, S. Ran, S. R. Saha, P. M. Neves, M. P. Zic, J. Paglione, S. Gladchenko, Q. Ye and J. A. Rodriguez-Rivera, npj Quant. Mater. 7, 39 (2022).
(29)
L. Jiao, S. Howard, S. Ran, Z. Wang, J. O. Rodriguez, M. Sigrist, Z. Wang, N. P. Butch, and V. Madhavan, Nature (London) 579, 523 (2020). |
Quark anti-quark expectation value in finite volume
Johan Bijnens
Karim Ghorbani111Speaker
Department of Theoretical Physics, Lund University, Sölvegatan 14A, SE 223 62, Sweden
Abstract
We have computed the quark anti-quark expectation value in finite volume at two loop
in chiral perturbation theory and compare it with a formula obtained in analogy to
the Lüscher formula for pion mass in finite volume. We observe
that due to the small finite size correction at two loop it is not possible to
obtain conclusions on the accuracy of the extended Lüscher formula.
LU TP 06-37
hep-lat/0610083
October 2006
I Introduction
Quantum chromo dynamics (QCD) is a strongly coupled theory at low energy.
Therefore it is difficult to study the physics below the QCD scale of about 1 GeV.
Lattice QCD on the other hand is an attempt to tackle this problem by taking the
Lagrangian and calculating (parts of) the generating functional numerically.
Although these computations with fairly small quark masses are now possible,
at the same time the finite volume effects are also becoming important.
In this work we employ chiral perturbation theory (ChPT) in order to study these
effects analytically. Chiral perturbation theory, in its modern form,
was first proposed by
Weinberg Weinberg and extended in GL1 ; BCE . ChPT is an effective field
theory to describe the strong interactions at low energy. To do the
computation in finite volume as suggested in GL2 one
imposes a periodic boundary condition on fields which result in
the momentum quantization and consequently modification on the quantum
corrections. Since the applicability of ChPT is limited to small momenta, in
finite volume this leads to
$$F_{\pi}L>1\,.$$
(1)
which holds for L larger than about 2 fm.
$F_{\pi}$ is the pion decay constant and L is the linear
size of the boundary. In addition we stay in a limit where the
Compton wavelength of the pion is smaller than the lattice size
which corresponds to
$$m_{\pi}^{2}F_{\pi}^{2}V>>1\,.$$
(2)
This is the so-called p-regime.
This condition ensures that perturbative treatment will work.
Expectation value of the quark condensate per se will be achieved by varying
the generating functional with respect to the scalar external field as follows
$$<\overline{q}T^{a}q>=\frac{1}{Z_{0}}(-i\frac{\delta}{\delta S_{a}})Z_{eff}[s_{%
a}]_{\arrowvert s_{a}=0}.$$
(3)
Where $T^{a}$ stands for the Gell-Mann matrices. We then carry out
the $\langle\overline{q}q\rangle$ at two loop in finite volume within
the framework of chiral perturbation theory. This quantity has already been evaluated at
one loop order in DG . Another way for the computation of the finite volume
effect is the Lüscher approach. In this method one obtains the leading finite size effects
for the pion mass to all order in perturbation theory from the scattering
amplitude Luscher1 ; Luscher2 . It has later been extended to evaluate the volume dependence
of the pion decay constantCH1 .
As the second goal we shall then obtain a formula for the the quark condensate volume
dependence by following the same line of reasoning as in the Lüscher work.
This work was published in Ref. BGH .
II A Lüscher formula for the vacuum condensate
We were inspired by the work of Lüscher, where he showed how the finite size effect on
the masses of spin-less particles are related to the forward elastic scattering amplitude
in infinite volume. This occurs when particles are enclosed in a lattice box
and as considered in this letter with temporal direction of the space-time extended to infinity.
To see explicitly where these effects stem from, one looks at the two-point correlation function
with fields subjected to a periodic boundary condition. The correlator takes on the form
$$<\phi(x)\phi(0>_{L}=\int\frac{dp_{0}}{2\pi}\sum_{\vec{p}}L^{-d}e^{-ip.x}G_{L}^%
{-1}(\vec{p}^{2},m_{0}^{2})$$
(4)
where the inverse propagator reads
$$G_{L}(\vec{p}^{2},m_{0}^{2})=p^{2}-m_{0}-\Sigma_{L}(p^{2})$$
(5)
and $\Sigma_{L}(p^{2})$ denotes the sum of the one-particle irreducible diagrams
to be evaluated in momentum space with discretized momenta and $\vec{p}=\frac{2\pi\vec{n}}{L}$,
with $\vec{n}$ a three dimensional vector of integers.
In practice the desired quantity to compute is the change in $\Sigma(p^{2})$ due to
the finite volume correction and the difficult task in Lüscher’s work was to show that
this consists of all self-energy diagrams with just one propagator in each diagram allowed
to wrap around the whole position space and therefore experiencing the boundary.
Using the Poisson summation formula the modified propagator can be obtained in terms of the
propagator in the infinite volume accompanied by an exponential factor as follows
$$G_{L}(q)=\sum_{\vec{n}}e^{-iq_{i}n_{i}L_{i}}G_{\infty}$$
(6)
where in three dimensions the length squared for a given
$\vec{n}$ reads $k=\vec{n}^{2}=n_{1}^{2}+n_{2}^{2}+n_{2}^{2}$.
For $\vec{n}=0$ it gives back the propagator in infinite volume.
The crucial point comes about in Euclidean space where $q_{0}\to-iq_{0}$
and, as shown below, in our result for $\langle\bar{q}q\rangle$
the exponential factor falls off rapidly
for large $\vec{n}^{2}$. Lüscher in his analysis kept only $\vec{n}^{2}=1$
but summation over all values shall be carried out numerically in our formula
as suggested in CH2 .
Out of three self-energy graphs shown in Ref. Luscher1 , only two are potentially relevant
to the $\langle\bar{q}q\rangle$ as depicted in Fig. 1.
The diagram (a) has however no contribution when we apply
ChPT to the case of parity even operators.
Extending our task to the evaluation of a general operator in the the diagram (b)
we arrive at the following relation
$$\langle O\rangle_{V}-\langle O\rangle_{\infty}=\int\frac{d^{d}q}{(2\pi)^{d}}%
\sum_{\vec{n}\neq 0}e^{-iLq.n}G_{\infty}\langle\phi|O|\phi\rangle\,.$$
(7)
Integration over momentum can be split into parallel and transverse components
with respect to the $\vec{n}$, followed by the continuation
$q_{\parallel}\to q_{\parallel}-is$.
For $s\to\infty$ the integration over this path is negligible.
Unlike the mass case it turns out that the computation of the diagram (b)
for the $\langle\bar{q}q\rangle$ is simpler mainly because the matrix
element $\langle\phi|O|\phi\rangle$ representing the amputated
vertex function has no external momenta in it and in fact it should be
evaluated at zero momentum transfer.
After isolating the poles and picking out the right one we, complete the integration
over $q_{\parallel}$ which finally leads to the following formula
$$\displaystyle\langle O_{V}\rangle-\langle O_{\infty}\rangle=-\langle\phi|O|\phi\rangle$$
(8)
$$\displaystyle\!\!\!\!\times\left(\sum_{\vec{n}\neq 0}\frac{1}{16\pi^{2}}\int_{%
0}^{\infty}\frac{q^{2}dq}{\sqrt{q^{2}+m^{2}}}e^{-\sqrt{\vec{n}^{2}(m^{2}+q^{2}%
)L^{2}}}\right)$$
As the last step we identify the remaining integral as the modified Bessel function
$K_{1}$
$$\langle O\rangle_{V}-\langle O\rangle_{\infty}=-\langle\phi|O|\phi\rangle\sum_%
{k=1,\infty}\frac{x(k)}{16\pi^{2}}\frac{m^{2}}{\sqrt{\zeta(k)}}K_{1}(\zeta(k))$$
(9)
where $\zeta(k)=\sqrt{(}k)mL$ and $x(k)$ is the multiplicity,i.e. $x(k)$
is the number of times the value $k=\vec{n}^{2}$ appears in the sum overt $\vec{n}$.
Deriving this formula also has the advantage that one can estimate the size of the
sigma term by evaluating the volume dependence of the
the quark anti-quark expectation value.
III The vacuum expectation value at two loop
The calculation of the quark condensate at two loop was done in ABP . We go back to
these computation while taking into account the finite size effects.
As mentioned before, this effect comes about as a modification in the
propagator. In Fig. 2 we show all the Feynman diagrams that contribute up to $p^{6}$.
In infinite volume the loop diagrams we deal with, contain the functions
$$\displaystyle A(m^{2})=\frac{i}{(2\pi)^{d}}\int d^{d}q\frac{1}{q^{2}-m^{2}}$$
$$\displaystyle B(m^{2})=\frac{i}{(2\pi)^{d}}\int d^{d}q\frac{1}{(q^{2}-m^{2})^{%
2}}$$
(10)
These integrals are solved in the dimensional regularization scheme
with $\epsilon=(4-d)/2$ .
$$\displaystyle A(m^{2})$$
$$\displaystyle=$$
$$\displaystyle\frac{m^{2}}{16\pi^{2}}\left(\frac{1}{\epsilon}-\gamma_{E}+log(4%
\pi)+1\right)$$
(11)
$$\displaystyle+\bar{A}(m^{2})+\epsilon A^{\epsilon}(m^{2})+o(\epsilon^{2})$$
$$\displaystyle B(m^{2})$$
$$\displaystyle=$$
$$\displaystyle\frac{m^{2}}{16\pi^{2}}\left(\frac{1}{\epsilon}-\gamma_{E}+log(4%
\pi)+1\right)$$
(12)
$$\displaystyle+\bar{B}(m^{2})+\epsilon B^{\epsilon}(m^{2})+o(\epsilon^{2}).$$
At finite volume integration over momenta should be replaced by a summation
over discretized momenta. It is also important to note that the pole structure
does not change its form, since the singularity arises when momenta in the
loop become very large and therefore the pole remains unchanged despite the replacement
of integration with summation. The final result for the integrals
at finite volume becomes
$$\displaystyle\bar{A}(m^{2})$$
$$\displaystyle=$$
$$\displaystyle-\frac{m^{2}}{16\pi^{2}}\log\left(\frac{m^{2}}{\mu^{2}}\right)$$
(13)
$$\displaystyle-\frac{1}{16\pi^{2}}\sum_{k=1,\infty}x(k)\frac{4m^{2}}{\zeta(k)}K%
_{1}(\zeta(k))$$
$$\displaystyle\bar{B}(m^{2})$$
$$\displaystyle=$$
$$\displaystyle-\frac{1}{16\pi^{2}}\log\left(\frac{m^{2}}{\mu^{2}}\right)-\frac{%
1}{16\pi^{2}}$$
(14)
$$\displaystyle+\frac{1}{16\pi^{2}}\sum_{k=1,\infty}x(k)2K_{o}(\zeta(k))$$
with $\zeta(k)=\sqrt{k}mL$.
It is evident from equations (13) and (14) that the simple relation
$\bar{A}(m^{2})=m^{2}(1/16\pi^{2}+\bar{B}(m^{2}))$ is no longer valid in
finite volume computations. After putting all the diagrams together and
applying the dimensional regularization scheme with renormalization
parameter $\mu$ we ensure that the divergent parts cancel and for finite part
we use the expressions in (13) and (14) instead of those in
(11) and (12).
The parts containing $A^{\epsilon}$ and $B^{\epsilon}$ cancel in the final
result for the vacuum expectation value.
We denote the lowest order masses for pions, kaons and etas as $\chi_{\pi}$, $\chi_{K}$ and
$\chi_{\eta}$ respectively. They are given in terms of the strange quark mass, $m_{s}$
and the average of up quark and down quark masses, $\hat{m}$ by
$$\displaystyle\chi_{\pi}=2B_{0}\hat{m}\,,\quad\chi_{K}=B_{0}(m_{s}+\hat{m})$$
$$\displaystyle\chi_{\eta}=2B_{0}(2m_{s}+\hat{m})/3\,.$$
(15)
We also define the following quantities
$$\displaystyle\langle\bar{u}u\rangle$$
$$\displaystyle=$$
$$\displaystyle-B_{0}F_{0}^{2}\left(1+\frac{v_{4}^{(u)}}{F_{0}}+\frac{v_{6}^{(u)%
}}{F_{0}^{2}}\right)\,,$$
$$\displaystyle\langle\bar{s}s\rangle$$
$$\displaystyle=$$
$$\displaystyle-B_{0}F_{0}^{2}\left(1+\frac{v_{4}^{(s)}}{F_{0}}+\frac{v_{6}^{(s)%
}}{F_{0}^{2}}\right)\,.$$
(16)
The analytical expressions for $v_{4}^{(u)}$,$v_{4}^{(s)}$,
$v_{6}^{(u)}$,$v_{6}^{(s)}$
are evaluated in ChPT and can be found in BGH . Numerical results are shown
in Fig. 3 and Fig. 4, where the ratio $R_{q}$ is defined as
$$R_{q}\equiv\frac{\langle\bar{q}q\rangle_{V}-\langle\bar{q}q\rangle_{\infty}}{%
\langle\bar{q}q\rangle_{\infty}}\,.$$
(17)
The $C^{r}_{i}$, the low energy coupling constants (LECs) from diagrams at $p^{6}$ BCE
cancel out in the numerator and we have ignored them in the denominator but the
$L^{r}_{i}$ will be considered as inputs and we use the values from fit 10 ofABT .
As expected the volume size effects are pronounced for small L and small pion mass.
But the full NNLO contribution to the finite volume effect
is much smaller than the total NNLO numerical correction.
In turn this total NNLO correction is smaller than the one
presented in ABP . This is due to the fact that in Ref. ABP
the physical value of pion decay constant $F_{\pi}$ and also physical masses for
mesons are put in the formula.
IV Conclusions
We compared the quark condensate in finite volume obtained in two different ways.
The method referred to as Lüscher’s approach yields the leading term of this quantity
to all orders in perturbation theory while only one propagator experiences the finite
volume. One the other hand, we try to do this computation up to $p^{6}$ rather directly
by letting the modification of all propagators be present at the same time.
The comparison of the direct calculation at one-loop with the extended formula is
trivial but at two-loop level, however, this is not the case.
This allows in principle a check of the accuracy of the extended Lüscher approach.
Unfortunately, because of the very small size of the finite volume effect at $p^{6}$,
no conclusion can be drawn concerning the accuracy level of the extended Lüscher
formula from our work.
Acknowledgements.
One of us, KG, wishes to thank the organizers of the IPM-LHP06 conference for the travel grant and
great hospitality.
This work is supported by the European Union RTN network,
Contract No.
MRTN-CT-2006-035482 (FLAVIAnet) and by
the European Community-Research Infrastructure
Activity Contract No. RII3-CT-2004-506078 (HadronPhysics).
KG acknowledges a fellowship from the Iranian Ministry of Science.
References
(1)
S. Weinberg,
PhysicaA 96 (1979) 327.
(2)
J. Gasser and H. Leutwyler,
Annals Phys. 158 (1984) 142;
Nucl. Phys. B 250 (1985) 465.
(3)
J. Bijnens, G. Colangelo and G. Ecker,
JHEP 9902 (1999) 020
[arXiv:hep-ph/9902437];
Annals Phys. 280 (2000) 100
[arXiv:hep-ph/9907333].
(4)
J. Gasser and H. Leutwyler,
Nucl. Phys. B 307 (1988) 763.
(5)
S. Descotes-Genon,
Eur. Phys. J. C 40 (2005) 81
[arXiv:hep-ph/0410233].
(6)
M. Lüscher,
Commun. Math. Phys. 104 (1986) 177.
(7)
M. Lüscher,
DESY 83/116,
Lecture given at Cargese Summer Institute,
Cargese, France, 1-15 September 1983.
(8)
G. Colangelo and C. Haefeli,
Phys. Lett. B 590 (2004) 258
[arXiv:hep-lat/0403025].
(9)
J. Bijnens and K. Ghorbani,
“Finite volume dependence of the quark-antiquark vacuum expectation
Phys. Lett. B 636 (2006) 51
[arXiv:hep-lat/0602019].
(10)
G. Colangelo and S. Durr,
Eur. Phys. J. C 33 (2004) 543
[arXiv:hep-lat/0311023].
(11)
G. Amorós, J. Bijnens and P. Talavera,
Nucl. Phys. B 585 (2000) 293
[Erratum-ibid. B 598 (2001) 665]
[arXiv:hep-ph/0003258].
(12)
G. Amorós, J. Bijnens and P. Talavera,
Nucl. Phys. B 602 (2001) 87
[arXiv:hep-ph/0101127]. |
Symmetry Reduction in Twisted Noncommutative Gravity
with Applications to Cosmology and Black Holes
Thorsten Ohl
Alexander Schenkel
Institut für Theoretische Physik und Astrophysik
Universität Würzburg, Am Hubland, 97074 Würzburg, Germany
e-mail: [email protected]: [email protected]
(October 2008)
Abstract
As a preparation for a mathematically consistent study of the
physics of symmetric spacetimes in a noncommutative setting, we
study symmetry reductions in deformed gravity. We focus on
deformations that are given by a twist of a Lie algebra acting on the
spacetime manifold. We derive conditions on those twists that allow a
given symmetry reduction. A complete classification of admissible
deformations is possible in a class of twists generated by
commuting vector fields. As examples, we explicitly construct the families of
vector fields that generate twists which are compatible with
Friedmann-Robertson-Walker cosmologies and Schwarzschild black
holes, respectively. We find nontrivial isotropic twists of FRW
cosmologies and nontrivial twists that are compatible with all
classical symmetries of black hole solutions.
1 Introduction
The study of noncommutative geometry is an active topic in both
theoretical physics and mathematics. From the mathematical perspective
it is a generalization of classical (commutative) geometry. From the
physics perspective it is suggested by the Gedankenexperiment
of localizing events in spacetime with a Planck scale
resolution [1]. In this
Gedankenexperiment, a sharp localization induces an
uncertainty in the spacetime coordinates, which can naturally be
described by a noncommutative spacetime. Furthermore, noncommutative
geometry and quantum gravity appear to be connected strongly and one
can probably model “low energy” effects of quantum gravity theories
using noncommutative geometry.
There have been many attempts to formulate scalar, gauge and gravity
theories on noncommutative spacetime, in particular using the simplest
example of a Moyal-Weyl spacetime having constant noncommutativity
between space and time coordinates, see
[2, 3] for reviews. Furthermore,
this framework had been applied to phenomenological particle physics
with [4, 5] and without Seiberg-Witten maps (see the
review [6] and references therein),
cosmology [7] and black hole physics (see the
review [8] and references therein).
Our work is based on the approach outlined
in [9, 10, 11], where a
noncommutative gravity theory based on an arbitrary twist deformation
is established. This approach has the advantages of being formulated
using the symmetry principle of deformed diffeomorphisms, being
coordinate independent and applicable to nontrivial topologies.
However, there is also the disadvantage that it does not match
the Seiberg-Witten limit of string theory [12].
Nevertheless, string theory is not the only candidate for a
fundamental theory of quantum gravity. Therefore, the investigation of
deformed gravity remains interesting on its own terms and it could very well
emerge from a fundamental theory of quantum gravity different from
string theory.
The outline of this paper is as follows. In section 2
we review the basics of the formalism of twisted noncommutative
differential geometry. For more details and the proofs we refer to the
original paper [10] and the
review [11]. We will work with a general twist and
do not restrict ourselves to the Moyal-Weyl deformation.
In section 3 we will study symmetry reduction in
theories based on twisted symmetries, such as the twisted
diffeomorphisms in our theory of interest. The reason is that we aim
to investigate which deformations of cosmological and black hole
symmetries are possible. We will derive the conditions that the twist
has to satisfy in order to be compatible with the reduced symmetry.
In section 4 we restrict the twists to the class of
Reshetikhin-Jambor-Sykora twists [13, 14],
that are twists generated by commuting vector fields and are convenient
for practical applications. Within this restricted class of twists we
can classify more explicitly the possible deformations of Lie algebra
symmetries acting on a manifold $\mathcal{M}$.
In section 5 and 6 we apply the
formalism to cosmological symmetries as well as the black hole. We
classify the possible Reshetikhin-Jambor-Sykora deformations of these models and
obtain physically interesting ones. In section 7 we
conclude and give an outlook to possible further investigations. In
particular possible applications to phenomenological cosmology and
black hole physics will be discussed.
2 Basics of Twisted Differential Geometry and Gravity
In order to establish notation, we will give a short summary of the framework of twisted differential geometry and gravity.
More details can be found in [9, 10, 11].
There is a quite general procedure for constructing noncommutative spaces and their corresponding symmetries
by using a twist. For this we require the following ingredients [11]:
1.
a Lie algebra $\mathfrak{g}$
2.
an action of the Lie algebra on the space we want to deform
3.
a twist element $\mathcal{F}$, constructed from the generators of the Lie algebra $\mathfrak{g}$
By a twist element we denote an invertible element of $U\mathfrak{g}\otimes U\mathfrak{g}$, where
$U\mathfrak{g}$ is the universal enveloping algebra of $\mathfrak{g}$. $\mathcal{F}$ has to fulfill some
conditions, which will be specified later. The basic idea in the following is to combine any bilinear map with the inverse twist
and therefore deform these maps. This leads to a mathematically consistent deformed theory covariant under
the deformed transformations. We will show this now for the deformation of diffeomorphisms.
For our purpose we are interested in the Lie algebra of vector fields $\Xi$ on a manifold $\mathcal{M}$.
The transformations induced by $\Xi$ can be seen as infinitesimal diffeomorphisms. A natural action of these
transformations on the algebra of tensor fields $\mathcal{T}:=\bigoplus\limits_{n,m}\bigotimes^{n}\Omega\otimes\bigotimes^{m}\Xi$
is given by the Lie derivative $\mathcal{L}$. $\Omega$ denotes the space of one-forms.
In order to deform this Lie algebra, as well as its action on tensor fields and the tensor fields themselves, we first have
to construct the enveloping algebra $U\Xi$. This is the associative
tensor algebra generated by the elements of $\Xi$ and the unit $1$, modulo the left and right ideals generated by
the elements $[v,w]-vw+wv$. This algebra can be seen as a Hopf algebra by using the following
coproduct $\Delta$, antipode $S$ and counit $\epsilon$ defined on the generators $u\in\Xi$ and $1$ by:
$$\displaystyle\begin{array}[]{ll}\Delta(u)=u\otimes 1+1\otimes u,&\Delta(1)=1%
\otimes 1~{},\\
\epsilon(u)=0,&\epsilon(1)=1~{},\\
S(u)=-u,&S(1)=1~{}.\end{array}$$
(1)
These definitions can be consistently carried over to the whole enveloping algebra demanding $\Delta$ and $\epsilon$
to be algebra homomorphisms and $S$ to be an anti-homomorphism, i.e. for any two elements $\eta,\xi\in U\Xi$, we require
$$\displaystyle\Delta(\eta\xi)=\Delta(\eta)\Delta(\xi)~{},$$
(2a)
$$\displaystyle\epsilon(\eta\xi)=\epsilon(\eta)\epsilon(\xi)~{},$$
(2b)
$$\displaystyle S(\eta\xi)=S(\xi)S(\eta)~{}.$$
(2c)
The action of the enveloping algebra on the tensor fields can be defined by extending the Lie derivative
$$\displaystyle\mathcal{L}_{\eta\xi}(\tau):=\mathcal{L}_{\eta}(\mathcal{L}_{\xi}%
(\tau))~{},~{}\forall\eta,\xi\in U\Xi~{},~{}\tau\in\mathcal{T}~{}.$$
(3)
This action is consistent with the Lie algebra properties, since $\mathcal{L}_{[u,v]}(\tau)=\mathcal{L}_{uv}(\tau)-\mathcal{L}_{vu}(\tau)$
for all $u,v\in\Xi$ by the properties of the Lie derivative.
The extension of the Lie algebra $\Xi$ to the Hopf algebra $(U\Xi,\cdot,\Delta,S,\epsilon)$, where $\cdot$ is the multiplication in $U\Xi$,
can now be used in order to construct deformations of it. For the deformations we restrict ourselves
to twist deformations, which is a wide class of possible deformations. The reason is that for twist deformations the
construction of deformed differential geometry and gravity can be performed explicitly by only using properties of the twist,
see [10]. Other deformations require further investigations.
In order to perform the deformation we require a twist element $\mathcal{F}=f^{\alpha}\otimes f_{\alpha}\in U\Xi\otimes U\Xi$
(the sum over $\alpha$ is understood) fulfilling the following conditions
$$\displaystyle\mathcal{F}_{12}(\Delta\otimes\mathrm{id})\mathcal{F}=\mathcal{F}%
_{23}(\mathrm{id}\otimes\Delta)\mathcal{F}~{},$$
(4a)
$$\displaystyle(\epsilon\otimes\mathrm{id})\mathcal{F}=1=(\mathrm{id}\otimes%
\epsilon)\mathcal{F}~{},$$
(4b)
$$\displaystyle\mathcal{F}=1\otimes 1+\mathcal{O}(\lambda)~{},$$
(4c)
where $\mathcal{F}_{12}:=\mathcal{F}\otimes 1$, $\mathcal{F}_{23}:=1\otimes\mathcal{F}$ and $\lambda$ is the deformation parameter.
The first condition will assure the associativity of the deformed products, the second will assure that deformed
multiplications with unit elements will be trivial and the third condition assures the existence of the undeformed
classical limit $\lambda\to 0$.
Furthermore, we can assume without loss of generality that $f_{\alpha}$ (and also $f^{\alpha}$) are linearly independent for all $\alpha$,
what can be assured by combining linearly dependent $f$.
Note that $\mathcal{F}$ is regarded as formal power series in $\lambda$, such as the deformation itself.
Strict (convergent) deformations will not be regarded here.
The simplest example is the twist on $\mathbb{R}^{n}$ given by
$\mathcal{F}_{\theta}:=\exp{\bigl{(}-\frac{i\lambda}{2}\theta^{\mu\nu}\partial_%
{\mu}\otimes\partial_{\nu}\bigr{)}}$ with $\theta^{\mu\nu}=\mathrm{const.}$
and antisymmetric, leading to the Moyal-Weyl deformation, but there are also more complicated ones.
From a twist, one can construct the twisted triangular Hopf algebra
$(U\Xi_{\mathcal{F}},\cdot,\Delta_{\mathcal{F}},S_{\mathcal{F}},\epsilon_{%
\mathcal{F}})$
with $R$-matrix $R:=\mathcal{F}_{21}\mathcal{F}^{-1}=:R^{\alpha}\otimes R_{\alpha}$, inverse $R^{-1}=:\bar{R}^{\alpha}\otimes\bar{R}_{\alpha}=R_{21}$
and
$$\displaystyle\Delta_{\mathcal{F}}(\xi):=\mathcal{F}\Delta(\xi)\mathcal{F}^{-1}%
~{},\quad\epsilon_{\mathcal{F}}(\xi):=\epsilon(\xi)~{},\quad S_{\mathcal{F}}(%
\xi):=\chi S(\xi)\chi^{-1}~{},$$
(5)
where $\chi:=f^{\alpha}S(f_{\alpha})$, $\chi^{-1}:=S(\bar{f}^{\alpha})\bar{f}_{\alpha}$ and $\bar{f}^{\alpha}\otimes\bar{f}_{\alpha}:=\mathcal{F}^{-1}$.
Furthermore, $\mathcal{F}_{21}:=f_{\alpha}\otimes f^{\alpha}$ and $R_{21}:=R_{\alpha}\otimes R^{\alpha}$. Again, we can assume without loss of generality that all summands of $\mathcal{F}^{-1}$, $R$ and $R^{-1}$ are linearly independent.
However, as explained in [10], it is simpler to use the triangular $\star$-Hopf algebra $\mathcal{H}_{\Xi}^{\star}=(U\Xi_{\star},\star,\Delta_{\star},S_{\star},%
\epsilon_{\star})$,
isomorphic to $(U\Xi_{\mathcal{F}},\cdot,\Delta_{\mathcal{F}},S_{\mathcal{F}},\epsilon_{%
\mathcal{F}})$. The operations
in this algebra on its generators $u,v\in\Xi$ (note that this algebra has the same generators as the classical Hopf algebra) are defined by
$$\displaystyle u\star v:=\bar{f}^{\alpha}(u)\bar{f}_{\alpha}(v)~{},$$
(6a)
$$\displaystyle\Delta_{\star}(u):=u\otimes 1+X_{\bar{R}^{\alpha}}\otimes\bar{R}_%
{\alpha}(u)~{},$$
(6b)
$$\displaystyle\epsilon_{\star}(u):=\epsilon(u)=0~{},$$
(6c)
$$\displaystyle S_{\star}^{-1}(u):=-\bar{R}^{\alpha}(u)\star X_{\bar{R}_{\alpha}%
}~{},$$
(6d)
where for all $\xi\in U\Xi$ we define $X_{\xi}:=\bar{f}^{\alpha}\xi\chi S^{-1}(\bar{f}_{\alpha})$. The action of the twist on the elements of $U\Xi$
is defined by extending the Lie derivative to the adjoint action [10]. Note that $U\Xi=U\Xi_{\star}$ as vector spaces. The $R$-matrix is
given by $R_{\star}:=X_{R^{\alpha}}\otimes X_{R_{\alpha}}$ and is triangular. The coproduct and antipode (6)
is defined consistently on $U\Xi_{\star}$ by using for all $\xi,\eta\in U\Xi_{\star}$ the definitions
$$\displaystyle\Delta_{\star}(\xi\star\eta):=\Delta_{\star}(\xi)\star\Delta_{%
\star}(\eta)~{},\quad~{}S_{\star}(\xi\star\eta):=S_{\star}(\eta)\star S_{\star%
}(\xi)~{}.$$
(7)
The next step is to define the $\star$-Lie algebra of deformed infinitesimal diffeomorphisms. It has been shown [10]
that for the twist deformation case the choice $(\Xi_{\star},[~{},~{}]_{\star})$, where $\Xi_{\star}=\Xi$ as vector spaces and
$$\displaystyle[u,v]_{\star}:=[\bar{f}^{\alpha}(u),\bar{f}_{\alpha}(v)]$$
(8)
is a natural choice for a $\star$-Lie algebra. It fulfills all conditions which are necessary for a sensible $\star$-Lie algebra given by
1.
$\Xi_{\star}\subset U\Xi_{\star}$ is a linear space, which generates $U\Xi_{\star}$
2.
$\Delta_{\star}(\Xi_{\star})\subseteq\Xi_{\star}\otimes 1+U\Xi_{\star}\otimes%
\Xi_{\star}$
3.
$[\Xi_{\star},\Xi_{\star}]_{\star}\subseteq\Xi_{\star}$
The advantage of using the $\star$-Hopf algebra $(U\Xi_{\star},\star,\Delta_{\star},S_{\star},\epsilon_{\star})$
instead of the $\mathcal{F}$-Hopf algebra $(U\Xi_{\mathcal{F}},\cdot,\Delta_{\mathcal{F}},S_{\mathcal{F}},\epsilon_{%
\mathcal{F}})$
is that the $\star$-Lie algebra of vector fields is isomorphic to $\Xi$ as a vector space.
For the $\mathcal{F}$-Hopf algebra this is not the case and the $\mathcal{F}$-Lie algebra consists in general of multidifferential
operators.
The algebra of tensor fields $\mathcal{T}$ is deformed by using the $\star$-tensor product [10]
$$\displaystyle\tau\otimes_{\star}\tau^{\prime}:=\bar{f}^{\alpha}(\tau)\otimes%
\bar{f}_{\alpha}(\tau^{\prime})~{},$$
(9)
where as basic ingredients the deformed algebra of functions $A_{\star}:=(C^{\infty}(M),\star)$ as well as the $A_{\star}$-bimodules of
vector fields $\Xi_{\star}$ and one-forms $\Omega_{\star}$ enter. We call $\mathcal{T}_{\star}$ the deformed algebra of tensor fields. Note that
$\mathcal{T}_{\star}=\mathcal{T}$ as vector spaces.
The action of the deformed infinitesimal diffeomorphisms on $\mathcal{T}_{\star}$ is defined by the $\star$-Lie derivative
$$\displaystyle\mathcal{L}^{\star}_{u}(\tau):=\mathcal{L}_{\bar{f}^{\alpha}(u)}(%
\bar{f}_{\alpha}(\tau))~{},~{}\forall\tau\in\mathcal{T}_{\star}~{},~{}u\in\Xi_%
{\star}~{},$$
(10)
which can be extended to all of $U\Xi_{\star}$ by $\mathcal{L}^{\star}_{\xi\star\eta}(\tau):=\mathcal{L}^{\star}_{\xi}(\mathcal{L%
}^{\star}_{\eta}(\tau))$.
Furthermore, we define the $\star$-pairing $\langle\cdot,\cdot\rangle_{\star}:\Xi_{\star}\otimes_{\mathbb{C}}\Omega_{\star%
}\to A_{\star}$ between
vector fields and one-forms as
$$\displaystyle\langle v,\omega\rangle_{\star}:=\langle\bar{f}^{\alpha}(v),\bar{%
f}_{\alpha}(\omega)\rangle~{},~{}\forall v\in\Xi_{\star},~{}\omega\in\Omega_{%
\star}~{},$$
(11)
where $\langle\cdot,\cdot\rangle$ is the undeformed pairing.
Based on the deformed symmetry principle one can define covariant derivatives, torsion and curvature. This leads to deformed Einstein
equations, see [10], which we do not have to review here, since we do not use them in the following.
3 Symmetry Reduction in Twisted Differential Geometry
Assume that we have constructed a deformed gravity theory based on a twist $\mathcal{F}\in U\Xi\otimes U\Xi$. Like in Einstein
gravity, the physical applications of this theory is strongly dependent on symmetry reduction. In this section we first define
what we mean by symmetry reduction of a theory covariant under a Lie algebraic symmetry (e.g. infinitesimal diffeomorphisms)
and then extend the principles to deformed symmetries and $\star$-Lie algebras.
In undeformed general relativity we often face the fact that the systems we want to describe have certain (approximate) symmetries.
Here we restrict ourselves to Lie group symmetries.
For example in cosmology one usually constrains oneself to fields invariant under certain symmetry groups $G$, like e.g. the euclidian
group $E_{3}$ for flat universes or the $SO(4)$ group for universes with topology $\mathbb{R}\times S_{3}$, where the spatial hypersurfaces
are 3-spheres.
For a non rotating black hole one usually demands the metric to be stationary and spherically symmetric. Practically, one
uses the corresponding Lie algebra $\mathfrak{g}$ of the symmetry group $G$, represents it faithfully on the Lie algebra
of vector fields $\Xi$ on the manifold $\mathcal{M}$ and demands the fields $\tau\in\mathcal{T}$, which occur in the theory,
to be invariant under these transformations, i.e. we demand
$$\displaystyle\mathcal{L}_{v}(\tau)=0~{},~{}\forall v\in\mathfrak{g}~{}.$$
(12)
Since the Lie algebra $\mathfrak{g}$ is a linear space we can choose a basis $\{t_{i}:i=1,\cdots,\mathrm{dim}(\mathfrak{g})\}$
and can equivalently demand
$$\displaystyle\mathcal{L}_{t_{i}}(\tau)=0~{},~{}\forall i=1,2,\cdots,\mathrm{%
dim}(\mathfrak{g})~{}.$$
(13)
The Lie bracket of the generators has to fulfill
$$\displaystyle[t_{i},t_{j}]=f_{ij}^{~{}~{}k}t_{k}~{},$$
(14)
where $f_{ij}^{~{}~{}k}$ are the structure constants.
One can easily show that if we combine two invariant tensors with the tensor product, the resulting tensor is invariant too
because of the trivial coproduct
$$\displaystyle\mathcal{L}_{t_{i}}(\tau\otimes\tau^{\prime})=\mathcal{L}_{t_{i}}%
(\tau)\otimes\tau^{\prime}+\tau\otimes\mathcal{L}_{t_{i}}(\tau^{\prime})~{}.$$
(15)
The same holds true for pairings $\langle v,\omega\rangle$ of invariant objects $v\in\Xi$ and $\omega\in\Omega$.
Furthermore, if a tensor is invariant under infinitesimal transformations, it is also invariant under (at least small) finite transformations,
since they are given by exponentiating the generators. The exponentiated generators are part of the enveloping algebra,
i.e. $\exp(\alpha^{i}t_{i})\in U\mathfrak{g}$, where $\alpha^{i}$ are parameters. For large finite transformations the topology of the
Lie group can play a role, such that the group elements may not simply be given by exponentiating the generators. In the following
we will focus only on small finite transformations in order to avoid topological effects.
We now generalize this to the case of $\star$-Hopf algebras and their corresponding $\star$-Lie algebras. Our plan is
as follows: we start with a suitable definition of a $\star$-Lie subalgebra constructed from the Lie algebra $(\mathfrak{g},[~{},~{}])$.
This definition is guided by conditions, which allow for deformed symmetry reduction using infinitesimal transformations. Then
we complete this $\star$-Lie subalgebra in several steps to a $\star$-enveloping subalgebra, a $\star$-Hopf subalgebra
and a triangular $\star$-Hopf subalgebra. We will always be careful that the dimension of the $\star$-Lie subalgebra remains the same
as the dimension of the corresponding classical Lie algebra. At each step we obtain several restrictions between the twist
and $(\mathfrak{g},[~{},~{}])$.
We start by taking the generators $\{t_{i}\}$ of $\mathfrak{g}\subseteq\Xi$
and representing their deformations in the $\star$-Lie algebra $(\Xi_{\star},[~{},~{}]_{\star})$ as
$$\displaystyle t_{i}^{\star}=t_{i}+\sum\limits_{n=1}^{\infty}\lambda^{n}t_{i}^{%
(n)}~{},$$
(16)
where $\lambda$ is the deformation parameter and $t_{i}^{(n)}\in\Xi_{\star}$.
The span of these deformed generators, together with the $\star$-Lie bracket, should form a $\star$-Lie subalgebra $(\mathfrak{g}_{\star},[~{},~{}]_{\star}):=(\mathrm{span}(t_{i}^{\star}),[~{},~%
{}]_{\star})$. Therefore $(\mathfrak{g}_{\star},[~{},~{}]_{\star})$ has
to obey certain conditions. Natural conditions are
$$\displaystyle[\mathfrak{g_{\star}},\mathfrak{g}_{\star}]_{\star}\subseteq%
\mathfrak{g}_{\star},$$
i.e. $$[t_{i}^{\star},t_{j}^{\star}]_{\star}=f_{ij}^{\star~{}k}t^{\star}_{k}$$ with $$f_{ij}^{\star~{}k}=f_{ij}^{~{}~{}k}+\mathcal{O}(\lambda)$$
(17a)
$$\displaystyle\Delta_{\star}(\mathfrak{g}_{\star})\subseteq\mathfrak{g}_{\star}%
\otimes 1+U\Xi_{\star}\otimes\mathfrak{g}_{\star},$$
which is equivalent to $$\bar{R}_{\alpha}(\mathfrak{g}_{\star})\subseteq\mathfrak{g}_{\star}~{}\forall_%
{\alpha}$$
(17b)
The first condition is a basic feature of a $\star$-Lie algebra.
The second condition implies that if we have two $\mathfrak{g}_{\star}$ invariant tensors $\tau,\tau^{\prime}\in\mathcal{T}_{\star}$, the
$\star$-tensor product of them is invariant as well
$$\displaystyle\mathcal{L}^{\star}_{t_{i}^{\star}}(\tau\otimes_{\star}\tau^{%
\prime})=\mathcal{L}^{\star}_{t_{i}^{\star}}(\tau)\otimes_{\star}\tau^{\prime}%
+\bar{R}^{\alpha}(\tau)\otimes_{\star}\mathcal{L}^{\star}_{\bar{R}_{\alpha}(t_%
{i}^{\star})}(\tau^{\prime})=0~{},$$
(18)
since $\bar{R}_{\alpha}(t_{i}^{\star})\in\mathfrak{g}_{\star}$.
The $\star$-pairings $\langle v,\omega\rangle_{\star}$ of two invariant objects $v\in\Xi_{\star}$ and $\omega\in\Omega_{\star}$ are also invariant
under the $\star$-action of $\mathfrak{g}_{\star}$.
These are important features if one wants to combine invariant objects to e.g. an invariant action.
Furthermore, the conditions are sufficient such that the following consistency relation is fulfilled for
any invariant tensor $\tau\in\mathcal{T}_{\star}$
$$\displaystyle 0=f_{ij}^{\star~{}k}\mathcal{L}^{\star}_{t^{\star}_{k}}(\tau)=%
\mathcal{L}^{\star}_{[t^{\star}_{i},t^{\star}_{j}]}(\tau)=\mathcal{L}^{\star}_%
{t^{\star}_{i}}(\mathcal{L}^{\star}_{t^{\star}_{j}}(\tau))-\mathcal{L}^{\star}%
_{\bar{R}^{\alpha}(t^{\star}_{j})}(\mathcal{L}^{\star}_{\bar{R}_{\alpha}(t^{%
\star}_{i})}(\tau))~{},$$
(19)
since $\bar{R}_{\alpha}(t_{i}^{\star})\in\mathfrak{g}_{\star}$.
Hence by demanding the two conditions (17) for the $\star$-Lie subalgebra
$(\mathrm{span}(t_{i}^{\star}),[~{},~{}]_{\star})$ we can consistently perform symmetry reduction by using deformed
infinitesimal transformations. In the classical limit $\lambda\to 0$ we obtain the classical Lie algebra
$(\mathfrak{g}_{\star},[~{},~{}]_{\star})\stackrel{{\scriptstyle\lambda\to 0}}{%
{\longrightarrow}}(\mathfrak{g},[~{},~{}])$.
Next, we consider the extension of the $\star$-Lie subalgebra $(\mathfrak{g}_{\star},[~{},~{}]_{\star})\subseteq(\Xi_{\star},[~{},~{}]_{\star})$
to the triangular $\star$-Hopf subalgebra $\mathcal{H}_{\mathfrak{g}}^{\star}=(U\mathfrak{g}_{\star},\star,\Delta_{\star}%
,S_{\star},\epsilon_{\star})\subseteq\mathcal{H}_{\Xi}^{\star}$.
This can be seen as extending the infinitesimal transformations to a quantum group. We will divide this path into several steps,
where in every step we have to demand additional restrictions on the twist.
Firstly, we construct the $\star$-tensor algebra generated by the elements of $\mathfrak{g}_{\star}$ and $1$. We take this tensor algebra
modulo the left and right ideals generated by the elements $[u,v]_{\star}-u\star v+\bar{R}^{\alpha}(v)\star\bar{R}_{\alpha}(u)$.
It is necessary that these elements are part of $U\mathfrak{g}_{\star}$, i.e. we require
$$\displaystyle\bar{R}^{\alpha}(\mathfrak{g}_{\star})\star\bar{R}_{\alpha}(%
\mathfrak{g}_{\star})\subseteq U\mathfrak{g}_{\star}~{}.$$
(20)
This leads to the algebra $(U\mathfrak{g}_{\star},\star)$, which is a subalgebra of $(U\Xi_{\star},\star)$.
Secondly, we extend this subalgebra to a $\star$-Hopf subalgebra. Therefore we additionally have to require that
$$\displaystyle\Delta_{\star}(U\mathfrak{g}_{\star})$$
$$\displaystyle\subseteq U\mathfrak{g}_{\star}\otimes U\mathfrak{g}_{\star}~{},$$
(21a)
$$\displaystyle S_{\star}(U\mathfrak{g}_{\star})$$
$$\displaystyle\subseteq U\mathfrak{g}_{\star}~{}.$$
(21b)
Note that we do not demand that $S^{-1}_{\star}$ (defined on $U\Xi_{\star}$)
closes in $U\mathfrak{g}_{\star}$, since this is in general not the case for a nonquasitriangular Hopf algebra and we do not want to
demand quasitriangularity at this stage.
Then the $\star$-Hopf algebra $\mathcal{H}_{\mathfrak{g}}^{\star}$ is a Hopf subalgebra of $\mathcal{H}_{\Xi}^{\star}$.
Thirdly, we additionally demand that there exists an $R$-matrix $R_{\star}\in U\mathfrak{g}_{\star}\otimes U\mathfrak{g}_{\star}$. It is natural
to take the $R$-matrix of the triangular $\star$-Hopf algebra $\mathcal{H}_{\Xi}^{\star}$ defined by $R_{\star}:=X_{R^{\alpha}}\otimes X_{R_{\alpha}}$.
This leads to the restrictions
$$\displaystyle X_{R^{\alpha}},X_{R_{\alpha}}\in U\mathfrak{g}_{\star}~{},~{}%
\forall_{\alpha}.$$
(22)
Since $R_{\star}$ is triangular, i.e. $R_{\star}^{-1}=\bar{R}_{\star}^{\alpha}\otimes\bar{R}_{\star\alpha}=R_{\star 2%
1}=R_{\star\alpha}\otimes R_{\star}^{\alpha}$, we also have
$X_{\bar{R}^{\alpha}},X_{\bar{R}_{\alpha}}\in U\mathfrak{g}_{\star}~{},~{}%
\forall_{\alpha}$.
If these conditions are fulfilled, $\mathcal{H}_{\mathfrak{g}}^{\star}$ is a triangular $\star$-Hopf subalgebra of $\mathcal{H}_{\Xi}^{\star}$
with the same $R$-matrix.
As we have seen, extending the $\star$-Lie subalgebra to a (triangular) $\star$-Hopf subalgebra gives severe restrictions on
the possible deformations, more than just working with the deformed infinitesimal transformations given by a $\star$-Lie subalgebra
or the finite transformations given by the $\star$-enveloping subalgebra $(U\mathfrak{g}_{\star},\star)$.
Now the question arises if we actually require the deformed finite transformations to form a (triangular) $\star$-Hopf algebra in order to
use them for a sensible symmetry reduction. Because $(U\mathfrak{g}_{\star},\star)$ describes deformed finite transformations and we
have the relation
$$\displaystyle\mathcal{L}^{\star}_{U\mathfrak{g}_{\star}\backslash\{1\}}(\tau)=%
\{0\}\Leftrightarrow\mathcal{L}^{\star}_{\mathfrak{g}_{\star}}(\tau)=\{0\}~{},$$
(23)
we can consistently demand tensors to be invariant under $(U\mathfrak{g}_{\star},\star)$, since we
require tensors to be invariant under $(\mathfrak{g}_{\star},[~{},~{}]_{\star})$.
Therefore, a well defined $(U\mathfrak{g}_{\star},\star)$ leads to a structure sufficient for symmetry reduction.
The equivalence (23) can be shown by using linearity
of the $\star$-Lie derivative and the property $\mathcal{L}^{\star}_{\xi\star\eta}(\tau)=\mathcal{L}^{\star}_{\xi}(\mathcal{L}%
^{\star}_{\eta}(\tau))$.
In order to better understand the different restrictions necessary for constructing the $\star$-Lie subalgebra $(\mathfrak{g}_{\star},[~{},~{}]_{\star})$,
the $\star$-enveloping subalgebra and the (triangular) $\star$-Hopf subalgebra $(U\mathfrak{g}_{\star},\star,\Delta_{\star},S_{\star},\epsilon_{\star})$,
we restrict ourselves in the following sections to the class of Reshetikhin-Jambor-Sykora twists [13, 14]. This is a
suitable nontrivial generalization of the Moyal-Weyl product, also containing e.g. $\kappa$ and $q$ deformations when applied to Poincaré symmetry.
4 The Case of Reshetikhin-Jambor-Sykora Twists
Let $\{V_{a}\in\Xi\}$ be an arbitrary set of mutually commuting vector fields,
i.e. $[V_{a},V_{b}]=0~{},~{}\forall_{a,b}$, on an $n$ dimensional manifold $\mathcal{M}$. Then the object
$$\displaystyle\mathcal{F}_{V}:=\exp\bigl{(}-\frac{i\lambda}{2}\theta^{ab}V_{a}%
\otimes V_{b}\bigr{)}\in U\Xi\otimes U\Xi$$
(24)
is a twist element, if $\theta$ is constant and antisymmetric [10, 13, 14].
We call (24) a Reshetikhin-Jambor-Sykora twist. Note that this twist is not restricted to the topology $\mathbb{R}^{n}$ for the manifold
$\mathcal{M}$.
Furthermore, we can restrict ourselves to $\theta$ with maximal rank and an even number of vector fields $V_{a}$,
since we can lower the rank of the Poisson structure afterwards by choosing some of the $V_{a}$ to be zero.
We can therefore without loss of generality use the standard form
$$\displaystyle\theta=\begin{pmatrix}0&1&0&0&\cdots\\
-1&0&0&0&\cdots\\
0&0&0&1&\cdots\\
0&0&-1&0&\cdots\\
\vdots&\vdots&\vdots&\vdots&\ddots\end{pmatrix}$$
(25)
by applying a suitable $GL(n)$ transformation on the $V_{a}$.
This twist element is easy to apply and in particular we have for the inverse and the $R$-matrix
$$\displaystyle\mathcal{F}_{V}^{-1}=\exp\bigl{(}\frac{i\lambda}{2}\theta^{ab}V_{%
a}\otimes V_{b}\bigr{)}~{},\quad~{}R=\mathcal{F}_{V,21}\mathcal{F}_{V}^{-1}=%
\mathcal{F}_{V}^{-2}=\exp\bigl{(}i\lambda\theta^{ab}V_{a}\otimes V_{b}\bigr{)}%
~{}.$$
(26)
Now let $(\mathfrak{g},[~{},~{}])\subseteq(\Xi,[~{},~{}])$ be the Lie algebra of the symmetry we want to deform. We choose a
basis of this Lie algebra $\{t_{i}:i=1,\cdots,\mathrm{dim}(\mathfrak{g})\}$ with $[t_{i},t_{j}]=f_{ij}^{~{}~{}k}t_{k}$.
Next, we discuss the symmetry reduction based on the $\star$-Lie subalgebra, as explained in section 3.
Therefore we make the ansatz (16) for the generators $t_{i}^{\star}$.
Furthermore, we evaluate the two conditions (17) the $t_{i}^{\star}$ have to satisfy.
We start with the coproduct condition (17b), which
is equivalent to $\bar{R}_{\alpha}(t_{i}^{\star})\in\mathrm{span}(t_{i}^{\star}),~{}\forall_{\alpha}$, where $\alpha$ is a multi index. Using the
explicit form of the inverse $R$-matrix (26) we arrive at the conditions
$$\displaystyle[V_{a_{1}},\cdots[V_{a_{n}},t_{i}^{\star}]\cdots]=\mathcal{N}_{a_%
{1}\cdots a_{n}i}^{\star j}t_{j}^{\star}~{},$$
(27)
where $\mathcal{N}_{a_{1}\cdots a_{n}i}^{\star j}:=\mathcal{N}_{a_{1}\cdots a_{n}i}^{%
j}+\sum\limits_{k=1}^{\infty}\lambda^{k}\mathcal{N}_{a_{1}\cdots a_{n}i}^{(k)~%
{}j}$ are constants.
The only independent condition in (27) is given by
$$\displaystyle[V_{a},t_{i}^{\star}]=\mathcal{N}^{\star j}_{ai}t^{\star}_{j}~{},$$
(28)
since it implies all the other ones by linearity.
In particular, the zeroth order in $\lambda$ of (28) yields
$$\displaystyle[V_{a},t_{i}]=\mathcal{N}_{ai}^{j}t_{j}~{}.$$
(29)
This leads to the following
Proposition 1.
Let $(\mathfrak{g},[~{},~{}])\subseteq(\Xi,[~{},~{}])$ be a classical Lie algebra and $(\Xi_{\star},[~{},~{}]_{\star})$ the
$\star$-Lie algebra of vector fields deformed by a Reshetikhin-Jambor-Sykora twist, constructed with vector fields $V_{a}$.
Then for a symmetry reduction respecting the minimal axioms (17), it is necessary that
the following Lie bracket relations hold true
$$\displaystyle[V_{a},\mathfrak{g}]\subseteq\mathfrak{g}~{},\forall_{a}~{}.$$
(30)
In other words, $(\mathrm{span}(t_{i},V_{a}),[~{},~{}])\subseteq(\Xi,[~{},~{}])$ forms a Lie algebra with ideal $\mathfrak{g}$.
Here $t_{i}$ are the generators of $\mathfrak{g}$.
Note that this gives conditions relating the classical Lie algebra $(\mathfrak{g},[~{},~{}])$ with the twist.
Next, we evaluate the $\star$-Lie bracket condition (17a).
Using the explicit form of the inverse twist (26) and (28) we obtain
$$\displaystyle\bar{f}_{\alpha_{(n)}}(t_{i}^{\star})=[V_{a_{1}},\cdots[V_{a_{n}}%
,t_{i}^{\star}]\cdots]=\bigl{(}\mathcal{N}^{\star}_{a_{n}}\cdots\mathcal{N}^{%
\star}_{a_{1}}\bigr{)}_{i}^{j}t_{j}^{\star}=:\mathcal{N}^{\star j}_{\alpha_{(n%
)}i}t_{j}^{\star}~{},$$
(31a)
$$\displaystyle\bar{f}^{\alpha_{(n)}}(t_{i}^{\star})=\Theta^{\beta_{(n)}\alpha_{%
(n)}}\bar{f}_{\beta_{(n)}}(t_{i}^{\star})~{},$$
(31b)
$$\displaystyle\Theta^{\beta_{(n)}\alpha_{(n)}}:=\frac{1}{n!}\left(\frac{i%
\lambda}{2}\right)^{n}\theta^{b_{1}a_{1}}\cdots\theta^{b_{n}a_{n}}~{},$$
(31c)
where $\alpha_{(n)},\beta_{(n)}$ are multi indices. This leads to
$$\displaystyle[t_{i}^{\star},t_{j}^{\star}]_{\star}=\Theta^{\beta_{(n)}\alpha_{%
(n)}}\mathcal{N}^{\star k}_{\beta_{(n)}i}\mathcal{N}^{\star l}_{\alpha_{(n)}j}%
[t_{k}^{\star},t_{l}^{\star}]~{}.$$
(32)
Note that in particular for the choice $t_{i}^{\star}=t_{i},~{}\forall_{i},$ the $\star$-Lie subalgebra closes with structure constants
$$\displaystyle[t_{i},t_{j}]_{\star}=\Theta^{\beta_{(n)}\alpha_{(n)}}\mathcal{N}%
^{k}_{\beta_{(n)}i}\mathcal{N}^{l}_{\alpha_{(n)}j}[t_{k},t_{l}]=\Theta^{\beta_%
{(n)}\alpha_{(n)}}\mathcal{N}^{k}_{\beta_{(n)}i}\mathcal{N}^{l}_{\alpha_{(n)}j%
}f_{kl}^{~{}~{}m}t_{m}=:f_{ij}^{\star~{}m}t_{m}~{},$$
(33)
where we have used the $\mathcal{N}$ defined in (29).
This leads to the following
Proposition 2.
Let $[V_{a},\mathfrak{g}_{\star}]\subseteq\mathfrak{g}_{\star}~{},\forall_{a}$. Then we can always construct a $\star$-Lie subalgebra
$(\mathfrak{g}_{\star},[~{},~{}]_{\star})\subseteq(\Xi_{\star},[~{},~{}]_{\star})$
by choosing the generators as $t_{i}^{\star}=t_{i}$ for all $i$. With this we have $\mathfrak{g}_{\star}=\mathfrak{g}$ as vector spaces
and the structure constants are deformed as
$$\displaystyle f_{ij}^{\star~{}m}=\Theta^{\beta_{(n)}\alpha_{(n)}}\mathcal{N}^{%
k}_{\beta_{(n)}i}\mathcal{N}^{l}_{\alpha_{(n)}j}f_{kl}^{~{}~{}m}~{}.$$
(34)
Since the condition (17b) together with the requirement $t_{i}^{\star}=t_{i}$, for all $i$, automatically fulfills
(17a), we choose $t_{i}^{\star}=t_{i}$, for all $i$, as a
canonical embedding. In general, other possible embeddings
require further constructions to fulfill condition
(17a) and are therefore less natural.
We will discuss possible differences between this and other embeddings later on, when we construct the $\star$-Hopf subalgebra and the
$\star$-Lie derivative action on $\star$-tensor fields.
In addition, we obtain that the necessary condition (20) for extending $\mathfrak{g}_{\star}$ to the $\star$-enveloping
subalgebra $(U\mathfrak{g}_{\star},\star)\subseteq(U\Xi_{\star},\star)$
is automatically fulfilled, since we have $\bar{R}_{\alpha_{(n)}}(\mathfrak{g}_{\star})\subseteq\mathfrak{g}_{\star}$
for all $\alpha_{(n)}$ and additionally
$$\displaystyle\bar{R}^{\alpha_{(n)}}(\mathfrak{g}_{\star})=(-2)^{n}\Theta^{%
\beta_{(n)}\alpha_{(n)}}\bar{R}_{\beta_{(n)}}(\mathfrak{g}_{\star})\subseteq%
\mathfrak{g}_{\star}~{},~{}\forall{\alpha}_{(n)}.$$
(35)
Next, we evaluate the conditions (21), which have to be fulfilled in order to construct the $\star$-Hopf subalgebra
$\mathcal{H}_{\mathfrak{g}}^{\star}\subseteq\mathcal{H}_{\Xi}^{\star}$. For the particular choice of the twist (24)
we obtain the following
Proposition 3.
Let $(U\mathfrak{g}_{\star},\star)\subseteq(U\Xi_{\star},\star)$ be a $\star$-enveloping subalgebra and let the deformation parameter
$\lambda\neq 0$.
Then in order to extend $(U\mathfrak{g}_{\star},\star)$ to the $\star$-Hopf subalgebra $\mathcal{H}_{\mathfrak{g}}^{\star}=(U\mathfrak{g}_{\star},\star,\Delta_{\star}%
,S_{\star},\epsilon_{\star})\subseteq\mathcal{H}_{\Xi}^{\star}$ the
condition
$$\displaystyle V_{a_{1}}\in\mathfrak{g}_{\star}~{},\quad\text{if }~{}[V_{a_{2}}%
,\mathfrak{g}_{\star}]\neq\{0\}$$
(36)
has to hold true for all pairs of indices $(a_{1},a_{2})$ connected by the antisymmetric matrix $\theta$ (25), i.e. $(a_{1},a_{2})\in\big{\{}(1,2),(2,1),(3,4),(4,3),\dots\big{\}}$.
Note that these conditions depend on the embedding $t_{i}^{\star}=t_{i}^{\star}(t_{j})$.
The proof of this proposition is shown in the appendix A.
Finally, if we demand $\mathcal{H}_{\mathfrak{g}}^{\star}$ to be a triangular $\star$-Hopf algebra (22) we obtain the stringent condition
$$\displaystyle V_{a}\in\mathfrak{g}_{\star}~{},~{}\forall_{a}~{}.$$
(37)
This can be shown by using $X_{R_{\alpha}}=R_{\alpha}$ and $V_{a}\star V_{b}=V_{a}V_{b}$, which holds true for the class of Reshetikhin-Jambor-Sykora twists.
As we have seen above, there are much stronger restrictions on the Lie algebra $(\mathfrak{g},[~{},~{}])$
and the twist, if we want to extend the deformed infinitesimal transformations $(\mathfrak{g}_{\star},[~{},~{}]_{\star})$
to the (triangular) $\star$-Hopf subalgebra $\mathcal{H}_{\mathfrak{g}}^{\star}$. In particular this extension restricts the $V_{a}$ themselves,
while for infinitesimal transformations and the finite transformations $(U\mathfrak{g}_{\star},\star)$ only the images of $V_{a}$
acting on $\mathfrak{g}_{\star}$ are important.
Next, we study the $\star$-action of the $\star$-Lie and Hopf algebra on the deformed tensor fields. The $\star$-action of the generators
$t_{i}^{\star}$ on $\tau\in\mathcal{T}_{\star}$ is defined by (10) and simplifies to
$$\displaystyle\mathcal{L}^{\star}_{t_{i}^{\star}}(\tau)=\Theta^{\alpha_{(n)}%
\beta_{(n)}}\mathcal{N}^{\star j}_{\alpha_{(n)}i}~{}\mathcal{L}_{t_{j}^{\star}%
}\bigl{(}\bar{f}_{\beta_{(n)}}(\tau)\bigr{)}~{}.$$
(38)
For invariant tensors, the $\star$-Lie derivative has to vanish to all orders in $\lambda$, since we work with formal power series.
If we now for explicitness take the natural choice $t_{i}^{\star}=t_{i}$ we obtain the following
Proposition 4.
Let $[V_{a},\mathfrak{g}_{\star}]\subseteq\mathfrak{g}_{\star}~{},\forall_{a}$ and $t_{i}^{\star}=t_{i},~{}\forall_{i}$. Then a tensor $\tau\in\mathcal{T}_{\star}$ is $\star$-invariant
under $(\mathfrak{g}_{\star},[~{},~{}]_{\star})$, if and only if it is invariant under the undeformed action of $(\mathfrak{g},[~{},~{}])$, i.e.
$$\displaystyle\mathcal{L}^{\star}_{\mathfrak{g}_{\star}}(\tau)=\{0\}~{}%
\Leftrightarrow~{}\mathcal{L}_{\mathfrak{g}}(\tau)=\{0\}~{}.$$
(39)
Proof.
For the proof we make the ansatz $\tau=\sum\limits_{n=0}^{\infty}\lambda^{n}\tau_{n}$ and investigate $\mathcal{L}^{\star}_{t_{i}}(\tau)$
order by order in $\lambda$, since we work with formal power series. By using (29) to reorder the
Lie derivatives such that $t_{i}$ is moved to the right, it can be shown recursively in powers of $\lambda$ that the proposition
holds true.
∎
Note that for $t_{i}^{\star}\neq t_{i}$ this does not necessarily hold true. We can not make statements for this case, since
we would require a general solution of (32), which we do not have yet. But we mention again that
we consider choosing $t_{i}^{\star}$ different from $t_{i}$ quite unnatural.
This proposition translates to the case of finite symmetry transformations with $t_{i}^{\star}=t_{i}$ because of the properties of
the $\star$-Lie derivative.
The framework developed in this section will now be applied to
cosmology and black holes in order to give some specific examples and
discuss possible physical implications.
5 Application to Cosmology
In this section we will investigate models with symmetry group $E_{3}$ in four spacetime dimensions with topology $\mathbb{R}^{4}$.
These are flat Friedmann-Robertson-Walker (FRW) universes.
The undeformed Lie algebra of this group is generated by the “momenta” $p_{i}$ and “angular momenta” $L_{i}$, $i\in\{1,2,3\}$,
which we can represent in the Lie algebra of vector fields as
$$\displaystyle p_{i}=\partial_{i}\quad,\quad L_{i}=\epsilon_{ijk}x^{j}\partial_%
{k}~{},$$
(40)
where $\epsilon_{ijk}$ is the Levi-Civita symbol.
The undeformed Lie bracket relations are
$$\displaystyle[p_{i},p_{j}]=0~{},\quad[p_{i},L_{j}]=-\epsilon_{ijk}p_{k}~{},%
\quad[L_{i},L_{j}]=-\epsilon_{ijk}L_{k}~{}.$$
(41)
We will work with the natural embedding $t_{i}^{\star}=t_{i}$, and therefore the $\star$-Lie subalgebra is given by
$\mathfrak{g}_{\star}=\mathfrak{e}_{3\star}=\mathfrak{e}_{3}=\mathrm{span}(p_{i%
},L_{i})$.
We can now explicitly evaluate the condition each twist vector field $V_{a}$ has to satisfy given by
$[V_{a},\mathfrak{e}_{3}]\subseteq\mathfrak{e}_{3}$ (cf. proposition 1).
Since the generators are at most linear in the spatial coordinates, $V_{a}$ can be at most quadratic in order to
fulfill this condition. If we make a quadratic ansatz with time dependent coefficients we obtain that each $V_{a}$
has to be of the form
$$\displaystyle V_{a}=V_{a}^{0}(t)\partial_{t}+c_{a}^{i}\partial_{i}+d_{a}^{i}L_%
{i}+f_{a}x^{i}\partial_{i}~{},$$
(42)
where $c_{a}^{i},~{}d_{a}^{i},~{}f_{a}\in\mathbb{R}$ and $V_{a}^{0}(t)\in C^{\infty}(\mathbb{R})$ in order to obtain hermitian deformations.
If all $V_{a}$ have the form (42), the $\star$-Lie algebra closes (cf. proposition 2).
Next, we have to find conditions such that the $V_{a}$ are mutually commuting. A brief calculation shows that the
following conditions have to be fulfilled:
$$\displaystyle d_{a}^{i}d_{b}^{j}\epsilon_{ijk}=0~{},\forall_{k}~{},$$
(43a)
$$\displaystyle c_{a}^{i}d_{b}^{j}\epsilon_{ijk}-c_{b}^{i}d_{a}^{j}\epsilon_{ijk%
}+f_{a}c_{b}^{k}-f_{b}c_{a}^{k}=0~{},\forall_{k}~{},$$
(43b)
$$\displaystyle[V_{a}^{0}(t)\partial_{t},V_{b}^{0}(t)\partial_{t}]=0~{}.$$
(43c)
As a first step, we will now work out all possible deformations of $\mathfrak{e}_{3}$ when twisted with two commuting vector fields.
We will classify the possible solutions. Therefore we divide the solutions into classes depending on the value of $d_{a}^{i}$ and $f_{a}$.
We use as notation for our cosmologies $\mathfrak{C}_{AB}$, where $A\in\{1,2,3\}$ and $B\in\{1,2\}$,
which will become clear later on, when we sum up the results in table 1.
Type $\mathfrak{C}_{11}$ is defined to be vector fields with $d_{1}^{i}=d_{2}^{i}=0$ and $f_{1}=f_{2}=0$, i.e.
$$\displaystyle V_{1(\mathfrak{C}_{11})}=V_{1}^{0}(t)\partial_{t}+c_{1}^{i}%
\partial_{i}~{},\quad V_{2(\mathfrak{C}_{11})}=V_{2}^{0}(t)\partial_{t}+c_{2}^%
{i}\partial_{i}~{}.$$
(44)
These vector fields fulfill the first two conditions (43a) and (43b). The solutions of the third condition
(43c) will be discussed later, since this classification we perform now does not depend on it.
Type $\mathfrak{C}_{21}$ is defined to be vector fields with $d_{1}^{i}=d_{2}^{i}=0$, $f_{1}\neq 0$ and $f_{2}=0$.
The first condition (43a) is trivially fulfilled and the second (43b) is fulfilled,
if and only if $c_{2}^{i}=0,~{}\forall_{i}$, i.e. type $\mathfrak{C}_{21}$ is given by the vector fields
$$\displaystyle\tilde{V}_{1(\mathfrak{C}_{21})}=V_{1}^{0}(t)\partial_{t}+c_{1}^{%
i}\partial_{i}+f_{1}x^{i}\partial_{i}~{},\quad\tilde{V}_{2(\mathfrak{C}_{21})}%
=V_{2}^{0}(t)\partial_{t}~{}.$$
(45)
These vector fields can be simplified to
$$\displaystyle V_{1(\mathfrak{C}_{21})}=c_{1}^{i}\partial_{i}+f_{1}x^{i}%
\partial_{i}~{},\quad V_{2(\mathfrak{C}_{21})}=V_{2}^{0}(t)\partial_{t}~{},$$
(46)
since both lead to the same twist (24).
Solutions with $d_{1}^{i}=d_{2}^{i}=0$, $f_{1}\neq 0$ and $f_{2}\neq 0$ lie in type $\mathfrak{C}_{21}$, since we can perform the twist conserving map
$V_{2}\to V_{2}-\frac{f_{2}}{f_{1}}V_{1}$, which transforms $f_{2}$ to zero. Furthermore $\mathfrak{C}_{31}$ is defined by
$d_{1}^{i}=d_{2}^{i}=0$, $f_{1}=0$ and $f_{2}\neq 0$ and is equivalent to $\mathfrak{C}_{21}$ by interchanging the labels of the vector fields.
Next, we go on to solutions with without loss of generality $\mathbf{d}_{1}\neq 0$ and $\mathbf{d}_{2}=0$ ($\mathbf{d}$ denotes the vector).
Note that this class contains also the class with $\mathbf{d}_{1}\neq 0$ and $\mathbf{d}_{2}\neq 0$.
To see this, we use the first condition (43a) and obtain that $\mathbf{d}_{1}$ and $\mathbf{d}_{2}$ have to be parallel,
i.e. $d^{i}_{2}=\kappa d^{i}_{1}$.
Then we can transform $\mathbf{d}_{2}$ to zero by using the twist conserving map $V_{2}\to V_{2}-\kappa V_{1}$.
Type $\mathfrak{C}_{12}$ is defined to be vector fields with $\mathbf{d}_{1}\neq 0$, $\mathbf{d}_{2}=0$ and $f_{1}=f_{2}=0$.
The first condition (43a) is trivially fulfilled, while the second condition (43b)
requires that $\mathbf{c}_{2}$ is parallel to $\mathbf{d}_{1}$, i.e. we obtain
$$\displaystyle V_{1(\mathfrak{C}_{12})}=V_{1}^{0}(t)\partial_{t}+c_{1}^{i}%
\partial_{i}+d_{1}^{i}L_{i}~{},\quad V_{2(\mathfrak{C}_{12})}=V_{2}^{0}(t)%
\partial_{t}+\kappa~{}d_{1}^{i}\partial_{i}~{},$$
(47)
where $\kappa\in\mathbb{R}$ is a constant.
Type $\mathfrak{C}_{22}$ is defined to be vector fields with $\mathbf{d}_{1}\neq 0$, $\mathbf{d}_{2}=0$, $f_{1}\neq 0$ and $f_{2}=0$.
Solving the second condition (43b) (therefore we have to use that the vectors are real!) we obtain
$$\displaystyle V_{1(\mathfrak{C}_{22})}=c_{1}^{i}\partial_{i}+d_{1}^{i}L_{i}+f_%
{1}x^{i}\partial_{i}~{},\quad V_{2(\mathfrak{C}_{22})}=V_{2}^{0}(t)\partial_{t%
}~{},$$
(48)
where we could set without loss of generality $V_{1}^{0}(t)$ to zero, as in type $\mathfrak{C}_{21}$. Note that $\mathfrak{C}_{21}$ is
contained in $\mathfrak{C}_{22}$ by violating the condition $\mathbf{d}_{1}\neq 0$.
Finally, we come to the last class, type $\mathfrak{C}_{32}$, defined by $\mathbf{d}_{1}\neq 0$, $\mathbf{d}_{2}=0$, $f_{1}=0$ and $f_{2}\neq 0$.
This class contains also the case $\mathbf{d}_{1}\neq 0$, $\mathbf{d}_{2}=0$, $f_{1}\neq 0$ and $f_{2}\neq 0$ by using the twist
conserving map $V_{1}\to V_{1}-\frac{f_{1}}{f_{2}}V_{2}$. The vector fields are given by
$$\displaystyle V_{1(\mathfrak{C}_{32})}=V_{1}^{0}(t)\partial_{t}+\frac{d_{1}^{j%
}c_{2}^{k}\epsilon_{jki}}{f_{2}}\partial_{i}+d_{1}^{i}L_{i}~{},\quad V_{2(%
\mathfrak{C}_{32})}=V_{2}^{0}(t)\partial_{t}+c_{2}^{i}\partial_{i}+f_{2}x^{i}%
\partial_{i}~{}.$$
(49)
Note that type $\mathfrak{C}_{11}$ and $\mathfrak{C}_{12}$ can be extended to a triangular $\star$-Hopf
subalgebra by choosing $V_{1}^{0}(t)=V_{2}^{0}(t)=0$ in each case.
For a better overview we additionally present the the results in table 1, containing all possible two vector field
deformations $\mathfrak{C}_{AB}$ of the Lie algebra of the euclidian group.
From this table the notation $\mathfrak{C}_{AB}$ becomes clear.
Next, we discuss solutions to the third condition (43c) $[V_{1}^{0}(t)\partial_{t},V_{2}^{0}(t)\partial_{t}]=0$.
It is obvious that choosing either
$V_{1}^{0}(t)=0$ or $V_{2}^{0}(t)=0$ and the other one arbitrary is a solution. Additionally, we consider solutions with $V_{1}^{0}(t)\neq 0$ and $V_{2}^{0}(t)\neq 0$.
Therefore there has to be some point $t_{0}\in\mathbb{R}$, such that without loss of generality $V_{1}^{0}(t)$ is unequal zero in some open region
$U\subseteq\mathbb{R}$ around $t_{0}$.
In this region we can perform the diffeomorphism $t\to\tilde{t}(t):=\int\limits_{t_{0}}^{t}dt^{\prime}\frac{1}{V_{1}^{0}(t^{%
\prime})}$
leading to $\tilde{V}_{1}^{0}(\tilde{t})=1$. With this the third condition (43c) becomes
$$\displaystyle 0=[V_{1}^{0}(t)\partial_{t},V_{2}^{0}(t)\partial_{t}]=[\tilde{V}%
_{1}^{0}(\tilde{t})\partial_{\tilde{t}},\tilde{V}_{2}^{0}(\tilde{t})\partial_{%
\tilde{t}}]=\Bigl{(}\partial_{\tilde{t}}\tilde{V}_{2}^{0}(\tilde{t})\Bigr{)}%
\partial_{\tilde{t}}~{}.$$
(50)
This condition is solved if and only if $\tilde{V}_{2}^{0}(\tilde{t})=\mathrm{const.}$ for $t\in U\subseteq\mathbb{R}$.
For the subset of analytical functions $C^{\omega}(\mathbb{R})\subset C^{\infty}(\mathbb{R})$ we can continue this condition
to all $\mathbb{R}$ and obtain the global relation $V_{2}^{0}(t)=\kappa V_{1}^{0}(t)$, with some constant $\kappa\in\mathbb{R}$.
For non analytic, but smooth functions, we can not continue these relations to all $\mathbb{R}$ and therefore only obtain local conditions restricting the functions in the overlap of their supports to be linearly dependent.
In particular non analytic functions with disjoint supports fulfill the condition (43c) trivially.
After characterizing the possible two vector field deformations of $\mathfrak{e}_{3}$ we briefly give a method how to obtain twists
generated by a larger number of vector fields. For this purpose we use the canonical form of $\theta$ (25).
Assume that we want to obtain deformations with e.g. four vector fields. Then of course all vector fields have to be of the form (42).
According to the form of $\theta$ we have two blocks of vector fields $(a,b)=(1,2)$ and $(a,b)=(3,4)$, in which the classification described
above for two vector fields can be performed. This means that all four vector field twists can be obtained by using two types
of two vector field twists.
We label the twist by using a tuple of types, e.g. $(\mathfrak{C}_{11},\mathfrak{C}_{22})$ means that $V_{1},V_{2}$ are of type $\mathfrak{C}_{11}$ and $V_{3},V_{4}$ of type $\mathfrak{C}_{22}$.
But this does only assure that $[V_{a},V_{b}]=0$ for $(a,b)\in\{(1,2),(3,4)\}$ and we have to demand further restrictions in
order to fulfill $[V_{a},V_{b}]=0$ for all $(a,b)$ and that all vector fields give independent contributions
to the twist. In particular twists constructed with linearly dependent vector fields can be reduced to a twist
constructed by a lower number of vector fields.
This method naturally extends to a larger number of vector fields, until we cannot find anymore independent and
mutually commuting vector fields. We will now give two examples for the $\mathfrak{e}_{3}$ case in order to clarify the construction.
As a first example we construct the four vector field twist $(\mathfrak{C}_{11},\mathfrak{C}_{11})$.
In this case all four vector fields commute without imprinting further restrictions. We assume that three of the four vectors
$\mathbf{c}_{a}$ are linearly independent, such that the fourth one, say $\mathbf{c}_{4}$, can be decomposed into the other ones.
If we now choose four linearly independent functions $V_{a}^{0}(t)$ (this means that they are non analytic) leads to a proper four vector field
twist.
As a second simple example we construct the four vector field twist $(\mathfrak{C}_{21},\mathfrak{C}_{21})$.
In order to have commuting vector fields we obtain the condition $c_{3}^{i}=\frac{f_{3}}{f_{1}}c_{1}^{i}$.
We therefore have $V_{3}=\frac{f_{3}}{f_{1}}V_{1}$ and the four vector field twist can be reduced to the two vector field twist
of type $\mathfrak{C}_{21}$ with $\tilde{V}_{1}=V_{1}$ and $\tilde{V}_{2}=V_{2}+\frac{f_{3}}{f_{1}}V_{4}$. This is an example
of an improper four vector field twist.
This method can be applied in order to investigate general combinations of two vector field twists, if one requires them.
Because this construction is straightforward and we do not require these twists for our discussions, we do not present them here.
At the end we calculate the $\star$-commutator of the linear coordinate functions $x^{\mu}\in A_{\star}$ for the various types of models
in first order in the deformation parameter $\lambda$. It is given by
$$\displaystyle c^{\mu\nu}:=[x^{\mu}\stackrel{{\scriptstyle\star}}{{,}}x^{\nu}]:%
=x^{\mu}\star x^{\nu}-x^{\nu}\star x^{\mu}=i\lambda\theta^{ab}V_{a}(x^{\mu})V_%
{b}(x^{\nu})+\mathcal{O}(\lambda^{2})~{}.$$
(51)
The results are given in appendix B and show that these commutators can be at most quadratic in
the spatial coordinates $x^{i}$.
Possible applications of these models will be discussed in the outlook, see section 7.
6 Application to Black Holes
In this section we investigate possible deformations of non rotating black holes.
We will do this in analogy to the cosmological models and therefore do not have to explain every single step.
The undeformed Lie algebra of the symmetry group $\mathbb{R}\times SO(3)$ of a non rotating black hole is generated by
the vector fields
$$\displaystyle p^{0}=\partial_{t}\quad,\quad L_{i}=\epsilon_{ijk}x^{j}\partial_%
{k}~{},$$
(52)
given in cartesian coordinates. We choose $t_{i}^{\star}=t_{i}$ for all $i$
and define $\mathfrak{g}_{\star}=\mathfrak{g}=\mathrm{span}(p^{0},L_{i})$.
It can be shown that each twist vector field $V_{a}$ has to be of the form
$$\displaystyle V_{a}=(c^{0}_{a}(r)+N_{a}^{0}t)\partial_{t}+d_{a}^{i}L_{i}+f_{a}%
(r)x^{i}\partial_{i}$$
(53)
in order to fulfill $[V_{a},\mathfrak{g}]\subseteq\mathfrak{g}$. Here $r=\|\mathbf{x}\|$ is the euclidian
norm of the spatial position vector.
The next task is to construct the two vector field deformations. Therefore we additionally have to
demand $[V_{a},V_{b}]=0,~{}\forall_{a,b}$, leading to the conditions
$$\displaystyle d_{a}^{i}d_{b}^{j}\epsilon_{ijk}=0~{},~{}\forall_{k}~{},$$
(54a)
$$\displaystyle(f_{a}(r)x^{j}\partial_{j}-N_{a}^{0})c^{0}_{b}(r)-(f_{b}(r)x^{j}%
\partial_{j}-N_{b}^{0})c^{0}_{a}(r)=0~{},$$
(54b)
$$\displaystyle f_{a}(r)f_{b}^{\prime}(r)-f_{a}^{\prime}(r)f_{b}(r)=0~{},$$
(54c)
where $f_{a}^{\prime}(r)$ means the derivative of $f_{a}(r)$.
Note that (54c) is a condition similar to (43c), and therefore
has the same type of solutions. Because of this, the functions $f_{1}(r)$ and $f_{2}(r)$ have to be parallel in the overlap
of their supports. From this we can always eliminate locally one $f_{a}(r)$
by a twist conserving map and simplify the investigation of the condition (54b).
At the end, the local solutions have to be glued together.
We choose without loss of generality $f_{1}(r)=0$ for our classification of local solutions.
The solution to (54a) is that the $\mathbf{d}_{a}$ have to be parallel. We use
$$\displaystyle\mathbf{d}_{a}=\kappa_{a}\mathbf{d}$$
(55)
with constants $\kappa_{a}\in\mathbb{R}$ and some arbitrary vector $\mathbf{d}\neq 0$.
We now classify the solutions to (54b) according to $N_{a}^{0}$ and $f_{2}(r)$ and label them by $\mathfrak{B}_{AB}$.
We distinguish between $f_{2}(r)$ being the zero function or not. The result is shown in table 2.
Other choices of parameters can be mapped by a twist conserving map into these classes.
Note that in particular for analytical functions $f_{a}(r)$ the twist conserving map transforming $f_{1}(r)$ to zero
can be performed globally, and with this also the classification of twists given in table 2.
In type $\mathfrak{B}_{32}$ we still have to solve a differential equation for $c_{1}^{0}(r)$ given by
$$\displaystyle c_{1}^{0}(r)=\frac{f_{2}(r)}{N_{2}^{0}}rc_{1}^{0\prime}(r)~{},$$
(56)
for an arbitrary given $f_{2}(r)$.
We will not work out the solutions to this differential equation, since type $\mathfrak{B}_{32}$ is a quite unphysical model,
in which the noncommutativity is increasing linear in time due to $N_{2}^{0}\neq 0$.
Note that $\mathfrak{B}_{11}$ can be extended to a triangular $\star$-Hopf algebra by choosing $c_{a}^{0}(r)=c_{a}^{0}$, for $a\in\{1,2\}$.
In addition, $\mathfrak{B}_{12}$ is a $\star$-Hopf algebra for $\kappa_{1}=\kappa_{2}=0$.
The $\star$-commutators $c^{\mu\nu}=[x^{\mu}\stackrel{{\scriptstyle\star}}{{,}}x^{\nu}]$ of the coordinate functions $x^{\mu}\in A_{\star}$ in order $\lambda^{1}$ for these
models are given in the appendix B. They can be used in order to construct sensible
physical models of a noncommutative black hole.
By using the method explained in the previous section, the two vector field twists can be extended to multiple vector field
twists. Since we do not require these twists in our work and their construction is straightforward, we do not
present them here.
7 Conclusion and Outlook
We have discussed symmetry reduction in noncommutative gravity using
the formalism of twisted noncommutative differential geometry. Our
motivation for these investigations derives from the fact that, for
most physical applications of gravity theories, including cosmology,
symmetry reduction is required due to the complexity of such models,
already in the undeformed case.
In section 3 we have presented a general method for
symmetry reduction in twisted gravity theories. As a result we have
obtained restrictions on the twist, depending on the structure of the
twisted symmetry group. In particular, we find that deforming the
infinitesimal symmetry transformations results in weaker restrictions
than deforming the finite transformations and demanding a quantum
group structure. In section 4 we have applied this
general method to gravity theories twisted by Reshetikhin-Jambor-Sykora twists.
These are twists constructed from commuting vector fields. In this
case we could give explicit conditions, which have to be fulfilled in
order to allow symmetry reduction of a given Lie group.
In sections 5 and 6 we have
investigated admissible deformations of FRW and black hole symmetries
by a Reshetikhin-Jambor-Sykora twist. In this class we have classified all
possible deformations. This lays the foundation for phenomenological
studies of noncommutative cosmology and black hole physics based on
twisted gravity.
In a forthcoming work [15] we will
investigate cosmological implications of twisted FRW models by
studying fluctuations of quantum fields living on twisted FRW
backgrounds. Quantum fields were already introduced in a twisted
framework in [16]. As we see from
proposition 4, the noncommutative backgrounds
are also invariant under the undeformed action of the classical
symmetry. This means that they have the same coordinate
representations with respect to the undeformed basis vectors as the
commutative fields in Einstein gravity. With this we have a construction
principle for noncommutative backgrounds, in their natural basis,
by representing the classical fields in the deformed basis.
A class of models of particular interest is
type $\mathfrak{C}_{22}$ in section 5
(cf. table 1). These twists break classical translation
invariance, but classical rotation invariance can be retained by
tuning $\mathbf{d}_{1}$ and $\mathbf{c}_{1}$ to small values. Furthermore,
the global factor $V_{2}^{0}(t)$ in the exponent of the twist can be used
in order to tune noncommutativity effects depending on time.
Obviously, enforcing a suitable $V_{2}^{0}(t)$ by hand leads to
phenomenologically valid models.
Since there is no natural choice of $V_{2}^{0}(t)$, it is interesting to
investigate the dynamics of $V_{2}^{0}(t)$ in a given field configuration
and study if it leads to a model consistent with cosmological
observations. In this case, the model would be physically
attractive. This will also be subject of future
work [15]. Dynamical noncommutativity has
already been studied in the case of scalar field theories on Minkowski
spacetime [17].
In the case of black hole physics, models of particular interest would
be $\mathfrak{B}_{11}$ with functions $c^{0}_{a}(r)$ decreasing
sufficiently quickly with $r$ and $\mathfrak{B}_{12}$ with $f_{2}(r)$
and $c_{2}^{0}(r)$ decreasing sufficiently quickly with $r$
(cf. table 2). It will again be interesting to
investigate the dynamics of these functions on a given field
configuration. Note that the type $\mathfrak{B}_{12}$
with $\kappa_{1}=\kappa_{2}=0$ is invariant under the classical black hole
symmetries, and therefore particularly interesting for physical
applications. On the other hand, models with nonvanishing $N_{a}^{0}$ are
of little physical interest, because the noncommutativity is growing
linearly in time, which would be unphysical.
Other avenues for future work are the classification of models on
nontrivial topologies (like, e.g., $\mathbb{R}\times S_{3}$ in
cosmology), investigating nontrivial
embeddings $t_{i}^{\star}=t_{i}^{\star}(t_{j})$ and using a wider class of twist
elements.
Acknowledgements
AS thanks Christoph Uhlemann and Julian Adamek for discussions and
comments on this work. This research is supported by Deutsche
Forschungsgemeinschaft through the Research Training Group 1147
Theoretical Astrophysics and Particle Physics.
Appendix A Proof of Proposition 3
In this appendix we show that for Reshetikhin-Jambor-Sykora twists (24) the conditions (21) necessary for
extending the $\star$-enveloping subalgebra $(U\mathfrak{g}_{\star},\star)$ to a $\star$-Hopf subalgebra are
equivalent to the simplified conditions of proposition 3.
The plan is as follows: we use (21a) and show that
it is equivalent to the conditions of proposition 3. In a second step, we show that (21b)
is automatically satisfied if (21a) is fulfilled, and thus does not lead to additional conditions.
We start with (21a) and show the identity
$$\displaystyle\Delta_{\star}(U\mathfrak{g}_{\star})\subseteq U\mathfrak{g}_{%
\star}\otimes U\mathfrak{g}_{\star}~{}\Leftrightarrow~{}\Delta_{\star}(%
\mathfrak{g}_{\star})\subseteq U\mathfrak{g}_{\star}\otimes U\mathfrak{g}_{%
\star}~{}.$$
(57)
The direction $\Rightarrow$ is trivial, since $\mathfrak{g}_{\star}\subset U\mathfrak{g}_{\star}$, and the direction $\Leftarrow$ can be shown
using that $\Delta_{\star}$ is a $\star$-algebra homomorphism and that $U\mathfrak{g}_{\star}$ closes under $\star$-multiplication.
Furthermore, using (6) we obtain
$$\displaystyle\Delta_{\star}(\mathfrak{g}_{\star})\subseteq U\mathfrak{g}_{%
\star}\otimes U\mathfrak{g}_{\star}~{}\Leftrightarrow~{}X_{\bar{R}^{\alpha}}%
\in U\mathfrak{g}_{\star}~{},~{}~{}\text{for all }\alpha\text{ with }\bar{R}_{%
\alpha}(\mathfrak{g}_{\star})\neq\{0\}~{}.$$
(58)
Therefore we have to use that all $\bar{R}^{\alpha}$ are linearly independent, that $X$ is a vector space isomorphism [10]
and that we have $\bar{R}_{\alpha}(\mathfrak{g}_{\star})\subseteq\mathfrak{g}_{\star}$,
due to the minimal axioms (17).
Additionally, we can show that $X_{\bar{R}^{\alpha}}=\bar{f}^{\beta}\bar{R}^{\alpha}\chi S^{-1}(\bar{f}_{\beta%
})=\bar{R}^{\alpha}$. This is done
by applying the explicit form of the twist (24) and using that the $V_{a}$ mutually commute.
Next, we show that
$$\displaystyle\bar{R}^{\alpha}\in U\mathfrak{g}_{\star}~{},~{}~{}\text{for all %
}\alpha\text{ with }\bar{R}_{\alpha}(\mathfrak{g}_{\star})\neq\{0\}~{}%
\Leftrightarrow~{}\theta^{ba}V_{b}\in\mathfrak{g}_{\star}~{},~{}~{}\text{for %
all }a\text{ with }[V_{a},\mathfrak{g}_{\star}]\neq\{0\}~{}.$$
(59)
The direction $\Rightarrow$ is trivial, since the RHS is a special case of the LHS.
The direction $\Leftarrow$ can be shown by using that the $V_{a}$ mutually commute and the explicit expression of the $R$-matrix
(26).
Finally, the RHS of (59) is equivalent to the condition of proposition 3 by
using the canonical form of $\theta$ (25).
Next, we show that (21b) is satisfied, if (21a) is fulfilled.
For this we use that for Reshetikhin-Jambor-Sykora twists we have $\chi=f^{\alpha}S(f_{\alpha})=1$, which leads to the identity
$$\displaystyle S_{\mathcal{F}}(\xi)=\chi S(\xi)\chi^{-1}=S(\xi)=S^{-1}(\xi)=S_{%
\mathcal{F}}^{-1}(\xi)~{},~{}\forall\xi\in U\Xi$$
(60)
for the antipode in the $\mathcal{F}$-Hopf algebra.
This property translates to the $\star$-Hopf algebra, since it is isomorphic to the $\mathcal{F}$-Hopf algebra and we obtain
the following equivalences of (21b)
$$\displaystyle S_{\star}(U\mathfrak{g}_{\star})\subseteq U\mathfrak{g}_{\star}~%
{}\Leftrightarrow~{}S_{\star}(\mathfrak{g}_{\star})\subseteq U\mathfrak{g}_{%
\star}~{}\Leftrightarrow~{}S^{-1}_{\star}(\mathfrak{g}_{\star})\subseteq U%
\mathfrak{g}_{\star}~{}.$$
(61)
For the first equivalence we had to use that $S_{\star}$ is a $\star$-anti homomorphism.
Using the RHS of (59), which is equivalent to (21a), and the definition of $S_{\star}^{-1}$
(6), we obtain
$$\displaystyle S^{-1}_{\star}(\mathfrak{g}_{\star})=-\sum_{n=0}^{\infty}\frac{(%
-i\lambda)^{n}}{n!}\theta^{a_{1}b_{1}}\cdots\theta^{a_{n}b_{n}}[V_{a_{1}},%
\cdots,[V_{a_{n}},\mathfrak{g}_{\star}]\cdots]~{}V_{b_{1}}\cdots V_{b_{n}}\in U%
\mathfrak{g}_{\star}~{},$$
(62)
where we have used $\xi\star V_{a}=\xi V_{b}$ for all $\xi\in U\Xi_{\star}$, since the action of the twist on $V_{a}$ is trivial.
Appendix B $\star$-Commutators of the Coordinate
Functions in FRW and Black Hole Models
In tables 3 and 4, we list the $\star$-commutators among the linear
coordinate functions to order $\lambda^{1}$ in the FRW and black hole
models. In these expressions, $(i\leftrightarrow j)$ denotes the
same term with $i$ and $j$ interchanged.
References
[1]
S. Doplicher, K. Fredenhagen and J. E. Roberts,
Phys. Lett. B331, 39 (1994);
S. Doplicher, K. Fredenhagen and J. E. Roberts,
Commun. Math. Phys. 172, 187 (1995)
[arXiv:hep-th/0303037].
[2]
R. J. Szabo,
Phys. Rept. 378, 207 (2003)
[arXiv:hep-th/0109162];
R. J. Szabo,
Class. Quant. Grav. 23, R199 (2006)
[arXiv:hep-th/0606233].
[3]
F. Muller-Hoissen,
AIP Conf. Proc. 977, 12 (2008)
[arXiv:0710.4418 [gr-qc]].
[4]
X. Calmet, B. Jurčo, P. Schupp, J. Wess and M. Wohlgenannt,
Eur. Phys. J. C23, 363 (2002)
[arXiv:hep-ph/0111115];
B. Melic, K. Passek-Kumericki, J. Trampetic, P. Schupp and M. Wohlgenannt,
Eur. Phys. J. C 42, 483 (2005)
[arXiv:hep-ph/0502249];
B. Melic, K. Passek-Kumericki, J. Trampetic, P. Schupp and M. Wohlgenannt,
Eur. Phys. J. C 42, 499 (2005)
[arXiv:hep-ph/0503064].
[5]
W. Behr, N. G. Deshpande, G. Duplancic, P. Schupp, J. Trampetic and J. Wess,
Eur. Phys. J. C 29, 441 (2003)
[arXiv:hep-ph/0202121];
G. Duplancic, P. Schupp and J. Trampetic,
Eur. Phys. J. C 32, 141 (2003)
[arXiv:hep-ph/0309138];
P. Schupp, J. Trampetic, J. Wess and G. Raffelt,
Eur. Phys. J. C 36, 405 (2004)
[arXiv:hep-ph/0212292];
P. Minkowski, P. Schupp and J. Trampetic,
Eur. Phys. J. C 37, 123 (2004)
[arXiv:hep-th/0302175];
B. Melic, K. Passek-Kumericki and J. Trampetic,
Phys. Rev. D72, 054004 (2005)
[arXiv:hep-ph/0503133];
Phys. Rev. D72, 057502 (2005)
[arXiv:hep-ph/0507231];
M. Buric, D. Latas, V. Radovanovic and J. Trampetic,
Phys. Rev. D75, 097701 (2007);
T. Ohl and J. Reuter,
Phys. Rev. D70, 076007 (2004)
[arXiv:hep-ph/0406098];
A. Alboteanu, T. Ohl and R. Rückl,
Phys. Rev. D74, 096004 (2006)
[arXiv:hep-ph/0608155];
eConf C0705302, TEV05 (2007)
[Acta Phys. Polon. B38, 3647 (2007)]
[arXiv:0709.2359 [hep-ph]];
Phys. Rev. D76, 105018 (2007)
[arXiv:0707.3595 [hep-ph]].
[6]
I. Hinchliffe, N. Kersting and Y. L. Ma,
Int. J. Mod. Phys. A19, 179 (2004)
[arXiv:hep-ph/0205040].
[7]
C. S. Chu, B. R. Greene and G. Shiu,
Mod. Phys. Lett. A16, 2231 (2001)
[arXiv:hep-th/0011241];
F. Lizzi, G. Mangano, G. Miele and M. Peloso,
JHEP 0206, 049 (2002)
[arXiv:hep-th/0203099];
R. Brandenberger and P. M. Ho,
Phys. Rev. D66, 023517 (2002)
[AAPPS Bull. 12N1, 10 (2002)]
[arXiv:hep-th/0203119];
Q. G. Huang and M. Li,
JHEP 0306, 014 (2003)
[arXiv:hep-th/0304203];
Q. G. Huang and M. Li,
JCAP 0311, 001 (2003)
[arXiv:astro-ph/0308458];
S. Tsujikawa, R. Maartens and R. Brandenberger,
Phys. Lett. B574, 141 (2003)
[arXiv:astro-ph/0308169];
A. H. Fatollahi and M. Hajirahimi,
Europhys. Lett. 75, 542 (2006)
[arXiv:astro-ph/0607257];
E. Akofor, A. P. Balachandran, S. G. Jo, A. Joseph and B. A. Qureshi,
JHEP 0805, 092 (2008)
[arXiv:0710.5897 [astro-ph]];
E. Akofor, A. P. Balachandran, A. Joseph, L. Pekowsky and B. A. Qureshi,
arXiv:0806.2458 [astro-ph];
S. Fabi, B. Harms and A. Stern,
Phys. Rev. D78, 065037 (2008)
[arXiv:0808.0943 [hep-th]].
[8]
P. Nicolini,
arXiv:0807.1939 [hep-th].
[9]
P. Aschieri, C. Blohmann, M. Dimitrijevic, F. Meyer, P. Schupp and J. Wess,
Class. Quant. Grav. 22, 3511 (2005)
[arXiv:hep-th/0504183].
[10]
P. Aschieri, M. Dimitrijevic, F. Meyer and J. Wess,
Class. Quant. Grav. 23, 1883 (2006)
[arXiv:hep-th/0510059].
[11]
P. Aschieri,
J. Phys. Conf. Ser. 53, 799 (2006)
[arXiv:hep-th/0608172].
[12]
L. Alvarez-Gaume, F. Meyer and M. A. Vazquez-Mozo,
Nucl. Phys. B 753, 92 (2006)
[arXiv:hep-th/0605113].
[13]
N. Reshetikhin,
Lett. Math. Phys. 20, 331 (1990).
[14]
C. Jambor and A. Sykora,
[arXiv:hep-th/0405268].
[15]
T. Ohl and A. Schenkel,
Cosmological and Black Hole Solutions in Twisted Noncommutative Gravity,
to be published;
Classical and Quantum Field Fluctuations on Noncommutative Backgrounds,
in preparation.
[16]
P. Aschieri, F. Lizzi and P. Vitale,
Phys. Rev. D77, 025037 (2008)
[arXiv:0708.3002 [hep-th]].
[17]
P. Aschieri, L. Castellani and M. Dimitrijevic,
Lett. Math. Phys. 85, 39 (2008)
[arXiv:0803.4325 [hep-th]]. |
On the origin of the radial flow in low energy heavy ion reactions
Regina Nebauer${}^{*+}$ and Jörg Aichelin${}^{*}$
${}^{*}$SUBATECH
Laboratoire de Physique Subatomique et des Technologies Associées
UMR Universitè de Nantes, IN2P3/CNRS, Ecole des Mines de Nantes
4, rue Alfred Kastler
F-44070 Nantes Cedex 03, France.
${}^{+}$ Universität Rostock, Rostock, Germany
Abstract
The average transverse energy of nucleons and intermediate mass fragments
observed in the heavy ion reaction Xe(50A MeV)+Sn shows the same linear increase as a
function of their mass as observed in heavy ion collisions up to the highest energies
available today and fits well into the systematics. At higher energies this
observation has been interpreted as a sign of a strong radial flow in an
otherwise thermalized system. Investigating the reaction with Quantum Molecular
dynamics simulations we find in between 50A MeV and 200A MeV a change in the
reaction mechanism. At 50A MeV the apparent radial flow is merely
caused by an in-plane flow and Coulomb repulsion. The average transverse
fragment energy does not
change in the course of the reaction and is equal to the initial fragment energy
due to the Fermi motion.
At 200A MeV, there are two kinds of fragments: those formed from spectator matter and those
from the center of the reaction. There the transverse energy is caused by the pressure
from the compressed nuclear matter. In both cases we observe a binary event structure,
even in central collisions. This demonstrates as well the non thermal character
of the reaction. The actual process which leads to multifragmentation is rather
complex and is discussed in detail.
I Introduction
It is known since long that for almost all particles observed in heavy ion
reactions between $30A~{}MeV$ and $200~{}A.GeV$
the transverse kinetic energy spectra have a
Maxwell-Boltzmann form, predicted for an emission from an equilibrated source.
However, the apparent temperature of the spectra and hence the average
kinetic energy of the particles is quite different for different hadrons and
fragments and increases with increasing mass and increasing energy. This
observation seemed to exclude an identification of the apparent temperature
with the real temperature of the system.
Recently it has been conjectured [1, 2, 3] that at all energies
between $50A~{}MeV$ and $200A~{}GeV$ the assumption of a strong radial flow can
reconcile the mass dependence of the apparent temperature with
thermodynamics. At relativistic
and ultra-relativistic energies this has been inferred by comparing transverse
pion, kaon and proton spectra [1]. At energies below
$500A~{}MeV$ the lever arm is still larger because one can include the
intermediate
mass fragments (IMF’s) of masses in between 2 and 10 [2, 3], emitted
at mid rapidity, to separate radial flow and temperature. The deviations in forward
and backward direction are usually interpreted as preequilibrium emission.
This observation has renewed the interest in the thermal analysis of heavy ion
reactions in the high as well as in the low energy heavy ion community
[2, 4]. It revealed a very large value for
the radial flow velocity (up to 40% of the speed of light) and a very peculiar
beam energy dependence
[4] which
has been discussed as a possible sign for a
transition from a hadronic phase to a Quark Gluon Plasma phase.
Especially at low beam energies the observed
value of the radial velocity in units of the speed of light of
$.05\leq\beta_{r}/c\leq 0.2$ [2] (the exact value depends on the event
selection) is hard to understand in terms of physical processes because
none of the other observables present evidence that the system is sufficiently
compressed for attributing the radial flow to an equation of state effect.
The very moderate in-plane flow is negative (as in deep inelastic collisions)
and points therefore to an other
origin than compression and subsequent release of the compressional energy,
in contradistinction to higher energies where the in-plane flow is positive.
Nevertheless, the low energy points follow smoothly the above mentioned
systematics.
This radial flow at 50A MeV has been reproduced[5]
by simulations using the Quantum Molecular Dynamics (QMD) [6]
approach. The same is true at higher energies
where QMD simulations have been performed by the FOPI collaboration
[3].
It is the purpose of this article to take advantage of this agreement and to
study the origin of the large increase of the transverse energy with the
mass and hence the origin of the radial flow. Here we concentrate
ourselves on the reaction at 50A MeV and study in detail how the fragments
are produced. The reaction mechanism at this low energy is quite different than
that at higher energies. To demonstrate this we present as well the same study
for a beam energy of 200A MeV.
A detailed study of the radial flow at higher energies, its equation of state
dependence and its physical origin will be the subject of a
forthcoming publication [7].
For our study at $(50A~{}MeV)$
we use simulations of the reaction $Xe+Sn$
which has recently been measured by the INDRA collaboration. The INDRA detector
at GANIL has been constructed to study
multifragmentation and therefore the angular coverage and the
energy thresholds
have been chosen to be better than that of any other $4\pi$ detector
elsewhere. Hence the data taken with this detector are most suitable
to confirm or disprove the theories embedded in the simulation programs.
A detailed comparison of our results with the experimental data will
be published elsewhere [5]. Here we mention that not only the mass
dependence of the average kinetic energy but also the kinetic energy spectra
themselves are in reasonable agreement with experiment.
For details about the QMD approach we refer to reference [6]. In
this program the nucleons are represented by Gaussian wave packets with a
constant width. The time evolution of the centers of these wave packets
is given by Euler Lagrange equations derived from the Lagrangian of the system.
The nucleons have an effective charge of Z/A. The fragment distribution becomes
stable after 240 fm/c [8]. From this point on we employ a Coulomb
trajectory
program for the fragments and nucleons until the Coulomb energy is released.
This second step is necessary at this low energy because a large fraction of
the kinetic energy of the fragments is due to the Coulomb energy.
II The radial flow at beam energies of 50A MeV
A first general idea of the time evolution of the collision can be obtained
from the time evolution of the density of the system.
If the maximal density is reached the nuclei have their maximal overlap, after
that the system
expands and the density decreases. This permits to find the time scale of
the reaction.
The total density is
the sum over all nucleons which are described by Gaussians:
$$\rho(\vec{r},t)\propto\sum_{i=1}^{A}e^{-\frac{(\vec{r}-\vec{r_{i}}(t))^{2}}{2{%
L}}}$$
(1)
The width of the Gaussians is $4L=4.33~{}fm^{2}$ and A is the number of nucleons
present in the system.
In figure 1, left hand side, we plot the time evolution of the
total density in the center of the reaction.
The maximum density is obtained at $\approx 50~{}fm/c$, on
the same time scale the
system expands and reaches at $120~{}fm/c$ a low density phase where the
fragments do not interact anymore.
On the right hand side of the same figure we display the density profile along the
beam (z) axis. Here we can follow the two nuclei. They occupy the same coordinate
space at
$50..60~{}fm/c$. The system expands after $120~{}fm/c$. We find
that this quasi-central ($b=3~{}fm$) collision is semi transparent. Projectile
and target pass through each other without being seriously decelerated.
For $b=0~{}fm$ we get the same result.
That binary character is confirmed by experiment:
In the center of mass system the experiment
shows
a flat
angular distribution ($dN/dcos\theta_{cm}$) between
$60^{\circ}\leq\theta_{cm}\leq 120^{\circ}$ as well as
a constant average kinetic
energy for fragments $Z\geq 3$ even in central collisions.
In forward and backward direction a strongly enhanced cross section
is observed. The INDRA collaboration made use of this observation and
presented their data in two angular bins:
$60^{o}~{}\leq~{}\theta_{CM}~{}\leq~{}120^{o}$ (IMF’s emitted in this angular range are called
mid-rapidity fragments (MRF’s)) and $\theta_{CM}~{}<~{}60^{o},\theta_{CM}~{}>~{}120^{o}$ called
projectile/target like fragments (PTF’s) [2].
How do the transverse fragment momenta reflect this passage through the other nucleus?
This is displayed in
figure 2 for a b = 3 fm reaction. Here we show the time evolution
of the average transverse momentum of
all fragments with $5\leq A\leq 10$ (left) and $A>10$ (right). For the
dotted points the transverse momentum is represented with respect to the beam
direction, for the triangles with respect to the largest eigenvector of the
momentum tensor which is tilted by the flow angle $\theta_{flow}$.
It is the first seminal result of this article that in the rotated system the
average transverse momentum is initially and finally the same. A very similar
result we obtain for b = 1 and b = 5 fm. During the reaction
the particles are accelerated in transverse direction but later they
feel a force into the opposite direction. The origin of this acceleration will
be discussed later.
As squares we display the time evolution of the in-plane
flow ${1\over N_{F}}\sum_{N_{F}}sign(p^{cm}_{z})p_{x}$. $N_{F}$ is the number of
fragments. We see that it remains moderate and is negative.
The average transverse energy of the fragments as a function of the fragment
charge is
displayed in fig. 3. On the left hand side we display the transverse
energy with respect to the beam axis, on the right hand side with respect to
the rotated system. The mean value is presented after the initialization, after
250 fm/c, when the fragment distribution gets stable, and finally after the
mutual Coulomb repulsion has ceased.
It is the second seminal result of this article that the average transverse
fragment energy initially is about 14 MeV and independent of the fragment size.
As Goldhaber
[9] has pointed out many years ago this is expected if the
fragment formation is a fast process. In this case the momentum distribution of
the fragments is a convolution of the momentum distribution of the
entrained nucleons and one expects for the
fragment momentum squared
$$P^{2}=<(\sum_{i}^{A}p_{i})^{2}>={3k_{Fermi}^{2}\over 5}A{N-A\over N-1}$$
where A is the number of nucleons entrained in the fragment and N is the
number of nucleons of the disintegrating nucleus. Hence the fragment
kinetic energy is
$$E_{A}={3E_{Fermi}\over 5}{N-A\over N-1}$$
and as well about 14 MeV in the QMD calculations. (Due to surface effects it
differs from the value of a Fermi gas in a square wall potential).
How it can happen that the fragments pass the reaction zone without
being heated up and are finally observed at midrapidity or in PTF’s
we will investigate in the second part of this article.
The nucleons interact via the potential
$$V=-124\frac{\rho}{\rho_{0}}+70.5(\frac{\rho}{\rho_{0}})^{2}~{}[MeV]$$
(2)
where the density is given by equation 1 and $\rho_{0}$ is the normal
nuclear matter density.
In order to reveal the physics which drives the reaction we display the relative
density of
those nucleons which are finally entrained in MRF’s or PTF’s as a function of time in
the x-z plane
$$\rho^{\tiny MRF/PTF}_{rel}(x,z,t)={\rho_{\tiny MRF/PTF}(x,z,t)\over\rho_{total%
}(x,z,t)}$$
(3)
and superimpose the gradient of the potential in the x-z plane as arrows where
x is the direction of the impact parameter.
For the sake of a clearer display we plot nucleons coming
from the projectile only. We start out with a discussion of the reaction
at zero impact parameter then we continue with $b=3fm$ reactions.
The motion of the nucleons in the potential of a nucleus is a sequence of acceleration
and deceleration. Nucleons on the surface are almost at rest, due to the density (and
thus the potential) gradient they become accelerated towards the center of the nucleus.
They
reach their maximal momentum when they pass
the center of the nucleus, climb up the potential on the
other side and are finally
at rest again when arriving at the surface.
When a heavy ion collision occurs, the position of
the nucleons in the projectile or target determines whether they ”feel”
the heavy ion collision right from the beginning or only when the
high density phase has already passed. We will show that the initial position
of the entrained nucleons decides
as
well whether the fragment is finally observed at mid rapidity or in forward/
backward direction.
In figure 4 we display the motion of the nucleons finally entrained in
PTF’s for a reaction at $b=0~{}fm$. The spatial distribution of those nucleons
is almost identical with
that of all nucleons present in the projectile.
In the first step of the collision the nucleons move away from the target
into the yet unperturbed part of the projectile. When they arrive
at the back end of the projectile they invert the direction of their momenta.
They are then accelerated in longitudinal direction towards the
center of the reaction. When they arrive finally there the high density
zone has disappeared already.
Hence the nucleons pass the center without a larger change of their initial
momentum. The initial correlations [10] among the nucleons which
finally form a fragment survive the reaction because
all potential gradients are small.
Nucleons finally emitted as MRF’s (fig. 5) are strongly
located at the front end of the nuclei. These are the nucleons which
climb the
nuclear potential created by the higher density in the reaction zone and which are
at rest before they are on the top of the potential wall.
Due to their position they are involved in
the collisions between projectile and target nucleons right from the beginning.
Collisions support the
deceleration. (Later collisions are to a large extend Pauli suppressed.)
They escape the barrier
in transverse direction.
As their (longitudinal as well as
transverse)momentum
is quite small, the nucleons stay longer in the
center of the reaction what favours the mixing of projectile and target
nucleons.
When leaving the reaction zone the fragments become decelerated due to the
potential interaction with the rest of the system. This
deceleration balances
the gain in energy due to the prior acceleration in transverse direction,
although the
physics of both processes is rather independent.
The same analysis for PTF’s for $b=3~{}fm$ is displayed in figure 6.
The general reaction mechanisms, even the time scales,
are the same as discussed before. In addition to the scenario at $b=0~{}fm$
asymmetry
effects occur here. Up to $40~{}fm/c$ the scenario is the same as at
zero impact parameter. When the high density phase occurs, the nucleons take the line of
least resistance, i.e. they follow the minimum of the potential on the right hand
side (for the projectile nucleons, the target nucleons take the inverse direction on the
other side). The larger part of the nucleons pass the reaction center when the
potential barrier has disappeared. As already discussed for zero impact
parameter case, the
nucleons emitted as fragments in forward/backward direction pass through the
reaction zone without an important change of their initial momenta.
For the reaction Xe(50A MeV)+Sn we could find two modes for the emission of fragments:
1.
The Fragments are already ”preformed” in the nuclei. The initial correlations
between the nucleons survive as they avoid the high density zone. For a finite impact
parameter a strong in-plane flow is observed. The fragments are quite big and emitted in
forward/backward direction.
2.
Fragments emitted in midrapidity are smaller. They are formed by mixing projectile
and target nucleons. Due to their initially small longitudinal momentum they stay in the
high density zone and emitted in transverse direction. These fragments are rather small as
they are formed via collisions.
III The radial flow at beam energies of 200A MeV
In the reaction at 200 A MeV the velocity of the nucleons due to
internal motion and the velocity of the two colliding nuclei in the center of mass frame
are now significantly different. As a consequence we expect a change of the reaction
mechanism.
That this is indeed the case can already be seen from the time evolution of the
transverse momenta of the fragments, figure 7. The collective directed in-plane
flow has the opposite sign as compared to 50A MeV. This is as well a sign for the
change of the reaction mechanism which develops between 50A MeV and 200A MeV from a
deep inelastic scenario
with a negative flow angle to a bounce off type reaction with a positive flow angle.
Despite of the higher beam
energy the transverse acceleration of the light
fragments is delayed as compared to the reaction at 50A MeV and sets on only
at the maximum overlap of the two nuclei. Then they are
accelerated and remain with a high transverse momentum after the expansion.
The large fragments have a completely different time evolution as compared to
50A MeV. First we see a deceleration in transverse direction of the
fragments. Then, when the highest density is reached this turns into an
acceleration. Finally, as at lower energies, the
fragments are decelerated again during the expansion of the system. Finally they have
a quite low transverse momentum.
To understand this behaviour it is useful to study the initial-final state correlations.
Particularly in momentum space we observe such strong correlations. In figure 8 we
display the initial distribution of the nucleons at t=0 (color level) and superimpose
the relative fraction of nucleons finally emitted in small (lhs) or big (middle, rhs) fragments.
Here we can clearly see that the nucleons finally emitted in heavy fragments have
an initial momentum which points away from the reaction partner. This initial - final -
state correlation, already
allusively present at 50A MeV increases with the mass of the fragment and the beam energy
[10].
For nucleons finally
emitted in small fragments the correlations are less pronounced but with preference of momenta
which point into the reaction zone what favours the mixing of projectile and
target nucleons in the fragments.
Nucleons emitted finally in small fragments are located in the center of the collision
partners and become only accelerated when a higher density is built up in the center of the
reaction. Following the density gradient they develop a radial flow. Finally they
disentangle form the system. This costs energy and therefore the momentum decreases.
We expect a constant acceleration of all fragment nucleons and therefore
a final transverse energy which increases linearly with the fragment mass.
The light fragments show only a small in-plane-flow which is positive in contrast to the
reaction at the lower energy. In contradistinction to 50A MeV we observe
for the light fragments a genuine acceleration in transverse direction independent of the
system in which it is measured. At 200A MeV the density built up in the center of the reaction
is sufficiently high for a noticeable acceleration of the fragments.
We see that especially for the heavy fragments that the initial transverse momentum in
the rotated system is quite different as compared to the beam system due to the initial
correlations in momentum space, discussed above.
For a better understanding of the time evolution of the momenta of the big fragments we
study as at 50A MeV the movement of their constituent nucleons in the nuclear potential,
figure 9. Here we display the gradient of the nuclear potential and the
relative fraction of nucleons emitted in fragments $A>10$.
In addition to their strong initial - final - state correlations in momentum space
the fragments have as well strong correlations in coordinate space. The big fragments
are at t=0 located at the surface of the collisions partners and have
therefore an average density well below the normal nuclear matter density. At the beginning
of the reaction
they are accelerated towards the center of the reaction because the density is
higher there.
This lowers the transverse momentum. When the high density at the center of the reaction is
built up, the acceleration changes sign. Thus we can already at 200A MeV and 3 fm
impact parameter distinguish between participants and spectators.
Now we can draw up two competing processes of fragment formation at 200A MeV:
1.
Large fragments are formed from spectator matter. The entrained nucleons do not pass
the region of high density and therefore the acceleration in transverse direction is moderate.
For the entrained nucleons we observe as well strong initial - final - state correlations
in momentum space. In other words, the more the nucleons avoid the high density zone the larger
is the change that they find themselves finally in big fragments. At 200A MeV the
probability that nucleons passes the high density zone without having a collision is well
reduced with respect to 50A MeV. Therefore any larger correlated part of nuclear matter which
passes the high density becomes destroyed by collisions and cannot form finally a fragment
anymore.
2.
Small correlated parts of nuclear matter can still survive the passage through the high
density zone. That will be different at still higher energies. These
nucleons form small fragments mixing projectile and target
via collisions. The nucleons finally emitted in this fragments are close to
the high density zone and accelerated due to the internal pressure. Their
transverse energy is high and expected to be proportional to their mass (radial flow).
These two competing processes allow to understand
the behaviour of the transverse energy as a function of the
charge (figure 10,lhs). At t=0, the beginning of the reaction, the strong
initial correlations in
momentum space (high transverse momentum for big fragments)
cause a linear increase of the transverse energy with the charge of the
fragment. For small fragments the process 2 dominates, finally the
transverse energy of fragments is large and increases with increasing charge (mass) of the
fragment.
For big fragments the process 1 is dominant, the average transverse
energy is small. In between we observe a transition phase.
On the rhs of the same figure we display as in figure 3 the transverse
energy of the fragments in the rotated systems which is almost identical with the left hand
side. Hence the radial energies are genuine and not artifacts due to a strong in-plane flow as
at 50A MeV.
IV Conclusion
In conclusion we have found that the kinetic energy of the fragments observed in
the heavy ion reaction $Xe+Sn$ at $50A~{}MeV$ reflects the initial Fermi energy of
the entrained nucleons. There is no hint that the system becomes equilibrated. Rather, in
agreement with experiment,
even in central collisions the reaction is semi transparent. We find that the final
longitudinal momentum of the fragments depends strongly on the time point when the
entrained nucleons pass the center of the reaction.
Nucleons which are passing early in the reaction are decelerated
due to collisions with the reaction partner and due to the strong potential gradient.
They are deviated into the transverse direction as their longitudinal momentum is not
high enough to
overcome the potential barrier formed in the reaction center in the phase of
highest density.
Nucleons which arrive later
do not encounter a strong potential gradient anymore and pass the reaction center
freely. Thus, they keep almost their initial momentum. We find that
compressional effects have little influence on the final momentum of the
fragments at this energy and the observed apparent radial flow is
not real. The linear increase of the fragment
kinetic energy with the mass for small fragments
finds its natural explication in terms of the initial Fermi motion,
the Coulomb barrier and a small in-plane flow.
At 200A MeV we found that the reaction mechanism has changed and approaches the
participant-spectator model as far as heavy fragments are concerned. Due to the smaller mean
free path heavy fragments cannot come anymore form the center of the reaction. Their
transverse energy remains therefore moderate and the in-plane flow weak.
Small fragments show a strong radial flow and have passed the center of the reaction
where the forces are much stronger than at 50A MeV.
At higher energies fragments formed from of participants become more and more rare due to
the shorter mean free path and due to stronger force gradients. Hence fragment formation
of spectator matter will dominate. Initial - final - state correlations in momentum space become
stronger and collisions start to contribute in addition to the radial
flow. How these different process influence the final radial flow and its dependence on the
nuclear equation of state is presently under investigation [7].
References
[1]
P. Braun-Munzinger et al, Phys. Lett. B344 (1995) 43 and
Phys. Lett. 365 (1996) 1
[2]
N. Marie, PhD Thesis , Université de Caen, France
[3]
W. Reisdorf, Nucl. Phys. A612 (1997) 493,
G. Poggi et al., Nucl Phys A586
(1995) 755, B. Hong et al. nucl-ex/9707001
[4]
N. Herrmann, Nucl. Phys. A610 (1996) 49c,
[5]
R. Nebauer at al., nucl-th/9810008
[6]
J. Aichelin, Phys. Rep. 202 (1991) 233.
[7]
C. Hartnack et al., to be published
[8]
R. Puri et al., Phys. Rev. C54 (1996) R28
[9]
A.S. Goldhaber Phys. Lett. B53 (1974) 306
[10]
P.B. Gossiaux and J. Aichelin, Phys. Rev. C56 (1997) 2109 |
Abstract
We consider in this paper a spray constituted of an incompressible
viscous gas and of small droplets which can breakup. This spray is
modeled by the coupling (through a drag force term) of the
incompressible Navier-Stokes equation and of the Vlasov-Boltzmann
equation, together with a fragmentation kernel. We first show at
the formal level that if the droplets are very small after the
breakup, then the solutions of this system converge towards the
solution of a simplified system in which the small droplets produced
by the breakup are treated as part of the fluid. Then, existence of
global weak solutions for this last system is shown to hold, thanks
to the use of the DiPerna-Lions theory for singular transport
equations.
Existence theory for the kinetic-fluid coupling when small droplets are
treated as part of the fluid
S. BENJELLOUN111CMLA, ENS Cachan & CNRS, 61 av du Président Wilson, 94235, Cachan, Cedex, France., L. DESVILLETTES111CMLA, ENS Cachan & CNRS, 61 av du Président Wilson, 94235, Cachan, Cedex, France., AND A. MOUSSA222UPMC Université Paris 06 & CNRS, UMR 7598, LJLL, F-75005, Paris & INRIA Paris-Rocquencourt, Équipe-Projet REO, BP 105, F-78153 Le Chesnay Cedex, France.
1 Introduction
Sprays are complex flows which are constituted of an underlying gas in
which a population of droplets (or dust specks) are dispersed,
cf. [18]. There are various possibilities for modeling
such flows, depending in particular on the volume fraction of the
liquid phase (cf. [9] for example).
We focus here on the case when the volume fraction occupied by the
droplets is small enough to be neglected in the equations (such sprays
are called thin sprays, cf. [18]), so that the
modeling of the liquid phase can be performed by the use of a pdf
(particles density function) which solves Vlasov-Boltzmann
equation (cf. [21, 1]). Denoting $f:=f(t,\mathbf{x},\boldsymbol{\xi},r)\geq 0$ the number density of droplets of radius $r$ which at
time $t$ and point $\mathbf{x}$ have velocity $\boldsymbol{\xi}$, the Vlasov equation
writes
$$\partial_{t}f+{\boldsymbol{\xi}}\cdot\nabla_{\mathbf{x}}f+\nabla_{\boldsymbol{%
\xi}}\cdot(f\mathbf{\Gamma})=Q(f),$$
(1)
where $\mathbf{\Gamma}$ represents the acceleration felt by the
droplets (resulting from the drag force exerted by the gas), and $Q$
is an operator taking into account the complex phenomena happening at
the level of the droplets (collisions, coalescences, breakup).
We also restrict ourselves to the case when the gas is incompressible
and viscous, which is for instance the usual framework when studying
the transport of sprays in the upper airways of the human lungs,
cf. [17].
Accordingly, the gas is modeled by the incompressible Navier-Stokes
equation
$$\displaystyle\nabla_{\mathbf{x}}\cdot\mathbf{u}=0,$$
(2)
$$\displaystyle\rho_{g}\left[\partial_{t}\mathbf{u}+\nabla_{\mathbf{x}}\cdot(%
\mathbf{u}\otimes\mathbf{u})\right]+\nabla_{\mathbf{x}}p-\mu\Delta_{\mathbf{x}%
}\mathbf{u}=\mathbf{F}_{\text{ret}},$$
(3)
where $\rho_{g}$ is the constant density of the gas, $\mathbf{u}:=\mathbf{u}(t,\mathbf{x})\in\mathbb{R}^{3}$ is its velocity, $\mu$ is its constant (dynamic) viscosity,
and $\mathbf{F}_{\text{ret}}$ is the retroaction of the drag force:
$$\mathbf{F}_{\text{ret}}(t,\mathbf{x})=-\int_{0}^{+\infty}\int_{\mathbb{R}^{3}}%
\frac{4}{3}\,\rho_{l}\,r^{3}\,f\,\mathbf{\Gamma}\,\mathrm{d}\boldsymbol{\xi}%
\mathrm{d}r.$$
(4)
Finally, $\rho_{l}$ is the constant density of the liquid (so that the
mass of the droplets of radius $r$ is $\frac{4}{3}\,\rho_{l}\,r^{3}$).
For particles with small Reynolds numbers, the drag force is given by the simple formula (known as “Stokes’ law”)
$$\mathbf{\Gamma}(t,\mathbf{x},\boldsymbol{\xi},r)=-\frac{9}{2}\frac{\mu}{\rho_{%
l}}\frac{\boldsymbol{\xi}-\mathbf{u}(t,\mathbf{x})}{r^{2}},$$
(5)
that we shall systematically use in the sequel.
The modeling of the breakup phenomena is important in the applications
and has led to various models appearing in the literature
(cf. [1]). A typical form of the breakup kernel is
obtained when assuming that the droplets after breakup have the same
velocities as before breakup. The operator writes then
$$\displaystyle Q(f)(t,\mathbf{x},\boldsymbol{\xi},r)=-\nu(\boldsymbol{\xi},r)\,%
f(t,\mathbf{x},\boldsymbol{\xi},r)+\int_{r^{*}>r}b(r^{*},r)\,\nu(\boldsymbol{%
\xi},r^{*})\,f(t,\mathbf{x},\boldsymbol{\xi},r^{*})\,dr^{*},$$
(6)
where $\nu:=\nu(\boldsymbol{\xi},r)\geq 0$ is the fragmentation rate, and $b:=b(r^{*},r)\geq 0$ is related to the probability of ending up with droplets of radius $r$ out of the breakup of droplets of radius $r^{*}$.
We finally obtain the following system
$$\displaystyle\nabla_{\mathbf{x}}\cdot\mathbf{u}=0,$$
(7)
$$\displaystyle\partial_{t}f+\boldsymbol{\xi}\cdot\nabla_{\mathbf{x}}f+\nabla_{%
\boldsymbol{\xi}}\cdot[f\mathbf{\Gamma}]=Q(f),$$
(8)
$$\displaystyle\rho_{g}\,\Big{[}\partial_{t}\mathbf{u}+\nabla_{\mathbf{x}}\cdot(%
\mathbf{u}\otimes\mathbf{u})\Big{]}+\nabla_{\mathbf{x}}p-\mu\Delta_{\mathbf{x}%
}\mathbf{u}=\mathbf{F}_{\text{ret}}.$$
(9)
Because of the dependence w.r.t. $r$ of $\mathbf{\Gamma}$ in eq. (5),
we see that the drag force acting on the small droplets leads to very quick equilibration
of their velocity with the velocity of the gas. It is therefore natural to try
to write a set of equations replacing (4) – (9),
in which the small droplets are considered as part of the gaseous phase.
Indeed, in the context of the numerical simulation of eq. (1) thanks to a particle method, the small droplets which are produced
because of the breakup can lead to a high computational cost.
Once integrated (at the level of the continuous equations) in the gaseous phase, they can be discretized along with the gas
thanks to a finite volume scheme, and their contribution to the computational cost remains in this way reasonable.
In order to perform a mathematical study of the system obtained by such an approximation, we consider the simplest
possible case, namely when the fragmentation rate $\nu$ takes the form $\nu(\boldsymbol{\xi},r):=\tau^{-1}\mathds{1}_{r>r_{2}}$, for some constants $r_{2}>0$ and $\tau>0$, the latter being seen as a characteristic time of fragmentation.
We also assume that the aerosol is bidispersed: only two possible radii $r_{1}>r_{2}$ exist for the droplets, and the result of the breakup of particles of radius $r_{1}$ are particles of radius $r_{2}$. This implies that the density $f$ splits in the following way
$$\displaystyle f(t,\mathbf{x},\boldsymbol{\xi},r)=f_{1}(t,\mathbf{x},%
\boldsymbol{\xi})\,\delta_{r=r_{1}}+f_{2}(t,\mathbf{x},\boldsymbol{\xi})\,%
\delta_{r=r_{2}},$$
and that system (4) – (9) rewrites, after having
normalized all the constants (except $r_{2}$):
$$\displaystyle[left=\empheqlbrace]$$
$$\displaystyle\nabla_{\mathbf{x}}\cdot\mathbf{u}=0,$$
(10)
$$\displaystyle\partial_{t}f_{1}+\boldsymbol{\xi}\cdot\nabla_{\mathbf{x}}f_{1}+%
\nabla_{\boldsymbol{\xi}}\cdot\left[{f_{1}\,(\mathbf{u}-\boldsymbol{\xi})}%
\right]=-{f_{1}},$$
(11)
$$\displaystyle\partial_{t}f_{2}+\boldsymbol{\xi}\cdot\nabla_{\mathbf{x}}f_{2}+%
\nabla_{\boldsymbol{\xi}}\cdot\left[\frac{f_{2}\,(\mathbf{u}-\boldsymbol{\xi})%
}{r_{2}^{2}}\right]=\frac{f_{1}}{r_{2}^{3}},$$
(12)
$$\displaystyle\partial_{t}\mathbf{u}+\nabla_{\mathbf{x}}\cdot(\mathbf{u}\otimes%
\mathbf{u})+\nabla_{\mathbf{x}}p-\Delta_{\mathbf{x}}\mathbf{u}=-\int_{\mathbb{%
R}^{3}}(\mathbf{u}-\boldsymbol{\xi})\,f_{1}\,\mathrm{d}\boldsymbol{\xi}-r_{2}%
\int_{\mathbb{R}^{3}}(\mathbf{u}-\boldsymbol{\xi})\,f_{2}\,\mathrm{d}%
\boldsymbol{\xi}.$$
(13)
As we already explained, we are interested in the asymptotic regime when the particles resulting from breakup are becoming smaller and smaller. In our bidispersed model, this reduces to study the limit $r_{2}\to 0$.
Denoting $\rho=r_{2}^{3}\displaystyle\int_{\mathbb{R}^{3}}f_{2}\,\mathrm{d}\boldsymbol{\xi}$, and noticing that
$$\displaystyle\frac{\mathrm{d}}{\mathrm{d}t}\bigg{\{}\int_{{\mathbb{T}}^{3}}%
\int_{\mathbb{R}^{3}}f_{1}\,\frac{|\boldsymbol{\xi}|^{2}}{2}\,\mathrm{d}%
\boldsymbol{\xi}\mathrm{d}\mathbf{x}$$
$$\displaystyle+r_{2}^{3}\,\int_{{\mathbb{T}}^{3}}\int_{\mathbb{R}^{3}}f_{2}\,%
\frac{|\boldsymbol{\xi}|^{2}}{2}\,\mathrm{d}\boldsymbol{\xi}\mathrm{d}\mathbf{%
x}+\int_{{\mathbb{T}}^{3}}\frac{|\mathbf{u}|^{2}}{2}\,\mathrm{d}\mathbf{x}%
\bigg{\}}$$
$$\displaystyle+\int_{{\mathbb{T}}^{3}}\int_{\mathbb{R}^{3}}f_{1}\,{|\boldsymbol%
{\xi}-\mathbf{u}|^{2}}\,\mathrm{d}\boldsymbol{\xi}\mathrm{d}\mathbf{x}$$
$$\displaystyle+\frac{1}{r_{2}^{2}}\,\int_{{\mathbb{T}}^{3}}\int_{\mathbb{R}^{3}%
}r_{2}^{3}f_{2}\,|\boldsymbol{\xi}-\mathbf{u}|^{2}\,\mathrm{d}\boldsymbol{\xi}%
\mathrm{d}\mathbf{x}+\int_{{\mathbb{T}}^{3}}|\nabla_{\mathbf{x}}\mathbf{u}|^{2%
}\,\mathrm{d}\mathbf{x}=0,$$
we see that (at the formal level) $r_{2}^{3}\,f_{2}(t,\mathbf{x},\boldsymbol{\xi})\to\rho(t,x)\,\delta_{%
\boldsymbol{\xi}=\mathbf{u}(t,\mathbf{x})}$ when $r_{2}\to 0$.
Integrating eq. (12) against $r_{2}^{3}\,\mathrm{d}\boldsymbol{\xi}$, we end up with
$$\partial_{t}\rho+\nabla_{\mathbf{x}}\cdot[\rho\,\mathbf{u}]=\int_{\mathbb{R}^{%
3}}f_{1}\,\mathrm{d}\boldsymbol{\xi}.$$
(14)
Then, integrating eq. (12) against $r_{2}^{3}\,\boldsymbol{\xi}\,\mathrm{d}\boldsymbol{\xi}$ and adding the result with eq. (13),
we obtain
$$\partial_{t}((1+\rho)\,\mathbf{u})+\nabla_{\mathbf{x}}\cdot((1+\rho)\,\mathbf{%
u}\otimes\mathbf{u})+\nabla_{\mathbf{x}}p-\Delta_{\mathbf{x}}\mathbf{u}=-\int_%
{\mathbb{R}^{3}}(\mathbf{u}-\boldsymbol{\xi})\,f_{1}\,\mathrm{d}\boldsymbol{%
\xi}+\int_{\mathbb{R}^{3}}f_{1}\,\boldsymbol{\xi}\,\mathrm{d}\boldsymbol{\xi}.$$
Combining this last equation with eq. (14) and replacing the notation $f_{1}$ by $f$, we write down the
system that we wish to study (we close it with periodic boundary conditions for the mathematical proof of existence):
$$\displaystyle[left=\empheqlbrace]$$
$$\displaystyle\nabla_{\mathbf{x}}\,\mathbf{u}=0,$$
(15)
$$\displaystyle\partial_{t}\rho+\nabla_{\mathbf{x}}\cdot[\rho\,\mathbf{u}]=\int_%
{\mathbb{R}^{3}}f\,\mathrm{d}\boldsymbol{\xi},$$
(16)
$$\displaystyle\partial_{t}f+\boldsymbol{\xi}\cdot\nabla_{\mathbf{x}}f+\nabla_{%
\boldsymbol{\xi}}\cdot[(\mathbf{u}-\boldsymbol{\xi})\,f]=-f,$$
(17)
$$\displaystyle(1+\rho)\Big{[}\partial_{t}\mathbf{u}+\nabla_{\mathbf{x}}\cdot(%
\mathbf{u}\otimes\mathbf{u})\Big{]}+\nabla_{\mathbf{x}}p-\Delta_{\mathbf{x}}%
\mathbf{u}=2\,\int_{\mathbb{R}^{3}}(\boldsymbol{\xi}-\mathbf{u})\,f\,\mathrm{d%
}\boldsymbol{\xi},$$
(18)
where $\rho:=\rho(t,\mathbf{x})\geq 0$, $\mathbf{u}:=\mathbf{u}(t,\mathbf{x})\in\mathbb{R}^{3}$, $p:=p(t,\mathbf{x})\geq 0$,
$f:=f(t,\mathbf{x},\boldsymbol{\xi})\geq 0$, and $t\geq 0$, $\mathbf{x}\in{\mathbb{T}}^{3}$, $\boldsymbol{\xi}\in\mathbb{R}^{3}$. Let us recall
that $\rho$ represents here the “added density” resulting from the very small particles. This is the reason why, though we normalized the fluid density in (13) by $\rho_{g}=1$, we have in (18) the term $(1+\rho)$ in front of the convective part of the fluid equation. For a more detailed version of the previous computation, see [5].
Our goal is to study the existence theory for this system, completed with the following initial data:
$$\displaystyle f(0,\mathbf{x},\boldsymbol{\xi})$$
$$\displaystyle=f_{\textnormal{in}}(\mathbf{x},\boldsymbol{\xi})\geq 0,\quad%
\mathbf{x}\in{\mathbb{T}}^{3},\quad\boldsymbol{\xi}\in{\mathbb{R}^{3}},$$
(19)
$$\displaystyle\rho(0,\mathbf{x})$$
$$\displaystyle=\rho_{\textnormal{in}}(\mathbf{x})\geq 0,\quad\mathbf{x}\in{%
\mathbb{T}}^{3},$$
(20)
$$\displaystyle\mathbf{u}(0,\mathbf{x})$$
$$\displaystyle=\mathbf{u}_{\textnormal{in}}(\mathbf{x})\in{\mathbb{T}}^{3},%
\quad\mathbf{x}\in{\mathbb{T}}^{3}.$$
(21)
Before stating our result, let us introduce a few notations. If there is no ambiguity on the considered time interval, we simply denote by $\textnormal{L}^{p}_{t}(\textnormal{L}^{q}_{\mathbf{x}})$ and $\textnormal{L}^{p}_{t}(\textnormal{L}^{q}_{\mathbf{x},\boldsymbol{\xi}})$ the spaces $\textnormal{L}^{p}\big{(}[0,T];\textnormal{L}^{q}({\mathbb{T}}^{3})\big{)}$ and $\textnormal{L}^{p}\big{(}[0,T];\textnormal{L}^{q}({\mathbb{T}}^{3}\times%
\mathbb{R}^{3})\big{)}$, for any pair of exponents $(p,q)\in[1,\infty]^{2}$. In particular
when $p=q$ we simply use the notation $\textnormal{L}^{p}$. We adopt the same convention for Sobolev spaces $\textnormal{W}^{1,p}_{\mathbf{x}}$ and $\textnormal{H}^{m}_{\mathbf{x}}$ ($m\in\mathbb{N}$) and denote by $\textnormal{H}^{-m}_{\mathbf{x}}$ (or equivalently $\textnormal{H}^{-m}({\mathbb{T}}^{3})$) the dual of the latter.
When the subscript “ div ” is added to any space,
the corresponding subspace are the divergence-free (in $\mathbf{x}$ only) elements of the ambient space. We denote by $\mathbb{P}$ the Leray projector onto divergence-free vector fields
$$\displaystyle\mathbb{P}=\textbf{Id}-\nabla_{\mathbf{x}}\Delta^{-1}_{\mathbf{x}%
}\textnormal{div}_{\mathbf{x}},$$
easily defined on $\mathscrbf{D}([0,T]\times{\mathbb{T}}^{3})$ thanks to Fourier analysis, extended by duality to $\mathscrbf{D}^{\prime}([0,T]\times{\mathbb{T}}^{3})$.
Finally, if $h$ is a scalar function defined on $\mathbb{R}_{+}\times{\mathbb{T}}^{3}\times\mathbb{R}^{3}$, we define the following moments for $h:=h(t,\mathbf{x},\boldsymbol{\xi})$ (and for $\alpha\in\mathbb{R}$)
$$\displaystyle m_{\alpha}h(t,\mathbf{x}):=\int_{{\mathbb{R}^{3}}}|\boldsymbol{%
\xi}|^{\alpha}\,h(t,\mathbf{x},\boldsymbol{\xi})\,\mathrm{d}\boldsymbol{\xi},%
\quad M_{\alpha}h(t):=\int_{{\mathbb{T}}^{3}}m_{\alpha}(h)(t,\mathbf{x})\,%
\mathrm{d}\mathbf{x},\quad\mathbf{m}_{1}h(t,\mathbf{x}):=\int_{{\mathbb{R}^{3}%
}}\boldsymbol{\xi}h(t,\mathbf{x},\boldsymbol{\xi})\,\mathrm{d}\boldsymbol{\xi}.$$
Thanks to those notations, we are able to write down our main
Theorem:
Theorem 1.1.
Let $T>0$ and assume that
$$f_{\textnormal{in}}\in\textnormal{L}^{\infty},(1+|\boldsymbol{\xi}|^{2})f_{%
\textnormal{in}}\in\textnormal{L}^{1},\rho_{\textnormal{in}}\in\textnormal{L}^%
{\infty},\mathbf{u}_{\textnormal{in}}\in\mathbf{L}_{\textnormal{div}}^{2}.$$
(22)
Then the system (15) – (21)
admits a global weak solution $(\rho\geq 0,\,f\geq 0,\,\mathbf{u}\in\mathbb{R}^{3})$ such that
$$\rho\in\textnormal{L}^{\infty}([0,T],\textnormal{L}^{5/3}({\mathbb{T}}^{3})),f%
\in\textnormal{L}^{\infty}([0,T]\times{\mathbb{T}}^{3}\times{\mathbb{R}^{3}}),%
\mathbf{u}\in\textnormal{L}^{2}([0,T],\textnormal{H}^{1}_{\textnormal{div}}({%
\mathbb{T}}^{3}))\cap\textnormal{L}^{\infty}([0,T],\textnormal{L}^{2}({\mathbb%
{T}}^{3})).$$
(23)
Moreover, the triplet $(\rho,\,f,\,\mathbf{u})$ satisfies the following energy estimate
$$\displaystyle\frac{1}{2}\left\{M_{2}f(t)+\|\sqrt{1+\rho(t)}\,\mathbf{u}(t)\|_{%
\textnormal{L}^{2}({\mathbb{T}}^{3})}^{2}\right\}$$
$$\displaystyle+\int_{0}^{t}\|\nabla_{\mathbf{x}}\mathbf{u}(s)\|_{\textnormal{L}%
^{2}({\mathbb{T}}^{3})}^{2}\mathrm{d}s+\frac{3}{2}\int_{0}^{t}\int_{{\mathbb{T%
}}^{3}\times\mathbb{R}^{3}}|\mathbf{u}-\boldsymbol{\xi}|^{2}f\mathrm{d}%
\boldsymbol{\xi}\mathrm{d}\mathbf{x}\mathrm{d}s$$
$$\displaystyle\leq\frac{1}{2}\left\{M_{2}f_{\text{\bf in}}+\|\sqrt{1+\rho_{%
\text{\bf in}}}\,\mathbf{u}_{\text{\bf in}}\|_{\textnormal{L}^{2}({\mathbb{T}}%
^{3})}^{2}\right\},$$
(24)
and the bound
$$\displaystyle\|f\|_{\textnormal{L}^{\infty}\left([0,\,T]\times{\mathbb{T}}^{3}%
\times{\mathbb{R}^{3}}\right)}\leq e^{2T}\|f_{\text{\bf in}}\|_{\textnormal{L}%
^{\infty}\left({\mathbb{T}}^{3}\times{\mathbb{R}^{3}}\right)}.$$
Remark 1.1.
We explain here the meaning of “ weak solutions ” in the above Theorem. For $p\in[1,\infty]$, when $\mathbf{u}\in\mathbf{L}^{p}$, $f\in\textnormal{L}^{\infty}$ and $\rho\in\textnormal{L}^{p^{\prime}}$,
equations (15)–(16) have a clear meaning in the distributional sense. However the meaning of (18) is not completely obvious, and we rewrite this equation, using both (18) and (17), as
$$\displaystyle\mathbb{P}\left\{\partial_{t}[(1+\rho)\mathbf{u}]+\textnormal{div%
}_{\mathbf{x}}\Big{[}(1+\rho)\mathbf{u}\otimes\mathbf{u}\Big{]}\right\}-\Delta%
_{\mathbf{x}}\mathbf{u}=\mathbb{P}\left\{2\mathbf{m}_{1}f-\mathbf{u}m_{0}f%
\right\}.$$
(25)
For the initial condition, we adopt the following general definition : consider $v_{\textnormal{in}}\in\mathscr{D}^{\prime}({\mathbb{T}}^{3})$ and $w$ in $\textnormal{L}_{\textnormal{loc}}^{1}([0,T];\textnormal{H}^{-m}({\mathbb{T}}^{%
3}))$ for some integer $m\in\mathbb{N}$. A
distribution $v\in\mathscr{D}^{\prime}\big{(}]-\infty,T[\times{\mathbb{T}}^{3}\big{)}$ is a solution of the Cauchy problem
$$\partial_{t}v=w,\qquad v(0)=v_{\textnormal{in}},$$
if the support of $v$ is included in $\mathbb{R}_{+}$ and if
$$\displaystyle\partial_{t}v$$
$$\displaystyle=\widetilde{w}+\delta_{0}\otimes v_{\textnormal{in}}\text{ in }%
\mathscr{D}^{\prime}\big{(}]-\infty,T[\times{\mathbb{T}}^{3}\big{)},$$
where $\widetilde{w}$ is the extension of $w$ by $0$ on $\mathbb{R}_{-}$.
The study of the existence of solutions to coupled (through drag force interaction) fluid-kinetic equations
is now a well-established subject.
As far as viscous equations are concerned for the modeling of the gaseous phase, we would like to quote the works on the Vlasov-Stokes equations in [12], on the Vlasov/incompressible Navier-Stokes equations in [2, 6, 20], on the Vlasov-Fokker-Planck/incompressible Euler equations in [8]
, on the Vlasov-Fokker-Planck/incompressible Navier-Stokes equations in [15], and on the Vlasov-Fokker-Planck/compressible Navier-Stokes equations in [16].
Without any viscosity or Fokker-Planck damping, the study is more difficult, and only smooth local solutions are known to exist, in the compressible setting (cf. [4] and [14]).
Our system (15) – (18) is reminiscent of
the Vlasov/incompressible Navier-Stokes equations with a variable density.
It is known that variable densities in the incompressible Navier-Stokes
equations lead to extra difficulties (w.r.t. constant densities), cf. [7, 13]. Here extra difficulties (w.r.t. [6]) appear in the final passage to the limit (that is, in the stability result for solutions satisfying the natural a priori estimates
of the problem).
They are linked with nonlinearities which are specific of our model, and which necessitate a
careful treatment using refined versions of Lemmas presented in [10].
The equations coming out of the theory of sprays are known to be difficult to approximate in a good way (that is, in a way in which the a priori estimates satisfied by the
equations themselves, such as the energy estimate, are also satisfied or well approximated by the approximating equations). This difficulty appears in this paper, and the approximating scheme which is used is rather complicated and necessitates successive steps, which are
reminiscent of those used in [6].
Our paper is structured as follows:
Section 2 is devoted to the Proof of existence of solutions to an approximated
version of this system. The approximation is
removed in Section 3 and Theorem 1.1
is proven there. Finally, we present in a short Appendix some autonomous results
which are used in the Proof of Theorem 1.1.
2 Existence for a regularized system
In this section, we begin the Proof of Theorem 1.1 by introducing a regularized system.
2.1 Definition of the regularized system
In order to do so,
we introduce a mollifier $\varphi_{\varepsilon}\in\mathscr{C}^{\infty}({\mathbb{T}}^{3})$ for all $\varepsilon>0$.
We also introduce a truncation in this way: we define $\gamma_{\varepsilon}$ a nonnegative element of $\mathscr{D}(\mathbb{R}^{3})$,
whose support lies in $\textnormal{{B}}(0,2/\varepsilon)$, bounded by $1$, and equal to $1$ on the ball $\textnormal{{B}}(0,1/\varepsilon)$.
Our regularized system writes
$$\displaystyle[left={\empheqlbrace}]$$
$$\displaystyle\textnormal{div}_{\mathbf{x}}\,\mathbf{u}_{\varepsilon}=0,$$
(26)
$$\displaystyle\partial_{t}\rho_{\varepsilon}+(\mathbf{u}_{\varepsilon}\star%
\varphi_{\varepsilon})\cdot\nabla_{\mathbf{x}}\,\rho_{\varepsilon}=m_{0}(f_{%
\varepsilon}\gamma_{\varepsilon}),$$
(27)
$$\displaystyle\partial_{t}f_{\varepsilon}+\boldsymbol{\xi}\cdot\nabla_{\mathbf{%
x}}f_{\varepsilon}+\nabla_{\boldsymbol{\xi}}\cdot[(\mathbf{u}_{\varepsilon}%
\star\varphi_{\varepsilon}-\boldsymbol{\xi})f_{\varepsilon}]=-f_{\varepsilon},$$
(28)
$$\displaystyle\mathbb{P}\Big{\{}(1+\rho_{\varepsilon})\big{[}\partial_{t}%
\mathbf{u}_{\varepsilon}+(\mathbf{u}_{\varepsilon}\star\varphi_{\varepsilon})%
\cdot\nabla_{\mathbf{x}}\mathbf{u}_{\varepsilon}\big{]}\Big{\}}-\Delta_{%
\mathbf{x}}\mathbf{u}_{\varepsilon}=2\mathbb{P}\Big{\{}\mathbf{m}_{1}(f_{%
\varepsilon}\gamma_{\varepsilon})-\mathbf{u}_{\varepsilon}m_{0}(f_{\varepsilon%
}\gamma_{\varepsilon})\Big{\}},$$
(29)
with initial conditions $f_{\textnormal{in}}^{\varepsilon}\in\mathscr{D}({\mathbb{T}}^{3}\times\mathbb{%
R}^{3})$, $\rho_{\textnormal{in}}^{\varepsilon}\in\mathscr{C}^{\infty}({\mathbb{T}}^{3})$ and $\mathbf{u}_{\textnormal{in}}^{\varepsilon}\in\mathscrbf{C}^{\infty}_{%
\textnormal{div}}({\mathbb{T}}^{3})$ approximating respectively $f_{\textnormal{in}}$ in all $\textnormal{L}_{\textnormal{loc}}^{p}({\mathbb{T}}^{3}\times\mathbb{R}^{3})$ ($p<\infty$) and in $\textnormal{L}^{1}({\mathbb{T}}^{3}\times\mathbb{R}^{3},(1+|\boldsymbol{\xi}|^%
{2})\,\mathrm{d}\mathbf{x}\,\mathrm{d}\boldsymbol{\xi})$, $\rho_{\textnormal{in}}$ in all $\textnormal{L}^{p}({\mathbb{T}}^{3})$ ($p<\infty$), and $\mathbf{u}_{\textnormal{in}}$ in $\mathbf{L}^{2}({\mathbb{T}}^{3})$.
Next subsection is devoted to the Proof of existence of solutions to this approximated system (that is, for a given
parameter $\varepsilon>0$).
2.2 Existence of solutions for the regularized system
In this Subsection, the parameter $\varepsilon$ is fixed, and we drop out the corresponding indices in order to make
the formulas more readable.
We fix a triplet of initial data $(\boldsymbol{\omega},h,\nu)\in\mathscrbf{C}^{\infty}_{\textnormal{div}}\times(%
\mathscr{C}^{1}\cap\textnormal{L}^{\infty})\times\mathscr{C}^{1}$ (the two last ones being nonnegative). The considered system (in which we do not explicitly write the initial data) is then the following:
$$\displaystyle[left={\empheqlbrace}]$$
$$\displaystyle\textnormal{div}_{\mathbf{x}}\,\mathbf{u}=0,$$
(30)
$$\displaystyle\partial_{t}\rho+(\mathbf{u}\star\varphi)\cdot\nabla_{\mathbf{x}}%
\,\rho=m_{0}(f\gamma),$$
(31)
$$\displaystyle\partial_{t}f+\boldsymbol{\xi}\cdot\nabla_{\mathbf{x}}f+\nabla_{%
\boldsymbol{\xi}}\cdot[(\mathbf{u}\star\varphi-\boldsymbol{\xi})f]=-f,$$
(32)
$$\displaystyle\mathbb{P}\Big{\{}(1+\rho)\big{[}\partial_{t}\mathbf{u}+(\mathbf{%
u}\star\varphi)\cdot\nabla_{\mathbf{x}}\mathbf{u}\big{]}\Big{\}}-\Delta_{%
\mathbf{x}}\mathbf{u}=2\mathbb{P}\Big{\{}\mathbf{m}_{1}(f\gamma)-\mathbf{u}m_{%
0}(f\gamma)\Big{\}}.$$
(33)
We are able to prove the
Proposition 2.1.
For any triplet of initial data $(\boldsymbol{\omega},h,\nu)\in\mathscrbf{C}^{\infty}_{\textnormal{div}}\times(%
\mathscr{C}^{1}\cap\textnormal{L}^{\infty})\times\mathscr{C}^{1}$ (the two last ones being nonnegative), there exists a weak solution (in the sense of Remark 1.1)
to the system (30) –
(33).
Proof of Proposition 2.1:
We shall show existence of a solution to the nonlinear system (30) – (33) by applying Schauder’s fixed point Theorem to the following mapping:
$$\displaystyle\textnormal{S}:\mathscr{C}^{0}_{t}(\mathbf{L}^{2}_{\textnormal{%
div}})$$
$$\displaystyle\longrightarrow\mathscr{C}^{0}_{t}(\mathbf{L}^{2}_{\textnormal{%
div}})$$
$$\displaystyle\mathbf{u}$$
$$\displaystyle\longmapsto\mathbf{u}^{\diamond},$$
where $\mathbf{u}^{\diamond}$ will be built thanks to the following steps :
(i)
We first consider $f^{\diamond}\in\mathscr{C}^{1}$, unique classical solution of
$$\displaystyle\partial_{t}f^{\diamond}+\boldsymbol{\xi}\cdot\nabla_{\mathbf{x}}%
f^{\diamond}+\nabla_{\boldsymbol{\xi}}\cdot[(\mathbf{u}\star\varphi-%
\boldsymbol{\xi})f^{\diamond}]=-f^{\diamond},$$
(34)
with $h$ as initial datum,
(ii)
then $\rho^{\diamond}\in\mathscr{C}^{1}$ is defined as the unique classical solution of
$$\displaystyle\partial_{t}\rho^{\diamond}+(\mathbf{u}\star\varphi)\cdot\nabla_{%
\mathbf{x}}\rho^{\diamond}=m_{0}(f^{\diamond}\gamma),$$
(35)
with $\nu$ as initial datum,
(iii)
and finally $\mathbf{u}^{\diamond}\in\mathscr{C}^{0}_{t}(\mathbf{L}^{2}_{\textnormal{div}})$ is built as the unique divergence-free weak solution of
$$\displaystyle\mathbb{P}\Bigg{\{}(1+\rho^{\diamond})\Big{[}\partial_{t}\mathbf{%
u}^{\diamond}+(\mathbf{u}\star\varphi)\cdot\nabla_{\mathbf{x}}\mathbf{u}^{%
\diamond}\Big{]}\Bigg{\}}-\Delta_{\mathbf{x}}\mathbf{u}^{\diamond}=2\mathbb{P}%
\Big{\{}\mathbf{m}_{1}(f^{\diamond}\gamma)-\mathbf{u}^{\diamond}m_{0}(f^{%
\diamond}\gamma)\Big{\}},$$
(36)
with $\boldsymbol{\omega}$ as initial datum.
In the sequel $\textnormal{P}_{T}(\boldsymbol{\omega},h,\nu)$ will denote any nonnegative function (which also may depend on $\varepsilon$, but as explained before we omit the corresponding index here) that splits into a finite sum of positively homogeneous functions of strictly positive degree of one of the following terms : $\|\boldsymbol{\omega}\|_{\mathbf{H}^{1}}$,$\|h\|_{\infty}$ or $\|\nu\|_{\infty}$. $\textnormal{P}_{T}(\boldsymbol{\omega},h,\nu)$ may change from one line to another,
but will always have the structure that we just described.
In the next paragraph, we show that $f^{\diamond}$, $u^{\diamond}$ and $\rho^{\diamond}$ are well defined.
2.2.1 Existence and uniqueness of $f^{\diamond}$, $\rho^{\diamond}$ and $\mathbf{u}^{\diamond}$
We know that $\mathbf{u}\star\varphi\in\mathscr{C}^{0}_{t}(\mathscrbf{C}^{1}_{\mathbf{x}})$ and $\nabla_{\mathbf{x}}(\mathbf{u}\star\varphi)\in\mathbf{L}^{\infty}$, so that the characteristic curves of (34) and (35)
are globally well-defined. We hence easily obtain the existence and uniqueness of two nonnegative
functions $f^{\diamond}$ and $\rho^{\diamond}$ (classically) solving
(34) and (35). Thanks to the maximum principle, we also know that
$$\displaystyle\|f^{\diamond}\|_{\infty}\leq e^{2T}\|h\|_{\infty},$$
from which we get
$$\displaystyle\|m_{0}(f^{\diamond}\gamma)\|_{\infty}+\|\mathbf{m}_{1}(f^{%
\diamond}\gamma)\|_{\infty}\leq C_{\gamma}e^{2T}\|h\|_{\infty}\leq\textnormal{%
P}_{T}(\boldsymbol{\omega},h,\nu),$$
(37)
and hence
$$\displaystyle\|\rho^{\diamond}\|_{\infty}$$
$$\displaystyle\leq\textnormal{P}_{T}(\boldsymbol{\omega},h,\nu).$$
(38)
We now consider $f^{\diamond}$ and $\rho^{\diamond}$ as given. The existence and uniqueness of $\mathbf{u}^{\diamond}\in\mathbf{H}^{1}_{t,\mathbf{x}}\subset\mathscr{C}_{t}^{0%
}(\mathbf{L}^{2}_{\mathbf{x}})$, divergence-free weak solution of (36) (with $\boldsymbol{\omega}$ for initial data) may be obtained thanks to the usual Galerkin method. Since this type of construction is rather standard, we won’t detail it here, and refer to [5] for a precise treatment of this procedure in our system, or for
instance to [3], Chapter 3, for a more generic approach.
In next subsection, we obtain natural estimates for the quantity $\mathbf{u}^{\diamond}$.
2.2.2 Estimates for $\mathbf{u}^{\diamond}$
Taking $\mathbf{u}^{\diamond}$ as test function in (36) and using (35),
we obtain the following estimate (since $\mathbf{u}\star\varphi$ is divergence free and $f^{\diamond}$ is nonnegative)
$$\displaystyle\frac{1}{2}\,\left\{\int_{{\mathbb{T}}^{3}}(1+\rho^{\diamond}(t,%
\mathbf{x}))|\mathbf{u}^{\diamond}(t,\mathbf{x})|^{2}\mathrm{d}\mathbf{x}\right\}$$
$$\displaystyle+\int_{0}^{t}\int_{{\mathbb{T}}^{3}}|\nabla_{\mathbf{x}}\mathbf{u%
}^{\diamond}(s,\mathbf{x})|^{2}\mathrm{d}\mathbf{x}\,\mathrm{d}s$$
$$\displaystyle\leq-\frac{3}{2}\,\int_{0}^{t}\int_{{\mathbb{T}}^{3}}|\mathbf{u}^%
{\diamond}(s,\mathbf{x})|^{2}m_{0}(f^{\diamond}\gamma)(s,\mathbf{x})\,\mathrm{%
d}\mathbf{x}\,\mathrm{d}s$$
$$\displaystyle+2\int_{0}^{t}\int_{{\mathbb{T}}^{3}}\mathbf{u}^{\diamond}(s,%
\mathbf{x})\cdot\mathbf{m}_{1}(f^{\diamond}\gamma)(s,\mathbf{x})\,\mathrm{d}%
\mathbf{x}\,\mathrm{d}s$$
$$\displaystyle+\frac{1}{2}\,\left\{\int_{{\mathbb{T}}^{3}}(1+\nu(\mathbf{x}))|%
\boldsymbol{\omega}(\mathbf{x})|^{2}\mathrm{d}\mathbf{x}\right\},$$
(39)
so that using $\rho^{\diamond}\geq 0$ and Grönwall’s lemma together with
estimates (37)–(38), we get
$$\displaystyle\|\mathbf{u}^{\diamond}\|_{\textnormal{L}^{\infty}_{t}(\mathbf{L}%
^{2}_{\mathbf{x}})}+\|\mathbf{u}^{\diamond}\|_{\textnormal{L}^{2}_{t}(\mathbf{%
H}^{1}_{\mathbf{x}})}\leq\textnormal{P}_{T}(\boldsymbol{\omega},h,\nu)\exp(%
\textnormal{P}_{T}(\boldsymbol{\omega},h,\nu)).$$
(40)
Taking then $\partial_{t}\mathbf{u}^{\diamond}$ as a test function in (36),
we get (using again $\rho^{\diamond}\geq 0$ and $f^{\diamond}\geq 0$)
$$\displaystyle\|\partial_{t}\mathbf{u}^{\diamond}\|_{2}^{2}+\frac{1}{2}\,\|%
\nabla_{\mathbf{x}}\mathbf{u}^{\diamond}(t)\|_{2}^{2}$$
$$\displaystyle\leq\|(1+\rho^{\diamond})(\mathbf{u}\star\varphi)\|_{\infty}\,\|%
\nabla_{\mathbf{x}}\mathbf{u}^{\diamond}\|_{2}\,\|\partial_{t}\mathbf{u}^{%
\diamond}\|_{2}+2\,\|\mathbf{m}_{1}(f^{\diamond}\gamma)\|_{2}\,\|\partial_{t}%
\mathbf{u}^{\diamond}\|_{2}$$
$$\displaystyle+2\,\|m_{0}(f^{\diamond}\gamma)\|_{\infty}\,\|\mathbf{u}^{%
\diamond}\|_{2}\,\|\partial_{t}\mathbf{u}^{\diamond}\|_{2}+\frac{1}{2}\,\|%
\nabla_{\mathbf{x}}\boldsymbol{\omega}\|_{2}^{2},$$
so that using Young’s inequality, estimate (40) above,
and estimates (37)–(38), we end up with
$$\displaystyle\|\partial_{t}\mathbf{u}^{\diamond}\|_{2}\leq\textnormal{P}_{T}(%
\boldsymbol{\omega},h,\nu)\,\exp(\textnormal{P}_{T}(\boldsymbol{\omega},h,\nu)%
)\,\Big{[}\|\mathbf{u}\|_{\textnormal{L}^{\infty}_{t}(\mathbf{L}^{2}_{\mathbf{%
x}})}+1\Big{]},$$
(41)
thanks to the elementary convolution inequality
$\|\mathbf{u}\star\varphi\|_{\infty}\leq\|\mathbf{u}\|_{\textnormal{L}^{\infty}%
_{t}(\mathbf{L}^{2}_{\mathbf{x}})}\|\varphi\|_{2}$.
Given any triplet $(\boldsymbol{\omega},h,\nu)\in\mathscrbf{C}^{\infty}_{\textnormal{div}}\times%
\mathscr{C}^{1}\cap\textnormal{L}^{\infty}\times\mathscr{C}^{1}$ of initial data, we can then define the mapping
$$\displaystyle\textnormal{S}:\mathscr{C}^{0}_{t}(\mathbf{L}^{2}_{\textnormal{%
div}})$$
$$\displaystyle\longrightarrow\mathscr{C}^{0}_{t}(\mathbf{L}^{2}_{\textnormal{%
div}})$$
$$\displaystyle\mathbf{u}$$
$$\displaystyle\longmapsto\mathbf{u}^{\diamond}.$$
In next paragraph, we use Schauder’s fixed point Theorem for this mapping.
2.2.3 Application of Schauder’s Theorem
According to the estimates of the previous paragraph, we see that
for any $\mathbf{u}\in\mathscr{C}^{0}_{t}(\mathbf{L}^{2}_{\textnormal{div}})$,
$$\displaystyle\|\textnormal{S}(\mathbf{u})\|_{\textnormal{L}^{\infty}_{t}(%
\mathbf{L}^{2}_{\mathbf{x}})}\leq\textnormal{P}_{T}(\boldsymbol{\omega},h,\nu)%
\exp(\textnormal{P}_{T}(\boldsymbol{\omega},h,\nu)):=R.$$
Hence, S sends $\textnormal{B}_{R}$ (the closed ball of radius $R$ of the normed space $\mathscr{C}^{0}_{t}(\mathbf{L}^{2}_{\textnormal{div}})$) to itself.
Thanks to estimates (40) and (41), a bounded subset of $\mathscr{C}^{0}_{t}(\mathbf{L}^{2}_{\textnormal{div}})$ is sent by S to a bounded subset of $\mathbf{H}^{1}_{t,\mathbf{x}}$, and Lemma 4.3 of the Appendix allows us to conclude that $\textnormal{S}(\textnormal{B}_{R})$ is relatively compact in $\mathscr{C}^{0}_{t}(\mathbf{L}^{2}_{\textnormal{div}})$.
It remains to study the continuity of S.
Notice first that for any $\mathbf{u}\in\mathscr{C}^{0}_{t}(\mathbf{L}^{2}_{\textnormal{div}})$, any weak bounded solution of (34) or (35) is a renormalized solution in the DiPerna-Lions sense introduced in [10]: the absorption coefficient ($=1$) of the first equation and the right-hand side of second one are bounded and one easily checks that the vector fields $(t,\mathbf{x},\boldsymbol{\xi})\mapsto(\boldsymbol{\xi},\mathbf{u}\star\varphi%
(t,\mathbf{x})-\boldsymbol{\xi})$ and $(t,\mathbf{x})\mapsto\mathbf{u}\star\varphi(t,\mathbf{x})$ satisfy all the desired assumptions given in [10].
Now if $(\mathbf{u}_{n})_{n}$ converges to $\mathbf{u}$ in $\mathscr{C}^{0}_{t}(\mathbf{L}^{2}_{\textnormal{div}})$, we know from the previous step that $(\textnormal{S}(\mathbf{u}_{n}))_{n\in\mathbb{N}}$ has a converging subsequence. Showing that the whole sequence converges to $\textnormal{S}(\mathbf{u})$ reduces hence to prove that it has only one accumulation point, namely $\textnormal{S}(\mathbf{u})$. Assume therefore that $\mathbf{v}\in\mathscr{C}^{0}_{t}(\mathbf{L}^{2}_{\textnormal{div}})$ is an accumulation point of $(\textnormal{S}(\mathbf{u}_{n}))_{n\in\mathbb{N}}$ and denote by $\sigma$ the corresponding extraction. Since $(\mathbf{u}_{\sigma(n)})_{n}$ is still bounded, using (40)–(41) and adding a subsequence if necessary,
we may assume that $\mathbf{v}$ is also the limit of $(\textnormal{S}(\mathbf{u}_{n}))_{n\in\mathbb{N}}$ in $\mathbf{H}^{1}_{t,\mathbf{x}}$, for the weak topology. On the other hand, since $(\mathbf{u}_{\sigma(n)})_{n}$ still converges to $\mathbf{u}$ in $\mathscr{C}^{0}_{t}(\mathbf{L}^{2}_{\textnormal{div}})$, DiPerna-Lions stability result of [10] ensures that $(f_{\sigma(n)})_{n}$ and $(\rho_{\sigma(n)})_{n}$ both converge in $\textnormal{L}^{p}_{\textnormal{loc}}$
(for all $p<\infty$) to $f$ and $\rho$, the corresponding (unique) weak and bounded solutions of the associated equations (with initial data $h$ and $\nu$ and vector field defined by $\mathbf{u}$). At this point, we have enough (strong) convergences to ensure that all the nonlinear terms of the fluid equation indeed
converge to the expected limit : $\mathbf{v}\in\mathbf{H}^{1}_{t,\mathbf{x}}$ is hence a weak solution of (36) with $\boldsymbol{\omega}$ as initial datum. We already mentioned the uniqueness of solutions for this equation, so that we eventually get $\mathbf{v}=\mathbf{u}^{\diamond}=\textnormal{S}(\mathbf{u})$.
S is hence a continuous map from a closed convex nonempty
set of a normed space to itself, and it has a relatively compact range.
Thanks to Schauder’s fixed point Theorem (see [11] for instance), S has a fixed point which is a
solution to (30) – (33) with initial data $(\boldsymbol{\omega},h,\nu)$.
This concludes the Proof of Proposition 2.1. $\square$
Reintroducing the parameter $\varepsilon>0$, we have obtained the existence (for any $T>0$, $\varepsilon>0$) of
a divergence-free vector field $\mathbf{u}_{\varepsilon}\in\textnormal{H}^{1}_{t,\mathbf{x}}$, and two nonnegative functions $f_{\varepsilon}\in\mathscr{C}^{1}_{t,\mathbf{x},\boldsymbol{\xi}}$ and $\rho_{\varepsilon}\in\mathscr{C}^{1}_{t,\mathbf{x}}$ solutions of system (26) – (29).
Note that since $\mathbf{u}_{\varepsilon}\star\varphi_{\varepsilon}\in\textnormal{L}^{\infty}_{%
t,\mathbf{x}}$ and $f_{\textnormal{in}}^{\varepsilon}$ is compactly supported, one can
show without difficulty that $f^{\varepsilon}(t,\cdot,\cdot)$ remains compactly supported for a.e. $t$ (with a support depending on $\varepsilon$).
In next section, we pass to the limit when $\varepsilon\to 0$ in the functions $\mathbf{u}_{\varepsilon}$, $f_{\varepsilon}$ and $\rho_{\varepsilon}$, and show that their limit is a (weak) solution of (15) – (18).
3 Passage to the limit and Proof of the main Theorem
We keep on proving Theorem 1.1 in this Section. We start with a local result of existence:
Proposition 3.1.
Under the assumptions on the initial data of Theorem 1.1, there exists
an interval $J_{\textnormal{in}}:=[0,t_{\star}]$ on which the system (15) – (18) admits a weak
solution $(\rho,\,f,\,\mathbf{u})$ (cf. Remark 1.1 for a definition of weak solutions).
Furthermore $\rho$ and $f$ are nonnegative, and
$$\displaystyle\rho$$
$$\displaystyle\in\textnormal{L}^{\infty}(J_{\textnormal{in}};\textnormal{L}^{5/%
3}({\mathbb{T}}^{3})),$$
$$\displaystyle f$$
$$\displaystyle\in\textnormal{L}^{\infty}(J_{\textnormal{in}}\times{\mathbb{T}}^%
{3}\times{\mathbb{R}^{3}}),$$
$$\displaystyle\mathbf{u}$$
$$\displaystyle\in\textnormal{L}^{2}(J_{\textnormal{in}};\textnormal{H}^{1}({%
\mathbb{T}}^{3}))\cap\textnormal{L}^{\infty}(J_{\textnormal{in}};\textnormal{L%
}^{2}({\mathbb{T}}^{3})).$$
Proof of Proposition 3.1 :
We begin with the Proof of bounds for the
solution of (26) – (29), which do not depend on $\varepsilon$.
In the sequel $\textnormal{P}_{T}(\mathbf{u}_{\textnormal{in}},\rho_{\textnormal{in}},f_{%
\textnormal{in}})$ [or more simply $\textnormal{P}_{T}^{\textnormal{in}}$] will denote a polynomial function (with positive coefficients) of the quantity $\|\sqrt{1+\rho_{\textnormal{in}}}\,\mathbf{u}_{\textnormal{in}}\|_{2}+\|f_{%
\textnormal{in}}\|_{\infty}+M_{2}f_{\textnormal{in}}$. Note
that $\textnormal{P}_{T}^{\textnormal{in}}$ will always be independent of $\varepsilon$, but may depend on $T$.
3.1 Uniform bounds with respect to $\varepsilon$
Let us first recall a classical Lemma linking velocity moments, the proof of which may be found in [6] (Lemma 1, Section 3.3):
Lemma 3.1.
Let $\gamma>0$ and $h$ be a nonnegative element of $\textnormal{L}^{\infty}([0,T]\times{\mathbb{T}}^{3}\times\mathbb{R}^{3})$, such that $m_{\gamma}h(t,\mathbf{x})<+\infty$ for a.e. $(t,\mathbf{x})$. The following estimate holds for $0\leq\alpha<\gamma$:
$$\displaystyle m_{\alpha}h(t,\mathbf{x})\leq\left(\frac{4}{3}\pi\|h(t,\mathbf{x%
},\cdot)\|_{\textnormal{L}^{\infty}(\mathbb{R}^{3}_{\boldsymbol{\xi}})}+1%
\right)m_{\gamma}h(t,\mathbf{x})^{\frac{\alpha+3}{\gamma+3}}.$$
Thanks to the maximum principle in eq. (28), we observe that
$$\displaystyle\|f_{\varepsilon}\|_{\infty}$$
$$\displaystyle\leq\|f_{\textnormal{in}}\|_{\infty}\,e^{2T},$$
(42)
from which, thanks to Lemma 3.1, we deduce
$$\displaystyle\left\|{m_{0}f_{\varepsilon}(t)}\right\|_{5/3}$$
$$\displaystyle\leq\textnormal{P}_{T}^{\textnormal{in}}\,M_{2}f_{\varepsilon}(t)%
^{3/5},$$
(43)
$$\displaystyle\left\|{\mathbf{m}_{1}f_{\varepsilon}(t)}\right\|_{5/4}$$
$$\displaystyle\leq\textnormal{P}_{T}^{\textnormal{in}}\,M_{2}f_{\varepsilon}(t)%
^{4/5}.$$
(44)
Since $f_{\varepsilon}$ is compactly supported and solves eq. (28) strongly, one gets by
multiplying this equation by $|\boldsymbol{\xi}|^{2}$ and integrating in $\mathbf{x},\boldsymbol{\xi}$:
$$\displaystyle\frac{\mathrm{d}}{\mathrm{d}t}M_{2}f_{\varepsilon}+3M_{2}f_{%
\varepsilon}=2\int_{{\mathbb{T}}^{3}}(\mathbf{u}_{\varepsilon}\star\varphi_{%
\varepsilon})\cdot\mathbf{m}_{1}(f_{\varepsilon})\,\mathrm{d}\mathbf{x}.$$
(45)
Using the previous estimates, we hence have
$$\displaystyle\frac{\mathrm{d}}{\mathrm{d}t}M_{2}f_{\varepsilon}+3M_{2}f_{\varepsilon}$$
$$\displaystyle\leq\left\|{\mathbf{u}_{\varepsilon}(t)}\right\|_{5}\,\stackrel{{%
\scriptstyle=1}}{{\overbrace{\left\|{\varphi_{\varepsilon}}\right\|_{1}}}}\,%
\left\|{\mathbf{m}_{1}f_{\varepsilon}(t)}\right\|_{5/4}$$
$$\displaystyle\leq\textnormal{P}_{T}^{\textnormal{in}}\,\left\|{\mathbf{u}_{%
\varepsilon}(t)}\right\|_{5}\,M_{2}f_{\varepsilon}(t)^{4/5},$$
from which we easily deduce
$$\displaystyle M_{2}f_{\varepsilon}(t)\leq\left\{M_{2}f_{\textnormal{in}}^{%
\varepsilon}+\textnormal{P}_{T}^{\textnormal{in}}\int_{0}^{t}\left\|{\mathbf{u%
}_{\varepsilon}(s)}\right\|_{5}\mathrm{d}s\right\}^{5}.$$
Using the Sobolev injection $\mathbf{H}^{1}({\mathbb{T}}^{3})\hookrightarrow\mathbf{L}^{5}({\mathbb{T}}^{3})$, one gets
$$\displaystyle M_{2}f_{\varepsilon}(t)\leq\textnormal{P}_{T}^{\textnormal{in}}%
\left\{1+\int_{0}^{t}\left\|{\mathbf{u}_{\varepsilon}(s)}\right\|_{2}\mathrm{d%
}s+\int_{0}^{t}\left\|{\nabla_{\mathbf{x}}\mathbf{u}_{\varepsilon}(s)}\right\|%
_{2}\mathrm{d}s\right\}^{5}.$$
(46)
Using the previous estimate with (43) – (44), we obtain
$$\displaystyle\left\|{m_{0}f_{\varepsilon}(t)}\right\|_{5/3}$$
$$\displaystyle\leq\textnormal{P}_{T}^{\textnormal{in}}\left\{1+\int_{0}^{t}%
\left\|{\mathbf{u}_{\varepsilon}(s)}\right\|_{2}\mathrm{d}s+\int_{0}^{t}\left%
\|{\nabla_{\mathbf{x}}\mathbf{u}_{\varepsilon}(s)}\right\|_{2}\mathrm{d}s%
\right\}^{3},$$
(47)
$$\displaystyle\left\|{\mathbf{m}_{1}f_{\varepsilon}(t)}\right\|_{5/4}$$
$$\displaystyle\leq\textnormal{P}_{T}^{\textnormal{in}}\left\{1+\int_{0}^{t}%
\left\|{\mathbf{u}_{\varepsilon}(s)}\right\|_{2}\mathrm{d}s+\int_{0}^{t}\left%
\|{\nabla_{\mathbf{x}}\mathbf{u}_{\varepsilon}(s)}\right\|_{2}\mathrm{d}s%
\right\}^{4}.$$
(48)
The energy estimate (39) is satisfied with $\nu:=\rho_{\textnormal{in}}^{\varepsilon}$, $\boldsymbol{\omega}:=\mathbf{u}_{\textnormal{in}}^{\varepsilon}$, and rewrites
$$\displaystyle\frac{1}{2}\left\{\int_{{\mathbb{T}}^{3}}(1+\rho_{\varepsilon}(t,%
\mathbf{x}))|\mathbf{u}_{\varepsilon}(t,\mathbf{x})|^{2}\mathrm{d}\mathbf{x}\right\}$$
$$\displaystyle+\int_{0}^{t}\int_{{\mathbb{T}}^{3}}|\nabla_{\mathbf{x}}\mathbf{u%
}_{\varepsilon}(s,\mathbf{x})|^{2}\,\mathrm{d}\mathbf{x}\,\mathrm{d}s$$
$$\displaystyle\leq-\frac{3}{2}\int_{0}^{t}\int_{{\mathbb{T}}^{3}}|\mathbf{u}_{%
\varepsilon}(s,\mathbf{x})|^{2}m_{0}(f_{\varepsilon}\gamma)(s,\mathbf{x})\,%
\mathrm{d}\mathbf{x}\,\mathrm{d}s$$
$$\displaystyle+2\int_{0}^{t}\int_{{\mathbb{T}}^{3}}\mathbf{u}_{\varepsilon}(s,%
\mathbf{x})\cdot\mathbf{m}_{1}(f_{\varepsilon}\gamma)(s,\mathbf{x})\,\mathrm{d%
}\mathbf{x}\,\mathrm{d}s$$
$$\displaystyle+\frac{1}{2}\left\{\int_{{\mathbb{T}}^{3}}(1+\rho_{\textnormal{in%
}}^{\varepsilon}(\mathbf{x}))\,|\mathbf{u}_{\textnormal{in}}^{\varepsilon}(%
\mathbf{x})|^{2}\,\mathrm{d}\mathbf{x}\right\},$$
so that (since $\rho_{\varepsilon}$ is nonnegative) we obtain
$$\displaystyle\frac{1}{2}\,\|\mathbf{u}_{\varepsilon}(t)\|_{2}^{2}+\int_{0}^{t}%
\|\nabla_{\mathbf{x}}\mathbf{u}_{\varepsilon}(s)\|_{2}^{2}\,\mathrm{d}s$$
$$\displaystyle\leq\frac{1}{2}\|\sqrt{1+\rho_{\textnormal{in}}^{\varepsilon}}\,%
\mathbf{u}_{\textnormal{in}}^{\varepsilon}\|_{2}^{2}+2\int_{0}^{t}\|\mathbf{u}%
_{\varepsilon}\cdot\mathbf{m}_{1}f_{\varepsilon}(s)\|_{1}\,\mathrm{d}s$$
$$\displaystyle\leq\textnormal{P}_{T}^{\textnormal{in}}+2\int_{0}^{t}\|\mathbf{u%
}_{\varepsilon}\cdot\mathbf{m}_{1}f_{\varepsilon}(s)\|_{1}\,\mathrm{d}s.$$
Using Hölder and Young inequalities, we get
$$\displaystyle\frac{1}{2}\|\mathbf{u}_{\varepsilon}(t)\|_{2}^{2}+\frac{1}{2}%
\int_{0}^{t}\|\nabla_{\mathbf{x}}\mathbf{u}_{\varepsilon}(s)\|_{2}^{2}\,%
\mathrm{d}s$$
$$\displaystyle\leq\textnormal{P}_{T}^{\textnormal{in}}+\textnormal{P}_{T}^{%
\textnormal{in}}\int_{0}^{t}\|\mathbf{u}_{\varepsilon}(s)\|_{2}^{2}\,\mathrm{d%
}s+\textnormal{P}_{T}^{\textnormal{in}}\int_{0}^{t}\|\mathbf{m}_{1}f_{%
\varepsilon}(s)\|_{5/4}^{2}\,\mathrm{d}s.$$
We then obtain thanks to (48)
$$\displaystyle\|\mathbf{u}_{\varepsilon}(t)\|_{2}^{2}+\int_{0}^{t}\|\nabla_{%
\mathbf{x}}\mathbf{u}_{\varepsilon}(s)\|_{2}^{2}\,\mathrm{d}s$$
$$\displaystyle\leq\textnormal{P}_{T}^{\textnormal{in}}+\textnormal{P}_{T}^{%
\textnormal{in}}\int_{0}^{t}\left\{\|\mathbf{u}_{\varepsilon}(s)\|_{2}^{2}+%
\int_{0}^{s}\left\|{\nabla\mathbf{u}_{\varepsilon}(\sigma)}\right\|_{2}^{2}\,%
\mathrm{d}\sigma\right\}^{4}\mathrm{d}s.$$
Denote by $\tilde{z}$ the maximal solution of the Cauchy problem
$$z^{\prime}(t)=\textnormal{P}_{T}^{\textnormal{in}}\,z(t)^{4},\qquad z(0)=%
\textnormal{P}_{T}^{\textnormal{in}},$$
defined on some maximal interval $I_{\textnormal{in}}$.
Then if $J_{\textnormal{in}}$ is the closure of $I_{\textnormal{in}}/2$, using the nonlinear Grönwall Lemma 4.1, we get for $t\in J_{\textnormal{in}}$:
$$\displaystyle\left\|{\mathbf{u}_{\varepsilon}(t)}\right\|_{2}^{2}+\int_{0}^{t}%
\left\|{\nabla\mathbf{u}_{\varepsilon}(s)}\right\|_{2}^{2}\mathrm{d}s\leq\|%
\tilde{z}\|_{\textnormal{L}^{\infty}}:=\textnormal{K}_{T}^{\textnormal{in}}.$$
Remark 3.1.
Notice that thanks to Lemma 4.2 (changing $\textnormal{P}_{T}^{\textnormal{in}}$ if necessary) we always have $|J_{\textnormal{in}}|\geq 1/\textnormal{P}_{T}^{\textnormal{in}}$.
We deduce from the previous local estimate that $(\mathbf{u}_{\varepsilon})_{\varepsilon}$ is bounded in $\textnormal{L}^{\infty}\big{(}J_{\textnormal{in}};\textnormal{L}^{2}({\mathbb{%
T}}^{3})\big{)}\cap\textnormal{L}^{2}\big{(}J_{\textnormal{in}};\textnormal{H}%
^{1}({\mathbb{T}}^{3})\big{)}$. Using (46), we hence get the boundedness of $(M_{2}f_{\varepsilon})_{\varepsilon}$ in $\textnormal{L}^{\infty}(J_{\textnormal{in}})$ and then,
with (43) and (44), we see that $(m_{0}f_{\varepsilon})_{\varepsilon}$ and $(\mathbf{m}_{1}f_{\varepsilon})_{\varepsilon}$ are respectively bounded in $\textnormal{L}^{\infty}\big{(}J_{\textnormal{in}};\textnormal{L}^{5/3}({%
\mathbb{T}}^{3})\big{)}$ and $\textnormal{L}^{\infty}\big{(}J_{\textnormal{in}};\textnormal{L}^{5/4}({%
\mathbb{T}}^{3})\big{)}$.
Thanks to a classical transport estimate, using (28), we see that $(\rho_{\varepsilon})_{\varepsilon}$ is bounded in $\textnormal{L}^{\infty}\big{(}J_{\textnormal{in}};\textnormal{L}^{5/3}({%
\mathbb{T}}^{3})\big{)}$.
From (46), we deduce that for all $\varepsilon>0$ and $t\in J_{\textnormal{in}}$,
$$\displaystyle M_{2}{f_{\varepsilon}}(t)\leq\textnormal{K}_{T}^{\textnormal{in}}.$$
(49)
3.2 Compactness properties
Recalling Proposition 2.1 of existence of a solution to the regularized problem (for a given $\varepsilon$), we deduce
from the previous (uniform in $\varepsilon$) bounds
the existence of $\rho\in\textnormal{L}^{\infty}(J_{\textnormal{in}};\textnormal{L}^{5/3}({%
\mathbb{T}}^{3}))$, $f\in\textnormal{L}^{\infty}(J_{\textnormal{in}};\textnormal{L}^{\infty}({%
\mathbb{T}}^{3}\times\mathbb{R}^{3}))$ such that,
up to a subsequence,
$$\displaystyle(\rho_{\varepsilon})_{\varepsilon}$$
$$\displaystyle\operatorname*{\rightharpoonup}_{\varepsilon\rightarrow 0}\rho%
\text{ in }\textnormal{L}^{\infty}(J_{\textnormal{in}};\textnormal{L}^{5/3}({%
\mathbb{T}}^{3}))-\star,$$
$$\displaystyle(f_{\varepsilon})_{\varepsilon}$$
$$\displaystyle\operatorname*{\rightharpoonup}_{\varepsilon\rightarrow 0}f\text{%
in }\textnormal{L}^{\infty}(J_{\textnormal{in}};\textnormal{L}^{\infty}({%
\mathbb{T}}^{3}\times\mathbb{R}^{3}))-\star.$$
Using the boundedness of $(M_{2}f_{\varepsilon})_{\varepsilon}$ in $\textnormal{L}^{\infty}(J_{\textnormal{in}})$ and adding a subsequence if necessary, one manages to also show that
$$\displaystyle(m_{0}f_{\varepsilon})_{\varepsilon}\text{ and }(m_{0}(f_{%
\varepsilon}\gamma_{\varepsilon}))_{\varepsilon}$$
$$\displaystyle\operatorname*{\rightharpoonup}_{\varepsilon\rightarrow 0}m_{0}f%
\text{ in }\textnormal{L}^{\infty}(J_{\textnormal{in}};\textnormal{L}^{5/3}({%
\mathbb{T}}^{3}))-\star,$$
$$\displaystyle(\mathbf{m}_{1}f_{\varepsilon})_{\varepsilon}\text{ and }(\mathbf%
{m}_{1}(f_{\varepsilon}\gamma_{\varepsilon}))_{\varepsilon}$$
$$\displaystyle\operatorname*{\rightharpoonup}_{\varepsilon\rightarrow 0}\mathbf%
{m}_{1}f\text{ in }\textnormal{L}^{\infty}(J_{\textnormal{in}};\mathbf{L}^{5/4%
}({\mathbb{T}}^{3}))-\star.$$
Finally, using the bounds on $(\mathbf{u}_{\varepsilon})_{\varepsilon}$ and the Sobolev injection $\mathbf{H}^{1}({\mathbb{T}}^{3})\hookrightarrow\mathbf{L}^{6}({\mathbb{T}}^{3})$, we get
the existence of $\mathbf{u}\in\textnormal{L}^{\infty}(J_{\textnormal{in}};\mathbf{L}^{2}({%
\mathbb{T}}^{3}))\cap\textnormal{L}^{2}(J_{\textnormal{in}};\mathbf{H}_{%
\textnormal{div}}^{1}({\mathbb{T}}^{3}))$ such that
$$\displaystyle(\mathbf{u}_{\varepsilon})_{\varepsilon}\operatorname*{%
\rightharpoonup}_{\varepsilon\rightarrow 0}\mathbf{u}$$
$$\displaystyle\text{ in }\textnormal{L}^{2}\big{(}J_{\textnormal{in}};\mathbf{H%
}^{1}({\mathbb{T}}^{3})\big{)},$$
$$\displaystyle(\mathbf{u}_{\varepsilon})_{\varepsilon}\operatorname*{%
\rightharpoonup}_{\varepsilon\rightarrow 0}\mathbf{u}$$
$$\displaystyle\text{ in }\textnormal{L}^{2}\big{(}J_{\textnormal{in}};\mathbf{L%
}^{6}({\mathbb{T}}^{3})\big{)},$$
$$\displaystyle(\mathbf{u}_{\varepsilon})_{\varepsilon}\operatorname*{%
\rightharpoonup}_{\varepsilon\rightarrow 0}\mathbf{u}$$
$$\displaystyle\text{ in }\textnormal{L}^{\infty}\big{(}J_{\textnormal{in}};%
\mathbf{L}^{2}({\mathbb{T}}^{3})\big{)}-\star,$$
and of course $(\mathbf{u}_{\varepsilon}\star\varphi_{\varepsilon})_{\varepsilon}$ converges also towards $\mathbf{u}$ for the same topologies.
3.3 Weak convergence of nonlinear terms
In view of the previous weak convergences, it only remains to check that the nonlinear terms
converge to the expected limits, so that the weak limit triplet $(f,\rho,\mathbf{u})$ will indeed be a solution of our system.
Notice that since all the bounds and weak extractions are performed on the local interval $J_{\textnormal{in}}$, the constructed solution will only be local in time. In the last subsection, we shall explain how to extend it. Until then, through all the current subsection, the index ${}_{t}$ will refer to the intervall $J_{\textnormal{in}}$ in the notation “$\textnormal{L}_{t}^{p}(\textnormal{E})$”.
Notice that the system can be written
$$\displaystyle\textnormal{div}_{\mathbf{x}}\mathbf{u}_{\varepsilon}=0,$$
(50)
$$\displaystyle\partial_{t}f_{\varepsilon}+\textnormal{div}_{\mathbf{x},%
\boldsymbol{\xi}}(\mathbf{a}_{\varepsilon}f_{\varepsilon})-2\,f_{\varepsilon}=0,$$
(51)
$$\displaystyle\partial_{t}\rho_{\varepsilon}+\textnormal{div}_{\mathbf{x}}(\rho%
_{\varepsilon}(\mathbf{u}_{\varepsilon}\star\varphi_{\varepsilon}))=m_{0}(f_{%
\varepsilon}\gamma_{\varepsilon}),$$
(52)
$$\displaystyle\mathbb{P}\Big{\{}\partial_{t}[(1+\rho_{\varepsilon})\mathbf{u}_{%
\varepsilon}]+\textnormal{div}_{\mathbf{x}}\big{[}(1+\rho_{\varepsilon})(%
\mathbf{u}_{\varepsilon}\star\varphi_{\varepsilon})\otimes\mathbf{u}_{%
\varepsilon}\big{]}\Big{\}}-\Delta_{\mathbf{x}}\mathbf{u}_{\varepsilon}=%
\mathbb{P}\Big{\{}2\mathbf{m}_{1}(f_{\varepsilon}\gamma_{\varepsilon})-\mathbf%
{u}_{\varepsilon}m_{0}(f_{\varepsilon}\gamma_{\varepsilon})\Big{\}},$$
(53)
where $\mathbf{a}_{\varepsilon}(t,\mathbf{x},\boldsymbol{\xi}):=(\boldsymbol{\xi},[%
\mathbf{u}_{\varepsilon}(t)\star\varphi_{\varepsilon}](\mathbf{x}))$.
In order to handle the nonlinear terms, we shall use several times the Proposition 4.1 of the Appendix.
We denote by $\mathscr{M}_{s}$ the vector space of bounded measures on $J_{\textnormal{in}}\times{\mathbb{T}}^{3}$.
All the coming facts and their proofs are true up to some (finite number of) extractions that we don’t mention in the sequel.
Fact 1 :
The products $(\rho_{\varepsilon}\,(\mathbf{u}_{\varepsilon}\star\varphi_{\varepsilon}))_{\varepsilon}$ and $(\rho_{\varepsilon}\,\mathbf{u}_{\varepsilon})_{\varepsilon}$ both tend to $\rho\,\mathbf{u}$ in $\mathscrbf{M}_{s}$.
Proof of Fact 1: Using Proposition 4.1 of the Appendix, these two terms are handled in the same way: the velocity term (convoluted by $\varphi_{\varepsilon}$ or not) plays the role of $(a_{\varepsilon})_{\varepsilon}$, and is bounded in $\textnormal{L}^{2}_{t}(\textnormal{W}^{1,2}_{\mathbf{x}})$, whereas $(\rho_{\varepsilon})_{\varepsilon}$ plays the role of $(b_{\varepsilon})_{\varepsilon}$ and is bounded in $\textnormal{L}^{\infty}_{t}(\textnormal{L}^{5/3}_{\mathbf{x}})$ ($(5/3)^{\prime}=5/2<2^{\star}=6$). Thanks to (52) and the previous bounds (see subsection 3.1), $(\partial_{t}\rho_{\varepsilon})_{\varepsilon}$ is bounded in $\textnormal{L}^{2}_{t}(\textnormal{H}^{-m}_{\mathbf{x}})$, (where $m$ is taken large enough). We hence have
$$\displaystyle(\rho_{\varepsilon}\,(\mathbf{u}_{\varepsilon}\star\varphi_{%
\varepsilon}))_{\varepsilon}$$
$$\displaystyle\operatorname*{\rightharpoonup}_{\varepsilon\rightarrow 0}\rho\,%
\mathbf{u},\text{ in }\mathscrbf{M}_{s}-\text{w}\star,$$
$$\displaystyle(\rho_{\varepsilon}\,\mathbf{u}_{\varepsilon})_{\varepsilon}$$
$$\displaystyle\operatorname*{\rightharpoonup}_{\varepsilon\rightarrow 0}\rho\,%
\mathbf{u},\text{ in }\mathscr{M}_{s}-\text{w}\star.$$
Fact 2 :
$\displaystyle\big{\langle}(\rho_{\varepsilon}+1)\mathbf{u}_{\varepsilon},%
\mathbf{u}_{\varepsilon}\big{\rangle}_{\mathbf{L}^{2}_{t,\mathbf{x}}}%
\operatorname*{\longrightarrow}_{\varepsilon\rightarrow 0}\big{\langle}(\rho+1%
)\mathbf{u},\mathbf{u}\big{\rangle}_{\mathbf{L}^{2}_{t,\mathbf{x}}}$.
Proof of Fact 2: We wish to prove that
$$\displaystyle\operatorname*{\lim}_{\varepsilon\rightarrow 0}\int_{J_{%
\textnormal{in}}}\int_{{\mathbb{T}}^{3}}(\rho_{\varepsilon}+1)\,|\mathbf{u}_{%
\varepsilon}|^{2}\,\mathrm{d}\mathbf{x}\,\mathrm{d}t=\int_{J_{\textnormal{in}}%
}\int_{{\mathbb{T}}^{3}}(\rho+1)\,|\mathbf{u}|^{2}\,\mathrm{d}\mathbf{x}\,%
\mathrm{d}t.$$
(54)
First write, since $\mathbf{u}_{\varepsilon}$ is divergence-free,
$$\displaystyle\int_{J_{\textnormal{in}}}\int_{{\mathbb{T}}^{3}}(\rho_{%
\varepsilon}+1)\,|\mathbf{u}_{\varepsilon}|^{2}\,\mathrm{d}\mathbf{x}\,\mathrm%
{d}t$$
$$\displaystyle=\int_{J_{\textnormal{in}}}\int_{{\mathbb{T}}^{3}}(\rho_{%
\varepsilon}+1)\,\mathbf{u}_{\varepsilon}\cdot\mathbf{u}_{\varepsilon}\,%
\mathrm{d}\mathbf{x}\,\mathrm{d}t$$
$$\displaystyle=\int_{J_{\textnormal{in}}}\int_{{\mathbb{T}}^{3}}\mathbb{P}\big{%
[}(\rho_{\varepsilon}+1)\mathbf{u}_{\varepsilon}\big{]}\cdot\mathbf{u}_{%
\varepsilon}\,\mathrm{d}\mathbf{x}\,\mathrm{d}t.$$
Harmonic analysis and singular integral theory allow to show that $\mathbb{P}$ is bounded from $\textnormal{L}^{q}_{t}(\textnormal{L}^{p}_{\mathbf{x}})$ to itself, for all $p\in]1,\infty[$ and $q\in[1,\infty]$ (see for instance [19]). Since the strong continuity of an operator implies its weak sequential continuity, we see that $\mathbb{P}$ is sequentially continuous from $\textnormal{L}^{2}_{t}(\textnormal{L}^{30/23}_{\mathbf{x}})$ to itself, equipped with the weak topology. But we have seen in Subsection 3.1 that $(\mathbf{u}_{\varepsilon})_{\varepsilon}$ is bounded in $\textnormal{L}^{2}_{t}(\mathbf{H}^{1}_{\mathbf{x}})$, which is embedded in $\textnormal{L}^{2}_{t}(\mathbf{L}^{6}_{\mathbf{x}})$ by Sobolev injection, and $(\rho_{\varepsilon})_{\varepsilon}$ is bounded in $\textnormal{L}^{\infty}_{t}(\textnormal{L}^{5/3}_{\mathbf{x}})$ so that by Hölder inequality, $((\rho_{\varepsilon}+1)\mathbf{u}_{\varepsilon})_{\varepsilon}$ is bounded in $\textnormal{L}^{2}_{t}(\mathbf{L}^{30/23}_{\mathbf{x}})$ and hence (up to a subsequence) $\Big{(}\mathbb{P}\big{[}(\rho_{\varepsilon}+1)\,\mathbf{u}_{\varepsilon}\big{]%
}\Big{)}_{\varepsilon}$ converges weakly to $\mathbb{P}\big{[}(\rho+1)\,\mathbf{u}\big{]}$ in $\textnormal{L}^{2}_{t}(\textnormal{L}^{30/23}_{\mathbf{x}})$. We now can use Proposition 4.2 of the Appendix, with $q=2$ and $r=\infty$. Indeed,
–
$(\mathbf{u}_{\varepsilon})_{\varepsilon}$ is bounded in $\textnormal{L}^{2}_{t}(\textnormal{W}^{1,2}_{\mathbf{x}})\cap\textnormal{L}^{%
\infty}_{t}(\textnormal{L}^{2}_{\mathbf{x}})$,
–
since $\partial_{t}\mathbb{P}=\mathbb{P}\partial_{t}$, (53) implies that $\Big{(}\partial_{t}\mathbb{P}\big{[}(\rho_{\varepsilon}+1)\mathbf{u}_{%
\varepsilon}\big{]}\Big{)}_{\varepsilon}$ is bounded in $\textnormal{L}^{1}_{t}(\textnormal{H}_{\mathbf{x}}^{-m})$ for $m$ large enough,
–
$(30/23)^{\prime}=30/7<2^{\star}=6$.
Hence the product $\Big{\{}\mathbb{P}\big{[}(\rho_{\varepsilon}+1)\,\mathbf{u}_{\varepsilon}\big{%
]}\cdot\mathbf{u}_{\varepsilon}\Big{\}}_{\varepsilon}$ converges to $\mathbb{P}\big{[}(\rho+1)\,\mathbf{u}\big{]}\cdot\mathbf{u}$ in $\mathscrbf{M}_{s}-\star$. In particular, using
$\mathds{1}_{J_{\textnormal{in}}\times{\mathbb{T}}^{3}}$ as a test function, we get Fact 2.
Fact 3
$(\mathbf{u}_{\varepsilon})_{\varepsilon}$ and $(\mathbf{u}_{\varepsilon}\star\varphi_{\varepsilon})_{\varepsilon}$ both strongly converge to $\mathbf{u}$ in $\mathbf{L}^{2}_{t,\mathbf{x}}$.
Proof of Fact 3: We estimate
$$\displaystyle\int_{J_{\textnormal{in}}}\int_{{\mathbb{T}}^{3}}|\mathbf{u}_{%
\varepsilon}-\mathbf{u}|^{2}\,\mathrm{d}\mathbf{x}\,\mathrm{d}t$$
$$\displaystyle\leq\int_{J_{\textnormal{in}}}\int_{{\mathbb{T}}^{3}}(1+\rho_{%
\varepsilon})\,|\mathbf{u}_{\varepsilon}-\mathbf{u}|^{2}\,\mathrm{d}\mathbf{x}%
\,\mathrm{d}t$$
$$\displaystyle=\int_{J_{\textnormal{in}}}\int_{{\mathbb{T}}^{3}}(1+\rho_{%
\varepsilon})\,|\mathbf{u}_{\varepsilon}|^{2}\,\mathrm{d}\mathbf{x}\,\mathrm{d%
}t+\int_{J_{\textnormal{in}}}\int_{{\mathbb{T}}^{3}}(1+\rho_{\varepsilon})\,|%
\mathbf{u}|^{2}\,\mathrm{d}\mathbf{x}\,\mathrm{d}t$$
$$\displaystyle-2\int_{J_{\textnormal{in}}}\int_{{\mathbb{T}}^{3}}(1+\rho_{%
\varepsilon})\,\mathbf{u}_{\varepsilon}\cdot\mathbf{u}\,\mathrm{d}\mathbf{x}\,%
\mathrm{d}t.$$
The first term of the second line converges to the same expression, but without $\varepsilon$: this
is exactly Fact 2 proven above. We have the same behavior for the second term of this line:
$(\rho_{\varepsilon})_{\varepsilon}$ converges weakly to $\rho$ in $\textnormal{L}^{\infty}_{t}(\textnormal{L}^{5/3}_{\mathbf{x}})-\star$ and $|\mathbf{u}|^{2}\in\textnormal{L}^{1}_{t}(\textnormal{L}^{3}_{\mathbf{x}})$. Eventually, for the third and last term, we use the (already proven) convergence of $[(1+\rho_{\varepsilon})\,\mathbf{u}_{\varepsilon}]_{\varepsilon}$ towards $(1+\rho)\,\mathbf{u}$ in $\textnormal{L}^{2}_{t}(\mathbf{L}^{30/23}_{\mathbf{x}})$,
and the embedding $\mathbf{u}\in\textnormal{L}^{2}_{t}(\mathbf{L}^{6}_{\mathbf{x}})%
\hookrightarrow\textnormal{L}^{2}_{t}(\mathbf{L}^{30/7}_{\mathbf{x}})$.
Strong convergence of $(\mathbf{u}_{\varepsilon})_{\varepsilon}$ is then directly transfered to $(\mathbf{u}_{\varepsilon}\star\varphi_{\varepsilon})_{\varepsilon}$.
Fact 4
$(\mathbf{u}_{\varepsilon})_{\varepsilon}$ converges strongly to $\mathbf{u}$ in all $\textnormal{L}^{c}_{t}(\mathbf{L}^{d}_{\mathbf{x}})$, for $c<2$ and $d<6$.
Proof of Fact 4: Thanks to Fact 3, we have strong convergence of $(\mathbf{u}_{\varepsilon})_{\varepsilon}$,
and hence almost everywhere convergence. Since this family is bounded $\textnormal{L}^{2}_{t}(\mathbf{L}^{6}_{\mathbf{x}})$,
we get the desired convergences.
Fact 5
$(f_{\varepsilon}\,(\mathbf{u}_{\varepsilon}\star\varphi_{\varepsilon}))_{\varepsilon}$ and
$(\mathbf{u}_{\varepsilon}\,m_{0}(f_{\varepsilon}\gamma_{\varepsilon}))_{\varepsilon}$ converges weakly in $\mathbf{L}^{1}$ to respectively $f\,\mathbf{u}$ and $\mathbf{u}\,m_{0}(f)$.
Proof of Fact 5: The family $(f_{\varepsilon})_{\varepsilon}$ converges weakly to $f$ in $\textnormal{L}^{\infty}_{t}(\textnormal{L}^{\infty}_{\mathbf{x},\boldsymbol{%
\xi}})-\star$, which with the strong convergence (Fact 3) of $(\mathbf{u}_{\varepsilon}\star\varphi_{\varepsilon})_{\varepsilon}$ in $\mathbf{L}^{2}_{t,\mathbf{x}}\hookrightarrow\mathbf{L}^{1}_{t,\mathbf{x}}$, ensures the weak $\mathbf{L}^{1}_{t,\mathbf{x},\boldsymbol{\xi}}$ convergence. Similarly, $(m_{0}(f_{\varepsilon}\,\gamma_{\varepsilon}))_{\varepsilon}$ converges weakly to $m_{0}f$ in $\textnormal{L}^{\infty}_{t}(\textnormal{L}^{5/3}_{\mathbf{x}})-\star$, and Fact 4 ensures for instance that $(\mathbf{u}_{\varepsilon})_{\varepsilon}$ converges strongly in $\textnormal{L}^{1}_{t}(\mathbf{L}^{5/2}_{\mathbf{x}})$.
Fact 6
$\big{[}(1+\rho_{\varepsilon})\,(\mathbf{u}_{\varepsilon}\star\varphi_{%
\varepsilon})\otimes\mathbf{u}_{\varepsilon}\big{]}_{\varepsilon}$ converges weakly in $\mathbf{L}^{1}_{t,\mathbf{x}}$ to $(1+\rho)\,\mathbf{u}\otimes\mathbf{u}$.
Proof of Fact 6: Once again, it’s a “ weak $\times$ strong ” type of convergence here. First notice
$$\displaystyle\frac{2}{11}=\frac{\theta}{6}+\frac{1-\theta}{2},$$
where $\theta:=21/22$. Hence, by Hölder’s inequality, we get the following interpolation result: $\textnormal{L}^{\infty}_{t}(\textnormal{L}^{2}_{\mathbf{x}})\cap\textnormal{L}%
^{2}_{t}(\textnormal{L}^{6}_{\mathbf{x}})\hookrightarrow\textnormal{L}^{2/%
\theta}_{t}(\textnormal{L}^{11/2}_{\mathbf{x}})$. This implies that both $(\mathbf{u}_{\varepsilon})_{\varepsilon}$ and $(\mathbf{u}_{\varepsilon}\star\varphi_{\varepsilon})_{\varepsilon}$ are bounded in $\textnormal{L}^{2/\theta}_{t}(\textnormal{L}^{11/2}_{\mathbf{x}})$, and the product $[(\mathbf{u}_{\varepsilon}\star\varphi_{\varepsilon})\otimes\mathbf{u}_{%
\varepsilon}]_{\varepsilon}$ (which converges almost everywhere to $\mathbf{u}\otimes\mathbf{u}$, see Fact 4) is hence bounded in $\textnormal{L}^{1/\theta}_{t}(\textnormal{L}^{11/4}_{\mathbf{x}})$. Since $\theta<1$ and $11/4>5/2$,
we get the strong convergence of this product to $\mathbf{u}\otimes\mathbf{u}$, in $\textnormal{L}^{1}_{t}(\textnormal{L}^{5/2}_{\mathbf{x}})$, and we already know that $(\rho_{\varepsilon})_{\varepsilon}$ converges weakly to $\rho$ in $\textnormal{L}^{\infty}_{t}(\textnormal{L}^{5/3}_{\mathbf{x}})-\star$.
Using all the weak convergences above, we end up with a local in time solution of our system, which
concludes the Proof of Proposition 3.1.
$\square$.
Next subsection is devoted to the prolongation of the local solutions defined above in solutions
defined on $[0,T]$.
3.4 Energy estimate and global existence
In order to prove the existence of global solutions to our system,
a standard strategy consists in reproducing the previous step at time $t_{\star}-\varepsilon$, obtain another local solution and paste it with the previous one, and so on and so forth.
For such a strategy to succeed, one must ensure that the sequence of local times of existence does not decrease too quickly. But, as noticed in Remark 3.1, the local time of existence is bounded
below by $1/\textnormal{P}_{T}^{\textnormal{in}}$, which is (by definition) a non-increasing function of $\|\sqrt{1+\rho_{\textnormal{in}}}\mathbf{u}_{\textnormal{in}}\|_{2}+\|f_{%
\textnormal{in}}\|_{\infty}+M_{2}f_{\textnormal{in}}$.
We already know (by maximum principle) that $\|f\|_{\infty}\leq e^{2t_{\star}}f_{\textnormal{in}}$, and this bound ensures
that all possible extension will always satisfy $\|f\|_{\infty}\leq e^{2T}f_{\textnormal{in}}$, since the sum of all local times of existence does not exceed $T$ (well if it does, we’re done !). It is hence sufficient to prove that, for almost all $t\in J_{\textnormal{in}}$,
$$\displaystyle M_{2}f(t)+\|\sqrt{1+\rho(t)}\,\mathbf{u}(t)\|_{2}\leq M_{2}f_{%
\textnormal{in}}+\|\sqrt{1+\rho_{\textnormal{in}}}\,\mathbf{u}_{\textnormal{in%
}}\|_{2}.$$
(55)
Indeed, such an estimate would propagate for each local solution and we may hence bound from below
all the corresponding times of existence, which means that our strategy would end in a finite number of steps. In fact (55) is a straightforward consequence of the following energy estimate:
Proposition 3.2.
The solution built in Proposition 3.1 satisfies for almost all $t\in J_{\textnormal{in}}$,
$$\displaystyle\frac{1}{2}\left\{M_{2}f(t)+\|\sqrt{1+\rho(t)}\,\mathbf{u}(t)\|_{%
\textnormal{L}^{2}({\mathbb{T}}^{3})}^{2}\right\}$$
$$\displaystyle+\int_{0}^{t}\|\nabla_{\mathbf{x}}\mathbf{u}(s)\|_{\textnormal{L}%
^{2}({\mathbb{T}}^{3})}^{2}\mathrm{d}s+\frac{3}{2}\int_{0}^{t}\int_{{\mathbb{T%
}}^{3}\times\mathbb{R}^{3}}|\mathbf{u}-\boldsymbol{\xi}|^{2}\,f\,\mathrm{d}%
\boldsymbol{\xi}\,\mathrm{d}\mathbf{x}\,\mathrm{d}s$$
$$\displaystyle\leq\frac{1}{2}\left\{M_{2}f_{\text{\bf in}}+\|\sqrt{1+\rho_{%
\text{\bf in}}}\,\mathbf{u}_{\text{\bf in}}\|_{\textnormal{L}^{2}({\mathbb{T}}%
^{3})}^{2}\right\}.$$
(56)
Proof of Proposition 3.2:
We first use inequality (39) for solutions of the regularized system,
and add the integral in time of (45)$\times\frac{1}{2}$ to get
$$\displaystyle\frac{1}{2}\left\{M_{2}f_{\varepsilon}(t)+\|\sqrt{1+\rho(t)}\,%
\mathbf{u}_{\varepsilon}(t)\|_{\textnormal{L}^{2}({\mathbb{T}}^{3})}^{2}\right\}$$
$$\displaystyle+\int_{0}^{t}\|\nabla_{\mathbf{x}}\mathbf{u}_{\varepsilon}(s)\|_{%
\textnormal{L}^{2}({\mathbb{T}}^{3})}^{2}\mathrm{d}s$$
$$\displaystyle+\frac{3}{2}\int_{0}^{t}\int_{{\mathbb{T}}^{3}\times\mathbb{R}^{3%
}}|\mathbf{u}_{\varepsilon}-\boldsymbol{\xi}|^{2}\,f_{\varepsilon}\,\mathrm{d}%
\boldsymbol{\xi}\,\mathrm{d}\mathbf{x}\,\mathrm{d}s$$
$$\displaystyle\leq\frac{1}{2}\,\left\{M_{2}f_{\text{\bf in}}+\|\sqrt{1+\rho_{%
\text{\bf in}}}\,\mathbf{u}_{\text{\bf in}}\|_{\textnormal{L}^{2}({\mathbb{T}}%
^{3})}^{2}\right\}+R_{\varepsilon}(t),$$
where
$$\displaystyle R_{\varepsilon}(t)=\frac{3}{2}\stackrel{{\scriptstyle R_{%
\varepsilon}^{1}(t)}}{{\overbrace{\int_{0}^{t}\int_{{\mathbb{T}}^{3}\times%
\mathbb{R}^{3}}f_{\varepsilon}\,|\mathbf{u}_{\varepsilon}|^{2}\,(1-\gamma_{%
\varepsilon}(\boldsymbol{\xi}))\,\mathrm{d}\boldsymbol{\xi}\,\mathrm{d}\mathbf%
{x}\,\mathrm{d}s}}}$$
$$\displaystyle+2\stackrel{{\scriptstyle R_{\varepsilon}^{2}(t)}}{{\overbrace{%
\int_{0}^{t}\int_{{\mathbb{T}}^{3}\times\mathbb{R}^{3}}f_{\varepsilon}\,%
\boldsymbol{\xi}\cdot\mathbf{u}_{\varepsilon}\,(\gamma_{\varepsilon}(%
\boldsymbol{\xi})-1)\,\mathrm{d}\boldsymbol{\xi}\,\mathrm{d}\mathbf{x}\,%
\mathrm{d}s}}}$$
$$\displaystyle+\stackrel{{\scriptstyle R_{\varepsilon}^{3}(t)}}{{\overbrace{%
\int_{0}^{t}\int_{{\mathbb{T}}^{3}\times\mathbb{R}^{3}}f_{\varepsilon}\,%
\boldsymbol{\xi}\cdot(\mathbf{u}_{\varepsilon}\star\varphi_{\varepsilon}-%
\mathbf{u}_{\varepsilon})\,\mathrm{d}\boldsymbol{\xi}\,\mathrm{d}\mathbf{x}\,%
\mathrm{d}s}}}.$$
Using Lemma 3.1, the maximum principle, and bound (49), we notice that for $\alpha>0$ small enough, $(m_{\alpha}f_{\varepsilon})_{\varepsilon}$ is bounded in $\textnormal{L}^{\infty}_{t}(\textnormal{L}^{3/2}_{\mathbf{x}})$. Now recall that $\gamma_{\varepsilon}$ is chosen so that $|1-\gamma_{\varepsilon}|(\boldsymbol{\xi})\leq\mathds{1}_{|\boldsymbol{\xi}|%
\gtrsim 1/\varepsilon}$, hence
$$\displaystyle R_{\varepsilon}^{1}(t)\lesssim\varepsilon^{\alpha}\|m_{\alpha}f_%
{\varepsilon}\|_{\textnormal{L}^{\infty}_{t}(\textnormal{L}^{3/2}_{\mathbf{x}}%
)}\|\mathbf{u}_{\varepsilon}\|_{\textnormal{L}^{2}_{t}(\textnormal{L}^{6}_{%
\mathbf{x}})}^{2},$$
which goes to $0$ with $\varepsilon$ since $(\mathbf{u}_{\varepsilon})_{\varepsilon}$ is bounded in $\textnormal{L}^{2}_{t}(\mathbf{H}^{1}_{\mathbf{x}})\hookrightarrow\textnormal{%
L}^{2}_{t}(\mathbf{L}^{6}_{\mathbf{x}})$. This shows that $(R_{\varepsilon}^{1})_{\varepsilon}\displaystyle\operatorname*{\longrightarrow%
}_{\varepsilon\rightarrow 0}0$ uniformly on $J_{\textnormal{in}}$, and a similar Proof applies for $(R_{\varepsilon}^{2})_{\varepsilon}$. Then, we have
$$\displaystyle R_{\varepsilon}^{3}(t)=\int_{0}^{t}\int_{{\mathbb{T}}^{3}\times%
\mathbb{R}^{3}}(f_{\varepsilon}\boldsymbol{\xi}\cdot\mathbf{v}_{\varepsilon})%
\,\mathrm{d}\boldsymbol{\xi}\,\mathrm{d}\mathbf{x}\,\mathrm{d}s,$$
where $(\mathbf{v}_{\varepsilon})_{\varepsilon}$ is bounded in $\textnormal{L}^{2}_{t}(\mathbf{L}^{6}_{\mathbf{x}})$ and converges to $0$ in $\textnormal{L}^{2}_{t,\mathbf{x}}\hookrightarrow\textnormal{L}^{1}_{t,\mathbf{%
x}}$. For values of $\boldsymbol{\xi}$ satisfying $|\boldsymbol{\xi}|\leq\displaystyle(\|\mathbf{v}_{\varepsilon}\|_{1}+%
\varepsilon)^{-1/2}$, we simply use the maximum principle for $(f_{\varepsilon})_{\varepsilon}$ to see that their contribution goes to $0$ with $\varepsilon$ (uniformly in time), and for the large values of $\boldsymbol{\xi}$ satisfying the opposite inequality, we handle them as we did for $R_{\varepsilon}^{1}$ and $R_{\varepsilon}^{2}$, using the $\textnormal{L}^{2}_{t}(\mathbf{L}^{6}_{\mathbf{x}})$ bound.
At this stage, we proved that $(R_{\varepsilon})_{\varepsilon}$ converges to $0$ uniformly in time. Let us treat the other terms of inequality (56). We first use the classical estimates of weak convergence to get
$$\displaystyle\int_{0}^{t}\|\nabla_{\mathbf{x}}\mathbf{u}(s)\|_{\textnormal{L}^%
{2}({\mathbb{T}}^{3})}^{2}\mathrm{d}s$$
$$\displaystyle\leq\operatorname*{\underline{\lim}}_{\varepsilon\rightarrow 0}%
\hskip 2.845276pt\int_{0}^{t}\|\nabla_{\mathbf{x}}\mathbf{u}_{\varepsilon}\|_{%
\textnormal{L}^{2}({\mathbb{T}}^{3})}^{2}\mathrm{d}s,$$
$$\displaystyle M_{2}f(t)$$
$$\displaystyle\leq\operatorname*{\underline{\lim}}_{\varepsilon\rightarrow 0}%
\hskip 2.845276ptM_{2}f_{\varepsilon}(t).$$
(57)
Furthermore, adapting the proof of estimate (54) used in Fact $2$,
we obtain for almost all times $t\in J_{\textnormal{in}}$,
$$\displaystyle\|\sqrt{1+\rho_{\varepsilon}(t)}\,\mathbf{u}_{\varepsilon}(t)\|_{%
\textnormal{L}^{2}({\mathbb{T}}^{3})}^{2}\operatorname*{\rightarrow}_{%
\varepsilon\rightarrow 0}\|\sqrt{1+\rho(t)}\,\mathbf{u}(t)\|_{\textnormal{L}^{%
2}({\mathbb{T}}^{3})}^{2},$$
so that it only remains to prove
$$\displaystyle\int_{0}^{t}\int_{{\mathbb{T}}^{3}\times\mathbb{R}^{3}}|\mathbf{u%
}-\boldsymbol{\xi}|^{2}f\mathrm{d}\boldsymbol{\xi}\mathrm{d}\mathbf{x}\mathrm{%
d}s$$
$$\displaystyle\leq\operatorname*{\underline{\lim}}_{\varepsilon\rightarrow 0}%
\int_{0}^{t}\int_{{\mathbb{T}}^{3}\times\mathbb{R}^{3}}|\mathbf{u}_{%
\varepsilon}-\boldsymbol{\xi}|^{2}f_{\varepsilon}\mathrm{d}\boldsymbol{\xi}%
\mathrm{d}\mathbf{x}\mathrm{d}s,$$
(58)
in order to conclude the Proof of Proposition 3.2. We first write
$$\displaystyle\int_{0}^{t}\int_{{\mathbb{T}}^{3}\times\mathbb{R}^{3}}|\mathbf{u%
}_{\varepsilon}-\boldsymbol{\xi}|^{2}f_{\varepsilon}\mathrm{d}\boldsymbol{\xi}%
\mathrm{d}\mathbf{x}\mathrm{d}s=\int_{0}^{t}\int_{{\mathbb{T}}^{3}\times%
\mathbb{R}^{3}}|\mathbf{u}_{\varepsilon}|^{2}f_{\varepsilon}\mathrm{d}%
\boldsymbol{\xi}\mathrm{d}\mathbf{x}\mathrm{d}s+\int_{0}^{t}M_{2}f_{%
\varepsilon}(s)\mathrm{d}s-2\int_{0}^{t}\int_{{\mathbb{T}}^{3}\times\mathbb{R}%
^{3}}\boldsymbol{\xi}\cdot\mathbf{u}_{\varepsilon}f_{\varepsilon}\mathrm{d}%
\boldsymbol{\xi}\mathrm{d}\mathbf{x}\mathrm{d}s.$$
(59)
Since $(\mathbf{u}_{\varepsilon})_{\varepsilon}$ converges strongly in $\mathbf{L}^{2}_{t,\mathbf{x}}$ and $(f_{\varepsilon})_{\varepsilon}$ converges weakly in $\textnormal{L}^{\infty}_{t,\mathbf{x},\boldsymbol{\xi}}-\star$, we have by Fatou’s Lemma
$$\displaystyle\int_{0}^{t}\int_{{\mathbb{T}}^{3}\times\mathbb{R}^{3}}|\mathbf{u%
}|^{2}f\mathds{1}_{|\boldsymbol{\xi}|\leq n}\mathrm{d}\boldsymbol{\xi}\,%
\mathrm{d}\mathbf{x}\,\mathrm{d}s\leq\operatorname*{\underline{\lim}}_{%
\varepsilon\rightarrow 0}\int_{0}^{t}\int_{{\mathbb{T}}^{3}\times\mathbb{R}^{3%
}}|\mathbf{u}_{\varepsilon}|^{2}f_{\varepsilon}\mathrm{d}\boldsymbol{\xi}\,%
\mathrm{d}\mathbf{x}\,\mathrm{d}s,$$
hence by monotone convergence,
$$\displaystyle\int_{0}^{t}\int_{{\mathbb{T}}^{3}\times\mathbb{R}^{3}}|\mathbf{u%
}|^{2}f\,\mathrm{d}\boldsymbol{\xi}\,\mathrm{d}\mathbf{x}\,\mathrm{d}s\leq%
\operatorname*{\underline{\lim}}_{\varepsilon\rightarrow 0}\int_{0}^{t}\int_{{%
\mathbb{T}}^{3}\times\mathbb{R}^{3}}|\mathbf{u}_{\varepsilon}|^{2}f_{%
\varepsilon}\,\mathrm{d}\boldsymbol{\xi}\,\mathrm{d}\mathbf{x}\,\mathrm{d}s.$$
(60)
Using another time Fatou’s Lemma with estimate (57) shows that
$$\displaystyle\int_{0}^{t}M_{2}f(s)\mathrm{d}s\leq\operatorname*{\underline{%
\lim}}_{\varepsilon\rightarrow 0}\int_{0}^{t}M_{2}f_{\varepsilon}(s)\mathrm{d}s.$$
(61)
On the other hand, since $v_{\varepsilon}:=\mathbf{u}_{\varepsilon}-\mathbf{u}$ is bounded in $\textnormal{L}^{2}_{t}(\mathbf{L}^{6}_{\mathbf{x}})$ and converges to $0$ in $\textnormal{L}^{1}_{t,\mathbf{x}}$, we have as before (see the study of $R_{\varepsilon}^{3}$)
$$\displaystyle\int_{0}^{t}\int_{{\mathbb{T}}^{3}\times\mathbb{R}^{3}}f_{%
\varepsilon}\,\boldsymbol{\xi}\cdot(\mathbf{u}_{\varepsilon}-\mathbf{u})\,%
\mathrm{d}\boldsymbol{\xi}\,\mathrm{d}\mathbf{x}\,\mathrm{d}s\operatorname*{%
\longrightarrow}_{\varepsilon\rightarrow 0}0,$$
and handling large velocities as we did for $R_{\varepsilon}^{1}$ and $R_{\varepsilon}^{2}$, we can show that
$$\displaystyle\int_{{\mathbb{T}}^{3}\times\mathbb{R}^{3}}(f\boldsymbol{\xi}%
\cdot\mathbf{u})\mathds{1}_{|\boldsymbol{\xi}|>n}\mathrm{d}\boldsymbol{\xi}\,%
\mathrm{d}\mathbf{x}\,\mathrm{d}s+\sup_{\varepsilon>0}\int_{0}^{t}\int_{{%
\mathbb{T}}^{3}\times\mathbb{R}^{3}}(f_{\varepsilon}\boldsymbol{\xi}\cdot%
\mathbf{u})\mathds{1}_{|\boldsymbol{\xi}|>n}\mathrm{d}\boldsymbol{\xi}\,%
\mathrm{d}\mathbf{x}\,\mathrm{d}s\operatorname*{\longrightarrow}_{n\rightarrow%
+\infty}0.$$
When $n\in\mathbb{N}$ is fixed, we have by weak $\textnormal{L}^{\infty}_{t,\mathbf{x},\boldsymbol{\xi}}-\star$ convergence of $(f_{\varepsilon})_{\varepsilon}$,
$$\displaystyle\int_{0}^{t}\int_{{\mathbb{T}}^{3}\times\mathbb{R}^{3}}(f_{%
\varepsilon}\,\boldsymbol{\xi}\cdot\mathbf{u})\mathds{1}_{|\boldsymbol{\xi}|%
\leq n}\,\mathrm{d}\boldsymbol{\xi}\,\mathrm{d}\mathbf{x}\,\mathrm{d}s%
\operatorname*{\longrightarrow}_{n\rightarrow+\infty}\int_{0}^{t}\int_{{%
\mathbb{T}}^{3}\times\mathbb{R}^{3}}(f_{\varepsilon}\,\boldsymbol{\xi}\cdot%
\mathbf{u})\,\mathrm{d}\boldsymbol{\xi}\,\mathrm{d}\mathbf{x}\,\mathrm{d}s,$$
so that the three last convergences imply together
$$\displaystyle\int_{0}^{t}\int_{{\mathbb{T}}^{3}\times\mathbb{R}^{3}}(f_{%
\varepsilon}\,\boldsymbol{\xi}\cdot\mathbf{u}_{\varepsilon})\,\mathrm{d}%
\boldsymbol{\xi}\,\mathrm{d}\mathbf{x}\,\mathrm{d}s\operatorname*{%
\longrightarrow}_{n\rightarrow+\infty}\int_{0}^{t}\int_{{\mathbb{T}}^{3}\times%
\mathbb{R}^{3}}(f\,\boldsymbol{\xi}\cdot\mathbf{u})\,\mathrm{d}\boldsymbol{\xi%
}\,\mathrm{d}\mathbf{x}\,\mathrm{d}s.$$
(62)
Using (60), (61) and (62) in (59),
we get (58) and this ends the Proof of Proposition 3.2.
$\square$
This also concludes the Proof of Theorem 1.1, thanks to the strategy of prolongation of local solutions explained at the beginning of the Subsection.
$\square$
4 Appendix
We present in the Appendix a few auxiliary Lemmas which are used in various parts of this work. We start
with a standard variant of Grönwall’s Lemma, that we recall for the sake of completeness.
4.1 A variant of Grönwall’s Lemma
Lemma 4.1.
Let $f\in\mathscr{C}^{1}(\mathbb{R}_{+})$ be a convex non-decreasing function and $\alpha\in\mathbb{R}$. Consider $(z,[0,t_{\star}[)$ the maximal solution of $z^{\prime}=f(z)$, $z(0)=\alpha$ on $\mathbb{R}_{+}$.
Let $a:\mathbb{R}_{+}\rightarrow\mathbb{R}_{+}$ be a continuous function such that
$$\displaystyle a(t)\leq\alpha+\int_{0}^{t}f(a(s))\,\mathrm{d}s.$$
Then, for all $t\in[0,t_{\star}[$,
$$\displaystyle a(t)\leq z(t).$$
Proof of Lemma 4.1:
The usual (linear) Grönwall lemma shows that if $u\in\mathscr{C}^{1}(\mathbb{R}_{+})$ satisfies $\displaystyle u(t)\leq\int_{0}^{t}b(s)u(s)\mathrm{d}s$ on $[0,t_{\star}[$, then $u$ is nonpositive on this interval.
We have here by convexity (on $[0,t_{\star}[$)
$$\displaystyle u(t):=a(t)-z(t)\leq\int_{0}^{t}[f(a(s))-f(z(s))]\,\mathrm{d}s%
\leq\int_{0}^{t}f^{\prime}(a(s))\,u(s)\,\mathrm{d}s,$$
and since $f$ is nondecreasing, $u\leq 0$ on $[0,t_{\star}[$.
$\square$
As a consequence, we get the
Lemma 4.2.
Let $A,\gamma>0$. The maximal solution of the Cauchy problem $z^{\prime}=A\,z^{1+\gamma}$, $z(0)=A$ is defined at least on $[0,(\gamma\,A^{\gamma+1})^{-1}[$.
Proof of Lemma 4.2
Indeed : $\displaystyle t\mapsto(A^{-\gamma}-A\,\gamma\,t)^{-1/\gamma}$ is well defined for $t<(\gamma\,A^{\gamma+1})^{-1}$, and solves the Cauchy problem.
$\square$
In next Subsection, we recall for the sake of completeness the following classical compactness Lemma:
4.2 A compactness lemma
Lemma 4.3.
The injection $\textnormal{H}^{1}_{t,\mathbf{x}}\hookrightarrow\mathscr{C}^{0}_{t}(%
\textnormal{L}^{2}_{\mathbf{x}})$ is compact.
Proof of Lemma 4.3:
Let $(f_{n})_{n}$ be a bounded sequence of $\textnormal{H}^{1}_{t,\mathbf{x}}$.
Step 1: $A=\{t\in[0,T]\,:\,\left\|{f_{n}(t)}\right\|_{\textnormal{H}^{1}_{\mathbf{x}}}%
\rightarrow+\infty\}$ is a null set with respect to the Lebesgue measure $\mu$. Indeed, if it were not, then for any $R>0$, we would have
$$\displaystyle R\,\mu(A)\leq\int_{A}\operatorname*{\underline{\lim}}_{n%
\rightarrow\infty}\left\|{f_{n}(t)}\right\|^{2}_{\textnormal{H}^{1}_{\mathbf{x%
}}}(t)\,\mathrm{d}t.$$
Fatou’s lemma would then imply
$$\displaystyle R\,\mu(A)\leq\operatorname*{\underline{\lim}}_{n\rightarrow%
\infty}\int_{A}\left\|{f_{n}(t)}\right\|^{2}_{\textnormal{H}^{1}_{\mathbf{x}}}%
\leq\operatorname*{\underline{\lim}}_{n\rightarrow\infty}\left\|{f_{n}}\right%
\|^{2}_{\textnormal{H}^{1}_{t,\mathbf{x}}},$$
which is impossible since the right-hand side is finite. By definition of $A$, for all $t$ in $[0,T]\backslash A$, the sequence $(\left\|{f_{n}(t)}\right\|_{\textnormal{H}^{1}_{\mathbf{x}}})_{n}$ has a bounded subsequence. Since $\mu(A)=0$, we can find a countable subset $[0,T]\backslash A$ that is dense in $[0,T]$. Let us denote this subset $\textnormal{B}=(t_{p})_{p\in\mathbb{N}}$. We can then extract a (diagonal) subsequence of $(f_{n})_{n}$ (sill denoted $(f_{n})_{n}$) such that for all $p\in\mathbb{N}$, the sequence $(f_{n}(t_{p}))_{n\in\mathbb{N}}$ is bounded in ${\textnormal{H}^{1}_{\mathbf{x}}}$.
Step 2: Since $(\partial_{t}f_{n})_{n}$ is bounded in $\textnormal{L}^{2}_{t,\mathbf{x}}$, $(f_{n})_{n}$ is uniformly equicontinuous w.r.t. $t$, with values in $\textnormal{L}_{\mathbf{x}}^{2}$.
Step 3: Since by Rellich’s Theorem, the injection $\textnormal{H}^{1}_{\mathbf{x}}\hookrightarrow\textnormal{L}^{2}_{\mathbf{x}}$ is compact, we can extract (again, diagonally) a subsequence (still denoted) $(f_{n})_{n}$ such that for all $p$, the sequence $(f_{n}(t_{p}))_{n}$ converges in $\textnormal{L}^{2}_{\mathbf{x}}$ to some element $f(t_{p})\in\textnormal{L}^{2}_{\mathbf{x}}$.
Step 4: $f$ is uniformly continuous on B, with values in $\textnormal{L}^{2}_{\mathbf{x}}$. Indeed, for any $t$, $s$ in B,
$$\displaystyle\left\|{f(t)-f(s)}\right\|_{\textnormal{L}^{2}_{\mathbf{x}}}\leq%
\left\|{f(t)-f_{n}(t)}\right\|_{\textnormal{L}^{2}_{\mathbf{x}}}+\left\|{f_{n}%
(t)-f_{n}(s)}\right\|_{\textnormal{L}^{2}_{\mathbf{x}}}+\left\|{f_{n}(s)-f(s)}%
\right\|_{\textnormal{L}^{2}_{\mathbf{x}}}.$$
The central term of the right-hand side goes to $0$ with $|t-s|$, independently of $n$ because of the uniform equicontinuity obtained in Step 2. When $t$ and $s$ are fixed elements of B, the two other terms of the right-hand side vanish when $n\rightarrow\infty$ because of the extraction of Step 3.
Step 5:
$f$ admits a unique continuous extension to $[0,T]$, which is uniformly continuous on this interval, with values in $\textnormal{L}^{2}_{\mathbf{x}}$.
Step 6: $f$ is actually the limit of $(f_{n})_{n}$ in $\mathscr{C}^{0}_{t}(\textnormal{L}^{2}_{\mathbf{x}})$. Indeed, for $\varepsilon>0$, choose $\delta$ a corresponding modulus of equicontinuity for $f$ and of uniform equicontinuity for $(f_{n})_{n}$. Then pick (B is dense in the compact set $[0,T]$) a finite number $t_{1},\dots,t_{N}$ of elements of B such
that $[0,T]\subseteq\cup_{i=1}^{N}]t_{i}-\delta,t_{i}+\delta[$. Eventually, for any $\sigma\in[0,T]$, if $\sigma\in]t_{i}-\delta,t_{i}+\delta[$, then
$$\displaystyle\left\|{f(\sigma)-f_{n}(\sigma)}\right\|_{\textnormal{L}^{2}_{%
\mathbf{x}}}$$
$$\displaystyle\leq\left\|{f(\sigma)-f(t_{i})}\right\|_{\textnormal{L}^{2}_{%
\mathbf{x}}}+\left\|{f(t_{i})-f_{n}(t_{i})}\right\|_{\textnormal{L}^{2}_{%
\mathbf{x}}}+\left\|{f_{n}(t_{i})-f_{n}(\sigma)}\right\|_{\textnormal{L}^{2}_{%
\mathbf{x}}}$$
$$\displaystyle\leq 2\varepsilon+\left\|{f(t_{i})-f_{n}(t_{i})}\right\|_{%
\textnormal{L}^{2}_{\mathbf{x}}},$$
so that $\left\|{f(\sigma)-f_{n}(\sigma)}\right\|_{\textnormal{L}^{2}_{\mathbf{x}}}\leq
3\varepsilon$ for $n$ large enough. This ends the proof of Lemma 4.3.
$\square$
4.3 Weak convergence of a product
We present here a result based on the method used in [10].
Proposition 4.1.
Let $q\in[1,\infty]$ and $p\in[1,3[$. Consider two families $(a_{\varepsilon})_{\varepsilon}$ and $(b_{\varepsilon})_{\varepsilon}$ respectively in $\textnormal{L}^{q}_{t}(\textnormal{W}^{1,p}_{\mathbf{x}})$ and $\textnormal{L}^{q^{\prime}}_{t}(\textnormal{L}^{s^{\prime}}_{\mathbf{x}})$, with $s<p^{\star}:=\frac{3p}{3-p}$. Assume the weak convergences
$$\displaystyle(a_{\varepsilon})_{\varepsilon}$$
$$\displaystyle\operatorname*{\rightharpoonup}_{\varepsilon\rightarrow 0}a,\text%
{ in}\quad\textnormal{L}^{q}_{t}(\textnormal{W}^{1,p}_{\mathbf{x}})-w\star,$$
$$\displaystyle(b_{\varepsilon})_{\varepsilon}$$
$$\displaystyle\operatorname*{\rightharpoonup}_{\varepsilon\rightarrow 0}b,\text%
{ in}\quad\textnormal{L}^{q^{\prime}}_{t}(\textnormal{L}^{s^{\prime}}_{\mathbf%
{x}})-w\star.$$
If $(\partial_{t}b_{\varepsilon})_{\varepsilon}$ is bounded in $\textnormal{L}^{q^{\prime}}_{t}(\textnormal{H}^{-m}_{\mathbf{x}})$ for some $m\in\mathbb{Z}$, then, up to a subsequence, we have the weak convergence
$$\displaystyle(a_{\varepsilon}\,b_{\varepsilon})_{\varepsilon}\operatorname*{%
\rightharpoonup}_{\varepsilon\rightarrow 0}a\,b\text{ in the sense of measures},$$
i.e. with test functions in $\mathscr{C}^{0}_{t,\mathbf{x}}$.
Proof of Proposition 4.1 :
Let us first notice that for all $\varepsilon>0$ $a_{\varepsilon}\,b_{\varepsilon}\in\textnormal{L}^{1}_{t,\mathbf{x}}$ (Sobolev injection) and that the sequences $(a_{\varepsilon})_{\varepsilon}$ and $(b_{\varepsilon})_{\varepsilon}$ are bounded in the spaces in which they converge weakly.
Step 1.
We have
$$\displaystyle a\,(b\star\varphi_{\eta})\operatorname*{\longrightarrow}_{\eta%
\rightarrow 0}a\,b,\text{ in }\textnormal{L}^{1}_{t,\mathbf{x}}\text{ strong }.$$
Step 2.
Since $(\partial_{t}b_{\varepsilon})_{\varepsilon}$ is bounded in $\textnormal{L}^{q^{\prime}}_{t}(\textnormal{H}^{-m}_{\mathbf{x}})$, $(b_{\varepsilon}\star\varphi_{\eta})_{\varepsilon}$ is bounded in $\textnormal{W}^{1,q^{\prime}}_{t}(\textnormal{W}^{1,p^{\prime}}_{\mathbf{x}})$ so that (thanks to Rellich’s Theorem), for all fixed $\eta$, $(b_{\varepsilon}\star\varphi_{\eta})_{\varepsilon}$ admits a (strongly) converging subsequence in $\textnormal{L}^{q^{\prime}}_{t}(\textnormal{L}^{p^{\prime}}_{\mathbf{x}})$, the limit being necessarily $b\star\varphi_{\eta}$ (this is due to the
uniqueness of the weak$-\star$ limit). In fact, we can choose (but we don’t write it explicitly) a common (diagonal) extraction for all $\eta$ after discretization ($\eta:=1/k$). Since $(a_{\varepsilon})_{\varepsilon}$ converges weakly in $\textnormal{L}^{q}_{t}(\textnormal{L}^{p}_{\mathbf{x}})$, we eventually get, for all fixed $\eta$,
$$\displaystyle(a_{\varepsilon}\,(b_{\varepsilon}\star\varphi_{\eta}))_{%
\varepsilon}\operatorname*{\rightharpoonup}_{\varepsilon\rightarrow 0}a\,(b%
\star\varphi_{\eta})\text{ in }\textnormal{L}^{1}_{t,\mathbf{x}}\text{ weak.}$$
Step 3.
We shall use the following “commutator Lemma”, the Proof of which is rather close to the usual Friedrichs Lemma (which is a key element of [10]).
Lemma 4.4.
Under the assumptions111The assumption on $(\partial_{t}b_{\varepsilon})_{\varepsilon}$ is obviously useless here. of Proposition 4.1, if $(\varphi_{\eta})_{\eta}$ is a sequence of even
mollifiers, then the commutator (convolution in $\mathbf{x}$ only)
$$\displaystyle S_{\varepsilon,\eta}$$
$$\displaystyle:=a_{\varepsilon}\,(b_{\varepsilon}\star\varphi_{\eta})-(a_{%
\varepsilon}\,b_{\varepsilon})\star\varphi_{\eta}$$
goes to $0$ in $\textnormal{L}^{1}_{t,\mathbf{x}}$ as $\eta\rightarrow 0$, uniformly in $\varepsilon$.
Proof of Lemma 4.4:
First recall the following standard fact : since $(a_{\varepsilon})_{\varepsilon}$ is bounded in $\textnormal{L}^{q}_{t}(\textnormal{W}^{1,p}_{\mathbf{x}})$ and $s<p^{\star}$, the sequence $(\tau_{\mathbf{h}}a_{\varepsilon}-a_{\varepsilon})_{\varepsilon}$ tends to $0$ in $\textnormal{L}^{q}_{t}(\textnormal{L}^{s}_{\mathbf{x}})$ as $\mathbf{h}\rightarrow 0$, uniformly in $\varepsilon$. We now write the following equality for the commutator
$$\displaystyle S_{\varepsilon,\eta}(t,\mathbf{x})=\int_{\textnormal{B}_{\eta}}%
\Big{[}a_{\varepsilon}(t,\mathbf{x})-a_{\varepsilon}(t,\mathbf{x}-\mathbf{y})%
\Big{]}\,b_{\varepsilon}(t,\mathbf{x}-\mathbf{y})\,\varphi_{\eta}(\mathbf{y})%
\,\mathrm{d}\mathbf{y},$$
whence thanks to Fubini’s Theorem,
$$\displaystyle\|S_{\varepsilon,\eta}\|_{1}\leq\|b_{\varepsilon}\|_{\textnormal{%
L}^{q^{\prime}}_{t}(\textnormal{L}^{s^{\prime}}_{\mathbf{x}})}\int_{%
\textnormal{B}_{\eta}}|\varphi_{\eta}(\mathbf{y})|\,\|\tau_{\mathbf{y}}a_{%
\varepsilon}-a_{\varepsilon}\|_{\textnormal{L}^{q}_{t}(\textnormal{L}^{s}_{%
\mathbf{x}})}\,\mathrm{d}\mathbf{y},$$
which yields the desired uniform convergence, and concludes the Proof
of Lemma 4.4. $\square$
Step 4.
We have
$$\displaystyle(a_{\varepsilon}b_{\varepsilon})\star\varphi_{\eta}-a_{%
\varepsilon}b_{\varepsilon}\operatorname*{\longrightarrow}_{\eta%
\rightharpoonup 0}0\text{ in the sense of measures,}$$
uniformly in $\varepsilon$ (with a fixed continuous test function). Indeed, if $\theta\in\mathscr{C}_{t,\mathbf{x}}^{0}$, since $\varphi_{\eta}$ is even, we know that
$$\displaystyle\langle(a_{\varepsilon}\,b_{\varepsilon})\star\varphi_{\eta}-a_{%
\varepsilon}\,b_{\varepsilon},\theta\rangle=\langle a_{\varepsilon}\,b_{%
\varepsilon},\theta\star\varphi_{\eta}-\theta\rangle,$$
and the right-hand side tends to $0$ with the desired uniformity because $(a_{\varepsilon}\,b_{\varepsilon})_{\varepsilon}$ is bounded in $\textnormal{L}^{1}_{t,\mathbf{x}}$, and $(\theta\star\varphi_{\eta}-\theta)_{\eta}$ goes to $0$ in $\textnormal{L}^{\infty}_{t,\mathbf{x}}$ ($\theta$ is uniformly continuous).
Step 5.
Write
$$\displaystyle a\,b-a_{\varepsilon}\,b_{\varepsilon}$$
$$\displaystyle=a\,b-a\,(b\star\varphi_{\eta})$$
$$\displaystyle+a\,(b\star\varphi_{\eta})-a_{\varepsilon}\,(b_{\varepsilon}\star%
\varphi_{\eta})$$
$$\displaystyle+a_{\varepsilon}\,(b_{\varepsilon}\star\varphi_{\eta})-(a_{%
\varepsilon}\,b_{\varepsilon})\star\varphi_{\eta}$$
$$\displaystyle+(a_{\varepsilon}\,b_{\varepsilon})\star\varphi_{\eta}-a_{%
\varepsilon}\,b_{\varepsilon}.$$
Fix $\theta\in\mathscr{C}^{0}_{t,\mathbf{x}}$. In the right-hand side, line number $i\in\{1,2,3,4\}$ corresponds to the Step $i$ proven previously. We choose first $\eta$ to handle (uniformly in $\varepsilon$) all the lines of the right-hand side, except the second one. Then, we
choose the appropriate $\varepsilon$ to handle the second line, thanks to Step 2.
This concludes the Proof of Proposition 4.1.
$\square$
Proposition 4.2.
In Proposition 4.1, when
$q<\infty$, the same conclusion holds assuming
only a bound in $\textnormal{L}^{1}_{t}(\textnormal{H}^{-m}_{\mathbf{x}})$ (instead of $\textnormal{L}^{q^{\prime}}_{t}(\textnormal{H}^{-m}_{\mathbf{x}})$) for $(\partial_{t}b_{\varepsilon})_{\varepsilon}$ if, in addition, we assume the convergence of $(a_{\varepsilon})_{\varepsilon}$ to $a$ in $\textnormal{L}^{r}_{t}(\textnormal{L}^{p}_{\mathbf{x}})$ weak$-\star$,
for some $r>q$.
Proof of Proposition 4.2:
The Proof is identical to the Proof of Proposition 4.1, except for the second step (the only one using the bound on $(\partial_{t}b_{\varepsilon})_{\varepsilon}$). For this step, we use a (diagonal) extraction such that,
for all fixed $\eta$, $(b_{\varepsilon}\star\varphi_{\eta})_{\varepsilon}$ converges strongly (and almost everywhere) to $b\star\varphi_{\eta}$ (but only) in $\textnormal{L}^{1}_{t}(\textnormal{L}^{p^{\prime}}_{\mathbf{x}})$. Since $(b_{\varepsilon})_{\varepsilon}$ (and hence $(b_{\varepsilon}\star\varphi_{\eta})_{\varepsilon}$) is bounded in $\textnormal{L}^{q^{\prime}}_{t}(\textnormal{L}^{p^{\prime}}_{\mathbf{x}})$, with $q^{\prime}>r^{\prime}\geq 1$, we see that $(b_{\varepsilon}\star\varphi_{\eta})_{\varepsilon}$ converges to $b\star\varphi_{\eta}$ strongly in $\textnormal{L}^{r^{\prime}}_{t}(\textnormal{L}^{p^{\prime}}_{\mathbf{x}})$, and the added assumption of weak-$\star$ convergence for $(a_{\varepsilon})_{\varepsilon}$ allows hence to get
$$\displaystyle(a_{\varepsilon}\,(b_{\varepsilon}\star\varphi_{\eta}))_{%
\varepsilon}\operatorname*{\rightharpoonup}_{\varepsilon\rightarrow 0}a\,(b%
\star\varphi_{\eta})\text{ in $\textnormal{L}^{1}_{t,\mathbf{x}}$ weak,}$$
which ends the Proof of Proposition 4.2.
$\square$
References
[1]
A.A. Amsden, P.J. O’Rourke, and T.D. Butler.
Kiva-II: A computer program for chemically reactive flows with
sprays.
Technical report, Los Alamos National University, 1989.
[2]
O. Anoshchenko and A. Boutet de Monvel-Berthier.
The existence of the global generalized solution of the system of
equations describing suspension motion.
Math. Meth. Appl. Sci., 20(6):495–519,
1997.
[3]
S.N. Antontsev, A.V. Kazhiktov, and V.N. Monakhov.
Boundary value problems in mechanics of nonhomogeneous fluids
.
Studies in Mathematics and Its Applications, 22, 1989, North Holland.
[4]
C. Baranger and L. Desvillettes.
Coupling Euler and Vlasov equations in the context of sprays: the
local-in-time, classical solutions.
J. Hyperbolic Differ. Equ., 3(1):1–26, 2006.
[5]
S. Benjelloun.
Quelques problèmes d’écoulements multi-fluide : analyse mathématique, modélisation numérique et simulation.
PhD Thesis, École Normale Supérieure de Cachan, 2012.
[6]
L. Boudin, L. Desvillettes, C. Grandmont, and A. Moussa.
Global existence of solutions for the coupled Vlasov and
Navier-Stokes equations.
Differential Integral Equations, 22(11-12):1247–1271,
2009.
[7]
F. Boyer and P. Fabrie.
Mathematical tools for the study of the incompressible
Navier-Stokes equations and related models, volume 183 of Applied Mathematical Sciences.
Springer-Verlag, New-York, 2013.
[8]
J.A. Carrillo, R. Duan and A. Moussa.
Global classical solutions close to equilibrium to the
Vlasov-Fokker-Planck-Euler system.
Kinet. Relat. Models, 1, 227-258, 2011.
[9]
L. Desvillettes.
Some Aspects of the Modeling at Different Scales of Multiphase Flows.
Comput. Meth. Appl. Mech. Eng., 199, 1265-1267, 2010 (special issue).
[10]
R. J. DiPerna and P. L. Lions.
Ordinary differential equations, transport theory and Sobolev spaces.
Inventiones Mathematicae, 98:511–547, 1989.
[11]
D. Gilbarg and N. S. Trudinger.
Elliptic Partial Differential Equations of Second Order.
Springer, 2001.
[12]
K. Hamdache.
Global existence and large time behaviour of solutions for the
Vlasov-Stokes equations.
Japan J. Industr. Appl. Math., 15:51–74,
1998.
[13]
P.L. Lions.
Mathematical Topics in Fluid Mechanics: Incompressible models.
Oxford Lecture Series in Mathematics and Its Applications. Clarendon
Press, 1996.
[14]
J. Mathiaud.
Local smooth solutions of a thin spray model with collisions.
Math. Models Methods Appl. Sci., 20(2):191–221, 2010.
[15]
A. Mellet and A. Vasseur.
Global weak solutions for a Vlasov-Fokker-Planck/Navier-Stokes system
of equations.
Math. Models Methods Appl. Sci.,
17(7):1039–1063, 2007.
[16]
A. Mellet and A. Vasseur.
Asymptotic analysis for a Vlasov-Fokker-Planck/ compressible
Navier-Stokes system of equations.
Commun. Math. Phys., 281(3):573–596, 2008.
[17]
A. Moussa.
Etude mathématique et numérique du transport des aéorosols dans
le poumon humain.
PhD Thesis, École Normale Supérieure de Cachan, 2009.
[18]
P. J. O’Rourke.
Collective drop effect on Vaporizing Liquid Sprays.
PhD Thesis, Los Alamos National University, 1981.
[19]
E. M. Stein and G. Weiss.
Introduction to Fourier Analysis on Euclidian Spaces.
Princeton University Press. 1971.
[20]
C. Yu.
Global weak solutions to the Navier-Stokes-Vlasov equations.
To appear in J. Math. Pures Appl.
[21]
F. A. Williams.
Combustion theory.
Benjamin Cummings. 1985. |
Revisiting the Issues On Netflow Sample and Export Performance
Hamed Haddadi, Raul Landa, Miguel Rio
Department of Electronic and Electrical Engineering
University College London
United Kingdom
Email: hamed,mrio,[email protected]
Saleem Bhatti
School of Computer Science
University of St. Andrews
United Kingdom
Email: [email protected]
Abstract
The high volume of packets and packet rates of traffic on some router links makes it exceedingly difficult for routers to examine every packet in order to keep detailed statistics about the traffic which is traversing the router. Sampling is commonly applied on routers in order to limit the load incurred by the collection of information that the router has to undertake when evaluating flow information for monitoring purposes. The sampling process in nearly all cases is a deterministic process of choosing 1 in every $N$ packets on a per-interface basis, and then forming the flow statistics based on the collected sampled statistics. Even though this sampling may not be significant for some statistics, such as packet rate, others can be severely distorted. However, it is important to consider the sampling techniques and their relative accuracy when applied to different traffic patterns.
The main disadvantage of sampling is the loss of accuracy in the collected trace when compared to the original traffic stream. To date there has not been a detailed analysis of the impact of sampling at a router in various traffic profiles and flow criteria. In this paper, we assess the performance of the sampling process as used in NetFlow in detail, and we discuss some techniques for the compensation of loss of monitoring detail.
I Introduction
Packet sampling is an integral part of passive network measurement on today’s Internet. The high traffic volumes on backbone networks and the pressure on routers has resulted in the need to control the consumption of resources in the measurement infrastructure. This has resulted in the definition and use of estimated statistics by routers, generated based on sampling packets in each direction of each port on the routers. The aims of this paper is to analyse the effects of the sampling process as operated by NetFlow, the dominant standard on today’s routers.
There are three constraints on a core router which lead to the use packet sampling: the size of the record buffer, the CPU speed and the record look-up time. In [6], it is noted that in order to manage and analyse the performance of a network, it is enough to look at the basic statistical measures and summary statistics such as average range, variance, and standard deviation. However, in this paper we analyse both analytically and practically the accuracy of the inference of original characteristics from the sampled stream when higher order statistics are used.
This paper focuses on the inference of original network traffic characteristics for flows from a sampled set of packets and examines how the sampling process can affect the quality of the results. In this context, a flow is identified specifically, as the tuple of the following five key fields: Source IP address, Destination IP address, Source port number, Destination port number, Layer 4 protocol type.
I-A NetFlow memory constraints
A router at the core of an internet link is carrying a large number of flows at any given time. this pressure on the router entails the use of strict rules in order to export the statistics and keep the router memory buffer and CPU resources available to deal with changes in traffic patterns by avoiding the handling of large tables of flow records. Rules for expiring NetFlow cache entries include:
•
Flows which have been idle for a specified time are expired and removed from the cache (15 seconds is default)
•
Long lived flows are expired and removed from the cache (30 minutes is default)
•
As the cache becomes full a number of heuristics are applied to aggressively age groups of flows simultaneously
•
TCP connections which have reached the end of byte stream (FIN) or which have been reset (RST) will be expired
I-B Sampling basics
Distributions studies have been done extensively in literature. In brief conclusion, internet traffic is believed to have Heavy-tailed distribution, self-similar nature, Long Range Dependence [2]. Sampling has the following effects on the flows:
•
It is easy to miss short flows [14]
•
Mis-ranking on high flows [4]
•
Sparse flow creation[14]
Packet sampling:
The inversion methods are of little to no use in practice for low sampling probability $q$, such as $q=0.01$ (1 packet in 100) or smaller, and become much worse as $q$ becomes smaller still. For example, on the Abilene network, 50% sampling was needed to detect the top flow correctly [4].
Flow sampling:
Preserves flows intact and the sampling is done on the flow records. In practice, any attempt to gather flow statistics involves classifying individual packets into flows. All packet meta-data has to be organised into flows before sampling can take place. This involves more CPU load and more memory if one uses the traditional hash table approach with one entry per flow. New flow classification techniques, such as bitmap algorithms, could be applied but there is no practical usage in this manner currently.
II Variation of Higher Order statistics
In this section we look at a more detailed analysis of the effect of sampling as performed by netflow on higher order statistics of the packet and flow size distributions. For the analysis of packet sampling application is used by NetFlow, we emulated the NetFlow operation on a 1 hour OC-48 trace, collected from the CAIDA link on $24^{th}$ of April 2003. This data set is available from the public repository at CAIDA [7]. The trace comprises of 84579462 packets with anonymised source and destination IP addresses. An important factor to rememberer in this work is the fact that the memory constraint on the router has been relaxed in generating the flows from the sampled stream. This means that there maybe more than tens of thousands of flow keys present at the memory at a given time, while in NetFlow, the export mechanism empties the buffer list regularly which can have a more severe impact on the resultant distribution of flow rates and statistics333The processing of the data was done using tools which are made available to the public by the authors..
II-A Effects of the short time-out imposed by memory constraints
Table I illustrates the data rates $d(t)$ per interval of measurement. Inverted data rates, by dividing $d(t)$ by the sampling probability $q$, are shown as $dn(t)$.
As observed in table I, the mean does not have a great variation, possibly because distributions of packet sizes within single flows do not exhibit high variability. The standard deviation of the estimated data rate is higher than the corresponding standard deviation for the unsampled data stream. In the absence of any additional knowledge about the higher level protocol, or the nature of the session level activity, in the unsampled data stream, each flow can be thought of as having packets of varying sizes that are more or less independent from one another. Thus, the whole traffic profile results from the addition of many independent random variables which, by the central limit theorem, tend to balance among themselves to produce a more predictable, homogeneous traffic aggregate. However, simple inversion eliminates this multiplicity of randomly distributed values by introducing a very strong correlation effect, whereby the size of all the packets in a reconstructed flow depend on the size of a very small set of sampled packets. This eliminates the possibility for balancing and thus increases the variability of the resulting stream, i.e. its standard deviation.
However, the skewness and kurtosis do change. Skewness is a measure of the asymmetry of the probability distribution of a real-valued random variable. Roughly speaking, a distribution has positive skew (right-skewed) if the right (higher value) tail is longer and negative skew (left-skewed) if the left (lower value) tail is longer (confusing the two is a common error). Skewness, the third standardised moment, is written as $\gamma_{1}$ and defined as:
$\gamma_{1}=\frac{\mu_{3}}{\sigma^{3}}$
where $\mu_{3}$ is the third moment about the mean and $\sigma$ is the standard deviation.
Kurtosis is more commonly defined as the fourth cumulant divided by the square of the variance of the probability distribution,
$\gamma_{2}=\frac{\kappa_{4}}{\kappa_{2}^{2}}=\frac{\mu_{4}}{\sigma^{4}}-3$
which is known as excess kurtosis. The ”minus 3” at the end of this formula is often explained as a correction to make the kurtosis of the normal distribution equal to zero. The skewness is a sort of measure of the asymmetry of the distribution function. The kurtosis measures the flatness of the distribution function compared to what would be expected from a Gaussian distribution.
Table II illustrates the packet rates $p(t)$ per interval of measurement. Inverted packet rates, by dividing $p(t)$ by the sampling probability $q$, are shown as $pn(t)$. The distributions before and after sampling are extremely close, and thus their difference tends to exaggerate those small difference that they do have. That is the reason of the enormous skewness and kurtosis that are observed. The skewness of the reconstructed stream is smaller than that of the unsampled stream this means that the reconstructed distribution is more symmetric, that is , it tends to diverge in a more homogeneous manner around the mean. Additionally, it is positive, meaning that in both cases the distribution tends to have longer tails towards large packets rather than towards short packets, concentrating its bulk on the smaller packets. If we concede that small flows (flows consisting of a small number of packets) tend to contain small packets, then it is clear that this smaller packets will be underrepresented and the distribution will shift its weight towards bigger packets (members of bigger flows). Thus, it will become more symmetric and hence less skewed.
The Kurtosis decreases in all of the considered examples. This means that the reconstructed streams are more homogeneous and less prone to outliers when compared with the original traces. Thus, more of the variance in the original traces in packet size can be attributed to infrequent packets that have inordinately big packets that were missed in the sampling process, and thus the variance in the reconstructed stream consists more of homogeneous differences and not large outliers. However, both the reconstructed and unsampled streams are leptokurtic and thus tend to have long, heavy tails.
II-B The two-sample KS test
The two-sample KS test is one of the most useful and general non-parametric methods for comparing two samples, as it is sensitive to differences in both location and shape of the empirical cumulative distribution functions of the two samples. A CDF was calculated for the number of packets per flow and the number of octets per flow for each of the 120 sampling intervals of 30 seconds each, both for the sampled/inverted and unsampled streams. Then, a Two-Sample Kolmogorov-Smirnov Test with $5\%$ significance level was performed between the 120 unsampled and the 120 sampled & inverted distributions. In every case the distributions before and after sampling and inversion were found to be significantly different, and thus it is very clear that the sampling and inversion
process significantly distorts the actual flow behaviour of the network.
III Practical Implications of Sampling
The effects of sampling on network traffic statistics can be measured from different perspectives. In this section we will cover the theories behind the sampling strategy and use some real data captures from CAIDA in an emulation approach to demonstrate the performance constraints of systematic sampling.
III-A Inversion errors on sampled statistics
The great advantage of sampling is the fact that the first order statistics do not show much variation when the sampling is done at consistent intervals and from a large pool of data. This enables the network monitoring to use the sampled statistics to form a relatively good measure of the aggregate measure of network performance. Figure 1 displays the data rates $d(t)$, in number of bytes seen per 30 second interval, on the one hour trace. The inverted data $d(t)$ is also shown with diamond notation, showing the statistics gathered after the sampled data is multiplied by the sampling rate. The black dots display the relative error per interval, $e(t)=\frac{d(t)-dn(t)}{d(t)}$.
Figure 2 displays the packet rates $p(t)$, the number of packets per 30 second interval, versus the sampled and inverted packet rates $pn(t)$. In this figure, it can be observed that the inversion does a very good job at nearly all times and the relative error is negligible. This is a characteristics of systematic sampling and is due to the central limit theorem.
It can be readily seen that the recovery of packet rates by simple inversion is much better than the recovery of data rates. This is because sampling one in a thousand packets deterministically can be trivially inverted by multiplying by the sampling rate (1000): we focus on packet level measurement, as opposed to a flow level measurement. If the whole traffic flow is collapsed into a single link, then if we sample one packet out every thousand and then multiply that by the sampling rate, we will get the total number of packets in that time window. We believe that the small differences that we can see in Figure 2 are due to the fact that at the end of the window some packets are lost (because their ‘representative’ was not sampled) or overcounted (a ‘representative’ for 1000 packets was sampled but the time interval finished before they had passed). We believe these errors happen between measurement windows in time, i.e. they are window-edge effects.
The inversion property described above does not hold for measuring the number of bytes in a sampling interval. Simple inversion essentially assumes that all packets in a given flow are the same size, and of course this assumption is incorrect. It is to be expected that the greater the standard deviation of packet size over an individual flow, the more inaccurate the recovery by simple inversion will be regarding the number of bytes per measurement interval. Figures 3 and 4 displays the standard error rate on data rate and packet rate recovery respectively, in different measurement intervals.
III-B Flow size and packet size distributions
Figure 5, displays the CDF of packet size distribution in all the flows formed from the sampled and unsampled streams. The little variation in the packet size distribution conforms to the findings of the previous section where it was discussed that the packet sampling has low impact on the packet size distribution.
Figure 6:1 shows the effect that the distribution of packet lengths can have on the distribution of flow lengths when periodic packet sampling is applied. As flows reconstructed from a sampled packet stream are predominantly formed by just one packet, their length distribution follows that of single packets (Figure 5). That is the reason for the sharp jump near 1500 octets, as this characteristic originates from the maximum frame size in ethernet networks.
From Figure 6:2 , it can be readily seen that, in the sampled stream, more than 90 percent of flows consist of a single packet, whereas in the unsampled case a much grater diversity in flow lengths exists for small flows. This is due to the fact that simple packet-based deterministic sampling under-represents short flows, and those short flows that are indeed detected by the procedure after sampling usually consist of a single packet. Thus, short flows are either lost or recovered as single packet flows, and long flows have their lengths reduced.
IV Conclusion
In this paper we have reviewed the effects of sampling and flow record creation, as done by NetFlow, on the traffic statistics which are reported by such a process. It is inevitable that systematic sampling can no longer provide a realistic picture of the traffic profile present on internet links. The emergence of applications such as video on-demand, file sharing, streaming applications and even on-line data processing packages prevents the routers from reporting an optimal measure of the traffic traversing them. In the inversion process, it is a mistake to assume that the inversion of statistics by multiplication by the sampling rate is an indicate of even the first order statistics such as packet rates.
An extension to this work and the inversion problem entails the use of more detailed statistics such as port numbers and TCP flags in order to be able to infer the original characteristics from the probability distribution functions of such variables. This will enable a more detailed recovery of original packet and data rates for different applications. The inference of such probabilities, plus use of methods such as Bayesian inference, would enable a forecasting method which would enable the inversion of the sampled stream in near real time.
In a related work, we will be looking at alternative flow synthesis schemes, looking at techniques replacing the NetFlow, such as use of hashing techniques using Bloom filters. The use of a light weight flow indexing system will allow for a larger number of flows to be present at the router, possibly increasing the memory constraints and allowing for a higher sampling rate, which will in turn lead to more accurate inversion.
V Related Work
There has been a great deal of worked done on analysis of sampling process and inversion problem. Choi et al. have explored the sampling error and measurement overhead of NetFlow in [11] though they have not looked at inversion process.
In [3], the authors have compared the Netflow reports with those obtained from SNMP statistics and packet level traces, but without using the sampling feature of NetFlow which is perhaps the dominant version in use nowadays. Estan et al. [5] have proposed a novel method of adapting the sampling rate at a NetFlow router in order to keep the memory resources at a constant level. This is done by upgrading the router firmware, which can be compromised by an attacker injecting varying traffic volume in order to take down the router. Also this work has not considered the flow length statistics which are the primary focus of our work.
Hohn et al. [10] have proposed a flow sampling model which can be used in an offline analysis of flow records formed from an unsampled packet stream. In this model the statistics of the original stream are recovered to a great extent. However the intensive computing and memory resources needed in this process prevents the implementation of such a scheme on highspeed routers. They prove it impossible to accurately recover statistics from a packet sampled stream, but based on the assumption of packets being independent and identically distributed
Roughan at [12] has looked at statistical processes of active measurement using Poisson and uniform sampling and has compared the theoretical performance of the two methods. Papagiannaki et al. at [8] have discussed the effect of sampling on tiny flows when looking at generation of traffic matrices.
Authors at [15] have been looking at anomaly detection using flow statistics, but without sampling. In [17] and [16], authors have looked at inferring the numbers and lengths of flows of original traffic that evaded sampling altogether. They have looked at inversion via multiplication.
Acknowledgements
The authors would like to acknowledge CAIDA[7] for providing the trace files. This work is conducted under the MASTS (EPSRC grant GR/T10503) and the 46PaQ project (EPSRC grant GR/S93707).
References
[1]
NetFlow Services Solutions Guide available at
http://www.cisco.com/univercd/cc/td/doc/cisintwk/intsolns/netflsol/nfwhite.htm
[2]
Will Leland, Murad Taqqu, Walter Willinger, and Daniel Wilson, On the Self-Similar Nature of Ethernet Traffic (Extended Version), IEEE/ACM Transactions on Networking, Vol. 2, No. 1, pp. 1-15, February 1994.
[3]
Sommer, R. and Feldmann, A. 2002. NetFlow: information loss or win?. In Proceedings of the 2nd ACM SIGCOMM Workshop on internet Measurment (Marseille, France, November 06 - 08, 2002). IMW ’02. ACM Press, New York, NY, 173-174. DOI= http://doi.acm.org/10.1145/637201.637226
[4]
Barakat, C., Iannaccone, G., and Diot, C. 2005. Ranking flows from sampled traffic. In Proceedings of the 2005 ACM Conference on Emerging Network Experiment and Technology (Toulouse, France, October 24 - 27, 2005). CoNEXT’05. ACM Press, New York, NY, 188-199.
[5]
Estan, C., Keys, K., Moore, D., and Varghese, G. 2004. Building a better NetFlow. SIGCOMM Comput. Commun. Rev. 34, 4 (Aug. 2004), 245-256. http://doi.acm.org/10.1145/1030194.1015495
[6]
Performance and Fault Management (Cisco Press Core Series) by Paul L Della Maggiora (Author), James M. Thompson (Author), Robert L. Pavone Jr. (Author), Kent J. Phelps (Author), Christopher E. Elliott (Editor), Publisher: Cisco press; 1ST edition, ISBN: 1578701805
[7]
CAIDA, the Cooperative Association for Internet Data Analysis: http://www.caida.org
[8]
K. Papagiannaki, N. Taft, and A. Lakhina. A Distributed Approach to Measure IP Traffic Matrices. In ACM Internet Measurement Conference, Taormina, Italy, October, 2004.
[9]
IETF PSAMP working Group: http://www.ietf.org/html.charters/psamp-charter.html
[10]
N. Hohn and D. Veitch. Inverting Sampled Traffic. In Internet Measurement Conference, 2003. http://citeseer.ist.psu.edu/hohn03inverting.html
[11]
Choi, B. and Bhattacharyya, S. 2005. Observations on Cisco sampled NetFlow. SIGMETRICS Perform. Eval. Rev. 33, 3 (Dec. 2005), 18-23. DOI= http://doi.acm.org/10.1145/1111572.1111579
[12]
A Comparison of Poisson and Uniform Sampling for Active Measurements, Matthew Roughan, accepted to appear in IEEE JSAC.
[13]
InMon Corporation (2004). sFlow accuracy and billing. Available at www.inmon.com/pdf/sFlowBilling.pdf
[14]
Sampling for Passive Internet Measurement: A Review, N.G. Duffield, Statistical Science,Vol. 19, No. 3, 472-498, 2004.
[15]
Brauckhoff, D., Tellenbach, B., Wagner, A., May, M., and Lakhina, A. 2006. Impact of packet sampling on anomaly detection metrics. In Proceedings of the 6th ACM SIGCOMM on internet Measurement (Rio de Janeriro, Brazil, October 25 - 27, 2006). IMC ’06. ACM Press, New York, NY, 159-164. DOI= http://doi.acm.org/10.1145/1177080.1177101
[16]
N. Duffield, C. Lund, and M. Thorup. Properties and Prediction of Flow Statistics from Sampled Packet Streams. In Proc. ACM SIGCOMM IMW’02, Marseille, France, Nov. 2002.
[17]
N. Duffield, C. Lund, and M. Thorup. Estimating Flow Distributions from Sampled Flow Statistics. In Proc. ACM SIGCOMM’03, Karlsruhe,
Germany, Aug. 2003. |
Suppression of excited-state effects in lattice determination of
nucleon electromagnetic form factors
${}^{a}$, S. Capitani${}^{a}$, D. Djukanovic${}^{b}$,
J. Hua${}^{a}$, B. Jäger${}^{c}$, P. Junnarkar${}^{a,b}$, H.B. Meyer${}^{a,b}$,
T.D. Rae${}^{a}$, H. Wittig${}^{a,b}$
${}^{a}$ PRISMA Cluster of Excellence and Institut für Kernphysik,
University of Mainz, 55099 Mainz, Germany
${}^{b}$ Helmholtz Institut Mainz,
University of Mainz, 55099 Mainz, Germany
${}^{c}$ Department of Physics, College of Science,
Swansea University, Swansea, SA2 8PP, UK
E-mail:
Abstract:
We study the ability of a variety of fitting techniques
to extract the ground state matrix elements of the vector current
from ratios of nucleon three- and two-point functions that contain
contaminations from excited states. Extending our high-statistics
study of nucleon form factors, we are able to demonstrate that the
treatment of excited-state contributions in conjunction with
approaching the physical pion mass has a significant impact on the
$Q^{2}$-dependence of the form factors.
1 Introduction
The persistent failure of the vast majority of lattice measurements of nucleon form factors to replicate the experimental value for the proton charge radius is one of the biggest current puzzles in lattice QCD. Possible explanations include the use of unphysically heavy pions, finite-volume effects and excited-state contributions (or indeed a combination of some or all of these). Our purpose here is to study how best to suppress excited-state effects on the isovector electromagnetic form factors of the nucleon in order to better understand their impact on a possible resolution of the puzzle.
The $eN$ scattering cross section is usually parameterized in terms of the Sachs form factors $G_{E}$, $G_{M}$, whereas the matrix element of the vector current between nucleon states has a natural decomposition in terms of the Dirac and Pauli form factors
$F_{1}$ and $F_{2}$,
$$\langle N(p^{\prime},s^{\prime})|V_{\mu}|N(p,s)\rangle=\overline{u}(p^{\prime}%
,s^{\prime})\left[\gamma_{\mu}F_{1}+i\frac{\sigma_{\mu\nu}q_{\nu}}{2m_{N}}F_{2%
}\right]u(p,s)\,.$$
(1)
The relationship between the two sets of form factors is given by
$$G_{E}(Q^{2})=F_{1}(Q^{2})-\frac{Q^{2}}{4m_{N}^{2}}F_{2}(Q^{2}),~{}~{}~{}G_{M}(%
Q^{2})=F_{1}(Q^{2})+F_{2}(Q^{2}).$$
(2)
In order to measure the nucleon form factors on the lattice, we compute the two- and three-point functions of nucleon interpolating operators $N^{\alpha}$
$$C_{2}(\mathbf{q},t_{s})=\sum_{\mathbf{x}}\textrm{e}^{i\mathbf{q}\cdot\mathbf{x%
}}\Gamma_{\alpha\beta}\langle\overline{N}^{\beta}(t_{s},\mathbf{x})N^{\alpha}(%
0)\rangle,~{}~{}~{}~{}~{}~{}~{}C_{3,\mathcal{O}}(\mathbf{q},t,t_{s})=\sum_{%
\mathbf{x},\mathbf{y}}\textrm{e}^{i\mathbf{q}\cdot\mathbf{y}}\Gamma_{\alpha%
\beta}\langle\overline{N}^{\beta}(t_{s},\mathbf{x})\mathcal{O}(t,\mathbf{y})N^%
{\alpha}(0)\rangle,$$
(3)
where $\Gamma=\frac{1}{2}(1+\gamma_{0})(1+i\gamma_{5}\gamma_{3})$ is a polarization matrix, from whose ratios
$$R_{V_{\mu}}(\mathbf{q},t,t_{s})=\frac{C_{3,V_{\mu}}(\mathbf{q},t,t_{s})}{C_{2}%
(\mathbf{0},t_{s})}\sqrt{\frac{C_{2}(\mathbf{q},t_{s}-t)C_{2}(\mathbf{0},t)C_{%
2}(\mathbf{0},t_{s})}{C_{2}(\mathbf{0},t_{s}-t)C_{2}(\mathbf{q},t)C_{2}(%
\mathbf{q},t_{s})}}$$
(4)
we can extract the Sachs form factors as
[1]
$$\displaystyle{\rm Re}\left[R_{V_{0}}(\mathbf{q},t,t_{s})\right]$$
$$\displaystyle=\sqrt{\frac{m_{N}+E_{\mathbf{q}}}{2E_{\mathbf{q}}}}G^{\rm eff}_{%
E}(Q^{2},t,t_{s})$$
(5)
$$\displaystyle{\rm Re}\left[R_{V_{i}}(\mathbf{q},t,t_{s})\right]_{i=1,2}$$
$$\displaystyle=\epsilon_{ij}q_{j}\frac{1}{\sqrt{2E_{\mathbf{q}}(m_{N}+E_{%
\mathbf{q}})}}G^{\rm eff}_{M}(Q^{2},t,t_{s}).$$
(6)
Similar relations hold for the axial and induced pseudoscalar form factors
[2, 3].
2 Excited-State Contributions
For $Q^{2}\not=0$, the excited states contributing to the incoming and outgoing nucleon channels are different in our kinematics (where the final nucleon is at rest), and ignoring finite-size effects we can parameterize the leading excited-state contaminations by
$$G^{\rm eff}_{X}(Q^{2},t,t_{s})=G_{X}(Q^{2})+c_{X,1}(Q^{2})\textrm{e}^{-\Delta t%
}+c_{X,2}(Q^{2})\textrm{e}^{-\Delta^{\prime}(t_{s}-t)}+\ldots$$
(7)
Since our aim is the extraction of the ground-state form factor $G_{X}(Q^{2})$, we need to find means to identify and remove, or else to suppress in some other way, the leading excited-state contributions. For this end we have considered three different methods
[4, 5]:
•
The well-known plateau method is based on the assumption that excited-state contributions in (7) are sufficiently suppressed, and proceeds by looking for a region in $t$ where the signal for $G^{\rm eff}_{X}(Q^{2},t,t_{s})$ for given $Q^{2}$ and $t_{s}$ has reached a plateau, which is then fitted to a constant identified as $G_{X}(Q^{2})$. While well-tried and simple, this method is beset with a number of problems, the most prominent of which is that in order for the plateau to be a good estimate for $G_{X}(Q^{2})$, the excited-state contributions from both sides need to have decayed by at least one or two e-foldings; in order to achieve this, large values for $t_{s}$ are required. Unfortunately, the signal-to-noise ratio becomes rather poor as $t_{s}$ is increased, rendering this difficult to achieve with present statistics. As we observe a systematic trend in the plateau values as a function of $t_{s}$ even for relatively large distances $t_{s}\sim 1.4$ fm (cf. fig. 1), the practicability of this method is seriously in doubt.
•
The by-now also well-established summation method
[6, 2]
proceeds by forming summed ratios
$$S_{X}(Q^{2},t_{s})=\sum_{t=0}^{t_{s}}G^{\rm eff}_{X}(Q^{2},t,t_{s})\rightarrow
c%
+t_{s}\left\{G_{X}(Q^{2})+\mathcal{O}\left(\textrm{e}^{-\Delta t_{s}}\right)\right\}$$
and fitting their $t_{s}$-dependence to a straight line, from the slope of which the desired ground-state form factor $G_{X}(Q^{2})$ can be extracted. The great advantage of the summation method is that it leads to a parametrically reduced excited-state contamination, since any $t$-dependence has been eliminated by summing over $t$, and $\Delta t_{s}\gg\Delta t$ by construction. The summation method’s major disadvantage is a significant increase in statistical errors when compared to the plateau method. One could therefore argue that the reduction in (hard to control) systematic error is paid for by a corresponding increase in (expensive to reduce) statistical error.
•
To have a further handle on the precise behaviour of the excited-state contributions, we have also considered two-state fits
[4, 7],
i.e. we have explicitly fit $G^{\rm eff}_{X}(Q^{2},t,t_{s})$ to leading excited-state contributions of the form (7) with $\Delta=m_{\pi}$, $\Delta^{\prime}=2m_{\pi}$, both
as a function of $t$, $t_{s}-t$ at each $t_{s}$ separately, and
as a function of $t_{s}$, $t$ at all $t_{s}$ simultaneously.
The advantage of these fits is that they are able to fully remove the leading excited-state contamination contained in the signal and project onto the ground state without any need to further increase the range of $t$ and $t_{s}$ under consideration. Their disadvantage lies in that they are somewhat dependent on our assumptions about the excited states (in particular the energy gap), making it somewhat difficult to assess the overall trustworthiness of the fitted results. In light of this latter disadvantage, we do not currently expect to quote final results from the two-state fits, but only to use them as a control to verify that excited-state contaminations do not seriously affect our final answers from the other methods.
In fig. 2, we compare the pion-mass dependence of the value of the electric Sachs form factor evaluated at $Q^{2}=0.1$ GeV${}^{2}$, $G_{E}(0.1~{}\textrm{GeV}^{2})$, as determined using the different methods. For the plateau method, there appears to be no appreciable pion-mass dependence, and the chiral extrapolation does not approach the experimental value as the pion mass is reduced. The summation method shows a slight improvement in so far as a chiral trend is visible, but an extrapolation would still miss a experimental value by a significant margin (although the discrepancy is less significant than for the plateau method, a large portion of this is due to the larger errors associated with the summation method). By contrast, the simultaneous two-state fits yield results that show a clear chiral trend and point towards the experimental value in the physical limit.
By parameterizing each Sachs form factor as a dipole
$$G_{X}(Q^{2})=G_{X}(0)\left(1+\frac{Q^{2}}{M_{X}^{2}}\right)^{-2}$$
(8)
where $G_{E}(0)=1$, $G_{M}(0)=\mu$, we can extract the charge radii and anomalous magnetic moment $\kappa$ via
$$\frac{1}{M_{E}^{2}}=\frac{r_{E}^{2}}{12}=\frac{r_{1}^{2}}{12}+\frac{\kappa}{8m%
_{N}^{2}}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}\frac{1}{M_{M}^{2}}=\frac{r_{M}^{2}}{12%
}=\frac{r_{1}^{2}+\kappa r_{2}^{2}}{12(1+\kappa)}$$
(9)
where $\kappa=\mu-1$.
Fig. 3 shows the data for $G_{E}$ and $G_{M}$ together with the dipole fits. The pion-mass dependence of the results obtained for the radii and the magnetic moment from the plateau and summation methods, respectively, is shown in fig. 4.
3 A Possible Scenario
In order to explain why we observe such a different chiral behaviour for the plateau and summation methods on the one hand, and the two-state fits on the other hand, it helps to consider a scenario in which we assume the two-state fits to be a true description of the data. Under this assumption, we can reconstruct the summed ratios that we would obtain at various value of $t_{s}$, and compare the slope obtained from fitting a straight line to the reconstructed values with the “true” value of the form factor.
An example of such a reconstruction on one of our ensembles is shown in fig. 5. In the left panel, we show the reconstructed ratios for the values of $t_{s}$ considered here as blue dots, with the straight-line fit to those reconstructed points shown in red. Also shown as a blue line is the functional form of the summed ratios as a function of $t_{s}$ over a much larger range of source-sink separations. It can be seen that the slope of the red line does not agree well with that of the blue curve as $t_{s}$ grows. In the right panel, we have added the asymptotic linear behaviour of the summed ratio (the dashed blue line). As can be seen from the inset, while the straight-line fit to the reconstructed ratios is indistinguishable from the true functional form at low values of $t_{s}$, it is quite different from the correct asymptotic value.
This behaviour is to be contrasted with the one observed for $g_{A}$
[2],
where the summation method has been found to be very effective in extracting the asymptotic behaviour. A crucial difference between the two cases is that for the form factors at $Q^{2}\not=0$, the gap appearing in the leading exponential corrections to the summed ratio is only $m_{\pi}$, rather than $2m_{\pi}$ as for the $Q^{2}=0$ case. This makes the approach to the asymptotic linear behaviour much slower, and (at least at first sight) seems to require twice as large a source-sink sparation $t_{s}$ in order to obtain the same reduction in systematic error.
4 Conclusions
A systematic trend in the value of $G_{E}$ obtained from the plateau method persists to $t_{s}\sim 1.4$ fm, and even with the summation method, $G_{E}$ systematically comes out too high; considering only the largest values of $t_{s}$ in the summation method brings the results closer to experiment, albeit at the expense of very large statistical errors.
The impact of excited states on the results of the plateau, and to a lesser extent also the summation method, appears to increase even further as the chiral limit is approached.
A possible reason for these observations is indicated by two-state fits: Given the appearance of the small gap $m_{\pi}$ for $Q^{2}\not=0$, the approach to the plateau is very slow, and the summed ratios still receive sizeable corrections to their asymptotic linear behaviour.
Acknowledgments:
We are grateful to our colleagues within the Coordinated Lattice Simulations
(CLS) initiative for sharing ensembles.
This work was granted access to the HPC resources of the Gauss Center for
Supercomputing at Forschungzentrum Jülich, Germany, made available within
the Distributed European Computing Initiative by the PRACE-2IP, receiving
funding from the European Community’s Seventh Framework Programme
(FP7/2007-2013) under grant agreement RI-283493 (project PRA039).
We are grateful for computer time allocated to project HMZ21 on the
JUQUEEN BG/Q computer at NIC, Jülich.
This work was supported by the DFG via SFB 1044
and grant HA 4470/3-1. We thank Dalibor Djukanovic and Christian Seiwerth
for technical support.
References
[1]
B. Jäger et al., PoS(LATTICE 2013)272
[arXiv:1311.5804].
[2]
S. Capitani et al., Phys. Rev. D 86 (2012) 074502
[arXiv:1205.0180].
[3]
P. Junnarkar et al., PoS(LATTICE 2014)150.
[4]
S. Capitani et al., PoS(LATTICE 2012)177
[arXiv:1211.1282].
[5]
S. Capitani et al., PoS(Confinement X)321.
[6]
L. Maiani et al., Nucl. Phys. B 293 (1987) 420.
[7]
J. Green et al., PoS(LATTICE 2011)157
[arXiv:1111.0255].
[8]
J. J. Kelly,
Phys. Rev. C 70 (2004) 068202. |
Spin-polarized currents driven by spin-dependent surface screening
Piotr Graczyk
[email protected]
Institute of Molecular Physics, Polish Academy of Sciences, M. Smoluchowskiego 17, 60-179 Poznan, Poland
Maciej Krawczyk
[email protected]
Faculty of Physics, Adam Mickiewicz University in Poznan, Umultowska 85, 61-614 Poznan, Poland
Abstract
We have examined spin polarization of the electron current in a ferromagnetic metal induced by the spin-dependent surface screening at the dielectric-ferromagnetic metal (D-FM) interface. In an applied ac voltage, the dynamic band splitting driven by the changes in the screening charge at the D-FM interface develops spin accumulation. The resultant spin accumulation gradient produces a time-dependent spin current. We have derived the rate of the spin accumulation dependent on the rate of the screening charge density accumulation within the Stoner band model. The spin-charge dynamics in the system is then modeled by a set of diffusive equations with the contributions from spin-dependent surface screening and spin-dependent conductivity. We show for MgO-Cu-Co-MgO system that the spin-dependent screening in thin Co film produces spin accumulation 7 times higher than that obtained by the spin-dependent conductivity in thick Co films. We propose an experimental approach to validate our numerical predictions and to distinguish between spin accumulation induced by the spin-dependent conductivity and by the spin-dependent surface screening.
The phenomenon of spin-dependent surface screening (SS) leads to the charge-mediated magnetoelectric effect (ME) which is the key mechanism for the electric control of magnetic properties in ultrathin multiferroic heterostructures Burton2012 ; Vaz2012 . The effect has been described first theoretically by Zhang in 1999 Zhang1999 and then confirmed by ab initio calculations Duan2008 ; Niranjan2009 ; Rondinelli2008 . It has been shown experimentally that SS may be utilized to change the magnetization, anisotropy or coercivity of ultrathin ferromagnetic film Molegraaf2009 ; Weisheit2007 ; Maruyama2009 ; Endo2010 ; Shiota2009 ; Shiota2011 ; Seki2011 ; Chiba2011 , to induce magnetic ordering transition Burton2009 ; Vaz2010 or even induce surface magnetism in nonmagnetic materials Sun2010 . In multiferroic tunnel junctions SS provides significant contribution to the tunneling resistance Zhuravlev2010 ; Yin2015 and opens possibilities to manipulate the spin polarization of tunneling electrons Pantel2012 ; Sen2019 . Owing to the surface magnetic anisotropy that is controlled by SS, spin waves may be excited parametrically or resonantly with the ac voltage Rana2017 ; Verba2014 ; Verba2017 ; Chen2017 ; Nozaki2012 .
Spin-dependent surface screening changes the absolute value of magnetization at the dielectric-ferromagnetic metal (D-FM) interface, in contrast to, studied more extensively, strain-driven ME effects Vaz2012 ; Graczyk2015 ; Graczyk2016 ; Graczyk2016a , which rotate the magnetization direction. Although strain-driven ME remains significant for relatively thick layers, it suffers from ageing effects, clamping and integration problems with current technology.
In this letter we study the generation of spin-polarized currents by spin-dependent surface screening effect in D-FM-D system subjected to the ac voltage. The change of the voltage results in the change of the screening charges at the insulator-ferromagnet interface. Since the screening is spin-dependent, this leads to the appearance of nonequilibrium spin density at the interface and then formed spin-density gradient generates diffusive spin current. We discuss this effect within the Stoner band model and derive the term that describes SS in drift-diffusive formalism. With this model we perform numerical simulations and find the spin current with the contributions from spin-dependent conductivity (SC) and spin-dependent surface screening (SS) for MgO-Cu-Co-MgO system. Finally, we propose the strategy to distinguish experimentally between that two current polarization mechanisms.
Considered systems are shown in Fig. 1. The multilayer consists of two 2-nm-thick MgO films separated by Co film of different thicknesses. In system A the Co layer is separated from MgO by 2-nm-thick nonmagnetic metal (Cu) at both interfaces while in the system B it is separated from MgO only from the top. Therefore, in the system A there is no D-FM interface, while in the system B it is present. Consequently, screening charges will appear in the ferromagnet only at the bottom interface in the system B when the voltage is applied to the multilayer structure through bottom and top electrodes (e2 and e1, respectively). It is assumed that charge and spin transport occur only perpendicular to the layers, along $x$ axis.
The coupled charge-spin dynamics driven by the prescribed ac voltage is considered on the base of the diffusive model Valet1993 ; Zhu2008 . The free charge density current is described in the metal layers by the equation:
$$J_{f}=\sigma E-D\frac{\partial n_{f}}{\partial x}+\beta D\frac{e}{\mu_{B}}%
\frac{\partial s}{\partial x},$$
(1)
which includes contributions from electron drift driven by electric field $E$, diffusion driven by the gradient of free charge density $n_{f}$ and diffusive spin polarisation (SC), respectively. Here, $\sigma$ is conductivity, $D$ is diffusion constant, $\beta$ is spin asymmetry coefficient, $\mu_{B}$ is Bohr magneton and $e$ is charge of electron. The displacement current is described by the equation:
$$J_{b}=\epsilon_{0}\epsilon_{r}\frac{\partial E}{\partial t},$$
(2)
where $\epsilon_{0}\epsilon_{r}$ is permittivity of the material. The conservation of free ($i=f$) and bound ($i=b$) charge density $n_{i}$ is described by continuity equation:
$$\frac{\partial n_{i}}{\partial t}=-\frac{\partial J_{i}}{\partial x}.$$
(3)
Electric potential $V$ is given by the Gauss Law:
$$\begin{split}\displaystyle\epsilon_{0}\frac{\partial^{2}V}{\partial x^{2}}&%
\displaystyle=n_{f}+n_{b},\\
\displaystyle E&\displaystyle=-\frac{\partial V}{\partial x},\end{split}$$
(4)
with the boundary conditions at the electrodes:
$$\begin{split}\displaystyle V|_{x=e1}&\displaystyle=U=U_{0}\sin(2\pi f_{0}t),\\
\displaystyle V|_{x=e2}&\displaystyle=0,\end{split}$$
(5)
where $f_{0}$ is the frequency of the oscillating voltage $U$.
The spin current $J_{s}$ in the ferromagnetic material is modeled by the equation Zhu2008 :
$$J_{s}=-D\frac{\partial s}{\partial x}-\beta\frac{\mu_{B}}{e}\left(\sigma E-D%
\frac{\partial n_{f}}{\partial x}\right),$$
(6)
which describes spin current driven by gradient of spin accumulation $s$ (diffusion) and spin-conductivity term.
The continuity equation for the spin accumulation $s$ is:
$$\frac{\partial s}{\partial t}=-\frac{\partial J_{s}}{\partial x}-\frac{s}{T_{1%
}}+f_{s},$$
(7)
which describes the rate of change of $s$ due to gradient of the spin current, the spin-flip relaxation with the characteristic time $T_{1}$. $f_{s}$ is an additional source from spin-dependent charge screening. The formula for $f_{s}$ is derived below. We use in the following paragraphs $n\equiv n_{f}$.
We will consider the voltage-driven magnetization and spin density change in cobalt within the Stoner model Stoner1938 . Due to the exchange splitting of the band d for the majority and minority spins and partially filled minority subband (Fig. 2a) there exists equilibrium spin density (magnetization) in the bulk of the ferromagnet which is equal to the difference between charge densities of the spin-up ($\uparrow$) and spin-down ($\downarrow$) subbands, $m=\mu_{B}(n^{\uparrow}-n^{\downarrow})/e$. The voltage applied to the system shifts the potential of the bands at the D-FM interfaces (Fig. 2b) leading to the accumulation of the screening charge density $n=\delta n^{\uparrow}_{0}+\delta n^{\downarrow}_{0}$ where $\delta n_{0}^{\sigma}$ is accumulated screening charge in the subband $\sigma$. Since the density of states are different between subbands at the Fermi level ($\Delta\rho=\rho^{\uparrow}-\rho^{\downarrow}<0$ for Co), then screening charge changes the value of the magnetization at the interface by $\delta m_{0}=\mu_{B}(\delta n^{\uparrow}_{0}-\delta n^{\downarrow}_{0})/e$ Duan2008 .
Another important feature of the screening in the ferromagnet is that the changes in the band occupancy influence the exchange splitting Niranjan2009 , i.e. the potential is spin-dependent Zhang1999 ; Zhuravlev2010 :
$$V^{\sigma}=V_{c}-\frac{I}{e}(\delta n^{\sigma}-\delta n^{-\sigma}),$$
(8)
where $I$ is exchange constant. Spin-dependent potential (8) means that the subbands shift relative to each other by $V_{s}=V^{\uparrow}-V^{\downarrow}=-2I\delta m$. (Fig. 2c). From Eqs. (5) and (11) in Ref. Zhang1999 we get the relation between $V_{s}$ and screening charge $n=\delta n^{\uparrow}+\delta n^{\downarrow}$ in the linearized Thomas-Fermi model:
$$V_{s}=-\frac{2I\Delta\rho}{\rho+4I\rho^{\uparrow}\rho^{\downarrow}}n,$$
(9)
where $\rho=\rho^{\uparrow}+\rho^{\downarrow}$. This shift has important consequences. Since $\rho^{\uparrow}\neq\rho^{\downarrow}$, the shift changes the equilibrium charge level with respect to the Fermi level by $E_{F}-s_{0}$ (Fig. 2d) where $s_{0}=-e(V^{\uparrow}\rho^{\uparrow}+V^{\downarrow}\rho^{\downarrow})/\rho$ is spin equilibrium level. This leads to the changes in equilibrium charge densities $\delta n^{\sigma}\neq\delta n_{0}^{\sigma}$ and potential $V_{c}$. Therefore, the surface magnetization value changes by $\delta m\neq\delta m_{0}$. Moreover, the screening length in ferromagnet $\lambda_{F}$ deviates from the conventional Thomas-Fermi value $\lambda=\left(\epsilon_{0}/e^{2}\rho\right)^{1/2}$ and thus the capacitance is reduced Zhang1999 . We can implement this effect to our numerical model where the screening length is defined as $\lambda=\left(\epsilon_{0}D/\sigma\right)^{1/2}$ by dividing drift term and multiplying diffusion term in Eq. (1) by $\gamma_{n}=\lambda_{F}/\lambda\approx 1.1$ for Co. We found that modification of the screening length does not influence the effect of the spin-polarized current generation which is considered here so it will be neglected in further part.
Since screening charge and thus $V_{s}$ (Eq. (9)) is time-dependent under the ac volatage, the change of the shift between subbands $\Delta V_{s}$ is the source of nonequilibrium spin density $\Delta s\propto\Delta V_{s}$. From Fig. 2d we see that with the respect to the $s_{0}$:
$$\begin{split}\displaystyle\Delta s&\displaystyle=\frac{\mu_{B}}{e}(\Delta n^{%
\uparrow}-\Delta n^{\downarrow})\\
&\displaystyle=\frac{\mu_{B}}{e}(-eV^{\uparrow}-s_{0})\rho^{\uparrow}-\frac{%
\mu_{B}}{e}(-eV^{\downarrow}-s_{0})\rho^{\downarrow}\\
&\displaystyle=-2\mu_{B}\Delta V_{s}\rho^{\uparrow}\rho^{\downarrow}/\rho e.%
\end{split}$$
(10)
Taking time derivatives we get from (9) and (10) the source term $f_{s}$ to be:
$$f_{s}=\frac{4I\rho^{\uparrow}\rho^{\downarrow}}{\rho}\frac{\Delta\rho}{\rho+4I%
\rho^{\uparrow}\rho^{\downarrow}}\frac{\mu_{B}}{e}\frac{\partial n}{\partial t%
}=\gamma_{s}\frac{\mu_{B}}{e}\frac{\partial n}{\partial t}.$$
(11)
In equilibrium the final change of magnetization $\delta m=\mu_{B}(\delta n^{\uparrow}-\delta n^{\downarrow})/e$ is a consequence of uniform potential shift of the bands and their relative exchange splitting (Fig. 2e).
Equations (1)-(7) with source term (11) are solved numerically by finite element method in Comsol Multiphysics with time-varying voltage applied to the electrodes with amplitude $U_{0}=8$ V and frequency $f_{0}=200$ MHz. The relative permittivity is $\epsilon_{r}=10$ for MgO and $\epsilon_{r}=1$ for metals. We assume the conductivity of metals $\sigma=1.2\cdot 10^{7}$ S/m and the diffusion constant $D=4\cdot 10^{-3}$ m${}^{2}$/s. The spin asymmetry coefficient for Co is $\beta=0.5$ Zhu2008 , the spin-flip relaxation time is $T_{1}=0.9$ ps Zhu2008 and the magnetoelectric coefficient calculated from Eq. (11) and parameters given in Ref. Zhang1999 is $\gamma_{s}=-0.25$. We assume also the magnetization of Co film lies in-plane of the film. Below we analize the distribution of amplitudes of spin accumulation $s$ and spin current $J_{s}$ between MgO layers with the contributions from SC and SS.
Figure 3 shows the calculated spatial distribution of the spin accumulation and the spin current (for the system B and thickness of Co film $L_{\text{F}}$=51 nm) at the moment when they reach maximum values, i.e., when $V$ crosses zero value and charge current flows in the +$x$ direction. We separated the contributions which come from SC and SS. For the current driven by the spin-dependent conductivity, since $\beta>0$ then $\tau_{-}>\tau_{+}$, i.e., the relaxation time for spin down is bigger than for spin up and there is more spin down electrons flowing than spin up electrons. Consequently, $J_{s}$ has negative sign (Fig. 3b, blue line) and its maximum is at the center of Cu/Co bilayer. Within the distance proportional to the spin-diffusion length ($L_{sf}=$60 nm for Co) from the ferromagnet surfaces there builds up a negative spin accumulation gradient (Fig. 3a, blue line) that stands against SC term (compare Eq. (6)) Zayets2012 . In the Cu layer there is weak exponential decay of $s$ from the Cu-Co interface.
For the current driven by the spin-dependent surface screening, since $\gamma_{s}<0$, the outflow of negative charges from the F-D interface generates negative spin accumulation (Fig. 3a, red line). The resulting spin accumulation gradient induces spin current in the direction +$x$ (Fig. 3b, red line). The maximum value of $J_{s}$ is close to the D-FM interface.
We have shown that the spin current driven by SS is a consequence of the spin accumulation gradient build-up. This is in contrast to the SC-polarized current, where spin accumulation gradient is a result of this current and stands against it.
The effects from SS and SC sum up to give resultant spin accumulation and spin current (black lines in Fig. 3). In Co ($\beta>0$, $\gamma_{s}<0$) they add up destructively and for Co thickness of 51 nm discussed above, that contributions balance out at the Cu layer as shown at the left-hand side of dash-dotted line in Fig. 3b.
It is possible to distinguish between SC and SS generated spin currents comparing samples where SS would be absent (system A) and samples where SS would be present (system B). Moreover, spin-dependent conductivity is a phenomenon that acts in a bulk while spin-dependent surface screening is an interface effect. Thus, in thick ferromagnetic layer SC shall dominate while SS should manifest itself in thin ferromagnetic layers.
We consider the values of spin accumulation and spin current in the top Cu layer and we change the thickness of Co layer. The black lines in Fig. 4 show $s$ and $J_{s}$ calculated for the system A. Since there is no D-FM interface, the spin polarization of the current is a result of spin-dependent conductivity. The flow of spin-polarized charges builds up a spin accumulation as discussed above. For a thin Co layer ($L_{\text{F}}<2L_{\text{sf}}$) the spin accumulation counteracts the spin current significantly. For the thickness larger than $2L_{\text{sf}}$ the influence of the interfacial accumulation loses its significance and the value of spin-polarized current reaches maximum value of 105 A/s in Cu (and the value of $J_{s}\approx-\beta J_{f}=7000$ A/s in the bulk of Co layer). The values of $s$ and $J_{s}$ saturate for thick Co films (see insets in Fig. 4).
The red lines in Fig. 4 show $s$ and $J_{s}$ for system B. Because of D-FM interface is present in the system, there are contributions both from SC and SS to the spin transport. Spin-dependent conductivity dominates for $L_{\text{F}}>L_{\text{sf}}$. If the distance from the D-FM interface to Cu layer is less than $2L_{\text{sf}}$, the currents induced by spin-dependent surface screening come into picture and the values of $s$ and $J_{s}$ deviates from those for system A. At $L_{\text{F}}=51$ nm SS balances out SC contribution in Cu and for $L_{\text{F}}<51$ nm spin current driven by SS dominates in the Cu layer. The maximum values obtained for $L_{F}=$ 2 nm are $s=$ 650 mA/m and $J_{s}=$730 A/s which are 7 times higher than those obtained for thick Co films by spin-dependent conductivity.
The dependences of $s$ and $J_{s}$ on $L_{\text{F}}$ in the systems considered above give us simple tool to distinguish between SS- and SC- driven currents. If SS contribution is absent in the system then $s\rightarrow 0$ and $J_{s}\rightarrow 0$ as $L_{\text{F}}\rightarrow 0$ for $L_{\text{F}}<L_{\text{sf}}$. If SS contribution is nonzero, $s$ and $J_{s}$ tend to nonzero values as $L_{\text{F}}\rightarrow 0$ for $L_{\text{F}}<L_{\text{sf}}$. It is rather crucial to choose ferromagnet with relatively long spin-diffusion length like Co, since the effect is hard to resolve for $L_{\text{F}}>L_{\text{sf}}$.
In summary, the mechanism of spin current generation by the spin-dependent surface screening effect has been described within the Stoner model. The mechanism has been implemented to the drift-diffusion model. Then the specific dielectric-ferromagnetic heterostructures have been considered numerically to show the properties of SS-driven spin currents. We found the spin accumulation and the spin polarization of the current driven by spin-dependent surface screening is 7 times higher for 2-nm-thick Co than those obtained by spin-dependent conductivity for thick Co layers. Since the strength of the effect is dependent on the density of the screening charge at the D-FM interface, it may be enhanced with the use of dielectrics of high permittivity. Therefore, it is interesting to consider the influence of SS-driven spin currents onto magnetization dynamics, in particular, the possibility to decrease spin wave damping through currents induced by that mechanism. The proposed systems and the method of extraction of SS-driven spin accumulation is suitable to the X-Ray magnetic circular dichroism experiment Kukreja2015 ; Li2016 , where the measurement of spin accumulation dynamics is element-specific and thus may be constrained to the signal from Cu layer and directly compared to the dependences shown in Fig. 4.
Acknowledgements.
The study has received financial support from the National Science Centre of Poland under grant 2018/28/C/ST3/00052.
References
(1)
J. D. Burton and E. Y. Tsymbal, “Magnetoelectric interfaces and spin
transport,” Philosophical Transactions of the Royal Society A:
Mathematical, Physical and Engineering Sciences, vol. 370, no. 1977,
pp. 4840–4855, 2012.
(2)
C. A. F. Vaz, “Electric field control of magnetism in multiferroic
heterostructures,” Journal of Physics: Condensed Matter, vol. 24,
p. 333201, 2012.
(3)
S. Zhang, “Spin-Dependent Surface Screening in Ferromagnets and Magnetic
Tunnel Junctions,” Physical Review Letters, vol. 83, no. 3,
pp. 640–643, 1999.
(4)
C.-g. Duan, J. P. Velev, R. F. Sabirianov, Z. Zhu, J. Chu, S. S. Jaswal, and
E. Y. Tsymbal, “Surface Magnetoelectric Effect in Ferromagnetic Metal
Films,” Physical Review Letters, vol. 101, p. 137201, 2008.
(5)
M. K. Niranjan, J. D. Burton, J. P. Velev, S. S. Jaswal, and E. Y. Tsymbal,
“Magnetoelectric effect at the SrRuO3/BaTiO3 (001) interface: An ab initio
study,” Applied Physics Letters, vol. 95, p. 052501, 2009.
(6)
J. M. Rondinelli, M. Stengel, and N. A. Spaldin, “Carrier-mediated
magnetoelectricity in complex oxide heterostructures,” Nature
Nanotechnology, vol. 3, no. January, pp. 46–50, 2008.
(7)
H. J. Molegraaf, J. Hoffman, C. A. Vaz, S. Gariglio, D. Van Der Morel, C. H.
Ahn, and J. M. Triscone, “Magnetoelectric effects in complex oxides with
competing ground states,” Advanced Materials, vol. 21, no. 34,
pp. 1–5, 2009.
(8)
M. Weisheit, S. Fähler, A. Marty, Y. Souche, C. Poinsignon, and
D. Givord, “Electric field-induced modification of magnetism in thin-film
ferromagnets,” Science, vol. 315, no. 5810, pp. 349–351, 2007.
(9)
T. Maruyama, Y. Shiota, T. Nozaki, K. Ohta, N. Toda, M. Mizuguchi, A. A.
Tulapurkar, T. Shinjo, M. Shiraishi, S. Mizukami, Y. Ando, and Y. Suzuki,
“Large voltage-induced magnetic anisotropy change in a few atomic layers of
iron,” Nature Nanotechnology, vol. 4, p. 158, jan 2009.
(10)
M. Endo, S. Kanai, S. Ikeda, F. Matsukura, and H. Ohno, “Electric-field
effects on thickness dependent magnetic anisotropy of sputtered
MgO/Co40Fe40B20/Ta structures,” Applied Physics Letters, vol. 96,
p. 212503, may 2010.
(11)
Y. Shiota, T. Maruyama, T. Nozaki, T. Shinjo, M. Shiraishi, and Y. Suzuki,
“Voltage-assisted magnetization switching in ultrathin Fe 80Co20 alloy
layers,” Applied Physics Express, vol. 2, no. 6, p. 063001, 2009.
(12)
Y. Shiota, T. Nozaki, F. Bonell, S. Murakami, T. Shinjo, and Y. Suzuki,
“Induction of coherent magnetization switching in a few atomic layers of
FeCo using voltage pulses,” Nature Materials, vol. 11, p. 39, nov
2011.
(13)
T. Seki, M. Kohda, J. Nitta, and K. Takanashi, “Coercivity change in an FePt
thin layer in a Hall device by voltage application,” Applied Physics
Letters, vol. 98, p. 212505, may 2011.
(14)
D. Chiba, S. Fukami, K. Shimamura, N. Ishiwata, K. Kobayashi, and T. Ono,
“Electrical control of the ferromagnetic phase transition in cobalt at room
temperature,” Nature Materials, vol. 10, p. 853, oct 2011.
(15)
J. D. Burton and E. Y. Tsymbal, “Prediction of electrically induced magnetic
reconstruction at the manganite/ferroelectric interface,” Physical
Review B, vol. 80, p. 174406, 2009.
(16)
C. A. F. Vaz, J. Hoffman, Y. Segal, J. W. Reiner, R. D. Grober, Z. Zhang, C. H.
Ahn, and F. J. Walker, “Origin of the magnetoelectric coupling effect in
PZT/LSMO multiferroic heterostructures,” Physical Review Letters,
vol. 104, no. March, p. 127202, 2010.
(17)
Y. Sun, J. D. Burton, and E. Y. Tsymbal, “Electrically driven magnetism on a
Pd thin film,” Physical Review B, vol. 81, p. 064413, 2010.
(18)
M. Y. Zhuravlev, S. Maekawa, and E. Y. Tsymbal, “Effect of spin-dependent
screening on tunneling electroresistance and tunneling magnetoresistance in
multiferroic tunnel junctions,” Physical Review B, vol. 81, no. 10,
p. 104419, 2010.
(19)
Y. W. Yin, M. Raju, W. J. Hu, J. D. Burton, Y.-M. Kim, a. Y. Borisevich, S. J.
Pennycook, S. M. Yang, T. W. Noh, a. Gruverman, X. G. Li, Z. D. Zhang, E. Y.
Tsymbal, and Q. Li, “Multiferroic tunnel junctions and ferroelectric
control of magnetic state at interface (invited),” Journal of Applied
Physics, vol. 117, p. 172601, 2015.
(20)
D. Pantel, S. Goetze, D. Hesse, and M. Alexe, “Reversible electrical
switching of spin polarization in multiferroic tunnel junctions,” Nature Materials, vol. 11, no. 4, pp. 289–293, 2012.
(21)
C. Sen, W. A. S. Aldulaimi, O. Mohammadmoradi, and I. B. Misirlioglu, “Loss
of spin polarization in ferromagnet / ferroelectric tunnel junctions due to
screening effects,” Journal of Physics D: Applied Physics, vol. 52,
p. 015305, 2019.
(22)
B. Rana, Y. Fukuma, K. Miura, H. Takahashi, and Y. Otani, “Excitation of
coherent propagating spin waves in ultrathin CoFeB film by voltage-controlled
magnetic anisotropy,” Applied Physics Letters, vol. 111, no. 5,
p. 052404, 2017.
(23)
R. Verba, V. Tiberkevich, I. Krivorotov, and A. Slavin, “Parametric
Excitation of Spin Waves by Voltage-Controlled Magnetic Anisotropy,” Physical Review Applied, vol. 1, no. 4, p. 044006, 2014.
(24)
R. Verba, M. Carpentieri, G. Finocchio, V. Tiberkevich, and A. Slavin,
“Excitation of Spin Waves in an In-Plane-Magnetized Ferromagnetic Nanowire
Using Voltage-Controlled Magnetic Anisotropy,” Physical Review
Applied, vol. 7, p. 064023, 2017.
(25)
Y.-j. Chen, H. K. Lee, R. Verba, J. A. Katine, I. Barsukov, V. Tiberkevich,
J. Q. Xiao, A. N. Slavin, and I. N. Krivorotov, “Parametric Resonance of
Magnetization Excited by Electric Field,” Nano Letters, vol. 17,
pp. 572–577, 2017.
(26)
T. Nozaki, Y. Shiota, S. Miwa, S. Murakami, F. Bonell, S. Ishibashi, H. Kubota,
K. Yakushiji, T. Saruya, A. Fukushima, S. Yuasa, T. Shinjo, and Y. Suzuki,
“Electric-field-induced ferromagnetic resonance excitation in an ultrathin
ferromagnetic metal layer,” Nature Physics, vol. 8, no. 6,
pp. 491–496, 2012.
(27)
P. Graczyk, R. Schäfer, and B. Mroz, “Magnetoelastic coupling between
NiFe thin film and LiCsSO4 studied by Kerr microscopy,” Journal of
Physics D: Applied Physics, vol. 48, no. 42, p. 425002, 2015.
(28)
P. Graczyk, A. Trzaskowska, K. Zaleski, and B. Mroz, “Tunable properties of
spin waves in magnetoelastic NiFe/GMO heterostructure,” Smart
Materials and Structures, vol. 25, no. 7, p. 075017, 2016.
(29)
P. Graczyk, A. Trzaskowska, K. Adrjanowicz, B. Peplinska, and B. Mroz, “Large
magnetoelastic coupling in NiFe/LiCsSO4 and NiFe/KH2PO4 heterostructures,”
Journal of Alloys and Compounds, vol. 656, pp. 825–829, 2016.
(30)
T. Valet and A. Fert, “Theory of the perpendicular magnetoresistance in
magnetic multilayers,” Physical Review B, vol. 48, no. 10,
pp. 7099–7113, 1993.
(31)
Y. H. Zhu, B. Hillebrands, and H. C. Schneider, “Signal propagation in
time-dependent spin transport,” Physical Review B, vol. 78, no. 5,
p. 054429, 2008.
(32)
E. C. Stoner, “Collective electron ferronmagnetism,” Proceedings of
the Royal Society of London. Series A. Mathematical and Physical Sciences,
vol. 165, no. 922, pp. 372–414, 1938.
(33)
V. Zayets, “Spin and charge transport in materials with spin-dependent
conductivity,” Physical Review B, vol. 86, no. 17, p. 174415, 2012.
(34)
R. Kukreja, S. Bonetti, Z. Chen, D. Backes, Y. Acremann, J. A. Katine, A. D.
Kent, H. A. Dürr, H. Ohldag, and J. Stöhr, “X-ray Detection of
Transient Magnetic Moments Induced by a Spin Current in Cu,” Physical
Review Letters, vol. 115, no. 9, p. 096601, 2015.
(35)
J. Li, L. R. Shelford, P. Shafer, A. Tan, J. X. Deng, P. S. Keatley, C. Hwang,
E. Arenholz, G. Van Der Laan, R. J. Hicken, and Z. Q. Qiu, “Direct
Detection of Pure ac Spin Current by X-Ray Pump-Probe Measurements,” Physical Review Letters, vol. 117, no. 7, p. 076602, 2016. |
Emotional Musical Prosody: Validated Vocal Dataset for Human Robot Interaction
Richard Savery
Georgia Tech Center for Music Technology
11email: [email protected]
Lisa Zahray
Georgia Tech Center for Music Technology
11email: [email protected]
Gil Weinberg
This material is based upon work supported by the National Science Foundation under Grant No. 1925178
Georgia Tech Center for Music Technology
11email: [email protected]
Abstract
Human collaboration with robotics is dependant on the development of a relationship between human and robot, without which performance and utilization can decrease. Emotion and personality conveyance has been shown to enhance robotic collaborations, with improved human-robot relationships and increased trust. One under-explored way for an artificial agent to convey emotions is through non-linguistic musical prosody. In this work we present a new 4.2 hour dataset of improvised emotional vocal phrases based on the Geneva Emotion Wheel. This dataset has been validated through extensive listening tests and shows promising preliminary results for use in generative systems.
Keywords:robotics, emotion, prosody, music
1 Introduction
As the use of robotics and artificial agents continue to expand, there is a growing need for better forms of communication between human and computer. While speech based mediums are common in home assistants and robots, we contend that for many human-agent collaborations, semantic meaning is not always required. One alternate to speech communication is non-linguistic emotional musical prosody where a robot communicates through musical phrases. Using musical phrases can avoid the challenges of uncanny valley, allowing robot interaction to have it’s own human-inspired form of communication [\citeauthoryearSavery, Rose\BCBL \BBA WeinbergSavery \BOthers.\APACyear2019\APACexlab\BCnt2]. In past work we have successfully shown the capabilities of prosody in robotic systems through simulations of a potential generative system. This includes increased trust in social robots [\citeauthoryearSavery, Rose\BCBL \BBA WeinbergSavery \BOthers.\APACyear2019\APACexlab\BCnt1] and industrial robots [\citeauthoryearSavery, Zahray\BCBL \BBA WeinbergSavery \BOthers.\APACyear2020]. In this paper we present the collection and validation of a emotional prosody dataset that can be used to improve future human robot interactions.
2 Custom Dataset
Our custom dataset with 4.2 hours of audio was created with the singer Mary Carter. After a pilot study collecting data, we decided to use the Geneva Emotion Wheel (GEW) [\citeauthoryearSacharin, Schlegel\BCBL \BBA SchererSacharin \BOthers.\APACyear2012], which is a circular model, containing 20 emotions with position corresponding to the circumplex model (see Figure 2).
One of the primary advantages of the GEW is that it includes 20 different emotions, but these emotions can also be reduced to four separate classes which align with a quadrant from the circumplex model. GEW also includes most of the Eckman’s basic emotions - fear, anger, disgust, sadness, happiness - only leaving out surprise. The ability to use different models of emotion allows for significant future use cases of the dataset [\citeauthoryearSavery \BBA WeinbergSavery \BBA Weinberg\APACyear2020].
This dataset only has one musician, and therefore only captures one person’s perspective on musical emotion. While the dataset can make no claim to represent all emotion and does not create a generalized emotion model, we believe using one person has advantages. By having only one vocalist generative systems have the possibility of recreating one person’s emotional style, instead of incorrectly aggregating multiple peoples to remove distinctive individual and stylistic features.
2.1 Process and Data
Carter was paid $500 and recorded the samples over a week long period at her home studio, using a template we created in Logic while maintaining the same microphone positioning. For the samples we requested phrases to be between 1 and 20 seconds, and to spend about 15 minutes on each emotion, allowing jumping between any order of the emotions. We allowed deletion of a phrase if she felt after singing it did not fit the emotion. The final recorded dataset includes 2441 phrases equalling 4.22 hours of data with an average of 122 for each emotion. Samples from the dataset can be heard online.111www.richardsavery.com/prosodycvae
2.2 Dataset Validation
To validate the dataset, we performed a study with 45 participants from Prolific and Mechanical Turk, paying each $3. Each question in the survey asked the participant to listen to a phrase and select a location on the wheel corresponding to the emotion and intensity they believed the phrase was trying to convey. Phrases fell under two categories of “best” and “all”, with each participant seeing 60 total phrases selected at random. The “best” category consisted of 5 phrases for each emotion that were hand-selected as best representing that emotion, ensuring an even distribution of phrase lengths in each emotion set. The “all” category consisted of a phrase sampled from all phrases in the dataset for that emotion, with a new phrase randomly selected for each participant. Rose plots of the validation results that combine the “best” and “all” categories can be seen in the appendix, separated into each Geneva Wheel quadrant.
2.3 Dataset to Midi
We converted each phrase’s audio into a midi representation to use as training data. We first ran the monophonic pitch detection algorithm CREPE [\citeauthoryearKim, Salamon, Li\BCBL \BBA BelloKim \BOthers.\APACyear2018] on each phrase, which outputs a frequency and a confidence value for a pitch being present every 0.01 seconds. As the phrases included breaths and silence, it was necessary to filter out pitches detected with low confidence. We applied a threshold followed by a median filter to the confidence values, and then forced each detected pitch region to be at least 0.04 seconds long. We then converted the frequencies to midi pitches. We found the most common pitch deviation for each phrase using a histogram of deviations, shifting the midi pitches by this deviation to tune each phrase.
3 Conclusion
We have so far analyzed the dataset by training a Vanilla VAE (without emotion labels) on the pitch alone. Figure 2 shows the latent space reduced from 100 dimensions through t-distributed stochastic neighbor embedding (t-SNE) [\citeauthoryearHinton \BBA RoweisHinton \BBA Roweis\APACyear2003]. This demonstrates that the latent space is able to separate by emotion, without any knowledge of the emotional labels. We then used a conditional VAE (conditioned on emotion labels) to generate phrases with playback through a separate audio sampler. These phrases were validated following the format of our dataset evaluation, and the results are shown in the appendix.
In the future we will run extensive studies with generated prosody applied to human-robot interactions. This will take place between varying group sizes from one human and robot, to groups of humans and robots with different embedded personalities. We expect for emotional musical prosody to enable many future collaborations between human and robot.
References
[\citeauthoryearHinton \BBA RoweisHinton \BBA Roweis\APACyear2003]
\APACinsertmetastarhinton2003stochastic{APACrefauthors}Hinton, G\BPBIE.\BCBT \BBA Roweis, S\BPBIT.
\APACrefYearMonthDay2003.
\BBOQ\APACrefatitleStochastic neighbor embedding Stochastic neighbor
embedding.\BBCQ
\BIn \APACrefbtitleAdvances in neural information processing systems
Advances in neural information processing systems (\BPGS 857–864).
\PrintBackRefs\CurrentBib
[\citeauthoryearKim, Salamon, Li\BCBL \BBA BelloKim \BOthers.\APACyear2018]
\APACinsertmetastarkim2018crepe{APACrefauthors}Kim, J\BPBIW., Salamon, J., Li, P.\BCBL \BBA Bello, J\BPBIP.
\APACrefYearMonthDay2018.
\BBOQ\APACrefatitleCREPE: A convolutional representation for pitch
estimation Crepe: A convolutional representation for pitch
estimation.\BBCQ
\BIn \APACrefbtitle2018 IEEE International Conference on Acoustics, Speech
and Signal Processing (ICASSP) 2018 ieee international conference on
acoustics, speech and signal processing (icassp) (\BPGS 161–165).
\PrintBackRefs\CurrentBib
[\citeauthoryearSacharin, Schlegel\BCBL \BBA SchererSacharin \BOthers.\APACyear2012]
\APACinsertmetastarsacharin2012geneva{APACrefauthors}Sacharin, V., Schlegel, K.\BCBL \BBA Scherer, K.
\APACrefYearMonthDay2012.
\BBOQ\APACrefatitleGeneva Emotion Wheel rating study (Report). Geneva,
Switzerland: University of Geneva Geneva emotion wheel rating study
(report). geneva, switzerland: University of geneva.\BBCQ
\APACjournalVolNumPagesSwiss Center for Affective Sciences.
\PrintBackRefs\CurrentBib
[\citeauthoryearSavery, Rose\BCBL \BBA WeinbergSavery \BOthers.\APACyear2019\APACexlab\BCnt1]
\APACinsertmetastarsavery2019establishing{APACrefauthors}Savery, R., Rose, R.\BCBL \BBA Weinberg, G.
\APACrefYearMonthDay2019\BCnt1.
\BBOQ\APACrefatitleEstablishing Human-Robot Trust through Music-Driven
Robotic Emotion Prosody and Gesture Establishing human-robot trust through
music-driven robotic emotion prosody and gesture.\BBCQ
\BIn \APACrefbtitle2019 28th IEEE International Conference on Robot and
Human Interactive Communication (RO-MAN) 2019 28th ieee international
conference on robot and human interactive communication (ro-man) (\BPGS 1–7).
\PrintBackRefs\CurrentBib
[\citeauthoryearSavery, Rose\BCBL \BBA WeinbergSavery \BOthers.\APACyear2019\APACexlab\BCnt2]
\APACinsertmetastarsavery_finding_2019{APACrefauthors}Savery, R., Rose, R.\BCBL \BBA Weinberg, G.
\APACrefYearMonthDay2019\BCnt2.
\BBOQ\APACrefatitleFinding Shimi’s voice: fostering human-robot
communication with music and a NVIDIA Jetson TX2 Finding Shimi’s
voice: fostering human-robot communication with music and a NVIDIA Jetson
TX2.\BBCQ
\APACjournalVolNumPagesProceedings of the 17th Linux Audio Conference5.
\PrintBackRefs\CurrentBib
[\citeauthoryearSavery \BBA WeinbergSavery \BBA Weinberg\APACyear2020]
\APACinsertmetastarsaverysurvey{APACrefauthors}Savery, R.\BCBT \BBA Weinberg, G.
\APACrefYearMonthDay2020.
\BBOQ\APACrefatitleA Survey of Robotics and Emotion: Classifications and
Models of Emotional Interaction A survey of robotics and emotion:
Classifications and models of emotional interaction.\BBCQ
\BIn \APACrefbtitleProceedings of the 29th International Conference on
Robot and Human Interactive Communication. Proceedings of the 29th
international conference on robot and human interactive communication.
\PrintBackRefs\CurrentBib
[\citeauthoryearSavery, Zahray\BCBL \BBA WeinbergSavery \BOthers.\APACyear2020]
\APACinsertmetastarsavery2020emotional{APACrefauthors}Savery, R., Zahray, L.\BCBL \BBA Weinberg, G.
\APACrefYearMonthDay2020.
\BBOQ\APACrefatitleEmotional Musical Prosody for the Enhancement of Trust
in Robotic Arm Communication Emotional musical prosody for the enhancement
of trust in robotic arm communication.\BBCQ
\BIn \APACrefbtitle29th IEEE International Conference on Robot & Human
Interactive Communication. 29th ieee international conference on robot &
human interactive communication.
\PrintBackRefs\CurrentBib
4 Appendix |
Dispersive Approach to Chiral Perturbation Theory
M. Zdráhal${}^{a,b,}[email protected], J. Novotný${}^{a,}[email protected]
${}^{a}$Institute of Particle and Nuclear Physics, Faculty of Mathematics and Physics,
Charles University, V Holešovičkách 2, CZ-180
00 Prague 8, Czech Republic
${}^{b}$Faculty of Physics, University of Vienna,
Boltzmanngasse 5, A-1090 Vienna, Austria
Abstract
We generalise the reconstruction theorem of Stern, Sazdjian,
and Fuchs based on the dispersion relations to the case of the
($2\rightarrow 2$) scattering of all the pseudoscalar octet mesons
($\pi,K,\eta$). We formulate it in a general way and include also a discussion of the assumptions of the theorem.
It is used to obtain the amplitudes of all such processes in the isospin
limit to the one-loop order (and can be straightforwardly extended
to two loops) independently on the particular power-counting scheme
of the chiral perturbation theory in question. The results in this
general form are presented.
1 Introduction
In low-energy QCD the chiral perturbation theory ($\chi\textrm{PT}$,
cf. [1]) has gained great prominence as a tool for
description interactions of the lightest (pseudoscalar) mesons,
$\pi$, $K$ and $\eta$. Within this framework these mesons are
understood as Goldstone bosons of the spontaneous breaking of the
chiral symmetry $SU(3)_{L}\times SU(3)_{R}$ (appearing if all three
lightest quarks were massless) down to $SU(3)_{V}$, gaining the masses
through the explicit breaking of the symmetry due to the nonzero
quark masses, whereas all the other hadrons (with mass at least of order
$\Lambda\sim 1\,\mathrm{GeV}$) are included only effectively. Using
the symmetry properties and the basic properties like analyticity,
unitarity, and the crossing symmetry, the Lagrangian of this effective
theory is constructed. It contains an infinite number of terms.
Nevertheless, the Weinberg power-counting scheme assigns to a given
diagram of this theory (and term of the Lagrangian) its importance
by means of the chiral dimension $O\!\left(p^{D}\right)$ (see [2]), and so
for computations to a given order in low energies it is sufficient
to use only a finite number of them.
However, we can try to find the amplitudes of a process using the
required properties directly with no need of the Lagrangian to be
explicitly written (with all the advantages and disadvantages that this
involves), which is the main goal of this paper. We extend the
reconstruction theorem given by Stern, Sazdjian, and Fuchs in
[3] and widely used in [4]. Their work was originally
motivated by the problem333The standard power counting is
based on the presumption that the value of $B_{0}$ is of order $\Lambda$.
So there also appeared approaches extending the $\chi\textrm{PT}$ to the case
including the possibility of small $B_{0}$ too, which have a lesser
predictive power connected with larger number of LEC - namely the
generalised $\chi\textrm{PT}$ [5]. of possible smallness of
the scalar condensate $B_{0}$ as the power-counting independent way
to compute the amplitude of $\pi\pi$ scattering in the 2 flavour
case. Even though the problem is not yet satisfactorily
closed444The only direct answer given by experiment is still
the confirmation of the standard power counting in the 2 flavour
case by the $K_{l4}$ measurement at BNL [6]., the method is
interesting also in itself. The extension of this method to the 3
flavour case of the $\pi K$ amplitude has been given by
Ananthanarayan and Büttiker [7]. We have
worked out a generalisation to the processes of all the pseudo-Goldstone bosons (pGB) in the
three flavour case.
In contrast to these previous papers, we have extracted from the
theorem the entire information about the isospin structure and
assumed just the needed crossing symmetry. The isospin properties
are used afterwards separately for particular processes.
Furthermore, the theorem is formulated formally generally by
pointing out all the properties essential for the theorem. This
should simplify the discussion of validity of the theorem and its
eventual further extensions to more complicated cases. In the simple
case of processes from the pure strong $\chi\textrm{PT}$ (conserving the strong
isospin), we have discussed the assumptions and showed their
reasonability and concluded with the construction of their
amplitudes to two loops (with results explicitly written just to the
one loop) in a way independent of the particular power-counting scheme
in use. In this aspect our work is thus also an extension of the work of Osborn [8]
to higher orders.
The plan of the paper is as follows. In Sec. 2 we establish
our notation; the fundamental content of the paper is Sec. 3, where
we reformulate the reconstruction theorem with its assumptions and
give the proof of it. In Sec. 4 we use the theorem (and Appendix A) to compute the amplitudes describing all the 2 pGB
scattering processes to the $O\!\left(p^{4}\right)$ chiral order (explicitly written in Appendix B). Finally, in Sec. 5 we provide a further
discussion of the assumptions and discussion of the results.
Appendix A summarises all the needed amplitudes to describe the
4-pGB processes and lists a general parametrisation of them to the
$O\!\left(p^{2}\right)$ order. Appendix B gives the results of the application of
the theorem to computation of amplitudes to the $O\!\left(p^{4}\right)$ order.
In Appendix C we give relations between our parameters and
renormalised LEC of the standard $\chi\textrm{PT}$ [9] in
a particular renormalisation scheme [10]. The
main points of the analysis of validity of dispersion relations and
the regions of analyticity of amplitudes that are used in the
theorem are briefly summarised in Appendix D.
2 Notation
We consider a scattering process of two pGB of the type
$$A(p_{A})B(p_{B})\rightarrow C(p_{C})D(p_{D})$$
(1)
and write its amplitude according to555This process depends
on two independent kinematical quantities only. It is convenient to
choose the total energy squared $s$ and the angle $\theta$ between
the momenta $p_{A}$ and $p_{C}$ in the center of mass system (CMS) or
two of three Mandelstam variables $s,t,u$. In order to express the crossing and Bose symmetry properties
in a simple way in what follows we keep writing the dependence on all three Mandelstam variables explicitly.
$$\left\langle C(p_{C})D(p_{D})|_{f}A(p_{A})B(p_{B})\right\rangle_{i}=\delta_{if%
}+i(2\pi)^{4}\delta^{(4)}(p_{C}+p_{D}-p_{A}-p_{B})A(s,t,u).$$
(2)
In the relations where we want to distinguish between more processes,
we write explicitly $A_{i\rightarrow f}(s,t,u)$ or $A_{AB\rightarrow CD}(s,t,u)$.
Later on, we will use the crossing symmetry, so let us define the
amplitude of the crossed processes. The amplitude of the direct
process will be denoted by $S(s,t,u)$, i.e. $A_{AB\rightarrow CD}(s,t,u)=S(s,t,u)$. The amplitude of the $T$-crossed channel will be
$A_{A\overline{C}\rightarrow\overline{B}D}(s,t,u)=f_{T}T(s,t,u)$,
where the phase factor $f_{T}$ is defined so that $T(s,t,u)$ fulfils
the crossing relation $S(s,t,u)=T(t,s,u)$ and similarly for the
$U$-crossed channel.
The transition from the CMS variables $s$ and $\cos\theta$ to the
Mandelstam variables $s,t,u$ for the process (1) is
possible using the relation
$$\cos\theta=\frac{s(t-u)+\Delta_{\scriptscriptstyle AB}\Delta_{%
\scriptscriptstyle CD}}{\lambda_{AB}^{1/2}(s)\lambda_{CD}^{1/2}(s)}\,,$$
(3)
where $\Delta_{\scriptscriptstyle ij}$ is a difference of the masses squared $\Delta_{\scriptscriptstyle ij}=m_{i}^{2}-m_{j}^{2}$ and
$$\lambda_{ij}(x)=\lambda(x,m_{i}^{2},m_{j}^{2})=x^{2}+m_{i}^{4}+m_{j}^{4}-2x(m_%
{i}^{2}+m_{j}^{2})-2m_{i}^{2}m_{j}^{2}$$
(4)
refers to the Källen’s quadratic form (also called the triangle
function).
The sum of squared masses of all the particles appearing in the
process is denoted as $\mathcal{M}$, i.e.
$$s+t+u=m_{A}^{2}+m_{B}^{2}+m_{C}^{2}+m_{D}^{2}=\mathcal{M}.$$
(5)
We will use the partial wave decomposition of the amplitudes in the form
$$A_{i\rightarrow f}(s,\cos\theta)=32\pi\sum_{l}A_{l}^{i\rightarrow f}(s)(2l+1)P%
_{l}(\cos\theta),$$
(6)
where $P_{l}$ are Legendre polynomials.
From the unitarity of S-matrix (and real analyticity of the physical
amplitude as well as $T$-invariance) there follows the unitarity
relations. Assuming that the only relevant intermediate states are those containing two
particles $\alpha$ and $\beta$ with masses $m_{\alpha}$ and $m_{\beta}$, its form for the partial wave of
amplitudes reads
$$\operatorname{Im}A_{l}^{i\rightarrow f}(s)=\sum_{\alpha,\beta}\frac{2}{S_{F}}%
\,\frac{\lambda^{1/2}_{\alpha\beta}(s)}{s}A_{l}^{i\rightarrow(\alpha,\beta)}(s%
)\left[A_{l}^{f\rightarrow(\alpha,\beta)}(s)\right]^{*}\theta(s-(m_{\alpha}+m_%
{\beta})^{2}).$$
(7)
Here, $S_{F}$ denotes the symmetry factor ($S_{F}=1$ if the two intermediate
states are distinguishable and $S_{F}=2$ if they are not).
Finally, we use the definition of the (signs of) meson fields given in the matrix
$$\begin{split}\displaystyle\phi(x)=\lambda_{a}\phi_{a}(x)&\displaystyle=\begin{%
pmatrix}\pi^{0}+\frac{1}{\sqrt{3}}\eta&-\sqrt{2}\pi^{+}&-\sqrt{2}K^{+}\\
\sqrt{2}\pi^{-}&-\pi^{0}+\frac{1}{\sqrt{3}}\eta&-\sqrt{2}K^{0}\\
\sqrt{2}K^{-}&-\sqrt{2}\overline{K^{0}}&-\frac{2}{\sqrt{3}}\eta\end{pmatrix}.%
\end{split}$$
(8)
The meson states are chosen so that the matrix element of mesons
gets one minus sign for each charged particle in the final state.
3 Reconstruction theorem
3.1 Statement of the theorem
In this section we prove the following theorem: The amplitude
$S(s,t;u)$ of a given process $A+B\rightarrow C+D$ fulfilling all
the conditions from the next subsection can be reconstructed to (and excluding)
$O\!\left(p^{8}\right)$ order just from the knowledge of imaginary parts of $s$ and
$p$ partial waves of all crossed amplitudes and some polynomial
$$\begin{split}\displaystyle S(s,t;u)&\displaystyle=R_{4}(s,t;u)+\Phi_{0}(s)+%
\left[s(t-u)+\Delta_{\scriptscriptstyle AB}\Delta_{\scriptscriptstyle CD}%
\right]\Phi_{1}(s)+\Psi_{0}(t)\\
&\displaystyle\quad+\left[t(s-u)+\Delta_{\scriptscriptstyle AC}\Delta_{%
\scriptscriptstyle BD}\right]\Psi_{1}(t)+\Omega_{0}(u)+\left[u(t-s)+\Delta_{%
\scriptscriptstyle AD}\Delta_{\scriptscriptstyle BC}\right]\Omega_{1}(u)+O\!%
\left(p^{8}\right),\end{split}$$
(9)
where $R_{4}(s,t;u)$ is a third-order polynomial in Mandelstam
variables (obeying exactly the same symmetries as the original
amplitude $S(s,t;u)$) and
$$\displaystyle\Phi_{0}(s)$$
$$\displaystyle=32s^{3}\int_{\Sigma}^{\Lambda^{2}}\frac{dx}{x^{3}}\frac{%
\operatorname{Im}S_{0}(x)}{x-s}\,,$$
(10)
$$\displaystyle\Phi_{1}(s)$$
$$\displaystyle=96s^{3}\int_{\Sigma}^{\Lambda^{2}}\frac{dx}{x^{3}}\frac{1}{x-s}%
\operatorname{Im}\frac{S_{1}(x)}{\lambda_{AB}^{1/2}(x)\lambda_{CD}^{1/2}(x)}\,,$$
(11)
$$\displaystyle\Psi_{0}(t)$$
$$\displaystyle=32t^{3}\int_{\tau}^{\Lambda^{2}}\frac{dx}{x^{3}}\frac{%
\operatorname{Im}T_{0}(x)}{x-t}\,,$$
(12)
$$\displaystyle\Psi_{1}(t)$$
$$\displaystyle=96t^{3}\int_{\tau}^{\Lambda^{2}}\frac{dx}{x^{3}}\frac{1}{x-t}%
\operatorname{Im}\frac{T_{1}(x)}{\lambda_{AC}^{1/2}(x)\lambda_{BD}^{1/2}(x)}\,,$$
(13)
$$\displaystyle\Omega_{0}(u)$$
$$\displaystyle=32u^{3}\int_{\Upsilon}^{\Lambda^{2}}\frac{dx}{x^{3}}\frac{%
\operatorname{Im}U_{0}(x)}{x-u}\,,$$
(14)
$$\displaystyle\Omega_{1}(u)$$
$$\displaystyle=96u^{3}\int_{\Upsilon}^{\Lambda^{2}}\frac{dx}{x^{3}}\frac{1}{x-u%
}\operatorname{Im}\frac{U_{1}(x)}{\lambda_{AD}^{1/2}(x)\lambda_{BC}^{1/2}(x)}\,,$$
(15)
$S_{0}$ and $S_{1}$ are $s$ and $p$ partial wave of the process in
question, similarly $T_{0}$ and $T_{1}$ and $U_{0}$ and $U_{1}$ the partial
waves of the $T$ and $U$ crossed processes. $\Sigma,\tau,\Upsilon$
are the minima of squared invariant mass of all possible
intermediate states in the $S$, $T$ and $U$ channels (see (18)).
Finally, $\Lambda$ is the scale from the assumptions.
3.2 Assumptions of the theorem
We use this theorem in the (strong part of) chiral perturbation
theory, but for better understanding of it, we are formulating all
the assumptions in the general way and then discussing their
validity in $\chi\textrm{PT}$ separately. The assumptions of the theorem are:
1.
There exists a threshold $\Lambda$ up to which we can regard the
theory under consideration to be complete; this means, for example, that
under this threshold there appears no particle of other types than
those which are already explicitly included in the theory and all the
influence of such extra particles is already taken into account effectively
or negligible under the threshold. In chiral perturbation theory the
threshold $\Lambda$ is the threshold of production of non-pGB
states, i.e. $\Lambda\sim 1\,\mathrm{GeV}$.
2.
There exists a well-behaved expansion of the amplitude in
powers of $p/\Lambda$. In $\chi\textrm{PT}$ it is enough to have any
chiral expansion, not necessarily the standard Weinberg one. As is
common in $\chi\textrm{PT}$, instead of $O\!\left((p/\Lambda)^{n}\right)$ we write just $O\!\left(p^{n}\right)$ for
short.
3.
We can write a three times666As discussed later on, the exact number of subtractions is not so relevant. subtracted dispersion relation for
the amplitude $S(s,t;u)$ in the complex $s$ plane for fixed value of
$u$ (in the form (16)). This is the essential
assumption of the proof - the theorem stands or falls by it. This assumption is also connected with the validity of the
partial wave expansion (for the sake of simplicity, we assume
that this expansion can be also analytically continued below the
physical threshold where needed). The legitimacy of these
assumptions in the case of $\chi\textrm{PT}$ is discussed in
Appendix D.
4.
We further assume that the amplitude considered as a function of a single Mandelstam variable (with the other fixed at some appropriate value) is analytic in some unempty open region. The regions where this and the previous assumptions are valid are discussed (and plotted) in Appendix D.
5.
The crossing relations given in the form of the previous section are
valid, i.e. $S(s,t,u)=T(t,s,u)=U(u,t,s)$. That is a widely accepted assumption, whose validity for a
general $2\rightarrow 2$ process has been proven from axiomatic theory by Bross, Epstein, and Glaser [11].
6.
The absorptive parts of the $l\geq 2$ partial amplitudes are suppressed to
the $O\!\left(p^{8}\right)$ order - thanks to that we can (up to this order) deal just
with the first two partial waves in the theorem. In our case that
follows from the (pseudo-)Goldstone boson character of the particles
under consideration.
In $\chi\textrm{PT}$ the amplitudes behave dominantly as $O\!\left(p^{2}\right)$ and do not
contain any bound state poles. The unitarity relation
(7) then implies that its imaginary part behaves as
$O\!\left(p^{4}\right)$, i.e. the amplitudes are dominantly real. According to the
analyticity of the amplitude, its leading $O\!\left(p^{2}\right)$ part should be a
polynomial in Mandelstam variables. Moreover, it has to be a
polynomial of the first order; otherwise its coefficient would grow
up as mass of the pGB went to zero, and that would contradict the
finiteness of the S-matrix in the chiral limit with the external
momenta fixed777We will see that the finiteness of the
S-matrix in the chiral limit deserves to be regarded as one of the
full-valued assumptions by itself.. However, the first-order
polynomial could not contribute to $l\geq 2$ partial waves, thus these
partial waves should behave in the chiral limit as $O\!\left(p^{4}\right)$, thereby
again using the unitarity relation, the imaginary parts of these partial
waves are at least of the $O\!\left(p^{8}\right)$ order.
3.3 Proof of the theorem
One of our assumptions is that for the amplitude we can write the
three times subtracted dispersion relation in the form:
$$\displaystyle S(s,t;u)=P_{3}(s,t;u)+\frac{s^{3}}{\pi}\int_{\Sigma}^{\infty}%
\frac{dx}{x^{3}}\frac{\operatorname{Im}S(x,\mathcal{M}-x-u;u)}{x-s}\\
\displaystyle+\frac{t^{3}}{\pi}\int_{\tau}^{\infty}\frac{dx}{x^{3}}\frac{%
\operatorname{Im}S(\mathcal{M}-x-u,x;u)}{x-t}\,,$$
(16)
where $P_{n}(s,t;u)$ is a polynomial of the $(n-1)$-th order in $s$
and $t$.888The coefficients of this polynomial depend on $u$. Thanks to the relation for the sum of the
Mandelstam variables, such polynomials could be written in the form
$P_{n}(s,t;u)=\alpha(u)+\beta(u)(s-t)+\gamma(u)(s-t)^{2}+\dots+\omega(u)(s-t)^{%
n-1}.$
(17)
$\Sigma$ is the minimum of the squared invariant mass of all the
possible intermediate states $(\alpha,\beta)$ in the S
channel (or rather the lowest mass
of the state with the same quantum numbers as the in and
out state in the S-channel)999In this relation we anticipate again that we will
consider the problems under conditions implying relevance of the
two-particle intermediate states only.,
$$\Sigma=\min_{(\alpha,\beta)}(m_{\alpha}+m_{\beta})^{2},$$
(18)
and analogically $\tau$ for the $T$ channel (later on we will also need
the same for the $U$ channel, what will be denoted by $\Upsilon$).
Thanks to the crossing symmetry we can replace the $S(s,t;u)$
amplitude in the second integral with the $T(t,s;u)$. Further, since
we assume that we know the complete theory only up to some threshold
$\Lambda$, we have to split the dispersion integral into two parts -
the low-energy $(x\leq\Lambda^{2})$ and the high-energy
$(x\geq\Lambda^{2})$ part. We consider the region where
$s\ll\Lambda^{2}$, i.e. in the high-energy part, $s\ll x$, we
have101010The fact that the remainder $\sum_{n\geq 4}s^{n}R_{n}(u)$
is of order at least $O\!\left(p^{8}\right)$ follows again from the finiteness of the
amplitude in the chiral limit (the importance of this assumption
stems from here) and the crossing symmetry.
$$\frac{s^{3}}{\pi}\int_{\Lambda^{2}}^{\infty}\frac{dx}{x^{3}}\left.\frac{%
\operatorname{Im}S(x,\mathcal{M}-x-u;u)}{x-s}\right|_{s<<x}=s^{3}H_{\Lambda}(u%
)+O\!\left(p^{8}\right),$$
(19)
where
$$H_{\Lambda}(u)=\frac{1}{\pi}\int_{\Lambda^{2}}^{\infty}\frac{dx}{x^{3}}\frac{%
\operatorname{Im}S(x,\mathcal{M}-x-u;u)}{x}$$
(20)
is a function of $u$ only. We can proceed similarly for the high-energy part of
the integration of $T(s,t;u)$. Hence, these high-energy parts can be
to the order $O\!\left(p^{8}\right)$ added to the subtraction polynomial now thereby
of the third order111111The consequence of that is also that we
can formally extend the validity of the original theory even beyond
threshold $\Lambda$ and compute the dispersion integral up to
infinity, pretending that the original theory is really complete
(naturally with the appropriate modification of the third-order
polynomial). Similar arguments would appear even if we introduced
the cutoff $\Lambda$ by other reasons, e.g. by reason of numerical
integration.. Afterwards, the dispersion relation is of the form
$$\displaystyle S(s,t;u)=P_{4}(s,t;u)+\frac{s^{3}}{\pi}\int_{\Sigma}^{\Lambda^{2%
}}\frac{dx}{x^{3}}\frac{\operatorname{Im}S(x,\mathcal{M}-x-u;u)}{x-s}\\
\displaystyle+\frac{t^{3}}{\pi}\int_{\tau}^{\Lambda^{2}}\frac{dx}{x^{3}}\frac{%
\operatorname{Im}T(x,\mathcal{M}-x-u;u)}{x-t}+O\!\left(p^{8}\right).$$
(21)
We decompose the (imaginary parts of) amplitudes into the partial
waves
$$\begin{split}\displaystyle S(s,t;u)&\displaystyle=32\pi\left(S_{0}(s)+3\cos%
\theta S_{1}(s)+S_{l\geq 2}(s,t;u)\right)\\
&\displaystyle=32\pi\left(S_{0}(s)+3\frac{s(t-u)+\Delta_{\scriptscriptstyle AB%
}\Delta_{\scriptscriptstyle CD}}{\lambda_{AB}^{1/2}(s)\lambda_{CD}^{1/2}(s)}S_%
{1}(s)+S_{l\geq 2}(s,t;u)\right).\end{split}$$
(22)
The terms $S_{l\geq 2}(s,t;u)$ incorporate contributions of all the
higher $(l\geq 2)$ partial waves and are suppressed to $O\!\left(p^{8}\right)$ according
to our assumptions.
From the integrals of the P partial waves, we can extract functions
depending just on $u$ and include them in the polynomial, e.g.
$$\displaystyle\int_{\Sigma}^{\Lambda^{2}}\frac{dx}{x^{3}}\frac{x(s+t-x-u)}{x-s}%
\operatorname{Im}\frac{S_{1}(x)}{\lambda_{AB}^{1/2}(x)\lambda_{CD}^{1/2}(x)}\\
\displaystyle=\int_{\Sigma}^{\Lambda^{2}}\frac{dx}{x^{3}}\frac{x(t-u)}{x-s}%
\operatorname{Im}\frac{S_{1}(x)}{\lambda_{AB}^{1/2}(x)\lambda_{CD}^{1/2}(x)}-%
\int_{\Sigma}^{\Lambda^{2}}\frac{dx}{x^{2}}\operatorname{Im}\frac{S_{1}(x)}{%
\lambda_{AB}^{1/2}(x)\lambda_{CD}^{1/2}(x)}.$$
(23)
Finally, we use this rearrangement of the fractions
$$\frac{x(t-u)}{x-s}=\frac{s(t-u)}{x-s}+t-u,$$
(24)
where the second term gives after integration a contribution in the
form $t-u$ times a function of $u$ only and therefore as a
polynomial in $t$ can also be included into the subtraction
polynomial $P_{4}$.
After that, the amplitude reads
$$\displaystyle S(s,t;u)=P_{4}(s,t;u)+\Phi_{0}(s)+[s(t-u)+\Delta_{%
\scriptscriptstyle AB}\Delta_{\scriptscriptstyle CD}]\Phi_{1}(s)\\
\displaystyle+\Psi_{0}(t)+[t(s-u)+\Delta_{\scriptscriptstyle AC}\Delta_{%
\scriptscriptstyle BD}]\Psi_{1}(t)+O\!\left(p^{8}\right),$$
(25)
where $\Phi_{0}(s),\Phi_{1}(s),\Psi_{0}(t)$, and $\Psi_{1}(t)$ are given by (10) – (13).
In order to implement the $s\leftrightarrow u$ crossing symmetry, we
add and subtract the following polynomial in $s$ and $t$ (with
coefficients depending on $u$)
$$\Omega_{0}(u)+[u(t-s)+\Delta_{\scriptscriptstyle AD}\Delta_{\scriptscriptstyle
BC%
}]\Omega_{1}(u),$$
(26)
with $\Omega_{0}(u)$ and $\Omega_{1}(u)$ given by (14) and
(15) and we include the subtraction of this polynomial in
the polynomial $P_{4}(s,t;u)$ and thereby get the amplitude in the
symmetric form (9).
Until now, we have proved the whole theorem except for the
$u$-dependence of the polynomial $P_{4}(s,t;u)$. However, we could write similar
relations also for the amplitudes $T(s,t;u)$ and $U(s,t;u)$ with the
polynomials in $(s,u)$ and the coefficients depending on variable
$t$. Thanks to the crossing symmetry and to the symmetric form of
the rest of the formula, these polynomials should obey the same
crossing symmetry. From this and from the analyticity of the
amplitudes in a neighborhood of $u$ for some fixed $s$ in our
kinematic region, we can conclude that $P_{4}(s,t;u)$ are
polynomials in respect to all Mandelstam variables.
4 Application of the theorem to the 4-meson processes to the $O\!\left(p^{4}\right)$ order
As we have already outlined, in the case of $2\rightarrow 2$
scattering processes of pseudoscalar octet mesons, we can use the
theorem and the unitarity relation to a selfconsistent construction
of all those amplitudes to (and excluding) $O\!\left(p^{8}\right)$ order. We need just a
parametrisation of $O\!\left(p^{2}\right)$ amplitudes as its input.
In the following we will consider the strong isospin conservation
limit121212Recall that in this case the validity of the required
dispersion relations can be proven directly from the axiomatic theory as discussed in Appendix D., where
Ward identities imply that there are just 7 independent processes.
From them the amplitudes of all the possible physical processes can
be determined as summarised in Appendix A. Since we want to keep the
independence on the specific power-counting scheme, we have
parameterised the $O\!\left(p^{2}\right)$ amplitudes in the most general form obeying
their symmetries (crossing, Bose, and isospin) - given also in Appendix A. Inserting them into the
unitarity relations (7) (assuming that we have
amplitudes of all the needed intermediate processes), we get the
imaginary parts to $O\!\left(p^{4}\right)$ order, which can be installed into the
theorem and we receive thereby the amplitudes in that order. With
the same procedure (though more technically involved), we can proceed
with the second iteration and obtain thereby the amplitudes to (and excluding)
$O\!\left(p^{8}\right)$ order. For the sake of simplicity, in this paper we confine
ourselves to the first iteration and thus to the results to $O\!\left(p^{4}\right)$
order.
It remains to show that we already have the amplitudes of all the
relevant intermediate processes, i.e. the self-consistency of
this procedure. We are concerned with the processes to (and excluding) $O\!\left(p^{8}\right)$
chiral order and (in accordance with the first assumption of the
theorem) in the region bellow the appearance of all non-Goldstone
particles and thereby away from poles and cuts of such intermediate
states. Therefore, their effect is included just in the polynomial
of the reconstruction theorem and in the higher orders. The
intermediate states with an odd number of Goldstone bosons are
forbidden by the even intrinsic parity of the effective Lagrangian -
because the effect of the axial anomaly enters at the $O\!\left(p^{4}\right)$ order
(Wess, Zumino and Witten [12]), such intermediate states
induce the contribution of the order at least $O\!\left(p^{8}\right)$. Furthermore, the contribution
of the states with more than two Goldstone bosons is also suppressed
to $O\!\left(p^{8}\right)$ order since the $n$-Goldstone-boson invariant phase space
scales like $p^{2n-4}$ and amplitudes with an arbitrary number of
external Goldstone boson legs behave dominantly as $O\!\left(p^{2}\right)$.
Consequently, the contribution of such ($n>2$) intermediate states
to the imaginary part of the amplitude is according to the unitarity
relation at least $O\!\left(p^{8}\right)$. Thus, we can consider the
two-Goldstone-boson intermediate states only.
Further simplification in the $O\!\left(p^{4}\right)$ order computation appears. Using
again the arguments of finiteness of S-matrix in the chiral limit
with the external momenta fixed, the coefficients of the polynomial
in the theorem should be at least $O\!\left(p^{0}\right)$, and so the polynomial in the
$O\!\left(p^{4}\right)$ amplitude is maximally of the second order in Mandelstam
variables.
4.1 The application schematically
The $O\!\left(p^{2}\right)$ amplitudes can be written (cf. with (17), and Appendix A with use of (5)) in the form131313Let us remind that in the $O\!\left(p^{2}\right)$ case $\alpha(s)=\alpha_{0}+\alpha_{1}s$, i.e. it is just the first order polynomial in $s$.
$$A=\frac{1}{F_{\pi}^{2}}\left(\alpha(s)+\beta(t-u)\right).$$
(27)
The decomposition into the partial waves can be obtained using (3) and (22)
$$\displaystyle A_{0}$$
$$\displaystyle=\frac{1}{32\pi}\frac{1}{F_{\pi}^{2}}\left(\alpha(s)-\beta\frac{%
\Delta_{\scriptscriptstyle AB}\Delta_{\scriptscriptstyle CD}}{s}\right),$$
(28)
$$\displaystyle A_{1}$$
$$\displaystyle=\frac{1}{32\pi}\frac{1}{3F_{\pi}^{2}}\beta\frac{\lambda^{1/2}_{%
AB}(s)\lambda^{1/2}_{CD}(s)}{s}\,.$$
(29)
The (right) discontinuity of our function $S_{0}$ and $S_{1}$ of the theorem, given by the unitarity relation,
is therefore in the $O\!\left(p^{4}\right)$ case very simple:
$$\displaystyle\begin{split}\displaystyle\operatorname{Im}S_{0}(s)&\displaystyle%
=\sum_{\begin{subarray}{c}\scriptscriptstyle k-\text{interm.}\\
\scriptscriptstyle\text{states}\end{subarray}}\frac{2}{S_{F_{k}}}\frac{\lambda%
_{k}^{1/2}(s)}{s}\left(\frac{1}{32\pi}\right)^{2}\frac{1}{F_{\pi}^{4}}\left[%
\left(\alpha_{ik}(s)-\beta_{ik}\frac{\Delta_{\scriptscriptstyle AB}\Delta_{k}}%
{s}\right)\left(\alpha_{kf}(s)-\beta_{kf}\frac{\Delta_{k}\Delta_{%
\scriptscriptstyle CD}}{s}\right)\right]\\
&\displaystyle\qquad\qquad\qquad\times\theta(s-s^{t}_{k}),\end{split}$$
(30)
$$\displaystyle\operatorname{Im}S_{1}(s)$$
$$\displaystyle=\sum_{\begin{subarray}{c}\scriptscriptstyle k-\text{interm.}\\
\scriptscriptstyle\text{states}\end{subarray}}\frac{2}{S_{F_{k}}}\frac{\lambda%
_{k}^{1/2}(s)}{s}\left(\frac{1}{32\pi}\right)^{2}\frac{1}{9F_{\pi}^{4}}\bigg{[%
}\beta_{ik}\frac{\lambda^{1/2}_{AB}(s)\lambda^{1/2}_{k}(s)}{s}\beta_{kf}\frac{%
\lambda^{1/2}_{k}(s)\lambda^{1/2}_{CD}(s)}{s}\bigg{]}\theta(s-s^{t}_{k}),$$
(31)
where the sums go over all the possible (two-pGB) intermediate states $k$ with its symmetry factor $S_{F_{k}}$ and all the objects with lower index $k$ are the quantities relating to this intermediate state, similarly indices $i$ and $f$ refer to initial and final state. Finally $s^{t}_{k}$ is a threshold of this intermediate state.
Thanks to the behaviour of square root of the triangle function, $\lambda^{1/2}(s)\rightarrow s$ for $s\rightarrow\infty$ and the fact that $\alpha(s)$ is the polynomial maximally of the first order in $s$, we know that not all the terms (in the brackets) need all the subtractions to give finite integrals (10) – (15) (e.g. the parts with negative power of $s$ in the polynomial in the brackets do not even need any subtraction). Any subtraction redundant in this aspect then gives just a polynomial, which can be included into the polynomial $R_{4}(s,t,t)$ of the theorem (9) and into the contributions of higher orders. Furthermore, if the integrals rising on both sides of rearrangement (24) have a good mathematical sense, we can pull out the polynomial in front of the integral again with possible change of the polynomial $R_{4}(s,t,t)$. Therefore, applying the reconstruction theorem, we can write $\Phi_{0}$ and $\Phi_{1}$ in the minimal141414With the minimal number of subtraction needed to get a finite nonpolynomial part, i.e. here with one or two subtractions. form (except for the polynomial included to $R_{4}(s,t,t)$):
$$\displaystyle\Phi_{0}(s)=\frac{1}{16\pi^{2}}\frac{1}{F_{\pi}^{4}}\sum_{k}\frac%
{1}{S_{F_{k}}}\Bigg{(}\alpha_{ik}(s)\alpha_{kf}(s)s\int_{s^{t}_{k}}^{\infty}%
\frac{dx}{x}\frac{1}{x-s}\frac{\lambda_{k}^{1/2}(x)}{x}\\
\displaystyle-\big{(}\alpha_{kf}(s)\beta_{ik}\Delta_{\scriptscriptstyle AB}+%
\alpha_{ik}(s)\beta_{kf}\Delta_{\scriptscriptstyle CD}\big{)}\Delta_{k}\int_{s%
^{t}_{k}}^{\infty}\frac{dx}{x}\frac{1}{x-s}\frac{\lambda_{k}^{1/2}(x)}{x}\\
\displaystyle+\beta_{ik}\beta_{kf}\Delta_{k}^{2}\Delta_{\scriptscriptstyle AB}%
\Delta_{\scriptscriptstyle CD}\int_{s_{k}^{t}}^{\infty}\frac{dx}{x^{2}}\frac{1%
}{x-s}\frac{\lambda_{k}^{1/2}(x)}{x}\Bigg{)}.$$
(32)
Similarly for the P-wave function $\Phi_{1}$ (in addition we have used $\lambda_{k}(x)=x^{2}-2x\Sigma_{k}+\Delta_{k}^{2}$, where $\Sigma_{k}$ is the sum of the second powers of masses of the two particles in the intermediate state $k$):
$$\Phi_{1}(s)=\frac{1}{16\pi^{2}}\frac{1}{3F_{\pi}^{4}}\sum_{k}\frac{1}{S_{F_{k}%
}}\beta_{ik}\beta_{kf}\Bigg{(}\big{(}s-2\Sigma_{k}\big{)}\int_{s^{t}_{k}}^{%
\infty}\frac{dx}{x}\frac{1}{x-s}\frac{\lambda_{k}^{1/2}(x)}{x}+\Delta_{k}^{2}%
\int_{s^{t}_{k}}^{\infty}\frac{dx}{x^{2}}\frac{1}{x-s}\frac{\lambda_{k}^{1/2}(%
x)}{x}\Bigg{)}.$$
(33)
In the integrals we recognize the once and twice subtracted dispersion integrals from Appendix B ((68) and (69)), i.e.
$$\displaystyle\begin{split}\displaystyle\Phi_{0}(s)&\displaystyle=\frac{1}{F_{%
\pi}^{4}}\sum_{k}\frac{1}{S_{F_{k}}}\Bigg{(}\beta_{ik}\beta_{kf}\frac{\Delta_{%
k}^{2}}{s^{2}}\Delta_{\scriptscriptstyle AB}\Delta_{\scriptscriptstyle CD}%
\overline{\overline{J}}_{k}(s)\\
&\displaystyle\qquad+\bigg{(}\alpha_{ik}(s)\alpha_{kf}(s)-(\alpha_{kf}(s)\beta%
_{ik}\Delta_{\scriptscriptstyle AB}+\alpha_{ik}(s)\beta_{kf}\Delta_{%
\scriptscriptstyle CD})\frac{\Delta_{\scriptscriptstyle k}}{s}\bigg{)}%
\overline{J}_{k}(s)\Bigg{)},\end{split}$$
(34)
$$\displaystyle\Phi_{1}(s)$$
$$\displaystyle=\frac{1}{3F_{\pi}^{4}}\sum_{k}\frac{1}{S_{F_{k}}}\beta_{ik}\beta%
_{kf}\bigg{(}\bigg{(}1-2\frac{\Sigma_{k}}{s}\bigg{)}\overline{J}_{k}(s)+\frac{%
\Delta_{k}^{2}}{s^{2}}\overline{\overline{J}}_{k}(s)\bigg{)}.$$
(35)
Using the change of the polynomial for the last time (in the replacement of $\frac{\overline{\overline{J}}(s)}{s}$ with $\frac{\overline{J}(s)}{s}$), we have the result:
$$\begin{split}&\displaystyle\Phi_{0}(s)+[s(t-u)+\Delta_{\scriptscriptstyle AB}%
\Delta_{\scriptscriptstyle CD}]\Phi_{1}(s)\\
&\displaystyle\quad=\frac{1}{3F_{\pi}^{4}}\sum_{k}\frac{1}{S_{F_{k}}}\Bigg{(}4%
\Delta_{\scriptscriptstyle AB}\Delta_{\scriptscriptstyle CD}\beta_{ik}\beta_{%
kf}\Delta_{k}^{2}\frac{\overline{\overline{J}}_{k}(s)}{s^{2}}\\
&\displaystyle\qquad+\bigg{(}\beta_{ik}\beta_{kf}\big{(}(t-u)\Delta_{k}^{2}-2%
\Delta_{\scriptscriptstyle AB}\Delta_{\scriptscriptstyle CD}\Sigma_{k}\big{)}-%
3(\alpha_{kf}(s)\beta_{ik}\Delta_{\scriptscriptstyle AB}+\alpha_{ik}(s)\beta_{%
kf}\Delta_{\scriptscriptstyle CD})\Delta_{k}\bigg{)}\frac{\overline{J}_{k}(s)}%
{s}\\
&\displaystyle\qquad+\bigg{(}3\alpha_{ik}(s)\alpha_{kf}(s)+\beta_{ik}\beta_{kf%
}\big{(}(t-u)(s-2\Sigma_{k})+\Delta_{\scriptscriptstyle AB}\Delta_{%
\scriptscriptstyle CD}\big{)}\bigg{)}\overline{J}_{k}(s)\Bigg{)}.\end{split}$$
(36)
We get similar results also for the $T$- and $U$-crossed parts151515If we carry out the same extraction of polynomial also for them, we get again these results by the change $s\leftrightarrow t$ respectively $s\leftrightarrow u$ and appropriate permutation in ABCD and then the symmetries of the polynomial of theorem remain the same.. Final results of this procedure for all the amplitudes are given in Appendix B.
5 Summary and discussions
In this paper we have worked out a generalisation of the dispersive
analysis from [3] to the processes involving all the light
octet mesons. We have formulated the reconstruction theorem for
their amplitudes free of the particular isospin structure of the
processes (which is used afterwards for the individual processes
separately) and have provided (an outline of) its proof. We have
also attempted to point out exactly the properties of the theory
which are needed for its proof mainly due to better understanding of
it as well as of its further generalisation on more complicated
cases (as below). In our particular case we have succeeded in
justifying most of them practically from the first principles of
axiomatic quantum field theory - at least in the regions from Fig. 1.
The amplitudes of 4-octet-meson processes (on condition of the
strong isospin conservation) form a selfconsistent system with
respect to the theorem and the unitarity relations, thanks to the fact that
we can simply construct these amplitudes to (and excluding161616Since we also want to include the power countings where the contributions of odd chiral orders appear (like generalised chiral power counting), we write instead of valid to $O\!\left(p^{6}\right)$ order (what will be case if odd orders do not appear) or to $O\!\left(p^{7}\right)$ order (where the odd orders are included) just valid to and excluding $O\!\left(p^{8}\right)$ order to avoid misunderstanding. We should also remind that in these odd chiral order power countings, the one-loop result is not $O\!\left(p^{4}\right)$ but $O\!\left(p^{5}\right)$ – we can get such a result by simple modification of the results presented keeping the $\beta_{\pi\eta}$ and $\gamma_{KK}$ parameters also in the intermediate processes.) $O\!\left(p^{8}\right)$ order (in this
work explicitly computed only to $O\!\left(p^{4}\right)$ order) independently on the
specific power-counting model in use. The amplitudes for the
particular model can be obtained with the specific choice of the parameters. For example, we can compare the nonpolynomial part of our results (with the standard choice of $O\!\left(p^{2}\right)$ constants from Appendix A) with that part of the standard chiral perturbation results (given by [10]) -- we see an agreement except for the different sign convention171717When
comparing, one should pay attention to their caption of the 4-kaon
process ${\overline{K^{0}}}K^{0}\rightarrow K^{+}K^{-}$, which gives us the false
illusion about the value of the Mandelstam variables, whereas the
right value would be better evoked by the caption ${\overline{K^{0}}}K^{0}\rightarrow K^{-}K^{+}$.. We can also use their result to get a relation between our $O\!\left(p^{4}\right)$ polynomial parameters and low-energy constants (LECs)
of the standard chiral perturbation theory in the particular renormalisation scheme used therein just by comparison of the polynomial parts, as is discussed and given in Appendix C.181818If we used the $O\!\left(p^{4}\right)$ parameters that has been obtained thereby also in the unitarity part, we would be already counting a part of the $O\!\left(p^{6}\right)$ corrections. It is worth mentioning that by subtracting the center of the Mandelstam triangle in the second-order polynomial part of our results, the parameters of the first-order polynomial, i.e. $\alpha$’s, $\beta$’s and $\gamma$’s, do not depend in the subtraction scheme used by [10] in $O\!\left(p^{4}\right)$ order on the LECs of operators with four derivatives from $\mathcal{L}_{4}$ in [9], i.e. $L_{1}^{r}$, $L_{2}^{r}$, and $L_{3}$.
Some of the computed amplitudes ($KK\rightarrow$ anything, processes
involving $\eta$’s) have not been computed in such a general form
yet. However, this pioneerness is connected with one problem of the
assumptions of the theorem. We have assumed the existence of a
threshold separating the processes containing just those 8 mesons
from the processes in which particles of other types (e.g. resonances) appear, and then we focus on a kinematical range far
below that threshold. However, in the case of the processes with the
exception of $\pi\pi\rightarrow\pi\pi$ and $\pi K\rightarrow\pi K$,
this range is very narrow or does not even exist (e.g. in the
process $KK\rightarrow KK$ with a kinematical threshold of
$990\,\mathrm{MeV}$ there can appear the $\rho$ resonance with its
mass of $770\,\mathrm{MeV}$). The significance of our result thereby
decreases. Nevertheless, beside the elegance of our construction due
to its intrinsic self-consistency, our results can become useful for
a check of complicated models or simulations, where one can separate
(or completely turn off) the effects of these resonances. (The
advantage is that even in the most general case in (to and excluding) $O\!\left(p^{8}\right)$ order, there
is a small number ($=47$) of parameters, coupling different processes
together.)
To that end it would be naturally preferable to perform also the
second iteration and get thereby the whole information achievable
from the reconstruction theorem. As has been already mentioned, the
completion of the second iteration is conceptually straightforward; however, there appear certain technical complications which are connected with the computation of partial waves of
amplitudes, where we need to compute integrals of the type $\int t^{n}\overline{J}(t)dt$ and then to “dispersively integrate” them. Because of the appearance of the three different masses, the results are much more complicated than in the 2 flavour $\pi\pi$ case (with only one mass) and so will be presented with all the technical details in a separate paper.
The careful reader has surely noticed that in the relation
(21) or (11) and the other similar (in
opposition to the verbal formulation of the theorem), we have
emphasised that the required imaginary part is not of the partial
wave but of the partial wave divided by
$\lambda_{AB}^{1/2}\lambda_{CD}^{1/2}$, i.e. in principle in the higher order there
could appear a problem under the physical threshold of the
amplitudes where $\lambda_{AB}^{1/2}\lambda_{CD}^{1/2}$ could be
imaginary. There presents itself a question whether we can obtain
this quantity also from the unitarity relations. But we should
remind that the origin of the relation (7) where the
imaginary part of the partial wave times the relevant Legendre
polynomial also occurs. So the only question is whether we can
believe the analytical continuation of the partial wave
decomposition and of the unitarity relation under the physical
threshold. If we do so, then the unitarity relation gives exactly
what we need to the theorem, i.e. what we call the imaginary part of
partial wave for simplicity.
The further important question is of the number of subtractions in
the dispersive relations. Jin and Martin [13] have shown
that thanks to the Froissart bound [14] it suffices to
consider just two subtractions. However, the Froissart bound deals
with the complete theory (full QCD), and in an effective theory one
does not know (or more exactly one does not deal with) what is above
$\Lambda$, and so the situation can occur that more than two subtractions are needed. In other words Froissart tells us two subtractions are
sufficient if we supply the dispersive integral above $\Lambda$ with
the complete theory. How to deal with this unknown, we have
already seen near the relation (19) – for $n$
subtractions we get a series of infinite order beginning with
the $n$-th power of Mandelstam variable. Fortunately, we have shown
that the terms with more than fourth power are at least of $O\!\left(p^{8}\right)$
order. And so, if we assume the finiteness of the S-matrix in the
chiral limit (and the other assumptions used there), we end up with
the third-order polynomial and the remainder of $O\!\left(p^{8}\right)$ order no matter
with how many subtraction we have begun.
From the possible extensions of the theorem we find especially
interesting two of them, the extension to the analysis of the meson
formfactors and the computation of amplitudes of $K\rightarrow 3\pi$
decays (with appearance of cusps). The latter of them is the subject
we are recently working on [15, 16].
Acknowledgment
We are grateful to K. Kampf for careful reading of the manuscript and helpful suggestions. We also thank to M. Knecht for inspiring discussions.
The work was supported by the Center for Particle Physics, Project
No. LC527 of the Ministry of Education of the Czech Republic and by the EU-RTN Programme, Contract No. MRTN–CT-2006-035482, “Flavianet.”
M.Z. would like also gratefully acknowledge an Early Stage
Researcher position supported by the EU-RTN Programme, Contract No. MRTN–CT-2006-035482, “Flavianet.”
Appendix A $O\!\left(p^{2}\right)$ amplitudes and the symmetry properties of amplitudes of the physical mesons
From the isospin Ward identities it follows that all the physical amplitudes could be expressed in terms of the 7 independent amplitudes. In our convention:
1.
$\eta\eta\rightarrow\eta\eta$
$$A(\eta\eta\rightarrow\eta\eta)=A_{\eta\eta}(s,t;u).$$
(37)
2.
$\pi\eta\rightarrow\pi\eta$
$$\displaystyle A(\pi^{\pm}\eta\rightarrow\pi^{\pm}\eta)$$
$$\displaystyle=A_{\pi\eta}(s,t;u),$$
(38)
$$\displaystyle A(\pi^{0}\eta\rightarrow\pi^{0}\eta)$$
$$\displaystyle=A_{\pi\eta}(s,t;u).$$
(39)
3.
$\pi\pi\rightarrow\pi\pi$
$$\displaystyle A(\pi^{+}\pi^{-}\rightarrow\pi^{0}\pi^{0})$$
$$\displaystyle=-A_{\pi\pi}(s,t;u),$$
(40)
$$\displaystyle A(\pi^{+}\pi^{-}\rightarrow\pi^{+}\pi^{-})$$
$$\displaystyle=A_{\pi\pi}(s,t;u)+A_{\pi\pi}(t,s;u),$$
(41)
$$\displaystyle A(\pi^{0}\pi^{0}\rightarrow\pi^{0}\pi^{0})$$
$$\displaystyle=A_{\pi\pi}(s,t;u)+A_{\pi\pi}(t,s;u)+A_{\pi\pi}(u,t;s).$$
(42)
4.
$KK\rightarrow\eta\eta$
$$A({\overline{K^{0}}}K^{0}\rightarrow\eta\eta)=-A(K^{-}K^{+}\rightarrow\eta\eta%
)=A_{\eta K}(s,t;u).$$
(43)
5.
$KK\rightarrow\pi\eta$
$$\displaystyle A(K^{-}K^{+}\rightarrow\pi^{0}\eta)$$
$$\displaystyle=-A_{\eta\pi K}(s,t;u),$$
(44)
$$\displaystyle A({\overline{K^{0}}}K^{0}\rightarrow\pi^{0}\eta)$$
$$\displaystyle=-A_{\eta\pi K}(s,t;u),$$
(45)
$$\displaystyle A(K^{-}K^{0}\rightarrow\pi^{-}\eta)$$
$$\displaystyle=-\sqrt{2}A_{\eta\pi K}(s,t;u),$$
(46)
$$\displaystyle A({\overline{K^{0}}}K^{+}\rightarrow\pi^{+}\eta)$$
$$\displaystyle=-\sqrt{2}A_{\eta\pi K}(s,t;u).$$
(47)
6.
$\pi\pi\rightarrow KK$
$$\displaystyle A(\pi^{0}\pi^{0}\rightarrow{\overline{K^{0}}}K^{0})$$
$$\displaystyle=-A(\pi^{0}\pi^{0}\rightarrow K^{-}K^{+})=A_{\pi K}(s,t;u),$$
(48)
$$\displaystyle A(\pi^{-}\pi^{0}\rightarrow K^{-}K^{0})$$
$$\displaystyle=\sqrt{2}B_{\pi K}(s,t;u),$$
(49)
$$\displaystyle A(\pi^{+}\pi^{0}\rightarrow{\overline{K^{0}}}K^{+})$$
$$\displaystyle=-\sqrt{2}B_{\pi K}(s,t;u),$$
(50)
$$\displaystyle A(\pi^{-}\pi^{+}\rightarrow K^{-}K^{+})$$
$$\displaystyle=A_{\pi K}(s,t;u)+B_{\pi K}(s,t;u),$$
(51)
$$\displaystyle A(\pi^{-}\pi^{+}\rightarrow{\overline{K^{0}}}K^{0})$$
$$\displaystyle=-A_{\pi K}(s,t;u)+B_{\pi K}(s,t;u).$$
(52)
For these processes there are two important amplitudes – the symmetric one $A_{\pi K}(s,t,u)$ and the antisymmetric one $B_{\pi K}(s,t,u)$. In fact, we can consider the only independent amplitude as (51) and then extract these amplitudes as the symmetric and the antisymmetric part of it in respect to the exchange of $t$ and $u$.
7.
$KK\rightarrow KK$
$$\displaystyle A(K^{-}K^{+}\rightarrow{\overline{K^{0}}}K^{0})$$
$$\displaystyle=-A_{KK}(s,t;u),$$
(53)
$$\displaystyle A(K^{-}K^{+}\rightarrow K^{-}K^{+})$$
$$\displaystyle=A({\overline{K^{0}}}K^{0}\rightarrow{\overline{K^{0}}}K^{0})=A_{%
KK}(s,t;u)+A_{KK}(t,s;u).$$
(54)
The most general form of the $O\!\left(p^{2}\right)$ ($O\!\left(p^{3}\right)$) amplitudes
The $O\!\left(p^{2}\right)$ amplitudes could be constructed as the most general
invariant amplitudes satisfying the crossing, Bose and isospin symmetries and using the
fact that they should be polynomials of the first order in the
Mandelstam variables obeying (5)191919In fact, it is the form of the $O\!\left(p^{3}\right)$
amplitudes, because this should hold also for them.. This form is
independent of the power counting used:
$$A_{\eta\eta}=-\frac{M_{\pi}^{2}}{3F_{\pi}^{2}}\alpha_{\eta\eta}\left(1-\frac{4%
M_{\eta}^{2}}{M_{\pi}^{2}}\right),$$
(55)
$$A_{\pi\eta}=\frac{1}{3F_{\pi}^{2}}\left(\beta_{\pi\eta}(3t-2M_{\eta}^{2}-2M_{%
\pi}^{2})+\alpha_{\pi\eta}M_{\pi}^{2}\right),$$
(56)
$$A_{\pi\pi}=\frac{1}{3F_{\pi}^{2}}\left(\beta_{\pi\pi}\left(3s-4M_{\pi}^{2}%
\right)+\alpha_{\pi\pi}M_{\pi}^{2}\right),$$
(57)
$$A_{\eta K}=\frac{1}{4F_{\pi}^{2}}\left[\beta_{\eta K}(3s-2M_{K}^{2}-2M_{\eta}^%
{2})+\alpha_{\eta K}\left(2M_{\eta}^{2}-\frac{2}{3}M_{K}^{2}\right)\right],$$
(58)
$$A_{\eta\pi K}=\frac{1}{4\sqrt{3}F_{\pi}^{2}}\left[\beta_{\eta\pi K}\left(3s-2M%
_{K}^{2}-M_{\pi}^{2}-M_{\eta}^{2}\right)-\left(2M_{K}^{2}-M_{\pi}^{2}-M_{\eta}%
^{2}+\alpha_{\eta\pi K}M_{\pi}^{2}\right)\right],$$
(59)
$$A_{\pi K}=\frac{1}{12F_{\pi}^{2}}\left[\beta_{\pi K}\left(3s-2M_{K}^{2}-2M_{%
\pi}^{2}\right)+2(M_{K}-M_{\pi})^{2}+4\alpha_{\pi K}M_{\pi}M_{K}\right],$$
(60)
$$B_{\pi K}=\frac{1}{4F_{\pi}^{2}}\gamma_{\pi K}(t-u),$$
(61)
$$A_{KK}=\frac{1}{6F_{\pi}^{2}}\left(\beta_{KK}\left(4M_{K}^{2}-3u\right)+3%
\gamma_{KK}(s-t)+2\alpha_{KK}M_{K}^{2}\right).$$
(62)
After a deeper analysis one can show that in $O\!\left(p^{2}\right)$ order the amplitude
$A_{\pi\eta}$ is $t$-independent, so the constant $\beta_{\pi\eta}$
is of the $O\!\left(p^{3}\right)$ order and similarly for $\gamma_{KK}$, which is
also equal to zero in the $O\!\left(p^{2}\right)$ case in an arbitrary power counting.
So, generally
$$\beta_{\pi\eta}\sim O\!\left(p^{1}\right)\qquad\text{and}\qquad\gamma_{KK}\sim
O%
\!\left(p^{1}\right)$$
(63)
The particular power countings differ in the particular values of
the (13+2) constants in those relations. For example, the standard
power counting gives (to $O\!\left(p^{3}\right)$):
$$\displaystyle\alpha_{\eta\eta}^{st}=\alpha_{\pi\eta}^{st}=\alpha_{\pi\pi}^{st}%
=\alpha_{\eta K}^{st}=\alpha_{\pi K}^{st}=\alpha_{KK}^{st}=1,$$
(64)
$$\displaystyle\beta_{\pi\pi}^{st}=\beta_{\eta K}^{st}=\beta_{\eta\pi K}^{st}=%
\beta_{\pi K}^{st}=\gamma_{\pi K}^{st}=\beta_{KK}^{st}=1,$$
(65)
$$\displaystyle\alpha_{\eta\pi K}^{st}=0,$$
(66)
$$\displaystyle\beta_{\pi\eta}^{st}=\gamma_{KK}^{st}=0.$$
(67)
Appendix B Results
In the results we have denoted the once subtracted integral
$$\overline{J}_{PQ}(s)=\frac{s}{16\pi^{2}}\int_{\Sigma}^{\infty}\frac{dx}{x}%
\frac{1}{x-s}\frac{\lambda^{1/2}_{PQ}(x)}{x}$$
(68)
in the S channel (analogically for the other crossed channels) and similarly the twice subtracted integral
$$\overline{\overline{J}}_{PQ}(s)=\frac{s^{2}}{16\pi^{2}}\int_{\Sigma}^{\infty}%
\frac{dx}{x^{2}}\frac{1}{x-s}\frac{\lambda^{1/2}_{PQ}(x)}{x}.$$
(69)
The amplitudes using the first iteration of the reconstruction
procedure read:
$\bullet\quad\eta\eta\rightarrow\eta\eta$
$$\begin{split}\displaystyle A_{\eta\eta}&\displaystyle=\frac{1}{3F_{\pi}^{2}}%
\alpha_{\eta\eta}\left(4M_{\eta}^{2}-M_{\pi}^{2}\right)+\frac{1}{3F_{\pi}^{4}}%
\delta_{\eta\eta}(s^{2}+t^{2}+u^{2})\\
&\displaystyle\quad+\left[\frac{1}{6F_{\pi}^{4}}\left(Z_{\eta\eta}^{\eta\eta}%
\overline{J}_{\eta\eta}(s)+Z^{\pi\pi}_{\eta\eta}\overline{J}_{\pi\pi}(s)+Z^{KK%
}_{\eta\eta}(s)\overline{J}_{KK}(s)\right)\right]+[s\leftrightarrow t]+[s%
\leftrightarrow u]+O\!\left(p^{5}\right).\end{split}$$
(70)
$$\displaystyle Z_{\eta\eta}^{\eta\eta}$$
$$\displaystyle=\frac{1}{3}\alpha_{\eta\eta}^{2}\left(M_{\pi}^{2}-4M_{\eta}^{2}%
\right)^{2},$$
(71)
$$\displaystyle Z^{\pi\pi}_{\eta\eta}$$
$$\displaystyle=\alpha_{\pi\eta}^{2}M_{\pi}^{4},$$
(72)
$$\displaystyle Z^{KK}_{\eta\eta}(s)$$
$$\displaystyle=\frac{3}{4}\left(3\beta_{\eta K}s-2\left(\beta_{\eta K}+\frac{1}%
{3}\alpha_{\eta K}\right)M_{K}^{2}+2\left(\alpha_{\eta K}-\beta_{\eta K}\right%
)M_{\eta}^{2}\right)^{2}.$$
(73)
$\bullet\quad\pi^{0}\eta\rightarrow\pi^{0}\eta$
$$\begin{split}\displaystyle A_{\pi\eta}&\displaystyle=\frac{1}{3F_{\pi}^{2}}%
\Big{(}\beta_{\pi\eta}(3t-2M_{\eta}^{2}-2M_{\pi}^{2})+M_{\pi}^{2}\alpha_{\pi%
\eta}\Big{)}+\frac{1}{3F_{\pi}^{4}}\left(\delta_{\pi\eta}(s^{2}+u^{2})+%
\varepsilon_{\pi\eta}t^{2}\right)\\
&\displaystyle\quad+\frac{1}{72F_{\pi}^{4}}\Big{(}Z^{KK}_{\pi\eta}(t)\overline%
{J}_{KK}(t)+Z^{\eta\eta}_{\pi\eta}\overline{J}_{\eta\eta}(t)+Z^{\pi\pi}_{\pi%
\eta}(t)\overline{J}_{\pi\pi}(t)\Big{)}\\
&\displaystyle\quad+\left[\frac{1}{9F_{\pi}^{4}}\left(Y_{\pi\eta}^{\pi\eta}%
\overline{J}_{\pi\eta}(s)+Y^{KK}_{\pi\eta}(s)\overline{J}_{KK}(s)\right)\right%
]+[s\leftrightarrow u]+O\!\left(p^{5}\right).\end{split}$$
(74)
$$\displaystyle\begin{split}\displaystyle Z^{KK}_{\pi\eta}(t)&\displaystyle=%
\left(9\beta_{\eta K}t-2(3\beta_{\eta K}+\alpha_{\eta K})M_{K}^{2}+6(\alpha_{%
\eta K}-\beta_{\eta K})M_{\eta}^{2}\right)\\
&\displaystyle\qquad\times\left(3\beta_{\pi K}t+2(1-\beta_{\pi K})(M_{K}^{2}+M%
_{\pi}^{2})+4(\alpha_{\pi K}-1)M_{\pi}M_{K}\right),\end{split}$$
(75)
$$\displaystyle Z^{\eta\eta}_{\pi\eta}$$
$$\displaystyle=-4\alpha_{\pi\eta}\alpha_{\eta\eta}M_{\pi}^{2}\left(M_{\pi}^{2}-%
4M_{\eta}^{2}\right),$$
(76)
$$\displaystyle Z^{\pi\pi}_{\pi\eta}(t)$$
$$\displaystyle=4\alpha_{\pi\eta}M_{\pi}^{2}\left(6\beta_{\pi\pi}t+(5\alpha_{\pi%
\pi}-8\beta_{\pi\pi})M_{\pi}^{2}\right),$$
(77)
$$\displaystyle Y_{\pi\eta}^{\pi\eta}$$
$$\displaystyle=\alpha_{\pi\eta}^{2}M_{\pi}^{4},$$
(78)
$$\displaystyle Y^{KK}_{\pi\eta}(s)$$
$$\displaystyle=\frac{3}{8}\left(3\beta_{\eta\pi K}s-2(1+\beta_{\eta\pi K})M_{K}%
^{2}+(1-\alpha_{\eta\pi K}-\beta_{\eta\pi K})M_{\pi}^{2}+(1-\beta_{\eta\pi K})%
M_{\eta}^{2}\right)^{2}.$$
(79)
$\bullet\quad\pi^{+}\pi^{-}\rightarrow\pi^{0}\pi^{0}$
$$\begin{split}\displaystyle A_{\pi\pi}&\displaystyle=\frac{1}{3F_{\pi}^{2}}%
\left(\beta_{\pi\pi}\left(3s-4M_{\pi}^{2}\right)+\alpha_{\pi\pi}M_{\pi}^{2}%
\right)+\frac{1}{F_{\pi}^{4}}\left(\delta_{\pi\pi}s^{2}+\varepsilon_{\pi\pi}(t%
^{2}+u^{2})\right)\\
&\displaystyle\quad+\frac{1}{72F_{\pi}^{4}}\left(Z_{\pi\pi}^{\pi\pi}(s)%
\overline{J}_{\pi\pi}(s)+Z^{\eta\eta}_{\pi\pi}\overline{J}_{\eta\eta}(s)+Z^{KK%
}_{\pi\pi}(s)\overline{J}_{KK}(s)\right)\\
&\displaystyle\quad+\left[\frac{1}{72F_{\pi}^{4}}\left(Y^{KK}_{\pi\pi}(t,u)%
\overline{J}_{KK}(t)+Y_{\pi\pi}^{\pi\pi}(t,u)\overline{J}_{\pi\pi}(t)\right)%
\right]+[t\leftrightarrow u]+O\!\left(p^{5}\right).\end{split}$$
(80)
$$\displaystyle Z_{\pi\pi}^{\pi\pi}(s)$$
$$\displaystyle=4\left(3\beta_{\pi\pi}s+(7\alpha_{\pi\pi}-4\beta_{\pi\pi})M_{\pi%
}^{2}\right)\left(3\beta_{\pi\pi}s+(\alpha_{\pi\pi}-4\beta_{\pi\pi})M_{\pi}^{2%
}\right),$$
(81)
$$\displaystyle Z^{\eta\eta}_{\pi\pi}$$
$$\displaystyle=4\alpha_{\pi\eta}^{2}M_{\pi}^{4},$$
(82)
$$\displaystyle Z^{KK}_{\pi\pi}(s)$$
$$\displaystyle=\left(3\beta_{\pi K}s+2(1-\beta_{\pi K})(M_{K}^{2}+M_{\pi}^{2})+%
4(\alpha_{\pi K}-1)M_{\pi}M_{K}\right)^{2},$$
(83)
$$\displaystyle Y^{KK}_{\pi\pi}(t,u)$$
$$\displaystyle=3\gamma_{\pi K}^{2}(t-4M_{K}^{2})(4M_{\pi}^{2}-t-2u),$$
(84)
$$\displaystyle Y_{\pi\pi}^{\pi\pi}(t,u)$$
$$\displaystyle=4\left(3\beta_{\pi\pi}^{2}t(t-u)+6\beta_{\pi\pi}(2\beta_{\pi\pi}%
u-\alpha_{\pi\pi}t)M_{\pi}^{2}+2(\alpha_{\pi\pi}^{2}+4\beta_{\pi\pi}(\alpha_{%
\pi\pi}-2\beta_{\pi\pi}))M_{\pi}^{4}\right).$$
(85)
$\bullet\quad{\overline{K^{0}}}K^{0}\rightarrow\eta\eta$
$$\begin{split}\displaystyle A_{\eta K}&\displaystyle=\frac{1}{4F_{\pi}^{2}}%
\left(\beta_{\eta K}(3s-2M_{K}^{2}-2M_{\eta}^{2})+\alpha_{\eta K}\left(2M_{%
\eta}^{2}-\frac{2}{3}M_{K}^{2}\right)\right)+\frac{1}{4F_{\pi}^{4}}\left(%
\delta_{\eta K}s^{2}+\varepsilon_{\eta K}(t^{2}+u^{2})\right)\\
&\displaystyle\quad+\frac{1}{24F_{\pi}^{4}}\left(Z_{\eta K}^{\eta\eta}(s)%
\overline{J}_{\eta\eta}(s)+Z_{\eta K}^{\pi\pi}(s)\overline{J}_{\pi\pi}(s)+Z_{%
\eta K}^{KK}(s)\overline{J}_{KK}(s)\right)\\
&\displaystyle\quad+\left[\frac{1}{32F_{\pi}^{4}}\left(Y_{\eta K}^{\eta K}(t,u%
)\overline{J}_{\eta K}(t)+Y_{\eta K}^{K\pi}(t,u)\overline{J}_{K\pi}(t)\right)%
\right.\\
&\displaystyle\qquad\quad-\frac{3}{32F_{\pi}^{4}}\frac{1}{t}\left(X_{\eta K}^{%
\eta K}(u)\overline{J}_{\eta K}(t)+X_{\eta K}^{K\pi}(u)\overline{J}_{K\pi}(t)%
\right)\\
&\displaystyle\quad\qquad\qquad+\left.\frac{3}{16F_{\pi}^{4}}\frac{1}{t^{2}}%
\Delta_{\scriptscriptstyle K\eta}^{2}\left(W_{\eta K}^{\eta K}\overline{%
\overline{J}}_{\eta K}(t)+W_{\eta K}^{K\pi}\overline{\overline{J}}_{K\pi}(t)%
\right)\right]+\left[t\leftrightarrow u\right]+O\!\left(p^{5}\right).\end{split}$$
(86)
$$\displaystyle W_{\eta K}^{\eta K}$$
$$\displaystyle=\Delta_{\scriptscriptstyle K\eta}^{2}\beta^{2}_{\eta K},$$
(87)
$$\displaystyle W_{\eta K}^{K\pi}$$
$$\displaystyle=\Delta_{\scriptscriptstyle K\pi}^{2}\beta^{2}_{\eta\pi K},$$
(88)
$$\displaystyle X_{\eta K}^{\eta K}(u)$$
$$\displaystyle=\left(\beta_{\eta K}u+2\left(\beta_{\eta K}-\frac{2}{3}\alpha_{%
\eta K}\right)M_{K}^{2}+2(\beta_{\eta K}+2\alpha_{\eta K})M_{\eta}^{2}\right)%
\beta_{\eta K}\Delta_{\scriptscriptstyle K\eta}^{2},$$
(89)
$$\displaystyle\begin{split}\displaystyle X_{\eta K}^{K\pi}(u)&\displaystyle=%
\beta_{\eta\pi K}\Big{(}\beta_{\eta\pi K}\left(M_{K}^{2}-M_{\pi}^{2}\right)^{2%
}u+2\left(\beta_{\eta\pi K}-2\right)M_{K}^{6}-2\left(2\beta_{\eta\pi K}-3%
\right)M_{K}^{4}M_{\eta}^{2}\\
&\displaystyle\qquad+2\left(3+\beta_{\eta\pi K}-\alpha_{\eta\pi K}\right)M_{K}%
^{4}M_{\pi}^{2}-2\left(\beta_{\eta\pi K}+1-\alpha_{\eta\pi K}\right)M_{\pi}^{4%
}M_{K}^{2}-2M_{\eta}^{4}M_{K}^{2}\\
&\displaystyle\qquad\quad+2\left(\alpha_{\eta\pi K}-4\right)M_{K}^{2}M_{\eta}^%
{2}M_{\pi}^{2}-2\left(\alpha_{\eta\pi K}-1\right)M_{\pi}^{4}M_{\eta}^{2}+2(%
\beta_{\eta\pi K}+1)M_{\eta}^{4}M_{\pi}^{2}\Big{)},\end{split}$$
(90)
$$\displaystyle Z_{\eta K}^{\eta\eta}(s)$$
$$\displaystyle=\left(3\beta_{\eta K}s-2(\beta_{\eta K}+\frac{1}{3}\alpha_{\eta K%
})M_{K}^{2}-2(\beta_{\eta K}-\alpha_{\eta K})M_{\eta}^{2}\right)\alpha_{\eta%
\eta}\left(4M_{\eta}^{2}-M_{\pi}^{2}\right),$$
(91)
$$\displaystyle Z_{\eta K}^{\pi\pi}(s)$$
$$\displaystyle=\alpha_{\pi\eta}M_{\pi}^{2}\left(3\beta_{\pi K}s+2(1-\beta_{\pi K%
})(M_{K}^{2}+M_{\pi}^{2})+4(\alpha_{\pi K}-1)M_{\pi}M_{K}\right),$$
(92)
$$\displaystyle\begin{split}\displaystyle Z_{\eta K}^{KK}(s)&\displaystyle=3%
\left(3\beta_{\eta K}s-2\left(\beta_{\eta K}+\frac{1}{3}\alpha_{\eta K}\right)%
M_{K}^{2}-2(\beta_{\eta K}-\alpha_{\eta K})M_{\eta}^{2}\right)\\
&\displaystyle\qquad\times\left(\frac{3}{2}\beta_{KK}s+2(\alpha_{KK}-\beta_{KK%
})M_{K}^{2}\right),\end{split}$$
(93)
$$\displaystyle\begin{split}\displaystyle Y_{\eta K}^{\eta K}(t,u)&\displaystyle%
=\frac{1}{9}\Big{(}27\beta_{\eta K}^{2}t(t-u)-6\left(8\alpha_{\eta K}^{2}-8%
\alpha_{\eta K}\beta_{\eta K}+39\beta_{\eta K}^{2}\right)M_{K}^{2}M_{\eta}^{2}%
\\
&\displaystyle\qquad+9\left(8\alpha_{\eta K}^{2}+8\alpha_{\eta K}\beta_{\eta K%
}+5\beta_{\eta K}^{2}\right)M_{\eta}^{4}+\left(8\alpha_{\eta K}^{2}-24\alpha_{%
\eta K}\beta_{\eta K}+45\beta_{\eta K}^{2}\right)M_{K}^{4}\\
&\displaystyle\quad\qquad+36\alpha_{\eta K}\beta_{\eta K}\left(M_{K}^{2}-3M_{%
\eta}^{2}\right)t+54\beta_{\eta K}^{2}\left(M_{K}^{2}+M_{\eta}^{2}\right)u\Big%
{)},\end{split}$$
(94)
$$\displaystyle\begin{split}\displaystyle Y_{\eta K}^{K\pi}(t,u)&\displaystyle=3%
\beta_{\eta\pi K}^{2}t(t-u)+\left(5\beta_{\eta\pi K}^{2}-8\beta_{\eta\pi K}+8%
\right)M_{K}^{4}+2\left(\beta_{\eta\pi K}^{2}+\beta_{\eta\pi K}+1\right)M_{%
\eta}^{4}\\
&\displaystyle\quad-8\left(2\beta_{\eta\pi K}^{2}+1\right)M_{K}^{2}M_{\eta}^{2%
}+\left(2\beta_{\eta\pi K}\left(1-\alpha_{\eta\pi K}\right)+2\left(1-\alpha_{%
\eta\pi K}\right)^{2}-\beta_{\eta\pi K}^{2}\right)M_{\pi}^{4}\\
&\displaystyle\quad+2\left(2\left(\beta_{\eta\pi K}^{2}+\beta_{\eta\pi K}+1-%
\alpha_{\eta\pi K}\right)-\alpha_{\eta\pi K}\beta_{\eta\pi K}\right)M_{\eta}^{%
2}M_{\pi}^{2}\\
&\displaystyle\quad-2\left(5\beta_{\eta\pi K}^{2}+2\alpha_{\eta\pi K}\beta_{%
\eta\pi K}+4\left(1-\alpha_{\eta\pi K}\right)\right)M_{K}^{2}M_{\pi}^{2}\\
&\displaystyle\quad-6\beta_{\eta\pi K}\left(M_{\eta}^{2}-2M_{K}^{2}-\left(%
\alpha_{\eta\pi K}-1\right)M_{\pi}^{2}\right)t+6\beta_{\eta\pi K}^{2}\left(M_{%
K}^{2}+M_{\pi}^{2}\right)u.\end{split}$$
(95)
$\bullet\quad{\overline{K^{0}}}K^{0}\rightarrow\pi^{0}\eta$
$$\begin{split}\displaystyle A_{\eta\pi K}&\displaystyle=\frac{1}{4\sqrt{3}F_{%
\pi}^{2}}V_{\eta\pi K}(s)+\frac{1}{4\sqrt{3}F_{\pi}^{4}}\left(\delta_{\eta\pi K%
}s^{2}+\varepsilon_{\eta\pi K}(t^{2}+u^{2})\right)\\
&\displaystyle\quad+\frac{1}{24\sqrt{3}F_{\pi}^{4}}\left(Z_{\eta\pi K}^{\eta%
\pi}(s)\overline{J}_{\eta\pi}(s)+Z_{\eta\pi K}^{KK}\overline{J}_{KK}(s)\right)%
V_{\eta\pi K}(s)\\
&\displaystyle\quad+\Big{[}-\frac{1}{96\sqrt{3}F_{\pi}^{4}}\left(Y_{\eta\pi K}%
^{\eta K}(t,u)\overline{J}_{\eta K}(t)+Y_{\eta\pi K}^{K\pi}(t,u)\overline{J}_{%
K\pi}(t)\right)\\
&\displaystyle\quad\qquad-\frac{1}{32\sqrt{3}F_{\pi}^{4}}\frac{1}{t}\left(X_{%
\eta\pi K}^{\eta K}(u)\overline{J}_{\eta K}(t)+X_{\eta\pi K}^{K\pi}(u)%
\overline{J}_{K\pi}(t)\right)\\
&\displaystyle\quad\qquad\qquad+\frac{1}{16\sqrt{3}F_{\pi}^{4}}\frac{1}{t^{2}}%
U_{\eta\pi K}\left.\left(W_{\eta\pi K}^{\eta K}\overline{\overline{J}}_{\eta K%
}(t)+W_{\eta\pi K}^{K\pi}\overline{\overline{J}}_{K\pi}(t)\right)\right]+[t%
\leftrightarrow u]+O\!\left(p^{5}\right).\end{split}$$
(96)
$$\displaystyle U_{\eta\pi K}$$
$$\displaystyle=\beta_{\eta\pi K}\Delta_{\scriptscriptstyle K\eta}\Delta_{%
\scriptscriptstyle K\pi},$$
(97)
$$\displaystyle V_{\eta\pi K}(s)$$
$$\displaystyle=\beta_{\eta\pi K}\left(3s-2M_{K}^{2}-M_{\pi}^{2}-M_{\eta}^{2}%
\right)-\left(2M_{K}^{2}-M_{\pi}^{2}-M_{\eta}^{2}+\alpha_{\eta\pi K}M_{\pi}^{2%
}\right),$$
(98)
$$\displaystyle Z_{\eta\pi K}^{\eta\pi}$$
$$\displaystyle=2\alpha_{\pi\eta}M_{\pi}^{2},$$
(99)
$$\displaystyle Z_{\eta\pi K}^{KK}(s)$$
$$\displaystyle=\frac{3}{2}\beta_{KK}s+2(\alpha_{KK}-\beta_{KK})M_{K}^{2},$$
(100)
$$\displaystyle\begin{split}\displaystyle Y_{\eta\pi K}^{\eta K}(t,u)&%
\displaystyle=9\beta_{\eta\pi K}\beta_{\eta K}(t(u-t)-2(M_{\eta}^{2}+M_{K}^{2}%
)u)-3(\beta_{\eta\pi K}+2)(2\alpha_{\eta K}+\beta_{\eta K})M_{\eta}^{4}\\
&\displaystyle\quad+\left(3\beta_{\eta K}(4-5\beta_{\eta\pi K})+4\alpha_{\eta K%
}(\beta_{\eta\pi K}-2)\right)M_{K}^{4}\\
&\displaystyle\quad+2(\alpha_{\eta K}(2-2\alpha_{\eta\pi K}+\beta_{\eta\pi K})%
+3\beta_{\eta K}(\alpha_{\eta\pi K}-1+4\beta_{\eta\pi K}))M_{K}^{2}M_{\pi}^{2}%
\\
&\displaystyle\quad+(\alpha_{\eta K}(28-10\beta_{\eta\pi K})+6\beta_{\eta K}(9%
\beta_{\eta\pi K}+1))M_{K}^{2}M_{\eta}^{2}\\
&\displaystyle\quad+6((\alpha_{\eta\pi K}-1)(2\alpha_{\eta K}+\beta_{\eta K})-%
\beta_{\eta\pi K}(\alpha_{\eta K}+2\beta_{\eta K}))M_{\eta}^{2}M_{\pi}^{2}\\
&\displaystyle\quad-3t\big{(}2(3\beta_{\eta K}+\alpha_{\eta K}\beta_{\eta\pi K%
})M_{K}^{2}-3(\beta_{\eta K}+2\alpha_{\eta K}\beta_{\eta\pi K})M_{\eta}^{2}+3%
\beta_{\eta K}(\alpha_{\eta\pi K}-1)M_{\pi}^{2}\big{)},\end{split}$$
(101)
$$\displaystyle\begin{split}\displaystyle Y_{\eta\pi K}^{K\pi}(t,u)&%
\displaystyle=3\beta_{\eta\pi K}t\left((10\gamma_{\pi K}-\beta_{\pi K})t+(%
\beta_{\pi K}+2\gamma_{\pi K})u\right)+3(\beta_{\pi K}-2\gamma_{\pi K}(3+2%
\beta_{\eta\pi K}))tM_{\eta}^{2}\\
&\displaystyle\quad-6\beta_{\eta\pi K}(\beta_{\pi K}+2\gamma_{\pi K})(M_{K}^{2%
}+M_{\pi}^{2})u-6(2\gamma_{\pi K}(4\beta_{\eta\pi K}-3)+\beta_{\pi K}-\beta_{%
\eta\pi K})tM_{K}^{2}\\
&\displaystyle\quad-3(2\beta_{\eta\pi K}(6\gamma_{\pi K}-1)+(1-\alpha_{\eta\pi
K%
})(6\gamma_{\pi K}-\beta_{\pi K}))tM_{\pi}^{2}\\
&\displaystyle\quad+(2+\beta_{\eta\pi K}-2\alpha_{\eta\pi K})(6\gamma_{\pi K}-%
\beta_{\pi K}-2)M_{\pi}^{4}\\
&\displaystyle\quad+\left(\beta_{\pi K}(4-5\beta_{\eta\pi K})+2\gamma_{\pi K}(%
21\beta_{\eta\pi K}-12)+8-4\beta_{\eta\pi K}\right)M_{K}^{4}\\
&\displaystyle\quad+2(\beta_{\eta\pi K}(4\beta_{\pi K}-6\gamma_{\pi K}-1)-%
\beta_{\pi K}+6\gamma_{\pi K}-2)M_{K}^{2}M_{\eta}^{2}\\
&\displaystyle\quad-2(\beta_{\pi K}(2\beta_{\eta\pi K}+1)-6\gamma_{\pi K}(3%
\beta_{\eta\pi K}+1)+2+\beta_{\eta\pi K})M_{\pi}^{2}M_{\eta}^{2}\\
&\displaystyle\quad+4(\alpha_{\pi K}-1)(V_{\eta\pi K}(t)+3(2M_{K}^{2}-M_{\pi}^%
{2}-M_{\eta}^{2}+\alpha_{\eta\pi K}M_{\pi}^{2}))M_{K}M_{\pi}\\
&\displaystyle\quad+2(3\beta_{\eta\pi K}(3\beta_{\pi K}+4\gamma_{\pi K}-1)+(%
\alpha_{\eta\pi K}+1)(\beta_{\pi K}-6\gamma_{\pi K}+2))M_{K}^{2}M_{\pi}^{2},%
\end{split}$$
(102)
$$\displaystyle\begin{split}\displaystyle X_{\eta\pi K}^{\eta K}(u)&%
\displaystyle=\Delta_{\scriptscriptstyle\eta K}\Big{(}3\beta_{\eta K}\beta_{%
\eta\pi K}(M_{\eta}^{2}-M_{K}^{2})u+3\beta_{\eta K}M_{\eta}^{4}+(2\alpha_{\eta
K%
}\beta_{\eta\pi K}+6\beta_{\eta K}\left(1-\beta_{\eta\pi K}\right))M_{K}^{4}\\
&\displaystyle\qquad\quad+3(\beta_{\eta K}(1-\alpha_{\eta\pi K})+2\beta_{\eta%
\pi K}(\alpha_{\eta K}+\beta_{\eta K}))M_{\eta}^{2}M_{\pi}^{2}\\
&\displaystyle\qquad\quad\quad+(3\beta_{\eta K}(\alpha_{\eta\pi K}-1)+2\beta_{%
\eta\pi K}(3\beta_{\eta K}-\alpha_{\eta K}))M_{K}^{2}M_{\pi}^{2}\\
&\displaystyle\qquad\quad\quad\quad-3(3\beta_{\eta K}+2\beta_{\eta\pi K}(\beta%
_{\eta K}+\alpha_{\eta K}))M_{K}^{2}M_{\eta}^{2}\Big{)},\end{split}$$
(103)
$$\displaystyle\begin{split}\displaystyle X_{\eta\pi K}^{K\pi}(u)&\displaystyle=%
\Delta_{\scriptscriptstyle\pi K}\Big{(}\beta_{\eta\pi K}(\beta_{\pi K}+2\gamma%
_{\pi K})\left(M_{\pi}^{2}-M_{K}^{2}\right)u+4\beta_{\eta\pi K}(1-\alpha_{\pi K%
})(M_{K}^{2}-M_{\eta}^{2})M_{K}M_{\pi}\\
&\displaystyle\qquad\quad+2\beta_{\eta\pi K}(2\gamma_{\pi K}-\beta_{\pi K}-1)(%
M_{K}^{4}-M_{K}^{2}M_{\eta}^{2}-M_{\pi}^{2}M_{\eta}^{2}+M_{K}^{2}M_{\pi}^{2})%
\\
&\displaystyle\qquad\quad\quad+(\beta_{\pi K}+2\gamma_{\pi K})(2M_{K}^{4}-M_{K%
}^{2}M_{\eta}^{2}+M_{\pi}^{2}M_{\eta}^{2}+(\alpha_{\eta\pi K}-3)M_{K}^{2}M_{%
\pi}^{2})\\
&\displaystyle\qquad\quad\quad\quad+(\beta_{\pi K}+2\gamma_{\pi K})(1-\alpha_{%
\eta\pi K})M_{\pi}^{4}\Big{)},\end{split}$$
(104)
$$\displaystyle W_{\eta\pi K}^{\eta K}$$
$$\displaystyle=3\Delta_{\scriptscriptstyle K\eta}^{2}\beta_{\eta K},$$
(105)
$$\displaystyle W_{\eta\pi K}^{K\pi}$$
$$\displaystyle=\Delta_{\scriptscriptstyle K\pi}^{2}(\beta_{\pi K}+2\gamma_{\pi K%
}).$$
(106)
$\bullet\quad\pi^{0}\pi^{0}\rightarrow{\overline{K^{0}}}K^{0}$
$$\begin{split}\displaystyle A_{\pi K}&\displaystyle=\frac{1}{12F_{\pi}^{2}}%
\left(\beta_{\pi K}\left(3s-2M_{K}^{2}-2M_{\pi}^{2}\right)+2(M_{K}-M_{\pi})^{2%
}+4\alpha_{\pi K}M_{\pi}M_{K}\right)\\
&\displaystyle\quad+\frac{1}{12F_{\pi}^{4}}\left(\delta_{\pi K}s^{2}+%
\varepsilon_{\pi K}(t^{2}+u^{2})\right)\\
&\displaystyle\quad+\frac{1}{72F_{\pi}^{4}}\Big{(}Z_{\pi K}^{\pi\pi}(s)%
\overline{J}_{\pi\pi}(s)+Z_{\pi K}^{\eta\eta}(s)\overline{J}_{\eta\eta}(s)+Z_{%
\pi K}^{KK}(s)\overline{J}_{KK}(s)\Big{)}\\
&\displaystyle\quad+\bigg{[}\frac{1}{96F_{\pi}^{4}}\left(Y_{\pi K}^{\eta K}(t,%
u)\overline{J}_{\eta K}(t)+Y_{\pi K}^{\pi K}(t,u)\overline{J}_{\pi K}(t)\right%
)\\
&\displaystyle\quad\qquad+\frac{1}{32F_{\pi}^{4}}\frac{1}{t}\left(X_{\pi K}^{%
\eta K}(u)\overline{J}_{\eta K}(t)+X_{\pi K}^{\pi K}(u)\overline{J}_{\pi K}(t)%
\right)\\
&\displaystyle\quad\qquad\qquad+\frac{1}{16F_{\pi}^{4}}\frac{1}{t^{2}}\Delta_{%
\scriptscriptstyle K\pi}^{2}\left(W_{\pi K}^{\eta K}\overline{\overline{J}}_{%
\eta K}(t)+W_{\pi K}^{\pi K}\overline{\overline{J}}_{\pi K}(t)\right)\bigg{]}+%
[t\leftrightarrow u]+O\!\left(p^{5}\right).\end{split}$$
(107)
$$\displaystyle\begin{split}\displaystyle Z_{\pi K}^{\pi\pi}(s)&\displaystyle=%
\left(3\beta_{\pi K}s+2(1-\beta_{\pi K})(M_{K}^{2}+M_{\pi}^{2})+4(\alpha_{\pi K%
}-1)M_{\pi}M_{K}\right)\\
&\displaystyle\qquad\times\left(6\beta_{\pi\pi}s+(5\alpha_{\pi\pi}-8\beta_{\pi%
\pi})M_{\pi}^{2}\right),\end{split}$$
(108)
$$\displaystyle Z_{\pi K}^{\eta\eta}(s)$$
$$\displaystyle=3\alpha_{\pi\eta}M_{\pi}^{2}\left(3\beta_{\eta K}s-2\left(\beta_%
{\eta K}+\frac{1}{3}\alpha_{\eta K}\right)M_{K}^{2}+2(\alpha_{\eta K}-\beta_{%
\eta K})M_{\eta}^{2}\right),$$
(109)
$$\displaystyle\begin{split}\displaystyle Z_{\pi K}^{KK}(s)&\displaystyle=3\left%
(3\beta_{\pi K}s+2(1-\beta_{\pi K})(M_{K}^{2}+M_{\pi}^{2})+4(\alpha_{\pi K}-1)%
M_{\pi}M_{K}\right)\\
&\displaystyle\qquad\times\left(\frac{3}{2}\beta_{KK}s+2(\alpha_{KK}-\beta_{KK%
})M_{K}^{2}\right),\end{split}$$
(110)
$$\displaystyle\begin{split}\displaystyle Y_{\pi K}^{\eta K}(t,u)&\displaystyle=%
3\beta_{\eta K}^{2}t(t-u)-6\beta_{\eta\pi K}\left(M_{\eta}^{2}-2M_{K}^{2}+(1-%
\alpha_{\eta\pi K})M_{\pi}^{2}\right)t+6\beta_{\eta\pi K}^{2}(M_{\eta}^{2}+M_{%
K}^{2})u\\
&\displaystyle\quad+\left(5\beta_{\eta\pi K}^{2}+8(1-\beta_{\eta\pi K})\right)%
M_{K}^{4}+\left(2+2\beta_{\eta\pi K}-\beta_{\eta\pi K}^{2}\right)M_{\eta}^{4}-%
2(5\beta_{\eta\pi K}^{2}+4)M_{K}^{2}M_{\eta}^{2}\\
&\displaystyle\quad+2\left(\alpha_{\eta\pi K}(\alpha_{\eta\pi K}-\beta_{\eta%
\pi K}-2)+\beta_{\eta\pi K}^{2}+\beta_{\eta\pi K}+1\right)M_{\pi}^{4}\\
&\displaystyle\quad-4(\beta_{\eta\pi K}(4\beta_{\eta\pi K}+\alpha_{\eta\pi K})%
+2(1-\alpha_{\eta\pi K}))M_{K}^{2}M_{\pi}^{2}\\
&\displaystyle\quad+2(2(\beta_{\eta\pi K}^{2}+\beta_{\eta\pi K}+1)-\alpha_{%
\eta\pi K}(\beta_{\eta\pi K}+2))M_{\pi}^{2}M_{\eta}^{2},\end{split}$$
(111)
$$\displaystyle\begin{split}\displaystyle Y_{\pi K}^{\pi K}(t,u)&\displaystyle=%
\frac{1}{3}\Big{(}3t\left(\left(\beta_{\pi K}^{2}+26\gamma_{\pi K}^{2}\right)t%
-\left(\beta_{\pi K}^{2}+2\gamma_{\pi K}^{2}\right)u\right)+6\left(\beta_{\pi K%
}^{2}+2\gamma_{\pi K}^{2}\right)\left(M_{K}^{2}+M_{\pi}^{2}\right)u\\
&\displaystyle\qquad-12\left(\left(8\gamma_{\pi K}^{2}+\beta_{\pi K}\right)%
\left(M_{K}^{2}+M_{\pi}^{2}\right)+2\beta_{\pi K}\left(\alpha_{\pi K}-1\right)%
M_{K}M_{\pi}\right)t\\
&\displaystyle\qquad\quad+2\left(16\alpha_{\pi K}\left(\alpha_{\pi K}-2\right)%
+\beta_{\pi K}\left(8-13\beta_{\pi K}\right)+6\left(4+13\gamma_{\pi K}^{2}%
\right)\right)M_{K}^{2}M_{\pi}^{2}\\
&\displaystyle\qquad\quad\quad+\left(5\beta_{\pi K}^{2}+8\left(\beta_{\pi K}+1%
\right)-30\gamma_{\pi K}^{2}\right)\left(M_{K}^{4}+M_{\pi}^{4}\right)\\
&\displaystyle\qquad\quad\quad\quad+16\left(\alpha_{\pi K}-1\right)\left(\beta%
_{\pi K}+2\right)M_{K}M_{\pi}\left(M_{K}^{2}+M_{\pi}^{2}\right)\Big{)},\end{split}$$
(112)
$$\displaystyle\begin{split}\displaystyle X_{\pi K}^{\eta K}(u)&\displaystyle=%
\beta_{\eta\pi K}\Big{(}-\beta_{\eta\pi K}\Delta_{\scriptscriptstyle K\eta}^{2%
}u+2\left(2-\beta_{\eta\pi K}\right)M_{K}^{6}+2M_{\eta}^{4}\left(\left(\beta_{%
\eta\pi K}+1\right)M_{K}^{2}-M_{\pi}^{2}\right)\\
&\displaystyle\qquad+2M_{K}^{2}M_{\pi}^{2}M_{\eta}^{2}\left(4-\alpha_{\eta\pi K%
}\right)+2M_{K}^{4}\left(\left(\alpha_{\eta\pi K}+2\beta_{\eta\pi K}-3\right)M%
_{\pi}^{2}-\left(\beta_{\eta\pi K}+3\right)M_{\eta}^{2}\right)\\
&\displaystyle\quad\qquad+2M_{\pi}^{4}\left(\left(\alpha_{\eta\pi K}-\beta_{%
\eta\pi K}-1\right)M_{\eta}^{2}+\left(1-\alpha_{\eta\pi K}\right)M_{K}^{2}%
\right)\Big{)},\end{split}$$
(113)
$$\displaystyle\begin{split}\displaystyle X_{\pi K}^{\pi K}(u)&\displaystyle=-%
\frac{1}{3}\Delta_{\scriptscriptstyle K\pi}^{2}\Big{(}\left(\beta_{\pi K}^{2}+%
2\gamma_{\pi K}^{2}\right)u+8\left(\alpha_{\pi K}-1\right)\beta_{\pi K}M_{K}M_%
{\pi}\\
&\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad+2\left(\beta_{%
\pi K}^{2}+2\beta_{\pi K}-6\gamma_{\pi K}^{2}\right)\left(M_{K}^{2}+M_{\pi}^{2%
}\right)\Big{)},\end{split}$$
(114)
$$\displaystyle W_{\pi K}^{\eta K}$$
$$\displaystyle=\Delta_{\scriptscriptstyle K\eta}^{2}\beta^{2}_{\eta\pi K},$$
(115)
$$\displaystyle W_{\pi K}^{\pi K}$$
$$\displaystyle=\frac{1}{3}\Delta_{\scriptscriptstyle K\pi}^{2}\left(\beta^{2}_{%
\pi K}+2\gamma_{\pi K}^{2}\right).$$
(116)
$\bullet\quad\pi^{-}\pi^{0}\rightarrow K^{-}K^{0}$
$$\begin{split}\displaystyle B_{\pi K}&\displaystyle=\frac{1}{4F_{\pi}^{2}}%
\gamma_{\pi K}(t-u)+\frac{1}{4F_{\pi}^{4}}\varphi_{\pi K}s(t-u)\\
&\displaystyle\quad+\frac{1}{48F_{\pi}^{4}}\gamma_{\pi K}(t-u)\left(2\overline%
{J}_{\pi\pi}(s)\beta_{\pi\pi}(s-4M_{\pi}^{2})+\overline{J}_{KK}(s)\beta_{KK}(s%
-4M_{K}^{2})\right)\\
&\displaystyle\quad+\bigg{[}\frac{1}{96F_{\pi}^{4}}\left(Y_{\pi Kch}^{\eta K}(%
t,u)\overline{J}_{\eta K}(t)+Y_{\pi Kch}^{\pi K}(t,u)\overline{J}_{\pi K}(t)%
\right)\\
&\displaystyle\quad\qquad+\frac{1}{32F_{\pi}^{4}}\frac{1}{t}\left(X_{\pi Kch}^%
{\eta K}(u)\overline{J}_{\eta K}(t)+X_{\pi Kch}^{\pi K}(u)\overline{J}_{\pi K}%
(t)\right)\\
&\displaystyle\quad\qquad\qquad+\frac{1}{16F_{\pi}^{4}}\frac{1}{t^{2}}\Delta_{%
\scriptscriptstyle K\pi}^{2}\left(W_{\pi Kch}^{\eta K}\overline{\overline{J}}_%
{\eta K}(t)+W_{\pi Kch}^{\pi K}\overline{\overline{J}}_{\pi K}(t)\right)\bigg{%
]}-[t\leftrightarrow u]+O\!\left(p^{5}\right).\end{split}$$
(117)
$$\displaystyle Y_{\pi Kch}^{\eta K}(t,u)$$
$$\displaystyle=Y_{\pi K}^{\eta K}(t,u),$$
(118)
$$\displaystyle\begin{split}\displaystyle Y_{\pi Kch}^{\pi K}(t,u)&\displaystyle%
=-\gamma_{\pi K}\Big{(}t\left((10\beta_{\pi K}-13\gamma_{\pi K})t+(2\beta_{\pi
K%
}+\gamma_{\pi K})u\right)\\
&\displaystyle\qquad\quad-4((4(\beta_{\pi K}-\gamma_{\pi K})+3)(M_{K}^{2}+M_{%
\pi}^{2})+6(\alpha_{\pi K}-1)M_{K}M_{\pi})t\\
&\displaystyle\qquad\quad\quad-2(2\beta_{\pi K}+\gamma_{\pi K})(M_{K}^{2}+M_{%
\pi}^{2})u+(14\beta_{\pi K}+5\gamma_{\pi K}+8)(M_{K}^{4}+M_{\pi}^{4})\\
&\displaystyle\qquad\quad\quad\quad+2(2\beta_{\pi K}-13\gamma_{\pi K}+8)M_{K}^%
{2}M_{\pi}^{2}+16(\alpha_{\pi K}-1)M_{K}M_{\pi}(M_{K}^{2}+M_{\pi}^{2})\Big{)},%
\end{split}$$
(119)
$$\displaystyle X_{\pi Kch}^{\eta K}(u)$$
$$\displaystyle=X_{\pi K}^{\eta K}(u),$$
(120)
$$\displaystyle\begin{split}\displaystyle X_{\pi Kch}^{\pi K}(u)&\displaystyle=-%
\frac{1}{3}\gamma_{\pi K}(M_{K}^{2}-M_{\pi}^{2})^{2}\Big{(}\left(2\beta_{\pi K%
}+\gamma_{\pi K}\right)u+8\left(\alpha_{\pi K}-1\right)M_{K}M_{\pi}\\
&\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad\qquad+2\left(2-2\beta_{\pi K%
}-3\gamma_{\pi K}\right)\left(M_{K}^{2}+M_{\pi}^{2}\right)\Big{)},\end{split}$$
(121)
$$\displaystyle W_{\pi Kch}^{\eta K}$$
$$\displaystyle=W_{\pi K}^{\eta K},$$
(122)
$$\displaystyle W_{\pi Kch}^{\pi K}$$
$$\displaystyle=\frac{1}{3}\gamma_{\pi K}\Delta_{\scriptscriptstyle K\pi}^{2}%
\left(2\beta_{\pi K}+\gamma_{\pi K}\right).$$
(123)
$\bullet\quad K^{-}K^{+}\rightarrow{\overline{K^{0}}}K^{0}$
$$\begin{split}\displaystyle A_{KK}&\displaystyle=\frac{1}{6F_{\pi}^{2}}\left(%
\beta_{KK}\left(4M_{K}^{2}-3u\right)+3\gamma_{KK}(s-t)+2\alpha_{KK}M_{K}^{2}%
\right)+\frac{1}{6F_{\pi}^{4}}\left(\delta_{KK}s^{2}+\varepsilon_{KK}t^{2}+%
\varphi_{KK}st\right)\\
&\displaystyle\quad+\frac{1}{288F_{\pi}^{4}}\Big{(}Z_{KK}^{KK}(s,t)\overline{J%
}_{KK}(s)+Z^{\pi\eta}_{KK}(s)\overline{J}_{\pi\eta}(s)+Z^{\eta\eta}_{KK}(s)%
\overline{J}_{\eta\eta}(s)+Z_{KK}^{\pi\pi}(s,t)\overline{J}_{\pi\pi}(s)\Big{)}%
\\
&\displaystyle\quad+\frac{1}{72F_{\pi}^{4}}\left(Y_{KK}^{KK}(s,t)\overline{J}_%
{KK}(t)+Y_{KK}^{\pi\eta}(t)\overline{J}_{\pi\eta}(t)+Y_{KK}^{\pi\pi}(s,t)%
\overline{J}_{\pi\pi}(t)\right)\\
&\displaystyle\quad+\frac{1}{36F_{\pi}^{4}}\overline{J}_{KK}(u)\left(3\beta_{%
KK}u-2(\alpha_{KK}+2\beta_{KK})M_{K}^{2}\right)^{2}+O\!\left(p^{5}\right).\end%
{split}$$
(124)
$$\displaystyle Z_{KK}^{KK}(s,t)$$
$$\displaystyle=8\Big{(}\left(3\beta_{KK}s+4(\alpha_{KK}-\beta_{KK})M_{K}^{2}%
\right)^{2}+3\beta_{KK}^{2}(s-4M_{K}^{2})(s+2t-4M_{K}^{2})\Big{)}$$
(125)
$$\displaystyle Z^{\pi\eta}_{KK}(s)$$
$$\displaystyle=-6\Big{(}3\beta_{\eta\pi K}s-2(1+\beta_{\eta\pi K})M_{K}^{2}+(1-%
\alpha_{\eta\pi K}-\beta_{\eta\pi K})M_{\pi}^{2}+(1-\beta_{\eta\pi K})M_{\eta}%
^{2}\Big{)}^{2}$$
(126)
$$\displaystyle Z^{\eta\eta}_{KK}(s)$$
$$\displaystyle=\left(9\beta_{\eta K}s-2(3\beta_{\eta K}+\alpha_{\eta K})M_{K}^{%
2}+6(\alpha_{\eta K}-\beta_{\eta K})M_{\eta}^{2}\right)^{2}$$
(127)
$$\displaystyle\begin{split}\displaystyle Z_{KK}^{\pi\pi}(s,t)&\displaystyle=3%
\Big{(}s\left((9\beta_{\pi K}^{2}-2\gamma_{\pi K}^{2})s-4\gamma_{\pi K}^{2}t%
\right)+4(1-\beta_{\pi K})^{2}(M_{K}^{4}+M_{\pi}^{4})\\
&\displaystyle\qquad+4s\left(3\beta_{\pi K}(1-\beta_{\pi K})+2\gamma_{\pi K}^{%
2}\right)(M_{K}^{2}+M_{\pi}^{2})\\
&\displaystyle\qquad\quad+8(2\gamma_{\pi K}^{2}t+((1-\beta_{\pi K})^{2}+2(1-%
\alpha_{\pi K})^{2}-4\gamma_{\pi K}^{2})M_{K}^{2})M_{\pi}^{2}\\
&\displaystyle\qquad\quad\quad+8(\alpha_{\pi K}-1)(3\beta_{\pi K}s+2(1-\beta_{%
\pi K})(M_{K}^{2}+M_{\pi}^{2}))M_{\pi}M_{K}\Big{)}\end{split}$$
(128)
$$\displaystyle Y_{KK}^{KK}(s,t)$$
$$\displaystyle=\frac{1}{16}Z_{KK}^{KK}(t,s)$$
(129)
$$\displaystyle Y_{KK}^{\pi\eta}(t)$$
$$\displaystyle=3\Big{(}3\beta_{\eta\pi K}t-2(1+\beta_{\eta\pi K})M_{K}^{2}+(1-%
\alpha_{\eta\pi K}-\beta_{\eta\pi K})M_{\pi}^{2}+(1-\beta_{\eta\pi K})M_{\eta}%
^{2}\Big{)}^{2}$$
(130)
$$\displaystyle Y_{KK}^{\pi\pi}(s,t)$$
$$\displaystyle=3\gamma_{\pi K}^{2}(2s+t-4M_{K}^{2})(t-4M_{\pi}^{2})$$
(131)
Appendix C Standard chiral perturbation theory $O\!\left(p^{4}\right)$ values of the polynomial parameters
In this appendix we give the values of our polynomial parameters which reproduce the $O\!\left(p^{4}\right)$ results of the standard chiral perturbation theory. To that end we have used the computation of [10], which contains all the considered amplitudes and its advantage is also that their results are in the unitary form.
The form of our results is in terms of physical observables and thus has to be scale-independent and the same for all the possible regularization schemes. The only thing which can change by an eventual change of the scheme is the relation between the parameters of our parametrization and
the (renormalised) constants of the Lagrangian theory (LEC). By the change of the scale, the values of the LEC change, but their combinations giving the value of our parameters remain scale-independent - what can be another test of the results obtained from the Lagrangian theory. A further difference between (Lagrangian theory) results of different authors can rise from a different choice of the way they parameterize the $O\!\left(p^{2}\right)$ constants (bare masses and decay constants) using the physical parameters, e.g. in [10] they expand $F_{K}$ and $F_{\eta}$ decay constants in terms of $F_{\pi}$, $L_{4}^{r}$ and $L_{5}^{r}$.
In [10] they have used the Gell-Mann-Okubo relation (GMO) to get their results more simplified. In the standard chiral power counting the GMO formula has correction of the $O\!\left(p^{4}\right)$ order and thus we could also use it to simplify our results in some places, where it would give only correction of the $O\!\left(p^{6}\right)$ order at least. However, to let these relations be closely connected to the results [10], we have not done it and the only place where we refer to the GMO formula (and its $O\!\left(p^{4}\right)$ order correction) is in those places where we want to emphasize the validity of the $O\!\left(p^{2}\right)$ values of the parameters from Appendix A. Nevertheless, the use of
$$\Delta_{\mathrm{GMO}}(M_{\eta}^{2}-M_{\pi}^{2})=4M_{K}^{2}-M_{\pi}^{2}-3M_{%
\eta}^{2},$$
(132)
which is in the standard power counting of the $O\!\left(p^{4}\right)$ order, can also be understood just as a (more complicated) notation of the right-hand side of this definition.
Other objects appearing in the relations are the chiral logarithms, given in accordance with [10] by
$$\mu_{i}=\frac{M_{i}^{2}}{32\pi^{2}F_{\pi}^{2}}\log\frac{M_{i}^{2}}{\mu^{2}}%
\qquad\text{with }i=\pi,K,\eta.$$
(133)
(As has been already stated, its dependence on the scale $\mu$ is compensated on the right-hand side of the following relations by the scale-dependence of LEC $L_{i}^{r}$ as listed in [9].)
By the comparison of [10] with our results from the previous appendix we get the following relations:
$\bullet\quad\eta\eta$
$$\displaystyle\delta_{\eta\eta}$$
$$\displaystyle=12(2L_{1}^{r}+2L_{2}^{r}+L_{3})-\frac{27F_{\pi}^{2}\mu_{K}}{4M_{%
K}^{2}}-\frac{27}{128\pi^{2}}\,,$$
(134)
$$\displaystyle\begin{split}\displaystyle F_{\pi}^{2}(\alpha_{\eta\eta}-1)(M_{%
\pi}^{2}-4M_{\eta}^{2})&\displaystyle=-\frac{4}{3}F_{\pi}^{2}\Delta_{\mathrm{%
GMO}}(M_{\eta}^{2}-M_{\pi}^{2})+96M_{\eta}^{4}(L_{4}^{r}-3L_{6}^{r})\\
&\displaystyle+8L_{5}^{r}\left(3M_{\pi}^{4}-10M_{\eta}^{2}M_{\pi}^{2}+13M_{%
\eta}^{4}\right)\\
&\displaystyle-192L_{7}\left(M_{\pi}^{4}-3M_{\eta}^{2}M_{\pi}^{2}+2M_{\eta}^{4%
}\right)-48L_{8}^{r}\left(2M_{\pi}^{4}-6M_{\eta}^{2}M_{\pi}^{2}+7M_{\eta}^{4}%
\right)\\
&\displaystyle+\frac{\left(7M_{\pi}^{4}-24M_{\pi}^{2}M_{\eta}^{2}-54M_{\eta}^{%
4}\right)\mu_{K}F_{\pi}^{2}}{3M_{K}^{2}}+\frac{\left(48M_{\eta}^{2}-7M_{\pi}^{%
2}\right)\mu_{\pi}F_{\pi}^{2}}{3}\\
&\displaystyle+\frac{\left(M_{\pi}^{4}-8M_{\eta}^{2}M_{\pi}^{2}+24M_{\eta}^{4}%
\right)\mu_{\eta}F_{\pi}^{2}}{M_{\eta}^{2}}+\frac{80M_{K}^{4}-176M_{\eta}^{2}M%
_{K}^{2}+103M_{\eta}^{4}}{32\pi^{2}}\,.\end{split}$$
(135)
$\bullet\quad\pi\eta$
$$\displaystyle\varepsilon_{\pi\eta}$$
$$\displaystyle=24L_{1}^{r}+4L_{3}-\frac{9F_{\pi}^{2}\mu_{K}}{4M_{K}^{2}}-\frac{%
9}{128\pi^{2}}\,,$$
(136)
$$\displaystyle\delta_{\pi\eta}$$
$$\displaystyle=12L_{2}^{r}+4L_{3}-\frac{9F_{\pi}^{2}\mu_{K}}{4M_{K}^{2}}-\frac{%
9}{128\pi^{2}}\,,$$
(137)
$$\displaystyle F_{\pi}^{2}\beta_{\pi\eta}$$
$$\displaystyle=8L_{4}^{r}\left(M_{\pi}^{2}+M_{\eta}^{2}\right)-\frac{M_{\pi}^{2%
}+M_{\eta}^{2}}{32\pi^{2}}-\frac{\left(M_{\pi}^{2}+3M_{\eta}^{2}\right)F_{\pi}%
^{2}\mu_{K}}{3M_{K}^{2}}-\frac{2F_{\pi}^{2}\mu_{\pi}}{3}\,,$$
(138)
$$\displaystyle\begin{split}\displaystyle F_{\pi}^{2}(\alpha_{\pi\eta}-1)M_{\pi}%
^{2}&\displaystyle=16L_{4}^{r}\left(M_{\pi}^{4}-4M_{\eta}^{2}M_{\pi}^{2}+M_{%
\eta}^{4}\right)+16M_{\eta}^{2}M_{\pi}^{2}(-L_{5}^{r}+6L_{6}^{r})\\
&\displaystyle+96L_{7}\left(M_{\pi}^{2}-M_{\eta}^{2}\right)M_{\pi}^{2}+48L_{8}%
^{r}M_{\pi}^{4}\\
&\displaystyle-\frac{\left(5M_{\pi}^{4}-28M_{\eta}^{2}M_{\pi}^{2}+6M_{\eta}^{4%
}\right)F_{\pi}^{2}\mu_{K}}{3M_{K}^{2}}+\frac{\left(-17M_{\pi}^{4}+9M_{\eta}^{%
2}M_{\pi}^{2}+4M_{\eta}^{4}\right)F_{\pi}^{2}\mu_{\pi}}{3\left(M_{\pi}^{2}-M_{%
\eta}^{2}\right)}\\
&\displaystyle+\frac{\left(M_{\pi}^{4}-M_{\eta}^{2}M_{\pi}^{2}+4M_{\eta}^{4}%
\right)M_{\pi}^{2}F_{\pi}^{2}\mu_{\eta}}{3M_{\eta}^{2}(M_{\pi}^{2}-M_{\eta}^{2%
})}-\frac{8M_{\pi}^{4}-11M_{\eta}^{2}M_{\pi}^{2}+6M_{\eta}^{4}}{96\pi^{2}}\,.%
\end{split}$$
(139)
$\bullet\quad\pi\pi$
$$\displaystyle\varepsilon_{\pi\pi}$$
$$\displaystyle=4L_{2}^{r}+\frac{1}{12}F_{\pi}^{2}\left(-\frac{\mu_{K}}{M_{K}^{2%
}}-\frac{8\mu_{\pi}}{M_{\pi}^{2}}\right)-\frac{7}{384\pi^{2}}\,,$$
(140)
$$\displaystyle\delta_{\pi\pi}$$
$$\displaystyle=8L_{1}^{r}+4L_{3}+\frac{1}{12}F_{\pi}^{2}\left(-\frac{\mu_{K}}{M%
_{K}^{2}}-\frac{8\mu_{\pi}}{M_{\pi}^{2}}\right)-\frac{13}{384\pi^{2}}\,,$$
(141)
$$\displaystyle F_{\pi}^{2}(\beta_{\pi\pi}-1)$$
$$\displaystyle=8\left(2L_{4}^{r}+L_{5}^{r}\right)M_{\pi}^{2}-\frac{F_{\pi}^{2}%
\mu_{K}M_{\pi}^{2}}{M_{K}^{2}}-\frac{5M_{\pi}^{2}}{32\pi^{2}}-4F_{\pi}^{2}\mu_%
{\pi},$$
(142)
$$\displaystyle F_{\pi}^{2}(\alpha_{\pi\pi}-1)M_{\pi}^{2}$$
$$\displaystyle=-16\left(2L_{4}^{r}+L_{5}^{r}-6L_{6}^{r}-3L_{8}^{r}\right)M_{\pi%
}^{4}-\frac{F_{\pi}^{2}\mu_{\eta}M_{\pi}^{4}}{3M_{\eta}^{2}}-\frac{F_{\pi}^{2}%
\mu_{K}M_{\pi}^{4}}{M_{K}^{2}}-\frac{7M_{\pi}^{4}}{96\pi^{2}}-F_{\pi}^{2}\mu_{%
\pi}M_{\pi}^{2}.$$
(143)
$\bullet\quad\eta K$
$$\displaystyle\varepsilon_{\eta K}$$
$$\displaystyle=16L_{2}^{r}+\frac{4}{3}L_{3}-\frac{F_{\pi}^{2}(2\mu_{K}+\mu_{\pi%
}+3\mu_{\eta})}{2\left(M_{K}^{2}-M_{\eta}^{2}\right)}+\frac{1}{64\pi^{2}}\,,$$
(144)
$$\displaystyle\delta_{\eta K}$$
$$\displaystyle=32L_{1}^{r}+\frac{40}{3}L_{3}+\frac{\left(9M_{\eta}^{2}-8M_{K}^{%
2}\right)F_{\pi}^{2}\mu_{K}}{2M_{K}^{2}\left(M_{K}^{2}-M_{\eta}^{2}\right)}+%
\frac{F_{\pi}^{2}\left(\mu_{\pi}-3\mu_{\eta}\right)}{4(M_{K}^{2}-M_{\eta}^{2})%
}-\frac{11}{64\pi^{2}}\,,$$
(145)
$$\displaystyle\begin{split}\displaystyle F_{\pi}^{2}(\beta_{\eta K}-1)&%
\displaystyle=\frac{32}{3}\left(M_{K}^{2}+M_{\eta}^{2}\right)L_{4}^{r}+8L_{5}^%
{r}M_{\pi}^{2}-\frac{M_{\eta}^{2}\left(7M_{K}^{2}-9M_{\eta}^{2}\right)F_{\pi}^%
{2}\mu_{K}}{3M_{K}^{2}(M_{K}^{2}-M_{\eta}^{2})}\\
&\displaystyle+\frac{\left(13M_{\eta}^{2}-11M_{K}^{2}\right)F_{\pi}^{2}\mu_{%
\pi}}{3\left(M_{K}^{2}-M_{\eta}^{2}\right)}+\frac{\left(4M_{K}^{4}-15M_{\eta}^%
{2}M_{K}^{2}+7M_{\eta}^{4}\right)F_{\pi}^{2}\mu_{\eta}}{3M_{\eta}^{2}(M_{K}^{2%
}-M_{\eta}^{2})}\\
&\displaystyle-\frac{17M_{\eta}^{2}+37M_{K}^{2}}{192\pi^{2}}\,,\end{split}$$
(146)
$$\displaystyle\begin{split}\displaystyle F_{\pi}^{2}(\alpha_{\eta K}-1)\left(3M%
_{\eta}^{2}-M_{K}^{2}\right)&\displaystyle=F_{\pi}^{2}\Delta_{\mathrm{GMO}}(M_%
{\eta}^{2}-M_{\pi}^{2})+32L_{4}^{r}\left(M_{K}^{4}-4M_{\eta}^{2}M_{K}^{2}+M_{%
\eta}^{4}\right)\\
&\displaystyle+192L_{6}^{r}M_{K}^{2}M_{\eta}^{2}+4L_{5}^{r}\left(-3M_{\pi}^{4}%
+6M_{K}^{2}M_{\pi}^{2}+M_{\eta}^{2}M_{\pi}^{2}-12M_{\eta}^{4}\right)\\
&\displaystyle+48L_{7}\left(M_{\pi}^{4}-4M_{\eta}^{2}M_{\pi}^{2}+3M_{\eta}^{4}%
\right)+24L_{8}^{r}\left(M_{\pi}^{4}-3M_{\eta}^{2}M_{\pi}^{2}+6M_{\eta}^{4}%
\right)\\
&\displaystyle-\frac{\left(14M_{K}^{6}-84M_{\eta}^{2}M_{K}^{4}+102M_{\eta}^{4}%
M_{K}^{2}-27M_{\eta}^{6}\right)F_{\pi}^{2}\mu_{K}}{3(M_{K}^{2}-M_{\eta}^{2})}%
\\
&\displaystyle-\frac{\left(M_{K}^{4}+12M_{\eta}^{2}M_{K}^{2}-14M_{\eta}^{4}%
\right)F_{\pi}^{2}\mu_{\pi}}{2(M_{K}^{2}-M_{\eta}^{2})}\\
&\displaystyle+\frac{\left(8M_{K}^{6}-41M_{\eta}^{2}M_{K}^{4}+104M_{\eta}^{4}M%
_{K}^{2}-78M_{\eta}^{6}\right)F_{\pi}^{2}\mu_{\eta}}{6M_{\eta}^{2}(M_{K}^{2}-M%
_{\eta}^{2})}\\
&\displaystyle-\frac{31M_{K}^{4}-55M_{\eta}^{2}M_{K}^{2}+39M_{\eta}^{4}}{96\pi%
^{2}}\,.\end{split}$$
(147)
$\bullet\quad\eta\pi K$
$$\displaystyle\varepsilon_{\eta\pi K}$$
$$\displaystyle=-4L_{3}+\frac{6\mu_{K}F_{\pi}^{2}}{M_{K}^{2}-M_{\pi}^{2}}-\frac{%
3\mu_{\pi}F_{\pi}^{2}}{2\left(M_{K}^{2}-M_{\pi}^{2}\right)}+\frac{3\mu_{\eta}F%
_{\pi}^{2}}{2\left(M_{K}^{2}-M_{\eta}^{2}\right)}+\frac{1}{64\pi^{2}}\,,$$
(148)
$$\displaystyle\delta_{\eta\pi K}$$
$$\displaystyle=8L_{3}+\frac{3\left(M_{\pi}^{2}-2M_{K}^{2}\right)\mu_{K}F_{\pi}^%
{2}}{2M_{K}^{2}\left(M_{K}^{2}-M_{\pi}^{2}\right)}-\frac{3\mu_{\pi}F_{\pi}^{2}%
}{4\left(M_{K}^{2}-M_{\pi}^{2}\right)}-\frac{3\mu_{\eta}F_{\pi}^{2}}{4\left(M_%
{K}^{2}-M_{\eta}^{2}\right)}-\frac{5}{64\pi^{2}}\,,$$
(149)
$$\displaystyle\begin{split}\displaystyle F_{\pi}^{2}(\beta_{\eta\pi K}-1)&%
\displaystyle=\frac{4}{3}L_{3}\Delta_{\mathrm{GMO}}(M_{\eta}^{2}-M_{\pi}^{2})+%
8M_{\pi}^{2}L_{5}^{r}\\
&\displaystyle+\frac{\left(-32M_{K}^{4}+2M_{K}^{2}(2M_{\pi}^{2}-3M_{\eta}^{2})%
+3M_{\pi}^{2}(M_{\pi}^{2}+M_{\eta}^{2})\right)F_{\pi}^{2}\mu_{K}}{6\left(M_{K}%
^{2}-M_{\pi}^{2}\right)}\\
&\displaystyle+\frac{\left(-40M_{K}^{2}+9M_{\eta}^{2}+33M_{\pi}^{2}\right)F_{%
\pi}^{2}\mu_{\pi}}{12\left(M_{K}^{2}-M_{\pi}^{2}\right)}-\frac{\left(M_{\pi}^{%
2}+5M_{\eta}^{2}\right)F_{\pi}^{2}\mu_{\eta}}{4(M_{K}^{2}-M_{\eta}^{2})}+\frac%
{-57M_{K}^{2}+4M_{\pi}^{2}+27M_{\eta}^{2}}{192\pi^{2}}\,,\end{split}$$
(150)
$$\displaystyle\begin{split}\displaystyle F_{\pi}^{2}\alpha_{\eta\pi K}M_{\pi}^{%
2}&\displaystyle=-\frac{F_{\pi}^{2}}{3}\Delta_{\mathrm{GMO}}(M_{\eta}^{2}-M_{%
\pi}^{2})+\frac{8}{3}L_{3}\left(8M_{K}^{2}+M_{\pi}^{2}+M_{\eta}^{2}\right)%
\Delta_{\mathrm{GMO}}(M_{\eta}^{2}-M_{\pi}^{2})\\
&\displaystyle+4\left(M_{\pi}^{4}-4M_{K}^{2}M_{\pi}^{2}+4M_{\eta}^{2}M_{\pi}^{%
2}-M_{\eta}^{4}\right)L_{5}^{r}+24\left(M_{\eta}^{4}-M_{\pi}^{4}\right)(2L_{7}%
+L_{8}^{r})\\
&\displaystyle\hskip-28.452756pt+\frac{\left(40M_{K}^{6}+8\left(M_{\eta}^{2}-1%
0M_{\pi}^{2}\right)M_{K}^{4}-3M_{\pi}^{2}(M_{\pi}^{2}+M_{\eta}^{2})^{2}+2M_{K}%
^{2}(17M_{\pi}^{4}+25M_{\pi}^{2}M_{\eta}^{2}-6M_{\eta}^{4})\right)F_{\pi}^{2}%
\mu_{K}}{6\left(M_{K}^{2}-M_{\pi}^{2}\right)}\\
&\displaystyle+\frac{\left(-70M_{K}^{4}+2\left(22M_{\pi}^{2}+19M_{\eta}^{2}%
\right)M_{K}^{2}-19M_{\pi}^{4}+9M_{\eta}^{4}-12M_{\pi}^{2}M_{\eta}^{2}\right)F%
_{\pi}^{2}\mu_{\pi}}{6\left(M_{K}^{2}-M_{\pi}^{2}\right)}\\
&\displaystyle+\frac{\left(14M_{K}^{4}-M_{\pi}^{4}+21M_{\eta}^{4}+6M_{\pi}^{2}%
M_{\eta}^{2}-2M_{K}^{2}\left(M_{\pi}^{2}+18M_{\eta}^{2}\right)\right)F_{\pi}^{%
2}\mu_{\eta}}{2\left(M_{K}^{2}-M_{\eta}^{2}\right)}\\
&\displaystyle+\frac{-408M_{K}^{4}+\left(73M_{\pi}^{2}+387M_{\eta}^{2}\right)M%
_{K}^{2}+8M_{\pi}^{4}-63M_{\eta}^{4}-M_{\pi}^{2}M_{\eta}^{2}}{192\pi^{2}}\,.%
\end{split}$$
(151)
$\bullet\quad\pi K$
$$\displaystyle\varphi_{\pi K}$$
$$\displaystyle=-4L_{3}+\frac{\left(M_{\pi}^{2}-4M_{K}^{2}\right)F_{\pi}^{2}\mu_%
{K}}{6M_{K}^{2}\left(M_{K}^{2}-M_{\pi}^{2}\right)}+\frac{\left(M_{\pi}^{2}-4M_%
{K}^{2}\right)F_{\pi}^{2}\mu_{\pi}}{12M_{\pi}^{2}\left(M_{K}^{2}-M_{\pi}^{2}%
\right)}-\frac{F_{\pi}^{2}\mu_{\eta}}{M_{\pi}^{2}-M_{\eta}^{2}}\,,$$
(152)
$$\displaystyle\varepsilon_{\pi K}$$
$$\displaystyle=12(4L_{2}^{r}+L_{3})-\frac{3F_{\pi}^{2}\mu_{K}}{M_{K}^{2}-M_{\pi%
}^{2}}+\frac{15F_{\pi}^{2}\mu_{\pi}}{2(M_{K}^{2}-M_{\pi}^{2})}+\frac{6F_{\pi}^%
{2}\mu_{\eta}}{M_{\pi}^{2}-M_{\eta}^{2}}+\frac{1}{64\pi^{2}}\,,$$
(153)
$$\displaystyle\begin{split}\displaystyle\delta_{\pi K}&\displaystyle=24(4L_{1}^%
{r}+L_{3})+\frac{3\left(3M_{\pi}^{2}-4M_{K}^{2}\right)F_{\pi}^{2}\mu_{K}}{2M_{%
K}^{2}\left(M_{K}^{2}-M_{\pi}^{2}\right)}+\frac{3\left(7M_{\pi}^{2}-8M_{K}^{2}%
\right)F_{\pi}^{2}\mu_{\pi}}{4M_{\pi}^{2}\left(M_{K}^{2}-M_{\pi}^{2}\right)}\\
&\displaystyle-\frac{3F_{\pi}^{2}\mu_{\eta}}{M_{\pi}^{2}-M_{\eta}^{2}}-\frac{2%
3}{64\pi^{2}}\,,\end{split}$$
(154)
$$\displaystyle\begin{split}\displaystyle F_{\pi}^{2}(\gamma_{\pi K}-1)&%
\displaystyle=8M_{\pi}^{2}L_{5}^{r}-\frac{2M_{K}^{2}F_{\pi}^{2}\mu_{K}}{M_{K}^%
{2}-M_{\pi}^{2}}-\frac{\left(M_{K}^{2}-5M_{\pi}^{2}\right)F_{\pi}^{2}\mu_{\pi}%
}{2(M_{K}^{2}-M_{\pi}^{2})}\\
&\displaystyle+\frac{\left(2M_{K}^{2}+M_{\pi}^{2}-3M_{\eta}^{2}\right)F_{\pi}^%
{2}\mu_{\eta}}{M_{\pi}^{2}-M_{\eta}^{2}}-\frac{5M_{K}^{2}+M_{\pi}^{2}}{192\pi^%
{2}}\,,\end{split}$$
(155)
$$\displaystyle\begin{split}\displaystyle F_{\pi}^{2}(\beta_{\pi K}-1)&%
\displaystyle=32\left(M_{K}^{2}+M_{\pi}^{2}\right)L_{4}^{r}+8L_{5}^{r}M_{\pi}^%
{2}-\frac{15M_{\pi}^{2}+19M_{K}^{2}}{64\pi^{2}F_{\pi}^{2}}\\
&\displaystyle+\frac{\left(3M_{\pi}^{4}+3M_{K}^{2}M_{\pi}^{2}-8M_{K}^{4}\right%
)F_{\pi}^{2}\mu_{K}}{M_{K}^{2}(M_{K}^{2}-M_{\pi}^{2})}-\frac{\left(3M_{\pi}^{4%
}-3M_{K}^{2}M_{\pi}^{2}-4M_{K}^{4}\right)F_{\pi}^{2}\mu_{\pi}}{2M_{\pi}^{2}(M_%
{K}^{2}-M_{\pi}^{2})}\\
&\displaystyle-\frac{\left(2M_{\pi}^{4}+9M_{\eta}^{4}+\left(4M_{K}^{2}+M_{\pi}%
^{2}\right)M_{\eta}^{2}\right)F_{\pi}^{2}\mu_{\eta}}{2M_{\eta}^{2}\left(M_{\pi%
}^{2}-M_{\eta}^{2}\right)}\,,\end{split}$$
(156)
$$\displaystyle\begin{split}\displaystyle F_{\pi}^{2}(\alpha_{\pi K}-1)M_{\pi}M_%
{K}&\displaystyle=16L_{4}^{r}\left(M_{K}^{4}-4M_{\pi}^{2}M_{K}^{2}+M_{\pi}^{4}%
\right)+4L_{5}^{r}M_{\pi}^{2}\left(M_{\pi}^{2}-5M_{K}^{2}\right)+48M_{K}^{2}M_%
{\pi}^{2}(2L_{6}^{r}+L_{8}^{r})\\
&\displaystyle-\frac{21M_{K}^{4}-25M_{\pi}^{2}M_{K}^{2}+21M_{\pi}^{4}}{192\pi^%
{2}}+\frac{\left(2M_{K}^{6}-4M_{K}^{4}M_{\pi}^{2}+6M_{\pi}^{4}M_{K}^{2}-3M_{%
\pi}^{6}\right)F_{\pi}^{2}\mu_{K}}{2M_{K}^{2}(M_{K}^{2}-M_{\pi}^{2})}\\
&\displaystyle-\frac{\left(8M_{K}^{6}-13M_{K}^{4}M_{\pi}^{2}+12M_{\pi}^{4}M_{K%
}^{2}-6M_{\pi}^{6}\right)F_{\pi}^{2}\mu_{\pi}}{4M_{\pi}^{2}(M_{K}^{2}-M_{\pi}^%
{2})}\\
&\displaystyle\hskip-22.762205pt\hskip-42.679134pt-\frac{\left(17M_{\pi}^{6}-2%
4M_{K}^{2}M_{\pi}^{4}+3\left(27M_{\eta}^{2}+7M_{\pi}^{2}-36M_{K}^{2}\right)M_{%
\eta}^{4}+\left(24M_{K}^{4}+12M_{\pi}^{2}M_{K}^{2}+M_{\pi}^{4}\right)M_{\eta}^%
{2}\right)F_{\pi}^{2}\mu_{\eta}}{24M_{\eta}^{2}\left(M_{\pi}^{2}-M_{\eta}^{2}%
\right)}\,.\end{split}$$
(157)
$\bullet\quad KK$
$$\displaystyle\varphi_{KK}$$
$$\displaystyle=24L_{2}^{r}-\frac{17F_{\pi}^{2}\mu_{K}}{4M_{K}^{2}}-\frac{F_{\pi%
}^{2}\mu_{\pi}}{4M_{\pi}^{2}}-\frac{7}{64\pi^{2}}\,,$$
(158)
$$\displaystyle\varepsilon_{KK}$$
$$\displaystyle=12(2L_{2}^{r}+L_{3})-\frac{F_{\pi}^{2}\mu_{K}}{4M_{K}^{2}}-\frac%
{F_{\pi}^{2}}{8}\left(\frac{2}{M_{\pi}^{2}}-\frac{27}{M_{K}^{2}-M_{\pi}^{2}}%
\right)\mu_{\pi}+\frac{9F_{\pi}^{2}\mu_{\eta}}{2\left(M_{\pi}^{2}-M_{\eta}^{2}%
\right)}-\frac{1}{64\pi^{2}}\,,$$
(159)
$$\displaystyle\begin{split}\displaystyle\delta_{KK}&\displaystyle=12(4L_{1}^{r}%
+L_{3})-\frac{11F_{\pi}^{2}\mu_{K}}{4M_{K}^{2}}-\frac{F_{\pi}^{2}}{16}\left(%
\frac{10}{M_{\pi}^{2}}+\frac{27}{M_{K}^{2}-M_{\pi}^{2}}\right)\mu_{\pi}\\
&\displaystyle-\frac{9}{8}F_{\pi}^{2}\left(\frac{3}{M_{\eta}^{2}}+\frac{2}{M_{%
\pi}^{2}-M_{\eta}^{2}}\right)\mu_{\eta}-\frac{29}{128\pi^{2}}\,,\end{split}$$
(160)
$$\displaystyle\begin{split}\displaystyle F_{\pi}^{2}\gamma_{KK}&\displaystyle=1%
6M_{K}^{2}L_{4}^{r}-\frac{7M_{K}^{2}}{64\pi^{2}}-\frac{3F_{\pi}^{2}\mu_{K}}{2}%
-\frac{3M_{K}^{2}\left(2M_{K}^{2}+M_{\pi}^{2}\right)F_{\pi}^{2}\mu_{\pi}}{8M_{%
\pi}^{2}(M_{K}^{2}-M_{\pi}^{2})}\\
&\displaystyle+\frac{\left(M_{\pi}^{4}-3M_{\eta}^{4}+2M_{\pi}^{2}M_{\eta}^{2}+%
3\left(M_{\eta}^{2}-3M_{\pi}^{2}\right)M_{K}^{2}\right)F_{\pi}^{2}\mu_{\eta}}{%
4M_{\eta}^{2}\left(M_{\pi}^{2}-M_{\eta}^{2}\right)}\,,\end{split}$$
(161)
$$\displaystyle\begin{split}\displaystyle F_{\pi}^{2}(\beta_{KK}-1)&%
\displaystyle=16M_{K}^{2}L_{4}^{r}+8L_{5}^{r}M_{\pi}^{2}-\frac{9M_{K}^{2}}{64%
\pi^{2}}-\frac{3F_{\pi}^{2}\mu_{K}}{2}-\frac{\left(6M_{K}^{4}+11M_{\pi}^{2}M_{%
K}^{2}-20M_{\pi}^{4}\right)F_{\pi}^{2}\mu_{\pi}}{8M_{\pi}^{2}(M_{K}^{2}-M_{\pi%
}^{2})}\\
&\displaystyle+\frac{\left(M_{\pi}^{4}-9M_{\eta}^{4}+8M_{\pi}^{2}M_{\eta}^{2}+%
M_{K}^{2}\left(11M_{\eta}^{2}-9M_{\pi}^{2}\right)\right)F_{\pi}^{2}\mu_{\eta}}%
{4M_{\eta}^{2}\left(M_{\pi}^{2}-M_{\eta}^{2}\right)}\,,\end{split}$$
(162)
$$\displaystyle\begin{split}\displaystyle F_{\pi}^{2}(\alpha_{KK}-1)M_{K}^{2}&%
\displaystyle=16M_{K}^{4}(-2L_{4}^{r}+6L_{6}^{r}+3L_{8}^{r})+8L_{5}^{r}M_{K}^{%
2}\left(M_{\pi}^{2}-3M_{K}^{2}\right)\\
&\displaystyle-\frac{7M_{K}^{4}}{192\pi^{2}}-\frac{3M_{K}^{2}F_{\pi}^{2}\mu_{K%
}}{2}-\frac{M_{K}^{2}\left(6M_{K}^{4}+17M_{\pi}^{2}M_{K}^{2}-20M_{\pi}^{4}%
\right)F_{\pi}^{2}\mu_{\pi}}{8M_{\pi}^{2}(M_{K}^{2}-M_{\pi}^{2})}\\
&\displaystyle\hskip-42.679134pt-\frac{\left(M_{\pi}^{6}+2M_{\eta}^{2}M_{\pi}^%
{4}-3M_{\eta}^{4}M_{\pi}^{2}+3M_{K}^{4}(9M_{\pi}^{2}-7M_{\eta}^{2})-6M_{K}^{2}%
\left(2M_{\pi}^{4}+5M_{\eta}^{2}M_{\pi}^{2}-7M_{\eta}^{4}\right)\right)F_{\pi}%
^{2}\mu_{\eta}}{12M_{\eta}^{2}\left(M_{\eta}^{2}-M_{\pi}^{2}\right)}\,.\end{split}$$
(163)
In [10] they believe $K^{+}K^{-}\rightarrow K^{+}K^{-}$ to be independent on $K^{+}K^{-}\rightarrow K^{0}{\overline{K^{0}}}$, but we know that isospin structure (Fierz-like identities) and the crossing symmetry dictate the relation between these two processes given by (54) together with (53). Therefore, we have used the values of our parameters obtained from their $K^{+}K^{-}\rightarrow K^{0}{\overline{K^{0}}}$ amplitude to explicitly check their $K^{+}K^{-}\rightarrow K^{+}K^{-}$ result $T_{ch}$.
Appendix D Short comment on the assumptions of the analyticity of the amplitude and the dispersion
relations
The most important assumptions of the theorem – the existence of
dispersion relations and the analyticity of the amplitude and its
absorptive parts are results of a complicated theory of analytic
properties of scattering amplitudes, which is even older than QCD
itself. A good and still valid summary of it is the article by Sommer
[17].
We will not address this theory in more detail, just summarising
results interesting for us (details can be found in [17]
and [18]).
From the principles of axiomatic field theory (even without use of the unitarity of
S-matrix), Lehmann has proven that the amplitude with $s$ fixed at some physical value is
holomorphic in some finite region of the $u$-plane (if $s$ is above
the physical threshold, this region is the so-called small Lehmann
ellipse). Absorptive parts of amplitudes ($\operatorname{Im}{A(s,u)}$ for
$s\geq\Sigma$ and similarly for $t$) are holomorphic in the large
Lehmann ellipses in the $u$-plane (depending on $s$) and there this
absorptive part has also a convergent partial wave decomposition.
Further, the $N$-times subtracted $u$-fixed dispersion relations can
be proven on the intersection of those large Lehmann ellipses for
$s\geq\Sigma$ and $t\geq\tau$ (if not empty). Since its semiminor
axis tends to zero for $s\rightarrow\infty$, this intersection is
just an interval on the negative real $u$-axis.
Taking into account the unitarity, Martin and others have succeeded in
enlarging the validity of the dispersion relations into the circle
$|u|<R$ with some fixed radius $R$.
There exist further methods to enlarge the region of validity as
well as the analyticity domain. Using the ones given in
[17] we end up with the results of Table IV there. We are
interested in their specific application for our processes. If we
assume the isospin conservation and that all the mesons $\pi,K,\eta$ are thereby stable (and forget about resonances) we get the
validity of fixed $u$ dispersion relations within the regions
depicted in Fig. 1 (the intersections of Lehmann ellipses are
marked by bold lines there). These regions are glued ellipses and have
been obtained analytically in [18] using procedures from
[17]. Let us remind that these regions are only the minimal
domains, where the dispersive relations are valid, proven directly
from the axiomatic theory. Using further methods we could also
extend these regions - e.g. we have not taken into account the
further specific crossing (and Bose) symmetries of some of the
amplitudes. Finally let us once more emphasise that the described
method is ineffective if we allow e.g. the $\eta\rightarrow 3\pi$
decay and such cases have to be proven for unphysical masses of $\eta$ and then analytically continued in them.
Another assumption of our theorem was the existence of a point $s$, where the amplitude is analytic with respect to
all the values of variable $u$, for which the theorem should be valid - but in all the cases one can show that there exists a value $s$ for
which the small Lehmann ellipse is larger than the regions from the Fig. 1 and so we can conclude that both the assumptions
of the theorem are valid within these regions.
References
[1]
J. Bijnens, arXiv:hep-ph/0604043.
[2]
S. Weinberg, Physica A 96 (1979) 327.
[3]
J. Stern, H. Sazdjian, N. H. Fuchs, Phys. Rev. D 47 (1993) 3814 [arXiv:hep-ph/9301244].
[4]
M. Knecht, B. Moussallam, J. Stern, N. H. Fuchs, Nucl. Phys. B 457 (1995) 513 [arXiv:hep-ph/9507319].
[5]
M. Knecht and J. Stern, arXiv:hep-ph/9411253.
[6]
S. Pislak et al. [BNL-E865 Collaboration],
Phys. Rev. Lett. 87 (2001) 221801
[arXiv:hep-ex/0106071].
[7]
B. Ananthanarayan and P. Buettiker, Eur. Phys. J. C 19 (2001) 517 [arXiv:hep-ph/0012023];
B. Ananthanarayan, P. Buettiker and B. Moussallam, Eur. Phys. J. C 22 (2001) 133 [arXiv:hep-ph/0106230].
[8]
H. Osborn, Nucl. Phys. B 15 (1970) 501.
[9]
J. Gasser and H. Leutwyler, Nucl. Phys. B 250 (1985) 465.
[10]
A. Gomez Nicola and J. R. Pelaez, Phys. Rev. D 65 (2002) 054009 [arXiv:hep-ph/0109056].
[11]
J. Bros, H. Epstein and V. Glaser, Comm. Math. Phys. 1, (1965) 240.
[12]
J. Wess, B. Zumino, Phys. Lett. B 37 (1971)
95;
E. Witten, Nucl. Phys. B 223 (1983) 422.
[13]
Y. S. Jin and A. Martin, Phys. Rev. 136
(1964) B1375.
[14]
M. Froissart,
Phys. Rev. 123 (1961) 1053.
[15]
K. Kampf, M. Knecht, J. Novotný, M. Zdráhal, arXiv:0810.1906 [hep-ph].
[16]
K. Kampf, M. Knecht, J. Novotný, M. Zdráhal, in preparation.
[17]
G. Sommer, Fortsch. Phys. 18 (1970) 577.
[18]
M. Zdráhal, Ph.D. thesis - in preparation, Charles University. |
Description of nuclear octupole and quadrupole deformation
close to the axial symmetry and phase transitions in the octupole mode
P.G. Bizzeti
A.M. Bizzeti-Sona
Dipartimento di Fisica, Universitá di Firenze
I.N.F.N., Sezione di Firenze
Via G. Sansone 1, 50019 Sesto Fiorentino (Firenze), Italy
(December 4, 2020)
Abstract
The dynamics of nuclear collective motion is investigated in the case
of reflection-asymmetric shapes.
The model is based on a
new parameterization of the octupole and quadrupole degrees of
freedom, valid for nuclei close to the axial symmetry.
Amplitudes of oscillation in other degrees of freedom different from
the axial ones are assumed to be small, but not frozen to zero.
The case of nuclei which already possess a permanent quadrupole
deformation is discussed in some more detail and a simple solution is
obtained at the critical point of the phase transition
between harmonic octupole oscillation and a permanent asymmetric shape.
The results are compared with experimental data of the Thorium
isotopic chain. The isotope ${}^{226}$Th is found to be close to
the critical point.
Octupole deformation. Phase transition.
pacs: 21.60.ev
I Introduction
Phase transitions in nuclear shapes have been recently observed
at the boundary between regions characterized by different intrinsic
shapes of quadrupole deformation.
It has been shown by Iachello Iachello (2000a, b, 2001)
that new dynamic symmetries, called E(5), X(5) and Y(5) hold,
respectively, at the critical point
between spherical shape and $\gamma$–unstable deformation,
between spherical and deformed
axially symmetric shape and between deformed axial and triaxial shape.
Here, we are mainly interested in the second case.
First examples of X(5)
symmetry in transitional nuclei have been found
in ${}^{152}$Sm Casten and Zamfir (2001) and ${}^{150}$Nd
Krucken et al. (2002).
Other candidates for the X(5)
symmetry in different nuclear regions have been reported
later Bizzeti and Bizzeti-Sona (2002); Hutter et al. (2003); Clark et al. (2003); Fransen et al. (2004); Bizzeti (2003); Bijker et al. (2003, 2004); Tonev et al. (2004).
From the theoretical point of view, slightly different potentials have
been explored by Caprio Caprio (2004) and by Bonatsos
et al. Bonatsos et al. (2004), while the evolution of the $s=2$ band
when the lower border of the square well potential is displaced from
zero has been investigated by Pietralla and Gorbachenko Pietralla and Gorbachenko (2004).
A similar phase transition (i.e.,
from shape oscillation to
permanent deformation, conserving the axial symmetry of the system)
could take place also in the octupole degree of freedom. We have
found a possible example of such a phase transition
in the Thorium isotope chain, with the critical
point close to the mass 226.
In this case, the octupole mode is combined with a stable
quadrupole deformation.
Preliminary results have been reported in
two recent Conferences Bizzeti and Bizzeti-Sona (2004); Bizzeti (2003).
In order to provide a theoretical frame where to discuss the different
aspects of the octupole motion, we introduce a new
parameterization of the quadrupole and octupole degrees of freedom,
valid in conditions close to the axial symmetry. In this limit, a model
similar to the classical one by Bohr Bohr (1952) has been developed.
This is the subject of the first part of this paper.
In the second part, the model is used to discuss the evolution of the
octupole mode along the isotopic chain of Thorium, and the results are
compared with the experimental data.
II Theoretical frame for combined
quadrupole and
octupole excitations
II.1 Previous investigations of the Octupole
plus Quadrupole deformation
Reflection–asymmetric nuclear shapes have been discussed in a number
of papers, either in terms of surface Quadrupole + Octupole
deformation (Bohr geometrical approach) or with an extended
Interacting Boson Model (algebraic approach). The latter,
proposed in 1985 by Engel and
Iachello Engel and Iachello (1985), has been recently used by Alonso et
al. Alonso et al. (1995), Raduta et al. Raduta and Ionescu (2003); Raduta et al. (1996a),
Zamfir and Kusnezov Zamfir and Kusnezov (2001, 2003).
An alternative approach assuming $\alpha$–cluster configurations
has been discussed by Shneidman et al. Shneidman et al. (2002).
In the frame of the geometrical approach, a
number of theoretical investigations of the octupole vibrations
around a stable quadrupole deformation have been reported in the
last 50 years Rohozinski (1988); Butler and Nazarewicz (1996); Lipas and Davidson (1961); Denisov and Dzyublik (1995); Jolos and von Brentano (1999); Minkov et al. (2000, 2004).
Most of them, however, are limited to the case of axial symmetry.
This approach has been criticized, e.g., by Donner and
Greiner Donner and Greiner (1966), who have stressed the fact that all terms of a
given tensor order must be taken into account for a consistent
treatment.
To do this, Donner and Greiner renounce to the use of an “intrinsic
frame” referred to the principal axes of the overall tensor of
inertia and choose to define the octupole amplitudes in the
“intrinsic frame” of the quadrupole mode alone.
In this approach, definite predictions have been obtained at the limit
where the octupole deformations are “small” in comparison with the
quadrupole ones Eisenberg and Greiner (1987).
II.2 The new parameterization
Here we adopt a different approach, which can be useful also in the
case of comparable octupole and quadrupole deformation, close to the
axial–symmetry limit. Namely, we
choose as “intrinsic” reference frame the principal axes of
the overall tensor of inertia, as it results from the combined
quadrupole and octupole deformation.
The definitions of quadrupole and octupole amplitudes
$a^{(\lambda)}_{\mu}$,
with $\lambda=2,3$ and
$a^{(\lambda)}_{-\mu}=(-)^{\mu}\ {a^{(\lambda)}_{\mu}}^{*}$,
are recalled in the Appendix A. All these amplitudes are
defined in the (non inertial) intrinsic frame. To this purpose, in the
case of the quadrupole mode alone, it is
enough to assume
$a^{(2)}_{2}\!=a\!^{(2)}_{-2}$ real and $a^{(2)}_{\pm 1}\!=\!0$, with
the standard parameterization in terms of $\beta_{2}$ and $\gamma_{2}$:
$$\displaystyle a^{(2)}_{0}$$
$$\displaystyle=$$
$$\displaystyle\beta_{2}\cos\gamma_{2}$$
(1a)
$$\displaystyle a^{(2)}_{1}$$
$$\displaystyle=$$
$$\displaystyle 0$$
(1b)
$$\displaystyle a^{(2)}_{2}$$
$$\displaystyle=$$
$$\displaystyle\sqrt{1/2}\ \beta_{2}\ \sin\gamma_{2}\ .$$
(1c)
For the octupole mode
alone, a parameterization suitable to this purpose has been proposed
in 1999 by Wexler and Dussel Wexler and Dussel (1999). We adopt here a very
similar one:
$$\displaystyle a^{(3)}_{0}$$
$$\displaystyle=$$
$$\displaystyle\beta_{3}\ \cos\gamma_{3}$$
(2a)
$$\displaystyle a^{(3)}_{1}$$
$$\displaystyle=$$
$$\displaystyle-(5/2)\ \left(X+iY\right)\ \sin\gamma_{3}$$
(2b)
$$\displaystyle a^{(3)}_{2}$$
$$\displaystyle=$$
$$\displaystyle\sqrt{1/2}\ \beta_{3}\ \sin\gamma_{3}$$
(2c)
$$\displaystyle a^{(3)}_{3}$$
$$\displaystyle=$$
$$\displaystyle X\left[\cos\gamma_{3}+(\sqrt{15}/2)\ \sin\gamma_{3}\right]$$
(2d)
$$\displaystyle+$$
$$\displaystyle i\ Y\left[\cos\gamma_{3}-(\sqrt{15}/2)\ \sin\gamma_{3}\right]\ .$$
With this choice, the tensor of inertia turns out to be diagonal (see
Appendix A).
In both cases, one has to consider, in addition to the intrinsic
variables ($\beta_{2},\ \gamma_{2}$ for the quadrupole, or
$\beta_{3},\ \gamma_{3}\ ,X\ ,Y$ for the octupole), the three Euler angles
defining the orientation of the intrinsic frame in the laboratory
frame, in order to reach a number of parameters equal to the number of
degrees of freedom (5 for the quadrupole, 7 for the octupole).
Unfortunately, the situation is not so simple when quadrupole and
octupole modes are considered together, as the intrinsic frames of the
two modes do not necessarily coincide.
We shall limit our discussion to situations close to the axial
symmetry limit – in which, obviously, the two frames coincide – and
define a parametrization which automatically sets to zero the three
products of inertia $\mathcal{J}_{\kappa,\kappa^{\prime}}$
($\kappa\neq\kappa^{\prime}$) up to the first order in the
amplitudes of non–axial modes.
To this purpose we put
$$a^{(\lambda)}_{\mu}=\bar{a}^{(\lambda)}_{\mu}+\tilde{a}^{(\lambda)}_{\mu}$$
(3)
where $\bar{a}^{(\lambda)}_{\mu}$ are defined according to the
eq.s (1,2) and $\tilde{a}^{(\lambda)}_{\mu}$ are
correction terms, which are assumed to be small compared to the axial
amplitudes $a^{(\lambda)}_{0}$, but of the same order of magnitude as the
other non-axial terms.
It will be enough to consider these corrections only
for those amplitudes which, according to eq.s (1,2),
are either zero or small of the second order: the imaginary part
of $a^{(2)}_{2},\ a^{(3)}_{2}$ and the
real and imaginary parts of $a^{(2)}_{1},\ a^{(3)}_{1}$.
The six “new” first–order amplitudes added to those of
eq.s (1,2) are, however, not independent from one
another, if we choose as the reference systeme the one in which
the three
products of inertia turn out to be zero.
The expressions of the inertia tensor as a function of the
deformation parameters, obtained with the Bohr
assumptions Bohr (1952) of not–too–big deformations and
irrotational flow, are given in the Appendix A. In order to
simplify the notations, from now on we consider the inertia parameters
$B_{2},\ B_{3}$ included in our definitions of the amplitudes $a^{(\lambda)}_{\mu}$, which therefore correspond to
$\sqrt{B_{\lambda}}\ a^{(\lambda)}_{\mu}$ in the original Bohr notations.
From the eq.s (56), and retaining only terms
of the first order
in the small amplitudes $a^{(\lambda)}_{\mu}$ with $\mu\neq 0$, we
obtain the conditions
$$\displaystyle\mathcal{J}_{12}$$
$$\displaystyle=$$
$$\displaystyle-2\sqrt{6}\left(\beta_{2}{\rm Im}\ \tilde{a}^{(2)}_{2}+\sqrt{5}%
\beta_{3}{\rm Im}\ \tilde{a}^{(3)}_{2}\right)=0$$
(4a)
$$\displaystyle\mathcal{J}_{13}$$
$$\displaystyle+$$
$$\displaystyle i\mathcal{J}_{23}\ =\ \sqrt{6}\ \left(\beta_{2}\tilde{a}^{(2)}_{%
1}+\sqrt{2}\beta_{3}\tilde{a}^{(3)}_{1}\right)=0$$
(4b)
which are satisfied (at the leading order) if we put
$$\begin{array}[c]{lp{0.1mm}lp{0mm}l}\tilde{a}^{(2)}_{1}=\!\frac{-\sqrt{2}\beta_%
{3}}{\sqrt{\beta_{2}^{2}+2\beta_{3}^{2}}}\left(\eta+i\zeta\right)&\hbox{}&%
\tilde{a}^{(3)}_{1}=\!\frac{\beta_{2}}{\sqrt{\beta_{2}^{2}+2\beta_{3}^{2}}}%
\left(\eta+i\zeta\right)\\
{\rm Im}\;\tilde{a}^{(2)}_{2}=\!\frac{-\sqrt{5}\beta_{3}}{\sqrt{\beta_{2}^{2}+%
5\beta_{3}^{2}}}\xi&\hbox{}&{\rm Im}\;\tilde{a}^{(3)}_{2}=\!\frac{\beta_{2}}{%
\sqrt{\beta_{2}^{2}+5\beta_{3}^{2}}}\xi\end{array}$$
(5)
with the new parameters $\eta,\ \zeta$ and $\xi$ small of the first
order.
It is clear that only the ratios of the relevant amplitudes are
constrained by the eq.s (4). The definition of the new variables
given in each line of eq.s (5) contain therefore an arbitrary
factor. Our choice (and in particular for the square–root factors at
the denominators) has some distinguished advantage, which will result
clear from the classical expression of the kinetic energy, discussed
in the next paragraph.
In the intrinsic reference frame, and at the same
order of approximation, the
values of the three principal moments of inertia can be derived from
eq.s (54,55):
$$\displaystyle\mathcal{J}_{1}$$
$$\displaystyle=$$
$$\displaystyle 3(\beta_{2}^{2}+2\beta_{3}^{2})+2\sqrt{3}(\beta_{2}^{2}\gamma_{2%
}+\sqrt{5}\beta_{3}^{2}\gamma_{3})$$
(6a)
$$\displaystyle\mathcal{J}_{2}$$
$$\displaystyle=$$
$$\displaystyle 3(\beta_{2}^{2}+2\beta_{3}^{2})-2\sqrt{3}(\beta_{2}^{2}\gamma_{2%
}+\sqrt{5}\beta_{3}^{2}\gamma_{3})$$
(6b)
$$\displaystyle\mathcal{J}_{3}$$
$$\displaystyle=$$
$$\displaystyle 4(\beta_{2}^{2}\gamma_{2}^{2}+\beta_{3}^{2}\gamma_{3}^{2})+18(X^%
{2}+Y^{2})$$
(6c)
$$\displaystyle+2(\eta^{2}+\zeta^{2})+8\xi^{2}\ .$$
With the amplitudes given by eq.s (1,2) the principal
axes of the quadrupole would coincide with those of the octupole.
It is not necessarily so with our more general assumptions.
When $a^{(\lambda)}_{1}\neq 0$, the
axis 3 of the tensor of inertia for the quadrupole mode alone does not
coincide with that of the octupole. If $a^{(\lambda)}_{1}=0$, but
Im$\ a^{(\lambda)}_{2}\neq 0$, the misalignment concerns the other two
principal axes perpendicular to the common axis 3.
II.3 The classical expression of the kinetic energy
Now it is possible to express the classical kinetic energy (as given
by Bohr hydrodynamical model) in terms of the new variables and
of the intrinsic components $q_{\kappa}$
of the angular velocity. The classical expression has the form
$$T=\frac{1}{2}\sum\dot{Q}_{\mu}\mathcal{G}_{\mu\nu}\dot{Q}_{\nu}$$
(7)
where $\dot{Q}\equiv\{\dot{\xi}_{1},\dot{\xi}_{2},...,\dot{\xi}_{9},q_{1},q_{2},q_{3}\}$, $\dot{\xi}_{\mu}\ (\mu=1,...,9)$ are the
time derivative of the nine parameters we have just defined, and
$q_{1},\ q_{2},\ q_{3}$ are the intrinsic components of the angular velocity
of the intrinsic system with respect to an inertial frame.
The elements of the matrix $\mathcal{G}$
(leading terms and
relevant first-order terms) are shown in the Table 1.
The determinant $G={\rm Det}\ \mathcal{G}$ takes the form
$$G\propto\beta_{2}^{2}\beta_{3}^{2}\big{(}\beta_{2}^{2}+2\beta_{3}^{2}\big{)}^{%
2}\big{(}\beta_{2}^{2}+5\beta_{3}^{2}\big{)}^{-1}\big{(}\beta_{2}^{2}\gamma_{2%
}+\sqrt{5}\ \beta_{3}^{2}\gamma_{3}\big{)}^{2}$$
and, at the limit $\beta_{3}\ll\beta_{2}$, results to be proportional
to $\beta_{2}^{8}$, and therefore
consistent with that of the Bohr model for a pure quadrupole motion.
This is a consequence of our choice of the normalization factors in
the eq.s (5). This choice has other advantages: all the non diagonal
terms involving the time derivatives of $\beta_{2}$ or $\beta_{3}$ and
either the derivative of one
of the other intrinsic amplitudes or $q_{3}$ turn out to be zero in the
present approximation. Other non diagonal elements are small (of the
first order) in the “small” amplitudes $\gamma_{2},\ \gamma_{3},\ X,\ Y,\ \xi,\ \eta,\ \zeta$. In situations close to the axial symmetry, they
have negligible effect on the results (see Appendix B),
with the only exception of
elements of the
last line and column. The latter, in fact, are still small of the
first order, but must be compared with the diagonal element ${\mathcal{J}}_{3}$,
which is small of the second order in the “small” non–axial
amplitudes.
These terms play an important role in the treatment of the intrinsic
component of the angular momentum along the approximate
axial–symmetry axis, which will be discussed in the next paragraph.
II.4 Intrinsic components of the angular momentum
According to the classical mechanics, the components $L_{1}$, $L_{2}$ and
$L_{3}$ of the angular momentum in the intrinsic frame
are obtained as the derivatives of the total kinetic energy
with respect to the corresponding intrinsic component of the
angular velocity:
$$L_{k}=\frac{\partial T}{\partial q_{k}}\ .$$
(8)
The part of the kinetic energy depending on the component $q_{k}$
has the form
$$T_{k}(q_{k})=\frac{1}{2}\mathcal{J}_{k}q_{k}^{2}+F_{k}q_{k}$$
(9)
where $F_{k}$ is a function of the dynamical variables $\xi_{i}$
and of their
derivatives with respect to the time, and is
small of the first order
according to our definition. As for the moments of inertia,
$\mathcal{J}_{3}$ is
small of the second order, while $\mathcal{J}_{1}$,
$\mathcal{J}_{2}$ are not small.
According to eq.(8), we have
$$\displaystyle L_{k}$$
$$\displaystyle=$$
$$\displaystyle\mathcal{J}_{k}q_{k}+F_{k}$$
$$\displaystyle q_{k}$$
$$\displaystyle=$$
$$\displaystyle\frac{L_{k}-F_{k}}{\mathcal{J}_{k}}$$
(10)
$$\displaystyle T_{k}(L_{k})$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{2}\mathcal{J}_{k}\ \left[\frac{1}{\mathcal{J}_{k}}\big{(%
}L_{k}-F_{k}\big{)}\right]^{2}+F_{k}\left[\frac{1}{\mathcal{J}_{k}}\big{(}L_{k%
}-F_{k}\big{)}\right]$$
(11)
$$\displaystyle=$$
$$\displaystyle\frac{L_{k}^{2}}{2\mathcal{J}_{k}}-\frac{F_{k}^{2}}{2\mathcal{J}_%
{k}}\ .$$
For $k=1,2$ the second term is small of the second order and can
be neglected.
It is not so for $k=3$, as $\mathcal{J}_{3}$ is also small and of
the same order as $F_{k}^{2}$.
In more detail, we have
$$\displaystyle L_{3}$$
$$\displaystyle=$$
$$\displaystyle\mathcal{J}_{3}q_{3}+\Big{[}\ \frac{\sqrt{8}\ \beta_{2}\beta_{3}}%
{\sqrt{\beta_{2}^{2}+5\beta_{3}^{2}}}\big{(}\gamma\dot{\xi}-\xi\dot{\gamma}%
\big{)}$$
(12)
$$\displaystyle+$$
$$\displaystyle 6\big{(}Y\dot{X}-X\dot{Y}\big{)}+2\big{(}\zeta\dot{\eta}-\eta%
\dot{\zeta}\big{)}\ \Big{]}$$
where we have put
$$\gamma=\sqrt{5}\ \gamma_{2}-\gamma_{3}\ .$$
(13)
At this point, it will be convenient to express the variables
$\gamma_{2}$ and $\gamma_{3}$ as linear combinations of two new variables,
one of which is, obviously, $\gamma=\sqrt{5}\gamma_{2}-\gamma_{3}$.
The other one, that we call $\gamma_{0}$, can be chosen proportional to
the linear combination which enters in the expression of the
determinant $G$,
$$\gamma_{0}=c\left(\beta_{2}^{2}\gamma_{2}+\sqrt{5}\ \beta_{3}^{2}\gamma_{3}%
\right).$$
(14)
With this choice, we obtain
$$\displaystyle\gamma_{2}$$
$$\displaystyle=$$
$$\displaystyle\frac{\gamma_{0}/c+\sqrt{5}\ \beta_{3}^{2}\gamma}{\beta_{2}^{2}+5%
\beta_{3}^{2}}$$
$$\displaystyle\gamma_{3}$$
$$\displaystyle=$$
$$\displaystyle\frac{\sqrt{5}\ \gamma_{0}/c-\beta_{2}^{2}\gamma}{\beta_{2}^{2}+5%
\beta_{3}^{2}}$$
(15)
and, at the leading order,
$$\beta_{2}^{2}\dot{\gamma}_{2}^{2}+\beta_{3}^{2}\dot{\gamma}_{3}^{2}=\frac{%
\beta_{2}^{2}\beta_{3}^{2}}{\beta_{2}^{2}+5\beta_{3}^{2}}\dot{\gamma}^{2}+%
\frac{1}{\beta_{2}^{2}+5\beta_{3}^{2}}\ (\dot{\gamma}_{0}/c\big{)}^{2}\ .$$
(16)
In deriving the eq. 16, the factor $1/c$ has been considered
constant. We may note,however, that the same result holds at
the leading order also if $1/c$ is a function of $\beta_{2}$ and/or
$\beta_{3}$. In fact, terms involving the time derivative of $1/c$ also
contain the “small” quantity $\gamma_{0}$, and their effect is
negligible in the present approximation (see Appendix B).
E.g., one could choose for $1/c$ a quadratic expression in
$\beta_{2},\ \beta_{3}$, in order to obtain for $\gamma_{0}$ an
adimensional quantity, like $\gamma_{2},\ \gamma_{3}$ and $\gamma$.
The same argument applies for possible redefinitions of other
“small” variables, like $\gamma$.
The expression (12) can be somewhat simplified with the
substitutions
$$\displaystyle X$$
$$\displaystyle=$$
$$\displaystyle w\sin\vartheta$$
(17a)
$$\displaystyle Y$$
$$\displaystyle=$$
$$\displaystyle w\cos\vartheta$$
(17b)
$$\displaystyle\eta$$
$$\displaystyle=$$
$$\displaystyle v\sin\varphi$$
(17c)
$$\displaystyle\zeta$$
$$\displaystyle=$$
$$\displaystyle v\cos\varphi$$
(17d)
$$\displaystyle\xi$$
$$\displaystyle=$$
$$\displaystyle u\sin\chi$$
(17e)
$$\displaystyle\gamma$$
$$\displaystyle=$$
$$\displaystyle\sqrt{2}\ \left(\sqrt{\beta_{2}^{2}+5\beta_{3}^{2}}\ /\ \beta_{2}%
\beta_{3}\right)u\cos\chi$$
(17f)
$$\displaystyle\frac{\gamma_{0}}{c}$$
$$\displaystyle=$$
$$\displaystyle f(\beta_{2},\beta_{3})\sqrt{\beta_{2}^{2}+5\beta_{3}^{2}}\ u_{0}$$
(17g)
which gives for the determinant of the matrix $\mathcal{G}$
$$G={\rm Det}\;\mathcal{G}=2304\big{(}\beta_{2}^{2}+2\beta_{3}^{2}\big{)}^{2}\ u%
_{0}^{2}\ v^{2}u^{2}w^{2}\ f^{2}(\beta_{2},\beta_{3}).$$
(18)
The choice of the function $f(\beta_{2},\beta_{3})$ is irrelevant for
what concerns the angular momentum. Non diagonal terms (small of the
first order) would depend on this choice, but their effect is
negligible (see Appendix B). As a criterion to define the
form of the function $f$, we observe that for permanent quadrupole
deformation $\beta_{2}=\bar{\beta}_{2}$ and at the limit $\beta_{3}^{2}\ll\beta_{2}^{2}$ our value of $G$ must agree with the result given, at this
limit, by Eisenberg and Greiner Eisenberg and Greiner (1987). This happens if the
function $f$ we have left undetermined tends to a constant when
$\beta_{2}\rightarrow\bar{\beta}_{2}$.
We adopt here the simplest possible choice,
i.e. $f(\beta_{2},\beta_{3})=1$, to obtain the matrix
$\mathcal{G}$
given in the Table 2, and (at the leading order)
$$\displaystyle G$$
$$\displaystyle=$$
$$\displaystyle{\rm Det}\;\mathcal{G}=2304\big{(}\beta_{2}^{2}+2\beta_{3}^{2}%
\big{)}^{2}\ u_{0}^{2}\ v^{2}u^{2}w^{2}$$
(19)
$$\displaystyle\mathcal{J}_{3}$$
$$\displaystyle=$$
$$\displaystyle 4u_{0}^{2}+2v^{2}+8u^{2}+18w^{2}$$
(20)
$$\displaystyle 2T$$
$$\displaystyle=$$
$$\displaystyle\dot{\beta}_{2}^{2}+\dot{\beta}_{3}^{2}+\dot{u}_{0}^{2}$$
(21)
$$\displaystyle+$$
$$\displaystyle 2\big{(}\dot{v}^{2}+v^{2}\dot{\varphi}^{2}\big{)}+2\big{(}\dot{u%
}^{2}+u^{2}\dot{\chi}^{2}\big{)}+2\big{(}\dot{w}^{2}+w^{2}\dot{\vartheta}^{2}%
\big{)}$$
$$\displaystyle+$$
$$\displaystyle 2q_{3}\left[2v^{2}\dot{\varphi}+4u^{2}\dot{\chi}+6w^{2}\dot{%
\vartheta}\right]$$
$$\displaystyle+$$
$$\displaystyle\mathcal{J}_{1}q_{1}^{2}+\mathcal{J}_{2}q_{2}^{2}+\mathcal{J}_{3}%
q_{3}^{2}\ .$$
II.5 Classification of elementary
excitations with respect to $K^{\pi}$
We can now deduce the intrinsic components of the angular momentum:
$$\displaystyle L_{1}$$
$$\displaystyle=$$
$$\displaystyle\mathcal{J}_{1}q_{1}$$
(22a)
$$\displaystyle L_{2}$$
$$\displaystyle=$$
$$\displaystyle\mathcal{J}_{2}q_{2}$$
(22b)
$$\displaystyle L_{3}$$
$$\displaystyle=$$
$$\displaystyle\mathcal{J}_{3}q_{3}+\left[2v^{2}\dot{\varphi}+4u^{2}\dot{\chi}+6%
w^{2}\dot{\vartheta}\right]$$
(22c)
$$\displaystyle=$$
$$\displaystyle 2v^{2}\big{(}q_{3}+\dot{\varphi}\big{)}+4u^{2}\big{(}2q_{3}+\dot%
{\chi}\big{)}+6w^{2}\big{(}3q_{3}+\dot{\vartheta}\big{)}$$
$$\displaystyle+4u_{0}^{2}q_{3}\ .$$
In the same way, we can obtain the classical moments conjugate to
$\chi$, $\vartheta$ and $\varphi$
(we observe that none of these variables appears in the expressions of
$G$ or $\mathcal{J}_{3}$):
$$\displaystyle p_{\varphi}$$
$$\displaystyle=$$
$$\displaystyle 2v^{2}(\dot{\varphi}+q_{3})$$
(22d)
$$\displaystyle p_{\chi}$$
$$\displaystyle=$$
$$\displaystyle 2u^{2}\left(\dot{\chi}+2q_{3}\right)$$
(22e)
$$\displaystyle p_{\vartheta}$$
$$\displaystyle=$$
$$\displaystyle 2w^{2}(\dot{\vartheta}+3q_{3})\ .$$
(22f)
Now we can solve the system of equations
(22) with respect to
the variables $q_{1},\ q_{2},\ q_{3},\ \dot{\varphi},\ \dot{\chi},\ \dot{\vartheta}$. We obtain
$$\displaystyle q_{1}$$
$$\displaystyle=$$
$$\displaystyle L_{1}/\mathcal{J}_{1}$$
(23a)
$$\displaystyle q_{2}$$
$$\displaystyle=$$
$$\displaystyle L_{2}/\mathcal{J}_{2}$$
(23b)
$$\displaystyle q_{3}$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{u_{0}}\big{(}L_{3}-p_{\varphi}-2p_{\chi}-3p_{\vartheta}%
\big{)}$$
(23c)
$$\displaystyle\dot{\varphi}$$
$$\displaystyle=$$
$$\displaystyle\frac{p_{\varphi}}{2v_{0}^{2}}-\frac{1}{u_{0}^{2}}\big{(}L_{3}-p_%
{\varphi}-2p_{\chi}-3p_{\vartheta}\big{)}$$
(23d)
$$\displaystyle\dot{\chi}$$
$$\displaystyle=$$
$$\displaystyle\frac{p_{\chi}}{2u^{2}}-\frac{2}{u_{0}^{2}}\big{(}L_{3}-p_{%
\varphi}-2p_{\chi}-3p_{\vartheta}\big{)}$$
(23e)
$$\displaystyle\dot{\vartheta}$$
$$\displaystyle=$$
$$\displaystyle\frac{p_{\vartheta}}{2\chi_{0}^{2}}-\frac{3}{u_{0}^{2}}\big{(}L_{%
3}-p_{\varphi}-2p_{\chi}-3p_{\vartheta}\big{)}$$
(23f)
The equations (23) have a very simple meaning in the case
where the potential energy does not depend on the variables $\varphi$,
$\chi$ or $\vartheta$. In such a case (a sort of model
$\varphi$-$\chi$-$\vartheta$–instable, in the sense of the
$\gamma$–instable model by Wilets and Jean Wilets and Jean (1956))
the conjugate moments of these three angular variables are
constants of the
motion, with integer eigenvalues $n_{\varphi}$, $n_{\chi}$ and $n_{\vartheta}$
(in units of $\hbar$).
Moreover, if we assume that $u_{0}\rightarrow 0$,
the third component $q_{3}$ of the angular velocity tends to
$\infty$ unless $L_{3}-p_{\varphi}-2p_{\chi}-3p_{\vartheta}=0$
(eq. (23c)).
In this case, the operator $L_{3}$ is diagonal, with eigenvalues
$K=n_{\varphi}+2n_{\chi}+3n_{\vartheta}$,
and the three degrees of freedom corresponding to
$\varphi$, $\chi$ and $\vartheta$ can be associated to non-axial
excitation modes with $K=$1, 2 and 3, respectively.
To investigate the character of the degree of freedom described by the
parameter $u_{0}$, and for a deeper understanding of the nature of
the other degrees of freedom, it is necessary to express the complete
Hamiltonian in the frame of a definite model which, although not
unique, is at least completely self-consistent at the limit close to
the axial symmetry.
In fact, it is now possible to use the Pauli prescriptions
Pauli (1933) to construct the
quantum operator $\hat{T}$ corresponding to the classical kinetic
energy $T$ of eq. (21). In doing this, we make use of the
partial inversion of the matrix $\mathcal{G}$ given by
the solution (23) of the linear system (22),
and note that none of the variables $\varphi$, $\chi$, $\vartheta$
or $q_{3}$
enter in the expression of $G$:
$$\displaystyle\frac{2}{\hbar^{2}}\ \hat{T}$$
$$\displaystyle=$$
$$\displaystyle-\left\{\frac{1}{\beta_{2}^{2}+2\beta_{3}^{2}}\frac{\partial}{%
\partial\beta_{2}}\left[\big{(}\beta_{2}^{2}+2\beta_{3}^{2}\big{)}\frac{%
\partial}{\partial\beta_{2}}\right]\right.$$
(24)
$$\displaystyle+$$
$$\displaystyle\frac{1}{\beta_{2}^{2}+2\beta_{3}^{2}}\frac{\partial}{\partial%
\beta_{3}}\left[\big{(}\beta_{2}^{2}+2\beta_{3}^{2}\big{)}\frac{\partial}{%
\partial\beta_{3}}\right]$$
$$\displaystyle+$$
$$\displaystyle\frac{1}{u_{0}}\frac{\partial}{\partial u_{0}}\left[u_{0}\frac{%
\partial}{\partial u_{0}}\right]+\frac{1}{2v}\frac{\partial}{\partial v}\left[%
v\frac{\partial}{\partial v}\right]$$
$$\displaystyle+$$
$$\displaystyle\frac{1}{2u}\frac{\partial}{\partial u}\left[u\frac{\partial}{%
\partial u}\right]+\frac{1}{2w}\frac{\partial}{\partial w}\left[w\frac{%
\partial}{\partial w}\right]$$
$$\displaystyle+$$
$$\displaystyle\left.\ \frac{1}{2v^{2}}\frac{\partial^{2}}{\partial\varphi^{2}}+%
\frac{1}{2u^{2}}\frac{\partial 2}{\partial\chi 2}+\frac{1}{2w^{2}}\frac{%
\partial 2}{\partial\vartheta 2}\right\}$$
$$\displaystyle+$$
$$\displaystyle\frac{1}{4u_{0}^{2}}\ \left[\hat{L}_{3}-\frac{\partial}{\partial%
\varphi}-2\frac{\partial}{\partial\chi}-3\frac{\partial}{\partial\vartheta}%
\right]^{2}$$
$$\displaystyle+$$
$$\displaystyle\frac{1}{2\mathcal{J}_{1}}\ \hat{L}_{1}^{2}+\frac{1}{2\mathcal{J}%
_{2}}\ \hat{L}_{2}^{2}\ .$$
This is, admittedly, only a semi-classical discussion. However, the
formal quantum treatment in the frame of the Pauli procedure,
shown in the Appendix C, gives exactly the same result.
Until now, no assumption has been made on the form of the
potential-energy operator, which will determine the particular model.
A few general remarks on this subject are contained in the Appendix
D. We now assume that the potential energy can be separated
in the sum of a term depending only on $u_{0}$ and another
containing the other dynamical variables. In this case the
differential equation in the variable $u_{0}$ is approximately
decoupled from the rest. One obtains the Schrödinger equation
$$\displaystyle\left\{\frac{1}{u_{0}}\frac{\partial}{\partial u_{0}}\left[u_{0}%
\frac{\partial}{\partial u_{0}}\right]\right.$$
$$\displaystyle+$$
$$\displaystyle\frac{2}{\hbar^{2}}\left[E_{u_{0}}-U(u_{0})\right]$$
$$\displaystyle-\ \frac{1}{u_{0}^{2}}$$
$$\displaystyle\left.\left[\frac{\Omega_{u_{0}}}{2}\right]^{2}\right\}\phi(u_{0}%
)=0$$
(25)
where we have put
$$\Omega_{u_{0}}=K-n_{\varphi}-2n_{\chi}-3n_{\vartheta}\ .$$
(26)
If we assume, for simplicity, a harmonic form for the potential
$U(u_{0})=\frac{1}{2}Cu_{0}^{2}$, the eq. (25) is the
radial equation of a bidimensional harmonic oscillator. For the
existence of a solution, it is required that
$$\displaystyle\Omega_{u_{0}}$$
$$\displaystyle=$$
$$\displaystyle 2n_{u_{0}}$$
$$\displaystyle E_{u_{0}}$$
$$\displaystyle=$$
$$\displaystyle(N_{u_{0}}+1)\hbar\omega_{u_{0}}$$
(27)
with $n_{u_{0}}$ positive or negative integer and
the integer $N_{u_{0}}\geq|n_{u_{0}}|$.
Excitations in the degree of freedom corresponding to the variable
$u_{0}$ carry, therefore, two units of angular momentum in the
direction of the 3rd axis of the intrinsic reference frame.
We could extend our model to include all the intrinsic variables
different from $\beta_{2},\ \beta_{3}$. We assume a potential energy
corresponding to the sum of independent harmonic potentials in the
variables $v,\ u,\ w$ and $u_{0}$, plus a term depending
on $\beta_{2}$ and $\beta_{3}$ (at the moment, we do not need to define the
form of this term). We also assume that the equations in
$\beta_{2},\ \beta_{3}$ can be approximately decoupled from those of the
other variables and that, in the latter, $\beta_{2}$ and $\beta_{3}$
can be replaced by their average values. It is easy to verify that the
differential equations in the pair of variables $v,\ \varphi$
(or $u,\ \chi$ or $w,\ \vartheta$) correspond again to a
bidimensional harmonic oscillator and that, as long as we
neglect the rotation–vibration coupling, the eigenvalue
$K$ of the intrinsic component $L_{3}$ of the angular momentum is given
by
$$K=n_{\varphi}+2n_{\chi}+3n_{\vartheta}+2n_{u_{0}}\ .$$
(28)
The energy eigenvalues are, for the equation in the variables
$v,\ \varphi$,
$$E_{v}=(N_{v}+1)\ \hbar\omega_{v}\hskip 22.762205pt{\rm with}\ N_{v}\geq|n_{%
\varphi}|$$
(29)
and have a similar form for the other two oscillators.
It remains to consider the character of the different dynamical
variables with respect to the parity operator.
We know that the parity of the amplitude $a^{(\lambda)}_{\mu}$ is
$(-1)^{\lambda}$. Therefore, $\beta_{2}$ and $\gamma_{2}$ are even, while
$\beta_{3}$, $X$ and $Y$ are odd. As for $\gamma_{3}$, we observe that
$\beta_{3}\sin\gamma_{3}$ is odd, and therefore $\gamma_{3}$ must be
even.
As a consequence, are also even the linear combinations $\gamma$ and
$\gamma_{0}/c$ defined in the eq. (13,14).
The new variables $\eta,\ \zeta$ and $\xi$ are odd, as (e.g.)
$\beta_{2}\xi$ is an octupole amplitude and therefore is odd, while
$\beta_{3}\xi$ must be even. Finally, on the basis of eq.s (17)
we realize that $v,\ u$ and $w$ must be odd
(while $\varphi,\chi$ and $\vartheta$ are even). We have therefore
identified elementary excitations corresponding to
$K^{\pi}=1^{-},2^{-},3^{-}$ and $2^{+}$. Excitations with $K=0$ (of positive or
negative parity) conserve the axial symmetry, and are related to the
variables $\beta_{2}$ and $\beta_{3}$. A particular example will be
discussed in the following Sections.
III A specific model:
Axial octupole vibrations in nuclei with
permanent quadrupole deformation
Specific assumptions on the form of the
potential–energy terms for all the variables describing the
quadrupole and octupole degrees of freedom are necessary in order to
obtain definite predictions, also if these are limited to the axial
modes.
We discuss here, as an example, the case of axial octupole excitations
in nuclei which already possess a stable quadrupole deformation. In
this case, one obtains relatively simple results, suitable for
comparison with experimental data. This comparison will be performed
in the next Section.
Following the usual treatment of vibration + rotation, we
put $\beta_{2}=\bar{\beta}_{2}+\beta_{2}^{\prime}$, with $\bar{\beta}_{2}=$
constant and $|\beta_{2}^{\prime}|\ll|\bar{\beta}_{2}|$.
The new variable $\beta_{2}^{\prime}$ is therefore assumed to be small
(of the first order) as all other variables, with the exception of
$\beta_{3}$.
With this choice, and assuming that the variables introduced in the
eq. (17) are suitable to describe the other degrees of freedom,
the matrix $\mathcal{G}$ takes a form similar to that of Table II (with
$\beta_{2}$ substituted by $\bar{\beta}_{2}$ and $\dot{\beta}_{2}$ by
$\dot{\beta}_{2}^{\prime}$) and – at the lowest significant order –
results to be diagonal with respect to the variables $\dot{\beta}_{3}$,
$\dot{\beta}_{2}^{\prime}$, $q_{1}$ and $q_{2}$.
Moreover, in our model, the amplitude of oscillation for all
degrees of freedom different from $\beta_{3}$ are constrained to very
small values: this fact implies strong restoring forces and,
therefore, oscillation frequencies much larger than for $\beta_{3}$.
In the limit of small amplitude of the octupole oscillations, this
case has been discussed, e.g., by Eisenberg and
Greiner Eisenberg and Greiner (1987).
In their approach, the “intrinsic” reference frame is chosen to
coincide with the principal axis of the quadrupole deformation tensor,
but the differences between their approach and ours tend to disappear
for $|\beta_{3}|\ll\bar{\beta}_{2}$. The model we try to develop should
be able to describe (axial) octupole vibrations of finite amplitude,
but its limit for $|\beta_{3}|\ll\bar{\beta}_{2}$ must obviously agree
with the results of Eisenberg and Greiner.
The quantum-mechanical equation of motion for $\beta_{3}$ can be
obtained with the Pauli prescription Pauli (1933), with the additional
assumption that the equations involving $\beta_{3}$
and $\beta_{2}^{\prime}$ or the angular-momentum components $L_{1},\ L_{2}$ are
effectively decoupled from those containing the other dynamical
variables and/or the $L_{3}$ operator. The latter equations could
possibly be complicated, and substantially coupled with one another
and with the angular momentum component $L_{3}$ along the (approximate)
symmetry axis. A short discussion of this subject, with some
simplifying assumptions, has been given in the previous
section II.4.
At the moment, we assume that terms involving $\beta_{2}^{\prime},\ L_{3}$
and other dynamical variables different from $\beta_{3}$ contribute to
the total energy with their own eigenvalue,
independent from the
eigenfunction in the $\beta_{3}$ degree of freedom, and we only consider
their lowest–energy state. We also assume that this state has $K=0$
(and neglect, as usual, the possible rotation-vibration coupling). In
this case the complete wavefunction has the form
$$\displaystyle\Psi$$
$$\displaystyle\propto$$
$$\displaystyle\psi(\beta_{3})\Phi_{0}(\beta_{2}^{\prime},\gamma_{0},w,\rho,\xi_%
{0},\vartheta,\chi,\varphi)\mathcal{D}^{J}_{M,0}$$
(30)
$$\displaystyle\propto$$
$$\displaystyle\psi(\beta_{3})\Phi_{0}\ Y_{JM}(\hat{\Omega})$$
where $\mathcal{D}^{J}_{M,M^{\prime}}$ are the Wigner matrices and
$\psi(-\beta_{3})=(-)^{J}\psi(\beta_{3})$. The differential equation
for the wavefunction $\psi(\beta_{3})$, obtained with the Pauli
prescription, has the form
$$\displaystyle\left[-\frac{\hbar^{2}}{2}\right.$$
$$\displaystyle G^{-1/2}$$
$$\displaystyle\frac{\rm d}{{\rm d}\beta_{3}}G^{1/2}\frac{\rm d}{{\rm d}\beta_{3%
}}+V(\beta_{3})$$
(31)
$$\displaystyle+$$
$$\displaystyle\left.\!\frac{\hbar^{2}}{2}\frac{J(J+1)}{\mathcal{J}_{1}}\right]%
\psi(\beta_{3})=E\psi(\beta_{3})$$
where $V(\beta_{3})$ is the potential-energy term.
The expression of the determinant $G$ is not uniquely defined, as it
also depends on the part of the matrix of inertia $\mathcal{G}$
involving all
other dynamical variables. This is a general problem for all models
where part of the degrees of freedom are ignored (and, in collective
models, a number of dynamical variables describing the details of
nucleon degrees of freedom, are certainly ignored).
With the present choice of dynamical variables (Table 2),
one gets
$G=$Det$\;\mathcal{G}=2304\ u_{0}^{2}\ v^{2}u^{2}w^{2}(\bar{\beta}_{2}^{2}+2\beta_{3}^%
{2})^{2}$,
and the eq. (31) becomes
$$\displaystyle\left[-\frac{\hbar^{2}}{2}\!\right.$$
$$\displaystyle\!\left(\frac{{\rm d}^{2}}{{\rm d}\beta_{3}^{2}}+\frac{4\beta_{3}%
}{\bar{\beta}_{2}^{2}+2\beta_{3}^{2}}\frac{\rm d}{{\rm d}\beta_{3}}\right)+V(%
\beta_{3})$$
(32)
$$\displaystyle+$$
$$\displaystyle\left.\frac{\hbar^{2}}{2}\frac{J(J+1)}{\mathcal{J}_{1}}\right]%
\psi(\beta_{3})=E\psi(\beta_{3})\ .$$
(33)
This is the equation we have used in ref.s Bizzeti and Bizzeti-Sona (2004); Bizzeti (2003).
A different choice of the dynamical variables would have brought to a
different result. E.g., with the variables used in the
Table 1,
one obtains
$G\propto\bar{\beta}_{2}^{2}\beta_{3}^{2}(\bar{\beta}_{2}^{2}+2\beta_{3}^{2})^{%
2}(\bar{\beta}_{2}^{2}+5\beta_{3}^{2})^{-1}(\bar{\beta}_{2}^{2}\gamma_{2}+%
\sqrt{5}\beta_{3}^{2}\gamma_{3})^{2}$.
However, in this case the limit $|\beta_{3}|\ll|\bar{\beta}_{2}|$
would not correspond to the Eisenberg–Greiner result, due to the
presence of the factor $\beta_{3}^{2}$ in the expression of the
determinant $G$. In fact, when the spherical symmetry is broken by the
permanent quadrupole deformation, and at the limit of small octupole
deformation, the octupole amplitudes $a^{(3)}_{\mu}$ are decoupled from
one another Eisenberg and Greiner (1987) and it would have been more reasonable
to chose a
dynamical variable $u_{3}=\beta_{3}\gamma_{3}$ in the place of $\gamma_{3}$.
With this substitution, the factor $\beta_{3}^{2}$ disappears.
It is convenient to express the differential equation (33) in
terms of the adimensional quantities
$x=\sqrt{2}\beta_{3}/\bar{\beta}_{2}$ and $\epsilon=\frac{1}{\hbar^{2}}\bar{\beta}_{2}^{2}E$,
$v=\frac{1}{\hbar^{2}}\bar{\beta}_{2}^{2}V$. One obtains
$$\frac{{\rm d}^{2}\psi(x)}{{\rm d}x^{2}}+\frac{2x}{1+x^{2}}\frac{\rm d\psi(x)}{%
{\rm d}x}+\left[\epsilon-\frac{J(J+1)}{6(1+x^{2})}-v(x)\right]\psi(x)\!=\!0$$
(34)
where $\psi(-x)=(-)^{J}\psi(x)$.
This equation reduces to that of Eisenberg and Greiner Eisenberg and Greiner (1987)
when $x\ll 1$.
As for the potential $v(x)$, we have explored two possible
forms: a quadratic term111After completion of this work, we
have been informed that a quadratic potential plus a centrifugal
term with variable moment of inertia has also been considered in the
model by Minkov et al. Minkov et al. (2004).
$v(x)=\frac{1}{2}cx^{2}$ or a
square–well potential, as it has been adopted Iachello (2000b)
at the critical point in the X(5) model ( $v(x)=0$ for
$|x|<b$ and $=+\infty$ outside, so that $\psi(\pm b)=0$ ).
In both cases, there is a free parameter to be determined from the
comparison with empirical data.
We now discuss in particular the second case.
We may note that for $V=0$ the eq. (34) is formally
equivalent to that of spheroidal oblate wavefunctions (see
eq. 21.6.3 of ref. Abramowitz and Stegun (1970)) with the parameters $m,\ \lambda$ and
$C$ redefined as $m=0$, $C^{2}=\epsilon$ and
$\lambda=J(J+1)/6-\epsilon$. Here, however, the solution is confined
in the interval $-b<x<b$ and the equation has been solved
numerically.
For a given $b$ and for every value of $J$, one obtains a complete set
of orthogonal eigenfunctions, with an integrating factor $(1+x^{2})$.
These eigenfunctions can be characterized by the quantum
number $s=\nu+1$, where $\nu$ is the number of zeroes in the open
interval $0<x<b$. A few examples of wavefunctions corresponding to
the square–well potential with $b=1.73$ and for different values of
$s$ and $J$ are depicted in the Fig. 1.
The dependence of the eigenvalues on the parameter $b$ is illustrated
by the Fig. 2, where the ratios $E(J)/E(2)$ are shown for the
g.s. band ($s=1$).
Other possible choices of the set of independent
dynamical variables would have brought to a different equation, but
the difference would have concerned the coefficient of the
first–derivative term the eq. (34),
with very small effect on the results, as long
as $b$ is in a range of “reasonable” values.
To exemplify the effect of this term, results obtained with the
coefficient of the first derivative put to zero are also shown – as
dotted lines – in the Figure 2. Differences between the two
sets of
results turn out to be very limited for small values of $b$ (at least,
up to $b\approx 2$).
There is at least one case (${}^{226}$Th) in which our results
with the square–well
potential are in good agreement with the level scheme, for low-lying
states of positive and negative parity, while for ${}^{228}$Th
a better agreement is obtained with the quadratic potential.
The possible interpretation of this result as evidence for a phase
transition in the octupole degree of freedom is discussed in the
next Section.
IV Evidence of phase transition
in the octupole degree of freedom
IV.1 The Radium and Thorium isotopic chain
A phase transition in the nuclear shape manifests itself as a
relatively sharp change of a proper order parameter –
e.g., the ratio $R=E(4^{+})/E(2^{+})$ – as a function of a
driving
parameter which can be, in our case, the number of neutrons in the
isotopes of a given element or the number of protons along an isotone
chain. Due to the finite number of degrees of freedom, the transition
region has a finite width around the critical point, and extends over
several nuclides in the chain. In the case of transitions between
spherical shape and axial quadrupole deformation, the X(5) symmetry,
valid at the critical point, predicts
$E(4^{+})/E(2^{+})=2.91$ and we can use this criterion to locate the
critical point of the phase transition.
The situation is more complex when the quadrupole and the octupole
degrees of freedom must be considered at the same time. Fig. 3
shows, as a function of the neutron number of Ra and Th isotopes, a
few parameters which can be used as indicators of the quadrupole and
octupole collectivity. As for the quadrupole mode, the decrease of
$E(2^{+})$ with increasing $N$ (Fig. 3$a$) shows a corresponding
increase of collectivity.
Moreover, in the Fig. 3$b$
we observe the transition between the vibrational (or not
collective) behavior of the lighter isotopes of the chain
and a clear rotational
behavior ($R\approx 10/3$) above $A=226$, with a critical point
which can be located around $A=224$.
The ratio $E(1^{-})/E(2^{+})$, depicted in Fig. 3$d$, shows that the
relative importance of the octupole collectivity increases with
decreasing $N$ and reaches its maximum in the region below $N=138$,
where the
critical point of the phase transition in the quadrupole mode could be
located on the basis of Fig. 3$b$.
Heavier isotopes show evidence of octupole vibrations
(of different $K$) around a quadrupole-deformed core Guntner et al. (2000); Raduta et al. (1996b).
Lighter isotopes ($N<132$)
appear not to be deformed in their lower-$J$ states.
However, at larger angular momentum a rotational-like band develops,
and this band has the alternate-parity pattern typical of a
stable octupole deformation Smith et al. (1995).
The model introduced in the first part of this paper, and developed in
the section III for the particular case of a permanent quadrupole
deformation, assumes that non–axial amplitudes are constrained to
very small values by the large restoring forces. This implies that
excitation of one of the non–axial degrees of freedom leads to high
excitation energy, compared to that of
the first $1^{-}$ level of the $K=0$ band. Experimental data of
Fig. 3$c$
show that this is actually the case for the light Thorium isotopes, at
least up to $A=228$. In fact, the first $1^{-}$ level is not far from
the first
$2^{+}$ and much lower than levels belonging to negative–parity bands
with $K\neq 0$ (as the lowest $2^{-}$ or the second $1^{-}$).
IV.2 Comparison with experimental
data for ${}^{226,228}$Th
Our model assumes a permanent quadrupole deformation. Therefore, it
can be useful only for relatively heavy Th isotopes
(Fig. 3$a,b$). The quadrupole-deformed
region extends above the mass 224 (which could correspond to the
critical point of the phase
transition, having $E(4^{+})/E(2^{+})\approx 2.91$). Heavier Th isotopes
(with $A\geq 230$) show negative-parity bands built on the different
states of octupole vibration, from $K=0$ to $K=3$, with band heads much
higher than the first $2^{+}$ (Fig. 3$c$).
Only for lower $A$, the $1^{-}$ band head
of the $K^{\pi}=0^{-}$ band decreases well below the band heads of all
other octupole bands, and higher levels of the $K^{\pi}=0^{-}$ band
merge with those
of positive-parity of the ground-state band (Fig. 4),
approaching (but not
reaching) the pattern expected for a rigid, reflection asymmetric
rotor.
The region of possible validity of our model is therefore restricted
to ${}^{226}$Th and ${}^{228}$Th.
The ratios $E(J^{\pi})/E(2^{+})$ for the low-lying states of
${}^{226}$Th and ${}^{228}$Th are depicted in the Fig. 4, and
compared with the predictions of different models. The rigid-rotor
model cannot account for the position of the lowest negative-parity
levels, and overestimates the excitation energy for all the high-spin
states. Instead, a rather good agreement is obtained with the present
model, if one assumes, in the case of ${}^{226}$Th, a square-well
potential (as the one hypothesized by Iachello in his X(5) model)
and, in the case of ${}^{228}$Th, a harmonic restoring force.
In both cases, the free parameter of the model has been adjusted to
reproduce the position of the $1^{-}$ level. As shown in the Fig.
4 (dotted line) a much better agreement with the high-spin
levels of ${}^{226}$Th is obtained with a slightly different value of
the parameter ($b=1.87$ instead of 1.73),
at the expense of a very limited discrepancy
for the $1^{-}$ level.
Therefore, for what concerns the level energies of the ground-state
band (including in it also the odd-$J$, negative-parity states),
${}^{226}$Th seems to present the expected behavior of a nucleus with
permanent quadrupole deformation and close to the critical point of the
phase transition in the octupole mode.
IV.3
Other possible tests of the critical-point behavior
A considerable amount of experimental information has been reported,
in the last few years, on possible
candidates Casten and Zamfir (2001); Krucken et al. (2002); Bizzeti and Bizzeti-Sona (2002); Hutter et al. (2003); Clark et al. (2003); Fransen et al. (2004); Bijker et al. (2003, 2004); Tonev et al. (2004)
for the dynamical
symmetry X(5)(phase transition point in the quadrupole mode).
It is now clear that the agreement between experimental and calculated
energies for the ground-state band does not automatically imply that a
similar agreement exists also for other observables, like the excitation
energy of the second $0^{+}$ level (the band head of the $s=2$ band)
and the in-band and inter-band transition probabilities.
In several transitional nuclei, the excitation energies in the
ground-state band are in excellent agreement with the X(5)
predictions but the calculated ratios of the $B$(E2) transition
probabilities fail to reproduce the experimental ones Clark et al. (2003),
unless an ad-hoc second-order term is included in the E2
transition operator Arias (2001); Caprio (2004).
It is therefore important to test the predictions of our models
also for what concerns such observables.
The low–lying level scheme of ${}^{226}$Th is shown in the
Fig. 5, together with the one resulting from the present
model, with the value of $b$ adjusted in order to reproduce the
empirical value of the ratio $E(1^{-})/E(2^{+})$.
At the moment, only the first two levels of the $s=2$ band
($0^{+}_{2},\ 2^{+}_{2}$)
are known and their excitation energies are somewhat higher than the
values predicted at the critical point.
We can observe, however, that also in the best X(5) nuclei
Casten and Zamfir (2001); Krucken et al. (2002); Tonev et al. (2004) the
position of the levels of the $s=2$ band deviates somewhat from the
model predictions (although in the opposite direction).
In our opinion, a similar
qualitative agreement is obtained also in the present case. The
negative–parity levels of the $s=2$ band are predicted to lie at
higher energies, and could be difficult to observe.
Absolute values of the transition strengths are not available for
${}^{226}$Th, but some relevant information is provided by the branching
ratios in the level decays. In the tables 3,4,
experimental ratios of the reduced transition strengths for E1 or E2
transitions coming from the same level are compared with the model
prediction at the critical point.
The electric dipole moment would vanish for the collective motion of
a fluid with uniform charge density, as the center of charge would
coincide with the center of mass. Therefore,the observed E1 transition
amplitudes are entirely due to the non uniformity of the
nuclear charge distribution Tsvekov et al. (2002).
To calculate the value of
$B$(E1)$=(i||\mathcal{M}(E1)||f)^{2}/(2J_{i}+1)$,
the E1 transition operator has been assumed to have the
form Bohr and Mottelson (1957, 1958); Strutinski (1957); Lipas (1963)
$$\mathcal{M}_{\mu}(E1)=C_{1}\beta_{2}\beta_{3}Y_{1,\mu}\ ,$$
(35)
with the constant factor $C_{1}$ depending on the nuclear charge
polarizability.
The E2 transition operator for the in-band transition and at the limit
close to the axial symmetry has been taken in the simple form
$$\mathcal{M}_{\mu}(E2)=C_{2}(\beta_{0})Y_{2,\mu}\ ,$$
(36)
neglecting the (weak) dependence on $\beta_{3}^{2}$.
Therefore, the theoretical ratio of the reduced strengths for
transitions of different multipolarity (E1 and E2) is determined apart
from a constant factor, which must be fixed by comparison with the
experimental data. The average of the ratios
$B$(E1)/$B$(E2) for the transitions coming from the $10^{+}$ and $11^{-}$
levels has been used for normalization of the theoretical values given
in the Table 4.
V Conclusions
A theoretical scheme for the description of quadrupole plus octupole
excitations close to the axial symmetry limit has been developed in
the Section II and specialized in the Section III to the
simpler case of a permanent (and axially symmetric) quadrupole
deformation. In principle, the model should be able to describe the
wide field of reflection–asymmetric nuclear shapes, close to (but
not coincident with!) the axial symmetry limit. Calculations of
nuclear shapes in the frame of the HBF-Cranking model Nazarewicz and Olanders (1985) for
nuclei of the Radium – Thorium region find a large variety of
results, including proper potentials
for quadrupole–octupole vibrations around a spherical
shape or for octupole vibrations around a deformed, reflection symmetric
shape, and also situations with a rather flat minimum of the potential
along a line at constant $\beta_{2}$ with $|\beta_{3}|<\beta_{3}^{max}$.
The latter case
is just what is expected for our “critical point” of the
phase transition in the octupole mode.
In the Section IV, we have investigated the
evolution of the nuclear shape along the isotopic chain of Thorium,
and shown that evidence of phase transition exists, not only in the
quadrupole mode but also in the octupole mode around a stable
quadrupole deformation. The model developed in the Section III
results to be able to account for the experimental data of ${}^{226}$Th
(at the critical point of the phase transition in the
octupole mode), and also of its
neighbor ${}^{228}$Th (characterized by axial octupole
vibrations). More and improved experimental data on E2 and E1
transition strengths would be necessary for a more stringent test of
the model predictions.
Further developments of the calculations are in program, to provide
detailed predictions for other significant cases: first of all, those
of ${}^{224}$Ra and ${}^{224}$Th, whose positive–parity levels show an
energy sequence very close to Iachello X(5) predictions
for the critical point of phase transition between
spherical shape and axially symmetric quadrupole deformation.
Appendix A Summary of the general formalism
The general formalism to describe collective states of
rotation/vibration in nuclei is discussed e.g. in
ref. Bohr (1952).
The nuclear surface is described in polar coordinates as
$$r(\theta,\phi)=R_{0}\left[1+\sum_{\lambda}\sum_{\mu=-\lambda,\lambda}\alpha^{(%
\lambda)}_{\mu}Y^{*}_{\lambda,\mu}(\theta,\phi)\right],$$
(37)
with the condition $\alpha^{(\lambda)}_{-\mu}=(-1)^{\mu}{\alpha^{(\lambda)}_{\mu}}^{*}$.
In the sum, the values of $\lambda$ are now limited to 2,3.
A term with $\lambda=1$ should be included in order to
maintain fixed
the position of the center of mass Eisenberg and Greiner (1987):
$$\displaystyle\alpha^{(1)}_{\mu}$$
$$\displaystyle=$$
$$\displaystyle\frac{3}{\sqrt{16\pi}}\ \sum_{\lambda,\lambda^{\prime}}\sqrt{(2%
\lambda+1)(2\lambda^{\prime}+1)}$$
(38)
$$\displaystyle\cdot\left(\begin{array}[]{ccc}\lambda&\lambda^{\prime}&1\\
0&0&0\end{array}\right)\ \left[\alpha^{(\lambda)}\otimes\alpha^{(\lambda^{%
\prime})}\right]^{(1)}_{\mu}\ .$$
This term, however, is inessential
for the following discussion, and will be omitted.
For not-too-large deformation, in fact, the amplitudes
$\alpha^{(1)}_{\mu}$ are much smaller than the others, and their effect
results to be negligible for most purposes (with the noticeable
exception of the E1 transition amplitudes). At the same level of
accuracy,
one can also neglect the slight variation of $R_{0}$ necessary to keep
the volume exactly constant.
In the Bohr model that we are considering here, the classical
expression of the collective kinetic energy is
$$T=\frac{1}{2}B_{2}\sum_{\mu=-2,2}\left|\dot{\alpha}^{(2)}_{\mu}\right|^{2}+%
\frac{1}{2}B_{3}\sum_{\mu=-3,3}\left|\dot{\alpha}^{(3)}_{\mu}\right|^{2}\ .$$
(39)
In order to simplify the notation in the following, we include the
inertia coefficient $B_{\lambda}$ in the definition of the collective
variables $a^{(\lambda)}_{\mu}$.
In the literature, this symbol is usually reserved
to the variables defined in the intrinsic reference frame and, from
now on, we will always use this reference frame for the collective
variables $a^{(\lambda)}_{\mu}$:
$$\sqrt{B_{\lambda}}\ \alpha^{(\lambda)}_{\mu}=\sum_{\nu}a^{(\lambda)}_{\nu}{D^{%
(\lambda)}_{\mu\nu}}^{*}(\theta_{i})$$
(40)
where $D^{(\lambda)}$ are the Wigner matrices and $\theta_{i}$ the Euler angles.
The axis of the intrinsic reference frame are defined along the
principal axis of the inertia tensor.
The expression of the kinetic energy
in terms of the time derivatives
of the intrinsic deformation variables $a^{(\lambda)}_{\mu}$
and of the angular velocity $\vec{q}$ of the intrinsic frame with
respect to an inertial frame is discussed, e.g., in the ref.s
Bohr (1952); Eisenberg and Greiner (1987).
If only the quadrupole mode is considered, the total kinetic energy
can be expressed as the sum of a vibrational term (in the intrinsic
frame) and a rotational term. If the octupole mode is also considered,
a rotation-vibration coupling term must be added Eisenberg and Greiner (1987); Wexler and Dussel (1999):
$$T=T_{\rm vib}+T_{\rm rot}+T_{\rm coup}$$
(41)
where
$$\displaystyle T_{\rm vib}$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{2}\sum_{\lambda,\mu}\left|\dot{a}^{(\lambda)}_{\mu}%
\right|^{2}$$
(42)
$$\displaystyle T_{\rm rot}$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{2}\sum_{\lambda,\nu,\nu^{\prime}}a^{(\lambda)}_{\nu}{a^{%
(\lambda)}_{\nu^{\prime}}}^{*}\sum_{k,k^{\prime}}q_{k}q_{k^{\prime}}\left(M^{%
\{\lambda\}}_{k}M^{\{\lambda\}}_{k^{\prime}}\right)_{\nu\nu^{\prime}}$$
(43)
$$\displaystyle\equiv$$
$$\displaystyle\frac{1}{2}\sum_{k,k^{\prime}}q_{k}q_{k^{\prime}}\mathcal{J}_{kk^%
{\prime}}\ ,$$
$$\displaystyle T_{\rm coup}$$
$$\displaystyle=$$
$$\displaystyle\frac{i}{2}\sum_{\lambda,\nu,\nu^{\prime},k}q_{k}\left[\dot{a}^{(%
\lambda)}_{\nu}{a^{(\lambda)}_{\nu^{\prime}}}^{*}-a^{(\lambda)}_{\nu}{{\dot{a}%
}^{(\lambda)*}_{\nu^{\prime}}}\right]\left(M^{\{\lambda\}}_{k}\right)_{\nu,\nu%
^{\prime}}$$
$$\displaystyle=$$
$$\displaystyle i$$
$$\displaystyle\!\sum_{\lambda}\!\sqrt{3(2\lambda\!+\!1)}\left[q^{(1)}\!\otimes%
\!\left[a^{(\lambda)}\!\otimes\dot{a}^{(\lambda)}\right]^{(1)}\right]^{(0)}_{0%
}\!.$$
(44)
Here, $q_{k}$ ($k=1,2,3$) are the Cartesian components of the angular
velocity $\vec{q}$ along the axes of the intrinsic frame, while
the $M^{\{\lambda\}}_{k}$ are $(2\lambda+1)$–dimensional matrices
giving the quantum–mechanical representation of the Cartesian
components of an angular momentum $M=\lambda$ in the intrinsic frame,
subject to commutation rules of the form
$$M^{\{\lambda\}}_{1}M^{\{\lambda\}}_{2}-M^{\{\lambda\}}_{2}M^{\{\lambda\}}_{1}=%
-iM^{\{\lambda\}}_{3}\ .$$
(45)
We assume, as usually,
$$\displaystyle\left(\right.$$
$$\displaystyle M^{\{\lambda\}}_{3}\left.\right)_{\nu\nu^{\prime}}=\nu\delta_{%
\nu,\nu^{\prime}}$$
$$\displaystyle\left(\right.$$
$$\displaystyle M^{\{\lambda\}}_{1}\left.\right)_{\nu\nu^{\prime}}=\frac{1}{2}\ %
\left(\sqrt{(\lambda-\nu)\ (\lambda+\nu+1)}\ \delta_{\nu^{\prime},\nu+1}\right.$$
$$\displaystyle+\left.\sqrt{(\lambda+\nu)\ (\lambda-\nu+1)}\ \delta_{\nu^{\prime%
},\nu-1}\right)$$
$$\displaystyle\left(\right.$$
$$\displaystyle M^{\{\lambda\}}_{2}\left.\right)_{\nu\nu^{\prime}}=\frac{1}{2}%
\left(-i\ \sqrt{(\lambda-\nu)\ (\lambda+\nu+1)}\ \delta_{\nu^{\prime},\nu+1}\right.$$
$$\displaystyle+\left.i\ \sqrt{(\lambda+\nu)\ (\lambda-\nu+1)}\ \delta_{\nu^{%
\prime},\nu-1}\right).$$
Taking into account the properties of the $M^{\{\lambda\}}_{k}$,
it is possible to obtain the explicit expression for the diagonal and
non diagonal elements of the tensor of inertia Bohr (1952); Wexler and Dussel (1999)
$$\mathcal{J}_{kk^{\prime}}=\frac{1}{2}\sum_{\lambda,\nu,\nu^{\prime}}a^{(%
\lambda)}_{\nu}{a^{(\lambda)}_{\nu^{\prime}}}^{*}\left(M^{\{\lambda\}}_{k}M^{%
\{\lambda\}}_{k^{\prime}}+M^{\{\lambda\}}_{k^{\prime}}M^{\{\lambda\}}_{k}%
\right)_{\nu\nu^{\prime}}\ .$$
(46)
The spin operators $M^{\{\lambda\}}_{k}$ transform
as the Cartesian components of a vector $\vec{M}^{\{\lambda\}}$ under
rotation in the ordinary space.
Taking into account
the commutation rule (45), we can define the irreducible
tensor components of $\vec{M}^{\{\lambda\}}$ as
$$\displaystyle M^{(1)\{\lambda\}}_{0}$$
$$\displaystyle=$$
$$\displaystyle M^{\{\lambda\}}_{3}$$
$$\displaystyle M^{(1)\{\lambda\}}_{\pm 1}$$
$$\displaystyle=$$
$$\displaystyle\pm\frac{1}{\sqrt{2}}\left(M^{\{\lambda\}}_{1}\mp i\ M^{\{\lambda%
\}}_{2}\right)$$
(47)
and express the products of two Cartesian components
as the sum of products of two tensor components.
To this purpose, we
define, for each value of $\lambda$,
the irreducible tensor product
$$\displaystyle T^{(J)}_{m}$$
$$\displaystyle=$$
$$\displaystyle\left[M^{(1)}\otimes M^{(1)}\right]^{(J)}_{m}$$
(48)
$$\displaystyle\equiv$$
$$\displaystyle\sum_{\nu,\nu^{\prime}}(1\nu 1\nu^{\prime}|Jm)M^{(1)}_{\nu}\ M^{(%
1)}_{\nu^{\prime}}$$
(where the common suffix $\{\lambda\}$ has been dropped,
for sake of simplicity).
We now introduce the reduced matrix elements
of the tensor operator $T^{(J)}$,
$$(\lambda||T^{(J)}||\lambda)=(-1)^{J}\sqrt{2J+1}\left\{\!\begin{array}[]{ccc}1&%
1&J\\
\lambda&\lambda&\lambda\end{array}\!\right\}\lambda(\lambda+1)(2\lambda+1)$$
(49)
to obtain
$$\displaystyle\left(M^{(1)}_{\mu}\ M^{(1)}_{\mu^{\prime}}\right)_{\nu,\nu^{%
\prime}}$$
$$\displaystyle=$$
$$\displaystyle\sum_{Jm}(1\mu 1\mu^{\prime}|Jm)\ \left(T^{(J)}_{m}\right)_{\nu,%
\nu^{\prime}}$$
$$\displaystyle=$$
$$\displaystyle\sum_{Jm}(1\mu 1\mu^{\prime}|Jm)(\lambda||T^{(J)}||\lambda)$$
$$\displaystyle\times\frac{(-1)^{\lambda-\nu+m}}{\sqrt{2J+1}}\ (\lambda\nu%
\lambda-\nu^{\prime}|Jm)\ .$$
We now observe that
$${a^{(\lambda)}_{\nu^{\prime}}}^{*}=(-1)^{\nu^{\prime}}a^{(\lambda)}_{-\nu^{%
\prime}}$$
(51)
and that
$$\left[a^{(\lambda)}\otimes a^{(\lambda)}\right]^{(J)}_{m}=\sum_{\nu,\nu^{%
\prime}}a^{(\lambda)}_{\nu}a^{(\lambda)}_{-\nu^{\prime}}\ (\lambda\nu\lambda-%
\nu^{\prime}|Jm)$$
(52)
to obtain
$$\displaystyle\sum_{\nu,\nu^{\prime}}$$
$$\displaystyle a^{(\lambda)}_{\nu}{a^{(\lambda)}_{\nu^{\prime}}}^{*}\left(M^{(1%
)\{\lambda\}}_{\mu}M^{(1)\{\lambda\}}_{\mu^{\prime}}\right)_{\nu\nu^{\prime}}$$
(53)
$$\displaystyle=\sum_{J,m}\frac{(-1)^{\lambda}}{\sqrt{2J+1}}(1\mu 1\mu^{\prime}|%
Jm)\ (\lambda||T^{(J)}||\lambda)$$
$$\displaystyle\ \times\ \left[a^{(\lambda)}\otimes a^{(\lambda)}\right]^{(J)}_{%
m}\ .$$
The rank of the tensor product $T^{(J)}$ of the two identical
vectors $M^{(1)\{\lambda\}}$ must be even, and therefore the possible
values of $J$ are limited to 0 and 2.
Now we can substitute, in the eq. (46),
the cartesian components of the
angular momentum with its tensor components
defined in the eq.s (47).
The possible values of $m$ (and $J$) contributing to
the sum are limited to $m=0$ (and therefore $J=$0 or 2) in the case of
$(M_{3})^{2}$, to $m=0$ or 2 ($J=$0 or 2) for
$(M_{1})^{2}$, $(M_{2})^{2}$ and
$(M_{1}M_{2})$, to $m=1$ ($J=2$)
for $(M_{3}M_{1})$ and $(M_{3}M_{2})$.
One obtains
$$\displaystyle\mathcal{J}_{1}$$
$$\displaystyle=$$
$$\displaystyle\sum_{\lambda}\left\{C_{0}(\lambda)\left[a^{(\lambda)}\otimes a^{%
(\lambda)}\right]^{(0)}_{0}\right.$$
(54a)
$$\displaystyle+$$
$$\displaystyle\frac{1}{\sqrt{6}}C_{2}(\lambda)\left[a^{(\lambda)}\otimes a^{(%
\lambda)}\right]^{(2)}_{0}$$
$$\displaystyle-$$
$$\displaystyle\left.C_{2}(\lambda){\rm Re}\left[a^{(\lambda)}\otimes a^{(%
\lambda)}\right]^{(2)}_{2}\right\}\ ,$$
$$\displaystyle\mathcal{J}_{2}$$
$$\displaystyle=$$
$$\displaystyle\sum_{\lambda}\left\{C_{0}(\lambda)\left[a^{(\lambda)}\otimes a^{%
(\lambda)}\right]^{(0)}_{0}\right.$$
(54b)
$$\displaystyle+$$
$$\displaystyle\frac{1}{\sqrt{6}}C_{2}(\lambda)\left[a^{(\lambda)}\otimes a^{(%
\lambda)}\right]^{(2)}_{0}$$
$$\displaystyle+$$
$$\displaystyle\left.C_{2}(\lambda){\rm Re}\left[a^{(\lambda)}\otimes a^{(%
\lambda)}\right]^{(2)}_{2}\right\}\ ,$$
$$\displaystyle\mathcal{J}_{3}$$
$$\displaystyle=$$
$$\displaystyle\sum_{\lambda}\left\{C_{0}(\lambda)\left[a^{(\lambda)}\otimes a^{%
(\lambda)}\right]^{(0)}_{0}\right.$$
(54c)
$$\displaystyle-$$
$$\displaystyle\sqrt{\frac{2}{3}}\left.C_{2}(\lambda)\left[a^{(\lambda)}\otimes a%
^{(\lambda)}\right]^{(2)}_{0}\right\}\ ,$$
$$\displaystyle\mathcal{J}_{12}$$
$$\displaystyle=$$
$$\displaystyle\sum_{\lambda}C_{2}(\lambda)\ {\rm Im}\left[a^{(\lambda)}\otimes a%
^{(\lambda)}\right]^{(2)}_{2}\ ,$$
(54d)
$$\displaystyle\mathcal{J}_{13}$$
$$\displaystyle=$$
$$\displaystyle\sum_{\lambda}C_{2}(\lambda)\ {\rm Re}\left[a^{(\lambda)}\otimes a%
^{(\lambda)}\right]^{(2)}_{1}\ ,$$
(54e)
$$\displaystyle\mathcal{J}_{23}$$
$$\displaystyle=$$
$$\displaystyle\sum_{\lambda}C_{2}(\lambda)\ {\rm Im}\left[a^{(\lambda)}\otimes a%
^{(\lambda)}\right]^{(2)}_{1}.$$
(54f)
where
$$\displaystyle C_{0}(\lambda)$$
$$\displaystyle=$$
$$\displaystyle(\!-\!1)^{\lambda}\ \frac{\lambda(\lambda+1)}{3}\sqrt{2\lambda+1}$$
$$\displaystyle C_{2}(\lambda)$$
$$\displaystyle=$$
$$\displaystyle(\!-\!1)^{\lambda+1}\ \sqrt{\frac{\lambda(\lambda\!+\!1)(2\lambda%
\!+\!\!3)(4\lambda^{2}-1)}{30}}$$
and therefore
$$\begin{array}[c]{lp{8mm}l}C_{0}(2)=2\ \sqrt{5}\ ,&\hbox{}&C_{0}(3)=-4\ \sqrt{7%
}\ ,\\
C_{2}(2)=-\sqrt{21}\ ,&\hbox{}&C_{2}(3)=3\ \sqrt{14}\ .\end{array}$$
(55)
If we chose as the intrinsic reference frame the principal axis of the
tensor of inertia, we must put $\mathcal{J}_{12}=\mathcal{J}_{13}=\mathcal{J}_{23}=0$. Using Eq.s (1,2,3)
we obtain, up to the first order
$$\displaystyle\mathcal{J}_{12}$$
$$\displaystyle=$$
$$\displaystyle-2\ \sqrt{21}\ {\rm Im}\ \left[\bar{a}^{(2)}\otimes\tilde{a}^{(2)%
}\right]^{(2)}_{2}$$
(56a)
$$\displaystyle+$$
$$\displaystyle 6\ \sqrt{14}\ {\rm Im}\ \left[\bar{a}^{(3)}\otimes\tilde{a}^{(3)%
}\right]^{(2)}_{2}=0,$$
$$\displaystyle\mathcal{J}_{13}$$
$$\displaystyle+$$
$$\displaystyle i\mathcal{J}_{23}=-2\ \sqrt{21}\ \left[\bar{a}^{(2)}\otimes%
\tilde{a}^{(2)}\right]^{(2)}_{1}$$
(56b)
$$\displaystyle+$$
$$\displaystyle 6\ \sqrt{14}\ \left[\bar{a}^{(3)}\otimes\tilde{a}^{(3)}\right]^{%
(2)}_{1}=0\ .$$
Here we make use of the fact that the zero–order terms are
automatically set to zero if the $\bar{a}^{(\lambda)}_{\mu}$ are defined
according to eq.s (1,2). By inserting these definitions
in the above equations, and retaining only the first–order terms,
one obtains
$$\displaystyle\left[\bar{a}^{(2)}\otimes\tilde{a}^{(2)}\right]^{(2)}_{1}$$
$$\displaystyle\approx$$
$$\displaystyle(2,0,2,1|2,1)\ \beta_{2}\ \tilde{a}^{(2)}_{1}$$
$$\displaystyle=$$
$$\displaystyle-\frac{1}{\sqrt{14}}\ \beta_{2}\ \tilde{a}^{(2)}_{1}\ ,$$
$$\displaystyle{\rm Im}\left[\bar{a}^{(2)}\otimes\tilde{a}^{(2)}\right]^{(2)}_{2}$$
$$\displaystyle\approx$$
$$\displaystyle(2,0,2,2|2,2)\ \beta_{2}\ {\rm Im}\ \tilde{a}^{(2)}_{2}$$
$$\displaystyle=$$
$$\displaystyle\sqrt{\frac{2}{7}}\ \beta_{2}\ {\rm Im}\ \tilde{a}^{(2)}_{2}\ ,$$
$$\displaystyle\left[\bar{a}^{(3)}\otimes\tilde{a}^{(3)}\right]^{(2)}_{1}$$
$$\displaystyle\approx$$
$$\displaystyle(3,0,3,1|2,1)\ \beta_{3}\ \tilde{a}^{(3)}_{1}$$
$$\displaystyle=$$
$$\displaystyle\sqrt{\frac{1}{42}}\ \beta_{3}\ \tilde{a}^{(3)}_{1},$$
$$\displaystyle{\rm Im}\left[\bar{a}^{(3)}\otimes\tilde{a}^{(3)}\right]^{(2)}_{2}$$
$$\displaystyle\approx$$
$$\displaystyle(3,0,3,2|2,2)\ \beta_{3}\ {\rm Im}\ \tilde{a}^{(3)}_{2}$$
$$\displaystyle=$$
$$\displaystyle-\sqrt{\frac{5}{21}}\ \beta_{3}\ {\rm Im}\ \tilde{a}^{(3)}_{2}.$$
We obtain, therefore, from eq.s (56)
$$\displaystyle\beta_{2}\tilde{a}^{(2)}_{1}$$
$$\displaystyle=$$
$$\displaystyle-\sqrt{2}\ \beta_{3}\tilde{a}^{(3)}_{1},$$
$$\displaystyle\beta_{2}{\rm Im}\ \tilde{a}^{(2)}_{2}$$
$$\displaystyle=$$
$$\displaystyle-\sqrt{5}\ \beta_{3}{\rm Im}\ \tilde{a}^{(3)}_{2}.$$
(57)
Appendix B Effect of non diagonal (first-order) terms
The matrix $\mathcal{G}$, as it results from the Table 1,
contains zero-order terms only in its principal diagonal (whose
last element, however, is small of the second order). Non diagonal
terms
have been expanded in series up to the first order in the “small”
dynamical variables (all of them, apart from $\beta_{2}$ and $\beta_{3}$).
We shall call $\xi_{k}$ a generic “small” variable, different from
$\beta_{\lambda}$ $(\lambda=2,3)$.
Terms of the last line and column (those related to the third intrinsic
component of the angular velocity) are discussed in the
Section II.4.
Here, we consider a simpler problem: the inversion of a
matrix $\mathcal{G}$ which has finite values for all terms
in the principal
diagonal, and only “small” values for all others.
In the zero-order approximation, the inverse matrix
$\mathcal{A=G}^{-1}$
is diagonal, with diagonal elements $A_{\mu\mu}=1/G_{\mu\mu}$.
The first–order approximation gives the non–diagonal elements
$$A_{\mu\nu}=A_{\nu\mu}=-\frac{G_{\mu\nu}}{G_{\mu\mu}G_{\nu\nu}}.$$
(58)
We are interested in particular on the effect of non diagonal terms on
the coefficients of the derivatives with respect to $\beta_{2}$ or
$\beta_{3}$.
Terms involving the derivatives with respect to
$\beta_{\lambda}$ and $\xi_{k}$ have the form
$$G^{-\frac{1}{2}}\frac{\partial}{\partial\xi_{k}}\left[G^{\frac{1}{2}}A_{k%
\lambda}\frac{\partial}{\partial\beta_{\lambda}}\right]+G^{-\frac{1}{2}}\frac{%
\partial}{\partial\beta_{\lambda}}\left[G^{\frac{1}{2}}A_{\lambda k}\frac{%
\partial}{\partial\xi_{k}}\right].$$
(59)
The non diagonal matrix elements of the matrix $\mathcal{A}$ are of the
first order in the “small” variables. They have therefore the form
$f^{(i)}_{\lambda k}(\beta_{2},\beta_{3})\xi_{i}$, where
$f^{(i)}_{\lambda k}(\beta_{2},\beta_{3})=\partial\mathcal{A}_{\lambda,k}/\ %
\partial\xi_{i}$,
or – possibly – are the sum of
several terms like that.
Substituting this expression in the
Eq. (59) one obtains
$$\displaystyle G$$
$${}^{-\frac{1}{2}}\frac{\partial}{\partial\xi_{k}}\left[G^{\frac{1}{2}}A_{k%
\lambda}\frac{\partial}{\partial\beta_{\lambda}}\right]+G^{-\frac{1}{2}}\frac{%
\partial}{\partial\beta_{\lambda}}\left[G^{\frac{1}{2}}A_{\lambda k}\frac{%
\partial}{\partial\xi_{k}}\right]$$
(60)
$$\displaystyle=$$
$$\displaystyle f^{(i)}_{\lambda k}(\beta_{2},\beta_{3})\left[2\xi_{i}\frac{%
\partial}{\partial\xi_{k}}+\frac{\xi_{i}}{2G}\ \left(\frac{\partial G}{%
\partial\xi_{k}}\right)+\delta_{ik}\right]\frac{\partial}{\partial\beta_{%
\lambda}}$$
$$\displaystyle+$$
$$\displaystyle\xi_{i}\left[f^{(i)}_{\lambda k}(\beta_{2},\beta_{3})\frac{1}{2G}%
\ \left(\frac{\partial G}{\partial\beta_{\lambda}}\right)+\left(\frac{\partial
f%
^{(i)}_{\lambda k}}{\partial\beta_{\lambda}}\right)\right]\frac{\partial}{%
\partial\xi_{k}}\ .$$
The last line of Eq. (60) only contains the partial-derivative
operator with respect to the small variable $\xi_{k}$.
Compared with the diagonal term involving the corresponding
second-derivative operator, it contains a
small factor $\xi_{i}$ more, and can be neglected.
In the first line
of the expression, all terms are potentially of the same order of
magnitude of
the leading ones, and in principle could not be neglected.
The first of the terms in the square brackets
contains the partial derivative with respect
to $\xi_{k}$. In the spirit of the adiabatic approach, we try to
estimate the expectation value of this term for the ground–state
wavefunction in all the variables, with the exception of $\beta_{2}$ and
$\beta_{3}$.
If we assume that the potential is harmonic for the ensemble of these
variables, the ground–state wavefunction can be expected to be close
to a multivariate Gaussian function. If, moreover, $\xi_{k}$
and $\xi_{i}$ are
uncorrelated, the expectation value of $\xi_{i}\partial/\partial\xi_{k}$
is zero, as $\left<\xi_{i}\right>=0$.
For the case $i=k$, instead, the expectation value is
$$\displaystyle\left<2\xi\frac{\partial}{\partial\xi}\right>$$
$$\displaystyle\approx$$
$$\displaystyle 2\ \frac{\int\exp(-\alpha\xi^{2}/2)\xi\frac{\partial}{\partial%
\xi}\exp(-\alpha\xi^{2}/2){\rm d}\xi}{\int\exp(-\alpha\xi^{2}){\rm d}\xi}$$
(61)
$$\displaystyle=$$
$$\displaystyle\frac{\int\exp(-\alpha\xi^{2}/2)\left(-2\alpha\xi^{2}\right)\exp(%
-\alpha\xi^{2}/2){\rm d}\xi}{\int\exp(-\alpha\xi^{2}){\rm d}\xi}$$
$$\displaystyle=$$
$$\displaystyle-1$$
(62)
and (in this approximation) cancels the third term, $\delta_{ik}$.
In conclusion, the coefficient of $\partial/\partial\beta_{\lambda}$
coming from the non–diagonal terms of the matrix ${\mathcal{G}}$
can be approximated with a sum of expressions like
$$\ f^{(i)}_{\lambda k}(\beta_{2},\beta_{3})\ \left<\frac{\xi_{i}}{G}\ \frac{%
\partial G}{\partial\xi_{k}}\right>$$
(63)
which approximately vanishes for $i\neq k$, and also
vanishes for $i=k$ unless both
$\mathcal{A}_{\lambda k}$ and $G$ depend explicitly on $\xi_{k}$.
Moreover, also in this latter case,
it is possible to eliminate non-diagonal elements of the matrix
$\mathcal{G}$ of the form $\mathcal{G}_{\lambda k}=\ g_{\lambda k}(\beta_{2},\beta_{3})\ \xi_{k}$, with a slight
change in the definition of $\beta_{\lambda}$, without any other
effect at the present order of approximation.
Namely, it is sufficient to substitute the dynamical variable
$\beta_{\lambda}$ with the new variable
$$\beta^{\rm o}_{\lambda}=\beta_{\lambda}+\frac{1}{2}g_{\lambda k}(\beta_{2},%
\beta_{3})\xi_{k}^{2}$$
(64)
to obtain
$$\dot{\beta}_{\lambda}=\dot{\beta}^{\rm o}_{\lambda}-g_{\lambda k}(\beta_{2},%
\beta_{3})\xi_{k}\dot{\xi}_{k}+0(\xi_{k}^{2})$$
(65)
and therefore, up to the second order in $\xi_{k}$,
$$\displaystyle\dot{\beta}_{\lambda}^{2}$$
$$\displaystyle+2g_{\lambda k}(\beta_{2},\beta_{3})\xi_{k}\dot{\beta}_{\lambda}%
\dot{\xi}_{k}$$
(66)
$$\displaystyle=$$
$$\displaystyle\left(\dot{\beta}^{\rm o}_{\lambda}-g_{\lambda k}(\beta_{2},\beta%
_{3})\xi_{k}\dot{\xi}_{k}\right)^{2}+2g_{\lambda k}(\beta_{2},\beta_{3})\xi_{k%
}\dot{\beta}_{\lambda}\dot{\xi}$$
$$\displaystyle\approx$$
$$\displaystyle\left(\dot{\beta}^{\rm o}_{\lambda}\right)^{2}.$$
This argument can be easily extended to the case in which the
dependence of $\mathcal{A}_{\lambda k}$ on $\xi_{k}$ comes from the
dependence on $\xi_{k}$ of one of the terms at the denominator in the
eq. (58).
Appendix C Quantization according to the Pauli rule
In this Appendix, we discuss some aspects of the quantization of the
kinetic energy expression given, e.g., in the
eq. (21),
by means of the Pauli procedure: namely, if
the classical expression of the kinetic energy in terms of the time
derivatives of the dynamical variables $\xi_{\mu}$ is
$$T=\frac{1}{2}\sum_{\mu,\nu}\mathcal{G}_{\mu,\nu}\dot{\xi}_{\mu}\dot{\xi}_{\nu}$$
(67)
the corresponding quantum operator has the form
$$\hat{T}=-\frac{\hbar^{2}}{2}G^{-1/2}\sum_{\mu,\nu}\frac{\partial}{\partial\xi_%
{\mu}}\ G^{1/2}\mathcal{A}_{\mu,\nu}\frac{\partial}{\partial\xi_{\nu}}$$
(68)
where $G={\rm Det}\ \mathcal{G}$ and $\mathcal{A}=\mathcal{G}^{-1}$.
The choice of the
“best set” of
dynamical variables is, in part, related to the expression of the
potential-energy term, and it is not obvious that it will eventually
coincide with the one discussed in the section II.5. However, it
can be useful to explore the properties of the kinetic-energy
operator in the particular model in which the matrix of coefficients
${\mathcal{G}}$ is exactly that of the Table 2, with the
non-diagonal terms confined in one single line (and column) of the
lower $6\times 6$ sub-matrix.
From now on, we limit our
discussion to the lowest $6\times 6$ sub-matrix $\mathcal{G}_{1}$ of the
matrix $\mathcal{G}$ (and of the matrices derived from $\mathcal{G}$,
Tables 6 ,7) as
the six corresponding variables – $\varphi,\ \chi,\ \vartheta,\ \theta_{1},\ \theta_{2},\ \theta_{3}$ – are effectively decoupled from
the others.
The formal procedure is discussed, e.g., in the Ch.s 5 and
6 of the ref. Eisenberg and Greiner (1987).
The first pass is the substitution of the intrinsic components
$q_{1},\ q_{2},\ q_{3}$ of the angular velocity with the time derivatives of
the three Euler angles $\theta_{1},\ \theta_{2},\ \theta_{3}$:
$$q_{k}=\sum_{i}V_{ki}\dot{\theta}_{i}$$
(69)
with
$$V=\left|\begin{array}[]{ccc}-\cos\theta_{3}\sin\theta_{2}&\sin\theta_{3}&0\\
\phantom{-}\sin\theta_{3}\sin\theta_{2}&\cos\theta_{3}&0\\
\cos\theta_{2}&0&1\\
\end{array}\right|\ .$$
(70)
As a consequence of this substitution, the $6\times 6$ matrix
$\mathcal{G}_{1}$ transforms according to the relation
$$\mathcal{G}_{1}^{\prime}=\widetilde{W}\ \ \mathcal{G}_{1}\ W,\hskip 56.905512%
ptW=\left|\begin{array}[]{cc}\{1\}&\{0\}\\
\{0\}&V\end{array}\right|$$
(71)
(where $\widetilde{W}$ is the transpose of the matrix $W$ and $\{1\}$,
$\{0\}$ are the $3\times 3$ unit matrix and null matrix) and
takes the form shown in the Table 5.
The next step is the inversion of the matrix $\mathcal{G}_{1}^{\prime}$.
In doing this, the explicit form (20) of $\mathcal{J}_{3}$
has been introduced
and second-order terms in the small amplitudes have been
neglected.
At this point, it is necessary to introduce the intrinsic components
of the angular momentum operator ($L_{1},\ L_{2},\ L_{3}$) in the place of
the derivatives with respect to the Euler angles. The expression of
$L_{k}$ in terms of $\partial/\partial\theta_{k}$ and vice-versa are
given, e.g. in the Ch. 5 of ref Eisenberg and Greiner (1987). One gets
$$-i\frac{\partial}{\partial\theta_{k}}=\sum V_{ik}\hat{L}_{i}$$
(72)
with the matrix $V$ given by the eq. (70).
It is also convenient to define the quantum operators
$$\displaystyle\hat{p}_{\varphi}$$
$$\displaystyle=$$
$$\displaystyle-i\partial\ /\ \partial\varphi$$
$$\displaystyle\hat{p}_{\chi}$$
$$\displaystyle=$$
$$\displaystyle-i\partial\ /\ \partial\chi$$
$$\displaystyle\hat{p}_{\vartheta}$$
$$\displaystyle=$$
$$\displaystyle-i\partial\ /\ \partial\vartheta$$
(73)
to obtain
$$\displaystyle-i$$
$$\displaystyle\left\{\frac{\partial}{\partial\varphi},\ \frac{\partial}{%
\partial\chi},\ \frac{\partial}{\partial\vartheta},\ \frac{\partial}{\partial%
\theta_{1}},\ \frac{\partial}{\partial\theta_{2}},\ \frac{\partial}{\partial%
\theta_{3}}\right\}$$
(74)
$$\displaystyle=\ \widetilde{W}\ \left\{\hat{p}_{\varphi},\ \hat{p}_{\chi},\ %
\hat{p}_{\vartheta},\ \hat{L}_{1},\ \hat{L}_{2},\ \hat{L}_{3}\right\}.$$
The introduction of a set of conjugate momenta $\hat{p}_{k}$ such that
$$-i\frac{\partial}{\partial\xi_{k}}=\sum_{i}W_{ik}\hat{p}_{i}$$
(75)
in the eq. (68) gives
$$\displaystyle\hat{T}$$
$$\displaystyle=$$
$$\displaystyle\frac{\hbar^{2}}{2}\left[\sum_{\mu,\nu}(W\mathcal{A}\widetilde{W}%
)_{\mu,\nu}\ \hat{p}_{\mu}\hat{p}_{\nu}\right.$$
(76)
$$\displaystyle\left.+i\ G^{-1/2}\sum_{\mu,\nu}\frac{\partial\ G^{1/2}(\mathcal{%
A}\widetilde{W})_{\mu,\nu}}{\partial\xi_{\mu}}\ \hat{p}_{\nu}\right].$$
In our case, $\mathcal{A}=(\mathcal{G}_{1}^{\prime} )^{-1}$,
$\mathcal{G}_{1}^{\prime}=\widetilde{W}\mathcal{G}_{1}W$, so that
$$W\mathcal{A}\widetilde{W}=W\left(W^{-1}\mathcal{G}_{1}^{-1}\widetilde{W}^{-1}%
\right)\widetilde{W}=\mathcal{G}_{1}^{-1}$$
(77)
and the first term of the eq. (76) takes the form given in the
Table 7.
The second term of the eq. (76) vanishes.
In fact, only the derivatives with respect to $\theta_{2}$ or
$\theta_{3}$ could give a contribution to the sum, as the other
variables do not appear in the elements of the sub-matrix. Moreover,
the determinant $G_{1}$ is simply proportional to $\sin^{2}\theta_{2}$
(with the proportionality factor depending on dynamical variables
which are outside the present sub-space), and
the term $G^{1/2}$ in the equation (76) can be replaced by
$\sin\theta_{2}$.
The matrix $i(\mathcal{G}_{1}^{\prime})^{-1}\widetilde{W}\sin\theta_{2}$
is given in the
Table 6. The sum of the derivatives of each element of
the 5th row, with respect to $\theta_{2}$, plus the corresponding one of
the 6th row, with respect to $\theta_{3}$, would give the coefficient of
the corresponding momentum operator, but
it is easy to verify they cancel each other.
Appendix D The possible forms of the potential term
The form of the approximate potential–energy expression depends on
the details of the underlying microscopic structure that the model
should
try to simulate. There are, however, some general rules to which the
expression of the potential energy must conform: it must be invariant
under space rotation, time reversal, and parity.
Irreducible tensors of different rank have been constructed with each of
the basic tensors $a^{(2)}$ and $a^{(3)}$, and scalar products of
tensors of equal rank have been considered.
In order to be invariant under time reversal, each term must
contain an even number of octupole amplitudes $a^{(3)}_{\mu}$.
Moreover, due to symmetry, only tensors
$T^{(K)}_{\lambda\lambda}=[a^{(\lambda)}\otimes a^{(\lambda)}]^{(K)}$
of even rank $K$ can be obtained
with the coupling of two identical tensors $a^{(\lambda)}$.
There are therefore two independent invariants of order 2,
$\big{(}a^{(2)}\cdot a^{(2)}\big{)}$ and $\big{(}a^{(3)}\cdot a^{(3)}\big{)}$,
two of order 3, $\big{(}a^{(2)}\cdot T^{(2)}_{22}\big{)}$
and $\big{(}a^{(2)}\cdot T^{(2)}_{33}\big{)}$,
and seven of order 4, namely
$\big{(}T^{(K)}_{22}\cdot T^{(K)}_{22}\big{)}$ ($K=2,4$),
$\big{(}T^{(K)}_{33}\cdot T^{(K)}_{33}\big{)}$ ($K=2,4,6$),
and $\big{(}T^{(K)}_{22}\cdot T^{(K)}_{33}\big{)}$ ($K=2,4$).
Additional fourth-order invariants of the form
$\big{(}[a^{(2)}\otimes a^{(3)}]^{(K)}\cdot[a^{(2)}\otimes a^{(3)}]^{(K)}\big{)}$ ($K=1-5$), are not independent from the above
ones, and can be expressed as linear combination of them with the
standard rules of angular momentum recoupling.
Invariant expressions up to the fourth order in the amplitudes
$a^{(\lambda)}_{\mu}$ ($\lambda=2,3$) are shown in the
Table 8, in terms of the dynamical variables
$\beta_{2}$, $\gamma_{2}$, $\beta_{3}$, $\gamma_{3}$, $w^{2}=(X^{2}+Y^{2})$,
$\xi$, and $v^{2}=(\eta^{2}+\zeta^{2})$. Expressions corresponding to
different choices of the dynamical variables can be easily obtained.
Here, only the terms up to the second order in the series expansion of
the “small” (non-axial) amplitudes are given. In this approximation,
the fourth-order invariants built with the quadrupole amplitudes
($C_{1},\ C_{2}$)
result to be proportional to each other and to the square of the
corresponding second-order invariant ($A_{1}^{2}$).
Moreover, as it is clear from the Table 8,
the three fourth-order invariants built with the octupole
amplitudes ($C_{3},\ C_{4},\ C_{5}$)
and the square of the corresponding second-order invariant ($A_{2}^{2}$)
are not linearly independent of one another, and provide only two
independent relations.
One can finally observe that the variables $X,\ Y,\ \eta$ and $\zeta$
always appear only in the combinations $(X^{2}+Y^{2})$,
$(\eta^{2}+\zeta^{2})$.
As long as the expression of the potential energy does only depend
on the invariants up to the fourth order,
the angles $\vartheta$ and $\varphi$ defined in the Eq. (17)
are not subject to a restoring force.
The situation is more complicated for the three variables related to
the $\mu=\pm 2$ components of $a^{\lambda}_{\mu}$. Also in this case,
however, it is possible to construct combinations of invariants that
contain only linear combinations of the squares of the variables
$\xi$, $\gamma=\sqrt{5}\gamma_{2}-\gamma_{3}$, and
$\gamma_{0}=\beta_{2}^{2}\gamma_{2}+\sqrt{5}\beta_{3}^{2}\gamma_{3}$.
Moreover, in this case, the first two of them do appear in the
combination
$\beta_{2}\beta_{3}\gamma^{2}+2\big{(}\beta_{2}^{2}+5\beta_{3}^{2}\big{)}\xi^{2}$.
It is therefore possible to imaging a situation in which the potential
energy is independent also of the angle $\chi$ (although this is is not
a direct consequence of the model, as it is the case for the angles
$\vartheta$ and $\varphi$).
Acknowledgements.
We have the pleasure to thank
Prof. F. Iachello and Prof. B.R. Mottelson for helpful discussions.
References
Iachello (2000a)
F. Iachello,
Phys. Rev. Lett. 85,
3580 (2000a).
Iachello (2000b)
F. Iachello,
Phys. Rev. Lett. 87,
052502 (2000b).
Iachello (2001)
F. Iachello,
Phys. Rev. Lett. 91,
132502 (2001).
Casten and Zamfir (2001)
R. F. Casten and
N. V. Zamfir,
Phys. Rev. Lett. 87,
052503 (2001).
Krucken et al. (2002)
R. Krucken,
B. Albanna,
C. Bialik,
R. F. Casten,
J. R. Cooper,
A. Dewald,
N. V. Zamfir,
C. J. Barton,
C. W. Beausang,
M. A. Caprio,
et al., Phys. Rev. Lett.
88, 232501
(2002).
Bizzeti (2003)
P. G. Bizzeti, in
Symmetries in Physics, edited by
A. Vitturi and
R. Casten
(World Scientific, Singapore, 2003).
Bizzeti and Bizzeti-Sona (2002)
P. G. Bizzeti and
A. M. Bizzeti-Sona,
Phys. Rev. C 66,
031301 (2002).
Hutter et al. (2003)
C. Hutter et al.,
Phys. Rev. C 67,
054315 (2003).
Clark et al. (2003)
R. M. Clark
et al., Phys. Rev. C
68, 037301
(2003).
Fransen et al. (2004)
C. Fransen,
N. Pietralla,
A. Linnemann,
V.Werner, and
R.Bijker, Phys. Rev. C
69, 014313
(2004).
Tonev et al. (2004)
D. Tonev,
A. Dewald,
T. Klug,
P. Petkov,
J. Jolie,
A. Fitzler,
O. Moller,
S. Heinze,
P. von Brentano,
and R. F.
Casten, Phys. Rev. C
69, 034334
(2004).
Bijker et al. (2003)
R. Bijker,
R. F. Casten,
N. V. Zamfir,
and E. A.
McCutchan, Phys. Rev. C
68, 064304
(2003).
Bijker et al. (2004)
R. Bijker,
R. F. Casten,
N. V. Zamfir,
and E. A.
McCutchan, Phys. Rev. C
69, 059901(E)
(2004).
Caprio (2004)
M. A. Caprio,
Phys. Rev.C 69,
044307 (2004).
Bonatsos et al. (2004)
D. Bonatsos,
D. Lenis,
N. Minkov,
D. Petrellis,
P. P. Raychev,
and P. A.
Terziev, Phys. Rev. C
70, 024305
(2004).
Pietralla and Gorbachenko (2004)
N. Pietralla and
O. Gorbachenko,
Phys. Rev. C 70,
011304 (2004).
Bizzeti and Bizzeti-Sona (2004)
P. G. Bizzeti and
A. M. Bizzeti-Sona,
Europ. J. Phys. A 20,
179 (2004).
Bohr (1952)
D. A. Bohr,
Dan. Mat. Fys. Medd. 26
(1952), (n.14).
Engel and Iachello (1985)
J. Engel and
F. Iachello,
Phys. Rev. Lett. 54,
1126 (1985).
Alonso et al. (1995)
C. E. Alonso,
J. M. Arias,
A. Frank,
H. M. Sofia,
S. M. Lenzi, and
A. Vitturi,
Nucl. Phys. A 586,
100 (1995).
Raduta and Ionescu (2003)
A. A. Raduta and
D. Ionescu,
Phys. Rev. C 67,
044312 (2003), (and
references therein).
Raduta et al. (1996a)
A. A. Raduta,
D. Ionescu,
I. Ursu, and
A. Faessler,
Nucl. Phys. A 720
(1996a).
Zamfir and Kusnezov (2001)
N. V. Zamfir and
D. Kusnezov,
Phys. Rev. C 63,
054306 (2001).
Zamfir and Kusnezov (2003)
N. V. Zamfir and
D. Kusnezov,
Phys. Rev. C 67,
014305 (2003).
Shneidman et al. (2002)
T. M. Shneidman,
G. G. Adamian,
N. V. Antonenko,
R. Jolos, and
W. Scheid,
Phys. Letters B 526,
322 (2002).
Rohozinski (1988)
S. Rohozinski,
Rep. Progr. Phys. 51,
541 (1988).
Butler and Nazarewicz (1996)
P. A. Butler and
W. Nazarewicz,
Rev. Mod. Phys. 68,
349 (1996), (and references
therein).
Lipas and Davidson (1961)
P. O. Lipas and
J. P. Davidson,
Nucl. Phys. 26,
80 (1961).
Denisov and Dzyublik (1995)
V. Y. Denisov and
A. Dzyublik,
Nucl. Phys. A 589,
17 (1995).
Jolos and von Brentano (1999)
R. V. Jolos and
P. von Brentano,
Phys. Rev. C 60,
064317 (1999).
Minkov et al. (2000)
N. Minkov,
S. B. Drenska,
P. P. Raychev,
R. P. Roussev,
and D. Bonatsos,
Phys. Rev. C 63,
044305 (2000).
Minkov et al. (2004)
N. Minkov,
S. Drenska,
P. Yotov, and
W. Scheid, in
Proceedings of the 23rd International Workshop on
Nuclear Theory (Rila, Bulgaria 2004), edited by
S. Dimitrova
(Heron Press, Sofia, 2004).
Donner and Greiner (1966)
W. Donner and
W. Greiner,
Z. Phys. 197,
440 (1966).
Eisenberg and Greiner (1987)
J. M. Eisenberg
and W. Greiner,
Nuclear Theory, vol. I
(North Holland, Amsterdam, 1987),
3rd ed.
Wexler and Dussel (1999)
C. Wexler and
G. G. Dussel,
Phys. Rev. C 60,
014305 (1999).
Wilets and Jean (1956)
L. Wilets and
M. Jean,
Phys. Rev. 102,
788 (1956).
Pauli (1933)
W. Pauli, in
Handbook der Physik, edited by
A. Smekal
(Springer, Berlin, 1933), vol.
XXIV/I.
Abramowitz and Stegun (1970)
M. Abramowitz and
I. A. Stegun,
Handbook of mathematical functions
(New York, 1970), section
21.6.
(39)
ENSDF data base, http//:www.nndc.bnl.gov.
Cocks et al. (1999)
J. C. F. Cocks
et al., Nucl. Phys. A
645, 61 (1999).
Guntner et al. (2000)
C. Guntner,
R. Krucken,
C. J. Barton,
C. W. Beausang,
R. F. Casten,
G. Cata-Danil,
J. R. Cooper,
J. Groeger,
J. Novak, and
N. V. Zamfir,
Phys. Rev. C 61,
064602 (2000).
Raduta et al. (1996b)
A. A. Raduta,
N. Lo Judice,
and I. Ursu,
Nucl. Phys. A 608
(1996b).
Smith et al. (1995)
J. F. Smith
et al., Phys. Rev. Lett.
75, 1050 (1995).
Arias (2001)
J. M. Arias,
Phys. Rev. C 63,
034308 (2001).
Tsvekov et al. (2002)
A. Tsvekov,
J. Kvasil, and
R. Nazmitdinov,
J. Phys. G: Nucl. Part. Phys.
28, 2187 (2002).
Bohr and Mottelson (1957)
D. A. Bohr and
B. R. Mottelson,
Nucl. Phys. 4
(1957).
Bohr and Mottelson (1958)
D. A. Bohr and
B. R. Mottelson,
Nucl. Phys. 9
(1958).
Strutinski (1957)
V. Strutinski,
J. Nucl. Energy 4
(1957).
Lipas (1963)
P. O. Lipas,
Nucl. Phys. 40
(1963).
Nazarewicz and Olanders (1985)
W. Nazarewicz and
P. Olanders,
Nucl. Phys. A 441,
420 (1985). |
Reconfiguration of Non-crossing Spanning Trees
Oswin Aichholzer
University of Technology Graz, Austria. [email protected]
Brad Ballinger
Cal Poly Humboldt, USA. [email protected]
Therese Biedl
University of Waterloo, Canada. {biedl,alubiw,jayson.lynch}@uwaterloo.ca
Mirela Damian
Villanova University, USA. [email protected]
Erik D. Demaine
MIT Computer Science and Artificial Intelligence Laboratory, USA. [email protected]
Matias Korman
Siemens Electronic Design Automation, USA. [email protected]
Anna Lubiw
University of Waterloo, Canada. {biedl,alubiw,jayson.lynch}@uwaterloo.ca
Jayson Lynch
University of Waterloo, Canada. {biedl,alubiw,jayson.lynch}@uwaterloo.ca
Josef Tkadlec
Harvard Unversity, USA. [email protected]
Yushi Uno
Osaka Metropolitan University, Japan. [email protected]
Abstract
For a set $P$ of $n$ points in the plane in general position,
a non-crossing spanning tree is a spanning tree of the points
where every edge is a straight-line segment between a pair of points and
no two edges intersect except at a common endpoint.
We study the problem of reconfiguring one non-crossing spanning tree of $P$ to another
using a sequence of flips where each flip removes one edge
and adds one new edge so that the result is again a non-crossing spanning tree of $P$.
There is a known upper bound of $2n-4$ flips [Avis and Fukuda, 1996] and a lower bound of $1.5n-5$ flips.
We give a reconfiguration algorithm that uses at most $2n-3$ flips but reduces that to $1.5n-2$ flips when one tree is a path and either: the points are in convex position; or the path is monotone in some direction.
For points in convex position, we prove an upper bound of
$2d-\Omega(\log d)$ where $d$ is half the size of the symmetric difference between the trees.
We also examine whether the happy edges (those common to the initial and final trees) need to flip,
and we find exact minimum flip distances for small point sets using exhaustive search.
1 Introduction
Let $P$ be a set of $n$ points in the plane in general position.
A non-crossing spanning tree is a spanning tree of $P$
whose edges are straight line segments between pairs of points such that no two edges intersect except at a common endpoint.
A reconfiguration step or flip removes one edge of a non-crossing spanning tree and adds one new edge so that the result is again a non-crossing spanning tree of $P$.
We study the problem of
reconfiguring one non-crossing spanning of $P$ to another
via a sequence of flips.
Researchers often consider three problems about reconfiguration, which are most easily expressed in terms of the reconfiguration graph that has a vertex for each configuration (in our case, each non-crossing spanning tree) and an edge for each reconfiguration step. The problems are:
(1) connectivity of the reconfiguration graph—is reconfiguration always possible?
(2) diameter of the reconfiguration graph—how many flips are needed for reconfiguration in the worst case? and (3) distance in the reconfiguration graph—what is the complexity of finding the minimum number of flips to reconfigure between two given configurations?
For reconfiguration of non-crossing spanning trees,
Avis and Fukuda [10, Section 3.7] proved
that reconfiguration is always possible,
and that at most $2n-4$ flips are needed.
Hernando et al. [20] proved a lower bound of $1.5n-5$ flips.
Even for the special case of points in convex position, there are no better upper or lower bounds known.
Our two main results
make some progress in reducing the diameter upper bounds.
(1) For points in general position, we give a reconfiguration algorithm
that uses at most $2n-3$ flips but reduces that to $1.5n-2$ flips
in two cases: (1) when the points are in convex position and one tree is a path; (2) for general point sets when one tree is a monotone path.
The algorithm first flips one tree
to a downward tree
(with each vertex connected to a unique higher vertex)
and the other
tree to an upward tree (defined symmetrically)
using $n-2$ flips—this is where we save when one tree is a path. After that, we give an algorithm to flip from a downward tree $T_{D}$ to an upward tree $T_{U}$ using at most $n-1$
“perfect” flips each of which removes an edge of $T_{D}$ and adds an edge of $T_{U}$. The algorithm is simple to describe, but proving that intermediate trees remain non-crossing is non-trivial.
We also show that $1.5n-5$ flips
may be required, even in the two special cases.
See Section 2.
(2) For points in convex position, we improve the upper bound on the number of required flips to
$2d-\Omega(\log d)$ where $d$ is half the size of the symmetric difference between the trees.
So $d$ flips are needed in any flip sequence, and $2d$ is an upper bound.
The idea is to find an edge $e$ of one tree that is crossed by at most (roughly) $d/2$ edges of the other tree,
flip all but one of the crossing edges temporarily to the convex hull (this will end up costing 2 flips per edge), and then flip the last crossing edge to $e$. Repeating this saves us one flip, compared to the $2d$ bound, for each of the (roughly) $\log d$ repetitions.
See Section 3.
Notably, neither of our algorithms uses the common—but perhaps limited—technique of
observing that the diameter is at most twice the radius, and bounding the radius of the reconfiguration
graph by identifying a “canonical” configuration that every other configuration can be flipped to. Rather, our
algorithms find reconfiguration sequences tailored to the specific input trees.
In hopes of
making further progress on the diameter and distance problems,
we address the question of which edges need to be involved in a minimum flip sequence from an initial non-crossing spanning tree $T_{I}$ to a final non-crossing spanning tree $T_{F}$.
We say that the edges of $T_{I}\cap T_{F}$ are happy edges, and we formulate the Happy Edge Conjecture that for points in convex position, there is a minimum flip sequence that never flips happy edges.
We prove the conjecture for happy convex hull edges.
See Section 4.
More generally, we say that a reconfiguration problem has the “happy element property” if elements that are common to the initial and final configurations can remain fixed in a minimum flip sequence.
Reconfiguration problems that satisfy the happy element property seem easier.
For example, the happy element property holds for reconfiguring
spanning trees in a graph (and indeed for matroids more generally), and the distance problem is easy.
On the other hand,
the happy element property
fails for reconfiguring triangulations of a point set in the plane, and for the problem of token swapping on a tree [12],
and in both cases,
this is the key to constructing gadgets
to prove that the distance problem is NP-hard [26, 31, 4].
As an aside, we note that
for reconfiguring triangulations of a set of points in convex position—where the distance problem is the famous open question of rotation distance in binary trees—the happy element property holds [35], which may be why no one has managed to prove that the distance problem is NP-hard.
Finally,
we implemented a combinatorial search program to compute the diameter (maximum reconfiguration distance between two trees) and radius of the reconfiguration graph for points in convex position.
For $6\leq n\leq 12$ the diameter is $\lfloor 1.5n-4\rfloor$ and the radius is $n-2$. In addition we provide the same information for the special case when the initial and final trees are non-crossing spanning paths, though intermediate configurations may be more general trees. We also verify
the Happy Edge Conjecture for $n\leq 10$ points in convex position.
See Section 5.
1.1 Background and Related Results
Reconfiguration is about changing one structure to another, either through continuous motion or through discrete changes.
In mathematics, the topic has a vast and deep history, for example in knot theory,
and the study of bounds on the simplex method for linear programming.
Recently, reconfiguration via discrete steps has become a focused research area, see the surveys by van den Heuvel [36] and Nishimura [29].
Examples include sorting a list by swapping pairs of adjacent elements, solving a Rubik’s cube, or changing one colouring of a graph to another.
With discrete reconfiguration steps, the reconfiguration graph is well-defined. Besides questions of diameter and distance in the reconfiguration graph, there is also research on enumeration via a Hamiltonian cycle in the reconfiguration graph, see the recent survey [27], and on mixing properties to find random configurations, see [33].
Our work is about reconfiguring one graph to another.
Various reconfiguration steps have been considered, for example exchanging one vertex for another (for reconfiguration of paths [14], independent sets [23, 11], etc.), or exchanging multiple edges
(for reconfiguration of matchings [13]).
However, we concentrate on elementary steps (often called edge flips) that exchange one edge for one other edge.
A main example of this is reconfiguring one spanning tree of a graph to another, which can always be accomplished by “perfect” flips that add an edge of the final tree and delete an edge of the initial tree—more generally, such a perfect exchange sequence is possible when reconfiguring bases of a matroid.
Our focus is on geometric graphs whose vertices are points in the plane
and whose edges are non-crossing line segments between the points.
In this setting, one well-studied problem is reconfiguring between triangulations of a point set in the plane, see the survey by Bose and Hurtado [15]. Here, a flip replaces
an edge in a convex quadrilateral by the other diagonal of the quadrilateral.
For the special case of $n$ points in convex position this is equivalent to rotation of an edge in a given (abstract) rooted binary tree, which is of interest in the study of splay trees, and the study of phylogenetic trees in computational biology.
While an upper bound for the reconfiguration distance of $2n-10$ is known to be tight for $n>12$ [32, 35],
the complexity of computing the shortest distance between
two triangulations of a convex point set
(equivalently between two given binary trees) is still unknown. See [6, 26, 31] for related hardness results for the flip-distance of triangulations of point sets and simple polygons.
Another well-studied problem for geometric graphs is reconfiguration of non-crossing perfect matchings. Here a flip typically exchanges matching and non-matching edges in non-crossing cycles, and there are results for a single cycle of unbounded length [21], and for multiple cycles [3, 34].
For points in convex position, the flip operation on a single cycle of length 4 suffices to connect the reconfiguration graph [19], but this is open for general point sets.
For a more general flip operation that connects any two disjoint matchings whose union is non-crossing, the reconfiguration graph is connected for points in convex position [1] (and, notably, the diameter is less than twice the radius), but connectivity is open for general point sets [3, 22].
The specific geometric graphs we study are non-crossing (or “plane”) spanning trees of a set of points in the plane.
For points in convex position, these have been explored for enumeration [30], and for a duality with quadrangulations and consequent lattice properties, e.g., see [9] and related literature.
For non-crossing spanning trees of a general point set in the plane,
there
are
several basic reconfiguration operations that can be used to transform these trees into each other. The one we use in this work is the simple edge exchange of an edge $e$ by an edge $e^{\prime}$ as described above. If
we require that
$e$ and $e^{\prime}$ do not cross, then this operation is called a compatible edge exchange. Even more restricted is an edge rotation, where
$e=uv$ and $e^{\prime}=uw$
share a common vertex $u$.
If the triangle $u,v,w$ is not intersected by an edge of the two involved trees, then this is called an empty triangle edge rotation. The most restricted operation is the edge slide (see also Section 4.3) where the edge $vw$ has to exist in both trees. The name comes from viewing this transformation as sliding one end of the edge $e$ along $vw$ to $e^{\prime}$ (and rotating the other end around $u$) and at no time crossing through any other edge of the trees.
For an overview and detailed comparison of the described operations see Nichols et al. [28].
It has been shown that the reconfiguration graph of non-crossing spanning trees is even connected
for the “as-local-as-possible” edge slide operation [2]. See also [7] where a tight bound of $\Theta(n^{2})$ steps for the diameter is show. This implies that also for the other reconfiguration operations the flip graph is connected.
For edge exchange, compatible edge exchange, and edge rotation a linear upper bound for the diameter is known [10], while for empty triangle edge rotations an upper bound of $O(n\log n)$ has been shown recently [28].
For all operations (except edge slides) the best known lower bound is $1.5n-5$ [20].
There are several variants for the reconfiguration of spanning trees. For example, the operations can be performed in parallel, that is, as long as the exchanges (slides etc) do not interfere with each other, they can be done in one step; see [28] for an overview of results.
In a similar direction of a more global operation we can say that two non-crossing spanning trees are compatible, if their union is still crossing free. A single reconfiguration step then transforms one such tree into the other. A lower bound of $\Omega(\log n/\log\log n)$ [16] and an upper bound of $O(\log n)$ [2] for the diameter of this reconfiguration graph has been shown.
Another variation is that the edges are labeled (independent of the vertex labels), and if an edge is exchanged the replacement edge takes that label. In this way two geometrically identical trees can have a rather large transformation distance.
For labeled triangulations
there is a good characterization of when reconfiguration is possible, and a polynomial bound on the number of steps required [25].
The reconfiguration of non-crossing spanning paths (where each intermediate configuration must be a path) has also been considered. For points in convex position, the diameter of the reconfiguration graph is $2n-6$ for $n\geq 5$ [8, 17]. Surprisingly, up to now it remains an open problem if the reconfiguration graph of non-crossing spanning paths is connected for general point sets [5].
1.2 Definitions and Terminology
Let $P$ be a set of $n$ points in
general position, meaning that no three points are collinear.
The points of $P$ are in convex position if the boundary of the convex hull of $P$ contains all the points of $P$.
An edge is a line segment joining two points of $P$, and a spanning tree $T$ of $P$ is a set of $n-1$ edges that form a tree. Two edges cross if they intersect but do not share a common endpoint.
A non-crossing spanning tree is a spanning tree such that no two of its edges cross.
When we write “two non-crossing spanning trees,” we mean that each tree is non-crossing but we allow edges of one tree to cross edges of the other tree.
We sometimes consider the special case where a non-crossing spanning tree of $P$ is a path. A path is monotone if there is a direction in the plane such that
the order of points along the path matches the order of points in that direction.
For a spanning tree of a graph,
a flip removes one edge and adds one new edge to obtain a new spanning tree, i.e., spanning trees $T$ and $T^{\prime}$ are related by a flip if $T^{\prime}=T\setminus\{e\}\cup\{e^{\prime}\}$, where $e\in T$ and $e^{\prime}\notin T$.
The same definition applies
to non-crossing spanning trees: If $T$ is a non-crossing spanning tree of $P$, and $T^{\prime}=T\setminus\{e\}\cup\{e^{\prime}\}$ is a non-crossing spanning tree of $P$, then we say that $T$ and $T^{\prime}$ are related by a flip. We allow $e$ and $e^{\prime}$ to cross.
Let $T_{I}$ and $T_{F}$ be initial and final non-crossing spanning trees of $P$. A flip sequence from $T_{I}$ to $T_{F}$ is a sequence of flips that starts with $T_{I}$ and ends with $T_{F}$ and such that each intermediate tree is a non-crossing spanning tree.
We say that $T_{I}$ can be reconfigured to $T_{F}$
using $k$ flips (or “in $k$ steps”) if there is a reconfiguration sequence of length at most $k$.
The flip distance from $T_{I}$ to $T_{F}$ is the minimum length of a flip sequence.
The edges of $T_{I}\cap T_{F}$ are called happy edges.
Thus, $T_{I}\cup T_{F}$ consists of the happy edges together with the symmetric difference $(T_{I}\setminus T_{F})\cup(T_{F}\setminus T_{I})$. We have $|T_{I}\setminus T_{F}|=|T_{F}\setminus T_{I}|$. A flip sequence of length $|T_{I}\setminus T_{F}|$ is called a perfect flip sequence. In a perfect flip sequence, every flip removes an edge of $T_{I}$ and adds an edge of $T_{F}$—these are called perfect flips.
2 A two-phase reconfiguration approach
In this section we give a new algorithm to reconfigure between two non-crossing spanning trees on $n$ points using at most $2n-3$ flips. This is basically the same as the upper bound of $2n-4$ originally achieved by Avis and Fukuda [10], but the advantage of our new algorithm is that it gives a bound of $1.5n-2$ flips when one tree is a path and either: (1) the path is monotone; or (2) the points are in convex position.
Furthermore, for these two cases, we show a lower bound of $1.5n-5$ flips, so the bounds are tight up to the additive constant.
Before proceeding, we mention one way in which our upper bound result differs from some other reconfiguration bounds.
Many of those bounds (i.e., upper bounds on the diameter $d$ of the reconfiguration graph) are actually bounds on the radius $r$ of the reconfiguration graph.
The idea is to identify a
“canonical” configuration and prove that its distance to any other configuration is at most $r$, thus proving that the diameter $d$ is at most $2r$. For example, Avis and Fukuda’s $2n$ bound is achieved via a canonical star centered at a convex hull point. As another example,
the bound of $O(n^{2})$ flips between triangulations of a point set can be proved using the Delaunay triangulation as a canonical configuration, and
the bound of $2n$ flips for the special case of points in convex position uses a canonical star triangulation [35].
For some reconfiguration graphs $d$ is equal to $2r$ (e.g., for the Rubik’s cube, because of the underlying permutation group). However, in general, $d$ can be less than $2r$, in which case,
using a canonical configuration will not give the best diameter bound.
Indeed, our result does not use a canonical configuration, and we do not bound the radius of the reconfiguration graph.
Our algorithm has two phases. In the first phase, we reconfigure the input trees in a total of at most $n-2$ flips so that one is “upward” and one is “downward” (this is where we save if one tree is a path). In the second phase we show that an upward tree can be reconfigured to a downward tree in $n-1$ flips. We begin by defining these terms.
Let $P$ be a set of $n$ points in general position. Order the points $v_{1},\ldots,v_{n}$ by increasing $y$-coordinate
(if necessary, we slightly perturb the point set to ensure that no two $y$-coordinates are identical).
Let $T$ be a non-crossing spanning tree of $P$.
Imagining the edges as directed upward,
we call a vertex $v_{i}$ a sink if there are no edges in $T$ connecting $v_{i}$ to a higher vertex $v_{j},j>i$, and we call $v_{i}$ a source if there are no edges connecting $v_{i}$ to a lower vertex $v_{k},k<i$.
We call $T$ a downward tree if it has only one sink (which must then be $v_{n}$) and we call $T$ an upward tree if it has only one source (which must then be $v_{1}$).
Observe that in a downward tree every vertex except $v_{n}$ has exactly one edge connected to a higher vertex, and in an upward tree every vertex except $v_{1}$ has exactly one edge connected to a lower vertex.
2.1 Phase 1: Reconfiguring to upward/downward trees
We first bound the number of flips needed to reconfigure a single tree $T$ to be upward or downward.
If a tree has $t$ sinks, then we need at least $t-1$ flips to reconfigure it to a downward tree—we show that $t-1$ flips suffice. Note that $t$ is at most $n-1$ since $v_{1}$ cannot be a sink (this bound is realized by a star at $v_{1}$).
Theorem 1.
Let $T$ be a non-crossing spanning tree with $s$ sources and $t$ sinks.
$T$ can be reconfigured to a downward tree with $t-1\leq n-2$ flips.
$T$ can be reconfigured to an upward tree with $s-1\leq n-2$ flips.
Furthermore, these reconfiguration sequences do not flip any edge of the form
$v_{i}v_{i+1}$ where $1\leq i<n$.
Proof.
We give the proof for a downward tree, since the other case is symmetric.
The proof is by induction on $t$. In the base case, $t=1$ and the tree is downward and so no flips are needed.
Otherwise, let $v_{i},1<i<n$ be a sink. The plan is to decrease $t$ by adding an edge going upward from $v_{i}$ and removing some edge $v_{k}v_{l},k<l$ from the resulting cycle while ensuring that $v_{k}$ does not become a sink.
If there is an edge $v_{i}v_{j}$ that does not cross any edge of $T$, we say that $v_{i}$ sees $v_{j}$.
We argue that $v_{i}$ sees some vertex $v_{j}$ with $j>i$.
If $v_{i}$ sees $v_{n}$, then choose $j=n$. Otherwise the upward ray directed from $v_{i}$ to $v_{n}$ hits some edge $e$ before it reaches $v_{n}$. Continuously rotate the ray around $v_{i}$ towards the higher endpoint of $e$ until the ray reaches the endpoint or is blocked by some other vertex.
In either case this gives us a vertex $v_{j}$ visible from $v_{i}$ and higher than $v_{i}$.
For example, in Figure 1, the sink $v_{5}$ sees $v_{7}$.
Adding the edge $v_{i}v_{j}$ to $T$ creates a cycle. Let $v_{k}$ be the lowest vertex in the cycle. Then $v_{k}$ has two upward edges, and in particular, $k<i$. Remove the edge $v_{k}v_{l}$ that goes higher up. Then $v_{k}$ does not become a sink, and furthermore, if the edge $v_{k}v_{k+1}$ is in $T$, then we do not remove it.
∎
Since no vertex is both a sink and a source, any tree has $s+t\leq n$, which yields the following result that will be useful later on when we reconfigure between a path and a tree.
Corollary 2.
Let $T$ be a non-crossing spanning tree. Then either $T$ can be reconfigured to a downward tree in $0.5n-1$ flips or $T$ can be reconfigured to an upward tree in $0.5n-1$ flips. Furthermore, these reconfiguration sequences do not flip any edge of the form
$v_{i}v_{i+1}$ where $1\leq i<n$.
We next bound the number of flips needed to reconfigure two given trees into
opposite trees,
meaning that one tree is upward and one is downward.
By Theorem 1, we can easily do this in at most $2n-4$ flips (using $n-2$ flips to reconfigure each tree independently). We now show that $n-2$ flips suffice to reconfigure the two trees into opposite trees.
Theorem 3.
Given two non-crossing spanning trees on the same point set,
we can flip them into opposite trees
in at most $n-2$ flips.
Proof.
Let the trees be $T_{1}$ and $T_{2}$, and
let $s_{i}$ and $t_{i}$ be the number of sources and sinks of $T_{i}$, for $i=1,2$.
Since $s_{i}+t_{i}\leq n$,
we have
$s_{1}+t_{1}+s_{2}+t_{2}\leq 2n$.
This implies that
$s_{1}+t_{2}\leq n$ or $t_{1}+s_{2}\leq n$.
In the former case use Theorem 1 to flip $T_{1}$ upward and $T_{2}$ downward in $s_{1}-1+t_{2}-1\leq n-2$ flips; otherwise flip $T_{1}$ downward and $T_{2}$ upward in $t_{1}-1+s_{2}-1\leq n-2$ flips.
∎
2.2 Phase 2: Reconfiguring an upward tree into a downward tree
In this section we show how to reconfigure
from an initial downward tree $T_{I}$ to a final upward tree $T_{F}$ on a general point set using only perfect flips.
Thus the total number of flips will be $|T_{I}\setminus T_{F}|$.
The sequence of flips is simple to describe, and it will be obvious that each flip yields a spanning tree. What needs proving is that each intermediate tree is non-crossing.
To simplify the description of the algorithm, imagine $T_{I}$ colored red and $T_{F}$ colored blue. Refer to Figure 2.
Recall that
$v_{1},\ldots,v_{n}$ is the ordering of the points by increasing $y$-coordinate.
Define $b_{i}$ to be the (unique) blue edge in $T_{F}$ going down from $v_{i}$, $i=2,\dots,n$.
An unhappy edge is an edge of $T_{I}\setminus T_{F}$, i.e., it is red but not blue.
Reconfiguration algorithm.
Let $T_{1}=T_{I}$.
For $i=2,\ldots,n$ we create a tree $T_{i}$ that contains $b_{2},\ldots,b_{i}$ and all the happy edges.
If $b_{i}$ is happy, then $b_{i}$ is already in the current tree and we simply set $T_{i}:=T_{i-1}$.
Otherwise, consider the cycle formed by adding $b_{i}$ to $T_{i-1}$ and, in this cycle,
let $r_{i}$ be the unhappy red edge with the lowest bottom endpoint.
Note that $r_{i}$ exists, otherwise all edges in the cycle would be blue.
Set $T_{i}:=T_{i-1}\cup\{b_{i}\}\setminus\{r_{i}\}$.
This reconfiguration algorithm, applied to the trees $T_{I}$ and $T_{F}$ from Figure 2, is depicted in Figure 3.
Theorem 4.
Given a red downward tree $T_{I}$ and a blue upward tree $T_{F}$ on a general point set, the reconfiguration algorithm described above flips $T_{I}$ to $T_{F}$ using $|T_{I}\setminus T_{F}|$ perfect flips.
Proof.
It is clear that each $T_{i}$ is a spanning tree, and that each flip is perfect, so the number of flips is $|T_{I}\setminus T_{F}|$. In particular, a happy edge is never removed, so $T_{i}$ contains all happy edges.
We must show that each $T_{i}$ is non-crossing. By induction, it suffices to show that if step $i\geq 2$ adds edge $b_{i}$ and removes edge $r_{i}$, then $b_{i}$ does not cross any edge of $T_{i-1}$ except possibly $r_{i}$. We examine the structure of $T_{i-1}$.
Let $B_{i-1}$ be the subtree with edges $b_{2},\ldots,b_{i-1}$.
Note that $B_{i-1}$ is connected.
By construction, $T_{i-1}$ contains $B_{i-1}$ and all the other edges of $T_{i-1}$ are red. Let $R_{i-1}$ consist of vertices $v_{i},\ldots,v_{n}$ and the edges of $T_{i-1}$ induced on those vertices. The edges of $R_{i-1}$ are red, and $R_{i-1}$ consists of some connected components (possibly isolated vertices).
In $T_{i-1}$ each component of $R_{i-1}$ has exactly one red connector-edge joining it to $B_{i-1}$.
Thus $T_{i-1}$ consists of $B_{i-1}$, $R_{i-1}$, and the connector-edges for the components of $R_{i-1}$.
Now consider the flip performed to create $T_{i}$ by adding edge $b_{i}$ and removing edge $r_{i}$. Since $b_{i}$ is a blue edge, it cannot cross any edge of $B_{i-1}$.
Since $b_{i}$’s topmost vertex is $v_{i}$, $b_{i}$ cannot cross any edge of $R_{i-1}$.
Furthermore, $b_{i}$ cannot cross a happy edge. Thus the only remaining possibility is for $b_{i}$ to cross an unhappy connector-edge.
We will prove:
Claim 5.
If $R_{i-1}$ is disconnected, then all connector-edges are happy.
Assuming the claim, we only need to show that $b_{i}$ is non-crossing when $R_{i-1}$ is connected. Then there is only one connector-edge $r$ joining $R_{i-1}$ to $B_{i-1}$. If $r$ is happy, then $b_{i}$ cannot cross it. So assume that $r$ is unhappy. Refer to Figure 4(a).
Now, the cycle $\gamma$ in $T_{i-1}\cup\{b_{i}\}$ contains $r$, since $b_{i}$ and $r$ are the only edges between $R_{i-1}$ and $B_{i-1}$. Of the red edges in $\gamma$, $r$ is the one with the lowest bottom endpoint. This implies that $r$ is chosen as $r_{i}$, and removed.
Therefore $b_{i}$ does not cross any edges of $T_{i}$, so $T_{i}$ is non-crossing.
It remains to prove the claim. Let $C$ be a connected component of $R_{i-1}$, and let $r$ be its connector-edge. Refer to Figure 4(b). We must prove that $r$ is happy.
In the initial red tree $T_{I}$, the vertices $v_{i},\ldots,v_{n}$ induce a connected subtree, so $C$ and $R_{i-1}\setminus C$ were once connected. Suppose they first became disconnected by the removal of red edge $r_{j}$ in step $j$ of the algorithm, for some $j<i$.
Consider the blue edge $b_{j}$ that was added in step $j$ of the algorithm, and the cycle $\gamma_{j}$ in $T_{j-1}\cup\{b_{j}\}$. Now $\gamma_{j}$ must contain another edge, call it $e$, with one endpoint in $C$ and one endpoint, $v_{k}$, not in $C$.
Note that $e$ is red since it has an endpoint in $C$, and note that $v_{k}$ is not in $R_{i-1}\setminus C$ otherwise $C$ and $R_{i-1}\setminus C$ would not be disconnected after the removal of $r_{j}$.
Therefore $v_{k}$ must lie in $B_{i-1}$, i.e., $k\leq i-1$. If $e$ is unhappy, then in step $j$ the algorithm would prefer to remove $e$ instead of $r_{j}$ since $e$ is a red edge in $\gamma_{j}$ with a lower bottom endpoint.
So $e$ is happy, which means that the algorithm never removes it, and it is contained in $T_{i-1}$. Therefore $e$ must be equal to $r$, the unique connector-edge in $T_{i-1}$ between $C$ and $B_{i-1}$. Therefore $r$ is happy.
∎
2.3 Two-phase reconfiguration algorithm
We can now combine the results of Sections 2.1 and 2.2 to develop a new two-phase reconfiguration algorithm between two non-crossing spanning trees $T_{I}$ and $T_{F}$:
1.
In the first phase we reconfigure $T_{I}$ into $T^{\prime}_{I}$ and $T_{F}$ into $T^{\prime}_{F}$ such that $T^{\prime}_{I}$ and $T^{\prime}_{F}$ are opposite trees (one upward and one downward), using
Theorem 3.
2.
In the second phase we reconfigure $T^{\prime}_{I}$ into $T^{\prime}_{F}$ using only perfect flips,
as given by Theorem 4.
(Note, however, that the happy edges in $T^{\prime}_{I}$ and $T^{\prime}_{F}$ may differ from the ones in $T_{I}$ and $T_{F}$, since the first phase does not preserve happy edges).
Finally, we concatenate the reconfiguration sequences from $T_{I}$ to $T^{\prime}_{I}$, from $T^{\prime}_{I}$ to $T^{\prime}_{F}$, and the reverse of the sequence from $T_{F}$ to $T^{\prime}_{F}$.
Theorem 6.
If $T_{I}$ and $T_{F}$ are non-crossing spanning trees on a general point set, then the algorithm presented above reconfigures $T_{I}$ to $T_{F}$ in at most
$2n-3$ flips.
Proof.
By Theorem 3, the first phase of the algorithm takes at most $n-2$ flips.
By Theorem 4, the second phase uses
at most $n-1$ flips.
It follows that the total number of flips is at most $2n-3$.
∎
Theorem 7.
For a general point set, if $T_{I}$ is a non-crossing spanning tree and $T_{F}$ is a non-crossing path that is monotone in some direction, then $T_{I}$ can be reconfigured to $T_{F}$ in
at most $1.5n-2-h$
flips, where $h=|T_{I}\cap T_{F}|$ is the number of happy edges.
Furthermore, there is a lower bound of $1.5n-5$ flips, even for points in convex position and if one tree is a monotone path.
Proof.
Rotate the plane so that $T_{F}$ is $y$-monotone. Note that $T_{F}$ is then
both an upward and a downward tree. We thus have the flexibility to turn $T_{I}$ into either an upward or a downward tree in the first phase of the algorithm. By Corollary 2,
$T_{I}$ can be turned into an upward or downward tree $T^{\prime}_{I}$ in at most $0.5n-1$ flips. Furthermore, since $T_{F}$ is a $y$-monotone path, any edge in
$T_{I}\cap T_{F}$ has the form $v_{i}v_{i+1}$ for some $1\leq i<n$, and thus, by
Corollary 2,
these edges do not flip, which implies that they are still in $T^{\prime}_{I}\cap T_{F}$, so $|T^{\prime}_{I}\cap T_{F}|\geq h$.
The second phase of the algorithm uses
$|T_{I}^{\prime}\setminus T_{F}|\leq n-1-h$
perfect flips to reconfigure $T^{\prime}_{I}$ into $T_{F}$. Hence the total number of flips is at most $1.5n-2-h$.
For the lower bound, see Lemma 8 below.
∎
Lemma 8.
On any set of $n\geq 4$ points in convex position, for $n$ even, there exists a non-crossing spanning tree $T_{I}$ and a
non-crossing path $T_{F}$ such that
reconfiguring $T_{I}$ to $T_{F}$ requires at least $1.5n-5$ flips, and this bound is tight.
Proof.
Our construction is depicted in Figure 5.
Note that the tree $T_{I}$ is the same as in the lower bound of
$1.5n-5$ proved by Hernando et al. [20], but their tree $T_{F}$ was not a path.
The construction is as follows (refer to Figure 5(a) and (c)). The points $v_{1}\ldots v_{n}$ (ordered by increasing $y$-coordinate) are placed in convex position on alternate sides of $v_{1}v_{n}$.
$T_{I}$ contains edges $v_{1}v_{2i+1}$ and $v_{n}v_{2i}$ for $i=1,\ldots,n/2-1$, and $v_{1}v_{n}$.
Note that $v_{1}$ and $v_{n}$ have degree $n/2$ each. $T_{F}$ is a path that connects vertices in order (so it includes edges $v_{i}v_{i+1}$, for $i=1,\ldots,n-1$).
Note that every non-hull edge of $T_{F}$ (in blue) crosses at least $n/2-1$ edges of $T_{I}$ (in red).
Indeed, edges of the form
$v_{2i+1}v_{2i}$ cross exactly $n/2$ edges:
$n/2-i-1$ edges incident to $v_{1}$, plus $i+1$ edges incident to $v_{n}$, minus $1$ because $v_{1}v_{n}$ is included in both counts.
Edges of the form $v_{2i}v_{2i+1}$ cross one less edge (specifically $v_{1}v_{2i+1}$).
Thus, any valid reconfiguration from $T_{I}$ to $T_{F}$ must flip $n/2-1$ edges of $T_{I}$ out of the way before the first of the $n-3$ non-hull edges of $T_{F}$ is added. After that, we need at least one flip for each of the remaining $n-4$ non-hull edges of $T_{F}$. Thus the total number of flips is at least $n/2-1+(n-4)=1.5n-5$.
We note that our two-phase reconfiguration algorithm uses $1.5n-3$ flips for this instance, but there is a flip sequence of length $1.5n-5$: first flip $v_{2}v_{n}$ to $v_{1}v_{2}$ to create a happy hull edge, then connect $v_{n}$ to all $v_{i}$ for odd $i\geq 7$ by performing the flips $v_{1}v_{i}$ to $v_{n}v_{i}$ in order of decreasing $i$. The number of flips thus far is $n/2-2$ (note that $v_{1}v_{3}$ and $v_{1}v_{5}$ stay in place). The resulting tree (shown in Figure 5b) has two happy hull edges. We show that this tree can be reconfigured into $T_{F}$ using perfect flips only (so the number of flips is $n-3$). Flip $v_{1}v_{n}$ to non-hull edge $v_{4}v_{5}$ and view the resulting tree as the union of an upward tree rooted at $v_{1}$ and a downward subtree rooted at $v_{n}$ (sharing $v_{4}v_{5}$). These two subtrees are separated by $v_{4}v_{5}$ and therefore can be independently reconfigured into their corresponding subtrees in $T_{F}$ using perfect flips, as given by Theorem 4. Thus the total number of flips is $(n/2-2)+(n-3)=1.5n-5$, proving this bound tight.
∎
Theorem 9.
For points in convex position, if $T_{I}$ is a non-crossing spanning tree and $T_{F}$ is a path, then $T_{I}$ can be reconfigured to $T_{F}$ in at most $1.5n-2-h$
flips, where $h=|T_{I}\cap T_{F}|$ is the number of happy edges.
Furthermore, there is a lower bound of $1.5n-5$ flips.
Proof.
When points are in convex position, two edges cross (a geometric property) if and only if their endpoints alternate in the cyclic ordering of points around the convex hull (a combinatorial property).
This insight allows us to show that the path $T_{F}$ is “equivalent to” a monotone path, which means that we can use the previous Theorem 7.
In particular, let the ordering of points in $T_{F}$ be $v_{1},\ldots,v_{n}$. We claim that the above algorithms can be applied using this ordering in place of the ordering of points by $y$-coordinate. Thus, a sink in $T_{I}$ is a point $v_{i}$ with no edge to a later vertex in the ordering, and etc.
One could justify this by examining the steps of the algorithms (we relied on geometry only to show that we can add a non-crossing edge “upward” from a sink, which becomes easy for points in convex position). As an alternative, we
make the argument formal by showing how to perturb the points so that $T_{F}$ becomes a monotone path while preserving the ordering of points around the convex hull—which justifies that the flips for the perturbed points are correct for the original points.
First adjust the points so that they lie on a circle with $v_{1}$ lowest at $y$-coordinate 1 and $v_{n}$ highest at $y$-coordinate $n$.
The convex hull separates into two chains
from $v_{1}$ to $v_{n}$.
Observe that $T_{F}$ visits the points of each chain in order from bottom to top (if $a$ appears before $b$ on one chain but $T_{F}$ visits $b$ before $a$, then the subpaths from $v_{1}$ to $b$ and from $a$ to $v_{n}$ would cross). Thus, we can place $v_{i}$ at $y$-coordinate $i$ while preserving the ordering of points around the circle.
We can now apply Theorem 7 to $T_{I}$ and $T_{F}$ on the perturbed points. This gives a sequence of at most $1.5n-h-2$ flips to reconfigure $T_{I}$ into $T_{F}$ and the flip sequence is still correct on the original points, thus proving the upper bound claimed by the theorem.
For the lower bound, note that the points in Figure 5 are in convex position and $T_{F}$ is a path. Thus Lemma 8 (which employs the example from Figure 5) settles the lower bound claim.
∎
3 Improving the Upper Bound for a Convex Point Set
In this section we
show that for $n$ points in convex position, reconfiguration between two non-crossing spanning trees can always be done with fewer than $2n$
flips.
Theorem 10.
There is an algorithm to reconfigure between
an initial non-crossing spanning tree $T_{I}$ and a final non-crossing spanning tree $T_{F}$
on $n$ points in convex position using
at most
$2d-\Omega(\log d)$ flips, where $d=|T_{I}\setminus T_{F}|$.
Before proving the theorem we note
that previous reconfiguration algorithms do not respect this bound.
Avis and Fukuda [10, Section 3.7] proved an upper bound of $2n$ minus a constant by reconfiguring each spanning tree $T_{i}$ to a star $S$ using $|S\setminus T_{i}|$ flips. When $T_{i}$ is a path, $|S\cap T_{i}|\leq 2$, so $|S\setminus T_{i}|\geq n-3$ and their method takes at least $2n-6$ flips. Similarly, the method of flipping both trees to a canonical path around the convex hull takes at least $2n-6$ flips when $T_{1}$ and $T_{2}$ are paths with only two edges on the convex hull.
Although paths behave badly for these canonicalization methods, they are actually easy cases
as we showed in
Section 2.
As in that section, we do not use a canonical tree to prove Theorem 10—instead, the flips are tailored to the specific initial and final trees.
Throughout this section, we assume points in convex position.
Consider
the
symmetric difference
$D=(T_{I}\setminus T_{F})\cup(T_{F}\setminus T_{I})$,
so $|D|=2d$.
It is easy to reconfigure $T_{I}$ to $T_{F}$ using $2d$ flips—we use $d$ flips to move the edges of $T_{I}\setminus T_{F}$ to the convex hull, giving an intermediate tree $T$, and, from the other end, use $d$ flips to move the edges of $T_{F}\setminus T_{I}$ to the same tree $T$. The plan is to save $\Omega(\log d)$ of these flips by using that many perfect flips
(recall that
a perfect flip exchanges an edge of $T_{I}\setminus T_{F}$ directly with an edge of $T_{F}\setminus T_{I}$).
In more detail, the idea is to find an edge $e\in D$ that is crossed by at most (roughly) $d/2$ edges of the other tree. We flip all but one of the crossing edges out of the way to the convex hull, and—if $e$ is chosen carefully—we show that we can perform one flip from $e$ to the last crossing edge, thus providing one perfect flip after at most $d/2$ flips.
Repeating this approach
$\log d$ times
gives our claimed bound.
We first show how to find an edge $e$ with not too many crossings. To do this, we define “minimal” edges.
An edge joining points $u$ and $v$
determines two subsets of points, those clockwise from $u$ to $v$ and those clockwise from $v$ to $u$ (both sets include $u$ and $v$). We call these the sides of the edge.
An edge is contained in a side if both endpoints are in the side.
We call a side minimal if it contains no edge of the symmetric difference $D$, and call an edge $e\in D$ minimal if at least one of its sides is minimal.
Note that if $D$ is non-empty then it contains at least one minimal edge (possibly a convex hull edge).
We need the following property of minimal edges.
Claim 11.
Let $e=uv$ be a minimal edge of $D$. Let $Q$ be a minimal side of $e$, and let $\bar{Q}$ be the other side. Suppose $e\in T_{F}$. Then
$T_{I}\cap Q$ consists of exactly two connected components, one containing $u$ and one containing $v$.
Proof.
The set $T_{F}\cap Q$ is a non-crossing tree
consisting of edge $e$ and two subtrees $T_{u}$ containing $u$ and $T_{v}$ containing $v$.
Since $e$ is minimal, there are no edges of $D$ in $Q$ except for $e$ itself.
This means that
$T_{I}\cap Q$
consists of $T_{u}$ and $T_{v}$,
and since $e\notin T_{I}$, these two components of $T_{I}\cap Q$ are disconnected in $Q$.
See Figure 6(a).
∎
Our algorithm will operate on a minimal edge $e$. To guarantee the savings in flips, we need a minimal edge with not too many crossings.
Lemma 12.
If $D$ is non-empty, then there is a minimal edge with at most
$\lfloor(d+3)/2\rfloor$ crossings.
Proof.
Clearly the lemma holds if some minimal edge is not crossed at all (e.g., a convex hull edge in $D$), so we assume that all minimal edges have a crossing.
Let $a$ be a minimal edge of $D$.
Suppose $a\in T_{F}$.
Let $A$ be a minimal side of $a$, and let $\bar{A}$ be the other side.
Our plan is to find a second minimal edge $b\in T_{F}$ such that $b$ is inside $\bar{A}$ and $b$ has
a minimal side $B$ that is contained in $\bar{A}$. We will then argue that $a$ or $b$ satisfies the lemma.
See Figure 6(b).
If $\bar{A}$ is minimal, then set $b:=a$ and $B=\bar{A}$.
Otherwise,
let $b$ be an edge of $T_{F}\setminus T_{I}$ in $\bar{A}$ whose $B$ side (the side in $\bar{A}$)
contains no other edge of $T_{F}\setminus T_{I}$.
Note that $b$ exists, and that all the edges of $T_{F}$ in $B$ (except $b$) lie in $T_{I}$.
If $b$ is not minimal, then $B$ contains
a minimal edge $c$ (which must then be in $T_{I}\setminus T_{F}$), and $c$ is not crossed by any edge of $T_{F}$,
because
such an edge would either have to cross $b$, which is impossible since $b\in T_{F}$, or lie in $B$, which is impossible since
all the edges of $T_{F}$ in $B\setminus\{b\}$
are in $T_{F}\cap T_{I}$.
But we assumed that all minimal edges are crossed, so $c$ cannot exist, and so $b$ must be minimal.
Let $k_{a}$ be the number of edges
of $T_{I}$ crossing $a$, let $k_{b}$ be the number of edges of $T_{I}$ crossing $b$, and let $k_{ab}$ be the number of edges of $T_{I}$ crossing both $a$ and $b$.
Observe that $k_{a}+k_{b}\leq d+k_{ab}$.
We claim that $k_{ab}\leq 3$. Then $k_{a}+k_{b}\leq d+3$ so $\min\{k_{a},k_{b}\}\leq(d+3)/2$, which
will complete
the proof since the number of crossings is an integer.
By Claim 11, $T_{I}\cap A$ has two connected components and $T_{I}\cap B$ has two connected components.
Now 4 connected components in a tree can have at most three edges joining them, which
implies that $k_{ab}\leq 3$.
(Note that this argument is correct even for $a=b$, though we get a sharper bound since $k_{a}=k_{b}=k_{ab}\leq 3$.)
∎
Algorithm.
Choose a minimal edge $e$ with $k$ crossings, where $0\leq k\leq(d+3)/2$ (as guaranteed by Lemma 12).
Suppose $e\in T_{F}$, so the crossing edges belong to $T_{I}$. The case where $e\in T_{I}$ is symmetric. In either case the plan is to perform some flips on $T_{I}$ and some on $T_{F}$ to reduce the difference $d$
by $k$ (or by 1, if $k=0$) and apply the algorithm recursively to the resulting instance. Note that the algorithm constructs a flip sequence by adding flips at both ends of the sequence.
If $k=0$ then add $e$ to $T_{I}$. This creates a cycle, and the cycle must have an edge $f$ in $T_{I}\setminus T_{F}$. Remove $f$. This produces a new tree $T_{I}$. We have performed one perfect flip and reduced $d$ by 1.
Now recurse.
Next suppose $k\geq 1$.
Let $e=uv$.
Let $Q$ be the minimal side of $e$ and let $\bar{Q}$ be the other side (both sets include $u$ and $v$).
Let $f_{1},\ldots,f_{k}$ be the edges that cross $e$. We will flip all but the last crossing edge to the convex hull.
For $i=1,\ldots k-1$ we flip $f_{i}$ as follows.
1.
Perform a flip in $T_{I}$ by removing $f_{i}$ and adding a convex hull edge $g$ that lies in $\bar{Q}$. (The existence of $g$ is proved below.)
2.
If $g\in T_{F}$ then this was a perfect flip and we have performed one perfect flip and reduced $d$ by 1.
3.
Otherwise (if $g\notin T_{F}$), perform a flip in $T_{F}$ by adding $g$ and removing an edge $h\in T_{F}\setminus T_{I}$,
that lies in $\bar{Q}$ and is not equal to $e$.
(The existence of $h$ is proved below.)
At this point, only $f_{k}$ crosses $e$.
Perform one flip in $T_{I}$ to remove $f_{k}$ and add $e$.
(Correctness proved below.)
Now, apply the algorithm recursively to the resulting $T_{I},T_{F}$.
This completes the description of the algorithm.
Correctness.
We must prove that $g$ and $h$ exist and that the final flip is valid.
First note that $e$ remains a minimal edge
after each flip performed inside the loop
because we never add or remove edges inside $Q$.
We need one more invariant of the loop.
Claim 13.
Throughout the loop $u$ and $v$ are disconnected in $T_{I}\cap{\bar{Q}}$.
Proof.
Suppose there is a path $\pi$ from $u$ to $v$ in $T_{I}\cap{\bar{Q}}$.
Now consider the edge $f_{k}$ which crosses $e$, say from $x\in Q$ to $y\in{\bar{Q}}$. Since $f_{k}$ cannot cross $\pi$, we have $y\in\pi$.
By Claim 11, $T_{I}\cap Q$ consists of two components, one containing $u$ and one containing $v$. Suppose, without loss of generality, that $x$ lies in the component containing $u$. Then there is a path from $x$ to $u$ in $T_{I}\cap Q$ and a path from $u$ to $y$ in $T_{I}\cap{\bar{Q}}$, and these paths together with $f_{k}$ make a cycle in $T_{I}$, a contradiction.
∎
First we prove that $g$ exists in Step 1. Removing $f_{i}$ from $T_{I}$ disconnects $T_{I}$ into two pieces. There are two convex hull edges that connect the two pieces. By Claim 11, $T_{I}\cap Q$ consists of two connected components, one containing $u$ and one containing $v$. Thus at most one of the convex hull edges lies in $Q$, so at least one lies in $\bar{Q}$.
Next we prove that $h$ exists in Step 3.
Adding $g$ to $T_{F}$ creates a cycle $\gamma$ in $T_{F}$ and this cycle must lie in $\bar{Q}$ (because $e\in T_{F}$) and must contain at least one edge of $T_{F}\setminus T_{I}$ (because $T_{I}$ does not contain a cycle). If $e$ were the only edge of $T_{F}\setminus T_{I}$ in $\gamma$, then $u$ and $v$ would be joined by a path in $T_{I}\cap{\bar{Q}}$, contradicting Claim 13. Thus $h$ exists.
Finally, we prove that the last flip in $T_{I}$ (to remove $f_{k}$ and add $e$) is valid. Removing $f_{k}$ leaves $u$ and $v$ disconnected in $Q$ by Claim 11 and disconnected in $\bar{Q}$ by Claim 13. Adding $e$ reconnects them, and yields a non-crossing spanning tree.
Analysis.
We now prove that the algorithm uses at most the claimed number of flips.
Observation 14.
In each recursive call:
if $k=0$, then the algorithm performs one perfect flip and reduces $d$ by 1;
and if $k>0$, then the algorithm performs at most $2k-1$ flips (one for $f_{k}$ and at most 2 for each other $f_{i}$) and reduces $d$ by $k$ (in each loop iteration, $g$ joins the happy set $T_{I}\cap T_{F}$ and in the final step $e$ joins the happy set).
Lemma 15.
The number of flips performed by the algorithm is at most $2d-\lfloor\log(d{+}3)\rfloor+1$.
Proof.
We prove this by induction on $d$. In the base case $d=0$ we perform $0=2d-\lfloor\log(d{+}3)\rfloor+1$ flips.
Now assume $d\geq 1$,
and consider what happens in the first recursive call of the algorithm, see Observation 14.
If the algorithm chooses an edge with $k=0$ crossings, then the algorithm performs one perfect flip.
The resulting instance has a difference set of size $d^{\prime}=d-1$ and induction applies, so in
the total number of flips we perform is at most
$$\displaystyle 1+2d^{\prime}-\lfloor\log(d^{\prime}{+}3)\rfloor+1=2d-\lfloor\log(d{+}2)\rfloor\leq 2d-\lfloor\log(d{+}3)\rfloor+1,$$
which proves the result in this case.
Now suppose that the algorithm chooses an edge with $k\geq 1$ crossings, where $k\leq\lfloor(d{+}3)/2\rfloor$.
The algorithm performs at most $2k-1$ flips and the resulting instance has a difference set
of size $d^{\prime}=d-k$ and therefore
$d^{\prime}{+}3\geq d{+}3-\lfloor(d{+}3)/2\rfloor=\lceil(d{+}3)/2\rceil\geq(d{+}3)/2$.
By induction, the total number
of flips that we perform is hence at most
$$\displaystyle(2k-1)+(2d^{\prime}-\lfloor\log(d^{\prime}{+}3)\rfloor+1)$$
$$\displaystyle\leq(2k-1)+(2(d-k)-\lfloor\log((d{+}3)/2)\rfloor+1)$$
$$\displaystyle\leq~{}2d-\lfloor\log((d{+}3)/2)\rfloor$$
$$\displaystyle\leq 2d-\lfloor\log(d{+}3)\rfloor+1$$
as desired.
∎
This completes the proof of Theorem 10.
4 The Happy Edge Conjecture
In this section we make some conjectures and prove some preliminary results in attempts to characterize which edges need to be flipped in minimum flip sequences for non-crossing spanning trees.
Recall that
an edge $e$ is happy if
$e$ lies in $T_{I}\cap T_{F}$.
We make the following
conjecture for points in convex position. In fact, we do not have a counterexample even for general point sets, though our guess is that the conjecture fails in the general case.
Conjecture 16.
[Happy Edge Conjecture for Convex Point Sets]
For any point set $P$ in convex position and any two non-crossing spanning trees $T_{I}$ and $T_{F}$ of $P$, there is a minimum flip sequence from $T_{I}$ to $T_{F}$ such that
no happy edge is flipped during the sequence.
In this section we first prove this conjecture for the case of happy edges on the convex hull. Then in Section 4.1 we make some stronger conjectures about which extra edges (outside $T_{I}$ and $T_{F}$) might be needed in minimum flip sequences. In Section 4.2 we show that even if no extra edges are needed, it may be tricky to find a minimum flip sequence—or, at least, a greedy approach fails.
Finally, in Section 4.3 we prove that the Happy Edge Conjecture fails if we restrict the flips to “slides” where one endpoint of the flipped edge is fixed and the other endpoint moves along an adjacent tree edge.
If the Happy Edge Conjecture is false then a minimum flip sequence might need to remove an edge and later add it back.
We are able to prove something about such “remove-add” subsequences, even for general point sets:
Proposition 17.
Consider any point set $P$ and any two non-crossing spanning trees $T_{I}$ and $T_{F}$ on $P$ and any minimum flip sequence from $T_{I}$ to $T_{F}$. If some edge $e$ is removed and later added back, then
some flip during that subsequence must
add an edge $f$ that crosses $e$.
Before proving this Proposition, we note the implication that the Happy Edge Conjecture is true for convex hull edges:
Corollary 18.
Conjecture 16 is true for happy edges on the convex hull.
Furthermore, every minimum flip sequence keeps the happy convex hull edges throughout the sequence.
Proof.
Let $e$ be a happy convex hull edge.
Suppose for a contradiction that there is a minimum flip sequence in which $e$ is removed. Note that $e$ must be added back, since it is in $T_{F}$. By Proposition 17, the flip sequence must use an edge $f$ that crosses $e$. But that is impossible because $e$ is a convex hull edge so nothing crosses it.
∎
Proof of Proposition 17.
Consider a flip sequence from $T_{I}$ to $T_{F}$ and suppose that an edge $e$ is removed and later added back, and that no edge crossing $e$ is added during that subsequence. We will make a shorter flip sequence. The argument is similar to the “normalization” technique used by Sleator et al. [35] to prove the happy edge result for flips in triangulations of a convex point set.
Let $T_{0},\ldots,T_{k}$ be the trees in the subsequence, where $T_{0}$ and $T_{k}$ contain $e$, but none of the intervening trees do. Suppose that none of the trees $T_{i}$ contains an edge that crosses $e$.
We will construct a shorter flip sequence from $T_{0}$ to $T_{k}$.
For each $i$, $0\leq i\leq k$ consider adding $e$ to $T_{i}$. For $i\neq 0,k$, this creates a cycle $\gamma_{i}$. Let $f_{i}$ be the first edge of $\gamma_{i}$ that is removed during the flip sequence from $T_{i}$ to $T_{k}$. Note that $f_{i}$ exists since $T_{k}$ contains $e$, so it cannot contain all of $\gamma_{i}$.
Define $N_{i}=T_{i}\cup\{e\}\setminus\{f_{i}\}$ for $1\leq i\leq k-1$, and define $N_{0}:=T_{0}$.
Observe that $N_{i}$ is a spanning tree, and is non-crossing because no edge of $T_{i}$ crosses $e$ by hypothesis.
Furthermore,
$N_{k-1}=T_{k}$
because the flip from $T_{k-1}$ to $T_{k}$ is exactly the same as the flip from $T_{k-1}$ to $N_{k-1}$.
We claim that $N_{0},\ldots,N_{k-1}$ is a flip sequence. This will complete the proof, since it is a shorter flip sequence from $T_{0}$ to $T_{k}$.
Consider $N_{i}$ and $N_{i+1}$. Suppose that the flip from $T_{i}$ to $T_{i+1}$ adds $g$ and removes $h$.
$$\begin{tikzcd}$$
Recall that $\gamma_{i}$ is the cycle containing $e$ in $T_{i}\cup e$.
If $h$ belongs to $\gamma_{i}$ then $f_{i}=h$, and then to get from $N_{i}$ to $N_{i+1}$ we add $g$ and remove $f_{i+1}$.
Next, suppose that $h$ does not belong to $\gamma_{i}$. Then the cycle $\gamma_{i}$ still exists in $T_{i+1}$. Now, $\gamma_{i+1}$ is the unique cycle in $T_{i+1}\cup e$. Thus $\gamma_{i+1}=\gamma_{i}$. Furthermore, $f_{i+1}$ is by definition the first edge removed from $\gamma_{i+1}$ in the flip sequence from $T_{i+1}$ to $T_{k}$. Thus $f_{i+1}=f_{i}$. Therefore, to get from $N_{i}$ to $N_{i+1}$ we add $g$ and remove $h$.
This shows that a single flip changes $N_{i}$ to $N_{i+1}$, which completes the proof.
∎
Note that the proof of Proposition 17 produces a strictly shorter flip sequence. But to prove the Happy Edge Conjecture (Conjecture 16) it would suffice to produce a flip sequence of the same length.
One possible approach is to consider how
remove-add pairs and add-remove pairs
interleave in a flip sequence.
Proposition 17 shows that
a remove-add pair for edge $e$ must contain an add-remove pair for $f$ inside it.
We may need to understand
how the order of flips can be rearranged in a flip sequence.
Such flip order rearrangements
are at the heart of results on triangulation flips, both for convex point sets [35, 32] and for general point sets [24].
4.1 Extra edges used in flip sequences
Any flip sequence from $T_{I}$ to $T_{F}$ must involve flips that remove edges of $T_{I}\setminus T_{F}$ and flips that add edges of $T_{F}\setminus T_{I}$.
Recall that in a perfect flip sequence, these are the only moves and they pair up perfectly, so the number of flips is $|T_{I}\setminus T_{F}|$.
Theorem 4 gives one situation where a perfect flip sequence is possible, but typically (e.g., in the example of Figure 5)
we must add edges not in $T_{F}$, and later remove them.
More formally, an edge outside $T_{I}\cup T_{F}$ that is used in a flip sequence is called a parking edge,
with the idea that we “park” edges there temporarily.
We make two further successively stronger conjectures. They may not hold, but disproving them would give more insight.
Conjecture 19.
For any point set $P$ in convex position and any two non-crossing spanning trees $T_{I}$ and $T_{F}$ of $P$ there is a minimum flip sequence from $T_{I}$ to $T_{F}$ that never
uses a parking edge that crosses an edge of $T_{F}$.
Conjecture 20.
For a point set $P$ in convex position and any two non-crossing spanning trees $T_{I}$ and $T_{F}$ on $P$ there is a minimum flip sequence from $T_{I}$ to $T_{F}$ that only uses
parking edges from the convex hull.
Our experiments verify Conjecture 20 for $n\leq 10$ points, (see Observation 22).
We note that Conjecture 20 cannot hold for general point sets (there just aren’t enough convex hull edges). However, we do not know if Conjecture 19 fails for general point sets.
Claim 21.
Conjecture 20 $\Longrightarrow$ Conjecture 19 $\Longrightarrow$ Conjecture 16.
Proof.
The first implication is clear.
For the second implication we use Proposition 17. Consider the minimum flip sequence promised by Conjecture 19. If there is a happy edge $e\in T_{I}\cap T_{F}$ that is removed during this flip sequence, then by Proposition 17, the flip sequence must add an edge $f$ that crosses $e$. But then $f$ is a parking edge that crosses an edge of $T_{F}$, a contradiction.
∎
4.2 Finding a perfect flip sequence—greedy fails
It is an open question whether there is a polynomial time algorithm to find [the length of] a minimum flip sequence between two given non-crossing spanning trees $T_{I}$ and $T_{F}$.
A more limited goal is testing whether there is a flip sequence of length $|T_{I}\setminus T_{F}|$—i.e., whether there is a perfect flip sequence. This is also open.
In Figure 7 we give an example to show that
a greedy approach to finding a perfect flip sequence may fail.
In this example there is a perfect flip sequence but a poor choice of perfect flips leads to a dead-end configuration where no further perfect flips are possible. Note that choosing perfect flips involves pairing edges of $T_{I}\setminus T_{F}$ with edges of $T_{F}\setminus T_{I}$ as well as ordering the pairs.
4.3 The Happy Edge Conjecture fails for edge slides
Researchers have examined various restricted types of flips for non-crossing spanning trees, see [28].
An edge slide is the most
restricted
flip operation possible: it keeps one endpoint of the flipped edge fixed and moves the
other one along an adjacent tree edge without intersecting any of the other edges or vertices of the tree. In other words, the edge that is removed, the edge that is inserted,
and the edge along which the slide takes place form an empty triangle. Aichholzer et al. [7] proved that for any set $P$ of $n$ points in the plane it is possible to transform between any two non-crossing spanning trees of $P$
using $O(n^{2})$ edge slides. The authors also give an example to show that $\Omega(n^{2})$ slides might be required even if the two spanning trees differ in only two edges. This example already implies that for point sets in general position the Happy Edge Conjecture fails for edge slides. We will show that this is also the case for points in convex position.
Figure 8 shows an example of two plane spanning trees $T_{I}$ and $T_{F}$ on $8$ points in convex position which can be transformed into each other with 8 slides, shown at the bottom of the figure. To obtain this short sequence we temporarily use an edge which is common to both trees to connect the two vertices $a$ and $b$. Thus this sequence contains a non-happy slide operation, that is, an edge that is common to both, $T_{I}$ and $T_{F}$ is moved. When flipping from tree $T_{I}$ to tree $T_{F}$ by using only happy slide operations there are some useful observations. First, there can not be an edge directly connecting $a$ and $b$, as this would cause a cycle. This implies that any edge which connects a vertex $v_{i}$, $1\leq i\leq 4$, with $b$ needs at least two slides to connect $a$ to some (possible different) vertex $v_{j}$. Moreover, the first of these edges that gets connected to $a$ needs at least three slides, as at the beginning this is the shortest path connecting $b$ to $a$. Thus in total we need at least $3+2+2+2=9$ happy slide operations. Figure 8(top) shows such a sequence. It is not hard to see that this example can be generalized to larger $n$ and implies that the Happy Edge Conjecture fails for points in convex position.
5 Exhaustive search over small point sets in convex position
For small point sets in convex position we investigated the minimum flip distance between non-crossing spanning trees by exhaustive computer search. Table 1 summarizes the obtained results. For $n=3,\ldots,12$ we give the number of non-crossing spanning trees (which is sequence A001764 in the On-Line Encyclopedia of Integer Sequences, https://oeis.org) and the number of reconfiguration
steps
between them. Moreover, we computed the maximum reconfiguration distance between two trees
(the diameter of the reconfiguration graph)
as well as the radius of the reconfiguration graph.
We provide the same information for the special case when the trees are non-crossing spanning paths. Note that in this case the intermediate graphs can still be non-crossing spanning trees. For the case where all intermediate graphs are also non-crossing spanning paths the diameter of the reconfiguration graph for points in convex position is known to be $2n-6$ for $n\geq 5$ [8, 17].
Results.
Our computations show that
for small sets in convex position the radius of each reconfiguration graph is strictly larger than half
the diameter.
More precisely, for $6\leq n\leq 12$ the diameter is $\lfloor 1.5n-4\rfloor$ and the radius is $n-2$ which would give an upper bound for the diameter of only $2n-4$.
But this might be an artefact of small numbers: compare for example the result of Sleator, Tarjan, and Thurston which give the upper bound of $2n-10$ for the rotation distance of binary trees which is tight only for $n\geq 13$ [32, 35]. That the radius seems not to be suitable for obtaining a tight bound for the diameter also supports our way of bounding the diameter of the reconfiguration graph by not using a central canonical tree.
In addition to the results shown in the table, we checked, for $n\leq 10$, which edges are exchanged, in order to test the Happy Edge Conjecture (Conjecture 16) and whether only parking edges on the convex hull are used (Conjecture 20).
Observation 22.
For $n\leq 10$ points in convex position (1) the Happy Edge Conjecture is true, and (2) there are always minimum flip sequences that only use parking edges on the convex hull.
Methods.
These computations are rather time consuming, as in principle for any pair of non-crossing spanning trees (paths) the flip distance has to be computed. For an unweighted and undirected graph $G$ with $n^{\prime}$ nodes (non-crossing spanning trees in our case) and $m^{\prime}$ edges (edge exchanges in our case) the standard algorithm to compute the diameter of $G$ is to apply breath first search (BFS) for each node of $G$. The time requirement for this simple solution is $O(n^{\prime}m^{\prime})$. There exist several algorithms which achieve a better running time for graphs of real world applications, see e.g., [18], but in the worst case they still need $O(n^{\prime}m^{\prime})$ time. The basic idea behind these approaches is to compute the eccentricity $e(v)$ of a node $v\in G$ (which is the radius as seen from this node $v$), and compare this with the largest distance $d$ between two nodes found so far. If $e(v)=d/2$ we know that the diameter of the graph is $d$ and the algorithm terminates. The difference between the various algorithms is how the nodes for computing the eccentricity and the lower bound for the diameter are chosen and the performance of the approaches are usually tested by applying them to a set of examples.
However, it turned out that by the structure of our reconfiguration graphs these approaches do not perform better than the simple
textbook
solution.
Because the radius of the reconfiguration graph is strictly larger than half the diameter in our test cases, no algorithmic shortcut is possible.
To still be able to compute the diameter of the rather large graphs (for $n=12$ the reconfiguration graph has $8~{}414~{}650$ nodes and $193~{}666~{}176$ edges) we make use of the inherent symmetries of our graphs. For every tree $T$ we can cyclically shift the labels of the vertices (by $1$ to $n-1$ steps) and/or mirror the tree to obtain another non-crossing spanning tree $T^{\prime}$ of the convex point set. All trees that can be obtained this way can be grouped together. While every tree is needed in the reconfiguration graph to correctly compute shortest reconfiguration distances, by symmetry a call of BFS for any tree from the same group will result in the same eccentricity. It is thus sufficient to call BFS only for one tree of each group. For $n$ points this reduces the number of calls by almost a factor of $2n$, as the size of the group can be up to $2n$ (some trees are self-symmetric in different ways, thus some groups have a cardinality less than $2n$).
For our experiments on which edges are exchanged (for Observation 22),
the computations get even more involved. The reason is that these properties of edges are defined by the initial and final tree. So it can happen that a short sub-path is valid only for some, but not all, pairs of trees where we would like to use it. Moreover, for similar reasons we can not make full use of the above described symmetry.
This is the reason why we have been able to test our conjectures only for sets with up to $n=10$ points.
6 Conclusions and Open Questions
We conclude with some open questions:
1.
We gave two algorithms to find flip sequences for non-crossing spanning trees and we bounded the length of the flip sequence. The algorithms run in polynomial time, but it would be good to optimize the run-times.
2.
A main open question is to close the gap between $1.5n$ and $2n$ for the leading term
of the diameter of the reconfiguration graph of non-crossing spanning trees.
3.
A less-explored problem is to find the radius of the reconfiguration graph (in the worst case, as a function of $n$, the number of points).
Is there a lower bound of $n-c$ on the radius of the reconfiguration graph for some small constant $c$?
4.
Prove or disprove the Happy Edge Conjecture.
5.
Is the distance problem (to find the minimum flip distance between two non-crossing spanning trees) NP-complete for general point sets? For convex point sets?
A first step towards an NP-hardness reduction would be to find instances where the Happy Edge Conjecture fails.
6.
An easier question is to test whether there is a perfect flip sequence between two non-crossing spanning trees. Can that be done in polynomial time, at least for points in convex position?
7.
Suppose the Happy Edge Conjecture turns out to be false. Is the following problem NP-hard?
Given two trees, is there a minimum flip sequence between them that does not flip happy edges?
8.
Suppose we have a minimum flip sequence that does not flip happy edges and does not use parking edges (i.e., the flips only involve edges of the difference set $D=(T_{I}\setminus T_{F})\cup(T_{F}\setminus T_{I})$). Is it a perfect flip sequence?
9.
All the questions above can be asked for the other versions of flips between non-crossing spanning trees (as discussed in Section 1.1 and surveyed in [28]).
10.
For the convex case, what if we only care about the cyclic order of points around the convex hull, i.e., we may freely relabel the points so long as we preserve the cyclic order of the labels. This “cyclic flip distance” may be less than the standard flip distance.
For example, two stars rooted at different vertices have cyclic flip distance $0$ but standard flip distance $n-2$.
Acknowledgements
This research was begun in the Fall 2019 Waterloo problem solving session run by Anna Lubiw and Therese Biedl. For preliminary progress on the Happy Edge Conjecture, we thank the participants, Pavle Bulatovic, Bhargavi Dameracharla, Owen Merkel, Anurag Murty Naredla, and Graeme Stroud.
Work on the topic continued in the 2020 and 2021 Barbados workshops run by Erik Demaine and we thank the other participants for many valuable discussions.
References
[1]
Oswin Aichholzer, Andrei Asinowski, and Tillmann Miltzow.
Disjoint compatibility graph of non-crossing matchings of points in
convex position.
The Electronic Journal of Combinatorics, 22:1–65, 2015.
doi:10.37236/4403.
[2]
Oswin Aichholzer, Franz Aurenhammer, and Ferran Hurtado.
Sequences of spanning trees and a fixed tree theorem.
Computational Geometry, 21(1-2):3–20, 2002.
doi:10.1016/S0925-7721(01)00042-6.
[3]
Oswin Aichholzer, Sergey Bereg, Adrian Dumitrescu, Alfredo García, Clemens
Huemer, Ferran Hurtado, Mikio Kano, Alberto Márquez, David Rappaport,
Shakhar Smorodinsky, Diane Souvaine, Jorge Urrutia, and David R. Wood.
Compatible geometric matchings.
Computational Geometry, 42(6):617–626, 2009.
doi:10.1016/j.comgeo.2008.12.005.
[4]
Oswin Aichholzer, Erik D Demaine, Matias Korman, Jayson Lynch, Anna Lubiw,
Zuzana Masárová, Mikhail Rudoy, Virginia Vassilevska Williams, and
Nicole Wein.
Hardness of token swapping on trees.
arXiv preprint, 2021.
doi:10.48550/arXiv.2103.06707.
[5]
Oswin Aichholzer, Kristin Knorr, Maarten Löffler, Zuzana
Masárová, Wolfgang Mulzer, Johannes Obenaus, Rosna Paul, and
Birgit Vogtenhuber.
Flipping Plane Spanning Paths.
In Proc. $38^{th}$ European Workshop on Computational Geometry
(EuroCG 2022), pages 66:1–66:7, Perugia, Italy, 2022.
doi:10.48550/arXiv.2202.10831.
[6]
Oswin Aichholzer, Wolfgang Mulzer, and Alexander Pilz.
Flip Distance Between Triangulations of a Simple Polygon is
NP-Complete.
Discrete & Computational Geometry, 54(2):368–389, 2015.
doi:10.1007/s00454-015-9709-7.
[7]
Oswin Aichholzer and Klaus Reinhardt.
A quadratic distance bound on sliding between crossing-free spanning
trees.
Computational Geometry, 37(3):155–161, 2007.
doi:10.1016/j.comgeo.2004.12.010.
[8]
Selim G. Akl, Md. Kamrul Islam, and Henk Meijer.
On planar path transformation.
Information Processing Letters, 104(2):59–64, 2007.
doi:10.1016/j.ipl.2007.05.009.
[9]
Nikos Apostolakis.
Non-crossing trees, quadrangular dissections, ternary trees, and
duality-preserving bijections.
Annals of Combinatorics, 25(2):345–392, 2021.
doi:10.1007/s00026-021-00531-w.
[10]
David Avis and Komei Fukuda.
Reverse search for enumeration.
Discrete Applied Mathematics, 65(1-3):21–46, 1996.
doi:10.1016/0166-218X(95)00026-N.
[11]
David Avis and Duc A Hoang.
On reconfiguration graph of independent sets under token sliding.
arXiv preprint, 2022.
doi:10.48550/arXiv.2203.16861.
[12]
Ahmad Biniaz, Kshitij Jain, Anna Lubiw, Zuzana Masárová, Tillmann
Miltzow, Debajyoti Mondal, Anurag Murty Naredla, Josef Tkadlec, and Alexi
Turcotte.
Token swapping on trees.
arXiv preprint, 2019.
doi:10.48550/arXiv.1903.06981.
[13]
Marthe Bonamy, Nicolas Bousquet, Marc Heinrich, Takehiro Ito, Yusuke Kobayashi,
Arnaud Mary, Moritz Mühlenthaler, and Kunihiro Wasa.
The perfect matching reconfiguration problem.
In 44th International Symposium on Mathematical Foundations of
Computer Science (MFCS 2019), volume 138 of Leibniz International
Proceedings in Informatics (LIPIcs), pages 80:1–80:14, Dagstuhl, Germany,
2019. Schloss Dagstuhl–Leibniz-Zentrum fuer Informatik.
doi:10.4230/LIPIcs.MFCS.2019.80.
[14]
Paul Bonsma.
The complexity of rerouting shortest paths.
Theoretical Computer Science, 510:1–12, 2013.
doi:10.1016/j.tcs.2013.09.012.
[15]
Prosenjit Bose and Ferran Hurtado.
Flips in planar graphs.
Computational Geometry, 42(1):60–80, 2009.
doi:10.1016/j.comgeo.2008.04.001.
[16]
Kevin Buchin, Andreas Razen, Takeaki Uno, and Uli Wagner.
Transforming spanning trees: A lower bound.
Computational Geometry, 42(8):724–730, 2009.
doi:10.1016/j.comgeo.2008.03.005.
[17]
Jou-Ming Chang and Ro-Yu Wu.
On the diameter of geometric path graphs of points in convex
position.
Information Processing Letters, 109(8):409–413, 2009.
doi:10.1016/j.ipl.2008.12.017.
[18]
Pilu Crescenzi, Roberto Grossi, Michel Habib, Leonardo Lanzi, and Andrea
Marino.
On computing the diameter of real-world undirected graphs.
Theoretical Computer Science, 514:84–95, 2013.
Graph Algorithms and Applications: in Honor of Professor Giorgio
Ausiello.
doi:10.1016/j.tcs.2012.09.018.
[19]
C. Hernando, F. Hurtado, and M. Noy.
Graphs of non-crossing perfect matchings.
Graphs and Combinatorics, 18:517–532, 2002.
doi:10.1007/s003730200038.
[20]
M.C. Hernando, F. Hurtado, A. Márquez, M. Mora, and M. Noy.
Geometric tree graphs of points in convex position.
Discrete Applied Mathematics, 93(1):51–66, 1999.
doi:10.1016/S0166-218X(99)00006-2.
[21]
Michael E Houle, Ferran Hurtado, Marc Noy, and Eduardo Rivera-Campo.
Graphs of triangulations and perfect matchings.
Graphs and Combinatorics, 21(3):325–331, 2005.
doi:10.1007/s00373-005-0615-2.
[22]
Mashhood Ishaque, Diane L. Souvaine, and Csaba D. Tóth.
Disjoint compatible geometric matchings.
Discrete & Computational Geometry, 49:89–131, 2013.
doi:10.1007/s00454-012-9466-9.
[23]
Takehiro Ito, Marcin Kamiński, Hirotaka Ono, Akira Suzuki, Ryuhei Uehara,
and Katsuhisa Yamanaka.
Parameterized complexity of independent set reconfiguration problems.
Discrete Applied Mathematics, 283:336–345, 2020.
doi:10.1016/j.dam.2020.01.022.
[24]
Iyad A. Kanj, Eric Sedgwick, and Ge Xia.
Computing the flip distance between triangulations.
Discrete & Computational Geometry, 58(2):313–344, 2017.
doi:10.1007/s00454-017-9867-x.
[25]
Anna Lubiw, Zuzana Masárová, and Uli Wagner.
A proof of the orbit conjecture for flipping edge-labelled
triangulations.
Discrete & Computational Geometry, 61(4):880–898, 2019.
doi:10.1007/s00454-018-0035-8.
[26]
Anna Lubiw and Vinayak Pathak.
Flip distance between two triangulations of a point set is
NP-complete.
Computational Geometry, 49:17–23, 2015.
doi:10.1016/j.comgeo.2014.11.001.
[27]
Torsten Mütze.
Combinatorial Gray codes-an updated survey.
arXiv preprint, 2022.
doi:10.48550/arXiv.2202.01280.
[28]
Torrie L. Nichols, Alexander Pilz, Csaba D. Tóth, and Ahad N. Zehmakan.
Transition operations over plane trees.
Discrete Mathematics, 343(8):111929, 2020.
doi:10.1016/j.disc.2020.111929.
[29]
Naomi Nishimura.
Introduction to reconfiguration.
Algorithms, 11(4):52, 2018.
doi:10.3390/a11040052.
[30]
Marc Noy.
Enumeration of noncrossing trees on a circle.
Discrete Mathematics, 180(1-3):301–313, 1998.
doi:10.1016/S0012-365X(97)00121-0.
[31]
Alexander Pilz.
Flip distance between triangulations of a planar point set is
APX-hard.
Computational Geometry, 47(5):589–604, 2014.
doi:10.1016/j.comgeo.2014.01.001.
[32]
Lionel Pournin.
The diameter of associahedra.
Advances in Mathematics, 259:13–42, 2014.
doi:10.1016/j.aim.2014.02.035.
[33]
Dana Randall.
Rapidly mixing markov chains with applications in computer science
and physics.
Computing in Science & Engineering, 8(2):30–41, 2006.
doi:10.1109/MCSE.2006.30.
[34]
Andreas Razen.
A lower bound for the transformation of compatible perfect matchings.
In Proc. 24th European Workshop on Computational Geometry
(EuroCG 2008), pages 115–118, 2008.
URL:
https://hal.inria.fr/inria-00595116/PDF/EuroCG08Abstracts.pdf#page=126.
[35]
Daniel D. Sleator, Robert E. Tarjan, and William P. Thurston.
Rotation distance, triangulations, and hyperbolic geometry.
Journal of the American Mathematical Society, 1(3):647–681,
1988.
doi:10.1090/S0894-0347-1988-0928904-4.
[36]
Jan van den Heuvel.
The complexity of change.
Surveys in Combinatorics, 409(2013):127–160, 2013.
doi:10.48550/arXiv.1312.2816. |
Embedding large graphs into a random graph
Asaf Ferber
Department of Mathematics, Yale University, and Department of Mathematics, MIT. Emails:
[email protected], and [email protected].
Kyle Luh
Department of Mathematics, Yale University. Email: [email protected].
Oanh Nguyen
Department of Mathematics, Yale University. Email: [email protected].
(December 5, 2020)
Abstract
In this paper we consider the problem of embedding almost spanning, bounded degree graphs in a random graph. In particular, let $\Delta\geq 5$ and let $H$ be a graph on $(1-o(1))n$ vertices and with maximum degree $\Delta$. We show that a random graph $G_{n,p}$ with high probability contains a copy of $H$, provided that $p\gg(n^{-1}\log^{1/\Delta}n)^{2/(\Delta+1)}$. Our assumption on $p$ is optimal (even up to the $polylogn$ factor, which is the conjectured one for the spanning case) and the result verifies the so-called Kahn-Kalai Conjecture for this class of graphs.
1 Introduction
Ever since its introduction by Erdős and Rényi in 1960 [8], the random graph model has been one of the main objects of
study in probabilistic combinatorics. Given a positive integer $n$
and a real number $p\in[0,1]$, the binomial random graph $G_{n,p}$ is a random variable taking values in the set
of all labeled graphs on the vertex set $[n]$. We can describe the
probability distribution of $G_{n,p}$ by saying that each pair of
elements of $[n]$ forms an edge in $G_{n,p}$ independently, with
probability $p$. Usually, the questions considered in this model
have the following generic form: given some graph property ${\cal P}$ and a function $p=p(n)\in[0,1]$, determine if $G_{n,p}$
satisfies ${\cal P}$ with high probability (whp), i.e., if the
probability that $G_{n,p}$ satisfies ${\cal P}$ tends to $1$ as $n$
tends to infinity.
The core meta-problem in the area is the study of
the evolution of $G_{n,p}$, i.e., analyzing how $G_{n,p}$
behaves with respect to certain graph properties as $p$ traverses
the interval $[0,1]$. This task is inherently connected with the
problem of finding thresholds guaranteeing almost sure
appearance (or almost sure non-appearance) of certain subgraphs in
$G_{n,p}$.
Recall that for a monotone property $\mathcal{P}_{n}$ (for example, “$G_{n,p}$ contains a fixed graph $H$”), a function $q(n)$ is a threshold function for $\mathcal{P}_{n}$ if and only if
$$\Pr\left[G_{n,p}\text{ satisfies }\mathcal{P}_{n}\right]\rightarrow\left\{%
\begin{array}[]{cc}1&\text{if }p(n)/q(n)=\omega(1)\\
0&\text{ if }p(n)/q(n)=o(1).\end{array}\right.$$
A classical result for random graphs is that a threshold for $G_{n,p}$ to
contain a given (fixed) graph $H$ as a subgraph, is
$n^{-1/m(H)}$, where $m(H)=\max\{|E(Hâ²)|/|V(Hâ²)|\mid Hâ²\subseteq H\}$.
The case where $H$ is “balanced” (that is, $m(H)=|E(H)|/|V(H)|$) had been proven in [8], and for a general $H$, one can find a proof in [5], pages 257-274.
Note that equivalently one can take as a threshold function the least $p(n)$ such that for each $H^{\prime}\subseteq H$, the expected number of copies of $H^{\prime}$ in $G_{n,p}$ is at least one. This particular function has been named as the expected threshold by Kahn and Kalai in [11].
Following [11], note that it also makes sense if, instead of a fixed $H$,
we consider a sequence $\{H_{n}\}_{n}$, of graphs with $|V(H_{n})|=n$. Formally,
for an arbitrary $H$, Kahn and Kalai defined $p_{E}(H)$ to be the least $p$ such that, for every spanning $H^{\prime}\subseteq H$, $(|V(H^{\prime})|!/|Aut(H^{\prime})|)p^{|E(H^{\prime})}|\geq 1$. Then, even for a fixed graph $H$, the expected threshold is the same as $p_{E}(H_{n})$ if we consider $H_{n}$ as $H$ plus $n-|V(H)|$ many isolated vertices.
For the class of large graphs (that is, whenever the number of vertices of $H$ is allowed to grow with $n$), one cannot expect $p_{E}(H)$ to capture the true threshold behavior. For example, suppose that $H_{n}$ are Hamilton cycles (that is, a cycle on $n$ vertices). In this case, on one hand, the expected number of Hamilton cycles in $G_{n,p}$ is clearly $\mu_{n}=\frac{(n-1)!}{2}p^{n}$, and therefore the expectation threshold is of order $1/n$. On the other hand, it is well known [13], that a threshold function for $H_{n}$ is of order $\log n/n$ (in fact, more precise results are known. The interested reader is referred to [5] and its relevant references). A similar phenomena holds for perfect matchings (that is, a collection of $\lfloor n/2\rfloor$ pairwise disjoint edges), as was proven in [8].
These examples lead Kahn and Kalai to the beautiful conjecture that in fact, one cannot lose more than a $\Theta(\log|V(H)|)$ factor in $p_{E}(H_{n})$. Specifically, they conjectured the following (Conjecture 2. in [11]):
Conjecture 1.1.
For any $\{H_{n}\}_{n}$, a threshold function for the property “$G_{n,p}$ contains $H_{n}$ ” is $O(p_{E}(H_{n})\log|V(H_{n})|)$.
Given a graph $H$ on $h$ vertices, we say that a graph $G$ on $n$ vertices contains an $H$-factor if it contains $\lfloor n/h\rfloor$ vertex disjoint copies of $H$. For example, if $H$ is an edge, then an $H$-factor is simply a matching of size $\lfloor n/2\rfloor$.
Perhaps the most impressive result supporting Conjecture 1.1 is due to Johansson, Kahn and Vu [10]. In their seminal paper, Johansson, Kahn and Vu determined the threshold behavior of $H$-factors. In particular, they showed that if $H=K_{\Delta+1}$ then
$$\displaystyle p(\Delta)=\left(n^{-1}\ln^{1/\Delta}n\right)^{\frac{2}{\Delta+1}}$$
(1)
is a threshold function for the existence of an $H$-factor. Their results naturally extend to hypergraphs and as a corollary they proved the long standing conjecture, which is known as Shamir’s Problem, of determining the threshold behavior of perfect matchings in random hypergraphs (this result garnered them the prestigious Fulkerson Prize!).
As their proof relies heavily on the facts that the connected components are small and that an $H$-factor has a lot of symmetries, it is natural to consider the following, more general, problem.
Problem 1.2.
Suppose that $H$ is an arbitrary graph on $n$ vertices with maximum degree $\Delta$. What is the threshold behavior of the property “contains $H$”?
Let us denote by $\mathcal{H}(n,\Delta)$ the family of all such graphs. A common belief is that the “worst case threshold” for $H\in\mathcal{H}(n,\Delta)$ is attained when $H$ is a $K_{\Delta+1}$ factor, as this is the the most “locally dense” graph in this family. It is thus natural to conjecture the following:
Conjecture 1.3.
Let $\Delta\in\mathbb{N}$, $H\in\mathcal{H}(n,\Delta)$, and $p=\omega\left(n^{-1}\ln^{1/\Delta}n\right)^{\frac{2}{\Delta+1}}$. Then, whp $G_{n,p}$ contains a copy of $H$.
Remark 1.4.
It is worth mentioning the following:
1.
If true, Conjecture 1.3 is optimal for the family $\mathcal{H}(n,\Delta)$ as it is tight for $K_{\Delta+1}$-factors.
2.
It may be the case that for some specific members $H\in\mathcal{H}(n,\Delta)$, one can do much better. For example, if $H\in\mathcal{H}(n,2)$ is a Hamilton cycle, as mentioned above, the threshold function for $H$ is roughly $\log n/n$, where from (1) we obtain a bound of roughly $n^{-2/3}$. Another example is the following remarkable recent result due to Montgomery [12]. He showed that for every $\Delta=O(1)$ and $H\in\mathcal{H}(n,\Delta)$ which is a tree, a threshold function for $H$ is of order of magnitude at most $\log^{5}n/n$.
In fact, we believe that something stronger is true. Let $\mathcal{F}$ be a family of graphs. A graph $G$ is called $\mathcal{F}$-universal if it contains every member of $\mathcal{F}$. We believe that the following is true.
Conjecture 1.5.
Let $\Delta\in\mathbb{N}$ and $p=\omega\left(n^{-1}\ln^{1/\Delta}n\right)^{\frac{2}{\Delta+1}}$. Then, whp $G_{n,p}$ is $\mathcal{H}(n,\Delta)$-universal.
As the problem of embedding large graphs in $G_{n,p}$ is fundamental, it has been studied quite extensively, particularly with respect to special graphs such as forests, planar graphs, and Hamilton cycles. Perhaps the first known result for general graphs from $\mathcal{H}(n,\Delta)$ is due to Alon and Füredi [3] from 1992. They introduced an embedding technique based on matchings, and showed that for a given
graph $H$ on $n$ vertices and with maximum degree $\Delta(H)=\Delta$, a
typical $G_{n,p}$ contains a copy of $H$, provided that
$p=\omega\left(n^{-1/\Delta}(\log n)^{1/\Delta}\right)$. Note that this bound on $p$ is natural, as in this range it is easy to see that whp every subset of size $\Delta$ has common neighbors in $G_{n,p}$. Therefore, this bound
has become a benchmark in the field and much subsequent work on
embedding spanning or almost-spanning structures of maximum degree
$\Delta$ in random graphs achieves this threshold.
Among all the examples we mention the following. Regarding the “universality problem”, Alon, Capalbo, Kohayakawa, Rödl, Ruciński and Szemerédi [2] showed that for any $\varepsilon>0$ and any natural number $\Delta$ there exists a constant $c>0$ such that the random graph $G_{np}$ is whp $\mathcal{H}((1-\varepsilon)n,\Delta)$-universal for $p\geq c(\log n/n)^{1/\Delta}$. This result was improved to $\mathcal{H}(n,\Delta)$ by Dellamonica, Kohayakawa, Rödl and Ruciński [7] and the $p$ has been slightly improved by Nenadov, Peter, and the first author [9] in case all graphs are “not locally dense”. Recently, Conlon, Nenadov, Skorić and the first author [6] managed to improve the edge probability to $p=n^{-1/(\Delta-1)}\log^{a}n$, for the family $\mathcal{H}((1-o(1))n,\Delta)$. Regarding the “embedding one graph” problem, it is worth mentioning the very nice result of Riordan [14], who managed to prove Conjecture 1.3 up to a factor of $n^{\Theta(1/\Delta^{2})}$ using the second moment method (note that his bound converges to the truth when $\Delta$ goes to infinity).
In this paper we solve Conjecture 1.3 for almost spanning graphs. That is, we prove the following:
Theorem 1.6.
Let $\varepsilon,\delta>0$ be any constants and let $\Delta\geq 5$ be an integer. Then, there exists an $N_{0}$ such that for all $n\geq N_{0}$ the following holds.
For every $H\in\mathcal{H}((1-\varepsilon)n,\Delta)$, with probability at least $1-\delta$ the random graph $G_{n,p}$ contains a copy of $H$, provided that $p=\omega\left(n^{-1}\log^{1/\Delta}n\right)^{2/(\Delta+1)}$
Note that for $\Delta=2$ it is relatively simple to obtain this result, and $\Delta=3$ is obtained from [6]. Our proof breaks down for $\Delta=4$, and it looks like a new idea is required for settling this case.
Another thing to be mentioned is that it seems like the $\log^{2/(\Delta(\Delta+1))}$ factor is redundant for the almost spanning case (for example, it is not needed if one tries to embed $(1-\varepsilon)\frac{n}{\Delta+1}$ copies of $K_{\Delta+1}$) but we have not managed to get rid of it.
Outline of the proof. Our proof strategy is relatively simple. Given a graph $H\in\mathcal{H}((1-o(1))n,\Delta)$, we first partition it into a “sparse” part $H^{\prime}$ and a “dense” part (see Section 2.2). The “sparse” part is going to be embedded using a result of Riordan [14] which is stated in Section 2.3. The “dense” part is basically the heart of our proof. While partitioning $H$, we make sure that the “dense” part consists of constant many “batches” of small graphs to be embedded iteratively. In each round, similar to the approach taken in [6], we make use of a hypergraph matching theorem of Aharoni and Haxell [1] (see Theorem 2.6) and Janson’s inequality to show that one can extend the current embedding.
1.1 Notation
Given two induced subgraph $S,S^{\prime}$ of a graph $G$, we let $dist_{G}(S,S^{\prime})$ denote the distance between them. That is, the length of the shortest path in $G$ connecting some vertex of $S$ to a vertex of $S^{\prime}$ (for example, if $V(S)\cap V(S^{\prime})\neq\emptyset$ then $dist_{G}(S,S^{\prime})=0$).
For a graph $G$, $V(G)$ and $E(G)$ denote its vertex set and edge set respectively. The number of vertices is denoted by $v(G)=|V(G)|$ and the number of edges $e(G)=|E(G)|$. For two vertices $x,y\in G$, if $xy\in E(G)$ then we sometimes abbreviate to $x\sim_{G}y$ (or $x\sim y$ if there is no risk of confusion).
For two sequences $f_{n}$ and $g_{n}$ of positive numbers, we say that $f_{n}=O(g_{n})$ or $g_{n}=\Omega(f_{n})$ if there exists a constant $C$ such that $f_{n}\leq Cg_{n}$ for all $n$. We say that $f_{n}=o(g_{n})$ or $g_{n}=\omega(f_{n})$ if $\lim_{n\to\infty}\frac{f_{n}}{g_{n}}=0$.
2 Auxiliary results
In this section we present some tools and auxiliary results to be used in the proof of our main result.
2.1 Probability estimates
We will use lower tail estimates for random
variables which count the number of copies of certain graphs in a
random graph. The following version of
Janson’s inequality, tailored for graphs, is the main estimate we will be using. This particular
statement follows immediately from Theorems $8.1.1$ and $8.1.2$ in
[4].
Theorem 2.1 (Janson’s inequality).
Let $p\in(0,1)$ and consider a family $\{H_{i}\}_{i\in\mathcal{I}}$ of subgraphs of the complete graph on the vertex set
$[n]$. For each $i\in\mathcal{I}$, let $X_{i}$
denote the indicator random variable for the event that $H_{i}\subseteq G_{n,p}$ and, for each ordered pair $(i,j)\in\mathcal{I}\times\mathcal{I}$ with $i\neq j$, write $H_{i}\sim H_{j}$ if $E(H_{i})\cap E(H_{j})\neq\emptyset$. Then, for
$$\displaystyle X$$
$$\displaystyle=\sum_{i\in\mathcal{I}}X_{i},$$
$$\displaystyle\mu$$
$$\displaystyle=\mathbb{E}[X]=\sum_{i\in\mathcal{I}}p^{e(H_{i})},\text{ and }$$
$$\displaystyle\delta$$
$$\displaystyle=\sum_{\begin{subarray}{c}(i,j)\in\mathcal{I}\times\mathcal{I}\\
H_{i}\sim H_{j}\end{subarray}}\mathbb{E}[X_{i}X_{j}]=\sum_{\begin{subarray}{c}%
(i,j)\in\mathcal{I}\times\mathcal{I}\\
H_{i}\sim H_{j}\end{subarray}}p^{e(H_{i})+e(H_{j})-e(H_{i}\cap H_{j})}$$
and any $0<\gamma<1$,
$$\Pr[X<(1-\gamma)\mu]\leq e^{-\frac{\gamma^{2}\mu^{2}}{2(\mu+\delta)}}.$$
2.2 Partitioning $H$
In order to embed a graph $H\in\mathcal{H}((1-o(1))n,\Delta)$ in $G_{n,p}$, we first wish to partition it “nicely” in a way which will be convenient for us to describe the embedding. Before stating it formally, we need the following definition.
Definition 2.2.
Let $H\in\mathcal{H}(m,\Delta)$, let $\mathcal{S}_{1},\ldots,\mathcal{S}_{k}$ be a collection of families of induced subgraphs of $H$ and let $H^{\prime}:=H\setminus\left(\cup_{i}\cup_{S\in\mathcal{S}_{i}}S\right)$. We say that the partition $(H^{\prime},\mathcal{S}_{1},\ldots,\mathcal{S}_{k})$ is $\alpha$-good for some $\alpha>0$ if and only if the following hold.
1.
For every $i=1,\dots,k$, all the graphs in $\mathcal{S}_{i}$ are isomorphic in $H$. More precisely, for any two graphs $S$ and $S^{\prime}$ in $\mathcal{S}_{i}$, there exists a way to label the vertices $V(S)=\{v_{1},\dots,v_{g}\}$ and $V(S^{\prime})=\{v_{1}^{\prime},\dots,v_{g}^{\prime}\}$ such that $v_{j}\sim v_{l}$ if and only if $v_{j}^{\prime}\sim v_{l}^{\prime}$ and $\deg_{H}(v_{j})=\deg_{H}(v_{j}^{\prime})$ for any $j$.
2.
Each $\mathcal{S}_{i}$ contains at most $\alpha n$ graphs.
3.
For every $1\leq i\leq k$ and every graph $S\in\mathcal{S}_{i}$, $S$ is minimal with respect to the property: $\frac{e(S)}{v(S)-2}>\frac{\Delta+1}{2}$. That is, $v(S)\geq 3$ and
$$\frac{e(S)}{v(S)-2}>\frac{\Delta+1}{2}\geq\frac{e(S^{\prime})}{v(S^{\prime})-2}$$
for all proper subgraphs $S^{\prime}\subset S$ with $v(S^{\prime})\geq 3$;
4.
Every two distinct graphs in $\cup_{i}\mathcal{S}_{i}$ are vertex disjoint;
5.
For every $i$ and every $S,S^{\prime}\in\mathcal{S}_{i}$ we have $dist_{H}(S,S^{\prime})\geq 3$;
6.
The graph $H^{\prime}$ has “small” density, namely, $\frac{e(S)}{v(S)-2}\leq\frac{\Delta+1}{2}$ for every subgraph $S\subseteq H^{\prime}$ with $v(S)\geq 3$.
Lemma 2.3.
Let $\alpha\in(0,1)$, $\Delta\in\mathbb{N}$ and let $H\in\mathcal{H}(m,\Delta)$ with $m$ being sufficiently large. Then, there exists an $\alpha$-good partition $(H^{\prime},\mathcal{S}_{1},\ldots,\mathcal{S}_{k})$ of $H$ with $k\leq k(\alpha,\Delta)$ where $k(\alpha,\Delta)$ is some constant depending only on $\alpha$ and $\Delta$.
Proof.
First, we find the graph $H^{\prime}$ greedily as follows.
Let $\mathcal{S}:=\emptyset$ and $H_{0}:=H$. As long as there exists an induced subgraph $S\subseteq H_{i}$ with at least three vertices and $e(S)/(v(S)-2)>(\Delta+1)/2$, pick a minimal (with respect to inclusion) such graph $S$ and update $\mathcal{S}:=\mathcal{S}\cup\{S\}$ and $H_{i+1}:=H_{i}\setminus V(S)$. Note that this procedure must terminate at some step $\ell$, and let $H^{\prime}:=H_{\ell}$. Observe that $e(S)/(v(S)-2)\leq(\Delta+1)/2$ for every $S\subseteq H^{\prime}$ (this verifies $6.$ in the definition of $\alpha$ goodness).
Second, note that since $\Delta(H)\leq\Delta$, it follows that for every subgraph $S\subseteq H$ we have $e(S)\leq v(S)\Delta/2$. In particular, for every $S\in\mathcal{S}$, together with the inequality $e(S)/(v(S)-2)>(\Delta+1)/2$ we conclude that $v(S)\leq 2\Delta+1$.
Third, define an auxiliary graph $\mathcal{A}$ with vertex set $\mathcal{S}$, where for $S$ and $S^{\prime}$ in $\mathcal{S}$, $SS^{\prime}$ is an edge in $\mathcal{A}$ if and only if $dist_{H}(S,S^{\prime})\leq 2$. Note that since $\Delta(H)\leq\Delta$ and since $v(S)\leq 2\Delta+1$ for every $S\in\mathcal{S}$, it follows that the maximum degree of $\mathcal{A}$ is $\Delta(\mathcal{A})\leq(2\Delta+1)\Delta^{2}\leq 3\Delta^{3}$. Therefore, one can (greedily) find a proper coloring of $\mathcal{S}$ using $c=3\Delta^{3}+1$ colors. Let $\mathcal{T}_{i}$ denote the color class $i$, and observe that $\mathcal{A}[\mathcal{T}_{i}]$ is an independent set in $\mathcal{A}$. In particular, we have that for every $i$ and every $S,S^{\prime}\in\mathcal{T}_{i}$, $dist_{H}(S,S^{\prime})\geq 3$.
Next, since $v(S)\leq 2\Delta+1$ for all $S\in\mathcal{S}$, it follows that there are at most (say) $s:=\Delta^{2\Delta+1}2^{{2\Delta+1\choose 2}}$ possible distinct such graphs (up to graph isomorphism in $H$), and therefore, one can further divide each of the classes $\mathcal{T}_{i}$ into at most $s$ subsets, each of which consists of isomorphic graphs.
Finally, if any of these subsets contain more than $\alpha n$ graphs, we further partition each of them into at most $\alpha^{-1}$ subsets. All in all, by relabeling, for $k\leq sc\alpha^{-1}$, one obtains an $\alpha$-good partitioning $\mathcal{S}_{1},\ldots,\mathcal{S}_{k}$ and $H^{\prime}$ as desired.
∎
2.3 Embedding the “sparse” part
Let $H\in\mathcal{H}(m,\Delta)$ and let $(H^{\prime},\mathcal{S}_{1},\ldots,\mathcal{S}_{k})$ be an $\alpha$-good partitioning of $H$ as obtained by Lemma 2.3. Note that, by definition of $\alpha$-goodness, for every $S\subseteq H$ with at least $3$ vertices, one has
$$e(S)/(v(S)-2)\leq(\Delta+1)/2.$$
In order to embed $H^{\prime}$, we make use of the following theorem due to Riordan (an immediate corollary of [14, Theorem 5]).
Theorem 2.4.
Let $\varepsilon,\delta>0$ and let $\Delta\geq 2$ be a positive integer. Then, there exists an $N_{0}$ such that for all $n\geq N_{0}$ the following holds.
For every $H\in\mathcal{H}(n,\Delta)$, let
$$\gamma(H)=\max_{S\subset H,v(S)\geq 3}\frac{e(S)}{v(S)-2}.$$
Then with probability at least $1-\delta$ the random graph $G_{n,M}$ contains a copy of $H$, provided that $M=p^{\prime}{n\choose 2}$ with $p^{\prime}\geq n^{-1/\gamma}\log\log n$.
Note that the property “containing $H$” is a monotone increasing property (that is, whenever a graph $G$ satisfies it, every graph $G^{\prime}\supseteq G$ satisfies it as well), and since whp $e(G_{n,p})\geq(1-o(1))\binom{n}{2}p$, it follows that the same conclusion of Theorem 2.4 holds for $G_{n,p}$, provided that (say) $p\geq 2n^{-1/\gamma}\log\log n$.
2.4 Embedding the small dense graphs
The following technical lemma is the primary ingredient in the proof of our main result. Roughly speaking, together with a hypergraph matching theorem (see Theorem 2.6), given an $\alpha$-good partitioning of $H$, we wish to find a tool which enables us to embed the graphs from the sets $\mathcal{S}_{i}$. Before making the exact statement, we need the following preparation.
Let $\varepsilon>0$ and $\Delta\geq 5$. Let $H\in\mathcal{H}((1-\varepsilon)n,\Delta)$, and let $(H^{\prime},\mathcal{S}_{1},\ldots,\mathcal{S}_{k})$ be an $\alpha$-good partition of $H$ where $\alpha=\varepsilon/(1000\Delta^{2})$ and $k\leq k(\alpha,\Delta)$(the existence of such a partition is ensured by Lemma 2.3). Our strategy in the proof of Theorem 1.6 is to expose $G_{n,p}$ in $k+1$ rounds $0\leq h\leq k$ as follows. In round $h=0$ we embed $H^{\prime}$ into a $G_{0}:=G_{n,q}$, where $q$ is the unique number $q\in(0,1)$ satisfying $(1-q)^{k+1}=1-p$. For each $0\leq h\leq k-1$, in round $h+1$ we extend our current embedding $f$ of $H_{h}:=H^{\prime}\cup\left(\cup_{i\leq h}\mathcal{S}_{i}\right)$ into an embedding of $H_{h+1}:=H_{h}\cup(\cup\mathcal{S}_{h+1})$ using the edge-set of a graph $G_{h+1}=G_{n,q}$ which is independent of all the other $G_{j}$s. Here we abuse notation and let $f$ denote the partial embedding at each stage. Our goal in this section is to prove the following key lemma which enables us to extend $f$.
Lemma 2.5.
Under the above setting, for every $h\leq k-1$, assuming that $f$ is an embedding of $H_{h}$ into $\cup_{i\leq h}G_{i}$, whp the embedding $f$ can be extended into an embedding of $H_{h+1}$ using the edges of $G_{h+1}$.
Clearly, given an embedding of $H^{\prime}$ into $G_{0}$, by iterating Lemma 2.5 $k$ times, one can (whp) extend $f$ into an embedding of $H$ into $\cup_{h=0}^{k}G_{h}$ which has the exact same distribution as a $G_{n,p}$.
In order to prove Lemma 2.5, we make use of the following Hall-type argument for finding large matchings in hypergraphs due to Aharoni and Haxel [1]. This idea is based on the one in [6], although here we have to be much more careful in order to obtain the correct $p$. Bellow, we describe how to use this theorem in our setting.
Theorem 2.6 (Hall’s criterion for hypergraphs, Corollary 1.2 [1]).
Let $\{L_{1}\dots L_{t}\}$ be a family of $s$-uniform hypergraphs on the same vertex set.
If, for every $\mathcal{I}\subseteq[t]$, the hypergraph $\bigcup_{i\in\mathcal{I}}L_{i}$ contains a matching of size greater than $s(|\mathcal{I}|-1)$, then there exists a function $g:[t]\rightarrow\bigcup_{i=1}^{t}E(L_{i})$ such that $g(i)\in E(L_{i})$ and $g(i)\cap g(j)=\emptyset$ for $i\neq j$.
In order to apply Theorem 2.6 to our setting we proceed as follows.
First, let us enumerate $\mathcal{S}_{h+1}=\{S_{1},\ldots,S_{t}\}$ and recall that since our partitioning is $\alpha$-good, all the graphs $S_{i}$ in $\mathcal{S}_{h+1}$ are isomorphic (henceforth, we refer to all the graphs $S_{i}$ as $S$). Moreover, letting $s=v(S)$, from the definition of $\alpha$-goodness and the proof of Lemma 2.3, we have
$$t\leq\alpha n,$$
(2)
$$s\leq 2\Delta+1,$$
(3)
and for any proper subgraph $S^{\prime}$ of $S$, one has
$$\frac{e(S^{\prime})}{v(S^{\prime})-2}\leq\frac{\Delta+1}{2}<\frac{e(S)}{s-2}.$$
(4)
Second, suppose that $V(S)=[s]$, and for every $i\leq t$, let us enumerate the vertex set of $S_{i}$ as $v_{i1},\dots,v_{is}$ in such a way that the map $\varphi(j)=v_{ij}$ is a graph isomorphism between $S$ and $S_{i}$, for all $i$. Let $D_{h}=[n]\setminus f(V(H_{h}))$. Since $v(H_{h})\leq(1-\varepsilon)n$, it follows that $|D_{h}|\geq\varepsilon n$. Let us choose
$$\alpha=\frac{\varepsilon}{2(2\Delta+1)^{2}}.$$
(5)
Clearly, we have that
$$|D_{h}|-s^{2}t\geq\varepsilon n-s^{2}\alpha n\geq\varepsilon n/2,$$
(6)
and this will be useful in Lemma 2.7 below.
Third, let $W_{ij}$ be the (image of the) set of neighbors of $v_{ij}$ in $H_{h}$. Note that all the $W_{ij}$s are fixed because $H_{h}$ has already been embedded and there are no edges between graphs in $\mathcal{S}_{h+1}$. Moreover, by $5.$ of Definition 2.2, the family $\{W_{ij}\}_{ij\in[t]\times[s]}$ satisfies
(i)
$W_{ij}\subseteq[n]\setminus D_{h}$, and $|W_{ij}|=|W_{i^{\prime}j}|\leq d_{j}:=\Delta-\text{indeg}(v_{ij})$ where $\text{indeg}(v_{ij})$ is the degree of $v_{ij}$ in $S_{i}$, and
(ii)
$W_{ij}\cap W_{i^{\prime}j^{\prime}}=\emptyset$ for all $i\neq i^{\prime}$.
Next, for each $1\leq i\leq t$, let $L_{i}$ be the $s$-uniform hypergraph on the vertex set $D_{h}$ where $(v_{1},\dots,v_{s})\in D_{h}^{s}$ is an hyperedge if and only if, in $G_{h+1}=G_{n,q}$ (the new random graph with edge-probabillity $q$ to be generated), $v_{j}$ is connected to every vertex in $W_{ij}$ and $v_{ij}v_{ij^{\prime}}\in E(S_{i})$ implies $v_{j}v_{j^{\prime}}\in E(G_{h+1})$. That is, every hyperedge $e$ of $L_{i}$ corresponds to an embedding of $S_{i}$ which extends $f$. A moment’s thought now reveals that extending $f$ into an embedding of $H_{h+1}$ is equivalent to showing that there exists a function $g:[t]\rightarrow\bigcup_{i=1}^{t}E(L_{i})$ such that $g(i)\in E(L_{i})$ and $g(i)\cap g(j)=\emptyset$ for every $i\neq j$. To this end, in order to check that the assumptions of Theorem 2.6 are satisfied (and therefore, such a $g$ exists), we need the following technical lemma.
Lemma 2.7.
For every $r\leq t$ and $1\leq h\leq k$, with probability at least $1-\exp\left(-\omega(r\log n)\right)$, the following holds. For every $D^{\prime}\subseteq D_{h}$
of size $|D^{\prime}|\geq|D_{h}|-s^{2}r$, and every $\mathcal{I}\subseteq[t]$ of size $|\mathcal{I}|=r$, $G_{h+1}$ contains a copy $S^{\prime}=(v_{1},\ldots,v_{s})$ of $S$ for which, $v_{j}\in D^{\prime}$ for all $1\leq j\leq s$ and for some $i\in\mathcal{I}$, $W_{ij}$ is contained in the neighborhood of $v_{j}$ for all $1\leq j\leq s$.
In the rest of this section we prove Lemma 2.7.
Proof of Lemma 2.7.
Let us fix $r\leq t$, $D^{\prime}$ and $\mathcal{I}$ as in the assumptions of the lemma. Note that $s^{2}r\leq s^{2}t\leq\epsilon n/2$ and $|D_{h}|\leq n$. Therefore, there are at most $\binom{n}{n-s^{2}r}\binom{n}{r}\leq\exp\left(s^{2}r\log n+r\log n\right)=\exp%
\left(\Theta(r\log n)\right)$ ways to choose $D^{\prime}$ and $\mathcal{I}$. Since the the tail probability we aim to show is $\exp\left(-\omega(r\log n)\right)$, this allows us to take a union bound.
Next, for a copy $S^{\prime}=(v_{1},\ldots,v_{s})$ of $S$, we define the graph $S^{\prime}\oplus i$ with vertex set and edge set
$$\displaystyle V(S\oplus i)=V(S)\cup\bigcup_{j=1}^{s}W_{ij},\quad\text{and}%
\quad E(S\oplus i)=E(S)\cup\{v_{j}v\mid j\in[s]\text{ and }v\in W_{ij}\}.$$
Furthermore, we decompose $D^{\prime}$ into $s$ disjoint sets $V_{1},\ldots,V_{s}\subseteq D^{\prime}$ each of which has size roughly $|D^{\prime}|/s$, and define the family of canonical graphs $\mathcal{S}$ as
$$\mathcal{S}:=\{S^{\prime}=(v_{1},\ldots,v_{s}):S^{\prime}\mbox{ is a copy of }%
S\mbox{ and }v_{j}\in V_{j}\;\text{for all}\;j\in[s]\}$$
and set
$$\mathcal{S}^{+}:=\{S^{\prime}\oplus i:S^{\prime}\in\mathcal{S}\;\text{and}\;i%
\in[r]\}.$$
Observe that if $G$ contains any graph from $\mathcal{S}^{+}$, then $G$ contains the desired copy. In order to bound the probability for $G$ not to contain such a graph, we apply Janson’s inequality (Theorem 2.1). Thus, we need to estimate the parameters $\mu$ and $\delta$ appearing in Theorem 2.1.
For every $S^{\prime}=(v_{1},\ldots,v_{s})\in\mathcal{S}$ we have
$$\Delta s=\sum_{j=1}^{s}(d_{j}+\text{indeg}(u_{j}))=\sum_{j=1}^{s}d_{j}+2e(S).$$
(7)
Note that each graph $S^{+}\in\mathcal{S}^{+}$ appears in $G_{h}$ with probability $q^{e(S^{+})}$, where $e(S^{+})=e(S)+\sum_{j=1}^{s}|W_{ij}|\leq e(S)+\sum_{j=1}^{n}d_{j}$. Clearly,
$$\mu=|\mathcal{S}^{+}|q^{e(S^{+})}=r\left(\frac{|D^{\prime}|}{s}\right)^{s}q^{e%
(S^{+})}.$$
It is worth observing that the last equation is not exact as the sizes of the sets $V_{j}$ may differ by $1$. However, this will not affect our calculations and we abstain from correcting this to avoid clutter in the notation.
Our goal is to show that $\mu=\omega(r\log n)$ and $\delta=o(\mu^{2}r^{-1}\log^{-1}n)$. Indeed, assuming so, by Janson’s inequality we obtain
$$\mathbb{P}(G\text{ does not contain any }S^{+}\in\mathcal{S}^{+})\leq\exp\left%
(-\frac{\mu^{2}}{4(\mu+\delta)}\right)=\exp(-\omega(r\log n))$$
as desired.
First, we show that $\mu=\omega(r\log n)$. By our choice of $\alpha$ in (5), it follows that $|D^{\prime}|\geq\varepsilon n/2$. Moreover, as $p=o(1)$, $k$ is a constant, and $(1-q)^{k+1}=1-p$, it follows that $q=\Theta(p)$. Therefore, we have that
$$\mu=\Omega(rn^{s}q^{e(S)+\sum_{j=1}^{s}d_{j}})=\omega\left(rn^{s-\frac{2e(S)+2%
\sum d_{j}}{\Delta+1}}(\log n)^{2(e(S)+\sum d_{j})/(\Delta(\Delta+1))}\right).$$
In order to estimate $\mu$, we show that
$$\Delta(\Delta+1)\leq 2e(S)+2\sum d_{j}\leq(\Delta+1)s,$$
(8)
from which we conclude that
$$\mu=\omega\left(r\log n)\right),$$
as desired.
By equation (7), (8) is equivalent to proving that
$$\Delta(\Delta+1-s)\leq\sum_{j=1}^{s}d_{j}\leq s.$$
(9)
Recall that by property (4) we have
$$\frac{e(S)}{s-2}>\frac{\Delta+1}{2}.$$
Therefore, combining this with (7), we obtain that
$$2e(S)>(\Delta+1)s-2(\Delta+1)=2e(S)+\sum d_{j}+s-2(\Delta+1),$$
which yields
$$s+\sum d_{j}\leq 2\Delta+1.$$
(10)
If $s\geq\Delta+1$, then clearly $0\leq\sum d_{j}\leq\Delta<s$, proving (9).
If $s=\Delta$, then by (10), $\sum d_{j}\leq\Delta+1$. From (7), $\sum d_{j}$ has the same parity as $\Delta s=\Delta^{2}$ and therefore it cannot be $\Delta+1$. Moreover, since $e(S)\leq{s\choose 2}$, from (7), we also find that $\sum d_{j}\geq\Delta$. Thus, $\sum d_{j}=\Delta$, proving (9).
Finally, we show that $s\leq\Delta-1$ cannot happen. Assume for contradiction that $s\leq\Delta-1$. If $s=3$, due to the assumptions $\frac{e(S)}{s-2}>\frac{\Delta+1}{2}$ and $\Delta\geq 5$, $e(S)$ has to be at least $4$ which is impossible. Hence, $4\leq s\leq\Delta-1$. Plugging $e(S)\leq{s\choose 2}$ into (4) gives $\frac{s(s-1)}{s-2}>\Delta+1$, or equivalently, $s^{2}-(\Delta+2)s+2(\Delta+1)>0$. Since $\Delta\geq 5$, the equation $x^{2}-(\Delta+2)x+2(\Delta+1)=0$ has two real roots
$$x_{1}=\frac{\Delta+2-\sqrt{\Delta^{2}-4\Delta-4}}{2}$$
and
$$x_{2}=\frac{\Delta+2+\sqrt{\Delta^{2}-4\Delta-4}}{2}.$$
By a simple calculation one can show that $x_{2}\geq\Delta-1\geq s$ when $\Delta\geq 5$ and $x_{1}<4\leq s$. Thus, $s$ is bounded between $x_{1}$ and $x_{2}$. This gives a contradiction and completes the proof of (9) and hence (8).
To prove that $\delta=o(\mu^{2}r^{-1}\log^{-1}n)$, we have
$$\delta=\sum_{i,j\in[r]}\sum_{\begin{subarray}{c}S^{\prime},S^{\prime\prime}\in%
\mathcal{S}\\
S^{\prime}\oplus i\sim S^{\prime\prime}\oplus j\end{subarray}}q^{e(S^{\prime}%
\oplus i)+e(S^{\prime\prime}\oplus j)-e((S^{\prime}\oplus i)\cap(S^{\prime%
\prime}\oplus j))}.$$
We consider the two cases $i=j$ and $i\neq j$.
First, let us now consider the case $S^{\prime}\oplus i\sim S^{\prime\prime}\oplus i$, for some $i\in[r]$ and distinct graphs $S^{\prime},S^{\prime\prime}\in\mathcal{S}$. Let $J:=S^{\prime}\cap S^{\prime\prime}$ and observe that $v(J)\geq 1$ and $v(J)\leq s-1$. Let $\mathcal{J}_{1}$ be the family consisting of all possible graphs of the form $S^{\prime}\cap S^{\prime\prime}$. That is,
$$\mathcal{J}_{1}:=\{J=S^{\prime}\cap S^{\prime\prime}:S^{\prime},S^{\prime%
\prime}\in\mathcal{S}\;\text{and}\;v(J)\in\{1,\ldots,s-1\}\}.$$
We have
$$e(S^{\prime}\oplus i\cap S^{\prime\prime}\oplus i)=e(J)+\sum_{j\in J}|W_{i,j}|%
\leq e(J)+\sum_{j\in J}d_{j}.$$
With these observations in hand, we can bound the contribution of such pairs to $\delta$ as follows:
$$\displaystyle\delta_{1}=\sum_{i\in[r]}\sum_{\begin{subarray}{c}S^{\prime},S^{%
\prime\prime}\in\mathcal{S}\\
v(S^{\prime}\cap S^{\prime\prime})\geq 1\end{subarray}}q^{e(S^{\prime}\oplus i%
)+e(S^{\prime\prime}\oplus i)-e(S^{\prime}\oplus i\cap S^{\prime\prime}\oplus i%
)}\leq\sum_{i\in[r]}\sum_{J\in\mathcal{J}_{1}}\sum_{\begin{subarray}{c}S^{%
\prime},S^{\prime\prime}\in\mathcal{S}\\
S^{\prime}\cap S^{\prime\prime}=J\end{subarray}}q^{2e({S^{+}})-(e(J)+\sum_{j%
\in J}d_{j})}$$
and so
$$\displaystyle\delta_{1}\leq r\sum_{J\in\mathcal{J}_{1}}\left(\frac{|D^{\prime}%
|}{s}\right)^{2(s-v(J))}q^{2e({S^{+}})-(e(J)+\sum_{j\in J}d_{j})}=\frac{\mu^{2%
}}{r}\sum_{J\in\mathcal{J}_{1}}\left(\frac{|D^{\prime}|}{s}\right)^{-2v(J)}q^{%
-(e(J)+\sum_{j\in J}d_{j})}.$$
(11)
Let us now show that the power $e(J)+\sum_{j\in J}d_{j}$ cannot be large. Since $I:=S\setminus J$ is a proper subgraph of $S$, we have from assumption (4),
$$\frac{e(I)}{v(I)-2}\leq\frac{\Delta+1}{2}<\frac{e(S)}{s-2}.$$
Rearranging, one gets
$$\displaystyle(\Delta+1)v(I)-2e(I)\geq 2(\Delta+1)>(\Delta+1)s-2e(S),$$
which gives
$$\Delta v(J)<2e(J,I)+2e(J)-v(J)$$
where $e(J,I)$ is the number of edges connecting $I$ and $J$.
Combining that with
$$\Delta v(J)=2e(J)+\sum_{j\in J}d_{j}+e(J,I),$$
yields
$$\displaystyle\sum_{j\in J}d_{j}\leq e(J,I)-v(J)-1$$
and so
$$\displaystyle e(J)+\sum_{j\in J}d_{j}\leq\frac{2e(J)+\sum_{j\in J}d_{j}+e(J,I)%
-v(J)-1}{2}=\frac{(\Delta-1)v(J)-1}{2}<\frac{(\Delta-1)v(J)}{2}.$$
Plugging this bound into (11), we obtain
$$\displaystyle\delta_{1}$$
$$\displaystyle\leq$$
$$\displaystyle\frac{\mu^{2}}{r}\sum_{J\in\mathcal{J}_{1}}\left(\frac{|D^{\prime%
}|}{s}\right)^{-2v(J)}q^{-\frac{\Delta-1}{2}v(J)}.$$
Note that there are at most $\binom{s}{v_{J}}\left(|D^{\prime}|/s\right)^{v_{J}}$ graphs $J\in\mathcal{J}_{1}$ on $v_{J}$ vertices. Thus, we have
$$\displaystyle\delta_{1}\leq\frac{\mu^{2}}{r}\sum_{v_{J}=1}^{s-1}\binom{s}{v_{J%
}}\left(\frac{|D^{\prime}|}{s}\right)^{-v_{J}}q^{-\frac{(\Delta-1)v(J)}{2}}=O(%
\mu^{2}r^{-1}n^{-\frac{2}{\Delta+1}})=o(\mu^{2}r^{-1}\log^{-1}n).$$
Next, if $S^{\prime}\oplus i\sim S^{\prime\prime}\oplus j$ for $i\neq j$ and $S^{\prime},S^{\prime\prime}\in\mathcal{S}$, then we have $(S^{\prime}\oplus i)\cap(S^{\prime\prime}\oplus j)=S^{\prime}\cap S^{\prime\prime}$. Let $J:=S^{\prime}\cap S^{\prime\prime}$ and
observe that $e(J)\geq 1$, as otherwise $S^{\prime}\oplus i$ and $S^{\prime\prime}\oplus j$ would not have any edges in common. Let $\mathcal{J}_{2}$ be the
family consisting of all possible graphs of the form $S^{\prime}\cap S^{\prime\prime}$,
$$\mathcal{J}_{2}:=\{J=S^{\prime}\cap S^{\prime\prime}:S^{\prime},S^{\prime%
\prime}\in\mathcal{S}\;\text{and}\;e(J)\geq 1\}.$$
The contribution of such pairs to $\delta$ is
$$\delta_{2}=\sum_{i\neq j}\sum_{\begin{subarray}{c}S^{\prime},S^{\prime\prime}%
\in\mathcal{S}\\
e(S^{\prime}\cap S^{\prime\prime})\geq 1\end{subarray}}q^{e(S^{\prime}\oplus i%
)+e(S^{\prime\prime}\oplus j)-e(S^{\prime}\cap S^{\prime\prime})}=\sum_{i\neq j%
}\sum_{J\in\mathcal{J}_{2}}\sum_{\begin{subarray}{c}S^{\prime},S^{\prime\prime%
}\in\mathcal{S}\\
S^{\prime}\cap S^{\prime\prime}=J\end{subarray}}q^{2e({S^{+}})-e(J)}$$
and can by bounded by
$$\delta_{2}\leq r^{2}\sum_{J\in\mathcal{J}_{2},2\leq v(J)\leq s}\left(\frac{|D^%
{\prime}|}{s}\right)^{2(s-v(J))}q^{2e({S^{+}})-e(J)}=\mu^{2}\sum_{J\in\mathcal%
{J}_{2},2\leq v(J)\leq s}\left(\frac{|D^{\prime}|}{s}\right)^{-2v(J)}q^{-e(J)}$$
which can be further split into two terms when $S^{\prime}\neq S^{\prime\prime}$ and when $S^{\prime}=S^{\prime\prime}$,
$$\delta_{2}\leq\mu^{2}\sum_{J\in\mathcal{J}_{2},2\leq v(J)<s}\left(\frac{|D^{%
\prime}|}{s}\right)^{-2v(J)}q^{-e(J)}+\mu^{2}\sum_{S\in\mathcal{S}}n^{-2s}q^{-%
e(S)}=:\delta_{2,1}+\delta_{2,2}.$$
For $3\leq v(J)<s$ and $J\subset S$ with $S\in\mathcal{S}$, by property (4), we have
$$e(J)\leq\frac{\Delta+1}{2}(v(J)-2).$$
When $v(J)=2$, $e(J)=1$ by the definition of $\mathcal{J}_{2}$.
Thus,
$$\displaystyle\delta_{2,1}$$
$$\displaystyle=O(\mu^{2})n^{-2}q^{-1}+O(\mu^{2})\sum_{v_{J}=3}^{s-1}n^{-v_{J}}q%
^{-\frac{\Delta+1}{2}(v(J)-2)}=o\left(\mu^{2}n^{-\frac{2\Delta}{\Delta+1}}%
\right)=o\left(\mu^{2}r^{-1}\log^{-1}n\right).$$
For $\delta_{2,2}$, we have
$$\delta_{2,2}=O\left(\mu^{2}n^{-s}q^{-e(S)}\right)=o\left(\mu^{2}n^{-s+\frac{2e%
(S)}{\Delta+1}}(\log n)^{-\frac{2e(S)}{\Delta(\Delta+1)}}\right).$$
Recall that we showed in the proof of the lower bound for $\mu$ that $s\geq\Delta$.
If $s\geq\Delta+1$ then
$$n^{-s+\frac{2e(S)}{\Delta+1}}\leq n^{-s+\frac{\Delta s}{\Delta+1}}=n^{-\frac{s%
}{\Delta+1}}\leq n^{-1}\leq r^{-1}.$$
In these estimates, unless $s=\Delta+1$ and $e(S)=s\Delta/2$, we have
$$n^{-s+\frac{2e(S)}{\Delta+1}}\leq n^{-1-\frac{1}{\Delta+1}}\leq r^{-1}\log^{-1%
}n.$$
Hence,
$$\delta_{2,2}=o\left(\mu^{2}r^{-1}\log^{-1}n\right).$$
In the case that $s=\Delta+1$ and $e(S)=s\Delta/2$, that is $S$ is a clique on $\Delta+1$ vertices, we still have
$$\delta_{2,2}=o\left(\mu^{2}n^{-s+\frac{2e(S)}{\Delta+1}}(\log n)^{-\frac{2e(S)%
}{\Delta(\Delta+1)}}\right)=o\left(\mu^{2}r^{-1}\log^{-1}n\right).$$
If $s=\Delta$ then
$$e(S)\leq{s\choose 2}=(\Delta-1)\Delta/2.$$
Thus,
$$n^{-s+\frac{2e(S)}{\Delta+1}}\leq n^{-\frac{2\Delta}{\Delta+1}}\leq n^{-1-%
\frac{\Delta-1}{\Delta+1}}\leq r^{-1}n^{-\frac{\Delta-1}{\Delta+1}}.$$
In all cases, we obtain
$$\delta_{2}=o(\mu^{2}r^{-1}\log^{-1}n).$$
To sum up, we have
$$\delta\leq\delta_{1}+\delta_{2}=O(\mu^{2}r^{-1}\log^{-1}n),$$
completing the proof of the lemma.
∎
Now we are ready to prove Lemma 2.5.
Proof.
All that is left to do is to check that the condition in Theorem 2.6 is satisfied. That is, for every $\mathcal{I}\subseteq[t]$, the hypergraph $\bigcup_{i\in\mathcal{I}}H_{i}$ contains a matching of size greater than $s(|\mathcal{I}|-1)$. Let $\mathcal{I}\subseteq[t]$ and let $r:=|\mathcal{I}|$. Assume that the conclusion of Lemma 2.7 holds and choose $D^{\prime}=D_{h}$. Clearly, one can find a hyperedge $e_{1}\in L_{i_{1}}$ for some $i_{1}\in\mathcal{I}$. Let $D^{\prime}:=D\setminus e$ and observe that $|D^{\prime}|\geq|D|-s$. Then, by Lemma 2.7, there exists another edge $e_{2}\in L_{i_{2}}$ for some $i_{2}\in\mathcal{I}$. Repeating this argument $s(r-1)$ times, using the fact that the set $D^{\prime}$ always has size at least $|D|-s^{2}(r-1)\geq|D|-s^{2}r$, gives us the desired result.
∎
3 Proof of Theorem 1.6
In this section we prove Theorem 1.6.
Proof.
Actually, we have already described more or less the whole proof in Section 2.4 but for completeness, we will repeat the main steps of the argument (for more details, the reader can read the first part of Section 2.4 again).
Let $H\in\mathcal{H}((1-\varepsilon)n,\Delta)$ and let $(H^{\prime},\mathcal{S}_{1},\ldots,S_{k})$ be an $\alpha$-good partition of $H$ as obtained by Lemma 2.3 with respect to $\alpha=\frac{\varepsilon}{2(2\Delta+1)^{2}}$ (recall that $k:=k(\Delta,\alpha)$ is a constant). Let us fix $p=\beta(n)(n^{-1}\log^{1/\Delta})^{2/(\Delta+1)}$, where $\beta(n)\rightarrow\infty$ arbitrarily slowly whenever $n$ goes to infinity. Note that as “containing $H$” is a monotone increasing property, it is enough to prove the theorem for this $p$.
Let $q$ be the unique number $0<q<1$ for which
$$(1-q)^{k+1}=1-p,$$
and observe that as $p=o(1)$ and $k$ is a constant, it follows that $q=(1+o(1))p/k=\Omega(p)$.
We embed $H$ in $k+1$ rounds, where in each round $0\leq h\leq k$ we expose the edges of a graph $G_{h}:=G_{n,q}$, independently of the previous exposures, and extend a current partial embedding $f$ of $H^{\prime}\cup\left(\bigcup_{1\leq i<h}\mathcal{S}_{i}\right)$, using the edges of $G_{h}$. Clearly, $\cup G_{h}$ can be coupled as $G_{n,p}$. Formally, we can describe the scheme as follows:
Round $0$. Embed $H_{0}:=H^{\prime}$ into $G_{0}$. In order to do so, observe that by property $6.$ of the definition of $\alpha$-goodness, using the notation in Theorem 2.4, we have that
$$\gamma(H^{\prime})\leq(\Delta+1)/2.$$
Therefore, using Theorem 2.4 we find (whp) a copy of $H^{\prime}$ in $G_{0}$.
Round $h+1$. Extend the embedding of $H_{h}$ into an embedding of $H_{h+1}:=H_{h}\cup\mathcal{S}_{h+1}$. This is obtained whp directly from Lemma 2.5, and one can iterate this $k+1$ times. This completes the proof.
∎
References
[1]
R. Aharoni and P. Haxell.
Hall’s theorem for hypergraphs.
Journal of Graph Theory, 35(2):83–88, 2000.
[2]
N. Alon, M. Capalbo, Y. Kohayakawa, V. Rödl, A. Ruciński, and
E. Szemerédi.
Universality and tolerance.
In Foundations of Computer Science, 2000. Proceedings. 41st
Annual Symposium on, pages 14–21. IEEE, 2000.
[3]
N. Alon and Z. Füredi.
Spanning subgraphs of random graphs.
Graphs and Combinatorics, 8(1):91–94, 1992.
[4]
N. Alon and J. H. Spencer.
The Probabilistic Method.
John Wiley & Sons, 2004.
[5]
B. Bollobás.
Random graphs.
Springer, 1998.
[6]
D. Conlon, A. Ferber, R. Nenadov, and N. Škorić.
Almost-spanning universality in random graphs.
Random Structures & Algorithms, arXiv preprint
arXiv:1503.05612v2.
[7]
D. Dellamonica Jr, Y. Kohayakawa, V. Rödl, and A. Ruciński.
Universality of random graphs.
In Proceedings of the nineteenth annual ACM-SIAM symposium on
Discrete algorithms, pages 782–788. Society for Industrial and Applied
Mathematics, 2008.
[8]
P. Erdős and A. Rényi.
On the evolution of random graphs.
Magyar Tud. Akad. Mat. Kutató Int. Közl., 5:17–61,
1960.
[9]
A. Ferber, R. Nenadov, and U. Peter.
Universality of random graphs and rainbow embedding.
to appear in Random Structure & Algorithms, arXiv preprint
arXiv:1311.7063.
[10]
A. Johansson, J. Kahn, and V. Vu.
Factors in random graphs.
Random Structures & Algorithms, 33(1):1–28, 2008.
[11]
J. Kahn and G. Kalai.
Thresholds and expectation thresholds.
Combinatorics, Probability and Computing, 16(03):495–502,
2007.
[12]
R. Montgomery.
Embedding bounded degree spanning trees in random graphs.
arXiv preprint arXiv:1405.6559, 2014.
[13]
L. Pósa.
Hamiltonian circuits in random graphs.
Discrete Mathematics, 14(4):359–364, 1976.
[14]
O. Riordan.
Spanning subgraphs of random graphs.
Combinatorics, Probability and Computing, 9(02):125–148, 2000. |
pacs: 13.20.He, 12.15.Mm, 12.60.Jv
EUROPEAN ORGANIZATION FOR NUCLEAR RESEARCH (CERN)
CERN-PH-EP-2013-123
LHCb-PAPER-2013-030
July 18, 2013
Search for the lepton-flavour violating decays $B^{0}_{s}\rightarrow e^{\pm}\mu^{\mp}$ and $B^{0}\rightarrow e^{\pm}\mu^{\mp}$
The LHCb collaboration
R. Aaij${}^{40}$,
B. Adeva${}^{36}$,
M. Adinolfi${}^{45}$,
C. Adrover${}^{6}$,
A. Affolder${}^{51}$,
Z. Ajaltouni${}^{5}$,
J. Albrecht${}^{9}$,
F. Alessio${}^{37}$,
M. Alexander${}^{50}$,
S. Ali${}^{40}$,
G. Alkhazov${}^{29}$,
P. Alvarez Cartelle${}^{36}$,
A.A. Alves Jr${}^{24,37}$,
S. Amato${}^{2}$,
S. Amerio${}^{21}$,
Y. Amhis${}^{7}$,
L. Anderlini${}^{17,f}$,
J. Anderson${}^{39}$,
R. Andreassen${}^{56}$,
J.E. Andrews${}^{57}$,
R.B. Appleby${}^{53}$,
O. Aquines Gutierrez${}^{10}$,
F. Archilli${}^{18}$,
A. Artamonov${}^{34}$,
M. Artuso${}^{58}$,
E. Aslanides${}^{6}$,
G. Auriemma${}^{24,m}$,
M. Baalouch${}^{5}$,
S. Bachmann${}^{11}$,
J.J. Back${}^{47}$,
C. Baesso${}^{59}$,
V. Balagura${}^{30}$,
W. Baldini${}^{16}$,
R.J. Barlow${}^{53}$,
C. Barschel${}^{37}$,
S. Barsuk${}^{7}$,
W. Barter${}^{46}$,
Th. Bauer${}^{40}$,
A. Bay${}^{38}$,
J. Beddow${}^{50}$,
F. Bedeschi${}^{22}$,
I. Bediaga${}^{1}$,
S. Belogurov${}^{30}$,
K. Belous${}^{34}$,
I. Belyaev${}^{30}$,
E. Ben-Haim${}^{8}$,
G. Bencivenni${}^{18}$,
S. Benson${}^{49}$,
J. Benton${}^{45}$,
A. Berezhnoy${}^{31}$,
R. Bernet${}^{39}$,
M.-O. Bettler${}^{46}$,
M. van Beuzekom${}^{40}$,
A. Bien${}^{11}$,
S. Bifani${}^{44}$,
T. Bird${}^{53}$,
A. Bizzeti${}^{17,h}$,
P.M. Bjørnstad${}^{53}$,
T. Blake${}^{37}$,
F. Blanc${}^{38}$,
J. Blouw${}^{11}$,
S. Blusk${}^{58}$,
V. Bocci${}^{24}$,
A. Bondar${}^{33}$,
N. Bondar${}^{29}$,
W. Bonivento${}^{15}$,
S. Borghi${}^{53}$,
A. Borgia${}^{58}$,
T.J.V. Bowcock${}^{51}$,
E. Bowen${}^{39}$,
C. Bozzi${}^{16}$,
T. Brambach${}^{9}$,
J. van den Brand${}^{41}$,
J. Bressieux${}^{38}$,
D. Brett${}^{53}$,
M. Britsch${}^{10}$,
T. Britton${}^{58}$,
N.H. Brook${}^{45}$,
H. Brown${}^{51}$,
I. Burducea${}^{28}$,
A. Bursche${}^{39}$,
G. Busetto${}^{21,q}$,
J. Buytaert${}^{37}$,
S. Cadeddu${}^{15}$,
O. Callot${}^{7}$,
M. Calvi${}^{20,j}$,
M. Calvo Gomez${}^{35,n}$,
A. Camboni${}^{35}$,
P. Campana${}^{18,37}$,
D. Campora Perez${}^{37}$,
A. Carbone${}^{14,c}$,
G. Carboni${}^{23,k}$,
R. Cardinale${}^{19,i}$,
A. Cardini${}^{15}$,
H. Carranza-Mejia${}^{49}$,
L. Carson${}^{52}$,
K. Carvalho Akiba${}^{2}$,
G. Casse${}^{51}$,
L. Castillo Garcia${}^{37}$,
M. Cattaneo${}^{37}$,
Ch. Cauet${}^{9}$,
R. Cenci${}^{57}$,
M. Charles${}^{54}$,
Ph. Charpentier${}^{37}$,
P. Chen${}^{3,38}$,
N. Chiapolini${}^{39}$,
M. Chrzaszcz${}^{25}$,
K. Ciba${}^{37}$,
X. Cid Vidal${}^{37}$,
G. Ciezarek${}^{52}$,
P.E.L. Clarke${}^{49}$,
M. Clemencic${}^{37}$,
H.V. Cliff${}^{46}$,
J. Closier${}^{37}$,
C. Coca${}^{28}$,
V. Coco${}^{40}$,
J. Cogan${}^{6}$,
E. Cogneras${}^{5}$,
P. Collins${}^{37}$,
A. Comerma-Montells${}^{35}$,
A. Contu${}^{15,37}$,
A. Cook${}^{45}$,
M. Coombes${}^{45}$,
S. Coquereau${}^{8}$,
G. Corti${}^{37}$,
B. Couturier${}^{37}$,
G.A. Cowan${}^{49}$,
D.C. Craik${}^{47}$,
S. Cunliffe${}^{52}$,
R. Currie${}^{49}$,
C. D’Ambrosio${}^{37}$,
P. David${}^{8}$,
P.N.Y. David${}^{40}$,
A. Davis${}^{56}$,
I. De Bonis${}^{4}$,
K. De Bruyn${}^{40}$,
S. De Capua${}^{53}$,
M. De Cian${}^{11}$,
J.M. De Miranda${}^{1}$,
L. De Paula${}^{2}$,
W. De Silva${}^{56}$,
P. De Simone${}^{18}$,
D. Decamp${}^{4}$,
M. Deckenhoff${}^{9}$,
L. Del Buono${}^{8}$,
N. Déléage${}^{4}$,
D. Derkach${}^{54}$,
O. Deschamps${}^{5}$,
F. Dettori${}^{41}$,
A. Di Canto${}^{11}$,
H. Dijkstra${}^{37}$,
M. Dogaru${}^{28}$,
S. Donleavy${}^{51}$,
F. Dordei${}^{11}$,
A. Dosil Suárez${}^{36}$,
D. Dossett${}^{47}$,
A. Dovbnya${}^{42}$,
F. Dupertuis${}^{38}$,
P. Durante${}^{37}$,
R. Dzhelyadin${}^{34}$,
A. Dziurda${}^{25}$,
A. Dzyuba${}^{29}$,
S. Easo${}^{48,37}$,
U. Egede${}^{52}$,
V. Egorychev${}^{30}$,
S. Eidelman${}^{33}$,
D. van Eijk${}^{40}$,
S. Eisenhardt${}^{49}$,
U. Eitschberger${}^{9}$,
R. Ekelhof${}^{9}$,
L. Eklund${}^{50,37}$,
I. El Rifai${}^{5}$,
Ch. Elsasser${}^{39}$,
A. Falabella${}^{14,e}$,
C. Färber${}^{11}$,
G. Fardell${}^{49}$,
C. Farinelli${}^{40}$,
S. Farry${}^{51}$,
V. Fave${}^{38}$,
D. Ferguson${}^{49}$,
V. Fernandez Albor${}^{36}$,
F. Ferreira Rodrigues${}^{1}$,
M. Ferro-Luzzi${}^{37}$,
S. Filippov${}^{32}$,
M. Fiore${}^{16}$,
C. Fitzpatrick${}^{37}$,
M. Fontana${}^{10}$,
F. Fontanelli${}^{19,i}$,
R. Forty${}^{37}$,
O. Francisco${}^{2}$,
M. Frank${}^{37}$,
C. Frei${}^{37}$,
M. Frosini${}^{17,f}$,
S. Furcas${}^{20}$,
E. Furfaro${}^{23,k}$,
A. Gallas Torreira${}^{36}$,
D. Galli${}^{14,c}$,
M. Gandelman${}^{2}$,
P. Gandini${}^{58}$,
Y. Gao${}^{3}$,
J. Garofoli${}^{58}$,
P. Garosi${}^{53}$,
J. Garra Tico${}^{46}$,
L. Garrido${}^{35}$,
C. Gaspar${}^{37}$,
R. Gauld${}^{54}$,
E. Gersabeck${}^{11}$,
M. Gersabeck${}^{53}$,
T. Gershon${}^{47,37}$,
Ph. Ghez${}^{4}$,
V. Gibson${}^{46}$,
L. Giubega${}^{28}$,
V.V. Gligorov${}^{37}$,
C. Göbel${}^{59}$,
D. Golubkov${}^{30}$,
A. Golutvin${}^{52,30,37}$,
A. Gomes${}^{2}$,
H. Gordon${}^{54}$,
M. Grabalosa Gándara${}^{5}$,
R. Graciani Diaz${}^{35}$,
L.A. Granado Cardoso${}^{37}$,
E. Graugés${}^{35}$,
G. Graziani${}^{17}$,
A. Grecu${}^{28}$,
E. Greening${}^{54}$,
S. Gregson${}^{46}$,
P. Griffith${}^{44}$,
O. Grünberg${}^{60}$,
B. Gui${}^{58}$,
E. Gushchin${}^{32}$,
Yu. Guz${}^{34,37}$,
T. Gys${}^{37}$,
C. Hadjivasiliou${}^{58}$,
G. Haefeli${}^{38}$,
C. Haen${}^{37}$,
S.C. Haines${}^{46}$,
S. Hall${}^{52}$,
B. Hamilton${}^{57}$,
T. Hampson${}^{45}$,
S. Hansmann-Menzemer${}^{11}$,
N. Harnew${}^{54}$,
S.T. Harnew${}^{45}$,
J. Harrison${}^{53}$,
T. Hartmann${}^{60}$,
J. He${}^{37}$,
T. Head${}^{37}$,
V. Heijne${}^{40}$,
K. Hennessy${}^{51}$,
P. Henrard${}^{5}$,
J.A. Hernando Morata${}^{36}$,
E. van Herwijnen${}^{37}$,
A. Hicheur${}^{1}$,
E. Hicks${}^{51}$,
D. Hill${}^{54}$,
M. Hoballah${}^{5}$,
C. Hombach${}^{53}$,
P. Hopchev${}^{4}$,
W. Hulsbergen${}^{40}$,
P. Hunt${}^{54}$,
T. Huse${}^{51}$,
N. Hussain${}^{54}$,
D. Hutchcroft${}^{51}$,
D. Hynds${}^{50}$,
V. Iakovenko${}^{43}$,
M. Idzik${}^{26}$,
P. Ilten${}^{12}$,
R. Jacobsson${}^{37}$,
A. Jaeger${}^{11}$,
E. Jans${}^{40}$,
P. Jaton${}^{38}$,
A. Jawahery${}^{57}$,
F. Jing${}^{3}$,
M. John${}^{54}$,
D. Johnson${}^{54}$,
C.R. Jones${}^{46}$,
C. Joram${}^{37}$,
B. Jost${}^{37}$,
M. Kaballo${}^{9}$,
S. Kandybei${}^{42}$,
W. Kanso${}^{6}$,
M. Karacson${}^{37}$,
T.M. Karbach${}^{37}$,
I.R. Kenyon${}^{44}$,
T. Ketel${}^{41}$,
A. Keune${}^{38}$,
B. Khanji${}^{20}$,
O. Kochebina${}^{7}$,
I. Komarov${}^{38}$,
R.F. Koopman${}^{41}$,
P. Koppenburg${}^{40}$,
M. Korolev${}^{31}$,
A. Kozlinskiy${}^{40}$,
L. Kravchuk${}^{32}$,
K. Kreplin${}^{11}$,
M. Kreps${}^{47}$,
G. Krocker${}^{11}$,
P. Krokovny${}^{33}$,
F. Kruse${}^{9}$,
M. Kucharczyk${}^{20,25,j}$,
V. Kudryavtsev${}^{33}$,
T. Kvaratskheliya${}^{30,37}$,
V.N. La Thi${}^{38}$,
D. Lacarrere${}^{37}$,
G. Lafferty${}^{53}$,
A. Lai${}^{15}$,
D. Lambert${}^{49}$,
R.W. Lambert${}^{41}$,
E. Lanciotti${}^{37}$,
G. Lanfranchi${}^{18}$,
C. Langenbruch${}^{37}$,
T. Latham${}^{47}$,
C. Lazzeroni${}^{44}$,
R. Le Gac${}^{6}$,
J. van Leerdam${}^{40}$,
J.-P. Lees${}^{4}$,
R. Lefèvre${}^{5}$,
A. Leflat${}^{31}$,
J. Lefrançois${}^{7}$,
S. Leo${}^{22}$,
O. Leroy${}^{6}$,
T. Lesiak${}^{25}$,
B. Leverington${}^{11}$,
Y. Li${}^{3}$,
L. Li Gioi${}^{5}$,
M. Liles${}^{51}$,
R. Lindner${}^{37}$,
C. Linn${}^{11}$,
B. Liu${}^{3}$,
G. Liu${}^{37}$,
S. Lohn${}^{37}$,
I. Longstaff${}^{50}$,
J.H. Lopes${}^{2}$,
N. Lopez-March${}^{38}$,
H. Lu${}^{3}$,
D. Lucchesi${}^{21,q}$,
J. Luisier${}^{38}$,
H. Luo${}^{49}$,
F. Machefert${}^{7}$,
I.V. Machikhiliyan${}^{4,30}$,
F. Maciuc${}^{28}$,
O. Maev${}^{29,37}$,
S. Malde${}^{54}$,
G. Manca${}^{15,d}$,
G. Mancinelli${}^{6}$,
J. Maratas${}^{5}$,
U. Marconi${}^{14}$,
P. Marino${}^{22,s}$,
R. Märki${}^{38}$,
J. Marks${}^{11}$,
G. Martellotti${}^{24}$,
A. Martens${}^{8}$,
A. Martín Sánchez${}^{7}$,
M. Martinelli${}^{40}$,
D. Martinez Santos${}^{41}$,
D. Martins Tostes${}^{2}$,
A. Massafferri${}^{1}$,
R. Matev${}^{37}$,
Z. Mathe${}^{37}$,
C. Matteuzzi${}^{20}$,
E. Maurice${}^{6}$,
A. Mazurov${}^{16,32,37,e}$,
B. Mc Skelly${}^{51}$,
J. McCarthy${}^{44}$,
A. McNab${}^{53}$,
R. McNulty${}^{12}$,
B. Meadows${}^{56,54}$,
F. Meier${}^{9}$,
M. Meissner${}^{11}$,
M. Merk${}^{40}$,
D.A. Milanes${}^{8}$,
M.-N. Minard${}^{4}$,
J. Molina Rodriguez${}^{59}$,
S. Monteil${}^{5}$,
D. Moran${}^{53}$,
P. Morawski${}^{25}$,
A. Mordà${}^{6}$,
M.J. Morello${}^{22,s}$,
R. Mountain${}^{58}$,
I. Mous${}^{40}$,
F. Muheim${}^{49}$,
K. Müller${}^{39}$,
R. Muresan${}^{28}$,
B. Muryn${}^{26}$,
B. Muster${}^{38}$,
P. Naik${}^{45}$,
T. Nakada${}^{38}$,
R. Nandakumar${}^{48}$,
I. Nasteva${}^{1}$,
M. Needham${}^{49}$,
S. Neubert${}^{37}$,
N. Neufeld${}^{37}$,
A.D. Nguyen${}^{38}$,
T.D. Nguyen${}^{38}$,
C. Nguyen-Mau${}^{38,o}$,
M. Nicol${}^{7}$,
V. Niess${}^{5}$,
R. Niet${}^{9}$,
N. Nikitin${}^{31}$,
T. Nikodem${}^{11}$,
A. Nomerotski${}^{54}$,
A. Novoselov${}^{34}$,
A. Oblakowska-Mucha${}^{26}$,
V. Obraztsov${}^{34}$,
S. Oggero${}^{40}$,
S. Ogilvy${}^{50}$,
O. Okhrimenko${}^{43}$,
R. Oldeman${}^{15,d}$,
M. Orlandea${}^{28}$,
J.M. Otalora Goicochea${}^{2}$,
P. Owen${}^{52}$,
A. Oyanguren${}^{35}$,
B.K. Pal${}^{58}$,
A. Palano${}^{13,b}$,
M. Palutan${}^{18}$,
J. Panman${}^{37}$,
A. Papanestis${}^{48}$,
M. Pappagallo${}^{50}$,
C. Parkes${}^{53}$,
C.J. Parkinson${}^{52}$,
G. Passaleva${}^{17}$,
G.D. Patel${}^{51}$,
M. Patel${}^{52}$,
G.N. Patrick${}^{48}$,
C. Patrignani${}^{19,i}$,
C. Pavel-Nicorescu${}^{28}$,
A. Pazos Alvarez${}^{36}$,
A. Pellegrino${}^{40}$,
G. Penso${}^{24,l}$,
M. Pepe Altarelli${}^{37}$,
S. Perazzini${}^{14,c}$,
E. Perez Trigo${}^{36}$,
A. Pérez-Calero Yzquierdo${}^{35}$,
P. Perret${}^{5}$,
M. Perrin-Terrin${}^{6}$,
L. Pescatore${}^{44}$,
G. Pessina${}^{20}$,
K. Petridis${}^{52}$,
A. Petrolini${}^{19,i}$,
A. Phan${}^{58}$,
E. Picatoste Olloqui${}^{35}$,
B. Pietrzyk${}^{4}$,
T. Pilař${}^{47}$,
D. Pinci${}^{24}$,
S. Playfer${}^{49}$,
M. Plo Casasus${}^{36}$,
F. Polci${}^{8}$,
G. Polok${}^{25}$,
A. Poluektov${}^{47,33}$,
E. Polycarpo${}^{2}$,
A. Popov${}^{34}$,
D. Popov${}^{10}$,
B. Popovici${}^{28}$,
C. Potterat${}^{35}$,
A. Powell${}^{54}$,
J. Prisciandaro${}^{38}$,
A. Pritchard${}^{51}$,
C. Prouve${}^{7}$,
V. Pugatch${}^{43}$,
A. Puig Navarro${}^{38}$,
G. Punzi${}^{22,r}$,
W. Qian${}^{4}$,
J.H. Rademacker${}^{45}$,
B. Rakotomiaramanana${}^{38}$,
M.S. Rangel${}^{2}$,
I. Raniuk${}^{42}$,
N. Rauschmayr${}^{37}$,
G. Raven${}^{41}$,
S. Redford${}^{54}$,
M.M. Reid${}^{47}$,
A.C. dos Reis${}^{1}$,
S. Ricciardi${}^{48}$,
A. Richards${}^{52}$,
K. Rinnert${}^{51}$,
V. Rives Molina${}^{35}$,
D.A. Roa Romero${}^{5}$,
P. Robbe${}^{7}$,
D.A. Roberts${}^{57}$,
E. Rodrigues${}^{53}$,
P. Rodriguez Perez${}^{36}$,
S. Roiser${}^{37}$,
V. Romanovsky${}^{34}$,
A. Romero Vidal${}^{36}$,
J. Rouvinet${}^{38}$,
T. Ruf${}^{37}$,
F. Ruffini${}^{22}$,
H. Ruiz${}^{35}$,
P. Ruiz Valls${}^{35}$,
G. Sabatino${}^{24,k}$,
J.J. Saborido Silva${}^{36}$,
N. Sagidova${}^{29}$,
P. Sail${}^{50}$,
B. Saitta${}^{15,d}$,
V. Salustino Guimaraes${}^{2}$,
C. Salzmann${}^{39}$,
B. Sanmartin Sedes${}^{36}$,
M. Sannino${}^{19,i}$,
R. Santacesaria${}^{24}$,
C. Santamarina Rios${}^{36}$,
E. Santovetti${}^{23,k}$,
M. Sapunov${}^{6}$,
A. Sarti${}^{18,l}$,
C. Satriano${}^{24,m}$,
A. Satta${}^{23}$,
M. Savrie${}^{16,e}$,
D. Savrina${}^{30,31}$,
P. Schaack${}^{52}$,
M. Schiller${}^{41}$,
H. Schindler${}^{37}$,
M. Schlupp${}^{9}$,
M. Schmelling${}^{10}$,
B. Schmidt${}^{37}$,
O. Schneider${}^{38}$,
A. Schopper${}^{37}$,
M.-H. Schune${}^{7}$,
R. Schwemmer${}^{37}$,
B. Sciascia${}^{18}$,
A. Sciubba${}^{24}$,
M. Seco${}^{36}$,
A. Semennikov${}^{30}$,
K. Senderowska${}^{26}$,
I. Sepp${}^{52}$,
N. Serra${}^{39}$,
J. Serrano${}^{6}$,
P. Seyfert${}^{11}$,
M. Shapkin${}^{34}$,
I. Shapoval${}^{16,42}$,
P. Shatalov${}^{30}$,
Y. Shcheglov${}^{29}$,
T. Shears${}^{51,37}$,
L. Shekhtman${}^{33}$,
O. Shevchenko${}^{42}$,
V. Shevchenko${}^{30}$,
A. Shires${}^{52}$,
R. Silva Coutinho${}^{47}$,
M. Sirendi${}^{46}$,
T. Skwarnicki${}^{58}$,
N.A. Smith${}^{51}$,
E. Smith${}^{54,48}$,
J. Smith${}^{46}$,
M. Smith${}^{53}$,
M.D. Sokoloff${}^{56}$,
F.J.P. Soler${}^{50}$,
F. Soomro${}^{18}$,
D. Souza${}^{45}$,
B. Souza De Paula${}^{2}$,
B. Spaan${}^{9}$,
A. Sparkes${}^{49}$,
P. Spradlin${}^{50}$,
F. Stagni${}^{37}$,
S. Stahl${}^{11}$,
O. Steinkamp${}^{39}$,
S. Stevenson${}^{54}$,
S. Stoica${}^{28}$,
S. Stone${}^{58}$,
B. Storaci${}^{39}$,
M. Straticiuc${}^{28}$,
U. Straumann${}^{39}$,
V.K. Subbiah${}^{37}$,
L. Sun${}^{56}$,
S. Swientek${}^{9}$,
V. Syropoulos${}^{41}$,
M. Szczekowski${}^{27}$,
P. Szczypka${}^{38,37}$,
T. Szumlak${}^{26}$,
S. T’Jampens${}^{4}$,
M. Teklishyn${}^{7}$,
E. Teodorescu${}^{28}$,
F. Teubert${}^{37}$,
C. Thomas${}^{54}$,
E. Thomas${}^{37}$,
J. van Tilburg${}^{11}$,
V. Tisserand${}^{4}$,
M. Tobin${}^{38}$,
S. Tolk${}^{41}$,
D. Tonelli${}^{37}$,
S. Topp-Joergensen${}^{54}$,
N. Torr${}^{54}$,
E. Tournefier${}^{4,52}$,
S. Tourneur${}^{38}$,
M.T. Tran${}^{38}$,
M. Tresch${}^{39}$,
A. Tsaregorodtsev${}^{6}$,
P. Tsopelas${}^{40}$,
N. Tuning${}^{40}$,
M. Ubeda Garcia${}^{37}$,
A. Ukleja${}^{27}$,
D. Urner${}^{53}$,
A. Ustyuzhanin${}^{52,p}$,
U. Uwer${}^{11}$,
V. Vagnoni${}^{14}$,
G. Valenti${}^{14}$,
A. Vallier${}^{7}$,
M. Van Dijk${}^{45}$,
R. Vazquez Gomez${}^{18}$,
P. Vazquez Regueiro${}^{36}$,
C. Vázquez Sierra${}^{36}$,
S. Vecchi${}^{16}$,
J.J. Velthuis${}^{45}$,
M. Veltri${}^{17,g}$,
G. Veneziano${}^{38}$,
M. Vesterinen${}^{37}$,
B. Viaud${}^{7}$,
D. Vieira${}^{2}$,
X. Vilasis-Cardona${}^{35,n}$,
A. Vollhardt${}^{39}$,
D. Volyanskyy${}^{10}$,
D. Voong${}^{45}$,
A. Vorobyev${}^{29}$,
V. Vorobyev${}^{33}$,
C. Voß${}^{60}$,
H. Voss${}^{10}$,
R. Waldi${}^{60}$,
C. Wallace${}^{47}$,
R. Wallace${}^{12}$,
S. Wandernoth${}^{11}$,
J. Wang${}^{58}$,
D.R. Ward${}^{46}$,
N.K. Watson${}^{44}$,
A.D. Webber${}^{53}$,
D. Websdale${}^{52}$,
M. Whitehead${}^{47}$,
J. Wicht${}^{37}$,
J. Wiechczynski${}^{25}$,
D. Wiedner${}^{11}$,
L. Wiggers${}^{40}$,
G. Wilkinson${}^{54}$,
M.P. Williams${}^{47,48}$,
M. Williams${}^{55}$,
F.F. Wilson${}^{48}$,
J. Wimberley${}^{57}$,
J. Wishahi${}^{9}$,
M. Witek${}^{25}$,
S.A. Wotton${}^{46}$,
S. Wright${}^{46}$,
S. Wu${}^{3}$,
K. Wyllie${}^{37}$,
Y. Xie${}^{49,37}$,
Z. Xing${}^{58}$,
Z. Yang${}^{3}$,
R. Young${}^{49}$,
X. Yuan${}^{3}$,
O. Yushchenko${}^{34}$,
M. Zangoli${}^{14}$,
M. Zavertyaev${}^{10,a}$,
F. Zhang${}^{3}$,
L. Zhang${}^{58}$,
W.C. Zhang${}^{12}$,
Y. Zhang${}^{3}$,
A. Zhelezov${}^{11}$,
A. Zhokhov${}^{30}$,
L. Zhong${}^{3}$,
A. Zvyagin${}^{37}$.
${}^{1}$Centro Brasileiro de Pesquisas Físicas (CBPF), Rio de Janeiro, Brazil
${}^{2}$Universidade Federal do Rio de Janeiro (UFRJ), Rio de Janeiro, Brazil
${}^{3}$Center for High Energy Physics, Tsinghua University, Beijing, China
${}^{4}$LAPP, Université de Savoie, CNRS/IN2P3, Annecy-Le-Vieux, France
${}^{5}$Clermont Université, Université Blaise Pascal, CNRS/IN2P3, LPC, Clermont-Ferrand, France
${}^{6}$CPPM, Aix-Marseille Université, CNRS/IN2P3, Marseille, France
${}^{7}$LAL, Université Paris-Sud, CNRS/IN2P3, Orsay, France
${}^{8}$LPNHE, Université Pierre et Marie Curie, Université Paris Diderot, CNRS/IN2P3, Paris, France
${}^{9}$Fakultät Physik, Technische Universität Dortmund, Dortmund, Germany
${}^{10}$Max-Planck-Institut für Kernphysik (MPIK), Heidelberg, Germany
${}^{11}$Physikalisches Institut, Ruprecht-Karls-Universität Heidelberg, Heidelberg, Germany
${}^{12}$School of Physics, University College Dublin, Dublin, Ireland
${}^{13}$Sezione INFN di Bari, Bari, Italy
${}^{14}$Sezione INFN di Bologna, Bologna, Italy
${}^{15}$Sezione INFN di Cagliari, Cagliari, Italy
${}^{16}$Sezione INFN di Ferrara, Ferrara, Italy
${}^{17}$Sezione INFN di Firenze, Firenze, Italy
${}^{18}$Laboratori Nazionali dell’INFN di Frascati, Frascati, Italy
${}^{19}$Sezione INFN di Genova, Genova, Italy
${}^{20}$Sezione INFN di Milano Bicocca, Milano, Italy
${}^{21}$Sezione INFN di Padova, Padova, Italy
${}^{22}$Sezione INFN di Pisa, Pisa, Italy
${}^{23}$Sezione INFN di Roma Tor Vergata, Roma, Italy
${}^{24}$Sezione INFN di Roma La Sapienza, Roma, Italy
${}^{25}$Henryk Niewodniczanski Institute of Nuclear Physics Polish Academy of Sciences, Kraków, Poland
${}^{26}$AGH - University of Science and Technology, Faculty of Physics and Applied Computer Science, Kraków, Poland
${}^{27}$National Center for Nuclear Research (NCBJ), Warsaw, Poland
${}^{28}$Horia Hulubei National Institute of Physics and Nuclear Engineering, Bucharest-Magurele, Romania
${}^{29}$Petersburg Nuclear Physics Institute (PNPI), Gatchina, Russia
${}^{30}$Institute of Theoretical and Experimental Physics (ITEP), Moscow, Russia
${}^{31}$Institute of Nuclear Physics, Moscow State University (SINP MSU), Moscow, Russia
${}^{32}$Institute for Nuclear Research of the Russian Academy of Sciences (INR RAN), Moscow, Russia
${}^{33}$Budker Institute of Nuclear Physics (SB RAS) and Novosibirsk State University, Novosibirsk, Russia
${}^{34}$Institute for High Energy Physics (IHEP), Protvino, Russia
${}^{35}$Universitat de Barcelona, Barcelona, Spain
${}^{36}$Universidad de Santiago de Compostela, Santiago de Compostela, Spain
${}^{37}$European Organization for Nuclear Research (CERN), Geneva, Switzerland
${}^{38}$Ecole Polytechnique Fédérale de Lausanne (EPFL), Lausanne, Switzerland
${}^{39}$Physik-Institut, Universität Zürich, Zürich, Switzerland
${}^{40}$Nikhef National Institute for Subatomic Physics, Amsterdam, The Netherlands
${}^{41}$Nikhef National Institute for Subatomic Physics and VU University Amsterdam, Amsterdam, The Netherlands
${}^{42}$NSC Kharkiv Institute of Physics and Technology (NSC KIPT), Kharkiv, Ukraine
${}^{43}$Institute for Nuclear Research of the National Academy of Sciences (KINR), Kyiv, Ukraine
${}^{44}$University of Birmingham, Birmingham, United Kingdom
${}^{45}$H.H. Wills Physics Laboratory, University of Bristol, Bristol, United Kingdom
${}^{46}$Cavendish Laboratory, University of Cambridge, Cambridge, United Kingdom
${}^{47}$Department of Physics, University of Warwick, Coventry, United Kingdom
${}^{48}$STFC Rutherford Appleton Laboratory, Didcot, United Kingdom
${}^{49}$School of Physics and Astronomy, University of Edinburgh, Edinburgh, United Kingdom
${}^{50}$School of Physics and Astronomy, University of Glasgow, Glasgow, United Kingdom
${}^{51}$Oliver Lodge Laboratory, University of Liverpool, Liverpool, United Kingdom
${}^{52}$Imperial College London, London, United Kingdom
${}^{53}$School of Physics and Astronomy, University of Manchester, Manchester, United Kingdom
${}^{54}$Department of Physics, University of Oxford, Oxford, United Kingdom
${}^{55}$Massachusetts Institute of Technology, Cambridge, MA, United States
${}^{56}$University of Cincinnati, Cincinnati, OH, United States
${}^{57}$University of Maryland, College Park, MD, United States
${}^{58}$Syracuse University, Syracuse, NY, United States
${}^{59}$Pontifícia Universidade Católica do Rio de Janeiro (PUC-Rio), Rio de Janeiro, Brazil, associated to ${}^{2}$
${}^{60}$Institut für Physik, Universität Rostock, Rostock, Germany, associated to ${}^{11}$
${}^{a}$P.N. Lebedev Physical Institute, Russian Academy of Science (LPI RAS), Moscow, Russia
${}^{b}$Università di Bari, Bari, Italy
${}^{c}$Università di Bologna, Bologna, Italy
${}^{d}$Università di Cagliari, Cagliari, Italy
${}^{e}$Università di Ferrara, Ferrara, Italy
${}^{f}$Università di Firenze, Firenze, Italy
${}^{g}$Università di Urbino, Urbino, Italy
${}^{h}$Università di Modena e Reggio Emilia, Modena, Italy
${}^{i}$Università di Genova, Genova, Italy
${}^{j}$Università di Milano Bicocca, Milano, Italy
${}^{k}$Università di Roma Tor Vergata, Roma, Italy
${}^{l}$Università di Roma La Sapienza, Roma, Italy
${}^{m}$Università della Basilicata, Potenza, Italy
${}^{n}$LIFAELS, La Salle, Universitat Ramon Llull, Barcelona, Spain
${}^{o}$Hanoi University of Science, Hanoi, Viet Nam
${}^{p}$Institute of Physics and Technology, Moscow, Russia
${}^{q}$Università di Padova, Padova, Italy
${}^{r}$Università di Pisa, Pisa, Italy
${}^{s}$Scuola Normale Superiore, Pisa, Italy
A search for the lepton-flavour violating decays $B^{0}_{s}\to e^{\pm}\mu^{\mp}$ and $B^{0}\to e^{\pm}\mu^{\mp}$ is
performed with a data sample, corresponding to an integrated luminosity
of 1.0 $\mbox{\,fb}^{-1}$ of $pp$ collisions at $\sqrt{s}=7$ TeV, collected by the LHCb experiment.
The observed number of $B^{0}_{s}\to e^{\pm}\mu^{\mp}$ and $B^{0}\to e^{\pm}\mu^{\mp}$ candidates is consistent with background expectations.
Upper limits on the branching fractions of both decays are determined to be ${\cal B}(B^{0}_{s}\to e^{\pm}\mu^{\mp})<1.1\,(1.4)\times 10^{-8}$ and
${\cal B}(B^{0}\to e^{\pm}\mu^{\mp})<2.8\,(3.7)\times 10^{-9}$ at 90% (95%) confidence level (C.L.).
These limits are a factor of twenty lower than those set by previous experiments. Lower bounds on the Pati-Salam leptoquark masses
are also calculated, $M_{\rm LQ}(B^{0}_{s}\to e^{\pm}\mu^{\mp})>107$ ${\mathrm{\,Te\kern-1.0ptV\!/}c^{2}}$ and $M_{\rm LQ}(B^{0}\to e^{\pm}\mu^{\mp})>126$ ${\mathrm{\,Te\kern-1.0ptV\!/}c^{2}}$ at 95% C.L.,
and are a factor of two higher than the previous bounds.
Submitted to Physical Review Letters
Rare decays that are forbidden in the Standard Model (SM) probe potential contributions from new processes and
particles at a scale beyond the reach of direct searches. The decays $B^{0}_{s}\to e^{\pm}\mu^{\mp}$ and $B^{0}\to e^{\pm}\mu^{\mp}$ and
their charged conjugate processes111Inclusion of charge conjugate processes are implied throughout this Letter.
are forbidden within the SM, in which lepton flavour is conserved.
These decays are allowed in some scenarios beyond the SM that include models
with heavy singlet Dirac neutrinos ilakovic ,
supersymmetric models susy and the Pati-Salam model patisalam .
The latter predicts a new interaction to mediate transitions between
leptons and quarks via exchange of spin$-$1 gauge bosons, called Pati-Salam leptoquarks (LQ), that carry both colour and lepton
quantum numbers.
Current limits from ATLAS atlas1 ; atlas2 ; atlas3 and CMS cms1 ; cms2 ; cms3 on the masses of first, second or
third generation leptoquarks are in the range $[0.4,0.9]$ ${\mathrm{\,Te\kern-1.0ptV\!/}c^{2}}$, depending on the value of the couplings and the decay channel.
These leptoquarks arise from a coupling between a quark and lepton of the same generation.
The decays $B^{0}_{s}\to e^{\pm}\mu^{\mp}$ and $B^{0}\to e^{\pm}\mu^{\mp}$ can be mediated by other leptoquarks which couple leptons and quarks that
are not necessarily
from the same generation valencia ; blanke , such as when the $\tau$ lepton couples to a first or second quark generation.
The previous best upper limits on the branching fraction of these decays come from the CDF collaboration bemu_cdf ,
${\cal B}(B^{0}_{s}\to e^{\pm}\mu^{\mp})$ $<2.0\;(2.6)\times 10^{-7}$ and ${\cal B}(B^{0}\to e^{\pm}\mu^{\mp})$ $<6.4\;(7.9)\times 10^{-8}$ at 90% (95%) confidence level (C.L.).
These limits correspond to bounds on the masses of the corresponding Pati-Salam
leptoquarks of $M_{\rm LQ}(B^{0}_{s}\to e^{\pm}\mu^{\mp})>47.8\,(44.9)$ ${\mathrm{\,Te\kern-1.0ptV\!/}c^{2}}$ and $M_{\rm LQ}(B^{0}\to e^{\pm}\mu^{\mp})>59.3\,(56.3)$ ${\mathrm{\,Te\kern-1.0ptV\!/}c^{2}}$ at 90 (95)% C.L. bemu_cdf .
This Letter presents a search for the $B^{0}_{s}\to e^{\pm}\mu^{\mp}$ and $B^{0}\to e^{\pm}\mu^{\mp}$ lepton-flavour violating (LFV) decays performed
with a data sample, corresponding to an integrated luminosity
of 1.0 $\mbox{\,fb}^{-1}$ of $pp$ collisions at $\sqrt{s}=7$ TeV, collected by the LHCb experiment in 2011 at the Large Hadron Collider.
To avoid any potential bias, events in the signal mass region $[5.1,5.5]\,{\mathrm{\,Ge\kern-1.0ptV\!/}c^{2}}$
were not examined until all analysis choices were finalized.
The LHCb detector is a single-arm forward spectrom-
eter covering the pseudorapidity range $2<\eta<5$, and is
described in detail in Ref. LHCbdetector .
Events were simulated for this analysis using the software described
in Refs. Sjostrand:2006za ; Lange:2001uf ; Allison:2006ve ; Agostinelli:2002hh ; Golonka:2005pn ; LHCb-PROC-2011-005 ; LHCb-PROC-2011-006 .
The trigger trigger consists of a hardware stage (L0), based on information from the calorimeter
and muon systems, followed by a
software stage (HLT) that applies a full event reconstruction, and is split into two stages called HLT1 and HLT2.
Candidate $B^{0}_{(s)}\to e^{\pm}\mu^{\mp}$ decays considered in this analysis must satisfy a hardware decision that
requires the presence of a muon candidate with transverse momentum
$p_{\rm T}>1.5$${\mathrm{\,Ge\kern-1.0ptV\!/}c}$.
All tracks considered in the HLT1 are required to have $p_{\rm T}>0.5$${\mathrm{\,Ge\kern-1.0ptV\!/}c}$.
The muon track of the $B^{0}_{(s)}\to e^{\pm}\mu^{\mp}$ candidates is required to have $p_{\rm T}>1.0$${\mathrm{\,Ge\kern-1.0ptV\!/}c}$ and impact parameter, ${\rm IP}$ $>0.1$ mm.
The HLT2 consists of
exclusive, cut-based triggers for $B^{0}_{(s)}$ two-body decays, and inclusive multivariate trigger ; BBDT $b$-hadron triggers.
The $B^{0}\to K^{+}\pi^{-}$ decay is used as the normalization channel and $B^{0}_{(s)}\to h^{+}{h}^{\prime-}$ ($h^{(\prime)}=K,\pi$) decays are used as a control channel,
since both have the same event topology as the signal.
The $B^{0}\to K^{+}\pi^{-}$ yield is computed from the yield of $B^{0}_{(s)}\to h^{+}{h}^{\prime-}$ decays, and the
fraction of $B^{0}\to K^{+}\pi^{-}$ in the $B^{0}_{(s)}\to h^{+}{h}^{\prime-}$ sample, as described in Ref. roadmap .
In order to minimize the bias introduced by the trigger requirements,
only $B^{0}_{(s)}\to h^{+}{h}^{\prime-}$ candidates that are triggered independently of the presence of either of the two signal hadrons at L0 and HLT1
are considered.
The $B^{0}_{(s)}\to e^{\pm}\mu^{\mp}$ candidates that pass the trigger selection criteria are further required to have well identified electron and
muon muonid candidates.
The measured momenta of the electrons are corrected to account for loss of momentum by bremsstrahlung in the detector using the photon
energy deposition in the electromagnetic calorimeter Kstee .
The signal candidates are required to be displaced with respect to any $pp$ collision vertex (PV), and
form a secondary vertex (SV) with $\chi^{2}$ per degree of freedom smaller than 9 and separated from the PV in the downstream direction by a flight distance significance greater than 15.
Only $B^{0}_{(s)}$ candidates with an impact parameter $\chi^{2}$ ($\chi^{2}_{\rm IP}$) less than 25 are considered.
The $\chi^{2}_{\rm IP}$ of a $B^{0}_{(s)}$ candidate
is defined as the difference between the $\chi^{2}$
of the PV reconstructed with and without the tracks forming the candidate.
When more than one PV is reconstructed, that giving the smallest
$\chi^{2}_{\rm IP}$ for the $B^{0}_{(s)}$ candidate is chosen.
Only $B^{0}_{(s)}$ candidates
with invariant mass in the range $[4.9,5.9]{\mathrm{\,Ge\kern-1.0ptV\!/}c^{2}}$ are kept for further analysis.
The selection criteria for the $B^{0}_{(s)}\to h^{+}{h}^{\prime-}$ and $B^{0}\to K^{+}\pi^{-}$ candidates are identical to those of the signal, apart from those used for particle identification.
A two-stage multivariate selection based on boosted decision trees (BDT) Breiman ; AdaBoost
is applied to the $B^{0}_{(s)}\to e^{\pm}\mu^{\mp}$ candidates following the same strategy as Ref. bsmumu .
The two multivariate discriminants are trained using simulated samples,
$B^{0}_{s}\to e^{\pm}\mu^{\mp}$ for signal and $b\bar{b}\to l^{\pm}l^{\prime\mp}X$ for background (where $l^{(\prime)}$ can either be a $\mu$ or an $e$ and
$X$ is any other set of particles),
which is dominated by simultaneous semileptonic decays of both $b$ and $\overline{b}$ hadrons within the same event.
The six variables entering the first multivariate discriminant, in decreasing order of their background rejection power, are:
the angle between the momentum of the $B^{0}_{(s)}$ candidate and the direction
defined by the vector joining the secondary and primary vertices;
the $B$ candidate IP and $\chi^{2}_{\rm IP}$; the minimum IP of the two daughter tracks with respect to any PV;
the distance of closest approach between the two daughter tracks and the $\chi^{2}$ of the SV.
The requirement on the first multivariate discriminant
removes 75 % of the background while retaining 93 % of signal, as determined from simulation.
The same selection is applied to the $B^{0}\to K^{+}\pi^{-}$ normalization channel
and the efficiencies of this requirement for the signal and the normalization channel are
equal within 0.2 %.
The surviving background mainly comprises random combinations of electrons and muons from semileptonic
$b\bar{b}\to e^{\pm}\mu^{\mp}X$ decays.
In total 5766 electron-muon pairs pass the trigger and offline selection requirements.
The selected candidates are classified in a binned two-dimensional space formed by the electron-muon invariant mass
and the output of a second BDT, for which nine variables are employed.
They are, in decreasing order of their background rejection power, the $B^{0}_{(s)}$ candidate IP and $\chi^{2}_{\rm IP}$,
the number of good two-track vertices a daughter track can make with other tracks
in the event, the $B^{0}_{(s)}$ candidate decay time, its $p_{\rm T}$
and degree of isolation cdf_iso ,
the distance of closest approach between the two tracks,
the minimum $p_{\rm T}$ of the tracks,
and the cosine of the angle between the muon momentum in the muon-electron
rest frame and the vector perpendicular to the $B$ candidate
momentum and to the beam axis.
The BDT output is independent of the invariant mass for signal
inside the search window. The output is transformed such that the signal is approximately uniformly
distributed between zero and one, while the background peaks at zero.
The probability for a signal event to have a given BDT value is
obtained from data using the $B^{0}_{(s)}\to h^{+}{h}^{\prime-}$ sample bsmumu_plb .
Simulated samples of $B^{0}_{(s)}\to e^{\pm}\mu^{\mp}$ and $B^{0}_{(s)}\to h^{+}{h}^{\prime-}$ decays have been used to check that the distributions of the variables entering in the BDT
that do not depend on the bremsstrahlung radiation are in good agreement. Corrections to the BDT shape due to the presence
of the radiation emitted by the electron of the $B^{0}_{(s)}\to e^{\pm}\mu^{\mp}$ decays have been evaluated using simulation.
The number of $B^{0}_{(s)}\to h^{+}{h}^{\prime-}$ signal events in each BDT bin is determined by
fitting the $h^{+}h^{\prime-}$ invariant mass distribution LHCbnote:carbone .
The systematic uncertainty on the signal BDT probability distribution function is taken to be the maximum spread
in the fractions of yields going into each bin, obtained by fitting the same $B^{0}_{(s)}\to h^{+}{h}^{\prime-}$ dataset with different signal and background fit models.
Corrections are applied to the BDT shape in order to take into account the effect of the
different trigger requirements used for the signal and the $B^{0}_{(s)}\to h^{+}{h}^{\prime-}$ control sample.
The invariant mass line shape of the signal events is described
by a Crystal Ball function (CB) crystalball with two tails, left and right,
defined by two parameters each.The values of the parameters
depend on the momentum resolution,
the momentum scale and the amount of bremsstrahlung radiation recovered.
The signal shape parameters are obtained from simulation, but need to be reweighted to account
for their dependency on the event multiplicity,
which affects the amount of bremsstrahlung radiation recovered and differs between data and simulation.
We use the number of hits in the scintillating pad detector ($N_{\rm SPD}$) as a measure of the event multiplicity. The distribution of
$N_{\rm SPD}$ for $B^{0}_{(s)}\to e^{\pm}\mu^{\mp}$ signal
candidates is obtained from a $B^{+}\to J/\psi(\mu^{+}\mu^{-})K^{+}$ data sample, which is selected with the same trigger conditions as the signal,
ensuring a similar distribution of $N_{\rm SPD}$.
The signal mass shape parameters are determined by reweighting the $B^{0}_{(s)}\to e^{\pm}\mu^{\mp}$ simulated events with the
$N_{\rm SPD}$ distribution measured in the $B^{+}\to J/\psi(\mu^{+}\mu^{-})K^{+}$ sample.
This reweighting technique is used also for a $J/\psi\to e^{+}e^{-}$ simulated sample and the reweighted parameters are then compared
with those obtained with a $J/\psi\to e^{+}e^{-}$ sample in data.
The difference between the mean values of the $J/\psi\to e^{+}e^{-}$ mass in data and simulation (+0.16%) is applied as a
systematic shift to the peak values of the $B^{0}\to e^{\pm}\mu^{\mp}$ and $B^{0}_{s}\to e^{\pm}\mu^{\mp}$ invariant mass in simulation. A systematic uncertainty is added to the $B^{0}_{(s)}\to e^{\pm}\mu^{\mp}$ mass parameters when the differences in the values of the other mass parameters for the $J/\psi\to e^{+}e^{-}$ sample in data and SPD-reweighted simulation
are larger than their statistical uncertainties.
The signal region, defined by the invariant mass window $[5.1,5.5]$${\mathrm{\,Ge\kern-1.0ptV\!/}c^{2}}$, retains
$(85.0\pm 0.1_{\rm stat}\pm 5.0_{\rm syst})\%$ and $(82.0\pm 0.1_{\rm stat}\pm 5.0_{\rm syst})\%$
of the $B^{0}_{s}\to e^{\pm}\mu^{\mp}$ and $B^{0}\to e^{\pm}\mu^{\mp}$ signal decays, respectively. The systematic uncertainties on these fractions are evaluated
with pseudo-experiments that fluctuate each parameter of the mass lineshape according to its uncertainty.
The width of the corresponding fraction distribution is taken as the systematic uncertainty.
The $B^{0}_{s}\to e^{\pm}\mu^{\mp}$ and $B^{0}\to e^{\pm}\mu^{\mp}$ yields are translated
into branching fractions according to
$$\displaystyle{\cal B}(B^{0}_{(s)}\to e^{\pm}\mu^{\mp})$$
$$\displaystyle=$$
$$\displaystyle\frac{{\cal B}_{\rm norm}\,{\rm\epsilon_{\rm norm}}\,f_{d}}{N_{%
\rm norm}\,{\rm\epsilon_{sig}}\,f_{d(s)}}\times N_{B^{0}_{(s)}\to e^{\pm}\mu^{%
\mp}}$$
(1)
$$\displaystyle=$$
$$\displaystyle\alpha_{B^{0}_{(s)}}\times N_{B^{0}_{(s)}\to e^{\pm}\mu^{\mp}},$$
where $N_{\rm norm}=10\,120\pm 920$ is the number of signal events in the normalization
channel, and the uncertainty is the sum in quadrature of the statistical and systematic uncertainties.
The branching fraction of the normalization channel is ${\cal B}_{\rm norm}$ $=(1.94\pm 0.96)\times 10^{-5}$ PDG2012
and $N_{B^{0}_{(s)}\to e^{\pm}\mu^{\mp}}$ is the number of observed signal events.
The factors $f_{d}$ and $f_{s}$ indicate the probabilities
that a $b$ quark fragments into a $B^{0}$ or $B^{0}_{s}$ meson, respectively.
We use $f_{s}/f_{d}=0.256\pm 0.020$
measured in $pp$ collision data at $\sqrt{s}=7$ TeV LHCb-PAPER-2012-037 .
The measured dependence of $f_{s}/f_{d}$ on the $B$ meson $p_{\rm T}$ LHCb-PAPER-2012-037
is found to be negligible for this analysis.
The efficiency ${\rm\epsilon_{sig(norm)}}$ for the signal (normalization) channel is
the product of the reconstruction efficiency of the final state particles
including the geometric detector acceptance,
the selection efficiency and the trigger efficiency.
The ratios of acceptance, reconstruction and selection efficiencies are computed with
simulation. A systematic uncertainty is assigned to these ratios, to take into account the difference between the tracking efficiencies measured in data and predicted in simulation.
Reweighting techniques are used to correct distributions in the simulation that do not match
those from data, in particular for those variables that depend on $N_{\rm SPD}$.
The trigger efficiency of L0 and HLT1 on signal decays is evaluated using data, while the HLT2 efficiency is evaluated
in simulation after validation with control samples.
The electron and muon identification efficiencies are evaluated from data using the $B^{+}\to J/\psi(\mu^{+}\mu^{-})K^{+}$ and $B^{+}\to J/\psi(e^{+}e^{-})K^{+}$ control samples.
The two normalization factors $\alpha_{B^{0}_{s}}$ and $\alpha_{B^{0}}$ are determined to be
$\alpha_{B^{0}_{s}}=(1.1\pm 0.2)\times 10^{-9}$ and
$\alpha_{B^{0}}=(2.8\pm 0.5)\times 10^{-10}$.
The BDT range is divided into eight bins with boundaries at $0.0,0.25,0.4,0.5,0.6,0.7,0.8,0.9$ and 1.0.
The number of expected combinatorial background events in each BDT bin and the invariant mass signal region
is determined from data by fitting to an exponential function events in
the mass sidebands, defined by $[4.9,5.0]{\mathrm{\,Ge\kern-1.0ptV\!/}c^{2}}$ and $[5.5,5.9]{\mathrm{\,Ge\kern-1.0ptV\!/}c^{2}}$.
The BDT output for signal and combinatorial background events is shown in Fig. 1.
In the exponential function both the slope and the normalization are allowed to vary.
The systematic uncertainty on the estimated number of combinatorial
background events in the signal regions is determined
by fluctuating the number of events observed in the sidebands according to a Poisson distribution,
and by varying the exponential slope according to its uncertainty.
As a cross-check, two other models,
the sum of two exponential functions and a single exponential fitted to the right sideband only,
have been used and provide consistent background estimates inside the signal region.
The low-mass sideband and the signal region are potentially polluted by exclusive backgrounds.
The background from $B^{+}_{c}\to J/\psi(\mu^{+}\mu^{-})e^{+}\nu_{e}$ and
$B^{+}_{c}\to J/\psi(e^{+}e^{-})\mu^{+}\nu_{\mu}$
decays is evaluated assuming the branching fraction
value from Ref. Abe:1998wi .
The decays $B^{0}\to\pi^{-}l^{+}\nu_{l}$, $B^{0}_{(s)}\to h^{+}{h}^{\prime-}$,
$B^{0}_{s}\to K^{-}l^{+}\nu_{l}$, $\mathchar 28931\relax_{b}^{0}\to pl^{-}\overline{\nu}_{l}$ and $B^{+}\to\pi^{+}l^{+}l^{-}$ (where $l^{\pm}=e^{\pm}$ or $\mu^{\pm}$)
are potential backgrounds if the hadrons are misidentified as electrons or muons.
The $B^{0}\to\pi^{-}l^{+}\nu_{l}$ and $B^{0}_{(s)}\to h^{+}{h}^{\prime-}$ branching fractions are taken from Ref. PDG2012 .
The $B^{+}\to\pi^{+}l^{+}l^{-}$ branching fraction is taken from Ref. Bpimumu .
The theoretical estimates of the $\mathchar 28931\relax_{b}^{0}\to pl^{-}\overline{\nu}_{l}$ and $B^{0}_{s}\to K^{-}l^{+}\nu_{l}$
branching fractions are taken from Refs. Wang and BsKmunu , respectively.
We use the $\mathchar 28931\relax_{b}^{0}$ fragmentation fraction $f_{\mathchar 28931\relax_{b}^{0}}$ measured by LHCb Aaij:2011jp
and account for its $p_{\rm T}$ dependence.
The mass and BDT distributions of these background modes are evaluated from simulated samples, using the probabilities of misidentifying kaon, pion and proton as muon or electron as functions of momenta and transverse
momenta, which are determined from $D^{*+}\to D^{0}(\to K^{-}\pi^{+})\pi^{+}$ and $\mathchar 28931\relax\to p\pi^{-}$
data samples. The mass lineshape of the $B^{0}_{(s)}\to h^{+}h^{{}^{\prime}-}\to e^{+}\mu^{-}$ peaking background is obtained from a simulated
sample of doubly-misidentified $B^{0}_{(s)}\to h^{+}{h}^{\prime-}$ events.
Apart from $B^{0}_{(s)}\to h^{+}{h}^{\prime-}$, all background modes are
normalized relative to the $B^{+}\to J/\psi(\mu^{+}\mu^{-})K^{+}$ decay. We assume $f_{u}=f_{d}$ where $f_{u}$ is the $B^{+}$ fragmentation fraction.
The $\mathchar 28931\relax_{b}^{0}\to pl^{-}\overline{\nu}_{l}$ and the $B^{+}_{c}\to J/\psi(\mu^{+}\mu^{-})e^{+}\nu_{e}$ and
$B^{+}_{c}\to J/\psi(e^{+}e^{-})\mu^{+}\nu_{\mu}$ modes are the dominant exclusive modes in the range
BDT$>0.5$, where the combinatorial background is reduced by a factor $\sim 500$ according to simulation.
These decay modes have an invariant mass distribution
that is compatible with an exponential in the region [4.9-5.9] ${\mathrm{\,Ge\kern-1.0ptV\!/}c^{2}}$, and hence are taken into account by the exponential
fit to the mass sidebands.
In the entire BDT and mass range ($[4.9,5.9]{\mathrm{\,Ge\kern-1.0ptV\!/}c^{2}}$), $4.4\pm 0.7$
doubly misidentified $B^{0}_{(s)}\to h^{+}{h}^{\prime-}$ decays are expected, with $(87.9\pm 0.1)\%$ lying in the signal mass interval of $[5.1,5.5]{\mathrm{\,Ge\kern-1.0ptV\!/}c^{2}}.$
The distribution in the BDT output is almost uniform, therefore $1.1\pm 0.3$ events of this background are expected in the last two BDT bins, to be compared
to $4.8^{+1.0}_{-0.7}$ background events obtained from the exponential fit
in the same mass and BDT ranges.
However, since this background peaks in the signal region, its BDT and mass distributions are accounted for in the limit computation.
For each BDT bin we count the number of candidates observed in the signal region,
and compare to the expected number of signal and background events.
The distribution of the observed candidates is shown in Fig. 2.
The systematic uncertainties in the background and signal predictions in each bin
are computed by varying the normalization factor, and the mass and BDT shapes
within their Gaussian uncertainties.
The results for the $B^{0}_{s}\to e^{\pm}\mu^{\mp}$ and $B^{0}\to e^{\pm}\mu^{\mp}$ decays
are summarized in Table 1. In the high BDT range, the observed number of candidates is in agreement with the
number of expected exclusive backgrounds in the signal region.
The compatibility of the observed distribution of events
with that expected for a given branching fraction
hypothesis is computed with the $\textrm{CL}_{\textrm{s}}$ method Read_02 .
The method provides $\textrm{CL}_{\textrm{s+b}}$, a measure of the
compatibility of the observed distribution with the signal plus background
hypothesis, $\textrm{CL}_{\textrm{b}}$, a measure of the compatibility with the background-only
hypothesis, and $\textrm{CL}_{\textrm{s}}=\textrm{CL}_{\textrm{s+b}}/\textrm{CL}_{\textrm{b}}$.
The expected and observed $\textrm{CL}_{\textrm{s}}$ values are shown in Fig. 3 for the $B^{0}_{s}\to e^{\pm}\mu^{\mp}$ and $B^{0}\to e^{\pm}\mu^{\mp}$ channels,
as a function of the assumed branching fraction.
The expected and measured limits for $B^{0}_{s}\to e^{\pm}\mu^{\mp}$ and $B^{0}\to e^{\pm}\mu^{\mp}$ at 90 %
and 95 % C.L. are shown in Table 2.
Note that since the same events are used to set limits for both $B^{0}_{s}$ and $B^{0}$ decays, the results are strongly correlated.
The inclusion of systematic uncertainties increases the expected $B^{0}\to e^{\pm}\mu^{\mp}$ and $B^{0}_{s}\to e^{\pm}\mu^{\mp}$ upper limits by $\sim 20\%$.
The observed limits are $\sim 1\,\sigma$ below the expectation due to the lower than expected numbers of observed events in the
fourth and last BDT bins.
The comparison of the distributions of observed events and expected
background events results in a p-value $(1-\textrm{CL}_{\textrm{b}})$ of 18 % (17 %)
for the $B^{0}_{s}\to e^{\pm}\mu^{\mp}$ ($B^{0}\to e^{\pm}\mu^{\mp}$) decay, where the $\textrm{CL}_{\textrm{b}}$ values are those
corresponding to $\textrm{CL}_{\textrm{s+b}}=0.5$.
In the framework of the Pati-Salam model, the relation linking the $B^{0}_{(s)}\to e^{\pm}\mu^{\mp}$ branching fractions and the leptoquark mass
($M_{\rm LQ}$) valencia is
$${\cal B}(B^{0}_{(s)}\to e^{\pm}\mu^{\mp})=\pi{\alpha_{S}^{2}(M_{\rm LQ})\over M%
_{\rm LQ}^{4}}F^{2}_{B^{0}_{(s)}}m^{3}_{B^{0}_{(s)}}R^{2}{\tau_{B^{0}_{(s)}}%
\over\hbar},$$
(2)
where
$$R={m_{B^{0}_{(s)}}\over m_{b}}{\left(\alpha_{S}(M_{\rm LQ})\over\alpha_{S}(m_{%
t})\right)}^{-{4\over 7}}{\left(\alpha_{S}(m_{t})\over\alpha_{S}(m_{b})\right)%
}^{-{12\over 23}}.$$
The $B^{0}$ and $B^{0}_{s}$ masses, $m_{B^{0}}$ and $m_{B^{0}_{s}}$, and the average lifetimes, $\tau_{B^{0}}$ and $\tau_{B^{0}_{s}}$, are taken from
Ref. PDG2012 .
The factors $F_{B^{0}}=0.190\pm 0.004$ GeV and $F_{B^{0}_{s}}=0.227\pm 0.004$ GeV
are the decay constants of the $B^{0}$ and $B^{0}_{s}$ mesons Davies , and $m_{b}$ and $m_{t}$
are the bottom and top quark masses PDG2012 , respectively, computed in the
$\overline{\rm MS}$ scheme Bardeen:1978yd .
The value of $\alpha_{s}$ at an arbitrary scale $M_{\rm LQ}$ is determined
using the software package rundec RunDec .
Using the limits on the branching fractions shown in Table 2, we find the following lower bounds for the leptoquark masses
if the leptoquark links the $\tau$ lepton to the first and second quark generation,
$M_{\rm LQ}(B^{0}_{s}\to e^{\pm}\mu^{\mp})>101\,(107){\mathrm{\,Te\kern-1.0ptV%
\!/}c^{2}}$ and $M_{\rm LQ}(B^{0}\to e^{\pm}\mu^{\mp})>135\,(126){\mathrm{\,Te\kern-1.0ptV\!/}c%
^{2}}$ at 90 (95) % C.L., respectively.
When the parameters entering in Eq. 2 are fluctuated within $\pm 1\,\sigma$, the limits on the leptoquark masses change by $\sim\pm 1$ TeV.
In summary, a search for the lepton-flavour violating decays $B^{0}_{s}\to e^{\pm}\mu^{\mp}$ and $B^{0}\to e^{\pm}\mu^{\mp}$ has been performed on a data sample, corresponding to an
integrated luminosity of 1.0$\mbox{\,fb}^{-1}$, collected in $pp$ collisions at $\sqrt{s}=7$ TeV.
The data are consistent with the background-only hypothesis.
Upper limits are set on the branching fractions, ${\cal B}(B^{0}_{s}\to e^{\pm}\mu^{\mp})$ $<1.1\,(1.4)\times 10^{-8}$ and
${\cal B}(B^{0}\to e^{\pm}\mu^{\mp})$ $<2.8\,(3.7)\times 10^{-9}$ at 90 (95) % C.L., that are the most restrictive to date.
These limits translate into lower bounds on the leptoquark masses in the Pati-Salam model valencia of $M_{\rm LQ}(B^{0}_{s}\to e^{\pm}\mu^{\mp})>101\,(107){\mathrm{\,Te\kern-1.0ptV%
\!/}c^{2}}$ and
$M_{\rm LQ}(B^{0}\to e^{\pm}\mu^{\mp})>135\,(126){\mathrm{\,Te\kern-1.0ptV\!/}c%
^{2}}$ at 90 (95) % C.L., respectively. These are a factor of two higher than the previous bounds.
I Acknowledgements
We thank Diego Guadagnoli for the theory inputs.
We express our gratitude to our colleagues in the CERN
accelerator departments for the excellent performance of the LHC. We
thank the technical and administrative staff at the LHCb
institutes. We acknowledge support from CERN and from the national
agencies: CAPES, CNPq, FAPERJ and FINEP (Brazil); NSFC (China);
CNRS/IN2P3 and Region Auvergne (France); BMBF, DFG, HGF and MPG
(Germany); SFI (Ireland); INFN (Italy); FOM and NWO (The Netherlands);
SCSR (Poland); MEN/IFA (Romania); MinES, Rosatom, RFBR and NRC
“Kurchatov Institute” (Russia); MinECo, XuntaGal and GENCAT (Spain);
SNSF and SER (Switzerland); NAS Ukraine (Ukraine); STFC (United
Kingdom); NSF (USA). We also acknowledge the support received from the
ERC under FP7. The Tier1 computing centres are supported by IN2P3
(France), KIT and BMBF (Germany), INFN (Italy), NWO and SURF (The
Netherlands), PIC (Spain), GridPP (United Kingdom). We are thankful
for the computing resources put at our disposal by Yandex LLC
(Russia), as well as to the communities behind the multiple open
source software packages that we depend on.
{mcitethebibliography}
10
\mciteSetBstSublistModen
\mciteSetBstMaxWidthFormsubitem)
\mciteSetBstSublistLabelBeginEnd\mcitemaxwidthsubitemform
(1)
A. Ilakovic, Lepton flavor violation
in the standard model extended by heavy singlet Dirac neutrinos,
Phys. Rev. D 62 (2000) 036010\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(2)
R. A. Diaz et al., Improving
bounds on flavor changing vertices in the two Higgs doublet model from
$B^{0}-\bar{B}^{0}$ mixing, Eur. Phys. J C41 (2005) 305\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(3)
J. C. Pati and A. Salam, Lepton
number as the fourth color, Phys. Rev. D 10 (1974) 275\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(4)
ATLAS collaboration, G. Aad et al.,
Search for second generation
scalar leptoquarks in $pp$ collisions at $\sqrt{s}=7$ TeV with the ATLAS
detector, Eur. Phys. J. C72 (2012) 2151,
arXiv:1203.3172\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(5)
ATLAS collaboration, G. Aad et al.,
Search for first generation scalar
leptoquarks in $pp$ collisions at $\sqrt{s}=7$ TeV with the ATLAS
detector, Phys. Lett. B709 (2012) 158, arXiv:1112.4828\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(6)
ATLAS Collaboration, G. Aad et al.,
Search for third generation scalar
leptoquarks in pp collisions at sqrt(s) = 7 TeV with the ATLAS detector,
JHEP 06 (2013)
033, arXiv:1303.0526\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(7)
CMS collaboration, S. Chatrchyan et al.,
Search for pair production of
first- and second-generation scalar leptoquarks in $pp$ collisions at
$\sqrt{s}=7$ TeV,
Phys. Rev. D 86 (2012) 052013, arXiv:1207.5406\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(8)
CMS collaboration, S. Chatrchyan et al.,
Search for third-generation
leptoquarks and scalar bottom quarks in $pp$ collisions at $\sqrt{s}=7$ TeV, JHEP 12
(2012) 055, arXiv:1210.5627\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(9)
CMS Collaboration, S. Chatrchyan et al.,
Search for pair production of
third-generation leptoquarks and top squarks in $pp$ collisions at
$\sqrt{s}=7$ TeV,
Phys. Rev. Lett. 110 (2013) 081801, arXiv:1210.5629\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(10)
G. Valencia and S. Willenbrock, Quark-lepton unification and rare meson decays,
Phys. Rev. D 50
(1994) 6843\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(11)
M. Blanke et al., Charged
lepton flavour violation and $(g-2)_{\mu}$ in the Littlest Higgs Model
with T-Parity: a clear distinction from Supersymmetry,
JHEP 05
(2007) 013, arXiv:0702136\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(12)
CDF collaboration, T. Aaltonen et al.,
Search for the decays
${B^{0}_{(s)}\to e^{+}\mu^{-}}$ and ${B^{0}_{(s)}\to e^{+}e^{-}}$ in CDF Run II,
Phys. Rev. Lett. 102 (2009) 201901, arXiv:0901.3803\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(13)
LHCb collaboration, A. A. Alves Jr. et al.,
The LHCb detector at the LHC,
JINST 3
(2008) S08005\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(14)
T. Sjöstrand, S. Mrenna, and P. Skands,
PYTHIA 6.4 physics and manual,
JHEP 05
(2006) 026, arXiv:hep-ph/0603175\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(15)
D. J. Lange, The EvtGen particle
decay simulation package,
Nucl. Instrum. Meth. A462 (2001) 152\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(16)
GEANT4 collaboration, J. Allison et al.,
GEANT4 developments and
applications, IEEE
Trans. Nucl. Sci. 53 (2006) 270\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(17)
GEANT4 collaboration, S. Agostinelli et al.,
GEANT4: a simulation toolkit,
Nucl. Instrum. Meth. A506 (2003) 250\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(18)
P. Golonka and Z. Was, PHOTOS Monte
Carlo: a precision tool for QED corrections in $Z$ and $W$ decays,
Eur. Phys. J. C45 (2006) 97, arXiv:hep-ph/0506026\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(19)
I. Belyaev et al., Handling of
the generation of primary events in Gauss, the LHCb simulation framework,
Nuclear Science
Symposium Conference Record (NSS/MIC) IEEE (2010) 1155\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(20)
M. Clemencic et al., The LHCb simulation application, Gauss: design, evolution and experience,
J. of Phys:
Conf. Ser. 331 (2011) 032023\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(21)
R. Aaij et al., The LHCb
trigger and its performance in 2011,
JINST 8
(2013) P04022, arXiv:1211.3055\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(22)
V. V. Gligorov and M. Williams, Efficient, reliable and fast high-level triggering using a bonsai boosted
decision tree,
JINST 8
(2013) P02013, arXiv:1210.6861\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(23)
LHCb collaboration, B. Adeva et al.,
Roadmap for selected key
measurements of LHCb, arXiv:0912.4179\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(24)
F. Archilli et al., Performance of the Muon Identification at LHCb,
arXiv:1306.0249, submitted to
JINST\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(25)
LHCb collaboration, R. Aaij et al.,
Measurement of the $B^{0}\rightarrow K^{*0}e^{+}e^{-}$ branching fraction at low dilepton mass,
JHEP 05 (2013)
159, arXiv:1304.3035\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(26)
L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stone, Classification
and regression trees, Wadsworth international group, Belmont, California,
USA, 1984\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(27)
R. E. Schapire and Y. Freund, A
decision-theoretic generalization of on-line learning and an application to
boosting, Jour. Comp. and Syst. Sc. 55 (1997) 119\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(28)
LHCb collaboration, R. Aaij et al.,
First evidence for the decay
$B^{0}_{s}\to\mu^{+}\mu^{-}$,
Phys. Rev. Lett. 110 (2012) 021801, arXiv:1211.2674\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(29)
CDF collaboration, A. Abulencia et al.,
Search for $B^{0}_{s}\to\mu^{+}\mu^{-}$
and $B^{0}_{d}\to\mu^{+}\mu^{-}$ decays in $p\bar{p}$ collisions with CDF II,
Phys. Rev. Lett. 95 (2005) 221805, arXiv:hep-ex/0508036\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(30)
LHCb collaboration, R. Aaij et al.,
Search for the rare decays $B^{0}_{s}\to\mu^{+}\mu^{-}$ and $B^{0}\to\mu^{+}\mu^{-}$,
Phys. Lett. B708 (2012) 55, arXiv:1112.1600\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(31)
A. Carbone et al., Invariant
mass line shape of $B^{0}_{(s)}\to h^{+}{h}^{\prime-}$ decays,
LHCb-PUB-2009-031 (2009)\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(32)
T. Skwarnicki, A study of the radiative cascade transitions between the
Upsilon-prime and Upsilon resonances, PhD thesis, Institute of Nuclear
Physics, Krakow, 1986,
DESY-F31-86-02\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(33)
Particle Data Group, J. Beringer et al.,
Review
of particle physics,
Phys. Rev. D86 (2012) 010001\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(34)
LHCb collaboration, R. Aaij et al.,
Measurement of the ratio of
fragmentation fractions $f_{s}/f_{d}$ and dependence on $B$ meson kinematics,
JHEP 04 (2013) 001, arXiv:1301.5286\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(35)
CDF collaboration, F. Abe et al.,
Observation of the $B_{c}$ meson in
$p\bar{p}$ collisions at $\sqrt{s}=1.8$ TeV,
Phys. Rev. Lett. 81 (1998) 2432, arXiv:hep-ex/9805034\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(36)
LHCb collaboration, R. Aaij et al.,
First observation of the decay
$B^{+}\to\pi^{+}\mu^{+}\mu^{-}$,
JHEP 12 (2012)
125, arXiv:1210.2645\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(37)
A. Khodjamirian, C. Klein, T. Mannel, and Y.-M. Wang,
Form factors and strong couplings
of heavy baryons from QCD light-cone sum rules,
JHEP 09 (2011)
106, arXiv:1108.2971\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(38)
W.-F. Wang and Z.-J. Xiao, The
semileptonic decays $B/B_{s}\to(\pi,K)(\ell^{+}\ell^{-},\ell\nu,\nu\bar{\nu})$
in the perturbative QCD approach beyond the leading-order,
Phys. Rev. D86 (2012) 114025, arXiv:1207.0265\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(39)
LHCb collaboration, R. Aaij et al.,
Measurement of $b$-hadron
production fractions in 7 TeV pp collisions, Phys. Rev. D85
(2012) 032008, arXiv:1111.2357\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(40)
A. L. Read, Presentation of search
results: the CL${}_{\rm s}$ technique,
J. Phys. G28
(2002) 2693\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(41)
C. Davies, Standard Model heavy
flavor physics on the lattice, PoS LATTICE2011 (2011) 019,
arXiv:1203.3862\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(42)
W. A. Bardeen, A. Buras, D. Duke, and T. Muta,
Deep inelastic scattering beyond
the leading order in asymptotically free gauge theories,
Phys. Rev. D18
(1978) 3998\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem
(43)
K. Chetyrkin, J. Kuhn, and M. Steinhauser,
RunDec: a Mathematica package for
running and decoupling of the strong coupling and quark masses ,
Comput. Phys. Comm. 133 (2000) 43, arXiv:hep-ph/0004189\mciteBstWouldAddEndPuncttrue\mciteSetBstMidEndSepPunct\mcitedefaultmidpunct
\mcitedefaultendpunct\mcitedefaultseppunct\EndOfBibitem |
Motion of Contact Line of a Crystal Over the Edge of Solid Mask in Epitaxial Lateral Overgrowth
M. Khenner
Department of Mathematics
State University of New York at Buffalo
Buffalo, NY 14260
email: [email protected]
Abstract
Mathematical model that allows for direct tracking of the homoepitaxial crystal growth out of
the window etched in the solid, pre-deposited layer on the substrate is described. The growth is governed
by the normal (to the crystal-vapor interface) flux from the vapor phase and by the interface diffusion.
The model accounts for possibly inhomogeneous energy of the mask surface
and for strong anisotropies of crystal-vapor interfacial energy and kinetic mobility.
Results demonstrate that the motion of the crystal-mask contact line slows down abruptly as radius of curvature
of the mask edge approaches zero.
Numerical procedure is suggested to overcome difficulties associated with ill-posedness of the evolution
problem for the interface with strong energy anisotropy.
Keywords: Thin films, epitaxy, MOCVD, surface diffusion,
interface dynamics, contact lines, rough surfaces, wetting, regularization of ill-posed evolution
problems.
PACS$\;$ codes: 68.10.Cr, 68.35.Ja, 68.45.Gd, 68.55.-a, 81.10.Aj.
1 Introduction
The Epitaxial Lateral Overgrowth (ELO) and Selective Area Epitaxy (SAG) are commonly used
to grow micro-scale semiconductor crystals and thin films.
The micrometer-scale selective growth is well-achievable by
chemical vapor deposition or liquid phase epitaxy [1]-[4]; the ELO and SAG can be used also to
grow nanostructures by the molecular beam epitaxy [5]-[7].
Recently, the mathematical models that allow to numerically study the ELO and SAG
from vapor [8] - [10] were introduced. These continuum, geometric crystal growth
models are formulated as
free-boundary problems in
two dimensions
(that is, normal to the substrate/mask and normal to the axis of long stripe openings
etched in the mask) and, therefore,
they are capable of explicitly accounting for the mask topography and the
interface
anisotropy. The crystal growth is by the normal
flux of particles from the vapor phase to the interface; the surface diffusion redistributes the material along the
interface during the growth. Evaporation-recondensation from both the mask surface and crystal-vapor interface, and
the diffusion of adatoms along the mask surface to the tri-junction crystal-mask-vapor
are accounted for by the models.
The details of the formulation, boundary conditions and numerical method can be found
in [8, 9]. In this paper we only address issues related to the mask surface modeling
and the crystal interaction with the edge of the mask (mask corner point). Note that
the crystal growth out of the trench occurs relatively early in the ELO process; however,
our previous modeling suggests [8, 9] that the perturbation introduced to the crystal-vapor
interface by the edge of the mask exists for a long time after the actual interaction of the advancing
edge of a crystal (contact line) and the mask edge took place. Thus it seems important to understand
the details of the contact line motion over the edge.
In [8, 9] the treatment of the contact line motion relies on the assumptions of
(a) the thermodynamic equilibrium
at the tri-junction (indeed, SAG and ELO are equilibrium processes), (b) the mask surface as the
rectangular, macroscopic step, and (c) the mask as the energetically homogeneous surface:
the mask-vapor surface energy $\gamma_{mv}=const.$ The assumptions (b) and (c)
limit the ability of any model to describe the overgrowth of the crystal line from the vertical onto
the horizontal part of the mask. Indeed, (a) implies that the equilibrium contact angle $\phi$
is given by the equation
$$\gamma_{mv}-\gamma_{cm}-\gamma_{cv}\cos{\phi}+\frac{\partial\gamma_{cv}}{%
\partial\phi}\sin{\phi}=0,$$
(1.1)
where $\gamma_{cm}$ and $\gamma_{cv}$ are the energies of the crystal-mask and crystal-vapor interfaces, respectively
(see Appendix A).
Then, (c) together with the natural assumptions of constant (at the tri-junction) $\gamma_{cm}$ and
$\gamma_{cv}$ imply that $\phi$ is a constant value
at the mask. However, due to (b) the mask edge is a singular point in the sense that the contact
angle is not well defined there. In [9], this situation was dealt with by instant readjustment
of the interface $-$ $x$-axis angle (angle of the tangent, $\phi_{x}$)
to a new value once the contact point
reaches the mask edge to preserve the equilibrium constant angle $\phi$. Such instant readjustment causes the short
unphysical retraction of the contact point into the vapor phase and influences the shape of the interface
111Angle of the tangent
is necessary for the determination of
the location of the contact point on the mask by coordinate’s extrapolation from marker particles
adjacent to the contact point, through
$\tan{\phi_{x}}=\frac{dy/ds}{dx/ds},$
(1.2)
where $s$ is the arc length along the curve (crystal-vapor interface).
.
In this paper we make an attempt to model
the overgrowth event self-consistently by approximating the mask edge by a circular arc
and restricting the contact point to movement along this arc. We numerically study the limit of zero arc radius.
Since in applications the mask is often a crystalline solid itself, we allow for the anisotropy
of $\gamma_{mv}$ and thus, by virtue of (1.1), for the variable $\phi$ along the mask.
(Quartz or tungsten masks are common [2].
Sometimes a mask is not deposited at all, but the long stripes are etched
in the crystalline GaAs or sapphire substrate itself [1, 11, 12]. The long crystalline mesas
left on a surface of a substrate in this latter case are more advantageous for the epitaxy of some semiconductor
materials than the
geometrically equivalent mesas of the foreign (to the substrate) mask material.)
Apart from the interest to this problem
that stems from the ELO technology, there is a broad class of phenomena in which the material-defect
interaction
is a key component. For instance, when liquid wets the microscopically rough surface
the contact line gets pinned on surface asperities and thus becomes unstable. The latter problem, both for a single
asperity and for the random ensemble of asperities has been
studied intensively from a physico-mechanical standpoint;
see, for instance, recent works [13, 14] and
the review article by de Gennes [15]
.
The somewhat different
situation where the surface defect is the macroscopic, organized structure
was considered, for example, by Oliver et al. in [16]. In that work, the defect is
the edge of a solid disk that supports
the liquid drop on its top.
The pinning effect of the edge has been
examined theoretically and Gibbs inequality condition for the equilibrium of a drop bound
by the edge has been confirmed experimentally. Also, the various conditions for drop stability
at the edge have been examined. Authors used a circular sapphire disk with the 90${}^{\circ}$ edge and the
aluminum disks with the edges subtending a range of angles.
In [17], the finite element stability analysis of an inclined pendant drop
was performed, as well as the experiment. The latter showed that both the contact line and the contact
angle can adjust around the capillary until the Young-Laplace equation
that balances the capillary pressure
with pressure forces due to gravity [15] is satisfied.
In [18], the asymptotic solution
to the model of the contact line pinning on a single, macroscopic spherical defect was derived.
However, our bibliography search for the study in which the direct numerical tracking of the contact line
motion over the defect is performed was unsuccessful.
Obviously, the ELO problem
for the homoepitaxial crystal is simpler
than any similar problem involving a liquid, since
there is at least no need to solve the Young-Laplace equation.
Heteroepitaxial and thermal stresses, if present in a
solid-on-solid system, may significantly complicate the analysis, but they
usually are incapable of modifying the equilibrium contact angle determined by the interfacial energies alone
[19] (however, as is well known, they may alter the dynamics of the free surface of a solid film).
The latter in the case of the SAG was demonstrated in [20],
where the morphological evolution of a heteroepitaxial thin
film growing on a patterned substrate was studied by the phase-field method.
2 Formulation
The Figure 1 shows the problem geometry.
As was already noted, the length of the open, unmasked stripes is assumed much larger
than their widths since large aspect ratios are common
in the applications; thus, the 2D crystal growth model can be used.
2.1 Equations of crystal growth
Based on the classical Mullins’ approach [21], the normal velocity of the
interface is formulated as a function of the local curvature and its derivatives.
In this work, we use
$$V_{q}=D\frac{\partial^{2}}{\partial s^{2}}\left[\left(\hat{\gamma}_{cv}(\alpha%
)+\frac{\partial^{2}\hat{\gamma}_{cv}(\alpha)}{\partial\alpha^{2}}\right)K%
\right]+J\hat{M}(\alpha),$$
(2.1)
which is the simplified expression from [9]. We neglect the evaporation-recondensation term
as being unimportant for this study and assume the isotropic interfacial diffusivity.
The equation (2.1) is in the non-dimensional form.
$D$ has the meaning of the constant effective interfacial diffusivity (see Table LABEL:Table2 in the Appendix B),
$K$ is the curvature and $J$ the composite parameter that involves, in particular,
the chemical potential of the vapor phase and the mean value of the interface mobility, $M_{0}$ (the
kinetic coefficient).
$\hat{\gamma}_{cv}(\alpha)$ and $\hat{M}(\alpha)$ are the anisotropy factors of the
interface energy and mobility, respectively.
These factors are specified in Section 2.4.
The
marker particles on the interface are numerically advanced using the parametric evolution equations
for the Cartesian coordinates of particles, viz.
$$\frac{\partial x}{\partial t}=V_{q}\ \frac{\partial y}{\partial s},$$
(2.2)
$$\frac{\partial y}{\partial t}=-V_{q}\ \frac{\partial x}{\partial s}.$$
These equations are fourth-order parabolic when $V_{q}$ is given by (2.1).
2.2 Mask Energy and Shape
To derive the expression for $\gamma_{mv}$ which accounts for the additional edge energy,
we use the method of
Xin & Wong [22, 23], which they employed in [24] for studies of the grain boundary
(GB) grooving in polycrystalline
thin films. In contrast to the case of dynamically evolving, faceted
interfaces such as GB grooves, the application to the static crystalline mask surface is straightforward
since the method requires an a priori knowledge of facets orientations and lengths.
Details of the method can be found in [22, 23].
For the rectangular crystal in the thermodynamic equilibrium that represents the mask on a substrate
(half of that crystal is shown in Fig. 1,2), the result is the following nondimensional
expression (units for the surface
energy are chosen $\mu_{m}h_{m}/\Omega$, where $\Omega$ is the atomic volume and $\mu_{m}$ is the equilibrium
chemical potential
of the mask surface):
$$\gamma_{mv}=r+\left[\bar{\ell}-r+2(r-\bar{\ell})F\left(\theta-\frac{\pi}{2}%
\right)\right]\cos{\theta}+2(1-r)\left[\frac{1}{2}-F\left(\theta-\pi\right)%
\right]\sin{\theta},\quad 0\leq\theta\leq\pi.$$
(2.3)
In (2.3), $r=R/H$ is the ratio of the nondimensional circular arc radius to the nondimensional
step height, $\bar{\ell}=d/H$, where $d$ is the nondimensional half-width of the mask, and
$F(u)$ is Heaviside function (all lengths are nondimensionalized by the half-width of the stripe, $L$).
For our modeling of the ELO, we are only interested in values of
$\theta$ in $[0,\pi/2]$. The function $\gamma_{mv}(\theta)$ is plotted in Fig. 3 for $\bar{\ell}=1$
and decreasing values
of $r$. Notice that $\gamma_{mv}$ is at maximum at $\theta=\pi/4$,
since the half-mask is taken square with the rounded edge.
The sharper is the edge, the larger is the maximum value. For the 90${}^{\circ}$
edge $r$ is exactly zero, $\theta$ takes on values 0 and $\pi/2$ only, and (2.3)
gives $\gamma_{mv}(0)=\gamma_{mv}(\pi/2)=1$.
To parametrize the mask surface, we take angle $\tilde{\theta}\in\left[-\tilde{\theta}_{1},\tilde{\theta}_{2}+\pi/2\right]$
as the independent
variable (see Fig. 2). From geometry,
$$\tilde{\theta}_{1}=\mbox{arctan}\frac{H-R}{R}=\mbox{arctan}\left(\frac{1}{r}-1%
\right),$$
(2.4)
$$\tilde{\theta}_{2}=\mbox{arctan}\frac{d-R}{R}=\mbox{arctan}\left(\frac{d}{R}-1%
\right).$$
Then,
$$\theta=\left\{\begin{array}[]{l}0,\mbox{ if}\;\;-\tilde{\theta}_{1}\leq\tilde{%
\theta}<0,\\
\tilde{\theta},\mbox{ if}\;\;0\leq\tilde{\theta}\leq\pi/2,\\
\pi/2,\mbox{ if}\;\;\pi/2<\tilde{\theta}\leq\pi/2+\tilde{\theta}_{2}.\end{%
array}\right.$$
(2.5)
Also,
$$x_{m}=\left\{\begin{array}[]{l}0,\mbox{ if}\;\;-\tilde{\theta}_{1}\leq\tilde{%
\theta}<0,\\
R\left(-1+\cos{\tilde{\theta}}\right),\mbox{ if}\;\;0\leq\tilde{\theta}\leq\pi%
/2,\\
-R\left(1+\tan{\tilde{\theta}}\right),\mbox{ if}\;\;\pi/2<\tilde{\theta}\leq%
\pi/2+\tilde{\theta}_{2},\end{array}\right.$$
(2.6)
$$y_{m}=\left\{\begin{array}[]{l}H-R\left(1+\tan{\tilde{\theta}}\right),\mbox{ %
if}\;\;-\tilde{\theta}_{1}\leq\tilde{\theta}<0,\\
H-R\left(1-\sin{\tilde{\theta}}\right),\mbox{ if}\;\;0\leq\tilde{\theta}\leq%
\pi/2,\\
H,\mbox{ if}\;\;\pi/2<\tilde{\theta}\leq\pi/2+\tilde{\theta}_{2}\end{array}\right.$$
(2.7)
are the nondimensional parametric functions that describe the mask surface. From (2.5)-(2.7),
the rounded portion of the mask is described by
$$\left\{\begin{array}[]{l}x_{m}^{(arc)}=R\left(-1+\cos{\theta}\right),\\
y_{m}^{(arc)}=H-R\left(1-\sin{\theta}\right),\;0\leq\theta=\tilde{\theta}\leq%
\pi/2.\end{array}\right.$$
(2.8)
Taking $\mu_{m}h_{m}/\Omega$ for the unit of $\gamma_{cm}$, and $\gamma_{0}$ (the mean value, see (2.16) below) for
the unit
of $\gamma_{cv}$, the equation (1.1) takes the form
$$E(\gamma_{mv}-\gamma_{cm})-\hat{\gamma}_{cv}(\alpha^{*})\cos{\phi}+\frac{%
\partial\hat{\gamma}_{cv}}{\partial\phi}(\alpha^{*})\sin{\phi}=0,$$
(2.9)
where $E=\mu_{m}h_{m}/(\Omega\gamma_{0})$, $\gamma_{mv}$ is given by (2.3),
$\gamma_{cm}$ now is the nondimensional energy of the crystal-mask interface (assumed constant and isotropic),
$\hat{\gamma}_{cv}$ is the anisotropy
factor of the interfacial energy, and $\alpha^{*}$ is value of the angle of the normal to the interface at the
tri-junction. Notice that
$$\alpha^{*}=\phi+\theta.$$
(2.10)
(2.9)
is solved once in the beginning of the computation and yields values of the imposed contact angle at every
$\tilde{\theta}_{i}\in\left[-\tilde{\theta}_{1},\tilde{\theta}_{2}+\pi/2\right]%
,i=1...N_{\tilde{\theta}}$,
where $N_{\tilde{\theta}}$ is the number of equidistant nodes in the interval.
$N_{\tilde{\theta}}$ is taken in the range 500…3000, the larger the smaller is the radius of the circular arc
(since for the accurate marching of the contact point along the arc it is important to have sufficient
number of nodes there).
2.3 Updating the contact point
In this section, the procedure which updates the position of the contact point on the mask is described.
The update is needed
after all marker particles on the interface but the contact point are moved one time step forward
using the evolution equations (2.2).
Let $x_{*}$ and $y_{*}$ are the “old” coordinates of the contact point. Using $x_{*},y_{*}$, the coordinates of the
two nearest nodes on the mask: $x_{m}(i),y_{m}(i),x_{m}(i+1),y_{m}(i+1)$, and $\theta_{i,i+1},\phi_{i,i+1}$,
we first compute values of $\theta$ and $\phi$ at $(x_{*},y_{*})$ by linear interpolation. Using the latter values,
the angle of the tangent to the interface at $(x_{*},y_{*})$ is found from
$$\phi_{x}=\phi+\theta-\pi/2.$$
(2.11)
Let $(x_{1},y_{1})$ and $(x_{2},y_{2})$ are the coordinates of the first two marker particles on the interface
closest to the contact point.
In case the contact point is on the vertical part of the mask, the updated location is $(0,\hat{y}_{*})$, where
$$\hat{y}_{*}=\frac{1}{3}\left(4y_{1}-y_{2}+\left(x_{2}-4x_{1}\right)\tan{\phi_{%
x}}\right).$$
(2.12)
This follows from the second-order, one-sided finite difference approximation of the contact angle condition (1.2).
In case the contact point is on the horizontal part of the mask, the updated location is $(\hat{x}_{*},H)$, where
$$\hat{x}_{*}=\frac{1}{3}\left(4x_{1}-x_{2}+\frac{3H-4y_{1}+y_{2}}{\tan{\phi_{x}%
}}\right).$$
(2.13)
In case the contact point is on the circular arc, we
solve the quadratic equation (ref. (2.8))
$$\frac{1}{3}\left(4y_{1}-y_{2}+\left(3x_{*}-4x_{1}+x_{2}\right)\tan{\phi_{x}}%
\right)=H-R\left(1-\sqrt{1-(1+x_{*}/R)^{2}}\right)$$
(2.14)
and find two $x$-coordinates $\hat{x}_{*}^{(1)},\hat{x}_{*}^{(2)}$ and corresponding $y$-coordinates
$\hat{y}_{*}^{(1)},\hat{y}_{*}^{(2)}$. Then, if both $\hat{x}_{*}^{(1)},\hat{x}_{*}^{(2)}\geq-R$ (that is, if
both possible updated locations are on the arc), we take the pair of coordinates closest to $(x_{*},y_{*})$
for the updated location.
If only one of $(\hat{x}_{*}^{(1},\hat{y}_{*}^{(1)}),\;(\hat{x}_{*}^{(2)},\hat{y}_{*}^{(2)})$ is on the arc and the other
is not, we choose
the pair that is on the arc (closer to the “old” location).
2.4 Anisotropy
The anisotropy factor of the kinetic mobility is chosen as in [25], where the following form is
suggested and the simulations of faceted growth are performed within the framework of phase field bulk
solidification model:
$$\hat{M}(\alpha)=1-\sigma+2\sigma\tanh{\frac{\chi}{|\tan{2\left(\alpha+\beta_{M%
}\right)}|}};\quad M=M_{0}\hat{M}(\alpha).$$
(2.15)
In (2.15), $\sigma$ and $\chi$ are constant parameters that control the width and flatness of
facets, respectively; $\beta_{M}$ is the phase angle. Facets are formed at regions on the interface
where $\alpha=\pi/4-\beta_{M}+n\pi/2,\;n=0,1,...$. It was shown in [25] that the results of computations
with the four-fold anisotropy (2.15) are in good agreement with the so-called kinetic Wulff shapes
(see references in [25]; specifically for the selective area epitaxy the kinetic Wulff shapes
were constructed and compared to the experimental shapes in [26]). For the simulations discussed
in the next section we chose $\sigma=0.95,\;\chi=2,\;\beta_{M}=0$.
The anisotropy factor of the crystal-vapor interfacial energy has standard, four-fold symmetric form
$$\hat{\gamma}_{cv}(\alpha)=1+\epsilon_{\gamma}\cos{\left[4\left(\alpha+\beta_{%
\gamma}\right)\right]};\quad\gamma_{cv}=\gamma_{0}\hat{\gamma}_{cv}(\alpha),$$
(2.16)
where the constant $\epsilon_{\gamma}>0$ determines the degree of anisotropy, and $\beta_{\gamma}$ is the phase angle.
In this work we allow for strong anisotropy of $\gamma_{cv}$, e.g. $\epsilon_{\gamma}>1/15$.
In this case, the interface stiffness $G=\hat{\gamma}_{cv}+\partial^{2}\hat{\gamma}_{cv}/\partial\alpha^{2}$ is
negative in
certain intervals of values of $\alpha$, and that manifests in the appearance of
corners on the equilibrium Wulff shape of a crystal [27];
in the dynamical case, the evolution equations become backward-parabolic and unstable
where $G<0$, see for example [28] - [30].
2.4.1 Regularization
To track interface evolution with the strongly anisotropic interfacial energy, the problem must be regularized
in order to penalize spatial oscillations as well as the tendency to form corners. The regularization
we use in this work is based on the addition of curvature dependence to $\gamma_{cv}$. Regularization by curvature
was first proposed
by Herring [27] for the equilibrium shape; see [31] for the modern mathematical analysis
of this
problem. Apart from being
useful as a modeling tool, the regularization by curvature has physical ground, namely the interaction
of steps on a crystal surface in the vicinity of a corner [29].
Unlike [29, 30] and other works where the $\delta\kappa^{2},\delta=const.$ ($\kappa$ is the
dimensional curvature) term is added to
the surface energy on the problem formulation stage,
here we regard the regularization only as a numerical technique intended to keep the computation alive, and
regularize by adding same
term to $\gamma_{cv}$ given by (2.16) only at the distinct moments and only at the locations on the
interface where the
instability (tendency to form a corner) occurs. This is done as follows.
First, we continuously check the
sign of $G$ at every marker location on the interface. Second, once at the certain location $G$ becomes negative
(the corner has been formed), we interrupt the computation, restore all variables including the local interface
shape
and curvature to their previous states at positive $G$, add $\Delta K^{2},\ \Delta=\delta/(\gamma_{0}L^{2})$ term
to the nondimensional interfacial energy
$\hat{\gamma}_{cv}(\alpha)$
at this location and
solve the
linear equation $G(\Delta)=0$. The result,
$$\Delta=\frac{15\epsilon_{\gamma}\cos{\left[4\left(\alpha+\beta_{\gamma}\right)%
\right]}-1}{K^{2}+2K\partial^{2}K/\partial\alpha^{2}+2(\partial K/\partial%
\alpha)^{2}},$$
(2.17)
is deflected by 0.1$\Delta$ to ensure positive $G$.
The computation is then restarted with the regularized
$\gamma_{cv}$; if needed, the procedure is repeated. This method allows to reduce the influence of the regularization
on interface shape to a minimum by (i) applying the regularization selectively and (ii) choosing value for the
regularization constant $\Delta$ no larger than needed to allow the computation to proceed.
The maximum value $\Delta\sim 10^{-5}$ comes out from the computation, which translates into $\delta\sim 10^{-11}$
erg; thus the additional corner energy necessary to regularize the problem is indeed very small.
222In [29], parameter $\delta$ is introduced as the second partial derivative of the crystalline
step energy
with respect to the first coordinate derivative of step density; this is difficult, if not impossible to extract
from experimental data. We are not aware of a single experiment there this or related quantity is measured.
Since value is unknown, we think it makes sense to find the suitable numerical
approximation to the nondimensional counterpart as explained.
It must be noted here that the numerical method itself introduces some minor smoothing, since the interface is remeshed
every time step with the help of cubic splines [8].
The regularization by curvature is not used at the contact point and its vicinity. Instead,
to ensure stable tracking of the sensitive contact point motion over the mask
edge in the simulations with strongly anisotropic surface energy, we had to resort to the method which is
described in the next subsection.
2.4.2 Contact point regularization
We take
$$\epsilon_{\gamma}=\epsilon_{\gamma}^{(0)}+\left(\Upsilon-\epsilon_{\gamma}^{(0%
)}\right)H_{\lambda}(s-s_{c}),$$
(2.18)
where $\epsilon_{\gamma}^{(0)}$ is constant value $\epsilon_{\gamma}$ takes on the interval of the arc
length $0\leq s<s_{c}-\lambda$,
$\Upsilon>\epsilon_{\gamma}^{(0)}$ is constant value $\epsilon_{\gamma}$ takes on the
interval $s_{c}+\lambda<s\leq S$, and
$H_{\lambda}(s-s_{c})$ is the smoothed Heaviside function (see, for instance, [32]),
viz.
$$H_{\lambda}(s-s_{c})=\left\{\begin{array}[]{l}0,\mbox{ if}\;s-s_{c}<-\lambda,%
\\
\frac{1}{2}\left(1+\frac{s-s_{c}}{\lambda}+\frac{1}{\pi}\sin{\frac{\pi(s-s_{c}%
)}{\lambda}}\right),\mbox{ if}\;-\lambda\leq s-s_{c}\leq\lambda,\\
1,\mbox{ if}\;s-s_{c}>\lambda.\end{array}\right.$$
(2.19)
In (2.19), $\lambda=const.$ is the small half-width of the transition region from $0$ to $1$ around
$s_{c}=const.\in(0,S)$. (2.18) and (2.19)
imply that the intervals of the interface $0\leq s<s_{c}-\lambda$ and $s_{c}-\lambda\leq s\leq s_{c}+\lambda$ evolve
with lesser degree of anisotropy
than the remaining interval $s_{c}+\lambda<s\leq S$. For instance, for the simulation shown in
Fig. 9, we chose $\epsilon_{\gamma}^{(0)}=0.01,s_{c}=0.08,\lambda=0.02$.
Of course,
such choice is difficult to justify physically, but nevertheless it allows to perform simulations and
obtain the right interface shape some (small) distance away from the contact point [33].
3 Results
Since the parameter space is very large, the overgrowth was computed
for the separate cases of evolution with the anisotropic mobility and crystal-vapor interfacial energy.
We studied the influence of parameters $r$ and $\gamma_{cm}$. Other parameters are fixed
at their values cited in Table LABEL:Table2.
3.1 Overgrowth with anisotropic mobility
It is assumed in this section that $\epsilon_{\gamma}=0$, thus $\hat{\gamma}_{cv}\equiv 1$ and
the contact angle $\phi$ is determined from the equation (2.9) where the third term identically
equals zero.
Fig. 4(a) shows the overgrowth over the blunt edge $r=0.5$ (the radius = 50 nm) of the energetically
inhomogeneous mask.
Also shown in Fig. 4(b)
is the actual contact angle $\phi$ vs. $\theta$. Value for $\gamma_{cm}$ is chosen 100 (500 erg/cm${}^{2}$), so that $\phi(\theta)>\pi/2\ \forall\ \theta$ (partial “wetting”).
One of accuracy checks is the comparison every
time step of the actual value of the contact angle with the imposed $\phi(\theta)$
given by (2.9). The difference is negligible for all runs performed.
Notice that the perfectly straight facet is formed in the direction 135${}^{\circ}$ to the substrate.
The feature (“bump”) is present at the junction of this facet and the horizontal, 90${}^{\circ}$ facet.
This bump is present on the surface even before the contact point reaches the mask edge, but the edge
makes it more pronounced. The bump gradually disappears after the edge is passed.
Fig. 5(a,b) compares the overgrowth over the sharper edge $r=0.0125$ (the radius = 1.25 nm)
of the energetically inhomogeneous
mask as in Fig. 4(b) with the overgrowth over
the edge of same sharpness of the energetically homogeneous mask. For the latter,
value for the constant $\phi$ is taken the average of the
smallest and largest values of $\phi$ in Fig. 4(b) ($\phi=1.956$). The overgrowth over the
energetically homogeneous mask is faster, and the bump is more pronounced in this case.
Nevertheless, the difference in the overgrowth of the mask edges of different energy is quite small.
Fig. 6(a,b) compares the overgrowth over the $r=0.0125$ edge
of the energetically inhomogeneous mask that supports different $\phi(\theta)$
(that is, the dynamics corresponding to $\phi(\theta)>\pi/2$ as in Fig. 4(b), and
$\phi(\theta)\leq\pi/2$ as in Fig. 6(c)).
Interestingly,
the bump and the crystal thickness are larger for the less wetting case (larger $\phi$’s)
when the contact point slides over the mask edge. However, immediately after the edge
the overgrowth of the more wetting crystal is faster, but it slows down as the overgrowth
on the mask progresses. This is expected since $\phi=\pi/2$ on the horizontal part of the mask.
Finally, the less wetting crystal takes over, but its
thickness in the $y$-direction is less than one of the more wetting crystal.
The distance from the contact point to the
foot point (0,0) of the mask vs. time is shown in Fig. 7(a,b) for
the overgrowth
over $r=0.5$ and $r=0.0125$ edge, respectively, and $\gamma_{cm}=100$.
For the blunt edge, the computational data is fitted well
by the cubic polynomial and thus the speed of the contact point vs. time
is the quadratic function that has
the minimum value $\approx 4\times 10^{-4}$ (4 nm/s) at $t=125$ (6.3 s) and the maximum value
$\approx 1.4\times 10^{-3}$ (14 nm/s) at $t=50,200$ (2.5,10 s). For the sharp edge,
the quadratic fit is better, and the speed of the contact point is linear function of time
(the min. value $\approx 3.25\times 10^{-4}$ (3.25 nm/s) at $t=100$ (5 s),
the max. value $\approx 5.5\times 10^{-4}$ (5.5 nm/s) at $t=86$ (4.3 s)).
This data,
as well as Figures 4-6 demonstrate
that de-pinning of the contact line of a crystal from the solid defect of finite size does
not have a threshold nature;
that is, the contact line is in constant motion over the defect, and the problem does not have a critical
eigenvalue such that the contact line is pinned when some parameter is smaller than this eigenvalue and
de-pinned (moving) when the parameter is larger.
In the language of [15], where the
analogy between perturbed (on a defect) liquid contact line and an elastic spring is used to phenomenologically
explain pinning (pages 836-38),
the total pinning force exerted by the defect on the line is such that graphs of this force and the restoring
linear elastic force (vs. position on a defect) are not intersecting, and thus there is no
equilibrium, pinning line position(s) on a defect.
Finally, the average nondimensional speed of the contact point’s motion over the edge
(naturally defined as the length of the circular segment of the mask, $\pi R/2$,
divided by the traversal time) is plotted in Fig. 8 vs. $r$.
There is an evident tendency for pinning of the contact point at the edge as $r\rightarrow 0$.
Data can be satisfactory fitted by a hyperbolic tangent curve (not shown).
In the limit $r=0$ (the mathematically sharp edge)
the true pinning is expected (the speed $=$ 0). The contact line starts to actually
“feel” the edge only when the latter has sufficiently small radius of curvature
($r\approx 0.1-0.15\;\Leftrightarrow\;$10-15 nm).
Note that despite such characteristic of the motion of the contact line over the edge as the speed
(instantaneous and average) is, of course, a function of all parameters of the system, the
pinning on the edge as its radius of curvature approaches zero is the universal tendency.
This was checked by means of multiple runs with different parameters. The computation fails
for $r<0.0075$ (the radius$<0.75$ nm) because as the number of discretizing nodes along the mask increases,
the nodes on the two straight segments of the mask come too close to each other and the
interpolation becomes inaccurate. In principle, the computation with smaller radii can be made successful
by using the nonuniform grid along the mask; we did not attempt to do that. It is also doubtful that
very sharp mask edges (having the radii of curvature of the order of a molecule size (1-2 nm))
are achievable by the lithography, etching and polishing used in the ELO/SAG practice
for the preparation of a mask. For comparison, the sapphire disks used for the late 1970s experiments with
the liquid drops by Oliver et al. [16] had the edge radius of curvature smaller than 50 nm,
and “much smaller” than that
value if the defects (chipped-off hollows) appear along the edge.
These disks were prepared by cutting from a fused sapphire boule.
The aluminum disks they used had very blunted edges with the radius of curvature 1 to 5 $\mu$m.
3.2 Overgrowth with strongly anisotropic $\gamma_{cv}$
It is assumed in this section that $\sigma=0$, thus $\hat{M}\equiv 1$.
The contact angle $\phi$ is determined from the equation (2.9), where the third term is nonzero
due to the anisotropy in the crystal-vapor surface energy.
Fig. 9 shows, as an example, the overgrowth over the blunt edge $r=0.5$
(the contact angle $\phi$ as in Fig. 4(b)). In Fig. 9(a),
the value of the anisotropy
parameter $\Upsilon$
(see (2.18)) is 0.07. This is just above the critical value 1/15. $\Upsilon=0.085$ in
Fig. 9(b). For both figures, the value of the phase angle $\beta_{\gamma}$ is chosen 1.67
(a little larger than $\pi/2$).
In Fig. 9(a), the corner forms slowly at the elevation, and it is followed by the
straight facet; other than that, the evolution resembles the isotropic evolution [8].
In Fig. 9(b), the interface instantly becomes unstable and the characteristic, coarsening
with time hill-and-valley structure
is formed [29, 30]. As the result of coarsening, the single elevated corner
(bump) followed by the $\alpha=45^{\circ}$
inclined facet are again formed. Thus this numerical experiment suggests that the strong
interfacial energy anisotropy may be one of the possible causes for the “bumpy” crystal shapes
routinely observed in ELO deposits [1, 4].
Fig. 10 shows the contact angle along the mask surface. In contrast to Figures 4(b) and
6(c), this function is not monotone. As $\epsilon_{\gamma}^{(0)}$ increases quite beyond value 0.01
chosen for computations in this section, the variations
in $\phi$ become more
pronounced and the computation becomes unstable.
Acknowledgements
The author thanks Professors Richard Braun (University of Delaware) and Brian Spencer (SUNY at Buffalo)
for the stimulating and useful discussions of this work. The reviewer is acknowledged for pointing out
the relevance of Herring’s equation to the problem considered in this work.
Appendix A On equation (1.1)
The equation (1.1) follows from the following equation derived by Herring [34]:
$$\gamma_{1}-\gamma_{2}\cos{\alpha_{2}}-\gamma_{3}\cos{\alpha_{3}}+\frac{%
\partial\gamma_{2}}{\partial\phi_{2}}\sin{\alpha_{2}}+\frac{\partial\gamma_{3}%
}{\partial\phi_{3}}\sin{\alpha_{3}}=0.$$
(A.1)
The equation (A.1) generalizes the Young-Dupré equation
$$\gamma_{2}\cos{\alpha_{2}}=\gamma_{1}-\gamma_{3}$$
(A.2)
to orientational dependence of surface energies (tensions) at the tri-junction (see Fig. 11).
To obtain (1.1) from (A.1), we identify the interface 1 in Fig. 11 with the mask-vapor interface,
the interface 2 with the crystal-vapor interface, the interface 3 with the crystal-mask interface and notice that
$\alpha_{2}=\phi$ and $\alpha_{3}=0$ since the slope of the mask surface is everywhere continuous (Fig. 2).
Appendix B Parameters of the model (characteristic of epitaxy of GaAs-like semiconductor at T$\approx$ 650${}^{\circ}$C
)
This appendix has been removed from online submission to arXiv.org in order to comply with
file size requirement; check the full version at
www.math.buffalo.edu/$\sim$mkhenner.
References
[1]
E.J. Thrush, J.P. Stagg, M.A. Gibbon, R.E. Mallard,
B. Hamilton, J.M. Jowett, E.M. Allen,
Mater. Sci. and Eng.$\;$ B21 (1993) 130.
[2]
M.G. Mauk, J.P. Curran, J. Crystal Growth$\;$
225 (2001) 348.
[3]
O.-K. Nam, T.S. Zheleva, M.D. Bremser, R.F. Davis,
J. Electron. Mater.$\;$ 27(4) (1998) 233.
[4]
J.E. Greenspan, X. Zhang, N. Puetz, B. Emmerstorfer,
J. Vac. Sci. Technol. A$\;$ 18(2) (2000) 648.
[5]
S.C. Lee, L.R. Dowson, S.R.J. Brueck,
J. Cryst. Growth$\;$ 240 (2002) 333-339.
[6]
Y. Yamazaki, J.H. Chang, M.W. Cho, T. Sekiguchi, T. Yao,
J. Cryst. Growth$\;$ 214/215 (2000) 202.
[7]
X. Mei, M. Blumin, D. Kim, Z. Wu, H.E. Ruda,
J. Cryst. Growth$\;$ 251 (2003) 253-257.
[8]
M. Khenner, R.J. Braun, M.G. Mauk,
J. Cryst. Growth$\;$ 235 (2002) 425.
[9]
M. Khenner, R.J. Braun, M.G. Mauk,
J. Cryst. Growth$\;$ 241 (2002) 330.
[10]
M. Khenner, “Enhancement of Epitaxial Lateral Overgrowth by Vapor-Phase Diffusion”,
in press.
[11]
R. Notzel, M. Ramsteiner, J. Menniger, A. Trampert,
H.-P. Schonherr,
L. Daveritz, K.H. Ploog, J. Appl. Phys.$\;$ 80(7) (1996) 4108-4111.
[12]
Y.P. Hsu, S.J. Chang, Y.K. Su, J.K. Sheu, C.T. Lee, T.C. Wen, L.W. Wu,
C.H. Kuo, C.S. Chang, S.C. Shei, J. Cryst. Growth$\;$ 261 (2004) 466-470.
[13]
D. Quéré, Physica A$\;$ 313 (2002) 32-46.
[14]
R. Golestanian, E. Raphael, Phys. Rev. E$\;$ 67 (2003) 031603.
[15]
P.G. de Gennes, Rev. Modern Phys.$\;$ 57(3) (1985) 827.
[16]
J.F. Oliver, C. Huh, S.G. Mason, J. Colloid Interface Sci.$\;$ 59 (1977) 568-581.
[17]
A. Laval, R.A. Brown, J. Colloid Interface Sci.$\;$ 89(2) (1982) 332.
[18]
S.B.G. O’Brien, J. Colloid Interface Sci.$\;$ 183 (1996) 51-56.
[19]
D.J. Srolovitz, S.H. Davis, Acta Mater.$\;$ 49(6) (2001) 1005-1007.
[20]
J.J. Eggleston, P.W. Voorhees, Appl. Phys. Lett.$\;$ 80(2) (2002) 306-308.
[21]
W.W. Mullins,
J. Appl. Phys.$\;$ 28(3) (1957)
333.
[22]
T. Xin, H. Wong, Surf. Sci. Lett.$\;$ 487 (2001) L529-L533.
[23]
T. Xin, H. Wong, Technical Report ME-MA1-03, Louisiana State University (2003).
[24]
T. Xin, H. Wong, Acta Mater.$\;$ 51 (2003) 2305-2317.
[25]
T. Uehara, R.F. Sekerka, J. Cryst. Growth$\;$ 254 (2003) 251.
[26]
S.H. Jones, L.K. Seidel, K.M. Lau, M. Harold, J. Cryst. Growth$\;$ 108 (1991) 73.
[27]
C. Herring, Phys. Rev.$\;$ 82 (1951) 87.
[28]
A. Di Carlo, M.E. Gurtin, P. Podio-Guidugli,
SIAM J. Appl. Math.$\;$ 52(4) (1992) 1111-1119.
[29]
A.A. Golovin, S.H. Davis, A.A. Nepomnyashchy,
Physica D$\;$ 122 (1998) 202-230.
[30]
T.V. Savina, A.A. Golovin, S.H. Davis, A.A. Nepomnyashchy, P.W. Voorhees,
Phys. Rev. E$\;$ 67 (2003) 021606-16.
[31]
B. J. Spencer, “Asymptotic solutions for the equilibrium crystal shape
with small corner energy regularization”,
in press.
[32]
M. Sussman, P. Smereka, S. Osher, J. Comput. Phys.$\;$ 114 (1994) 146.
[33]
Question still remains whether it is possible to have the contact angle in the
unstable range of angles and still move along the mask, or just remain attached (R.J. Braun, private communication).
This difficulty is physical as well as mathematical.
We plan to address this issue in future publications.
[34]
C. Herring, in Physics of Powder Metallurgy,$\;$ p. 143. Ed. W.E. Kingston,
NY, McGraw-Hill 1951. |
TCDMATH 19-15
$T\overline{T}$, $\widetilde{J}J$, $JT$ and $\widetilde{J}T$ deformations
Sergey Frolov222 Correspondent fellow at
Steklov Mathematical Institute, Moscow.333email:
[email protected]
School of Mathematics and Hamilton Mathematics Institute,
Trinity College, Dublin 2,
Ireland
Abstract
The light-cone gauge approach to ${T\overline{T}}$ deformed models is generalised to models deformed by U(1) conserved currents $J^{\alpha}$, $\widetilde{J}^{\alpha}$, stress-energy tensor $T^{\alpha}{}_{\beta}$, and their various quadratic combinations of the form $\epsilon_{\alpha\beta}K_{1}^{\alpha}K_{2}^{\beta}$. It is then applied
to derive a ten-parameter deformed Hamiltonian for a system of scalars with an arbitrary potential, the flow equations for the Hamiltonian density, and the flow equations for the energy of the deformed model.
The flow equations disagree with the ones recently proposed
in arXiv:1903.07606.
The results obtained are applied to analyse a CFT with left- and right-moving conserved currents deformed by these operators. It is shown that with a proper choice of the parameter of the ${T\overline{T}}$ deformation the deformed CFT Hamiltonian density is independent of the parameters of the $J\Theta$ and $\bar{J}\overline{\Theta}$ deformations. This leads to the existence of two extra relations which generalise the
$J\Theta=0$ and $\bar{J}\overline{\Theta}=0$ relations of the undeformed CFT. The spectrum of the deformed CFT is found and shown to satisfy the flow equations.
Contents
1 Introduction
2 String model and the transformation
2.1 Extended string model
2.2 Parameter dependent canonical transformation
3 Deformed Hamiltonian and the flow equations
3.1 Deformed Hamiltonian
3.2 Flow equations for the density of the Hamiltonian
3.3 Flow equations for the energy
3.4 Homogeneous inviscid Burgers equation
4 Flow equations in CFT conventions
4.1 Flow equations for the Hamiltonian density
4.2 Flow equations for the energy in CFT conventions
4.3 Deformed CFT with left- and right-moving currents
5 Conclusions
A Canonical transformation matrices
B Virasoro constraints coefficients
C Examples of deformed Hamiltonians
D Examples of the deformed spectrum
1 Introduction
The ${T\overline{T}}$ deformation of 2d field theories introduced in [1] admits various generalisations if the model under consideration possesses additional conserved currents. Examples include the Lorentz invariance preserving higher-spin deformations of integrable models [2], and the ${\rm J}\overline{T}$ type deformations [3] which break Lorentz invariance of an undeformed model. Any such a deformation is obtained by adding to the Hamiltonian of an undeformed model an operator of the form
${\cal O}_{K_{1}K_{2}}\sim\epsilon_{\alpha\beta}K^{\alpha}_{1}K^{\beta}_{2}$ where $K_{1}$ and $K_{2}$ are conserved currents.
It is hoped that the spectrum of any model deformed by these operators is completely fixed by the spectrum of the undeformed model just as it is for the ${T\overline{T}}$ deformation [1]. Indeed, the spectrum of
a CFT with left- and right-moving conserved currents
deformed by ${\rm J}{\overline{T}}$ was derived in
[4], completing the results of [3], see also [5], and the spectrum of such a CFT deformed by ${T\overline{T}}$, ${\rm J}{\overline{T}}$, ${\bar{{\rm J}}T}$ was conjectured in [6] by relating the problem to deformations of holographically dual strings on $AdS_{3}$. A very general nine-parameter deformation of a CFT by the stress energy tensor $T$, the conserved currents ${\rm J}$ and $\bar{\rm J}$, and their various quadratic combinations was studied in [7] where a set of flow equations with respect to the deformation parameters was proposed, and then used to conjecture the spectrum of a deformed CFT.
In this paper we generalise the light-cone gauge approach [8, 9] to ${T\overline{T}}$ deformed models
to the case of the most general ten-parameter deformation
by conserved currents $J^{\alpha}$, $\widetilde{J}^{\alpha}$, the canonical Noether stress-energy tensor $T^{\alpha}{}_{\beta}$, and their various quadratic combinations of the form $\epsilon_{\alpha\beta}K^{\alpha}_{1}K^{\beta}_{2}$. The light-cone gauge approach was used in [9]
to derive the ${T\overline{T}}$ deformed action for a very general system of any number of scalars, fermions and chiral bosons with an arbitrary potential.
Basically all of the ${T\overline{T}}$ deformed models [10, 11, 12, 13, 8, 14, 15, 16, 17] studied before and after [9] are particular cases of this system which includes various Lorentz invariant systems of bosons and fermions, in particular supersymmetric sigma models, and some of non-Lorentz invariant systems, e.g. the chiral SYK model and the nonlinear matrix
Schrödinger model.
The light-cone gauge approach is based on the observation that
the homogeneous inviscid Burgers equation which determines the spectrum of a ${T\overline{T}}$ deformed model with zero momentum coincides with the gauge invariance condition of the target space-time energy and momentum of a non-critical string theory quantised in a parameter dependent light-cone gauge introduced in [18].
The light-cone gauge-fixed Hamiltonian can be thought of as the Hamiltonian of a deformed model, and the deformation parameter can be identified with the light-cone gauge parameter, see [9] for a detailed discussion.
The deformed Hamiltonian ${\cal H}_{\alpha}$ can be used to derive the flow equation,
$\partial_{\alpha}{\cal H}_{\alpha}=\,$${T\overline{T}}$, with respect to the parameter $\alpha$ of the ${T\overline{T}}$ deformation.
The flow equation is then used to get the inviscid Burgers equation which governs the spectrum of the deformed model.
In this paper we only consider a system of $n$ scalars with an arbitrary potential.
We assume that the bosonic model is invariant under shifts of one of its fields, say $x^{1}$, and $J^{\alpha}$ is its canonical Noether current due to the symmetry. The current
$\widetilde{J}^{\alpha}$ is a topological current associated to the field $x^{1}$: $\widetilde{J}^{\alpha}=\epsilon^{\alpha\beta}\partial_{\beta}x^{1}$. In the case of a CFT the left- and right-moving conserved currents are linear combinations of $J^{\alpha}$ and $\widetilde{J}^{\alpha}$.
We begin with the usual action for bosonic strings invariant under shifts of three isometry coordinates $x^{a}$, $a=+,-,1$. The undeformed model is obtained by imposing the standard light-cone gauge $x^{+}=\tau\,,\,p_{-}=1$ where $p_{-}$ is the momentum conjugate to $x^{-}$.
It appears that to describe the most general ten-parameter deformation we need to introduce pairs of auxiliary non-dynamical co-vectors and scalars $(V^{a}_{\alpha}\,,\,\tilde{X}_{a})$ and $({\cal U}_{a\alpha}\,,\,\Upsilon^{a})$ associated with the conserved currents $J_{a}^{\alpha}$ and $\widetilde{J}^{a\alpha}$, respectively. We then perform a ten-parameter canonical transformation $A$ which involves the three coordinates $x^{a}$ and their momenta $p_{a}$, and the coordinates and momenta of the auxiliary fields. Finally, we impose the light-cone gauge $x^{+}=\tau\,,\,p_{-}=1$ on the transformed coordinates and momenta, solve the Virasoro and Gauss-law constraints, and identify the light-cone gauge-fixed Hamiltonian ${\cal H}_{A}=-p_{+}$ with the deformed one.
The derivatives of the deformed Hamiltonian with respect to the ten parameters
give rise to flow equations (3.9-3.13) which are valid for any bosonic model of the type we study. We believe that the flow equations are universal, and apply
to any Lorentz and non-Lorentz invariant model of bosons and fermions.
The flow equations disagree with the ones recently proposed in [7].
The main difference is that some of the coefficients in front of the deforming operators on the r.h.s. of our flow equations depend on the deformation parameters of the quadratic operators while those of the flow equations of [7] do not.
The reason for the disagreement is simple. The authors of [7] only analysed the model of a single massless free scalar which is a CFT with left- and right-moving conserved currents. As we show, for such a model there is a choice of the ${T\overline{T}}$ deformation parameter such that the deformed Hamiltonian does not depend on the parameters of the ${\rm J}\Theta$ and $\bar{\rm J}\overline{\Theta}$ deformations.
The flow equations with respect to these parameters then become
extra constraints (4.11) which generalise the relations ${\rm J}\Theta=0$ and $\bar{\rm J}\overline{\Theta}=0$ of the undeformed CFT. Taking into account these constraints one can make a choice of the deformation parameters which removes the parameter dependent terms from the flow equations. Such a choice exists only for a CFT with left- and right-moving conserved currents, and for any other model one has to deal with the flow equations (3.9-3.13).
The flow equations for the deformed Hamiltonian can be used to find a system of flow equations (3.17, 3.18) for the spectrum of the deformed model. The system does not involve derivatives of the energy with respect to the parameters of the linear deformations by the space components of the currents $J^{\alpha}$ and $\tilde{J}^{\alpha}$. These derivatives were used in [7] to find the expectation values of $J^{1}$ and $\tilde{J}^{1}$. As we explain in the paper, these expectation values are given by the derivatives of the energy with respect to the charges $\widetilde{P}^{1}$ and $P_{1}$ of the currents $J^{\alpha}$ and $\tilde{J}^{\alpha}$. Thus, we do not really need the two parameters. The system of flow equations is complicated and at the moment we do not know how to find a proper solution of the system for a generic state. Nevertheless, we show that if a state satisfies level-matching conditions which we derive then the energy of such a state obeys a version of the
homogeneous inviscid Burgers equation.
The flow equations for the energy however can be solved for a deformed CFT with left- and right-moving conserved currents. The solution (4.18) we find reduces to the known ones in the particular cases of one- and two-parameter deformations. It seems to agree with the one proposed in [7]
at least if one switches off their parameters of the deformation by ${\rm Q}_{\pm}^{1}$, and parameters of the ${\rm J}\Theta$ and $\bar{\rm J}\overline{\Theta}$ deformations.
To get a precise agreement it might be necessary to perform an extra redefinition of the parameter of the ${T\overline{T}}$ deformation. We have not tried to do it.
In the case of the ${T\overline{T}}$, ${\rm J}{\overline{T}}$, ${\bar{{\rm J}}T}$ deformation our solution seems different from the one proposed in [6] but we believe that the two solutions may be equivalent after a proper redefinition of the deformation parameters.
The plan of the paper is as follows. In section 2.1 we introduce an extended action for bosonic strings propagating in a target manifold possessing (at least) three
abelian isometries and determine all the constraints and gauge conditions necessary to recover the undeformed model. In section 2.2 we generate a ten-parameter deformation of the model as a chain of one-parameter canonical transformations producing different deformations. In section 3.1 we impose the light-cone gauge on the transformed coordinates and momenta, solve all the constraints, and find the deformed Hamiltonian. In section 3.2 we use the Hamiltonian to derive the flow equations with respect to the deformation parameters. Then in section 3.3 the flow equations are converted into the flow equations for the spectrum of the deformed model.
In section 3.4 we determine level-matching conditions physical states must satisfy, and derive the
homogeneous inviscid Burgers equations for the energy of those states.
The flow equations for the Hamiltonian density and the spectrum are then rewritten in the CFT conventions in section 4 where we also consider a very special case of a CFT with left- and right-moving conserved currents, and show that for such a CFT there are extra constraints (4.11) replacing the relations
${\rm J}\Theta=0$ and $\bar{\rm J}\overline{\Theta}=0$ of the undeformed CFT. In section 4.3 we propose a formula for the spectrum of such a deformed CFT which satisfies the flow equations.
In Conclusions we discuss open questions and generalisations of the light-cone gauge approach. Finally, in Appendices we collect some explicit formulae, and consider examples of deformed Hamiltonians, and deformed spectrum of CFT with left- and right-moving conserved currents.
2 String model and the transformation
2.1 Extended string model
In this section we follow the same approach as in [9] but simplify a bit their consideration. We begin with bosonic strings propagating in a $n+2-$dimensional target Minkowski manifold ${\cal M}$ possessing
(at least) three abelian isometries realised by shifts of the isometry
coordinates.
We denote coordinates of ${\cal M}$ by $x^{M}$, $M=+,-,1,\ldots,n$,
and choose the isometry
coordinates to be $x^{+}$, $x^{-}$ and $x^{1}$.
The “transversal” coordinates $x^{\mu}$, $\mu=1,\ldots,n$ are the fields of the model we wish to deform.
Obviously, the target-space metric $G_{MN}$ of ${\cal M}$ does not depend on $x^{\pm}$ and $x^{1}$.
Assuming for simplicity that the $B$-field vanishes but making no assumption on the form of $G_{MN}$, we write
the initial string action in the standard form
$$\displaystyle S=\int_{0}^{R}\,{\rm d}\sigma{\rm d}\tau\,{\cal L}\,,\quad{\cal L%
}=-{1\over 2}\gamma^{\alpha\beta}\partial_{\alpha}x^{M}\partial_{\beta}x^{N}G_%
{MN}\,,$$
(2.1)
where $\gamma^{\alpha\beta}=h^{\alpha\beta}\sqrt{-h}$ is the Weyl-invariant
combination of the world-sheet
metric $h_{\alpha\beta}$ with $\det\gamma=-1$, and $\epsilon^{01}=-\epsilon_{01}=1$. The range $R$ of the world-sheet space coordinate $\sigma$ will be fixed by a generalised uniform light-cone gauge.
The string action invariance under the shifts of $x^{\pm}$ and $x^{1}$
leads to the existence of the three conserved currents
$$\displaystyle J_{a}^{\alpha}={\partial{\cal L}\over\partial\partial_{\alpha}x^%
{a}}\,,\quad\partial_{\alpha}J_{a}^{\alpha}=0\,,\quad a=+,-,1\,.$$
Then, to any coordinate $x^{M}$ we can associate a topological conserved current
$$\displaystyle\widetilde{J}^{M\alpha}=\epsilon^{\alpha\beta}\partial_{\beta}x^{%
M}\,,$$
and the deformations we discuss in this paper involve the topological current $\tilde{J}^{\alpha}\equiv\tilde{J}^{1\alpha}$
which for a free massless scalar field is dual to $J^{\alpha}\equiv J_{1}^{\alpha}$. We denote the corresponding conserved charges as
$$\displaystyle P_{a}=\int_{0}^{R}{\rm d}\sigma J_{a}^{0}\,,\quad\widetilde{P}^{%
a}=\int_{0}^{R}{\rm d}\sigma\widetilde{J}^{a0}\,.$$
It is clear that $\widetilde{P}^{a}$ does not vanish only if $x^{a}$ has a nonzero winding number.
To get the most general deformation we introduce pairs of non-dynamical co-vectors and scalars $(V^{a}_{\alpha}\,,\,\tilde{X}_{a})$ and $({\cal U}_{a\alpha}\,,\,\Upsilon^{a})$ associated with currents $J_{a}^{\alpha}$ and $\tilde{J}^{a\alpha}$, respectively. Then, the extended string action together with the new fields has the form
$$\displaystyle S=\int_{0}^{R}\,{\rm d}\sigma{\rm d}\tau\,\Big{(}-{1\over 2}%
\gamma^{\alpha\beta}\partial_{\alpha}x^{M}\partial_{\beta}x^{N}G_{MN}-\epsilon%
^{\alpha\beta}V_{\alpha}^{a}\partial_{\beta}\tilde{X}_{a}-\epsilon^{\alpha%
\beta}{\cal U}_{a\alpha}\partial_{\beta}\Upsilon^{a}\Big{)}\,.$$
(2.2)
The action (2.1) is obviously invariant under reparametrisations, the six U(1) gauge symmetries
$$\displaystyle V_{\alpha}^{a}\to V_{\alpha}^{a}+\partial_{\alpha}\xi^{a}\,,%
\quad\tilde{X}_{a}\to\tilde{X}_{a}\,,\qquad{\cal U}_{a\alpha}\to{\cal U}_{a%
\alpha}+\partial_{\alpha}\zeta_{a}\,,\quad\Upsilon^{a}\to\Upsilon^{a}\,,$$
(2.3)
and the constant shifts of the coordinates $\tilde{X}_{a}$ and $\Upsilon^{a}$
$$\displaystyle\tilde{X}_{a}\to\tilde{X}_{a}+c_{a}\,,\quad\Upsilon^{a}\to%
\Upsilon^{a}+\eta^{a}\,.$$
The last symmetry leads to the existence of six conserved currents
$$\displaystyle\partial_{\alpha}(\epsilon^{\alpha\beta}V_{\beta}^{a})=0\,,\quad%
\partial_{\alpha}(\epsilon^{\alpha\beta}{\cal U}_{a\beta})=0\,.$$
These are equations of motion for $V_{\alpha}^{a}$ and ${\cal U}_{a\alpha}$ which imply that
$$\displaystyle V_{\alpha}^{a}=\delta_{1\alpha}v^{a}+\partial_{\alpha}\xi^{a}\,,%
\quad{\cal U}_{a\alpha}=\delta_{1\alpha}u_{a}+\partial_{\alpha}\zeta_{a}\,,$$
where $v^{a}$ and $u_{a}$ are time-independent zero modes of $V_{1}^{a}$ and ${\cal U}_{a1}$, respectively. Thus, one gets a family of models parametrised by $v^{a}$ and $u_{a}$, and if they vanish then we get back to the original string model (2.1). We are going to impose a gauge condition which depends on ten parameters, and consider the resulting gauge-fixed action as a deformation of our favourite one.
The simplest way to impose such a gauge condition is to switch to the Hamiltonian formalism. Introducing the momenta canonically-conjugate to the coordinates $x^{M}$
$$\displaystyle p_{M}={\delta S\over\delta\dot{x}^{M}}=-\gamma^{0\beta}\partial_%
{\beta}x^{N}\,G_{MN}\,,\quad\dot{x}^{M}\equiv\partial_{0}x^{M}\,,$$
we bring the
string action (2.2) to the first-order form
$$\displaystyle S=\int_{0}^{R}\,{\rm d}\sigma{\rm d}\tau\,\left(p_{a}\dot{x}^{a}%
+V^{a}\dot{\tilde{X}}_{a}+{\cal U}_{a}\dot{\Upsilon}^{a}+p_{k}\dot{x}^{k}+{%
\gamma^{01}\over\gamma^{00}}C_{1}+{1\over\gamma^{00}}C_{2}-V_{0}^{a}C_{a}^{V}-%
{\cal U}_{a0}C^{a}_{\cal U}\right)\,.$$
(2.4)
Here $a=+,-,1$, $k=2,\ldots,n$, and $V^{a}\equiv V^{a}_{1}$, $\,{\cal U}_{a}\equiv{\cal U}_{a1}$. Then,
$$\displaystyle C_{1}$$
$$\displaystyle=p_{M}x^{\prime M}=p_{a}x^{\prime a}+p_{k}x^{\prime k}\,,\quad x^%
{\prime M}\equiv\partial_{1}x^{M}\,.$$
$$\displaystyle C_{2}$$
$$\displaystyle={1\over 2}G^{ab}p_{a}p_{b}+{1\over 2}x^{\prime a}x^{\prime b}G_{%
ab}+G^{ak}p_{a}p_{k}+x^{\prime a}x^{\prime k}G_{ak}+\widetilde{\cal H}_{x}\,,$$
are the Virasoro constraints,
and $\widetilde{\cal H}_{x}$
depends only on the transversal fields $x^{k}$ and $p_{k}$
$$\displaystyle\widetilde{\cal H}_{x}={1\over 2}G^{kl}p_{k}p_{l}+{1\over 2}x^{%
\prime k}x^{\prime l}\,G_{kl}\,,\quad k,l=2,\ldots,n\,.$$
(2.5)
Finally,
$$\displaystyle C_{a}^{V}=\tilde{X}_{a}^{\prime}\,,\quad C^{a}_{\cal U}=\Upsilon%
^{\prime a}\,,\quad a=+,-,1\,,$$
are the constraints which generate the U(1) gauge transformations (2.3). It is clear from (2.4) that $V^{a}$ and ${\cal U}_{a}$ are the momenta canonically-conjugate to the coordinates $\tilde{X}_{a}$ and $\Upsilon^{a}$. It is worthwhile to mention that on the constraints surface one can replace
$$\displaystyle C_{1}\to\tilde{C}_{1}=p_{a}x^{\prime a}+V^{a}\tilde{X}_{a}^{%
\prime}+{\cal U}_{a}\Upsilon^{\prime a}+p_{k}x^{\prime k}\,.$$
(2.6)
This simple observation will be useful later.
If we do nothing with the coordinates and momenta of the extended model, and just impose the light-cone gauge
conditions
$$\displaystyle x^{+}=\tau\,,\quad p_{-}=1\,,\quad V^{\prime a}=0\,,\quad{\cal U%
}^{\prime}_{a}=0\,,$$
(2.7)
then, solving the constraints $C_{a}^{V}=0$ and $C^{a}_{\cal U}=0$, we get that $V^{a}=v^{a}$, $\tilde{X}_{a}=\tilde{x}_{a}$, ${\cal U}_{a}=u_{a}$ and $\Upsilon^{a}=\mbox{\scriptsize$\Upsilon$}^{a}$ are $\sigma$-independent.
Using the Virasoro
constraint $C_{1}$ to find $x^{\prime-}$
$$\displaystyle x^{\prime-}=-p_{\mu}x^{\prime\mu}\equiv-px^{\prime}\,,$$
(2.8)
we bring $C_{2}$ to the form
$$\displaystyle C_{2}$$
$$\displaystyle={G^{++}\over 2}p_{+}^{2}+(G^{-+}+G^{\mu+}p_{\mu})p_{+}+{G^{--}%
\over 2}+G^{\mu-}p_{\mu}+{G_{--}\over 2}(px^{\prime})^{2}-G_{\mu-}x^{\prime\mu%
}px^{\prime}+{\cal H}_{x}\,,$$
(2.9)
$$\displaystyle{\cal H}_{x}$$
$$\displaystyle={1\over 2}G^{\mu\nu}p_{\mu}p_{\nu}+{1\over 2}x^{\prime\mu}x^{%
\prime\nu}\,G_{\mu\nu}\,.$$
Solving the resulting equation $C_{2}=0$ for $p_{+}$,
we find the gauge-fixed action of the extended model
$$\displaystyle S_{0}=\int{\rm d}\tau\,\left(v^{a}\dot{\tilde{x}}_{a}+u_{a}\dot{%
\mbox{\scriptsize$\Upsilon$}}^{a}\right)+\int_{0}^{R}\,{\rm d}\sigma{\rm d}%
\tau\,\left(p_{\mu}\dot{x}^{\mu}\,-\,{\cal H}_{0}\right)\,,\quad{\cal H}_{0}\,%
=\,-p_{+}(p_{\mu},x^{\mu},x^{\prime\mu})\,.$$
(2.10)
Here the integration range $R=P_{-}$ is found by integrating the gauge condition $p_{-}=1$ over $\sigma$, and
the density ${\cal H}_{0}$ of the world-sheet Hamiltonian depends on the periodic
transversal fields. Thus, the gauge-fixed string
action describes a two-dimensional model on a cylinder of
circumference $R=P_{-}$. Obviously, $v^{a}$ and $u_{a}$ are time-independent, and setting them to any constants, one gets the usual light-cone gauge string action.
The gauge-fixed model in general is not Lorentz invariant, and, as was shown in [9], to get a Lorentz invariant model one has to choose the target space metric of the form
$$\displaystyle ds^{2}=dx^{+}dx^{-}-2Vdx^{+}dx^{+}+G_{\mu\nu}dx^{\mu}dx^{\nu}\,,$$
(2.11)
where $V$ is an arbitrary function of the transversal coordinates. Then, ${\cal H}_{0}$ becomes the Hamiltonian density of a sigma-model of $n$ scalar fields with the
potential $V$
$$\displaystyle{\cal H}_{0}={\cal H}_{x}+V(x)\,.$$
(2.12)
We are going to perform a parameter dependent canonical transformation of the coordinates $x^{a}$, $\tilde{X}_{a}$, $\Upsilon^{a}$ and momenta $p_{a}$, $V^{a}$, ${\cal U}_{a}$, and impose the light-cone gauge conditions (2.7) on the transformed $x^{+}$, $p_{-}$, $V^{a}$ and ${\cal U}_{a}$.
Then, solving the constraints, one gets a parameter-dependent gauge-fixed Hamiltonian density ${\cal H}_{A}$ where $A$ labels the parameters.
We want to think about this gauge-fixed model as a deformation of the model with the Hamiltonian density ${\cal H}_{0}$, and consider the parameters of the canonical transformation as the deformation parameters.
The deformed two-dimensional model with the Hamiltonian density ${\cal H}_{A}$ is invariant under the shifts of
the world-sheet coordinates $\tau$ and $\sigma$, and its canonical stress-energy tensor is given by
$$\displaystyle T^{0}{}_{0}$$
$$\displaystyle={\cal H}_{A}\,,\quad T^{1}{}_{0}=-{\partial{\cal H}_{A}\over%
\partial x^{\prime\mu}}{\partial{\cal H}_{A}\over\partial p_{\mu}}\,,$$
$$\displaystyle T^{0}{}_{1}$$
$$\displaystyle=p_{\mu}x^{\prime\mu}\,,\quad T^{1}{}_{1}={\cal H}_{A}-{\partial{%
\cal H}_{A}\over\partial x^{\prime\mu}}x^{\prime\mu}-p_{\mu}{\partial{\cal H}_%
{A}\over\partial p_{\mu}}\,.$$
In particular, the
world-sheet energy and momentum are conserved
$$\displaystyle E=\int_{0}^{R}\,{\rm d}\sigma\,{\cal H}_{A}=-P_{+}\,,\quad P=-%
\int_{0}^{R}\,{\rm d}\sigma\,p_{\mu}x^{\prime\mu}\,.$$
Then, the Hamiltonian density is invariant under a shift of the coordinate $x^{1}$, and the corresponding conserved current is
$$\displaystyle J^{0}=p_{1}\,,\quad J^{1}=-{\partial{\cal H}_{A}\over\partial x^%
{\prime 1}}\,.$$
Finally, the topological current in the Hamiltonian formalism is given by
$$\displaystyle\widetilde{J}^{0}=x^{\prime 1}\,,\quad\widetilde{J}^{1}=-\dot{x}^%
{1}=-{\partial{\cal H}_{A}\over\partial p_{1}}\,.$$
2.2 Parameter dependent canonical transformation
The simplest way to find a canonical transformation generating a multi-parameter deformation of ${\cal H}_{0}$ is to realise it as a chain of one-parameter transformations producing different deformations.
Analysing infinitesimal transformations, one can identify suitable one-parameter ones. We find convenient to use the following consecutive ones
1. Transformation, ${\cal T}_{T\bar{T}}$, generating the ${T\overline{T}}$deformation [9]
$$\displaystyle{\cal T}_{T\overline{T}}:\quad x^{+}\to x^{+}-a_{+-}x^{-}\,,\quad
x%
^{-}\to x^{-}\,,\quad p_{+}\to p_{+}\,,\quad p_{-}\to p_{-}+p_{+}a_{+-}\,.$$
It is easy to check that for any target space metric $G_{MN}$ the gauge-fixed Hamiltonian density ${\cal H}_{a_{+-}}$ satisfies the flow equation
$$\displaystyle{\partial{\cal H}_{a_{+-}}\over\partial a_{+-}}={T\overline{T}}\,%
,\quad{T\overline{T}}\equiv T^{0}{}_{1}T^{1}{}_{0}-T^{1}{}_{1}T^{0}{}_{0}=-%
\epsilon_{\alpha\beta}T^{\alpha}{}_{1}T^{\beta}{}_{0}=T_{1}T_{0}\,.$$
2. Transformation, ${\cal T}_{\tilde{J}T_{0}}$, generating the $-\widetilde{J}T_{0}$ deformation
$$\displaystyle{\cal T}_{\tilde{J}T_{0}}:\quad x^{+}\to x^{+}-a_{+1}x^{1}\,,%
\quad x^{1}\to x^{1}\,,\quad p_{+}\to p_{+}\,,\quad p_{1}\to p_{1}+p_{+}a_{+1}\,,$$
$$\displaystyle{\partial{\cal H}_{a_{+1}}\over\partial a_{+1}}=-\widetilde{J}T_{%
0}\,,\quad\widetilde{J}T_{0}\equiv\widetilde{J}^{0}T^{1}{}_{0}-\widetilde{J}^{%
1}T^{0}{}_{0}=-\epsilon_{\alpha\beta}\widetilde{J}^{\alpha}T^{\beta}{}_{0}\,.$$
3. Transformation, ${\cal T}_{JT_{1}}$, generating the $JT_{1}$ deformation
$$\displaystyle{\cal T}_{JT_{1}}:\quad x^{1}\to x^{1}-a_{1-}x^{-}\,,\quad x^{-}%
\to x^{-}\,,\quad p_{1}\to p_{1}\,,\quad p_{-}\to p_{-}+p_{1}a_{1-}\,,$$
$$\displaystyle{\partial{\cal H}_{a_{1-}}\over\partial a_{1-}}=JT_{1}\,,\quad JT%
_{1}\equiv J^{0}T^{1}{}_{1}-J^{1}T^{0}{}_{1}=-\epsilon_{\alpha\beta}J^{\alpha}%
T^{\beta}{}_{1}\,.$$
The sequence ${\cal T}_{JT_{1}}{\cal T}_{\tilde{J}T_{0}}{\cal T}_{T\overline{T}}$ of these three transformations can be represented by a single canonical transformation
$$\displaystyle x^{a}\to(A^{-1})^{a}{}_{b}\,x^{b}\,,\quad p_{a}\to p_{b}\,A^{b}{%
}_{a}\,,\quad a,b=+,-,1\,.$$
Here $A$ is the following matrix
$$\displaystyle A=\left(\begin{array}[]{ccc}1&a_{+-}&a_{+1}\\
0&1&0\\
0&a_{1-}&1\\
\end{array}\right)\,,\ A^{-1}=\left(\begin{array}[]{ccc}1&-\tilde{a}_{+-}&{-a_%
{+1}}\\
&{1}&0\\
0&-a_{1-}&1\\
\end{array}\right)\,,$$
where
$$\displaystyle\tilde{a}_{+-}=a_{+-}-a_{+1}a_{1-}\,.$$
In general $A$ can be any nondegenerate real matrix with entries $A_{ij}=\delta_{ij}+a_{ij}$, and one can easily show that
(i) $A_{++}$ deforms by $-T^{0}{}_{0}$:
$A_{++}\partial_{A_{++}}{\cal H}_{A_{++}}=-T^{0}{}_{0}=-{\cal H}_{A_{++}}$;
(ii) $A_{--}$ deforms by $T^{1}{}_{1}$:
$A_{--}\partial_{A_{--}}{\cal H}_{A_{--}}=T^{1}{}_{1}$;
(iii) $A_{11}$ deforms by $\widetilde{J}J$:
$A_{11}{\partial{\cal H}_{A_{11}}\over\partial A_{11}}={\widetilde{J}J}$;
(iv) $a_{-+}$ deforms by $1$:
$\partial_{a_{-+}}{\cal H}_{a_{-+}}=1$;
(v) $a_{-1}$ deforms by $-\widetilde{J}^{1}$:
$\partial_{a_{-1}}{\cal H}_{a_{-1}}=-\widetilde{J}^{1}$;
(vi) $a_{1+}$ deforms by $J^{0}=p_{1}$:
$\partial_{a_{1+}}{\cal H}_{a_{1+}}=J^{0}$.
We do not need the deformations by $T^{0}{}_{\alpha}$, $J^{0}$ and $\widetilde{J}^{0}$ because they just change the energy by the corresponding charges.
We could have used the transformations with $A_{--}$ and $A_{11}$ at the next steps but this leads to complicated flow equations when all ten parameters are nonvanishing, and a careful analysis shows that it is better to use these transformations (which commute) at the very end. As to the transformation with $a_{-1}$, it appears to be easier to use the auxiliary non-dynamical fields to generate the deformation by $\widetilde{J}^{1}$ and the remaining four deformations.
The remaining transformations are
4. Transformation, ${\cal T}_{T^{1}{}_{0}}$, generating the $-T^{1}{}_{0}$ deformation
$$\displaystyle{\cal T}_{T^{1}{}_{0}}:\quad x^{\prime+}\to x^{\prime+}-V^{+}\,,%
\quad\tilde{X}^{\prime}_{+}\to\tilde{X}^{\prime}_{+}+p_{+}\,,\quad p_{+}\to p_%
{+}\,,\quad V^{+}\to V^{+}\,,$$
$$\displaystyle{\partial{\cal H}_{v^{+}}\over\partial v^{+}}=-T^{1}{}_{0}\,.$$
This transformation obviously represents gauging the shift symmetry of $x^{+}$, and due to the light-cone gauge condition $x^{+}=\tau$ it is natural that the zero mode of $V^{+}$ generates the deformation by $-T^{1}{}_{0}$.
5. Transformation, ${\cal T}_{J^{1}}$, generating the $J^{1}$ deformation
$$\displaystyle{\cal T}_{J^{1}}:\quad x^{\prime 1}\to x^{\prime 1}-V^{1}\,,\quad%
\tilde{X}^{\prime}_{1}\to\tilde{X}^{\prime}_{1}+p_{1}\,,\quad p_{1}\to p_{1}\,%
,\quad V^{1}\to V^{1}\,,$$
$$\displaystyle{\partial{\cal H}_{v^{1}}\over\partial v^{1}}=J^{1}\,.$$
Gauging the shift symmetry of $x^{1}$ leads to the dependence of ${\cal H}_{v^{1}}$ on $x^{\prime 1}-v^{1}$.
6. Transformation, ${\cal T}_{JT_{0}}$, generating the $JT_{0}$ deformation
$$\displaystyle{\cal T}_{JT_{0}}:\quad x^{+}\to x^{+}-b_{+1}\tilde{X}_{1}\,,%
\quad\tilde{X}_{1}\to\tilde{X}_{1}\,,\quad p_{+}\to p_{+}\,,\quad V^{1}\to V^{%
1}+p_{+}b_{+1}\,,$$
$$\displaystyle{\partial{\cal H}_{b_{+1}}\over\partial b_{+1}}=JT_{0}\,,\quad JT%
_{0}\equiv J^{0}T^{1}{}_{0}-J^{1}T^{0}{}_{0}=-\epsilon_{\alpha\beta}J^{\alpha}%
T^{\beta}{}_{0}\,.$$
This transformation must be performed after the transformation ${\cal T}_{J^{1}}$.
7. Transformation, ${\cal T}_{\tilde{J}^{1}}$, generating the $\widetilde{J}^{1}$ deformation
$$\displaystyle{\cal T}_{\widetilde{J}^{1}}:\quad x^{1}\to x^{1}\,,\quad\Upsilon%
^{1}\to\Upsilon^{1}+x^{1}\,,\quad p_{1}\to p_{1}-{\cal U}_{1}\,,\quad{\cal U}_%
{1}\to{\cal U}_{1}\,,$$
$$\displaystyle{\partial{\cal H}_{u_{1}}\over\partial u_{1}}=\widetilde{J}^{1}\,.$$
This transformation in fact represents gauging the symmetry generated by
$$\displaystyle\widetilde{Q}(\xi)=\int d\sigma\widetilde{J}^{0}\xi=\int d\sigma x%
^{\prime 1}\xi\,,$$
which is trivial unless $\xi$ depends on $\sigma$.
This leads to the dependence of ${\cal H}_{u_{1}}$ on $p_{1}-u_{1}$.
8. Transformation, ${\cal T}_{\widetilde{J}T_{1}}$, generating the $\widetilde{J}T_{1}$ deformation
$$\displaystyle{\cal T}_{\widetilde{J}T_{1}}:$$
$$\displaystyle\tilde{X}_{-}\to\tilde{X}_{-}+p_{-}-{b_{--}\over c_{1-}}(x^{-}-V^%
{-})+b_{--}(x^{\prime 1}-V^{1})\,,\quad\tilde{X}_{1}\to\tilde{X}_{1}+c_{1-}%
\tilde{X}_{-}\,,$$
$$\displaystyle x^{\prime-}\to x^{\prime-}-V^{-}-c_{1-}(x^{\prime 1}-V^{1})\,,$$
$$\displaystyle p_{-}\to p_{-}+V^{-}{b_{--}\over c_{1-}}+b_{--}(x^{\prime 1}-V^{%
1})\,,\quad p_{1}\to p_{1}-c_{1-}\tilde{X}^{\prime}_{-}-c_{1-}\,,$$
$$\displaystyle V^{-}\to V^{-}+c_{1-}(x^{\prime 1}-V^{1})\,,$$
$$\displaystyle{\partial{\cal H}_{b_{--},c_{1-}}\over\partial b_{--}}=\widetilde%
{J}T_{1}\,,\quad{\partial{\cal H}_{b_{--},c_{1-}}\over\partial c_{1-}}=0\,,%
\quad\widetilde{J}T_{1}\equiv\widetilde{J}^{0}T^{1}{}_{1}-\widetilde{J}^{1}T^{%
0}{}_{1}=-\epsilon_{\alpha\beta}\widetilde{J}^{\alpha}T^{\beta}{}_{1}\,,$$
where in the gauge-fixed Hamiltonian we set $v^{-}=0$.
Note that this transformation requires both $b_{--}$ and $c_{1-}$ to be nonvanishing. This is the most complicated transformation. It can be represented as a combination of three simpler transformations
$$\displaystyle{\cal T}_{\tilde{J}T_{1}}={\cal T}_{3}{\cal T}_{2}{\cal T}_{1}\,,$$
where ${\cal T}_{1}$ gauges the shift symmetry of $x^{-}$
$$\displaystyle{\cal T}_{1}:\quad x^{\prime-}\to x^{\prime-}-V^{-}\,,\quad\tilde%
{X}^{\prime}_{-}\to\tilde{X}^{\prime}_{-}+p_{-}\,,\quad p_{-}\to p_{-}\,,\quad
V%
^{-}\to V^{-}\,,$$
${\cal T}_{2}$ is a linear transformation
$$\displaystyle{\cal T}_{2}:\quad\tilde{X}^{\prime}_{-}\to\tilde{X}^{\prime}_{-}%
-{b_{--}\over c_{1-}}x^{-}\,,\quad p_{-}\to p_{-}+V^{-}{b_{--}\over c_{1-}}\,,$$
and ${\cal T}_{3}$ is a transformation we call twisting $V^{-}$ with $x^{\prime 1}$
$$\displaystyle{\cal T}_{3}:$$
$$\displaystyle\tilde{X}_{1}\to\tilde{X}_{1}+c_{1-}\tilde{X}_{-}\,,\quad p_{1}%
\to p_{1}-c_{1-}\tilde{X}^{\prime}_{-}-c_{1-}\,,\quad V^{-}\to V^{-}+c_{1-}(x^%
{\prime 1}-V^{1})\,.$$
We could not find a simpler way to generate the $\widetilde{J}T_{1}$ deformation.
However, as we will see soon, setting $v^{-}=0$, one can take the limit $c_{1-}\to 0$ in the gauge-fixed Hamiltonian density, and this simplifies the flow equations drastically.
9. Transformation, ${\cal T}_{\widetilde{J}J}$, generating the $\widetilde{J}J$ deformation
$$\displaystyle{\cal T}_{\widetilde{J}J}:$$
$$\displaystyle x^{1}\to{1\over A_{11}}x^{1}\,,\quad p_{1}\to A_{11}p_{1}\,,%
\quad\tilde{X}_{1}\to A_{11}\tilde{X}_{1}\,,\quad V^{1}\to{1\over A_{11}}V^{1}\,,$$
(2.13)
$$\displaystyle\Upsilon^{1}\to{1\over A_{11}}\Upsilon^{1}\,,\quad{\cal U}_{1}\to
A%
_{11}{\cal U}_{1}\,,$$
$$\displaystyle A_{11}{\partial{\cal H}_{A_{11}}\over\partial A_{11}}={%
\widetilde{J}J}\,,\quad\widetilde{J}J\equiv\widetilde{J}^{0}J^{1}-\widetilde{J%
}^{1}J^{0}=-\epsilon_{\alpha\beta}\widetilde{J}^{\alpha}J^{\beta}\,.$$
The rescaling of $\tilde{X}_{1}$, $V^{1}$ and $\Upsilon^{1}$, ${\cal U}^{1}$ is necessary for a gauge-fixed Hamiltonian density to depend on the differences $x^{\prime 1}-v^{1}$ and $p_{1}-u_{1}$.
10. Transformation, ${\cal T}_{T^{1}{}_{1}}$, generating the $T^{1}{}_{1}$ deformation
$$\displaystyle{\cal T}_{T^{1}{}_{1}}:\quad x^{-}\to{1\over A_{--}}x^{-}\,,\quad
p%
_{-}\to p_{-}A_{--}\,,$$
$$\displaystyle A_{--}{\partial{\cal H}_{A_{--}}\over\partial A_{--}}=T^{1}{}_{1%
}\,.$$
Due to the light-cone gauge condition $p_{-}=1$ the deformation with $A_{--}$ is equivalent to a change of the circumference of the cylinder the gauge-fixed model lives on.
Now, we can generate a ten-parameter transformation by using the following sequence of the one-parameter ones
$$\displaystyle{\cal T}_{A}={\cal T}_{T^{1}{}_{1}}{\cal T}_{\widetilde{J}J}{\cal
T%
}_{\widetilde{J}T_{1}}{\cal T}_{\widetilde{J}^{1}}{\cal T}_{JT_{0}}{\cal T}_{J%
^{1}}{\cal T}_{T^{1}{}_{0}}{\cal T}_{JT_{1}}{\cal T}_{\widetilde{J}T_{0}}{\cal
T%
}_{T\overline{T}}\,.$$
(2.14)
We see that to generate all these deformations we need all three pairs of the auxiliary fields
$(V^{a}_{\alpha}\,,\,\tilde{X}_{a})$ and the pair $({\cal U}_{1\alpha}\,,\,\Upsilon^{1})$. Thus, in what follows we set
$${\cal U}_{+\alpha}=0\,,\,\Upsilon^{+}=0\,,\quad{\cal U}_{-\alpha}=0\,,\,%
\Upsilon^{-}=0\,.$$
The transformation (2.14) can be represented by a single canonical transformation.
Introducing a column of coordinates
$$\displaystyle{\cal X}=\big{\{}{\cal X}^{\hat{1}}\,{\cal X}^{\hat{2}},\ldots,{%
\cal X}^{\hat{7}}\big{\}}=\big{\{}x^{+},x^{-},x^{1},\tilde{X}_{+},\tilde{X}_{-%
},\tilde{X}_{1},\Upsilon^{1}\big{\}}\,,$$
and a column of momenta
$$\displaystyle{\cal P}=\big{\{}{\cal P}_{\hat{1}},{\cal P}_{\hat{2}},\ldots,{%
\cal P}_{\hat{7}}\big{\}}=\big{\{}p_{+},p_{-},p_{1},V^{+},V^{-},V^{1},{\cal U}%
_{1}\big{\}}\,,$$
the transformation can be written in the form
$$\displaystyle{\cal X}^{\prime}\to A_{x}{\cal X}^{\prime}+B_{x}{\cal P}+C_{x}\,%
,\quad{\cal P}\to A_{p}^{T}{\cal P}+B_{p}^{T}{\cal X}^{\prime}+C_{p}\,.$$
(2.15)
Here the matrices $A_{x}$, $B_{x}$ , $A_{p}$ and $B_{p}$, and the columns $C_{x}$, $C_{p}$ satisfy the identities
$$\displaystyle A_{x}A_{p}+B_{x}B_{p}={\mathbb{I}}\,,\quad A_{x}B^{T}_{x}+B_{x}A%
_{x}^{T}=0\,,\quad A_{p}^{T}B_{p}+B_{p}^{T}A_{p}=0\,,$$
which are necessary for the transformation to be canonical, and additional identities
$$\displaystyle A_{p}A_{x}+B_{x}^{T}B_{p}^{T}$$
$$\displaystyle={\mathbb{I}}\,,\quad A_{p}B_{x}+(A_{p}B_{x})^{T}=0\,,\quad B_{p}%
A_{x}+(B_{p}A_{x})^{T}=0\,,$$
$$\displaystyle C_{p}^{T}A_{x}{\cal X}^{\prime}+{\cal X}^{\prime T}B_{p}C_{x}$$
$$\displaystyle=-{c_{1-}\over A_{11}}x^{\prime 1}\,,\quad C_{p}B_{x}{\cal P}+{%
\cal P}^{T}A_{p}C_{x}=0\,,\quad C_{p}^{T}C_{x}=0\,,$$
which lead to ${\cal P}^{T}{\cal X}^{\prime}\to{\cal P}^{T}{\cal X}^{\prime}-{c_{1-}\over A_{%
11}}x^{\prime 1}$.
The explicit form of the matrices can be found in appendix A.
3 Deformed Hamiltonian and the flow equations
3.1 Deformed Hamiltonian
Now, we can derive the gauge-fixed Hamiltonian density ${\cal H}_{A}$ which we consider as a deformation of ${\cal H}_{0}$.
First, we note that up to total derivatives the kinetic term in the action (2.4) does not change under the canonical transformation (2.15).
Then, as was mentioned before on the constraints surface the constraint $C_{1}$ is equivalent to $\tilde{C}_{1}$ (2.6). Clearly,
under (2.15), $\tilde{C}_{1}$ changes as
$$\displaystyle\tilde{C}_{1}\overset{{\cal T}_{A}}{\longrightarrow}\tilde{C}_{1}%
={\cal P}^{T}{\cal X}^{\prime}-{c_{1-}\over A_{11}}x^{\prime 1}+p_{k}x^{\prime
k%
}\,.$$
Then, we obviously have
$$\displaystyle C_{a}^{V}\overset{{\cal T}_{A}}{\longrightarrow}C_{a}^{V}=(A_{x}%
{\cal X}^{\prime}+B_{x}{\cal P}+C_{x})^{\hat{3}+\hat{a}}=(A_{x})^{\hat{3}+\hat%
{a}}{}_{r}{\cal X}^{\prime r}+(B_{x})^{\hat{3}+\hat{a},r}{\cal P}_{r}+C_{x}^{%
\hat{3}+\hat{a}}\,.$$
$$\displaystyle C^{1}_{\cal U}\overset{{\cal T}_{A}}{\longrightarrow}C^{1}_{\cal
U%
}=(A_{x}{\cal X}^{\prime}+B_{x}{\cal P})^{\hat{7}}=(A_{x})^{\hat{7}}{}_{r}{%
\cal X}^{\prime r}+(B_{x})^{\hat{7},r}{\cal P}_{r}\,,\quad r=\hat{1},\hat{2},%
\ldots,\hat{7}\,.$$
Finally, a simple calculation gives the transformed Virasoro constraint $C_{2}$
$$\displaystyle C_{2}\overset{{\cal T}_{A}}{\longrightarrow}C_{2}$$
$$\displaystyle={1\over 2}\tilde{G}^{rs}{\cal P}_{r}{\cal P}_{s}+{1\over 2}{\cal
X%
}^{\prime r}{\cal X}^{\prime s}\tilde{G}_{rs}+\tilde{G}^{rk}{\cal P}_{r}p_{k}+%
{\cal X}^{\prime r}x^{\prime k}\tilde{G}_{rk}$$
(3.1)
$$\displaystyle+\hat{G}^{r}{}_{s}{\cal P}_{r}{\cal X}^{\prime s}+\check{G}^{r}{}%
_{k}{\cal P}_{r}x^{\prime k}+{\cal X}^{\prime r}p_{k}\check{G}_{r}{}^{k}+%
\tilde{\cal H}_{x}\,.$$
The coefficients in this equation are listed in appendix B.
Now, we impose the light-cone gauge conditions (2.7) on the transformed coordinates and momenta. Then, from the analysis of
the canonical transformation (2.15) we know that the zero mode $v^{-}$ does not generate any of the deformations. Thus, in addition to (2.7) we can set it to 0 (and we have already set ${\cal U}_{\pm\alpha}=\Upsilon^{\pm}=0$)
$$\displaystyle v^{-}=0\,.$$
Next, we solve the constraints $\tilde{C}_{1}=0$, $C_{a}^{V}=0$ and $C^{1}_{\cal U}=0$ for
$$x^{\prime-}\,,\,\tilde{X}^{\prime}_{+}\,,\,\tilde{X}^{\prime}_{-}\,,\,\tilde{X%
}^{\prime}_{1}\,,\,\Upsilon^{\prime 1}\,.$$
The solution can be written in the form
$$\displaystyle{\cal X}^{\prime r}=\varkappa^{r}\,p_{+}+\chi^{r}\,,$$
(3.2)
where we also have ${\cal X}^{\prime\hat{1}}=x^{\prime+}=0$ and ${\cal X}^{\prime\hat{3}}=x^{\prime 1}$. The coefficients $\varkappa^{r}$ and $\chi^{r}$ can be read off from eqs.(B.1).
Using the solution,
we bring $C_{2}$ to the form $C_{2}={\cal G}_{2}\,p_{+}^{2}+{\cal G}_{1}\,p_{+}+{\cal G}_{0}$,
where the coefficients ${\cal G}_{i}$ are listed in appendix B.
One then checks that even though the transformation (2.15) is singular in the limit $c_{1-}\to 0$, the coefficients ${\cal G}_{i}$ are regular at $c_{1-}=0$. In fact it is necessary to have $v^{-}=0$ for the regularity. Thus, in what follows we set $c_{1-}=0$, and only discuss this case. If $c_{1-}\neq 0$ then there should exist a redefinition of the parameters which makes the gauge-fixed Hamiltonian independent of $c_{1-}$.
The solution of the quadratic equation
$$\displaystyle C_{2}$$
$$\displaystyle={\cal G}_{2}\,p_{+}^{2}+{\cal G}_{1}\,p_{+}+{\cal G}_{0}=0\,\,,$$
(3.3)
which reduces to $-{\cal H}_{0}$ when all the parameters vanish gives us the deformed Hamiltonian and action
$$\displaystyle S_{A}=\int_{0}^{R}\,{\rm d}\sigma{\rm d}\tau\,\left(p_{\mu}\dot{%
x}^{\mu}\,-\,{\cal H}_{A}\right)\,,\quad{\cal H}_{A}\,=\,-p_{+}(p_{\mu},x^{\mu%
},x^{\prime\mu})\,,\quad R={P_{-}}\,.$$
(3.4)
We consider some deformations of a model with the Hamiltonian
(2.12) in appendix C.
3.2 Flow equations for the density of the Hamiltonian
To simplify the understanding of the origin of various terms in the flow equations in this subsection we use the notations
$$\displaystyle\alpha_{\mbox{\tiny${T\overline{T}}$}}\equiv a_{+-}\,,\quad\alpha%
_{\mbox{\tiny${\widetilde{J}T_{0}}$}}\equiv-a_{+1}\,,\quad\alpha_{\mbox{\tiny$%
{JT_{1}}$}}\equiv a_{1-}\,,\quad\alpha_{\mbox{\tiny${JT_{0}}$}}\equiv b_{+1}\,%
,\quad\alpha_{\mbox{\tiny${\widetilde{J}T_{1}}$}}\equiv b_{--}\,,\quad v\equiv%
-v^{+}\,,$$
(3.5)
so that the flow equations at the leading order would have the form
$$\displaystyle{\partial{\cal H}_{A}\over\partial\alpha_{\cal O}}={\cal O}\,,%
\quad{\partial{\cal H}_{A}\over\partial v}=T^{1}{}_{0}\,.$$
The Hamiltonian density ${\cal H}_{A}$ (or, better to say the defining equation (3.3)) can be used to derive the flow equations with respect to the ten deformation parameters.
First of all, it is easy to check that the coefficients ${\cal G}_{i}$, and, therefore, ${\cal H}_{A}$, depend on the differences $x^{\prime 1}-v^{1}$ and $p_{1}-u_{1}$. Thus,
$$\displaystyle{\partial{\cal H}_{A}\over\partial v^{1}}=J^{1}\,,\quad{\partial{%
\cal H}_{A}\over\partial u_{1}}=\widetilde{J}^{1}\,.$$
(3.6)
Moreover, all the operators appearing on the right hand side of the flow equations also must depend on $x^{\prime 1}-v^{1}$ and $p_{1}-u_{1}$. Let us introduce the following improved currents ${\mathbb{J}}^{\alpha}$, $\widetilde{\mathbb{J}}^{\alpha}$ which depend on those differences
$$\displaystyle{\mathbb{J}}^{0}=p_{1}-u_{1}=J^{0}-u_{1}\,,\quad{\mathbb{J}}^{1}=%
J^{1}\,,$$
(3.7)
$$\displaystyle\widetilde{\mathbb{J}}^{0}=x^{\prime 1}-v^{1}=\widetilde{J}^{0}-v%
^{1}\,,\quad\widetilde{\mathbb{J}}^{1}=\widetilde{J}^{1}\,,$$
and the improved stress-energy tensor ${\mathbb{T}}^{\alpha}{}_{\beta}$
$$\displaystyle{\mathbb{T}}^{0}{}_{0}=T^{0}{}_{0}\,,\quad{\mathbb{T}}^{1}{}_{0}=%
T^{1}{}_{0}\,,$$
$$\displaystyle{\mathbb{T}}^{0}{}_{1}$$
$$\displaystyle=(p_{1}-u_{1})(x^{\prime 1}-v^{1})+p_{k}x^{\prime k}-v\,T^{0}{}_{%
0}=T^{0}{}_{1}-v\,T^{0}{}_{0}-v^{1}J^{0}-u_{1}\widetilde{J}^{0}+u_{1}v^{1}\,,$$
(3.8)
$$\displaystyle{\mathbb{T}}^{1}{}_{1}$$
$$\displaystyle={\cal H}_{A}-{\partial{\cal H}_{A}\over\partial x^{\prime k}}x^{%
\prime k}-p_{k}{\partial{\cal H}_{A}\over\partial p_{k}}-{\partial{\cal H}_{A}%
\over\partial x^{\prime 1}}(x^{\prime 1}-v^{1})-(p_{1}-u_{1}){\partial{\cal H}%
_{A}\over\partial p_{1}}-v\,T^{1}{}_{0}$$
$$\displaystyle=T^{1}{}_{1}-v\,T^{1}{}_{0}-v^{1}J^{1}-u_{1}\widetilde{J}^{1}\,,$$
whose components ${\mathbb{T}}^{\alpha}{}_{1}$ are also shifted by $v\,T^{\alpha}{}_{0}$.
Then, in terms of the improved currents (3.7, 3.8) the flow equations take the form (in the limit $c_{1-}\to 0$)
$$\displaystyle{\partial{\cal H}_{A}\over\partial v}={\mathbb{T}}^{1}{}_{0}=T^{1%
}{}_{0}\,,$$
(3.9)
$$\displaystyle A_{--}{\partial{\cal H}_{A}\over\partial A_{--}}={\mathbb{T}}^{1%
}{}_{1}=T^{1}{}_{1}-vT^{1}{}_{0}-v^{1}J^{1}-u_{1}\widetilde{J}^{1}\,,$$
(3.10)
$$\displaystyle A_{11}{\partial{\cal H}_{A}\over\partial A_{11}}$$
$$\displaystyle=\widetilde{\mathbb{J}}{\mathbb{J}}=\widetilde{J}J-u_{1}J^{1}+v^{%
1}\widetilde{J}^{1}\,,$$
(3.11)
$$\displaystyle A_{--}{\partial{\cal H}_{A}\over\partial\alpha_{\mbox{\tiny${T%
\overline{T}}$}}}={\mathbb{T}}\overline{\mathbb{T}}\,,$$
(3.12)
$$\displaystyle{A_{--}}{\partial{\cal H}_{A}\over\partial\alpha_{\mbox{\tiny${JT%
_{1}}$}}}$$
$$\displaystyle=A_{11}{\mathbb{J}}{\mathbb{T}}_{1}\,,$$
(3.13)
$$\displaystyle{\partial{\cal H}_{A}\over\partial\alpha_{\mbox{\tiny${JT_{0}}$}}}$$
$$\displaystyle=A_{11}{\mathbb{J}}{\mathbb{T}}_{0}+{\alpha_{\mbox{\tiny${%
\widetilde{J}T_{1}}$}}\over A_{--}}{\mathbb{T}}\overline{\mathbb{T}}\,,$$
$$\displaystyle A_{--}{\partial{\cal H}_{A}\over\partial\alpha_{\mbox{\tiny${%
\widetilde{J}T_{1}}$}}}$$
$$\displaystyle={1\over A_{11}}\widetilde{\mathbb{J}}{\mathbb{T}}_{1}\,,$$
$$\displaystyle{\partial{\cal H}_{A}\over\partial\alpha_{\mbox{\tiny${\widetilde%
{J}T_{0}}$}}}$$
$$\displaystyle={1\over A_{11}}\widetilde{\mathbb{J}}{\mathbb{T}}_{0}+{\alpha_{%
\mbox{\tiny${JT_{1}}$}}\over A_{--}}{\mathbb{T}}\overline{\mathbb{T}}\,.$$
A simple analysis of eqs. (3.6), (3.9) and (3.10) shows that the following rescaling of $v^{1},v,u_{1}$
$$\displaystyle v^{1}\to A_{--}v^{1}\,,\quad v\to A_{--}v\,,\quad u_{1}\to A_{--%
}u_{1}\,,$$
brings the equations to the form
$$\displaystyle A_{--}{\partial{\cal H}_{A}\over\partial v^{1}}=J^{1}\,,\quad A_%
{--}{\partial{\cal H}_{A}\over\partial u_{1}}=\widetilde{J}^{1}\,,\quad A_{--}%
{\partial{\cal H}_{A}\over\partial v}=T^{1}{}_{0}\,,\quad A_{--}{\partial{\cal
H%
}_{A}\over\partial A_{--}}=T^{1}{}_{1}\,.$$
(3.14)
This rescaling, however, breaks the dependence of ${\cal H}_{A}$ on $x^{\prime 1}-v^{1}$ and $p_{1}-u_{1}$. To restore the dependence we can rescale the world-sheet space coordinate $\sigma$
$$\displaystyle\sigma\to\sigma/A_{--}\ \Rightarrow\ S_{A}\to S_{A}=\int_{0}^{RA_%
{--}}\,{{\rm d}\sigma\over A_{--}}{\rm d}\tau\,\left(p_{\mu}\dot{x}^{\mu}\,-\,%
{\cal H}_{A}(p_{\mu},x^{\mu},A_{--}x^{\prime\mu})\right)\,.$$
We see that to restore the canonical Poisson structure we also need to rescale $p_{\mu}$
$$\displaystyle p_{\mu}\to A_{--}p_{\mu}\ \Rightarrow\ S_{A}\to S_{A}=\int_{0}^{%
RA_{--}}\,{\rm d}\sigma{\rm d}\tau\,\left(p_{\mu}\dot{x}^{\mu}\,-\,{1\over A_{%
--}}{\cal H}_{A}(A_{--}p_{\mu},x^{\mu},A_{--}x^{\prime\mu})\right)\,.$$
One can then check by using the coefficients ${\cal G}_{i}$ that
$$\displaystyle{1\over A_{--}}{\cal H}_{A}(A_{--}p_{\mu},x^{\mu},A_{--}x^{\prime%
\mu})={\cal H}_{A}(p_{\mu},x^{\mu},x^{\prime\mu})\big{|}_{A_{--}=1}\,.$$
Thus, without loss of generality one can set $A_{--}=1$, and consider $R$ as the tenth deformation parameter. In quantum theory the flow equation (3.10) with respect to $A_{--}$ would imply the following flow equation for the energy with respect to $R$
$$\displaystyle{dE_{n}\over dR}=\langle n|T^{1}{}_{1}|n\rangle\,,$$
where $|n\rangle$ is an eigenstate of $H_{A}$ with the energy $E_{n}$.
The $\widetilde{J}J$ deformation can be also easily understood. It is clear from the way it was generated,
see eq.(2.13), and can be checked explicitly by using the defining equation (3.3) that the deformed Hamiltonian depends on $A_{11}$ only through the combinations
$$\displaystyle A_{11}(p_{1}-u_{1})\,,\quad{1\over A_{11}}(x^{\prime 1}-v^{1})\,.$$
Thus, the energy of a deformed model with parameters $A_{11}$, $u_{1}$, $v^{1}$ is related to the energy of the deformed model with $A_{11}=1$ as
$$\displaystyle E_{n}(A_{11},P_{1}-Ru_{1},\widetilde{P}^{1}-Rv^{1})=E_{n}(1,A_{1%
1}(P_{1}-Ru_{1}),(\widetilde{P}^{1}-Rv^{1})/A_{11})\,,$$
where $\widetilde{P}^{1}=x^{1}(R)-x^{1}(0)$ is basically an effective range of $x^{1}$ which may depend on winding numbers of several scalars and target space-time metric of the model, see the discussion below eq.(4.2). This can be also seen from the flow equations. Indeed, rescaling $u_{1}$ and $v^{1}$ as $u_{1}\to u_{1}/A_{11}$, $v^{1}\to A_{11}v^{1}$, and then rescaling the currents as $J^{\alpha}\to J^{\alpha}/A_{11}$, $\widetilde{J}^{\alpha}\to A_{11}\widetilde{J}^{\alpha}$ removes completely the $A_{11}$ dependence from the flow equations. Thus, we may set $A_{11}=1$.
3.3 Flow equations for the energy
Since the Hamiltonian ${\bf H}_{A}$, the world-sheet momentum ${\bf P}$, the U(1) charge ${\bf P}_{1}$, and the dual charge $\widetilde{\bf P}^{1}$ are mutually commuting,
the energy $E_{n}$ of their common eigenstate $|n\rangle$ is a function of $R$, and the eigenvalues $P$, $P_{1}$, and $\widetilde{P}^{1}$ of ${\bf P}$, ${\bf P}_{1}$, and $\widetilde{\bf P}^{1}$, and the deformation parameters.
Then, the flow equations for the Hamiltonian density, and the Hellmann-Feynman
theorem lead to the following relations
$$\displaystyle\langle n|T^{0}{}_{0}|n\rangle$$
$$\displaystyle={E_{n}\over R}\,,\quad\langle n|T^{0}{}_{1}|n\rangle=-{P\over R}%
\,,\quad\langle n|J^{0}|n\rangle={P_{1}\over R}\,,\quad\langle n|\widetilde{J}%
^{0}|n\rangle={\widetilde{P}^{1}\over R}\,,$$
(3.15)
$$\displaystyle\langle n|T^{1}{}_{1}|n\rangle$$
$$\displaystyle={dE_{n}\over dR}\,,\quad\langle n|T^{1}{}_{0}|n\rangle={1\over R%
}{\partial E_{n}\over\partial v}\,,\quad\langle n|J^{1}|n\rangle={1\over R}{%
\partial E_{n}\over\partial v^{1}}\,,\quad\langle n|\widetilde{J}^{1}|n\rangle%
={1\over R}{\partial E_{n}\over\partial u_{1}}\,.$$
In addition, assuming factorisation, for an operator of the form
${\cal O}_{K_{1}K_{2}}=-\epsilon_{\alpha\beta}K^{\alpha}_{1}K^{\beta}_{2}$, one has
$$\displaystyle\langle n|{\cal O}_{K_{1}K_{2}}|n\rangle=\langle n|K^{0}_{1}|n%
\rangle\langle n|K^{1}_{2}|n\rangle-\langle n|K^{1}_{1}|n\rangle\langle n|K^{0%
}_{2}|n\rangle\,,$$
where $K^{\alpha}_{i}$ are conserved currents, and $|n\rangle$ is an eigenvector of their charges.
By using these relations, it is straightforward to derive flow equations for the energy of the state.
There is, however, a better form of flow equations for the energy.
The Hamiltonian density depends on $p_{1}-u_{1}=J^{0}-u_{1}$, and $x^{\prime 1}-v^{1}=\widetilde{J}^{0}-v^{1}$, and therefore it is more natural to consider the energy $E_{n}$ as a function of the improved charges ${\mathbb{P}}_{1}=P_{1}-Ru_{1}$, and $\widetilde{\mathbb{P}}^{1}=\widetilde{P}^{1}-Rv^{1}$.
To make sure that the energy would depend on $u_{1}$ and $v^{1}$ only through ${\mathbb{P}}_{1}$ and $\widetilde{\mathbb{P}}^{1}$, one also should consider $E_{n}$ as a function of
$$\displaystyle\check{P}\equiv P+{P_{1}\widetilde{P}^{1}\over R}\,.$$
Indeed, the world-sheet momentum can be represented as
$$\displaystyle P=-\int_{0}^{R}d\sigma\,p_{\mu}x^{\prime\mu}=-\int_{0}^{R}d%
\sigma\,\big{(}({P_{1}\over R}+\check{p}_{1})({\widetilde{P}^{1}\over R}+%
\check{x}^{\prime 1})+p_{k}x^{\prime k}\big{)}=-{P_{1}\widetilde{P}^{1}\over R%
}+\check{P}\,,$$
where $\check{p}_{1}$ and $\check{x}^{\prime 1}$ satisfy $\int_{0}^{R}\check{p}_{1}=0$, $\check{x}^{1}(R)=\check{x}^{1}(0)$. Clearly, this formula demonstrates the splitting of the world-sheet momentum into its intrinsic part $\check{P}$, and the part due to the rotation and winding in the $x^{1}$ direction.
Thus, we can consider the energy as a function of $R$, $\check{P}$, ${\mathbb{P}}_{1}$ and $\widetilde{\mathbb{P}}^{1}$, and the deformation parameters $\alpha_{\cal O}$, $v$, $A_{11}$
$$\displaystyle E_{n}=E_{n}(R\,,\check{P}\,,{\mathbb{P}}_{1}\,,\widetilde{%
\mathbb{P}}^{1}\,,\alpha_{\cal O}\,,v\,,A_{11})=E_{n}(R\,,\check{P}\,,P_{1}-Ru%
_{1}\,,\widetilde{P}^{1}-Rv^{1}\,,\alpha_{\cal O}\,,v\,,A_{11})\,.$$
It is now clear that the last two relations in (3.15) take the form
$$\displaystyle\langle n|J^{1}|n\rangle=-{\partial E_{n}\over\partial\widetilde{%
\mathbb{P}}^{1}}\,,\quad\langle n|\widetilde{J}^{1}|n\rangle=-{\partial E_{n}%
\over\partial{\mathbb{P}}_{1}}\,.$$
Then, the flow equations for the Hamiltonian density depend on the improved currents and stress-energy tensor (3.7, 3.8), and their expectation values are
$$\displaystyle\langle n|{\mathbb{T}}^{0}{}_{1}|n\rangle$$
$$\displaystyle=-{1\over R}\big{(}\check{P}-{{\mathbb{P}}_{1}\widetilde{\mathbb{%
P}}^{1}\over R}+vE_{n}\big{)}\,,\quad\langle n|{\mathbb{T}}^{1}{}_{1}|n\rangle%
={\partial E_{n}\over\partial R}-{v\over R}{\partial E_{n}\over\partial v}\,,$$
(3.16)
$$\displaystyle\langle n|{\mathbb{J}}^{0}|n\rangle$$
$$\displaystyle={{\mathbb{P}}_{1}\over R}\,,\quad\langle n|\widetilde{\mathbb{J}%
}^{0}|n\rangle={\widetilde{\mathbb{P}}^{1}\over R}\,.$$
By using these relations, one gets the following flow equations for the energy of the state $|n\rangle$ (with $A_{--}=1$)
$$\displaystyle A_{11}{\partial E_{n}\over\partial A_{11}}$$
$$\displaystyle={\mathbb{P}}_{1}{\partial E_{n}\over\partial{\mathbb{P}}_{1}}-%
\widetilde{\mathbb{P}}^{1}{\partial E_{n}\over\partial\widetilde{\mathbb{P}}^{%
1}}\,,$$
(3.17)
$$\displaystyle{\partial E_{n}\over\partial\alpha_{\mbox{\tiny${T\overline{T}}$}}}$$
$$\displaystyle=-E_{n}{\partial E_{n}\over\partial R}-{\check{P}-{{\mathbb{P}}_{%
1}\widetilde{\mathbb{P}}^{1}\over R}\over R}{\partial E_{n}\over\partial v}\,,$$
(3.18)
$$\displaystyle{1\over A_{11}}{\partial E_{n}\over\partial\alpha_{\mbox{\tiny${%
JT_{1}}$}}}$$
$$\displaystyle={\mathbb{P}}_{1}\big{(}{\partial E_{n}\over\partial R}-{v\over R%
}{\partial E_{n}\over\partial v}\big{)}-\big{(}\check{P}-{{\mathbb{P}}_{1}%
\widetilde{\mathbb{P}}^{1}\over R}+vE_{n}\big{)}{\partial E_{n}\over\partial%
\widetilde{\mathbb{P}}^{1}}\,,$$
$$\displaystyle{1\over A_{11}}{\partial E_{n}\over\partial\alpha_{\mbox{\tiny${%
JT_{0}}$}}}$$
$$\displaystyle={{\mathbb{P}}_{1}\over R}{\partial E_{n}\over\partial v}+E_{n}{%
\partial E_{n}\over\partial\widetilde{\mathbb{P}}^{1}}+{\alpha_{\mbox{\tiny${%
\widetilde{J}T_{1}}$}}\over A_{11}}{\partial E_{n}\over\partial\alpha_{\mbox{%
\tiny${T\overline{T}}$}}}\,,$$
$$\displaystyle A_{11}{\partial E_{n}\over\partial\alpha_{\mbox{\tiny${%
\widetilde{J}T_{1}}$}}}$$
$$\displaystyle=\widetilde{\mathbb{P}}^{1}\big{(}{\partial E_{n}\over\partial R}%
-{v\over R}{\partial E_{n}\over\partial v}\big{)}-\big{(}\check{P}-{{\mathbb{P%
}}_{1}\widetilde{\mathbb{P}}^{1}\over R}+vE_{n}\big{)}{\partial E_{n}\over%
\partial{\mathbb{P}}_{1}}\,,$$
$$\displaystyle A_{11}{\partial E_{n}\over\partial\alpha_{\mbox{\tiny${%
\widetilde{J}T_{0}}$}}}$$
$$\displaystyle={\widetilde{\mathbb{P}}^{1}\over R}{\partial E_{n}\over\partial v%
}+E_{n}{\partial E_{n}\over\partial{\mathbb{P}}_{1}}+A_{11}\alpha_{\mbox{\tiny%
${JT_{1}}$}}{\partial E_{n}\over\partial\alpha_{\mbox{\tiny${T\overline{T}}$}}%
}\,.$$
It is unclear how to solve this system of equations. However, the equation (3.17) with respect to $A_{11}$ for the $\widetilde{J}J$ deformation can be easily solved. Its solution is
$$\displaystyle E_{n}({\mathbb{P}}_{1}\,,\widetilde{\mathbb{P}}^{1}\,,A_{11})=E_%
{n}(A_{11}{\mathbb{P}}_{1}\,,{\widetilde{\mathbb{P}}^{1}\over A_{11}}\,,1)\,,$$
(3.19)
where we kept the energy dependence only on the essential variables. Thus, without loss of generality we can set $A_{11}=1$ in the remaining five equations.
3.4 Homogeneous inviscid Burgers equation
The deformed and undeformed models are obtained from the same extended string sigma model by applying different light-cone type gauge conditions.
The physical quantities, therefore, must be the same in both models. However, as usual in a light-cone gauge, physical states must satisfy additional conditions which follow from the requirement that the strings are closed. We will refer to these conditions as to generalised level-matching conditions. In this section we determine which states of the deformed model are physical, and discuss a relation between the energy of physical states in the deformed and undeformed models. We set $A_{--}=1$ and use $R$ as a parameter.
Let us denote $x^{a}_{0}$, $\tilde{X}_{a}^{0}$, $\Upsilon^{a}_{0}$, $p_{a}^{0}$, $V^{a}_{0}$ and ${\cal U}_{a}^{0}$ the coordinates and momenta of the extended string model (2.4) before one applies the canonical transformation (2.15). They are related to the transformed coordinates and momenta as
in eqs.(2.15). The transformed coordinates and momenta satisfy the light-cone gauge conditions and the constraints. Substituting the solution of the gauge conditions and the constraints into the relations (2.15), one finds for the coordinates
$$\displaystyle x^{\prime-}_{0}$$
$$\displaystyle=-{\mathbb{T}}^{0}{}_{1}\,,$$
(3.20)
$$\displaystyle x^{\prime 1}_{0}$$
$$\displaystyle=\alpha_{\mbox{\tiny${JT_{1}}$}}{\mathbb{T}}^{0}{}_{1}+{%
\widetilde{\mathbb{J}}^{0}\over A_{11}}+\alpha_{\mbox{\tiny${JT_{0}}$}}{\cal H%
}_{A}\,,$$
$$\displaystyle x^{\prime+}_{0}$$
$$\displaystyle=(\alpha_{\mbox{\tiny${T\overline{T}}$}}+\alpha_{\mbox{\tiny${JT_%
{1}}$}}\alpha_{\mbox{\tiny${\widetilde{J}T_{0}}$}}+\alpha_{\mbox{\tiny${%
\widetilde{J}T_{1}}$}}\alpha_{\mbox{\tiny${JT_{0}}$}}){\mathbb{T}}^{0}{}_{1}+%
\alpha_{\mbox{\tiny${\widetilde{J}T_{0}}$}}\big{(}{\widetilde{\mathbb{J}}^{0}%
\over A_{11}}+\alpha_{\mbox{\tiny${JT_{0}}$}}{\cal H}_{A}\big{)}+A_{11}\alpha_%
{\mbox{\tiny${JT_{0}}$}}{\mathbb{J}}^{0}+v\,,$$
$$\displaystyle\tilde{X}^{\prime}_{a}{}^{0}$$
$$\displaystyle=0\,,\quad\Upsilon^{\prime}{}^{a}_{0}=0\,,\quad a=+,-,1\,,$$
and for the momenta
$$\displaystyle p_{+}^{0}$$
$$\displaystyle=-{\cal H}_{A}\,,$$
(3.21)
$$\displaystyle p_{-}^{0}$$
$$\displaystyle=1-\alpha_{\mbox{\tiny${T\overline{T}}$}}{\cal H}_{A}+A_{11}%
\alpha_{\mbox{\tiny${JT_{1}}$}}{\mathbb{J}}^{0}+\alpha_{\mbox{\tiny${%
\widetilde{J}T_{1}}$}}{\widetilde{\mathbb{J}}^{0}\over A_{11}}+\alpha_{\mbox{%
\tiny${JT_{1}}$}}\alpha_{\mbox{\tiny${\widetilde{J}T_{1}}$}}{\mathbb{T}}^{0}{}%
_{1}\,,$$
$$\displaystyle p_{1}^{0}$$
$$\displaystyle=A_{11}{\mathbb{J}}^{0}+\alpha_{\mbox{\tiny${\widetilde{J}T_{0}}$%
}}{\cal H}_{A}+\alpha_{\mbox{\tiny${\widetilde{J}T_{1}}$}}{\mathbb{T}}^{0}{}_{%
1}\,,$$
$$\displaystyle V^{+}_{0}$$
$$\displaystyle=-v\,,$$
$$\displaystyle V^{1}_{0}$$
$$\displaystyle=v^{1}/A_{11}-\alpha_{\mbox{\tiny${JT_{0}}$}}{\cal H}_{A}\,,$$
$$\displaystyle{\cal U}_{1}^{0}$$
$$\displaystyle=v^{1}/A_{11}-\alpha_{\mbox{\tiny${JT_{0}}$}}A_{11}u_{1}{\cal H}_%
{A}\,,$$
where the improved currents and stress-energy tensor are given by (3.7, 3.8).
Eqs.(3.20) show that the generalised level-matching conditions are
$$\displaystyle x^{-}_{0}(R)-x^{-}_{0}(0)$$
$$\displaystyle=-\int_{0}^{R}d\sigma\,{\mathbb{T}}^{0}{}_{1}={\mathbb{P}}=0\,,$$
(3.22)
$$\displaystyle x^{1}_{0}(R)-x^{1}_{0}(0)$$
$$\displaystyle=\widetilde{P}_{0}^{1}=\int_{0}^{R}d\sigma\,({\widetilde{\mathbb{%
J}}^{0}\over A_{11}}+\alpha_{\mbox{\tiny${JT_{0}}$}}{\cal H}_{A})={\widetilde{%
\mathbb{P}}^{1}\over A_{11}}+\alpha_{\mbox{\tiny${JT_{0}}$}}E_{A}=w{\cal R}^{1%
}_{0}\,,$$
$$\displaystyle x^{+}_{0}(R)-x^{+}_{0}(0)$$
$$\displaystyle=\alpha_{\mbox{\tiny${\widetilde{J}T_{0}}$}}w{\cal R}^{1}_{0}+%
\int_{0}^{R}d\sigma\,(A_{11}\alpha_{\mbox{\tiny${JT_{0}}$}}{\mathbb{J}}^{0}+v)%
=\alpha_{\mbox{\tiny${\widetilde{J}T_{0}}$}}w{\cal R}^{1}_{0}+A_{11}\alpha_{%
\mbox{\tiny${JT_{0}}$}}{\mathbb{P}}_{1}+vR=0\,,$$
where $w\in{\mathbb{Z}}$ is the winding number of $x^{1}_{0}$ and $x^{1}$, and the improved world-sheet momentum ${\mathbb{P}}$, and charges ${\mathbb{P}}_{1}$ and $\widetilde{\mathbb{P}}^{1}$ are given by
$$\displaystyle{\mathbb{P}}$$
$$\displaystyle=P+v\,E+v^{1}P_{1}+u_{1}\widetilde{P}^{1}-u_{1}v^{1}R\,,$$
(3.23)
$$\displaystyle{\mathbb{P}}_{1}$$
$$\displaystyle=P_{1}-u_{1}R\,,$$
$$\displaystyle\widetilde{\mathbb{P}}^{1}$$
$$\displaystyle=\widetilde{P}^{1}-v^{1}R=w{\cal R}^{1}-v^{1}R\,.$$
Solving the generalised level-matching conditions for $P$, $v$ and $v^{1}$, one gets
$$\displaystyle P$$
$$\displaystyle=w\big{(}{{\cal R}_{0}^{1}\over R}\alpha_{\mbox{\tiny${\widetilde%
{J}T_{0}}$}}\,E_{A}+{A_{11}{\cal R}^{1}_{0}\over R}{\mathbb{P}}_{1}-{{\cal R}^%
{1}\over R}P_{1}\big{)}\,,$$
(3.24)
$$\displaystyle vR$$
$$\displaystyle=-\alpha_{\mbox{\tiny${\widetilde{J}T_{0}}$}}w{\cal R}^{1}_{0}-A_%
{11}\alpha_{\mbox{\tiny${JT_{0}}$}}{\mathbb{P}}_{1}\,,$$
$$\displaystyle v^{1}R$$
$$\displaystyle=A_{11}\alpha_{\mbox{\tiny${JT_{0}}$}}E_{A}+w({\cal R}^{1}-A_{11}%
{\cal R}^{1}_{0})\,.$$
Having found the generalised level-matching conditions, we can now relate the energies of physical states of the deformed and undeformed models.
Integrating the relations (3.21) over $\sigma$, one gets
$$\displaystyle E_{0}$$
$$\displaystyle=E_{A}\,,$$
(3.25)
$$\displaystyle R_{0}$$
$$\displaystyle=R-\alpha_{\mbox{\tiny${T\overline{T}}$}}E_{A}+\alpha_{\mbox{%
\tiny${\widetilde{J}T_{1}}$}}{\widetilde{\mathbb{P}}^{1}\over A_{11}}+\alpha_{%
\mbox{\tiny${JT_{1}}$}}(A_{11}{\mathbb{P}}_{1}-\alpha_{\mbox{\tiny${\widetilde%
{J}T_{1}}$}}{\mathbb{P}})\,,$$
$$\displaystyle P_{1}^{0}$$
$$\displaystyle=A_{11}{\mathbb{P}}_{1}-\alpha_{\mbox{\tiny${\widetilde{J}T_{1}}$%
}}{\mathbb{P}}+\alpha_{\mbox{\tiny${\widetilde{J}T_{0}}$}}E_{A}\,,$$
where $R_{0}\equiv P_{-}^{0}=\int_{0}^{R}p_{-}^{0}$ is the circumference of the cylinder, $P_{1}^{0}=\int_{0}^{R}p_{1}^{0}$ is the $U(1)$ charge, and $E_{0}$ is the energy
of the undeformed model which is equal to the energy of the deformed model for physical states.
These equations generalise the integrated form of the homogeneous inviscid Burgers equation which describes the spectrum of a ${T\overline{T}}$ deformed model for states with vanishing world-sheet momentum. For physical states one can set ${\mathbb{P}}=0$ and ${\widetilde{\mathbb{P}}^{1}}=w{\cal R}^{1}-A_{11}\alpha_{\mbox{\tiny${JT_{0}}$%
}}E_{A}$ but we prefer to keep them in (3.25) because they might be important for states which do not satisfy the generalised level-matching conditions. Due to the momentum $P_{1}$ quantisation
$$\displaystyle P_{1}^{0}={2\pi m\over{\cal R}^{1}_{0}}\,,\quad P_{1}={2\pi m%
\over{\cal R}^{1}}\,,\quad m\in{\mathbb{Z}}\,,$$
and the last equation in (3.25) can be considered as a relation between the ranges ${\cal R}_{0}^{1}$ and ${\cal R}^{1}$ of $x^{1}_{0}$ and $x^{1}$.
4 Flow equations in CFT conventions
To analyse the deformations of a CFT it is convenient to use the light-cone world-sheet coordinates
$$\displaystyle\sigma^{\pm}=\sigma\pm\tau\,,\quad\partial_{\pm}={1\over 2}(%
\partial_{\sigma}\pm\partial_{\tau})\,,\quad\partial_{\alpha}J^{\alpha}=%
\partial_{+}J^{+}+\partial_{-}J^{-}\,.$$
The light-cone components of a vector $V^{\alpha}$, a co-vector $U_{\alpha}$, and the stress-energy tensor $T^{\alpha}{}_{\beta}$ are related to the $0,1$ components as
$$\displaystyle V^{\pm}$$
$$\displaystyle=V^{1}\pm V^{0}\,,\quad V^{0}={1\over 2}(V^{+}-V^{-})\,,\quad V^{%
1}={1\over 2}(V^{+}+V^{-})\,,$$
$$\displaystyle U_{\pm}$$
$$\displaystyle={1\over 2}(U_{1}\pm U_{0})\,,\quad U_{0}=U_{+}-U_{-}\,,\quad U_{%
1}=U_{+}+U_{-}\,,$$
$$\displaystyle T^{+}{}_{+}$$
$$\displaystyle={1\over 2}(T^{0}{}_{0}+T^{1}{}_{1}+T^{0}{}_{1}+T^{1}{}_{0})\,,%
\quad T^{-}{}_{-}={1\over 2}(T^{0}{}_{0}+T^{1}{}_{1}-T^{0}{}_{1}-T^{1}{}_{0})\,,$$
$$\displaystyle T^{-}{}_{+}$$
$$\displaystyle={1\over 2}(T^{1}{}_{1}-T^{0}{}_{0}-T^{0}{}_{1}+T^{1}{}_{0})\,,%
\quad T^{+}{}_{-}={1\over 2}(T^{1}{}_{1}-T^{0}{}_{0}+T^{0}{}_{1}-T^{1}{}_{0})\,.$$
In a CFT one has $T^{0}{}_{0}+T^{1}{}_{1}=0$, $T^{0}{}_{1}+T^{1}{}_{0}=0$, and therefore $T^{+}{}_{+}=T^{-}{}_{-}=0$.
In terms of the light-cone components the deformation operators become
$$\displaystyle T\overline{T}$$
$$\displaystyle=-\epsilon_{\alpha\beta}T^{\alpha}{}_{1}T^{\beta}{}_{0}=-2%
\epsilon_{\alpha\beta}T^{\alpha}{}_{-}T^{\beta}{}_{+}=T^{+}{}_{-}T^{-}{}_{+}-T%
^{-}{}_{-}T^{+}{}_{+}\,,$$
$$\displaystyle VT_{1}$$
$$\displaystyle=-\epsilon_{\alpha\beta}V^{\alpha}T^{\beta}{}_{1}={1\over 2}\big{%
(}V^{+}(T^{-}{}_{+}+T^{-}{}_{-})-V^{-}(T^{+}{}_{+}+T^{+}{}_{-})\big{)}\,,$$
$$\displaystyle VT_{0}$$
$$\displaystyle=-\epsilon_{\alpha\beta}V^{\alpha}T^{\beta}{}_{0}={1\over 2}\big{%
(}V^{+}(T^{-}{}_{+}-T^{-}{}_{-})-V^{-}(T^{+}{}_{+}-T^{+}{}_{-})\big{)}\,,$$
$$\displaystyle VT_{\pm}$$
$$\displaystyle=-\epsilon_{\alpha\beta}V^{\alpha}T^{\beta}{}_{\pm}={1\over 2}%
\big{(}V^{+}T^{-}{}_{\pm}-V^{-}T^{+}{}_{\pm}\big{)}\,,$$
where $V^{\alpha}$ could be either $J^{\alpha}$ or $\widetilde{J}^{\alpha}$.
4.1 Flow equations for the Hamiltonian density
Introducing new deformation parameters (and setting $A_{--}=1$)
$$\displaystyle\alpha_{\pm}=\alpha_{\mbox{\tiny${JT_{1}}$}}\pm\alpha_{\mbox{%
\tiny${JT_{0}}$}}\,,\quad\tilde{\alpha}_{\pm}=\alpha_{\mbox{\tiny${\widetilde{%
J}T_{1}}$}}\pm\alpha_{\mbox{\tiny${\widetilde{J}T_{0}}$}}\,,$$
we rewrite the flow equations (3.13) in the form
$$\displaystyle{\partial{\cal H}_{A}\over\partial\alpha_{\pm}}$$
$$\displaystyle=A_{11}{\mathbb{J}}{\mathbb{T}}_{\pm}\pm\,{\tilde{\alpha}_{+}+%
\tilde{\alpha}_{-}\over 4}{\mathbb{T}}\overline{\mathbb{T}}\,,$$
(4.1)
$$\displaystyle{\partial{\cal H}_{A}\over\partial\tilde{\alpha}_{\pm}}$$
$$\displaystyle={1\over A_{11}}\widetilde{\mathbb{J}}{\mathbb{T}}_{\pm}\pm{%
\alpha_{+}+\alpha_{-}\over 4}{\mathbb{T}}\overline{\mathbb{T}}\,.$$
For a generic model (even a CFT) there is no reason to mix the currents $J$ and $\widetilde{J}$.
However, the situation is different if we want to deform a Lorentz-invariant model with the Lagrangian
$$\displaystyle{\cal L}_{0}=-{1\over 2}\partial_{\alpha}x^{\mu}\partial^{\alpha}%
x^{\nu}\,G_{\mu\nu}-V(x)\,,$$
(4.2)
which is invariant under shifts of $x^{1}$, and whose target-space metric components $G_{1\nu}$ are independent of $x$. Then, shifting and rescaling $x^{1}$
$$\displaystyle x^{1}\to{1\over\sqrt{G_{11}}}(x^{1}-G_{1k}x^{k})\,,$$
we (almost) decouple $x^{1}$ from the other fields
$$\displaystyle{\cal L}_{0}\to{\cal L}_{0}=-{1\over 2}\partial_{\alpha}x^{1}%
\partial^{\alpha}x^{1}-{1\over 2}\partial_{\alpha}x^{k}\partial^{\alpha}x^{l}%
\,\tilde{G}_{kl}-V(x)\,.$$
(4.3)
We see that $x^{1}$ is a free massless scalar whose only knowledge of the other fields is in its winding numbers. For such a scalar the topological current $\widetilde{J}$ is dual to $J^{\alpha}=-\partial^{\alpha}x^{1}$
$$\displaystyle\widetilde{J}^{0}=x^{\prime 1}=-J^{1}\,,\quad\widetilde{J}^{1}=-%
\dot{x}^{1}=p_{1}=-J^{0}\,,$$
and introducing
$$\displaystyle{\rm J}_{\pm}^{\alpha}=-{1\over 2}(J^{\alpha}\pm\widetilde{J}^{%
\alpha})\,,\quad{\rm J}_{\pm}^{0}=\mp\partial_{\pm}x^{1}\,,\quad{\rm J}_{\pm}^%
{1}=\partial_{\pm}x^{1}\,,$$
$$\displaystyle{\rm J}_{+}^{-}=2\partial_{+}x^{1}\,,\quad{\rm J}_{+}^{+}=0\,,%
\quad{\rm J}_{-}^{+}=2\partial_{-}x^{1}\,,\quad{\rm J}_{-}^{-}=0\,,$$
we see that the current ${\rm J}^{\alpha}\equiv{\rm J}_{+}^{\alpha}$ is purely left-moving (holomorphic) while $\bar{\rm J}^{\alpha}\equiv{\rm J}_{-}^{\alpha}$ is purely right-moving (anti-holomorphic). Clearly, a generic deformation would ruin these nice properties but one can still hope that these currents ${\rm J}^{\alpha}$ and $\bar{\rm J}^{\alpha}$ would suit better to describe the deformation. In what follows discussing the deformation of a model with Lagrangian (4.2), we first decouple the scalar $x^{1}$, and then deform the model.
Coming back to the flow equations (4.1), we see that in the presence of the $\tilde{J}J$-deformation the currents ${\rm J}^{\alpha}$ and
$\bar{\rm J}^{\alpha}$ should be defined as
$$\displaystyle{\rm J}_{\pm}^{\alpha}=-{1\over 2}({A_{11}}J^{\alpha}\pm{1\over A%
_{11}}\widetilde{J}^{\alpha})\,,\quad\tilde{J}J=-2\epsilon_{\alpha\beta}{\rm J%
}_{+}^{\alpha}{\rm J}_{-}^{\beta}\,.$$
(4.4)
The appearance of $A_{11}$ in this formula is expected because the $\widetilde{J}J$-deformation (2.13) rescales the coordinate $x^{1}$ and momentum $p_{1}$, and the factors of $A_{11}$ just reverse this rescaling.
Now, introducing the deformation parameters
$$\displaystyle\beta_{\pm}=-{1\over 2}(\alpha_{\pm}+\tilde{\alpha}_{\pm})\,,%
\quad\bar{\beta}_{\pm}=-{1\over 2}(\alpha_{\pm}-\tilde{\alpha}_{\pm})\,,$$
we bring the flow equations (4.1) to the form
$$\displaystyle{\partial{\cal H}_{A}\over\partial\beta_{\pm}}=2{\cal J}{\mathbb{%
T}}_{\pm}\pm\,{\beta_{+}+\beta_{-}\over 2}{\mathbb{T}}\overline{\mathbb{T}}\,,$$
(4.5)
$$\displaystyle{\partial{\cal H}_{A}\over\partial\bar{\beta}_{\pm}}=2\bar{\cal J%
}{\mathbb{T}}_{\pm}\mp{\bar{\beta}_{+}+\bar{\beta}_{-}\over 2}{\mathbb{T}}%
\overline{\mathbb{T}}\,.$$
Here ${\cal J}^{\alpha}\equiv{\cal J}^{\alpha}_{+}$ and $\bar{\cal J}\equiv{\cal J}^{\alpha}_{-}$ are the improved currents
$$\displaystyle{\cal J}_{\pm}^{\alpha}=-{1\over 2}({A_{11}}{\mathbb{J}}^{\alpha}%
\pm{1\over A_{11}}\tilde{\mathbb{J}}^{\alpha})\,,$$
(4.6)
and up to a normalisation in the Euclidean CFT terminology $JT_{+}\leftrightarrow J\Theta$, $JT_{-}\leftrightarrow J\overline{T}$,
$\bar{J}T_{+}\leftrightarrow\bar{J}T$, $\bar{J}T_{-}\leftrightarrow\bar{J}\overline{\Theta}$.
A nice feature of these equations is that in the absence of linear deformations ($v^{1}=v=u_{1}=0$) the flow equations for the left- and right-moving currents are coupled to each other only through the ${T\overline{T}}$ operator. Nevertheless, it still may not be the best form of the equations because they all involve the ${T\overline{T}}$ operator. Indeed, we have the flow equation for the ${T\overline{T}}$ deformation, and we can make nonlinear changes of the parameter $a_{+-}$. Performing
the following transformation
$$\displaystyle a_{+-}=\alpha+{\varkappa_{+-}\over 2}\beta_{+}\beta_{-}+{%
\varkappa_{+}\over 4}\beta_{+}^{2}+{\varkappa_{-}\over 4}\beta_{-}^{2}+{\bar{%
\varkappa}_{+-}\over 2}\bar{\beta}_{+}\bar{\beta}_{-}+{\bar{\varkappa}_{+}%
\over 4}\bar{\beta}_{+}^{2}+{\bar{\varkappa}_{-}\over 4}\bar{\beta}_{-}^{2}\,,$$
one finds that the equations (3.12, 4.5) take the form
$$\displaystyle{\partial{\cal H}_{A}\over\partial\alpha}={\mathbb{T}}\overline{%
\mathbb{T}}\,,$$
(4.7)
$$\displaystyle{\partial{\cal H}_{A}\over\partial\beta_{+}}=2{\cal J}{\mathbb{T}%
}_{+}\,+\,{\beta_{+}(1+\varkappa_{+})+\beta_{-}(1+\varkappa_{+-})\over 2}{%
\mathbb{T}}\overline{\mathbb{T}}\,,$$
(4.8)
$$\displaystyle{\partial{\cal H}_{A}\over\partial\beta_{-}}=2{\cal J}{\mathbb{T}%
}_{-}\,-\,{\beta_{+}(1-\varkappa_{+-})+\beta_{-}(1-\varkappa_{-})\over 2}{%
\mathbb{T}}\overline{\mathbb{T}}\,,$$
$$\displaystyle{\partial{\cal H}_{A}\over\partial\bar{\beta}_{+}}=2\bar{\cal J}{%
\mathbb{T}}_{+}-{\bar{\beta}_{+}(1-\bar{\varkappa}_{+})+\bar{\beta}_{-}(1-\bar%
{\varkappa}_{+-})\over 2}{\mathbb{T}}\overline{\mathbb{T}}\,,$$
$$\displaystyle{\partial{\cal H}_{A}\over\partial\bar{\beta}_{-}}=2\bar{\cal J}{%
\mathbb{T}}_{-}+{\bar{\beta}_{+}(1+\bar{\varkappa}_{+-})+\bar{\beta}_{-}(1+%
\bar{\varkappa}_{-})\over 2}{\mathbb{T}}\overline{\mathbb{T}}\,.$$
It is clear that natural values of the parameters $\varkappa$’s are $0,\pm 1$. A particular choice of $\varkappa$’s depends on the model under study. If the model does not have left- and right-moving conserved currents then the original choice with all $\varkappa$’s equal to 0 seems to be the best one.
However, if a CFT has left- and right-moving currents then there are better choices.
Since in such a CFT $JT_{+}=0$ and $\bar{J}T_{-}=0$,
it seems reasonable to choose $\varkappa$’s so that the flow equations for $JT_{-}$ and $\bar{J}T_{+}$ would take the simplest form.
Thus, one would choose $\varkappa_{-}=1$, $\varkappa_{+-}=1$, $\bar{\varkappa}_{+}=1$, $\bar{\varkappa}_{+-}=1$. Then, the remaining two flow equations take the simplest form if
one chooses $\varkappa_{+}=-1$ and $\bar{\varkappa}_{-}=-1$. With this choice
the flow equations with respect to $\beta_{\pm}$ and $\bar{\beta}_{\pm}$ are
$$\displaystyle\varkappa_{+}=-1\,,\quad\varkappa_{-}=1\,,\quad\bar{\varkappa}_{+%
}=1\,,\quad\bar{\varkappa}_{-}=-1\,,\quad\varkappa_{+-}=1\,,\quad\bar{%
\varkappa}_{+-}=1\,,$$
(4.9)
$$\displaystyle{\partial{\cal H}_{A}\over\partial\beta_{-}}$$
$$\displaystyle=2{\cal J}{\mathbb{T}}_{-}\,,\qquad{\partial{\cal H}_{A}\over%
\partial\beta_{+}}=2{\cal J}{\mathbb{T}}_{+}\,+\,\beta_{-}{\mathbb{T}}%
\overline{\mathbb{T}}\,,$$
(4.10)
$$\displaystyle{\partial{\cal H}_{A}\over\partial\bar{\beta}_{+}}$$
$$\displaystyle=2\bar{\cal J}{\mathbb{T}}_{+}\,,\qquad{\partial{\cal H}_{A}\over%
\partial\bar{\beta}_{-}}=2\bar{\cal J}{\mathbb{T}}_{-}\,+\,\bar{\beta}_{+}{%
\mathbb{T}}\overline{\mathbb{T}}\,.$$
Calculating the coefficients ${\cal G}_{i}$ with this choice of $\varkappa$’s for a deformed CFT which had left- and right-moving conserved currents before the deformation, see appendix C, one finds that they are independent of $\beta_{+}$ and $\bar{\beta}_{-}$ (in fact with all the ten parameters switched on), and, therefore, the Hamiltonian density ${\cal H}_{A}$ is independent of these two parameters too. Thus, the flow equations with respect to $\beta_{+}$ and $\bar{\beta}_{-}$ lead to the existence of the following two relations
$$\displaystyle 2{\cal J}{\mathbb{T}}_{+}\,+\,\beta_{-}{\mathbb{T}}\overline{%
\mathbb{T}}$$
$$\displaystyle=0\,,$$
(4.11)
$$\displaystyle 2\bar{\cal J}{\mathbb{T}}_{-}+\bar{\beta}_{+}{\mathbb{T}}%
\overline{\mathbb{T}}$$
$$\displaystyle=0\,,$$
in a deformed CFT. One might think that these relations imply the existence of improved left- and right-moving currents but it is so only in the case of a single $J\overline{T}$ or $\bar{J}T$ deformation.
4.2 Flow equations for the energy in CFT conventions
To write the flow equations for the energy in CFT conventions we introduce the charges ${\rm Q}_{\pm}$ and ${\cal Q}_{\pm}$ of the left- and right-moving currents (4.4) and (4.6)
$$\displaystyle{{\rm Q}_{\pm}\over R}=\langle n|{\rm J}_{\pm}^{0}|n\rangle=-{1%
\over 2R}(A_{11}P_{1}\pm{1\over A_{11}}\widetilde{P}^{1})\,,$$
and similar expressions for the improved charges ${\cal Q}_{\pm}$ of the currents (4.6).
Then, we get
$$\displaystyle{\partial E_{n}\over\partial{\rm Q}_{\pm}}={\partial E_{n}\over%
\partial{\cal Q}_{\pm}}=-\big{(}{1\over A_{11}}{\partial E_{n}\over\partial{%
\mathbb{P}}_{1}}\pm A_{11}{\partial E_{n}\over\partial\widetilde{\mathbb{P}}^{%
1}}\big{)}\,,$$
and therefore
$$\displaystyle\langle n|{\rm J}_{\pm}^{1}|n\rangle=\langle n|{\cal J}_{\pm}^{1}%
|n\rangle={1\over 2}\big{(}A_{11}{\partial E_{n}\over\partial\widetilde{P}^{1}%
}\pm{1\over A_{11}}{\partial E_{n}\over\partial P_{1}}\big{)}=\mp{1\over 2}{%
\partial E_{n}\over\partial{\cal Q}_{\pm}}\,.$$
We also need
$$\displaystyle\langle n|{\mathbb{T}}^{1}{}_{\pm}|n\rangle={1\over 2}\big{(}%
\langle n|{\mathbb{T}}^{1}{}_{1}|n\rangle\pm\langle n|{\mathbb{T}}^{1}{}_{0}|n%
\rangle\big{)}={1\over 2}\big{(}{\partial E_{n}\over\partial R}\pm{1\over R}(1%
\mp v){\partial E_{n}\over\partial v}\big{)}\,,$$
$$\displaystyle\langle n|{\mathbb{T}}^{0}{}_{\pm}|n\rangle={1\over 2}\big{(}%
\langle n|{\mathbb{T}}^{0}{}_{1}|n\rangle\pm\langle n|{\mathbb{T}}^{0}{}_{0}|n%
\rangle\big{)}={1\over 2R}\big{(}-\check{P}+{{\cal Q}_{+}^{2}-{\cal Q}_{-}^{2}%
\over R}\pm(1\mp v)E_{n}\big{)}\,.$$
By using these relations and eqs.(4.10), one gets the following flow equations for the energy of the state $|n\rangle$ (with $A_{--}=1$)
$$\displaystyle A_{11}{\partial E_{n}\over\partial A_{11}}$$
$$\displaystyle={\cal Q}_{+}{\partial E_{n}\over\partial{\cal Q}_{-}}+{\cal Q}_{%
-}{\partial E_{n}\over\partial{\cal Q}_{+}}\,,$$
(4.12)
$$\displaystyle{\partial E_{n}\over\partial\alpha}$$
$$\displaystyle=-E_{n}{\partial E_{n}\over\partial R}-{\check{P}-{{\cal Q}_{+}^{%
2}-{\cal Q}_{-}^{2}\over R}\over R}{\partial E_{n}\over\partial v}\,,$$
(4.13)
$$\displaystyle{\partial E_{n}\over\partial\beta_{-}}$$
$$\displaystyle={{\cal Q}_{+}}\big{(}{\partial E_{n}\over\partial R}-{1\over R}(%
1+v){\partial E_{n}\over\partial v}\big{)}-{1\over 2}{\partial E_{n}\over%
\partial{\cal Q}_{+}}\big{(}\check{P}-{{\cal Q}_{+}^{2}-{\cal Q}_{-}^{2}\over R%
}+(1+v)E_{n}\big{)}\,,$$
$$\displaystyle{\partial E_{n}\over\partial\bar{\beta}_{+}}$$
$$\displaystyle={{\cal Q}_{-}}\big{(}{\partial E_{n}\over\partial R}+{1\over R}(%
1-v){\partial E_{n}\over\partial v}\big{)}+{1\over 2}{\partial E_{n}\over%
\partial{\cal Q}_{-}}\big{(}\check{P}-{{\cal Q}_{+}^{2}-{\cal Q}_{-}^{2}\over R%
}-(1-v)E_{n}\big{)}\,,$$
$$\displaystyle{\partial E_{n}\over\partial\beta_{+}}$$
$$\displaystyle={{\cal Q}_{+}}\big{(}{\partial E_{n}\over\partial R}+{1\over R}(%
1-v){\partial E_{n}\over\partial v}\big{)}-{1\over 2}{\partial E_{n}\over%
\partial{\cal Q}_{+}}\big{(}\check{P}-{{\cal Q}_{+}^{2}-{\cal Q}_{-}^{2}\over R%
}-(1-v)E_{n}\big{)}+{\beta_{-}}{\partial E_{n}\over\partial\alpha}\,,$$
$$\displaystyle{\partial E_{n}\over\partial\bar{\beta}_{-}}$$
$$\displaystyle={{\cal Q}_{-}}\big{(}{\partial E_{n}\over\partial R}-{1\over R}(%
1+v){\partial E_{n}\over\partial v}\big{)}+{1\over 2}{\partial E_{n}\over%
\partial{\cal Q}_{-}}\big{(}\check{P}-{{\cal Q}_{+}^{2}-{\cal Q}_{-}^{2}\over R%
}+(1+v)E_{n}\big{)}+{\bar{\beta}_{+}}{\partial E_{n}\over\partial\alpha}\,.$$
As was discussed in the previous subsection, in the case of a CFT with left- and right-moving conserved currents ${\partial E_{n}\over\partial\beta_{+}}={\partial E_{n}\over\partial\bar{\beta}%
_{-}}=0$.
4.3 Deformed CFT with left- and right-moving currents
In this subsection we use the defining equation (3.3) for the Hamiltonian density to find a solution of the flow equations (4.13) for a deformed CFT with left- and right-moving conserved currents. The undeformed Lagrangian is given by (4.2), and before deforming the model we shift and rescale $x^{1}$ as discussed in section 4.1 to reduce (4.2) to the canonical form (4.3) with the undeformed Hamiltonian density
$$\displaystyle{\cal H}_{0}={1\over 2}p_{1}^{2}+{1\over 2}(x^{\prime 1})^{2}+{1%
\over 2}G^{kl}(x)p_{k}p_{l}+{1\over 2}x^{\prime k}x^{\prime l}\,G_{kl}(x)%
\equiv{\cal K}\,.$$
(4.14)
We set $A_{11}=1$ because the effect of the $\tilde{J}J$ deformation was discussed in section
3.3, and the energy of a $\tilde{J}J$ deformed model is related to the energy of the model with $A_{11}=1$ by eq.(3.19). To simplify the notations we also set the parameters $v^{1}$, $u_{1}$ to zero. The dependence of these parameters can be easily restored by shifting $p_{1}$ and $x^{\prime 1}$ in the Hamiltonian density, or by replacing the charges ${\rm Q}_{\pm}$ with the improved charges ${\cal Q}_{\pm}$ in the energy eigenvalues.
Let us assume that ${\cal H}_{A}$ and the coefficients ${\cal G}_{i}$ of the defining equation (3.3) are operators, and let us rewrite (3.3) in the following form
$$\displaystyle\langle n|\,{\cal G}_{2}\,{\cal H}_{A}^{2}-{\cal G}_{1}\,{\cal H}%
_{A}+{\cal G}_{0}\,|n\rangle=0\,\,.$$
(4.15)
Clearly, this equation does not really make sense, because the products of operators require careful definitions. In particular, ${\cal G}_{1}$ given by (C.9) depends on $px^{\prime}$, $p_{1}$ and $x^{\prime 1}$ which do not commute with ${\cal H}_{A}$. This can be cured because ${\cal G}_{2}$ does not depend on any fields, and therefore, shifting ${\cal H}_{A}$ by a linear combination of $px^{\prime}$, $p_{1}$ and $x^{\prime 1}$, one can remove all field-dependent terms from ${\cal G}_{1}$.
Performing this shift, we bring eq.(4.15) to the form111This is also the simplest form to check (C.12).
$$\displaystyle{\cal G}_{2}\,\langle n|({\cal H}_{A}-{\mathscr{S}}_{A})^{2}|n%
\rangle-\widetilde{\cal G}_{1}\,\langle n|({\cal H}_{A}-{\mathscr{S}}_{A})|n%
\rangle+\langle n|\widetilde{\cal G}_{0}\,|n\rangle=0\,\,,$$
where
$$\displaystyle{\cal G}_{2}$$
$$\displaystyle=\alpha(1-v^{2})+{1\over 2}\beta_{-}^{2}(1+v)^{2}+{1\over 2}\bar{%
\beta}_{+}^{2}(1-v)^{2}\,,$$
$$\displaystyle\widetilde{\cal G}_{1}$$
$$\displaystyle=1-v^{2}\,,$$
$$\displaystyle\widetilde{\cal G}_{0}={\cal K}+$$
$$\displaystyle{1\over{\cal G}_{2}}\Big{[}\frac{1}{2}px^{\prime}\left(\bar{\beta%
}_{+}^{2}(1-v)^{2}-\beta_{-}^{2}(1+v)^{2}\right)-(1-v^{2})\big{(}\beta_{-}{\rm
J%
}_{+}^{0}(1+v)+\bar{\beta}_{+}{\rm J}_{-}^{0}(1-v)\big{)}$$
$$\displaystyle+(px^{\prime})^{2}\big{(}\bar{\beta}_{+}^{2}\beta_{-}^{2}-\alpha^%
{2}\big{)}+2{\rm J}_{+}^{0}px^{\prime}\big{(}\alpha\beta_{-}(1+v)+\bar{\beta}_%
{+}^{2}\beta_{-}(1-v)\big{)}$$
$$\displaystyle-2{\rm J}_{-}^{0}px^{\prime}\big{(}\alpha\bar{\beta}_{+}(1-v)+%
\bar{\beta}_{+}\beta_{-}^{2}(1+v)\big{)}-\big{(}\beta_{-}{\rm J}_{+}^{0}(1+v)+%
\bar{\beta}_{+}{\rm J}_{-}^{0}(1-v)\big{)}^{2}\Big{]}\,,$$
$$\displaystyle{\mathscr{S}}_{A}$$
$$\displaystyle={1\over{\cal G}_{2}}\Big{(}\frac{1}{2}px^{\prime}\left(-2\alpha v%
-\bar{\beta}_{+}^{2}(1-v)+\beta_{-}^{2}(1+v)\right)+\beta_{-}{\rm J}_{+}^{0}(1%
+v)+\bar{\beta}_{+}{\rm J}_{-}^{0}(1-v)\Big{)}\,.$$
Now, to get an equation for the spectrum of the deformed CFT with left- and right-moving conserved currents we use the following relations
$$\displaystyle\langle n|{\cal H}_{A}|n\rangle={E_{n}\over R}\,,\quad\langle n|{%
\rm J}_{\pm}^{0}|n\rangle={{\rm Q}_{\pm}\over R}\,,$$
(4.16)
$$\displaystyle\langle n|px^{\prime}|n\rangle=-{P\over R}=-{\check{k}-{\rm Q}_{+%
}^{2}+{\rm Q}_{-}^{2}\over R^{2}}\,,\quad\langle n|{\cal K}|n\rangle={E_{n}^{(%
0)}\over R}={\check{{\cal E}}_{n}^{(0)}+{\rm Q}_{+}^{2}+{\rm Q}_{-}^{2}\over R%
^{2}}\,.$$
(4.17)
Here, as was discussed in section 3.3, we
split of the world-sheet momentum $P=k/R$ into its intrinsic part $\check{P}=\check{k}/R$, and the part due to the rotation and winding in the $x^{1}$ direction, and do the same with the eigenvalues $E_{n}^{(0)}={\cal E}_{n}^{(0)}/R$ of the Hamiltonian (4.14). It is worthwhile to mention that only $\check{{\cal E}}_{n}^{(0)}$ always coincides with the corresponding part of the spectrum of the undeformed model. As to ${\rm Q}_{\pm}$ dependent part, it depends on the point of view on the deformation. One may say that the range of $x^{1}$ does not change with the deformation and then ${\rm Q}_{\pm}$ do not change either, or one may use that the deformation was generated by the chain of transformations (2.14) which explicitly changed the range of $x^{1}$ and, therefore, ${\rm Q}_{\pm}$. In any case ${\rm Q}_{\pm}$ are the charges measured in the deformed theory.
Finally, we use the following replacement rule
$$\displaystyle\langle n|K_{1}K_{2}|n\rangle\to\langle n|K_{1}|n\rangle\langle n%
|K_{2}|n\rangle\,,$$
for products of any two operators in (4.16, 4.17). The only justification for this rule we have is that it leads to the deformed eigenvalues $E_{n}$ which satisfy the flow equations (4.13).
By using these rules we find that the spectrum of the deformed CFT with left- and right-moving conserved currents is given by
$$\displaystyle{\cal A}\big{(}{E_{n}\over R}-{\cal S}_{E}\big{)}^{2}-\big{(}{E_{%
n}\over R}-{\cal S}_{E}\big{)}+{\cal C}=0\,,$$
(4.18)
$$\displaystyle{E_{n}\over R}={\cal S}_{E}+{1-\sqrt{1-4{\cal A}{\cal C}}\over 2{%
\cal A}}\,.$$
Here
$$\displaystyle{\cal A}$$
$$\displaystyle=\alpha+{b_{-}^{2}\over 2}+{\bar{b}_{+}^{2}\over 2}\,,$$
(4.19)
$$\displaystyle{\cal A}{\cal C}$$
$$\displaystyle={\cal A}\,{{\cal E}_{n}^{(0)}\over{\cal R}^{2}}+\frac{k}{2{\cal R%
}^{2}}\left(b_{-}^{2}-\bar{b}_{+}^{2}\right)-\frac{1}{{\cal R}}\big{(}b_{-}{%
\rm Q}_{+}+\bar{b}_{+}{\rm Q}_{-}\big{)}+\frac{k^{2}}{{\cal R}^{4}}\big{(}\,{%
\bar{b}_{+}^{2}b_{-}^{2}}-\alpha^{2}\big{)}$$
$$\displaystyle-2{k\,{\rm Q}_{+}\over{\cal R}^{3}}b_{-}\big{(}\alpha+{\bar{b}_{+%
}^{2}}\big{)}+2{k\,{\rm Q}_{-}\over{\cal R}^{3}}\bar{b}_{+}\big{(}\alpha+{b_{-%
}^{2}}\big{)}-\frac{1}{{\cal R}^{2}}\big{(}b_{-}{\rm Q}_{+}+\bar{b}_{+}{\rm Q}%
_{-}\big{)}^{2}\,,$$
$$\displaystyle{\cal S}_{E}$$
$$\displaystyle={k\,v\over{\cal R}^{2}}+{1\over{\cal A}}\big{(}\frac{k}{2{\cal R%
}^{2}}\big{(}\bar{b}_{+}^{2}-b_{-}^{2}\big{)}+\frac{1}{{\cal R}}b_{-}{\rm Q}_{%
+}+\frac{1}{{\cal R}}\bar{b}_{+}{\rm Q}_{-}\big{)}\,.$$
where
$$\displaystyle b_{-}=\beta_{-}\,\sqrt{1+v\over 1-v}\,,\quad\bar{b}_{+}=\bar{%
\beta}_{+}\,\sqrt{1-v\over 1+v}\,,\quad{\cal R}=R\,\sqrt{1-v^{2}}\,.$$
(4.20)
We see that all the dependence of $v$ is absorbed in the rescaling of $\beta_{-}$, $\bar{\beta}_{+}$ and $R$, and a shift of the energy by $Pv/(1-v^{2})$. This is in agreement with the consideration in appendix C.
Our solution (4.18, 4.19)
seems to agree with the one proposed in [7]
at least if one switches off their parameters of the deformation by ${\rm J}_{\pm}^{1}$, and parameters of the ${\rm J}\Theta$ and $\bar{\rm J}\overline{\Theta}$ deformations. We see in particular that the parameter dependence of our coefficient ${\cal A}$ and their coefficient $A$ is similar. To check a precise agreement it
would be necessary to match the conventions, and it also
might be necessary to perform an extra redefinition of the parameter of the ${T\overline{T}}$ deformation. We have not tried to do it.
Then, at $v=0$ our solution is very similar to the one proposed in [6] for closely related single-trace deformations by using dual strings on deformed $AdS_{3}$. In the case of ${{\rm J}\overline{T}}$ deformation our formula agrees with theirs. Still, there are some differences. The easiest one to notice is the parameter dependence of our coefficient ${\cal A}$ and their coefficient $A$. To reproduce their dependence we would have to shift $\alpha$ by a term proportional to $\beta_{-}\bar{\beta}_{+}$. Such a shift would ruin the nice structure of the flow equations (4.10), and introduce ${T\overline{T}}$ terms on the r.h.s. of these equations. It is also unclear how their charges are related to ours. By these reasons we have not manage to find an exact map between our and their formulae but we believe they may agree. In appendix D we presents the spectrum deformed by one or two of the operators ${T\overline{T}}$ , ${{\rm J}\overline{T}}$ , and ${\bar{{\rm J}}T}$.
5 Conclusions
In this paper the light-cone gauge approach [8, 9] to the ${T\overline{T}}$ deformed two-dimensional models has been extended to include
the most general ten-parameter deformation by U(1) conserved currents, stress-energy tensor and their various quadratic combinations.
We have found an explicit ten-parameter deformed Hamiltonian for a rather general system of scalars with an arbitrary potential, and used it to derive the flow equations with respect to the parameters, and the spectrum of a deformed CFT with left- and right-moving conserved currents.
There are many open questions and generalisations of the approach. We mention just a few of them.
Clearly, the light-cone gauge approach can also be used to study multi-parameter deformations of the system of scalars, fermions and chiral bosons introduced in [9]. Since these deformations breaks Lorentz symmetry it is natural to analyse non-Lorentz invariant models such as the nonlinear matrix Schrödinger model.
The system of flow equations for the spectrum of a deformed model is complicated, and admits many solutions. We could not find its solution which would be completely determined by the spectrum of the undeformed model. It would be interesting to understand if such a solution exists for a generic model.
Even in the case of a CFT but without left- and right-moving conserved currents it is unclear how to find a unique solution.
If a model is integrable then there is a generalisation of the ${T\overline{T}}$ deformation to more general higher-spin deformations introduced in [2].
It is expected that they can be described in the light-cone gauge approach by coupling strings to W gravity, see [19] for a review.
Higher-spin conserved currents can be also used to study ${{\rm J}\overline{T}}$ type deformations which break Lorentz invariance of an undeformed model, see [20, 21] for a recent discussion. Clearly, as soon as the higher-spin deformations of [2] are understood in the light-cone gauge approach, one can consider multi-parameter deformations involving the higher-spin currents.
Acknowledgements
I would like to thank Tristan McLoughlin for useful discussions.
Appendix A Canonical transformation matrices
The matrices $A_{x}$, $B_{x}$, $A_{p}$ and $B_{p}$ appearing in the canonical transformation (2.15) are
$$\displaystyle A_{x}=\left(\begin{array}[]{ccccccc}1&-\frac{\tilde{a}_{+-}}{A_{%
--}}&\frac{c_{1-}\tilde{a}_{+-}-a_{+1}}{A_{11}}&0&-c_{1-}b_{+1}&-b_{+1}A_{11}&%
0\\
0&\frac{1}{A_{--}}&-\frac{c_{1-}}{A_{11}}&0&0&0&0\\
0&-\frac{a_{1-}}{A_{--}}&\frac{1+c_{1-}a_{1-}}{A_{11}}&0&0&0&0\\
0&0&0&1&0&0&0\\
0&-\frac{b_{--}}{c_{1-}A_{--}}&\frac{b_{--}}{A_{11}}&0&1&0&0\\
0&0&0&0&0&A_{11}&0\\
0&0&\frac{1}{A_{11}}&0&0&0&\frac{1}{A_{11}}\\
\end{array}\right)\,,$$
$$\displaystyle B_{x}=\left(\begin{array}[]{ccccccc}a_{+1}b_{+1}&0&0&-1&\tilde{a%
}_{+-}&\frac{a_{+1}-c_{1-}\tilde{a}_{+-}}{A_{11}}&0\\
0&0&0&0&-1&\frac{c_{1-}}{A_{11}}&0\\
-b_{+1}&0&0&0&a_{1-}&-\frac{1+c_{1-}a_{1-}}{A_{11}}&0\\
1&0&0&0&0&0&0\\
0&A_{--}&0&0&\frac{b_{--}}{c_{1-}}&-\frac{b_{--}}{A_{11}}&0\\
0&0&A_{11}&0&0&0&-A_{11}\\
0&0&0&0&0&0&0\\
\end{array}\right)\,,$$
$$\displaystyle A_{p}=\left(\begin{array}[]{ccccccc}1&a_{+-}&a_{+1}&0&0&b_{+1}&0%
\\
0&A_{--}&0&0&0&0&0\\
0&a_{1-}A_{11}&A_{11}&0&0&0&0\\
0&0&0&1&0&0&0\\
0&\frac{b_{--}}{c_{1-}}&0&0&1&0&0\\
0&-\frac{b_{--}}{A_{11}}&0&0&-\frac{c_{1-}}{A_{11}}&\frac{1}{A_{11}}&0\\
0&-a_{1-}A_{11}&-A_{11}&0&0&0&A_{11}\\
\end{array}\right)\,,$$
$$\displaystyle B_{p}=\left(\begin{array}[]{ccccccc}0&0&0&0&0&0&0\\
0&0&0&0&0&0&0\\
0&\frac{b_{--}}{A_{11}}&0&0&\frac{c_{1-}}{A_{11}}&0&0\\
0&0&0&0&0&0&0\\
0&-c_{1-}a_{1-}&-c_{1-}&0&0&0&0\\
0&0&0&0&0&0&0\\
0&0&0&0&0&0&0\\
\end{array}\right)\,,$$
$$\displaystyle C_{x}=\left(\begin{array}[]{c}0\\
0\\
0\\
0\\
0\\
-c_{1-}\\
0\\
\end{array}\right)\,,\quad C_{p}=\left(\begin{array}[]{c}0\\
-c_{1-}a_{1-}\\
-c_{1-}\\
0\\
0\\
0\\
0\\
\end{array}\right)$$
where
$$\displaystyle\tilde{a}_{+-}={a_{+-}-a_{+1}a_{1-}}\,.$$
Appendix B Virasoro constraints coefficients
The coefficients $\tilde{G}^{rs}$ and so on appearing in the transformed Virasoro constraint $C_{2}$
(3.1)
are given by
$$\displaystyle\tilde{G}^{rs}$$
$$\displaystyle=G^{ab}(A_{p})^{r}{}_{\hat{a}}(A_{p})^{s}{}_{\hat{b}}+G_{ab}(B_{x%
})^{{\hat{a}}r}(B_{x})^{{\hat{b}}s}\,,$$
$$\displaystyle\tilde{G}_{rs}$$
$$\displaystyle=G_{ab}(A_{x})^{\hat{a}}{}_{r}(A_{x})^{\hat{b}}{}_{s}+G^{ab}(B_{p%
})_{r{\hat{a}}}(B_{p})_{s{\hat{b}}}\,,$$
$$\displaystyle\hat{G}^{r}{}_{s}$$
$$\displaystyle=G^{ab}(A_{p})^{r}{}_{\hat{a}}(B_{p})_{s{\hat{b}}}+G_{ab}(B_{x})^%
{{\hat{a}}r}(A_{x})^{\hat{b}}{}_{s}\,,$$
$$\displaystyle\tilde{G}^{rk}$$
$$\displaystyle=G^{ak}(A_{p})^{r}{}_{\hat{a}}\,,\quad\tilde{G}_{rk}=G_{ak}(A_{x}%
)^{\hat{a}}{}_{r}\,,$$
$$\displaystyle\check{G}^{r}{}_{k}$$
$$\displaystyle=G_{ak}(B_{x})^{{\hat{a}}r}\,,\quad\check{G}_{r}{}^{k}=G^{ak}(B_{%
p})_{r{\hat{a}}}\,.$$
Here the indices $a,b=+,-,1$ are in one-to-one correspondence with ${\hat{a}},{\hat{b}}=\hat{1},\hat{2},\hat{3}$, and $r,s=\hat{1},\ldots,\hat{7}$.
The solution (3.2) to the constraints equations $\tilde{C}_{1}=0$, $C_{a}^{V}=0$ and $C^{1}_{\cal U}=0$, and the light-cone gauge conditions $x^{+}=\tau$, $p_{-}=1$, $v^{-}=0$ is given by
$$\displaystyle{\cal X}^{\prime\hat{2}}$$
$$\displaystyle=x^{\prime-}=v^{+}p_{+}-\widetilde{px}^{\prime}+{c_{1-}\over A_{1%
1}}(x^{\prime 1}-v^{1})\,,$$
(B.1)
$$\displaystyle{\cal X}^{\prime\hat{4}}$$
$$\displaystyle=\tilde{X}^{\prime}_{+}=-p_{+}\,,$$
$$\displaystyle{\cal X}^{\prime\hat{5}}$$
$$\displaystyle=\tilde{X}^{\prime}_{-}=\frac{b_{--}}{c_{1-}A_{--}}(v^{+}p_{+}-%
\widetilde{px}^{\prime})-A_{--}+({1\over A_{--}}-1){b_{--}\over A_{11}}(x^{%
\prime 1}-v^{1})\,,$$
$$\displaystyle{\cal X}^{\prime\hat{6}}$$
$$\displaystyle=\tilde{X}^{\prime}_{1}=-p_{1}+u_{1}+{c_{1-}\over A_{11}}\,,$$
$$\displaystyle{\cal X}^{\prime\hat{7}}$$
$$\displaystyle=\Upsilon^{\prime 1}=-x^{\prime 1}\,,$$
where
$$\displaystyle\widetilde{px}^{\prime}=(p_{1}-u_{1})(x^{\prime 1}-v^{1})+p_{k}x^%
{\prime k}\,.$$
The coefficients ${\cal G}_{i}$ appearing in (3.3)
are given by
$$\displaystyle{\cal G}_{2}$$
$$\displaystyle={1\over 2}\tilde{G}^{++}+{1\over 2}\varkappa^{r}\varkappa^{s}%
\tilde{G}_{rs}+\hat{G}^{+}{}_{s}\varkappa^{s}\,,$$
$$\displaystyle{\cal G}_{1}$$
$$\displaystyle=\tilde{G}^{+s^{\prime}}{\cal P}_{s^{\prime}}+{1\over 2}\varkappa%
^{r}\chi^{s}\tilde{G}_{rs}+\tilde{G}^{r^{\prime}k}{\cal P}_{r^{\prime}}p_{k}+%
\varkappa^{r}x^{\prime k}\tilde{G}_{rk}$$
$$\displaystyle+\hat{G}^{+}{}_{s}\chi^{s}+\hat{G}^{r^{\prime}}{}_{s}{\cal P}_{r^%
{\prime}}\varkappa^{s}+\check{G}^{+}{}_{k}x^{\prime k}+\varkappa^{r}p_{k}%
\check{G}_{r}{}^{k}\,.$$
$$\displaystyle{\cal G}_{0}$$
$$\displaystyle={1\over 2}\tilde{G}^{r^{\prime}s^{\prime}}{\cal P}_{r^{\prime}}{%
\cal P}_{s^{\prime}}+{1\over 2}\chi^{r}\chi^{s}\tilde{G}_{rs}+\tilde{G}^{r^{%
\prime}k}{\cal P}_{r^{\prime}}p_{k}+\chi^{r}x^{\prime k}\tilde{G}_{rk}$$
$$\displaystyle+\hat{G}^{r^{\prime}}{}_{s}{\cal P}_{r^{\prime}}\chi^{s}+\check{G%
}^{r^{\prime}}{}_{k}{\cal P}_{r^{\prime}}x^{\prime k}+\chi^{r}p_{k}\check{G}_{%
r}{}^{k}+\tilde{\cal H}_{x}\,,$$
where $k=2,\ldots,n$ are the indices of the remaining transversal coordinates and momenta, and $r^{\prime},s^{\prime}=\hat{2},\ldots,\hat{7}$.
Appendix C Examples of deformed Hamiltonians
In this appendix we consider deformations of a model with the Hamiltonian density
$$\displaystyle{\cal H}_{0}={1\over 2}G^{\mu\nu}(x)p_{\mu}p_{\nu}+{1\over 2}x^{%
\prime\mu}x^{\prime\nu}\,G_{\mu\nu}(x)+V(x)={\cal K}+V\,,$$
(C.1)
where ${\cal K}$ is the kinetic energy, and $G_{\mu\nu}$ and $V$ do not depend on $x^{1}$. The target space metric of the string sigma model is given by (2.11). Models of this type include also models with rotational symmetries where one should first use spherical coordinates to bring the model’s Hamiltonian to the form (C.1).
In all considered examples we switch off the deformations by $J^{1}$, $\widetilde{J}^{1}$, $T^{1}{}_{1}$ and $\widetilde{J}J$, that is we set the parameters $v^{1}$, $u_{1}$ to zero, and $A_{--}$, $A_{11}$ to one. The dependence of these parameters can be easily restored if desired. We use the notations (3.5)
$$\displaystyle\alpha_{\mbox{\tiny${T\overline{T}}$}}\equiv a_{+-}\,,\quad\alpha%
_{\mbox{\tiny${\widetilde{J}T_{0}}$}}\equiv-a_{+1}\,,\quad\alpha_{\mbox{\tiny$%
{JT_{1}}$}}\equiv a_{1-}\,,\quad\alpha_{\mbox{\tiny${JT_{0}}$}}\equiv b_{+1}\,%
,\quad\alpha_{\mbox{\tiny${\widetilde{J}T_{1}}$}}\equiv b_{--}\,,\quad v\equiv%
-v^{+}\,.$$
Deformation by ${T\overline{T}}$ and $T^{1}{}_{0}$
In this case the only nonvanishing parameters are $\alpha\equiv\alpha_{\mbox{\tiny${T\overline{T}}$}}$ and $v\equiv-v^{+}$. Calculating the coefficients ${\cal G}_{i}$, one gets
$$\displaystyle{\cal G}_{2}$$
$$\displaystyle=\tilde{\alpha}(1-v^{2})\,,\quad\tilde{\alpha}=\alpha(1+\alpha V)\,,$$
(C.2)
$$\displaystyle{\cal G}_{1}$$
$$\displaystyle=(1-v^{2})(1+2\alpha V)-2v\,\tilde{\alpha}\,px^{\prime}\,,\quad px%
^{\prime}\equiv p_{\mu}x^{\prime\mu}\,,$$
$$\displaystyle{\cal G}_{0}$$
$$\displaystyle={\cal K}+V(1-v^{2})-\tilde{\alpha}(px^{\prime})^{2}-v(1+2\alpha V%
)px^{\prime}\,.$$
Solving the defining equation (3.3), one gets the deformed Hamiltonian density
$$\displaystyle{\cal H}_{\alpha,v}={1\over 2\tilde{\alpha}}\left(1-\sqrt{1-{4%
\tilde{\alpha}{\cal K}\over 1-v^{2}}+\frac{4\tilde{\alpha}^{2}(px^{\prime})^{2%
}}{\left(1-v^{2}\right)^{2}}}\,\right)+{V\over 1+\alpha V}-\frac{v\,px^{\prime%
}}{1-v^{2}}\,.$$
(C.3)
The 2-parameter deformed model is in fact related to the ${T\overline{T}}$ deformed one. Indeed, if one performs the following rescaling of the momenta, and the world-sheet space coordinate $\sigma$
$$\displaystyle p_{\mu}\to\sqrt{1-v^{2}}\,p_{\mu}\,,\quad\sigma\to{\sigma\over%
\sqrt{1-v^{2}}}\,,$$
(C.4)
then the deformed action (3.4) transforms as
$$\displaystyle S_{A}\to S_{A}=\int_{0}^{R\sqrt{1-v^{2}}}\,{\rm d}\sigma{\rm d}%
\tau\,\big{(}p_{\mu}\dot{x}^{\mu}\,-\,{1\over\sqrt{1-v^{2}}}({\cal H}_{\alpha,%
0}-v\,px^{\prime})\big{)}\,.$$
(C.5)
The last term in the formula is just proportional to the world-sheet momentum $P$, and, therefore, the energy $E_{\alpha,v}(R,P_{v})$ of a state with momentum $P_{v}$ in the 2-parameter deformed model on a circle of circumference $R$ is related to the energy $E_{\alpha,0}(R\sqrt{1-v^{2}}\,,P_{0})$ of the corresponding state with momentum $P_{0}$ in the ${T\overline{T}}$ deformed model on a circle of circumference $R\sqrt{1-v^{2}}$ as
$$\displaystyle E_{\alpha,v}(R,P_{v})={1\over\sqrt{1-v^{2}}}(E_{\alpha,0}(R\sqrt%
{1-v^{2}}\,,P_{0})+v\,P_{0})\,.$$
(C.6)
Due to the momentum quantisation $P_{v}$ and $P_{0}$ are related as $P_{v}=\sqrt{1-v^{2}}\,P_{0}$ which also follows from the transformation (C.4).
Clearly, this simple relation exists only because of the Lorentz invariance of the ${T\overline{T}}$ model.
Deformation by ${T\overline{T}}$, $JT_{0}$, $JT_{1}$, $\tilde{J}T_{0}$ and $\tilde{J}T_{1}$
In this case we set the parameter $v$ of the $T^{1}{}_{0}$ to 0: $v=0$. Calculating the coefficients ${\cal G}_{i}$, one gets
$$\displaystyle{\cal G}_{2}$$
$$\displaystyle=\alpha_{\mbox{\tiny${T\overline{T}}$}}+{1\over 2}\alpha_{\mbox{%
\tiny${JT_{0}}$}}^{2}G_{11}+{1\over 2}\alpha_{\mbox{\tiny${\widetilde{J}T_{0}}%
$}}^{2}G^{11}+(\alpha_{\mbox{\tiny${T\overline{T}}$}}^{2}-\alpha_{\mbox{\tiny$%
{\widetilde{J}T_{0}}$}}^{2}\alpha_{\mbox{\tiny${JT_{0}}$}}^{2})V\,,$$
(C.7)
$$\displaystyle{\cal G}_{1}$$
$$\displaystyle=1+\alpha_{\mbox{\tiny${JT_{1}}$}}\big{(}p_{1}+(\alpha_{\mbox{%
\tiny${\widetilde{J}T_{1}}$}}-\alpha_{\mbox{\tiny${JT_{0}}$}}G_{11})px^{\prime%
}\big{)}$$
$$\displaystyle-\alpha_{\mbox{\tiny${\widetilde{J}T_{0}}$}}(p_{\mu}G^{\mu 1}-%
\alpha_{\mbox{\tiny${JT_{0}}$}}px^{\prime}+\alpha_{\mbox{\tiny${\widetilde{J}T%
_{1}}$}}G^{11}px^{\prime})+\alpha_{\mbox{\tiny${\widetilde{J}T_{1}}$}}x^{%
\prime 1}-\alpha_{\mbox{\tiny${JT_{0}}$}}G_{1\mu}x^{\prime\mu}\big{]}$$
$$\displaystyle+4V\big{[}\alpha_{\mbox{\tiny${T\overline{T}}$}}\big{(}1+\alpha_{%
\mbox{\tiny${\widetilde{J}T_{0}}$}}\alpha_{\mbox{\tiny${JT_{0}}$}}px^{\prime}+%
\alpha_{\mbox{\tiny${JT_{1}}$}}(p_{1}+\alpha_{\mbox{\tiny${\widetilde{J}T_{1}}%
$}}px^{\prime})+\alpha_{\mbox{\tiny${\widetilde{J}T_{1}}$}}x^{\prime 1}\big{)}$$
$$\displaystyle+\alpha_{\mbox{\tiny${\widetilde{J}T_{0}}$}}\alpha_{\mbox{\tiny${%
JT_{0}}$}}\big{(}\alpha_{\mbox{\tiny${JT_{0}}$}}(p_{1}+\alpha_{\mbox{\tiny${%
\widetilde{J}T_{1}}$}}px^{\prime})+\alpha_{\mbox{\tiny${\widetilde{J}T_{0}}$}}%
(\alpha_{\mbox{\tiny${JT_{1}}$}}px^{\prime}+x^{\prime 1})\big{)}\,,$$
$$\displaystyle{\cal G}_{0}$$
$$\displaystyle={\cal K}+{1\over 2}px^{\prime}\big{[}\alpha_{\mbox{\tiny${JT_{1}%
}$}}^{2}G_{11}px^{\prime}+\alpha_{\mbox{\tiny${\widetilde{J}T_{1}}$}}^{2}G^{11%
}px^{\prime}-2\alpha_{\mbox{\tiny${JT_{0}}$}}(p_{1}+\alpha_{\mbox{\tiny${%
\widetilde{J}T_{1}}$}}px^{\prime})$$
$$\displaystyle\qquad\quad-2(-\alpha_{\mbox{\tiny${\widetilde{J}T_{1}}$}}p_{\mu}%
G^{\mu 1}+\alpha_{\mbox{\tiny${JT_{1}}$}}\alpha_{\mbox{\tiny${\widetilde{J}T_{%
0}}$}}px^{\prime}+\alpha_{\mbox{\tiny${T\overline{T}}$}}px^{\prime}+\alpha_{%
\mbox{\tiny${\widetilde{J}T_{0}}$}}x^{\prime 1})+2\alpha_{\mbox{\tiny${JT_{1}}%
$}}G_{1\mu}x^{\prime\mu}\big{]}$$
$$\displaystyle\qquad\quad+V\big{[}\big{(}1+\alpha_{\mbox{\tiny${JT_{1}}$}}(p_{1%
}+\alpha_{\mbox{\tiny${\widetilde{J}T_{1}}$}}px^{\prime})+\alpha_{\mbox{\tiny$%
{\widetilde{J}T_{1}}$}}x^{\prime 1}\big{)}^{2}$$
$$\displaystyle\qquad\quad-\big{(}\alpha_{\mbox{\tiny${JT_{1}}$}}\alpha_{\mbox{%
\tiny${\widetilde{J}T_{0}}$}}px^{\prime}+\alpha_{\mbox{\tiny${T\overline{T}}$}%
}px^{\prime}+\alpha_{\mbox{\tiny${JT_{0}}$}}(p_{1}+\alpha_{\mbox{\tiny${%
\widetilde{J}T_{1}}$}}px^{\prime})+\alpha_{\mbox{\tiny${\widetilde{J}T_{0}}$}}%
x^{\prime 1}\big{)}^{2}\big{]}\,.$$
The Hamiltonian density ${\cal H}_{A}$ is then found by solving the defining equation (3.3). The resulting Hamiltonian is very complicated, and it is unclear how one could find its spectrum even in the CFT limit where $V=0$. The situation becomes better if $G_{11}=1$ (or a constant), and $G_{1k}=0\,,\,k=2,\ldots,n$. Then, the model has left- and right-moving conserved currents, and one can redefine the deformation parameters to simplify the Hamiltonian. This is done in the next subsection.
Deformed CFT with left- and right-moving currents
In this subsection we use parameters $\alpha\equiv\alpha_{\mbox{\tiny${T\overline{T}}$}}$, $v\equiv-v^{+}$ and $\beta_{\pm}$, $\bar{\beta}_{\pm}$, and consider deformations of a CFT with the Hamiltonian
$$\displaystyle{\cal H}_{0}={1\over 2}p_{1}^{2}+{1\over 2}(x^{\prime 1})^{2}+{1%
\over 2}G^{kl}(x)p_{k}p_{l}+{1\over 2}x^{\prime k}x^{\prime l}\,G_{kl}(x)%
\equiv{\cal K}\,,$$
(C.8)
by ${T\overline{T}}$, $JT_{+}\leftrightarrow J\Theta$, $JT_{-}\leftrightarrow J\overline{T}$,
$\bar{J}T_{+}\leftrightarrow\bar{J}T$, $\bar{J}T_{-}\leftrightarrow\bar{J}\overline{\Theta}$ and $T^{1}{}_{0}$.
Calculating the coefficients ${\cal G}_{i}$ by choosing $\varkappa$’s as in (4.9), one gets
$$\displaystyle{\cal G}_{2}$$
$$\displaystyle=\alpha(1-v^{2})+{1\over 2}\beta_{-}^{2}(1+v)^{2}+{1\over 2}\bar{%
\beta}_{+}^{2}(1-v)^{2}\,,$$
(C.9)
$$\displaystyle{\cal G}_{1}$$
$$\displaystyle=1-v^{2}-2v\alpha px^{\prime}+\big{(}\beta_{-}^{2}(1+v)-\bar{%
\beta}_{+}^{2}(1-v)\big{)}px^{\prime}$$
$$\displaystyle-\bar{\beta}_{+}(p_{1}-x^{\prime 1})(1-v)-\beta_{-}(p_{1}+x^{%
\prime 1})(1+v)\,,$$
$$\displaystyle{\cal G}_{0}$$
$$\displaystyle={\cal K}-\alpha(px^{\prime})^{2}+px^{\prime}\big{(}\bar{\beta}_{%
+}(p_{1}-x^{\prime 1})-\beta_{-}(p_{1}+x^{\prime 1})+{1\over 2}(\bar{\beta}_{+%
}^{2}+\beta_{-}^{2})px^{\prime}-v\big{)}\,.$$
We see that ${\cal G}_{i}$, and therefore the deformed Hamiltonian, do not depend on $\beta_{+}$ and $\bar{\beta}_{-}$.
Thus, for a CFT with left- and right-moving conserved currents there is no deformation by $JT_{+}\leftrightarrow J\Theta$ and $\bar{J}T_{-}\leftrightarrow\bar{J}\overline{\Theta}$ if one chooses correctly the parameter of the ${T\overline{T}}$ deformation. Let us also mention that if the potential $V$ does not vanish then ${\cal G}_{i}$ do depend on $\beta_{+}$ and $\bar{\beta}_{-}$. The Hamiltonian of the model is obviously given by
$$\displaystyle{\cal H}_{A}={\cal H}(R,\alpha,\beta_{-},\bar{\beta}_{+},v)={{%
\cal G}_{1}-\sqrt{{\cal G}_{1}^{2}-4{\cal G}_{2}{\cal G}_{0}}\over 2{\cal G}_{%
2}}\,,$$
(C.10)
but its surprising feature is that its $v$-dependence is very similar to the one in (C.5).
Indeed it is easy to check that the deformed action
$$\displaystyle S_{A}=\int_{0}^{R}\,{\rm d}\sigma{\rm d}\tau\,\big{(}p_{\mu}\dot%
{x}^{\mu}\,-\,{\cal H}(\alpha,\beta_{-},\bar{\beta}_{+},v)\big{)}\,,$$
(C.11)
satisfies
$$\displaystyle S_{A}=\int_{0}^{R\sqrt{1-v^{2}}}\,{\rm d}\sigma{\rm d}\tau\,\big%
{(}p_{\mu}\dot{x}^{\mu}\,-\,{1\over\sqrt{1-v^{2}}}(\,{\cal H}(\alpha,b_{-},%
\bar{b}_{+},0)-v\,px^{\prime}\,)\big{)}\,.$$
(C.12)
where
$$\displaystyle b_{-}=\beta_{-}\,\sqrt{1+v\over 1-v}\,,\quad\bar{b}_{+}=\bar{%
\beta}_{+}\,\sqrt{1-v\over 1+v}\,.$$
(C.13)
Thus, up to the rescaling of $\beta_{-}$, $\bar{\beta}_{+}$, the energy of the model with nonvanishing $v$ is related to its energy with $v=0$ as in
(C.6).
Let us also calculate the coefficients ${\cal G}_{i}$ with another interesting
choice of $\varkappa$’s
$$\displaystyle\varkappa_{+}=-1\,,\quad\varkappa_{-}=-1\,,\quad\bar{\varkappa}_{%
+}=-1\,,\quad\bar{\varkappa}_{-}=-1\,,\quad\varkappa_{+-}=1\,,\quad\bar{%
\varkappa}_{+-}=1\,.$$
(C.14)
Then, the flow equations with respect to $\beta_{-}$ and $\bar{\beta}_{+}$ take the form
$$\displaystyle{\partial{\cal H}_{A}\over\partial\beta_{-}}=2{\cal J}{\mathbb{T}%
}_{-}\,-\,{\beta_{-}}{\mathbb{T}}\overline{\mathbb{T}}\,,\qquad{\partial{\cal H%
}_{A}\over\partial\bar{\beta}_{+}}=2\bar{\cal J}{\mathbb{T}}_{+}-{\bar{\beta}_%
{+}}{\mathbb{T}}\overline{\mathbb{T}}\,.$$
(C.15)
The coefficients ${\cal G}_{i}$ with $\varkappa$’s as in (C.14) are given by
$$\displaystyle{\cal G}_{2}$$
$$\displaystyle=\alpha(1-v^{2})+\beta_{-}^{2}(1+v)v-\bar{\beta}_{+}^{2}(1-v)v\,,$$
(C.16)
$$\displaystyle{\cal G}_{1}$$
$$\displaystyle=1-v^{2}-2v\alpha px^{\prime}+\big{(}\beta_{-}^{2}(1+2v)-\bar{%
\beta}_{+}^{2}(1-2v)\big{)}px^{\prime}$$
$$\displaystyle-\bar{\beta}_{+}(p_{1}-x^{\prime 1})(1-v)-\beta_{-}(p_{1}+x^{%
\prime 1})(1+v)\,,$$
$$\displaystyle{\cal G}_{0}$$
$$\displaystyle={\cal K}-\alpha(px^{\prime})^{2}+px^{\prime}\big{(}\bar{\beta}_{%
+}(p_{1}-x^{\prime 1})-\beta_{-}(p_{1}+x^{\prime 1})+(\bar{\beta}_{+}^{2}+%
\beta_{-}^{2})px^{\prime}-v\big{)}\,.$$
Now, setting $v=0$, one gets
$$\displaystyle{\cal G}_{2}$$
$$\displaystyle=\alpha\,,$$
(C.17)
$$\displaystyle{\cal G}_{1}$$
$$\displaystyle=1+\big{(}\beta_{-}^{2}-\bar{\beta}_{+}^{2}\big{)}px^{\prime}-%
\bar{\beta}_{+}(p_{1}-x^{\prime 1})-\beta_{-}(p_{1}+x^{\prime 1})\,,$$
$$\displaystyle{\cal G}_{0}$$
$$\displaystyle={\cal K}-\alpha(px^{\prime})^{2}+px^{\prime}\big{(}\bar{\beta}_{%
+}(p_{1}-x^{\prime 1})-\beta_{-}(p_{1}+x^{\prime 1})+(\bar{\beta}_{+}^{2}+%
\beta_{-}^{2})px^{\prime}\big{)}\,.$$
Finally, setting $\alpha=0$, and solving the defining equation (3.3), one gets the deformed Hamiltonian
$$\displaystyle{\cal H}_{\beta_{-},\bar{\beta}_{+}}={{\cal G}_{0}\over{\cal G}_{%
1}}={{\cal K}-\alpha(px^{\prime})^{2}+px^{\prime}\big{(}\bar{\beta}_{+}(p_{1}-%
x^{\prime 1})-\beta_{-}(p_{1}+x^{\prime 1})+(\bar{\beta}_{+}^{2}+\beta_{-}^{2}%
)px^{\prime}\big{)}\over 1+\big{(}\beta_{-}^{2}-\bar{\beta}_{+}^{2}\big{)}px^{%
\prime}-\bar{\beta}_{+}(p_{1}-x^{\prime 1})-\beta_{-}(p_{1}+x^{\prime 1})}\,,$$
(C.18)
which satisfies the flow equations (C.15).
The flow equations are now more complicated but
a curious feature of this choice is that setting $v=\alpha=0$, one gets Hamiltonian (C.18) which is a rational function of the coordinates and momenta. It suggests that
it is the ${T\overline{T}}$ deformation which is responsible for the square root form of the deformed Hamiltonian.
Appendix D Examples of the deformed spectrum
In this appendix we specialise eqs.(4.18, 4.19)
to the one- and two-parameter deformations.
Deformation by ${T\overline{T}}$
Setting $v=0$, $\beta_{-}=0$, $\bar{\beta}_{+}=0$, one gets
$$\displaystyle{E_{n}\over R}$$
$$\displaystyle={1-\sqrt{1-{4\alpha\over R}E_{n}^{(0)}+\frac{4\alpha^{2}}{R^{2}}%
P^{2}}\over 2\alpha}\,,$$
(D.1)
which is a well-known result.
Deformation by ${{\rm J}\overline{T}}$ and $\bar{J}T$
Setting $v=0$, $\alpha=0$, $\bar{\beta}_{+}=0$, one gets
$$\displaystyle{E_{n}+P\over R}$$
$$\displaystyle=\frac{2{\rm Q}_{+}}{R\,\beta_{-}}+{1-\sqrt{\big{(}1+{2\beta_{-}%
\over R}{\rm Q}_{+}\big{)}^{2}-{2\beta_{-}^{2}\over R}(E_{n}^{(0)}+P)}\over%
\beta_{-}^{2}}\,,$$
(D.2)
which agrees with [4].
Let us also give the formula for the $\bar{J}T$ deformation
$$\displaystyle{E_{n}^{\bar{J}T}-P\over R}$$
$$\displaystyle=\frac{2{\rm Q}_{-}}{R\,\beta_{+}}+{1-\sqrt{\big{(}1+{2\beta_{+}%
\over R}{\rm Q}_{-}\big{)}^{2}-{2\beta_{+}^{2}\over R}(E_{n}^{(0)}-P)}\over%
\beta_{-}^{2}}\,.$$
(D.3)
Deformation by ${T\overline{T}}$+${{\rm J}\overline{T}}$
Setting $v=0$, $\bar{b}_{+}=0$, one gets
$$\displaystyle{E_{n}+P\over R}$$
$$\displaystyle=\frac{2\alpha P+2\beta_{-}{\rm Q}_{+}}{(2\alpha+\beta_{-}^{2})R}%
+{1-\sqrt{\big{(}1+{2\alpha\over R}P+{2\beta_{-}\over R}{\rm Q}_{+}\big{)}^{2}%
-{2(2\alpha+\beta_{-}^{2})\over R}(E_{n}^{(0)}+P)}\over 2\alpha+\beta_{-}^{2}}\,,$$
(D.4)
which agrees with [6, 7].
Deformation by ${{\rm J}\overline{T}}$+$\bar{J}T$
Setting $v=0$, $\bar{b}_{+}=0$, one gets
$$\displaystyle{E_{n}\over R}=\frac{(\bar{\beta}_{+}^{2}-\beta_{-}^{2})P+2\beta_%
{-}{\rm Q}_{+}+2\beta_{+}{\rm Q}_{-}}{(\beta_{-}^{2}+\bar{\beta}_{+}^{2})R}$$
(D.5)
$$\displaystyle+{1-\sqrt{\big{(}1+{2\beta_{-}{\rm Q}_{+}+2\beta_{+}{\rm Q}_{-}%
\over R}\big{)}^{2}-{2\beta_{-}^{2}(E_{n}^{(0)}+P)+2\beta_{+}^{2}(E_{n}^{(0)}-%
P)\over R}-{4\beta_{m}\bar{\beta}_{+}P(\beta_{m}\bar{\beta}_{+}P+2\beta_{-}{%
\rm Q}_{-}-2\bar{\beta}_{+}{\rm Q}_{+})\over R^{2}}}\over\beta_{-}^{2}+\bar{%
\beta}_{+}^{2}}\,.$$
References
[1]
A. B. Zamolodchikov,
“Expectation value of composite field T anti-T in two-dimensional quantum field theory,”
hep-th/0401146.
[2]
F. A. Smirnov and A. B. Zamolodchikov,
“On space of integrable quantum field theories,”
Nucl. Phys. B 915 (2017) 363
doi:10.1016/j.nuclphysb.2016.12.014
[arXiv:1608.05499 [hep-th]].
[3]
M. Guica,
“An integrable Lorentz-breaking deformation of two-dimensional CFTs,”
SciPost Phys. 5 (2018) no.5, 048
doi:10.21468/SciPostPhys.5.5.048
[arXiv:1710.08415 [hep-th]].
[4]
S. Chakraborty, A. Giveon and D. Kutasov,
“$J\overline{T}$ deformed CFT${}_{2}$ and string theory,”
JHEP 1810 (2018) 057
doi:10.1007/JHEP10(2018)057
[arXiv:1806.09667 [hep-th]].
[5]
A. Bzowski and M. Guica,
“The holographic interpretation of $J\bar{T}$-deformed CFTs,”
JHEP 1901 (2019) 198
doi:10.1007/JHEP01(2019)198
[arXiv:1803.09753 [hep-th]].
[6]
S. Chakraborty, A. Giveon and D. Kutasov,
“$T\bar{T}$, $J\bar{T}$, $T\bar{J}$ and String Theory,”
arXiv:1905.00051 [hep-th].
[7]
B. Le Floch and M. Mezei,
“Solving a family of $T\bar{T}$-like theories,”
arXiv:1903.07606 [hep-th].
[8]
M. Baggio, A. Sfondrini, G. Tartaglino-Mazzucchelli and H. Walsh,
“On $T\bar{T}$ deformations and supersymmetry,”
arXiv:1811.00533 [hep-th].
[9]
S. Frolov,
“TTbar deformation and the light-cone gauge,”
arXiv:1905.07946 [hep-th].
[10]
L. F. Alday, G. Arutyunov and S. Frolov,
“New integrable system of 2dim fermions from strings on AdS,”
JHEP 0601 (2006) 078, hep-th/0508140.
[11]
A. Cavaglià, S. Negro, I. M. Szécsényi and R. Tateo,
“$T\bar{T}$-deformed 2D Quantum Field Theories,”
JHEP 1610 (2016) 112
doi:10.1007/JHEP10(2016)112
[arXiv:1608.05534 [hep-th]].
[12]
G. Bonelli, N. Doroud and M. Zhu,
“$T\bar{T}$-deformations in closed form,”
JHEP 1806 (2018) 149
doi:10.1007/JHEP06(2018)149
[arXiv:1804.10967 [hep-th]].
[13]
R. Conti, L. Iannella, S. Negro and R. Tateo,
“Generalised Born-Infeld models, Lax operators and the $\mathrm{T}\overline{\mathrm{T}}$ perturbation,”
JHEP 1811 (2018) 007
doi:10.1007/JHEP11(2018)007
[arXiv:1806.11515 [hep-th]].
[14]
H. Jiang, A. Sfondrini and G. Tartaglino-Mazzucchelli,
“$T\bar{T}$ deformations with $\mathcal{N}=(0,2)$ supersymmetry,”
arXiv:1904.04760 [hep-th].
[15]
C. K. Chang, C. Ferko and S. Sethi,
“Supersymmetry and $T\overline{T}$ deformations,”
JHEP 1904 (2019) 131
doi:10.1007/JHEP04(2019)131
[arXiv:1811.01895 [hep-th]].
[16]
C. K. Chang, C. Ferko, S. Sethi, A. Sfondrini and G. Tartaglino-Mazzucchelli,
“$T\bar{T}$ Flows and (2,2) Supersymmetry,”
arXiv:1906.00467 [hep-th].
[17]
E. A. Coleman, J. Aguilera-Damia, D. Z. Freedman and R. M. Soni,
“$T\bar{T}$-Deformed Actions and (1,1) Supersymmetry,”
arXiv:1906.05439 [hep-th].
[18]
G. Arutyunov, S. Frolov and M. Zamaklar,
“Finite-size effects from giant magnons,”
Nucl. Phys. B 778 (2007) 1,
hep-th/0606126.
[19]
C. M. Hull,
“Lectures on W gravity, W geometry and W strings,”
In *Trieste 1992, Proceedings, High energy physics and cosmology* 76-142 and London Queen Mary and Westfield Coll. - QMW-93-02 (93/02) 54 p. (305933)
[hep-th/9302110].
[20]
R. Conti, S. Negro and R. Tateo,
“Conserved currents and $\text{T}\bar{\text{T}}_{s}$ irrelevant deformations of 2D integrable field theories,”
arXiv:1904.09141 [hep-th].
[21]
B. Le Floch and M. Mezei,
“KdV charges in $T\bar{T}$ theories and new models with super-Hagedorn behavior,”
arXiv:1907.02516 [hep-th]. |
Hadronic $B$ Decay
T. E. Browder
Physics Department, University of Hawaii at Manoa,
Honolulu, HI, 96822, USA
Abstract
We review recent experimental
results from CLEO and LEP
experiments on hadronic decays of hadrons containing $b$
quarks[1]. We discuss charm counting and the
semileptonic branching fraction in $B$ decays and
the color suppressed amplitude in $B$ decay.
1 Charm counting and the semileptonic branching fraction
A complete picture of inclusive
$B$ decay is beginning to emerge from recent measurements
by CLEO II and the LEP experiments.[1]
These measurements can be used to address
the question of whether the hadronic decay of the $B$ meson is compatible with
its semileptonic branching fraction.
Three facts emerge from the experimental examination of inclusive $B$
decay at the $\Upsilon(4S)$:
$$n_{c}=1.10\pm 0.05$$
(1)
where $n_{c}$ is the number of charm quarks
produced per $B$ decay from recent CLEO II
results[3]
and using
${\cal B}(D^{0}\to K^{-}\pi^{+})=(3.91\pm 0.08\pm 0.17\%)$.[4]
$${\cal B}(B\to X\ell\nu)=10.23\pm 0.39\%.$$
(2)
This value is the
average of the CLEO and ARGUS model independent measurements using
dileptons.[2]
We note that the value used by the LEP Electroweak Working Group
for ${\cal B}(b\to X\ell\nu)=11.16\pm 0.20\%$ is only marginally
consistent with the $\Upsilon(4S)$ results.
The third quantity, ${\cal B}(b\to c\bar{c}s)$,
is calculated from the inclusive $B\to D_{s}$,
$B\to(c\bar{c})X$, and $B\to\Xi_{c}$ branching fractions, and is
$${\cal B}(b\to c\bar{c}s)=14.0\pm 2.8\%.$$
(3)
The above value is determined assuming no contribution from $B\to D$ decays,
an assumption which can be checked using data and is discussed in further
detail below.
In the parton model, it is
difficult to accomodate a low semileptonic branching fraction unless the
hadronic width of the $B$ meson is increased.[5]
The explanations for the semileptonic branching fraction
which have been proposed can be distinguished by expressing
the hadronic width of the $B$ meson in terms of three components:
$$\Gamma_{hadronic}(b)=\Gamma(b\to c\bar{c}s)+\Gamma(b\to c\bar{u}d)+\Gamma(b\to
s%
~{}g).$$
If the semileptonic
branching fraction is to be reduced to the observed level, then one of the
components must be enhanced.
A large number of explanations for the low semileptonic branching fraction
and charm yield have been proposed in the last few years.
These explanations can be logically classified as follows:
1.
An enhancement of $b\to c\bar{c}s$ due to large QCD corrections
or the breakdown of local duality. A variety of possible
experimental signatures
have been suggested.[6]${}^{,}$[7]${}^{,}$
[8]${}^{,}$[9]${}^{,}$[10]${}^{,}$[11]
2.
An enhancement of $b\to c\bar{u}d$ due to non-perturbative effects.
[12]${}^{,}$[13]${}^{,}$[14]${}^{,}$[15]
3.
An enhancement of $b\to s~{}g$ or $b\to d~{}g$
from New Physics.[16]${}^{,}$[17]${}^{,}$[18]
4.
The cocktail solution: For example,
if both the $b\to c\bar{c}s$ and the
$b\to c\bar{u}d$ mechanisms are increased,
this could suffice to explain the inclusive observations.
5.
There might also be a
systematic experimental problem in the determination of either $n_{c}$,
${\cal B}(b\to c\bar{c}s)$, or ${\cal B}(B\to X\ell\nu)$.[19]
Inclusive charm particle-lepton correlations can be used to probe
the $B$ decay mechanism and give further insight into this problem.
The correlation of the lepton charge and the charm particle flavor
distinguishes between different production mechanisms.
High momentum leptons, $p_{\ell}>1.4$ GeV, are used
to tag the flavor of the $B$. The angular correlation between the meson
and the lepton is then employed to select events in which the tagging lepton
and meson are from different $B$s. When the lepton and meson
originate from the same $B$ meson they tend to be back to back, whereas when
the meson and leptons come
from different $B$ mesons they are uncorrelated.
After this separation is performed, wrong sign
charge correlations from $B-\bar{B}$
mixing must be subtracted. Since the mixing rate is well measured,
this correction is straightforward and has little uncertainty.
This technique has been applied previously to several types of
correlations of charmed hadrons and leptons.
For example, the sign of $\Lambda_{c}$-lepton
correlations distinguishes between the
$b\to c\bar{u}d$ and the $b\to c\bar{c}s$ mechanisms.
It was found that the $b\to c\bar{c}s$ mechanism comprises
$19\pm 13\pm 4\%$ of $B\to\Lambda_{c}$ decays[20].
This observation effectively ruled out one proposed source of
additional $b\to c\bar{c}s$ decays.[6]
Similiarly, examination of the sign of
$D_{s}$-lepton correlations shows that
most $D_{s}$ mesons originate from $b\to c\bar{c}s$ rather than
from $b\to c\bar{u}d$ with $s\bar{s}$ quark popping at the lower vertex.
In this case, it was found that $17.2\pm 7.9\pm 2.6\%$ of $D_{s}$ mesons
originate from the latter mechanism[21].
The same experimental
technique has now been applied to $D$-lepton correlations.
The conventional $b\to c\bar{u}d$ mechanism which was previously assumed to be
responsible for all $D$ production in $B$ decay will give $D\ell^{-}$
correlations.
If a significant fraction of $D$ mesons
arise from $b\to c\bar{c}s$ with light quark popping at the
upper vertex as proposed by Buchalla, Dunietz,
and Yamamoto significant wrong sign $D\ell^{+}$ correlations will be
observed.[7]
Final results of this study have been presented by CLEO II which finds,
$\Gamma(B\to D~{}X)/\Gamma(B\to\bar{D}X)=0.100\pm 0.026\pm 0.016$.[22]
This implies a new contribution to the $b\to c\bar{c}s$ width
$${\cal B}(B\to DX)=7.9\pm 2.2\%$$
ALEPH finds evidence for semi-inclusive
$B\to D^{0}\bar{D^{0}}X+D^{0}D^{\mp}X$ decays with a somewhat larger branching
fraction of $12.8\pm 2.7\pm 2.6\%$.[23]
DELPHI reports the observation of
$B\to D^{*+}D^{*-}X$ decays with a branching fraction
of $1.0\pm 0.2\pm 0.3\%$.[29]
Additional and quite compelling evidence that these
signals are due to $B\to D^{(*)}\bar{D}^{(*)}K^{(*)}$ decays
has been presented by CLEO[24], which has observed
fully reconstructed signals in exclusive modes:
$${\cal B}(\bar{B}^{0}\to D^{*+}\bar{D}^{0}K^{-})=0.45^{+0.25}_{-0.19}\pm 0.08\%$$
$${\cal B}(B^{-}\to D^{*0}\bar{D}^{0}K^{-})=0.54^{+0.33}_{-0.24}\pm 0.12\%$$
$${\cal B}(\bar{B}^{0}\to D^{*+}\bar{D}^{*0}K^{-})=1.30^{+0.61}_{-0.47}\pm 0.27\%$$
$${\cal B}(B^{-}\to D^{*0}\bar{D}^{*0}K^{-})=1.45^{+0.78}_{-0.58}\pm 0.36\%$$
The rates observed by ALEPH and DELPHI are
consistent with the rate of wrong sign
$D$-lepton correlation reported by CLEO.[22]
It is possible that these channels are actually resonant modes
of the form $B\to D{D}_{s}^{**}$
decays, where the p-wave $D_{s}^{**}$ or radially excited $D_{s}^{{}^{\prime}}$ state
decays to $\bar{D}^{(*)}\bar{K}$.[30] A direct
search by CLEO has ruled
out the possibility of narrow $B\to D_{s_{1}}X$ decay:
${\cal B}(B\to D_{s1}^{+}X)<0.95\%$ at the 90% confidence level.[31]
There are other implications of these observations.
A $B$ decay mechanism with a ${\cal O}(10\%)$ branching
fraction has been found
which was not previously included in the CLEO or LEP
Monte Carlo simulations of $B$ decay. This may have consequences
for other analyses of particle-lepton correlations. For example,
CLEO has re-examined the model independent dilepton measurement
of ${\cal B}(B\to X\ell\nu)$. Due to the lepton threshold of 0.6 GeV
and the soft spectrum of leptons, the CLEO measurement
is fortuitously unchanged. It is also important to check the size
of this effect in LEP measurements of the $B$ semileptonic branching fraction
using dileptons.
We can now recalculate[22]
$${\cal B}(b\to c\bar{c}s)=21.9\pm 3.7\%$$
which would suggest a somewhat larger charm yield ($n_{c}\sim 1.22$).
This supports hypothesis (1), large QCD corrections in $b\to c\bar{c}s$
BUT the charm yield $n_{c}$ as computed in the usual way
is unchanged. Moreover, the contribution of
$B\to D\bar{D}KX$ decays was properly accounted for
in the computation of $n_{c}$. This suggests that the experimental
situation is still problematic.
One possibility that must be addressed is whether
there could be an error in
the normalization ${\cal B}(D^{0}\to K^{-}\pi^{+})$.[19]
This branching fraction calibrates the inclusive measurements of
$B\to D^{0}$, $B\to D^{+}$, and $B\to D_{s}$ rates as well as
$n_{c}$. Historically, a flaw in
${\cal B}(D^{0}\to K^{-}\pi^{+})$ has been the culprit in other consistency
problems with charm counting.
The most precise measurements of ${\cal B}(D^{0}\to K^{-}\pi^{+})$
are obtained by fitting the $p_{T}$ spectrum
of soft pions in charm jets. An examination of Table 1 shows
that these measurements are statistically precise but systematics
dominated. The Particle Data Group world average is currently dominated by
the ALEPH measurement.[25]
Using $B$ decay data it is also possible to make consistency checks on the
calibration branching fraction. For example, the ratio
$${{\cal B}(B\to D^{*}\ell\nu)_{\rm partial}}\over{{\cal B}(B\to D^{*}\ell\nu)_{%
\rm full}}$$
where the decay in the
numerator is observed without reconstructing the $D$ decay gives
a measurement of the calibration branching fraction with very different
systematic effects. In CLEO data, this method
gives
${\cal B}(D^{0}\to K^{-}\pi^{+})=3.81\pm 0.15\pm 0.16\%$.[26]
Another quantity which can be examined is
$${{\cal B}(B\to DX\ell\nu)}\over{{\cal B}(B\to X\ell\nu)}$$
which should be unity modulo small corrections for semileptonic $D_{s}$ and
baryon production as well as for processes
which do not produce charm. Applied
to CLEO data, this method gives
${\cal B}(D^{0}\to K^{-}\pi^{+})=3.69\pm 0.08\pm 0.17\%$.[22]
Neither method is sufficiently precise yet to
conclusively demonstrate that either the
$D$ branching fraction scale is correct or that it has a systematic flaw.
Another possibility is enhanced ${\cal B}(b\to cu\bar{d})$.
On the theoretical side, Bagan et al. find that at next to leading order,
$$r_{ud}={{{\cal B}(b\to c\bar{u}d)}\over{{\cal B}(b\to c\ell\nu)}}=4.0\pm 0.4$$
The value of ${\cal B}(b\to c\bar{u}d)$
can be checked using measurements
of inclusive $B$ decay from the $\Upsilon(4S)$ experiments:
$${\cal B}(b\to c\bar{u}d)_{exp}={\cal B}(B\to DX)+{\cal B}(B\to\Lambda_{c}X)$$
$$-{\cal B}(B\to DD_{s}X)-2{\cal B}(B\to D\bar{D}KX)-2.25{\cal B}(b\to c\ell\nu)$$
$$=(0.871\pm 0.035)+(0.036\pm 0.020)$$
$$-(0.10\pm 0.027)-2\times(0.079\pm 0.022)-(0.236\pm 0.010)$$
$${\cal B}(b\to c\bar{u}d)_{exp}=0.41\pm 0.07$$
In the above calculation, a small correction ($0.004$) has been applied
to the $B\to\Lambda_{c}X$ branching fraction to account
for $b\to c\bar{c}s$ production in baryonic $B$ decay.
The factor of $2.25$ accounts for phase space suppression in
$b\to c\tau\nu$ decay.
The experimental result is consistent with the theoretical
expectation,
$${\cal B}(b\to c\bar{u}d)_{theory}=0.42\pm 0.04$$
However, the present experimental accuracy is not quite sufficient
to completely rule out $b\to c\bar{u}d$ as the cause of the discrepancy.
We note that ALEPH and OPAL have recently reported a value for
$n_{c}$ in $Z\to b\bar{b}$ decay.[27],[28]
ALEPH finds $n_{c}^{Z}=1.230\pm 0.036\pm 0.038\pm 0.053$.
The rate of $D_{s}$ and $\Lambda_{c}$ production is significantly higher
than what is observed at the $\Upsilon(4S)$. It is not clear whether
the quantity being measured is the same as $n_{c}$ at the $\Upsilon(4S)$, which
would be the case if the spectator model holds and if the contribution
from the other $b$-hadrons, $B_{s}$ and $\Lambda_{b}$, could be neglected.
OPAL reports a somewhat lower value of $n_{c}=1.10\pm 0.045\pm 0.060\pm 0.037$
after correcting for unseen charmonium states.
OPAL assumes no contribution from $\Xi_{c}$ production while ALEPH includes
a very large contribution from this source. The
contribution of $B\to$baryon decays to charm counting
as well as the $\Lambda_{c}$, $\Xi_{c}$ branching fraction scales are still poorly
measured and definitely merit further investigation.
2 Exclusive Hadronic Decays
Recent progress has been made on partial reconstruction of hadronic
$B$ decays.[32]
For example, the decay chain
$$B\to D^{*}\pi_{f},D^{*}\to(D)\pi_{s}$$
can be measured without reconstructing the $D$ meson. In this reaction,
there are five particles ($B,D^{*},D,\pi_{s},\pi_{f}$) with five 4-momenta
give 20 unknowns. The 4-momenta of the
$\pi_{s},\pi_{f}$ are measured which gives 8 constraints.
The $B,D,D^{*}$ masses and beam energy are known and gives 4 constraints.
Then energy-momentum conservation in the $B\to D^{*}\pi_{f}$ and $D^{*}\to D\pi_{s}$
decay chains gives 8 additional constraints. Thus, one can
perform a $20-8-8-4=0$ C fit.
Two variables are used to extract the signal:
$\cos\Theta_{D}^{*}$, the angle between the $p_{\pi_{s}}$ and $p_{B}$
in the $D^{*}$ rest frame, and
$\cos\theta_{B}^{*}$, the angle between the $p_{\pi_{F}}$ and $P_{B}$
in the $B$ rest frame.
This method gives the most precise measurements of two exclusive branching
fractions:
$${\cal B}(\bar{B}^{0}\to D^{*+}\pi^{-})=(2.81\pm 0.11\pm 0.21\pm 0.05)\times 10%
^{-3}$$
$${\cal B}(B^{-}\to D^{*0}\pi^{-})=(4.81\pm 0.42\pm 0.40\pm 0.21)\times 10^{-3}.$$
The second systematic error is from the $D^{*}$ branching fractions.
A similiar partial reconstruction
analysis has been applied to the $B^{-}\to D^{**}(2420)^{0}\pi^{-}$
and $B^{-}\to D^{**}(2460)^{0}\pi^{-}$ decay modes.[33]
The event yields from fitting these distributions are substantial:
$281\pm 56$ $D^{**}(2420)$, $165\pm 61$ $D^{**}(2460)$, although
there are also large background subtractions. These correspond
to branching fractions,
$${\cal B}(B^{-}\to D_{1}(2420)\pi^{-})=(1.17\pm 0.24\pm 0.16\pm 0.03)\times 10^%
{-3}$$
$${\cal B}(B^{-}\to D_{2}^{*}(2460)\pi^{-})=(2.1\pm 0.8\pm 0.3\pm 0.05)\times 10%
^{-3}$$
The former mode was previously
observed using a similiar technique by ARGUS. The latter mode is
observed for the first time by CLEO. As noted by J. Gronberg and
H. Nelson, the partial reconstruction technique may also be useful for
observing a time dependent CP asymmetry in $\bar{B}^{0}\to D^{*+}\pi^{-}$.
2.1 The sign of the color suppressed amplitude and lifetimes
The sign
and magnitude of the color suppressed amplitude can be determined using
several classes of decay modes in charm and bottom mesons. The numerical
determination assumes factorization and uses form factors from various
phenemonological models.
For $D$ decay one
uses exclusive modes such as $D\to K\pi$, $D\to K\rho$ etc.,
and obtains
$$a_{1}=1.10\pm 0.03,~{}a_{2}=-0.50\pm 0.03$$
The destructive interference observed in two body $D^{+}$
decays leads to the $D^{+}$-$D^{0}$ lifetime difference.[1]
For $B$ decay, one can find the magnitude of $|a_{1}|$ from
the branching fractions for the decay modes
$\bar{B}^{0}\to D^{(*)+}\pi^{-}$, $\bar{B}^{0}\to D^{+(*)}\rho^{-}$.
This gives $|a_{1}|=1.06\pm 0.03\pm 0.06$.
One can also extract $|a_{1}|$
from measurements of branching fractions
$B\to D^{+,(0)}D_{s}^{(*)-}$[34].
The magnitude $|a_{2}|$ can be determined from the branching
fractions for $B\to\psi K^{(*)}$. This yields $|a_{2}|=0.23\pm 0.01\pm 0.01$.
The value of $a_{2}/a_{1}$ can be found by comparing
$B^{-}$ decays where both the external and spectator diagrams
contribute to $\bar{B}^{0}$ decays where only the external spectator
decays contribute. For example,
the model of Neubert et al. predicts the following ratios:
$$R_{1}={{\cal B}(B^{-}\to D^{0}\pi^{-})\over{\cal B}(\bar{B^{0}}\to D^{+}\pi^{-%
})}=(1+1.23a_{2}/a_{1})^{2}$$
(4)
$$R_{2}={{\cal B}(B^{-}\to D^{0}\rho^{-})\over{\cal B}(\bar{B^{0}}\to D^{+}\rho^%
{-})}=(1+0.66a_{2}/a_{1})^{2}$$
(5)
$$R_{3}={{\cal B}(B^{-}\to D^{*0}\pi^{-})\over{\cal B}(\bar{B^{0}}\to D^{*+}\pi^%
{-})}=(1+1.29a_{2}/a_{1})^{2}$$
(6)
$$R_{4}={{\cal B}(B^{-}\to D^{*0}\rho^{-})\over{\cal B}(\bar{B^{0}}\to D^{*+}%
\rho^{-})}\approx(1+0.75a_{2}/a_{1})^{2}$$
(7)
Improved measurements of these exclusive branching fractions
with better background subtraction
and additional data have recently been
presented by CLEO (see Fig. 2).[34]
Using the latest branching fractions,
$$a_{2}/a_{1}=0.21\pm 0.03\pm 0.03^{+0.13}_{-0.12},$$
where the third
error is a conservative estimte of the uncertainty ($\sim 20\%$)
in the relative production of $B^{+}$ and
$B^{0}$ mesons at the $\Upsilon(4S)$. There are a number of
additional theoretical uncertainties
which could significantly modify the magnitude of $a_{2}/a_{1}$ but
not its sign. For example,
the ratios of some heavy-to-heavy to heavy-to-light form factors is needed
(e.g. $B\to\pi/B\to D$).
Comparing the value of $a_{2}/a_{1}$ determined using
form factors from the model of Neubert et al. with the value obtained
using form factors from the
model of Deandrea et al. shows that this uncertainty is small.
The effect
of including the $B\to VV$ mode for which the form factors have somewhat
larger theoretical uncertainties is also small.
It is important to remember that
the determination of $a_{2}/a_{1}$ also assumes the factorization hypothesis.
The large error on the relative production of $B^{+}$ and $B^{0}$ mesons
is the most significant experimental
uncertainty in the determination of $a_{2}/a_{1}$.
The value of $a_{2}/a_{1}$ determined above
is consistent with the ratio $|a_{2}|$/$|a_{1}|$ where $|a_{2}|$ is computed
from $B\to\psi$ modes and $|a_{1}|$ is computed from $\bar{B}^{0}\to D^{(*)}\pi,D^{(*)}\rho$ modes. Although the result is surprisingly different
from what is observed in hadronic charm decay (where the interference
is destructive) and from what is expected in the $1/N_{c}$ expansion,
Buras claims that the result can be accomodated
in NLO QCD calculations.[35]
If the constructive interference which is observed in these
$B^{+}$ decays is present in all $B^{+}$ decays, then we expect
a significant $B^{+}$-$B^{0}$ lifetime difference ($\tau_{B}^{+}<\tau_{B^{0}}$),
of order $15-20\%$, in a direction opposite
to the $D^{+}-D^{0}$ lifetime difference. This
scenario is only marginally consistent
with experimental measurements of lifetimes;
the world average computed by the LEP lifetime working group
in August 1997 is
$$\tau_{B^{+}}/\tau_{B^{0}}=1.07\pm 0.04$$
It is possible
that the hadronic $B^{+}$ decays that have been observed to date are
atypical. The remaining higher multiplicity $B^{+}$ decays could
have destructive interference or no interference.[37]
Or perhaps there is a mechanism which also enhances the $\bar{B}^{0}$
width to compensate for the increase in the $B^{+}$ width
and which maintains the $B^{+}/B^{0}$ lifetime ratio near unity.
Such a mechanism would be relevant to the charm counting and
semileptonic branching fraction problem.
In either case, there will be experimental consequences in the
pattern of hadronic $B$ branching fractions.[38]
Experimentally one can compare other $B^{-}$ and $B^{0}$ decays including
$D^{**}\pi^{-}$ and $D^{**}\rho^{-}$ as well
decays to $D^{(*)}a_{1}^{-}$, $a_{1}^{-}\to\rho^{0}\pi^{-}$
and $D^{(*)}b_{1}^{-}$, $b_{1}^{-}\to\omega\pi^{-}$ to check the first
possibility.
3 Conclusion
The charm counting and semileptonic $B$ branching
fraction problem persists. Three possible
solutions are still experimentially viable.
These are (1) a systematic problem in the $D$
branching fraction, (2) an enhancement of $b\to c\bar{u}d$ or
(3) an enhancement of $b\to sg$. Any proposed solution must also satisfy
the experimental constraints on ${\cal B}(b\to c\bar{c}s)$ and
${\cal B}(b\to c\bar{u}d)$.
The sign of $a_{2}/a_{1}$ is found to be positive in the low multiplicity hadronic
$B$ decays that have been observed so far. This indicates
constructive interference in hadronic $B^{+}$ decay.
It will be interesting to see whether this pattern persists
as higher multiplicity $B$ decay modes are measured.
References
References
[1]
T.E. Browder, K. Honscheid, and D. Pedrini, UH-515-848-96,
OHSTPY-HEP-E-96-006, 1996 edition of
Annual Reviews of Nuclear and Particle Science.
[2]
T.E. Browder and K. Honscheid, Progress in Nuclear
and Particle Physics, Vol. 35, ed. K. Faessler, p. 81-220 (1995).
[3]
L. Gibbons et al. (CLEO Collaboration),
Phys. Rev. D 56, 3783 (1997).
[4]
D. Akerib et al. (CLEO Collaboration),
Phys. Rev. Lett. 71, 3070 (1993).
[5]
I.I. Bigi, B. Blok, M. Shifman, A. Vainshtein,
Phys. Lett. B 323, 408 (1994).
[6]
A. Falk, M. Wise, I. Dunietz,
Phys. Rev. D 51, 1183 (1995); Phys. Lett. B 73, 1075 (1995).
[7]
M. Buchalla, I. Dunietz, H. Yamamoto,
Phys. Lett. B 364, 188 (1995).
[8]
E. Bagan, P. Ball, V. Braun, P. Gosdzinsky,
Nucl. Phys. B 432, 3 (1994); Phys. Lett. B 342, 362 (1995) and Erratum;
Phys. Lett. B 374, 363 (1996).
[9]
W.F.Palmer and B. Stech, Phys. Rev. D 48, 4174 (1993).
[10]
I. Dunietz, J. Incandela, F.D. Snider,
and H. Yamamoto, hep-ph/9612421 Eur. Phys. J. C1, 211 (1998).
[11]
A. Lenz, U. Nierste, and G. Ostermaier,
hep-ph/9706501, Phys. Rev. D 56, 7228 (1997).
[12]
K. Honscheid, K.R. Schubert, and R. Waldi,
Z. Phys. C 63, 117 (1994).
[13]
M. Neubert and
C.T. Sachrajda,Nucl. Phys. B 483, 339 (1997).
Also see the recent review, M. Neubert, hep-ph/9801269 to appear
in the Proceedings of the 1997 Jerusalem Europhysics Conference.
[14]
G. Altarelli, G. Martinelli,
S. Petrarca, and F. Rapuano, Phys. Lett. B 382, 409 (1996).
[15]
I.L. Grach, I.M. Narodetskii, G. Simula, and K.A. Ter-Martirosyan,
hep-ph/9603239.
[16]
A. L. Kagan, Phys. Rev. D 51, 6196 (1995);
A. L. Kagan and J. Rathsman, hep-ph/9701300.
[17]
L. Roszkowski, M. Shifman, Phys. Rev. D 53, 404 (1996).
[18]
B. Grzadowski and W.S. Hou, Phys. Lett. B 272, 383 (1992).
[19]
I. Dunietz, FERMILAB-PUB-96/104-T, hep-ph/9606247.
[20]
R. Ammar et al. (CLEO Collaboration),
Phys. Rev. D 55, 13 (1997); R. Barish et al. (CLEO Collaboration),
Phys. Rev. Lett. 79, 3599 (1997).
[21]
X. Fu et al. (CLEO Collaboration), CLEO-CONF 95-11.
[22]
T.E. Coan et al. (CLEO Collaboration), CLNS-97-1516,
to appear in Phys. Rev. Lett.
[23]
ALEPH Collaboration, ICHEP96 PA05-060
[24]
CLEO Collaboration, CLEO CONF 97-26.
[25]
R. Barate et al. (ALEPH Collaboration),
Phys. Lett. B 405, 191 (1997).
[26]
M. Artuso et al. (CLEO Collaboration),
CLNS 97/1517, submitted to Phys. Rev. Lett.
[27]
D. Buskulic et al. (ALEPH Collaboration),
Phys. Lett. B 388, 648 (1996).
[28]
G. Alexander et al. (OPAL Collaboration),
Z. Phys. C 72, 1 (1996).
[29]
DELPHI Collaboration, ICHEP96 PA01-108,
DELPHI 96-97 CONF 26. DELPHI has also given
a preliminary result on the rate of double charm production
using inclusive vertexing, $r_{2C}=16.6\pm 6\%$,
with $D_{s}$ present in $84\pm 16\%$ of the events.
See EPS 448, contributed paper for the 1997 Jerusalem EPS conference.
[30]
B. Blok, M. Shifman, and N. Uraltsev,
Nucl. Phys. B 494, 237 (1997).
[31]
M. Bishai et al. (CLEO Collaboration),
Cornell preprint CLNS 97/1513.
[32]
G. Brandenburg et al. (CLEO Collaboration),
CLNS 97/1485, to appear in Phys. Rev. Lett.
[33]
J. Gronberg et al. (CLEO Collaboration),
CLEO CONF 96-25.
[34]
J. Rodriguez, hep-ex/9801028,
contribution to the Proceedings
IInd International Conference on B Physics and CP Violation,
Honolulu, HI 1997 (World Scientific).
[35]
A. J. Buras, Nucl. Phys. B 434, 606 (1995).
[36]
B. Stech, contribution to the Proceedings
IInd International Conference on B Physics and CP Violation,
Honolulu, HI 1997 (World Scientific).
[37]
M. Neubert, hep-ph/9707368 to appear
in the Proceedings of the Montpellier QCD97 Conference.
[38]
F. E. Close and H. J. Lipkin,
Phys. Lett. B 372, 306 (1996). |
A Survey of Binary Code Similarity
Irfan Ul Haq
IMDEA Software Institute &
Universidad Politécnica de Madrid
[email protected]
Juan Caballero
IMDEA Software Institute
[email protected]
Abstract
Binary code similarity approaches compare two or more pieces of binary code to identify their similarities and differences.
The ability to compare binary code enables many real-world applications
on scenarios where source code may not be available
such as patch analysis, bug search, and malware detection and analysis.
Over the past 20 years numerous binary code similarity approaches have been proposed,
but the research area has not yet been systematically analyzed.
This paper presents a first survey of binary code similarity.
It analyzes 61 binary code similarity approaches,
which are systematized on four aspects:
(1) the applications they enable,
(2) their approach characteristics,
(3) how the approaches are implemented, and
(4) the benchmarks and methodologies used to evaluate them.
In addition, the survey discusses the scope and origins of the area,
its evolution over the past two decades, and the challenges that lie ahead.
Code diffing, Code search, Cross-architecture, Program executables
I Introduction
Binary code similarity approaches compare two or more pieces of binary code e.g., basic blocks, functions, or whole programs,
to identify their similarities and differences.
Comparing binary code is fundamental in scenarios
where the program source code is not available, which happens with
commercial-of-the-shelf (COTS) programs, legacy programs, and malware.
Binary code similarity has a wide list of real-world applications such as
bug search [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15],
malware clustering [16, 17, 18],
malware detection [19, 20, 21],
malware lineage [22, 23, 24],
patch generation [25],
patch analysis [26, 27, 28, 29, 8, 30, 31],
porting information across program
versions [32, 26, 27], and
software theft detection [33].
Identifying binary code similarity is challenging because
much program semantics are lost due to the compilation process including
function names, variable names, source comments, and
data structure definitions.
Additionally, even when the program source code does not change,
the binary code may change if the source is recompiled,
due to secondary changes introduced by the compilation process.
For example, the resulting binary code
can significantly change when
using different compilers,
changing compiler optimizations,
and selecting different target operating systems and CPU architectures.
Furthermore, obfuscation transformations can be applied on
both the source code and the generated binary code,
hiding the original code.
Given its applications and challenges, over the past 20 years numerous binary code similarity approaches have been proposed.
However, as far as we know there does not exist a systematic survey of this
research area.
Previous surveys deal with
binary code obfuscation techniques in packer tools [34],
binary code type inference [35], and
dynamic malware analysis techniques [36].
Those topics are related because binary code similarity may need to tackle obfuscation,
binary code type inference may leverage similar binary analysis platforms,
and malware is often a target of binary code similarity approaches.
But, binary code similarity is well-separated from those topics
as shown by previous surveys having
no overlap with this paper on the set of papers analyzed.
Other surveys have explored similarity detection on any binary data,
i.e., not specific to code,
such as hashing for similarity search [37] and
similarity metrics on numerical and binary feature
vectors [38, 39].
In contrast, this survey focuses on approaches that compare
binary code, i.e., that disassemble the executable byte stream.
This paper presents a first survey of binary code similarity.
It first identifies 61 binary code similarity approaches through a
systematic selection process that examines over a hundred papers
published in research venues
from different computer science areas such as computer security,
software engineering, programming languages, and machine learning.
Then, it systematizes four aspects of those 61 approaches:
(1) the applications they enable,
(2) their approach characteristics,
(3) how the approaches have been implemented, and
(4) the benchmarks and methodologies used to evaluate them.
In addition, it discusses the scope and origin of binary code similarity,
its evolution over the past two decades, and the challenges that lie ahead.
Binary code similarity approaches widely vary in their approach,
implementation, and evaluation.
This survey systematizes each of those aspects, and
summarizes the results in easy to access
tables that compare the 61 approaches across multiple dimensions,
allowing beginners and experts
to quickly understand their similarities and differences.
For example, the approach systematization includes, among others,
the number of input pieces of binary code being compared
(e.g., one-to-one, one-to-many, many-to-many);
the granularity of the pieces of binary code analyzed
(e.g., basic blocks, functions, programs);
whether the comparison happens at the syntactical representation,
the graph structure, or the code semantics;
the type of analysis used
(e.g., static, dynamic, symbolic), and
the techniques used for scalability
(e.g., hashing, embedding, indexing).
The implementation systematization includes
the binary analysis platforms used to build the approach,
the programming languages used to code it,
the supported architectures for the input pieces of binary code being compared,
and whether the approach is publicly released.
The evaluation systematization covers the datasets on which the approaches
are evaluated and the evaluation methodology including
the type of evaluation (e.g., accuracy, comparison wih prior works, performance) and how the robustness of the approach is evaluated in face of
common code transformations such as compiler and compilation option changes,
different architectures, and obfuscation.
Beyond the systematization,
this survey also discusses how binary code similarity has evolved from binary code diffing to
binary code search and how the focus has moved from a single architecture to
cross-architecture approaches.
It shows that the present of the field is vibrant as many new approaches are
still being proposed.
It discusses technical challenges that remain open, but
concludes that the future of the area is bright with important applications
scenarios under way such as those related to binary code search engines and the
Internet-of-Things.
Paper structure.
This paper is organized as follows.
Section II provides an overview of binary code similarity.
Section III details the scope of the survey and
the paper selection process.
Section IV summarizes applications of binary code similarity and
Section §V the evolution of the field over
the last two decades.
Section VI systematizes the characteristics of the
61 binary code similarity approaches,
Section VII their implementation, and
Section VIII their evaluation.
Finally, we discuss future research directions in Section IX,
and conclude in Section X.
II Overview
In this section, we first provide background on
the compilation process (§II-A).
Then, we present an overview of the binary code similarity problem (§II-B).
II-A Compilation Process
Binary code refers to the machine code that is produced by the
compilation process and that can be run directly by a CPU.
The standard compilation process takes as input the
source code files of a program.
It compiles them with a chosen compiler and optimization level and
for a specific platform (defined by the architecture, word size, and OS)
producing object files.
Those object files are then linked into a binary program,
either a stand-alone executable or a library.
Binary code similarity approaches typically deal with an
extended compilation process, illustrated in Figure 1,
which adds two optional steps to the standard compilation process:
source code and binary code transformations.
Both types of transformations are typically semantics-preserving
(i.e., do not change the program functionality) and are
most commonly used for obfuscation,
i.e., to hamper reverse-engineering of the distributed binary programs.
Source code transformations happen pre-compilation.
Thus, their input and output are both source code.
They can be applied regardless of the target platform, but may be
specific to the programming language used to write the program.
On the other hand, binary code transformations happen post-compilation.
Thus, their input and output are binary code.
They are independent of the programming language used,
but may be specific to the target platform.
Obfuscation is a fundamental step in malware, but can also be applied
to benign programs, e.g., to protect their intellectual property.
There exist off-the-shelf obfuscation tools that use
source code transformations (e.g., Tigress [40]),
as well as binary code transformations (e.g., packers [41]).
Packing is a binary code transformation widely used by malware.
Once a new version of a malware family is ready, the malware authors
pack the resulting executable to hide its functionality and
thus bypass detection by commercial malware detectors. The packing process
takes as input an executable and produces another executable with the
same functionality, but with the original code hidden
(e.g., encrypted as data and unpacked at runtime).
The packing process is typically applied many times
to the same input executable, creating polymorphic variants
of exactly the same source code,
which look different to malware detectors.
Nowadays, the majority of malware is packed
and malware often uses custom packers for which off-the-shelf unpackers
are not available [41].
A main challenge in binary code similarity is that the compilation process
can produce different binary code representations for the same source code.
An author can modify any of the grey boxes in
Figure 1 to produce a different,
but semantically-equivalent, binary program
from the same source code.
Some of these modifications may be due to the standard compilation process.
For example, to improve program efficiency an author may vary the
compiler’s optimization level, or change the compiler altogether.
Both changes will transform the produced binary code,
despite the source code remaining unchanged.
An author may also change the target platform to obtain a version of the
program suitable for a different architecture.
In this case, the produced binary code may radically differ if the
new target architecture uses a different instruction set.
An author may also deliberately apply obfuscation transformations to
produce polymorphic variants of the same source code.
The produced variants will typically have the same functionality defined by
the original source code.
A desirable goal for binary code similarity approaches is that they are able to identify
the similarity of binary code that corresponds to the same source code
having undergone different transformations.
The robustness of a binary code similarity approach captures
the compilation and obfuscation transformations that it can handle,
i.e., the transformations despite which it can still detect similarity.
II-B Binary Code Similarity Overview
Binary code similarity approaches compare pieces of binary code.
The three main characteristics of binary code similarity approaches are:
(1) the type of the comparison
(identical, similar, equivalent),
(2) the granularity of the pieces of binary code being compared
(e.g., instructions, basic blocks, functions), and
(3) the number of input pieces being compared
(one-to-one, one-to-many, many-to-many).
We detail these three characteristics next.
For simplicity, we describe the comparison type and comparison granularity
for two inputs and then generalize to multiple inputs.
Comparison type.
Two (or more) pieces of binary code are identical if they have
the same syntax, i.e., the same representation.
The binary code can be represented in different ways such as
an hexadecimal string of raw bytes,
a sequence of disassembled instructions, or
a control-flow graph.
Determining if several pieces of binary code are identical is
a Boolean decision (either they are identical or not)
that it is easy to check:
simply apply a cryptographic hash
(e.g., SHA256)
to the contents of each piece.
If the hash is the same, the pieces are identical.
However, such straightforward approach fails to detect similarity
in many cases.
For example, compiling the same program source code twice,
using the same compilation parameters
(i.e., same compiler version, same optimization level,
same target platform)
produces two executables with different file hash.
This happens because the executable may include metadata
that differs in both compilations such as the
compilation date, which is automatically computed and included into the
header of the generated executable.
Two pieces of binary code are equivalent if they have the
same semantics,
i.e., if they offer exactly the same functionality.
Equivalence does not care about the syntax of the binary code.
Clearly, two identical pieces of binary code will have the same semantics,
but different pieces of binary code may as well.
For example, mov %eax,$0 and xor %eax,%eax
are semantically equivalent x86 instructions
because both set the value of register EAX to zero.
Similarly, the same source code compiled for two different target
architectures should produce equivalent executables, whose syntax
may be completely different if the architectures use different instruction sets.
Proving that two arbitrary programs are functionally equivalent is an
undecidable problem that reduces to solving the halting problem [42].
In practice, determining binary code equivalence is a very
expensive process that can only be performed for small pieces of binary code.
Two pieces of binary code can be considered similar if their
syntax, structure, or semantics are similar.
Syntactic similarity compares the code representation.
For example, clone detection approaches consider that a target piece of binary code is a clone of some source binary code if their syntax are similar.
Structural similarity compares graph representations of binary code
(e.g., control flow graphs, call-graphs).
It sits between syntactic and semantic similarity.
The intuition is that the control flow of the binary code captures
to some extent its semantics, e.g., the decisions taken on the data.
Furthermore, the graph representation captures
multiple syntactic representations of the same functionality.
However, it is possible to modify the graph structure without affecting
the semantics, e.g., by inlining functions.
Semantic similarity compares the code functionality.
A simple approach to semantic similarity compares
the interaction between the program and its environment through OS APIs or
system calls.
But, two programs with similar system calls can perform significantly
different processing on their output,
so more fine-grained semantic similarity approaches focus on a
syntax-independent comparison of the code.
Generally speaking, the more robust an approach is,
i.e., the more transformations it can capture,
the more expensive it also is.
Syntactic similarity approaches are cheapest to compute,
but least robust.
They are sensitive to simple changes in the binary code,
e.g., register reallocation, instruction reordering,
replacing instructions with semantically equivalent ones.
Structural similarity sits in the middle.
It is robust against multiple syntactical transformations,
but sensitive to transformations that change code structure such as
code inlining or removal of unused function parameters.
Semantic similarity is robust against semantics-preserving transformations,
despite changes to the code syntax and structure,
but it is very expensive to compute for large pieces of binary code.
Comparison granularity.
Binary code similarity approaches can be applied at different granularities.
Common granularities are
instructions;
basic blocks;
functions; and
whole programs.
To perform a comparison at a coarser granularity,
some approaches use a different comparison at a finer granularity,
and then combine the finer granularity results.
For example, to compare whether two programs are similar,
an approach could determine the fraction of identical functions
between both programs.
Thus, we differentiate between the input granularity,
i.e., the granularity of the input pieces of binary code that the approach compares,
and the approach granularities,
i.e., the granularities of the different comparisons in the approach.
Applying a specific comparison at a finer granularity may restrict the type of
comparison that can be performed at a coarser granularity,
as illustrated in Figure 2.
The figure shows that computing whether two pieces of binary code are
identical at a finer granularity (e.g., basic block) can be used to
compute that the coarser granularity pieces that encompass them
(e.g., their functions) are identical, equivalent, or similar.
However, similarity at a finer granularity cannot be used to
infer that the coarser granularity code is equivalent or identical.
For example, when comparing two functions,
just because all their basic blocks are similar,
it cannot be concluded that the functions are identical or equivalent.
On the other hand, similarity is the most general type of comparison
and any finer granularity comparison type can be used to infer it.
Number of inputs.
Binary code similarity approaches can compare two or more pieces of binary code.
Those that compare more than two pieces can further be split into comparing one piece to the rest or
comparing each piece to all other pieces.
Thus, we identify three types of approaches based on the number of inputs and
how they are compared:
one-to-one ($OO$), one-to-many ($OM$), and many-to-many ($MM$).
The source of the input pieces is application-dependent.
They may come
from the same program version
(e.g., two functions of the same executable),
from two versions of the same program,
and from two different programs.
One-to-one approaches compare an original piece of binary code (also called source, old, plaintiff, or reference) to a target piece of binary code (also called new, patched, or upgrade).
Most $OO$ approaches perform binary code diffing,
i.e., they diff two consecutive, or close, versions of the
same program to identify what was added, removed, or modified in the
target (patched) version.
The granularity of binary code diffing is most often functions and the
diffing tries to obtain a mapping between a function in the original
program version and another function in the target program version.
Added functions are original functions that cannot be mapped to a
target function;
removed functions are target functions that cannot be mapped to an
original function; and
modified functions are mapped functions that are not identical.
One-to-many approaches compare a query piece of binary code to many target pieces of binary code.
Most $OM$ approaches perform binary code search,
i.e., they search if the query piece is
similar to any of the target pieces and
return the top $k$ most similar target pieces of binary code.
The target pieces may come from multiple versions of the same program
(different than the version the query piece comes from),
from different programs compiled for the same architecture, or
from programs compiled for different architectures.
In contrast to $OO$ and $OM$ approaches,
many-to-many approaches do not distinguish between source and target pieces.
All input pieces are considered equal and compared against each other.
These approaches typically perform binary code clustering,
i.e., they output groups of similar pieces of binary code called clusters.
III Scope & Paper Selection
To keep our survey of the state-of-the-art focused and manageable
it is important to define what is, and what is not, within scope.
Overall, the main restriction is that we focus on works that compare
binary code.
This restriction, in turn, introduces the following four constraints:
1.
We exclude approaches that require access to the source code,
namely source-to-source (e.g., [43]) and
source-to-binary (e.g., [44]) similarity approaches.
2.
We exclude approaches that operate on bytecode
(e.g., [45, 46]).
3.
We exclude behavioral approaches that compare similarity exclusively
on the interaction of a program with its environment through
system calls or OS API calls
(e.g., [47, 48, 49, 50]).
4.
We exclude approaches that consider binary code as a sequence of
raw bytes with no structure such as
file hashes (e.g., [51]),
fuzzy hashes (e.g., [52, 53]), and
signature-based approaches (e.g., [54, 55]).
Approaches need to disassemble raw bytes into instructions to be considered.
While we do not include the papers describing byte-level approaches,
we do examine the use of some of those techniques (e.g., fuzzy hashing)
by the analyzed approaches.
In addition, we introduce the following constraints to keep the
scope of the survey manageable:
5.
We limit the survey to papers published on peer-reviewed venues
and technical reports from academic institutions.
Thus, we do not analyze tools,
but rather the research works describing their approach
(e.g., [26, 27] for BinDiff).
6.
We exclude papers that do not propose a new binary code similarity approach or technique,
but simply apply off-the-shelf binary code similarity tools as a step
towards their goal.
Paper selection.
To identify candidate papers, we first systematically
examined all papers published in the last 20 years in 14 top venues
for computer security and software engineering:
IEEE S&P, ACM CCS, USENIX Security, NDSS, ACSAC, RAID, ESORICS, ASIACCS, DIMVA, ICSE, FSE, ISSTA, ASE, and MSR. Not all relevant binary code similarity approaches have been published in those venues,
which is especially true for early approaches.
To identify candidate papers in other venues,
we extensively queried specialized search engines such as Google Scholar
using terms related to binary code similarity and its applications,
e.g., code search, binary diffing, bug search.
We also carefully examined the references of the candidate papers
for any further papers we may have missed.
This exploration identified over a hundred candidate papers.
We then read each candidate paper to determine if they
proposed a binary code similarity approach that satisfied the above scope constraints.
In the end, we identified the 61 binary code similarity research works in Table I,
whose approaches are systematized.
The first three columns of Table I capture
the name of the approach,
the year of publication, and
the venue where the work was published.
The research works are sorted by publication date
creating a timeline of the development of the field.
Papers are identified by their system name, if available, otherwise
by the initials of each author’s last name and the year of publication.
The 61 papers have been published in 37 venues.
Binary code similarity is quite multidisciplinary;
while most papers appear in computer security venues (36 papers in 20 venues),
there are works in software engineering (13 papers in 8 venues),
systems (6 papers in 4 venues), and
machine learning (2 papers in 2 venues).
The venues with most binary code similarity papers are:
DIMVA (6), ASE (4), CCS (3), USENIX Security (3), and PLDI (3).
IV Applications
This section motivates the importance of binary code similarity by
describing the applications it enables.
Of the 61 papers analyzed, 36 demonstrate an application,
i.e., present a quantitative evaluation, or case studies,
of at least one application.
The other 23 papers present generic binary code similarity capabilities
that can be used for multiple applications such as
binary diffing tools (e.g., [84, 85, 86]),
binary code search platforms (e.g., [72, 62, 71]), and
binary clone detection approaches
(e.g., [74, 73, 56, 65]).
Table II summarizes the eight applications identified.
Most of the 36 papers demonstrate a single application, although a few
(e.g., F2004, BinSequence)
demonstrate multiple.
One property of an application is whether the application compares
different versions of the same program
(patch analysis, patch generation, porting information, malware lineage),
different programs (malware clustering, malware detection,
software theft detection),
or can be applied to both cases (bug search).
Next, we detail those applications.
1.
Bug search –
Arguably the most popular application of binary code similarity is finding a known bug in a large repository of target
pieces of binary code [2, 3, 4, 6, 11, 12, 9, 8, 7, 15, 5, 10, 13, 14, 1].
Due to code reuse, the same code may appear in multiple programs,
or even in multiple parts of the same program.
Thus, when a bug is found, it is important to identify similar code that
may have reused the buggy code and contain the same, or a similar, bug.
Bug search approaches take as input a query buggy piece of binary code and search for
similar pieces of binary code in a repository.
A variant of this problem is cross-platform bug search,
where the target pieces of binary code in the repository may be compiled for
different platforms (e.g., x86, ARM, MIPS) [3, 4, 6, 12, 9, 7, 15, 10, 13].
2.
Malware detection –
Binary code similarity can be used to detect malware by comparing a given
executable to a set of previously known malware samples.
If similarity is high then the input sample is likely a variant of a previously known malware family.
Many malware detection approaches are purely behavioral,
comparing system or API call behaviors
(e.g., [47, 48]).
However, as described in Section III,
we focus on approaches that use
binary code similarity [19, 20, 21].
3.
Malware clustering –
An evolution of malware detection is
clustering similar, known malicious, executables into families.
Each family cluster should contain executables
from the same malicious program,
which can be different versions of the malicious program,
as well as polymorphic (e.g., packed) variants of a version.
Similar to malware detection,
we focus on approaches that compare binary code [16]
and exclude purely behavioral approaches
based on system calls and network traffic
(e.g., [50, 87, 88]).
4.
Malware lineage –
Given a set of executables known to belong to the same program,
lineage approaches build a graph where nodes are program versions and edges
capture the evolution of the program across versions.
Lineage approaches are most useful with malware because version information is
typically not available [22, 23, 24, 89].
Since input samples should belong to the same family,
malware lineage often builds on the results of malware clustering.
5.
Patch generation and analysis –
The earliest binary code similarity application, and one of the most popular,
is to diff two consecutive, or close, versions of the
same program to identify what was patched in the newer version.
This is most useful with proprietary programs where the vendor
does not disclose patch details.
The diffing produces small binary code patches that can be efficiently
shipped to update the program [25].
It can also be used to automatically identify security patches that
fix vulnerabilities [30],
analyze those security patches [26, 27, 28, 29, 31, 8], and
generate an exploit for the old vulnerable version [90].
6.
Porting information –
Binary code similarity can be used for porting information between two
close versions of the same program.
Since close versions typically share a large amount of code,
the analysis done for one version may be largely reusable for a newer version.
For example, early binary code similarity approaches ported
profiling data [32, 91] and analysis results obtained during
malware reverse engineering [26, 27].
7.
Software theft detection –
Binary code similarity can be used for identifying unauthorized reuse of code from
a plantiff’s program such as
the source code being stolen, its binary code reused,
a patented algorithm being reimplemented without license,
or the license being violated (e.g., GPL code in a commercial application).
Early approaches for detecting such infringements used software birthmarks,
i.e., signatures that capture inherent functionality of the plaintiff
program [92, 93].
However, as described in Section III,
we exclude signature-based approaches and focus on approaches using
binary code similarity [33].
V Binary Code Similarity Evolution
This section describes the origins of binary code similarity and its evolution over
the last 20 years, highlighting some noteworthy approaches.
The origins.
The origins of binary code similarity are in the problem of generating a patch (or delta)
that captures the differences between
two consecutive (or close) versions of the same program.
Text diffing tools had existed since the 1970’s
(the popular UNIX diff was released in 1974) and
had been integrated in early source code versioning systems such as
SCCS [94] (1975) and RCS [95] (1985).
The increasing popularity of low bandwidth communication networks
(e.g., wireless)
and the limited resources in some devices,
raised interest in techniques that would increase efficiency by transmitting
a small patch that captured the differences
between two versions of a binary, i.e., non-text, file,
instead of transmitting the whole file.
In 1991, Reichenberger proposed a diffing technique
for generating patches between arbitrary binary files
without any knowledge about the file structure [96].
The approach worked at the byte-level,
instead of the line-level granularity of text diffing,
and efficiently identified the
byte sequences that should appear in the patch because they only appeared
in the updated version and thus were not available in the original file
to be patched.
Several tools for generating and applying binary patches soon appeared
such as RTPatch [97],
BDiff95 [98], and
xDelta [99].
Those tools worked at byte-level and could diff any type of file.
The first binary code similarity approaches are from 1999.
That year, Baker et al. proposed an approach for compressing differences
of executable code and built a prototype diffing tool
called Exediff [25],
which generated patches for DEC Alpha executables.
Their intuition was that many of the changes when diffing two executable
versions of the same program represent secondary changes due to
the compilation process, as opposed to direct changes in the source code.
One example they mentioned is register allocation
that may change at recompilation.
Another example they mentioned is that code added in the newer version
would displace parts of the old code,
and thus the compilation process would have to adjust
pointer values in the displaced code to point to the correct addresses.
Their idea was to reconstruct secondary changes at patch time,
so that they would not need to be included in the patch,
reducing the patch size.
Exediff is the earliest approach we have identified
that focused on computing similarity between binary code,
taking advantage of the code structure by disassembling the raw bytes
into instructions.
Also in 1999, Wang et al. presented BMAT [32], a tool that aligned two versions of a Windows DLL library executable
to propagate profile information from the older (extensively profiled) version
to a newer version, thus reducing the need for re-profiling.
Their approach is the first to compare functions and basic blocks
(Exediff compared two sequences of instructions).
It first matched functions in the two executables and then
matched similar blocks within each matched function.
It used a hashing technique to compare blocks.
The hashing removed relocation information to handle pointer changes,
but was order-sensitive.
The first decade.
After the initial works of Exediff and BMAT,
we only identify 7 binary code similarity approaches in the next decade (2000-2009).
However, some of these are highly influential as
they extend binary code similarity from purely syntactical to also include semantics;
they widen the scope from binary code diffing (OO) to also include
binary code clustering (MM) and binary code search (OM);
and they apply binary code similarity to malware.
In 2004, Thomas Dullien (alias Halvar Flake) proposed a
graph-based binary code diffing approach that focused on the structural properties
of the code by heuristically constructing a callgraph isomorphism
that aligns functions in two versions of the same
binary program [26].
This is the first approach to handle instruction reordering inside a
function introduced by some compiler optimizations.
A followup work [27]
extended the approach to also match basic blocks inside matched functions
(as done in BMAT)
and introduced the Small Primes Product (SPP) hash
to identify similar basic blocks despite instruction reordering.
These two works are the basis for the
popular BinDiff binary code diffing plugin for the Ida disassembler [84].
In 2005, Kruegel et al. proposed a graph coloring technique to detect
polymorphic variants of a malware.
This is the first approach that performed semantic similarity
and MM comparison.
They categorized instructions with similar functionality into 14
semantic classes.
Then, they colored the inter-procedural control-flow graph (ICFG)
using those classes.
The graph coloring is robust against syntactical obfuscations such as
junk insertion, instruction reordering, and instruction replacement.
In 2008, Gao et al. proposed BinHunt [28]
to identify semantic differences between two versions of the same program.
This is the first approach that checks code equivalence.
It uses symbolic execution and a constraint solver to check if two basic blocks
provide the same functionality.
In 2009, Xin et al. presented Smit [16],
an approach that given a malware sample finds similar malware in
a repository. Smit is the first OM and binary code search approach.
It indexes malware callgraphs in a database and uses
graph edit distance to find malware with similar callgraphs.
The last decade.
The last decade (2010-2019) has seen a huge increase in the popularity of
binary code similarity, with 52 approaches identified.
The focus on this decade has been on binary code search approaches,
with an emphasis since 2015 on its cross-architecture version
(16 approaches),
and in recent years on machine learning approaches. In 2013, Wei et al. proposed Rendezvous,
a binary code search engine that given the binary code of a query function,
finds other functions in a repository with similar syntax and
structural properties.
Reducing the search granularity from whole programs (Smit) to
smaller pieces of binary code such as functions
enables an array of applications such as clone detection and bug search.
Most binary code search approaches target the bug search application.
This application was first addressed on source code in 2012 by
Jang et al. [100].
In 2014, David et al. proposed Tracy [1],
the first binary code search approach focused on bug search.
Tracy used the concept of tracelets,
an execution path in a CFG,
to find functions similar to a vulnerable function.
In 2015, Pewny et al. presented Multi-MH [3],
the first cross-architecture binary code search approach.
Multi-MH indexed functions by their input-output semantics.
Given a function compiled for one CPU architecture (e.g., x86)
Multi-MH can find similar functions compiled for other
architectures (e.g., MIPS).
This problem quickly gained traction due to the popularity of embedded devices.
In 2016, Lageman et al. [72] trained a neural network to decide
if two functions were compiled from the same source code.
The use of deep learning has picked up in the last two years,
e.g., $\alpha$Diff (2018), InnerEye (2019), and Asm2Vec (2019).
VI Approaches
In this section, we systematize the approaches of binary code similarity,
describing the Approach Characteristics columns in
Table I.
We recommend the reader to print Table I in
a separate page to have it in hand while reading this section.
VI-A Comparison Type
Columns: Input Comparison; Approach Comparison
This section discusses the type of comparison between the approach inputs,
as well as the finer granularity comparisons that may be used by the approach.
Input comparison.
All 61 works analyzed compare their inputs to identify similarity.
That is, no approach identifies identical inputs
(since a hash suffices for that) or input equivalence
since it is an undecidable problem [101]
that can only be solved efficiently for small pieces of binary code.
Thus, we classify binary code similarity approaches based on their input
comparison as:
one-to-one (OO, 21 approaches), one-to-many (OM, 30 approaches), and
many-to-many (MM, 10 approaches).
The dominance of $OM$ approaches is due to the high interest in binary code search in the last decade.
It is always possible to build an $OM$ or $MM$
approach from an $OO$ approach.
For example, a simple implementation of an $OM$ approach
is to compare the given query piece of binary code with each of the $n$ targets
using an $OO$ approach that returns the similarity between both inputs.
Then, simply rank the $n$ targets by decreasing similarity
and return the top $k$ entries or the entries above a similarity threshold.
However, most $OM$ approaches avoid this simple implementation
since it is inefficient.
The two main solutions to improve performance are
extracting a feature vector from each input and
storing the target pieces of binary code in a repository with indices.
Obtaining a feature vector for each input
allows to perform the feature extraction only once per input.
This offers significant benefits when the feature extraction is expensive,
e.g., in the case of Blex whose feature extraction requires executing
a piece of binary code multiple times with different inputs.
Once the feature vectors have been extracted,
a similarity metric between two feature vectors is used.
This similarity metric is typically cheap to compute as feature vectors
are often numerical or Boolean.
A common source of confusion is that some approaches propose similarity
metrics, while others propose distance metrics.
It is important to keep in mind that when the metrics are normalized
between zero and one, the distance is simply one minus the similarity.
The other solution used by $OM$ approaches is adding indices
on a subset of the features in the feature vector.
The indices can be used to reduce the number of comparisons by
applying the similarity metric only between the feature vector of
the input piece of binary code and the feature vectors of selected targets
more likely to be similar.
Approach comparison.
Most approaches use a single type of comparison:
similarity (42 approaches), equivalence (5), and identical (2).
Note that even if only one type of comparison is used in the approach,
it may differ from the input comparison.
For example, Exediff looks for identical instructions in the process of
diffing two programs.
There are 12 approaches that use multiple comparison types at
different granularities.
Of those, six use identical comparison at finer granularities
to quickly identify the same pieces of binary code (BMAT, DR2005, SWPQS2006, BinClone)
or to reduce expensive comparisons such as graph isomorphism (Smit, Spain).
The other six use equivalence comparisons at finer granularities
to capture semantic similarity.
VI-B Granularity
Columns: Input Granularity; Approach Granularities
We separate the input granularity from
the granularities of the pieces of binary code compared in the approach
(i.e., approach granularities)
since it is common to use finer granularities (e.g., functions) to
compare coarser input granularity (e.g., whole programs).
We have identified 8 comparison granularities:
instruction (I),
set of related instructions (I*),
basic block (B),
set of related basic blocks (B*),
function (F),
set of related functions (F*),
trace (T), and
whole program (P).
Instruction, basic block, function, and whole program are
standard granularities that require no explanation.
Related instructions (I*) are either consecutive (e.g., n-gram) or
share a property (e.g., data dependent).
They may belong to different basic blocks,
and even to different functions.
For example, Tracy groups instructions that collectively
impact an output variable.
Related basic blocks (B*)
share structural properties (e.g., graphlets in a CFG)
or belong to the same execution path.
Basic blocks in a set may belong to the same or multiple functions.
Related functions (F*) implement a program component such as
a library, a class, or a module.
Trace granularity compares the execution trace of two binary programs
on the same input.
The most common input granularity is
function (26 approaches) followed by whole program (25) and
related basic blocks (4).
Whole program is the preferred input granularity for $OO$ approaches
(16/21 )
since most binary code diffing approaches try to establish a one-to-one mapping
between all functions in the input programs,
and also for $MM$ (7/10 ) approaches that tend to cluster input programs.
On the other hand, function is the preferred granularity
for binary code search approaches (21/30 ).
Another four binary code search approaches use B* to
identify code reuse that covers only a subset of a function or
crosses function boundaries.
The most common approach granularity is function (30 approaches)
followed by basic block (20).
The majority of approaches (47/61) use different input and
approach granularities,
i.e., use finer approach granularities to compare coarser input granularity.
Most approaches with the same input and approach granularity perform
function searches (12/14).
The 11 approaches that perform equivalence comparisons
do so at fine granularities due to its low efficiency:
six have B granularity,
five I*, and one F. Some approaches accumulate features at a fine granularity that
are never directly compared and thus do not show in the
approach granularities column.
For instance, Gemini accumulates basic block features
to generate a numerical vector at function granularity.
Thus, only functions are compared.
VI-C Syntactic Similarity
Column: Syntactic similarity
Syntactic approaches capture similarity of the code representation,
more especifically they compare sequences of instructions.
Most commonly, the instructions in a sequence are consecutive
in the virtual address space and belong to the same function.
The instructions in the sequence may first be normalized,
e.g., considering only the mnemonic, only the opcode, or
normalizing the operands into classes.
We detail instruction normalization in Section VI-J
and simply refer to instruction sequences in the rest of this subsection.
The instruction sequences may be of fixed or variable length.
Fixed-size sequences are obtained by sliding a window
over the instruction stream,
e.g., over the linearly-ordered instructions in a function.
This process is characterized by the window size,
i.e., the number of instructions in the sequence,
and the stride, i.e., the number of instructions to slide the
start of the window to produce the next sequence.
When the stride is smaller than the window size,
consecutive sequences overlap.
When the stride is one, the resulting sequence is called an n-gram.
For example, given the sequence of instruction mnemonics
$\{mov,push,add\}$ two 2-grams will be extracted:
$\{mov,push\}$ and $\{push,add\}$.
There are 7 works that use n-grams:
Idea, MBC, Rendezvous, MutantX-S, Exposé, iLine, and Kam1n0.
Fixed-size sequences are also used by SWPQS2006 with a configurable stride larger than one.
Rendezvous, in addition to n-grams, also uses n-perms,
unordered n-grams that capture instruction reordering within the sequence.
An n-perm may capture multiple n-grams,
e.g., 2-perm $\{mov,push\}$ captures 2-grams $\{mov,push\}$ and $\{push,mov\}$.
The most common methods to compare instruction sequences are
hashing, embedding, and alignment.
Hashing is used by 6 approaches
(BMAT, DR2005, SWPQS2006, Smit, BinClone, Spain)
to obtain a fixed-length value out of a variable-length instruction sequence.
If the hash values are the same, the sequences are similar.
Five approaches generate an embedding from n-gram sequences
(Idea, MBC, MutantX-S, Exposé, Kam1n0).
Three approaches (Exediff, Tracy, BinSequence) align two sequences to
produce a mapping between them by inserting gaps in either sequence
to account for inserted, removed, and modified instructions.
These approaches define a similarity score when instructions are aligned,
and a gap score when an instruction aligns with a gap.
Other less common comparison methods are using
vectors of Boolean features (iLine) and encoding sequences as strings for indexing (Rendezvous).
VI-D Semantic Similarity
Column: Semantic similarity
Semantic similarity captures if the code being compared has similar
effects, as opposed to syntactic similarity that captures
similarity in the code representation.
The semantics of a piece of binary code can be described by the changes it produces in the
process state, i.e., updates to the content of registers and memory.
We identify 26 approaches computing semantic similarity.
Most of them capture semantics at basic block granularity
because basic blocks are straight-line code without control flow.
Three methods are used to capture semantics:
instruction classification,
input-output pairs, and
symbolic formulas.
Instruction classification.
The first approach to introduce semantics into binary code similarity was KKMRV2005,
which classified instructions into 14 classes
(e.g., arithmetic, logic, data transfer) and used a 14-bit value to
capture the classes of the instructions in a basic block.
This semantic $color$ bitvector captures the effects of the basic block.
This approach was later adopted by
BMM2006, Beagle, Fossil, and Sigma.
Instruction classification can be used to compute semantic similarity,
but cannot determine if two pieces of binary code are, or are not, equivalent.
Input-output pairs.
Intuitively, two pieces of binary code are functionally equivalent if given the same input
they produce the same output, for all possible inputs.
Such equivalence is independent of the code representation and
compares the final state after the code is executed,
ignoring intermediate program states.
This approach was proposed by Jiang et al.
on source code [102] and later used by numerous binary code similarity approaches:
BinHash, MXW2015, Blex, Multi-MH, BinGo, CACompare, Spain, KS2017 and IMF-sim.
It involves executing both pieces of binary code with the same input and
comparing their output, repeating the process many times.
If the output differs for any input, then the two pieces of binary code are not equivalent.
Unfortunately, to determine that both pieces are equivalent, the approach
would require testing all possible inputs,
which is not realistic for any non-trivial piece of binary code.
Thus, in practice, this approach can only determine
that two pieces of binary code are likely equivalent, with a confidence proportional
to the fraction of inputs that have been tested, or
that they are not equivalent (with certainty).
The tested inputs are typically selected randomly,
although it is possible to use other selection rules,
e.g., taking values from the program data section (CACompare).
It is generally a dynamic approach,
but some approaches (e.g., Multi-MH) evaluate concrete inputs on
statically-extracted symbolic formulas.
Symbolic formula.
A symbolic formula is an assignment statement in which the
left side is an output variable and the
right side is a logical expression of input variables and literals
that captures how to derive the output variable.
For instance, the instruction add %eax,%ebx
can be represented with the symbolic formula
EBX2 = EAX + EBX1
where EBX2 and EBX1 are symbols representing the values of
the EBX register before and after executing the instruction.
Eleven approaches use symbolic formulas:
BinHunt, iBinHunt, BinHash, Exposé, Tracy, RMKNHLLP2014, Tedem,
CoP, Multi-MH, Esh, and Xmatch.
Of those, eight approaches extract symbolic formulas at basic block granularity,
Xmatch and Exposé extract formulas for the return values of a function, and
BinSim extracts symbolic formulas from an execution trace that capture how
the arguments of a system call were derived.
Three methods are used to compare symbolic formulas:
using a theorem prover to check for equivalence,
comparing the semantic hash of the formulas to check for equivalence,
and computing the similarity of the graph representation of the formulas.
Theorem prover –
BinHunt introduced the idea of using
theorem provers such as STP [103] or Z3 [104]
to check if two symbolic formulas are equivalent,
i.e., whether the output variable always contains the same value
after the execution of both formulas,
assuming that the input variables share the same values.
The main limitation of this approach
is that it is computationally expensive because
it can only perform pairwise equivalence queries,
the solving time quickly increases with formula sizes, and
the solver may fail to return an answer for some queries.
Note that a piece of binary code may have multiple outputs
(registers and variables in memory),
each represented by its own symbolic formula.
These approaches need to try all pair-wise comparisons
and check if there exists a permutation of variables
such that all matched variables contain the same value.
Semantic hashes –
An alternative to using a theorem prover is to check if two symbolic formulas
have the same hash, after normalizing the formulas
(e.g., using common register names) and
simplifying them (e.g., applying constant propagation).
The intuition is that if the two symbolic formulas have the same hash
they should be equivalent.
Three approaches use semantic hashes: BinHash, BinJuice, and GitZ.
Semantic hashes are efficient, but are limited in that
it is possible for two equivalent formulas to have different hashes
even after normalization and simplification.
For example, reordering of symbolic terms in one of the formulas
(e.g., due to instruction reordering) results in a different hash.
Graph distance –
Xmatch and Tedem represent the symbolic formula of a basic block as a tree,
and compute their similarity by applying graph/tree edit distance.
Computing the graph/tree edit distance is more expensive than
comparing semantic hashes, but the graph representation has the advantage
over semantic hashes that it can handle term reordering.
VI-E Structural Similarity
Column: Structural similarity
Structural similarity computes similarity on graph representations of
binary code.
It sits between syntactic and semantic similarity since
a graph may capture multiple syntactic representations of the same code and
may be annotated with semantic information.
Structural similarity can be computed on different graphs.
The three most common are the intra-procedural control flow graph (CFG),
the inter-procedural control flow graph (ICFG),
and the callgraph (CG).
All three are directed graphs.
In the CFG and ICFG, nodes are basic blocks and
an edge indicates a control flow transition (e.g., branch, jump).
Basic blocks in a CFG belong to a single function;
each function has its own CFG.
Basic blocks in the ICFG belong to any program function;
there is one ICFG per program.
In the CG, nodes are functions and
edges capture caller-callee relationships.
The intuition behind structural approaches is that CG, ICFG, and CFGs
are fairly stable representations whose structure varies little
for similar code.
Approaches that operate on the CG or ICFG have a
whole program input granularity,
while those that operate on CFGs may have function granularity,
or use function similarity as a step towards whole program similarity.
Structural similarity approaches may use labeled graphs.
For example, F2004 and DR2005 use node labels in the CG to capture
structural information about a function’s CFG
(e.g., number of instructions and edges).
Other approaches label basic blocks in the CFG/ICFG with a feature vector that
captures the semantics of the basic clock (e.g., KKMRV2005, BinJuice)
or its embedding (Genius, Gemini, see §VI-F).
Edge labels can be used
to capture the type of control flow transfer (BMM2006)
or to aggregate the semantic labels of source and destination nodes (Fossil).
Structural similarity is used by 27 approaches.
14 approaches operate only on CFGs;
five on both CFGs and the CG;
four only on the ICFG; and
two only on the CG.
There are also three approaches that use non-standard graphs:
Sigma proposes a semantic integrated graph
that combines information from the CFG, CG, and register flow graph,
while QSM2015 and libv use the
execution dependence graph [105].
The remainder of this subsection discusses different approaches
used to compute graph similarity.
(Sub)Graph isomorphism –
Most structural similarity approaches check for variations of graph isomorphism.
An isomorphism of two graphs $G$ and $H$ is an edge-preserving bijection $f$
between their node sets such that if any two nodes $u$, $v$ are adjacent in $G$,
then $f(u)$ and $f(v)$ are also adjacent in $H$.
Graph isomorphism requires that the node set cardinality
is the same in both graphs, which is too strict for binary code similarity.
Thus, approaches instead check for subgraph isomorphism,
which determines if $G$ contains a subgraph isomorphic to $H$.
Subgraph isomorphism is a known NP-complete problem.
Other approaches check for the maximum common subgraph isomorphism (MCS),
which finds the largest subgraph isomorphic to two graphs and is also
NP-Complete.
Given the high complexity of both problems, approaches try to reduce the
number of graph pairs that need to be compared,
as well as the size of the compared graphs.
For example, DR2005 avoids comparing CFGs with the same hash
(match) and CFGs with very different number of nodes and edges
(unlikely match).
iBinHunt reduces the number of nodes to consider
by assigning taint labels to basic blocks.
Only nodes with the same taint label are considered in the
subgraph isomorphism.
For candidate graph pairs that pass the filtering,
approximate algorithms are used that
can be grouped into greedy and backtracking.
Greedy –
These approaches perform neighborhood exploration.
An initial set of matching nodes is first identified.
Then, the matching is recursively expanded by checking only the neighbors
(i.e., parents or children) of already matched nodes.
BMAT, F2004, DR2005, LKI2013, Tedem, Multi-MH, KLKI2016,
Kam1n0, BinSequence, and BinArm use this approach.
A limitation of greedy algorithms is that early errors propagate,
significantly reducing the accuracy.
Backtracking –
Backtracking algorithms fix a wrong matching by revisiting the solution,
and if the new matching does not improve the overall matching it is reverted
(BMM2006, BinHunt, iBinHunt, MXW2015, QSM2015, discovRE).
Backtracking is more expensive,
but can improve accuracy by avoiding local optimal matching.
Optimization.
An alternative used by four approaches
(Smit, BinSlayer, CXZ2014, Genius)
is to model graph similarity as an optimization problem.
Given two CFGs and a cost function between two basic blocks,
they find a bijective mapping between the two CFGs with minimum cost.
Such bipartite matching ignores graph structure,
i.e., does not use edge information.
To address this, Smit and BinSlayer assign lower cost to connected
basic blocks.
To perform the matching,
Smit, BinSlayer, and CXZ2014 use the $O(n^{3})$ Hungarian algorithm,
while Genius uses a genetic algorithm.
K-subgraph matching.
KKMRV2005 proposed to divide a graph into $k$-subgraphs,
where each subgraph contains only $k$ connected nodes.
Then, generate a fingerprint for each k-subgraph and the
similarity of two graphs corresponds to the maximum number
of $k$-subgraphs matched.
Four other approaches later leveraged this approach:
Beagle, CXZ2014, Rendezvous, and Fossil.
Path similarity.
There are three approaches (CoP, Sigma, BinSequence)
that convert function similarity into a path similarity comparison.
First, they extract a set of executions paths from a CFG,
then define a path similarity metric between execution paths, and
finally combine the path similarity into a function similarity.
Graph embedding.
Another method used by Genius, VulSeeker, and Gemini,
detailed in §VI-F,
is to extract a real-valued feature vector from each graph and then
compute the similarity of the feature vectors.
VI-F Feature-Based Similarity
Column: Feature-based, Machine learning
A common method (28 approaches) to compute similarity is to
represent a piece of binary code as a vector or a set of features
such that similar pieces of binary code have similar
feature vectors or feature sets.
A feature captures a syntactic, semantic, or structural property
of the binary code.
Features can be Boolean, numeric, or categorical.
Categorical features have discrete values,
e.g., the mnemonic of an instruction.
A feature vector typically has all numeric or all Boolean features,
the latter is called a bitvector.
Categorical features are typically first encoded into Boolean features using
one-hot encoding or into real-valued features using an embedding.
Of the 28 approaches, 21 use numeric feature vectors,
six use feature sets, and
BinClone uses bitvectors.
Once the features have been extracted,
a similarity metric between feature vectors or feature sets
is used to compute the similarity. Common similarity metrics are
the Jaccard index for feature sets,
dot product for bitvectors,
and the Euclidean or cosine distance for numeric vectors.
Figure 3 shows two alternative methods for
feature-based similarity.
The top method comprises of two steps:
feature selection and feature encoding.
Feature selection is a manual process
where an analyst uses domain knowledge to identify representative features.
The alternative approach showed below is learning to automatically
generate real-valued feature vectors,
called embeddings, from training data.
Embeddings are used by eight recent approaches
(Genius, Gemini, $\alpha$Diff, VulSeeker, RLZ2019, InnerEye, Asm2Vec, SAFE).
Embeddings are used in natural language processing (NLP)
to encode categorical features using real-valued numbers,
which helps deep learning algorithms by reducing the dimensionality
and increasing the density of feature vectors compared to one-hot encoding.
Embeddings enable automatic feature extraction
and efficient similarity computation.
But, features in embeddings do not provide information about what
has been learnt.
Binary code similarity embeddings can be classified by the properties they captured
and their granularity.
The first binary code similarity approach using embeddings was Genius,
later followed by VulSeeker and Gemini.
All the three approaches build a graph embedding for the ACFG of a
function, i.e., a CFG with nodes annotated with selected basic block features.
While Genius uses clustering and graph edit distance to compute the embedding,
VulSeeker and Gemini improve efficiency by training a neural network that avoids
expensive graph operations.
Later approaches
($\alpha$Diff, RLZ2019, InnerEye, Asm2Vec, SAFE)
avoid manually selected features by focusing on instruction,
or raw byte, co-ocurrencence.
In NLP, it is common to extract a word embedding that captures word co-occurrence
(e.g., word2vec) and then build a sentence embedding that builds upon it.
RLZ2019 and InnerEye use an analogous approach by considering
instructions as words and basic blocks as sentences.
SAFE uses similar approach to create functions embedding than basic blocks.
Also related is Asm2Vec that obtains a function embedding by
combining path embeddings capturing instruction co-occurrence
along different execution paths in the function.
Instead of using instruction co-ocurrence, $\alpha$Diff computes a function embedding directly from the sequence of raw bytes
of a function using a convolutional network.
Machine learning.
We identify three uses of machine learning in binary code similarity approaches:
(1) to generate an embedding as explained above,
(2) to cluster similar pieces of binary code using unsupervised learning (BinHash, MutantX-S, iLine, RMKNHLLP2014, KLKI2016, Genius), and
(3) to classify with a probability if the pieces of binary code are being compiled from the same source code
(BinDnn, IMF-sim).
BinDnn is the first use of neural networks for binary code similarity.
Instead of generating an embedding,
BinDnn directly uses a neural network classifier to determine if two functions
are compiled from the same source code.
Surprisingly, BinDnn is not cited
by later binary code similarity approaches including those using neural networks to build embeddings.
VI-G Hashing
Column: Locality sensitive hashing
A hash is a function that maps data of arbitrary size to a fixed-size value.
Hash values are compact to store, efficient to compute, and
efficient to compare, which makes them great for indexing.
Hashes operate at the raw-byte level.
They are not especifically designed for binary code,
but rather for arbitrary binary data.
However, three classes of hashes have been used for binary code similarity:
cryptographic hashes, locality-sensitive hashes, and executable file hashes.
Cryptographic hashes capture identical, rather than similar, inputs.
They are used by some binary code similarity approaches to quickly identify duplicates at
fine granularity (e.g., basic block).
Locality-sensitive hashes produce similar hash values for similar inputs,
as oppossed to cryptographic hashes where a small difference in the input
generates a completely different hash value.
Executable file hashes take as input an executable file,
but the hash computation only considers parts of the executable
such as the import table or selected fields of the executable’s header.
Their goal is to output the same hash value for polymorphic variants
of the same malware.
Locality sensitive hashing (LSH).
LSH produces similar hash values for similar inputs,
efficiently approximating a nearest neighbor search.
LSH algorithms typically apply multiple hash functions on the given input
and produce the same hash value for similar inputs,
i.e., they increase collision probability for similar inputs.
LSH is used in binary code similarity to boost performance.
For example, it is used by $OM$ approaches for indexing pieces of binary code,
enabling efficient binary code search (Kam1n0, Gemini, InnerEye).
Of the 11 approaches that use LSH, seven use MinHash [106]
(BinHash, Multi-MH, Genius, BinSequence, BinShape, CACompare, BinSign),
two do not specify the algorithm used (Gemini, InnerEye),
SWPQS2006 uses the algorithm by Gionis et al. [107],
and Kam1n0 proposes its own Adaptive Locality Sensitive Hashing (ALSH).
Fuzzy hashing is a popular type of LSH used
to compute similarity of arbitrary files.
For example, the VirusTotal file analysis service [108]
reports ssdeep [52] hashes for submitted files.
Other fuzzy hashes include tlsh [53],
sdhash [109], and
mrsh-v2 [110].
None of the 61 approaches use them,
but we briefly discuss them because
they are often applied to executable files.
When applied on executables, fuzzy hashes may capture similarity
of the binary code, but also similarity of data present in the executables.
This issue has been recently examined by Pagani et al. [111].
Their results show that when applied only on the code of the
.text section they work significantly worse than
when used for whole program similarity.
In fact, they show that a byte change at the right position,
extra nop instructions, and instruction swapping
can degrade similarity significantly (in some cases bring it down to zero).
They observe that when compiling the same source code with
different optimizations,
the data sections of the executables remain the same,
which seems to be a key reason fuzzy hashes work better on whole executables.
Executable file hashes.
This class of hashes are especifically designed for executable files.
They are designed to output the same hash value for
polymorphic variants of the same malware.
The hash computation only considers parts of the executable that are less
likely to change when simply repacking, or resigning, the same executable.
They are not used by any of the 61 approaches,
but we briefly describe three popular hashes for completeness.
peHash [51] hashes selected fields of a PE executable
that are less susceptible to changes during compilation and packing,
e.g., initial stack size, heap size.
ImpHash hashes the import table of an executable.
Since the functionality of packed variants is the same, their imported
functions should be the same as well.
Unfortunately, it gives false positives with unrelated executables packed
with the same packer, if the packer reconstructs the original import table
at runtime.
Authentihash is the hash of a PE executable ignoring its
Windows Authenticode code signing data.
It enables identifying the same executable signed by different publishers.
It hashes the whole PE executable except three pieces:
the Authenticode data, pointers to the Authenticode data, and
the file checksum.
VI-H Supported Architectures
Column: Cross-architecture
A cross-architecture approach can compare pieces of binary code for
different CPU architectures, e.g., x86, ARM, and MIPS.
This differs from architecture-independent approaches (e.g., F2004)
that support different architectures,
but cannot cross-compare among them,
i.e., they can compare two x86 inputs and two MIPS inputs,
but cannot compare an x86 input with a MIPS input.
There are 16 cross-architecture approaches, all proposed since 2015.
A common application is given a buggy piece of binary code,
to search for similar pieces of binary code, compiled for other architectures,
which may also contain the bug.
For example, to search for programs in firmware images
where a version of OpenSSL vulnerable to Heartbleed
has been statically compiled.
The code syntax for different architectures may significantly differ as
they may use separate instruction sets with different instruction mnemonics,
sets of registers, and default calling conventions.
Thus, cross-architecture approaches compute semantic similarity.
Cross-architecture approaches employ one of two techniques.
Seven approaches lift the
binary code to an architecture-independent intermediate representation (IR):
Multi-MH, MockingBird, BinGo, Xmatch, CACompare, GitZ, FirmUp.
Then, identical analysis can be peformeed on the IR,
regardless of the origin architecture.
The advantage is that the analysis only depends on the IR and the IR design
can be outsourced to a separate group.
Section VII details the specific architectures
supported by each approach and the IRs they use.
An alternative approach used by 9 approaches is to
use feature-based similarity (discussed in §VI-F).
These approaches use a separate module for each architecture to obtain
a feature vector that captures the semantics of the binary code
(discovRE, BinDnn, Genius, Gemini, $\alpha$Diff, VulSeeker, RLZ2019,
InnerEye, SAFE).
VI-I Type of Analysis
Column: Static analysis; Dynamic analysis; Dataflow analysis
Binary code similarity approaches can use static analysis, dynamic analysis, or both.
Static analysis examines the disassembled binary code, without executing it.
Instead, dynamic analysis examines code executions by running the
code on selected inputs.
Dynamic analysis can be performed online,
as the code executes in a controlled environment,
or offline on traces of the execution.
Fifty one approaches use only static analysis,
four use only dynamic analysis, and six combine both.
The dominance of static analysis for binary code similarity is due to most applications
requiring all the input code to be compared.
This is easier with static analysis as it provides complete code coverage.
Dynamic analysis examines one execution at a time and
can only determine similarity of the code run in that execution.
To increase the coverage of dynamic analysis,
three approaches (iBinHunt, Blex, IMF-sim) run the code on multiple inputs
covering different execution paths and combine the results.
However, it is infeasible to run any non-trivial piece of binary code on inputs that cover all possible execution paths.
One advantage of dynamic analysis is simplicity,
as memory addresses, operand values, and control-flow targets
are known at runtime, which sidesteps static analysis challenges such as
memory aliasing and indirect jumps.
Another advantage is that it can handle some obfuscations and does not
require disassembly of the code,
also difficult with obfuscated code [112].
Section VIII details which approaches have evaluated its
robustness on obfuscated code.
Overall, dynamic analysis has been used in binary code similarity for
malware unpacking (Smit, Beagle, MutantX-S, CXZ2014, BinSim),
for operating on trace granularity (BinSim, KS2017), and for
collecting runtime values for semantic similarity (Blex, BinSim, IMF-sim).
Dataflow analysis is a common type of analysis that
examines how values propagate through the code.
It comprises of data sources to be tracked
(e.g., registers or memory locations holding specific variables),
propagation rules defining how values are propagated by
different instructions or IR statements, and sinks,
i.e., program points where to check the values reaching them.
Of the 19 approaches that use dataflow analysis,
16 use symbolic execution to extract a set of symbolic formulas
to compute semantic similarity (§VI-D).
Spain uses taint analysis to summarize the patterns of
vulnerabilities and their security patches.
And, iBinHunt uses both taint analysis and symbolic execution.
It first uses taint analysis as a filter to find pieces of binary code that process the
same user input, restricting the expensive subgraph isomorphism computation
to those with the same taint label.
And, it computes basic block similarity using symbolic formulas.
IMF-sim uses backward taint analysis to infer pointer
arguments of a function from dereference errors.
VI-J Normalization
Column: Normalization
Syntactic similarity approaches often normalize instructions,
so that two instructions that are normalized to the same form
are considered similar despite some syntactic differences,
e.g., different registers being used.
Overall, there are 33 approaches that use instruction normalization.
They apply the following three types of instruction normalization:
•
Operand removal –
A normalization used by nine approaches is to abstract an
instruction only by its mnemonic or opcode, ignoring all operands.
For example, $add\,\%eax,\%ebx$ and $add\,[\%ecx],\%edx$ would be both
represented by $add$ and considered similar, despite both using
different operands.
•
Operand normalization –
A normalization used by 17 approaches is to replace
instruction operands with symbols that capture the operand type
such as REG for register, MEM for memory, and
IMM for immediate values. For example, $add\,\%eax,\%ebx$ and $add\,\%ecx,\%edx$ would be both
represented as $add\,REG,REG$,
matching the instructions despite different register allocations
used by the compiler. Operand normalization abstracts less than operand removal.
For example, $add\,[\%ecx],\%edx$ would be represented
as $add\,MEM,REG$ and thus be considered different from the above.
Some approaches also use different symbols for
general purpose registers and segment registers, and
for operands of different sizes (e.g., RegGen8 and RegGen32 in BinClone).
•
Mnemonic normalization –
A normalization used by 3 approches (BMAT, Exposé and iLine) is to
represent multiple mnemonics by the same symbol.
For example, both BMAT and iLine represent all conditional jumps
(e.g., je, jne) with the same symbol to account for the compiler
modifying jump conditions.
Another type of normalization is to ignore code that should not
affect the semantics.
For example, some instruction sets contain no-op instructions
that do not change the process state.
And, compilers often use no-op equivalent instructions for padding
such as instructions that move a register to itself,
e.g., $mov\,\%eax,\%eax$.
No-op equivalent instructions do not matter for semantic similarity and
structural similarity approaches,
but since they change the syntax they may affect syntactic
similarity approaches.
Three approaches remove no-op instructions (iLine, MXW2015, QSM2015).
A few approaches also remove unreachable dead code (BMM2006, FirmUp),
which may be introduced by obfuscations;
function epilogue and prologue (Exposé) instructions,
which may make small unrelated functions to look similar; and
functions added by the compiler to load the program,
not present in the source code (BinGo, Sigma).
VII Implementations
This section systematizes the implementation of the 61 approaches.
For each approach, Table III shows
the static and dynamic platforms it builds on,
the programming language used to implement the approach,
whether the implementation supports distributing the analysis,
the supported target program architectures and operating systems, and
how the approach is released.
A dash (-) indicates that we could not find the information (i.e., unknown),
while a cross ( ✗) means unsupported/unused.
Building a binary code similarity approach from scratch requires significant effort.
Thus, all approaches build on top of previously
available binary analysis platforms or tools, which provide functionality
such as disassembly and control flow graphs for static analysis, or
instruction-level monitoring for dynamic analysis.
However, the implementation of an approach may only use
a subset of the functionality offered by the underlying platform.
The most popular static platform is Ida (42 approaches),
followed by BAP (4), DynInst (3), and McSema (2).
Ida main functionalities are disassembly and building control flow graphs.
Its popularity comes from supporting a large number of architectures.
Some binary analysis platforms already support using Ida as their
disassembler, so it is not uncommon to combine Ida and another platform
(6 approaches).
Among dynamic approaches,
PIN is the most popular with 5 approaches,
followed by Temu and Anubis with 3 approaches each.
Previous work has analyzed binary analysis platforms used by binary code
type inferece approaches [35].
Since most platforms overlap,
we refer the reader to that work for platform details,
but provide an extended version of their table in the Appendix
(Table V) with six extra platforms.
A few approaches build on top of previous binary code similarity approaches.
One case is that both approaches have overlapping authors.
For example, RMKNHLLP2014 is based on BinJuice,
and MXW2015 extends iBinHunt, which already shared components with BinHunt.
The other case is using a previously released approach.
For example, BinSlayer and Spain use BinDiff as a first stage filter to
find matched functions.
The most popular programming language is Python (20 approaches),
followed by C++ (14).
One reason for this is that the majority of approaches use Ida for disassembly
and Ida supports both C++ and Python plugins.
Moreover, two approaches (Kam1n0, BinSign) have used distributed
platforms, such as Hadoop, to distribute their analysis.
Binary code analysis can operate directly on a particular instruction set
(e.g., x86, ARM) or convert the instruction set into an
intermediate representation (IR).
Using an IR has two main advantages.
First, it is easier to reuse the analysis built on top of an IR.
For example, supporting a new architecture only requires adding a new front-end
to translate into the IR, but the analysis code can be reused.
Second, complex instruction sets (e.g., x86/x86-64) can be converted
into a smaller set of IR statements and expressions,
which make explicit any side-effects such as implicit operands or
conditional flags.
There are 15 approaches that use an IR and most use the IR
provided by the underlying platform.
Out of 15 approaches, 5 use VINE provided by BitBlaze (BinHunt, iBinHunt, CoP, MXW2015, BinSim),
another 5 use VEX provided with Valgrind (Multi-MH, MockingBird, CACompare, GitZ, FirmUp), and
two use LLVM-IR (Esh, Xmatch).
The remaining three approaches use
SAGE III (SWPQS2006), METASM (Tedem) and REIL (BinGo).
It is surprising that only 6 of the 16 cross-architecture approaches use an IR
(Multi-MH, Xmatch, MockingBird, CACompare, GitZ, FirmUp).
The remaining approaches provide separate analysis modules for each architecture
that typically output a feature vector with a common format.
The most supported architectures for the target programs to be analyzed
are x86/x86-64 (59 approaches),
followed by ARM (16), and MIPS (10).
The only two approaches that do not support x86/x86-64 are Exediff that targets Digital Alpha, and BinArm that targets ARM.
There are 16 cross-architecture approaches and 9
approaches that support firmware.
The first cross-architecture approach was Multi-MH,
which added support for ARM in 2015.
Since then, ARM support has become very prevalent due to the popularity
of mobile and IoT devices.
It is worth noting that even if there are 42 approaches that use Ida,
which supports more than 60 processor families,
most approaches built on top of Ida only analyze x86/x86-64 programs.
The most supported OS is Linux (41 approaches)
followed by Windows (35).
Only 4 approaches support MacOS.
Early approaches that focused on x86/x86-64 often used Ida to obtain support
for both PE/Windows and ELF/Linux executables.
Most recently, all approaches that leverage ARM support Linux,
which is used by Android and also by many IoT devices.
Of the 61 approaches, only 12 are open source
(BinSlayer, Tracy, QSM2015, Esh, Genius, Kam1n0, Gemini, Asm2Vec,
VulSeeker, RLZ2019, InnerEye, SAFE).
DR2005 was implemented in the BinDiff commercial tool,
which is now available as a free binary.
The remaining approaches have not been released in any form,
although the platforms they build on may be open source.
Moreover, 4 of the 12 open-source approaches (Esh, Genius, RLZ2019, InnerEye)
have partially released their source code and machine learning models.
VIII Evaluations
This section systematizes the evaluation of the 61 binary code similarity approaches.
For each approach, Table IV summarizes the
datasets used (§VIII-A) and
the evaluation methodology (§VIII-B).
VIII-A Datasets
The left side of Table IV describes the datasets used
by each approach.
It first shows
the total number of executables used in the evaluation and
their split into benign and malicious executables.
Executables may come from different programs or correspond to
multiple versions of the same program,
e.g., with varying compiler and compilation options.
Then, for approaches that have function granularity, it captures the total
number of functions evaluated,
and for approaches that analyze firmware,
the number of images from where the executables are obtained.
A dash (–) in a column means that we could not find the number in the paper.
For example, SWPQS2006 evaluates on system library files in Windows XP,
but the total number of executables is not indicated.
Most approaches use custom datasets,
the one popular benchmark is Coreutils used by 18 approaches.
In addition, approaches that evaluate on firmware use
two openly available firmwares
(ReadyNAS [113] and DD-WRT [114]).
Over half of the approaches evaluate on
less than 100 executables,
7 on less than 1K,
8 on less than 10K, and
only 8 on over 10K.
Out of 61 approaches,
37 have evaluated only on benign programs,
8 only on malware, and 16 on both.
This indicates that binary code similarity is also popular for malware analysis.
However, of the 24 approaches evaluated on malware,
only five use packed malware samples
(Smit, Beagle, MutantX-S, CXZ2014, BinSim).
These approaches first unpack the malware using a
custom unpacker (Smit) or a generic (write-and-execute) unpacker
(Beagle, MutantX-S, CXZ2014, BinSim), and then compute binary code similarity.
The rest have access to the malware’s source code or to
unpacked samples.
For binary code search approaches that use function granularity,
the number of functions in the repository better captures the dataset size.
The largest dataset is by Genius,
which evaluates on 420M functions extracted from 8,126 firmwares.
Prior approaches had evaluated on at most 0.5M functions,
which demonstrates the scalability gains from its embedding approach.
Five other approaches have evaluated on over 1M functions:
Fossil (1.5M),
BinSequence (3.2M),
BinArm (3.2M),
FirmUp (40M), and
Gemini (420M), which uses the same dataset as Genius.
In addition, InnerEye has evaluated on a repository of 1.2M basic blocks.
VIII-B Methodology
The right side of Table IV describes four aspects of
the evaluation methodology used by each approach:
robustness,
accuracy,
performance,
and comparison with prior approaches.
Robustness.
The first 8 columns on the right side of Table IV
capture how authors evaluate the robustness of their approaches,
i.e., their ability to capture similarity despite transformations
applied to the input programs.
First, it shows whether they use each of the four compilers we have
observed being employed to compile the programs in the dataset:
GCC, ICC, Visual Studio (MSVS), and Clang.
Then, it captures whether the authors evaluate similarity between
programs compiled with different compilation options (cross-optimization),
between programs compiled with different compilers (cross-compiler), and
between programs compiled for different architectures (cross-architecture).
Finally, it captures whether the authors evaluate similarity when
obfuscation transformations are applied to the input programs.
There are 34 approaches that evaluate robustness
(at least one ✓in the last four robustness columns) and 27 that do not.
Many early works did not evaluate robustness.
This evaluation has become increasingly popular as approaches mature.
The most popular robustness evaluation is
cross-optimization (23 approaches), followed by
cross-compiler (19), cross-architecture (16), and obfuscation (13).
There are 9 approaches that have evaluated cross-optimization, cross-compiler, and
cross-architecture.
Approaches that evaluate cross-compiler also typically evaluate
cross-optimization, as it is a simpler case.
Simlarly, approaches that evaluate cross-architecture
typically also evaluate cross-compiler,
as cross-architecture programs may be produced using different compilers.
Note that it is possible for approaches to compile programs with multiple
compilers, but not perform cross-compiler evaluation,
i.e., not compare similarity between programs compiled with different
compilers.
There are 13 approaches that have evaluated on obfuscated programs.
Of those, two use only source code transformations (CoP, Asm2Vec),
three use only binary code transformations (KKMRV2005, Tpm, BinShape),
five use packed malware (Smit, Beagle, MutantX-S, CXZ2014, BinSim), and
three evaluate both source code and binary code transformations
(BinSim, IMF-sim, Fossil).
Accuracy evaluation and comparison.
There are 49 approaches that perform a quantitative evaluation
of their accuracy using some ground truth (✓), and
12 that perform qualitative accuracy evaluation through case studies ( ✗).
Quantitative evaluation most often uses standard accuracy metrics such as
true positives, false positives, precision, and recall.
However, two approaches propose novel application-specific accuracy
metrics (iLine, KS2017).
There are 33 approaches that compare with prior approaches.
All of them compare accuracy and six also compare runtime.
The top target for comparison is BinDiff (13 approaches compare with it),
followed by Tracy (5), discovRE (4), and Multi-MH (4).
Comparing accuracy across binary code similarity approaches is challenging for
multiple reasons.
First, only a small fraction of the proposed approaches have publicly released
their code (Section VII).
Since most approaches are not publicly available,
comparison is often performed by re-implementing previous approaches,
which may require significant effort.
One advantage of reimplementation is that approaches can be compared
on new datasets.
The alternative to re-implementation is to evaluate the new approach on
the same dataset used by a prior approach,
and compare with the reported results.
This method is only used by 6 approaches
(Genius, BinGo, Xmatch, CACompare, Gemini, and BinArm)
likely because most datasets are custom and not publicly available.
Fortunately, we observe that public code release has become more
common in recent approaches.
Second, the input comparison and input granularity may differ among approaches
making it nearly impossible to perform a fair comparison.
For instance, it is hard to compare in a fair manner an approach that
identifies program similarity using callgraphs (e.g., Smit)
with an approach comparing basic blocks (e.g., InnerEye).
Third, even when the input comparison and input granularity are the same,
the evaluation metrics and methodology may differ,
significantly impacting the measured accuracy.
The latter challenge is best illustrated on binary code search approaches operating at
function granularity.
These approaches find the most similar functions in a repository to a
given function.
They return multiple entries ranked in descending similarity order and
count a true positive if one of the top-k most
similar entries is a true match.
Unfortunately, the values of $k$ vary across approaches and significantly
impact the accuracy,
e.g., a 98% precision on top 10 is significantly worse than a
98% precision on top 3.
Thus, it becomes hard to compare accuracy numbers obtained with different
$k$ values and it becomes tempting to raise $k$ until a sufficiently
high accuracy number is achieved.
Furthermore, many approaches do not describe the similarity threshold
used to determine that no similar entry exists in the repository.
This means that they always find some similar entry in the repository,
even if the similarity may be really low.
Performance.
It is common (49/61) to measure the runtime performance of an approach.
Runtime is typically measured end-to-end,
but a few approaches report it for each approach component (e.g., BinSim).
Four approaches report their asymptotic complexity
(BMM2006, SWPQS2006, iLine, MockingBird).
IX Discussion
This section discusses open challenges and possible future research directions.
Small pieces of binary code.
Many binary code similarity approaches ignore small pieces of binary code,
setting a threshold on the minimum number of instructions or basic blocks
to be considered.
Oftentimes, only pieces of binary code with a handful of instructions are ignored,
e.g., functions with less than 5 instructions,
but some approaches use large thresholds like 100 basic blocks in Tracy.
Small pieces of binary code are challenging because they are common,
may comprise of a single basic block that prevents structural analysis, and
may have identical syntax despite different semantics.
For example, setter functions that update the value of a field in an object
have nearly identical syntax and structure,
simply setting a memory variable with the value of a parameter.
But, they may have very different semantics,
e.g., setting a security level or updating a performance counter,
Furthermore, semantic similarity techniques like instruction classification,
symbolic formulas, and input-output pairs may fail to capture their
differences, e.g., if they do not distinguish different memory variables.
Similarity of small pieces of binary code remains an open challenge.
One potential avenue would be to further incorporate context.
Some structural approaches already consider the callgraph to match functions
(e.g., F2004), but this does not always suffice.
We believe that it may be possible to further incorporate other context
like locality (e.g., how close the binary code is in the program structure) or
data references (e.g., whether they use equivalent variables).
Source-to-binary similarity.
Some applications like plagiarism detection
may require source code to be compared with binary code.
Early approaches for source-to-binary similarity used software birthmarks
that capture inherent functionality of the
source code [92, 93].
Recently, source-to-binary similarity has been applied for
searching if a known bug in open source code exists in
some target binary code [44].
The availability of source code provides more semantics
(e.g., variable and function names, types, comments)
potentially improving accuracy compared to simply
compiling the source code and performing binary code similarity.
We believe other applications remain that require
determining if a target piece of binary code has been compiled, or has evolved,
from some given source code.
For example, there may be programs for which source code is only available
for an old version and there is a need to understand how newer binary versions
have evolved from the original source code.
Data similarity.
This survey has focused on binary code similarity,
but programs comprise both code and data.
There may be situations where the data used by the code is as important as
the code, e.g., when the only change between versions of a program is in the
data used such as changing parameters for a machine learning classifier.
Furthermore, data may be stored in complex data structures that may be
key to the functionality.
There exists a long history of type inference techniques on
binary code [35],
which we believe could be combined with binary code similarity to compare the data structures
used by different pieces of binary code, or how two pieces of binary code use the same data structure.
Semantic relationships.
A different aspect of semantic similarity is to identify binary code
with related functionality,
e.g., cryptographic or networking functions.
This is challenging as code that is related in its functionality
may not have the same semantics.
For example, a decryption function is clearly related to its
encryption function, but performs opposite operations.
So far, most work has focused on domain specific techniques such as those
for identifying cryptographic functions
(e.g., [115, 116]).
But, recently some approaches have started exploring domain-agnostic
techniques [117].
We believe further work is needed to better define such semantic relationships
and their identification.
Scalability.
The largest binary code similarity evaluation so far is by Genius and Gemini on 420M functions
from 8,126 firmware images (500K functions per image).
While that is a significant step from prior approaches,
if we consider instead 100K unique firmware,
a conservative number since it is expected that there will be
20 Billion IoT devices connected to the Internet by 2020 [118],
we need a binary code similarity approach that can handle 50 Billion functions.
Thus, further improvements on scalability will be needed to realize the
vision of binary code search engines.
Obfuscation.
Many challenges still remain for binary code similarity on obfuscated code.
For example, a recent study has shown that state of the art
unpacking techniques can miss 20%–60% of the original code,
e.g., due to incomplete function detection [89].
And, no binary code similarity approach currently handles virtualization-based packers
such as Themida [119] and VMProtect [120],
which generate a random bytecode instruction set,
transform an input piece of binary code into that bytecode instruction set, and
attach an interpreter (or virtual machine)
from the generated bytecode to native code.
Furthermore, obfuscation is best addressed with semantic similarity,
which has a challenge with obfuscation transformations that do not
respect the original semantics of the code, but still perform its main goals.
Approach comparison.
The variety of datasets and methodologies used to evaluate the approaches
(e.g., top $k$ evaluation for OM approaches),
together with the absence of source code for many of them,
makes it hard to perform a fair comparison to understand
their benefits and limitations.
We believe the field would greatly benefit from open datasets for benchmarking,
and independent validation under the same experimental conditions.
Furthermore, there is a need to improve the comparison of approaches
that handle obfuscation, beyond changes of compiler and compiler options.
Building a dataset that covers real-world obfuscations is fundamental
given the huge space of possible obfuscation transformations and that
each approach supports a different subset.
X Conclusion
During the last 20 years,
researchers have proposed many approaches to perform binary code similarity and
have applied them to address important problems such as
patch analysis, bug search, and malware detection and analysis.
The field has evolved
from binary code diffing to binary code search;
from syntactic similarity to incorporate structural and semantic similarity;
to cover multiple code granularities,
to strengthen the robustness of the comparison
(i.e., cross-optimization, cross-compiler, cross-OS, cross-architecture,
obfuscation);
to scale the number of inputs compared
(e.g., through hashing, embedding, indexing); and
to automatically learn similarity features.
Despite its popularity, the area had not yet been systematically analyzed.
This paper has presented a first survey of binary code similarity.
The core of the paper has systematized 61 binary code similarity approaches on four dimensions:
the applications they enable,
their approach characteristics,
how the approaches are implemented, and
the benchmarks and evaluation methodologies used to evaluate them.
It has discussed the advantages and limitations of
different approaches, their implementation, and their evaluation.
It has summarized the results into easy to access tables useful
for both beginners and experts.
Furthermore, it has discussed the evolution of the area and
outlined the challenges that remain open,
showing that the area has a bright future.
Acknowledgments.
This research was partially supported
by the Regional Government of Madrid through the
BLOQUES-CM (S2018/TCS-4339) grant; and
by the Spanish Government through the
SCUM (RTI2018-102043-B-I00) grant.
This research is also supported by the EIE Funds of the European Union.
All opinions, findings and conclusions, or recommendations expressed herein
are those of the authors and do not necessarily reflect the views
of the sponsors.
References
[1]
Y. David and E. Yahav, “Tracelet-based Code Search in Executables,” in
Conference on Programming Language Design and Implementation. ACM, 2014.
[2]
J. Pewny, F. Schuster, L. Bernhard, T. Holz, and C. Rossow, “Leveraging
Semantic Signatures for Bug Search in Binary Programs,” in Annual
Computer Security Applications Conference. ACM, 2014.
[3]
J. Pewny, B. Garmany, R. Gawlik, C. Rossow, and T. Holz, “Cross-architecture
bug search in binary executables,” in IEEE Symposium on Security and
Privacy. IEEE Computer Society, 2015.
[4]
S. Eschweiler, K. Yakdan, and E. Gerhards-Padilla, “discovRE: Efficient
Cross-Architecture Identification of Bugs in Binary Code,” in Network
and Distributed System Security Symposium, 2016.
[5]
Y. David, N. Partush, and E. Yahav, “Statistical Similarity of Binaries,”
in Conference on Programming Language Design and Implementation. ACM, 2016.
[6]
Q. Feng, R. Zhou, C. Xu, Y. Cheng, B. Testa, and H. Yin, “Scalable
Graph-based Bug Search for Firmware Images,” in ACM Conference on
Computer and Communications Security. ACM, 2016.
[7]
M. Chandramohan, Y. Xue, Z. Xu, Y. Liu, C. Y. Cho, and H. B. K. Tan, “BinGo:
Cross-architecture cross-OS Binary Search,” in International
Symposium on Foundations of Software Engineering. ACM, 2016.
[8]
H. Huang, A. M. Youssef, and M. Debbabi, “BinSequence: Fast, Accurate and
Scalable Binary Code Reuse Detection,” in Asia Conference on Computer
and Communications Security. ACM,
2017.
[9]
Q. Feng, M. Wang, M. Zhang, R. Zhou, A. Henderson, and H. Yin, “Extracting
Conditional Formulas for Cross-Platform Bug Search,” in Asia
Conference on Computer and Communications Security. ACM, 2017.
[10]
Y. David, N. Partush, and E. Yahav, “Similarity of Binaries Through
Re-optimization,” in Conference on Programming Language Design and
Implementation. ACM, 2017.
[11]
J. Gao, X. Yang, Y. Fu, Y. Jiang, and J. Sun, “VulSeeker: A Semantic Learning
Based Vulnerability Seeker for Cross-platform Binary,” in Conference
on Automated Software Engineering, ser. ASE 2018, 2018.
[12]
X. Xu, C. Liu, Q. Feng, H. Yin, L. Song, and D. Song, “Neural Network-based
Graph Embedding for Cross-Platform Binary Code Similarity Detection,” in
ACM Conference on Computer and Communication Security. ACM, 2017.
[13]
Y. David, N. Partush, and E. Yahav, “FirmUp: Precise Static Detection of
Common Vulnerabilities in Firmware,” in Conference on Architectural
Support for Programming Languages and Operating Systems. ACM, 2018.
[14]
P. Shirani, L. Collard, B. L. Agba, B. Lebel, M. Debbabi, L. Wang, and
A. Hanna, “BINARM: Scalable and Efficient Detection of Vulnerabilities in
Firmware Images of Intelligent Electronic Devices,” in Conference on
Detection of Intrusions and Malware, and Vulnerability Assessment, 2018.
[15]
B. Liu, W. Huo, C. Zhang, W. Li, F. Li, A. Piao, and W. Zou, “$\alpha$Diff:
Cross-version Binary Code Similarity Detection with DNN,” in
Conference on Automated Software Engineering. ACM, 2018.
[16]
X. Hu, T.-c. Chiueh, and K. G. Shin, “Large-scale Malware Indexing Using
Function-call Graphs,” in ACM Conference on Computer and
Communications Security. ACM, 2009.
[17]
X. Hu, K. G. Shin, S. Bhatkar, and K. Griffin, “MutantX-S: Scalable Malware
Clustering Based on Static Features,” in USENIX Annual Technical
Conference, 2013.
[18]
T. Kim, Y. R. Lee, B. Kang, and E. G. Im, “Binary executable file similarity
calculation using function matching,” The Journal of Supercomputing,
Dec 2016.
[19]
C. Kruegel, E. Kirda, D. Mutz, W. Robertson, and G. Vigna, “Polymorphic Worm
Detection Using Structural Information of Executables,” in Symopsium
on Recent Advance in Intrusion Detection. Springer Berlin Heidelberg, 2005.
[20]
D. Bruschi, L. Martignoni, and M. Monga, “Detecting Self-mutating Malware
Using Control-flow Graph Matching,” in Detection of Intrusions and
Malware and Vulnerability Assessment. Springer-Verlag, 2006.
[21]
S. Cesare, Y. Xiang, and W. Zhou, “Control Flow-Based Malware Variant
Detection,” IEEE Transactions on Dependable and Secure Computing,
vol. 11, no. 4, pp. 307–317, 2014.
[22]
M. Lindorfer, A. Di Federico, F. Maggi, P. M. Comparetti, and S. Zanero,
“Lines of Malicious Code: Insights into the Malicious Software Industry,”
in Annual Computer Security Applications Conference. ACM, 2012.
[23]
J. Jang, M. Woo, and D. Brumley, “Towards Automatic Software Lineage
Inference,” in 22nd USENIX Security Symposium, 2013.
[24]
J. Ming, D. Xu, and D. Wu, “Memoized semantics-based binary diffing with
application to malware lineage inference,” in IFIP International
Information Security Conference. Springer, 2015.
[25]
B. S. Baker, U. Manber, and R. Muth, “Compressing Differences of Executable
Code,” in ACMSIGPLAN Workshop on Compiler Support for System
Software, 1999.
[26]
H. Flake, “Structural comparison of executable objects,” in Detection
of Intrusions and Malware & Vulnerability Assessment, 2004.
[27]
T. Dullien and R. Rolles, “Graph-based Comparison of Executable Objects,”
in Symposium Sur La Securite Des Technologies De L’Information Et Des
Communications, 2005.
[28]
D. Gao, M. K. Reiter, and D. Song, BinHunt: Automatically Finding
Semantic Differences in Binary Programs. Springer Berlin Heidelberg, 2008.
[29]
Y. Hu, Y. Zhang, J. Li, and D. Gu, “Cross-architecture binary semantics
understanding via similar code comparison,” in Conference on Software
Analysis, Evolution, and Reengineering. IEEE, 2016.
[30]
Z. Xu, B. Chen, M. Chandramohan, Y. Liu, and F. Song, “SPAIN: Security Patch
Analysis for Binaries Towards Understanding the Pain and Pills,” in
International Conference on Software Engineering. IEEE Press, 2017.
[31]
U. Kargén and N. Shahmehri, “Towards Robust Instruction-level Trace
Alignment of Binary Code,” in International Conference on Automated
Software Engineering. IEEE Press,
2017.
[32]
Z. Wang, K. Pierce, and S. McFarling, “BMAT - A Binary Matching Tool,” in
In Second ACM Workshop on Feedback-Directed and Dynamic Optimization,
1999.
[33]
L. Luo, J. Ming, D. Wu, P. Liu, and S. Zhu, “Semantics-based
Obfuscation-resilient Binary Code Similarity Comparison with Applications to
Software Plagiarism Detection,” in International Symposium on
Foundations of Software Engineering. ACM, 2014.
[34]
K. A. Roundy and B. P. Miller, “Binary-Code Obfuscations in Prevalent Packer
Tools,” ACM Computing Surveys (CSUR), vol. 46, no. 1, p. 4, 2013.
[35]
J. Caballero and Z. Lin, “Type Inference on Executables,” ACM
Computing Surveys, vol. 48, no. 4, pp. 1–35, May 2016.
[36]
M. Egele, T. Scholte, E. Kirda, and C. Kruegel, “A Survey on Automated
Dynamic Malware-Analysis Techniques and Tools,” ACM computing surveys
(CSUR), vol. 44, no. 2, p. 6, 2012.
[37]
J. Wang, H. T. Shen, J. Song, and J. Ji, “Hashing for Similarity Search: A
Survey,” Tech. Rep., 2014.
[38]
S.-H. Cha, “Comprehensive Survey on Distance/Similarity Measures between
Probability Density Functions,” vol. 1, 01 2007.
[39]
S. seok Choi and S. hyuk Cha, “A survey of Binary similarity and distance
measures,” Journal of Systemics, Cybernetics and Informatics, pp.
43–48, 2010.
[40]
C. Collberg, S. Martin, J. Myers, and J. Nagra, “Distributed Application
Tamper Detection via Continuous Software Updates,” in 28th Annual
Computer Security Applications Conference, 2012.
[41]
X. Ugarte-Pedrero, D. Balzarotti, I. Santos, and P. G. Bringas, “SOK: Deep
packer inspection: A longitudinal study of the complexity of run-time
packers,” in IEEE Symposium on Security and Privacy, 2015.
[42]
A. Lakhotia, M. D. Preda, and R. Giacobazzi, “Fast Location of Similar Code
Fragments Using Semantic ’Juice’,” in Program Protection and Reverse
Engineering Workshop. ACM, 2013.
[43]
T. Kamiya, S. Kusumoto, and K. Inoue, “CCFinder: a multilinguistic
token-based code clone detection system for large scale source code,”
IEEE Transactions on Software Engineering, vol. 28, no. 7, pp.
654–670, 2002.
[44]
H. Zhang and Z. Qian, “Precise and Accurate Patch Presence Test for
Binaries,” in 27th USENIX Security Symposium, Baltimore, MD, 2018.
[45]
J. Crussell, C. Gibler, and H. Chen, “AnDarwin: Scalable Detection of
Semantically Similar Android Applications,” in European Symposium on
Research in Computer Security. Springer Berlin Heidelberg, 2013.
[46]
K. Chen, P. Liu, and Y. Zhang, “Achieving accuracy and scalability
simultaneously in detecting application clones on android markets,” in
Proceedings of the 36th International Conference on Software
Engineering. ACM, 2014.
[47]
S. Forrest, S. A. Hofmeyr, A. Somayaji, and T. A. Longstaff, “A sense of self
for unix processes,” in IEEE Symposium on Security and Privacy. IEEE, 1996.
[48]
E. Kirda, C. Kruegel, G. Banks, G. Vigna, and R. Kemmerer, “Behavior-based
spyware detection.” in USENIX Security Symposium, 2006.
[49]
M. Christodorescu, S. Jha, and C. Kruegel, “Mining Specifications of
Malicious Behavior,” in 6th Joint Meeting of the European Software
Engineering Conference and the ACM SIGSOFT Symposium on The Foundations of
Software Engineering. ACM, 2007.
[50]
U. Bayer, P. M. Comparetti, C. Hlauschek, C. Kruegel, and E. Kirda,
“Scalable, Behavior-Based Malware Clustering,” in Network and
Distributed System Security Symposium, 2009.
[51]
G. Wicherski, “peHash: A Novel Approach to Fast Malware Clustering,” in
Conference on Large-scale Exploits and Emergent Threats: Botnets,
Spyware, Worms, and More. USENIX
Association, 2009.
[52]
J. Kornblum, “Identifying Almost Identical Files Using Context Triggered
Piecewise Hashing,” Digital Investigation, vol. 3, pp. 91–97, Sep.
2006.
[53]
J. Oliver, C. Cheng, and Y. Chen, “TLSH – A Locality Sensitive Hash,” in
Cybercrime and Trustworthy Computing Workshop. IEEE Computer Society, 2013.
[54]
J. Newsome, B. Karp, and D. Song, “Polygraph: Automatically generating
signatures for polymorphic worms,” in IEEE Symposium on Security and
Privacy, 2005.
[55]
Z. Li, M. Sanghi, Y. Chen, M.-Y. Kao, and B. Chavez, “Hamsa: Fast signature
generation for zero-day polymorphic worms with provable attack resilience,”
in IEEE Symposium on Security and Privacy, 2006.
[56]
A. Sæbjørnsen, J. Willcock, T. Panas, D. Quinlan, and Z. Su,
“Detecting Code Clones in Binary Executables,” in Symposium on
Software Testing and Analysis. ACM,
2009.
[57]
I. Santos, F. Brezo, J. Nieves, Y. K. Penya, B. Sanz, C. Laorden, and P. G.
Bringas, “Idea: Opcode-Sequence-Based Malware Detection,” in
Symposium on Engineering Secure Software and Systems. Springer Berlin Heidelberg, 2010.
[58]
B. Kang, T. Kim, H. Kwon, Y. Choi, and E. G. Im, “Malware Classification
Method via Binary Content Comparison,” in Research in Applied
Computation Symposium. ACM, 2012.
[59]
J. Ming, M. Pan, and D. Gao, “iBinHunt: Binary Hunting with Inter-procedural
Control Flow,” in International Conference on Information Security
and Cryptology. Springer-Verlag,
2013.
[60]
W. Jin, S. Chaki, C. Cohen, A. Gurfinkel, J. Havrilla, C. Hines, and
P. Narasimhan, “Binary function clustering using semantic hashes,” in
International Conference on Machine Learning and Applications. IEEE, 2012.
[61]
M. Bourquin, A. King, and E. Robbins, “BinSlayer: Accurate Comparison of
Binary Executables,” in Program Protection and Reverse Engineering
Workshop. ACM, 2013.
[62]
W. M. Khoo, A. Mycroft, and R. Anderson, “Rendezvous: a Search Engine for
Binary Code,” in Working Conference on Mining Software Repositories,
2013.
[63]
B. H. Ng and A. Prakash, “Expose: Discovering potential binary code
re-use,” in Annual Computer Software and Applications
Conference. IEEE, 2013.
[64]
Y. R. Lee, B. Kang, and E. G. Im, “Function Matching-based Binary-level
Software Similarity Calculation,” in Research in Adaptive and
Convergent Systems. ACM, 2013.
[65]
M. R. Farhadi, B. C. M. Fung, P. Charland, and M. Debbabi, “BinClone:
Detecting Code Clones in Malware,” in International Conference on
Software Security and Reliability. IEEE Computer Society, 2014.
[66]
B. Ruttenberg, C. Miles, L. Kellogg, V. Notani, M. Howard, C. LeDoux,
A. Lakhotia, and A. Pfeffer, “Identifying Shared Software Components to
Support Malware Forensics,” in Detection of Intrusions and Malware,
and Vulnerability Assessment. Springer International Publishing, 2014.
[67]
M. Egele, M. Woo, P. Chapman, and D. Brumley, “Blanket Execution: Dynamic
Similarity Testing for Program Binaries and Components,” in 23rd
USENIX Security Symposium, 2014.
[68]
L. Luo, J. Ming, D. Wu, P. Liu, and S. Zhu, “Semantics-based
obfuscation-resilient binary code similarity comparison with applications to
software and algorithm plagiarism detection,” IEEE Transactions on
Software Engineering, no. 12, pp. 1157–1177, 2017.
[69]
S. Alrabaee, P. Shirani, L. Wang, and M. Debbabi, “Sigma: A semantic
integrated graph matching approach for identifying reused functions in binary
code,” 2015.
[70]
J. Qiu, X. Su, and P. Ma, “Library functions identification in binary code by
using graph isomorphism testings,” in Conference on Software
Analysis, Evolution and Reengineering, 2015.
[71]
Y. Xiao, S. Cao, Z. Cao, F. Wang, F. Lin, J. Wu, and H. Bi, “Matching Similar
Functions in Different Versions of a Malware,” in
Trustcom/BigDataSE/ISPA, 2016 IEEE. IEEE, 2016.
[72]
N. Lageman, E. D. Kilmer, R. J. Walls, and P. D. McDaniel, “BinDNN: Resilient
Function Matching Using Deep Learning,” in International Conference
on Security and Privacy in Communication Systems. Springer, 2016.
[73]
S. H. Ding, B. C. Fung, and P. Charland, “Kam1N0: MapReduce-based Assembly
Clone Search for Reverse Engineering,” in International Conference on
Knowledge Discovery and Data Mining. ACM, 2016.
[74]
Y. Hu, Y. Zhang, J. Li, and D. Gu, “Binary Code Clone Detection Across
Architectures and Compiling Configurations,” in International
Conference on Program Comprehension. IEEE Press, 2017.
[75]
L. Nouh, A. Rahimian, D. Mouheb, M. Debbabi, and A. Hanna, BinSign:
Fingerprinting Binary Functions to Support Automated Analysis of Code
Executables. Springer International
Publishing, 2017.
[76]
P. Shirani, L. Wang, and M. Debbabi, “BinShape: Scalable and Robust Binary
Library Function Identification Using Function Shape,” in Detection
of Intrusions and Malware, and Vulnerability Assessment. Springer International Publishing, 2017.
[77]
J. Ming, D. Xu, Y. Jiang, and D. Wu, “BinSim: Trace-based Semantic Binary
Diffing via System Call Sliced Segment Equivalence Checking,” in 26th
USENIX Security Symposium, 2017.
[78]
S. Wang and D. Wu, “In-memory Fuzzing for Binary Code Similarity Analysis,”
in International Conference on Automated Software Engineering. IEEE Press, 2017.
[79]
S. Alrabaee, P. Shirani, L. Wang, and M. Debbabi, “FOSSIL: A Resilient and
Efficient System for Identifying FOSS Functions in Malware Binaries,”
IEEE Transactions on Privacy and Security, vol. 21, no. 2, pp.
8:1–8:34, Jan. 2018.
[80]
K. Redmond, L. Luo, and Q. Zeng, “A Cross-Architecture Instruction Embedding
Model for Natural Language Processing-Inspired Binary Code Analysis,” in
Workshop on Binary Analysis Research, San Diego, CA, USA, 2019.
[81]
F. Zuo, X. Li, Z. Zhang, P. Young, L. Luo, and Q. Zeng, “Neural Machine
Translation Inspired Binary Code Similarity Comparison beyond Function
Pairs,” in Network and Distributed System Security Symposium, 2019.
[82]
S. H. H. Ding, B. C. M. Fung, and P. Charland, “Asm2vec: Boosting static
representation robustness for binary clone search against code obfuscation
and compiler optimization,” in IEEE Symposium on Security and
Privacy. San Francisco, CA: IEEE
Computer Society, 2019.
[83]
L. Massarelli, G. A. Di Luna, F. Petroni, R. Baldoni, and L. Querzoni, “SAFE:
Self-Attentive Function Embeddings for Binary Similarity,” in
Detection of Intrusions and Malware, and Vulnerability Assessment,
2019.
[84]
“BinDiff,” 2018, https://www.zynamics.com/bindiff.html.
[85]
“Diaphora,” 2018, http://diaphora.re/.
[86]
“DarunGrim,” 2018, http://www.darungrim.org/.
[87]
R. Perdisci, W. Lee, and N. Feamster, “Behavioral Clustering of HTTP-Based
Malware and Signature Generation Using Malicious Network Traces,” in
USENIX Symposium on Networked Systems Design and Implementation, 2010.
[88]
M. Z. Rafique and J. Caballero, “FIRMA: Malware Clustering and Network
Signature Generation with Mixed Network Behaviors,” in Symposium on
Research in Attacks, Intrusions and Defenses, St. Lucia, October 2013.
[89]
I. U. Haq, S. Chica, J. Caballero, and S. Jha, “Malware Lineage in the
Wild,” Elsevier Computer & Security, vol. 78, pp. 347–363, 2018.
[90]
Brumley, David and Poosankam, Pongsin and Song, Dawn and Zheng, Jiang,
“Automatic patch-based exploit generation is possible: Techniques and
implications,” in IEEE Symposium on Security and Privacy, 2008.
[91]
Z. Wang, K. Pierce, and S. McFarling, “BMAT – A Binary Matching Tool for
Stale Profile Propagation,” The Journal of Instruction-Level
Parallelism, vol. 2, p. 2000, 2000.
[92]
G. Myles and C. Collberg, “K-gram Based Software Birthmarks,” in
Symposium on Applied Computing. ACM, 2005.
[93]
S. Choi, H. Park, H.-I. Lim, and T. Han, “A Static Birthmark of Binary
Executables Based on API Call Structure,” in Asian Computing Science
Conference on Advances in Computer Science: Computer and Network
Security. Springer-Verlag, 2007.
[94]
M. J. Rochkind, “The Source Code Control System,” IEEE Transactions
on Software Engineering, no. 4, pp. 364–370, 1975.
[95]
W. F. Tichy, “RCS - A System for Version Control,” Software: Practice
and Experience, vol. 15, no. 7, pp. 637–654, 1985.
[96]
C. Reichenberger, “Delta Storage for Arbitrary Non-Text Files,” in
Proceedings of the 3rd ACM International Workshop on Software
Configuration Management, 1991.
[97]
“RTPatch,” 1991, http://pocketsoft.com/rtpatch_binary_diff.html.
[98]
K. Coppieters, “A Cross-platform Binary Diff,” 1995.
[99]
“Xdelta3,” 1998, http://xdelta.org/.
[100]
J. Jang, A. Agrawal, and D. Brumley, “ReDeBug: Finding Unpatched Code Clones
in Entire OS Distributions,” in IEEE Symposium on Security and
Privacy, 2012.
[101]
V. A. Zakharov, “The Equivalence Problem for Computational Models: Decidable
and Undecidable Cases,” in International Conference on Machines,
Computations, and Universality, 2001.
[102]
L. Jiang and Z. Su, “Automatic Mining of Functionally Equivalent Code
Fragments via Random Testing,” in Symposium on Software Testing and
Analysis. ACM, 2009.
[103]
V. Ganesh and D. L. Dill, “A Decision Procedure for Bit-vectors and
Arrays,” in Conference on Computer Aided Verification. Springer-Verlag, 2007.
[104]
L. De Moura and N. Bjørner, “Z3: An Efficient SMT Solver,” in
Conference on Tools and Algorithms for the Construction and Analysis of
Systems. Springer-Verlag, 2008.
[105]
C. B. Zilles, “Master/slave Speculative Parallelization And Approximate
Code,” Univerisity of Wisconsin, USA, Tech. Rep., 2002.
[106]
A. Z. Broder, M. Charikar, A. M. Frieze, and M. Mitzenmacher, “Min-wise
independent permutations,” Journal of Computer and System Sciences,
vol. 60, no. 3, pp. 630–659, 2000.
[107]
A. Gionis, P. Indyk, R. Motwani et al., “Similarity search in high
dimensions via hashing,” in International Conference on Very Large
Data Bases (VLDB), 1999.
[108]
“Virustotal,” https://virustotal.com/.
[109]
V. Roussev, “Data fingerprinting with similarity digests,” in IFIP
International Conference on Digital Forensics. Springer, 2010.
[110]
F. Breitinger and H. Baier, “Similarity preserving hashing: Eligible
properties and a new algorithm mrsh-v2,” in Conference on Digital
Forensics and Cyber Crime. Springer,
2012.
[111]
F. Pagani, M. Dell’Amico, and D. Balzarotti, “Beyond Precision and Recall:
Understanding Uses (and Misuses) of Similarity Hashes in Binary Analysis,”
in Conference on Data and Application Security and Privacy. ACM, March 2018.
[112]
C. Linn and S. Debray, “Obfuscation of Executable Code to Improve Resistance
to Static Disassembly,” in ACM Conference on Computer and
Communications Security. ACM, 2003.
[113]
“ReadyNAS2014 Firmware Image v6.1.6,” 2019,
http://www.downloads.netgear.com/files/GDC/READYNAS-100/ReadyNASOS-6.1.6-arm.zip.
[114]
“DD-WRT2013 Firmware Image r21676,” 2019,
ftp://ftp.dd-wrt.com/betas/2013/05-27-2013-r21676/senao-eoc5610/linux.bin.
[115]
Z. Wang, X. Jiang, W. Cui, X. Wang, and M. Grace, “ReFormat: Automatic
Reverse Engineering of Encrypted Messages,” in European Symposium on
Research in Computer Security, 2009.
[116]
D. Xu, J. Ming, and D. Wu, “Cryptographic Function Detection in Obfuscated
Binaries via Bit-Precise Symbolic Loop Mapping,” in IEEE Symposium on
Security and Privacy, 2017.
[117]
V. Karande, S. Chandra, Z. Lin, J. Caballero, L. Khan, and K. Hamlen, “BCD:
Decomposing Binary Code Into Components Using Graph-Based Clustering,” in
ACM ASIA Conference on Information, Computer and Communications
Security, 2018.
[118]
“Gartner says 8.4 billion connected things will be in use in 2017, up 31
percent from 2016,” 2017,
https://www.gartner.com/newsroom/id/3598917.
[119]
“Themida,” 2018, https://www.oreans.com/themida.php/.
[120]
“VMProtect,” 2018, http://vmpsoft.com/products/vmprotect/.
[121]
“Angr: The Binary Analysis Framework,” 2016, http://angr.io/.
[122]
D. Brumley, I. Jager, T. Avgerinos, and E. J. Schwartz, “BAP: A binary
analysis platform,” in Conference on Computer Aided Verification,
2011.
[123]
“BeaEngine,” 2018, http://www.beaengine.org/.
[124]
“Bitblaze: Binary Analysis For Computer Security,” 2008,
http://bitblaze.cs.berkeley.edu/.
[125]
“Boomerang decompiler,” 2004, http://boomerang.sourceforge.net/.
[126]
“Automated Malware Analysis,” 2012, https://cuckoosandbox.org/.
[127]
“Dyninst: Putting the Performance in High Performance Computing,” 2009,
http://www.dyninst.org/.
[128]
“IDA,” 2018, https://www.hex-rays.com/products/ida/.
[129]
“The LLVM Compiler Infrastructure,” 2004, http://llvm.org/.
[130]
“IDA Processor Support,” 2019, http://www.fysnet.net/newbasic.htm.
[131]
“McSema,” 2012, https://github.com/trailofbits/mcsema/.
[132]
“Pin - A Dynamic Binary Instrumentation Tool,” 2005,
https://software.intel.com/en-us/articles/pin-a-dynamic-binary-instrumentation-tool.
[133]
“QEMU: an open source processor emulator,” 2006,
http://www.qemu.org/.
[134]
“ROSE compiler infrastructure,” 2000, http://www.rosecompiler.org/.
[135]
“Valgrind,” 2007, http://valgrind.org/.
[136]
A. Srivastava, A. Edwards, and H. Vo, “Vulcan: Binary Transformation In A
Distributed Environment,” Tech. Rep., 2001.
Appendix
Implementation platforms.
Table V shows
the intermediate representation (IR) used by the platform,
whether it supports static and dynamic analysis,
the target architectures and operating systems it supports, and
how it is released (open source or free binary).
Among the 17 platforms, 12 support static analysis and
7 dynamic analysis.
The functionality provided by static analysis platforms widely varies.
Ida is a disassembler,
Boomerang is a decompiler, and the rest
offer diverse static analysis functionality such as disassembly,
building control flow graphs and call graphs,
IR simplifications, and data flow propagation.
All dynamic analysis platforms can run unmodified binaries
(i.e., no need to recompile or relink).
QEMU is a whole system emulator that can run a full guest OS (e.g., Windows)
on a potentially different host OS (e.g., Linux).
Dyninst, PIN, and Valgrind execute an unmodified target program with
customizable instrumentation (e.g., through binary rewriting)
on the CPU of the host system.
BAP and BitBlaze build their dynamic analysis components on top of PIN and QEMU. |
On the evaluation of the ac-Stark shift and Kramers-Heisenberg dispersion formula
R. Radhakrishnan
[email protected]
Department of Theoretical Physics,
Guindy Campus, University of Madras, Chennai-25, India.
Ramesh Babu Thayyullathil
Department of physics, Cochin university of Science and
Technology, Cochin 682022, India.
(November 21, 2020)
Abstract
We describe a unified approach for the determination of ac-Stark
correction and Kramers-Heisenberg dispersion formula. In both cases
the contribution from infinite intermediate summation appearing in
the expression for the corresponding matrix elements are evaluated
exactly in the dipole approximation for the ground state of hydrogen
atom using a variation of the Dalgarno-Lewis method. The analytical
expressions obtained can be efficiently used for the numerical
evaluation of matrix element for all values of incident photon
energy and comparison is made with results obtained by different
methods.
pacs: 32.80.Rm, 42.50.Hz
I Introduction
In the presence of intense laser field, atomic or molecular levels
are displaced (shifted or broadened) and these stimulated radiative
corrections are alternatively known as ac-Stark shift, dynamic Stark
effect or light shift cohen . Being a fundamental process
ac-stark effect is very well studied theoretically as well as
experimentally in wide area of atomic and molecular physics. This
effect is also very important in the area of laser trapping and
cooling and is a basic mechanism behind many nonlinear optical
effects bec . It is a promising tool for the field of optical
communications (Optical switching methods) and plays a very
decisive role in many current research areas
dyn ; schu ; glib ; shin ; zou ; sch ; rud ; bach ; apa ; ago ; chen ; yak . In this
work we have described an efficient method to obtain a closed form
analytical expression for two of the basic process viz. ac stark
shift and elastic scattering cross section of photons for the atomic
hydrogen in the ground state.
Our aim is to get an analytical expression for dipole dynamical
polarizability $\tau^{(2)}(\omega)$ from which one can calculate
the level shift $\delta E^{(2)}$ and level width $\Gamma^{(2)}$.
The hydrogen atom being the simplest of all atomic systems, plays
a very special role in this respect because it has closed-form
wavefunctions. Since these wave functions are analytically known,
both the dipole dynamic polarizability and Kramer’s-Heisenberg
dispersion formula for hydrogen atom can be written in a closed form.
The level shift for atomic hydrogen was previously
calculated by both perturbative and nonperturbative methods
pan ; maq:83 ; kar:04 ; bak ; chu:coo ; gon:89 ; maq:90 ; sha:tan ; ben:kri ; flo:pat .
In the perturbative method, evaluation of the higher order
matrix element is the key problem and it is achieved through
Greens function formalism pan ; maq:83 ; kar:04 or by
solving a system of second-order differential equations
gon:89 .
The Kramers-Heisenberg formula alternatively known as
dispersion formula plays a very important role in the determination of
scattering cross-sections hei .
Considering its importance as an age-old relation
there were only very few attempts to evaluate it
analytically mit ; gav . There were some generalization
of it and alternate forms are also available in the
literature hea:woo ; hol ; bre ; tul .
Recently a low-energy expansion of the Kramers-Heisenberg
formula for Rayleigh scattering process in atomic hydrogen was reported lee .
Similar to the ac Stark shift this also involves the
evaluation of the second order matrix element.
The exact calculation of higher order process
in the perturbative formulation is nontrivial because of the presence
of infinite summation over intermediate states in the expression
for the higher order matrix element. In our approach the matrix element
containing infinite summation over the whole hydrogenic
spectrum (discrete and continuum) is performed by using an
implicit summation technique due to Dalgarno Lewis dar:lew ,
which reduces the evaluation of the infinite summation
to finding a solution of some inhomogeneous differential
equation. The closed form expression which we have obtained
is very simple and also very convenient for analytical
continuation. Thus with this method, we can very easily obtain
the relevant matrix element for radiation with energy larger
than ionization energy (above threshold ionization),
while other methods need some kind of
approximation like Pade approximation in the
case of Coulomb Greens function formalism.
II Radiative correction in the ground state
The level shift depends on the intensity $I$ and the frequency $\omega$ of the radiation and the complex second order shift
$\Delta^{(2)}$ is given by maq:83 ; kar:04 ; chu:coo ; gon:89
$$\Delta^{(2)}(\omega)=\delta E^{(2)}-i\Gamma^{(2)}=-\frac{I}{I_{0}}\tau^{(2)}(\omega)$$
(1)
where the real part $\delta E^{(2)}$ is the energy shift and the
imaginary part $\Gamma^{(2)}$ gives the level width and
$I_{0}=7.016\times 10^{16}$ W/cm${}^{2}$ is the characteristic atomic field
strength intensity.
Here $\tau^{(2)}(\omega)$ is the dipole dynamic polarizability
and in the atomic unit it can be written as
$$\tau^{(2)}=\sum_{n}\left\{\frac{\langle g\mid\epsilon^{*}\cdot{\bf r}\mid n%
\rangle\langle n\mid\epsilon\cdot{\bf r}\mid g\rangle}{(E_{g}-E_{n}+\omega)}+%
\frac{\langle g\mid\epsilon\cdot{\bf r}\mid n\rangle\langle n\mid\epsilon^{*}%
\cdot{\bf r}\mid g\rangle}{(E_{g}-E_{n}-\omega)}\right\}$$
(2)
where $E_{g}$ is the atomic ground state energy
and $\epsilon$ is
the polarization of the radiation. First term in the bracket is
the absorption-emission term and the second is the
emission-absorption term. The infinite summation over the complete
set of intermediate state in equation (2) can be performed
exactly by defining a set of operators $F$ and $\widetilde{F}$
such that
$$\mathbf{\epsilon}\cdot{\bf r}\left|g\right\rangle=\left(FH_{0}-H_{0}F+\omega F%
\right)\left|g\right\rangle$$
(3)
$$\mathbf{\epsilon}^{*}\cdot{\bf r}\left|g\right\rangle=\left(\widetilde{F}H_{0}%
-H_{0}\widetilde{F}-\omega\widetilde{F}\right)\left|g\right\rangle$$
(4)
where $H_{0}=-\nabla^{2}/2-1/r$ is the unperturbed Hamiltonian for
atomic hydrogen and $|g\rangle=e^{-r}/\sqrt{\pi}$ is the ground
state wave function in atomic units. With these definitions and the closure relation
$\sum_{n}\mid n\rangle\langle n\mid\ =\hat{I}$ the expression in
equation (2) for dipole dynamical polarizability will be reduced to
$$\tau^{(2)}(\omega)=\langle g\mid\mathbf{\epsilon}^{*}\cdot{\bf r}F\mid g%
\rangle+\langle g\mid\mathbf{\epsilon}\cdot{\bf r}\widetilde{F}\mid g\rangle.$$
(5)
Thus the infinite summation over the intermediate states are
reduced to the determination of the operator $F$ and
$\widetilde{F}$. Using co-ordinate space representation and writing
$F=\mathbf{\epsilon}\cdot\mathbf{r}f(r)$ and
$\widetilde{F}=\mathbf{\epsilon}^{*}\cdot\mathbf{r}\widetilde{f}(r)$,
the equations (3) and (4) become
$$\displaystyle r\frac{d^{2}}{dr^{2}}f(r)+(4-2r)\frac{d}{dr}f(r)+(2\omega r-2)f(%
r)=2r$$
(6)
$$\displaystyle r\frac{d^{2}}{dr^{2}}\widetilde{f}(r)+(4-2r)\frac{d}{dr}%
\widetilde{f}(r)-(2\omega r+2)\widetilde{f}(r)=2r$$
(7)
By the method of Laplace transform inc for the
solution of differential equation we can obtain the solutions to
the above differential equations as
$$f(r)=\frac{1}{\omega}-\frac{1}{2\omega^{3}}\Phi(1,1,\lambda,r)$$
(8)
and
$$\widetilde{f}(r)=\frac{1}{\omega}+\frac{1}{2\omega^{3}}\widetilde{\Phi}(1,1,%
\widetilde{\lambda},r)$$
(9)
with
$$\Phi(p,q,\lambda,r)=\int\limits_{\lambda}^{1}ds\,e^{-r(s-1)}K(p,q,\lambda,s)$$
(10)
$$\widetilde{\Phi}(p,q,\lambda,r)=\int\limits_{1}^{\lambda}ds\,e^{-r(s-1)}%
\widetilde{K}(p,q,\lambda,s)$$
(11)
$$K(p,q,\lambda,s)=\left(\frac{1-\lambda}{1+\lambda}\right)^{\frac{1}{\lambda}}%
\left(s+\lambda\right)^{p+\frac{1}{\lambda}}\left(s-\lambda\right)^{q-\frac{1}%
{\lambda}}$$
(12)
$$\widetilde{K}(p,q,\lambda,s)=\left(\frac{\lambda-1}{\lambda+1}\right)^{\frac{1%
}{\lambda}}\left(\lambda+s\right)^{p+\frac{1}{\lambda}}\left(\lambda-s\right)^%
{q-\frac{1}{\lambda}}$$
(13)
where $\lambda=\sqrt{1-2\omega}$ and
$\widetilde{\lambda}=\sqrt{1+2\omega}$. In this work we obtain the solutions
with integer values of $p$ and $q$. But in general for the purpose of analytic
continuation it can be complex and $\lambda$ also can become complex depending
on the value of the frequency $\omega$.
Now using equations (10) and (11)
in equation (5) the final form of the dipolar polarizability
becomes
$$\tau^{(2)}(\omega)=\frac{2}{3\omega^{3}}\int\limits_{0}^{\infty}dr\,e^{-2r}\,r%
^{4}\{\widetilde{\Phi}(1,1,\widetilde{\lambda},r)-\Phi(1,1,\lambda,r)\}$$
(14)
The limit $\omega\rightarrow 0$ of the expression for $\tau^{(2)}(\omega)$
can be shown to approach the value corresponding to the static dipolar
polarizability which is $9/2$.
III Elastic scattering of photons
The differential scattering cross section for low energy
elastic scattering of photons with frequency $\omega$,
by atoms is given by the dispersion formula hei ; mit ; gav
$$\frac{d\sigma}{d\Omega}=a_{0}^{2}\left(\epsilon\cdot\epsilon^{\prime}\right)^{%
2}\mid M(\omega)\mid^{2}=a_{0}^{2}\left(\epsilon\cdot\epsilon^{\prime}\right)^%
{2}\mid 1-P(\omega)-P(-\omega)\mid^{2}~{}$$
(15)
where $a_{0}$ is the Bohr radius, $M$ is the Kramers-Heisenberg
matrix element, $\epsilon$ and $\epsilon^{\prime}$ respectively are the
initial and final polarization of photons and $P(\omega)$ in atomic
units is given as gav
$$P(\omega)=-\frac{2}{3}\sum_{n}\frac{\langle g\mid\mathbf{p}\mid n\rangle\cdot%
\langle n\mid\mathbf{p}\mid g\rangle}{E_{g}-E_{n}+\omega}.$$
(16)
Here $\mathbf{p}$ is the momentum
operator and the summation is over the complete set of states including continuum
states. It is useful to note from equation (15) that the
differential cross section for coherent scattering of photons is
just the Thompson cross section modified by the dynamic polarizability.
In a similar fashion we can consider Raman scattering where
initial and final states are different. The analytical expressions
for $P(\omega)$ were derived earlier using Schwartz and Teinman method
mit and Coulomb Greens function (CGF) formalism gav .
Using a slight variation of the formalism in the previous section
we can easily calculate $P(\omega)$. In this case the infinite
summation over the intermediate state in equation (16) is
performed by defining a set of operators $\mathbf{U}$ and
${\widetilde{\mathbf{U}}}$ such that,
$$\displaystyle\mathbf{p}\left|g\right\rangle=\left(\mathbf{U}H_{0}-H_{0}\mathbf%
{U}+\omega\mathbf{U}\right)\left|g\right\rangle$$
(17)
$$\displaystyle\mathbf{p}\left|g\right\rangle=\left(\widetilde{\mathbf{U}}H_{0}-%
H_{0}\widetilde{\mathbf{U}}-\omega\mathbf{\widetilde{U}}\right)\left|g\right\rangle$$
(18)
and the expression for $M$ will become
$$M(\omega)=1+\frac{2}{3}\left(\langle g\mid\mathbf{p}\cdot\mathbf{U}\mid g%
\rangle+\langle g\mid\mathbf{p}\cdot\mathbf{\widetilde{U}}\mid g\rangle\right)$$
(19)
Now the evaluation of the infinite summation is reduced to the
evaluation of the operators $\mathbf{U}$ and
$\widetilde{\mathbf{U}}$. To obtain a similar expression as in
the previous section, instead of going to momentum space
representation mit ; gav , we use the coordinate space
representation of $\mathbf{U}$ and $\widetilde{\mathbf{U}}$.
This is done by taking
$\mathbf{U}=\mathbf{r}u(r)$ and
$\widetilde{\mathbf{U}}=\mathbf{r}\widetilde{u}(r)$ and the
equations (17) and (18) become
$$\displaystyle r\frac{d^{2}}{dr^{2}}u(r)+(4-2r)\frac{d}{dr}u(r)+(2\omega r-2)u(%
r)=2i$$
(20)
$$\displaystyle r\frac{d^{2}}{dr^{2}}\widetilde{u}(r)+(4-2r)\frac{d}{dr}%
\widetilde{u}(r)-(2\omega r+2)\widetilde{u}(r)=2i$$
(21)
These differential equations has the same form which appeared in the
previous section and the solutions can be written as
$$u(r)=-\frac{i}{2\omega^{2}}\Phi(1,1,\lambda,r)$$
(22)
and
$$\widetilde{u}(r)=\frac{i}{2\omega^{2}}\widetilde{\Phi}(1,1,\widetilde{\lambda}%
,r)$$
(23)
Using this in the definition of $U$ and $\widetilde{U}$ and substituting it
in equation (19)
we get a closed form expression for the Kramers-Heisenberg matrix
element. It now takes the form
$$M(\omega)=1-\frac{4}{3\omega^{2}}\int\limits_{0}^{\infty}dr\,e^{-2r}\,r^{3}%
\left\{\Phi(1,1,\lambda,r)-\widetilde{\Phi}(1,1,\widetilde{\lambda},r)\right\}$$
(24)
If the incident photons are unpolarized and the polarization of
the scattered photons are not observed the differential scattering
cross section will take the standard form
$$d\sigma=r_{0}^{2}\frac{1}{2}(1+cos^{2}\theta)|M(\omega)|^{2}d\Omega$$
(25)
IV Discussion and Conclusion
The radial integrals in equations (14) and (24) can be
done exactly isg . For the numerical evaluation of various
integrals it is very convenient to define the following
$$\displaystyle I(p,q,\lambda,n)$$
$$\displaystyle=$$
$$\displaystyle\int\limits_{0}^{\infty}dr\,e^{-2r}r^{n}\,\Phi(p,q,\lambda,r)$$
$$\displaystyle=$$
$$\displaystyle n!\left(\frac{1-\lambda}{1+\lambda}\right)^{\frac{1}{\lambda}}%
\int\limits_{\lambda}^{1}ds\frac{\left(s+\lambda\right)^{p+\frac{1}{\lambda}}%
\left(s-\lambda\right)^{q-\frac{1}{\lambda}}}{(1+s)^{n+1}}$$
and
$$\displaystyle\widetilde{I}(p,q,\lambda,n)$$
$$\displaystyle=$$
$$\displaystyle\int\limits_{0}^{\infty}dr\,e^{-2r}r^{n}\,\widetilde{\Phi}(p,q,%
\lambda,r)$$
$$\displaystyle=$$
$$\displaystyle n!\left(\frac{\lambda-1}{\lambda+1}\right)^{\frac{1}{\lambda}}%
\int\limits_{1}^{\lambda}ds\frac{\left(\lambda+s\right)^{p+\frac{1}{\lambda}}%
\left(\lambda-s\right)^{q-\frac{1}{\lambda}}}{(1+s)^{n+1}}.$$
With these, the Kramers-Heisenberg matrix element and the dipole
dynamic polarizability respectively will take the following forms
$$M(\omega)=1-\frac{4}{3\omega^{2}}\left\{I(1,1,\lambda,3)-\widetilde{I}(1,1,%
\widetilde{\lambda},3)\right\}$$
(28)
$$\tau^{(2)}(\omega)=\frac{2}{3\omega^{3}}\left\{\widetilde{I}(1,1,\widetilde{%
\lambda},4)-I(1,1,\lambda,4)\right\}$$
(29)
This is the final form of the analytic expression for the matrix elements.
We did not attempt to rewrite these in terms of well known special
functions because the remaining integrals in equations (IV) and (IV)
can be done very efficiently by a simple numerical integration.
The integrals in
equation (10) is defined only for Re$(q-1/\lambda)>-1$. But in order to
calculate the matrix elements in equations (28) and (29) for
all values of $\omega$, we require an analytical continuation of the
definition of $\Phi(.)$ given in equation (10).
It can be achieved by performing a partial integration in the definition given
in equation (10) and this gives
$$\displaystyle\Phi(p,q,\lambda,r)$$
$$\displaystyle=$$
$$\displaystyle\frac{1}{q+1-\frac{1}{\lambda}}\left[K(p,q+1,\lambda)+r\Phi(p,q+1%
,\lambda,r)\right.$$
$$\displaystyle\left.+\left(p+\frac{1}{\lambda}\right)\Phi(p-1,q+1,\lambda,r)%
\right].$$
This is a very important recurrence relation because (IV) can be used to analytically
continue the definition of these function for Re$(q-1/\lambda)>-2$.
We can also repeat this procedure to analytically continue the definition
to other values of Re$(q-1/\lambda)$. And for the
numerical evaluation of various integrals as a function of
radiation frequency $\omega$ we have extensively used this
relation. It is also very important to note the appearance of
simple poles in the definition of $\Phi(.)$ when $q=-1+1/\lambda$.
Since $\lambda=\sqrt{1-2\omega}$, this is due to the presence of
intermediate resonances (when $2\omega=1-1/n^{2}$, where
$n=2,3,\ldots$) in the expression for second order matrix element
in the perturbation theory. This also leads to a strong frequency
dependence of ac stark effect and scattering cross section.
Similarly for above one photon ionization threshold i.e. $\omega>1/2$, a simple analytic continuation makes $\lambda$
to be a purely imaginary number. Thus by this simple method, with the same
analytic expression for the matrix element, we can numerically
evaluate it for the whole physical range of photon frequency $\omega$.
Displayed in table 1 are
$\tau^{(2)}(\omega)$ as a function of photon frequency in atomic units and are
compared with values obtained by other methods.
It is purely real for $\omega<1/2$.
For $\omega>1/2$ one photon ionization is possible and this makes the
intermediate virtual state to lie in the continuum and $\tau^{(2)}(\omega)$
becomes complex.
The dipole dynamic polarizability for photon
energy below and above one photon ionization threshold is given in table 1. Our values are compared
with values obtained by Arnous et al. maq:83 using Coulomb Greens function
method, but with an additional factor of $\omega^{-2}$
in their result. It is very easy to see that this factor is
missing from their result by taking the $\omega\rightarrow 0$ limit.
In this limit our result approaches the dc polarizability, which is 9/2.
The values given in maq:83 also approaches to the same limit
provided their results are multiplied by an overall factor of $\omega^{-2}$.
Using the analytic continuation described in equation (IV) we can calculate
$\tau^{(2)}(\omega)$ for $\omega$ very close to intermediate resonance also.
Our result also agrees with the second order
level shift and width reported by Pan et al. pan .
From the values given in tables 1 and 2
it is useful to note the change
in sign of the level shift when the photon frequency cross these resonance
values. The values of Kramers-Heisenberg matrix element are given in tables
2 and 3. They are in good agreement
with the results of Gavrila gav .
Acknowledgements.Authors acknowledge the support from UGC
through DSA-COSIST Scheme
References
References
(1)
C. Cohen-Tannoudji, J. Dupont-Roc and G. Grynberg
Atom Photon Interactions 1992 (New York: Wiley)
(2)
S. A. Moskalenko and D. W. Snoke 2000
Bose-Einstein Condensation of Excitons and Biexcitons
(Cambridge University Press) pp 201
(3)
J. F. Dynes, M. D. Frogley, M. Beck, J. Faist, and C. C. Phillips
Phys. Rev. Lett. 94 157403(2005)
(4)
D. I. Schuster et al., Phys. Rev. Lett. 94 123602 (2005)
(5)
S. Gilb et al., Phys. Rev. A71 042709 (2005)
(6)
Y. Shin et al., Phys. Rev. Lett. 92 050405 (2004)
(7)
Zhou Shu-Yu, Xu Zhen, Zhou Shan-Yu and Wang Yu-Zhu
Chinese Phys. Lett. 22 1672 (2005)
(8)
T. W. Schmidt, R. B. López-Martens and G. Roberts
J. Phys. B: At. Mol. Opt. Phys. 37 1125 (2004)
and the reference there in
(9)
T. G. Rudolph, H. S. Freedhoff and Z. Ficek
Phys. Rev. A58 1296 (1998)
(10)
H. Bachau
J. Phys. B: At. Mol. Opt. Phys. 35 509 (2002)
(11)
A. Apalategui, A. Saenz and P. Lambropoulos
J. Phys. B: At. Mol. Opt. Phys. 33 2791 (2000)
(12)
P. Agostini et al.
J. Phys. B: At. Mol. Opt. Phys. 22 1971 (1989)
(13)
Chen Jing-Biao, Wang Feng-Zhi, Yang Dong-Hai and Wang Yi-Qiu
Chinese Phys. Lett. 18 202(2001)
(14)
V. Yakhontov, R. Santra and K. Jungmann
J. Phys. B: At. Mol. Opt. Phys. 32 1615(1999)
(15)
J. S. Bakos
Phys. Reports 31 209(1977)
(16)
L. Pan, K. T. Taylor and C. W. Clark
Phys. Rev. Lett. 61 2673, 1991 Phys. Rev. A43 6272(1988)
(17)
E. Arnous, J. Bastian and A. Maquet
Phys. Rev. A27 977(1983)
(18)
E. Karule and A. Gailitis
J. Phys. B: At. Mol. Opt. Phys. 21 829(2004)
(19)
S. Chu and J. Cooper
Phys. Rev. A32 2769(1985)
(20)
Y. Gontier and M. Trahin
Phys. Rev. A40 1351(1989)
(21)
Y. Justum, A. Maquet and Y. Heno
Phys. Rev. A41 2791(1990)
(22)
R. Shakeshaft and X. Tang
Phys. Rev. A36 3193(1987)
(23)
J. D. Benkenstein and J. B. Krieger
jmp 11 2721(1970)
(24)
V. Florescu and S. Patrascu
JMP 31 2165(1990)
(25)
W. Heitler
The Quantum Theory of Radiation
(New York:Oxford University(1954)) p 192
(26)
M. H. Mittleman and Fread A Wolf
Phys. Rev. A128 2686(1962)
(27)
M. Gavrila
Phys. Rev. A163 147(1967)
(28)
M. Gavrila and A. Costescu
Phys. Rev. A2 1752(1970)
(29)
W. P. Healy Phys. Rev. A16 1568(1977),
W. P. Healy and R. G. Woolley
J. Phys. B: At. Mol. Opt. Phys. 11, 1131(1978)
(30)
B. R. Holsteina
Eur. Phys. J. A 18 227(2003)
(31)
J. Bremer
J. Phys. B: At. Mol. Opt. Phys. 12 2797(1979)
(32)
J. Tulkki and T. Aberg
J. Phys. B: At. Mol. Opt. Phys. 15 L435(1982)
(33)
Hee-Won Lee and Hee Il Kim
Mon. Not. R. Astron. Soc. 347 802(2004)
(34)
A. Dalgarno and J. T. Lewis
Proc. R. Soc. Lond. Ser. A 233 70(1955)
(35)
R. B. Thayyullathil, R. Radhakrishnan and M. Seema
JPA 36 8473(2003)
(36)
E. L. Ince,
Ordinary Differential equations(New York: Dover Publications 1956 ) p 187
(37)
I. S. Gradshteyn and I. M. Ryzhik
Table of Integrals, Series and Products(New York: Academic(1980))p 860 |
Abstract
The analytical representation of the longitudinal hadronic shower development
from the face of a calorimeter is presented and compared with
experimental data.
The suggested formula is particularly useful at designing,
testing and calibration of huge calorimeter complex like in ATLAS at LHC.
Analytical Representation
of the Longitudinal Hadronic Shower Development
Y.A. Kulchitsky
Institute of Physics, National Academy of Sciences, Minsk, Belarus
&
JINR, Dubna, Russia
V.B. Vinogradov
JINR, Dubna, Russia
One of the important questions of hadron calorimetry is the question
of the longitudinal development of hadronic showers
[1].
For analysing of energy depositions in various depths of calorimeters
it is necessary to have the analytical representation of longitudinal shower
development from the face of calorimeter.
Such formula is particularly useful at designing, testing and calibration
of huge calorimeter complex like in ATLAS [2] at LHC.
There is well known parameterisation of longitudinal shower development
from shower origin suggested by Bock et al. [3]:
$$\frac{dE_{s}(x)}{dx}=k\ \Biggl{\{}w\ \biggl{(}\frac{x}{X_{o}}\biggr{)}^{a-1}\ %
e^{-b\frac{x}{X_{o}}}\ +\ (1-w)\ \biggl{(}\frac{x}{\lambda_{I}}\biggr{)}^{a-1}%
\ e^{-d\frac{x}{\lambda_{I}}}\Biggr{\}},$$
(1)
where
$X_{o}$ is the radiation length,
$\lambda_{I}$ is the interaction length,
$a,\ b,\ d,\ w$ are parameters,
$k$ is the normalisation factor.
In this work we use the values of parameters obtained in
[4]:
$a=0.786+0.427\cdot lnE$,
$b=0.29$,
$d=0.978$,
$\omega=1.032-0.159\cdot lnE$.
In practice, the calorimeter longitudinal segmentation is unsufficient
for precise determination of shower vertex position.
Because of it is necessary to integrate the longitudinal profile from shower
origin over the shower position:
$$\frac{dE(x)}{dx}=\int\limits_{0}^{x}\frac{dE_{s}(x-x_{v})}{dx}\ e^{-\frac{x_{v%
}}{\lambda_{I}}}\ dx_{v},$$
(2)
where $x_{v}$ is a coordinate of the shower vertex.
In given work the analytical representation of the hadronic shower
longitudinal development from the calorimeter face has been derived.
It turns out that
$$\displaystyle\frac{dE(x)}{dx}$$
$$\displaystyle=$$
$$\displaystyle N\ \Biggl{\{}\frac{wX_{o}}{a}\biggl{(}\frac{x}{X_{o}}\biggr{)}^{%
a}e^{-b\frac{x}{X_{o}}}{}_{1}F_{1}\biggl{(}1,a+1,\biggl{(}b-\frac{X_{o}}{%
\lambda_{I}}\biggr{)}\frac{x}{X_{o}}\biggr{)}$$
(3)
$$\displaystyle+\ \frac{(1-w)\lambda_{I}}{a}\biggl{(}\frac{x}{\lambda_{I}}\biggr%
{)}^{a}e^{-d\frac{x}{\lambda_{I}}}{}_{1}F_{1}\biggl{(}1,a+1,-\bigl{(}1-d\bigr{%
)}\frac{x}{\lambda_{I}}\biggr{)}\Biggr{\}},$$
where
${}_{1}F_{1}(\alpha,\beta,z)$
is the confluent hypergeometric function [5] and
$N$ is the normalisation factor.
The confluent hypergeometric function in the first term
of formula (3) can be calculated by using the following
relation:
$${}_{1}F_{1}(1,a+1,z)=az^{-a}e^{z}\ \gamma(a,z),$$
(4)
where $\gamma(a,z)$ is the incomplete gamma function [5].
This function can be calculated by using the entry G106 in the program library
[6].
The confluent hypergeometric function in the second term
can be calculated by using the series development:
$${}_{1}F_{1}(1,a+1,-z)=1-\frac{z}{a+1}+\frac{z^{2}}{(a+1)(a+2)}+\dots.$$
(5)
This series has the remarkable property: at $a=0$ it corresponds to the
function $e^{-z}$.
For real calorimeters with longitudinal size usually less then 10 $\lambda_{I}$
the value of $z=(1-d)\frac{x}{\lambda_{I}}$ does not exceed $0.2$.
Therefore it is sufficient only 3 terms in order to calculate the
function ${}_{1}F_{1}(1,a+1,-z)$ with the precision of 0.1%.
The normalisation factor is
$$N=E_{beam}\biggr{/}\int\limits_{0}^{\infty}\frac{dE(x)}{dx}dx,$$
(6)
where $E_{beam}$ is the beam energy.
The integral in the denominator can be calculated by using the following
relation [7]:
$$\int\limits_{0}^{\infty}z^{\nu}e^{-\lambda z}{}_{1}F_{1}(\alpha,\gamma,kz)\ dz%
=\Gamma(\nu+1){\lambda}^{-\nu-1}{}_{2}F_{1}\biggl{(}\alpha,\nu+1,\gamma,\frac{%
k}{\lambda}\biggr{)},$$
(7)
where ${}_{2}F_{1}(\alpha,\beta,\gamma,z)$
is the hypergeometric function [5],
which in our case have the very simple form:
$${}_{2}F_{1}(1,a+1,a+1,z)=\frac{1}{1-z}.$$
(8)
As a result we obtain
$$N=\frac{E_{beam}}{\lambda_{I}\ \Gamma(a)\ \bigl{(}w\ X_{o}\ b^{-a}+(1-w)\ %
\lambda_{I}\ d^{-a}\bigr{)}}.$$
(9)
In Fig. 1 the calculations by formula (3) are shown
and compared with the experimental data at 20 – 140 GeV obtained by
using the conventional iron-scintillator calorimeter [4] and
the one with innovative feature — the longitudinal orientation of the
scintillating tiles [8].
The good agreement is observed.
So, now there is useful analytical formula which gives the possibility
immediately to obtain the longitudinal shower energy deposition
from calorimeter face.
References
[1]
R.K. Bock and A. Vasilescu, The Particle Detector Briefbook,
http://www.cern.ch/Physics/ParticleDetector/BriefBook.
[2]
ATLAS Collaboration, ATLAS Technical Proposal for a
General-Purpose pp experiment at the Large Hadron Collider,
CERN/LHCC/ 94-93,
CERN, Geneva, Switzerland.
ATLAS Collaboration,
ATLAS Liquid Argon Calorimeter Technical Design Report,
CERN/LHCC/96-41, ATLAS TDR 2, 1996,
CERN, Geneva, Switzerland.
ATLAS Collaboration,
ATLAS TILE Calorimeter Technical Design Report,
CERN/LHCC/96-42, ATLAS TDR 3, 1996,
CERN, Geneva, Switzerland.
[3]
R.K. Bock et al., NIM 186 (1981) 533.
[4]
E. Hughes, Proceedings of the First Int. Conf. on Calorimetry in HEP,
p. 525, FNAL, Batavia, 1990.
[5]
M. Abramovitz and I.A. Stegun eds.,
Handbook of Mathematical Functions, National Bureau of Standards,
Applied mathematics,
N.Y., Columbia Univ. Press, 1964.
[6]
Application Software Group, CERN Program Library,
1995, CERN, Geneva, Switzerland.
[7]
I.S. Gradshteyn and I.M. Ryzhik,
Table of Integrals, Series, and Products, 5th ed., Academic Press, 1994.
[8]
J.A. Budagov, Y.A. Kulchitsky, V.B. Vinogradov et al.,
Preprint JINR, E1-97-318, 1997, Dubna, Russia;
ATLAS Internal note, TILECAL-No-127, 1997, CERN, Geneva, Switzerland. |
The Evolutionary Sequence of Post-Starburst Galaxies
C. L. Wilkinson,${}^{1}$
K. A. Pimbblet,${}^{1,2}$
J. P. Stott${}^{3,4}$
${}^{1}$E.A. Milne Centre for Astrophysics, University of Hull, Cottingham Road, Kingston Upon Hull, HU6 7RX, UK
${}^{2}$School of Physics and Astronomy, Monash University, Clayton, VIC 3800, Australia
${}^{3}$Department of Physics, Lancaster University, Lancaster LA1 4YB, UK
${}^{4}$Department of Physics, University of Oxford, Keble Road, Oxford, OX1 3RH, UK
E-mail: [email protected] (CLW)
(Accepted XXX. Received YYY; in original form ZZZ)
Abstract
There are multiple ways in which to select post-starburst galaxies in the literature.
In this work, we present a study into how two well-used selection techniques have consequences on observable post-starburst galaxy parameters, such as colour, morphology and environment and how this affects interpretations of their role in the galaxy duty cycle. We identify a master sample of H$\delta$ strong (EW${}_{H\delta}$ > 3Å) post-starburst galaxies from the value-added catalogue in the 7th data release of the Sloan Digital Sky Survey (SDSS DR7) over a redshift range 0.01 < $z$ < 0.1. From this sample we select two E+A subsets, both having a very little [OII] emission (EW${}_{[OII]}$ $>-2.5$Å) but one having an additional cut on EW${}_{H\alpha}$ ($>-3$Å). We examine the differences in observables and AGN fractions to see what effect the H$\alpha$ cut has on the properties of post-starburst galaxies and what these differing samples can tell us about the duty cycle of post-starburst galaxies. We find that H$\delta$ strong galaxies peak in the ‘blue cloud’, E+As in the ‘green valley’ and pure E+As in the ‘red sequence’. We also find that pure E+As have a more early-type morphology and a higher fraction in denser environments compared with the H$\delta$ strong and E+A galaxies. These results suggest that there is an evolutionary sequence in the post-starburst phase from blue disky galaxies with residual star formation to passive red early-types.
keywords:
galaxies: evolution - galaxies: interactions - galaxies: clusters: general
††pubyear: 2017††pagerange: The Evolutionary Sequence of Post-Starburst Galaxies–The Evolutionary Sequence of Post-Starburst Galaxies\savesymbol
iint
\restoresymbolTXFiint
1 Introduction
Post-starburst galaxies are a vital link between star-forming spirals and quiescent E/S0 galaxies. E+A galaxies, also known as k+a galaxies, are specific type of post-starburst galaxy that were identified by Dressler &
Gunn (1983) as having an elliptical morphology but rich in A-class stars. Dressler &
Gunn (1983) find that 26 E/S0 galaxies display spectra showing no [OII]${}_{\lambda 3727}$ or H$\alpha$ emission but have deep Balmer absorption lines. The absence of [OII] indicates that there is no ongoing star formation whilst the presence of H$\delta$ absorption is a typical sign of young and recently formed A-class stars. This mix of spectral lines and morphology suggests a recent starburst approximately 1-3 Gyr ago, that has been quenched by an ‘unknown’ process (Dressler &
Gunn 1982; Couch &
Sharples 1987).
E+A galaxies are thought to be a transitionary phase between star-forming disks and quiescent early-type galaxies. Since their discovery, there have been many studies on E+As, but there is a large variance on how they are defined. Most observational work adopts a lack of [OII] emission and strong absorption in the H$\delta$ Balmer line to classify E+As (Poggianti
et al. 2009; Vergani
et al. 2010; Rodríguez Del Pino et al. 2014). Some have combined the Balmer absorption lines such as H$\delta$ with H$\gamma$ and/or H$\beta$ (Zabludoff et al. 1996; Norton et al. 2001; Chang et al. 2001; Blake et al. 2004; Tran et al. 2004; Yang et al. 2004) in order to maximise their E+A samples.
However, there are a number of studies that use a lack of H$\alpha$ emission (an indicator of star-formation) in their selection, such as Quintero
et al. (2004), Hogg et al. (2006), Goto (2007a) and Goto (2007b). While this method eliminates dusty star-formers it simultaneously excludes those E+As that contain H$\alpha$ caused by an active galactic nucleus. With different lines being used to select post-starburst galaxies, we ask if this has an impact on their characteristics and what can this tell us about their star formation duty cycle?
In earlier works E+A galaxies have been observed to be blue (Dressler &
Gunn 1983; Couch &
Sharples 1987). This disagrees with the originally belief that galaxies turn red immediately after a starburst (Larson
et al. 1980). Poggianti et al. (1999); Poggianti
et al. (2009) select E+As on the basis of having EW${}_{[OII]}$ emission < 5Å and EW${}_{H\delta}$ absorption > 3Å and find that there are two populations of E+As; blue and red, although without a discussion of their origin. Further works by Tran et al. (2004) and Vergani
et al. (2010), who select E+A primarily on [OII] emission and Balmer absorption lines, find that their location on a colour magnitude diagram is towards the redder end of the blue cloud and into the green valley. However, Quintero
et al. (2004) and Hogg et al. (2006), who select E+As without H$\alpha$ emission, find that E+As are located towards the bluer end of the red sequence and into the green valley on the colour-magnitude diagram. This suggests that the lack of H$\alpha$ emission could be the cause of the red E+A galaxies found in Poggianti et al. (1999).
Work on the morphology of E+A galaxies reveals them to be generally bulge dominated and in some cases having an underlying disk component (Tran et al. 2004, Quintero
et al. 2004). This combination of bulge and disk components replicates the S0 type morphology and reinforces the link between late- and early-type galaxies. This may suggest that the starburst triggering the E+A phase is centrally located, and not in the disc. The latter study opts for different selection methods yet their results on morphology are unanimous: H$\alpha$ has no effect on the morphology of post-starburst galaxies. Poggianti et al. (1999) find that the majority of E+A galaxies contain signs of spiral morphology; this indicates that the time-scale or process responsible for quenching star-formation is different from the process responsible for a morphological transformation. In turn, this could indicate multiple processes involved in triggering the E+A phase. Some have been shown to have a disturbed morphology, which is an indicator of a recent merger (Yang et al. 2004). By identifying E+A galaxies we are essentially catching them in the act of transition. Not all starbursts are strong enough to trigger the E+A phase (McIntosh
et al. 2014), but it is estimated that 30% of elliptical galaxies have passed through the E+A phase at some point during their evolutionary duty cycle (Tran et al. 2003; Goto 2005) making the E+A phase an important and significant part of galaxy evolution.
The properties of E+A galaxies are found to vary with different environments. For example, cluster E+A galaxies contain a more prominent disk component than those outside a cluster (Tran et al. 2003) and are found to be 1.5$\sigma$ bluer than those in the field at $z$ $\sim$ 0.1. This finding goes against the typical colour-density relation that suggests galaxies in clusters should be redder, not bluer (Hogg et al. 2003). This difference in colour is attributed to differing galaxy masses in both field and cluster E+A galaxies. Clusters have been found to contain much smaller E+As than in the field (Tran et al. 2003).
E+A galaxies are found both in a cluster environment as well as a field environment, however, different studies show different environmental preferences. From a sample of 21 nearby E+A galaxies (selected by [OII] and Balmer lines) from the Las Campanas Redshift Survey, Zabludoff et al. (1996) finds 75% of the E+As are in the field. They find 5 E+A galaxies within their sample display tidal features, a typical sign of galaxy-galaxy interactions and mergers. This suggests that mergers and galaxy-galaxy interactions could be the trigger for the E+A phase. The type of mergers found to dominate in low density environments are gas-rich (Bekki
et al. 2001; Lin et al. 2010; Sánchez-Blázquez
et al. 2009) which are most likely to trigger a starburst strong enough of producing an E+A signature.
When H$\alpha$ is included into the selection criteria different results are found; Mahajan (2013) find that E+A galaxies prefer a weak-group environment ranging from 4 to 10 galaxies. For local X-ray bright galaxy clusters at 0.02$<z<$0.06, they find 86% of local clusters contained some form of substructure on a weak group-scale (4-10 members) and 91.4% of these weak-groups contained E+A galaxies. This result suggests that star-formation is quenched due to a galaxy being pre-processed in a weak-group environment before being accreted into a cluster environment.
Opposite results are found by Poggianti
et al. (2009): at higher redshifts ($0.4<z<0.8$) E+A galaxies are predominantly in clusters where the star-formation has been quenched on a short time scale of $\ll$ 1 Gyr. Poggianti
et al. (2009) find only a few E+A galaxies are located in other environments such as the field and in weak-groups and that the variation in E+A fraction is found to be dependant on environment. Mahajan (2013). Regardless of whether H$\alpha$ has been used in the selection criteria of E+As there is a discrepancy in what different works report with regards to environmental preference. Lemaux et al. (2016) suggests that ‘true’ E+As (those with no H$\alpha$) have a higher fraction in a cluster environment compared to those selected by ‘traditional’ methods (using only Balmer absorption lines and the absence of [OII]). Could environment be driving changes in H$\alpha$ emission or are we seeing two different stages of the E+A duty cycle?
A typical sign that star-formation has ceased is a lack of H$\alpha$ emission and as E+As galaxies are no longer producing stars, one would expect to see no significant H$\alpha$ emission. Although not commonly used as a criteria for E+A selection, this work aims to test whether including a cut in H$\alpha$ could explain some of the differences found in other studies and connect up these potentially different phases of post-starburst evolution. Here we present a study of the properties of low redshift post-starburst galaxies using different selection criteria from the Sloan Digital Sky Survey Data Release 7 (SDSS DR7; Abazajian
et al. 2009) with the broad aim of testing the star formation duty cycle of post-starburst galaxies.
In section 2 we describe the data used in this study and the different selection cuts we impose. In section 3 we analyse and discuss the results. Our conclusions are presented in in section 4. We assume the following cosmological parameters for a flat Universe: $H_{0}=\rm 69.3\,kms^{-1}Mpc^{-1},\Omega_{M}=0.238\ and\ \Omega_{\Lambda}=0.762$ (Spergel
et al. 2007).
2 Sample Selection
Using the MPA/JHU value-added catalogue from SDSS DR7 (Abazajian
et al. 2009), we create a master sample of 234 H$\delta$ strong (EW${}_{H\delta}$ > 3Å; note that positive/negative values for equivalent widths refer to absorption/emission respectively) post-starburst galaxies with 0.01 $<z<$ 0.10. This limit on H$\delta$ is comparable to studies such as Poggianti et al. (1999), Bekki
et al. (2001), Vergani
et al. (2010) and Rodríguez Del Pino et al. (2014). The lower limit on the redshift range ensures we exclude any stars from our sample. Although the SDSS spectroscopy aims to be 100% complete at r=17.77 (Strauss
et al. 2002), we find that our sample is $\sim$80% complete at this level, as shown in Fig.1.
Most studies of post-starbursts use a typical S/N of 3-5 in the continuum, as evidenced by Pracy et al. (2005) (S/N$>$5); Couch &
Sharples (1987) (S/N$>$3) and Brown et al. (2009) (S/N$>$4), hence we adopt a minimum S/N of 5.
Our complete S/N $>$ 5 H$\delta$ strong sample contains 192 galaxies. Using this sample we apply additional cuts to create a further two sub-samples; E+As and ‘pure’ E+As (E+As with low H$\alpha$ emission) in order to test whether different criteria for post-starburst galaxies can explain the differences found in literature and to gain a better insight into the post-starburst duty cycle.
Our first sub-sample is based on the Poggianti et al. (1999) definition of E+A galaxies. With the primary selection of EW${}_{H\delta}$ > 3Å, we apply an additional cut of $>-2.5$Å in the equivalent width of [OII]${}_{\lambda 3727}$ (Zabludoff et al. (1996); Yang et al. 2004). This sample of E+As contains 67 galaxies.
To account for this, our second sub-sample ensures there is no ongoing (potentially dusty) star-formation in the E+A sample by applying a further cut in H$\alpha$ emission (EW${}_{H\alpha}$ > -3Å) (Goto 2007b). Doing this we obtain 25 pure E+A galaxies. We list the selection criteria for our three post-starburst sub-samples below:
H$\delta$ Strong:
1.
EW${}_{\delta}$ > 3Å
E+A:
1.
EW${}_{\delta}$ > 3Å
2.
EW${}_{[OII]}$ > -2.5Å
Pure E+A:
1.
EW${}_{H\delta}$ > 3Å
2.
EW${}_{[OII]}$ > -2.5Å
3.
EW${}_{H\alpha}$ > -3Å
We test whether or not we would expect to see H$\alpha$ given the limit on [OII] emission stated above. We use a simple approach by plotting EW${}_{[OII]}$ against EW${}_{H\alpha}$ as shown in Fig. 2. In this plot we witness a significant amount of scatter (regression analysis gives R${}^{2}$ to be 0.3), however, there is an underlying trend present and we are able to fit a regression line using a robust linear model (Huber 1981). From this we see at the EW${}_{[OII]}$ limit we would expect EW${}_{H\alpha}$ to be -2.1Å in emission. On average the [OII] cut should remove any significant star formation indicated by H$\alpha$. However, the large scatter means that there is a population of galaxies with up to EW=-15Å of H$\alpha$ emission at the [OII] limit (See fig. 2). We also calculate the star formation rate derived from the H$\alpha$ flux, SFR${}_{H\alpha}$ from Kennicutt (1998) and apply an extinction correction (10${}^{A/-2.5}$) from Treyer
et al. (2007). Assuming $Av=0.5$ we find the average SFR${}_{H\alpha}$ for H$\delta$ strong galaxies that didn’t make the E+A cut to be 0.07M${}_{\odot}$/yr${}^{-1}$. For the E+A that are not classed as pure we find SFR${}_{H\alpha}$ to be 0.06M${}_{\odot}$/yr${}^{-1}$ and pure E+A have an average SFR${}_{H\alpha}$ of 0.01M${}_{\odot}$/yr${}^{-1}$. Stacked spectra for our three sub-samples can be found in Fig. 3. We note here that emission-filling could potentially affect our H$\delta$ measurements. We believe this would be a small effect (Goto et al. 2003; Blake et al. 2004). However, the usual method of measuring this in conjunction with, e.g., H$\alpha$ to determine the strength of its effect is unable to be achieved due to our sample selection approach.
3 Colour
In this section we investigate whether the differences in selection criteria, based on spectral lines, has an impact on the colours of the post-starburst population. This will allow us to determine why Poggianti et al. (1999); Poggianti
et al. (2009) finds two populations of E+A galaxies that peak in the red sequence and blue cloud.
To study the differences in colour between our samples we first plot a colour magnitude diagram, shown in Fig. 4. Using the galaxies from the SDSS DR7 catalogue, we are able to locate the red sequence and blue cloud regions and over plot our samples to see which regions they reside (Fig.4).
We include a colour histogram (Fig. 5) that shows the bi-modality of the colour distribution. Using Fig. 4 and Fig. 5 we find that the H$\delta$ strong sample peaks in the blue cloud with a median colour of 1.36$\pm 0.04$. The E+A sample peaks in the green valley with a median colour of 1.46$\pm 0.04$. We find that the pure E+As predominantly reside in the red sequence with a median colour of 1.66$\pm 0.04$.
Hogg et al. (2006) select E+A galaxies using a cut in H$\alpha$ and find that E+As are located in the red sequence towards the green valley. While E+As have a significant connection to the green valley (Vergani
et al. 2010), we find that by selecting E+As using H$\alpha$ emission we are predominantly selecting those towards the red sequence as star formation has ceased and the galaxies are truly passive.
To test whether the E+A and pure E+A samples are distinguishable we apply a two sample Kolmogorov-Smirnov statistical (abbreviated to KS onwards) test. When comparing the E+A sample with the pure E+As the KS test returns a p-value of 0.007 which demonstrated that E+As and pure E+As are unlikely to be drawn from the same parent distribution. In this study we consider any p-value less than 0.01 to be statistically different, whilst p-value over 0.1 we consider to be the same. P-values between these limits we consider to have a weak difference.
To test the distinguishability between the E+A sample and the H$\delta$ strong sample, we apply the same test and output a p-value of 0.110. These distributions are likely to be from the same parent distribution. When we take H$\delta$ strong galaxies that are not defined as E+As and compare to the E+A sample we get a p-value of 0.005. This means there is a statistical difference between non-E+A H$\delta$ strong galaxies and E+A galaxies. Fig. 5 clearly shows that the pure E+A sample and the H$\delta$ strong sample are statistically different. To quantify the distinguishability of the H$\delta$ strong sample and pure E+As we include a p-value which is found to be 10${}^{-6}$.
4 Morphology and Morphological Parameters
To explore the differences in findings on morphology we test the morphological fractions of each sample using the Galaxy Zoo (GZ) catalogue (Lintott
et al. 2008, 2011). The GZ project provides an extensive catalogue of galaxy morphologies and is based on the data from SDSS. GZ is a citizen science project in which volunteers view images of galaxies and cast their vote as to whether the galaxy is elliptical, spiral, merger or unknown. We assign a classification of elliptical, spiral or merger to our galaxies if that classification takes 50% or more of the votes. The minimum number of votes per galaxy in our sample is 16 and the mean number or counts our sample has is 39.5. We note here that S0 galaxies are not easily distinguishable from the elliptical and spiral morphologies in the GZ catalogue. Whilst the majority of S0 galaxies are within the elliptical classification there will certainly be S0 contamination in the spiral classifications (Bamford
et al. 2009).
We find that from our H$\delta$ strong sample we are able to match 185 (96.4%) galaxies to the GZ catalogue. This in turn means we matched 62 E+As (92.5%) and 21 pure E+As (84.0%) to the catalogue. The morphology fractions for each of the samples are shown in Table 1. We find that there is no statistically significant merger fraction in either of the three samples, this is due to the very low numbers that were identified. We find that $\sim$55% of the H$\delta$ strong sample have a spiral morphology and $\sim$30% have an elliptical morphology. There is a clear majority here for a preference of disc morphologies. This result somewhat changes in the E+A sample in which there is a $\approx$50/50 (42%/44%) split between the elliptical and spiral morphologies for those matched to GZ. However, in the pure E+A sample we see $\sim$80% of the galaxies are classed as ellipticals and no significant spiral fraction. This is in agreement with the findings that pure E+As are redder than the H$\delta$ strong and E+A galaxies and occupy the red sequence with the majority of early-type. Interestingly we note that mergers make up less than 5% of any of the post-starburst categories and therefore do not seem to be important for this phase of galaxy evolution.
As well as using the GZ catalogue to test morphology fractions we also look at the $fracDeV$ parameter, which corresponds to the Sérsic index (a generalization of the de Vaucouleur profile, de Vaucouleurs 1948; Sersic 1968) and links together galaxy shape with surface brightness profiles (Fig. 6). $fracDeV$ is the fraction of light fit by the de Vaucouleur profile versus an exponential profile. When $fracDeV$=1, this represents a pure de Vaucouleur profile and is appropriate for an elliptical morphology. When $fracDev$=0, this represents a pure exponential profile and is appropriate for a disk morphology as seen in spiral galaxies. The $fracDev$ parameter is listed in the SDSS database for all $u,g,r,i,z$ filters (Abazajian
et al. 2009).
Fig. 6 shows the normalised distribution of the $fracDeV$ parameter and broadly agrees with the morphological classifications in Table 1. For the pure E+A the $fracDeV$ distribution tends to one meaning an elliptical morphology. For the H$\delta$ strong sample and the E+A sample we visually see little difference in distribution and that both distributions are dominated by a disk-like profile. We calculate their median $fracDeV$ values to be 0.086$\pm 0.025$ for H$\delta$ strong galaxies, 0.238$\pm 0.046$ for E+As and 0.779$\pm 0.063$ for pure E+As.To test this statistically we apply a KS test to all samples so determine their distinguishability. We find that when comparing the H$\delta$ strong sample to the E+A sample we obtain a p-value of 0.338; we conclude that these two samples are not statistically different. To determine whether the E+A sample and the pure E+As emission are distinguishable we apply the same test and obtain a p-value of 0.003. We can therefore claim that there is a significant difference between E+As and pure E+As. To complete these tests, a p-value of $\sim$10${}^{-6}$ was achieved when comparing the H$\delta$ strong sample to the pure E+As.
We examine the radii of our samples to further quantify our morphology results and to compare to the trends found in Shen et al. (2003). Here we use the effective radii based on the deVaucouleur profile in the SDSS r-band. We note here we have included a first order seeing correction to the radii, as shown in Equation 1 assuming the average SDSS seeing of 2".
$$r_{corrected}=(r_{observed}^{2}-seeing^{2})^{1/2}$$
(1)
Fig. 7 shows the normalised radius distributions of the three samples. The majority of the pure E+A galaxies are mainly under 10kpc in radius and are therefore relatively small galaxies. We find the following median radii for the H$\delta$ strong, E+A and pure E+A galaxies respectively: 11.4kpc$\pm$0.5, 9.8kpc$\pm$0.8 and 4.1kpc$\pm$1.2. When a KS test is applied to the H$\delta$ strong sample and the E+A sample we obtain a p-value of 0.176. When comparing the E+As and H$\delta$ strong sample to the pure E+As we obtain the p-values 0.006 and $\sim$10${}^{-6}$ respectively. This results shows that our third sample are statistically smaller in radii than the other two samples. This is also shown in Fig. 8 which shows the relation between radius and absolute magnitude with the corresponding trends from Shen et al. (2003). We look at the median separation between R${}_{50}$ and the trends from Shen et al. (2003) to quantify which line fits which population. From the H$\delta$ strong sample we find that the median separation from the early-type trend is 1.21kpc and 0.03kpc from the late-type trend. Looking at the E+A sample we find that the median separation is 0.48kpc from both the early-type trend and late-type trend, this demonstrate the finds from Table 1. Finally, we find that the median separation of the pure E+A sample to the early-type trend is 0.39kpc and 0.85kpc to the late-type trend. This shows the H$\delta$ strong galaxies follow the trend for late-type galaxies, the E+As an even mix of early- and late-type and the pure E+As resemble an early-type relation. This again confirms our findings from the colour and morphology analysis.
5 Environment
Whilst there is little evidence in the literature that selection criteria affect the environmental results, we aim to test whether there are any environmental differences between our sub-samples.
In order to explore the environments of the galaxies in our samples we use the Yang group and cluster catalogue (Yang et al. 2007). We directly match 176 galaxies in our sample (91.2%); this corresponds to 64 E+As (94.0%) and 22 pure E+As (88.0%). The Yang catalogue is based on the 4th data release of SDSS and selects galaxies with a redshift completeness $C>0.7$, so therefore may not match totally to the 7th data release.
First, we examine whether our galaxies are in groups/clusters, if so, we look at the membership of that group or cluster to determine the size. Our results are displayed in Table 2. We split our environment classifications into field galaxies, pairs, weak groups (3$<$N$<$10), rich groups (10$<$N$<$50) and clusters (N$>$50).
Table 2 shows that $\sim$70% and $\sim$60% of the H$\delta$ strong galaxies and E+A galaxies respectively are field galaxies with a strong preference for this type of environment; this supports the findings of Zabludoff et al. (1996) who find that 75% of E+A galaxies to be in a field environment. We find that only$\sim$40% of the pure E+As are in the field environment whilst there is a significant amount of pure E+A galaxies in the group environments. When combining these environments together into all field (field and pair, where a pair is two galaxies that are gravitationally bound) and all cluster (weak groups, rich groups and clusters) environments we see the following splits in preference (field/cluster); 80/20, 70/30 and 50/50 for H$\delta$ strong, E+A and pure E+As respectively. This suggests that pure E+As are more likely to be in denser environments than the other sub-samples of post-starburst galaxies.
6 AGN Connection
Hydrodynamical simulations show that major mergers of spirals contribute heavily to the populations of E+A galaxies (Snyder et al. 2011). When these gas-rich spirals merge, high levels of star-formation occur and to quench this star-formation a feedback mechanism is required, potentially AGN driven feedback.
During a merger gas is funnelled into the nucleus of a merging galaxy by tidal torques and and can trigger a nuclear starburst (Mihos &
Hernquist 1996; Hopkins et al. 2008a; Hopkins et al. 2008b) and therefore a potential progenitor of the E+A phase. With an excess of gas in the nucleus of the galaxy, this provides fuel for the supermassive black hole in the centre and can trigger an active galactic nuclei (AGN). AGN can produce jets, winds or radiation that heat or remove cold gas from galaxies, removing the fuel for further star formation.
De Propris &
Melnick (2014) show in their study of 10 K+A galaxies that at any age there is no presence of AGN. They suggest, however, that a quasar phase immediately follows the quenching mechanism and this leads onto the K+A phase.
In this section we investigate the E+A/AGN link further by examining the AGN fractions within our samples and comparing to the fractions found in SDSS. We start by plotting a BPT diagram (Baldwin
et al. 1981; Veilleux &
Osterbrock 1987) shown in Fig. 9 which includes the Kauffman (Kauffmann
et al. 2003), Kewley (Kewley
et al. 2001) and Schawinski lines (Schawinski
et al. 2014). Using these lines we locate the seyfert, LINER, composite and star formation regions. Starting with a control sample derived from SDSS we find that seyfert galaxies make up 10.8% of the general galaxy population, LINERs 11.5%, composites 20.7% and star formers 57.0%.
We plot the H$\delta$ strong and E+A samples in Fig. 9 on the BPT diagram and find that the majority these two samples are within the star forming region (65.4%$\pm$6.4 and 57.1%$\pm$11.7 respectively). We find that composites makes up 23.0%$\pm$3.8 of the H$\delta$ strong sample and 35.7%$\pm$9.2 of the E+A sample. Seyferts make up 8.2%$\pm$2.3 of the H$\delta$ strong sample and 4.8%$\pm$3.4 of the E+As whilst LINERs make up 3.1%$\pm$1.4 of the H$\delta$ strong sample and 2.4%$\pm$2.4 of the E+As. These results show similar fraction in both samples of star formers compared to the general galaxy population, however we witness an enhancement in E+A composite galaxies and decline in E+A AGN.
We also look at AGN fractions using a WHAN diagram (Fig. 10), described by Cid Fernandes
et al. (2010); Cid Fernandes et al. (2011) which unlike the BPT can identify ‘fake’ AGN (i.e. retired galaxies) from those with weakly active nuclei from the BPT LINER region. Using the SDSS control sample we find that 46.0% of the general galaxy population is made up of star forming galaxies, 27.2% passive galaxies, 19.3% strong AGN and 7.4% weak AGN.
We find that the H$\delta$ strong sample is comprised of 59.7%$\pm$6.1 star forming galaxies, 15.1%$\pm$3.1 passive galaxies, 16.4%$\pm$3.2 sAGN and 8.8%$\pm$2.3 wAGN. The E+A sample is found to be comprised of 54.2%$\pm$10.6 star forming galaxies, 27.1%$\pm$7.5 passive galaxies, 12.5%$\pm$5.1 sAGN and 8.3%$\pm$4.2 wAGN. These fractions are similar to those found in the SDSS control sample, however, we witness a decline in passive galaxies and slight enhancement of star formers within the H$\delta$ strong sample compared to SDSS galaxies, which is to be expected.
7 Discussion
7.1 Links to Previous Studies
We have discussed in section 1 how different studies have used different selection methods to select post-starburst galaxies. There are two main methods for selection; the most common selecting on the lack of [OII] emission and the presence Balmer absorption whilst the other selects on H$\alpha$ emission, as well as [OII] emission and Balmer absorption.
Poggianti et al. (1999); Poggianti
et al. (2009) find that there are two populations of E+As when they select using [OII] emission and H$\delta$ Balmer absorption; these two populations can be classed as red and blue. We have seen from Fig. 4 and Fig. 5 that the H$\delta$ galaxies peak in the blue cloud, the E+As peak in the green valley and the pure E+As peak in the red sequence. This result matches with that of Hogg et al. (2006) who place H$\alpha$ selected E+As within the red sequence. This means the cut in H$\alpha$ is removing galaxies with residual star formation, therefore, we are intrinsically selecting the red E+As mentioned in Poggianti et al. (1999) and Poggianti
et al. (2009).
Along with colour we have also tested the morphologies of our three samples. The literature tells us E+As are bulge dominated with an underlying disc component (Tran et al. 2004; Quintero
et al. 2004), similar to the S0 morphology. This was not seen to change with different selection methods mentioned above. We find in Table 1 and Fig. 6 that there are significant fractions in both the elliptical and spiral morphologies for H$\delta$ strong galaxies and E+As. However, we see that pure E+As are predominantly elliptical.
Regardless of what selection method is used in the literature there is a great deal of discrepancy in the environmental preferences of post-starburst galaxies. Table 2 shows that $\sim$70% and $\sim$60% of H$\delta$ strong galaxies and E+As are within the field, this result agrees with the findings of Zabludoff et al. (1996) at low redshifts $z<0.1$. Lemaux et al. (2016) states that pure E+As have a higher fraction in clusters than E+As selected only on [OII] emission and Balmer absorption, we find the same trend when examining the fractions of post-starburst galaxies in various environments. Mahajan (2013) find that E+As prefer a weak-group environment, whilst we do see a systematically higher fraction in the weak-group environment compared with rich environments this result is not statistically significant.
7.2 Duty Cycle of Post-Starburst Galaxies
In section 3 we find that the colours of H$\delta$ strong galaxies peak in the blue cloud, the E+As in the green valley and the pure E+As in the red sequence (Fig. 4 and Fig. 5). This result is backed up morphologically in section 4. We find that the majority of H$\delta$ strong galaxies have a spiral morphology whereas the E+As have a mix of spiral and elliptical morphologies . The majority of pure E+As display an elliptical morphology (Table 1). This is confirmed again by the light profiles of our samples (Fig. 6), we find that H$\delta$ strong and E+A galaxies have a disky light profile whereas the pure E+As have a de Vaucouleur light profile, typical of an early-type morphology. This difference in morphology is seen in the radii of our galaxies (Fig. 7 and Fig.8), in which the H$\delta$ strong and E+A galaxies follow the late-type relationship found by Shen et al. (2003) and the pure E+As follow the early-type relationship. The lack of merging morphologies suggests that major mergers do not play an important role in the post-starburst phase.
We speculate that the H$\delta$ strong galaxies, E+A and pure E+As are part of an evolutionary sequence in which star forming galaxies are quenched into blue disky H$\delta$ strong galaxies. They are then quenched further into green valley E+As before, finally, quenching fully into red elliptical pure E+As. If this is true, how does the morphology transition from disk to elliptical? This could be due to a fading disk with time as young stars die or possibly due to any mergers. We add here that we witness no evidence for mergers in the morphologies (Table 1) so the trigger would have to be a minor merger or a mergerless interaction.
Another factor we consider is environment, although we have low-number statistics, we see a higher proportion of pure E+As in denser environments than the H$\delta$ strong and E+A galaxies. Whilst in the H$\delta$ strong and E+A galaxies we see a strong preference for a field environment. We note this is not a migration of galaxies from the field to a cluster environment as the migration time scale is much longer than that of the evolution of our galaxies. Instead, this may point to an accelerated evolution from H$\delta$ strong galaxies to pure E+As in dense environments (Hatch et al. 2011). A dense environment would enhance processes such as interactions and minor mergers. Also the hot atmosphere of the intracluster medium would help quench and fade the disk by starving the galaxy of the fuel to replenish its cold gas supply.
We see from looking at AGN fractions using BPT and WHAN diagnostics that there is a higher fraction of AGN in the H$\delta$ strong sample compared with the E+As. This could suggest that E+As are the product of H$\delta$ strong galaxies that have been quenched by AGN feedback, however, as the E+A cut selects against line emission this is purely speculative and we have no solid evidence that AGN are important in quenching star formation after a burst.
8 Conclusion
The post-starburst phase in galaxy evolution is an important and significant link between star-forming discs and quiescent early-type galaxies. Throughout the literature there are several different selection techniques. We split these different techniques into two categories; those that select on the presence of Balmer absorption and the absence of [OII] emission and those which also select against H$\alpha$ emission. There are advantages and disadvantages in using either method, one includes those galaxies that have H$\alpha$ due to the presence of AGN and the other excludes dusty starburst galaxies. In this work we have explored how the different selection techniques can affect the results found in colour, morphology, environment and AGN fraction. Here we list our principal findings and suggest what we feel is an ideal selection method:
1.
By locating our samples on a colour magnitude diagram and colour histogram, we have been able to find where each population peaks. We find that H$\delta$ strong galaxies peak in the blue cloud, E+A galaxies in the green valley and pure E+As in the red sequence.
2.
Using the GZ catalogue we find that H$\delta$ strong galaxies are predominantly spiral in morphology with a fraction of 54%. We find that E+As have approximately an even split between the elliptical and spiral morphologies and pure E+As have an elliptical fraction of 81%. This results is confirmed when we look at the $fracdev$ parameter which shows the H$\delta$ strong galaxies and E+As have a disk morphology and the pure E+As have an early-type morphology. This is again demonstrated when we look at the radii of our galaxies, we see that the H$\delta$ strong galaxies and E+As follow the late-type trend found in Shen et al. (2003), whilst the pure E+As follow the early-type trend.
3.
We find that the majority of H$\delta$ strong galaxies and E+A galaxies have a strong preference for the low density environment, however, there is an even split between the low and high density environments within the pure E+A population. We see that there is a higher fraction of pure E+As in denser environments than the other two populations.
4.
In this study we have also tested the AGN fractions of our samples by using BPT and WHAN diagrams. Both diagrams suggest that the emission lines of the H$\delta$ strong and E+A populations are excited by star formation and not AGN activity. The AGN fraction of the H$\delta$ strong galaxies is typical of the general galaxy population, so provides no evidence that AGN play a role in quenching star-formation in post-starburst galaxies. However, we see a significant decrease in AGN fraction compared with the general population which could be a consequence of AGN fading into the E+A phase assuming AGN feedback leads to the E+A phase. However, we note that the E+A sample is selected to not have strong emission lines (albeit in [OII]) and so it is not possible to draw strong conclusions.
We postulate that different selection criteria allow us to probe different stages of post-starburst galaxies. After the initial starburst has ended the galaxies are seen as blue spirals/disks with a young stellar spectra, indicated by deep Balmer absorption features. However, $\sim$65% of these disks are still forming stars as seen by their [OII] and/or H$\alpha$ emission lines. If we cut on [OII] emission (and therefore star formation) we see a population of green valley disk galaxies suggestive of the next stage in the post-starburst evolution as galaxies quench towards the red sequence. If we make a further cut on star formation using a lack of H$\alpha$ emission then we find a population of early-type red sequence galaxies. These are galaxies that still show deep Balmer absorption but with no ongoing star formation and represent the final post starburst phase before the underlying stellar population becomes dominated by old stars.
The lack of any significant enhancement in the merger or AGN fractions for the post-starbursts indicates that, while we cannot rule out a scenario whereby the initial starburst was triggered by a major merger and then initially quenched by an AGN, it is not these processes that continue to quench the galaxies down to full quiescence. One puzzle in our proposed evolutionary sequence is how the morphologies have changed from disky to early-type without evidence of major merging. We speculate that this morphological evolution could be due to the secular fading of the star-forming disk leading to a bulge dominated system or to a series of minor interactions not detected in the imaging. There also seems to be a suggestion that the passive, pure E+As are more prevalent in denser environments, which perhaps suggests that the quenching that follows a starburst proceeds more quickly in dense environments perhaps due to the enhancement of minor interactions in groups or the removal of a cold gas supply by the hot intracluster medium of the most massive systems.
Acknowledgements
JPS gratefully acknowledges support through the Hintze Research Fellowship.
Funding for the SDSS and SDSS-II has been provided by the Alfred P. Sloan Foundation, the Participating Institutions, the National Science Foundation, the U.S. Department of Energy, the National Aeronautics and Space Administration, the Japanese Monbukagakusho, the Max Planck Society, and the Higher Education Funding Council for England. The SDSS Web Site is http://www.sdss.org/.
The SDSS is managed by the Astrophysical Research Consortium for the Participating Institutions. The Participating Institutions are the American Museum of Natural History, Astrophysical Institute Potsdam, University of Basel, University of Cambridge, Case Western Reserve University, University of Chicago, Drexel University, Fermilab, the Institute for Advanced Study, the Japan Participation Group, Johns Hopkins University, the Joint Institute for Nuclear Astrophysics, the Kavli Institute for Particle Astrophysics and Cosmology, the Korean Scientist Group, the Chinese Academy of Sciences (LAMOST), Los Alamos National Laboratory, the Max-Planck-Institute for Astronomy (MPIA), the Max-Planck-Institute for Astrophysics (MPA), New Mexico State University, Ohio State University, University of Pittsburgh, University of Portsmouth, Princeton University, the United States Naval Observatory, and the University of Washington
References
Abazajian
et al. (2009)
Abazajian K. N., et al., 2009, ApJS,
182, 543
Baldwin
et al. (1981)
Baldwin J. A., Phillips M. M., Terlevich R., 1981, PASP, 93, 5
Bamford
et al. (2009)
Bamford S. P., et al., 2009, MNRAS, 393, 1324
Bekki
et al. (2001)
Bekki K., Shioya Y., Couch W. J., 2001, ApJ,
547, L17
Blake et al. (2004)
Blake C., et al., 2004, Monthly Notices of the Royal Astronomical
Society, 355, 713
Brown et al. (2009)
Brown M. J. I., et al., 2009, ApJ,
703, 150
Chang et al. (2001)
Chang T.-C., van Gorkom J. H., Zabludoff A. I., Zaritsky D., Mihos J. C.,
2001, AJ, 121, 1965
Cid Fernandes
et al. (2010)
Cid Fernandes R., Stasińska G., Schlickmann M. S., Mateus A.,
Vale Asari N., Schoenell W., Sodré L., 2010, MNRAS, 403, 1036
Cid Fernandes et al. (2011)
Cid Fernandes R., Stasińska G., Mateus A., Vale Asari N., 2011,
MNRAS, 413, 1687
Couch &
Sharples (1987)
Couch W. J., Sharples R. M., 1987, MNRAS, 229, 423
De Propris &
Melnick (2014)
De Propris R., Melnick J., 2014, MNRAS,
439, 2837
Dressler &
Gunn (1982)
Dressler A., Gunn J. E., 1982, ApJ, 263, 533
Dressler &
Gunn (1983)
Dressler A., Gunn J. E., 1983, ApJ, 270, 7
Goto (2005)
Goto T., 2005, MNRAS, 357, 937
Goto (2007a)
Goto T., 2007a, MNRAS, 377, 1222
Goto (2007b)
Goto T., 2007b, MNRAS, 381, 187
Goto et al. (2003)
Goto T., et al., 2003, PASJ, 55, 771
Hatch et al. (2011)
Hatch N. A., Kurk J. D., Pentericci L., Venemans B. P., Kuiper E., Miley
G. K., Röttgering H. J. A., 2011, MNRAS, 415, 2993
Hogg et al. (2003)
Hogg D. W., et al., 2003, The Astrophysical Journal, 585, L5
Hogg et al. (2006)
Hogg D. W., Masjedi M., Berlind A. A., Blanton M. R., Quintero A. D.,
Brinkmann J., 2006, ApJ, 650, 763
Hopkins et al. (2008a)
Hopkins P. F., Hernquist L., Cox T. J., Kereš D., 2008a, ApJS, 175, 356
Hopkins et al. (2008b)
Hopkins P. F., Cox T. J., Kereš D., Hernquist L., 2008b, ApJS, 175, 390
Huber (1981)
Huber P. J., 1981, Robust statistics
Kauffmann
et al. (2003)
Kauffmann G., et al., 2003, MNRAS, 346, 1055
Kennicutt (1998)
Kennicutt Jr. R. C., 1998, ARA&A,
36, 189
Kewley
et al. (2001)
Kewley L. J., Heisler C. A., Dopita M. A., Lumsden S., 2001, ApJS, 132, 37
Larson
et al. (1980)
Larson R. B., Tinsley B. M., Caldwell C. N., 1980, ApJ, 237, 692
Lemaux et al. (2016)
Lemaux B. C., Tomczak A. R., Lubin L. M., Wu P.-F., Gal R. R., Rumbaugh
N., Kocevski D. D., Squires a. G. K., 2016, preprint (arXiv:1608.00973)
Lin et al. (2010)
Lin L., et al., 2010, ApJ, 718, 1158
Lintott
et al. (2008)
Lintott C. J., et al., 2008, MNRAS, 389, 1179
Lintott
et al. (2011)
Lintott C., et al., 2011, MNRAS,
410, 166
Mahajan (2013)
Mahajan S., 2013, MNRAS, 431, L117
McIntosh
et al. (2014)
McIntosh D. H., et al., 2014, MNRAS, 442,
533
Mihos &
Hernquist (1996)
Mihos J. C., Hernquist L., 1996, ApJ, 464, 641
Norton et al. (2001)
Norton S. A., Gebhardt K., Zabludoff A. I., Zaritsky D., 2001, ApJ, 557, 150
Poggianti et al. (1999)
Poggianti B. M., Smail I., Dressler A., Couch W. J., Barger A. J., Butcher
H., Ellis R. S., Oemler Jr. A., 1999, ApJ,
518, 576
Poggianti
et al. (2009)
Poggianti B. M., et al., 2009, ApJ,
693, 112
Pracy et al. (2005)
Pracy M. B., Couch W. J., Blake C., Bekki K., Harrison C., Colless M.,
Kuntschner H., de Propris R., 2005, MNRAS, 359, 1421
Quintero
et al. (2004)
Quintero A. D., et al., 2004, ApJ, 602, 190
Rodríguez Del Pino et al. (2014)
Rodríguez Del Pino B., Bamford S. P., Aragón-Salamanca A.,
Milvang-Jensen B., Merrifield M. R., Balcells M., 2014, MNRAS, 438, 1038
Sánchez-Blázquez
et al. (2009)
Sánchez-Blázquez P., Gibson B. K., Kawata D., Cardiel N., Balcells
M., 2009, MNRAS, 400, 1264
Schawinski
et al. (2014)
Schawinski K., et al., 2014, MNRAS, 440, 889
Sersic (1968)
Sersic J. L., 1968, Atlas de galaxias australes
Shen et al. (2003)
Shen S., Mo H. J., White S. D. M., Blanton M. R., Kauffmann G., Voges W.,
Brinkmann J., Csabai I., 2003, MNRAS, 343, 978
Snyder et al. (2011)
Snyder G. F., Cox T. J., Hayward C. C., Hernquist L., Jonsson P., 2011,
ApJ, 741, 77
Spergel
et al. (2007)
Spergel D. N., et al., 2007, ApJS, 170, 377
Strauss
et al. (2002)
Strauss M. A., et al., 2002, AJ, 124, 1810
Tran et al. (2003)
Tran K.-V. H., Franx M., Illingworth G., Kelson D. D., van Dokkum P.,
2003, ApJ, 599, 865
Tran et al. (2004)
Tran K.-V. H., Franx M., Illingworth G. D., van Dokkum P., Kelson D. D.,
Magee D., 2004, ApJ, 609, 683
Treyer
et al. (2007)
Treyer M., et al., 2007, The Astrophysical Journal Supplement Series, 173, 256
Veilleux &
Osterbrock (1987)
Veilleux S., Osterbrock D. E., 1987, ApJS, 63,
295
Vergani
et al. (2010)
Vergani D., et al., 2010, A&A, 509,
A42
Yang et al. (2004)
Yang Y., Zabludoff A. I., Zaritsky D., Lauer T. R., Mihos J. C., 2004,
ApJ, 607, 258
Yang et al. (2007)
Yang X., Mo H. J., van den Bosch F. C., Pasquali A., Li C., Barden M.,
2007, ApJ, 671, 153
Zabludoff et al. (1996)
Zabludoff A. I., Zaritsky D., Lin H., Tucker D., Hashimoto Y., Shectman
S. A., Oemler A., Kirshner R. P., 1996, ApJ,
466, 104
de Vaucouleurs (1948)
de Vaucouleurs G., 1948, Annales d’Astrophysique, 11, 247 |
High-resolution transmission spectroscopy of four hot inflated gas giant exoplanets
Jiří Žák
Astronomical Institute, Czech Academy of Sciences,
Fričova 298, 25165, Ondřejov, Czech Republic
Department of Theoretical Physics and Astrophysics, Masaryk University, Kotlářská 2, 61137 Brno, Czech Republic
Petr Kabath
Astronomical Institute, Czech Academy of Sciences,
Fričova 298, 25165, Ondřejov, Czech Republic
Henri M.J. Boffin
ESO, Karl-Schwarzschild-str. 2, 85748 Garching, Germany
Valentin D. Ivanov
ESO, Karl-Schwarzschild-str. 2, 85748 Garching, Germany
Marek Skarka
Astronomical Institute, Czech Academy of Sciences,
Fričova 298, 25165, Ondřejov, Czech Republic
Department of Theoretical Physics and Astrophysics, Masaryk University, Kotlářská 2, 61137 Brno, Czech Republic
(Received: January 3, 2019; Revised: ; Accepted: July 15,2019)
Abstract
The technique of transmission spectroscopy allows us to constrain the chemical composition of the atmospheres of transiting exoplanets. It relies on very high signal-to-noise spectroscopic (or spectrophotometric) observations and is thus most suited for bright exoplanet host stars. In the era of TESS, NGST and PLATO, more and more suitable targets, even for mid-sized telescopes, are discovered. Furthermore, a wealth of archival data is available that could become a basis for long-term monitoring of exo-atmospheres.
We analyzed archival HARPS spectroscopic time series of four host stars to transiting bloated gas exoplanets, namely WASP-76b, WASP-127b, WASP-166b and KELT-11b, searching for traces of sodium (sodium doublet), hydrogen (H$\alpha$, H$\beta$), and lithium (670.8 nm).
The archival data sets include spectroscopic time series taken during transits. Comparing in- and out-of-transit spectra we can filter out the stellar lines and investigate the absorption from the planet. Simultaneously, the stellar activity is monitored using the Mg I and Ca I lines.
We independentely detect sodium in the atmosphere of WASP-76b at a 7–9 $\sigma$ level. Furthermore, we report also at 4–8 $\sigma$ level of significance the detection of sodium in the atmosphere of WASP-127b, confirming earlier result based on low-resolution spectroscopy. The data show no sodium nor any other atom at high confidence levels for WASP-166b nor KELT-11b, hinting at the presence of thick high clouds.
planetary systems — techniques: spectroscopic — planets and satellites: atmospheres —
planets and satellites: individual: WASP-76b, WASP-127b, WASP-166b, KELT-11b
††journal: AJ††software: HARPS Data Reduction Software, iSpec (Blanco-Cuaresma et al. 2014), IRAF (Tody 1986)
1 Introduction
Transmission spectroscopy has the power to probe the physical conditions in
the atmospheres of exoplanets, as it allows detecting the various absorption features that are imprinted on the
stellar light that passes through their atmospheres. Not surprisingly, the
first detections of features in exo-atmospheres were made from space
(Charbonneau et al., 2002; Vidal-Madjar et al., 2003), while ground-based transmission spectroscopy delivered its
first secure detection of sodium in extrasolar atmospheres only about a decade ago
(Redfield et al., 2008; Snellen et al., 2008). If from space, only low-resolution spectroscopy can be used, from the ground, both low-resolution spectrophotometry and high-resolution spectroscopy can and have been used to detect atoms and molecules in the atmospheres of several exoplanets. In particular, in most recent years, high-resolution spectroscopy has been more and more successful. Moreover, Wyttenbach et al. (2015) used HARPS at the ESO 3.6-m telescope to be able to detect
sodium in HD 189733b, demonstrating that spectrographs attached to moderate-size telescopes
can also be used for this purpose. In the past only a handful of suitable, bright enough targets were
available for such kind of study, but this is now changing rapidly with the discoveries from
the TESS mission (Ricker et al., 2015; Kabath et al., 2019).
The high resolving power, i.e.,
R$\geq$40 000–60 000, removes the need for a reference star that is at the same
time close on sky to the target, and of similar brightness – two rarely met
conditions that are limiting the low-resolution spectrophotometry studies. In
contrast, the high resolving power solely requires accurate telluric correction and
a careful analysis of the radial velocities of the star-planet system (e.g., Astudillo-Defru & Rojo, 2013; Wyttenbach et al., 2017).
Another advantage of high-resolution spectroscopy is its ability to resolve and characterize individual planetary absorption lines. In particular, the
Na D lines are well suited for transmission spectroscopy due to the strong expected signal and the fact that they are very narrow features, in comparison to the broad molecular absorption bands of, e.g., $\rm{CH_{4},\ H_{2}O,\ TiO\ and\ VO.}$
If sodium, and in some cases, potassium measurements
based on low-resolution spectra were reported
(e.g., for some of the most recent ones, Sedaghati et al., 2016; Chen et al., 2017; Pallé et al., 2017; Murgas et al., 2017; Lendl et al., 2017; Chen et al., 2018; Pearson et al., 2019), only a few studies detected sodium using high-resolution transmission spectroscopy:
Wyttenbach et al. (2017) registered strong absorption signal in WASP-49b.
Using HARPS-N mounted on the TNG, Casasayas-Barris et al. (2017) and Casasayas-Barris et al. (2018) concluded to the presence of sodium in the atmospheres of WASP-69b and MASCARA-2b. Khalafinejad et al. (2018) tentatively detected sodium in WASP-17b using the MIKE instrument on the 6.5m Magellan Telescope. Finally, Jensen et al. (2018) detected sodium and hydrogen in WASP-12b using the Hobby-Eberly Telescope and Deibert et al. (2019) found sodium in HAT-P-12b using the Subaru Telescope.
These are not the only elements to have been discovered in the atmospheres of exoplanets, as there are published detection of hydrogen (Ehrenreich et al., 2008; Lecavelier Des Etangs et al., 2010), helium
(Nortmann et al., 2018; Spake et al., 2018), calcium, scandium (Astudillo-Defru & Rojo, 2013), TiO and VO
(Désert et al., 2008; Sedaghati et al., 2017), CO (Snellen et al., 2010), H${}_{2}$O and CH${}_{3}$
(Tinetti et al., 2007; Swain et al., 2008; Birkby et al., 2013). Some of these detections were disputed by
Gibson et al. (2011) who discusses in great detail the complex systematics that affect
the results, even those obtained from space. This issue was addressed in recent
years by paying closer attention to the systematic effects (e.g., Sedaghati et al., 2015) and by
introducing new stable instruments.
It is also important to note that there is a tremendous amount of archival spectra – obtained during transits for other purposes, such as measuring the spin-orbit alignment measurements via the Rossiter-McLaughlin effect (Queloz et al., 2000) – that are available for further analysis. They can set the foundation of long-term exoclimate studies. Therefore, the method of high-resolution transmission spectroscopy from the ground is likely to become even more widely used than today, and to gain further importance for the characterization of exoatmospheres.
Here we investigate the atmospheric chemistry of two inflated super-Neptunes and two inflated hot-Jupiters with archival HARPS data111Data were obtained under Programme IDs: 090.C-0540(F), 098.C-0304(A) and 0100.C-0750(A). We study the instrument systematics that can influence the results and the effects of the stellar activity (see Klocová et al., 2017, for further discussion) using the calcium lines as activity indicators (Khalafinejad et al., 2017). This work is part of a larger effort to characterize systematically and uniformly a large number of exoplanetary atmospheres, and to shed more light onto their chemical composition. We also aim to identify interesting targets for CHEOPS and JWST follow-up.
2 Data sets and their analysis
2.1 Data summary
Table 2 shows the various datasets we used, while Table 1 presents the main properties of our targets. For illustration purpose we show part of the spectrum around the sodium Na D region of each of our targets, along with the telluric spectrum in Fig. 1.
WASP-76b:
The HARPS data are covering 3 transit events, on 11/12 November 2012, 24/25 October 2017 and 22/23 November 2017. The total number of frames was $176$ with $105$ in-transit, and the exposure times were between 300-600 seconds.
WASP-127b:
The HARPS data are covering 2 transit events during the nights of 27/28 February and 19/20 March 2017. The total number of frames was $82$, of which 52 were in-transit.
WASP-166b:
The HARPS data are covering three transits during the nights of 13/14 January, 03/04 March and 14/15 March 2017. The total number of frames was $188$, of which 112 were in-transit.
KELT-11b:
The HARPS data are covering 3 transit events and 4 nights of out-of-transit on 01/02 February (transit), 14/15 Februray, 15/16 February (transit), 16/17 February, 05/06 March, 06/07 March (transit) and 07/08 March 2017. The total number of frames was 342, of which 144 were in-transit.
2.2 Data Reduction
We retrieved the publicly available HARPS (Mayor et al., 2003) reduced spectroscopic data sets from the ESO Science data archive for the four planets WASP-76b, WASP-127b, WASP-166b and KELT-11b (see Table 1). The retrieved data are fully reduced products obtained with the HARPS Data Reduction Software (DRS version 3.5). In more detail, each spectrum is provided as a merged 1D spectrum with 0.01 $\AA$ wavelength step. The reduced spectrum covers the wavelength range between 380 nm and 690 nm, and has a spectral resolution R $\approx$ 115 000 corresponding to 2.7 km s${}^{-1}$ per resolving element. Spectra are corrected to the barycentric frame of reference. The sodium doublet is well-centered on the 16th order of HARPS.
2.3 Normalization
We obtained the radial velocities of the star (see below) from the FITS header of the spectra and subsequently shifted both during transit $f(\lambda,t_{\rm{in}})$ and out of transit $f(\lambda,t_{\rm{out}})$ frames to the stellar rest frame. We have made a cut of the region of interest and performed a normalization for the obtained spectra $f(\lambda,t_{\rm{in}})\rightarrow\tilde{f}(\lambda,t_{\rm{in}})$ and $f(\lambda,t_{\rm{out}})\rightarrow\tilde{f}(\lambda,t_{\rm{out}})$, where the tilde represents normalization. Normalization has been carried out using the standard IRAF routine continuum. More precisely, a Legendre polynomial has been used to perform the normalization of continuum regions without strong lines within the regions centred around the Na D lines (583 to 596 nm), H$\alpha$ (645 to 670 nm), H$\beta$ (475 to 495 nm) and lithium (666 to 674 nm).
2.4 Telluric correction
As the next step, the Earth’s atmospheric signature needs to be accounted for. There are several ways to remove telluric lines from the spectra. One of the methods is to simultaneously observe an early type rapidly rotating star to obtain the telluric spectrum. However, this method has a disadvantage of reducing the observing time spent on the primary target. Another method is described in Astudillo-Defru & Rojo (2013) and Wyttenbach et al. (2015) where they use the fact that the variation of the telluric lines depth follows the airmass variation linearly. We decided to follow the method used in Casasayas-Barris et al. (2017). We took the one-dimensional telluric spectrum constructed from the line list of HITRAN (Rothman et al., 2013) and matched its resolution to the resolution of the data. Then we corrected the reference telluric spectrum for the radial velocity to the same frame as the observed spectrum. The telluric lines absorption depth is changing with the airmass. Thus we used a Least Square Method on the unblended telluric lines to scale the telluric model to the same airmass (depth) as the observed spectrum. Afterwards, we divided each observed spectrum by the telluric spectrum scaled for the given airmass using standard IRAF commands resulting in the removal of the telluric lines. As discussed in Casasayas-Barris et al. (2017), the scaling of the telluric spectra is not perfect as line depth variation is not equal for all the telluric lines, hence possibly introducing small residuals into the final transmission spectrum. As an illustration,
the RMS of a single-night transmission spectrum in the telluric region (589.9 to 590.4 nm) before applying the telluric correction is 0.00339. The RMS of the same region was decreased to 0.00196 after applying the correction. For comparison, the RMS of the spectral region without telluric features (587 to 587.5 nm) is 0.00193.
2.5 Transmission spectrum
After the telluric correction, we averaged all out of transit spectra to create a “master out” frame $\tilde{F}_{\rm out}(\lambda)=\sum_{\rm{out}}\tilde{f}_{\rm{out}}(\lambda,t_{%
\rm{out}})$. Next, we divided each in-transit frame $\tilde{f}_{\rm in}(\lambda,t_{\rm{in}})$ by $\tilde{F}_{\rm out}(\lambda)$ to create the individual transmission spectra. Each of these frames was subsequently corrected for the planetary motion as the planetary signal is shifted from the stellar reference frame. We created a model of radial velocity taking only out-of-transit data222As the planetary signal is shifted by as much as tens of kilometers from the rest frame, using the observed radial velocities would imply correcting for the possible Rossiter-McLaughlin effect and would produce systematic effects in the final transmission spectrum Cegla et al. (2017); Wyttenbach et al. (2017).. Finally, we summed up all the individual transmission spectra and normalized this sum to unity and then subtracted unity to retrieve the final transmission spectrum, $\tilde{\mathcal{R}}(\lambda)$:
$$\tilde{\mathcal{R}}(\lambda)=\sum_{\rm{in}}\left.\frac{\tilde{f}_{\rm{in}}(%
\lambda,t_{\rm{in}})}{\tilde{F}_{\rm{out}}(\lambda)}\right|_{\rm{Planet\ RV\ %
shift}}-1$$
(1)
2.6 Error estimates
The uncertainties of the exoatmospheric Na D lines measurements
were estimated with a simple simulation from the data itself. This
approach has the advantage of preserving as much as possible the
systematic errors of the data.
The starting point of the simulation is our actual measurement –
the fit that reproduces the observed spectrum with a flat line and
two Gaussians (see Sec. 2.5). In the
following we only consider a region of the spectrum containing 190
pixel elements, centred on the Na doublet. We converted the fit
within this region into a noiseless spectrum with the same
wavelength sampling as the original observational data.
The first step of our error analysis is to calculate the residuals
of the observational data with respect to the fit. Second, the
residuals were shifted by one pixel and then added to the best fit
model spectrum. Thus, the residual from the first pixel of the data
was added to the second pixel of the fit, and the residual from
the last pixel of the data was added to the first pixel of the best
fit. This created a new realization of artificial data that have
the same Na D signal and a different noise for each pixel. However,
the overall statistical properties of the noise are preserved: for
example if there is a systematic trend of increasing noise toward
the end of the observations, because of increasing airmass, it will
be preserved. The third step is to remeasure the Na absorptions in
the new artificial spectrum.
The last two steps were repeated 189 times, creating 189 unique new
realizations – each time the shift of the residual pattern was
increased by one, so for the second realization the first deviation
was added to the third pixel and the last deviation was added to
the second pixel. For the third realization the first deviation was
added to the fourth pixel and the last deviation – to the third
pixel. In general, to create the i${}^{\rm th}$ realization we added
the first deviation to the (i+1)${}^{\rm th}$ pixel of the noiseless
model spectrum and the last realization – to the i${}^{\rm th}$
pixel of the noiseless model spectrum. Summarizing, this exercise
is a simple shift of the residuals in a circular manner. It has
been used before, e.g. by Cáceres et al. (2009, 2011).
The measurement errors were derived from the distributions (over
all realizations) of the fitting parameters determined for each
realization in the third step of the process described above.
These distributions are shown in Fig. 2. We adopt as
final measurements and as their errors the means and the FWHMs
values of these distributions.
2.7 Stellar activity
We also checked for possible false-positive detections. As suggested in Barnes et al. (2017), stellar activity could produce signatures in lines originating mainly from the stellar chromosphere. To rule out any such fake signals we investigated the vicinity of Mg I (581.36 nm) and Ca I (612.22 and 616.22 nm) lines. The lack of any features in these regions shows that the increase of absorption in the sodium Na D lines is not caused by the chromospheric activity or by systematic effects in our analysis.
2.8 Rossiter–McLaughlin effect
When looking at the presence of any atomic or molecular species in the spectrum of a planet, it is important to take into account the Rossiter–McLaughlin effect (R-M; Queloz et al., 2000; Ohta et al., 2005), as shown by Wyttenbach et al. (2017). In Fig. 4, we show the measured radial velocities of our targets as obtained with HARPS as well as the expected orbital motion of the star due to the presence of the planet. Any deviation from this could be due to the R-M effect. This is indeed the case for WASP-166, where the R-M effect is clearly visible (as also shown by Hellier et al. (2018)).
For the three other targets, the R-M effect is not measurable.
It is possible to have an order of magnitude estimate of the R-M effect (Triaud, 2018):
$$\Delta v_{r}\simeq\frac{2}{3}\gamma^{2}V\sin i\sqrt{1-b^{2}},$$
(2)
where $V\sin i$ is the rotational velocity of the star, $b$ is the impact parameter and $\gamma=\frac{R_{p}}{R_{s}}$, with $R_{p}$ being the radius of the planet and $R_{s}$ that of the star.
Estimating these values for the 4 targets we get the following values for $\Delta v_{r}$: 2.3 m s${}^{-1}$ (WASP-76), 2 m s${}^{-1}$ (WASP-127), 8 m s${}^{-1}$ (WASP-166), and 4 m s${}^{-1}$ (KELT-11). These are in agreement with what we see in Fig. 4.
Moreover, in all cases, the R-M effect is too small to have any real impact on the depth of the Na D lines (Nortmann et al., 2018). We thus conclude here that the R-M effect is not important in our analysis.
3 Results
Each data set was analyzed for Na D, H${}_{\rm\alpha}$, H${}_{\rm\beta}$ to see the absorption in the exo-atmosphere and for Ca I (612.2217 nm and 616.2173 nm) and Mg I (518.3604 nm) to monitor the stellar activity. The results are described for each star separately.
WASP-76b:
The scale height of the atmosphere of WASP-76b is estimated to be about 1212 km, making it an excellent system for the analysis by transmission spectroscopy (Kabath et al., 2019). This exoplanet was also suggested as one of the primary targets for JWST observations (Mollière et al., 2017). Fig.5 presents the final divided in/out transit spectra corrected for planetary radial velocity, displaying clearly visible planetary sodium (Na D lines) absorption. The RMS of the Na D region is 0.0027 and the RMS of the binned spectrum by 20 is 0.0013. The sodium line depths are $0.65\,\%$ and $0.57\,\%$, respectively, therefore, corresponding to 9.5 $\sigma$ and 7.1 $\sigma$ significance. This is the first detection of sodium in the upper atmosphere of WASP-76b. Details of the detected peaks are summarized in Table 3.
Figure 3 shows the non-significant detections in Ca I and Mg I lines (RMS of the data points is 0.0007), showing that the stellar activity is low and any planetary atmosphere sodium (Na D lines) detection should be significant and not influenced by the star. Furthermore, the star WASP-76 is a slow rotator as reported by Brown et al. (2017).
The D1 line is apparently stronger than the D2 line, which would imply different mixing ratios and compositions of the atmospheres of different planets because all other detections reported were with stronger D2 line. The explanation for the strength of the lines can be found in the literature for other Na lines in redder regions (Civiš et al., 2012), but such a measurement is missing in the optical. Civiš et al. (2012) present a laboratory spectrum of sodium lines demonstrating
that different atomic states are shaping the spectral lines and their strengths.
Furthermore, Slanger et al. (2005) measured the D2/D1 ratio in the Earth’s
atmosphere and found it to be varying between 1.2-1.8. They explained that the variability of the
ratio is originating from a competition between Oxygen reacting with
NaO($A^{3}\Sigma^{+}$), produced from the reaction of sodium with Oxygen. In other words, the line strength and the mixing ratio depend on the composition of the surrounding
atmosphere and it can even vary over time.
In the course of the review process of this paper, an independent detection of the sodium in the atmosphere of WASP-76b planet by Seidel et al. (2019) was reported for the same HARPS data set. Their results are in agreement with ours, confirming the methodology. We note, however, that our FWHM of the D2 line is about 30 % smaller compared to their values.
WASP-127b:
WASP-127b is a heavily bloated super-Neptune. Its scale height is $2500\pm 400$ km, making it another ideal target for transmission spectroscopy. Indeed, a cloudless sky was observed with many features such as sodium, potassium and lithium detected and hints of TiO absorption were reported with GTC instrumentation (Pallé et al., 2017; Chen et al., 2018).
Fig. 5 presents the final divided in/out transit spectrum. The RMS of the data is 0.0042 and by binning by 20 points we reach a final RMS of 0.0019. We were able to confirm only sodium (Na D lines) with a high significance of 4.2-$\sigma$ (D2) and 8.33-$\sigma$ (D1), respectively. We did not detect lithium (670.8 nm), which is most likely due to aperture differences between the GTC (10 m) and the 3.6 m telescope that hosts HARPS, while as mentioned earlier, potassium (766.5 & 769.9 nm) is out of range of HARPS. However, our confirmation of sodium is unique because this is one of the first proofs for observations using different instruments as well as two different methods.
The stellar activity of WASP-127 star can be monitored with the Ca I and Mg I lines. We were not able to see any activity in these lines, as shown in Fig. 3.
WASP-166b:
WASP-166b is a bloated super-Neptune with a scale height of 1103 km. Fig. 5 presents the final divided in/out transit spectrum of all regions of interest. The RMS of the data is 0.0025 and, after binning by 20 points, 0.0014. However, we are not able to detect any feature in the sodium (Na D) lines. Therefore, we can only put an upper limit of 0.14 $\%$ on the detection.
Fig. 3 displays no signatures in the Mg I and Ca I lines. Therefore, we can assume that the star is not active.
KELT-11b:
The scale height of KELT-11b is about 2500 km and since the planetary atmosphere is bloated, this is another excellent system for detection of exo-atmosphere (Kabath et al., 2019).
Fig. 5 presents the final divided in/out transit spectrum of all regions of interest. The RMS of the data is 0.0012 (unbinned) and 0.0006 (binned by 20 points). The upper limit for a sodium (Na D lines) detection is therefore 0.06 $\%$. This is a very stringent constraint on the presence of sodium in the atmosphere and possibly hints at the presence of clouds in the atmosphere of KELT-11b.
We checked the stellar activity on Ca I and Mg I lines. There was no significant detection of activity observed in our data (see Fig. 3).
3.1 Previous Na D detections
We have investigated all the previous Na D line detections in exoplanets based on high-resolution spectroscopy. We have summarized them in Table 4. As can be seen, WASP-76b is the first reported planet having potentially an absorption ratio of D2 to D1 lines less than one333One should note, however, that the deviation from unity is less than 1-$\sigma$, and we therefore do not claim much about this yet..
The unusual reported values make this planet an ideal target for future JWST follow-up to further clarify this discovery.
4 Conclusions
We investigated a group of four inflated gas giant planets, among which two inflated hot-Jupiters and two super-Neptunes. All four targets are ideal targets for transmission spectroscopy, given their magnitudes and their large scale heights. We reported for the first time the detection of sodium in WASP-76b with a 7–9 $\sigma$ significance.
We have been able to fully resolve the Na D lines. We report for the first time an absorption ratio of D2 to D1 lines less than one, although this is within 1-$\sigma$ of unity. If confirmed, this possibly hints at an unusual composition of the exo-atmosphere.
There was no hydrogen (H$\alpha$ and H$\beta$ lines) detected for WASP-76b. Another planet, WASP-127b was already characterized in previous studies and we could confirm the presence of sodium in the atmosphere. This is to our knowledge the first time such a detection is confirmed, and provides further credence to the transmission spectroscopy method given that two different instruments on two very different aperture telescopes (3.6-m vs. 10-m GTC) provide similar results. For WASP-166b and KELT-11b, we were not able to detect any feature, despite having extremely precise data, thus hinting at cloudy skies or simply the absence of sodium and other elements. We note that non-detections are usually not reported, although they provide useful information as well.
The data presented here are available in the ESO archive under ESO Prog.IDs: 090.C-0540(F), 098.C-0304(A) and 0100.C-0750(A).
PK and JZ would like to acknowledge the support from the GACR international grant 17-01752J. MS acknowledges financial support of Postdoc@MUNI project CZ$.02.2.69/0.0/0.0/16\_027/$ 0008360. This work benefited from support for mobility by the DAAD-18-08 programme, Project-ID: 57388159. We acknowledge the usage of IRAF and iSpec. We would like to
thank the anonymous referee for providing comments that helped to improve the manuscript.
References
Anderson et al. (2014)
Anderson, D. R., Collier Cameron, A.,
Delrez, L., et al. 2014, MNRAS, 445, 1114
Astudillo-Defru & Rojo (2013)
Astudillo-Defru, N., & Rojo, P. 2013, A&A, 557, 56
Barnes et al. (2017)
Barnes, J. R., Jeffers, S. V.,
Anglada-Escudé, G., et al. 2017, MNRAS, 466, 1733
Birkby et al. (2013)
Birkby, J. L.; de Kok, R. J.; Brogi, M. et
al. 2013, MNRAS, 436, L35
Blanco-Cuaresma et al. (2014)
Blanco-Cuaresma, S., Soubiran,
C., Heiter, U., & Jofré, P. 2014, A&A, 569, 111
Brown et al. (2017)
Brown, D. J. A., Triaud, A. H. M. J., Doyle,
A. P., et al. 2017, MNRAS, 464, 810
Cáceres et al. (2009)
Cáceres, C., Ivanov,
V. D., Minniti, D., et al. 2009, A&A, 507, 481
Cáceres et al. (2011)
Cáceres, C., Ivanov,
V. D., Minniti, D., et al. 2011, A&A, 530, 5
Casasayas-Barris et al. (2017)
Casasayas-Barris, N., Palle, E.,
Nowak, G., et al. 2017, A&A, 608, 135
Casasayas-Barris et al. (2018)
Casasayas-Barris, N., Palle, E.,
Yan, F., et al. 2018, A&A, 616, 151
Cegla et al. (2017)
Cegla, H. M., Lovis, C., Bourrier, V., et al. 2017, A&A, 598, L3
Charbonneau et al. (2002)
Charbonneau, D., Brown, T. M., Noyes,
R. W., & Gilliland, R. L. 2002, ApJ, 568, 377
Chen et al. (2017)
Chen, G., Pallé, E., Nortmann, L., et al. 2017, A&A, 600, L11
Chen et al. (2018)
Chen, G., Pallé, E., Welbanks, L., et al. 2018, A&A, 616, 145
Cameron, A., et al. 2015, AJ, 150, 18.
Hellier et al. (2018)
Hellier, C., Anderson, D. R.,
Triaud, A. H. M. J., et al. 2018, arXiv e-prints , arXiv:1811.05292
Civiš et al. (2012)
Civiš, S., Ferus, M., Kubelík,
P., et al. 2012, A&A, 542, 35
Deibert et al. (2019)
Deibert, E. K., de Mooij, E. J W.,
Jayawardhana, R., et al. 2019, AJ, 157, 58D
Désert et al. (2008)
Désert, J.-M., Vidal-Madjar, A.,
Lecavelier Des Etangs, A. et al. 2008, A&A, 492, 585
Ehrenreich et al. (2008)
Ehrenreich, D., Lecavelier Des
Etangs, A., Hébrard, G. et al. 2008, A&A, 483, 933
Gibson et al. (2011)
Gibson, N. P., Pont, F. & Aigrain, S.
2011, MNRAS, 411, 2199
Hellier et al. (2018)
Hellier, C., Anderson, D. R., Triaud,
A. H. M. J., et al. 2018, ArXiv e-prints , arXiv:1811.05292
Jensen et al. (2012)
Jensen, A. G., Redfield S., Endl M., Cochran, W. D., et al. 2012, 751, 86J ApJ
Jensen et al. (2018)
Jensen, A. G., Cauley, P., Redfield S., Cochran, W. D., Endl, M. 2018, 156, 154 AJ
Kabath et al. (2019)
Kabath, P., Zak, J., Boffin, H., et al., accepted PASP
Khalafinejad et al. (2017)
Khalafinejad, S., von Essen, C.,
Hoeijmakers, H. J., et al. 2017, A&A, 598, 131
Khalafinejad et al. (2018)
Khalafinejad, S., Salz, M., Cubillos, P., et al. 2018, A&A, 618, 98
Klocová et al. (2017)
Klocová, T., Czesla, S.,
Khalafinejad, S., et al. 2017, A&A, 607, A66
Lam et al. (2017)
Lam, K. W. F., Faedi, F., Brown, D. J. A., et al. 2017, A&A, 599, A3
Lendl et al. (2017)
Lendl, M., Cubillos, P. E., Hagelberg, J. et
al. 2017, A&A, 606, 18
Lecavelier Des Etangs et al. (2010)
Lecavelier Des Etangs,
A., Ehrenreich, D., Vidal-Madjar, A. et al. 2010, A&A, 514, 72
Mayor et al. (2003)
Mayor, M., Pepe, F., Queloz, D., et al. 2003,
The Messenger, 114, 20
Mollière et al. (2017)
Mollière, P., van Boekel, R.,
Bouwman, J., et al. 2017, A&A, 600, A10
Murgas et al. (2017)
Murgas, F., Pallé, E., Parviainen, H.,
et al. 2017, A&A, 605, 114
Nortmann et al. (2018)
Nortmann, L., Pallé, E., Salz, M. et
al. 2018, Sci, 362, 1388
Ohta et al. (2005)
Ohta, Y., Taruya, A., & Suto, Y. 2005, ApJ,
622, 1118
Pallé et al. (2017)
Pallé, E., Chen, G., Prieto-Arranz, J., et
al. 2017, A&A, 602, L15
Pearson et al. (2019)
Pearson, K. A., Griffith, C. A.,
Robert, T. Z., et al. 2019, AJ, 127, 21P
Pepper et al. (2017)
Pepper, J., Rodriguez, J. E.,
Collins, K. A., et al. 2017, AJ, 153, 215
Redfield et al. (2008)
Redfield, S., Endl, M., Cochran,
W. D., & Koesterke, L. 2008, ]apj, 673, L87
Ricker et al. (2015)
Ricker, G.R., Winn, J.N., Vanderspek, R., et
al. 2015, Journal of Astronomical Telescopes, Instruments, and Systems 1(1), 014003
Rothman et al. (2013)
Rothman, L. S., Gordon, I. E., Babikov, Y., et
al. 2013, J. Quant. Spec. Radiat. Transf., 130, 4
Sedaghati et al. (2015)
Sedaghati, E., Boffin, H. M. J., Csizmadia,
S., et al. 2015, A&A, 576, L11.
Sedaghati et al. (2016)
Sedaghati, E., Boffin, H. M. J., Jeřabková, T., et al. 2016, A&A, 596, A47
Sedaghati et al. (2017)
Sedaghati, E., Boffin, H. M. J., MacDonald,
R. J., et al., 2017, Nature, 549, 238
Seidel et al. (2019)
Seidel, J. V., Ehrenreich, D., Wyttenbach, A., et al. 2019, arXiv e-prints , arXiv:1902.00001
Slanger et al. (2005)
Slanger, T., Cosby, T., Huestis, D. L., et al. 2005,
Journal of Geophysical Research, Vol. 110, D23302.
Smette et al. (2015)
Smette, A., Sana, H., Noll, S., et al. 2015,
A&A, 576, A77.
Snellen et al. (2008)
Snellen, I. A. G., de Kok, R. J., de
Mooij, E. J. W. & Albrecht, S. 2008, Nature, 465, 1049
Snellen et al. (2010)
Snellen, I. A. G., Albrecht, S., de
Mooij, E. J. W., et al. 2008, A&A, 487, 357
Spake et al. (2018)
Spake, J. J., Sing, D. K., Evans, T. M. et
al. 2018, Nature, 557, 68
Swain et al. (2008)
Swain, M. R., Vasisht, G. & Tinetti, G.
2008, Nature, 452, 329
Tinetti et al. (2007)
Tinetti, G., Vidal-Madjar, A., Liang,
M.-C. et al. 2007, Nature, 448, 169
Tinetti et al. (2007)
Tinetti, G., Vidal-Madjar, A., Liang,
M.-C. et al. 2007, Nature, 448, 169
Tody (1986)
Tody, D. 1986, Proc. SPIE, 627, 733
Triaud (2018)
Triaud, A. H. M. J. 2018, Handbook of
Exoplanets, 2
Queloz et al. (2000)
Queloz, D., Eggenberger, A., Mayor, M. et
al. 2000, A&A, 359, L13
Vidal-Madjar et al. (2003)
Vidal-Madjar, A., Lecavelier des
Etangs, A., Désert, J.-M., et al. 2003, Nature, 422, 143
West et al. (2016)
West, R. G., Hellier, C., Almenara,
J.-M., et al. 2016, A&A, 585, 126
Wyttenbach et al. (2015)
Wyttenbach, A., Ehrenreich, D., Lovis, C.,
Udry, S., & Pepe, F., 2015, A&A, 577, 62
Wyttenbach et al. (2017)
Wyttenbach, A., Lovis, C., Ehrenreich, D.,
et al. 2017, A&A, 602, 36 |
Exploring Linguistic Style Matching in Online Communities: The Role of Social Context and Conversation Dynamics
Aparna Ananthasubramaniam, Hong Chen11footnotemark: 1, Jason Yan11footnotemark: 1, Kenan Alkiek11footnotemark: 1, Jiaxin Pei11footnotemark: 1,
Agrima Seth11footnotemark: 1, Lavinia Dunagan11footnotemark: 1, Minje Choi11footnotemark: 1, Benjamin Litterer11footnotemark: 1 and David Jurgens
School of Information, University of Michigan
{akananth,hongcc,jasonyan,kalkiek,pedropei,
agrima,laviniad,minje,blitt,jurgens}@umich.edu
denotes equal contribution
Abstract
Linguistic style matching (LSM) in conversations can be reflective of several aspects of social influence such as power or persuasion. However, how LSM relates to the outcomes of online communication on platforms such as Reddit is an unknown question.
In this study, we analyze a large corpus of two-party conversation threads in Reddit where we identify all occurrences of LSM using two types of style: the use of function words and formality. Using this framework, we examine how levels of LSM differ in conversations depending on several social factors within Reddit: post and subreddit features, conversation depth, user tenure, and the controversiality of a comment. Finally, we measure the change of LSM following loss of status after community banning.
Our findings reveal the interplay of LSM in Reddit conversations with several community metrics, suggesting the importance of understanding conversation engagement when understanding community dynamics.
1 Introduction
Social influence can be subtle. When two persons converse, their interpersonal dynamics can lead to one person adopting the language of the other. For example, in settings where one person has higher status or power, the lower-status person may unconsciously begin mirroring the language of the other (Danescu-Niculescu-Mizil et al., 2012). This process has been described as accommodation Giles and Ogay (2007) or linguistic style matching (LSM) Niederhoffer and Pennebaker (2002) and can reflect the underlying influence that individuals have on each other Chartrand and Bargh (1999). Past work has primarily focused on how linguistic influence changes relative to the identities of the speakers. However, the larger social context in which a conversation happens also plays a role in determining whether an individual may be influential. Here, we perform a large-scale study of linguistic influence to test how specific types of social context influence the level of accommodation.
Past work in the social sciences has studied accommodation to understand the influence and social power dynamics in specific settings, like job interviews (applicants and interviewers) Willemyns et al. (1997) and academic context (students and faculty)Jones et al. (1999). Also, LSM has been studied to understand group dynamics (Gonzales et al., 2010) and negotiations (Ireland and Henderson, 2014). Work in NLP has operationalized these theories to test accommodation theory in new domains. Typically, these works adopt some tests for measuring influence in language and have shown these measures correlate with known social differences. However, it is yet unknown how LSM occurs in conversations in online community platforms and differs by community dynamics.
Our work examines the larger context in which linguistic influence occurs. Using a large sample of 2.3 million conversations from Reddit and two measures of linguistic influence, we test how the level of linguistic influence correlates with conversational outcomes, such as conversation length and even the continued presence of a person in a community. Further, we examine how specific social and contextual factors influence the rates of linguistic influence. For instance, we discover that the controversy level of the parent comment can lead to different dynamics of style matching in the conversation threads.
This paper offers the following three contributions. First, we systematically compare complementary measures of accommodation, showing clear evidence of style accommodation in Reddit conversations. Second, we draw the relationships of several social factors that affect LSM, including levels of engagement, the popularity of the content, and tenure within a subreddit.
Third, we demonstrate the use of LSM to measure the loss of status through the banning of subreddits. We have released all code and data for full reproducibility.111https://github.com/davidjurgens/style-influence
2 Accommodation and its Measurement
In this section, we discuss communication accommodation theory and associated sociolinguistic research to outline the accommodation of communicative behavior based on perceived social power dynamics. Subsequently, we explore the concept of linguistic style matching and methods adopted by researchers to quantify this phenomenon. We also investigate various factors that contribute to LSM variations and their strategic uses.
2.1 Accommodation Theory as Social Influence
When two individuals engage in social interaction, they may either converge or diverge in their communicative behavior. The Communication Accommodation Theory (CAT) suggests that the degree of convergence or divergence is affected by the relative social power between the interlocutors Xu et al. (2018). Asymmetric convergence is more likely to occur in situations where there is a power imbalance between the interlocutors. Individuals with lower social power or status are more likely to adapt their communication style to align with those in higher or dominant positions Muir et al. (2016). For instance, Puerto Ricans in New York City during the 1970s, who were perceived to hold less power than African Americans, adopted the dialect of African Americans to converge with their more powerful counterparts Wolfram (1974).
Social power has been often found to be an important determinant of degrees of accommodation Giles et al. (1991); Ng and Bradac (1993) and interactants of differential social power or social status can act in a complementary fashion Street (1991).
2.2 Linguistic Style Matching
Linguistic alignment is a pervasive phenomenon that occurs in human communication where interactants unconsciously coordinate their language usage. This coordination, described as convergence in the psycholinguistic theory of communication accommodation, involves aspects such as word choice, syntax, utterance length, pitch, and gestures Giles et al. (1991). Linguistic style matching (LSM) is a specific manifestation of linguistic alignment, wherein individuals unconsciously match their speaking or writing styles during conversations Ireland et al. (2011). Unlike content accommodation, LSM focuses on stylistic accommodation, examining how things are communicated rather than what they communicate.
Individuals strategically negotiate their language style to decrease social distance, seek approval, and accommodate each other. LSM can also reflect the level of common understanding and conceptualization of the conversation topic between speakers. The degree of LSM can indicate social power dynamics as indicated by Giles and Ogay (2007). Empirical evidence from recent studies Danescu-Niculescu-Mizil et al. (2012) showed that participants with less power (such as lawyers or non-administrative roles in Wikipedia) exhibit greater coordination in conversational behavior than participants with high power (such as justices or administrators). Additionally, Noble and Fernández (2015) identified a positive correlation between linguistic accommodation and social network centrality, which effect can be greater than the effect of power status distinction. Studies by Muir et al. (2016, 2017) further show that individuals in a lower position of power tend to accommodate their linguistic style to match that of their higher-power counterparts during face-to-face communication as well as computer-mediated communication.
The variance in LSM can be attributed to various social and psychological factors and can be triggered for different purposes. Linguistic alignment may signal likability and agreement, relate to seeking approval or arise from social desirability. Higher levels of accommodation in social behaviors are found to be associated with increased feelings of affiliation, liking, and successful interpersonal relationships Bayram and Ta (2019).
Thus, linguistic alignment can be strategically employed to establish relationship initiation and stability Ireland et al. (2011), increase group cohesion, and task performance Gonzales et al. (2010), and assist in negotiations Taylor and Thomas (2008). Furthermore, alignment has been found to enhance persuasiveness, motivating listeners to adopt healthier practices Cialdini (2001) while in some cases like presidential debates, it has been perceived as more aggressive Romero et al. (2015). The degree of matching may differ based on context and individual factors.
3 Data
Reddit is a popular social media platform with a forum-based interface. It allows users to interact with dispersed individuals who share similar experiences or topics of interest. Our dataset to study LSM spans from July 2019 to December 2022 and includes 35M users and 500K subreddits.
Using the Pushshift Reddit Dataset which contains the full history of comments aggregated on a monthly basis (Baumgartner et al., 2020), we construct conversation threads from the comments and filter those that satisfy the following conditions: (1) the conversation chain consists of exactly two users; (2) the beginning of the conversation chain must be a root comment which does not have a parent comment; and (3) the lengths of a conversation chain must between 3 and 100.
These conditions allow us to capture conversation dynamics between exactly two users without any interference. Our resulting dataset contains 16,893,013 conversation turns (or comments) across 2,305,775 conversation chains from 68,788 subreddits.
4 How should we measure linguistic influence?
Computational work has proposed multiple approaches for both what to measure and how to measure linguistic influence. In this section, we aim to build intuition for what the two measures of accommodation—using function words and formality—are operationalizing.
4.1 Linguistic Style Markers
Our study measures linguistic influence with two complementary style markers. We use the notation $m$ to refer to a marker throughout.
Marker 1: Function Words
Function words (e.g. pronouns, prepositions, articles, and auxiliary words) are primarily employed unconsciously and frequently and incorporate social knowledge for comprehension and usage Meyer and Bock (1999); Ireland and Pennebaker (2010).
Prior computational studies of linguistic accommodation have measured linguistic influence by tracking the relative frequencies of function words across conversation turns (Danescu-Niculescu-Mizil et al., 2011; Babcock et al., 2014; Gonzales et al., 2010).
Function words reflect how content is expressed, rather than what specific content is expressed (e.g., content words) and are thought to be a better proxy for unconscious language processing Tausczik and Pennebaker (2010). Here, we use the function words defined by the Linguistic Inquiry and Word Count (LIWC) lexicon Pennebaker et al. (2001); Pennebaker and Chung (2007).
Marker 2: Formality
Individuals adopt a specific register that is appropriate to their position in the social context, real or desired Niederhoffer and Pennebaker (2002). A commonly varied register is the level of formality used when speaking to another. The level of formality shown by a speaker is known to reflect the speaker’s opinion towards a topic or their closeness to the listener (Hovy, 1987).
Unlike function words, variation in formality often requires conscious processing to select the appropriate phrasing in a given circumstance. As a result, it offers a complementary view into how a speaker influences another through shifting the conversation towards a more formal or informal register.
Here, we measure formality using a supervised classification model. The model is a fine-tuned RoBERTa-based classifier (Liu et al., 2019) trained on the GYAFC (Rao and Tetreault, 2018) and Online Formality Corpus (Pavlick and Tetreault, 2016) datasets; we use the model available from the Hugging Face API222https://huggingface.co/s-nlp/roberta-base-formality-ranker. Both datasets contain social media text and the reported model performance is high for both blogs and Q&A text (Spearman’s $\rho$>0.7).
Using this classifier, each comment’s formality is measured on a continuous scale in [0,1].
Importantly, these style variables are related; function word frequency also changes in more formal contexts, where articles and prepositions typically become more common while pronouns and interjections become less common (Heylighen and Dewaele, 1999).
Content word-based measures of style and function word counts are thought to capture the same latent style variables, i.e., they are interchangeable at a stylometric level (Grieve, 2023).
4.2 Measuring Linguistic Influence
At a high-level, linguistic influence (also referred to as LSM or accommodation in this paper) is measured by testing whether the value for some measure $m$ of a comment made by user $a$ is predictive of the value of $m$ in the reply to that comment by user $b$.
Therefore, one straightforward way to measure accommodation is with linear regression: $m_{b}\sim\beta_{0}+\beta_{1}m_{a}$ where $\beta_{0}$ reflects the baseline level of the measure (e.g., the average formality) and $\beta_{1}$ measures the level of accommodation (e.g., the average increase in formality associated with a 1-unit increase in the formality of the parent comment).
However, as Xu et al. (2018) note, the characteristics of a comment are likely influenced by other unrelated factors such as the length of the comment or the number of turns in the conversation. Indeed, they show that unless one controls for such factors, linguistic influence may be overestimated.
Therefore, we used a mixed-effects regression to control for comment $a$ and $b$’s length in tokens (fixed effects $L_{a}$, $L_{b}$), the number of replies $r_{b\rightarrow a}$ that $b$ has made to $a$ so far in the conversation.
To capture individual and community-level variation, we include random effects to control for the effect of the subreddit $s$; these random effects let us control for differences in the norms of communities (e.g., some communities are more/less formal) to test for relative changes in $m$. Linguistic accommodation is modeled as
$$\displaystyle m_{b}\sim$$
$$\displaystyle\beta_{0}+\beta_{1}m_{a}+$$
$$\displaystyle\beta_{2}L_{a}+\beta_{3}L_{b}+\beta_{4}r_{b\rightarrow a}+(1|s)$$
where $\beta_{1}$ measures the level of accommodation.
4.3 Results
We first observe clear evidence of accommodation in both style markers: parent comments with more function words receive replies with more function words (Figure 0(a)), and more formal parent comments receive more formal replies (Figure 0(b)).
For comments where we have the text of the original post, we observe accommodation even after controlling for the author and original post’s style markers, suggesting that users may accommodate to the style of the person they are interacting with in the comment thread. However, this effect plateaus when the parent comment has above-average levels of a style marker, suggesting a potential threshold for the impact of parent comment style on reply style. This attenuation of effect may be the result of several mechanisms, including regression to the mean or an author modulating their replies according to their own personal style (i.e., a more extreme parent comment may trigger greater modulation).
Second, the two style markers are almost perfectly uncorrelated, suggesting that they measure distinct constructs. In order to calculate the correlation between these two measures, we randomly sample 1,000 subsets of the conversation turns and calculate the extent of accommodation in function words and formality in that subset.
The correlation between the function-word- and formality-based accommodation scores is -0.00171.
Third, accommodation in the two style markers seems to occur via fundamentally distinct psychological processes. Accommodation can occur either 1) through a subconscious priming mechanism, where the speaker instinctively repeats what they hear; or 2) through a more conscious, strategic act with communicative intent (Doyle and Frank, 2016).
Figure 2 suggests that function-word-accommodation seems to be an unconscious form of relating to the audience, while formality-accommodation seems to be more intentional and strategic. Commenters exhibit greater accommodation in function words when they take less time to reply to the prior comment (1(a)) and greater accommodation in formality when they reply more slowly (1(b)). These results are consistent with prior work, suggesting that accommodation of function words occurs subconsciously (reflexively, takes less time) and builds on this work to show that accommodation in other style markers, like formality, occurs strategically (intentionally, takes more time).
Fourth, there is little variation in accommodation across subreddit characteristics. Figure 3 shows the levels of accommodation across ten different types of subreddits, using an existing taxonomy of popular subreddits.333https://www.reddit.com/r/ListOfSubreddits/wiki/listofsubreddits/ While certain types of subreddits (e.g., lifestyle) tend to have higher levels of accommodation than others (e.g., technology, entertainment), most differences are only weakly significant ($p>0.01$) with a small effect size. Moreover, Figure 4 shows the relationship between subreddit size and variation in linguistic style, for 300 subreddits sampled based on their number of subscribers. To calculate variation in linguistic style, we use Biber (1988)’s comprehensive set of linguistic features. Linguistic variation within each subreddit is estimated as the mean Shannon Entropy of each Biber tag frequency at the subreddit level.
Despite expectations that larger communities may exhibit greater diversity in language use (Kocab et al., 2019), we find no relationship between community size and linguistic variation.
Overall, these findings point to the nuanced dynamics of LSM in online interactions, indicating that factors such as function word usage and formality in the parent comment are associated with the linguistic style and tone of replies.
5 What factors about a comment influence the degree of accommodation?
LSM can be affected by many factors and existing studies have pointed out the roles of not only linguistic characteristics but also the contextual factors affecting LSM (Niederhoffer and Pennebaker, 2002). In this section, we study the connection between LSM and a series of contextual factors where the comment is posted (i.e., comment depth) and the “success” of a comment (i.e., comment Karma and parent comment Karma).
5.1 Experimental Setup
To test for heterogeneity in the level of accommodation with respect to several covariates (e.g., depth, Karma), we run a mixed effects regression similar to Section 4.2, but include an interaction term to test whether accommodation changes significantly with respect to some covariate (say, Karma $K$):
$$\displaystyle m_{b}\sim$$
$$\displaystyle\beta_{0}+\beta_{1}m_{a}+\beta_{2}K+\beta_{3}m_{a}*K+$$
$$\displaystyle\beta_{4}L_{a}+\beta_{5}L_{b}+\beta_{6}r_{b\rightarrow a}+(1|b)+(1|s)$$
Here, $\beta_{1}$ measures the level of accommodation when $K=0$ and $\beta_{3}$ measures the increase in accommodation when $K$ increases by one point; if $\beta_{3}$ is significantly different from $0$, then we have evidence that accommodation is heterogeneous with respect to Karma.
In order to visualize these effects, we fit the model in the above equation to estimate accommodation at different values of Karma. In order to appropriately represent uncertainty in this model, we sample 100,000 conversation turns at each value of Karma 10 times and use this to obtain 10 different estimates of accommodation for each value of the covariate. To visualize the association between Karma and accommodation, we plot Karma on the x-axis and the LSM estimates on the y-axis.
5.2 Results
As shown in Figure 5, various factors of comments are related to LSM.
Comment depth
Comment depth reflects the position of a comment in the conversation tree. Deeper comments are usually posted in longer conversations and when the users are more engaged in the dialogue. As shown in Figure 4(a) and Figure 4(c), comment depth is positively correlated with LSM. However, accommodation in formality drops off for very deep comments. LSM happen more when the comment is deeper in the conversation tree, suggesting that users tend to match not only the content but also the structural aspects of their language in response to their interlocutor. Such a trend could be due to greater investment in the conversation. When two users are involved in longer and deeper conversations, they are more likely to be engaged in the conversation, which may lead to higher subconscious but lower conscious LSM.
Comment Karma
A key feature of Reddit is the ability for users to upvote or downvote comments, which determines the comment’s karma - a measure of its popularity within the community. In figure 5, we observe several non-linear associations between karma, comment characteristics, and LSM. In terms of comment karma, users’ LSM tends to remain relatively constant, except for cases where the comment has very high karma, which is associated with an increase in LSM. This finding implies that highly popular comments may foster greater linguistic alignment between users.
We also see that comments with low karma have lower levels of LSM than comments with high karma (Figure 4(d)), which makes sense since we’d expect users to respond better to comments whether the author is mirroring their interlocutor. Notably, this upward trend reverses in comments with very high karma – which have lower levels of LSM than comments with lower levels of karma. The reversal of the LSM trend in comments with high karma warrants further exploration. One possible explanation for this phenomenon is that highly upvoted comments may exhibit unconventional linguistic styles that deviate from the norm, which could be seen as novel by the Reddit community. Another explanation may be that comments with high karma are more likely to be popular in larger, diverse communities where users may have a wider range of linguistic styles. Additionally, it is possible that comments with high karma receive a higher volume of comments and interactions, which may dilute the overall LSM score due to the presence of diverse linguistic styles from multiple interlocutors.
6 What effect does accommodation have on the conversation itself?
Linguistic accommodation is usually associated with positive social benefits Taylor and Thomas (2008); Gonzales et al. (2010). Here, we test whether linguistic accommodation is associated with two positive behaviors in social media: sustained conversation and length of participation in a subreddit.
6.1 Experimental Setup
We fit a linear regression on conversational dyads following the LSM measure in Section 4.2. Following the procedure from the prior section, we estimate the level of accommodation for comments around a particular covariate by sampling 100,000 conversation turns at or near the respective value of the covariate. Once again, we verify that differences between covariates are significant, by introducing interaction terms in the regression and testing for a statistically significance effect.
6.2 Results
Figures 6a and 6b compare the effect of alignment when conditioned on the total length of the conversation thread. For both functions words and formality, we observe from the fitted lines that accommodation is more likely to happen from longer conversations, but only up to a certain length of approximately 30-40. This suggests the possibility of LSM being an earlier indicator of how engaged the users will be in a conversation. On the other hand, the likelihood of accommodation in formality decreases when the conversation becomes longer than a certain threshold, which suggests that speakers may stop consciously trying to accommodate once the conversation becomes sufficiently long.
Figures 6c and 6d compare accommodation likelihoods at a given turn within a conversation. Interestingly, we can observe that LSM starts off highest at the beginning of a conversation and decreases as the number of turns increases. Combining the two results, we can conjecture that while the degree of LSM generally decreases within a conversation thread, the initial levels of LSM observed at the early stages of a conversation can indicate how engaged the speakers will be, which one can use to estimate the overall conversation length.
How does LSM differ by tenure and number of subsequent posts in a subreddit? Figure 7 shows that, for both style markers, users who have a longer tenure in the subreddit or who post more in the subreddit in the next month tend to display higher subconscious and lower conscious LSM. We consider these results as evidence of the “lifespan” of a user’s engagement toward conversations held within that subreddit, and ultimately engagement toward the subreddit itself, which has been noted in prior work (Danescu-Niculescu-Mizil et al., 2013).
7 What effect does the social context have on accommodation: Controversiality?
In this section, we examine whether LSM differs by social contexts that arise during conversations. Specifically, we focus on the controversy level of the parent comment.
In contrast to non-controversial issues, controversial issues lead to competitive disagreement, where the goal of the groups involved in argumentation is to convince the opponent group(s) of the validity of one’s point of view Ilie (2021). The arguments on controversial issues tend to invite strong emotions with negative affect Mejova et al. (2014) and deteriorate the deliberation in the public sphere because interactions often turn uncivil Doxtader (1991).
7.1 Experimental Setup
Following the procedure from the prior section, we estimate the level of accommodation for comments at each covariate, separately for controversial and non-controversial comments.
When a comment or post receives a substantial number of upvotes and downvotes, Reddit automatically designates it as controversial. The exact method used by Reddit to determine controversy remains private. However, the Reddit API offers a binary label indicating whether a comment is controversial or non-controversial Koncar et al. (2021). Approximately 1.30% (n=218,899) of the comments in our sample are labeled as controversial.
We test that differences between conditions are significant with a three-way interaction term in the regression between the parent-comment style, the comment’s Karma (or other covariates) and the comment’s controversiality: $m_{a}\times K\times C$.
7.2 Results
Figure 8 reveals that LSM occurs differently in controversial and non-controversial comments. For both function words and formality, LSM is less likely to occur in controversial rather than non-controversial comments when the conversation length is below a certain threshold (12-14). Interestingly, we see that this trend is strengthened as the conversation length increases. One possible explanation is that controversial comments generate more initial interest that promotes users to engage more in conversations. However, this initial effect is washed away as the conversation takes further turns, and the conversation is less likely to continue due to reasons such as incivility. Non-controversial comments, on the other hand, enjoy less of this initial boost and is more likely to carry on if the users have accommodated each other’s language during their conversation.
With the addition of Karma, we can observe a more complex trend that plays out differently for each style marker. For function words, conversations in controversial comments have a nonlinear relationship that drops as the parent comment’s Karma increases, whereas a weak positive correlation can be observed for non-controversial comments and levels of Karma. In contrast, for formality, LSM occurs most at comments with about 0-5 Karma and decreases for higher Karma for both controversial and non-controversial comments.
Overall, we observe that social contexts that are defined by the community platform such as Karma or controversy have complex, nonlinear effects on how LSM occurs in conversations.
8 Loss of Status via Community Banning
Reddit bans specific subreddit communities as a result of policy violations, such as repeated posting of highly offensive content or lack of moderator oversight Chandrasekharan et al. (2017). When users are highly active in such communities, the ban potentially results in a loss of status, as they are forced to find new communities to participate in.
Here, we test the extent to which users change how they are linguistically influenced by others after such a ban. While prior work has studied how users change after gaining status (e.g., Danescu-Niculescu-Mizil et al., 2012), our unique setting allows us to perform a novel study of the potentially humbling effects of status loss. In addition, a study of the subreddit r/changemyview suggests that formality is (weakly) associated with more effective persuasion on Reddit (Dayter and Messerli, 2022); we hypothesize that users who recently experienced a ban may have multiple pragmatic reasons to accommodate more.
8.1 Experimental Setup
We test for changes to linguistic influence using a pseudo-causal difference-in-difference analysis Lechner et al. (2011). Subreddit ban dates were determined by identifying all banned subreddits and then using the last date of a post in that subreddit. Our sample includes 1,024 subreddits banned between July 2019 and December 2022. We identify 16,686 users in our sample who made at least one comment in these subreddits in the 30 days before their ban. Each user from a banned subreddit is considered as treated and matched with a control user who did not participate in that subreddit.
Three analyses of the effect of the ban are performed, controlling for user-level and temporal factors. First, we estimate the effect of commenting in a banned subreddit, by comparing posts made in banned subreddits $t$ months before the ban to posts made by the same users at the same time, in other subreddits. Second, using a difference-in-differences approach, we estimate the effect of banning a subreddit on authors’ use of accommodation in (unbanned) subreddits they were active in for $t$ months before and after the ban. This second analysis measures the spill-over effects of the ban on users’ behaviors in other subreddits; the difference-in-differences estimator uses users active in these subreddits at the same time, but not in a banned subreddit, as a control for temporal and subreddit-level effects. Third, we calculate the effect of the ban on commenting behavior in subreddits users migrated to (i.e., newly joined) after the ban was enacted. The difference-in-differences estimator compares accommodation in comments in the banned subreddits to comments in the subreddits these users migrated to; to isolate the effect of migration, the difference between the comments in the migrated and banned subreddits are compared against the spill-over effects in other subreddits that users were a part of during this time.
8.2 Results
Our results suggest that policy actions on Reddit, such as banning, have an effect on the level of accommodation by users. First, the level of subconscious accommodation tends to be lower in banned subreddits than other subreddits the users comment in during the 30 days before the ban (the effects are all below 0 in Figure 8(c) ($p<2e-16$).
Second, following the banning of a subreddit, users tend to change their LSM levels in other subreddits: Figure 9 shows that function-word-mirroring (Figure 8(a)) and formality-mirroring (Figure 8(b)) increase after a subreddit is banned. Our results suggest that users who had previously been active in banned subreddits may have been making an effort to index agreeableness by accommodating (e.g., to avoid losing status in another community).
Third, changes in accommodation are initially amplified in subreddits that these users migrate to after their original community was banned. The comments left by these users in banned subreddits exhibit higher levels of accommodation than would be expected immediately before the ban and maintain higher subconscious accommodation in subreddits they migrated (Figures 8(c) and 8(d)$p<2e-16$). Since function-word mirroring is likely subconscious and formality-mirroring strategic (Section 4), our results suggest that users who had previously been active in banned subreddits may have, intrinsically, indexed agreeableness by accommodating (e.g., to gain status in their new community) but without making a conscious effort (e.g., because they were upset about the loss of a status). These users also increased LSM in the subreddit immediately before it was banned (e.g., perhaps to index agreeableness when warnings about the ban were issued).
9 Discussion and Conclusion
In this study, we performed a large-scale computational analysis on Reddit conversations to understand when LSM occurs and its effect on platform engagement. Overall, do our findings indicate that LSM frequently occurs in online conversations within Reddit, and that it exhibits complex nonlinear relationships with conversation metrics such as Karma, conversation lengths, or controversy scores, which suggests linguistic influence can affect conversation dynamics. Furthermore, we show that the degree of accommodation in conversations is related to greater levels of engagement both at conversation and platform levels. Our findings highlight the possibility of identifying LSM as an indicator of engagement and civil conversations and suggest ideas for building and maintaining online communities that promote constructive discourse.
In our experiments, we have assumed LSM as a unidirectional concept by measuring the exhibition of a particular style conditioned on the previous turn. However, LSM can occur in several different directions, such as the two speakers converging into a single style or even diverging to separate styles. While not in the scope of this study, the existence of such types of LSM in Reddit conversation threads can be studies in future research.
10 Ethical Considerations
This study was conducted only on observational data and did not require any human intervention. We did not use any information that could identify individuals or specific demographic groups, and all of our presented results were obtained through aggregation from millions of users and comments.
Acknowledgments
This material is based in part upon work supported by the National Science Foundation under Grant No IIS-2143529.
References
Babcock et al. (2014)
Meghan J Babcock, Vivian P Ta, and William Ickes. 2014.
Latent semantic similarity and language style matching in initial
dyadic interactions.
Journal of Language and Social Psychology, 33(1):78–88.
Baumgartner et al. (2020)
Jason Baumgartner, Savvas Zannettou, Brian Keegan, Megan Squire, and Jeremy
Blackburn. 2020.
The pushshift reddit
dataset.
Bayram and Ta (2019)
A Burcu Bayram and Vivian P Ta. 2019.
Diplomatic chameleons: Language style matching and agreement in
international diplomatic negotiations.
Negotiation and Conflict Management Research, 12(1):23–40.
Biber (1988)
Douglas Biber. 1988.
Variation across Speech and Writing.
Cambridge University Press.
Chandrasekharan et al. (2017)
Eshwar Chandrasekharan, Umashanthi Pavalanathan, Anirudh Srinivasan, Adam
Glynn, Jacob Eisenstein, and Eric Gilbert. 2017.
You can’t stay here: The efficacy of reddit’s 2015 ban examined
through hate speech.
Proceedings of the ACM on Human-Computer Interaction,
1(CSCW):1–22.
Chartrand and Bargh (1999)
Tanya L Chartrand and John A Bargh. 1999.
The chameleon effect: The perception–behavior link and social
interaction.
Journal of personality and social psychology, 76(6):893.
Cialdini (2001)
Robert B Cialdini. 2001.
The science of persuasion.
Scientific American, 284(2):76–81.
Danescu-Niculescu-Mizil et al. (2011)
Cristian Danescu-Niculescu-Mizil, Michael Gamon, and Susan Dumais. 2011.
Mark my words! linguistic style accommodation in social media.
In Proceedings of the 20th international conference on World
wide web, pages 745–754.
Danescu-Niculescu-Mizil et al. (2012)
Cristian Danescu-Niculescu-Mizil, Lillian Lee, Bo Pang, and Jon Kleinberg.
2012.
Echoes of power: Language effects and power differences in social
interaction.
In Proceedings of the 21st international conference on World
Wide Web, pages 699–708.
Danescu-Niculescu-Mizil et al. (2013)
Cristian Danescu-Niculescu-Mizil, Robert West, Dan Jurafsky, Jure Leskovec, and
Christopher Potts. 2013.
No country for old
members: User lifecycle and linguistic change in online communities.
In Proceedings of the 22nd International Conference on World
Wide Web, WWW ’13, page 307–318, New York, NY, USA. Association for
Computing Machinery.
Dayter and Messerli (2022)
Daria Dayter and Thomas C. Messerli. 2022.
Persuasive language and features of formality on the r/changemyview
subreddit.
Internet Pragmatics, 5(1):165–195.
Doxtader (1991)
Erik W Doxtader. 1991.
The entwinement of argument and rhetoric: A dialectical reading of
habermas’ theory of communicative action.
Argumentation and Advocacy, 28(2):51–63.
Doyle and Frank (2016)
Gabriel Doyle and Michael C Frank. 2016.
Investigating the sources of linguistic alignment in conversation.
In Proceedings of the 54th Annual Meeting of the Association
for Computational Linguistics (Volume 1: Long Papers), pages 526–536.
Giles et al. (1991)
Howard Giles, Nikolas Coupland, and IUSTINE Coupland. 1991.
1. accommodation theory: Communication, context, and.
Contexts of accommodation: Developments in applied
sociolinguistics, 1.
Giles and Ogay (2007)
Howard Giles and Tania Ogay. 2007.
Communication accommodation theory.
In Explaining Communication: Contemporary Theories and
Exemplars, chapter 16. Lawrence Erlbaum Associates, Inc.
Gonzales et al. (2010)
Amy L Gonzales, Jeffrey T Hancock, and James W Pennebaker. 2010.
Language style matching as a predictor of social dynamics in small
groups.
Communication Research, 37(1):3–19.
Grieve (2023)
Jack Grieve. 2023.
Register
variation explains stylometric authorship analysis.
Corpus Linguistics and Linguistic Theory, 19(1):47–77.
Heylighen and Dewaele (1999)
Francis Heylighen and Jean-Marc Dewaele. 1999.
Formality of language: definition, measurement and behavioral
determinants.
Technical report, Center “Leo Apostel”, Free University of
Brussels.
Hovy (1987)
Eduard Hovy. 1987.
Generating natural language under pragmatic constraints.
Journal of Pragmatics, 11(6):689–719.
Ilie (2021)
Cornelia Ilie. 2021.
Discussion, dispute or controversy? paradigms of conflict-driven
parliamentary practices.
Journal of Language Aggression and Conflict, 9(2):237–270.
Ireland and Henderson (2014)
Molly E. Ireland and Marlone D. Henderson. 2014.
Language
style matching, engagement, and impasse in negotiations.
Negotiation and Conflict Management Research, 7(1):1–16.
Ireland and Pennebaker (2010)
Molly E Ireland and James W Pennebaker. 2010.
Language style matching in writing: synchrony in essays,
correspondence, and poetry.
Journal of personality and social psychology, 99(3):549.
Ireland et al. (2011)
Molly E Ireland, Richard B Slatcher, Paul W Eastwick, Lauren E Scissors, Eli J
Finkel, and James W Pennebaker. 2011.
Language style matching predicts relationship initiation and
stability.
Psychological science, 22(1):39–44.
Jones et al. (1999)
Elizabeth Jones, Cynthia Gallois, Victor Callan, and Michelle Barker. 1999.
Strategies of accommodation: Development of a coding system for
conversational interaction.
Journal of Language and Social Psychology, 18(2):123–151.
Kocab et al. (2019)
Annemarie Kocab, Jayden Ziegler, and Jesse Snedeker. 2019.
It takes a village: The role of community size in linguistic
regularization.
Cognitive Psychology, 114:101227.
Koncar et al. (2021)
Philipp Koncar, Simon Walk, and Denis Helic. 2021.
Analysis and prediction of multilingual controversy on reddit.
In 13th ACM Web Science Conference 2021, pages 215–224.
Lechner et al. (2011)
Michael Lechner et al. 2011.
The estimation of causal effects by difference-in-difference methods.
Foundations and Trends® in Econometrics,
4(3):165–224.
Liu et al. (2019)
Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer
Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019.
Roberta: A robustly optimized bert pretraining approach.
arXiv preprint arXiv:1907.11692.
Mejova et al. (2014)
Yelena Mejova, Amy X Zhang, Nicholas Diakopoulos, and Carlos Castillo. 2014.
Controversy and sentiment in online news.
arXiv preprint arXiv:1409.8152.
Meyer and Bock (1999)
Antje S Meyer and Kathryn Bock. 1999.
Representations and processes in the production of pronouns: Some
perspectives from dutch.
Journal of Memory and Language, 41(2):281–301.
Muir et al. (2016)
Kate Muir, Adam Joinson, Rachel Cotterill, and Nigel Dewdney. 2016.
Characterizing the linguistic chameleon: Personal and social
correlates of linguistic style accommodation.
Human Communication Research, 42(3):462–484.
Muir et al. (2017)
Kate Muir, Adam Joinson, Rachel Cotterill, and Nigel Dewdney. 2017.
Linguistic style accommodation shapes impression formation and
rapport in computer-mediated communication.
Journal of Language and Social Psychology, 36(5):525–548.
Ng and Bradac (1993)
Sik Hung Ng and James J Bradac. 1993.
Power in language: Verbal communication and social influence.
Sage Publications, Inc.
Niederhoffer and Pennebaker (2002)
Kate G Niederhoffer and James W Pennebaker. 2002.
Linguistic style matching in social interaction.
Journal of Language and Social Psychology, 21(4):337–360.
Noble and Fernández (2015)
Bill Noble and Raquel Fernández. 2015.
Centre stage: How social network position shapes linguistic
coordination.
In Proceedings of the 6th workshop on cognitive modeling and
computational linguistics, pages 29–38.
Pavlick and Tetreault (2016)
Ellie Pavlick and Joel Tetreault. 2016.
An empirical analysis
of formality in online communication.
Transactions of the Association for Computational Linguistics,
4:61–74.
Pennebaker and Chung (2007)
James W Pennebaker and Cindy K Chung. 2007.
Expressive writing, emotional upheavals, and health.
In Foundations of health psychology. Oxford University Press.
Pennebaker et al. (2001)
James W Pennebaker, Martha E Francis, and Roger J Booth. 2001.
Linguistic inquiry and word count: Liwc 2001.
Mahway: Lawrence Erlbaum Associates, 71(2001):2001.
Rao and Tetreault (2018)
Sudha Rao and Joel Tetreault. 2018.
Dear sir or madam, may
I introduce the GYAFC dataset: Corpus, benchmarks and metrics for
formality style transfer.
In Proceedings of the 2018 Conference of the North American
Chapter of the Association for Computational Linguistics: Human Language
Technologies, Volume 1 (Long Papers), pages 129–140, New Orleans,
Louisiana. Association for Computational Linguistics.
Romero et al. (2015)
Daniel M Romero, Roderick I Swaab, Brian Uzzi, and Adam D Galinsky. 2015.
Mimicry is presidential: Linguistic style matching in presidential
debates and improved polling numbers.
Personality and Social Psychology Bulletin, 41(10):1311–1319.
Street (1991)
RL Street. 1991.
Accommodation in medical consultations.
Contexts of accommodation: Developments in applied
sociolinguistics, pages 131–156.
Tausczik and Pennebaker (2010)
Yla R Tausczik and James W Pennebaker. 2010.
The psychological meaning of words: Liwc and computerized text
analysis methods.
Journal of language and social psychology, 29(1):24–54.
Taylor and Thomas (2008)
Paul J Taylor and Sally Thomas. 2008.
Linguistic style matching and negotiation outcome.
Negotiation and conflict management research, 1(3):263–281.
Willemyns et al. (1997)
Michael Willemyns, Cynthia Gallois, Victor J Callan, and Jeffery Pittam. 1997.
Accent accommodation in the job interview: Impact of interviewer
accent and gender.
Journal of Language and Social Psychology, 16(1):3–22.
Wolfram (1974)
Walt Wolfram. 1974.
Sociolinguistic Aspects of Assimilation: Puerto Rican English
in New York City.
Center for Applied Linguistics.
Xu et al. (2018)
Yang Xu, Jeremy Cole, and David Reitter. 2018.
Not that much power: Linguistic alignment is influenced more by
low-level linguistic features rather than social power.
In Proceedings of the 56th Annual Meeting of the Association
for Computational Linguistics (Volume 1: Long Papers), pages 601–610. |
Acoustic radiation force and torque on an absorbing particle
Glauber T. Silva
Abstract
Exact formulas of the radiation force and torque exerted on an absorbing particle in the Rayleigh scattering limit
caused by an arbitrary harmonic acoustic wave are presented.
They are used to analyze the trapping conditions by a single-beam
acoustical tweezer based on a spherically focused ultrasound beam.
Results reveal that particle’s absorption has a pivotal role in single-beam trapping.
Furthermore, we obtain the radiation torque induced by a Bessel beam in an on-axis particle.
Physical Acoustics Group, Instituto de Física,
Universidade Federal de Alagoas, Maceió, AL 57072-970, Brazil.
[email protected]
PACS numbers: 43.25.Qp, 43.20.Fn, 43.55.Ev
1 Introduction
An increasing interest on acoustic radiation force has arisen after
the concept of acoustical tweezers was developed 1.
An acoustical tweezer can be accomplished by means of an ultrasound
standing wave 1 or a focused beam 2.
Understanding on how Rayleigh particles (i.e. particles much smaller than the incident wavelength) are
trapped is relevant, for instance, in lab-on-a-chip tweezers design 3.
Commonly, particle trapping is described through Gor’kov’s radiation force theory 4.
In turn, this theory is valid when the particle’s viscous boundary layer $\delta$
is much smaller than its radius $a$ 5, where $\delta=\sqrt{2\nu/\omega}$,
where $\nu$ is the kinematic viscosity of the host fluid and $\omega$ is the wave frequency.
Another important effect in radiation force generation is the ultrasound absorption inside the particle.
This effect has been analyzed in the radiation force exerted on a particle by
a plane wave 6.
The purpose of this letter is to extend Gor’kov’s radiation force theory with
absorption effects beyond plane wave results.
Ultrasound absorption is assumed to obey a frequency power-law model 7.
Shear viscosity inside the particle is not taken into consideration and $\delta\ll a$ is assumed .
Apart from radiation force, we also obtain the radiation torque induced
by an arbitrary ultrasound beam on the particle.
Exact formulas for the radiation force and torque are derived from the partial-wave expansion
up to the quadrupole moment 8, 9.
The developed theory is applied in the analysis of the radiation force due to a spherically focused beam.
Results show that transverse trapping depends on particle’s compressibility and density with respect to
the host fluid, while axial trapping also reckon on the particle’s absorption.
In addition, we study the radiation torque produced by a Bessel vortex beam in an on-axis particle.
It is demonstrated that only the first-order Bessel vortex beam can generate radiation torque in this configuration.
2 Rayleigh scattering
Consider an incident acoustic wave of angular frequency $\omega$ propagating
in a nonviscous fluid of density $\rho_{0}$ and speed of sound $c_{0}$.
The wave is scattered by a compressional fluid particle of radius $a$, density $\rho_{1}$,
and speed of sound $c_{1}$.
The incident wave can be represented by the pressure $p$ and the velocity particle ${\bf v}$,
while
the scattered pressure $p_{\text{s}}$ is described in spherical coordinates $(r,\theta,\varphi)$ with respect to
the particle’s center
through the partial-wave expansion method.
In the Rayleigh scattering limit, i.e. the particle is much smaller that the incident wavelength,
only monopole and dipole terms are relevant in the partial-wave expansion.
Accordingly, 10
$$p_{\text{s}}=p_{0}\left\{s_{0}a_{0}^{0}h_{0}^{(1)}(kr)Y_{0}^{0}(\theta,\varphi%
)+s_{1}\left[a_{1}^{-1}Y_{1}^{-1}(\theta,\varphi)+a_{1}^{0}Y_{1}^{0}(\theta,%
\varphi)+a_{1}^{1}Y_{1}^{1}(\theta,\varphi)\right]h_{1}^{(1)}(kr)\right\},$$
(1)
where $p_{0}$ is incident peak pressure magnitude, $k=\omega/c_{0}$ is the wavenumber,
$h_{n}^{(1)}$ the $n$th-order
spherical Hankel function of first-type,
and $Y_{n}^{m}$ is the $n$th-order and $m$th-degree
spherical harmonic.
The quantities $a_{n}^{m}$ and $s_{n}$ are the beam-shape and scattering coefficients, respectively.
The beam-shape coefficient is explained in the Appendix A.
The scattering coefficient is obtained by applying the continuity condition of pressure and particle
velocity across the particle’s surface.
These conditions lead to following scattering coefficient
$$s_{n}=-\det\left[\begin{matrix}\gamma j_{n}(ka)&j_{n}(k_{1}a)\\
j_{n}^{\prime}(ka)&j_{n}^{\prime}(k_{1}a)\end{matrix}\right]\det\left[\begin{%
matrix}\gamma h_{n}^{(1)}(ka)&j_{n}(k_{1}a)\\
{h_{n}^{(1)}}^{\prime}(ka)&j_{n}^{\prime}(k_{1}a)\end{matrix}\right]^{-1},$$
(2)
where
$k_{1}$ in the wavenumber inside the particle, $\gamma=\rho_{0}k_{1}/(\rho_{1}k)$, and
the prime symbol indicates differentiation.
It is assumed that the ultrasound absorption within the particle obeys a frequency power-law
as follows 7
$\alpha=\alpha_{\upsilon}\omega^{\upsilon},$
where $\alpha_{\upsilon}>0$, and $0<\upsilon\leq 2$.
Thus, the inner wavenumber of the particle is expressed as
$k_{1}=\omega_{1}/c_{1}+i\alpha_{\upsilon}\omega^{\upsilon}.$
Our analysis is limited to weak-absorption, thus, $\alpha_{\upsilon}\omega^{\upsilon}\ll\omega/c_{1}$ is taken.
Now, we may expand the scattering coefficients in Eq. (2) for $a\rightarrow 0$, $\rho_{1}/\rho_{0}=O(1)$,
and $c_{0}/c_{1}=O(1)$.
Hence, we find the first relevant terms of this expansion as
$$\displaystyle s_{0}$$
$$\displaystyle=-\left(\frac{2\alpha_{\upsilon}\omega^{\upsilon-1}\rho_{0}c_{0}^%
{2}}{\rho_{1}c_{1}}+if_{0}\right)\frac{(ka)^{3}}{3}-f_{0}^{2}\frac{(ka)^{6}}{9},$$
(3)
$$\displaystyle s_{1}$$
$$\displaystyle=-if_{1}\frac{(ka)^{3}}{6}-\frac{2\rho_{0}\rho_{1}c_{1}\alpha_{%
\upsilon}\omega^{\upsilon-1}}{5(\rho_{0}+2\rho_{1})^{2}}(ka)^{5}-f_{1}^{2}%
\frac{(ka)^{6}}{36},$$
(4)
where $f_{0}=1-\rho_{0}c_{0}^{2}/\rho_{1}c_{1}^{2}$ and $f_{1}=2(\rho_{1}-\rho_{0})/(2\rho_{1}+\rho_{0})$ are the monopole and the dipole scattering factors, respectively.
3 Radiation force
The linear momentum of the incident wave might be transferred
to the suspended particle generating the so-called acoustic radiation force.
This force can be expressed as
${\bf f}=\pi a^{2}(I_{0}/c_{0}){\bf Y}$, where
$I_{0}=p_{0}^{2}/2\rho_{0}c_{0}$ is the averaged incident intensity and
${\bf Y}$ is the dimensionless radiation force.
In the Rayleigh scattering limit, the radiation force vector has the following Cartesian components 8, 11
$$\displaystyle Y_{x}+iY_{y}$$
$$\displaystyle=\frac{i}{2\pi(ka)^{2}}\biggl{[}\sqrt{\frac{2}{3}}\bigl{[}(s_{0}+%
s_{1}^{*}+2s_{0}s_{1}^{*})a_{0}^{0}a_{1}^{1*}+(s_{0}^{*}+s_{1}+2s_{0}^{*}s_{1}%
)a_{0}^{0*}a_{1}^{-1}\bigr{]}$$
$$\displaystyle+\sum_{m=-1}^{1}\sqrt{\frac{(2+m)(3+m)}{15}}\bigl{(}s_{1}a_{1}^{m%
}a_{2}^{m+1*}+s_{1}^{*}a_{1}^{-m*}a_{2}^{-m-1}\bigr{)}\biggr{]},$$
(5)
$$\displaystyle Y_{z}$$
$$\displaystyle=\frac{1}{\pi(ka)^{2}}\textrm{Im}\left[\sqrt{\frac{1}{3}}(s_{0}+s%
_{1}^{*}+2s_{0}s_{1}^{*})a_{0}^{0}a_{1}^{0*}+\sum_{m=-1}^{1}\sqrt{\frac{(2-m)(%
2+m)}{15}}s_{1}a_{1}^{m}a_{2}^{m*}\right].$$
(6)
where the symbol ${}^{*}$ means complex conjugation and ‘Im’ signifies the imaginary-part of.
Note that the quadrupole mode of the incident wave $a_{2}^{m}$ is necessary to compute the radiation force.
By substituting the equations in (24) into
Eqs. (5) and (6), one obtains after some algebra
$${\bf f}=-\frac{2\pi}{k^{2}c_{0}}\text{Re}\bigg{[}\frac{3is_{1}}{k}c_{0}\nabla%
\left(\frac{\rho_{0}\|{\bf v}(0)\|^{2}}{2}-\frac{|p(0)|^{2}}{2\rho_{0}c_{0}^{2%
}}\right)+(s_{0}+3s_{1}+2s_{0}s_{1}^{*})p(0){\bf v}^{*}(0)\bigg{]}$$
(7)
where ‘Re’ means the real-part of.
We remark that the acoustic fields are evaluated in the particle’s center ${\bf r}=0$.
Equation (7) can be further developed by using Eqs. (3) and (4).
In doing so, we find that
the radiation force can be represented as the sum of a gradient, a scattering,
and a absorption component.
The gradient radiation force is given by
$${\bf f}_{\text{grad}}=-\nabla U(0),$$
(8)
where
$$U=\pi a^{3}\left(f_{0}\frac{|p|^{2}}{3\rho_{0}c_{0}^{2}}-f_{1}\frac{\rho_{0}\|%
{\bf v}\|^{2}}{2}\right)$$
(9)
is the radiation force potential function as obtained by Gor’kov 4.
This force does not depend on the ultrasound absorption by the particle.
Moreover, it might be important whenever the incident beam has a stationary spatial distribution
such as standing waves or focused beams.
On the other hand, the scattering radiation force reads
$${\bf f}_{\text{sca}}=\frac{4\pi a^{2}}{9c_{0}}\left(f_{0}^{2}+f_{0}f_{1}+\frac%
{3f_{1}^{2}}{4}\right)(ka)^{4}\overline{\bf I}(0),$$
(10)
where $\overline{\bf I}=(1/2)\text{Re}\{p{\bf v}^{*}\}$ is the
incident intensity averaged in time.
Equation (10) generalizes the previous result
obtained by Gor’kov for the radiation force produced by a plane traveling wave on a
non-absorbing particle 4.
Finally, the absorption radiation force is expressed as
$${\bf f}_{\text{abs}}=\frac{8\pi a^{3}}{3}\frac{\rho_{0}c_{0}^{\upsilon}}{\rho_%
{1}c_{1}}\alpha_{\upsilon}k^{\upsilon}\overline{\bf I}(0).$$
(11)
This force varies with frequency like the absorption itself.
We see that both the scattering and the absorbing radiation forces depend on the averaged incident intensity.
Therefore, they will be relevant whenever the incident wave has a traveling part, such as a plane wave, for instance.
For a standing wave, the scattering and the absorption radiation forces are much smaller than their gradient counterpart.
According to Eqs. (10) and (11), the magnitude ratio of the absorbing force to the scattering force
is given by
$$\left\|\frac{{\bf f}_{\text{abs}}}{{\bf f}_{\text{sca}}}\right\|=\frac{6\rho_{%
0}c_{0}^{\upsilon+1}\alpha_{\upsilon}k^{\upsilon-4}}{a^{3}\rho_{1}c_{1}(f_{0}^%
{2}+f_{0}f_{1}+3f_{1}^{2}/4)}.$$
(12)
Therefore, the importance of the absorbing radiation force decreases with frequency as $\omega^{\upsilon-4}$.
Though, this force increases with the particle’s radius to the power-third as it approaches to zero.
When $\upsilon=2$ and the incident beam is a plane wave, we recover the result given in Eq. (52) in Ref. 6,
which in turn has a typo (an extra factor of 1/3).
4 Radiation torque
An acoustic beam may set the particle to spin around its axis.
This happens because the beam’s angular momentum with respect
to the particle’s center might be transferred to the particle giving rise to the acoustic radiation torque.
In the Rayleigh scattering limit, the radiation torque is given in terms
of the monopole and the dipole scattering modes only.
Accordingly, we have 9
${\bf N}=\pi a^{3}(I_{0}/c_{0})\bm{\tau}$, where $\bm{\tau}$ is the
dimensionless radiation torque whose Cartesian components are
$$\displaystyle\tau_{x}+i\tau_{y}$$
$$\displaystyle=-\frac{\sqrt{2}}{\pi(ka)^{3}}\left(\frac{s_{1}+s_{1}^{*}}{2}+|s_%
{1}|^{2}\right)\left(a_{1}^{-1}a_{1}^{0*}+a_{1}^{0}a_{1}^{1*}\right),$$
(13)
$$\displaystyle\tau_{z}$$
$$\displaystyle=-\frac{1}{\pi(ka)^{3}}\left(\frac{s_{1}+s_{1}^{*}}{2}+|s_{1}|^{2%
}\right)\left(\left|a_{1}^{1}\right|^{2}-\left|a_{1}^{-1}\right|^{2}\right).$$
(14)
Substituting the beam-shape coefficients given in (24) into these equations results
$$\bm{\tau}=-\frac{12\rho_{0}^{3}\rho_{1}c_{0}^{\upsilon+1}c_{1}\alpha_{\upsilon%
}k^{\upsilon+1}a^{2}}{5p_{0}^{2}(\rho_{0}+2\rho_{1})^{2}}[i{\bf v}(0)\times{%
\bf v}^{*}(0)].$$
(15)
It is clear that no torque is produced in a non-absorbing particle suspended in an inviscid fluid.
Furthermore,
the radiation torque increases with frequency as $\omega^{\upsilon+1}$.
5 Spherically focused beam
Consider that a spherically focused transducer
with diameter $2b=50\,\mathrm{mm}$ and curvature radius $z_{0}=70\,\mathrm{mm}$ ($f$-number of 1.4) generates
an ultrasound beam of frequency $1\,\mathrm{MHz}$ in water at room temperature ($\rho_{0}=1000\,\mathrm{kg/m^{3}}$ and $c_{0}=1480\,\mathrm{m/s}$). The peak pressure at the transducer is $p_{0}=10\,\mathrm{kPa}$.
We chose these parameters because such transducer could be readily manufactured for experiments.
Moreover, the generated ultrasound beam can be described in the paraxial approximation.
The scattering particle has radius $a=10\,\mathrm{\mu m}$ and
hence $ka=0.04$.
It is worth to notice that the viscous boundary layer is $\delta=0.6\,\mathrm{\mu m}$; thus, $\delta\ll a$.
The pressure amplitude of the beam in the focal plane and along the axial direction are given
in cylindrical coordinates $(\varrho,z)$, respectively, by 12
$$\displaystyle p(\varrho,z_{0})$$
$$\displaystyle=\frac{ip_{0}b}{\varrho}\exp\left[ikz_{0}\left(\frac{\varrho^{2}}%
{z_{0}^{2}}+1\right)\right]J_{1}\left(\frac{k\varrho b}{z_{0}}\right),$$
(16)
$$\displaystyle p(0,z)$$
$$\displaystyle=\frac{p_{0}e^{ikz}}{z/z_{0}-1}\left[1-\exp\left(-ikb^{2}\frac{z-%
z_{0}}{2zz_{0}}\right)\right].$$
(17)
where $J_{n}$ is the $n$th-order Bessel function.
The particle might be transversely trapped if the radiation force potential has
a minimum at the beam’s axis, i.e. at $\varrho=0$.
Hence, substituting Eq. (16) into Eq. (9), one encounters
$$\lim_{\varrho\rightarrow 0}U(\varrho,z_{0})=\frac{\pi a^{3}I_{0}k^{2}b^{4}f_{0%
}}{6c_{0}z_{0}^{2}}.$$
(18)
Therefore, the particle will be transversely trapped if $f_{0}<0$, otherwise it
will be pushed away from the beam’s axis.
The situation is different in the axial direction.
Consider the limiting case $b=z_{0}/2$ ($f$-number equal to 1).
Then
and at the focus point, the radiation force potential becomes
$$\lim_{z\rightarrow z_{0}}U(0,z)=\frac{\pi a^{3}I_{0}k^{2}b^{4}}{2c_{0}z_{0}^{2%
}}\left(\frac{f_{0}}{3}-\frac{f_{1}}{2}\right),\quad\frac{z_{0}}{2b}\geq 1.$$
(19)
Thus, the radiation force potential has a minimum only when one of the following conditions is met:
$f_{0}<0$ and $f_{1}>0$, $f_{0}<0$ and $f_{1}<2f_{0}/3$, and
$f_{0}>0$ and $f_{1}>2f_{0}/3$.
Note that the last condition allows the particle to be axially trapped if $f_{0}>0$,
but this implies, according to Eq. (18), that no transverse trap occurs.
Another important issue is even if the radiation force potential has a minimum, the particle
may not be trapped.
This happens because
the absorption and the scattering radiation forces push the particle
away from the potential minimum.
In Fig. 1, we depict the radiation force due to the focused transducer
on a polyurethane (PU) particle with
$\rho_{1}=1130\,\mathrm{kg/m}$, $c_{1}=1468\,\mathrm{m/s}$, and
$\alpha_{\nu}/2\pi=1.43\,\mathrm{Np\cdot MHz/m}$; and a red blood cell (RBC) with
$\rho_{1}=1576\,\mathrm{kg/m}$, $c_{1}=1047\,\mathrm{m/s}$, and
$\alpha_{\nu}/2\pi=0.72\,\mathrm{Np\cdot MHz^{1.47}/m}$.
The PU particle has $f_{0}=0.10$, so it is not transversely trapped.
In axial direction, the absorption radiation force is dominant and pushes
the PU particle away from the focus point $kz_{0}=297$.
Note that the axial radiation force on this particle is rather different when absorption is neglected.
With $f_{0}=-0.26$ and $f_{1}=0.27$, the RBC is three-dimensionally trapped on the
transducer’s focus point.
Furthermore, trapping may also occur in a series of points (depicted by vertical arrows) prior to the focal region.
It is also noticeable that the absorption radiation force on the RBC can be neglected and the scattering radiation force is smaller
than the backward gradient force.
6 Bessel vortex beam
We derive here the radiation torque induced by a Bessel vortex beam on a particle
located in the beam’s axis.
Assume that the beam propagates along the $z$-axis.
Hence, the pressure amplitude of the beam is given, in cylindrical coordinates, by
$$p=p_{0}J_{\ell}(k\varrho\sin\beta)e^{i(\ell{\varphi}+kz\cos\beta)},\quad\ell=0%
,\pm 1,\pm 2,\dots,$$
(20)
where $\beta$ is the beam’s half-cone angle.
The index $\ell$ is known as the orbital angular momentum of the beam.
After calculating the particle velocity of the Bessel beam,
we find that only the axial radiation
torque remains.
Therefore, we have
$$\tau_{z}=\frac{6\ell\rho_{0}\rho_{1}c_{1}c_{0}^{\upsilon-1}\alpha_{\upsilon}k^%
{\upsilon+1}a^{2}}{5(\rho_{0}+2\rho_{1})^{2}}\sin^{2}\beta,\quad|\ell|=1.$$
(21)
When $|\ell|\neq 1$ no radiation torque is produced on the particle.
This happens because radiation torque is induced by
the dipole mode only as shown in Eqs. (13) and (14).
This mode with respect to beam’s axis is not present in all but the first-order Bessel vortex beam.
Equation (21) can also be obtained directly from Eq. (14) in Ref. 13.
7 Conclusions
An extension of Gor’kov’s radiation force theory to include absorption effects inside
the particle was presented.
With this theory, the stability of axial and transverse particle trapping by
a spherically focused ultrasound beam was analyzed.
Moreover, an exact expression for the radiation torque on a suspended particle was derived
and was applied to obtain the torque caused by a Bessel vortex beam in an on-axis particle.
In conclusion, we believe that the closed-form expressions i the limit of non-viscous fluids developed here might be useful
in analyzing the trapping stability of single-beam acoustical tweezers.
Appendix A Beam-shape coefficients
The beam-shape coefficients are given by 10
$$a_{n}^{m}=\frac{1}{p_{0}j_{n}(kR)}\int_{\Omega}p_{\text{i}}(kR,\theta,\varphi)%
Y_{n}^{m*}(\theta,\varphi)d\Omega,$$
(22)
where $\Omega$ is the unit sphere, $d\Omega$ is the differential solid angle, and $R$ is the radius of a control spherical
region in which the incident beam propagates.
Expanding the incident pressure in the origin up to the second-order approximation,
yields
$$p({\bf r})=p(0)+ik{\bf r}\cdot{\bf v}(0)+\frac{ik}{2}{\bf r}\cdot\left[{\bf r}%
\cdot\nabla{\bf v}(0)\right]\,$$
(23)
where ${\bf v}=(v_{x},v_{y},v_{z})$.
Substituting Eq. (23) into Eq. (22) along with
${\bf r}=r(\sin\theta\cos\varphi,\sin\theta\sin\varphi,\cos\theta)$,
we obtain the beam-shape coefficients up to the quadrupole approximation as
$$\displaystyle a_{0}^{0}$$
$$\displaystyle=\frac{\sqrt{4\pi}}{p_{0}}p(0),$$
$$\displaystyle a_{1}^{-1}$$
$$\displaystyle=\frac{\rho_{0}c_{0}}{p_{0}}\sqrt{6\pi}[iv_{x}(0)+v_{y}(0)],$$
$$\displaystyle a_{1}^{0}$$
$$\displaystyle=\frac{2i\rho_{0}c_{0}}{p_{0}}\sqrt{3\pi}v_{z}(0),$$
$$\displaystyle a_{1}^{1}$$
$$\displaystyle=\frac{\rho_{0}c_{0}}{p_{0}}\sqrt{6\pi}[-iv_{x}(0)+v_{y}(0)],$$
$$\displaystyle a_{2}^{-2}$$
$$\displaystyle=\frac{\rho_{0}c_{0}}{kp_{0}}\sqrt{30\pi}\left[i\left(\frac{%
\partial v_{x}(0)}{\partial x}-\frac{\partial v_{y}(0)}{\partial y}\right)+2%
\frac{\partial v_{x}(0)}{\partial y}\right],$$
(24)
$$\displaystyle a_{2}^{-1}$$
$$\displaystyle=\frac{2\rho_{0}c_{0}}{kp_{0}}\sqrt{30\pi}\left[i\frac{\partial v%
_{x}(0)}{\partial z}+\frac{\partial v_{y}(0)}{\partial z}\right],$$
$$\displaystyle a_{2}^{0}$$
$$\displaystyle=\frac{2i\rho_{0}c_{0}}{kp_{0}}\sqrt{5\pi}\left[i\left(\frac{%
\partial v_{x}(0)}{\partial x}+\frac{\partial v_{y}(0)}{\partial y}\right)-2%
\frac{\partial v_{z}(0)}{\partial z}\right],$$
$$\displaystyle a_{2}^{1}$$
$$\displaystyle=\frac{2\rho_{0}c_{0}}{kp_{0}}\sqrt{30\pi}\left[-i\frac{\partial v%
_{x}(0)}{\partial z}+\frac{\partial v_{y}(0)}{\partial z}\right],$$
$$\displaystyle a_{2}^{2}$$
$$\displaystyle=\frac{\rho_{0}c_{0}}{kp_{0}}\sqrt{30\pi}\left[i\left(\frac{%
\partial v_{x}(0)}{\partial x}-\frac{\partial v_{y}(0)}{\partial y}\right)-2%
\frac{\partial v_{x}(0)}{\partial y}\right].$$
References and links
1
J. Wu, “Acoustical tweezers,” J. Acoust. Soc. Am. 89, 2140–2143
(1991).
2
J. Lee, S. Teh, A. Lee, H. Kim, C. Lee, and K. Shung, “Single beam acoustic
trapping,” Appl. Phys. Lett. 95, 073701 (2009).
3
X. Ding, S.-C. S. Lin, B. Kiraly, H. Yue, S. Li, S. J. B. I-K. Chiang, J. Shi,
and T. J. Huang, “On-chip manipulation of single microparticles, cells, and
organisms using surface acoustic waves,” Proc. Natl. Acad. Sci. 109,
11105–11109 (2012).
4
L. P. Gor’kov, “On the forces acting on a small particle in an acoustic field
in an ideal fluid,” Sov. Phys. Dokl. 6, 773–775 (1962).
5
A. A. Doinikov, “Acoustic radiation pressure on a compressible sphere in a
viscous fluid,” J. Fluid Mech. 267, 1–21 (1994).
6
R. Löfstedt and S. Putterman, “Theory of long wavelength acoustic radiation
pressure,” J. Acoust. Soc. Am. 90, 2027–2033 (1991).
7
T. L. Szabo, “Time domain wave equations for lossy media obeying a frequency
power law,” J. Acoust. Soc. Am. 96, 491–500 (1994).
8
G. T. Silva, “An expression for the radiation force exerted by an acoustic
beam with arbitrary wavefront,” J. Acoust. Soc. Am. 130, 3541–3545
(2011).
9
G. T. Silva, T. P. Lobo, and F. G. Mitri, “Radiation torque produced by an
arbitrary acoustic wave,” Europhys. Phys. Lett. 97, 54003 (2012).
10
G. T. Silva, “Off-axis scattering of an ultrasound Bessel beam by a
sphere,” IEEE Trans. Ultrason. Ferroelec. Freq. Contr. 58, 298–304
(2011).
11
G. T. Silva, J. H. Lopes, and F. G. Mitri, “Off-axial acoustic radiation force
of repulsor and tractor Bessel beams on a sphere,” IEEE Trans. Ultras.
Ferroel. Freq. Control 60, 1207–1212 (2013).
12
B. G. Lucas and T. G. Muir, “The field of a focusing source,” J. Acoust. Soc.
Am. 72, 1289–1289 (1982).
13
F. G. Mitri, T. P. Lobo, and G. T. Silva, “Axial acoustic radiation torque of
a Bessel vortex beam on spherical shells,” Phys. Rev. E 85, 026602
(2012). |
Semi-Supervised Classification and Segmentation on High Resolution Aerial Images
Sahil Khose${}^{*}$ Abhiraj Tiwari${}^{*}$ Ankita Ghosh${}^{*}$
Manipal Institute of Technology
Manipal
{sahil.khose, abhiraj.tiwari1, ankita.ghosh1}@learner.manipal.edu
Abstract
FloodNet is a high-resolution image dataset acquired by a small UAV platform, DJI Mavic Pro quadcopters, after Hurricane Harvey. The dataset presents a unique challenge of advancing the damage assessment process for post-disaster scenarios using unlabeled and limited labeled dataset. We propose a solution to address their classification and semantic segmentation challenge. We approach this problem by generating pseudo labels for both classification and segmentation during training and slowly incrementing the amount by which the pseudo label loss affects the final loss. Using this semi-supervised method of training helped us improve our baseline supervised loss by a huge margin for classification, allowing the model to generalize and perform better on the validation and test splits of the dataset. In this paper, we compare and contrast the various methods and models for image classification and semantic segmentation on the FloodNet dataset.
††
${}^{*}$Authors have contributed equally to this work and share first authorship
1 Introduction
The frequency and severity of natural disasters threaten human health, infrastructure and natural systems. It is extremely crucial to have accurate, timely and understandable information to improve our disaster management systems. Rapid data collection from remote areas can be easily facilitated using small unmanned aerial systems which provide high-resolution images. Visual scene understanding of these collected images is vital for quick response and large scale recovery post-natural disaster. Classification and segmentation tasks are fitting in such situations as they can provide scene information to help the task force make decisions.
One of the major challenges with generating a vision dataset is the cost of labeling the data, especially for semantic segmentation. This often leads to labels only for a small percentage of the data which gives rise to the need for semi-supervised methods that can produce results that are at par with supervised methods. Another challenge that we face apart from the lack of labeled dataset is the heavy class imbalance. Lack of labeled data coupled with class imbalance makes it a very challenging task to solve. Through our approach, we try to tackle these problems and produce creditable results.
Our contribution in this paper is two folds: semi-supervised classification and semi-supervised semantic segmentation. We discuss the existing literature in these fields based on which we have crafted our approach in Section 2. The subsequent sections throw light upon the pipeline of classification 3 and segmentation 4 respectively. Our paper describes the preprocessing and augmentation techniques used on the dataset, the methodology followed, the models and hyperparameters we have experimented with and the results obtained. Finally, we conclude by summarizing our results and discussing the scope of improvement in future 5.
2 Related Works
Supervised Classification:
Classification has been one of the earliest undertakings in the deep learning domain. Over the years, several architectures and methods have emerged which have leveraged extensive datasets like ImageNet [5] to produce state of the art results using supervised learning. The architectures that our paper makes use of are ResNet [6] and EfficientNet [14].
ResNet proposes residual connection architecture which makes it feasible to train networks with a large number of layers without escalating the training error percentage. Using the technique of skip connections, it resolves the issue of vanishing gradient.
EfficientNet proposes a simple but highly effective scaling method that can be used to scale up any model architecture to any target resource constraints while maintaining model efficiency. They observed the effects of model scaling and identified that carefully balancing network depth, width and resolution can lead to better performance.
Supervised Semantic Segmentation: Segmentation aids in extracting the maximum amount of information from an image. Instance segmentation treats every occurrence of a class as a unique object. Semantic segmentation is more coherent and associates every pixel of an image with a class label. Deep learning models like UNet[13], PSPNet[15] and DeepLabV3+[3] have provided exceptional results for this task.
The architecture of UNet is divided into two parts: contracting path and expansive path. The contracting path follows the generic framework of a convolutional network while the expansive path undergoes deconvolution to reconstruct the segmented image.
PSPNet exploits the capability of global context information using different region-based context aggregation by introducing a pyramid pooling module along with the proposed pyramid scene parsing.
DeepLabV3+ is a refinement of DeepLabV3 which uses atrous convolution. Atrous convolution is a powerful tool to explicitly adjust the filter’s field-of-view as well as control the resolution of feature responses computed by Deep Convolution Neural Network.
Semi-supervised Approach:
Pseudo-Label [9] proposes a simple semi-supervised learning approach. The idea is to train the neural network in a supervised fashion with both labeled and unlabeled data simultaneously. For unlabeled data, pseudo labels are generated by selecting the class which has maximum predicted probability. This is in effect equivalent to Entropy Regularization [2]. It favors a low-density separation between classes, a commonly assumed prior for semi-supervised learning.
3 Classification
The dataset has 2343 images of dimensions $3000\times 4000\times 3$ divided into training (1445), validation (450) and test (448) splits. Out of the 1445 training images 398 are labeled and 1047 are unlabeled. The labels for the classification task are Flooded and Non-Flooded.
In this section we describe our approach for classifying the FloodNet dataset [11] into 2 classes, Flooded and Non-Flooded.
3.1 Data and Preprocessing
In the given dataset 398 sample images were labeled out of which 51 samples were flooded and 347 were non-flooded. The large class imbalance prevents the model from achieving a good F1 score while training with the labeled dataset. To prevent this we used a weighted sampling strategy while loading the data in the model as inspired from [10]. Both the classes were sampled equally during batch generation.
The labeled dataset was heavily augmented to get more images for training the model under supervision. The image samples were randomly cropped, shifted, resized and flipped along the horizontal and vertical axes.
We downsized the image from $3000\times 4000$ to $300\times 400$ dimensions to strike a balance between processing efficiency gained by the lower dimensional images and information retrieval of the high-resolution images.
3.2 Methodology
ResNet18 with a binary classification head was used for semi-supervised training on the dataset. The model was trained for $E$ epochs out of which only the labeled samples were used for $E^{\alpha}_{i}$ epochs after which pseudo labels were used to further train the model. $\alpha$ has an initial value of $\alpha_{i}$ that increases up to $\alpha_{f}$ from epoch $E^{\alpha}_{i}$ to $E^{\alpha}_{f}$ as described in Algorithm 1.
A modified form of Binary Cross-Entropy (BCE) was used as the loss function as shown in line 10 in Algorithm 1 where $l$ is the label of a sample, $\hat{l}$ is the predicted class for labeled sample and $u_{epoch}$ is the predicted class for an unlabeled sample in the current epoch. This loss function was optimized using Stochastic Gradient Descent (SGD) [12].
3.3 Experiments
We used ResNet18 as it is computationally efficient. We experimented with Adam [8] optimizer and SGD. Optimizing using SGD was much more stable and the optimizer was less susceptible to overshooting. Different values of $\alpha$ were experimented with and it was found that a slow and gradual increase in alpha was better for training the model. Our best performing model uses $\alpha_{i}=0$ and $\alpha_{f}=1$. The value of $\alpha$ increases from epoch $E^{\alpha}_{i}=10$ to $E^{\alpha}_{f}=135$. The model was trained on batch size of 64.
3.4 Results
Our system performed significantly better than all the classification baseline results mentioned in the FloodNet paper while having a considerably smaller architecture (half the number of parameters) as shown in Table 1. Our best model achieves 98.10% F1 and 96.70% accuracy on the test set.
4 Segmentation
In this section, we detail our approach for training a model which generates multi-class segmentation masks for given images. The semantic labels for the task is a 10 pixel-level class segmentation mask consisting of Background, Building-flooded, Building non-flooded, Road-flooded, Road non-flooded, Water, Tree, Vehicle, Pool and Grass classes. They are mapped from 0 to 9 respectively.
4.1 Data and Preprocessing
To expedite the process of feature extraction for the deep learning model, we apply bilateral filter to the image, followed by two iterations of dilation and one iteration of erosion. For image augmentation we perform shuffling, rotation, scaling, shifting and brightness contrast. The images and masks are resized to $512\times 512$ dimensions while training since high-resolution images preserve useful information.
4.2 Methodology
The dataset contains labeled masks of dimension $3000\times 4000\times 3$ with pixel values ranging from 0 to 9, each denoting a particular semantic label. These are one-hot encoded to generate labels with 10 channels, where $i^{th}$ channel contains information about $i^{th}$ class.
We experiment with various encoder-decoder and pyramid pooling based architectures to train our model, the details of which are mentioned in Section 4.3. The loss function used is a weighted combination of Binary Cross-Entropy loss (BCE) and Dice loss as it provides visually cleaner results.
We apply a semi-supervised approach and generate pseudo masks for the unlabeled images. While training the model for $E$ epochs, the labeled samples were used for $E^{\alpha}_{i}$ epochs where Adam is used as an optimizer. After that pseudo masks were used to further train the model as described in Algorithm 1. $\alpha$ has an initial value of $\alpha_{i}$ that increases upto $\alpha_{f}$ from epoch $E^{\alpha}_{i}$ to $E^{\alpha}_{f}$. SGD optimizer with 0.01 LR is used when pseudo masks are introduced to the model.
4.3 Experiments
We apply three state-of-the-art semantic segmentation models on the FloodNet dataset. We adopt one encoder-decoder based network named UNet, one pyramid pooling module based network PSPNet and the last network model DeepLabV3+ employs both encoder-decoder and pyramid pooling based module. We train all of them in a supervised fashion.
For UNet, PSPNet and DeepLabV3+ the backbones used were ResNet34 , ResNet101 and EfficientNet-B3 respectively.
For UNet the learning rate was $0.01$ with step LR scheduler set at intervals [10,30,50] and decay factor $\gamma$ set to $0.1$. For PSPNet the learning rate was $0.001$ without any LR decay. For DeepLabV3+ the learning rate was 0.001 with step LR scheduler set at intervals [7,20] and $\gamma$ set to $0.1$.
Adam optimizer and batch size of 24 was used for all the models with Mean Intersection over Union (MIoU) as the evaluation metric. We observed the best results when we weighed the BCE loss and Dice loss equally.
Once we recognized the best performing model on the task, we trained a DeepLabV3+ model with EfficientNet-B3 as the backbone and used SGD optimizer instead of Adam optimizer in a semi-supervised fashion. Due to computation and memory constraints we randomly sampled unlabeled data with the ratio of $1:10$ for generating the pseudo masks.
4.4 Results
Table 2 showcases the comparison of the best models we achieved for each of the 3 architectures. The best test set result was achieved by a DeepLabV3+ architecture with EfficientNet-B3 backbone. A few example images of the predictions of the model against the ground truth are provided in Figure 1. It is evident from the table that small objects like vehicles and pools are the most difficult tasks for our models.
5 Conclusion
In this work, we have explored methods to approach semi-supervised classification and segmentation along with handling the class imbalance problem on high-resolution images. We have conducted a range of experiments to obtain the best possible technique and models to optimize for the tasks.
Our classification framework achieves laudable results with just 398 labeled images and also utilizes the entirety of the unlabeled data. Our segmentation framework shows an increase of 0.19% on using the unlabeled data as pseudo labels. This provides a wide scope of improvement. The amount of unlabeled data is approximately three times the amount of labeled data which if employed efficiently can produce superior results.
We foresee multiple opportunities for future research. Training the model in an unsupervised fashion and fine-tuning it with the labeled data followed by distillation as presented in SimCLRv2 [4] is a very promising method. Training with a contrastive loss [7] has been incremental at times. With the emergence of Visual Transformers, self-supervised Vision Transformers [1] could also be explored for this task.
6 Acknowledgement
We would like to thank Research Society Manipal for their valuable inputs and research guidance.
References
[1]
Mathilde Caron, Hugo Touvron, Ishan Misra, Hervé Jégou, Julien Mairal,
Piotr Bojanowski, and Armand Joulin.
Emerging properties in self-supervised vision transformers.
arXiv preprint arXiv:2104.14294, 2021.
[2]
Olivier Chapelle, Bernhard Scholkopf, and Alexander Zien, editors.
Semi-Supervised Learning.
The MIT Press, Sept. 2006.
[3]
Liang-Chieh Chen, Yukun Zhu, George Papandreou, Florian Schroff, and Hartwig
Adam.
Encoder-decoder with atrous separable convolution for semantic image
segmentation.
In Proceedings of the European Conference on Computer Vision
(ECCV), September 2018.
[4]
Ting Chen, Simon Kornblith, Kevin Swersky, Mohammad Norouzi, and Geoffrey E.
Hinton.
Big self-supervised models are strong semi-supervised learners.
CoRR, abs/2006.10029, 2020.
[5]
Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei.
Imagenet: A large-scale hierarchical image database.
In 2009 IEEE Conference on Computer Vision and Pattern
Recognition, pages 248–255, 2009.
[6]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun.
Deep residual learning for image recognition.
In 2016 IEEE Conference on Computer Vision and Pattern
Recognition (CVPR), pages 770–778, 2016.
[7]
Prannay Khosla, Piotr Teterwak, Chen Wang, Aaron Sarna, Yonglong Tian, Phillip
Isola, Aaron Maschinot, Ce Liu, and Dilip Krishnan.
Supervised contrastive learning.
CoRR, abs/2004.11362, 2020.
[8]
Diederik P. Kingma and Jimmy Ba.
Adam: A method for stochastic optimization.
In Yoshua Bengio and Yann LeCun, editors, 3rd International
Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May
7-9, 2015, Conference Track Proceedings, 2015.
[9]
Dong-Hyun Lee.
Pseudo-label : The simple and efficient semi-supervised learning
method for deep neural networks.
ICML 2013 Workshop : Challenges in Representation Learning
(WREPL), 07 2013.
[10]
Sidharth R, Abhiraj Tiwari, Parthivi Choubey, Saisha Kashyap, Sahil Khose,
Kumud Lakara, Nishesh Singh, and Ujjwal Verma.
Bert based transformers lead the way in extraction of health
information from social media, 2021.
[11]
Maryam Rahnemoonfar, Tashnim Chowdhury, Argho Sarkar, Debvrat Varshney, Masoud
Yari, and Robin R. Murphy.
Floodnet: A high resolution aerial imagery dataset for post flood
scene understanding.
CoRR, abs/2012.02951, 2020.
[12]
Herbert Robbins and Sutton Monro.
A Stochastic Approximation Method.
The Annals of Mathematical Statistics, 22(3):400 – 407, 1951.
[13]
Olaf Ronneberger, Philipp Fischer, and Thomas Brox.
U-net: Convolutional networks for biomedical image segmentation.
In Nassir Navab, Joachim Hornegger, William M. Wells, and
Alejandro F. Frangi, editors, Medical Image Computing and
Computer-Assisted Intervention – MICCAI 2015, pages 234–241, Cham, 2015.
Springer International Publishing.
[14]
Mingxing Tan and Quoc Le.
EfficientNet: Rethinking model scaling for convolutional neural
networks.
In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning,
volume 97 of Proceedings of Machine Learning Research, pages
6105–6114. PMLR, 09–15 Jun 2019.
[15]
Hengshuang Zhao, Jianping Shi, Xiaojuan Qi, Xiaogang Wang, and Jiaya Jia.
Pyramid scene parsing network.
In Proceedings of the IEEE Conference on Computer Vision and
Pattern Recognition (CVPR), July 2017. |
On the waiting time distribution for continuous stochastic systems
Robert Gernert${}^{1}$
[email protected]
Clive Emary${}^{2}$
[email protected]
Sabine H. L. Klapp${}^{1}$
[email protected]
${}^{1}$ Institut für Theoretische Physik, Sekr. EW 7–1,
Technische Universität Berlin, Hardenbergstrasse 36,
D-10623 Berlin, Germany
${}^{2}$ Department of Physics and Mathematics,
University of Hull, Kingston-upon-Hull, United Kingdom
(November 23, 2020)
Abstract
The waiting time distribution (WTD) is a common tool for analysing discrete stochastic processes in classical and quantum systems. However, there are many physical examples
where the dynamics is continuous and only approximately discrete, or where it is favourable to discuss the dynamics on a discretized and a continuous level in parallel. An example is the hindered motion of particles through potential landscapes with barriers. In the present paper we propose a consistent generalisation of the WTD from the discrete case
to situations where the particles perform continuous barrier-crossing characterised by a finite duration. To this end, we introduce a recipe to calculate the WTD
from the Fokker-Planck/Smoluchowski equation. As an application, we consider the paradigmatic example of an overdamped particle diffusing through a washboard potential. For sufficiently large energy barriers, the WTD defined via the Smoluchowski equation becomes consistent not only with that obtained from direct simulation of the underlying Langevin equation, but also with the WTD resulting from the analytical solution of a (two-state) master equation model for the short-time dynamics developed previous by us [PRE 86, 061135 (2012)]. Thus, our approach “interpolates” between these two types of stochastic motion. We illustrate our approach for both symmetric systems and systems under constant force.
pacs:
05.10.Gg, 05.60.-k, 05.40.-a
November 23, 2020
I Introduction
Waiting time distributions (WTD) measure the distribution of delay times between subsequent hopping events (“jumps”) in a discrete stochastic process.
The concept of a WTD has been introduced already half a century ago for the description of diffusion processes on lattices
by continuous time random walks montroll65 ; scher73 ; haus87 .
Since then, the analysis of transport processes via WTDs has become common for a wide range of classical and quantum systems
whose dynamics can be viewed, at least approximately, as a discrete stochastic process. Recent applications include, e.g. in the biophysical area,
the opening kinetics of ion channels higgins09 , the (un-)binding of chemical bonds qian13 , and the “turnover” events of enzyme molecules saha12 ; english06 .
In econophysics, WTDs are used to analyse stock market fluctuations mainardi00 . A further important field are diffusion processes of atoms, molecules, or colloids on
corrugated surfaces skaug13 ; prager06 ; ganapathy10 ; hanes12 .
Although the processes considered there are actually continuous (particles on real surfaces are not bound to lattice sites),
the effective barriers characterising rough surfaces are often much larger than the thermal energy; thus
the motion can be approximately viewed as a sequence of jumps.
A similar situation occurs in supercooled liquids and glasses,
where the stochastic processes correspond to hopping of particles between cages (formed by other particles)
jung05 ; chaud08 , to glass rearrangements ahn13 or, on a more abstract level,
to jumps between “metabasins” in configuration space doliwa03 ; heuer05 ; rubner08 .
Finally, within the area of quantum transport, typical applications include
electron transport in quantum-dot nanostructures brandes08 ; flindt09 ; rajabi13 and tunnelling events in dynamical single-electron emitters albert11 .
Despite these wide and crossdisciplinary applications of WTDs there is, for many systems, no unique recipe to actually calculate this quantity.
Indeed, a straightforward definition is only possible for truly discrete stochastic systems where the dynamics is described by a (Markovian)
master equation (ME). Here, the jumps are a priori discrete and have (once they occur) zero duration. Given a sufficiently simple structure of the ME, it is often
even possible to calculate the WTD analytically (see recent studies in quantum transport flindt09 ; brandes08 ; albert11 and ion channel kinetics higgins09 ).
For continuous systems, however, the calculation and
even the very definition of a WTD are not straightforward: and ultimately this is tied to the unavoidable ambiguity in defining a discrete
“jump” for a continuous evolving system.
A frequently-used strategy in such situations is to obtain the WTD numerically from an analysis of (particle) trajectories generated either experimentally or by many-particle computer simulations. Indeed, state-of-the-art experiments allow real time tracking of the motion of colloids on random surfaces hanes12 ; evers13 , on solid-liquid interfaces skaug13 ,
or during surface growth ganapathy10 . Similar trajectory analysis has been performed
in computer simulations (such as Molecular or Brownian Dynamics) of glassy or otherwise frustrated systems (see, e.g., jung05 ; heuer05 ).
For a continuous system with “quasi-hopping” motion, typical trajectories reveal the existence of “traps” (in real or configuration space), where the trajectory spends most time,
and fast motion (barrier crossing) in between. However, to extract from that a WTD, one necessarily needs to discretise the space (i.e., to define regions which trajectories can enter or leave),
yielding some ambiguity in the definition of jumps. In particular, depending on these details of the discretisation,
the resulting jumps can have a finite duration. Moreover, in many practical applications the measured or simulation-generated trajectories render quite noisy WTDs;
extracting smooth results (and relevant time constants) then requires extensive simulation or experimental time. In view of these difficulties it seems desirable to construct a WTD in a yet alternative way, namely
on the basis of a noise-averaged description of the stochastic process. The natural framework is, of course, the Fokker-Planck equation for the one-body probability density.
In this spirit, we propose in the present paper a simple recipe to calculate the WTD from a Fokker-Planck equation.
Our approach is inspired from the calculation risken84 of the first passage time distribution; indeed the two quantities are closely related when we assume the dynamics to be
Markovian (i.e., time-homogeneous). Considering a particular model system we validate our WTD definition by comparing
with the WTD following from direct simulation of the underlying Langevin equation of motion.
Specifically, we consider the paradigm example of an overdamped (“Brownian”) particle in a periodic “washboard” (e.g., sinusoidal) potential,
a generic model with relevance in many areas such as particle transport on topologically or energetically
structured surfaces Tierno10 ; Dalle11 , the motion of motor proteins xing05 , particles in optical lattices Siler10 and optical line traps Lopez08 ,
biased Josephson junctions zapata96 ; shah07 ; augello10 , and more generally in biophysical processes bonnet11 ; xing05 ; kay07 ; Ros05 .
In an earlier study emary12 we have analysed in detail the short-time dynamics of that system, particularly the plateaus appearing in the mean-squared displacement and the corresponding non-Gaussian parameters for large values of the ratio between barrier height and thermal energy.
We have demonstrated that, in this limit of deep wells, the continuous dynamics described by the Smoluchowski equation (SE)
can be well described by a ME approach involving two states per well emary12 . One attractive feature of this ME model is that it is analytically solvable; i.e., properties such as cumulants
of the density distribution can be calculated by minimal effort.
Here we use this ME approach as a third, analytical route to calculate the WTD. We show that the WTD obtained from the SE and BD approach
becomes consistent with that of the discrete case in the appropriate limit (deep wells).
As wells become shallower, the ME approach becomes less good, an indicator being that the duration of the transitions between different wells increases. In summary, we suggest a consistent generalisation of WTD from the discrete, zero-duration jump case
to the continuous, extended-jump one.
The rest of this paper is organised as follows. In Sec. II we specify the equations of motion of our model and discuss representative trajectories. The three approaches to the WTD are introduced in Sec. III, supplemented by two appendices where we present detailed information on the numerical calculation. Numerical results are presented and discussed in Sec. IV, and in Sec. V we summarise our findings.
II Model
We begin by discussing a concrete model as an illustration of the nature of trajectories and jumps in the class of systems we investigate here. We consider a particle in the tilted washboard potential
$$u(x)=u_{0}\sin^{2}\frac{\pi x}{a}-Fx,$$
(1)
where $a$ is the period of the potential, and the parameters $u_{0}$ and $F$
describe the depth and tilt of the potential, respectively. Under overdamped conditions (strong friction),
the equation of motion of the particle in one dimension is given by
$$\gamma\dot{x}(t)=-u^{\prime}(x)+\sqrt{2\gamma k_{\mathrm{B}}T}\xi(t).$$
(2)
The thermal energy is denoted by $k_{\mathrm{B}}T$, $\gamma=k_{\mathrm{B}}T/D_{0}$ is the friction constant (involving the diffusion constant
$D_{0}$ of the free system), $-u^{\prime}(x)=-du/dx$ is the force, and $\xi(t)$ represents Gaussian white noise
with properties $\langle\xi(t)\rangle=0$ and $\langle\xi(t)\xi(t^{\prime})\rangle=\delta(t-t^{\prime})$.
In the present study we focus on cases where the form of the potential is such that it can be meaningfully divided in a number of wells (defined between subsequent maxima).
This occurs when the difference between
the potential minima and maxima, $\Delta u(u_{0},F)$, is large compared to the
thermal energy, i.e., $u_{0}\gg k_{\mathrm{B}}T$, and when the driving force $F$ is smaller than the so-called critical force $F_{\mathrm{c}}=u_{0}\pi/a$, related to the diffusion maximum
Reimann01 (for $F>F_{\mathrm{c}}$ the potential minima vanish).
In order to illustrate the motion in such situations we calculate representative particle trajectories
from Eq. (2) by using a Brownian Dynamics (BD) algorithm ermak75 .
Examples for two different values of $u_{0}/k_{\mathrm{B}}T$ and $F\!=\!0$ are shown in Fig. 1a). The trajectories in Fig. 1a) clearly reflect that the particle is “trapped” for certain times in the regions around the potential minima $x/a\in\mathbb{Z}$. To facilitate the comparison between the trajectories corresponding to different $u_{0}$ they are plotted in Fig. 1b) as functions of the relative time. This latter is defined as the difference between the actual time and the time $t_{\mathrm{jump}}$, which is the arithmetic mean of the times related to the beginning and end of a jump. Here, “beginning” and “end” refer to leaving of a potential minimum (without going back) and arriving at a different minimum (this definition of a jump will later be called minimum-based definition).
From Fig. 1b) we see that even for the largest barrier considered, the particle needs a finite time to cross the barrier. Thus, the motion is still not perfectly discrete.
However, as we may note already here, a different picture on the degree of “discreteness” arises when we rescale the time axis with the corresponding Kramers’ rate $r_{\text{K}}(u_{0})=D_{0}/(2\pi k_{\mathrm{B}}T)\sqrt{-u^{\prime\prime}(x_{%
\mathrm{min}})u^{\prime\prime}(x_{\mathrm{max}})}\,\exp(-\Delta u/(k_{\mathrm{%
B}}T))$ risken84 ; emary12 ; kramers40 , see Fig. 2.
By multiplying the time with this well-known quantity, we take into account that the increase of the potential height alone already leads to an increase of the escape time reimann99 . As a consequence the intervals between jumps in the rescaled time become independent of $u_{0}$ in the long time limit.
Indeed, after rescaling, the data for $u_{0}=15k_{\mathrm{B}}T$ reflects an essentially discrete motion. However, at $u_{0}=5k_{\mathrm{B}}T$ we still observe a finite duration of the jump. This clearly demonstrates that the case $u_{0}=5k_{\mathrm{B}}T$ is at the boundary between discrete and continuous motion.
III Routes to calculate the WTD
III.1 Direct evaluation via BD simulations
To define the WTD we first need to consider the possible types of jumps. To this end, we note that the stochastic process $x(t)$ [see Eq. (2)] is a Markov process. Thus, the jump characteristics are independent of the history. Further, because the motion is one-dimensional, there are only two directions in which the particle can jump. Finally, Eq. (2) obeys the translational symmetry $x\to x\pm a$. Because of these three properties the jumps can be grouped into only two types, namely “to the right” and “to the left”, independent of when or where the jump began.
We label these jump types by the index “J”. The WTD $w^{J}(t)$ is then the probability density for the time between an arbitrary jump and a subsequent jump of type $J$.
In the context of BD simulations [i.e., direct evaluation of Eq. (2)], the WTD $w^{J}(t)$ is extracted from a histogram of waiting times extracted from the trajectories pertaining to a given realisation of noise. We note that the WTDs $w^{J}(t)$ must fulfil the normalisation condition
$$\displaystyle 1=\int_{0}^{\infty}\mathrm{d}t\,\sum_{J}w^{J}(t),$$
(3)
which expresses the fact that the particle leaves its minimum for certain. In terms of the survival probability
$$\displaystyle S(t)=\int_{t}^{\infty}\mathrm{d}t^{\prime}\,\sum_{J}w^{J}(t^{%
\prime})$$
(4)
the normalisation is expressed as $S(0)=1$. The sum runs over all jumps types that leave a potential minimum, in our case “to the right” and “to the left”.
In our BD calculations of $w^{J}(t)$ we consider two definitions of a jump which are illustrated in Fig. 3. Within the first definition, a jump occurs if the particle crosses a potential maximum (maximum-based definition). Note that this automatically defines the jump to be instantaneous.
Within the second definition, a jump involves the entire motion between two neighbouring minima of the potential (minimum-based definition). As a consequence, the jump has a finite duration defined as the time between leaving a minimum and arriving at a neighbouring one.
We call the corresponding probability distribution the jump duration distribution.
Further, we define the waiting time as the time between the end times of two subsequent jumps.
Numerical details of the calculation are given in Appendix A.
III.1.1 Discussion of the jump definitions
To investigate the role of the two jump definitions we consider the resulting WTDs for the symmetric case $F\!=\!0$, $u_{0}\!=\!5k_{\mathrm{B}}T$, plotted in Fig. 4. Because of the symmetry of the potential at $F\!=\!0$, there is only one relevant WTD. Clearly, the choice of the jump definition has a strong impact on the WTD.
In particular, we find that only the minimum-based definition yields a smooth shape of $w^{J}(t)$.
On the contrary, the maximum-based jump definition poses several problems. First, during the analysis of the trajectories we noticed that the particle crosses the maximum many times subsequently which generates a lot of spurious (maximum-based) jumps. We deal with this problem by introducing a “dead time” after recording a jump which must pass until the next jump can be recorded. We set the “dead time” to $1\tau$ (i.e., the time the particle needs to diffuse the distance $\sqrt{\tau/D_{0}}=a$) to ensure that it leaves the maximum during that time. As a consequence of introducing the dead time, the corresponding WTD is defined to be zero for $t<1\tau$.
Further, we see in Fig. 4 that by using the maximum-based jump definition, waiting times in the range $1\tau\lesssim t\lesssim 5\tau$ are more likely than by using the minimum-based jump definition. The reason is that the particle, once it has gone uphill, is counted to have made a jump in the maximum-based jump definition. On the contrary, the same particle is only counted in the minimum-based jump definition if it goes downhill in the same direction (and not backwards). Thus, the maximum-based jump definition frequently records jumps although the particle does not change the potential well. This contradicts our idea of a jump to cross the potential barrier. This problem could be solved by keeping track of the direction in which the particle crosses the maximum. However, the first problem remains (i.e. accessing short waiting times). We therefore choose the minimum-based jump definition for the further investigation.
III.2 Definition from the Smoluchowski equation
The Fokker-Planck equation corresponding to Eq. (2) is given by the Smoluchowski equation (SE)
$$\displaystyle\partial_{t}p(x,t)$$
$$\displaystyle=D_{0}\frac{\partial^{2}}{\partial x^{2}}p(x,t)+\frac{D_{0}}{k_{%
\mathrm{B}}T}\frac{\partial}{\partial x}\Big{(}p(x,t)u^{\prime}(x)\Big{)}\,,$$
(5)
where $p(x,t)\Delta x$ is the probability that the particle is in the interval $[x,x+\Delta x]$ at time $t$.
Again, we are interested in the distribution of waiting times $t$.
In Sec. III.1 we defined $w^{J}(t)$ as the WTD for the jump $J$. Here we associate $J$ with the jumps from an initial position $B$ to a target position $C$. Typically $B$ and $C$ are neighbouring minima of the potential. In order to realise the requirement that the WTD describes the time between subsequent jumps we need to exclude processes that lead the particle towards the other neighbouring minimum, which we call $A$.
Correspondingly, we describe the WTD by the function $w_{A}^{B\to C}(t)$ where $B\to C$ stands for $J$ and the subscript $A$ indicates the excluded position. The definition is as follows:
Definition: The quantity $w_{A}^{B\to C}(t)\,\Delta t$, where $\Delta t$ is a small time interval, is the probability that
•
the particle was for certain at position $$B$$ at $$t\!=\!0$$,
(6a)
•
the particle reaches position $$C$$ in the interval $$[t,t+\Delta t]$$, and
(6b)
•
the particle neither arrived at $$C$$ nor $$A$$ within the time interval $$[0,t]$$.
(6c)
Note that we do not care about what happened at earlier times $t<0$, that is, whether the particle moved to $B$ via a jump process or whether it was placed there “by hand”.
This crucial assumption reflects that we are working in a Markovian situation, where the motion from $B$ to $C$ is independent of the history.
To formulate a recipe to calculate the WTD, we recall that the SE, Eq. (5), can also be written as a continuity equation, that is,
$$\displaystyle\partial_{t}p(x,t)+\partial_{x}j(x,t)$$
$$\displaystyle=0,$$
(7)
where $j(x,t)=-D_{0}\left(\partial_{x}p(x,t)+p(x,t)u^{\prime}(x)/(k_{\mathrm{B}}T)\right)$ is the current density. Equation (7) further shows that
$j(x,t)\,\Delta t$ (or $-j(x,t)\,\Delta t$, respectively) can be interpreted as the probability that the particle crosses position $x$ into positive (negative) $x$-direction during the time interval $[t,t+\Delta t]$.
Consequently $|j(C,t)\Delta t|$ is the probability that the particle crosses position $C$ in forward or backward direction.
This probability is similar in character to that mentioned in condition (6b). However, $|j(C,t)\,\Delta t|$ includes the probability that the particle went across $C$ (or $A$), came back, and crosses $C$ again. Such a process violates condition (6c).
Based on these considerations, we propose the following recipe to calculate the WTD $w_{A}^{B\to C}(t)$:
We initialise the probability density as
$$\displaystyle p(x,t=0)$$
$$\displaystyle=\delta(x-B)$$
(8)
according to condition (6a).
Then, we let $p(x,t)$ evolve in time according to Eq. (5). However, during this time evolution
we exclude processes from the ensemble of realisations (i.e. possible trajectories) where the particle has already reached position $C$ or $A$ [see condition (6c)].
Therefore $p(x,t)$, which is an average over this reduced set of realisations, fulfils the boundary conditions
$$\displaystyle p(x,t)$$
$$\displaystyle=0,\,x\geq\max(A,C)\quad\forall\,t$$
(9a)
$$\displaystyle p(x,t)$$
$$\displaystyle=0,\,x\leq\min(A,C)\quad\forall\,t\,.$$
(9b)
Equation (9) expresses what is often called “absorbing wall boundary conditions”. Clearly, there is some arbitrariness in defining a suitable position $A$ (the boundary “on the other side”) for a continuous system in general. However, given that we consider a continuous potential with well-defined wells (as it is the case here) one can give a clear physical meaning to the positions $A$, $B$, and $C$.
To summarise, we calculate the WTD from the relation
$$\displaystyle w_{A}^{B\to C}(t)=|j^{B}_{A,C}(C,t)|\,,$$
(10)
where the notation $j^{B}_{A,C}$ expresses the dependency of the current density on the initial and boundary conditions given in Eqs. (8) and (9), respectively.
By using the absolute value in Eq. (10) we take into account the fact that the sign of $j$ depends on the direction of motion. Indeed, motion into the positive (negative) x-direction
implies a positive (negative) sign of $j$. The numerical calculation of the WTD via the SE route is described in appendix B.
Finally, we note that the WTDs defined according to Eq. (10) also fulfil the normalisation condition Eq. (3), that is
$$\displaystyle\int_{0}^{\infty}\mathrm{d}t\left(w_{A}^{B\to C}(t)+w^{B\to A}_{C%
}(t)\right)=1.$$
(11)
By definition risken84 , the first passage time distribution (FPTD) is the probability distribution for the time $t$ the particle needs to leave a given domain for the first time. This definition implicitly assumes that the spatial probability distribution $q(x,\tau)$ to find the particle at position $x$ at an earlier time $\tau<0\leq t$ is known. In the present context, where we are considering a Markovian system, the precise value of $\tau$ is irrelevant. We thus consider the FPTD $F_{D}(t;q)$ where the domain $D$, in an one-dimensional system, is given by an interval, e.g. $D=[A,C]$. Further, if the particle was for certain at a position $B$ at $t=0$, we can specialise $q(x)=q^{0}_{B}(x)\colonequals\delta(x-B)$. The relation between the FPTD and the WTDs defined before is then given by
$$\displaystyle F_{[A,C]}(t;q^{0}_{B})=w_{A}^{B\to C}(t)+w_{C}^{B\to A}(t).$$
(12)
Equation (12) reflects the fact that the FPTD does not contain information about the direction in which the particle left the domain, whereas the WTD does.
III.3 Master equation
In Ref. emary12 , we introduced a simple master equation model of this system.
In this model, the location of the particle is described by two discrete indices: an index $n$ that describes which of the potential wells the particle finds itself in, and an internal index $\alpha=L,R$ that corresponds to whether the particle is localised in the left or right side of the well.
Relative to the centre of the well at $x=0$, the loci of these localised states are denoted $x_{L}$ and $x_{R}$.
Transitions between these states are assumed to occur at rates $\gamma^{\pm}$ to describe hopping within a single well, and $\Gamma^{\pm}$ to describe hopping between the wells. In both cases, superscript $+$ indicates movement to the right and $-$, movement to the left.
We choose the inter-well rates, $\Gamma^{\pm}$, to be twice the corresponding Kramers’ rate emary12 ; risken84 ; kramers4thorder ; the intra-well rates, $\gamma^{\pm}$, as well as the two positions, $x_{L,R}$, are set by fitting the short-time behaviour of the the first two position cumulants to the behaviour of the SE. See Ref. emary12 for details.
With the probabilities $p_{\alpha}^{(n)}$ that the particle is in state $\alpha=L,R$ of the $n$th well arranged into the vector
$$\displaystyle|\rho^{(n)}\rangle\!\rangle\equiv\left(\begin{array}[]{c}p^{(n)}_%
{L}\\
p^{(n)}_{R}\end{array}\right),$$
(13)
the system can be described with the master equation
$$\displaystyle\!\!\!\!\!\!\!\!\!\frac{d}{dt}|\rho^{(n)}\rangle\!\rangle=%
\mathcal{W}_{0}|\rho^{(n)}\rangle\!\rangle+\mathcal{W}_{+}|\rho^{(n-1)}\rangle%
\!\rangle+\mathcal{W}_{-}|\rho^{(n+1)}\rangle\!\rangle,$$
(14)
where matrices
$$\displaystyle\mathcal{W}_{+}=\left(\begin{array}[]{cc}0&\Gamma^{+}\\
0&0\end{array}\right)~{}~{}\text{and}~{}~{}\mathcal{W}_{-}=\left(\begin{array}%
[]{cc}0&0\\
\Gamma^{-}&0\end{array}\right)$$
(15)
describe jump processes between the wells and
$$\displaystyle\mathcal{W}_{0}$$
$$\displaystyle=$$
$$\displaystyle\left(\begin{array}[]{cc}-\gamma^{+}-\Gamma^{-}&\gamma^{-}\\
\gamma^{+}&-\gamma^{-}-\Gamma^{+}\end{array}\right)$$
(16)
describes intra-well processes and normalisation.
The waiting time distribution(s) in the ME approach are easy to define but before we do so, we need some more notation. We define the ‘expectation value’
$\langle\!\langle\mathcal{A}\rangle\!\rangle\equiv\langle\!\langle 1|\mathcal{A%
}|\rho_{0}\rangle\!\rangle$
where trace and intra-well stationary vectors are defined by
$$\displaystyle\langle\!\langle 1|\left(\mathcal{W}_{0}+\mathcal{W}_{+}+\mathcal%
{W}_{-}\right)$$
$$\displaystyle=$$
$$\displaystyle 0$$
$$\displaystyle\left(\mathcal{W}_{0}+\mathcal{W}_{+}+\mathcal{W}_{-}\right)|\rho%
_{0}\rangle\!\rangle$$
$$\displaystyle=$$
$$\displaystyle 0.$$
(17)
Following brandes08 , we then obtain the WTDs as
$$\displaystyle w_{ss^{\prime}}(\tau)\equiv\frac{\langle\!\langle\mathcal{W}_{s}%
e^{\mathcal{W}_{0}\tau}\mathcal{W}_{s^{\prime}}\rangle\!\rangle}{\langle\!%
\langle\mathcal{W}_{s^{\prime}}\rangle\!\rangle};\quad s,s^{\prime}=\pm$$
(18)
There are four different waiting times defined here, corresponding to the four different combinations of jumps: $s,s^{\prime}=\pm$.
In comparing with the results from BD/SE, we will only consider the diagonal WTDs $w_{ss}(\tau)$ as these are found to be the closest analogues of the definitions used in the continuous system. The reason for this is that in off-diagonal WTDs, the second jump can occur immediately after the first, and thus $\lim_{\tau\to 0}w_{ss^{\prime}}(\tau)$ for $s^{\prime}\neq s$ is finite. Clearly, this is not the case in the full dynamics and so we consider only the diagonal WTDs.
This difference between on- and off-diagonal WTDs only occurs at very short times and, as we will see in the next section, the ME description is unreliable in this regime, anyway.
IV Numerical results
The main new finding obtained so far is our definition of the WTD via the continuous SE route. In the following we compare results obtained numerically from this definition with those from BD simulations (using the minimum-based jump definition), as well as with results from the ME approach.
To be specific, we choose the positions $A$, $B$, and $C$ to be adjacent minima of the potential $u(x)$ [see Eq. (1)].
In the subsequent section IV.1,
we examine the case of zero drive ($F\!=\!0$) for two amplitudes of the periodic potential. Both symmetric and asymmetric initial (and boundary)
conditions are considered. In Sec. IV.2
we fix $u_{0}$ to $15k_{\mathrm{B}}T$ and study driven systems for two values $F<F_{c}$ where $F_{c}=u_{0}\pi/a$ is the “critical” driving force beyond which the minima in $u(x)$ are eliminated.
IV.1 Zero drive ($F=0$)
We start by considering fully “symmetric” cases, where the potential is untilted ($F=0$), and the particle jumps from the centre of a minimum (say, $x=0$) to the right ($0\to a$) or to the left ($0\to-a$).
Clearly, these jumps are characterised by the same WTD, i.e.,
$w_{-a}^{0\to a}(t)=w_{a}^{0\to-a}(t)$. Moreover, in such fully symmetric situations [with respect to the initial- and boundary conditions, and to $u(x)$], the WTD is proportional to the first passage distribution, i.e. $w_{-a}^{0\to a}(t)=1/2\,F_{[-a,a]}(t;q_{0}^{0})$ [see Eq. (12)]. However, as we will demonstrate, the SE approach is more versatile when we deviate from the symmetric case.
We thus discuss the symmetric case as a reference.
Results for the corresponding WTD at $u_{0}=5k_{\mathrm{B}}T$ and $u_{0}=15k_{\mathrm{B}}T$ are plotted in Figs. 5
and 6, respectively.
These values of $u_{0}$ are representative since, as we recall from our earlier discussion of particle trajectories (see Figs. 1 and 2), the case $u_{0}=5k_{\mathrm{B}}T$ is at the boundary between continuous and discrete motion while at $u_{0}=15k_{\mathrm{B}}T$, the motion is essentially discrete.
Figures 5 and 6 contain data from all three approaches (SE, BD, ME). All curves share the same general structure in that
the WTD is essentially (but not strictly) zero at very short times ($t\ll\tau$) and then grows with time up to a value which remains nearly constant over a range of intermediate times
(note the logarithmic time axis). Finally, at very large times the WTD decays rapidly to zero. The extension of the different regions and the actual values of the WTD depend on the potential height,
as we will analyse in detail below. From a methodological perspective we can state already here that the SE approach represents at all times a very accurate,
smooth “fit” of the (somewhat noisy) BD data. The ME approach displays deviations at short times, to which we will come back at the end of this subsection.
The growth region of the WTD can be further analysed by inspection of the jump duration distribution obtained from BD, which is included in Figs. 5
and 6. At both potential amplitudes this distribution displays a peak located at the (finite) time where the WTD grows most strongly. Moreover, the width
of the peak corresponds approximately to that of the entire growth region. We can interpret these findings as follows.
First, the fact that the peak occurs at a finite time already signals that we are looking at (more or less) continuous stochastic processes where the particles needs a finite time
to cross a barrier (see Fig. 1b)). This is also the reason that we find a growth region in the WTD at all. As we have seen from the trajectories in Fig. 1,
the jump time tends to decrease with increasing $u_{0}$. This explains the shift of the peak of the corresponding distribution towards earlier times in Fig. 6.
Second, regarding the width of the jump duration distribution, we note that only realisations of the random force with a strong bias can push the particle against the potential ascent. The larger $u_{0}$, the smaller the fraction of appropriate noise realisations and hence the sharper the distribution. This also has a direct influence on the width of the growth region of the WTD: For very short times only very short jumps can contribute to the WTD, hence the WTD rises coincidently with the jump duration distribution.
Interestingly, it is also possible to calculate the jump duration distribution via the SE approach.
To this end we just need to adjust the boundary and initial conditions. During a jump the particle leaves a minimum, say $x\!=\!0$, crosses the barrier and arrives at the next minimum, $x\!=\!a$. The difference to the jump $0\to a$ we considered above is that the particle actually leaves the minimum at $t\!=\!0$, i.e., it does not return. The essential step to the jump duration distribution is to consider an ensemble of realisations without processes where the particle comes back. Following Sec. III.2 this condition corresponds to an absorbing boundary at $x\!=\!0$. The second absorbing boundary must be located at the point of arrival, $x\!=\!a$. This suggests the construction of the WTD $w_{0}^{\epsilon\to a}(t)$ where the starting position $\epsilon$ is close to $0$ because the particle just left the minimum.
Through the normalisation Eq. (11) $w_{0}^{\epsilon\to a}(t)$ is connected with $w_{a}^{\epsilon\to 0}(t)$, which states how likely the particle comes back to $x\!=\!0$. We do not have any information about this quantity in the BD simulation, with which we compare. Therefore we need to remove this information. The total probability that the particle returns is given by $\int_{0}^{\infty}\mathrm{d}t\,w_{a}^{\epsilon\to 0}(t)$. This leads us to the following definition of the normalised jump duration distribution.
$$\displaystyle w_{j}(t)$$
$$\displaystyle=\frac{w_{0}^{\epsilon\to a}(t)}{1-\int_{0}^{\infty}\mathrm{d}t^{%
\prime}\,w_{a}^{\epsilon\to 0}(t^{\prime})}=\frac{w_{0}^{\epsilon\to a}(t)}{%
\int_{0}^{\infty}\mathrm{d}t^{\prime}\,w_{0}^{\epsilon\to a}(t^{\prime})}$$
(19)
A comparison between the BD and SE results for the distribution is given in Fig. 7. Clearly, the SE route yields very good results.
At intermediate times the WTD has a broad maximum which, for deep wells, takes the form of a plateau (see Fig. 6).
Numerical values for the WTD maxima at the two potential amplitudes are given in Table 1, where we have included values of Kramers’ escape rate $r_{\text{K}}$ in the most precise form risken84 , i.e., without the Gaussian approximation of the integrals,
$$\displaystyle r_{\text{K},\infty}^{-1}=D_{0}^{-1}\int_{-a/2}^{a/2}\mathrm{d}x%
\,e^{-\beta u(x)}\int_{0}^{a}\mathrm{d}y\,e^{\beta u(y)}\,.$$
(20)
This close relation can be understood on the basis of our SE approach where we have identified the WTD as a current
(i.e., a rate in one dimension). This current has been calculated with absorbing boundary conditions. A very similar calculation, namely by using one absorbing boundary and an infinite soft potential barrier on the other side together with the stationarity approximation risken84 ; haenggi90 , leads to Kramers’ rate. Of course, for short and long times the probability distribution is not stationary; therefore the connection between the values of the WTD and Kramers’ rate only holds at intermediate times.
At times beyond the maximum (or plateau, respectively), the WTD rapidly decays to zero. In Fig. 8 we replot the corresponding behaviour with a linear time axis and a
logarithmic $y$-axis. From that it is clearly seen that the decay can be described by a (single) exponential, i.e., $w_{-a}^{0\to a}(t)\sim e^{-\lambda t}$ in this range of times.
The corresponding decay constants $\lambda$
(as obtained from the SE approach) are listed in Table 1. We find that, for both values of $u_{0}$ considered, $\lambda$ is approximately
twice as large as Kramers’ escape rate (or plateau height, respectively). This is expected since the probability density is petering out simultaneously into both directions. Therefore the decay constant is expected to be the sum of Kramers’ rates for every barrier, in the symmetric case $\lambda\approx 2r_{\text{K}}$.
We now turn to the ME approach. As described in Sec. III.3, this approach assumes a discretised
two-state-per-well model, that is, the particle can (only) take the positions $x_{L}$ and $x_{R}$ within each well. Leaving, e.g., position
$x_{L}$, the particle performs either an intra-well jump (to $x_{R}$) or an inter-well jump to $x_{R}-a$ or to $x_{L}+a$ (via $x_{R}$). Here we are only interested in processes of the latter type, where the particle actually crosses a barrier. Further, because of translational symmetry and $F\!=\!0$ we only need to distinguish between the WTD for a “long” jump
$w_{\mathrm{long}}(t)=w_{x_{R}-a}^{x_{L}\to x_{L}+a}(t)=w_{x_{L}+a}^{x_{R}\to x%
_{R}-a}(t)$, and the WTD for a “short” jump, $w_{\mathrm{short}}(t)=w_{x_{R}-a}^{x_{R}\to x_{L}+a}(t)=x_{x_{L}+a}^{x_{L}\to x%
_{R}-a}(t)$.
The ME results shown in Figs. 5, 6, and Fig. 8 pertain to a long jump. It is seen that the
ME data become indeed consistent with those from the SE and BD approach, if one considers times beyond the growth region of the WTD.
Regarding the short-time behaviour, we find from Figs. 5 and 6
that the growth region of the WTD occurs at smaller times in the ME approach as compared to the SE and BD results. In other words, the WTD
predicts non-zero waiting times already below the mean jump duration predicted by BD. More precisely, the inflexion point of the WTD occurs at (roughly) the inverse
of the transition rate $\gamma\!=\!\gamma^{+}+\gamma^{-}$ for intra-well transitions. In a linear approximation with respect to $x_{R}$ and $x_{L}$ of Eq. (25) in emary12 one has $1/\gamma\!\approx\!\tau/(2u_{0}\pi^{2})$,
which yields a good estimate of the inflexion point. We conclude that it is the intra-well relaxation
(which, in turn, is governed by the potential amplitude) which is the main ingredient determining the short time behaviour of the WTD given by the ME. Indeed, because the ME is based on
a discrete model (and thus neglects travel times) we would not expect that the corresponding WTD is connected to the mean jump duration (as it was the case within the continuous SE approach).
We finally note that asymmetric initial conditions such as those used in ME ($x_{L/R}\neq 0$)
can also be incorporated into the SE approach. As described before, we can distinguish between the WTD for a “long” jump,
$w_{\mathrm{long}}(t)$, and the WTD for a “short” jump $w_{\mathrm{short}}(t)$.
In Fig. 9 we show these two WTDs together with our earlier SE result for $w_{-a}^{0\to a}(t)$.
At short times, $w_{\mathrm{short}}(t)$ has a global maximum
which exceeds the corresponding values of $w_{\mathrm{long}}(t)$ and the maximum of $w_{-a}^{0\to a}(t)$, that is, Kramers’ escape rate. At subsequent time
the two WTDs $w_{\mathrm{long/short}}$ then merge at a value slightly below $w_{-a}^{0\to a}$. This is because the blip of $w_{\mathrm{short}}$ reduces the survival probability which in turn reduces the escape rate. The inset of Fig. 9 shows the probability densities $p(x,t)$ corresponding to $w_{\mathrm{long}}$ and $w_{\mathrm{short}}$ at $t\!=\!0.02\tau$ where the WTDs deviate from each other. At $t\!=\!0.02\tau$ the asymmetrically initialised densities have been broadened by diffusion to reach the boundaries, but yet not strong enough to eliminate the asymmetry.
IV.2 Driven system ($F>0$)
We now consider driven systems. Again, minima of the potential occur every $a$, the first being located at $x=m\colonequals a/(2\pi)\arcsin(F/F_{c})$. Because of the tilt of the potential,
we have to distinguish between jumps to the left, $w^{-}(t)=w_{m+a}^{m\to m-a}(t)$, and jumps to the right, $w^{+}(t)=w_{m-a}^{m\to m+a}(t)$. Clearly, these WTDs provide
directional information, which cannot be extracted from the first passage time distribution.
Results for the WTDs in the “deep well” case are shown in Fig. 10.
It is seen that the general structure of the WTD (consisting of a growth region, a plateau, and then a rapid decay) is the same as that at $F=0$ (see Fig. 6).
One main effect of increasing $F$ from zero is that the global maximum of $w^{+}$ increases, while the time range corresponding to the plateau shortens. In other words,
the average waiting time decreases and it occurs with larger probability. This is plausible, since a non-zero driving force leads to a lower effective barrier
$\Delta u(u_{0},F)=u_{0}(\sqrt{1-(F/F_{c})^{2}}-F/F_{c}\arccos(F/F_{c}))$ in the driving direction. The opposite effects occurs against the driven direction, as reflected by the decreasing maxima
in $w^{-}(t)$. We also note that, for each $F$, the effects in the two WTDs are coupled via the the normalisation condition, i.e., the maximum in one of the WTDs can only grow
at the expense of the other one.
Comparing the different methods, we see from Fig. 10 that the WTDs calculated with the SE agree with the BD results at all times. The WTDs given by the ME are consistent with SE and BD results for intermediate and long times. Again we observe a deviation for short times because, as discussed in the previous section, the short time scale in the two-state-per-well model describes intra-well relaxation and there is no time scale connected to the mean jump duration.
V Conclusion
In present paper we have introduced and compared several routes to calculate the WTD in a system
with continuous, Markovian dynamics. Specifically, we have considered
the example of a Brownian particle in an one-dimensional (tilted) washboard potential, and focused on cases where the potential barriers are large
against the thermal energy.
Traditionally, the WTD in such a situation is calculated by analysing trajectories obtained from BD simulations of the Langevin equation.
One main new result of the present paper is a definition of the WTD on the basis of the corresponding Smoluchowski equation (SE).
Here, the WTD is identified as a current calculated from the SE with adsorbing boundaries.
The resulting WTD is closely related to the first passage time distribution; however, the WTD contains directional information which is absent in the FPTD. This becomes particularly important in spatially asymmetric situations. Moreover,
our definition of the WTD is more versatile in that it allows for unusual (asymmetric) initial and/or boundary conditions.
Analysing a variety of systems with different initial conditions, as well as with and without external drive, we find in all cases full quantitative agreement
between the SE and BD results for the WTD. We also stress that, due to our rather general definition of the WTD in the SE approach, we are able
to calculate additional quantities such as the jump duration distribution. The latter is crucial for understanding the growth behaviour of the WTD at short times.
In addition to the SE (and BD) method, we have also presented a fully analytical master equation approach to the WTD. The ME approach is based on the discrete,
two-state-per well model introduced by us in an earlier study emary12 . Comparing the results with those from SE and BD it turns out that the ME yields a very accurate
WTD at intermediate times, where it exhibits a plateau, as well as at long times, where it has an exponential tail. Only the short-time behaviour differs due to the
fact that the ME model involves an “intra-well” relaxation time which is absent in the continuous approaches.
From a computational point of view the BD route to the WTD seems, at first glance, to be the most straightforward one. However, closer inspection shows that the results
are quite dependent on the definition of “jumps”. Furthermore, the results are often quite noisy. Here, the SE approach where noise is averaged out a priori, is clearly superior.
We stress, however, the solution is computationally expensive and “fragile” when using a standard solver. In Appendix B
we have therefore sketched an alternative route to obtain the WTD from the SE based on a Fourier transform.
Finally, our SE approach to the WTD can be easily generalised to systems characterised by a different potential, to interacting systems,
and to systems with higher dimensionality. For example, for processes involving more than one spatial dimension, one would
simply replace the absorbing boundary value by an absorbing surface and calculate the WTD
as an integral over the probability current over a (part of the) surface. Another generalisation concerns the character of the dynamics, which we here assumed to be Markovian.
We claim that even for a non-Markovian SE (see guill99 ; lichtner12 ; loos14 for examples), which would involve memory kernels and thus has higher dimension in time,
the identification of the WTD with the probability current remains intact. Finally, our generalisation of the WTD for continuous systems opens the route to calculate other quantities typically reserved for discrete (Markovian or non-Markovian) systems, such as the idle-time distribution albert12 . Work in these directions is in progress.
Acknowledgements.
We gratefully acknowledge stimulating discussions with T. Brandes.
This work was supported by the Deutsche Forschungsgemeinschaft
through SFB 910 (projects B2 and A7).
Appendix A WTD from BD simulations
To obtain a reliable WTD the recorded histograms must have a high statistical quality at all times where the WTD is nonzero. Any deviation influences the normalisation and therefore shifts the entire WTD. Typically, the domain of non-zero WTD spans several orders of magnitude. Moreover, for large times the WTD decays exponentially (see Fig. 8) which further enhances the computational effort due to the need of many long simulations. These considerations imply that the histogram bins must be sufficiently small to capture the time variations of the WTD and sufficiently large to yield a good statistical average. Here we use a piece-wise constant bin size distribution, which we adjust manually.
To achieve the necessary accuracy in the generated trajectories we choose a time step of $\Delta t=10^{-5}\tau$. This sets the mean size of a fluctuation to $4\cdot 10^{-3}a$ which is necessary to sample the potential sufficiently accurate.
The simulation starts with the particle at the first minimum $m=a/(2\pi)\arcsin(F/F_{c})$. During the simulation the trajectory is analysed to record leavings and arrivals of the particle at other minima $m\pm ka,\,k\in\mathbb{Z}$. Because one simulation alone cannot produce enough jumps in a reasonable computation time, we run several simulations in parallel. The least amount of jumps occur at $u_{0}\!=\!15k_{\mathrm{B}}T$ and $F\!=\!0$. In this case we run $832$ simulations in parallel or successively, each computing $500$ jumps, to achieve a reasonable resolution for the long time part of the WTD.
Short waiting times occur even less frequently than large ones. To calculate the WTD at these short times a new set of simulations is started. Contrary to the regular trajectory simulation, the new simulations stop if the jump does not happen in a given time interval $u$ (with $\log_{10}(u/\tau)\in\{-1,\dots,2\}$). Each histogram for the Figs. 5, 6 and 10 is made of about $10^{4}$ of those short and $10^{5}$ long waiting times.
Appendix B WTD from the SE
To calculate the WTD from the SE Eq. (5) we mainly use the FTCS (forward-time-centered-space) discretisation scheme fletcher91 . Because the whole process is determined by a very small current $j$ over the barrier and, correspondingly, small values for the probability density (see inset in Fig. 9) we need very small discretisation steps $\Delta x$, a typical example being $\Delta x=0.0016a$. Nevertheless, the FTCS method works for all parameter sets except for $u_{0}\!=\!15k_{\mathrm{B}}T$, $F\!=\!0$ where the solution of the SE just takes too long. To calculate the long-time part of the WTD in this particular case we use a different method involving a Fourier transform. To this end we write the probability density as
$$\displaystyle p(x,t)$$
$$\displaystyle=\sum_{n=1}^{N}p_{n}(t)\,s_{n}(x)\,.$$
(21a)
The sum involves the first $$N$$ modes
$$\displaystyle p_{n}(t)$$
$$\displaystyle=\frac{1}{a}\int_{-a}^{a}\mathrm{d}x\,s_{n}(x)\,p(x,t)\,,$$
(21b)
where we exploit that for the special case $$F\!=\!0$$ with $$A$$, $$B$$, $$C$$ being successive minima of $$u(x)$$ the basis functions $$s_{n}$$ can be chosen as
$$\displaystyle s_{n}(x)$$
$$\displaystyle=\sin\left(\pi n\frac{x+a}{2a}\right)\,.$$
(21c)
Applying the Fourier transform (21) to Eq. (5) yields
$$\displaystyle\dot{p}_{n}(t)=\frac{\pi^{2}}{4a^{2}}\Big{(}-n^{2}p_{n}$$
$$\displaystyle+nU(p_{n-4}\Theta(n\!-\!4)-p_{n+4}-p_{4-n}\Theta(4\!-\!n))\Big{)}\,,$$
(22)
where $\Theta(x)$ is the Heaviside-step function. The numerical solution of Eq. (22) takes even longer than FTCS. Therefore we introduce the matrix based formulation $\dot{\mathbf{p}}(t)=A\cdot\mathbf{p}$
where $\mathbf{p}(t)$ contains all $p_{n}(t)$. The solution reads
$$\displaystyle\mathbf{p}(t)=\exp(tA)\cdot\mathbf{p}(0)\,,$$
(23)
which we calculate by decomposing the matrix $A$ into $S\cdot D\cdot S^{-1}$. In the latter expression the diagonal matrix $D$ consists of the eigenvalues of $A$ and $S$ contains the corresponding eigenvectors column by column. Inserting the diagonalisation in Eq. (23) we arrive at
$$\displaystyle\mathbf{p}(t)$$
$$\displaystyle=e^{t\,S\cdot D\cdot S^{-1}}\mathbf{p}(0)=S\cdot e^{tD}\cdot S^{-%
1}\cdot\mathbf{p}(0)\,,$$
(24)
where $e^{tD}$ is also diagonal. The computation of $\mathbf{p}(t)$ via Eq. (24) is fast and reliable. Specifically, for the case $u_{0}\!=\!15k_{\mathrm{B}}T$ we find that $100$ Fourier modes are sufficient (computations finish in minutes even for $10^{4}$ modes).
A further advantage of the diagonalisation is that it provides a simple access to the long time dynamics. This is because the limit $\lim_{t\to\infty}\partial_{t}\ln w(t)$ gives the largest eigenvalue $d$
, and thus $w(t)\sim e^{dt}$. Note that $d$ is negative and close to zero.
References
(1)
E. W. Montroll and G. H. Weiss, J. Math. Phys. 6, 167 (1965).
(2)
H. Scher and M. Lax, Phys. Rev. B 7, 4491 (1973).
(3)
J. W. Haus and K. W. Kehr, Phys. Rep. 150, 263 (1987).
(4)
E. R. Higgins, H. Schmidle, and M. Falcke, J. Theo. Bio. 259, 338 (2009).
(5)
J. Qian, Y. Lin, H. Jiang, and H. Yao, Appl. Phys. Lett. 103, 223702 (2013).
(6)
S. Saha, A. Sinha and A. Dua, J. Chem. Phys. 137, 045102 (2012).
(7)
B. P. English, W. Min, A. M. van Oijen, K. T. Lee, G. Luo, H. Sun, B. J. Cherayil, S. C. Kou, and
X. S. Xie, Nat. Chem. Biol. 2, 87 (2006).
(8)
F. Mainardi, M. Raberto, R. Gorenflo, and E. Scalas, Physica A 287, 468 (2000).
(9)
M. J. Skaug, J. Mabry, and D. K. Schwartz, Phys. Rev. Lett. 110, 256101 (2013).
(10)
T. Prager and L. Schimansky-Geier, J. Stat. Phys. 123, 391 (2006).
(11)
R. Ganapathy, M. R. Buckley, S. J. Gerbode, and I. Cohen, Science 22, 445 (2010).
(12)
R. D. L. Hanes, C. Dalle-Ferrier, M. Schmiedeberg, M. C. Jenkins, and S. U. Egelhaaf, Soft Matter 8, 2714 (2012).
(13)
Y. Jung, J. P. Garrahan, and D. Chandler, J. Chem. Phys. 123, 084509 (2005).
(14)
P. Chaudhuri, Y. Gao, L. Berthier, M. Kilfoil, and W. Kob,
J. Phys.: Condens. Matter 20, 244126 (2008).
(15)
J. W. Ahn, B. Falahee, C. D. Piccolo, M. Vogel, and D. Bingemann, J. Chem. Phys. 138, 12A527 (2013).
(16)
B. Doliwa and A. Heuer, Phys. Rev. E 67, 030501(R).
(17)
A. Heuer, B. Doliwa, and A. Saksaengwijit, Phys. Rev. E 72, 021503 (2005).
(18)
O. Rubner and A. Heuer, Phys. Rev. E 78, 011504 (2008).
(19)
T. Brandes, Ann. Phys. (Berlin) 17, 477 (2008).
(20)
C. Flindt, C. Fricke, F. Hohls, T. Novotny, K. Netocny, T. Brandes, and R. J. Haug, Proc. Nat. Acad. Sci. 106, 10116 (2009).
(21)
L. Rajabi, C. Pöltl, and M. Governale, Phys. Rev. Lett. 111, 067002 (2013).
(22)
M. Albert, C. Flindt, and M. Büttiker, Phys. Rev. Lett. 107, 086805 (2011).
(23)
F. Evers, R. D. L. Hanes, C. Zunke, R. F. Capellmann, J. Bewerunge, C. Dalle-Ferrier, M. C. Jenkins, I. Ladadwa, A. Heuer, R. Castañeda-Priego, and S. U. Egelhaaf, Eur. Phys. J. Spec. Top. 222, 2995 (2013).
(24)
H. Risken, The Fokker-Planck Equation (Springer, Berlin,
1984).
(25)
P. Tierno, P. Reimann, T. H. Johansen, and F. Sagues,
Phys. Rev. Lett. 105, 230602 (2010).
(26)
C. Dalle-Ferrier, M. Krüger, R. D. L. Hanes, S. Walta, M. C. Jenkins, and S. U. Egelhaaf,
Soft Matter 7, 2064 (2011).
(27)
J. Xing, H. Wang, and G. Oster, Biophys. J. 89, 1551 (2005).
(28)
M. Siler and P. Zemanek,
New J. Phys. 12, 083001 (2010).
(29)
B. J. Lopez, N. J. Kuwada, E. M. Craig, B. R. Long, and H. Linke,
Phys. Rev. Lett. 101, 220601 (2008).
(30)
I. Zapata, R. Bartussek, F. Sols, and P. Hänggi, Phys. Rev. Lett. 77, 2292 (1996).
(31)
M. A. Shahzamanian, M. Eatesami, and H. Yavary, Supercond. Sci. Technol. 20, 640 (2007).
(32)
G. Augello, D. Valenti, B. Spagnolo, Eur. Phys. J. B 78, 225 (2010).
(33)
I. Bonnet and P. Desboilles, Eur. Phys. J. E 34, 25 (2011).
(34)
E. R. Kay, D. A. Leigh, and F. Zerbetto, Angew. Chem. Int. Ed. 46, 72 (2007).
(35)
A. Ros, R. Eichhorn, J. Regtmeier, T. T. Duong, P. Reimann, and D. Anselmetti,
Nature 436, 928 (2005).
(36)
P. Reimann, C. Van den Broeck, H. Linke, P. Hänggi, J. M. Rubi, and A. Pérez-Madrid,
Phys. Rev. Lett. 87, 010602 (2001).
(37)
C. Emary, R. Gernert, and S. H. L. Klapp, Phys. Rev. E 86, 061135 (2012).
(38)
D. L. Ermak, J. Chem. Phys. 62, 4189 (1975).
(39)
H. A. Kramers, Physica 7, 284 (1940).
(40)
P. Reimann, G. J. Schmid, and P. Hänggi, Phys. Rev. E 60, R1 (1999).
(41)
In this context, we have used the $4^{th}$ order approximation for Kramers’ rate risken84 .
(42)
P. Hänggi, P. Talkner, M. Borkovec, Rev. Mod. Phys. 62, 251 (1990).
(43)
S. Guillouzic, I. L’Heureux, and A. Longtin, Phys. Rev. E 59, 3970 (1999).
(44)
K. Lichtner, A. Pototsky, and S. H. L. Klapp, Phys. Rev. E 86, 051405 (2012).
(45)
S. A. M. Loos, R. Gernert, S. H. L. Klapp, Phys. Rev. E 89, 052136 (2014).
(46)
M. Albert, G. Haack, C. Flindt, M. Büttiker, Phys. Rev. Lett. 108, 186806 (2012).
(47)
C. A. J. Fletcher, Computational techniques for Fluid Dynamics, Volume 1 (2nd edition, Springer, 1991). |
Matrix Theory on ALE Spaces and Wrapped Membranes
David Berenstein and Richard Corrado
Research supported in part by the Robert A. Welch
Foundation and NSF Grant PHY 9511632.
[email protected]
[email protected]
(March 5, 1998)
††preprint: UTTG–02–98
hep-th/9803048
\oneaddress
Theory Group, Department of Physics
University of Texas at Austin
Austin TX 78712 USA
\Abstract
We study the properties of wrapped membranes in matrix
theory on ALE spaces. We show that the only BPS bound states of wrapped
membranes that can form are roots of the $A$-$D$-$E$ group. We
determine a bound on the energy of a bound state and find the correct
dependence on the blow-up parameters and longitudinal momentum
expected from M-Theory. For the $A_{n-1}$ series, we
construct explicit classical solutions for the wrapped membrane bound
states. These states have a very rich structure and have a natural
interpretation in terms of noncommutative geometry.
In the $A_{1}$ case, we examine the spectrum of excitations
around the wrapped membrane solution and provide an explicit
calculation of their energies. The results agree exactly with supergravity
calculations.
1 Introduction
The M(atrix)-Theory [1] proposal for a
non-perturbative description of M-Theory has been demonstrated to
properly capture M-Theory physics in a variety of settings (see
[2, 3, 4] for
recent reviews). However, several important areas where our
understanding of matrix theory is incomplete still remain. For
example, the fact that matrix theory captures the physics of
linearized supergravity is fairly well
understood [5]. The reasons for its apparent
failure to capture supergravity results in other
situations [6, 7, 8, 9] seem to be subtleties in our understanding
of the discrete light-cone quantization (DLCQ) of M-Theory, namely
that the low-energy description of DLCQ M-Theory is not quite
supergravity [2, 10, 3].
Another area which requires further study is the
compactification of matrix theory, in particular on curved manifolds.
One promising case is that of “compactification” on ALE spaces,
which has been conjectured to be described by the theory of D0-brane
partons moving on the ALE space [11, 6, 12, 13].
In this paper we continue the study of matrix theory on ALE spaces. In
Section 2, we give a very brief review of ALE matrix
theory. Our emphasis will be to further examine the characteristics of
the description of membranes which are wrapped around homology
2-cycles as recently described
in [14, 15, 16]. In
Section 3, we focus on the existence of wrapped membrane
BPS bound states in the wrapped membrane matrix model presented
in [16]. We demonstrate that such bound states
must be roots under the $A$-$D$-$E$ group. We then derive a bound on
the energy of such a bound state. For the case of the $A_{n-1}$
series, we explicitly construct the bound states for all roots. In
Section 4, we
briefly discuss how the membrane solution fits into the framework of
noncommutative geometry and thereby satisfies the properties of
spherical membranes [17, 18]. In
Section 5, we discuss the spectrum of excitations
of the wrapped membrane and calculate the energies of excitations
around the $A_{1}$ solution, finding agreement with supergravity expectations.
2 A Brief Review of ALE Matrix Theory
A matrix description of M-Theory on an ALE space must possess several
crucial ingredients if it is to be considered both correct and
useful. A candidate formalism for such a description has been provided
by the worldvolume effective theories describing D0-branes moving on
an ALE
space [19, 20, 21].
Let us briefly review the construction of these models and examine the
spacetime features of M-Theory that they capture.
The ALE matrix models are given by the dimensional reduction to
quantum mechanics of the six-dimensional gauged supersymmetric sigma
models appearing in the hyperkähler quotient
construction [22, 23] of the ALE
space. The field content is summarized by a quiver diagram, which is
based on the $A$-$D$-$E$ extended Dynkin diagram, such as the
$A_{n-1}$ diagram shown in Figure 1.
For each vertex of the diagram there is a gauge group $U(N_{i})$, where
$N_{i}=Nk_{i}$, with $N$ the number of D0-branes and $k_{i}$ the Dynkin
label of the vertex, as well as a vector multiplet, $(A_{i},a_{i})$ (in
terms of $d=4$, $\CN=1$ vector and chiral superfields), in the
$(1,\ldots,\text{ad}(U(N_{i})),\ldots,1)$ representation. For each edge
of the diagram there is an associated hypermultiplet,
$(x_{i,i+1},y_{i+1,i})$, in the fundamental–anti-fundamental
representation, $(1,\ldots,1,N_{i},\overline{N}_{i+1},1,\ldots,1)$, of
the neighboring gauge groups.
The Lagrangian is the most general one compatible with
the gauge symmetry and $d=6$, $\CN=1$ SUSY. However, since we are in
$d=1$ dimensions, the question of the mass dimension of the fields and
coupling constants is an issue. We will choose string units, with
$T_{A}=1$. All fields have mass dimension $-1$ and the coupling constant
has dimension $-\frac{1}{2}$. This requires us to scale the
hypermultiplets so that an inverse square of a coupling constant
appears out in front of the terms in the Lagrangian that they appear
in. We will choose this to be the “average” coupling constant, defined by
$$\frac{1}{g^{2}}=\sum_{i}\frac{k_{i}}{g_{i}^{2}}.$$
(2.1)
We also scale the D and F-terms by $2/g^{2}$. In terms of
$d=4$, $\CN=1$ superfields, our Lagrangian is
$$\begin{split}\displaystyle\CL=\sum_{i}&\displaystyle\left[\left(\frac{1}{16N_{%
i}g_{i}^{2}}\int d^{2}\theta\,W_{i}^{2}+\text{c.c.}\right)+\frac{1}{g_{i}^{2}}%
\int d^{4}\theta\,\bar{a}_{i}e^{A_{i}}a_{i}e^{-A_{i}}\right.\\
&\displaystyle+\frac{1}{g^{2}}\int d^{4}\theta\,\left(\bar{x}_{i,i+1}e^{A_{i}}%
x_{i,i+1}e^{-A_{i+1}}+\bar{y}_{i+1,i}e^{A_{i+1}}y_{i+1,i}e^{-A_{i}}\right)\\
&\displaystyle+\left(\frac{1}{g^{2}}\int d^{2}\theta\,\left(y_{i+1,i}a_{i}x_{i%
,i+1}-x_{i-1,i}a_{i}y_{i,i-1}\right)+\text{c.c.}\right)\\
&\displaystyle\left.+\frac{2}{g^{2}}\int d^{4}\theta\,d_{i}A_{i}+\left(\frac{2%
}{g^{2}}\int d^{2}\theta\,f_{i}a_{i}+\text{c.c.}\right)\right].\end{split}$$
(2.2)
The coupling constant satisfies
$$\frac{1}{g^{2}}=T_{D0}.$$
(2.3)
Then it is clear that the $1/g_{i}^{2}$ are the masses of the fractional
D0-branes [20, 11, 15] associated to each vertex.
The coefficients of the D and F-terms are constrained by the
requirement that a supersymmetric ground state exists. This requires
that
$$\sum_{i}k_{i}f_{i}=\sum_{i}k_{i}d_{i}=0.$$
(2.4)
For $N_{i}=Nk_{i}$, there are zero-energy ground states. Under the
$SU(2)_{R}$ R-symmetry, the $\vec{\zeta}_{i}=(f_{i},d_{i})$ form
triplets. In fact, according to the hyperkähler quotient
construction, the $\vec{\zeta}_{i}$ are the blow-up parameters for the
ALE space. With the normalization chosen in (2.2), the
area of the $i^{th}$ $\BP^{1}$ is given by $4\pi|\vec{\zeta}_{i}|^{2}$.
For simplicity, one typically chooses all the gauge
couplings to be the same, but this is not required. As we will see
later, these coupling constants have an interpretation as Wilson lines
for the $A$-$D$-$E$ gauge group in the DLCQ M-Theory description.
Applying the matrix prescription [1], one is lead
to conjecture that, in the limit $N\rightarrow\infty$, $g\rightarrow\infty$, the D0-branes described by (2.2) are the
partons of the infinite-momentum frame description of M-Theory on the
ALE space [11, 6]. On the other hand,
the finite $N$, finite $g$ matrix models would be conjectured to
provide a description of the DLCQ of
M-Theory [24] on the ALE
space [8].
Several pieces of evidence support the M-Theory interpretation of
these ALE matrix models. First, the models contain the geometry of the
ALE space. For $N=1$, we have the standard $U(1)$ gauged
supersymmetric sigma model. After taking the hyperkähler
quotient [22],
we recover the target space $\BR^{5}\times\CM_{\vec{\zeta}}$.
Here we use $\CM_{\vec{\zeta}}$ to denote the ALE space parameterized
by the blowup parameters $\vec{\zeta}$. For larger $N$,
the moduli space of the theory is
$$(\BR^{5}\times\CM_{\vec{\zeta}})^{N}/S_{N},$$
(2.5)
so that the ALE geometry is recovered along the flat directions of the
classical ground states of the system.
The models have the equivalent of $\CN=1$ supersymmetry in six
dimensions, or eight supercharges, which is the same amount of
supersymmetry present in light-cone M-Theory on an ALE space. The
gauge group $\prod_{i}U(N_{i})$ contains an overall $U(1)$ factor which is a
linear combination of the $U(1)$s for each $U(N_{i})$ factor. The vector
associated to this $U(1)$ is decoupled from the rest of the dynamics. In the
matrix model, these decoupled degrees of freedom are associated to the
center-of-mass motion in the transverse space. The amount of
supersymmetry provides for a 256-fold degeneracy in the continuum
spectrum of the theory, which allows these states to be
identified with the gravity multiplet. The SUSY
might also allow for the existence of non-renormalization theorems
that could lead to correct results for interactions.
It is crucial that each finite mass BPS object in M-Theory have an
explicit description as some state in the matrix quantum
mechanics. Furthermore, whenever BPS branes come together, or a
2-cycle which has a membrane wrapped around it shrinks to zero-size,
an enhanced gauge symmetry must appear in the quantum dynamics.
The mechanism by which wrapped membranes and enhanced gauge symmetry
appear in the model was introduced by
Douglas [11] and elaborated upon
in [6, 14, 8, 15, 16].
From the Kronheimer construction [23], the vertices
of the extended Dynkin diagram are associated to the homology
2-cycles (which are $\BP^{1}$s) of the ALE
space. In [14], it was proposed that, in
the blow-down, states which correspond to wrapped membranes exist on
the Coulomb branch of the quiver gauge theory. In particular,
a configuration describing a membrane wrapped on the $i^{th}$ $\BP^{1}$
would be described by a metastable state for which the hypermultiplets
corresponding to that vertex vanished, $x_{i-1,i}=x_{i,i+1}=\cdots=0$. This picture was
further discussed in [15], where a
proposal was made for its extension to the blown-up ALE spaces.
A different approach was taken in [16], where
it was proposed that the matrix description of membranes wrapped on
these $\BP^{1}$s could be obtained by considering the ALE matrix models
described above, but now relaxing the condition that $N_{i}=Nk_{i}$. Instead, one should take
$$N_{i}=Nk_{i}+r_{i},$$
(2.6)
where $N$ is chosen such that the set of $r_{i}>0$ are as small as
possible. The resulting matrix model describes $N$ D0-branes
propagating on a ALE space with $r_{i}$ membranes wrapped around the
$i^{th}$ $\BP^{1}$. It is this latter approach which we will consider in
the rest of the paper.
If all of the $r_{i}=nk_{i}$ for some integer $n$ (we can include the
possibility that $n=0$, in which case $N_{i}=Nk_{i}$), then the
corresponding homology 2-cycle is trivial, and the configuration of
wrapped membranes can decay into a collection of D0-branes. A single
wrapped membrane corresponds to a simple root of the spacetime gauge
group, while the other roots have been conjectured to be given by
bound states of these simple roots [11]. We will
demonstrate this result explicitly in Section 3.
The objective of the following sections is to give a detailed account
of several features of wrapped membranes in the ALE matrix models. We
give a complete description of what BPS bound states of wrapped membranes
form in the quantum mechanical system and then construct explicit bound
state solutions for the $A_{n-1}$ series. We give a short discussion
of the noncommutative geometry properties of the solutions and exhibit
the spherical membrane properties for the explicit $A_{1}$
solution. For the $A_{1}$ ALE space, we examine the spectrum of
excitations of the wrapped membrane, providing an explicit calculation
of their energies. We find that the large $N$ behavior of the excitations
matches supergravity expectations.
3 Bound States of Wrapped Membranes
We would like to determine the conditions under which a
(non-threshold) BPS bound state of membranes can form. These are ground
states of the interacting part of the theory that, classically,
completely break the gauge symmetry group. Since the argument is
classical, any solution of the F-terms for the hypermultiplets can be
deformed to set the gauge fields equal to zero. In this way, the
symmetry breaking pattern can be chosen to preserve the $SO(5)$
symmetry of the directions transverse to the ALE.
In order to determine the masses of the gauge fields, the argument can
be extended to the six-dimensional quiver gauge theory. The Higgs
mechanism requires that the extra degrees of freedom that give mass to
a vector multiplet come from “eating” a hypermultiplet. The number
of vector multiplets that can be Higgsed can be obtained by counting
the total number of vector multiplet degrees of freedom
($4\sum_{i}N_{i}^{2}$) and subtracting the number of hypermultiplet
($4\sum_{i,j\,\text{adj.}}N_{i}N_{j}$) and decoupled $U(1)$ degrees of
freedom (4). The resulting number must be smaller than zero or otherwise
there remains a flat direction corresponding to a vector multiplet, in
which case the membrane configuration can be separated classically
into two separate bound states. We therefore need $N_{i}$ such that
$$0\geq 4\left(\sum_{i}N_{i}^{2}-\sum_{i,j\,\text{adj.}}N_{i}N_{j}-1\right)=2%
\left(\sum_{i,j}\hat{C}_{ij}N_{i}N_{j}-2\right),$$
(3.1)
where $\hat{C}_{ij}$ is the extended Cartan matrix of the
corresponding $A$-$D$-$E$ group. One immediately recognizes the
quantity $\sum_{i,j}\hat{C}_{ij}N_{i}N_{j}$ as the squared norm of the
vector $N_{i}$ in the lattice of the algebra. Since the
inequality (3.1) is satisfied whenever
$\sum_{i,j}\hat{C}_{ij}N_{i}N_{j}\leq 2$, we see
that the state must either be neutral under the Cartan subalgebra (and
is therefore a collection of D0-branes) or it is a root.
It is fairly straightforward, in general, to
determine a bound on the energy of the state. Wrapped membrane
bound states should form massive vector multiplets in seven spacetime
dimensions. Since the fermionic zero-modes corresponding to the
decoupled $U(1)$ center-of-mass motion already give rise to a 16-fold
degeneracy, the ground state of the interacting part of the theory must be
non-degenerate. This is difficult to prove in general, but we can
still obtain a bound on the energy rather easily. We will then show
that these states exist explicitly in the case of the $A_{n-1}$ ALE spaces.
Let us assume that the F-terms are zero, then, after computing the traces,
the Hamiltonian for the D-terms we obtain from (2.2) is
$$H_{D}=\sum_{i}\left(\frac{N_{i}D_{i}^{2}}{2g_{i}^{2}}-\frac{2N_{i}d_{i}D_{i}}{%
g^{2}}\right).$$
(3.2)
Since the decoupled $U(1)$ is the sum of the $U(1)$s at each vertex, the
corresponding D-term, $D_{\text{dec.}}$, is given by the sum
$$D_{\text{dec.}}=\frac{1}{\sum_{i}\frac{N_{i}}{g_{i}^{2}}}\sum_{i}\frac{N_{i}D_%
{i}}{g_{i}^{2}}.$$
(3.3)
Therefore, each D-term has an expansion as
$$D_{i}=D_{\text{dec.}}+D_{i,\text{int.}},$$
(3.4)
where the $D_{i,\text{int.}}$ are orthogonal to each other and
$D_{\text{dec.}}$ and satisfy
$$\sum_{i}\frac{N_{i}D_{i,\text{int.}}}{g_{i}^{2}}=0.$$
(3.5)
Using the expansion (3.4) in (3.2), we
find that the Hamiltonian for the decoupled $U(1)$ is
$$H_{\text{dec.}}=D_{\text{dec.}}\sum_{i}N_{i}\left(\frac{D_{\text{dec.}}}{2g_{i%
}^{2}}-\frac{2d_{i}}{g^{2}}\right).$$
(3.6)
This is minimized by
$$D_{\text{dec.}}=-\frac{2\sum_{i}N_{i}d_{i}}{g^{2}\sum_{i}\frac{N_{i}}{g_{i}^{2%
}}},$$
(3.7)
which leads to a bound on the energy
$$E\geq\frac{2\left(\sum_{i}N_{i}d_{i}\right)^{2}}{g^{4}\sum_{i}\frac{N_{i}}{g_{%
i}^{2}}}.$$
(3.8)
From this expression for the bound, we see that the
D0-brane charge of the state is proportional to
$\sum_{i}\frac{N_{i}}{g_{i}^{2}}$. If we take the $N_{i}$ as in (2.6)
and apply (2.4), we find that
$$E\geq\frac{2\left(\sum_{i}r_{i}d_{i}\right)^{2}}{g^{4}\sum_{i}\frac{N_{i}}{g_{%
i}^{2}}}.$$
(3.9)
A similar calculation works when the F-terms are non-zero. Defining
$F_{\text{dec.}}$ in a manner analogous to (3.3), we find
the solution
$$F_{\text{dec.}}=-\frac{2\sum_{i}N_{i}f_{i}}{g^{2}\sum_{i}\frac{N_{i}}{g_{i}^{2%
}}}.$$
(3.10)
Putting the results together, we find a bound
$$E\geq\frac{2}{g^{4}\sum_{i}\frac{N_{i}}{g_{i}^{2}}}\left(\bigl{(}\sum_{i}r_{i}%
d_{i}\bigr{)}^{2}+\big{|}\sum_{i}r_{i}f_{i}\bigr{|}^{2}\right)=\frac{2\bigl{|}%
\sum_{i}r_{i}\vec{\zeta}_{i}\bigr{|}^{2}}{g^{4}\sum_{i}\frac{N_{i}}{g_{i}^{2}}}.$$
(3.11)
This is the light-cone energy of an object with mass
$$m=\frac{2\bigl{|}\sum_{i}r_{i}\vec{\zeta}_{i}\bigr{|}}{g^{2}}$$
(3.12)
and longitudinal momentum
$$p^{+}=\sum_{i}\frac{N_{i}}{g_{i}^{2}}.$$
(3.13)
In particular, it is
clear that by tuning the $g_{i}$, one assigns different amounts of
longitudinal momentum to each simple root. In the DLCQ, the parameters
responsible for this are the Wilson lines of the spacetime gauge group
around the light-like circle.
In the large $N$ limit, we obtain from (3.11)
$$E\geq\frac{2\bigl{|}\sum_{i}r_{i}\vec{\zeta}_{i}\bigr{|}^{2}}{Ng^{2}}\left(1-O%
(1/N)\right).$$
(3.14)
We will show explicitly below that there are solutions that satisfy
the bound for any $N$. Moreover, from the DLCQ interpretation
of the coupling constants, we know that at large $N$ the effects of
the Wilson lines should disappear. In particular this implies that the
ratios $g_{i}/g_{j}$ become unimportant in the large $N$ limit and
therefore that some of these coupling constants decouple from the
dynamics in the large $N$ limit. This can serve as a very useful
calculational tool, and might serve to produce some
non-renormalization theorems in the large $N$ limit.
As a check on our results, we can calculate the tension of
the membrane from (3.12). For a single membrane state
$$m=\frac{2\bigl{|}\vec{\zeta}\bigr{|}}{g^{2}}=T^{(2)}A,$$
(3.15)
Since the area of the 2-sphere is $A=4\pi|\vec{\zeta}|$,
from (2.3)
we recover (in string units) the well known
result [25]
$$T^{(2)}=\frac{T_{D0}}{2\pi}.$$
(3.16)
Now we will construct the bound states for all roots in the $A_{n-1}$
matrix models. First, as the $k_{i}=1$ for the $A_{n-1}$, the
$r_{i}=0~{}\text{or}~{}1$. For a state to be a root and non-trivial in
homology, all of the $r_{i}=1$ must be adjacent and there must be at
least one $r_{i}=0$.
We will begin with the simplifying assumption that the F-terms are set
to zero. Later, we will describe how to obtain solutions when both
F and D-terms are present.
In order to solve the F-term equations of the hypermultiplets, we set
all of the $a_{i}=0$ in the vector multiplets. To solve the F-term
equations of the vector multiplets, we make the ansatz
$$x_{i,i+1}y_{i+1,i}=y_{i+1,i}x_{i,i+1}=0.$$
(3.17)
Now we must minimize the D-terms. The states we are interested in
should be excited states under the decoupled $U(1)$, but should form
supersymmetric bound states of the internal part of the
theory. Therefore, we will set $D_{i,\text{int.}}=0$ and seek a
solution to the equations
$$\bar{x}_{i-1,i}x_{i-1,i}-y_{i,i-1}\bar{y}_{i,i-1}-x_{i,i+1}\bar{x}_{i,i+1}+%
\bar{y}_{i+1,i}y_{i+1,i}=\left(2d_{i}+\frac{g^{2}}{g_{i}^{2}}D_{\text{dec.}}%
\right)\,\Bid_{N_{i}}.$$
(3.18)
With our ansatz, the operators $\bar{x}_{i-1,i}x_{i-1,i}$ and
$y_{i,i-1}\bar{y}_{i,i-1}$ commute. As they are self-adjoint, they
can be simultaneously diagonalized, and they moreover multiply to
zero. Therefore they are fully determined as the projections to the
positive and negative spectra of the operator
$$O_{i}=\bar{x}_{i-1,i}x_{i-1,i}-y_{i,i-1}\bar{y}_{i,i-1}$$
(3.19)
Now, $O_{i}$ and the operator
$$O^{\prime}_{i}=x_{i-1,i}\bar{x}_{i-1,i}-\bar{y}_{i,i-1}y_{i,i-1}$$
(3.20)
which appears in the D-term at the $(i-1)^{th}$ vertex, are
isospectral, except for an extra zero eigenvalue on one of them
whenever $r_{i}=r_{i-1}\pm 1$, so that $N_{i}=N_{i-1}\pm 1$. Imposing an
ordering on the eigenvalues, $O_{i}=\text{diag}(O_{i}^{1},\ldots,O^{N_{i}}_{i})$,
with $O_{i}^{1}\geq\cdots\geq O^{N_{i}}_{i}$, the D-term
equations (3.18) become
$$O_{i}^{\ell}-O_{i+1}^{\prime\ell}=\left(2d_{i}+\frac{g^{2}}{g_{i}^{2}}D_{\text%
{dec.}}\right)$$
(3.21)
Now, if the state is nontrivial, then, without loss of generality, we
can label the first vertex where there is a membrane wrapped as the
$0^{th}$, so that on the quiver diagram we will have the sequence
$\ldots,r_{n-1}=0,r_{0}=1,r_{1}=1,\ldots,r_{\iota}=1,r_{\iota+1}=0,\ldots$
(i.e., there are $\iota$ 1s and $n-\iota$ 0s on the diagram). Then
$O_{0}^{N+1}=0$, so that (3.21) can be solved to find
$O_{1}^{\prime N+1}=-(2d_{0}+\frac{g^{2}}{g_{0}^{2}}D_{\text{dec.}})$. But then
$O_{1}^{N+1}=-(2d_{0}+\frac{g^{2}}{g_{0}^{2}}D_{\text{dec.}})$ as well. By
repeated use of (3.21), we can follow these
eigenvalues until we get to the transition $r_{\iota}=1\rightarrow r_{\iota+1}=0$, where we jump to the next non-zero eigenvalue, so that
$$O_{\iota+1}^{N}=O_{\iota+1}^{\prime N+1}=-\sum_{j=1}^{\iota+1}\left(2d_{j-1}+%
\frac{g^{2}}{g_{j-1}^{2}}D_{\text{dec.}}\right).$$
(3.22)
By iterating this process around to the $0^{th}$ vertex, we
find
$$\begin{split}\displaystyle O_{0}^{N}&\displaystyle=-\sum_{j=1}^{n}\left(2d_{j-%
1}+\frac{g^{2}}{g_{j-1}^{2}}D_{\text{dec.}}\right)=-{D_{\text{dec.}}}\\
\displaystyle O_{1}^{N}&\displaystyle=-2d_{0}-\left(1+\frac{g^{2}}{g_{0}^{2}}%
\right)D_{\text{dec.}}.\end{split}$$
(3.23)
By continuing this process of circuiting the diagram, we determine all
of the eigenvalues. We find the general expression
$$O^{\ell}_{k}=-\sum_{j=1}^{k}\left(2d_{j-1}+\frac{g^{2}}{g_{j-1}^{2}}D_{\text{%
dec.}}\right)-\begin{cases}(N+1-\ell)D_{\text{dec.}}&\text{for}~{}0\leq k\leq%
\iota\\
(N-\ell)D_{\text{dec.}}&\text{for}~{}\iota<k<n.\end{cases}$$
(3.24)
The explicit form for $D_{\text{dec.}}$ is given by the
solution (3.7). For $n=1$, this formula is in agreement
with the $A_{1}$ solution presented in [16].
Now let us consider the situation when both F and D-terms are
present. Here, in addition to solving the D-terms (3.18), we
must also solve the F-term equations, which are of the form
$$y_{i,i-1}x_{i-1,i}-x_{i,i+1}y_{i+1,i}=\left(2f_{i}+\frac{g^{2}}{g_{i}^{2}}F_{%
\text{dec.}}\right)\,\Bid_{N_{i}}.$$
(3.25)
The simplest ansatz for a solution would clearly be one for which the
above method of tracing eigenvalues around the quiver diagram would
work. This will require that the individual terms in (3.18)
and (3.25) commute amongst themselves, so that they are
simultaneously diagonalizable. We must therefore demand that
$$\begin{split}\displaystyle\bar{y}_{i,i-1}y_{i,i-1}x_{i-1,i}&\displaystyle=x_{i%
-1,i}y_{i,i-1}\bar{y}_{i,i-1}\\
\displaystyle y_{i,i-1}x_{i-1,i}\bar{x}_{i-1,i}&\displaystyle=\bar{x}_{i-1,i}x%
_{i-1,i}y_{i,i-1}.\end{split}$$
(3.26)
If we then define
$$\begin{split}\displaystyle P_{i}&\displaystyle=y_{i,i-1}x_{i-1,i}\\
\displaystyle P_{i}^{\prime}&\displaystyle=x_{i-1,i}y_{i,i-1},\end{split}$$
(3.27)
then it is a straightforward exercise to show that
$$[O_{i},P_{i}]=[P_{i},\bar{P}_{i}]=0.$$
(3.28)
From (3.26), we see that $x_{i-1,i}$ generates a sort of
“supersymmetry” for the matrices $\bar{y}_{i,i-1}y_{i,i-1}$ and
$y_{i,i-1}\bar{y}_{i,i-1}$. That is, it acts as an intertwiner between
their eigenspaces, so that given an eigenstate, $|\alpha\rangle$, of
$y_{i,i-1}\bar{y}_{i,i-1}$, the state $x_{i}|\alpha\rangle$ is an
eigenstate of $\bar{y}_{i,i-1}y_{i,i-1}$ with the same eigenvalue.
Similarly,
$y_{i,i-1}$ is a supersymmetry for $x_{i-1,i}\bar{x}_{i-1,i}$ and
$\bar{x}_{i-1,i}x_{i-1,i}$, while both $x_{i-1,i}$ and $y_{i,i-1}$
are supersymmetries for $P_{i}$ and $P_{i}^{\prime}$. We can therefore
conclude that, as before, $O_{i}$ and
$O_{i}^{\prime}$ have the same spectrum, except, perhaps, for a
zero-mode. Similarly, the spectra of $P_{i}$ and $P_{i}^{\prime}$ can differ
only by a zero-mode.
Solutions may be explicitly obtained by solving the
eigenvalue equations (3.21) and
$$P_{i}^{\ell}-P_{i+1}^{\prime\ell}=\left(2f_{i}+\frac{g^{2}}{g_{i}^{2}}F_{\text%
{dec.}}\right),$$
(3.29)
via following the eigenvalues around the quiver diagram, as outlined
above. Then we find that the $O_{i}$ are given by (3.24),
as before, while the
$$P^{\ell}_{k}=-\sum_{j=1}^{k}\left(2f_{j-1}+\frac{g^{2}}{g_{j-1}^{2}}F_{\text{%
dec.}}\right)-\begin{cases}(N+1-\ell)F_{\text{dec.}}&\text{for}~{}0\leq k\leq%
\iota\\
(N-\ell)F_{\text{dec.}}&\text{for}~{}\iota<k<n,\end{cases}$$
(3.30)
where $F_{\text{dec.}}$ is given by (3.10).
We note that if the F-terms are in fact absent, we recover our old
ansatz (3.17).
Within the ansatz (3.26) the solutions obtained
from (3.24) and (3.30) are
unique. Explicit expressions for the $x_{i,i+1}$ and $y_{i+1,i}$ can
be obtained by choosing, say, the $x_{i,i+1}$ to have positive real
entries. This can
always be achieved by a gauge transformation. The solution obtained in
this manner is a classical solution to the equations for a
supersymmetric vacuum (of the internal theory), and therefore there is
a corresponding quantum mechanical state whose wavefunction is
localized near the classical solution. It is clear that all of the
hypermultiplets have a mass gap and it is reasonable to believe that
the solution breaks the gauge group completely.
4 Noncommutative Geometry and Spherical Membranes
It is worthwhile to illustrate the
geometric nature of the membrane solutions we have constructed. We
will show in this section that the natural set of gauge-invariant
coordinates derived from the quiver theory are elevated by these
solutions to noncommuting coordinates111For a recent discussion
of noncommutative geometry and matrix theory,
see [26] and references therein.. These
noncommuting coordinates
will satisfy the constraints imposed by the ALE space geometry to
leading order in the $1/N$ expansion.
In the case of the $A_{n-1}$, we know that $\BZ_{n}$-invariant
coordinates are given by invariant products of the coordinates of the
$\BC^{n}$ being quotiented. In terms of the quiver theory describing a
single membrane wrapped on the $0^{th}$ $\BP^{1}$, we can
consider the $N_{0}\times N_{0}$ matrices
$$\begin{split}\displaystyle U&\displaystyle=x_{01}\cdots x_{n-1,0}\\
\displaystyle V&\displaystyle=y_{0,n-1}\dots y_{10}\\
\displaystyle W&\displaystyle=x_{01}y_{10}.\end{split}$$
(4.1)
Coordinates $(u,v,w)$ on an ALE space of the same shape will satisfy
$$uv=P(w),$$
(4.2)
where $P(w)$ is an $n^{th}$-order polynomial whose coefficients are
determined by the $f_{i}$.
Now, when the $f_{i}=0$ (e.g. on the blow-down), $P(w)=w^{n}$. Also,
from (3.30) we see that
$W=0$. Since (3.24) tells us that $O_{0}$ is
positive-definite (for $\sum_{i}r_{i}d_{i}>0$, $D_{\text{dec.}}$ is
negative-definite), we know that
$\bar{y}_{01}y_{01}=0$, whence $y_{01}=V=0$. On the other hand,
$O_{1}$ is negative-definite, so that $x_{01}\bar{x}_{01}=0$ and
$x_{01}=U=0$ as well. Therefore the membrane is indeed localized at
the singularity, as $u=v=w=0$ is the singular locus of (4.2).
By means of the F-term equations (3.25), one can show that
for large $N$
$$UV=P(W)+O(1/N).$$
(4.3)
Moreover, one can also show that
$$\begin{split}\displaystyle[W,U]&\displaystyle=F_{\text{dec.}}U=-\frac{2\sum_{i%
}d_{i}f_{i}}{N}U+O(1/N^{2})\\
\displaystyle[W,V]&\displaystyle=-F_{\text{dec.}}V=\frac{2\sum_{i}d_{i}f_{i}}{%
N}V+O(1/N^{2}).\end{split}$$
(4.4)
This is reminiscent of the angular momentum
commutation relations, and shows that the intrinsic geometry of the
wrapped membranes that we have constructed is noncommutative. The
membranes are spherical, but if we were to probe the membrane locally,
where it approaches the flat membrane
solution, we would find an effective “Planck constant” of order
$1/N$ times the area of the wrapped membrane. This comes as no
surprise when we consider the way membranes were first constructed in
[17, 1], as well as the discussion
of spherical membranes in matrix theory by Kabat and
Taylor [18].
As an illustration of the noncommutative properties of the solutions,
it is illuminating to check that single membranes have properties
which are analogous to those discussed
in [18]. In the $A_{1}$ case, we can set the F-terms
to zero and use (3.24) to solve for the hypermultiplets.
Then $X_{01}=Y_{01}=0$, while the other hypermultiplet components are
off-diagonal (up to a $U(N)\times U(N+1)$ gauge transformation)
$$\begin{split}\displaystyle X_{10}&\displaystyle=\sqrt{\frac{4d}{2N+1}}\begin{%
pmatrix}\sqrt{N}&0&\cdots&\cdots&0\\
0&\sqrt{N-1}&0&\cdots&\vdots\\
\vdots&0&\ddots&0&\vdots\\
0&\cdots&0&1&0\end{pmatrix}\\
\displaystyle Y_{10}&\displaystyle=\sqrt{\frac{4d}{2N+1}}\begin{pmatrix}0&1&0&%
\cdots&0\\
\vdots&0&\ddots&0&\vdots\\
\vdots&\cdots&0&\sqrt{N-1}&0\\
0&\cdots&\cdots&0&\sqrt{N}\end{pmatrix},\end{split}$$
(4.5)
where we have set $g_{0}=g_{1}$ for convenience. As coordinates on the ALE
space are gauge-invariant combinations of the quiver theory
hypermultiplet components, we may consider the quadratic combinations
$$\begin{split}\displaystyle J_{0}&\displaystyle=\frac{1}{8\sqrt{d}}\bigl{(}X_{1%
0}\bar{X}_{10}-Y_{10}\bar{Y}_{10}\bigr{)}\\
&\displaystyle=\frac{\sqrt{d}}{2N+1}\begin{pmatrix}N-1&0&\cdots&\cdots&0\\
0&N-3&0&\cdots&\vdots\\
\vdots&0&\ddots&0&\vdots\\
\vdots&\cdots&0&-(N-3)&0\\
0&\cdots&\cdots&0&-(N-1)\end{pmatrix}\end{split}$$
(4.6)
$$J_{+}=\frac{1}{4\sqrt{d}}Y_{10}\bar{X}_{10}=\frac{2\sqrt{d}}{2N+1}\begin{%
pmatrix}0&\sqrt{N-1}&0&\cdots&0\\
\vdots&0&\sqrt{2(N-2)}&0&\vdots\\
\vdots&\cdots&0&\ddots&0\\
\vdots&\cdots&\cdots&0&\sqrt{N-1}\\
0&\cdots&\cdots&\cdots&0\\
\end{pmatrix}$$
(4.7)
$$J_{-}=\frac{1}{4\sqrt{d}}X_{10}\bar{Y}_{10}=\frac{2\sqrt{d}}{2N+1}\begin{%
pmatrix}0&\cdots&\cdots&\cdots&0\\
\sqrt{N-1}&0&\cdots&\cdots&\vdots\\
0&\sqrt{2(N-2)}&0&\cdots&\vdots\\
\vdots&0&\ddots&0&\vdots\\
0&\cdots&0&\sqrt{N-1}&0\\
\end{pmatrix}.$$
(4.8)
These matrices are proportional to the generators of
the N of $SU(2)$, so they commute according to the
$SU(2)$ algebra
$$\begin{split}&\displaystyle[J_{+},J_{-}]=2\left(\frac{2\sqrt{d}}{2N+1}\right)J%
_{0}\\
&\displaystyle[J_{\pm},J_{0}]=\mp\left(\frac{2\sqrt{d}}{2N+1}\right)J_{\pm}%
\end{split}$$
(4.9)
and the sum over their squares is
$$\sum_{i}J_{i}^{2}=J_{0}^{2}+\frac{1}{2}\{J_{+},J_{-}\}=\frac{4(N^{2}-1)d}{(2N+%
1)^{2}}\Bid_{N}=r^{2}\Bid_{N}.$$
(4.10)
Evidently, (4.6)–(4.8) are the redundant set
of coordinates
which parameterize the surface of the membrane. We note that the set
of operators $\tilde{J}_{+}=\bar{X}_{10}Y_{10}/4\sqrt{d},\ldots$
also form a set of membrane coordinates, this time in
the $\mathbf{N+1}$ representation. That there is a pair of good
membrane coordinates can be useful in calculations.
As in [18], we can ask if there is a regime in which
these membranes approach the flat membrane solution. At large $N$,
$$[J_{+},J_{-}]=\frac{d}{N}\begin{pmatrix}\Bid&0\\
0&-\Bid\end{pmatrix}+O(1/N^{2}),$$
(4.11)
which, when restricted to the upper-left quadrant, resembles the
commutation relations of the coordinates of the flat membrane.
In the semi-classical correspondence between Poisson brackets and
commutators of [17, 1], we find
the “Planck constant”
$$\hbar=\frac{d}{N},$$
(4.12)
which, as promised above, is proportional to the area of the membrane.
In the large $N$ limit, from (4.10), we also find that
$r$ is the radius of the membrane,
$$r=\sqrt{d}\left(1-\frac{1}{2N}\right).$$
(4.13)
Our choice of normalization in (4.6)–(4.8)
was made specifically
to obtain this leading-order behavior. These results agree nicely with
those of [18], once the different $N$ dependence of
the membrane longitudinal momentum here is taken into account. We note
that the corrections to the radius (4.13) are stronger than
those found for the membranes in [18], being of
$O(1/N)$ as compared to $O(1/N^{2})$.
5 The Spectrum of Excitations of the Wrapped Membrane
We will now concentrate our efforts on finding the spectrum of
excitations around the membrane solution for the simplest possible
case, namely the $A_{1}$ singularity. We will first examine a toy model,
in order to establish a framework for understanding the problem. Then
we utilize the rotational invariance of the $A_{1}$ solution to
determine the allowed representations for perturbations around the
membrane. We then calculate the mass spectrum for the P-wave
excitations and, finally, for all modes.
We note that the perturbative approach we take to the calculation of
the spectrum of fluctuations is well-suited only for membranes which
are large as compared to the 11-dimensional Planck length. It is only
in this limit that we may reliably expect that the fluctuations have
energies small compared to the mass of the membrane and are thereby
long-lived.
We want to describe the linearized spectrum of excitations of a large
membrane wrapped around a sphere. Consider a toy model consisting of a
massless free scalar field on the space $S^{2}\times\BR$, where $\BR$
is time and the sphere is of radius $r$. It is a straightforward
exercise to show that the normal modes are the spherical harmonics,
with masses given by
$$M_{\ell}=\frac{\sqrt{\ell(\ell+1)}}{r}$$
(5.1)
for $\ell=0,1,$ etc. This toy model reproduces the linearized degrees
of freedom for a BPS membrane wrapped on a 2-sphere, if we give the
action maximal supersymmetry. Instead of a single scalar, we can
consider $U(1)$ SYM with maximal supersymmetry as our starting point.
This is the natural effective worldvolume theory of a D2-brane (see,
e.g., [27, 28]) and,
in particular, is valid for the large radius limit we are considering.
As we are considering a free $U(1)$ theory, there is no need to
quantize the magnetic flux through the sphere. This quantum number is
associated to the compactness of the 11th dimension. In order to get
an explicitly $O(6)$-invariant Lagrangian, as one would expect from
M-Theory, one can dualize the vector into a scalar.
Since half of the supersymmetries are broken in the ALE space, we must also
explicitly break half of the supersymmetry of our toy model. This can
be achieved by twisting one complex scalar. In particular, we identify
one of the scalars as a section of the normal bundle of the
2-sphere sitting in the ALE space. The resulting theory is
supersymmetric if the fermions are sections of the bundle
$\CO(-1)\oplus\CO(1)$. There are six bosonic zero modes, which
correspond to the motion of the membrane in the transverse $\BR^{6}$.
The corresponding eight fermionic zero modes generate the required
16-fold degeneracy of the ground state. For the massive modes, there
are eight bosonic and eight fermionic modes for each spherical
harmonic. This gives us the
information we require in order to interpret the calculation we will
now make in the matrix model.
For the $A_{1}$ singularity there is an $SU(2)$
rotational invariance which aids in the solution. Namely, since there
is a pair of hypermultiplets connecting
the same two vertices, the action (and so too the D and F-terms) is
invariant under the $SU(2)_{r}$ transformation
$$\begin{pmatrix}x_{01}\\
y_{01}\end{pmatrix}\rightarrow\begin{pmatrix}x_{01}^{\prime}\\
y_{01}^{\prime}\end{pmatrix}=R\begin{pmatrix}x_{01}\\
y_{01}\end{pmatrix},~{}~{}~{}\begin{pmatrix}x_{10}\\
y_{10}\end{pmatrix}\rightarrow\begin{pmatrix}x_{10}^{\prime}\\
y_{10}^{\prime}\end{pmatrix}=R\begin{pmatrix}x_{10}\\
y_{10}\end{pmatrix},$$
(5.2)
for $R\in SU(2)_{r}$, which mixes the hypermultiplets.
Clearly there is a $U(N)\times U(N+1)$ gauge
transformation,
$$\begin{pmatrix}U&0\\
0&U\end{pmatrix}R\begin{pmatrix}X_{10}\\
Y_{10}\end{pmatrix}V=\begin{pmatrix}X_{10}\\
Y_{10}\end{pmatrix},$$
(5.3)
where $U\in U(N)$ and $V\in U(N+1)$, that undoes the action
of (5.2) on the membrane ground
state (4.5). In other words,
the solution (4.5) is invariant under the product
$U\otimes R\otimes V$, which is a consequence of the fact that the
ground state is spherically symmetric. Obviously $U$ and $V$ must
themselves correspond to (inverse) elements of $SU(2)_{r}$.
Now the action of $U\otimes R\otimes V$ on any perturbations
around (4.5) should be faithful. Since $R$ acts as
the 2 of $SU(2)_{r}$ on
$\Bigl{(}\begin{smallmatrix}\delta x_{10}\\
\delta y_{10}\end{smallmatrix}\Bigr{)}$
and $U$ and $V$ act as the N and N+1,
respectively, $U\otimes R\otimes V$ must act as the representation
$$\textbf{2}\otimes\textbf{N}\otimes(\textbf{N+1})=\textbf{1}\oplus\textbf{3}%
\oplus\textbf{3}\oplus\textbf{5}\oplus\textbf{5}\oplus\cdots\oplus\textbf{2N--%
1}\oplus\textbf{2N--1}\oplus\textbf{2N+1}.$$
(5.4)
On the other hand, under this gauge transformation,
the vector multiplets transform in the
$$\begin{split}\displaystyle\textbf{N}\otimes\textbf{N}&\displaystyle=\textbf{1}%
\oplus\textbf{3}\oplus\cdots\oplus\textbf{2N--1}\\
\displaystyle\textbf{(N+1)}\otimes(\textbf{N+1})&\displaystyle=\textbf{1}%
\oplus\textbf{3}\oplus\cdots\oplus\textbf{2N+1}\end{split}$$
(5.5)
representations.
One immediately notices that the dimension of the representation of
the vectors (5.5) is greater by a singlet than that
of the hypermultiplets in (5.4). This singlet is
simply the decoupled $U(1)$. The absence of a singlet for the
decoupled $U(1)$ in the hypermultiplet representation protects against
the possibility of Higgsing away this element of the gauge
symmetry. Another important observation
about (5.4) and (5.5) is that
only even spherical harmonics will appear in the multipole
expansion of the perturbations. This is a considerable simplification,
and we will be able to compute the full spectrum of excitations of the
wrapped membrane.
As a start, let us now calculate the mass spectrum for the P-wave
excitations, i.e., those in the 3 of $SU(2)_{r}$. The
calculation will be illustrative of the techniques we will use to
compute the spectrum for generic $j$.
Since the masses are generated via the Higgs mechanism, it will be
sufficient to calculate the masses of the vector multiplets, since
these are related to those
of the hypers in an obvious way. Furthermore, we can exploit the
$SO(5)$ invariance of the scalars in the vector multiplet, so that we
have $a_{0i}=a^{\alpha}_{0}\lambda^{(N+1)}_{\alpha}$,
$a_{1i}=a^{\alpha}_{1}\lambda^{(N)}_{\alpha}$, for each $i$, where
the $\lambda^{(N)}_{\alpha}$ generate the N
of $SU(2)_{r}$. Finally, we note from the $SU(2)_{r}$ invariance that we
can take $a^{\alpha}_{0}=a_{0}$, $a^{\alpha}_{1}=a_{1}$, for each $\alpha$. We
then can compute the mass of $a_{0}$ and $a_{1}$ for the simplest case,
namely the Cartan generator, $\lambda_{0}$.
The quadratic part of the Lagrangian for $a_{0}$ and $a_{1}$ is
$$\begin{split}\displaystyle\CL=&\displaystyle\frac{1}{2g_{0}^{2}}\,\text{Tr}%
\bigl{(}\lambda^{(N+1)}_{0}\bigr{)}^{2}\dot{a}_{0}^{2}+\frac{1}{2g_{1}^{2}}\,%
\text{Tr}\bigl{(}\lambda^{(N)}_{0}\bigr{)}^{2}\dot{a}_{1}^{2}\\
&\displaystyle-\frac{1}{g^{2}}\left(\text{Tr}\left[\bigl{(}\lambda^{(N+1)}_{0}%
\bigr{)}^{2}(\bar{X}_{10}X_{10}+\bar{Y}_{10}Y_{10})\right]a_{0}^{2}+\text{Tr}%
\left[\bigl{(}\lambda^{(N)}_{0}\bigr{)}^{2}(X_{10}\bar{X}_{10}+Y_{10}\bar{Y}_{%
10})\right]a_{1}^{2}\right.\\
&\displaystyle\phantom{+\frac{1}{g^{2}}}~{}\left.+2\,\text{Tr}\left[\lambda^{(%
N)}_{0}X_{10}\lambda^{(N+1)}_{0}\bar{X}_{10}+\lambda^{(N)}_{0}Y_{10}\lambda^{(%
N+1)}_{0}\bar{Y}_{10}\right]a_{0}a_{1}\right).\end{split}$$
(5.6)
From (4.5), if we generalize to arbitrary coupling
constants at each vertex, we have
$$\begin{split}\displaystyle X_{10}\bar{X}_{10}+Y_{10}\bar{Y}_{10}&\displaystyle%
=(N+1)|D_{\text{dec.}}|\,\Bid_{N}\\
\displaystyle\bar{X}_{10}X_{10}+\bar{Y}_{10}Y_{10}&\displaystyle=N|D_{\text{%
dec.}}|\,\Bid_{N+1}\\
\displaystyle X_{10}\bar{X}_{10}-Y_{10}\bar{Y}_{10}&\displaystyle=|D_{\text{%
dec.}}|\,\lambda^{(N)}_{0}\\
\displaystyle\bar{X}_{10}X_{10}-\bar{Y}_{10}Y_{10}&\displaystyle=|D_{\text{dec%
.}}|\,\lambda^{(N+1)}_{0}.\end{split}$$
(5.7)
To simplify the cross-terms
in (5.6), we use the “commutation” relations
$$\begin{split}\displaystyle\lambda^{(N)}_{0}X_{10}-X_{10}\lambda^{(N+1)}_{0}+X_%
{10}&\displaystyle=0\\
\displaystyle\lambda^{(N)}_{0}Y_{10}-Y_{10}\lambda^{(N+1)}_{0}-Y_{10}&%
\displaystyle=0.\end{split}$$
(5.8)
These relations encode, in part, the invariance of the membrane
solution (4.5) under the rotation group of the
sphere (5.2).
One may also readily calculate
$$\text{Tr}\bigl{(}\lambda^{(N)}_{0}\bigr{)}^{2}=\frac{(N-1)N(N+1)}{3}.$$
(5.9)
Using these relations in (5.6) yields
$$\CL=\frac{1}{2}\bigl{(}\dot{\tilde{a}_{0}}^{2}+\dot{\tilde{a}_{1}}^{2}\bigr{)}%
-\frac{|D_{\text{dec.}}|}{g^{2}}\left[Ng_{0}^{2}\tilde{a}_{0}^{2}+2\sqrt{(N-1)%
(N+2)}\,g_{0}g_{1}\tilde{a}_{0}\tilde{a}_{1}+(N+1)g_{1}^{2}\tilde{a}_{1}^{2}%
\right],$$
(5.10)
where we have defined the normalized variables
$$\begin{split}\displaystyle\tilde{a}_{0}&\displaystyle=\sqrt{\text{Tr}\bigl{(}%
\lambda^{(N+1)}_{0}\bigr{)}^{2}}\,\frac{a_{0}}{g_{0}}\\
\displaystyle\tilde{a}_{1}&\displaystyle=\sqrt{\text{Tr}\bigl{(}\lambda^{(N)}_%
{0}\bigr{)}^{2}}\,\frac{a_{1}}{g_{1}}.\end{split}$$
(5.11)
The eigenvalues of the mass matrix are the energies for the oscillator modes
$$\begin{split}\displaystyle\omega_{\pm}^{2}&\displaystyle=\left[\left(Ng_{0}^{2%
}+(N+1)g_{1}^{2}\right)\pm\sqrt{\left(Ng_{0}^{2}+(N+1)g_{1}^{2}\right)^{2}-8g_%
{0}^{2}g_{1}^{2}}\right]\frac{|D_{\text{dec.}}|}{g^{2}}\\
&\displaystyle\sim\frac{2g_{0}^{2}g_{1}^{2}d}{g^{4}}\left[1\pm\left(1-\frac{4g%
^{4}}{g_{0}^{2}g_{1}^{2}}\,\frac{1}{N^{2}}+\cdots\right)\right].\end{split}$$
(5.12)
Therefore, in the large $N$ limit, we find two modes, one whose energy
is independent of $N$ at leading order (and therefore decouples) and
another whose energy is
$$\omega_{-}=\frac{2\sqrt{2d}}{N}$$
(5.13)
and survives as the membrane excitation as $N\rightarrow\infty$.
Notice that the energy of this mode only depends on $d$,
providing some evidence for the hypothesis made in
Section 3 that the ratio $g_{0}/g_{1}$
decouples in the large $N$ limit . The mass of this mode, $\delta m$,
can be calculated from
$$\omega_{-}=\frac{(m+\delta m)^{2}-m^{2}}{2p^{+}}\sim\frac{m\,\delta m}{p^{+}},$$
(5.14)
where $m$ and $p^{+}$ are given by (3.12)
and (3.13), respectively. In this way, we find
$$\delta m=\sqrt{\frac{2}{d}}=\frac{\sqrt{2}}{r},$$
(5.15)
where $r$ is the radius of the membrane. This result agrees precisely
with our toy model result (5.1) for the P-wave, $\ell=1$.
Now, for generic excitations appearing in the $\mathbf{2j+1}$
representation under the decomposition (5.4), we may
again exploit symmetry arguments. In general, these reduce the problem
to computing the masses of the scalars in the vector multiplets, which
may be taken to be of the form
$$\begin{split}\displaystyle a_{0i}&\displaystyle=a_{0}\bigl{(}\lambda^{(N+1)}_{%
+}\bigr{)}^{j}+\bar{a}_{0}\bigl{(}\lambda^{(N+1)}_{-}\bigr{)}^{j}\\
\displaystyle a_{1i}&\displaystyle=a_{1}\bigl{(}\lambda^{(N)}_{+}\bigr{)}^{j}+%
\bar{a}_{1}\bigl{(}\lambda^{(N)}_{-}\bigr{)}^{j},\end{split}$$
(5.16)
for each $i=1,\ldots,5$. As a motivation for why we have chosen this
particular form, we note that these are the highest weight states for
their respective $SU(2)_{r}$ representations. This stems from the fact
that, as the representations are irreducible, all of the operators
acting in a given representation are constructed from polynomials in
the generators.
After dropping the terms which vanish after
taking traces, the quadratic Lagrangian for these modes is
$$\begin{split}\displaystyle\CL=&\displaystyle 2\,\text{Tr}\left[\bigl{(}\lambda%
^{(N+1)}_{+}\bigr{)}^{j}\bigl{(}\lambda^{(N+1)}_{-}\bigr{)}^{j}\right]\,\left(%
\frac{1}{g_{0}^{2}}|\dot{a}_{0}|^{2}-\frac{N|D_{\text{dec.}}|}{g^{2}}|a_{0}|^{%
2}\right)\\
&\displaystyle+2\,\text{Tr}\left[\bigl{(}\lambda^{(N)}_{+}\bigr{)}^{j}\bigl{(}%
\lambda^{(N)}_{-}\bigr{)}^{j}\right]\,\left(\frac{1}{g_{1}^{2}}|\dot{a}_{1}|^{%
2}-\frac{(N+1)|D_{\text{dec.}}|}{g^{2}}|a_{1}|^{2}\right)\\
&\displaystyle-\frac{2}{g^{2}}a_{0}\bar{a}_{1}\left[\bigl{(}\lambda^{(N)}_{+}%
\bigr{)}^{j}X_{10}\bigl{(}\lambda^{(N+1)}_{-}\bigr{)}^{j}\bar{X}_{10}+\bigl{(}%
\lambda^{(N)}_{+}\bigr{)}^{j}Y_{10}\bigl{(}\lambda^{(N+1)}_{-}\bigr{)}^{j}\bar%
{Y}_{10}\right]\\
&\displaystyle-\frac{2}{g^{2}}\bar{a}_{0}a_{1}\left[\bigl{(}\lambda^{(N)}_{-}%
\bigr{)}^{j}X_{10}\bigl{(}\lambda^{(N+1)}_{+}\bigr{)}^{j}\bar{X}_{10}+\bigl{(}%
\lambda^{(N)}_{-}\bigr{)}^{j}Y_{10}\bigl{(}\lambda^{(N+1)}_{+}\bigr{)}^{j}\bar%
{Y}_{10}\right].\end{split}$$
(5.17)
The cross-terms here can be simplified through the use of the
“commutation” relations
$$\begin{split}\displaystyle X_{10}\lambda^{(N+1)}_{+}-\lambda^{(N)}_{+}X_{10}-Y%
_{10}&\displaystyle=0\\
\displaystyle X_{10}\lambda^{(N+1)}_{-}-\lambda^{(N)}_{-}X_{10}&\displaystyle=%
0\\
\displaystyle Y_{10}\lambda^{(N+1)}_{+}-\lambda^{(N)}_{+}Y_{10}&\displaystyle=%
0\\
\displaystyle Y_{10}\lambda^{(N+1)}_{-}-\lambda^{(N)}_{-}Y_{10}-X_{10}&%
\displaystyle=0.\end{split}$$
(5.18)
As in the case of the relations (5.8), these encode the
rotational invariance of the membrane solution. By repetitive
application of (5.18), one may establish the useful relationship
$$(N-j)\,\text{Tr}\left[\bigl{(}\lambda^{(N+1)}_{+}\bigr{)}^{j}\bigl{(}\lambda^{%
(N+1)}_{-}\bigr{)}^{j}\right]=(N+j+1)\,\text{Tr}\left[\bigl{(}\lambda^{(N)}_{+%
}\bigr{)}^{j}\bigl{(}\lambda^{(N)}_{-}\bigr{)}^{j}\right].$$
(5.19)
In terms of the normalized variables
$$\begin{split}\displaystyle\tilde{a}_{0}&\displaystyle=\sqrt{2\,\text{Tr}\left[%
\bigl{(}\lambda^{(N+1)}_{+}\bigr{)}^{j}\bigl{(}\lambda^{(N+1)}_{-}\bigr{)}^{j}%
\right]}\,\frac{a_{0}}{g_{0}}\\
\displaystyle\tilde{a}_{1}&\displaystyle=\sqrt{2\,\text{Tr}\left[\bigl{(}%
\lambda^{(N)}_{+}\bigr{)}^{j}\bigl{(}\lambda^{(N)}_{-}\bigr{)}^{j}\right]}\,%
\frac{a_{1}}{g_{1}},\end{split}$$
(5.20)
the Lagrangian (5.17) becomes
$$\begin{split}\displaystyle\CL=&\displaystyle|\dot{\tilde{a}_{0}}|^{2}+|\dot{%
\tilde{a}_{1}}|^{2}-\frac{|D_{\text{dec.}}|}{g^{2}}\left(Ng_{0}^{2}|\tilde{a}_%
{0}|^{2}+(N+1)g_{1}^{2}|\tilde{a}_{1}|^{2}\right)\\
&\displaystyle-\frac{|D_{\text{dec.}}|}{g^{2}}\sqrt{(N-j)(N+j+1)}\,g_{0}g_{1}%
\left(\tilde{a}_{0}\bar{\tilde{a}_{1}}+\bar{\tilde{a}_{0}}\tilde{a}_{1}\right)%
,\end{split}$$
(5.21)
From this, we find the energies of the modes as the eigenvalues of
the mass matrix
$$\begin{split}\displaystyle\omega_{\pm}^{2}&\displaystyle=\left[\left(Ng_{0}^{2%
}+(N+1)g_{1}^{2}\right)\pm\sqrt{\left(Ng_{0}^{2}+(N+1)g_{1}^{2}\right)^{2}-4j(%
j+1)g_{0}^{2}g_{1}^{2}}\right]\frac{|D_{\text{dec.}}|}{g^{2}}\\
&\displaystyle\sim\frac{2g_{0}^{2}g_{1}^{2}d}{g^{4}}\left[1\pm\left(1-\frac{2j%
(j+1)g^{4}}{g_{0}^{2}g_{1}^{2}}\,\frac{1}{N^{2}}+\cdots\right)\right].\end{split}$$
(5.22)
We note that, in the case $j=1$, this reduces properly
to (5.12).
As in the case of the P-waves above, we can consider the behavior
at large $N$. There are again two modes, one which decouples and
another with energy
$$\omega_{-}=\frac{\sqrt{2j(j+1)d}}{N}$$
(5.23)
and mass
$$\delta m=\sqrt{\frac{j(j+1)}{d}}=\frac{\sqrt{j(j+1)}}{r},$$
(5.24)
which agrees exactly with the toy model result (5.1). In the
limit of large radius, the effective field theory on the membrane and
the perturbation expansion in the matrix model agree.
6 Conclusions
We have given quite a bit of evidence that the matrix model proposed
in [19, 16] captures many of the
essential features of M-Theory membranes which are wrapped around the
homology 2-cycles of an ALE space. In [16], we
gave an explicit derivation of the solution describing the wrapped
membrane in the $A_{1}$ case. We found that its energy, as well as the
leading order membrane-antimembrane potential, had the properties
necessary for its interpretation as a wrapped membrane.
In Section 3, we gave a counting argument that set the
correspondence between BPS bound states of membranes and roots of the
$A$-$D$-$E$ group. We derived a general bound on the light-cone energy
of a membrane bound state and found that it had the necessary
dependence on the blow-up parameters and longitudinal momentum. We
also gave an explicit solution for all membrane bound states for the
$A_{n-1}$ series. The 16-fold degeneracy of these solutions is
consistent with the requirement that they form BPS vector multiplets
in seven-dimensional spacetime physics.
We were able to obtain all of the solutions that we expected from the
M-Theory interpretation of the model, but we have not been able to
show that we have given a unique solution. The existence of other
solutions would probably lead to a greater than 16-fold degeneracy
that would be hard to reconcile with the expected seven-dimensional
physics.
The solutions we found exhibit a very rich
structure, and fit naturally into a noncommutative geometry framework.
In section 4, we showed that there exist gauge-invariant
coordinates for the membrane in the $A_{1}$ model that satisfy the
relations appropriate to spherical membranes [18].
We then characterized the representations of excitations around the $A_{1}$
membrane in Section 5, finding that, as expected,
only even spherical harmonics contribute in the multipole
expansion. We then calculated the energy spectrum of the excitations
and found the exact agreement with the expected result from the
consideration of a large membrane, where we could trust both a
calculation in a toy model, as well as in matrix model perturbation theory.
The end result is very promising. The perturbation analysis gives the
correct results in the large radius limit, and some coupling constants
seem to decouple in the large $N$ limit. If this decoupling is a
generic property of the large $N$ dynamics, this opens up the possibility of
proving non-renormalization theorems.
Despite our success in solving for the membrane bound state
solutions for the $A_{n-1}$ series, we have not been able to
generalize them to the $D$ and $E$ series. This is basically due to
the fact that our method of following the eigenvalues around the
quiver diagram, while perfect for closed quivers, does not work for
the $D$ and $E$ open quivers. By inspection of a few cases where $N$
is taken very small, one may also see that not all of
the matrices in these cases can be diagonalized simultaneously. This
is bound to complicate finding membrane solutions.
However, one should obtain similar results for the membrane properties
in these cases. For example, the association of the BPS bound states with
roots of the $A$-$D$-$E$ algebra and the bound on the energy of a
bound state discussed in Section 3 are both completely
general, so they apply equally well to the $D$ and $E$ series. Also,
at large $N$, one expects that the couplings $g_{i}/g_{j}$
should still decouple from the low energy degrees of freedom (whose
energies scale as $1/N$), as the states that feel these couplings
should have energies of order 1. Of course, even given a solution, in
these cases the mass matrix is more complicated and should prove
difficult to diagonalize.
There are still more tests that the matrix models for the ALE spaces should
be put to. For example, it is particularly interesting to reproduce
the Coulomb
and velocity-dependent potentials between membranes and
gravitons [29, 30, 5, 18].
These and related matters are currently under investigation.
Acknowledgements
We would like to thank Philip Candelas, Dan Kabat, and especially
Willy Fischler for enlightening discussions about these and related
issues. We thank Jacques Distler for his collaboration at an early
stage of this work and many important conversations.
References
[1]
T. Banks, W. Fischler, S. H. Shenker, and L. Susskind, “M Theory as a
matrix model: A conjecture,” Phys. Rev. D55 (1997) 5112–5128,
hep-th/9610043.
[2]
T. Banks, “Matrix theory,” hep-th/9710231.
[3]
D. Bigatti and L. Susskind, “Review of matrix theory,”
hep-th/9712072.
[4]
W. Taylor, “Lectures on D-branes, gauge theory and M(atrices),”
hep-th/9801182.
[5]
D. Kabat and W. Taylor, “Linearized supergravity from matrix theory,”
hep-th/9712185.
[6]
M. R. Douglas, H. Ooguri, and S. H. Shenker, “Issues in M(atrix) theory
compactification,” Phys. Lett. B402 (1997) 36–42,
hep-th/9702203.
[7]
M. Dine and A. Rajaraman, “Multigraviton scattering in the matrix model,”
hep-th/9710174.
[8]
M. R. Douglas and H. Ooguri, “Why matrix theory is hard,”
hep-th/9710178.
[9]
K. Becker and M. Becker, “On graviton scattering amplitudes in M theory,”
hep-th/9712238.
[10]
S. Hellerman and J. Polchinski, “Compactification in the lightlike limit,”
hep-th/9711037.
[11]
M. R. Douglas, “Enhanced gauge symmetry in M(atrix) theory,” JHEP
Elec. J. 7 (1997), no. 4,
hep-th/9612126.
[12]
W. Fischler and A. Rajaraman, “M(atrix) string theory on K3,” Phys.
Lett. B411 (1997) 53–58,
hep-th/9704123.
[13]
M. R. Douglas, “D branes and matrix theory in curved space,” in Strings
’97 (Amsterdam, June 16-21, 1997).
hep-th/9707228.
[14]
D.-E. Diaconescu and J. Gomis, “Duality in matrix theory and three-dimensional
mirror symmetry,” hep-th/9707019.
[15]
D.-E. Diaconescu, M. R. Douglas, and J. Gomis, “Fractional branes and wrapped
branes,” hep-th/9712230.
[16]
D. Berenstein, R. Corrado, and J. Distler, “Aspects of ALE matrix models and
twisted matrix strings,” hep-th/9712049.
[17]
B. de Wit, J. Hoppe, and H. Nicolai, “On the quantum mechanics of
supermembranes,” Nucl. Phys. B305 (1988) 545.
[18]
D. Kabat and W. Taylor, “Spherical membranes in matrix theory,” Adv.
Theor. Math. Phys. 2 (1998), no. 1,
hep-th/9711078.
[19]
M. R. Douglas and G. Moore, “D-branes, quivers, and ALE instantons,”
hep-th/9603167.
[20]
J. Polchinski, “Tensors from K3 orientifolds,” Phys. Rev. D55
(1997) 6423–6428, hep-th/9606165.
[21]
C. V. Johnson and R. C. Myers, “Aspects of type IIB theory on ALE
spaces,” Phys. Rev. D55 (1997) 6382–6393,
hep-th/9610140.
[22]
N. J. Hitchin, A. Karlhede, U. Lindstrom, and M. Rocek, “Hyperkähler
metrics and supersymmetry,” Commun. Math. Phys. 108 (1987) 535.
[23]
P. B. Kronheimer, “The construction of ALE spaces as hyper-Kähler
quotients,” J. Diff. Geom. 19 (1989) 665–683.
[24]
L. Susskind, “Another conjecture about M(atrix) theory,”
hep-th/9704080.
[25]
M. B. Green, C. M. Hull, and P. K. Townsend, “D-brane Wess-Zumino actions,
T-duality and the cosmological constant,” Phys. Lett. B382
(1996) 65–72, hep-th/9604119.
[26]
A. Connes, M. R. Douglas, and A. Schwarz, “Noncommutative geometry and matrix
theory: Compactification on tori,”
hep-th/9711162.
[27]
J. Polchinski, S. Chaudhuri, and C. V. Johnson, “Notes on D-branes,”
hep-th/9602052.
[28]
J. Polchinski, “TASI Lectures on D-branes,”
hep-th/9611050.
[29]
O. Aharony and M. Berkooz, “Membrane dynamics in M(atrix) theory,” Nucl. Phys. B491 (1997) 184–200,
hep-th/9611215.
[30]
G. Lifschytz and S. D. Mathur, “Supersymmetry and membrane interactions in
M(atrix) theory,” Nucl. Phys. B507 (1997) 621–644,
hep-th/9612087. |
LC-PHSM-2004-006
March 2004
Study of Higgs Boson Pair Production
at Linear Collider
K. Desch${}^{a}$, T. Klimkovich ${}^{a,b}$, T. Kuhl${}^{b}$,
A. Raspereza${}^{b}$
${}^{a}$ Universität Hamburg, Institut für Experimentalphysik
Notkestrasse 85, D-22607 Hamburg, Germany.
${}^{b}$ DESY, Notkestrasse 85, D-22607 Hamburg, Germany.
We study the potential of the TESLA linear collider operated at
a center-of-mass energy of 500 to 1000${\mathrm{\ Ge\kern-1.0ptV}}$
for the measurement of the neutral Higgs boson
properties within the framework of the MSSM. The process of associated
Higgs boson production with subsequent decays of Higgs
bosons into b-quark and $\tau$-lepton pairs is considered.
An integrated luminosity of 500 fb${}^{-1}$ is assumed at each energy.
The Higgs boson masses and production cross sections
are measured by reconstructing the $\mathrm{b\bar{b}b\bar{b}}$ and $\mathrm{b\bar{b}\tau^{+}\tau^{-}}$ final states.
The precision of these measurements
is evaluated in dependence of the Higgs boson masses.
Under the assumed experimental conditions
a statistical accuracy ranging from 0.1 to 1.0 GeV is achievable
on the Higgs boson mass.
The topological cross section $\mathrm{\sigma(\mathrm{e^{+}e^{-}}\rightarrow HA\rightarrow b\bar{b}b\bar{b})}$ can be determined with
the relative precision of 1.5 – 6.6 % and cross sections $\mathrm{\sigma(\mathrm{e^{+}e^{-}}\rightarrow HA\rightarrow b\bar{b}\tau^{+}%
\tau^{-})}$ and $\mathrm{\sigma(\mathrm{e^{+}e^{-}}\rightarrow HA\rightarrow\tau^{+}\tau^{-}b%
\bar{b})}$ with precision of 4 – 30 %.
Constraints on the Higgs boson widths can be set exploiting
$\mathrm{b\bar{b}\tau^{+}\tau^{-}}$ channel. The 5$\sigma$ discovery limit corresponds to the Higgs mass of around 385 GeV for the degenerate Higgs boson masses in
the $\mathrm{HA\to b\bar{b}b\bar{b}}$ channel at $\sqrt{s}$ = 800 GeV with integrated luminosity
of 500 fb${}^{-1}$. The potential of the Higgs mass
determination for the benchmark point SPS 1a for the process $\mathrm{HA\to b\bar{b}b\bar{b}}$
at $\sqrt{s}$ = 1 TeV and luminosity 1000 fb${}^{-1}$ is investigated.
1 Introduction
Elucidating the mechanism of electroweak symmetry breaking will be one of the central
tasks at a future linear $\mathrm{e^{+}e^{-}}$ collider. In the minimal
Standard Model (SM) breaking of electroweak symmetry
is provided by one Higgs doublet, which introduces an
additional spin-0 particle, the Higgs boson [1].
Among models with an extended Higgs sector, two Higgs doublet
models (2HDM) [2] are of particular
interest, since this structure of the Higgs sector is required
in the Minimal Supersymmetric Standard Model (MSSM) [3], the
most seriously considered extension of the SM.
The Higgs sector of the CP-conserving MSSM comprises the
following physical states:
two CP-even Higgs bosons, the lighter of which is denoted h and
the heavier H, one CP-odd Higgs boson A and two charged
bosons H${}^{\pm}$. Two crucial parameters,
which influence the Higgs boson phenomenology, are $\tan\beta$, the ratio of vacuum expectation values
of the two Higgs doublets, and the angle $\alpha$, which
controls mixing in the CP-even sector.
The cross sections of the Higgs-strahlung
and fusion processes, involving CP-even Higgs bosons,
$$\mathrm{e^{+}e^{-}\to hZ(HZ),\hskip 19.916929pte^{+}e^{-}\to h(H)\nu_{e}\bar{%
\nu}_{e},\hskip 19.916929pte^{+}e^{-}\to h(H)e^{+}e^{-},}$$
(1)
are defined in terms of the strength of the Higgs couplings to
weak bosons at tree level. These depend on the parameters $\alpha$
and $\beta$ in the following way:
$$\mathrm{g_{hZZ,hWW}\sim\sin(\beta-\alpha),\hskip 56.905512ptg_{HZZ,HWW}\sim%
\cos(\beta-\alpha).}$$
(2)
In the MSSM, the set of the tree level couplings between
neutral Higgs particles and weak bosons is extended by
two additional couplings,
$$\mathrm{g_{hAZ}\sim\cos(\beta-\alpha),\hskip 56.905512ptg_{HAZ}\sim\sin(\beta-%
\alpha).}$$
(3)
As a consequence, in $\mathrm{e^{+}e^{-}}$ collisions, the Higgs-strahlung
and fusion processes will be complemented by the associated
Higgs boson pair production mechanisms:
$$\mathrm{e^{+}e^{-}\to hA(HA).}$$
(4)
To fully explore the MSSM Higgs sector, the processes (4)
must be studied, since
they directly probe $\mathrm{g_{hAZ}}$ and $\mathrm{g_{HAZ}}$ trilinear couplings,
provide access to the A boson properties
and supplement the study of the CP-even Higgs boson
properties via Higgs-strahlung and fusion mechanisms.
The study of the Higgs pair production process is motivated
by a so-called decoupling limit of the MSSM in which
the h boson approaches the properties of the SM Higgs
boson. The closer MSSM scenario moves towards the decoupling
limit the more difficult it becomes to distinguish the Higgs sector
from the SM. In such a scenario the detection of heavier neutral
Higgs bosons would be crucial for establishing an extended Higgs sector.
The decoupling limit is approached relatively fast for
large values of the H and A boson masses, $m_{\mathrm{A}},m_{\mathrm{H}}$
$>$ 200 GeV, in a large portion of the MSSM parameter space.
The distinct feature of this scenario is vanishing coupling of heavy CP-even
Higgs boson to weak bosons, $\cos(\beta-\alpha)\rightarrow 0$.
As a consequence, the H boson production via
the fusion and Higgs-strahlung processes is significantly
suppressed, whereas the cross section of the $\mathrm{e^{+}e^{-}\to HA}$ process
reaches its maximal value making associated heavy Higgs pair
production a promising channel for the detection of the H and A
bosons at a future linear $\mathrm{e^{+}e^{-}}$ collider.
It should be also emphasized that in the decoupling limit the H and A
bosons are almost degenerate in mass and have similar decay
properties111
These features will be exploited later on in our analysis..
In this note the study of Higgs boson pair production
at a future linear $\mathrm{e^{+}e^{-}}$ collider is presented. This analysis
extends the previous studies [4, 5]. The analyzed topologies
include the $\mathrm{b\bar{b}b\bar{b}}$ and $\mathrm{b\bar{b}\tau^{+}\tau^{-}}$ final states.
In the following we will assume the $\mathrm{e^{+}e^{-}\to HA}$ production process.
However, this assumption does not restrict the generality
of our study and the analyses developed in this
note are applicable to the process $\mathrm{e^{+}e^{-}\to hA}$ as well as to the processes
of associated Higgs pair production in CP-violating MSSM scenarios
where the neutral Higgs bosons are no longer CP eigenstates [6].
This note is organized as follows.
In Section 2 the assumed experimental conditions and
main features of the detector are discussed.
Section 3 describes the simulation of the signal and the
relevant background processes.
In Section 4 the analysis tools, employed in our study, are briefly
reviewed. Section 5 describes the analysis
of the $\mathrm{b\bar{b}b\bar{b}}$ and $\mathrm{b\bar{b}\tau^{+}\tau^{-}}$ final states.
In Section 6, we discuss the prospects of
measuring the Higgs boson properties using
the $\mathrm{b\bar{b}b\bar{b}}$ and $\mathrm{b\bar{b}\tau^{+}\tau^{-}}$ topologies. In Appendix the interpretation of this analysis is done for the SPS 1a benchmark point.
2 Experimental Conditions and Detector
Simulations
The study is performed for a linear collider operated
at a center-of-mass energies of 500${\mathrm{\ Ge\kern-1.0ptV}}$ to 1000${\mathrm{\ Ge\kern-1.0ptV}}$ and
an event sample corresponding to an integrated luminosity
of 500 fb${}^{-1}$.
The detector used in the simulation follows the proposal
for the TESLA collider presented in
the TDR [4]. Studies were also performed in Asia and North America [7, 8].
The interaction region is surrounded by a tracker consisting
of a multi-layered pixel micro-vertex detector (VTX) as the innermost part
and a large time projection chamber (TPC) supplemented with additional silicon strip detector in the forward region (FTD) and two bands between VTX and TPC. In radial direction follow
an electromagnetic calorimeter, a hadron calorimeter,
the coils of a superconducting magnet
and an instrumented iron flux return yoke.
The solenoidal magnetic field is 4 Tesla.
The envisaged impact parameter resolution
provided by the vertex detector is
$$\delta(IP_{r\phi,z})\leq 5\mu m\oplus\frac{10\mu m\hskip 2.845276ptGeV/c}{p%
\sin^{3/2}\theta},$$
(5)
where $p$ is the track momentum and $\theta$ is the polar angle.
The overall central tracker momentum resolution is
$$\sigma\left(\frac{1}{p_{t}}\right)\leq 5\cdot 10^{-5}~{}[{\mathrm{\ Ge\kern-1.%
0ptV}}/c]^{-1},$$
(6)
where $p_{t}$ is the transverse momentum. To enable efficient separation
of neutral and charged particles, highly granular
electromagnetic and hadronic calorimeters are foreseen for the
TESLA detector. Their energy resolutions are:
$$\frac{\sigma_{E_{el}}}{E_{el}}=\frac{10\%}{\sqrt{E}_{el}}\oplus 1\%,\hskip 56.%
905512pt\frac{\sigma_{E_{h}}}{E_{h}}=\frac{50\%}{\sqrt{E}_{h}}\oplus 4\%,$$
(7)
where $E_{el}$ and $E_{h}$ are the energies
measured for electrons and hadrons in the corresponding
calorimeters. The polar angle coverage of the central tracker
maintaining the resolution is
$|\cos\theta|<0.85$, above this range the tracking
resolution deteriorates. The electromagnetic and hadron
calorimeters cover $|\cos\theta|<0.996$
maintaining the resolution over the whole angular range.
The detector response is simulated with parametric fast
simulation program SIMDET [9].
Beamstrahlung is accounted for using CIRCE [10].
The event reconstruction is done in terms of energy flow objects.
First, tracks are measured with tracking system and associated
to calorimetric clusters to define charged energy flow objects
(electrons, muons and charged hadrons). Since momentum measurement
with tracking system is much more accurate than direction and energy
measurements with calorimeters, the tracking information
is used for estimation of the four-momentum of the charged objects.
Calorimetric clusters with no associated tracks are regarded as
neutral energy flow objects (photons and neutral hadrons).
Measurements of the four-momentum of neutral objects are solely
based on the calorimetric information.
3 Physics Processes and Monte Carlo
Samples
Samples of $\mathrm{e^{+}e^{-}\to HA}$ events are generated
for several Higgs boson mass hypotheses
with PYTHIA 6.2 [11], including initial state radiation.
The tree level cross section of the HA production
is related to the Higgs-strahlung cross section in the SM,
$\mathrm{\sigma_{HZ}^{SM}}$, in the following way [2]:
$$\mathrm{\mathrm{\sigma_{HA}}=g^{2}_{HAZ}\bar{\lambda}_{HA}\mathrm{\sigma_{HZ}^%
{SM}},}$$
(8)
where
$$\mathrm{\bar{\lambda}_{HA}=\frac{\lambda^{3/2}_{HA}}{\lambda^{1/2}_{HZ}[12m^{2%
}_{Z}/s+\lambda_{HZ}]}}$$
(9)
accounts for the correct suppression of the P-wave cross section near the kinematic threshold. The quantity $\lambda_{ij}=[1-(m_{i}+m_{j})^{2}/s][1-(m_{i}-m_{j})^{2}/s]$ is the usual momentum factor of the two particle phase space. The Higgs-strahlung cross section in the SM is computed using
the program HPROD [12].
For the case of maximal
allowed $\mathrm{g_{HAZ}}$ coupling, $\sin(\beta-\alpha)$ = 1, which is
assumed for these studies, signal cross sections are given in Table 1.
The branching fractions of the Higgs bosons into b-quarks and $\tau$-lepton
pairs are set to their typical values in the MSSM :
$\mathrm{Br(H\to b\bar{b})}$ = $\mathrm{Br(A\to b\bar{b})}$ = 90%, $\mathrm{Br(H\to\tau\tau)}$ = $\mathrm{Br(A\to\tau\tau)}$ = 10%.
For the background estimation, the following processes are generated
using PYTHIA 6.2: $\mathrm{\mathrm{e^{+}e^{-}}\to WW}$, $\mathrm{\mathrm{e^{+}e^{-}}\to ZZ}$,
$\mathrm{\mathrm{e^{+}e^{-}}\to\mathrm{q\bar{q}}}$222The $\mathrm{\mathrm{e^{+}e^{-}}\to t\bar{t}}$ events are also taken into account,
$\mathrm{\mathrm{e^{+}e^{-}}\to Z\mathrm{e^{+}e^{-}}}$ and $\mathrm{\mathrm{e^{+}e^{-}}\to We\nu}$. The cross sections
for the most important background processes
are given in Table 2.
4 Analysis tools
4.1 B-tagging
Identification of b-quarks plays a crucial role in this analysis. Efficient tagging of jets containing
heavy flavour hadrons will be achieved with a highly granular micro-vertex
detector, allowing for precise reconstruction of track parameters in the
vicinity of the primary interaction point.
The procedure of tagging b-jets exploits information from the single track as well as secondary vertex information.
Secondary vertices are searched within jets using the
package ZVTOP [13]
developed for the SLD experiment, where tracks are described
as probability tubes and seed vertices are defined as regions
where these tubes overlap. Afterwards an attempt is made to assign
additional tracks by an iterative procedure. For each found vertex
the invariant total mass and momentum are calculated from the four-momenta of energy flow
objects assigned to the vertex. Three dimensional decay length and
decay length significance are also computed.
For a jet flavour separation a neural network is developed.
The detailed description of the neural network implementation
for the full simulation of the TESLA detector with the
BRAHMS program [14]
can be found in [15]. Three different neural networks
are introduced for jets with no, one and more than one secondary
vertices found. For jets without secondary vertex,
the impact parameter joint probability [16] and the two
highest impact parameter
significances are used as inputs for the neural network. If jet contains
at least one secondary vertex, an additional information,
including vertex masses, momenta, decay lengths and decay
length significances, is fed into the neural network.
The neural networks are trained on event samples simulated with SIMDET
using the same variables and jet classification as in BRAHMS. For the
analysis, a jet-wise tag, referred hereafter as jet b-tag variable,
is used. For a jet with neural network output $x$ it is defined as
$$B(x)=\frac{f_{b}(x)}{f_{b}(x)+f_{udsc}(x)},$$
(10)
where
$f_{b}(x)$ and $f_{udsc}(x)$ are probability density functions
of neural network output in a samples of b– and udsc–jets, respectively.
Tagging of c-jets proceeds in a similar way.
Fig. 1 shows b-tag and c-tag
purity versus efficiency curves for $\mathrm{Z\rightarrow q\bar{q}}$
events simulated at
a center-of-mass energy of $\sqrt{s}=91.2\,\mathrm{GeV}$.
Results obtained with SIMDET and BRAHMS are compared in this figure.
The c-tag performance agrees within 5% over the entire range of efficiency.
Some discrepancy in modelling of the b-tag
is observed in the region of high efficiency and low purity.
This discrepancy occurs due to not adequate modelling of the
resolution tails in the impact parameter joint probability
distribution by SIMDET.
However, both $\mathrm{b\bar{b}b\bar{b}}$ and $\mathrm{b\bar{b}\tau^{+}\tau^{-}}$ analyses impose b-tag requirements
strong enough not to be sensitive to this discrepancy.
Some flavour tag related systematic studies have been performed.
It has been found that the b-tag performance is nearly independent of
the center-of-mass energy in the range from $\sqrt{s}=91.2\,\mathrm{GeV}$
to $\sqrt{s}=500\,\mathrm{GeV}$. Furthermore, possible changes in
micro-vertex detector configuration are found to have
marginal impact on the b-tag performance. For example,
removing innermost silicon layer changes the selection efficiency by not
more than 5%. These studies confirm the stability of b-tagging.
However, c-tag and e.g. b-quark charge tagging are depending on the
innermost layer.
4.2 Kinematic Fit
The mass resolution of the reconstructed Higgs bosons is improved
by means of a kinematic fit. In the $\mathrm{b\bar{b}b\bar{b}}$ analysis,
conservation of four-momentum is required, leading
in total to four constraints. In the case of the $\mathrm{b\bar{b}\tau^{+}\tau^{-}}$ topology,
the energies of $\tau$-leptons are not measurable due to undetectable
neutrinos. However, the flight direction of highly boosted $\tau$-leptons
can be approximated with good accuracy by the direction of the
total momentum of its visible decay products. Exploiting
this approximation together with the the four-momentum conservation
requirement, we arrive at a two constraint fit.
The jet angular and momentum resolution functions and the
angular resolution of $\tau$-leptons
are derived from Monte Carlo studies.
The kinematic fit is performed
using the code developed by DELPHI [17].
For the samples of the $\mathrm{HA\to b\bar{b}b\bar{b}}$ and $\mathrm{HA\to b\bar{b}\tau^{+}\tau^{-}}$ events
with ($m_{\mathrm{H}},m_{\mathrm{A}}$) = (150,100) GeV, the performance of the kinematic
fit is illustrated in Fig. 2 and 3,
respectively.
5 Analysis Procedures
5.1 The $\mathrm{HA\to b\bar{b}b\bar{b}}$ Channel
Events of the $\mathrm{b\bar{b}b\bar{b}}$ topology are characterized by four high
multiplicity hadronic jets, containing decay products
of b-hadrons. A cut-based technique is employed to separate signal
from background. Selection criteria are optimized separately for
500 GeV and 800 GeV center-of-mass energies. Each event is required
to pass the following cuts:
1.
Total energy deposited in the detector, the visible
energy $\mathrm{E_{vis}}$, must be greater than 340 GeV (600 GeV)
for 500 GeV (800 GeV) center-of-mass energies.
2.
Each event is forced into four jets using the DURHAM algorithm [18] and the number of tracks per jet is required to be greater than three.
3.
To separate centrally produced H and A bosons from the
WW and ZZ events, peaking in forward/backward direction,
we apply a cut on the polar angle of the thrust vector [11], $|\cos\theta_{T}|<0.8$.
4.
Further suppression of the WW and ZZ backgrounds is achieved by
requiring the event thrust value to be less than 0.88.
5.
Two fermion background is suppressed by applying a cut on
the
DURHAM jet resolution parameter, for which the event changes from four
to three jets, $\log_{10}{y_{34}}\geq-2.9$.
6.
High multiplicity six-jet events originating from the $\mathrm{\mathrm{e^{+}e^{-}}\to t\bar{t}}$
production are reduced by requiring the number of enflow (energy flow) objects in the event to be less than 130. This cut is applied only at $\sqrt{s}$ = 500 GeV.
7.
The background from $\mathrm{\mathrm{e^{+}e^{-}}\to t\bar{t}}$ events is further reduced by applying a
cut on the jet resolution parameter, for which the event changes from
six to five jets, $\log_{10}{y_{56}}$ $\leq$ -3.1 (-2.8) at
$\sqrt{s}$ = 500 GeV (800 GeV).
8.
Finally, we make use of the b-tag information to enhance the purity
of the selected event sample. First, the b-tag variable for each jet
is calculated as described in Section 4. The four b-tag variables are sorted in descending order,
$B_{1}>B_{2}>B_{3}>B_{4}$. Two quantities $B_{12}$, $B_{34}$ are
then defined as
$$B_{12}=\frac{B_{1}B_{2}}{B_{1}B_{2}+(1-B_{1})(1-B_{2})},$$
$$B_{34}=\frac{B_{3}B_{4}}{B_{3}B_{4}+(1-B_{3})(1-B_{4})}.$$
The value of $B_{12}$ has to be greater than 0.75 (0.6)
at $\sqrt{s}$ = 500 GeV (800 GeV). The value of $B_{34}$ is required to be greater
than 0.05 independent of the center-of-mass energy.
The numbers of expected signal and background events, retained after selection, and signal efficiency for the example points with ($m_{\mathrm{H}}$,$m_{\mathrm{A}}$) = (150,100) GeV at $\sqrt{s}$ = 500 GeVand ($m_{\mathrm{H}}$,$m_{\mathrm{A}}$) = (300,250) at $\sqrt{s}$ = 800 GeVare presented in Tables 3, 4. Fig. 4 and 5 show the distributions of
the selection variables for the different sources of
background and for the signal at $\sqrt{s}$ = 500 GeV. The cutflow table for this case shows that such cuts like those on visible energy, number of tracks per jet, polar angle on the thrust vector and b-tag cuts are very effective to separate signal from the background. Especially should be stressed that after all kinematic selection i.e. before b-tag cuts the signal efficiency is 54% while the background is drastically reduced (1.6% from the initial value). After b-tag cuts only 1.3% of the background events, left after kinematic selection, remains, while the signal efficiency is 43%. This confirms, that the new tool for b-tagging is very powerful. The signal efficiencies, number of signal events and total background for different Higgs boson masses after selection cuts are given in Table 5.
Events accepted in the final sample are subjected to a 4C kinematic fit.
For each of the three possible di-jet pairings, the di-jet mass sum and
the di-jet mass difference are reconstructed. For the example point the mass sum and the mass difference presented in Fig. 2 show that due to the kinematic fit the mass reconstruction resolution can be improved significantly.
5.2 The $\mathrm{HA\to b\bar{b}\tau^{+}\tau^{-},\tau^{+}\tau^{-}b\bar{b}}$ Channels
The signature of the $\mathrm{b\bar{b}\tau^{+}\tau^{-}}$ final state consists of two high-multiplicity hadronic jets enriched with the decay
products of b-hadrons and two low-multiplicity jets initiated
by $\tau$-leptons.
Initially, a preselection of high-multiplicity events compatible
with the four-jet topology is applied. Events are required to fulfill
the following criteria.
1.
The leptonic and two photon events are rejected by applying
lower cut on the number of energy flow objects, $\mathrm{N_{enflow}}$ $>$ 30.
2.
Genuine six-jet events, resulting from the $\mathrm{\mathrm{e^{+}e^{-}}\to t\bar{t}}$
production and characterized by very large particle multiplicity,
are partially suppressed by an upper cut on the number of energy flow
objects, $\mathrm{N_{enflow}}$ $<$ 120.
3.
The visible energy $\mathrm{E_{vis}}$ must be greater than
250 GeV (400 GeV) at $\sqrt{s}$ = 500 GeV (800 GeV). This cut suppresses background processes
characterized by large missing energy: $\mathrm{ZZ\to q\bar{q}\nu\bar{\nu}}$ events, radiative
returns to the Z resonance with a photon escaping into the beam pipe and $\mathrm{\mathrm{e^{+}e^{-}}\to We\nu}$, $\mathrm{\mathrm{e^{+}e^{-}}\to Z\mathrm{e^{+}e^{-}}}$ processes with electrons escaping undetected
in the forward direction. An upper cut $\mathrm{E_{vis}}$ $<$ 500 GeV (760 GeV) at
$\sqrt{s}$ = 500 GeV (800 GeV) allows partial
rejection of the fully hadronic $\mathrm{WW\to q\bar{q}q\bar{q}}$ and $\mathrm{ZZ\to q\bar{q}q\bar{q}}$ events without
significant loss in the signal efficiency.
4.
The event thrust value has to be less than 0.95. This cut disentangles
more spheric HA events from the WW, ZZ and $\mathrm{q\bar{q}}$ backgrounds in
which the high boost of the final state bosons/quarks results in large values
of the event thrust.
5.
To separate centrally produced H and A bosons from the WW
and ZZ processes in which weak bosons tend to peak in the forward/backward
directions, a cut on the polar angle of the thrust vector,
$|\cos\theta_{T}|$ $<$ 0.9, is applied.
6.
The event is discarded if it contains an $\mathrm{e^{+}e^{-}}$ or $\mathrm{\mu^{-}\mu^{+}}$ pair
with invariant mass compatible with the Z mass within 5 GeV.
This cut suppresses the $\mathrm{ZZ\to q\bar{q}\ell^{+}\ell^{-}}$ background.
Events satisfying these criteria are resolved into four-jet topology
and $\tau$-lepton identification is performed. Two identification
categories for $\tau$-leptons are introduced. A low multiplicity jet is assigned
to the first category if it fulfills one of the following tight conditions:
•
A jet contains only one charged track,
identified as an electron or muon. No neutral clusters are assigned to the jet.
•
A jet contains one charged track, identified
as hadron, and not more than four neutral clusters
(one-prong hadronic decay of $\tau$-lepton).
•
A jet contains three charged tracks with unit total charge and
has not more than two neutral calorimetric clusters (three-prong
hadronic decay of $\tau$-lepton).
To enhance the acceptance of signal events, the second identification
category, imposing looser criteria, is introduced. According to these
criteria, a jet is identified as $\tau$-lepton if it has not more than two
charged tracks and less than five neutral clusters.
An event is accepted if at least one of the low multiplicity jets satisfies the tight identification criteria.
Signal efficiencies and numbers of signal and background
events remaining after preselection are given in
Table 6.
In the next step, events are selected into the final sample
by means of a binned likelihood technique. The signal likelihood, $\mathrm{L_{HA}}$
is built from the following variables:
•
event thrust, T;
•
the polar angle of the thrust vector, $\cos\theta_{T}$;
•
minimal opening angle between any two of the four jets;
•
b-tag variables of the two hadronic jets, $\mathrm{B_{1}}$ and $\mathrm{B_{2}}$
(hadronic jets are ordered by their energy);
•
missing energy;
•
number of energy flow objects, $\mathrm{N_{enflow}}$.
Fig. 6 and 7 show
the distributions of these variables. The distribution of
$\mathrm{L_{HA}}$ is shown in Fig. 8 for
($m_{\mathrm{H}}$,$m_{\mathrm{A}}$) = (150,100) GeV at $\sqrt{s}$ = 500 GeV. An event is accepted in the
final sample if the value of $\mathrm{L_{HA}}$ is greater than certain
threshold optimized separately for each Higgs boson
mass hypothesis considered to yield maximal significance of the
signal. The number of signal and background events and the
signal efficiencies after final selection are given in
Table 7. Signal efficiency ranges between 30 and
45%, and the number of background events contributing
to the final sample between 2000 and 6000, depending on the Higgs mass.
For events accepted in the final sample,
the di-jet and di-tau invariant masses are reconstructed
exploiting the 2C kinematic fit, described in Section 4.
6 Results
In the final step of the analysis, the spectra
of the di-jet mass sum and difference, obtained in the $\mathrm{HA\to b\bar{b}b\bar{b}}$ channel,
and distributions of the di-jet and di-tau masses,
reconstructed in the $\mathrm{HA\to b\bar{b}\tau^{+}\tau^{-},\tau^{+}\tau^{-}b\bar{b}}$ channels, are used to determine
Higgs boson properties.
6.1 Cross section and Mass
First, the analysis is performed assuming that the natural widths
of the Higgs bosons are small compared to the detector
resolution.
As an example Fig. 9 shows the distributions of the di-jet mass sum and di-jet mass difference obtained after selection cuts and kinematic fit in the $\mathrm{HA\to b\bar{b}b\bar{b}}$
channel for the Higgs boson mass hypothesis of
($m_{\mathrm{H}}$,$m_{\mathrm{A}}$) = (300,250) GeV at $\sqrt{s}$ = 800 GeV. Three entries per event contribute to these distributions, corresponding to three possibilities to
pair jets in the four-jet events. Two entries form a so-called combinatorial background. Fig. 10 demonstrates the final di-jet mass sum and di-jet mass difference after the cut on the di-jet mass difference sum, respectively, as indicated by arrows in Fig. 9. The signal efficiencies, number of signal events and total background for different Higgs boson masses after cuts on di-jet mass sum and difference are given in Table 5. The mass distributions are fitted separately
with the signal normalization, the sum and the difference
of the Higgs boson masses as free parameters. The shapes of the
signal distributions are parametrised using high statistics signal samples.
Background is approximated with a polynomial function. From the fit,
the sum and the difference of the Higgs boson masses
and errors on these quantities are determined.
The di-jet mass sum and difference are found to be weakly correlated
quantities. Hence, the statistical errors on the masses of the
Higgs bosons can be calculated as follows:
$$\delta m_{\mathrm{H}}=\delta m_{\mathrm{A}}=\frac{1}{2}\sqrt{(\delta\Sigma^{2}%
+\delta\Delta^{2})},$$
where $\delta\Sigma$ and $\delta\Delta$ are statistical
uncertainties in determination of the Higgs boson mass sum and difference.
The error on the topological cross section is calculated as
$$\delta\sigma/\sigma=\sqrt{N_{S}+N_{B}}/N_{S}.$$
The notations $N_{S}$ and $N_{B}$ stand for
the number of background and signal entries
within the windows in the di-jet mass sum and di-jet mass
difference distributions, where the signal is accumulated. The boundaries
of these windows are optimized to yield a minimal relative error on
the topological cross section.
In the $\mathrm{HA\to b\bar{b}\tau^{+}\tau^{-},\tau^{+}\tau^{-}b\bar{b}}$ channels, in the case of large mass splitting between
Higgs bosons, the signal exhibits itself as two peaks in the
reconstructed di-jet and di-tau mass spectra as illustrated in
Fig. 11.
The spectra are fitted simultaneously with the superposition of background
and signal distributions. The fit is performed with
four free parameters : two normalization factors for the
$\mathrm{HA\to b\bar{b}\tau^{+}\tau^{-}}$ and $\mathrm{HA\to\tau^{+}\tau^{-}b\bar{b}}$ samples and two Higgs boson masses.
Background distribution is assumed to be well measured and therefore fixed.
The shapes of signal peaks are parametrised using high statistics
MC samples. When H and A are degenerate in mass,
signal distributions overlap and
the contributions from the $\mathrm{HA\to b\bar{b}\tau^{+}\tau^{-}}$ and $\mathrm{HA\to\tau^{+}\tau^{-}b\bar{b}}$ samples
cannot be disentangled. To stabilize the fitting procedure we assume that
the numbers of the $\mathrm{HA\to b\bar{b}\tau^{+}\tau^{-}}$ and $\mathrm{HA\to\tau^{+}\tau^{-}b\bar{b}}$ events, selected
into the final sample, are equal.
This assumption is validated by the properties of the
decoupling limit. As was mentioned in the introductory section,
in the decoupling limit the H and A bosons have almost identical
properties: $m_{\mathrm{H}}$ $\approx$ $m_{\mathrm{A}}$, $\mathrm{Br(H\to b\bar{b})}$ $\approx$ $\mathrm{Br(A\to b\bar{b})}$,
$\mathrm{Br(H\to\tau\tau)}$ $\approx$ $\mathrm{Br(A\to\tau\tau)}$, $\mathrm{\Gamma_{H}}$ $\approx$ $\mathrm{\Gamma_{A}}$.
Consequently, the $\mathrm{HA\to b\bar{b}\tau^{+}\tau^{-}}$ and $\mathrm{HA\to\tau^{+}\tau^{-}b\bar{b}}$ events are expected
to contribute equally to the final selected sample. This
reduces the number of free parameters to three : a common normalization factor for the two signal samples and two Higgs
boson masses. As an example, Fig. 12
shows the result of the fit for
($m_{\mathrm{H}}$,$m_{\mathrm{A}}$) = (300,300) GeV.
Table 8 summarizes the
statistical accuracy of the mass measurements for the $\mathrm{HA\to b\bar{b}b\bar{b}}$ and $\mathrm{HA\to b\bar{b}\tau^{+}\tau^{-},\tau^{+}\tau^{-}b\bar{b}}$ channels and their combination. The statistical accuracy of the combined
mass measurement is $\sim$ 100 MeV for the Higgs pair production far
above kinematic threshold and degrades to $\sim$ 1 GeV with aproaching
kinematic limit. At the same time the statistical errors for the $\sqrt{s}$ = 800 GeV are in general twice as bigger than those for the 500 GeV for the both channels. The combination of the two channels helps to improve the mass determination accuracy.
The statistical uncertainty on the topological cross
sections is reported in Table 9. In the $\mathrm{HA\to b\bar{b}b\bar{b}}$ channel, the topological
cross sections can be measured with relative precision between 1.5 and 6.6% while in the
$\mathrm{HA\to b\bar{b}\tau^{+}\tau^{-},\tau^{+}\tau^{-}b\bar{b}}$ channels between 5 to 30%.
A big part of SUSY parameter space leads to degenerate H and A Higgs boson masses. For this case the discovery significance as a function of $m_{H}$ (=$m_{A}$) (Fig. 13) is calculated for the $\mathrm{HA\to b\bar{b}b\bar{b}}$ channel at $\sqrt{s}$ = 800 GeV. Approaching the kinematic limit, the significance drops below 5$\sigma$ between 380 GeV to 390. The whole range of the significances for the Higgs masses is from 28.2 to 3.4. The 5$\sigma$ discovery reach and $\eta^{2}$ as a function of $m_{H}$ (=$m_{A}$) where $\eta^{2}$ is assumed $\mathrm{e^{+}e^{-}\to HA}$ cross section relative to that for $\sin^{2}(\beta-\alpha)=1$ (as shown in Fig. 13).
6.2 Width
We utilize the $\mathrm{HA\to b\bar{b}\tau^{+}\tau^{-},\tau^{+}\tau^{-}b\bar{b}}$ channels to measure the Higgs boson widths.
Fig. 14 illustrates the impact of the natural widths of
the Higgs bosons on the reconstructed di-jet and di-tau mass spectra.
We consider separately two cases :
1.
large mass splitting between H and A,
2.
mass degeneracy.
In the first case, the strategy of the measurement is the following.
The masses of Higgs bosons are fixed to the values measured in the
$\mathrm{HA\to b\bar{b}b\bar{b}}$ channel.
The reconstructed di-jet and di-tau mass line shapes are parametrised
as a function
of $\mathrm{\Gamma_{H}}$ and $\mathrm{\Gamma_{A}}$ using MC samples of the $\mathrm{HA\to b\bar{b}\tau^{+}\tau^{-}}$ and
$\mathrm{HA\to\tau^{+}\tau^{-}b\bar{b}}$ events generated with different Higgs boson widths. A simultaneous
log-likelihood fit of the parametrised mass line shapes
to the actual reconstructed di-jet and di-tau mass distributions is performed
with four free parameters : $\mathrm{\Gamma_{H}}$, $\mathrm{\Gamma_{A}}$
and two normalization factors for the $\mathrm{HA\to b\bar{b}\tau^{+}\tau^{-}}$ and $\mathrm{HA\to\tau^{+}\tau^{-}b\bar{b}}$ samples.
Fig. 15 shows the dependence of the log-likelihood
on the probed Higgs boson widths for the case of $\mathrm{\Gamma_{H}}$ = $\mathrm{\Gamma_{A}}$ = 5 GeV and
($m_{\mathrm{H}}$,$m_{\mathrm{A}}$) = (200,150) GeV at $\sqrt{s}$ = 500 GeV. The dependence is shown for the case when only di-jet mass or di-tau mass spectrum are fitted and for the simultaneous fit to the two distributions. The feasible accuracies of the
width measurements for three representative Higgs boson mass hypotheses
are given in Table 10. The precision on the width determination ranges from 600 MeV to 4.0 GeV.
In the case of very close Higgs boson masses, the broadening
of the signal peak in the mass distributions can be caused by two factors :
the finite natural widths of the Higgs bosons and non-zero mass
difference, $\delta M=m_{\mathrm{H}}-m_{\mathrm{A}}$. The width measurement in this case proceeds
as follows. The sum of the Higgs boson masses is fixed to the value
measured in the $\mathrm{HA\to b\bar{b}b\bar{b}}$ channel, $m_{\mathrm{H}}+m_{\mathrm{A}}=2M$.
From the $\mathrm{HA\to b\bar{b}\tau^{+}\tau^{-}}$ and $\mathrm{HA\to\tau^{+}\tau^{-}b\bar{b}}$ MC
samples with $\mathrm{\Gamma_{H}}$ = $\mathrm{\Gamma_{A}}$ = 0 GeV, the experimental mass resolution functions
$R_{bb,\tau\tau}(x)$ are derived, which approximate the reconstructed
di-jet and di-tau mass distributions in the limit of infinitely small
Higgs widths. To ensure stability of the fitting procedure, the H and A
boson are assumed to have identical decay properties,
$\mathrm{Br(H\to b\bar{b})}$ = $\mathrm{Br(A\to b\bar{b})}$, $\mathrm{Br(H\to\tau\tau)}$ = $\mathrm{Br(A\to\tau\tau)}$, $\mathrm{\Gamma_{H}}$ = $\mathrm{\Gamma_{A}}$ = $\Gamma$,
as expected in the decoupling limit.
The distributions of the invariant di-jet and di-tau
masses are fitted simultaneously with the functions
$$\begin{array}[]{ccccccc}F_{bb,\tau\tau}(x)&=&B_{bb,\tau\tau}(x)&+&N\cdot BW(x,%
M\pm\delta M/2,\Gamma)\otimes R_{bb,\tau\tau}(x)&+\\
&&&&N\cdot BW(x,M\mp\delta M/2,\Gamma)\otimes R_{bb,\tau\tau}(x)&\\
\end{array}$$
The fitting functions consist of three terms. The first one,
$B_{bb,\tau\tau}(x)$, describes the background spectrum. The other
two describe two signal distributions and represent
the convolution of the Breit-Wigner function
$$BW(x,M,\Gamma)=\frac{xM\Gamma}{(x^{2}-M^{2})^{2}+M^{2}\Gamma^{2}}$$
with the experimental resolution functions $R_{bb,\tau\tau}(x)$.
The fit is performed with the common normalization factor, $N$,
the Higgs boson width, $\Gamma$, and the Higgs boson mass difference,
$\delta M$, as free parameters. The Higgs width errors
obtained from the fitting procedure are presented in
Table 11 for three representative Higgs mass hypotheses. The precision on the width determination ranges from 1 GeV to 4.5 GeV.
7 Conclusion
We examined the potential of a future linear collider detector for the
determination of the MSSM Higgs boson properties exploiting
associated Higgs boson pair production followed by the Higgs decays
to b-quarks and $\tau$-leptons. It is shown that combining
the $\mathrm{HA\to b\bar{b}b\bar{b}}$ and $\mathrm{HA\to b\bar{b}\tau^{+}\tau^{-},\tau^{+}\tau^{-}b\bar{b}}$ channels the Higgs boson
masses can be measured with an accuracy of up to several hundred MeV for
Higgs pair production far above the kinematic threshold.
The precision deteriorates to about 1 GeV with approaching the kinematic
threshold. The topological
cross section $\mathrm{\sigma(\mathrm{e^{+}e^{-}}\rightarrow HA\rightarrow b\bar{b}b\bar{b})}$ can be measured with a relative precision
varying between 1.5 and 6.6%. The error on the topological cross sections
$\mathrm{\sigma(\mathrm{e^{+}e^{-}}\rightarrow HA\rightarrow b\bar{b}\tau^{+}%
\tau^{-})}$ and
$\mathrm{\sigma(\mathrm{e^{+}e^{-}}\rightarrow HA\rightarrow\tau^{+}\tau^{-}b%
\bar{b})}$ is estimated to range between 4 and 30%.
Moderate constraints on the Higgs boson widths can be set from
the analysis of the reconstructed di-jet and di-tau mass spectra in the
$\mathrm{HA\to b\bar{b}\tau^{+}\tau^{-}}$ and $\mathrm{HA\to\tau^{+}\tau^{-}b\bar{b}}$ channels.
The 5$\sigma$ discovery limit corresponds to the Higgs mass of around
385 GeV for the degenerate Higgs boson masses in
the $\mathrm{HA\to b\bar{b}b\bar{b}}$ channel at $\sqrt{s}$ = 800 GeV with integrated luminosity
of 500 fb${}^{-1}$.
Appendix A SPS 1a
The present analysis is applied to one of the so-called benchmark points SPS 1a for SUSY searches [19]. SPS 1 is a typical mSUGRA scenario which consists of a point with an intermediate value of $\tan\beta$ and a model line attached to it (SPS 1a) and of a ”typical” mSUGRA point with relatively high $\tan\beta$ (SPS 1b). The parameters for the SPS 1a point are $m_{0}$ = 100 GeV, $m_{1/2}$ = 250 GeV, $A_{0}$ = -100 GeV, $\tan\beta$ = 10, $\mu>0$. For this point the Higgs masses are $m_{h^{0}}$ = 113.7 GeV, $m_{A^{0}}$ = 394.65 GeV, $m_{H^{0}}$ = 394.9, $m_{H^{\pm}}$ = 403.6 GeV according to the Hdecay and Feynhiggsfast programs [20, 21].
The analysis is done for the center-of-mass energy of $\sqrt{s}=1\,\mathrm{TeV}$, at which the cross section for the process $\mathrm{e^{+}e^{-}\to HA}$ is 2.5 fb. The luminosity assumed is 1000 fb${}^{-1}$. The branching ratio for the H (A) Higgs boson to $b\bar{b}$ is 0.64 (0.40), $\Gamma_{tot}$ = 0.785 GeV (1.251 GeV).
The results presented in Fig. 16 are the mass sum and the mass difference after the selection cuts, kinematic fit and the final cuts for the mass difference and the mass sum respectively. The masses can be measured with precision of 1.3 GeV. The signal efficiency is 29% after selection cuts and 24% after cuts on di-jet mass sum and difference. The cross section can be measured with the relative uncertainty of 9%.
References
[1]
P.W. Higgs, Phys. Lett. 12 (1964) 132, Phys. Rev. Lett. 13 (1964) 508
and Phys. Rev. 145 (1966) 1156 ;
F. Englert and R. Brout, Phys. Rev. Lett. 13 (1964) 321;
G.S. Guralnik, C.R. Hagen and T.W.B. Kibble, Phys. Rev. Lett. 13 (1964) 585.
[2]
J.F. Gunion, H.E. Haber, G. Kane, S. Dawson, ”The Higgs Hunter’s Guide”, Addison Wesley, 1990.
[3]
H.P. Nilles, Phys. Rep. 110 (1984) 1;
H.E. Haber and G.L. Kane, Phys. Rep. 117 (1985) 75;
R. Barbieri, Riv. Nuovo Cim., 11 n${}^{\circ}$4 1.
[4]
T. Behnke, S. Bertolucci, R.-D. Heuer, R. Settles,
”TESLA : The Superconducting Electron-Positron Linear Collider
with an Integrated X-Ray Laser Laboratory. Technical Design Report,
Part IV : A Detector for TESLA”, DESY 2001-011 and ECFA 2001-209 (2001).
[5]
A. Andreazza, C. Troncon, ”Study of HA Production in $e^{+}e^{-}$ Collisions at $\sqrt{s}$ = 800 GeV”. DESY-123-E, p. 417.
[6]
A. Pilaftsis and C.E.M. Wagner, Nucl. Phys. B553 (1999) 3.
[7]
S. Kiyoura, S. Kanemura, K. Odagiri, Y. Okada, E. Senaha, S. Yamashita and
Y. Yasui,
arXiv:hep-ph/0301172.
[8]
T. Abe et al.,
”Linear Collider Physics Resource Book for Snowmass 2001”,
BNL-52627, CLNS 01/1729, FERMILAB-Pub-01/058-E, LBNL-47813, SLAC-R-570, UCRL-ID-143810-DR, LC-REV-2001-074-US.
[9]
SIMDET V4.0, M. Pohl and H.J. Schreiber, ”SIMDET : a Parametric
Monte Carlo for a TESLA Detector”, DESY-02-061 LC-DET-2002-005 (2002).
[10]
CIRCE V6, T. Ohl, Comp. Phys. Comm. 94 (1996) 53.
[11]
T. Sjöstrand, L. Lonnblad, S. Mrenna, ”PYTHIA 6.2: Physics and Manual”, hep-ph/0108264.
[12]
A. Djouadi, ”HPROD: A Program for SM and MSSM Higgs Boson Production in $e^{+}e^{-}$ Collisions”,
http://w3.lpm.univ-montp2.fr/$\tilde{\phantom{1}}$djouadi/GDR/programs/hprod.html
[13]
D. Jackson, Nucl. Instr. and Meth. A388 (1997) 247.
[14]
T. Behnke et al., ”BRAHMS : a Monte Carlo for a Detector
at a 500/800 GeV Linear Collider”, LC-TOOL-2001-005.
[15]
S. Xella-Hansen, D.J. Jackson, R. Hawkings, C.J.S. Damerell,
”Flavour Tagging Studies for the TESLA Linear Collider”,
LC-PHSM-2001-024.
[16]
R. Barate et al. ”A Measurement of $R_{b}$ using a Lifetime-Mass Tag”,
CERN-PPE/97-017.
[17]
N. Kjaer and R. Moller, ”Reconstruction of invariant
masses in multi-jet events”,
DELPHI Note 91-17 PHYS 88;
N. Kjaer, private communication.
[18]
S. Catani, Yu.L. Dokshitzer, M. Olsson, G. Turnock and B.R. Webber, Phys. Lett. B269 (1991) 432.
[19]
B.C. Allanach, M. Battaglia, G.A. Blair, M. Carena et al., hep-ph/0202233.
[20]
A. Djouadi, J. Kalinowski, M. Spira, ”Hdecay: a program for Higgs boson decays in the Standard Model and its supersymmetric extension”,
Comput. Phys. Commun. 108:56-74,1998; hep-ph/9704448.
[21]
S. Heinemeyer, W. Hollik, G. Weiglein, ”Feynhiggs and Feynhiggsfast: programs for higher order calculations in the neutral CP even Higgs boson sector of the MSSM”, LC-TH-2001-065, KA-TP-5-2001, 2001. In 2nd ECFA/DESY Study 1998-2001 2393-2421. |
Subsets and Splits