text
stringlengths
6
128k
We present the discovery of five quasars at z~6 selected from 260 deg^2 of the Sloan Digital Sky Survey (SDSS) southern survey, a deep imaging survey obtained by repeatedly scanning a stripe along the Celestial Equator. The five quasars with 20<z_{AB}<21 are 1-2 magnitudes fainter than the luminous z~6 quasars discovered in the SDSS main survey. One of them was independently discovered by the UKIRT Infrared Deep Sky Survey. These quasars, combined with another z~6 quasar known in this region, make a complete flux- limited quasar sample at z_{AB}<21. The sample spans the redshift range 5.85<z<6.12 and the luminosity range -26.5<M_{1450}<-25.4 (H_0=70 km s^{-1} Mpc^{-1}, Omega_{m}=0.3, and Omega_{Lambda}=0.7). We use the 1/V_{a} method to determine that the comoving quasar spatial density at <z>=6.0 and <M_{1450}>=-25.8 is (5.0+/-2.1) x 10^{-9} Mpc^{-3} mag^{-1}. We model the bright-end quasar luminosity function (QLF) at z~6 as a power law Phi(L_{1450}) \propto L_{1450}^{beta}. The slope beta calculated from a combination of our sample and the luminous SDSS quasar sample is -3.1+/-0.4, significantly steeper than the slope of the QLF at z~4. Based on the derived QLF, we find that the quasar/AGN population cannot provide enough photons to ionize the intergalactic medium (IGM) at z~6 unless the IGM is very homogeneous and the luminosity (L*_{1450}) at which the QLF power law breaks is very low.
Four classic criteria used to the classification of complex flows are discussed here. These criteria are useful to identify regions of the flow related to shear, elongation or rigid-body motion. These usual criteria, namely $Q$, $\Delta$, $\lambda_{2}$ and $\lambda_{cr}/\lambda_{ci}$, use the fluid's rate-of-rotation tensor, which is known to vary with respect to a reference frame. The advantages of using objective (invariant with respect to a general transformation on the reference frame) criteria are discussed in the present work. In this connection, we construct versions of classic criteria replacing standard vorticity, a non-objective quantity, by effective vorticity, a rate of rotation tensor with respect to the angular velocity of the eigenvectors of the strain rate tensor. The classic criteria and their corresponding objective versions are applied to classify two complex flows: the transient ABC flow and the flow through the abrupt 4:1 contraction. It is shown that the objective versions of the criteria provide richer information on the kinematics of the flow
We propose to use a quantum ratchet to transport quantum information in a chain of atoms trapped in an optical superlattice. The quantum ratchet is created by a continuous modulation of the optical superlattice which is periodic in time and in space. Though there is zero average force acting on the atoms, we show that indeed the ratchet effect permits atoms on even and odd sites to move along opposite directions. By loading the optical lattice with two-level bosonic atoms, this scheme permits to perfectly transport a qubit or entangled state imprinted in one or more atoms to any desired position in the lattice. From the quantum computation point of view, the transport is achieved by a smooth concatenation of perfect swap gates. We analyze setups with noninteracting and interacting particles and in the latter case we use the tools of optimal control to design optimal modulations. We also discuss the feasibility of this method in current experiments.
Motivated by the dynamics in the deep interiors of many stars, we study the interaction between overshooting convection and the large-scale poloidal fields residing in radiative zones. We have run a suite of 3D Boussinesq numerical calculations in a spherical shell that consists of a convection zone with an underlying stable region that initially compactly contains a dipole field. By varying the strength of the convective driving, we find that, in the less turbulent regime, convection acts as turbulent diffusion that removes the field faster than solely molecular diffusion would do. However, in the more turbulent regime, turbulent pumping becomes more efficient and partially counteracts turbulent diffusion, leading to a local accumulation of the field below the overshoot region. These simulations suggest that dipole fields might be confined in underlying stable regions by highly turbulent convective motions at stellar parameters. The confinement is of large-scale field in an average sense and we show that it is reasonably modeled by mean-field ideas. Our findings are particularly interesting for certain models of the Sun, which require a large-scale, poloidal magnetic field to be confined in the solar radiative zone in order to explain simultaneously the uniform rotation of the latter and the thinness of the solar tachocline.
For Banach spaces $X,Y,$ we consider a distance problem in the space of bounded linear operators $\mathcal{L}(X,Y).$ Motivated by a recent paper \cite{RAO21}, we obtain sufficient conditions so that for a compact operator $T\in\mathcal{L}(X,Y)$ and a closed subspace $Z\subset Y,$ the following equation holds, which relates global approximation with local approximation: \[d(T,\mathcal{L}(X,Z))=\sup\{d(Tx,Z):x\in X,\|x\|=1\}.\] In some cases, we show that the supremum is attained at an extreme point of the corresponding unit ball. Furthermore, we obtain some situations when the following equivalence holds: $$T\perp_B \mathcal{L}(X,Z)\Leftrightarrow T^{**}x_0^{**}\perp_B Z^{\perp\perp}\Leftrightarrow T^{**}\perp_B\mathcal{L}(X^{**},Z^{\perp\perp}),$$ for some $x_0^{**}\in X^{**}$ satisfying $\|T^{**}x_0^{**}\|=\|T^{**}\|\|x_0^{**}\|,$ where $Z^\perp$ is the annihilator of $Z.$ One such situation is when $Z$ is an $L^1-$predual space and an $M-$ideal in $Y$ and $T$ is a multi-smooth operator of finite order. Another such situation is when $X$ is an abstract $L_1-$space and $T$ is a multi-smooth operator of finite order. Finally, as a consequence of the results, we obtain a sufficient condition for proximinality of a subspace $Z$ in $Y.$
Angular momentum transport in protostellar discs can take place either radially, through turbulence induced by the magnetorotational instability (MRI), or vertically, through the torque exerted by a large-scale magnetic field that threads the disc. Using semi-analytic and numerical results, we construct a model of steady-state discs that includes vertical transport by a centrifugally driven wind as well as MRI-induced turbulence. We present approximate criteria for the occurrence of either one of these mechanisms in an ambipolar diffusion-dominated disc. We derive ``strong field'' solutions in which the angular momentum transport is purely vertical and ``weak field'' solutions that are the stratified-disc analogues of the previously studied MRI channel modes; the latter are transformed into accretion solutions with predominantly radial angular-momentum transport when we implement a turbulent-stress prescription based on published results of numerical simulations. We also analyze ``intermediate field strength'' solutions in which both modes of transport operate at the same radial location; we conclude, however, that significant spatial overlap of these two mechanisms is unlikely to occur in practice. To further advance this study, we have developed a general scheme that incorporates also the Hall and Ohm conductivity regimes in discs with a realistic ionization structure.
Convolutional neural networks (CNNs) are typically over-parameterized, bringing considerable computational overhead and memory footprint in inference. Pruning a proportion of unimportant filters is an efficient way to mitigate the inference cost. For this purpose, identifying unimportant convolutional filters is the key to effective filter pruning. Previous work prunes filters according to either their weight norms or the corresponding batch-norm scaling factors, while neglecting the sequential dependency between adjacent layers. In this paper, we further develop the norm-based importance estimation by taking the dependency between the adjacent layers into consideration. Besides, we propose a novel mechanism to dynamically control the sparsity-inducing regularization so as to achieve the desired sparsity. In this way, we can identify unimportant filters and search for the optimal network architecture within certain resource budgets in a more principled manner. Comprehensive experimental results demonstrate the proposed method performs favorably against the existing strong baseline on the CIFAR, SVHN, and ImageNet datasets. The training sources will be publicly available after the review process.
This paper gives a construction, using heat kernels, of differential forms on the moduli space of metrised ribbon graphs, or equivalently on the moduli space of Riemann surfaces with boundary. The construction depends on a manifold with a bundle of Frobenius algebras, satisfying various conditions. These forms satisfy gluing conditions which mean they form an open topological conformal field theory, i.e. a kind of open string theory. If the integral of these forms converged, it would yield the purely quantum part of the partition function of a Chern-Simons type gauge theory. Yang-Mills theory on a four manifold arises as one of these Chern-Simons type gauge theories.
The application of data-intensive automatic speech recognition (ASR) technologies to dysarthric and elderly adult speech is confronted by their mismatch against healthy and nonaged voices, data scarcity and large speaker-level variability. To this end, this paper proposes two novel data-efficient methods to learn homogeneous dysarthric and elderly speaker-level features for rapid, on-the-fly test-time adaptation of DNN/TDNN and Conformer ASR models. These include: 1) speaker-level variance-regularized spectral basis embedding (VR-SBE) features that exploit a special regularization term to enforce homogeneity of speaker features in adaptation; and 2) feature-based learning hidden unit contributions (f-LHUC) transforms that are conditioned on VR-SBE features. Experiments are conducted on four tasks across two languages: the English UASpeech and TORGO dysarthric speech datasets, the English DementiaBank Pitt and Cantonese JCCOCC MoCA elderly speech corpora. The proposed on-the-fly speaker adaptation techniques consistently outperform baseline iVector and xVector adaptation by statistically significant word or character error rate reductions up to 5.32% absolute (18.57% relative) and batch-mode LHUC speaker adaptation by 2.24% absolute (9.20% relative), while operating with real-time factors speeding up to 33.6 times against xVectors during adaptation. The efficacy of the proposed adaptation techniques is demonstrated in a comparison against current ASR technologies including SSL pre-trained systems on UASpeech, where our best system produces a state-of-the-art WER of 23.33%. Analyses show VR-SBE features and f-LHUC transforms are insensitive to speaker-level data quantity in testtime adaptation. T-SNE visualization reveals they have stronger speaker-level homogeneity than baseline iVectors, xVectors and batch-mode LHUC transforms.
The conditions to induce appreciable CP-and T-odd effects in neutrino oscillations are discussed. The propagation in matter leads to fake CP-and CPT-odd asymmetries, besides a Bohm-Aharonov type modification of the interference pattern. We study the separation of fake and genuine CP violation by means of energy and distance dependence.
We obtain some general restrictions on the continuous endomorphisms of a profinite group G under the assumption that G has only finitely many open subgroups of each index (an assumption which automatically holds, for instance, if G is finitely generated). In particular, given such a group G and a continuous endomorphism phi we obtain a semidirect decomposition of G into a 'contracting' normal subgroup and a complement on which phi induces an automorphism; both the normal subgroup and the complement are closed. If G is isomorphic to a proper open subgroup of itself, we show that G has an infinite abelian normal pro-p subgroup.
We present a versatile electric trap for the exploration of a wide range of quantum phenomena in the interaction between polar molecules. The trap combines tunable fields, homogeneous over most of the trap volume, with steep gradient fields at the trap boundary. An initial sample of up to 10^8 CH3F molecules is trapped for as long as 60 seconds, with a 1/e storage time of 12 seconds. Adiabatic cooling down to 120 mK is achieved by slowly expanding the trap volume. The trap combines all ingredients for opto-electrical cooling, which, together with the extraordinarily long storage times, brings field-controlled quantum-mechanical collision and reaction experiments within reach.
Over the past half-decade, many methods have been considered for neural architecture search (NAS). Bayesian optimization (BO), which has long had success in hyperparameter optimization, has recently emerged as a very promising strategy for NAS when it is coupled with a neural predictor. Recent work has proposed different instantiations of this framework, for example, using Bayesian neural networks or graph convolutional networks as the predictive model within BO. However, the analyses in these papers often focus on the full-fledged NAS algorithm, so it is difficult to tell which individual components of the framework lead to the best performance. In this work, we give a thorough analysis of the "BO + neural predictor" framework by identifying five main components: the architecture encoding, neural predictor, uncertainty calibration method, acquisition function, and acquisition optimization strategy. We test several different methods for each component and also develop a novel path-based encoding scheme for neural architectures, which we show theoretically and empirically scales better than other encodings. Using all of our analyses, we develop a final algorithm called BANANAS, which achieves state-of-the-art performance on NAS search spaces. We adhere to the NAS research checklist (Lindauer and Hutter 2019) to facilitate best practices, and our code is available at https://github.com/naszilla/naszilla.
In the case of cyclic quiver we prove that the deformed Harish-Chandra map whose existence was conjectured by Etingof and Ginzburg is well defined. As an application we prove Kirillov-type formula for the cyclotomic Bessel function.
We study the Voronoi Diagram of Rotating Rays, a Voronoi structure where the input sites are rays and the distance function between a point and a site/ray, is the counterclockwise angular distance. This novel Voronoi diagram is motivated by illumination or coverage problems, where a domain must be covered by floodlights/wedges of uniform angle, and the goal is to find the minimum angle necessary to cover the domain. We study the diagram in the plane, and we present structural properties, combinatorial complexity bounds, and a construction algorithm. If the rays are induced by a convex polygon, we show how to construct the Voronoi diagram within this polygon in linear time. Using this information, we can find in optimal linear time the Brocard angle, the minimum angle required to illuminate a convex polygon with floodlights of uniform angle.
Improvements in main memory storage density are primarily driven by process technology scaling, which negatively impacts reliability by exacerbating various circuit-level error mechanisms. To compensate for growing error rates, both memory manufacturers and consumers use error-mitigation mechanisms that improve manufacturing yield and allow system designers to meet reliability targets. Developing effective error mitigations requires understanding the errors' characteristics (e.g., worst-case behavior, statistical properties). Unfortunately, we observe that proprietary on-die Error-Correcting Codes (ECC) used in modern memory chips introduce new challenges to efficient error mitigation by obfuscating CPU-visible error characteristics in an unpredictable, ECC-dependent manner. This dissertation builds a detailed understanding of how on-die ECC obfuscates the statistical properties of main memory error mechanisms using a combination of real-chip experiments and statistical analyses. We experimentally study memory errors, examine how on-die ECC obfuscates their statistical characteristics, and develop new testing techniques to overcome the obfuscation. Our results show that the obfuscated error characteristics can be recovered using new memory testing techniques that exploit the interaction between on-die ECC and the statistical characteristics of memory error mechanisms to expose physical cell behavior. We conclude by discussing the critical need for transparency in DRAM reliability characteristics in order to enable DRAM consumers to better understand and adapt commodity DRAM chips to their system-specific needs. We hope and believe that the analysis, techniques, and results we present in this dissertation will enable the community to better understand and tackle current and future reliability challenges as well as adapt commodity memory to new advantageous applications.
The sensing and positioning capabilities foreseen in 6G have great potential for technology advancements in various domains, such as future smart cities and industrial use cases. Channel charting has emerged as a promising technology in recent years for radio frequency-based sensing and localization. However, the accuracy of these techniques is yet far behind the numbers envisioned in 6G. To reduce this gap, in this paper, we propose a novel channel charting technique capitalizing on the time of arrival measurements from surrounding Transmission Reception Points (TRPs) along with their locations and leveraging sensor fusion in channel charting by incorporating laser scanner data during the training phase of our algorithm. The proposed algorithm remains self-supervised during training and test phases, requiring no geometrical models or user position ground truth. Simulation results validate the achievement of a sub-meter level localization accuracy using our algorithm 90% of the time, outperforming the state-of-the-art channel charting techniques and the traditional triangulation-based approaches.
The BL Lac PKS 1413+135 was observed by the Large Survey Project "MeerKAT Absorption Line Survey" (MALS) in the L-band, at 1139 MHz and 1293-1379 MHz, targeting the HI and OH lines in absorption at z = 0.24671. The radio continuum is thought to come from a background object at redshift lower than 0.5, as suggested by the absence of gravitational images. The HI absorption line is detected at high signal-to-noise, with a narrow central component, and a red wing, confirming previous results. The OH 1720 MHz line is clearly detected in (maser) emission, peaking at a velocity shifted by -10 to -15 km/s with respect to the HI peak. The 1612 MHz line is lost due to radio interferences. The OH 1667 MHz main line is tentatively detected in absorption, but not the 1665 MHz one. Over 30 years, a high variability is observed in optical depths, due to the rapid changes of the line of sight, caused by the superluminal motions of the radio knots. The HI line has varied by 20 per cent in depth, while the OH-1720 MHz depth has varied by a factor 4. The position of the central velocity and the widths also varied. The absorbing galaxy is an early-type spiral (maybe S0) seen edge-on, with a prominent dust lane, covering the whole disk. Given the measured mass concentration, and the radio continuum size at centimeter wavelengths (100 mas corresponding to 400 pc at z = 0.25), the width of absorption lines from the nuclear regions are expected up to 250 km/S. The narrowness of the observed lines (< 15 km/s) suggest that the absorption comes from an outer gas ring, as frequently observed in S0 galaxies. The millimetric lines are even narrower (< 1 km/s), which corresponds to the continuum size restricted to the core. The core source is covered by individual 1 pc molecular clouds, of column density a few 10^22 cm-2, which is compatible with the gas screen detected in X-rays.
To reduce the size of recommendation models, there have been many studies on compressing recommendation models using knowledge distillation. In this paper, we decompose recommendation models into three layers, i.e., the input layer, the intermediate layer, and the output layer, and address deficiencies layer by layer. First, previous methods focus only on two layers, neglecting the input layer. Second, in the intermediate layer, existing methods ignore the inconsistency of user preferences induced by the projectors. Third, in the output layer, existing methods use only hard labels rather than soft labels from the teacher. To address these deficiencies, we propose \textbf{M}ulti-layer \textbf{K}nowledge \textbf{D}istillation (MKD), which consists of three components: 1) Distillation with Neighbor-based Knowledge (NKD) utilizes the teacher's knowledge about entities with similar characteristics in the input layer to enable the student to learn robust representations. 2) Distillation with Consistent Preference (CPD) reduces the inconsistency of user preferences caused by projectors in the intermediate layer by two regularization terms. 3) Distillation with Soft Labels (SLD) constructs soft labels in the output layer by considering the predictions of both the teacher and the student. Our extensive experiments show that MKD even outperforms the teacher with one-tenth of the model size.
Optical precision experiments are a powerful tool to explore hidden sectors of a variety of standard-model extensions with potentially tiny couplings to photons. An important example is given by extensions involving an extra light U(1) gauge degree of freedom, so-called paraphotons, with gauge-kinetic mixing with the normal photon. These models naturally give rise to minicharged particles which can be searched for with optical experiments. In this paper, we study the effects of paraphotons in such experiments. We describe in detail the role of a magnetic field for photon-paraphoton oscillations in models with low-mass minicharged particles. In particular, we find that the upcoming light-shining-through-walls experiments are sensitive to paraphotons and can distinguish them from axion-like particles.
The notion of Total Interference Degree (TID) is traditionally used to estimate the intensity of prevalent interference in a Multi-RadioMulti-ChannelWirelessMesh Network (MRMC WMN). Numerous Channel Assignment (CA) approaches, linkscheduling algorithms and routing schemes have been proposed for WMNs which rely entirely on the concept of TID estimates. They focus on minimizing TID to create a minimal interference scenario for the network. In our prior works [1] and [2], we have questioned the efficacy of TID estimate and then proposed two reliable interference estimation metrics viz., Channel Distribution Across Links Cost (CDALcost) and Cumulative X-Link-Set Weight (CXLSwt). In this work, we assess the ability of these interference estimation metrics to replace TID as the interferenceminimizing factor in a CA scheme implemented on a grid MRMC WMN. We carry out a comprehensive evaluation on ns-3 and then conclude from the results that the performance of the network increases by 10-15% when the CA scheme uses CXLSwt as the underlying Interference Mitigation Function (IMF) when compared with CA using TID as IMF. We also confirm that CDALcost is not a better IMF than TID and CXLSwt.
Pulsar data analysis pipelines have historically been comprised of bespoke software systems, supporting the off-line analysis of data. However modern data acquisition systems are making off-line analyses impractical. They often output multiple simultaneous high volume data streams, significantly increasing data capture rates. This leads to the accumulation of large data volumes, which are prohibitively expensive to retain. To maintain processing capabilities when off-line analysis becomes infeasible due to cost, requires a shift to on-line data processing. This paper makes four contributions facilitating this shift with respect to the search for radio pulsars: i) it characterises for the modern era, the key components of a pulsar search science (not signal processing) pipeline, ii) it examines the feasibility of implementing on-line pulsar search via existing tools, iii) problems preventing an easy transition to on-line search are identified and explained, and finally iv) it provides the design for a new prototype pipeline capable of overcoming such problems. Realised using Commercial off-the-shelf (COTS) software components, the deployable system is open source, simple, scalable, and cheap to produce. It has the potential to achieve pulsar search design requirements for the Square Kilometre Array (SKA), illustrated via testing under simulated SKA loads.
We study the intrinsic structure of parametric minimal discs in metric spaces admitting a quadratic isoperimetric inequality. We associate to each minimal disc a compact, geodesic metric space whose geometric, topological, and analytic properties are controlled by the isoperimetric inequality. Its geometry can be used to control the shapes of all curves and therefore the geometry and topology of the original metric space. The class of spaces arising in this way as intrinsic minimal discs is a natural generalization of the class of Ahlfors regular discs, well-studied in analysis on metric spaces.
Observations of young star-forming regions suggest that star clusters are born completely mass segregated. These initial conditions are, however, gradually lost as the star cluster evolves dynamically. For star clusters with single stars only and a canonical initial mass function, it has been suggested that traces of these initial conditions vanish at a time $\tau_\mathrm{v}$ between 3 and $3.5\,t_\mathrm{rh}$ (initial half-mass relaxation times). Since a significant fraction of stars are observed in binary systems and it is widely accepted that most stars are born in binary systems, we aim to investigate what role a primordial binary population (even up to $100\,\%$ binaries) plays in the loss of primordial mass segregation of young star clusters. We used numerical $N$-body models similar in size to the Orion Nebula Cluster (ONC) -- a representative of young open clusters -- integrated over several relaxation times to draw conclusions on the evolution of its mass segregation. We also compared our models to the observed ONC. We found that $\tau_\mathrm{v}$ depends on the binary star fraction and the distribution of initial binary parameters that include a semi-major axis, eccentricity, and mass ratio. For instance, in the models with $50\,\%$ binaries, we find $\tau_\mathrm{v} = (2.7 \pm 0.8)\,t_\mathrm{rh}$, while for $100\,\%$ binary fraction, we find a lower value $\tau_\mathrm{v} = (2.1 \pm 0.6)\,t_\mathrm{rh}$. We also conclude that the initially completely mass segregated clusters, even with binaries, are more compatible with the present-day ONC than the non-segregated ones.
Lattice simulations can play an important role in the study of dynamical electroweak symmetry breaking by providing quantitative results on the nonperturbative dynamics of candidate theories. For this programme to succeed, it is crucial to identify the questions that are relevant for phenomenology, and develop the tools that will provide robust answers to these questions. The existence of a conformal window for nonsupersymmetric gauge theories, and its characterization, is one of the phenomenologically important problems that can be studied on the lattice. We summarize the recent results from studies of IR fixed points by numerical simulations, discuss their current limitations, and analyze the future perspectives.
In this paper, we introduce a bilevel optimization framework for addressing inverse mean-field games, alongside an exploration of numerical methods tailored for this bilevel problem. The primary benefit of our bilevel formulation lies in maintaining the convexity of the objective function and the linearity of constraints in the forward problem. Our paper focuses on inverse mean-field games characterized by unknown obstacles and metrics. We show numerical stability for these two types of inverse problems. More importantly, we, for the first time, establish the identifiability of the inverse mean-field game with unknown obstacles via the solution of the resultant bilevel problem. The bilevel approach enables us to employ an alternating gradient-based optimization algorithm with a provable convergence guarantee. To validate the effectiveness of our methods in solving the inverse problems, we have designed comprehensive numerical experiments, providing empirical evidence of its efficacy.
Alfven waves created by sub-photospheric motions or by magnetic reconnection in the low solar atmosphere seem good candidates for coronal heating. However, the corona is also likely to be heated more directly by magnetic reconnection, with dissipation taking place in current sheets. Distinguishing observationally between these two heating mechanisms is an extremely difficult task. We perform 1.5-dimensional MHD simulations of a coronal loop subject to each type of heating and derive observational quantities that may allow these to be differentiated.
The class of Cressie-Read empirical likelihoods are constructed with weights derived at a minimum distance from the empirical distribution in the Cressie-Read family of divergences indexed by $\gamma$ under the constraint of an unbiased set of $M$-estimating equations. At first order, they provide valid posterior probability statements for any given prior, but the bias in coverage of the resulting empirical quantile is inversely proportional to the asymptotic efficiency of the corresponding $M$-estimator. The Cressie-Read empirical likelihoods based on the maximum likelihood estimating equations bring about quantiles covering with $O(n^{-1})$ accuracy at the underlying posterior distribution. The choice of $\gamma$ has an impact on the variance in small samples of the posterior quantile function. Examples are given for the $M$-type estimating equations of location and for the quasi-likelihood functions in the generalized linear models.
We propose a modified density estimation problem that is highly effective for detecting anomalies in tabular data. Our approach assumes that the density function is relatively stable (with lower variance) around normal samples. We have verified this hypothesis empirically using a wide range of real-world data. Then, we present a variance-stabilized density estimation problem for maximizing the likelihood of the observed samples while minimizing the variance of the density around normal samples. To obtain a reliable anomaly detector, we introduce a spectral ensemble of autoregressive models for learning the variance-stabilized distribution. We have conducted an extensive benchmark with 52 datasets, demonstrating that our method leads to state-of-the-art results while alleviating the need for data-specific hyperparameter tuning. Finally, we have used an ablation study to demonstrate the importance of each of the proposed components, followed by a stability analysis evaluating the robustness of our model.
Due to the powerful computing capability of quantum computers, cryptographic researchers have applied quantum algorithms to cryptanalysis and obtained many interesting results in recent years. In this paper, we study related-key attack in the quantum setting, and proposed a specific related-key attack which can recover the key of block ciphers efficiently, as long as the attacked block ciphers satisfy certain conditions. The attack algorithm employs Bernstein-Vazirani algorithm as a subroutine and requires the attacker to query the encryption oracle with quantum superpositions. Afterwards, we rigorously demonstrate the validity of the attack and analyze its complexity. Our work shows that related-key attack is quite powerful when combined with quantum algorithms, and provides some guidance for the design of block ciphers that are secure against quantum adversaries.
In this paper we provide a (negative) solution to a problem posed by Stanis{\l}aw Krajewski. Consider a recursively enumerable theory U and a finite expansion of the signature of U that contains at least one predicate symbol of arity $\ge$ 2. We show that, for any finite extension $\alpha$ of U in the expanded language that is conservative over U, there is a conservative extension $\beta$ of U in the expanded language, such that $\alpha\vdash\beta$ and $\beta\nvdash\alpha$. The result is preserved when we consider either extensions or model-conservative extensions of U in stead of conservative extensions. Moreover, the result is preserved when we replace $\vdash$ as ordering on the finitely axiomatized extensions in the expanded language by a special kind of interpretability, to wit interpretability that identically translates the symbols of the U-language. We show that the result fails when we consider an expansion with only unary predicate symbols for conservative extensions of U ordered by interpretability that preserves the symbols of U.
We employ a continuous dynamical decoupling scheme to suppress the decoherence induced by elastic collisions of cold atoms. Using a continuous echo pulse we achieve a thirty-fold increase in the coherence time of Rb87 atoms trapped in a dipole trap. Coherence times of more than 100ms are demonstrated for an ensemble with an optical depth of 120.
We compute quantum corrections to the Raychaudhuri equation, by replacing classical geodesics with quantal (Bohmian) trajectories, and show that they prevent focusing of geodesics, and the formation of conjugate points. We discuss implications for the Hawking-Penrose singularity theorems, and for curvature singularities.
Boundary value problems for nonlocal fractional elliptic equations with parameter in Banach spaces are studied. Uniform $L_p$-separability properties and sharp resolvent estimates are obtained for elliptic equations in terms of fractional derivatives. Particularly, it is proven that the fractional ellipitic operator generated by these equations is sectorial and also is a generator of an analytic semigroup. Moreover, maximal regularity properties of nonlocal fractional abstract parabolic equation are established. As an application, the nonlocal anisotropic fractional differential equations and the system of nonlocal fractional differential equations are studied.
In this paper we show that BF topological superconductors (insulators) exibit phase transitions between different topologically ordered phases characterized by different ground state degeneracy on manifold with non-trivial topology. These phase transitions are induced by the condensation (or lack of) of topological defects. We concentrate on the (2+1)-dimensional case where the BF model reduce to a mixed Chern-Simons term and we show that the superconducting phase has a ground state degeneracy $k$ and not $k^2$. When the symmetry is $U(1) \times U(1)$, namely when both gauge fields are compact, this model is not equivalent to the sum of two Chern-Simons term with opposite chirality, even if naively diagonalizable. This is due to the fact that U(1) symmetry requires an ultraviolet regularization that make the diagonalization impossible. This can be clearly seen using a lattice regularization, where the gauge fields become angular variables. Moreover we will show that the phase in which both gauge fields are compact is not allowed dynamically.
The magnetic and iron vacancy orders in superconducting (Tl,Rb)2Fe4Se5 single-crystals are investigated by using a high-pressure neutron diffraction technique. Similar to the temperature effect, the block antiferromagnetic order gradually decreases upon increasing pressure while the Fe vacancy superstructural order remains intact before its precipitous disappearance at the critical pressure Pc = 8.3 GPa. Combined with previously determined Pc for superconductivity, our phase diagram under pressure reveals the concurrence of the block AFM order, the iron vacancy order and superconductivity for the 245 superconductor. A synthesis of current experimental data in a coherent physical picture is attempted.
We consider ${\cal N}=(n,0)$ supersymmetric AdS$_3$ vacua of type II supergravity realising the superconformal algebra $\mathfrak{osp}(n|2)$ for $n>4$. For the cases $n=6$ and $n=5$, one can realise these algebras on backgrounds that decompose as foliations of AdS$_3\times \mathbb{CP}^3$ ( squashed $\mathbb{CP}^3$ for $n=5$) over an interval. We classify such solutions with bi-spinor techniques and find the local form of each of them: They only exist in (massive) IIA and are defined locally in terms of an order 3 polynomial $h$ similar to the AdS$_7$ vacua of (massive) IIA. Many distinct local solutions exist for different tunings of $h$ that give rise to bounded (or semi infinite) intervals bounded by physical behaviour. We show that it is possible to glue these local solutions together by placing D8 branes on the interior of the interval without breaking supersymmetry, which expands the possibilities for global solutions immensely. We illustrate this point with some simple examples. Finally we also show that AdS$_3$ vacua for $n=7,8$ only exist in $d=11$ supergravity and are all locally AdS$_4\times$S$^7$.
This paper proposes the creation of different interfaces in the mobile operating system for different age groups. The different age groups identified are kids, elderly people and all others. The motive behind creating different interfaces is to make the smartphones of today's world usable to all age groups.
In this paper we propose the first better than second order accurate method in space and time for the numerical solution of the resistive relativistic magnetohydrodynamics (RRMHD) equations on unstructured meshes in multiple space dimensions. The nonlinear system under consideration is purely hyperbolic and contains a source term, the one for the evolution of the electric field, that becomes stiff for low values of the resistivity. For the spatial discretization we propose to use high order $\PNM$ schemes as introduced in \cite{Dumbser2008} for hyperbolic conservation laws and a high order accurate unsplit time discretization is achieved using the element-local space-time discontinuous Galerkin approach proposed in \cite{DumbserEnauxToro} for one-dimensional balance laws with stiff source terms. The divergence free character of the magnetic field is accounted for through the divergence cleaning procedure of Dedner et al. \cite{Dedneretal}. To validate our high order method we first solve some numerical test cases for which exact analytical reference solutions are known and we also show numerical convergence studies in the stiff limit of the RRMHD equations using $\PNM$ schemes from third to fifth order of accuracy in space and time. We also present some applications with shock waves such as a classical shock tube problem with different values for the conductivity as well as a relativistic MHD rotor problem and the relativistic equivalent of the Orszag-Tang vortex problem. We have verified that the proposed method can handle equally well the resistive regime and the stiff limit of ideal relativistic MHD. For these reasons it provides a powerful tool for relativistic astrophysical simulations involving the appearance of magnetic reconnection.
In this article I attempt to collect some ideas,opinions and formulae which may be useful in solving the problem of gauge/ string / space-time correspondence This includes the validity of D-brane representation, counting of gauge-invariant words, relations between the null states and the Yang-Mills equations and the discussion of the strong coupling limit of the string sigma model. The article is based on the talk given at the "Odyssey 2001" conference.
Let $C,A$ be countable abelian groups. In this paper we determine the complexity of classifying extensions $C$ by $A$, in the cases when $C$ is torsion-free and $A$ is a $p$-group, a torsion group with bounded primary components, or a free $R$-module for some subring $R\subseteq \mathbb{Q}$. Precisely, for such $C$ and $A$ we describe in terms of $C$ and $A$ the potential complexity class in the sense of Borel complexity theory of the equivalence relation $\mathcal{R}_{\mathbf{Ext}\left( C,A\right) }$ of isomorphism of extensions of $C$ by $A$. This complements a previous result by the same author, settling the case when $C$ is torsion and $A$ is arbitrary. We establish the main result within the framework of Borel-definable homological algebra, recently introduced in collaboration with Bergfalk and Panagiotopoulos. As a consequence of our main results, we will obtain that if $C$ is torsion-free and $A$ is either a free $R$-module or a torsion group with bounded components, then an extension of $C$ by $A$ splits if and only if it splits on all finite-rank subgroups of $C$. This is a purely algebraic statements obtained with methods from Borel-definable homological algebra.
Global dimensions for fusion categories defined by a pair (G,k), where G is a Lie group and k a positive integer, are expressed in terms of Lie quantum superfactorial functions. The global dimension is defined as the square sum of quantum dimensions of simple objects, for the category of integrable modules over an affine Lie algebra at some level. The same quantities can also be defined from the theory of quantum groups at roots of unity or from conformal field theory WZW models. Similar results are also presented for those associated module-categories that can be obtained via conformal embeddings (they are "quantum subgroups" of a particular kind). As a side result, we express the classical (or quantum) Weyl denominator of simple Lie groups in terms of products of classical (or quantum) factorials calculated for the exponents of the group. Some calculations use the correspondence existing between periodic quivers for simply-laced Lie groups and fusion rules for module-categories (alias nimreps) of type SU(2).
This work concerns random dynamics of hyperbolic entire and meromorphic functions of finite order and whose derivative satisfies some growth condition at infinity. This class contains most of the classical families of transcendental functions and goes much beyond. Based on uniform versions of Nevanlinna's value distribution theory we first build a thermodynamical formalism which, in particular, produces unique geometric and fiberwise invariant Gibbs states. Moreover, spectral gap property for the associated transfer operator along with exponential decay of correlations and a central limit theorem are shown. This part relies on our construction of new positive invariant cones that are adapted to the setting of unbounded phase spaces. This setting rules out the use of Hilbert's metric along with the usual contraction principle. However these cones allow us to apply a contraction argument stemming from Bowen's initial approach.
We establish several new bounds for the number of conjugacy classes of a finite group, all of which involve the maximal number c of conjugacy classes of a normal subgroup fixed by some element of a suitable subset of the group. To apply these formulas effectively, the parameter c, which in general is hard to control, is studied in some important situations. These results are then used to provide a new, shorter proof of the most difficult case of the well-known k(GV)-problem, which occurs for p=5 and V induced from the natural module of a 5-complement of GL(2,5). We also show how, for large p, the new results reduce the k(GV)-problem to the primitive case, thereby improving previous work on this. Furthermore, we discuss how they can be used in tackling the imprimitive case of the as of yet unsolved noncoprime k(GV)-problem.
This paper studies the transmission of Gaussian sources through additive white Gaussian noise (AWGN) channels in bandwidth expansion regime, i.e., the channel bandwidth is greater than the source bandwidth. To mitigate the error propagation phenomenon of conventional digital transmission schemes, we propose in this paper a new capacity-approaching joint source channel coding (JSCC) scheme based on partially block Markov superposition transmission (BMST) of nested lattice codes. In the proposed scheme, first, the Gaussian source sequence is discretized by a lattice-based quantizer, resulting in a sequence of lattice points. Second, these lattice points are encoded by a short systematic group code. Third, the coded sequence is partitioned into blocks of equal length and then transmitted in the BMST manner. Main characteristics of the proposed JSCC scheme include: 1) Entropy coding is not used explicitly. 2) Only parity-check sequence is superimposed, hence, termed partially BMST (PBMST). This is different from the original BMST. To show the superior performance of the proposed scheme, we present extensive simulation results which show that the proposed scheme performs within 1.0 dB of the Shannon limits. Hence, the proposed scheme provides an attractive candidate for transmission of Gaussian sources.
When subjected to electro-mechanical loading, ferroelectrics see their polarization evolve through the nucleation and evolution of domains. Existing mesoscale phase-field models for ferroelectrics are typically based on a gradient-descent law for the evolution of the order parameter. While this implicitly assumes that domain walls evolve with linear kinetics, experiments instead indicate that domain wall kinetics is nonlinear. This, in turn, is an important feature for the modeling of rate-dependent effects in polarization switching. We propose a new multiple-phase-field model for ferroelectrics, which permits domain wall motion with nonlinear kinetics, with applications in other solid-solid phase transformation problems. By means of analytical traveling wave solutions, we characterize the interfacial properties (energy and width) and the interface kinetics of straight domain walls, as furnished by the general kinetics model, and compare them to those of the classical Allen--Cahn model. We show that the proposed model propagates domain walls with arbitrarily chosen nonlinear kinetic relations, which can be tuned to differ for the different types of domain walls in accordance with experimental evidence.
Fetal neuroinflammation and prenatal stress (PS) may contribute to lifelong neurological disabilities. Astrocytes and microglia, among the brain's non-neuronal glia cell populations, play a pivotal role in neurodevelopment, predisposition to and initiation of disease throughout lifespan. One of the most common neurodevelopmental disorders manifesting between 1-4 years of age is autism spectrum disorder (ASD). A pathological glial-neuronal interplay is thought to increase the risk for clinical manifestation of ASD in at-risk children, but the mechanisms remain poorly understood and integrative, multi-scale models are needed. We propose a model that integrates the data across the scales of physiological organization, from genome to phenotype, and provides a foundation to explain the disparate findings on the genomic level. We hypothesize that via gene-environment interactions, fetal neuroinflammation and PS may reprogram glial immunometabolic phenotypes that impact neurodevelopment and neurobehavior. Drawing on genomic data from the recently published series of ovine and rodent glial transcriptome analyses with fetuses exposed to neuroinflammation or PS, we conduct an analysis on the Simons Foundation Autism Research Initiative (SFARI) Gene database. We confirm 21 gene hits. Using unsupervised statistical network analysis, we then identify six clusters of probable protein-protein interactions mapping onto the immunometabolic and stress response networks and epigenetic memory. These findings support our hypothesis. We discuss the implications for ASD etiology, early detection, and novel therapeutic approaches. We conclude with delineation of the next steps to verify our model on the individual gene level in an assumption-free manner.
We have measured magnetic trap lifetimes of ultra-cold Rb87 atoms at distances of 5-1000 microns from surfaces of conducting metals with varying resistivity. Good agreement is found with a theoretical model for losses arising from near-field magnetic thermal noise, confirming the complications associated with holding trapped atoms close to conducting surfaces. A dielectric surface (silicon) was found in contrast to be so benign that we are able to evaporatively cool atoms to a Bose-Einstein condensate by using the surface to selectively adsorb higher energy atoms.
Prostate cancer is a highly prevalent cancer and ranks as the second leading cause of cancer-related deaths in men globally. Recently, the utilization of multi-modality transrectal ultrasound (TRUS) has gained significant traction as a valuable technique for guiding prostate biopsies. In this study, we propose a novel learning framework for clinically significant prostate cancer (csPCa) classification using multi-modality TRUS. The proposed framework employs two separate 3D ResNet-50 to extract distinctive features from B-mode and shear wave elastography (SWE). Additionally, an attention module is incorporated to effectively refine B-mode features and aggregate the extracted features from both modalities. Furthermore, we utilize few shot segmentation task to enhance the capacity of classification encoder. Due to the limited availability of csPCa masks, a prototype correction module is employed to extract representative prototypes of csPCa. The performance of the framework is assessed on a large-scale dataset consisting of 512 TRUS videos with biopsy-proved prostate cancer. The results demonstrate the strong capability in accurately identifying csPCa, achieving an area under the curve (AUC) of 0.86. Moreover, the framework generates visual class activation mapping (CAM), which can serve as valuable assistance for localizing csPCa. These CAM images may offer valuable guidance during TRUS-guided targeted biopsies, enhancing the efficacy of the biopsy procedure.The code is available at https://github.com/2313595986/SmileCode.
In this article, we separate the vector and axialvector components of the tensor diquark operators explicitly, construct the axialvector-axialvector type and vector-vector type scalar tetraquark currents and scalar-tensor type tensor tetraquark current to study the scalar, vector and axialvector tetraquark states with the QCD sum rules in a consistent way. The present calculations do not favor assigning the $Z_c(4100)$ to be a scalar or vector tetraquark state. If the $Z_c(4100)$ is a scalar tetraquark state without mixing effects, it should have a mass about $3.9\,\rm{GeV}$ or $4.0\,\rm{GeV}$ rather than $4.1\,\rm{GeV}$; on the other hand, if the $Z_c(4100)$ is a vector tetraquark state, it should have a mass about $4.2\,\rm{GeV}$ rather than $4.1\,\rm{GeV}$. However, if we introduce mixing, a mixing scalar tetraquark state can have a mass about $4.1\,\rm{GeV}$. As a byproduct, we obtain an axialvector tetraquark candidate for the $Z_c(4020)$.
The formation of nanostructures during metalorganic vapor-phase epitaxy on patterned (001)/(111)B GaAs substrates is reviewed. The focus of this review is on the seminal experiments that revealed the key kinetic processes during nanostructure formation and the theory and modelling that explained the phenomenology in successively greater detail. Experiments have demonstrated that V-groove quantum wires and pyramidal quantum dots result from self-limiting concentration profiles that develop at the bottom of V-grooves and inverted pyramids, respectively. In the 1950s, long before the practical importance of patterned substrates became evident, the mechanisms of capillarity during the equilibration of non-planar surfaces were identified and characterized. This was followed, from the late 1980s by the identification of growth rate anisotropies (i.e. differential growth rates of crystallographic facets) and precursor decomposition anisotropies, with parallel developments in the fabrication of V-groove quantum wires and pyramidal quantum dots. The modelling of these growth processes began at the scale of facets and culminated in systems of coupled reaction-diffusion equations, one for each crystallographic facet that defines the pattern, which takes account of the decomposition and surface diffusion kinetics of the group-III precursors and the subsequent surface diffusion and incorporation of the group-III atoms released by these precursors. Solutions of the equations with optimized parameters produced concentration profiles that provided a quantitative interpretation of the time-, temperature-, and alloy-concentration dependence of the self-ordering process seen in experiments.
The study of topology in solids is undergoing a renaissance following renewed interest in the properties of ferroic domain walls as well as recent discoveries regarding topological insulators and skyrmionic lattices. Each of these systems possess a property that is `protected' in a symmetry sense, and is defined rigorously using a branch of mathematics known as topology. In this article we review the formal definition of topological defects as they are classified in terms of homotopy theory, and discuss the precise symmetry-breaking conditions that lead to their formation. We distinguish topological defects from geometric defects, which arise from the details of the stacking or structure of the material but are not protected by symmetry. We provide simple material examples of both topological and geometric defect types, and discuss the implications of the classification on the resulting material properties.
Security of a continuous-variable quantum key distribution protocol based on noisy coherent states and channel is analyzed. Assuming the noise of coherent states is induced by Fred, a neutral party relative to others, we prove that the prepare and measurement scheme and entanglement-based scheme are equivalent. Then, we show that this protocol is secure against Gaussian collective attacks even if the channel is lossy and noisy, and further, a lower bound to the secure key rate is derived.
We discuss a sample of over 3000 candidate RR Lyrae stars selected by various methods using Sloan Digital Sky Survey data for about 1000 deg^2 of sky. These stars probe the halo structure out to ~100 kpc from the Galactic center. Their spatial and radial velocity distributions are very inhomogeneous, with the most prominent features tracing the Sgr dwarf tidal stream. Outside the Sgr dwarf tidal stream, the spatial distribution in the 5-60 kpc range of Galactocentric radius R is well described by an R^{-3} power law.
We show that if an open arc J of the boundary of a Jordan domain $\Omega$ is rectifiable, then the derivative $\Phi$' of the Riemann map $\Phi: D\rightarrow \Omega$ from the open unit disk D onto $\Omega$ behaves as an $H^1$ function when we approach the arc $\Phi^{-1}(J^{\prime})$,where $J^{\prime}$ is any compact subarc of $J$. "
We present Hubble Space Telescope imaging of 14 gas-rich, low surface brightness galaxies in the field at distances of 25-36 Mpc, with mean effective radii and $g$-band central surface brightnesses of 1.9 kpc and 24.2 mag arcsec$^{-2}$. Nine meet the standard criteria to be considered ultra-diffuse galaxies (UDGs). An inspection of point-like sources brighter than the turnover magnitude of the globular cluster luminosity function and within twice the half-light radii of each galaxy reveals that, unlike those in denser environments, gas-rich, field UDGs host very few old globular clusters (GCs). Most of the targets (nine) have zero candidate GCs, with the remainder having one or two candidates each. These findings are broadly consistent with expectations for normal dwarf galaxies of similar stellar mass. This rules out gas-rich, field UDGs as potential progenitors of the GC-rich UDGs that are typically found in galaxy clusters. However, some in galaxy groups may be directly accreted from the field. In line with other recent results, this strongly suggests that there must be at least two distinct formation pathways for UDGs, and that this sub-population is simply an extreme low surface brightness extension of the underlying dwarf galaxy population. The root cause of their diffuse stellar distributions remains unclear, but the formation mechanism appears to only impact the distribution of stars (and potentially dark matter), without strongly impacting the distribution of neutral gas, the overall stellar mass, or the number of GCs.
Tongue cancer is a common oral cavity malignancy that originates in the mouth and throat. Much effort has been invested in improving its diagnosis, treatment, and management. Surgical removal, chemotherapy, and radiation therapy remain the major treatment for tongue cancer. The survival of patients determines the treatment effect. Previous studies have identified certain survival and risk factors based on descriptive statistics, ignoring the complex, nonlinear relationship among clinical and demographic variables. In this study, we utilize five cutting-edge machine learning models and clinical data to predict the survival of tongue cancer patients after treatment. Five-fold cross-validation, bootstrap analysis, and permutation feature importance are applied to estimate and interpret model performance. The prognostic factors identified by our method are consistent with previous clinical studies. Our method is accurate, interpretable, and thus useable as additional evidence in tongue cancer treatment and management.
In this paper, we consider testing the martingale difference hypothesis for high-dimensional time series. Our test is built on the sum of squares of the element-wise max-norm of the proposed matrix-valued nonlinear dependence measure at different lags. To conduct the inference, we approximate the null distribution of our test statistic by Gaussian approximation and provide a simulation-based approach to generate critical values. The asymptotic behavior of the test statistic under the alternative is also studied. Our approach is nonparametric as the null hypothesis only assumes the time series concerned is martingale difference without specifying any parametric forms of its conditional moments. As an advantage of Gaussian approximation, our test is robust to the cross-series dependence of unknown magnitude. To the best of our knowledge, this is the first valid test for the martingale difference hypothesis that not only allows for large dimension but also captures nonlinear serial dependence. The practical usefulness of our test is illustrated via simulation and a real data analysis. The test is implemented in a user-friendly R-function.
We report the discovery of HAT-P-17b,c, a multi-planet system with an inner transiting planet in a short-period, eccentric orbit and an outer planet in a 4.8 yr, nearly circular orbit. The inner planet, HAT-P-17b, transits the bright V = 10.54 early K dwarf star GSC 2717-00417, with an orbital period P = 10.338523 +/- 0.000009 d, orbital eccentricity e = 0.346 +/- 0.007, transit epoch T_c = 2454801.16945 +/- 0.00020, and transit duration 0.1691 +/- 0.0009 d. HAT-P-17b has a mass of 0.530 +/- 0.018 M_J and radius of 1.010 +/- 0.029 R_J yielding a mean density of 0.64 +/- 0.05 g cm^-3. This planet has a relatively low equilibrium temperature in the range 780-927 K, making it an attractive target for follow-up spectroscopic studies. The outer planet, HAT-P-17c, has a significantly longer orbital period P_2 = 1797^+58_-89 d and a minimum mass m_2 sin i_2 = 1.4^+1.1_-0.4 M_J. The orbital inclination of HAT-P-17c is unknown as transits have not been observed and may not be present. The host star has a mass of 0.86 +/- 0.04 M_Sun, radius of 0.84 +/- 0.02, effective temperature 5246 +/- 80 K, and metallicity [Fe/H] = 0.00 +/- 0.08. HAT-P-17 is the second multi-planet system detected from ground-based transit surveys.
Future microwave sky surveys will have the sensitivity to detect the kinematic Sunyaev-Zeldovich signal from moving galaxy clusters, thus providing a direct measurement of their line-of-sight peculiar velocity. We show that cluster peculiar velocity statistics applied to foreseeable surveys will put significant constraints on fundamental cosmological parameters. We consider three statistical quantities that can be constructed from a cluster peculiar velocity catalog: the probability density function, the mean pairwise streaming velocity, and the pairwise velocity dispersion. These quantities are applied to an envisioned data set which measures line-of-sight cluster velocities with normal errors of 100 km/s for all clusters with masses larger than $10^{14}$ solar masses over a sky area of up to 5000 square degrees. A simple Fisher matrix analysis of this survey shows that the normalization of the matter power spectrum and the dark energy equation of state can be constrained to better than 10 percent, and the Hubble constant and the primordial power spectrum index can be constrained to a few percent, independent of any other cosmological observations. We also find that the current constraint on the power spectrum normalization can be improved by more than a factor of two using data from a 400 square degree survey and WMAP third-year priors. We also show how the constraints on cosmological parameters changes if cluster velocities are measured with normal errors of 300 km/s.
We develop a Bayesian hierarchical semiparametric model for phenomena related to time series of counts. The main feature of the model is its capability to learn a latent pattern of heterogeneity in the distribution of the process innovation rates, which are softly clustered through time with the help of a Dirichlet process placed at the top of the model hierarchy. The probabilistic forecasting capabilities of the model are put to test in the analysis of crime data in Pittsburgh, with favorable results.
We demonstrate quantum entanglement of two trapped atomic ion qubits using a sequence of ultrafast laser pulses. Unlike previous demonstrations of entanglement mediated by the Coulomb interaction, this scheme does not require confinement to the Lamb-Dicke regime and can be less sensitive to ambient noise due to its speed. To elucidate the physics of an ultrafast phase gate, we generate a high entanglement rate using just 10 pulses, each of $\sim20$ ps duration, and demonstrate an entangled Bell-state with $(76\pm1)$% fidelity. These results pave the way for entanglement operations within a large collection of qubits by exciting only local modes of motion.
In mixed-initiative co-creation tasks, wherein a human and a machine jointly create items, it is important to provide multiple relevant suggestions to the designer. Quality-diversity algorithms are commonly used for this purpose, as they can provide diverse suggestions that represent salient areas of the solution space, showcasing designs with high fitness and wide variety. Because generated suggestions drive the search process, it is important that they provide inspiration, but also stay aligned with the designer's intentions. Additionally, often many interactions with the system are required before the designer is content with a solution. In this work, we tackle these challenges with an interactive constrained MAP-Elites system that leverages emitters to learn the preferences of the designer and then use them in automated steps. By learning preferences, the generated designs remain aligned with the designer's intent, and by applying automatic steps, we generate more solutions per user interaction, giving a larger number of choices to the designer and thereby speeding up the search. We propose a general framework for preference-learning emitters (PLEs) and apply it to a procedural content generation task in the video game Space Engineers. We built an interactive application for our algorithm and performed a user study with players.
Cells perform directed motion in response to external stimuli that they detect by sensing the environment with their membrane protrusions. In particular, several biochemical and biophysical cues give rise to tactic migration in the direction of their specific targets. This defines a multi-cue environment in which cells have to sort and combine different, and potentially competitive, stimuli. We propose a non-local kinetic model for cell migration in presence of two external factors both influencing cell polarization: contact guidance and chemotaxis. We propose two different sensing strategies and we analyze the two resulting models by recovering the appropriate macroscopic limit in different regimes, in order to see how the size of the cell, with respect to the variation of both external fields, influences the overall behavior. Moreover, we integrate numerically the kinetic transport equation in a two-dimensional setting in order to investigate qualitatively various scenarios.
The design of gate drivers is an important topic in power converter topologies that can help reduce switching losses and increase power density. Gate driving techniques that offer zero-voltage switching and/or zero current switching have recently been successfully proposed for different modular multilevel converters such as the cascaded H bridge. Previous papers on other multilevel converters such as the multi-active bridge, however, do not sufficiently assess the topics of gate driver design for this topology. This work presents a novel isolated gate driver architecture tailored to the multi-active bridge topology. Zero voltage switching is then achieved using two multi-winding transformers. The advantages of the proposed topology are not only a reduction of switching losses but also reduced component count. The topology is evaluated on a prototype using experimental results. It was shown using simulation and experiments that the proposed topology has a high efficiency while providing compact power packaging. Especially for converters with many levels, the proposed topology is therefore advantageous compared to existing solutions.
In this paper we give an algorithm for solving a main case of the conjugacy problem in the braid groups. We also prove that half-twists satisfy a special root property which allows us to reduce the solution for the conjugacy problem in half-twists into the free group. Using this algorithm one is able to check conjugacy of a given braid to one of E. Artin's generators in any power, and compute its root. Moreover, the braid element which conjugates a given half-twist to one of E. Artin's generators in any power can be restored. The result is applicable to calculations of braid monodromy of branch curves and verification of Hurwitz equivalence of braid monodromy factorizations, which are essential in order to determine braid monodromy type of algebraic surfaces and symplectic 4-manifolds.
The paper has been withdrawn by the author, due to a fatal error. A horse stumbles that has four legs.
In the presence of P,T-violating interactions, the exchange of axion-like particles between electrons and nucleons in atoms and molecules induces electric dipole moments (EDMs) of atoms and molecules. We perform calculations of such axion-exchange-induced atomic EDMs using the relativistic Hartree-Fock-Dirac method including electron core polarisation (RPA) corrections. We present analytical estimates to explain the dependence of these induced atomic EDMs on the axion mass and atomic parameters. From the experimental bounds on the EDMs of atoms and molecules, including $^{133}$Cs, $^{205}$Tl, $^{129}$Xe, $^{199}$Hg, $^{171}$Yb$^{19}$F, $^{180}$Hf$^{19}$F$^+$ and $^{232}$Th$^{16}$O, we constrain the P,T-violating scalar-pseudoscalar nucleon-electron and electron-electron interactions mediated by a generic axion-like particle of arbitrary mass. Our limits improve on existing laboratory bounds from other experiments by many orders of magnitude for $m_a \gtrsim 10^{-2}~\textrm{eV}$. We also place constraints on CP violation in certain types of relaxion models.
We prove an asymptotic formula for the Fourier transform of the arithmetic surface measure associated to the Waring--Goldbach problem and provide several applications, including bounds for discrete spherical maximal functions along the primes and distribution results such as ergodic theorems.
Deep neural networks currently deliver promising results for microscopy image cell segmentation, but they require large-scale labelled databases, which is a costly and time-consuming process. In this work, we relax the labelling requirement by combining self-supervised with semi-supervised learning. We propose the prediction of edge-based maps for self-supervising the training of the unlabelled images, which is combined with the supervised training of a small number of labelled images for learning the segmentation task. In our experiments, we evaluate on a few-shot microscopy image cell segmentation benchmark and show that only a small number of annotated images, e.g. 10% of the original training set, is enough for our approach to reach similar performance as with the fully annotated databases on 1- to 10-shots. Our code and trained models is made publicly available
It ia well-known that surface plasmon wave propagates along a straight line, but this common sense was broken by the artificial curved light - plasmon Airy beam. In this paper we introduce a new class of curved surface plasmon wave - the photonic hook plasmon. It propagates along wavelength scaled curved trajectory with radius less than surface plasmon polariton wavelength, and can exist despite the strong energy dissipation at the metal surface.
In this paper we extend a previous investigation by us regarding an iterative construction of irreducible polynomials over finite fields of odd characteristic. In particular, we show how it is possible to iteratively construct irreducible polynomials by means of two families of transforms, which we call the $Q_k$ and $\hat{Q}_k$-transforms, related to certain degree two isogenies over elliptic curves, which split the multiplication-by-$2$ map.
Industrial computing devices, in particular cyber-physical, real-time and safety-critical systems, focus on reacting to external events and the need to cooperate with other devices to create a functional system. They are often implemented with languages that focus on a simple, local description of how a component reacts to external input data and stimuli. Despite the trend in modern software architectures to structure systems into largely independent components, the remaining interdependencies still create rich behavioural dynamics even for small systems. Standard and industrial programming approaches do usually not model or extensively describe the global properties of an entire system. Although a large number of approaches to solve this dilemma have been suggested, it remains a hard and error-prone task to implement systems with complex interdependencies correctly. We introduce multiple coupled finite state machines (McFSMs), a novel mechanism that allows us to model and manage such interdependencies. It is based on a consistent, well-structured and simple global description. A sound theoretical foundation is provided, and associated tools allow us to generate efficient low-level code in various programming languages using model-driven techniques. We also present a domain specific language to express McFSMs and their connections to other systems, to model their dynamic behaviour, and to investigate their efficiency and correctness at compile-time.
The real numbers, it is taught at universities, correspond to our idea of a continuum, although the hyperreal numbers are located ``in between'' the real numbers. The number $x + dx$, where $dx$ should be an infinitesimal number and $x$ real, is infinitesimally close to $x$ but ``infinitely'' far away from all other real numbers. Analogously: If $f'(x_0)$ and $f(x_0)$ are given for a differentiable function $f:\mathbb{R}\rightarrow\mathbb{R}$ at $x_0\in\mathbb{R}$, we can not determine $f(x)$ at {\em any} point $x\in \mathbb{R}$ different from $x_0$. These points seem to be ``infinitely'' far away. That is one conceptual problem of solving differential equations in numerical mathematics. In this article, we will present a numerical algorithm to solve very simple initial value problems. However, the change of paradigm is, that we will not ``leave'' the point $x_0$. Solving ordinary differential equations is like searching for ``recipes'' $f$. Instead of trying to find these recipes for values $x\in\mathbb{R}$, we will learn them from special relations in the ``monad'' of $x_0$.
We consider the challenging problem of tracking multiple objects using a distributed network of sensors. In the practical setting of nodes with limited field of views (FoVs), computing power and communication resources, we develop a novel distributed multi-object tracking algorithm. To accomplish this, we first formalise the concept of label consistency, determine a sufficient condition to achieve it and develop a novel \textit{label consensus approach} that reduces label inconsistency caused by objects' movements from one node's limited FoV to another. Second, we develop a distributed multi-object fusion algorithm that fuses local multi-object state estimates instead of local multi-object densities. This algorithm: i) requires significantly less processing time than multi-object density fusion methods; ii) achieves better tracking accuracy by considering Optimal Sub-Pattern Assignment (OSPA) tracking errors over several scans rather than a single scan; iii) is agnostic to local multi-object tracking techniques, and only requires each node to provide a set of estimated tracks. Thus, it is not necessary to assume that the nodes maintain multi-object densities, and hence the fusion outcomes do not modify local multi-object densities. Numerical experiments demonstrate our proposed solution's real-time computational efficiency and accuracy compared to state-of-the-art solutions in challenging scenarios. We also release source code at https://github.com/AdelaideAuto-IDLab/Distributed-limitedFoV-MOT for our fusion method to foster developments in DMOT algorithms.
Higher-order accuracy (order of $k+1$ in the $L^2$ norm) is one of the well known beneficial properties of the discontinuous Galerkin (DG) method. Furthermore, many studies have demonstrated the superconvergence property (order of $2k+1$ in the negative norm) of the semi-discrete DG method. One can take advantage of this superconvergence property by post-processing techniques to enhance the accuracy of the DG solution. A popular class of post-processing techniques to raise the convergence rate from order $k+1$ to order $2k+1$ in the $L^2$ norm is the Smoothness-Increasing Accuracy-Conserving (SIAC) filtering. In addition to enhancing the accuracy, the SIAC filtering also increases the inter-element smoothness of the DG solution. The SIAC filtering was introduced for the DG method of the linear hyperbolic equation by Cockburn et al. in 2003. Since then, there are many generalizations of the SIAC filtering have been proposed. However, the development of SIAC filtering has never gone beyond the framework of using spline functions (mostly B-splines) to construct the filter function. In this paper, we first investigate the general basis function (beyond the spline functions) that can be used to construct the SIAC filter. The studies of the general basis function relax the SIAC filter structure and provide more specific properties, such as extra smoothness, etc. Secondly, we study the basis functions' distribution and propose a new SIAC filter called compact SIAC filter that significantly reduces the original SIAC filter's support size while preserving (or even improving) its ability to enhance the accuracy of the DG solution. We show that the proofs of the new SIAC filters' ability to extract the superconvergence and provide numerical results to confirm the theoretical results and demonstrate the new finding's good numerical performance.
We propose a unifying rheological framework for dense suspensions of non-Brownian spheres, predicting the onsets of particle friction and particle inertia as distinct shear thickening mechanisms, while capturing quasistatic and soft particle rheology at high volume fractions and shear rates respectively. Discrete element method simulations that take suitable account of hydrodynamic and particle-contact interactions corroborate the model predictions, demonstrating both mechanisms of shear thickening, and showing that they can occur concurrently with carefully selected particle surface properties under certain flow conditions. Microstructural transitions associated with frictional shear thickening are presented. We find very distinctive divergences of both the microstructural and dynamic variables with respect to volume fraction in the thickened and non-thickened states.
Fluctuations of conserved quantum numbers are associated with the corresponding susceptibilities because of the symmetry of the system. The underlying fact is that these fluctuations as defined through the static correlators become identical to the direct calculation of these susceptibilities defined through the thermodynamic derivatives, due to the fluctuation-dissipation theorem. Through a rigorous exercise we explicitly show that a diagrammatic calculation of the static correlators associated with the conserved quark number fluctuations and the corresponding susceptibilities are possible in case of mean field theories, if the implicit dependence of the mean fields on the quark chemical potential are taken into account appropriately. As an aside we also give an analytical prescription for obtaining the implicit dependence of the mean fields on the quark chemical potential.
A novel framework for statistical learning is introduced which combines ideas from regularization and ensembling. This framework is applied to learn an ensemble of logistic regression models for high-dimensional binary classification. In the new framework the models in the ensemble are learned simultaneously by optimizing a multi-convex objective function. To enforce diversity between the models the objective function penalizes overlap between the models in the ensemble. Measures of diversity in classifier ensembles are used to show how our method learns the ensemble by exploiting the accuracy-diversity trade-off for ensemble models. In contrast to other ensembling approaches, the resulting ensemble model is fully interpretable as a logistic regression model, asymptotically consistent, and at the same time yields excellent prediction accuracy as demonstrated in an extensive simulation study and gene expression data applications. The models found by the proposed ensemble methodology can also reveal alternative mechanisms that can explain the relationship between the predictors and the response variable. An open-source compiled software library implementing the proposed method is briefly discussed.
IGR~J19149+1036 is a high mass X-ray binary detected by INTEGRAL in 2011 in the hard X-ray domain. We have analyzed the BAT survey data of the first 103 months of the Swift mission detecting this source at a significance level of ~30 standard deviations. The timing analysis on the long term BAT light curve reveals the presence of a strong sinusoidal intensity modulation of 22.25+/- 0.05 d, that we interpret as the orbital period of this binary system. A broad band (0.3-150 keV) spectral analysis was performed combining the BAT spectrum and the XRT spectra from the pointed follow up observations. The spectrum is adequately modeled with an absorbed power law with a high energy cutoff at ~24 keV and an absorption cyclotron feature at ~31 keV. Correcting for the gravitational redshift, the inferred magnetic field at the neutron star surface is B_surf ~ 3.6 x 10^12 gauss.
We analyze the implications of a Higgs discovery on possible ``new-physics'' scenarios, for $m_H$ up to $\sim 700$ GeV. For this purpose we critically review lower and upper limits on the Higgs mass in the SM and in the MSSM, respectively. Furthermore, we discuss the general features of possible ``heavy'' ($m_H \gsim 2 m_Z$) Higgs scenarios by means of a simple heavy-fermion condensate model.
Studying the accretion process in very low-mass objects has important implications for understanding their formation mechanism. Many nearby late-M dwarfs that have previously been identified in the field are in fact young brown dwarf members of nearby young associations. Some of them are still accreting. They are therefore excellent targets for further studies of the accretion process in the very low-mass regime at different stages. We aim to search for accreting young brown dwarf candidates in a sample of 85 nearby late-M dwarfs. Using photometric data from DENIS, 2MASS, and WISE, we constructed the spectral energy distribution of the late-M dwarfs based on BT-Settl models to detect infrared excesses. We then searched for lithium and H$\alpha$ emission in candidates that exhibit infrared excesses to confirm their youth and the presence of accretion. Among the 85 late-M dwarfs, only DENIS-P J1538317$-$103850 (M5.5) shows strong infrared excesses in WISE bands. The detection of lithium absorption in the M5.5 dwarf and its Gaia trigonometric parallax indicate an age of $\sim$1 Myr and a mass of 47 $M_{\rm J}$. The H$\alpha$ emission line in the brown dwarf shows significant variability that indicates sporadic accretion. This 1 Myr-old brown dwarf also exhibits intense accretion bursts with accretion rates of up to $10^{-7.9}$$M_{\odot}$ yr$^{-1}$. Our detection of sporadic accretion in one of the youngest brown dwarfs might imply that sporadic accretion at early stages could play an important role in the formation of brown dwarfs. Very low-mass cores would not be able to accrete enough material to become stars, and thus they end up as brown dwarfs.
Measuring the similarity between data points often requires domain knowledge, which can in parts be compensated by relying on unsupervised methods such as latent-variable models, where similarity/distance is estimated in a more compact latent space. Prevalent is the use of the Euclidean metric, which has the drawback of ignoring information about similarity of data stored in the decoder, as captured by the framework of Riemannian geometry. We propose an extension to the framework of variational auto-encoders allows learning flat latent manifolds, where the Euclidean metric is a proxy for the similarity between data points. This is achieved by defining the latent space as a Riemannian manifold and by regularising the metric tensor to be a scaled identity matrix. Additionally, we replace the compact prior typically used in variational auto-encoders with a recently presented, more expressive hierarchical one---and formulate the learning problem as a constrained optimisation problem. We evaluate our method on a range of data-sets, including a video-tracking benchmark, where the performance of our unsupervised approach nears that of state-of-the-art supervised approaches, while retaining the computational efficiency of straight-line-based approaches.
Recurrent neural networks (RNNs) are known to be difficult to train due to the gradient vanishing and exploding problems and thus difficult to learn long-term patterns and construct deep networks. To address these problems, this paper proposes a new type of RNNs with the recurrent connection formulated as Hadamard product, referred to as independently recurrent neural network (IndRNN), where neurons in the same layer are independent of each other and connected across layers. Due to the better behaved gradient backpropagation, IndRNN with regulated recurrent weights effectively addresses the gradient vanishing and exploding problems and thus long-term dependencies can be learned. Moreover, an IndRNN can work with non-saturated activation functions such as ReLU (rectified linear unit) and be still trained robustly. Different deeper IndRNN architectures, including the basic stacked IndRNN, residual IndRNN and densely connected IndRNN, have been investigated, all of which can be much deeper than the existing RNNs. Furthermore, IndRNN reduces the computation at each time step and can be over 10 times faster than the commonly used Long short-term memory (LSTM). Experimental results have shown that the proposed IndRNN is able to process very long sequences and construct very deep networks. Better performance has been achieved on various tasks with IndRNNs compared with the traditional RNN, LSTM and the popular Transformer.
We present the biggest up-to-date sample of edge-on galaxies with B/PS bulges and X-structures. The sample was prepared using images from the DESI Legacy catalogue and contains about 2000 galaxies. To find suitable candidates in catalogue, we made the assumption that the residues (original images minus model) of galaxies with B/PS bulges should exhibit a characteristic X-shape. Galaxies with such features were selected by eye and then used as input data for a neural network training, which was applied to a bigger sample of edge-on galaxies. Using the available data and the photometric models from the literature, we investigated the observational and statistical properties of the sample created. Comparing the $B/D$ ratios for galaxies with and without B/PS bulges, we found that the $B/D$ ratio for galaxies from our sample is statistically higher, with typical values in the range $\approx 0.2-0.5$ depending on the decomposition procedure. We studied how the opening angles $\varphi$ of the X-structure and the length of its rays are distributed in the formed sample and found them to be consistent with previous measurements and predictions from $N$-body models, e.g. $\varphi \gtrsim 25~\deg$, but measured here for a much larger number of galaxies. We found a sharp increase in the B/PS bulge fraction for stellar masses $\log M_{\star} \gtrsim 10.4$, but for edge-on galaxies, which complements the results of previous works. The sample can be used in future work to test various bar models and their relationship with B/PS bulges, as well as to study their stability and evolution.
We study the nonlinear Schr\"odinger equation with an arbitrary real potential $V(x)\in (L^1+L^\infty)(\Gamma)$ on a star graph $\Gamma$. At the vertex an interaction occurs described by the generalized Kirchhoff condition with strength $-\gamma<0$. We show the existence of ground states $\varphi_{\omega}(x)$ as minimizers of the action functional on the Nehari manifold under additional negativity and decay conditions on $V(x)$. Moreover, for $V(x)=-\dfrac{\beta}{x^\alpha}$, in the supercritical case, we prove that the standing waves $e^{i\omega t}\varphi_{\omega}(x)$ are orbitally unstable in $H^{1}(\Gamma)$ when $\omega$ is large enough. Analogous result holds for an arbitrary $\gamma\in\mathbb{R}$ when the standing waves have symmetric profile.
For more than two decades it has been known that any compact Stein surface (of real dimension four) admits a compatible Lefschetz fibration over a two-disk. More recently, Giroux and Pardon have generalized this result by giving a complex geometric proof for the existence of compatible Lefschetz fibrations on Stein domains of any even dimension. As a preparatory step in proving the former, Akbulut and Ozbagci have shown that there exist infinitely many pairwise non-equivalent Lefschetz fibrations on the four-ball by using a result of Lyon constructing fibrations on the complements of (p,q)-torus links in the three-sphere. In this paper, we first extend this result to obtain compatible Lefschetz fibrations on the six-ball whose pages are (p, q, 2)-Brieskorn varieties, and then construct a compatible 'relative' Lefschetz fibrations on any Stein domain (of dimension six) which admit a certain ('admissible') 'relative Stein pair' structure. In particular, we provide a purely topological proof for the existence of Lefschetz fibrations on specific 6-dimensional Stein domains.
In this thesis we develop generalized versions of the Chung-Feller theorem for lattice paths constrained in the half plane. The beautiful cycle method which was developed by Devoretzky and Motzkin as a means to prove the ballot problem is modified and applied to generalize the classical Chung-Feller theorem. We use Lagrange inversion to derive the generalized formulas. For the generating function proof we study various ways of decomposing lattice paths. We also show some results related to equidistribution properties in terms of Narayana and Catalan generating functions. We then develop generalized Chung-Feller theorems for Motzkin and Schroeder paths. Finally we study generalized paths and the analogue of the Chung-Feller theorem for them.
This paper studies the light-tailed asymptotics of the stationary tail probability vectors of a Markov chain of M/G/1 type. Almost all related studies have focused on the typical case, where the transition block matrices in the non-boundary levels have a dominant impact on the decay rate of the stationary tail probability vectors and their decay is aperiodic. In this paper, we study not only the typical case but also atypical cases such that the stationary tail probability vectors decay periodically and/or their decay rate is determined by the tail distribution of jump sizes from the boundary level. We derive light-tailed asymptotic formulae for the stationary tail probability vectors by locating the dominant poles of the generating function of the sequence of those vectors. Further we discuss the positivity of the dominant terms of the obtained asymptotic formulae.
A method of solving the time-dependent Schr\"odinger equation is presented, in which a finite region of space is treated explicitly, with the boundary conditions for matching the wave-functions on to the rest of the system replaced by an embedding term added on to the Hamiltonian. This time-dependent embedding term is derived from the Fourier transform of the energy-dependent embedding potential, which embeds the time-independent Schr\"odinger equation. Results are presented for a one-dimensional model of an atom in a time-varying electric field, the surface excitation of this model atom at a jellium surface in an external electric field, and the surface excitation of a bulk state.
The pioneer work of Krim and Widom unveiled the origin of the viscous nature of friction at the atomic scale. This generated extensive experimental and theoretical activity. However, fundamental questions remain open like the relation between sliding friction and the topology of the substrate, as well as the dependence on the temperature of the contact surface. Here we present results, obtained using molecular dynamics, for the phononic friction coefficient ($\eta_{ph}$) for a one dimensional model of an adsorbate-substrate interface. Different commensuration relations between adsorbate and substrate are investigated as well as the temperature dependence of $\eta_{ph}$. In all the cases we studied $\eta_{ph}$ depends quadratically on the substrate corrugation amplitude, but is a non-trivial function of the commensuration ratio between substrate and adsorbate. The most striking result is a deep and wide region of small values of $\eta_{ph}$ for substrate-adsorbate commensuration ratios between $\approx 0.6-0.9$. Our results shed some light on contradictory results for the relative size of phononic and electronic friction found in the literature.
We show that if $K$ is an L-space twisted torus knot $T^{l,m}_{p,pk \pm 1}$ with $p \ge 2$, $k \ge 1$, $m \ge 1$ and $1 \le l \le p-1$, then the fundamental group of the $3$-manifold obtained by $\frac{r}{s}$-surgery along $K$ is not left-orderable whenever $\frac{r}{s} \ge 2 g(K) -1$, where $g(K)$ is the genus of $K$.
Some well-known and less well-known or new notions related to group actions are surveyed. Some of these notions are used to generalize affine spaces. Actions are seen as functions with values in transformation monoids
Let $ \chi $ be a virtual (generalized) character of a finite group $ G $ and $ L=L(\chi)$ be the image of $ \chi $ on $ G-\lbrace 1 \rbrace $. The pair $ (G, \chi) $ is said to be sharp of type $ L $ if $|G|=\prod _{ l \in L} (\chi(1) - l) $. If the principal character of $G$ is not an irreducible constituent of $\chi$, the pair $(G,\chi)$ is called normalized. In this paper, we first provide some counterexamples to a conjecture that was proposed by Cameron and Kiyota in $1988$. This conjecture states that if $(G,\chi)$ is sharp and $|L|\geq 2$, then the inner product $(\chi,\chi)_G$ is uniquely determined by $ L $. We then prove that this conjecture is true in the case that $(G,\chi) $ is normalized, $\chi$ is a character of $ G $, and $ L $ contains at least an irrational value.
APN functions play a fundamental role in cryptography against attacks on block ciphers. Several families of quadratic APN functions have been proposed in the recent years, whose construction relies on the existence of specific families of polynomials. A key question connected with such constructions is to determine whether such APN functions exist for infinitely many dimensions or not. In this paper we consider a family of functions recently introduced by Li et al. in 2021 showing that for any dimension $m\geq 3$ there exists an APN function belonging to such a family. Our main result is proved by a combination of different techniques arising from both algebraic varieties over finite fields connected with linearized permutation rational functions and {partial vector space partitions}, together with investigations on the kernels of linearized polynomials.
The coupling of a dilaton to the $SU(2)$-Yang-Mills field leads to interesting non-perturbative static spherically symmetric solutions which are studied by mixed analitical and numerical methods. In the abelian sector of the theory there are finite-energy magnetic and electric monopole solutions which saturate the Bogomol'nyi bound. In the nonabelian sector there exist a countable family of globally regular solutions which are purely magnetic but have zero Yang-Mills magnetic charge. Their discrete spectrum of energies is bounded from above by the energy of the abelian magnetic monopole with unit magnetic charge. The stability analysis demonstrates that the solutions are saddle points of the energy functional with increasing number of unstable modes. The existence and instability of these solutions are "explained" by the Morse-theory argument recently proposed by Sudarsky and Wald.
Interaction graphs, such as those recording emails between individuals or transactions between institutions, tend to be sparse yet structured, and often grow in an unbounded manner. Such behavior can be well-captured by structured, nonparametric edge-exchangeable graphs. However, such exchangeable models necessarily ignore temporal dynamics in the network. We propose a dynamic nonparametric model for interaction graphs that combine the sparsity of the exchangeable models with dynamic clustering patterns that tend to reinforce recent behavioral patterns. We show that our method yields improved held-out likelihood over stationary variants, and impressive predictive performance against a range of state-of-the-art dynamic interaction graph models.
This addendum provides results complementary to those obtained in [J. Phys. G49, 055001 (2022)]. Specifically, an equivalent form of the relation, which binds together the "spacelike" kernel functions for the hadronic vacuum polarization contribution to the muon anomalous magnetic moment $a^{\text{HVP}}_{\mu}$, is obtained. It is shown that the infrared limiting value of the "spacelike" and "timelike" kernel functions, which enter the representations for $a^{\text{HVP}}_{\mu}$ involving the Adler function and the $R$-ratio, is identical to the corresponding QED contribution to the muon anomalous magnetic moment of the preceding order in the electromagnetic coupling. The next-to-leading order contributions $a^{\text{HVP}(3b)}_{\mu}$ (which includes the leptonic and hadronic insertions) and $a^{\text{HVP}(3c)}_{\mu}$ (which includes the double hadronic insertion), are studied. The three kernel functions appearing in the representations for $a^{\text{HVP}(3b)}_{\mu}$, which involve the hadronic vacuum polarization function, Adler function, and the $R$-ratio, are presented for the cases of the electron and $\tau$-lepton loop insertions.
We report on experimental determinations of the temperature field in the interior (bulk) of turbulent Rayleigh-Benard convection for a cylindrical sample with aspect ratio (diameter over height) of 0.50, both in the classical and in the ultimate state. The Prandtl number was close to 0.8. We find a "logarithmic layer" in which the temperature varies as A*ln(z/L) + B with the distance z from the bottom plate of the sample. The amplitude A varies with radial position r. In the classical state these results are in good agreement with direct numerical simulations (DNS); in the ultimate state there are as yet no DNS. A close analogy between the temperature field in the classical state and the "Law of the Wall" for the time-averaged down-stream velocity in shear flow is discussed.
The dynamic nature of system gives rise to dynamical features of epidemic spreading, such as oscillation and bistability. In this paper, by studying the epidemic spreading in growing networks, in which susceptible nodes may adaptively break the connections with infected ones yet avoid getting isolated, we reveal a new phenomenon - \emph{epidemic reemergence}, where the number of infected nodes is incubated at a low level for a long time and then bursts up for a short time. The process may repeat several times before the infection finally vanishes. Simulation results show that all the three factors, namely the network growth, the connection breaking and the isolation avoidance, are necessary for epidemic reemergence to happen. We present a simple theoretical analysis to explain the process of reemergence in detail. Our study may offer some useful insights helping explain the phenomenon of repeated epidemic explosions.