text
stringlengths
6
128k
We compare two different implementations of fault-tolerant entangling gates on logical qubits. In one instance, a twelve-qubit trapped-ion quantum computer is used to implement a non-transversal logical CNOT gate between two five qubit codes. The operation is evaluated with varying degrees of fault tolerance, which are provided by including quantum error correction circuit primitives known as flagging and pieceable fault tolerance. In the second instance, a twenty-qubit trapped-ion quantum computer is used to implement a transversal logical CNOT gate on two [[7,1,3]] color codes. The two codes were implemented on different but similar devices, and in both instances, all of the quantum error correction primitives, including the determination of corrections via decoding, are implemented during runtime using a classical compute environment that is tightly integrated with the quantum processor. For different combinations of the primitives, logical state fidelity measurements are made after applying the gate to different input states, providing bounds on the process fidelity. We find the highest fidelity operations with the color code, with the fault-tolerant SPAM operation achieving fidelities of 0.99939(15) and 0.99959(13) when preparing eigenstates of the logical X and Z operators, which is higher than the average physical qubit SPAM fidelities of 0.9968(2) and 0.9970(1) for the physical X and Z bases, respectively. When combined with a logical transversal CNOT gate, we find the color code to perform the sequence--state preparation, CNOT, measure out--with an average fidelity bounded by [0.9957,0.9963]. The logical fidelity bounds are higher than the analogous physical-level fidelity bounds, which we find to be [0.9850,0.9903], reflecting multiple physical noise sources such as SPAM errors for two qubits, several single-qubit gates, a two-qubit gate and some amount of memory error.
We present a path planning framework that takes into account the human's safety perception in the presence of a flying robot. The framework addresses two objectives: (i) estimation of the uncertain parameters of the proposed safety perception model based on test data collected using Virtual Reality (VR) testbed, and (ii) offline optimal control computation using the estimated safety perception model. Due to the unknown factors in the human tests data, it is not suitable to use standard regression techniques that minimize the mean squared error (MSE). We propose to use a Hidden Markov model (HMM) approach where human's attention is considered as a hidden state to infer whether the data samples are relevant to learn the safety perception model. The HMM approach improved log-likelihood over the standard least squares solution. For path planning, we use Bernstein polynomials for discretization, as the resulting path remains within the convex hull of the control points, providing guarantees for deconfliction with obstacles at low computational cost. An example of optimal trajectory generation using the learned human model is presented. The optimal trajectory generated using the proposed model results in reasonable safety distance from the human. In contrast, the paths generated using the standard regression model have undesirable shapes due to overfitting. The example demonstrates that the HMM approach has robustness to the unknown factors compared to the standard MSE model.
We investigate the photonic bands of an atomic Bose-Einstein condensate with a triangular vortex lattice. Index contrast between the vortex cores and the bulk of the condensate is achieved through the enhancement of the index via atomic coherence. Frequency dependent dielectric function is used in the calculations of the bands, resulting in photonic band gap widths of a few MHz.
The O vacancy (Ov) formation energy, $E_\textrm{Ov}$, is an important property of a metal-oxide, governing its performance in applications such as fuel cells or heterogeneous catalysis. These defects are routinely studied with density functional theory (DFT). However, it is well-recognized that standard DFT formulations (e.g. the generalized gradient approximation) are insufficient for modeling the Ov, requiring higher levels of theory. The embedded cluster method offers a promising approach to compute $E_\textrm{Ov}$ accurately, giving access to all electronic structure methods. Central to this approach is the construction of quantum(-mechanically treated) clusters placed within suitable embedding environments. Unfortunately, current approaches to constructing the quantum clusters either require large system sizes, preventing application of high-level methods, or require significant manual input, preventing investigations of multiple systems simultaneously. In this work, we present a systematic and general quantum cluster design protocol that can determine small converged quantum clusters for studying the Ov in metal-oxides with accurate methods such as local coupled cluster with single, double and perturbative triple excitations [CCSD(T)]. We apply this protocol to study the Ov in the bulk and surface planes of rutile TiO2 and rocksalt MgO, producing the first accurate and well-converged determinations of $E_\textrm{Ov}$ with this method. These reference values are used to benchmark exchange-correlation functionals in DFT and we find that all studied functionals underestimate $E_\textrm{Ov}$, with the average error decreasing along the rungs of Jacob's ladder. This protocol is automatable for high-throughput calculations and can be generalized to study other point defects or adsorbates.
In this paper, we treat an open problem related to the number of periodic orbits of Hamiltonian diffeomorphisms on closed symplectic manifolds, so-called generic Conley conjecture. Generic Conley conjecture states that generically Hamiltonian diffeomorphisms have infinitely many simple contractible periodic orbits. We prove generic Conley conjecture for very wide classes of symplectic manifolds.
In 1999, Pitman and Stanley introduced the polytope bearing their name along with a study of its faces, lattice points, and volume. The Pitman-Stanley polytope is well-studied due to its connections to probability, parking functions, the generalized permutahedra, and flow polytopes. Its lattice points correspond to plane partitions of skew shape with entries 0 and 1. Pitman and Stanley remarked that their polytope can be generalized so that lattice points correspond to plane partitions of skew shape with entries $0,1, \ldots , m$. Since then, this generalization has been untouched. We study this generalization and show that it can also be realized as a flow polytope of a grid graph. We give multiple characterizations of its vertices in terms of plane partitions of skew shape and integer flows. For a fixed skew shape, we show that the number of vertices of this polytope is a polynomial in $m$ whose leading term, in certain cases, counts standard Young tableaux of a skew shifted shape. Moreover, we give formulas for the number of faces, as well as generating functions for the number of vertices.
As a step towards the structure theory of Lie algebras in symmetric monoidal categories we establish results involving the Killing form. The proper categorical setting for discussing these issues are symmetric ribbon categories.
We theoretically analyze equilibrium fluctuations of persistent current (PC) in nanorings. We demonstrate that these fluctuations persist down to zero temperature provided the current operator does not commute with the total Hamiltonian of the system. For a model of a quantum particle on a ring we explicitly evaluate PC noise power which has the form of sharp peaks at frequencies set by the corresponding interlevel distances. In rings with many conducting channels a much smoother and broader PC noise spectrum is expected. A specific feature of PC noise is that its spectrum can be tuned by an external magnetic flux indicating the presence of quantum coherence in the system.
This paper studies problems on locally stopping distributed consensus algorithms over networks where each node updates its state by interacting with its neighbors and decides by itself whether certain level of agreement has been achieved among nodes. Since an individual node is unable to access the states of those beyond its neighbors, this problem becomes challenging. In this work, we first define the stopping problem for generic distributed algorithms. Then, a distributed algorithm is explicitly provided for each node to stop consensus updating by exploring the relationship between the so-called local and global consensus. Finally, we show both in theory and simulation that its effectiveness depends both on the network size and the structure.
Entity relation extraction consists of two sub-tasks: entity recognition and relation extraction. Existing methods either tackle these two tasks separately or unify them with word-by-word interactions. In this paper, we propose HIORE, a new method for unified entity relation extraction. The key insight is to leverage the high-order interactions, i.e., the complex association among word pairs, which contains richer information than the first-order word-by-word interactions. For this purpose, we first devise a W-shape DNN (WNet) to capture coarse-level high-order connections. Then, we build a heuristic high-order graph and further calibrate the representations with a graph neural network (GNN). Experiments on three benchmarks (ACE04, ACE05, SciERC) show that HIORE achieves the state-of-the-art performance on relation extraction and an improvement of 1.1~1.8 F1 points over the prior best unified model.
Modern AI techniques open up ever-increasing possibilities for autonomous vehicles, but how to appropriately verify the reliability of such systems remains unclear. A common approach is to conduct safety validation based on a predefined Operational Design Domain (ODD) describing specific conditions under which a system under test is required to operate properly. However, collecting sufficient realistic test cases to ensure comprehensive ODD coverage is challenging. In this paper, we report our practical experiences regarding the utility of data simulation with deep generative models for scenario-based ODD validation. We consider the specific use case of a camera-based rail-scene segmentation system designed to support autonomous train operation. We demonstrate the capabilities of semantically editing railway scenes with deep generative models to make a limited amount of test data more representative. We also show how our approach helps to analyze the degree to which a system complies with typical ODD requirements. Specifically, we focus on evaluating proper operation under different lighting and weather conditions as well as while transitioning between them.
We present a theory of the thermal Hall effect in insulating quantum magnets, where the heat current is totally carried by charge-neutral objects such as magnons and spinons. Two distinct types of thermal Hall responses are identified. For ordered magnets, the intrinsic thermal Hall effect for magnons arises when certain conditions are satisfied for the lattice geometry and the underlying magnetic order. The other type is allowed in a spin liquid which is a novel quantum state since there is no order even at zero temperature. For this case, the deconfined spinons contribute to the thermal Hall response due to Lorentz force. These results offer a clear experimental method to prove the existence of the deconfined spinons via a thermal transport phenomenon.
Venus has no known satellites, but has four known co-orbitals: (322756) 2001 CK32, 2002 VE68, 2012 XE133, and 2013 ND15. Here, we present numerical evidence suggesting that 2015 WZ12 is a possible Venus co-orbital; it might have been until recently a transient Trojan. Follow-up observations of this target in the near future will be difficult, though.
We introduce the theory of non-linear cosmological perturbations using the correspondence limit of the Schr\"odinger equation. The resulting formalism is equivalent to using the collisionless Boltzman (or Vlasov) equations which remain valid during the whole evolution, even after shell crossing. Other formulations of perturbation theory explicitly break down at shell crossing, e.g. Eulerean perturbation theory, which describes gravitational collapse in the fluid limit. This paper lays the groundwork by introducing the new formalism, calculating the perturbation theory kernels which form the basis of all subsequent calculations. We also establish the connection with conventional perturbation theories, by showing that third order tree level results, such as bispectrum, skewness, cumulant correlators, three-point function are exactly reproduced in the appropriate expansion of our results. We explicitly show that cumulants up to N=5 predicted by Eulerian perturbation theory for the dark matter field $\delta$ are exactly recovered in the corresponding limit. A logarithmic mapping of the field naturally arises in the Schr\"odinger context, which means that tree level perturbation theory translates into (possibly incomplete) loop corrections for the conventional perturbation theory. We show that the first loop correction for the variance is $\sigma^2 = \sigma_L^2+ (-1.14+n)\sigma_L^4$ for a field with spectral index $n$. This yields 1.86 and 0.86 for $n=-3,-2$ respectively, and to be compared with the exact loop order corrections 1.82, and 0.88. Thus our tree-level theory recovers the dominant part of first order loop corrections of the conventional theory, while including (partial) loop corrections to infinite order in terms of $\delta$.
Identifying the factors that determine academic performance is an essential part of educational research. Existing research indicates that class attendance is a useful predictor of subsequent course achievements. The majority of the literature is, however, based on surveys and self-reports, methods which have well-known systematic biases that lead to limitations on conclusions and generalizability as well as being costly to implement. Here we propose a novel method for measuring class attendance that overcomes these limitations by using location and bluetooth data collected from smartphone sensors. Based on measured attendance data of nearly 1,000 undergraduate students, we demonstrate that early and consistent class attendance strongly correlates with academic performance. In addition, our novel dataset allows us to determine that attendance among social peers was substantially correlated ($>$0.5), suggesting either an important peer effect or homophily with respect to attendance.
Let $R$ be the ring of $S$-integers in a number field $K$. Let $\mathcal{B}=\{\beta, \beta^{\ast}\}$ be the multi-set of roots of a nonzero quadratic polynomial over $R$. There are varieties $V(\mathcal{B})_{N,k}$ defined over $R$ parametrizing periodic continued fractions $[b_1,\ldots , b_N,\overline{a_1,\ldots ,a_k}]$ for $\beta$ or $\beta^{\ast}$. We study the $R$-points on these varieties, finding contrasting behavior according to whether groups of units are infinite or not. If $R$ is the rational integers or the ring of integers in an imaginary quadratic field, we prove that the $R$-points of $V(\mathcal{B})_{N,k}$ are not Zariski dense. On the other hand, suppose that $\beta\not\in K\cup\{\infty\}$, $R^\times$ is infinite, and that there are infinitely many units in the (left) order $R_\beta$ of $\beta R+R\subseteq K(\beta)$ with norm to $K$ equal to $(-1)^k$. Then we prove that the $R$-points on $V(\mathcal{B})_{1,k}$ are Zariski dense for $k\geq 8$ and the $R$-points on $V(\mathcal{B})_{0,k}$ are Zariski dense for $k\geq 9$. We also prove that $V(\mathcal{B})_{1,k}$ and $V(\mathcal{B})_{0,k}$ are $K$-rational irreducible varieties for $k$ sufficiently large.
The aim of this article is to investigate the well-posedness, stability and convergence of solutions to the time-dependent Maxwell's equations for electric field in conductive media in continuous and discrete settings. The situation we consider would represent a physical problem where a subdomain is emerged in a homogeneous medium, characterized by constant dielectric permittivity and conductivity functions. It is well known that in these homogeneous regions the solution to the Maxwell's equations also solves the wave equation which makes calculations very efficient. In this way our problem can be considered as a coupling problem for which we derive stability and convergence analysis. A number of numerical examples validate theoretical convergence rates of the proposed stabilized explicit finite element scheme.
This is a pedagogical review of the recent observational data obtained from type Ia supernova surveys that support the accelerating expansion of the universe. The methods for the analysis of the data are reviewed and some of the theoretical implications obtained from their analysis are discussed.
We report yes-go and no-go results on consistent cross-couplings for a collection of gravitons. Motivated by the search of theories where multiplets of massless spin-two fields cross-interact, we look for all the consistent deformations of a positive sum of Pauli-Fierz actions. We also investigate the problem of deforming a (positive and negative) sum of linearized Weyl gravity actions and show explicitly that there exists multi-Weyl-graviton theories. As the single-graviton Weyl theory, these theories do not have an energy bounded from below.
We study the problem of computing optimal prices for a version of the Product-Mix auction with budget constraints. In contrast to the ``standard'' Product-Mix auction, the objective is to maximize revenue instead of social welfare. We prove correctness of an algorithm proposed by Paul Klemperer and DotEcon which is sufficiently efficient in smaller markets.
Web crawling, snowball sampling, and respondent-driven sampling (RDS) are three types of network sampling techniques used to contact individuals in hard-to-reach populations. This paper studies these procedures as a Markov process on the social network that is indexed by a tree. Each node in this tree corresponds to an observation and each edge in the tree corresponds to a referral. Indexing with a tree (instead of a chain) allows for the sampled units to refer multiple future units into the sample. In survey sampling, the design effect characterizes the additional variance induced by a novel sampling strategy. If the design effect is some value $DE$, then constructing an estimator from the novel design makes the variance of the estimator $DE$ times greater than it would be under a simple random sample with the same sample size $n$. Under certain assumptions on the referral tree, the design effect of network sampling has a critical threshold that is a function of the referral rate $m$ and the clustering structure in the social network, represented by the second eigenvalue of the Markov transition matrix, $\lambda_2$. If $m < 1/\lambda_2^2$, then the design effect is finite (i.e. the standard estimator is $\sqrt{n}$-consistent). However, if $m > 1/\lambda_2^2$, then the design effect grows with $n$ (i.e. the standard estimator is no longer $\sqrt{n}$-consistent). Past this critical threshold, the standard error of the estimator converges at the slower rate of $n^{\log_m \lambda_2}$. The Markov model allows for nodes to be resampled; computational results show that the findings hold in without-replacement sampling. To estimate confidence intervals that adapt to the correct level of uncertainty, a novel resampling procedure is proposed. Computational experiments compare this procedure to previous techniques.
A promising approach to overcome decoherence in quantum computing schemes is to perform active quantum error correction using topology. Topological subsystem codes incorporate both the benefits of topological and subsystem codes, allowing for error syndrome recovery with only 2-local measurements in a two-dimensional array of qubits. We study the error threshold for topological subsystem color codes under very general external noise conditions. By transforming the problem into a classical disordered spin model, we estimate using Monte Carlo simulations that topological subsystem codes have an optimal error tolerance of 5.5(2)%. This means there is ample space for improvement in existing error-correcting algorithms that typically find a threshold of approximately 2%.
Understanding how the birthplace of stars affects planet-forming discs is important for a comprehensive theory of planet formation. Most stars are born in dense star-forming regions where the external influence of other stars, particularly the most massive stars, will affect the survival and enrichment of their planet-forming discs. Simulations suggest that stellar dynamics play a central role in regulating how external feedback affects discs, but comparing models to observations requires an estimate of the initial stellar density in star-forming regions. Structural analyses constrain the amount of dynamical evolution a star-forming region has experienced; regions that maintain substructure and do not show mass segregation are likely dynamically young, and therefore close to their birth density. In this paper, we present a structural analysis of two clusters in the Carina Nebula, Tr14 and Tr16. We show that neither cluster shows evidence for mass segregation or a centrally concentrated morphology, suggesting that both regions are dynamically young. This allows us to compare to simulations from Nicholson et al. (2019) who predict disc survival rates in star-forming regions of different initial densities. The surviving disc fractions in Tr14 and Tr16 are consistent with their predictions (both are $\sim 10$%), supporting a growing body of evidence that the star-forming environment plays an important role in the survival and enrichment of protoplanetary discs.
We use the Suita conjecture (now a theorem) to prove that for any domain $\Omega \subset \mathbb{C}$ its Bergman kernel $K(\cdot, \cdot)$ satisfies $K(z_0, z_0) = \hbox{Volume}(\Omega)^{-1}$ for some $z_0 \in \Omega$ if and only if $\Omega$ is either a disk minus a (possibly empty) closed polar set or $\mathbb{C}$ minus a (possibly empty) closed polar set. When $\Omega$ is bounded with $C^{\infty}$-boundary, we provide a simple proof of this using the zero set of the Szeg\"o kernel. Finally, we show that this theorem fails to hold in $\mathbb{C}^n$ for $n > 1$ by constructing a bounded complete Reinhardt domain (with algebraic boundary) which is strongly convex and not biholomorphic to the unit ball $\mathbb{B}^n \subset \mathbb{C}^n$.
In this paper, we introduce a new extended version of the shallow water equations with surface tension which is skew-symmetric with respect to the L2 scalar product and allows for large gradients of fluid height. This result is a generalization of the results published by P. Noble and J.-P. Vila in [SIAM J. Num. Anal. (2016)] and by D. Bresch, F. Couderc, P. Noble and J.P. Vila in [C.R. Acad. Sciences Paris (2016)] which are restricted to quadratic forms of the capillary energy respectively in the one dimensional and two dimensional setting.This is also an improvement of the results by J. Lallement, P. Villedieu et al. published in [AIAA Aviation Forum 2018] where the augmented version is not skew-symetric with respect to the L2 scalar product. Based on this new formulation, we propose a new numerical scheme and perform a nonlinear stability analysis.Various numerical simulations of the shallow water equations are presented to show differences between quadratic (w.r.t the gradient of the height) and general surface tension energy when high gradients of the fluid height occur.
Two main physical mechanisms are used to explain supernova explosions: thermonuclear explosion of a white dwarf(Type Ia) and core collapse of a massive star (Type II and Type Ib/Ic). Type Ia supernovae serve as distance indicators that led to the discovery of the accelerating expansion of the Universe. The exact nature of their progenitor systems however remain unclear. Radio emission from the interaction between the explosion shock front and its surrounding CSM or ISM provides an important probe into the progenitor star's last evolutionary stage. No radio emission has yet been detected from Type Ia supernovae by current telescopes. The SKA will hopefully detect radio emission from Type Ia supernovae due to its much better sensitivity and resolution. There is a 'supernovae rate problem' for the core collapse supernovae because the optically dim ones are missed due to being intrinsically faint and/or due to dust obscuration. A number of dust-enshrouded optically hidden supernovae should be discovered via SKA1-MID/survey, especially for those located in the innermost regions of their host galaxies. Meanwhile, the detection of intrinsically dim SNe will also benefit from SKA1. The detection rate will provide unique information about the current star formation rate and the initial mass function. A supernova explosion triggers a shock wave which expels and heats the surrounding CSM and ISM, and forms a supernova remnant (SNR). It is expected that more SNRs will be discovered by the SKA. This may decrease the discrepancy between the expected and observed numbers of SNRs. Several SNRs have been confirmed to accelerate protons, the main component of cosmic rays, to very high energy by their shocks. This brings us hope of solving the Galactic cosmic ray origin's puzzle by combining the low frequency (SKA) and very high frequency (Cherenkov Telescope Array: CTA) bands' observations of SNRs.
This paper will review a new technique of detecting companion stars in LMXBs and X-ray transients in outburst using the Bowen fluorescence lines at 4634-4640 Angs. These lines are very efficiently reprocessed in the atmospheres of the companion stars, and thereby provide estimates of the K2 velocities and mass functions. The method has been applied to Sco X-1, X1822-371 and GX339-4 which, in the latter case, provides the first dynamical evidence for the presence of an accreting black hole. Preliminary results from a VLT campaign on V801 Ara, V926 Sco and XTE J1814-338 are also presented.
We demonstrate a systematic method for solving the Hamilton-Jacobi equation for general relativity with the inclusion of matter fields. The generating functional is expanded in a series of spatial gradients. Each term is manifestly invariant under reparameterizations of the spatial coordinates (``gauge-invariant''). At each order we solve the Hamiltonian constraint using a conformal transformation of the 3-metric as well as a line integral in superspace. This gives a recursion relation for the generating functional which then may be solved to arbitrary order simply by functionally differentiating previous orders. At fourth order in spatial gradients, we demonstrate solutions for irrotational dust as well as for a scalar field. We explicitly evolve the 3-metric to the same order. This method can be used to derive the Zel'dovich approximation for general relativity.
The goal of this paper is to lay the foundations for a combinatorial study, via orthogonal functions and intertwining operators, of category O for the rational Cherednik algebra of type G(r,p,n). As a first application, we give a self-contained and elementary proof of the analog for the groups G(r,p,n), with r>1, of Gordon's theorem (previously Haiman's conjecture) on the diagonal coinvariant ring. We impose no restriction on p; the result for p<r has been proved by Vale using a technique analogous to Gordon's. Because of the combinatorial application to Haiman's conjecture, the paper is logically self-contained except for standard facts about complex reflection groups. The main results should be accessible to mathematicians working in algebraic combinatorics who are unfamiliar with the impressive range of ideas used in Gordon's proof of his theorem.
A multisymplectic setting for classical field theories subjected to non-holonomic constraints is presented. The infinite dimensional setting in the space of Cauchy data is also given.
The Hamiltonian structures of several hybrid kinetic-fluid models are identified explicitly, upon considering collisionless Vlasov dynamics for the hot particles interacting with a bulk fluid. After presenting different pressure-coupling schemes for an ordinary fluid interacting with a hot gas, the paper extends the treatment to account for a fluid plasma interacting with an energetic ion species. Both current-coupling and pressure-coupling MHD schemes are treated extensively. In particular, pressure-coupling schemes are shown to require a transport-like term in the Vlasov kinetic equation, in order for the Hamiltonian structure to be preserved. The last part of the paper is devoted to studying the more general case of an energetic ion species interacting with a neutralizing electron background (hybrid Hall-MHD). Circulation laws and Casimir functionals are presented explicitly in each case.
We describe regularized methods for image reconstruction and focus on the question of hyperparameter and instrument parameter estimation, i.e. unsupervised and myopic problems. We developed a Bayesian framework that is based on the \post density for all unknown quantities, given the observations. This density is explored by a Markov Chain Monte-Carlo sampling technique based on a Gibbs loop and including a Metropolis-Hastings step. The numerical evaluation relies on the SPIRE instrument of the Herschel observatory. Using simulated and real observations, we show that the hyperparameters and instrument parameters are correctly estimated, which opens up many perspectives for imaging in astrophysics.
We prove a formula for the speed of distance stationary random sequences. A particular case is the classical formula for the largest Lyapunov exponent of an i.i.d. product of two by two matrices in terms of a stationary measure on projective space. We apply this result to Poisson-Delaunay random walks on Riemannian symmetric spaces. In particular, we obtain sharp estimates for the asymptotic behavior of the speed of hyperbolic Poisson-Delaunay random walks when the intensity of the Poisson point process goes to zero. This allows us to prove that a dimension drop phenomena occurs for the harmonic measure associated to these random walks. With the same technique we give examples of co-compact Fuchsian groups for which the harmonic measure of the simple random walk has dimension less than one.
The Vietnamese Power system is expected to expand considerably in upcoming decades. However, pathways towards higher shares of renewables ought to be investigated. In this work, we investigate a highly renewable Vietnamese power system by jointly optimising the expansion of renewable generation facilities and the transmission grid. We show that in the cost-optimal case, highest amounts of wind capacities are installed in southern Vietnam and solar photovoltaics (PV) in central Vietnam. In addition, we show that transmission has the potential to reduce levelised cost of electricity by approximately 10%.
Though the truths of logic and pure mathematics are objective and independent of any contingent facts or laws of nature, our knowledge of these truths depends entirely on our knowledge of the laws of physics. Recent progress in the quantum theory of computation has provided practical instances of this, and forces us to abandon the classical view that computation, and hence mathematical proof, are purely logical notions independent of that of computation as a physical process. Henceforward, a proof must be regarded not as an abstract object or process but as a physical process, a species of computation, whose scope and reliability depend on our knowledge of the physics of the computer concerned.
Convective processes are crucial in shaping exoplanetary atmospheres but are computationally expensive to simulate directly. A novel technique of simulating moist convection on tidally locked exoplanets is to use a global 3D model with a stretched mesh. This allows us to locally refine the model resolution to 4.7 km and resolve fine-scale convective processes without relying on parameterizations. We explore the impact of mesh stretching on the climate of a slowly rotating TRAPPIST-1e-like planet, assuming it is 1:1 tidally locked. In the stretched-mesh simulation with explicit convection, the climate is 5 K colder and 25% drier than that in the simulations with parameterized convection (with both stretched and quasi-uniform meshes)}. This is due to the increased cloud reflectivity - because of an increase of low-level cloudiness - and exacerbated by the diminished greenhouse effect due to less water vapor. At the same time, our stretched-mesh simulations reproduce the key characteristics of the global climate of tidally locked rocky exoplanets, without any noticeable numerical artifacts. Our methodology opens an exciting and computationally feasible avenue for improving our understanding of 3D mixing in exoplanetary atmospheres. Our study also demonstrates the feasibility of a global stretched mesh configuration for LFRic-Atmosphere, the next-generation Met Office climate and weather model.
The novel weak-value-amplification (WVA) scheme of precision metrology is deeply rooted in the quantum nature of destructive interference between the pre- and post-selection states. And, an alternative version, termed as joint WVA (JWVA), which employs the difference-signal from the post-selection accepted and rejected results, has been found possible to achieve even better sensitivity (two orders of magnitude higher) under some technical limitations (e.g. misalignment errors). In this work, after erasing the quantum coherence, we analyze the difference-signal amplification (DSA) technique, which serves as a classical counterpart of the JWVA, and show that similar amplification effect can be achieved. We obtain a simple expression for the amplified signal, carry out characterization of precision, and point out the optimal working regime. We also discuss how to implement the post-selection of a classical mixed state. The proposed classical DSA technique holds similar technical advantages of the JWVA and may find interesting applications in practice.
In this note, we consider the observational constraints on some cosmological models by using the 307 Union type Ia supernovae (SNIa), the 32 calibrated Gamma-Ray Bursts (GRBs) at $z>1.4$, the updated shift parameter $R$ from WMAP 5-year data (WMAP5), and the distance parameter $A$ of the measurement of the baryon acoustic oscillation (BAO) peak in the distribution of SDSS luminous red galaxies with the updated scalar spectral index $n_s$ from WMAP5. The tighter constraints obtained here update the ones obtained previously in the literature.
This paper describes a method for scheduling the events of a switched system to achieve an optimal performance. The approach has guarantees on convergence and computational complexity that parallel derivative-based iterative optimization but in the infinite dimensional, integer constrained setting of mode scheduling. In comparison to methods relying on mixed integer programming, the presented approach does not require a priori discretizations of time or state. Furthermore, in comparison to embedding and relaxation methods, every iteration of the algorithm returns a dynamically feasible solution. A large class of problems call for optimal mode scheduling. This paper considers a vehicle tracking problem and a high dimensional multimachine power network synchronization problem. For the power network example, both single horizon and receding horizon approaches prevent instability of the network, and the receding horizon approach does so at near real-time speeds on a single processor.
In this work, we have examined how the multi-vacancy defects induced in the horizontal direction change the energetics and the electronic structure of semiconducting Single-Walled Carbon Nanotubes (SWCNTs). The electronic structure of SWCNTs is computed for each deformed configuration by means of real space, Order(N) Tight Binding Molecular Dynamic (O(N) TBMD) simulations. Energy band gap is obtained in real space through the behavior of electronic density of states (eDOS) near the Fermi level. Vacancies can effectively change the energetics and hence the electronic structure of SWCNTs. In this study, we choose three different kinds of semiconducting zigzag SWCNTs and determine the band gap modifications. We have selected (12,0), (13,0) and (14,0) zigzag SWCNTs according to n (mod 3) = 0, n (mod 3) = 1 and n (mod 3) = 2 classification. (12,0) SWCNT is metallic in its pristine state. The application of vacancies opens the electronic band gap and it goes up to 0.13 eV for a di- vacancy defected tube. On the other hand (13,0) and (14,0) SWCNTs are semiconductors with energy band gap values of 0.44 eV and 0.55 eV in their pristine state, respectively. Their energy band gap values decrease to 0.07 eV and 0.09 eV when mono-vacancy defects are induced in their horizontal directions. Then the di-vacancy defects open the band gap again. So in both cases, the semiconducting-metallic - semiconducting transitions occur. It is also shown that the band gap modification exhibits irreversible characteristics, which means that band gap values of the nanotubes do not reach their pristine values with increasing number of vacancies.
There are several works \cite{De} (and \cite{St}), \cite{En}, \cite{Co} and \cite{Va} enumerating four-dimensional parallelotopes. In this work we give a new enumeration showing that any four-dimensional parallelotope is either a zonotope or the Minkowski sum of a zonotope with the regular 24-cell $\{3,4,3\}$. Each zonotopal parallelotope is the Minkowski sum of segments whose generating vectors form a unimodular system. There are exactly 17 four-dimensional unimodular systems. Hence, there are 17 four-dimensional zonotopal parallelotopes. Other 35 four-dimensional parallelotopes are: the regular 24-cell $\{3,4,3\}$ and 34 sums of the regular parallelotope with non-zero zonotopal parallelotopes. For the nontrivial enumerating of the 34 sums we use a theorem discribing necessary and sufficient conditions when the Minkowski sum of a parallelotope with a segment is a parallelotope.
Coherent superpositions of the 49s and 48s Rydberg states of cold Rb atoms were studied near the surface of an atom chip. The superpositions were created and manipulated using microwaves resonant with the two-photon 49s-48s transition. Coherent behavior was observed using Rabi flopping, Ramsey sequences, spin-echo and spin-locking. These results are discussed in the context of Rydberg atoms as electric field noise sensors. We consider the coherence of systems quadratically coupled to noise fields with 1/f^k power spectral densities (k \approx 1).
In this paper we argue that when gauge invariance is taken into consideration, there is no consistent geometric framework of Finsler class that can accommodate Randers type spaces. In this context, an alternative non-Finslerian framework for Randers spacetimes compatible with gauge invariance is introduced.
Individually addressed Er$^{3+}$ ions in solid-state hosts are promising resources for quantum repeaters, because of their direct emission in the telecom band and compatibility with silicon photonic devices. While the Er$^{3+}$ electron spin provides a spin-photon interface, ancilla nuclear spins could enable multi-qubit registers with longer storage times. In this work, we demonstrate coherent coupling between the electron spin of a single Er$^{3+}$ ion and a single $I=1/2$ nuclear spin in the solid-state host crystal, which is a fortuitously located proton ($^1$H). We control the nuclear spin using dynamical decoupling sequences applied to the electron spin, implementing one- and two-qubit gate operations. Crucially, the nuclear spin coherence time exceeds the electron coherence time by several orders of magnitude, because of its smaller magnetic moment. These results provide a path towards combining long-lived nuclear spin quantum registers with telecom-wavelength emitters for long-distance quantum repeaters.
A subset $S$ of a group $G$ invariably generates $G$ if, when each element of $S$ is replaced by an arbitrary conjugate, the resulting set generates $G.$ An invariable generating set $X$ of $G$ is called minimal if no proper subset of $X$ invariably generates $G.$ We will address several questions related to the behaviour of minimal invariable generating sets of a finite group.
A domain wall separating two oppositely magnetized regions in a ferromagnetic semiconductor exhibits, under appropriate conditions, strongly nonlinear I-V characteristics similar to those of a p-n diode. We study these characteristics as functions of wall width and temperature. As the width increases or the temperature decreases, direct tunneling between the majority spin bands decreases the effectiveness of the diode. This has important implications for the zero-field quenched resistance of magnetic semiconductors and for the design of a recently proposed spin transistor.
New results are reported from a measurement of $\pi^0$ electroproduction near threshold using the $p(e,e^{\prime} p)\pi^0$ reaction. The experiment was designed to determine precisely the energy dependence of $s-$ and $p-$wave electromagnetic multipoles as a stringent test of the predictions of Chiral Perturbation Theory (ChPT). The data were taken with an electron beam energy of 1192 MeV using a two-spectrometer setup in Hall A at Jefferson Lab. For the first time, complete coverage of the $\phi^*_{\pi}$ and $\theta^*_{\pi}$ angles in the $p \pi^0$ center-of-mass was obtained for invariant energies above threshold from 0.5 MeV up to 15 MeV. The 4-momentum transfer $Q^2$ coverage ranges from 0.05 to 0.155 (GeV/c)$^2$ in fine steps. A simple phenomenological analysis of our data shows strong disagreement with $p-$wave predictions from ChPT for $Q^2>0.07$ (GeV/c)$^2$, while the $s-$wave predictions are in reasonable agreement.
In 2007 an interesting phenomenon was discovered: a thread of water, the so-called water bridge (WB), can hang between two glass beakers filled with deionized water if voltage is applied to them. We analyze the available explanations of the WB stability and propose a completely different one: the force that supports the WB is the surface tension of water and the role of electric field is not to allow the WB to reduce its surface energy by means of breaking into separate drops.
Accurate short-term traffic prediction plays a pivotal role in various smart mobility operation and management systems. Currently, most of the state-of-the-art prediction models are based on graph neural networks (GNNs), and the required training samples are proportional to the size of the traffic network. In many cities, the available amount of traffic data is substantially below the minimum requirement due to the data collection expense. It is still an open question to develop traffic prediction models with a small size of training data on large-scale networks. We notice that the traffic states of a node for the near future only depend on the traffic states of its localized neighborhoods, which can be represented using the graph relational inductive biases. In view of this, this paper develops a graph network (GN)-based deep learning model LocaleGN that depicts the traffic dynamics using localized data aggregating and updating functions, as well as the node-wise recurrent neural networks. LocaleGN is a light-weighted model designed for training on few samples without over-fitting, and hence it can solve the problem of few-sample traffic prediction. The proposed model is examined on predicting both traffic speed and flow with six datasets, and the experimental results demonstrate that LocaleGN outperforms existing state-of-the-art baseline models. It is also demonstrated that the learned knowledge from LocaleGN can be transferred across cities. The research outcomes can help to develop light-weighted traffic prediction systems, especially for cities lacking historically archived traffic data.
We theoretically study electromagnetically induced transparency (EIT) in reflection spectra of V-type system at the gas-solid interface. In addition to a narrow dip arising from the EIT effect, we find the other particular saturation effect induced by pump field, which does not exist in $\Lambda$ or $\Xi$ -type system reflection spectra. The saturation effect only induces an intensity decrement in the reflection spectra, and there is no influence on the narrow dip arising from the EIT effect. We detailedly calculate and analyze the dependence of V-type system reflection spectra on probe field intensity, pump field intensity, coherent decay rate, and the initial population after the collision between atoms and the interface.
Magnons are viewed as local deviations from the ordered state. Usually, the spin magnetic moment of magnons is considered. In a 3D-confined structure of a magnetic insulator with magnetodipolar mode (MDM) oscillations, an orbital angular momentum (OAM) as well as a spin angular momentum (SAM) can be observed along a static magnetic field. In such a confined structure as quasi-2D ferrite disk, energy levels of MDM oscillations are quantized. Quantum confinement is characterized by a half-integer internal OAM, which is also associated with a circulating energy flow. The observation of MDM resonances in the 3D-confined structure of a magnetic insulator is due to the interaction of two subsystems: ferromagnetic and electric polarization orders. The coupling states of these two concurrent orders, caused by OAMs, are considered as magnetoelectric (ME) states. The fields in the vicinity of MDM resonators are characterized by simultaneous violation of time reversal and inversion symmetry. This plays a significant role in the problems of strong light-matter interaction regime and quantum atmosphere. The analysis of SAM and OAM in 3D-confined magnetic insulators becomes very important for the realization of ME meta-atomic structures. Current interest lies in considering such artificial systems as subwavelength ME quantum emitters of electromagnetic radiation.
We define dynamic striping as the ability to assign different Lustre striping characteristics to contiguous segments of a file as it grows. In this paper, we evaluate the effects of dynamic striping using a watermark-based strategy where the stripe count or width is increased once a file's size exceeds one of the chosen watermarks. To measure the performance of this strategy we used a modified version of the IOR benchmark, a netflow analysis workload, and the blastn algorithm from NCBI BLAST. The results indicate that dynamic striping is beneficial to tasks with unpredictable data file size and large sequential reads, but are less conclusive for workloads with significant random read phases.
The effective action for gravity at high curvatures is likely to contain higher derivative terms. These corrections may have profound consequences for the singularity structure of space-time and for early Universe cosmology. In this contribution, recent work is reviewed which demonstrates that it is possible to construct a class of effective gravitational actions for which all solutions with sufficient symmetries have limited curvature and are nonsingular. Near the limiting curvature, the coupling between matter and gravity goes to zero and in this sense the theory is asymptotically free.
A single-sort continuum Curie-Weiss system of interacting particles is studied. The particles are placed in the space $\mathbb{R}^d$ divided into congruent cubic cells. For a region $V\subset \mathbb{R}^d$ consisting of $N\in \mathbb{N}$ cells, every two particles contained in $V$ attract each other with intensity $J_1/N$. The particles contained in the same cell are subjected to binary repulsion with intensity $J_2>J_1$. For fixed values of the temperature, the interaction intensities, and the chemical potential the thermodynamic phase is defined as a probability measure on the space of occupation numbers of cells, determined by a condition typical of Curie-Weiss theories. It is proved that the half-plane $J_1\,\times\,$\textit{chemical potential} contains phase coexistence points at which there exist two thermodynamic phases of the system. An equation of state for this system is obtained.
Using the 3D smoothed particle hydrodynamics code PHANTOM, we investigate the evolution of the orbital properties of massive black hole binaries embedded in massive discs where gravitational instabilities (GIs) triggered by the disc self-gravity are the only source of angular momentum transport. In particular, we investigate the evolution of binaries with different initial eccentricities $e_0=0.05,\,0.5,\,0.8$ and mass ratios $q=0.1,\,0.3,\,0.9$. Our simulations suggest that there might not be a unique value of critical eccentricity. We find initially more eccentric binaries to tend to higher asymptotic eccentricity values than more circular ones. This implies that there is a range of critical eccentricity values, that depends on the initial condition of the system. In particular, we find the width of this range to be narrower for more unequal binaries. We furthermore measure preferential accretion onto our binaries, finding more accretion onto the primary only for mass ratio $q=0.3$ and eccentricity $e=0.8$. We discuss how this might have implications for the amplitude of the gravitational wave background detected by Pulsar Timing Arrays (PTA) experiments. We finally measure the corresponding value of the viscosity parameter $\alpha$ in our simulations and discuss how this depends on the binary properties.
We describe a phase transition for long-range entanglement in a three-dimensional cluster state affected by noise. The partially decohered state is modeled by the thermal state of a suitable Hamiltonian. We find that the temperature at which the entanglement length changes from infinite to finite is nonzero. We give an upper and lower bound to this transition temperature.
An outstanding problem in gravitation theory and relativistic astrophysics today is to understand the final outcome of an endless gravitational collapse. Such a continual collapse would take place when stars more massive than few times the mass of the sun collapse under their own gravity on exhausting their nuclear fuel. According to the general theory of relativity, this results either in a black hole, or a naked singularity- which can communicate with faraway observers in the universe. While black holes are (almost) being detected and are increasingly used to model high energy astrophysical phenomena, naked singularities have turned into a topic of active discussion, aimed at understanding their structure and implications. Recent developments here are reviewed, indicating future directions.
We use representation theory to construct integral formulas for solutions to the quantum Toda lattice in general type. This result generalizes work of Givental for SL(n)/B in a uniform way to arbitrary type and can be interpreted as a kind of mirror theorem for the full flag variety G/B. We also prove the existence of a totally positive critical point of the 'superpotential' in every mirror fiber.
We study the nature of tunneling phase time for various quantum mechanical structures such as networks and rings having potential barriers in their arms. We find the generic presence of Hartman effect, with superluminal velocities as a consequence, in these systems. In quantum networks it is possible to control the `super arrival' time in one of the arms by changing the parameters on another arm which is spatially separated from it. This is yet another quantum nonlocal effect. Negative time delays (time advancement) and `ultra Hartman effect' with negative saturation times have been observed in some parameter regimes. In presence and absence of Aharonov-Bohm (AB) flux quantum rings show Hartman effect. We obtain the analytical expression for the saturated phase time. In the opaque barrier regime this is independent of even the AB flux thereby generalizing the Hartman effect. We also briefly discuss the concept of "space collapse or space destroyer" by introducing a free space in between two barriers covering the ring. Further we show in presence of absorption the reflection phase time exhibits Hartman effect in contrast to the transmission phase time.
Autonomous mobile robot competitions judge based on a robot's ability to quickly and accurately navigate the game field. This means accurate localization is crucial for creating an autonomous competition robot. Two common localization methods are odometry and computer vision landmark detection. Odometry provides frequent velocity measurements, while landmark detection provides infrequent position measurements. The state can also be predicted with a physics model. These three types of localization can be "fused" to create a more accurate state estimate using an Extended Kalman Filter (EKF). The EKF is a nonlinear full-state estimator that approximates the state estimate with the lowest covariance error when given the sensor measurements, the model prediction, and their variances. In this paper, we demonstrate the effectiveness of the EKF by implementing it on a 4-wheel mecanum-drive robot simulation. The position and velocity accuracy of fusing together various combinations of these three data sources are compared. We also discuss the assumptions and limitations of an EKF.
In real-world video surveillance applications, person re-identification (ReID) suffers from the effects of occlusions and detection errors. Despite recent advances, occlusions continue to corrupt the features extracted by state-of-art CNN backbones, and thereby deteriorate the accuracy of ReID systems. To address this issue, methods in the literature use an additional costly process such as pose estimation, where pose maps provide supervision to exclude occluded regions. In contrast, we introduce a novel Holistic Guidance (HG) method that relies only on person identity labels, and on the distribution of pairwise matching distances of datasets to alleviate the problem of occlusion, without requiring additional supervision. Hence, our proposed student-teacher framework is trained to address the occlusion problem by matching the distributions of between- and within-class distances (DCDs) of occluded samples with that of holistic (non-occluded) samples, thereby using the latter as a soft labeled reference to learn well separated DCDs. This approach is supported by our empirical study where the distribution of between- and within-class distances between images have more overlap in occluded than holistic datasets. In particular, features extracted from both datasets are jointly learned using the student model to produce an attention map that allows separating visible regions from occluded ones. In addition to this, a joint generative-discriminative backbone is trained with a denoising autoencoder, allowing the system to self-recover from occlusions. Extensive experiments on several challenging public datasets indicate that the proposed approach can outperform state-of-the-art methods on both occluded and holistic datasets
The incomplete beta function is an important special function in statistics. In modern theory of hypergeometric functions, we regard hypergeometric functions as pairings of twisted cycles and twisted cocycles. However, the incomplete beta function cannot be understood in this scheme; in other words, the domain of the integration is not cycle (incomplete). We will generalize the theory of A-hypergeometric systems for incomplete functions. We give a general study as well as a detailed study on an incomplete Gauss hypergeometric function.
We analyze the pattern formation in systems of active particles with chiral forces in the context of pedestrian dynamics. To describe the interparticle interactions, we use the standard social force model and supplement it with a new type of force that reflects chirality. We perform numerical simulations of two pedestrian flows moving in opposite directions along a corridor. We observe two dynamic phase transitions that occur for varying number densities of particles and strengths of the chirality force: one from disordered motion to multi-lane motion and another from multi-lane to two-lane motion. We develop a qualitative theory that describes the demarcation lines for these phase transitions in the phase diagram chirality-density. The results of our analysis agree fairly well with the simulation data. A comparison with previously reported experimental data has been provided. Our findings may find applications in urban and transportation-planning problems.
Laboratory formation of large carbon clusters with m C atoms where m could be up to few thousand, in carbonaceous plasma, has been studied by using an especially designed ion source. Carbon is introduced into the glow discharge plasma by sputtering of the graphite electrode. Soot dominated plasma is created whose constituents are carbon clusters. It produces and recycles cluster containing plasma. Regenerative sooting plasma creates the environment in which the entire spectrum of clusters that contain the linear chains, rings and fullerenes. Velocity spectra of the extracted clusters have been measured with an ExB filter. These spectra indicate and identify the mechanisms operating in the soot.
Existing 3D pose datasets of object categories are limited to generic object types and lack of fine-grained information. In this work, we introduce a new large-scale dataset that consists of 409 fine-grained categories and 31,881 images with accurate 3D pose annotation. Specifically, we augment three existing fine-grained object recognition datasets (StanfordCars, CompCars and FGVC-Aircraft) by finding a specific 3D model for each sub-category from ShapeNet and manually annotating each 2D image by adjusting a full set of 7 continuous perspective parameters. Since the fine-grained shapes allow 3D models to better fit the images, we further improve the annotation quality by initializing from the human annotation and conducting local search of the pose parameters with the objective of maximizing the IoUs between the projected mask and the segmentation reference estimated from state-of-the-art deep Convolutional Neural Networks (CNNs). We provide full statistics of the annotations with qualitative and quantitative comparisons suggesting that our dataset can be a complementary source for studying 3D pose estimation. The dataset can be downloaded at http://users.umiacs.umd.edu/~wym/3dpose.html.
Multi-dimensional distributions of discrete data that resemble ellipsoids arise in numerous areas of science, statistics, and computational geometry. We describe a complete algebraic algorithm to determine the quadratic form specifying the equation of ellipsoid for the boundary of such multi-dimensional discrete distribution. In this approach, the equation of ellipsoid is reconstructed using a set of matrix equations from low-dimensional projections of the input data. We provide a Mathematica program realizing the full implementation of the ellipsoid reconstruction algorithm in an arbitrary number of dimensions. To demonstrate its many potential uses, the fast reconstruction method is applied to quasi-Gaussian statistical distributions arising in elementary particle production at the Large Hadron Collider.
This paper looks in detail at how an evolutionary algorithm attempts to solve instances from the multimodal problem generator. The paper shows that in order to consistently reach the global optimum, an evolutionary algorithm requires a population size that should grow at least linearly with the number of peaks. It is also shown a close relationship between the supply and decision making issues that have been identified previously in the context of population sizing models for additively decomposable problems. The most important result of the paper, however, is that solving an instance of the multimodal problem generator is like solving a peak-in-a-haystack, and it is argued that evolutionary algorithms are not the best algorithms for such a task. Finally, and as opposed to what several researchers have been doing, it is our strong belief that the multimodal problem generator is not adequate for assessing the performance of evolutionary algorithms.
Using a perturbation approach, we make rigorous the formal boundary layer asymptotic analysis of Turcotte, Spence and Bau from the early eighties for the vertical flow of an internally heated Boussinesq fluid in a vertical channel with viscous dissipation and pressure work. A key point in our proof is to establish the non-degeneracy of a special solution of the Painlev\'{e}-I transcendent. To this end, we relate this problem to recent studies for the ground states of the focusing nonlinear Schr\"{o}dinger equation in an annulus. We also relate our result to a particular case of the well known Lazer-McKenna conjecture from nonlinear analysis.
This paper explores the use of Large Language Models (LLMs) and in particular ChatGPT in programming, source code analysis, and code generation. LLMs and ChatGPT are built using machine learning and artificial intelligence techniques, and they offer several benefits to developers and programmers. While these models can save time and provide highly accurate results, they are not yet advanced enough to replace human programmers entirely. The paper investigates the potential applications of LLMs and ChatGPT in various areas, such as code creation, code documentation, bug detection, refactoring, and more. The paper also suggests that the usage of LLMs and ChatGPT is expected to increase in the future as they offer unparalleled benefits to the programming community.
Codes for rank modulation have been recently proposed as a means of protecting flash memory devices from errors. We study basic coding theoretic problems for such codes, representing them as subsets of the set of permutations of $n$ elements equipped with the Kendall tau distance. We derive several lower and upper bounds on the size of codes. These bounds enable us to establish the exact scaling of the size of optimal codes for large values of $n$. We also show the existence of codes whose size is within a constant factor of the sphere packing bound for any fixed number of errors.
The $\eta'$ mass reduction in the nuclear medium is expected from the degeneracy of the pseudoscalar-singlet and octet mesons when chiral symmetry is manifest. In this study, we investigate the $\eta'N$ 2body interaction which is the foundation of the in-medium $\eta'$ properties using the linear sigma model as a chiral effective model. The $\eta'N$ interaction in the linear sigma model comes from the scalar meson exchange with U$_A$(1) symmetry effect and is found to be fairly strong attraction. Moreover, the $\eta N$ transition is included in our calculation, and is important for the imaginary part of the $\eta'$-optical potential. The transition amplitude of $\eta'N$ to the $\eta N$ channel is relatively small compared to that of elastic channel. From the analysis of the $\eta'N$ 2body system, we find a $\eta'N$ bound state with the binding energy $12.3-3.3i$MeV. We expect that this strongly attractive two body interaction leads to a deep and attractive optical potential.
Online reinforcement learning (RL) algorithms are increasingly used to personalize digital interventions in the fields of mobile health and online education. Common challenges in designing and testing an RL algorithm in these settings include ensuring the RL algorithm can learn and run stably under real-time constraints, and accounting for the complexity of the environment, e.g., a lack of accurate mechanistic models for the user dynamics. To guide how one can tackle these challenges, we extend the PCS (Predictability, Computability, Stability) framework, a data science framework that incorporates best practices from machine learning and statistics in supervised learning (Yu and Kumbier, 2020), to the design of RL algorithms for the digital interventions setting. Further, we provide guidelines on how to design simulation environments, a crucial tool for evaluating RL candidate algorithms using the PCS framework. We illustrate the use of the PCS framework for designing an RL algorithm for Oralytics, a mobile health study aiming to improve users' tooth-brushing behaviors through the personalized delivery of intervention messages. Oralytics will go into the field in late 2022.
Sea-level rise and associated flood hazards pose severe risks to the millions of people globally living in coastal zones. Models representing coastal adaptation and impacts are important tools to inform the design of strategies to manage these risks. Representing the often deep uncertainties influencing these risks poses nontrivial challenges. A common uncertainty characterization approach is to use a few benchmark cases to represent the range and relative probabilities of the set of possible outcomes. This has been done in coastal adaptation studies, for example, by using low, moderate, and high percentiles of an input of interest, like sea-level changes. A key consideration is how this simplified characterization of uncertainty influences the distributions of estimated coastal impacts. Here, we show that using only a few benchmark percentiles to represent uncertainty in future sea-level change can lead to overconfident projections and underestimate high-end risks as compared to using full ensembles for sea-level change and socioeconomic parametric uncertainties. When uncertainty in future sea level is characterized by low, moderate, and high percentiles of global mean sea-level rise, estimates of high-end (95th percentile) damages are underestimated by between 18% (SSP1-2.6) and 46% (SSP5-8.5). Additionally, using the 5th and 95th percentiles of sea-level scenarios underestimates the 5-95% width of the distribution of adaptation costs by a factor ranging from about two to four, depending on SSP-RCP pathway. The resulting underestimation of the uncertainty range in adaptation costs can bias adaptation and mitigation decision-making.
We use three strong lensing clusters to constrain the cosmological parameters Omega_m and Omega_lambda. Recent HST observations of galaxy clusters reveal a large number of multiple images, which are predicted to be at different redshifts. We showed in a previous work that if it is possible to measure spectroscopically the redshift of many multiple images then one can constrain (Omega_m,Omega_lambda) through ratios of angular diameter distances independently of any external assumptions. Using three strong lensing clusters, our combined results lead to tight constraints.
We have investigated the magnetic behavior of the nanocrystalline form of a well-known Laves phase compound, ErCo2 - the bulk form of which has been known to undergo an interesting first-order ferrimagnetic ordering near 32 K - synthesized by high-energy ball-milling. It is found that, in these nanocrystallites, Co exhibits ferromagnetic order at room temperature as inferred from the magnetization data. However, the magnetic transition temperature for Er sublattice remains essentially unaffected as though the (Er)4f-Co(3d) coupling is weak on Er magnetism. The net magnetic moment as measured at high fields, sat at 120 kOe, is significantly reduced with respect to that for the bulk in the ferrimagnetically ordered state and possible reasons are outlined. We have also compared the magnetocaloric behavior for the bulk and the nano particles.
Here, we provide a simple Hubbard-like model of spin-$1/2$ fermions that gives rise to the SU(2) symmetric Thirring model that is equivalent, in the low-energy limit, to Yang-Mills-Chern-Simons model. First, we identify the regime that simulates the SU(2) Yang-Mills theory. Then, we suitably extend this model so that it gives rise to the SU(2) level $k$ Chern-Simons theory with $k\geq2$ that can support non-Abelian anyons. This is achieved by introducing multiple fermionic species and modifying the Thirring interactions, while preserving the SU(2) symmetry. Our proposal provides the means to theoretically and experimentally probe non-Abelian SU(2) level $k$ topological phases.
In this thesis, we settle the computational complexity of some fundamental questions in polynomial optimization. These include the questions of (i) finding a local minimum, (ii) testing local minimality of a point, and (iii) deciding attainment of the optimal value. Our results characterize the complexity of these three questions for all degrees of the defining polynomials left open by prior literature. Regarding (i) and (ii), we show that unless P=NP, there cannot be a polynomial-time algorithm that finds a point within Euclidean distance $c^n$ (for any constant $c$) of a local minimum of an $n$-variate quadratic program. By contrast, we show that a local minimum of a cubic polynomial can be found efficiently by semidefinite programming (SDP). We prove that second-order points of cubic polynomials admit an efficient semidefinite representation, even though their critical points are NP-hard to find. We also give an efficiently-checkable necessary and sufficient condition for local minimality of a point for a cubic polynomial. Regarding (iii), we prove that testing whether a quadratically constrained quadratic program with a finite optimal value has an optimal solution is NP-hard. We also show that testing coercivity of the objective function, compactness of the feasible set, and the Archimedean property associated with the description of the feasible set are all NP-hard. We also give a new characterization of coercive polynomials that lends itself to a hierarchy of SDPs. In our final chapter, we present an SDP relaxation for finding approximate Nash equilibria in bimatrix games. We show that for a symmetric game, a $1/3$-Nash equilibrium can be efficiently recovered from any rank-2 solution to this relaxation. We also propose SDP relaxations for NP-hard problems related to Nash equilibria, such as that of finding the highest achievable welfare under any Nash equilibrium.
The far-from-equilibrium low-temperature dynamics of ultra-thin magnetic films is analyzed by using Monte Carlo numerical simulations on a two dimensional Ising model with competing exchange ($J_0$) and dipolar ($J_d$) interactions. In particular, we focus our attention on the low temperature region of the $(\delta,T)$ phase diagram (where $\delta= J_0/J_d$) for the range of values of $\delta$ where striped phases with widths $h=1$ ($h1$) and $h=2$ ($h2$) are present. The presence of metastable states of the phase $h2$ in the region where the phase $h1$ is the thermodynamically stable one and viceversa was established recently. In this work we show that the presence of these metastable states appears as a blocking mechanism that slows the dynamics of magnetic domains growth when the system is quenched from a high temperature state to a low temperature state in the region of metastability.
The patterns of R violation resulting from imposition of a gauged U(1) horizontal symmetry, on the minimal supersymmetric standard model are systematically analyzed. We concentrate on a class of models with integer U(1) charges chosen to reproduce the quark masses and mixings as well as charged lepton masses exactly or approximately. The U(1) charges are further restricted from the requirement that very large bilinear lepton number violating terms should not be allowed in the super-potential. It is shown that all the trilinear $\lambda'_{ijk}$ and all but at most two trilinear $\lambda_{ijk}$ couplings vanish or are enormously suppressed.
We show the results of a study using the spectral synthesis technique study for the full MaNGA sample showing their Chemical Enrichment History (ChEH) as well as the evolution of the stellar mass-metallicity relation (MZR) over cosmic time. We find that the more massive galaxies became enriched first and the lower mass galaxies did so later, producing a change in the MZR which becomes shallower in time. Separating the sample into morphology and star-forming status bins some particularly interesting results appear: The mass dependency of the MZR becomes less relevant for later morphological types, to the extent that it inverts for Sd/Irr galaxies, suggesting that morphology is at least as important a factor as mass in chemical evolution. The MZR for the full sample shows a flattening at the high-mass end and another at the low-mass range, but the former only appears for retired galaxies while the latter only appears for star-forming galaxies. We also find that the average metallicity gradient is currently negative for all mass bins but for low mass galaxies it was inverted at some point in the past, before which all galaxies had a positive gradient. We also compare how diverse the ChEHs are in the different bins considered as well as what primarily drives the diversity: How much galaxies become enriched or how quickly they do so.
Labeling neural network submodules with human-legible descriptions is useful for many downstream tasks: such descriptions can surface failures, guide interventions, and perhaps even explain important model behaviors. To date, most mechanistic descriptions of trained networks have involved small models, narrowly delimited phenomena, and large amounts of human labor. Labeling all human-interpretable sub-computations in models of increasing size and complexity will almost certainly require tools that can generate and validate descriptions automatically. Recently, techniques that use learned models in-the-loop for labeling have begun to gain traction, but methods for evaluating their efficacy are limited and ad-hoc. How should we validate and compare open-ended labeling tools? This paper introduces FIND (Function INterpretation and Description), a benchmark suite for evaluating the building blocks of automated interpretability methods. FIND contains functions that resemble components of trained neural networks, and accompanying descriptions of the kind we seek to generate. The functions span textual and numeric domains, and involve a range of real-world complexities. We evaluate methods that use pretrained language models (LMs) to produce descriptions of function behavior in natural language and code. Additionally, we introduce a new interactive method in which an Automated Interpretability Agent (AIA) generates function descriptions. We find that an AIA, built from an LM with black-box access to functions, can infer function structure, acting as a scientist by forming hypotheses, proposing experiments, and updating descriptions in light of new data. However, AIA descriptions tend to capture global function behavior and miss local details. These results suggest that FIND will be useful for evaluating more sophisticated interpretability methods before they are applied to real-world models.
Motivated by the physics of coherently coupled, ultracold atom-molecule mixtures, we investigate a classical model possessing the same symmetry -- namely a $U(1)\times \mathbb{Z}_2$ symmetry, associated with the mass conservation in the mixture ($U(1)$ symmetry), times the $\mathbb{Z}_2$ symmetry in the phase relationship between atoms and molecules. In two spatial dimensions the latter symmetry can lead to a finite-temperature Ising transition, associated with (quasi) phase locking between the atoms and the molecules. On the other hand, the $U(1)$ symmetry has an associated Berezinskii-Kosterlitz-Thouless (BKT) transition towards quasi-condensation of atoms or molecules. The existence of the two transitions is found to depend crucially on the population imbalance (or detuning) between atoms and molecules: when the molecules are majority in the system, their BKT quasi-condensation transition occurs at a higher temperature than that of the atoms; the latter has the unconventional nature of an Ising (quasi) phase-locking transition, lacking a finite local order parameter below the critical temperature. When the balance is gradually biased towards the atoms, the two transitions merge together to leave out a unique BKT transition, at which both atoms and molecules acquire quasi-long-range correlations, but only atoms exhibit conventional BKT criticality, with binding of vortex-antivortex pairs into short-range dipoles. The molecular vortex-antivortex excitations bind as well, but undergo a marked crossover from a high-temperature regime in which they are weakly bound, to a low-temperature regime of strong binding, reminiscent of their transition in the absence of atom-molecule coupling.
Building prediction models from mass-spectrometry data is challenging due to the abundance of correlated features with varying degrees of zero-inflation, leading to a common interest in reducing the features to a concise predictor set with good predictive performance. In this study, we formally established and examined regularized regression approaches, designed to address zero-inflated and correlated predictors. In particular, we describe a novel two-stage regularized regression approach (ridge-garrote) explicitly modelling zero-inflated predictors using two component variables, comprising a ridge estimator in the first stage and subsequently applying a nonnegative garrote estimator in the second stage. We contrasted ridge-garrote with one-stage methods (ridge, lasso) and other two-stage regularized regression approaches (lasso-ridge, ridge-lasso) for zero-inflated predictors. We assessed the predictive performance and predictor selection properties of these methods in a comparative simulation study and a real-data case study to predict kidney function using peptidomic features derived from mass-spectrometry. In the simulation study, the predictive performance of all assessed approaches was comparable, yet the ridge-garrote approach consistently selected more parsimonious models compared to its competitors in most scenarios. While lasso-ridge achieved higher predictive accuracy than its competitors, it exhibited high variability in the number of selected predictors. Ridge-lasso exhibited slightly superior predictive accuracy than ridge-garrote but at the expense of selecting more noise predictors. Overall, ridge emerged as a favourable option when variable selection is not a primary concern, while ridge-garrote demonstrated notable practical utility in selecting a parsimonious set of predictors, with only minimal compromise in predictive accuracy.
There is a long-standing controversy about the convergence of the dipole moment of the galaxy angular distribution (the so-called clustering dipole). We study the growth of the clustering dipole of galaxies as a function of the limiting flux of the sample from the Two Micron All Sky Survey (2MASS). Contrary to some earlier claims, we find that the dipole does not converge before the completeness limit of the 2MASS Extended Source Catalog, i.e. up to 13.5 mag in the near-infrared K_s band (equivalent to an effective distance of 300 Mpc/h). We compare the observed growth of the dipole with the theoretically expected, conditional one (i.e., given the velocity of the Local Group relative to the CMB), for the LambdaCDM power spectrum and cosmological parameters constrained by WMAP. The observed growth turns out to be within 1-sigma confidence level of its theoretical counterpart once the proper observational window of the 2MASS flux-limited catalog is included. For a contrast, if the adopted window is a top-hat, then the predicted dipole grows significantly faster and converges to its final value for a distance of about 300 Mpc/h. By comparing the observational windows, we show that for a given flux limit and a corresponding distance limit, the 2MASS flux-weighted window passes less large-scale signal than the top-hat one. We conclude that the growth of the 2MASS dipole for effective distances greater than 200 Mpc/h is only apparent. On the other hand, for a distance of 80 Mpc/h (mean depth of the 2MASS Redshift Survey) and the LambdaCDM power spectrum, the true dipole is expected to reach only ~80% of its final value. Eventually, since for the window function of 2MASS the predicted growth is consistent with the observed one, we can compare the two to evaluate beta = (Omega_m)^0.55 / b. The result is beta = 0.38+-0.04, which leads to an estimate of the density parameter Omega_m = 0.20+-0.08.
Efficiently and accurately translating a corpus into a low-resource language remains a challenge, regardless of the strategies employed, whether manual, automated, or a combination of the two. Many Christian organizations are dedicated to the task of translating the Holy Bible into languages that lack a modern translation. Bible translation (BT) work is currently underway for over 3000 extremely low resource languages. We introduce the eBible corpus: a dataset containing 1009 translations of portions of the Bible with data in 833 different languages across 75 language families. In addition to a BT benchmarking dataset, we introduce model performance benchmarks built on the No Language Left Behind (NLLB) neural machine translation (NMT) models. Finally, we describe several problems specific to the domain of BT and consider how the established data and model benchmarks might be used for future translation efforts. For a BT task trained with NLLB, Austronesian and Trans-New Guinea language families achieve 35.1 and 31.6 BLEU scores respectively, which spurs future innovations for NMT for low-resource languages in Papua New Guinea.
This work introduces a model of Future Technology Transformations for the power sector (FTT:Power), a representation of global power systems based on market competition, induced technological change (ITC) and natural resource use and depletion. It is the first component of a family of sectoral bottom-up models of technology, designed for integration into the global macroeconometric model E3MG. ITC occurs as a result of technological learning produced by cumulative investment and leads to highly nonlinear, irreversible and path dependent technological transitions. The model uses a dynamic coupled set of logistic differential equations. As opposed to traditional bottom-up energy models based on systems optimisation, such differential equations offer an appropriate treatment of the times and structure of change involved in sectoral technology transformations, as well as a much reduced computational load. Resource use and depletion are represented by local cost-supply curves, which give rise to different regional energy landscapes. The model is explored for a single global region using two simple scenarios, a baseline and a mitigation case where the price of carbon is gradually increased. While a constant price of carbon leads to a stagnant system, mitigation produces successive technology transitions leading towards the gradual decarbonisation of the global power sector.
Semi-supervised learning by self-training heavily relies on pseudo-label selection (PLS). The selection often depends on the initial model fit on labeled data. Early overfitting might thus be propagated to the final model by selecting instances with overconfident but erroneous predictions, often referred to as confirmation bias. This paper introduces BPLS, a Bayesian framework for PLS that aims to mitigate this issue. At its core lies a criterion for selecting instances to label: an analytical approximation of the posterior predictive of pseudo-samples. We derive this selection criterion by proving Bayes optimality of the posterior predictive of pseudo-samples. We further overcome computational hurdles by approximating the criterion analytically. Its relation to the marginal likelihood allows us to come up with an approximation based on Laplace's method and the Gaussian integral. We empirically assess BPLS for parametric generalized linear and non-parametric generalized additive models on simulated and real-world data. When faced with high-dimensional data prone to overfitting, BPLS outperforms traditional PLS methods.
In a recent paper (Phys. Rev. Lett. 109, 160501 (2012). arXiv:1201.0849), it is claimed that any quantum protocol for classical two-sided computation between Alice and Bob can be proven completely insecure for Alice if it is secure against Bob. Here we show that the proof is not sufficiently general, because the security definition it based on is only a sufficient condition but not a necessary condition.
In the first part of this thesis we study baryonic U(1) symmetries dual to Betti multiplets in the AdS_4/CFT_3 correspondence for M2 branes at Calabi-Yau 4-fold singularities. We begin by focusing on isolated toric singularities without vanishing 6-cycles, which we classify, and propose for them field theory duals. We then study in detail the cone over Q^111 and find agreement between the spectrum of baryonic operators in this theory and M5 branes wrapping 5-cycles in the Q^111 space. The physics of vacua in which these symmetries are spontaneously broken precisely matches a dual gravity analysis involving resolutions of the singularity, where we are able to match condensates, Goldstone bosons and global strings. We then study the implications of turning on a torsion 4-form flux. This flux non-trivially affects the supergravity dual of Higgsing, and we show that the supergravity and field theory analyses precisely match in an example based on Y^12(CP^2). We then explain how the choice of M-theory circle can result in exotic renormalization group flows. We also argue that theories where the resolutions have 6-cycles are expected to receive non-perturbative corrections from M5 instantons. We give a general formula relating the instanton action to normalizable harmonic 2-forms. In the second part of this thesis we study the breaking of baryonic symmetries in the AdS_5/CFT_4 correspondence. This leads, for particular vacuum expectation values, to the emergence of baryonic symmetries during the renormalization group flow. We identify these vacuum expectation values with critical values of the B-field moduli in the dual supergravity backgrounds. We study in detail the C^3/Z_3 orbifold theory and the dual supergravity backgrounds that correspond to the breaking of the emerging baryonic symmetries, and identify the expected Goldstone bosons and global strings in the IR.
Both symmetry and organized breaking of symmetry have a pivotal r\^ole in our understanding of structure and pattern formation in physical systems, including the origin of mass in the Universe and the chiral structure of biological macromolecules. Here we report on a new symmetry breaking phenomenon that takes place in all biologically active proteins, thus this symmetry breaking relates to the inception of life. The unbroken symmetry determines the covalent bond geometry of a sp3 hybridized carbon atom. It dictates the tetrahedral architecture of atoms around the central carbon of an amino acid. Here we show that in a biologically active protein this symmetry becomes broken. Moreover, we show that the pattern of symmetry breaking is in a direct correspondence with the local secondary structure of the folded protein.
The inference of topological principles is a key problem in structured reconstruction. We observe that wrongly predicted topological relationships are often incurred by the lack of holistic geometry clues in low-level features. Inspired by the fact that massive signals can be compactly described with frequency analysis, we experimentally explore the efficiency and tendency of learning structure geometry in the frequency domain. Accordingly, we propose a frequency-domain feature learning strategy (F-Learn) to fuse scattered geometric fragments holistically for topology-intact structure reasoning. Benefiting from the parsimonious design, the F-Learn strategy can be easily deployed into a deep reconstructor with a lightweight model modification. Experiments demonstrate that the F-Learn strategy can effectively introduce structure awareness into geometric primitive detection and topology inference, bringing significant performance improvement to final structured reconstruction. Code and pre-trained models are available at https://github.com/Geo-Tell/F-Learn.
We continue our study of BPS equations and supersymmetric configurations in the Bagger-Lambert theory. The superalgebra allows three different types of central extensions which correspond to compounds of various M-theory objects: M2-branes, M5-branes, gravity waves and Kaluza-Klein monopoles which intersect or have overlaps with the M2-branes whose dynamics is given by the Bagger-Lambert action. As elementary objects they are all 1/2-BPS, and multiple intersections of $n$-branes generically break the supersymmetry into $1/2^n$, as it is well known. But a particular composite of M-branes can preserve from 1/16 up to 3/4 of the original ${\cal N}=8$ supersymmetries as previously discovered. In this paper we provide the M-theory interpretation for various BPS equations, and also present explicit solutions to some 1/2-BPS equations.
We present a straightforward integration method to compute the abundance and temperature evolution in explosive scenarios. In this approach the thermal equation is implicitely coupled with chemical equations in order to avoid instabilities and ensure a gentle transition from the normal combustion regime to the quasi (QSE) and complete nuclear statistical equilibrium (NSE). Two nuclear networks, with 14 nuclei (alpha-network) and 86 nuclei (including protons and neutrons) respectively, have been considered. The scheme is suitable to cope with a variety of explosive burning regimes.
The ability to design the control of heat flow has innumerable benefits in the design of electronic systems such as thermoelectric energy harvesters, solid-state lighting, and thermal imagers, where the thermal design plays a key role in performance and device reliability. However, to realize one advanced control function of thermal flux, one needs to design one sophisticated, multilayered and inhomogeneous thermal structure with different composition/shape at different regions of one device. In this work, we employ one identical sensu-unit with facile natural composition to experimentally realize a new class of thermal metamaterials for controlling thermal conduction (e.g., thermal concentrator, focusing/resolving, uniform heating), only resorting to positioning and locating the same unit element of sensu-shape structure. The thermal metamaterial unit and the proper arrangement of multiple identical units are capable of transferring, redistributing and managing thermal energy in a versatile fashion. It is also shown that our sensu-shape unit elements can be used in manipulating dc currents without any change in the layout for the thermal counterpart. The proposed scheme can also be applied to control dc electric currents and dc magnetic fields that governed by Laplace equation. These could markedly enhance the capabilities in thermal sensing, thermal imaging, thermal-energy storage, thermal packaging, thermal therapy, and more domains beyond.
We set up a framework of 2-Hilbert bundles, which allows to rigorously define the "stringor bundle", a higher differential geometric object anticipated by Stolz and Teichner in an unpublished preprint about 20 years ago. Our framework includes an associated bundle construction, allowing us to associate a 2-Hilbert bundle with a principal 2-bundle and a unitary representation of its structure 2-group. We prove that the Stolz-Teichner stringor bundle is canonically isomorphic to the 2-Hilbert bundle obtained from applying our associated bundle construction to a string structure on a manifold and the stringor representation of the string 2-group that we discovered in earlier work. This establishes a perfect analogy to spin manifolds, representations of the spin groups, and spinor bundles.
This thesis examines some of the applications of scaling relations in understanding non linear structure formation.
We present a phenomenological study of photon-initiated (PI) lepton production at the LHC, as implemented in the structure function (SF) approach. We provide detailed predictions for multi-differential lepton pair production, and show that the impact on observables sensitive to the the weak mixing angle, $\sin^2 \theta_W$, and $W$ boson mass, $M_W$, as well as PDFs can be non-negligible, in particular given the high precision being aimed for. The SF calculation can provide percent level precision in the corresponding production cross sections, and is therefore well positioned to account for LHC precision requirements. We consider the pure $\gamma\gamma$ channel, and compare in detail to the NLO collinear calculation. We in addition include initial-state $Z$ as well as mixed $\gamma /Z + q$ contributions, and assess their impact. We also consider photon-initiated lepton-lepton scattering, and again find the SF approach can provide high precision predictions for this process in way that can straightforwardly account for any fiducial cuts imposed. Finally, we provide a publicly available Monte Carlo generator, SFGen, for PI lepton pair production and lepton-lepton scattering within the SF approach, for use by the community.
The doped perovskite cobaltite La1-xSrxCoO3 (LSCO) has been advanced as a model system for studying intrinsic magnetic phase separation. We have employed a first-order reversal curve (FORC) method to probe the amount of irreversible switching in bulk polycrystalline LSCO as a function of Sr doping, field cooling procedure, and temperature. The value of the FORC distribution, rho, is used as a measure of the extent of irreversible switching. For x < 0.18, the small values of rho and its ridge-like distribution along local coercivity (Hc) and zero bias (Hb), are characteristic of non-interacting single domain particles. This is consistent with the formation of an array of isolated nanoscopic ferromagnetic clusters, as observed in previous work. For x >= 0.18, the much larger values of rho, the tilting of its distribution towards negative bias field, and the emergence of regions with negative rho, are consistent with increased long-range ferromagnetic ordering. The FORC distributions display little dependence on the cooling procedure. With increasing temperature, the fraction of irreversible switching determined from the FORC distribution follows closely the ferromagnetic phase fraction measured by La nuclear magnetic resonance. Our results furthermore demonstrate that the FORC method is a valuable first-pass characterization tool for magnetic phase separation.
The two most fundamental processes describing change in biology, development and evolu-tion, occur over drastically different timescales, difficult to reconcile within a unified framework. Development involves temporal sequences of cell states controlled by hierarchies of regulatory structures. It occurs over the lifetime of a single individual, and is associated to the gene expression level change of a given genotype. Evolution, by contrast entails genotypic change through the acquisition/loss of genes and changes in the network topology of interactions among genes. It involves the emergence of new, environmentally selected phenotypes over the lifetimes of many individuals. Here we present a model of regulatory network evolution that accounts for both timescales. We extend the framework of Boolean models of gene regulatory networks (GRN)-currently only applicable to describing development to include evolutionary processes. As opposed to one-to-one maps to specific attractors, we identify the phenotypes of the cells as the relevant macrostates of the GRN. A phenotype may now correspond to multiple attractors, and its formal definition no longer requires a fixed size for the genotype. This opens the possibility for a quantitative study of the phenotypic change of a genotype, which is itself changing over evolutionary timescales. We show how the realization of specific phenotypes can be controlled by gene duplication events (used here as an archetypal evolutionary event able to change the genotype), and how successive events of gene duplication lead to new regulatory structures via selection. At the same time, we show that our generalized framework does not inhibit network controllability and the possibility for network control theory to describe epigenetic signaling during development.
A introduction to the syntax and Semantics of Answer Set Programming intended as an handout to [under]graduate students taking Artificial Intlligence or Logic Programming classes.