text
stringlengths
6
128k
In this Letter, both the manipulation of valley-polarized currents and the optical-like behaviors of Dirac fermions are theoretically explored in polycrystalline graphene. When strain is applied, the misorientation between two graphene domains separated by a grain boundary can result in a mismatch of their electronic structures. Such a discrepancy manifests itself in a strong breaking of the inversion symmetry, leading to perfect valley polarization in a wide range of transmission directions. In addition, these graphene domains act as different media for electron waves, offering the possibility to modulate and obtain negative refraction indexes.
Partial observability and uncertainty are common problems in sequential decision-making that particularly impede the use of formal models such as Markov decision processes (MDPs). However, in practice, agents may be able to employ costly sensors to measure their environment and resolve partial observability by gathering information. Moreover, imprecise transition functions can capture model uncertainty. We combine these concepts and extend MDPs to robust active-measuring MDPs (RAM-MDPs). We present an active-measure heuristic to solve RAM-MDPs efficiently and show that model uncertainty can, counterintuitively, let agents take fewer measurements. We propose a method to counteract this behavior while only incurring a bounded additional cost. We empirically compare our methods to several baselines and show their superior scalability and performance.
In this paper we present an algorithm, called conauto-2.0, that can efficiently compute a set of generators of the automorphism group of a graph, and test whether two graphs are isomorphic, finding an isomorphism if they are. This algorithm uses the basic individualization/refinement technique, and is an improved version of the algorithm conauto, which has been shown to be very fast for random graphs and several families of hard graphs. In this paper, it is proved that, under some circumstances, it is not only possible to prune the search space (using already found generators of the automorphism group), but also to infer new generators without the need of explicitly finding an automorphism of the graph. This result is especially suited for graphs with regularly connected components, and can be applied in any isomorphism testing and canonical labeling algorithm (that use the individualization/refinement technique) to significantly improve its performance. Additionally, a dynamic target cell selection function is used to adapt to different graphs. The resulting algorithm preserves all the nice features of conauto, but reduces the time for testing graphs with regularly connected components and other hard graph families. We run extensive experiments, which show that the most popular algorithms (namely, nauty, bliss, Traces, and saucy) are slower than conauto-2.0, among others, for the graph families based on components.
Silicon photonics enables wafer-scale integration of optical functionalities on chip. A silicon-based laser frequency combs could significantly expand the applications of silicon photonics, by providing integrated sources of mutually coherent laser lines for terabit-per-second transceivers, parallel coherent LiDAR, or photonics-assisted signal processing. Here, we report on heterogeneously integrated laser soliton microcombs combining both InP/Si semiconductor lasers and ultralow-loss silicon nitride microresonators on monolithic silicon substrate. Thousands of devices are produced from a single wafer using standard CMOS techniques. Using on-chip electrical control of the microcomb-laser relative optical phase, these devices can output single-soliton microcombs with 100 GHz repetition rate. Our approach paves the way for large-volume, low-cost manufacturing of chip-based frequency combs for next-generation high-capacity transceivers, datacenters, space and mobile platforms.
Synchronization is the process of achieving identical dynamics among coupled identical units. If the units are different from each other, their dynamics cannot become identical; yet, after transients, there may emerge a functional relationship between them -- a phenomenon termed "generalized synchronization." Here, we show that the concept of transient uncoupling, recently introduced for synchronizing identical units, also supports generalized synchronization among nonidentical chaotic units. Generalized synchronization can be achieved by transient uncoupling even when it is impossible by regular coupling. We furthermore demonstrate that transient uncoupling stabilizes synchronization in the presence of common noise. Transient uncoupling works best if the units stay uncoupled whenever the driven orbit visits regions that are locally diverging in its phase space. Thus, to select a favorable uncoupling region, we propose an intuitive method that measures the local divergence at the phase points of the driven unit's trajectory by linearizing the flow and subsequently suppresses the divergence by uncoupling.
We examine the sensitivity of flavor changing neutral current (FCNC) processes to anomalous triple gauge boson couplings. We show that in the non-linear realization of the electroweak symmetry breaking sector these processes are very sensitive to two CP conserving anomalous couplings. A clean separation of their effects is possible in the next round of experiments probing $b\to s\gamma$ and $b\to s\ell^+\ell^-$ processes, as well as kaon decays such as $K^+\to\pi^+\nu\bar\nu$. The obtained sensitivity is found to be competitive with that of direct measurements at high energy colliders. In particular, for one of the $WWZ$ couplings the one-loop FCNC effects are enhanced by a logarithmic dependence on the scale of new physics. We also explore the potential signals of CP violating anomalous triple gauge boson couplings in rare $B$ decays.
This letter establishes a novel relationship between a class of recurrent neural networks and certain evolutionary dynamics that emerge in the context of population games. Specifically, it is shown that the output of a recurrent neural network, in the context of classification problems, coincides with the evolution of the population state in a population game. This connection is established with dynamic payoffs and under replicator evolutionary dynamics. The connection provides insights into the neural network's behavior from both dynamical systems and game-theoretical perspectives, aligning with recent literature that suggests that neural network outputs may resemble the Nash equilibria of suitable games. It also uncovers potential connections between the neural network classification problem and mechanism design. To illustrate our results, we present different numerical experiments in the context of classification problems.
Signed distance field (SDF) is a prominent implicit representation of 3D meshes. Methods that are based on such representation achieved state-of-the-art 3D shape reconstruction quality. However, these methods struggle to reconstruct non-convex shapes. One remedy is to incorporate a constructive solid geometry framework (CSG) that represents a shape as a decomposition into primitives. It allows to embody a 3D shape of high complexity and non-convexity with a simple tree representation of Boolean operations. Nevertheless, existing approaches are supervised and require the entire CSG parse tree that is given upfront during the training process. On the contrary, we propose a model that extracts a CSG parse tree without any supervision - UCSG-Net. Our model predicts parameters of primitives and binarizes their SDF representation through differentiable indicator function. It is achieved jointly with discovering the structure of a Boolean operators tree. The model selects dynamically which operator combination over primitives leads to the reconstruction of high fidelity. We evaluate our method on 2D and 3D autoencoding tasks. We show that the predicted parse tree representation is interpretable and can be used in CAD software.
We show that if g is a generic (in the sense of Baire category) isometry of a generic subspace of the Urysohn metric space U, then g does not extend to a full isometry of U. The same holds for the Urysohn sphere S. Let M be a Fraisse L-structure, where L is a relational countable language and M has no algebraicity. We provide necessary and sufficient conditions for the following to hold: for a generic substructure A of M, every automorphism f in Aut(A) extends to a full automorphism f in Aut(M). From our analysis, a dichotomy arises and some structural results are derived that, in particular, apply to omega-stable Fraisse structures without algebraicity.
We realize Leavitt ultragraph path algebras as partial skew group rings. Using this realization we characterize artinian ultragraph path algebras and give simplicity criteria for these algebras.
In this article we briefly discuss the finite generation of fiber rings of invariant k-jets of holomorphic curves in a complex projective manifold, using differential Galois theory.
We analyze the effects of an in-medium broadening of nucleon resonances on the exclusive photoproduction of mesons on nuclei as well as on the total photoabsorption cross sections in a transport calculation. We show that the resonance widths observed in semi-inclusive photoproduction on nuclei are insensitive to an in-medium broadening of nucleon resonances. This is due to a simple effect: the sizeable width of the nuclear surface and Fermi motion.
Previous analyses of mid-infrared water spectra from young protoplanetary disks observed with the Spitzer-IRS found an anti-correlation between water luminosity and the millimeter dust disk radius observed with ALMA. This trend was suggested to be evidence for a fundamental process of inner disk water enrichment, used to explain properties of the Solar System 40 years ago, in which icy pebbles drift inward from the outer disk and sublimate after crossing the snowline. Previous analyses of IRS water spectra, however, were uncertain due to the low spectral resolution that blended lines together. We present new JWST-MIRI spectra of four disks, two compact and two large with multiple radial gaps, selected to test the scenario that water vapor inside the snowline is regulated by pebble drift. The higher spectral resolving power of MIRI-MRS now yields water spectra that separate individual lines, tracing upper level energies from 900 K to 10,000 K. These spectra clearly reveal excess emission in the low-energy lines in compact disks, compared to the large disks, demonstrating an enhanced cool component with $T \approx$ 170-400 K and equivalent emitting radius $R_{\rm{eq}}\approx$ 1-10 au. We interpret the cool water emission as ice sublimation and vapor diffusion near the snowline, suggesting that there is indeed a higher inwards mass flux of icy pebbles in compact disks. Observation of this process opens up multiple exciting prospects to study planet formation chemistry in inner disks with JWST.
Pulsars are special objects whose positions can be determined independently from timing, radio interferometric, and Gaia astrometry at sub-milliarcsecond (mas) precision; thus, they provide a unique way to monitor the link between dynamical and kinematic reference frames. We aimed to assess the orientation consistency between the dynamical reference frame represented by the planetary ephemeris and the kinematic reference frames constructed by Gaia and VLBI through pulsar positions. We identified 49 pulsars in Gaia Data Release 3 and 62 pulsars with very long baseline interferometry (VLBI) positions from the PSR$\pi$ and MSPSR$\pi$ projects and searched for the published timing solutions of these pulsars. We then compared pulsar positions measured by timing, VLBI, and Gaia to estimate the orientation offsets of the ephemeris frames with respect to the Gaia and VLBI reference frames by iterative fitting. We found orientation offsets of $\sim$10 mas in the DE200 frame with respect to the Gaia and VLBI frame. Our results depend strongly on the subset used in the comparison and could be biased by underestimated errors in the archival timing data, reflecting the limitation of using the literature timing solutions to determine the frame rotation.
In previous theoretical research, we inferred that cancer stem cells (CSCs), the cells that presumably drive tumor growth and resistance to conventional cancer treatments, are not uniformly distributed in the bulk of a tumorsphere. To confirm this theoretical prediction, we cultivated tumorspheres enriched in CSCs, and performed immunofluorecent detection of the stemness marker SOX2 using a confocal microscope. In this article, we present a method developed to process the images that reconstruct the amount and location of the CSCs in the tumorspheres. Its advantage is the use of a statistical criterion to classify the cells in stem and differentiated instead of setting an arbitrary threshold. From the analysis of the results of the methods using graph theory and computational modeling, we concluded that the distribution of Cancer Stem Cells in an experimental tumorsphere is non-homogeneous. This method is independent of the tumorsphere assay being useful for analyzing images in which several different kinds of cells are stained with different markers.
The future of 3D printing utilizing unmanned aerial vehicles (UAVs) presents a promising capability to revolutionize manufacturing and to enable the creation of large-scale structures in remote and hard- to-reach areas e.g. in other planetary systems. Nevertheless, the limited payload capacity of UAVs and the complexity in the 3D printing of large objects pose significant challenges. In this article we propose a novel chunk-based framework for distributed 3D printing using UAVs that sets the basis for a fully collaborative aerial 3D printing of challenging structures. The presented framework, through a novel proposed optimisation process, is able to divide the 3D model to be printed into small, manageable chunks and to assign them to a UAV for partial printing of the assigned chunk, in a fully autonomous approach. Thus, we establish the algorithms for chunk division, allocation, and printing, and we also introduce a novel algorithm that efficiently partitions the mesh into planar chunks, while accounting for the inter-connectivity constraints of the chunks. The efficiency of the proposed framework is demonstrated through multiple physics based simulations in Gazebo, where a CAD construction mesh is printed via multiple UAVs carrying materials whose volume is proportionate to a fraction of the total mesh volume.
Superconductivity forms out of the condensation of Cooper pairs of electrons. The mechanism by which Cooper pairs are created in non-conventional superconductors is often elusive because experimental signatures that connect a specific pairing mechanism to the properties of superconducting state are rare. The recently discovered superconducting oxide-insulator/KTaO$_3$ interface may offer clues about its origins. Here we observe distinct dependences of the superconducting transition temperature Tc on carrier density n$_{2D}$ for electron gases formed at KTaO$_3$ (111), (001) and (110) interfaces. For the KTaO$_3$ (111) interface, a remarkable linear dependence of Tc on n$_{2D}$ is observed over a range of nearly one order of magnitude. Further, our study of the dependence of superconductivity on gate electric fields reveals the role of the interface in mediating superconductivity, which also allows for a reversible electric switching of superconductivity at T = 2 K. We found that the extreme sensitivity of superconductivity to crystallographic orientation can be explained by Cooper pairing via inter-orbital interactions induced by the inversion-breaking transverse optical (TO1) phonons and quantum confinement. This mechanism is also consistent with the dependence of Tc on n$_{2D}$ at the KTaO$_3$ (111) interface. Our study may shed light on the pairing mechanism in other superconducting quantum-paraelectrics.
By using a method, previously established to calculate electromagnetic fields, we compute the force of light upon a metallic particle. This procedure is based on both Maxwell's Stress Tensor and the Couple Dipole Method. With these tools, we study the force when the particle is over a flat dielectric surface. The multiple interaction of light between the particle and the surface is fully taken into account. The wave illuminating the particle is either evanescent or propagating depending an whether or not total internal reflection takes place. We analyze the behaviour of this force on either a small or a large particle in terms of the wavelength. A remarkable result obtained for evanescent field illumination, is that the force on a small silver particle can be either attractive or repulsive depending on the wavelength. This behaviour also varies as the particle becomes larger.
Except for Koshy who devotes seven pages to applications of Fibonacci Numbers to electric circuits, most books and the Fibonacci Quarterly have been relatively silent on applications of graphs and electric circuits to Fibonacci numbers. This paper continues a recent trend of papers studying the interplay of graphs, circuits, and Fibonacci numbers by presenting and studying the Circuit Array, an infinite 2-dimensional array whose entries are electric resistances labelling edge values of circuits associated with a family of graphs. The Circuit Array has several features distinguishing it from other more familiar arrays such as the Binomial Array and Wythoff Array. For example, it can be proven modulo a strongly supported conjecture that the numerators of its left-most diagonal do not satisfy any linear, homogeneous, recursion, with constant coefficients (LHRCC). However, we conjecture with supporting numerical evidence an asymptotic formula involving $\pi$ satisfied by the left-most diagonal of the Circuit Array.
We present a multilevel Monte Carlo (MLMC) method for the uncertainty quantification of variably saturated porous media flow that are modeled using the Richards' equation. We propose a stochastic extension for the empirical models that are typically employed to close the Richards' equations. This is achieved by treating the soil parameters in these models as spatially correlated random fields with appropriately defined marginal distributions. As some of these parameters can only take values in a specific range, non-Gaussian models are utilized. The randomness in these parameters may result in path-wise highly nonlinear systems, so that a robust solver with respect to the random input is required. For this purpose, a solution method based on a combination of the modified Picard iteration and a cell-centered multigrid method for heterogeneous diffusion coefficients is utilized. Moreover, we propose a non-standard MLMC estimator to solve the resulting high-dimensional stochastic Richards' equation. The improved efficiency of this multilevel estimator is achieved by parametric continuation that allows us to incorporate simpler nonlinear problems on coarser levels for variance reduction while the target strongly nonlinear problem is solved only on the finest level. Several numerical experiments are presented showing computational savings obtained by the new estimator compared to the original MC estimator.
The Fanaroff-Riley class II radio galaxy Cygnus A hosts jets which produce radio emission, X-ray cavities, cocoon shocks, and X-ray hotspots where the jet interacts with the ICM. Surrounding one hotspot is a peculiar "hole" feature which appears as a deficit in X-ray emission. We use relativistic hydrodynamic simulations of a collimated jet interacting with an inclined interface between lobe and cluster plasma to model the basic processes which may lead to such a feature. We find that the jet reflects off of the interface into a broad, turbulent flow back out into the lobe, which is dominated by gas stripped from the interface at first and from the intracluster medium itself at later times. We produce simple models of X-ray emission from the ICM, the hotspot, and the reflected jet to show that a hole of emission surrounding the hotspot as seen in Cygnus A may be produced by Doppler de-boosting of the emission from the reflected jet as seen by an observer with a sight line nearly along the axis of the outgoing material.
As crowdsourcing emerges as an efficient and cost-effective method for obtaining labels for machine learning datasets, it is important to assess the quality of crowd-provided data, so as to improve analysis performance and reduce biases in subsequent machine learning tasks. Given the lack of ground truth in most cases of crowdsourcing, we refer to data quality as annotators' consistency and credibility. Unlike the simple scenarios where Kappa coefficient and intraclass correlation coefficient usually can apply, online crowdsourcing requires dealing with more complex situations. We introduce a systematic method for evaluating data quality and detecting spamming threats via variance decomposition, and we classify spammers into three categories based on their different behavioral patterns. A spammer index is proposed to assess entire data consistency and two metrics are developed to measure crowd worker's credibility by utilizing the Markov chain and generalized random effects models. Furthermore, we showcase the practicality of our techniques and their advantages by applying them on a face verification task with both simulation and real-world data collected from two crowdsourcing platforms.
This paper concerns the dynamical behavior of weakly reversible, deterministically modeled population processes near the facets (codimension-one faces) of their invariant manifolds and proves that the facets of such systems are "repelling." It has been conjectured that any population process whose network graph is weakly reversible (has strongly connected components) is persistent. We prove this conjecture to be true for the subclass of weakly reversible systems for which only facets of the invariant manifold are associated with semilocking sets, or siphons. An important application of this work pertains to chemical reaction systems that are complex-balancing. For these systems it is known that within the interior of each invariant manifold there is a unique equilibrium. The Global Attractor Conjecture states that each of these equilibria is globally asymptotically stable relative to the interior of the invariant manifold in which it lies. Our results pertaining to weakly reversible systems imply that this conjecture holds for all complex-balancing systems whose boundary equilibria lie in the relative interior of the boundary facets. As a corollary, we show that the Global Attractor Conjecture holds for those systems for which the associated invariant manifolds are two-dimensional.
Active learning (AL) techniques reduce labeling costs for training neural machine translation (NMT) models by selecting smaller representative subsets from unlabeled data for annotation. Diversity sampling techniques select heterogeneous instances, while uncertainty sampling methods select instances with the highest model uncertainty. Both approaches have limitations - diversity methods may extract varied but trivial examples, while uncertainty sampling can yield repetitive, uninformative instances. To bridge this gap, we propose HUDS, a hybrid AL strategy for domain adaptation in NMT that combines uncertainty and diversity for sentence selection. HUDS computes uncertainty scores for unlabeled sentences and subsequently stratifies them. It then clusters sentence embeddings within each stratum using k-MEANS and computes diversity scores by distance to the centroid. A weighted hybrid score that combines uncertainty and diversity is then used to select the top instances for annotation in each AL iteration. Experiments on multi-domain German-English datasets demonstrate the better performance of HUDS over other strong AL baselines. We analyze the sentence selection with HUDS and show that it prioritizes diverse instances having high model uncertainty for annotation in early AL iterations.
This thesis investigates the quality of randomly collected data by employing a framework built on information-based complexity, a field related to the numerical analysis of abstract problems. The quality or power of gathered information is measured by its radius which is the uniform error obtainable by the best possible algorithm using it. The main aim is to present progress towards understanding the power of random information for approximation and integration problems.
We show that large language models (LLMs) are remarkably good at working with interpretable models that decompose complex outcomes into univariate graph-represented components. By adopting a hierarchical approach to reasoning, LLMs can provide comprehensive model-level summaries without ever requiring the entire model to fit in context. This approach enables LLMs to apply their extensive background knowledge to automate common tasks in data science such as detecting anomalies that contradict prior knowledge, describing potential reasons for the anomalies, and suggesting repairs that would remove the anomalies. We use multiple examples in healthcare to demonstrate the utility of these new capabilities of LLMs, with particular emphasis on Generalized Additive Models (GAMs). Finally, we present the package $\texttt{TalkToEBM}$ as an open-source LLM-GAM interface.
We propose a multi-criteria Composite Index Method (CIM) to compare the performance of alternative approaches to solving an optimization problem. The CIM is convenient in those situations when neither approach dominates the other when tested on different sizes of problem instances. The CIM takes problem instance size and multiple performance criteria into consideration within a weighting scheme to produce a single number that measures the relative improvement of one alternative over the other. Different weights are given to each dimension based on their relative importance as determined by the end user. We summarize the successful application of the CIM to an NP-hard combinatorial optimization problem known as the backhaul profit maximization problem (BPMP). Using the CIM we tested a series of eleven techniques for improving solution time using CPLEX to solve two different BPMP models proposed in the literature.
In this paper, we study Markov dynamics on unitary duals of compact quantum groups. We construct such dynamics from characters of quantum groups. Then we show that the dynamics have generators, and we give an explicit formula of the generators using the representation theory. Moreover, we construct Markov dynamics on the unitary dual of an inductive limit of compact quantum groups.
We report on the status of a variety of radiative B decays studied by the CLEO detector with $9.7\times 10^6$ $B\bar{B}$ pairs.
In this paper we investigate solvability of a partial integral equation in the space $L_2(\Omega\times\Omega),$ where $\Omega=[a,b]^\nu.$ We define a determinant for the partial integral equation as a continuous function on $\Omega$ and for a continuous kernels of the partial integral equation we give explicit description of the solution.
We propose a new, unifying framework that yields an array of cryptographic primitives with certified deletion. These primitives enable a party in possession of a quantum ciphertext to generate a classical certificate that the encrypted plaintext has been information-theoretically deleted, and cannot be recovered even given unbounded computational resources. - For X \in {public-key, attribute-based, fully-homomorphic, witness, timed-release}, our compiler converts any (post-quantum) X encryption to X encryption with certified deletion. In addition, we compile statistically-binding commitments to statistically-binding commitments with certified everlasting hiding. As a corollary, we also obtain statistically-sound zero-knowledge proofs for QMA with certified everlasting zero-knowledge assuming statistically-binding commitments. - We also obtain a strong form of everlasting security for two-party and multi-party computation in the dishonest majority setting. While simultaneously achieving everlasting security against all parties in this setting is known to be impossible, we introduce everlasting security transfer (EST). This enables any one party (or a subset of parties) to dynamically and certifiably information-theoretically delete other participants' data after protocol execution. We construct general-purpose secure computation with EST assuming statistically-binding commitments, which can be based on one-way functions or pseudorandom quantum states. We obtain our results by developing a novel proof technique to argue that a bit b has been information-theoretically deleted from an adversary's view once they output a valid deletion certificate, despite having been previously information-theoretically determined by the ciphertext they held in their view. This technique may be of independent interest.
We propose generalizations of a number of standard network models, including the classic random graph, the configuration model, and the stochastic block model, to the case of time-varying networks. We assume that the presence and absence of edges are governed by continuous-time Markov processes with rate parameters that can depend on properties of the nodes. In addition to computing equilibrium properties of these models, we demonstrate their use in data analysis and statistical inference, giving efficient algorithms for fitting them to observed network data. This allows us, for instance, to estimate the time constants of network evolution or infer community structure from temporal network data using cues embedded both in the probabilities over time that node pairs are connected by edges and in the characteristic dynamics of edge appearance and disappearance. We illustrate our methods with a selection of applications, both to computer-generated test networks and real-world examples.
We study the interplay between coherent transport by tunneling and diffusive transport through classically chaotic phase-space regions, as it is reflected in the Floquet spectrum of the periodically driven quartic double well. The tunnel splittings in the semiclassical regime are determined with high numerical accuracy, and the association of the corresponding doublet states to either chaotic or regular regions of the classical phase space is quantified in terms of the overlap of the Husimi distribution with the chaotic layer along the separatrix. We find a strong correlation between both quantities. They show an increase by orders of magnitude as chaotic diffusion between the wells starts to dominate the classical dynamics. We discuss semiclassical explanations for this correlation.
The possible origin of the R-parity violating interactions in the minimal supersymmetric standard model and its connection to the radiative symmetry breaking mechanism (RSBM) is investigated. In the context of the simplest model where the implementation of the RSBM is possible, we find that in the majority of the parameter space R-parity is spontaneously broken at the low-scale. These results hint at the possibility that R-parity violating processes will be observed at the Large Hadron Collider, if Supersymmetry is realized in nature.
Using a model heat engine, we show that neural network-based reinforcement learning can identify thermodynamic trajectories of maximal efficiency. We consider both gradient and gradient-free reinforcement learning. We use an evolutionary learning algorithm to evolve a population of neural networks, subject to a directive to maximize the efficiency of a trajectory composed of a set of elementary thermodynamic processes; the resulting networks learn to carry out the maximally-efficient Carnot, Stirling, or Otto cycles. When given an additional irreversible process, this evolutionary scheme learns a previously unknown thermodynamic cycle. Gradient-based reinforcement learning is able to learn the Stirling cycle, whereas an evolutionary approach achieves the optimal Carnot cycle. Our results show how the reinforcement learning strategies developed for game playing can be applied to solve physical problems conditioned upon path-extensive order parameters.
We investigate the unusual properties of quasirelativistic massless fermions under a magnetic or electric field by means of nonminimal couplings. Within this approach, the spin-orbit coupling (SOC) effects are properly generated in the energy spectrum of the quasiparticles. By including a magnetic field, $B$, we show that the spin splitting of Landau Levels (LL) obeys a $\sqrt{B}$ linear dependence with SOC, typical of relativistic particles. Moreover, our calculated spectrum of LLs resembles the behavior of the three-dimensional (3D) massless Kane fermions. Using a nonminimal coupling with an external electric field, we demonstrate that a Rashba-like SOC naturally appears into the relativistic equations and apply to the case of two-dimensional (2D) massless Dirac fermions. Still considering our proposed approach, the Hall conductivity is also computed for the 2D case under transverse electric field both at zero and finite temperatures for a general chemical potential. The results feature a typical quantization of the Hall conductivity at low temperatures, when the absolute value of the gap opened by the electric field is larger than the considered chemical potential.
The electronic structure of a vortex line trapped by an insulating columnar defect in a type-II superconductor is analysed within the Bogolubov-de Gennes theory. For quasiparticle trajectories with small impact parameters defined with respect to the vortex axis the normal reflection of electrons and holes at the defect surface results in the formation of an additional subgap spectral branch. The increase in the impact parameter at this branch is accompanied by the decrease of the excitation energy. When the impact parameter exceeds the radius of the defect this branch transforms into the Caroli--de Gennes--Matricon one. As a result, the minigap in the quasiparticle spectrum increases with the increase in the defect radius. The scenario of the spectrum transformation is generalized for the case of arbitrary vorticity.
Using a nonlinear Schr\"odinger equation including short-range two-body attraction and three-body repulsion, we investigate the spatial distribution of indirect excitons in semiconductor coupled quantum wells. The results obtained can interpret the experimental phenomenon that annular exciton cloud first contracts then expands when the number of confined excitons is increased in impurity potential well, as observed by Lai \emph{et al.} [Lai $et al.$, Science \textbf{303}, 503 (2004)]. In particular, the model reconciles the patterns of exciton rings reported by Butov \emph{et al.} [Butov $et al.$, Nature \textbf{418}, 751 (2002)]. At higher densities, the model predicts much richer patterns, which could be tested by future experiments.
We study the velocity dispersion profiles of the nuclei of NGC 1326, 2685, 5273 and 5838 in the CO first overtone band. There is evidence for a black hole (BH) in NGC 1326 and 5838. Gas is seen flowing out of the nuclear region of NGC 5273. We put upper limits on the nuclear BHs responsible for its activity and that of NGC 2685.
In this paper, we investigate the effect of TDD, as compared to a non-TDD approach, as well as its retainment (or retention) over a time span of (about) six months. To pursue these objectives, we conducted a (quantitative) longitudinal cohort study with 30 novice developers (i.e., third-year undergraduate students in Computer Science). We observed that TDD affects neither the external quality of software products nor developers' productivity. However, we observed that the participants applying TDD produced significantly more tests, with a higher fault-detection capability than those using a non-TDD approach. As for the retainment of TDD, we found that TDD is retained by novice developers for at least six months.
Two approximations of the integral of a class of sinusoidal composite functions, for which an explicit form does not exist, are derived. Numerical experiments show that the proposed approximations yield an error that does not depend on the width of the integration interval. Using such approximations, definite integrals can be computed in almost real-time.
The proposed stochastic model for pedestrian dynamics is based on existing approaches using cellular automata, combined with substantial extensions, to compensate the deficiencies resulting of the discrete grid structure. This agent motion model is extended by both a grid-based path planning and mid-range agent interaction component. The stochastic model proves its capabilities for a quantitative reproduction of the characteristic shape of the common fundamental diagram of pedestrian dynamics. Moreover, effects of self-organizing behavior are successfully reproduced. The stochastic cellular automata approach is found to be adequate with respect to uncertainties in human motion patterns, a feature previously held by artificial noise terms alone.
We prove, using the celebrated result by Spitzer about winding of planar Brownian motion, and the existence of harmonic morphisms $f:M\to{\mathbb S}^1$ representing cohomology classes in $\text{H}^1(M,\mathbb Z)$, that there is a stochastic process $H_t:{\mathcal C}(M)\to{\text{Hom}(\text{H}^1(M;\mathbb R), \mathbb R)}\simeq{\text{H}_1(M;\mathbb R)}$ ($t\in[0,\infty)$), where ${\mathcal C}(M)= \{ \alpha:[0, \infty) \to M :\alpha \,\, \text{is continuous} \}$, which has a multivariate Cauchy distribution i.e. such that for each nontrivial cohomology class $[\omega]\in{\text{H}^1(M;\mathbb R), \mathbb R)}$, represented by a closed 1-form $\omega$, in the de Rham cohomology, the process $A^\omega_t:{\mathcal C}(M)\to\mathbb R\,$ ($t\in[0,\infty)$) with $A^\omega_t(B)=H_t(B)([\omega]),\, B\in{\mathcal C}(M)$ converges in distribution, with respect to Wiener measure on ${\mathcal C}(M)$, to a Cauchy's distribution, with parameter 1. The process describes the ``homological winding" of the Brownian paths in $M$, thus it can be regarded as a generalization of Spitzer result. The last section discusses the asymptotic behavior of holonomy along Brownian paths.
We present a systematic study of how vortices in superfluid films interact with the spatially varying Gaussian curvature of the underlying substrate. The Gaussian curvature acts as a source for a geometric potential that attracts (repels) vortices towards regions of negative (positive) Gaussian curvature independently of the sign of their topological charge. Various experimental tests involving rotating superfluid films and vortex pinning are first discussed for films coating gently curved substrates that can be treated in perturbation theory from flatness. An estimate of the experimental regimes of interest is obtained by comparing the strength of the geometrical forces to the vortex pinning induced by the varying thickness of the film which is in turn caused by capillary effects and gravity. We then present a non-perturbative technique based on conformal mappings that leads an exact solution for the geometric potential as well as the geometric correction to the interaction between vortices. The conformal mapping approach is illustrated by means of explicit calculations of the geometric effects encountered in the study of some strongly curved surfaces and by deriving universal bounds on their strength.
We define cryptographic assumptions applicable to two mistrustful parties who each control two or more separate secure sites between which special relativity guarantees a time lapse in communication. We show that, under these assumptions, unconditionally secure coin tossing can be carried out by exchanges of classical information. We show also, following Mayers, Lo and Chau, that unconditionally secure bit commitment cannot be carried out by finitely many exchanges of classical or quantum information. Finally we show that, under standard cryptographic assumptions, coin tossing is strictly weaker than bit commitment. That is, no secure classical or quantum bit commitment protocol can be built from a finite number of invocations of a secure coin tossing black box together with finitely many additional information exchanges.
We investigate the logical structure of intuitionistic Kripke-Platek set theory IKP, and show that the first-order logic of IKP is intuitionistic first-order logic IQC.
The purpose of this paper is to present some functionalities of the HyperPro System. HyperPro is a hypertext tool which allows to develop Constraint Logic Programming (CLP) together with their documentation. The text editing part is not new and is based on the free software Thot. A HyperPro program is a Thot document written in a report style. The tool is designed for CLP but it can be adapted to other programming paradigms as well. Thot offers navigation and editing facilities and synchronized static document views. HyperPro has new functionalities such as document exportations, dynamic views (projections), indexes and version management. Projection is a mechanism for extracting and exporting relevant pieces of code program or of document according to specific criteria. Indexes are useful to find the references and occurrences of a relation in a document, i.e., where its predicate definition is found and where a relation is used in other programs or document versions and, to translate hyper-texts links into paper references. It still lack importation facilities.
The interest in active matter stimulates the need to generalize thermodynamic description and relations to active matter systems, which are intrinsically out of equilibrium. One important example is the Jarzynski relation, which links the exponential average of work done in an arbitrary process connecting two equilibrium states with the difference of the free energies of these states. Using a simple model system, a single thermal active Ornstein-Uhlenbeck particle in a harmonic potential, we show that if the standard stochastic thermodynamics definition of work is used, the Jarzynski relation is not generally valid for processes between stationary states of active matter systems.
We propose a new factorization pattern for tree-level Yang-Mills (YM) amplitudes, where they decompose into a sum of products of two lower-point amplitudes by setting specific two-point non-planar Mandelstam variables within a rectangular configuration to zero. This approach manifests the hidden zeros of YM amplitudes recently identified. Furthermore, by setting specific Lorentz products involving polarization vectors to zero, the amplitudes further reduce to a sum of products of three currents. These novel factorizations provide a fresh perspective on the structure of YM amplitudes, potentially enhancing our understanding and calculation of these fundamental quantities.
In this report, we report some fundamental results and bounds on the number of messages and storage required to implement barriers using futuristic on-chip optical and RF networks. We prove that it is necessary to maintain a count to at least N (number of threads) in memory, broadcast the barrier id at least once, and if we elect a co-ordinator, we can reduce the number of messages by a factor of O(N ).
We draw an explicit connection between the statistical properties of an entangled two-mode continuous variable (CV) resource and the amount of entanglement that can be dynamically transferred to a pair of non-interacting two-level systems. More specifically, we rigorously reformulate entanglement transfer process by making use of covariance matrix formalism. When the resource state is Gaussian, our method makes the approach to the transfer of quantum correlations much more flexible than in previously considered schemes and allows the straightforward inclusion of the effects of noise affecting the CV system. Moreover, the proposed method reveals that the use of de-Gaussified two-mode states is almost never advantageous for transferring entanglement with respect to the full Gaussian picture, despite the entanglement in the non-Gaussian resource can be much larger than in its Gaussian counterpart. We can thus conclude that the entanglement-transfer map overthrows the "ordering" relations valid at the level of CV resource states.
We present an extensive study of the radiative transfer in dusty galaxies based on Monte Carlo simulations. The main output of these simulations are the attenuation curves ${\cal A}_\lambda$ (i.e. the ratio between the observed, dust extinguished, total intensity to the intrinsic unextinguished one of the galaxy as a function of wavelength). We have explored the dependence of ${\cal A}_\lambda$ on a conspicuous set of quantities (Hubble type, inclination, dust optical thickness, dust distribution and extinction properties) for a large wavelength interval, ranging from 1250\AA to the K band, thus finally providing a comprehensive atlas of dust extinction in galaxies, which is electronically available. This study is particularly suitable for inclusion into galaxy formation evolution models and to directly interpret observational data on high redshift galaxies.
Despite no new physics so far at the LHC, a $Z'$ boson with $m_{Z'} \sim 100$ GeV could still emerge via Drell-Yan (DY) production, $q \bar q \to Z' \to \mu^+ \mu^-$, in the next few years. To unravel the nature of the $Z'$ coupling, we utilize the $c$- and $b$-tagging algorithms developed by ATLAS and CMS to investigate $cg \to c Z'$ at 14 TeV LHC. While light-jet contamination can be eliminated, mistagged $b$-jets cannot be rejected in any of the tagging schemes we adopt. On the other hand, for nonzero $bbZ'$ coupling, far superior $b$-tagging could discover the $bg \to b Z'$ process, where again light-jet mistag can be ruled out, but mistagged $c$-jets cannot yet be excluded. Provided that DY production is discovered soon enough, we find that a simultaneous search for $c g \to c Z'$ and $b g \to b Z'$ can conclusively discern the nature of $Z'$ couplings involved.
We demonstrate that two-time correlation functions, which are generalizations of out-of-time-ordered correlators (OTOCs), can show 'false-flags' of chaos by exhibiting behaviour predicted by random matrix theory even in a system with classically regular dynamics. In particular, we analyze a system of bosons trapped in a double-well potential and probed by a quantum dot which is coupled to the bosons dispersively. This is an integrable system (considered both as separate parts and in total). Despite the continuous time evolution generated by the actual Hamiltonian, we find that the n-fold two-time correlation function for the probe describes an effective stroboscopic or Floquet dynamics whereby the bosons appear to be alternately driven by two different non-commuting Hamiltonians in a manner reminiscent of the Trotterized time evolution that occurs in digital quantum simulation. The classical limit of this effective dynamics can have a nonzero Lyapunov exponent, while the effective level statistics and return probability show traditional signatures of chaotic behaviour. In line with several other recent studies, this work highlights the fact that the behavior of OTOCs and their generalizations must be interpreted with some care.
Growth and roughness of the interface of deposited polymer chains driven by a field onto an impenetrable adsorbing surface are studied by computer simulations in (2+1) dimensions. The evolution of the interface width W shows a crossover from short-time growth described by the exponent beta1 to a long-time growth with exponent beta2 (>beta1). The saturated width increases, i.e., the interface roughens, with the molecular weight Lc, but the roughness exponent alpha (from Ws~L^alpha) becomes negative in contrast to models for particle deposition; alpha depends on the chain length--a nonuniversal scaling with the substrate length L. Roughening and deroughening occur as the field E and the temperature T compete such that Ws=(A+BT)E^-1/2.
We consider the generalization of a matrix integral with arbitrary spectral curve $\rho_0(E)$ to a 0+1D theory of matrix quantum mechanics (MQM). Using recent techniques for 1D quantum systems at large-$N$, we formulate a hydrodynamical effective theory for the eigenvalues. The result is a simple 2D free boson BCFT on a curved background, describing the quantum fluctuations of the eigenvalues around $\rho_0(E)$, which is now the large-$N$ limit of the quantum expectation value of the eigenvalue density operator $\hat{\rho}(E)$. The average over the ensemble of random matrices becomes a quantum expectation value. Equal-time density correlations reproduce the results (including non-perturbative corrections) of random matrix theory. This suggests an interpretation of JT gravity as dual to a $\textit{one-time-point}$ reduction of MQM. As an application, we compute the R\'enyi entropy associated to a bipartition of the eigenvalues. We match a previous result by Hartnoll and Mazenc for the $c=1$ matrix model dual to two-dimensional string theory and extend it to arbitrary $\rho_0(E)$. The hydrodynamical theory provides a clear picture of the emergence of spacetime in two dimensional string theory. The entropy is naturally finite and displays a large amount of short range entanglement, proportional to the microcanonical entropy. We also compute the reduced density matrix for a subset of $n<N$ eigenvalues.
We argue that an ensemble of backgrounds best understands hydrodynamic dispersion relations in a medium with few degrees of freedom and is therefore subject to strong thermal fluctuations. In the linearized regime, dispersion relations become describeable by polynomials with random coefficients. We give a short review of this theory and perform a numerical study of the distribution of the roots of polynomials whose coefficients are of the order of a Knudsen series but fluctuate in accordance with canonical fluctuations of temperature. We find that, remarkably, the analytic structure of the poles of fluctuating dispersion relations is very different from deterministic ones, particularly regarding the distribution of imaginary parts with respect to real components. We argue that this provides evidence that hydrodynamic behavior persists, and is enhanced, by non-perturbative background fluctuations.
Considered are the large $N$, or large intensity, forms of the distribution of the length of the longest increasing subsequences for various models. Earlier work has established that after centring and scaling, the limit laws for these distributions relate to certain distribution functions at the hard edge known from random matrix theory. By analysing the hard to soft edge transition, we supplement and extend results of Baik and Jenkins for the Hammersley model and symmetrisations, which give that the leading correction is proportional to $z^{-2/3}$, where $z^2$ is the intensity of the Poisson rate, and provides a functional form as derivates of the limit law. Our methods give the functional form both in terms of Fredholm operator theoretic quantities, and in terms of Painlev\'e transcendents. For random permutations and their symmetrisations, numerical analysis of exact enumerations and simulations gives compelling evidence that the leading corrections are proportional to $N^{-1/3}$, and moreover provides an approximation to their graphical forms.
Domain Generalization (DG) endeavors to create machine learning models that excel in unseen scenarios by learning invariant features. In DG, the prevalent practice of constraining models to a fixed structure or uniform parameterization to encapsulate invariant features can inadvertently blend specific aspects. Such an approach struggles with nuanced differentiation of inter-domain variations and may exhibit bias towards certain domains, hindering the precise learning of domain-invariant features. Recognizing this, we introduce a novel method designed to supplement the model with domain-level and task-specific characteristics. This approach aims to guide the model in more effectively separating invariant features from specific characteristics, thereby boosting the generalization. Building on the emerging trend of visual prompts in the DG paradigm, our work introduces the novel \textbf{H}ierarchical \textbf{C}ontrastive \textbf{V}isual \textbf{P}rompt (HCVP) methodology. This represents a significant advancement in the field, setting itself apart with a unique generative approach to prompts, alongside an explicit model structure and specialized loss functions. Differing from traditional visual prompts that are often shared across entire datasets, HCVP utilizes a hierarchical prompt generation network enhanced by prompt contrastive learning. These generative prompts are instance-dependent, catering to the unique characteristics inherent to different domains and tasks. Additionally, we devise a prompt modulation network that serves as a bridge, effectively incorporating the generated visual prompts into the vision transformer backbone. Experiments conducted on five DG datasets demonstrate the effectiveness of HCVP, outperforming both established DG algorithms and adaptation protocols.
Fuzzy implication functions have been widely investigated, both in theoretical and practical fields. The aim of this work is to continue previous works related to fuzzy implications constructed by means of non necessarily associative aggregation functions. In order to obtain a more general and flexible context, we extend the class of implications derived by fuzzy negations and t-norms, replacing the latter by general overlap functions. We also investigate their properties, characterization and intersections with other classes of fuzzy implication functions.
The world-sheet S-matrix of the string in AdS5 x S5 has been shown to admit a q-deformation that relates it to the S-matrix of a generalization of the sine-Gordon theory, which arises as the Pohlmeyer reduction of the superstring. Whilst this is a fascinating development the resulting S-matrix is not explicitly unitary. The problem has been known for a long time in the context of S-matrices related to quantum groups. A braiding relation often called "unitarity" actually only corresponds to quantum field theory unitarity when the S-matrix is Hermitian analytic and quantum group S-matrices manifestly violate this. On the other hand, overall consistency of the S-matrix under the bootstrap requires that the deformation parameter is a root of unity and consequently one is forced to perform the "vertex" to IRF, or SOS, transformation on the states to truncate the spectrum consistently. In the IRF formulation unitarity is now manifest and the string S-matrix and the S-matrix of the generalised sine-Gordon theory are recovered in two different limits. In the latter case, expanding the Yang-Baxter equation we find that the tree-level S-matrix of the Pohlmeyer-reduced string should satisfy a modified classical Yang-Baxter equation explaining the apparent anomaly in the perturbative computation. We show that the IRF form of the S-matrix meshes perfectly with the bootstrap equations.
The Higgs boson of 125 GeV requires large stop masses, leading to the large $\mu$-parameter in most cases of gauge mediation. On the other hand, the explanation for the muon $g-2$ anomaly needs small slepton and neutralino/chargino masses. Such disparity in masses may be obtained from a mass splitting of colored and non-colored messenger fields. However, even if the required small slepton and neutralino/chargino masses are realized, all parameter regions consistent with the muon g-2 are excluded by the recent updated ATLAS result on the wino search in the case that the messenger fields are in ${\bf 5}+\bar {\bf 5}$ representations of SU(5). It is also revealed that the messenger fields in ${\bf 10} + \overline{\bf 10}$ or ${\bf 24}$ representation can not explain the muon g-2 anomaly. We show, giving a simple example model, that the above confliction is solved if there is an additional contribution to the Higgs soft mass which makes the $\mu$-parameter small. We also show that the required Higgs B-term for the electroweak symmetry breaking is consistently generated by radiative corrections from gaugino loops.
We use sensitivity analysis to design $\textit{optimality-based}$ discretization (cutting-plane) methods for the global optimization of nonconvex semi-infinite programs (SIPs). We begin by formulating the optimal discretization of SIPs as a max-min problem and propose variants that are more computationally tractable. We then use parametric sensitivity theory to design an efficient method for solving these max-min problems to local optimality and argue this yields valid discretizations without sacrificing global optimality guarantees. Finally, we formulate optimality-based $\textit{generalized}$ discretization of SIPs as max-min problems and design efficient local optimization algorithms to solve them approximately. Numerical experiments on test instances from the literature demonstrate that our new optimality-based discretization methods can significantly reduce the number of iterations for convergence relative to the classical feasibility-based method.
We present Jordan-Brans-Dicke and general scalar-tensor gravitational theory in extra dimensions in an asymptotically flat or anti de Sitter spacetime. We consider a special gravitating, boson field configuration, a $q$-star, in 3, 4, 5 and 6 dimensions, within the framework of the above gravitational theory and find that the parameters of the stable stars are a few per cent different from the case of General Relativity.
Microresonator-based frequency combs (microcombs or Kerr-combs) can potentially miniaturize the numerous applications of conventional frequency combs. A priority is the realization of broad-band (ideally octave spanning) spectra at detectable repetition rates for comb self referencing. However, access to these rates involves pumping larger mode volumes and hence higher threshold powers. Moreover, threshold power sets both the scale for power per comb tooth and also the optical pump. Along these lines, it is shown that a class of resonators having surface-loss-limited Q factors can operate over a wide range of repetition rates with minimal variation in threshold power. A new, surface-loss-limited resonator illustrates the idea. Comb generation on mode spacings ranging from 2.6 GHz to 220 GHz with overall low threshold power (as low as 1 mW) is demonstrated. A record number of comb lines for a microcomb (around 1900) is also observed with pump power of 200 mW. The ability to engineer a wide range of repetition rates with these devices is also used to investigate a recently observed mechanism in microcombs associated with dispersion of subcomb offset frequencies. We observe high-coherence, phase-locking in cases where these offset frequencies are small enough so as to be tuned into coincidence. In these cases, a record-low microcomb phase noise is reported at a level comparable to an open-loop, high-performance microwave oscillator.
In 2003, H\'{e}thelyi and K\"{u}lshammer proposed that if $G$ is a finite group and $p$ is a prime dividing the group order, then $k(G)\geq 2\sqrt{p-1}$, and they proved this conjecture for solvable $G$ and showed that it is sharp for those primes $p$ for which $\sqrt{p-1}$ is an integer. This initiated a flurry of activity, leading to many generalizations and variations of the result; in particular, today the conjecture is known to be true for all finite groups. In this note, we put forward a natural new and stronger conjecture, which is sharp for all primes $p$, and we prove it for solvable groups, and when $p$ is large, also for arbitrary groups.
The superconducting instability in a non-Fermi liquid in $ d>1$ is considered. For a particular form of the single particle spectral function with homogeneous scaling $A(\Lambda k, \Lambda \omega) = \Lambda^{\alpha} A(k, \omega)$ it is shown that the pair susceptibility is also a scaling function of temperature with power defined by $\alpha$. We find three different regimes depending on the scaling constant. The BCS result is recovered for $\alpha = -1$ and it corresponds to a marginal scaling of the coupling constant. For $\alpha > -1$ the superconducting transition happens above some critical coupling. In the opposite case of $\alpha < -1$ for any fixed coupling the system undergoes a transition at low temperatures. Possible implications for theories of high-$T_c$ with a superconducting transition driven by the interlayer Josephson tunneling are discussed. 1 ps file for fig is attached at the bottom of the tex file.
Let $S_g$ be the closed oriented surface of genus g and let $\text{Mod}^{\pm}(S_g)$ be the extended mapping class group of $S_g$. When the genus is at least 5, we prove that $\text{Mod}^{\pm}(S_g)$ can be generated by two torsion elements. One of these generators is an order 2 element, and the other one is an order 4g+2 element.
The challenge of understanding the collective behaviors of social systems can benefit from methods and concepts from physics [1-6], not because humans are similar to electrons, but because certain large-scale behaviors can be understood without an understanding of the small-scale details [7], in much the same way that sound waves can be understood without an understanding of atoms. Democratic elections are one such behavior. Over the past few decades, physicists have explored scaling patterns in voting and the dynamics of political opinion formation, e.g. [8-13]. Here, we define the concepts of negative representation, in which a shift in electorate opinions produces a shift in the election outcome in the opposite direction, and electoral instability, in which an arbitrarily small change in electorate opinions can dramatically swing the election outcome, and prove that unstable elections necessarily contain negatively represented opinions. Furthermore, in the presence of low voter turnout, increasing polarization of the electorate can drive elections through a transition from a stable to an unstable regime, analogous to the phase transition by which some materials become ferromagnetic below their critical temperatures. Empirical data suggest that United States presidential elections underwent such a phase transition in the 1970s and have since become increasingly unstable.
Electron captures by atomic nuclei in dense matter are among the most important processes governing the late evolution of stars, limiting in particular the stability of white dwarfs. Despite considerable progress in the determination of the equation of state of dense Coulomb plasmas, the threshold electron Fermi energies are still generally estimated from the corresponding $Q$ values in vacuum. Moreover, most studies have focused on nonmagnetized matter. However, some white dwarfs are endowed with magnetic fields reaching $10^9$ G. Even more extreme magnetic fields might exist in super Chandrasekhar white dwarfs, the progenitors of overluminous type Ia supernovae like SN 2006gz and SN 2009dc. The roles of the dense stellar medium and magnetic fields on the onset of electron captures and on the structure of white dwarfs are briefly reviewed. New analytical formulas are derived to evaluate the threshold density for the onset of electron captures for arbitrary magnetic fields. Their influence on the structure of white dwarfs is illustrated by simple analytical formulas and numerical calculations.
We present radiation hydrodynamics simulations of the collapse of massive pre-stellar cores. We treat frequency dependent radiative feedback from stellar evolution and accretion luminosity at a numerical resolution down to 1.27 AU. In the 2D approximation of axially symmetric simulations, it is possible for the first time to simulate the whole accretion phase (up to the end of the accretion disk epoch) for the forming massive star and to perform a broad scan of the parameter space. Our simulation series show evidently the necessity to incorporate the dust sublimation front to preserve the high shielding property of massive accretion disks. While confirming the upper mass limit of spherically symmetric accretion, our disk accretion models show a persistent high anisotropy of the corresponding thermal radiation field. This yields to the growth of the highest-mass stars ever formed in multi-dimensional radiation hydrodynamics simulations, far beyond the upper mass limit of spherical accretion. Non-axially symmetric effects are not necessary to sustain accretion. The radiation pressure launches a stable bipolar outflow, which grows in angle with time as presumed from observations. For an initial mass of the pre-stellar host core of 60, 120, 240, and 480 Msun the masses of the final stars formed in our simulations add up to 28.2, 56.5, 92.6, and at least 137.2 Msun respectively.
X-ray timing observations of accreting stellar mass black holes have shown that they can produce signals with such short time scales that we must be probing very close to the innermost stable circular orbit that is predicted by the theory of General Relativity (GR). These signals are quasi-periodic oscillations (QPOs), and both the high-frequency variety (HFQPOs, which have frequencies in the 40-450 Hz range) as well as the 0.1-10 Hz low-frequency type have the potential to provide tests of GR in the strong field limit. An important step on the path to GR tests is to constrain the physical black hole properties, and the straightforward frequency measurements that are possible with X-ray timing may provide one of the cleanest measurements of black hole spins. While current X-ray satellites have uncovered these phenomenona, the HFQPOs are weak signals, and future X-ray timing missions with larger effective area are required for testing the candidate theoretical QPO mechanisms. Another main goal in the study of accreting black holes is to understand the production of relativistic jets. Here, we have also made progress during the past decade by finding clear connections between the radio emission that traces the strength of the jet and the properties of the X-ray emission. With new radio capabilities just coming on-line, continuing detailed X-ray studies of accreting black holes is crucial for continuing to make progress.
We discuss the application of various concepts from the theory of topological dynamical systems to Delone sets and tilings. We consider in particular, the maximal equicontinuous factor of a Delone dynamical system, the proximality relation and the enveloping semigroup of such systems.
We analyze divergencies in 2-point and 3-point functions for noncommutative $\theta$-expanded SU(2)-gauge theory with massless fermions. We show that, after field redefinition and renormalization of couplings, one divergent term remains.
We study the a, b and c coefficients of the isobaric-multiplet mass equation using a macroscopic-microscopic approach developed by P. Moeller and his collaborators in ADNDT 59, 185 (1995) and ADNDT 109-110, 1 (2016). We show that already the macroscopic part of the finite-range liquid-drop model (FRLDM) describes the general trend of the a and b coefficients relatively well, while the staggering behavior of b coefficients for doublets and quartets can be understood in terms of the difference of average proton and neutron pairing energies. The sets of isobaric masses, predicted by the full macroscopic-microscopic approaches, are used to explore the general trends of IMME coefficients up to A=100. We conclude that while the agreement for a coefficients is quite satisfactory, the global approaches have less sensitivity to predict the staggering pattern observed for b coefficients of doublets and quartets. The best set of theoretical b coefficients for multiplets up to about A=100 is used to predict masses of proton-rich nuclei based on the known experimental masses of neutron-rich mirror partners, and subsequently to investigate their one- and two-proton separation energies. The estimated position of the proton-drip line is in fair agreement with known experimental data. These masses are important for simulations of the astrophysical rp-process.
Behavior prediction in dynamic, multi-agent systems is an important problem in the context of self-driving cars, due to the complex representations and interactions of road components, including moving agents (e.g. pedestrians and vehicles) and road context information (e.g. lanes, traffic lights). This paper introduces VectorNet, a hierarchical graph neural network that first exploits the spatial locality of individual road components represented by vectors and then models the high-order interactions among all components. In contrast to most recent approaches, which render trajectories of moving agents and road context information as bird-eye images and encode them with convolutional neural networks (ConvNets), our approach operates on a vector representation. By operating on the vectorized high definition (HD) maps and agent trajectories, we avoid lossy rendering and computationally intensive ConvNet encoding steps. To further boost VectorNet's capability in learning context features, we propose a novel auxiliary task to recover the randomly masked out map entities and agent trajectories based on their context. We evaluate VectorNet on our in-house behavior prediction benchmark and the recently released Argoverse forecasting dataset. Our method achieves on par or better performance than the competitive rendering approach on both benchmarks while saving over 70% of the model parameters with an order of magnitude reduction in FLOPs. It also outperforms the state of the art on the Argoverse dataset.
A procedure of solving nonstationary Schredinger equations in the exact analytic form is elaborated on the basis of exactly solvable stationary models. The exact solutions are employed to study the nonadiabatic geometric phase.
An accurate understanding of a user's query intent can help improve the performance of downstream tasks such as query scoping and ranking. In the e-commerce domain, recent work in query understanding focuses on the query to product-category mapping. But, a small yet significant percentage of queries (in our website 1.5% or 33M queries in 2019) have non-commercial intent associated with them. These intents are usually associated with non-commercial information seeking needs such as discounts, store hours, installation guides, etc. In this paper, we introduce Joint Query Intent Understanding (JointMap), a deep learning model to simultaneously learn two different high-level user intent tasks: 1) identifying a query's commercial vs. non-commercial intent, and 2) associating a set of relevant product categories in taxonomy to a product query. JointMap model works by leveraging the transfer bias that exists between these two related tasks through a joint-learning process. As curating a labeled data set for these tasks can be expensive and time-consuming, we propose a distant supervision approach in conjunction with an active learning model to generate high-quality training data sets. To demonstrate the effectiveness of JointMap, we use search queries collected from a large commercial website. Our results show that JointMap significantly improves both "commercial vs. non-commercial" intent prediction and product category mapping by 2.3% and 10% on average over state-of-the-art deep learning methods. Our findings suggest a promising direction to model the intent hierarchies in an e-commerce search engine.
We give a quasi-complete solution of the (\Delta,N) problem for two well-known families of digraphs used as good models for large interconnection networks. In our study we also relate both families, the New Amsterdam and Manhattan digraphs, with the double-step graphs (or circulant graphs with degree two).
Neutron stars, and magnetars in particular, are known to host the strongest magnetic fields in the Universe. The origin of these strong fields is a matter of controversy. In this preliminary work, via numerical simulations, we study, for the first time in non-ideal general relativistic magnetohydrodynamic (GRMHD) regime, the growth of the magnetic field due to the action of the mean-field dynamo due to sub-scale, unresolved turbulence. The dynamo process, combined with the differential rotation of the (proto-)star, is able to produce an exponential growth of any initial magnetic seed field up to the values required to explain the observations. By varying the dynamo coefficient we obtain different growth rates. We find a quasi-linear dependence of the growth rates on the intensity of the dynamo. Furthermore, the time interval in which exponential growth occurs and the growth rates also seems to depend on the initial configuration of the magnetic field.
We investigate non-standard interaction effects in antineutrino-electron scattering experiments with baselines short enough to ignore standard oscillation phenomena. The setup is free of ambiguities from the interference between new physics and oscillation effects and is sensitive to both semileptonic new physics at the source and purely leptonic new physics in the weak interaction scattering at the detector. We draw on the TEXONO experiment as the model system, extending its analysis of non-standard interaction effects at the detector to include the generally allowed non-standard interaction phase at the detector and both non-universal and flavor changing new physics at the reactor source. We confirm that the current data allows for new physics constraints at the detector of the same order as those currently published, but we find that constraints on the source new physics are at least an order of magnitude weaker. The new physics phase effects are at the 5% level, noticeable in the 90% C.L. contour plots but not significantly affecting the conclusions. Based on projected increase in sensitivity with an upgraded TEXONO experiment, we estimate the improvement of sensitivity to both source and detector non-standard interactions. We find that the bounds on source parameters improve by an order of magnitude, but do not reach parameter space beyond current limits. On the other hand, the detector new physics sensitivity would push current limits by factors 5 to 10 smaller.
Safety-critical systems with neural network components require strong guarantees. While existing neural network verification techniques have shown great progress towards this goal, they cannot prove the absence of software faults in the network implementation. This paper presents NeuroCodeBench - a verification benchmark for neural network code written in plain C. It contains 32 neural networks with 607 safety properties divided into 6 categories: maths library, activation functions, error-correcting networks, transfer function approximation, probability density estimation and reinforcement learning. Our preliminary evaluation shows that state-of-the-art software verifiers struggle to provide correct verdicts, due to their incomplete support of the standard C mathematical library and the complexity of larger neural networks.
By means of fixed point index theory for multi-valued maps, we provide an analogue of the classical Birkhoff--Kellogg Theorem in the context of discontinuous operators acting on affine wedges in Banach spaces. Our theory is fairly general and can be applied, for example, to eigenvalues and parameter problems for ordinary differential equations with discontinuities. We illustrate in details this fact for a class of second order boundary value problem with deviated arguments and discontinuous terms. In a specific example, we explicitly compute the terms that occur in our theory.
Microbial colonies cultured on agar Petri dishes have become a model system to study biological evolution in populations expanding in space. Processes such as clonal segregation and gene surfing have been shown to be affected by interactions between microbial cells and their environment. In this work we investigate the role of mechanical interactions such as cell-surface adhesion. We compare two strains of the bacterium E. coli: a wild-type strain and a "shaved" strain that adheres less to agar. We show that the shaved strain has a selective advantage over the wild type: although both strains grow with the same rate in liquid media, the shaved strain produces colonies that expand faster on agar. This allows the shaved strain outgrow the wild type when both strains compete for space. We hypothesise that, in contrast to a more common scenario in which selective advantage results from increased growth rate, the higher fitness of the shaved strain is caused by reduced adhesion and friction with the agar surface.
We have monitored a Type II outburst of the Be/X-ray binary MXB 0656-072 in a series of pointed RXTE observations during October through December 2003. The source spectrum shows a cyclotron resonance scattering feature at 32.8 +/- 0.5 keV, corresponding to a magnetic field strength of (3.67 +/- 0.06) x 10^12 G and is stable through the outburst and over the pulsar spin phase. The pulsar, with an average pulse period of 160.4 +/- 0.4 s, shows a spin-up of 0.45 s over the duration of the outburst. From optical data, the source distance is estimated to be 3.9 +/- 0.1 kpc and this is used to estimate the X-ray luminosity and a theoretical prediction of the pulsar spin-up during the outburst.
We revisit the two-site Hubbard-Holstein model by using extended phonon coherent states. The nontrivial singlet bipolaron is studied exactly in the whole coupling regime. The ground-state (GS) energy and the double occupancy probability are calculated. The linear entropy is exploited successfully to quantify bipartite entanglement between electrons and their environment phonons, displaying a maximum entanglement of the singlet-bipolaron in strong coupling regime. A dramatic drop in the crossover regime is observed in the GS fidelity and its susceptibility. The bipolaron properties is also characterized classically by correlation functions. It is found that the crossover from a two-site to single-site bipolaron is more abrupt and shifts to a larger electron-phonon coupling strength as electron-electron Coulomb repulsion increases.
A novel quickest detection setting is proposed which is a generalization of the well-known Bayesian change-point detection model. Suppose \{(X_i,Y_i)\}_{i\geq 1} is a sequence of pairs of random variables, and that S is a stopping time with respect to \{X_i\}_{i\geq 1}. The problem is to find a stopping time T with respect to \{Y_i\}_{i\geq 1} that optimally tracks S, in the sense that T minimizes the expected reaction delay E(T-S)^+, while keeping the false-alarm probability P(T<S) below a given threshold \alpha \in [0,1]. This problem formulation applies in several areas, such as in communication, detection, forecasting, and quality control. Our results relate to the situation where the X_i's and Y_i's take values in finite alphabets and where S is bounded by some positive integer \kappa. By using elementary methods based on the analysis of the tree structure of stopping times, we exhibit an algorithm that computes the optimal average reaction delays for all \alpha \in [0,1], and constructs the associated optimal stopping times T. Under certain conditions on \{(X_i,Y_i)\}_{i\geq 1} and S, the algorithm running time is polynomial in \kappa.
We consider random walks in Dirichlet environment (RWDE) on $\Z ^d$, for $ d \geq 3 $, in the sub-ballistic case. We associate to any parameter $ (\alpha_1, ..., \alpha_{2d}) $ of the Dirichlet law a time-change to accelerate the walk. We prove that the continuous-time accelerated walk has an absolutely continuous invariant probability measure for the environment viewed from the particle. This allows to characterize directional transience for the initial RWDE. It solves as a corollary the problem of Kalikow's 0-1 law in the Dirichlet case in any dimension. Furthermore, we find the polynomial order of the magnitude of the original walk's displacement.
We show that the Euler system is not exactly controllable by a finite-dimensional external force. The proof is based on the comparison of the Kolmogorov epsilon-entropy for Holder spaces and for the class of functions that can be obtained by solving the 2D Euler equations with various right-hand sides.
The theory of small-amplitude waves propagating across a blood vessel junction has been well established with linear analysis. In this study we consider the propagation of large-amplitude, nonlinear waves (i.e. shocks and rarefactions) through a junction from a parent vessel into two (identical) daughter vessels using a combination of three approaches: numerical computations using a Godunov method with patching across the junction, analysis of a nonlinear Riemann problem in the neighbourhood of the junction and an analytical theory which extends the linear analysis to the following order in amplitude. A unified picture emerges: an abrupt (prescribed) increase in pressure at the inlet to the parent vessel generates a propagating shock wave along the parent vessel which interacts with the junction. For modest driving, this shock wave divides into propagating shock waves along the two daughter vessels and reflects a rarefaction wave back towards the inlet. However, for larger driving the reflected rarefaction wave becomes transcritical, generating an additional shock wave. Just beyond criticality this new shock wave has zero speed, pinned to the junction, but for further increases in driving this additional shock divides into two new propagating shock waves in the daughter vessels.
We present optical spectroscopic and Swift UVOT/XRT observations of the X-ray and UV/optical bright tidal disruption event (TDE) AT 2018fyk/ASASSN-18ul discovered by ASAS-SN. The Swift lightcurve is atypical for a TDE, entering a plateau after $\sim$40 days of decline from peak. After 80 days the UV/optical lightcurve breaks again to decline further, while the X-ray emission becomes brighter and harder. In addition to broad H, He and potentially O/Fe lines, narrow emission lines emerge in the optical spectra during the plateau phase. We identify both high ionisation (O III) and low ionisation (Fe II) lines, which are visible for $\sim$45 days. We similarly identify Fe II lines in optical spectra of ASASSN-15oi 330 d after discovery, indicating that a class of Fe-rich TDEs exists. The spectral similarity between AT 2018fyk, narrow-line Seyfert 1 galaxies and some extreme coronal line emitters suggests that TDEs are capable of creating similar physical conditions in the nuclei of galaxies. The Fe II lines can be associated with the formation of a compact accretion disk, as the emergence of low ionisation emission lines requires optically thick, high density gas. Taken together with the plateau in X-ray and UV/optical luminosity this indicates that emission from the central source is efficiently reprocessed into UV/optical wavelengths. Such a two-component lightcurve is very similar to that seen in the TDE candidate ASASSN-15lh, and is a natural consequence of a highly relativistic orbital pericenter.
Janus colloids propelled by light, e.g., thermophoretic particles, offer promising prospects as artificial microswimmers. However, their swimming behavior and its dependence on fluid properties and fluid-colloid interactions remain poorly understood. Here, we investigate the behavior of a thermophoretic Janus colloid in its own temperature gradient using numerical simulations. The dissipative particle dynamics method with energy conservation is used to investigate the behavior in non-ideal and ideal-gas like fluids for different fluid-colloid interactions, boundary conditions, and temperature-controlling strategies. The fluid-colloid interactions appear to have a strong effect on the colloid behavior, since they directly affect heat exchange between the colloid surface and the fluid. The simulation results show that a reduction of the heat exchange at the fluid-colloid interface leads to an enhancement of colloid's thermophoretic mobility. The colloid behavior is found to be different in non-ideal and ideal fluids, suggesting that fluid compressibility plays a significant role. The flow field around the colloid surface is found to be dominated by a source-dipole, in agreement with the recent theoretical and simulation predictions. Finally, different temperature-control strategies do not appear to have a strong effect on the colloid's swimming velocity.
The physical characterization of exoplanets will require to take spectra at several orbital positions. For that purpose, a direct imaging capability is necessary. Direct imaging requires an efficient stellar suppression mechanism, associated with an ultrasmooth telescope. We show that before future large space missions (interferometer, 4-8 m class coronograph, external occulter or Fresnel imager), direct imaging of giant planets and close-by super-Earth are at the cross-road of a high scientific interest and a reasonable feasibility. The scientific interest lies in the fact that super-Earths share common geophysical attributes with Earths. They already begin to be detected by radial velocity (RV) and, together with giant planets, they have a larger area than Earths, making them detectable with a 1.5-2 m class telescope in reflected light. We propose such a (space) telescope be a first step before large direct imaging missions.
Motivated by recent questions about the extension of Courant's nodal domain theorem, we revisit a theorem published by C. Sturm in 1836, which deals with zeros of linear combination of eigenfunctions of Sturm-Liouville problems. Although well known in the nineteenth century, this theorem seems to have been ignored or forgotten by some of the specialists in spectral theory since the second half of the twentieth-century. Although not specialists in History of Sciences, we have tried to put these theorems into the context of nineteenth century mathematics.
Mounts and hills played a predominant role in all pre-Hispanic Andean cultures, especially for the Inca culture. Through the use of georeferenced orthophotography, we found that the Inca site, Ruinas de Chada, represents the origin of a radial ceque system with alignments connecting, at one end high peaks of mounts at the Andes, and on the other end the summit of small hills in which important shrines were built. These alignments extend over two hundred kilometers, thus we propose that the information codified on this shrine was based on an ancient geodetic science. A sacred geometric relationship is encrypted in the pattern formed by the position of all shrines with high accuracy, in which the Andean Chakana symbol is represented. These findings suggest that the valley of Chada could have been the sacred center of Collasuyu.
Of the roughly 3000 neutron stars known, only a handful have sub-stellar companions. The most famous of these are the low-mass planets around the millisecond pulsar B1257+12. New evidence indicates that observational biases could still hide a wide variety of planetary systems around most neutron stars. We consider the environment and physical processes relevant to neutron star planets, in particular the effect of X-ray irradiation and the relativistic pulsar wind on the planetary atmosphere. We discuss the survival time of planet atmospheres and the planetary surface conditions around different classes of neutron stars, and define a neutron star habitable zone. Depending on as-yet poorly constrained aspects of the pulsar wind, both Super-Earths around B1257+12 could lie within its habitable zone.
Speech quality in online conferencing applications is typically assessed through human judgements in the form of the mean opinion score (MOS) metric. Since such a labor-intensive approach is not feasible for large-scale speech quality assessments in most settings, the focus has shifted towards automated MOS prediction through end-to-end training of deep neural networks (DNN). Instead of training a network from scratch, we propose to leverage the speech representations from the pre-trained wav2vec-based XLS-R model. However, the number of parameters of such a model exceeds task-specific DNNs by several orders of magnitude, which poses a challenge for resulting fine-tuning procedures on smaller datasets. Therefore, we opt to use pre-trained speech representations from XLS-R in a feature extraction rather than a fine-tuning setting, thereby significantly reducing the number of trainable model parameters. We compare our proposed XLS-R-based feature extractor to a Mel-frequency cepstral coefficient (MFCC)-based one, and experiment with various combinations of bidirectional long short term memory (Bi-LSTM) and attention pooling feedforward (AttPoolFF) networks trained on the output of the feature extractors. We demonstrate the increased performance of pre-trained XLS-R embeddings in terms a reduced root mean squared error (RMSE) on the ConferencingSpeech 2022 MOS prediction task.
The recent paper by Byrd & Lipton (2019), based on empirical observations, raises a major concern on the impact of importance weighting for the over-parameterized deep learning models. They observe that as long as the model can separate the training data, the impact of importance weighting diminishes as the training proceeds. Nevertheless, there lacks a rigorous characterization of this phenomenon. In this paper, we provide formal characterizations and theoretical justifications on the role of importance weighting with respect to the implicit bias of gradient descent and margin-based learning theory. We reveal both the optimization dynamics and generalization performance under deep learning models. Our work not only explains the various novel phenomenons observed for importance weighting in deep learning, but also extends to the studies where the weights are being optimized as part of the model, which applies to a number of topics under active research.
This study introduces Polyp-DDPM, a diffusion-based method for generating realistic images of polyps conditioned on masks, aimed at enhancing the segmentation of gastrointestinal (GI) tract polyps. Our approach addresses the challenges of data limitations, high annotation costs, and privacy concerns associated with medical images. By conditioning the diffusion model on segmentation masks-binary masks that represent abnormal areas-Polyp-DDPM outperforms state-of-the-art methods in terms of image quality (achieving a Frechet Inception Distance (FID) score of 78.47, compared to scores above 83.79) and segmentation performance (achieving an Intersection over Union (IoU) of 0.7156, versus less than 0.6694 for synthetic images from baseline models and 0.7067 for real data). Our method generates a high-quality, diverse synthetic dataset for training, thereby enhancing polyp segmentation models to be comparable with real images and offering greater data augmentation capabilities to improve segmentation models. The source code and pretrained weights for Polyp-DDPM are made publicly available at https://github.com/mobaidoctor/polyp-ddpm.
We show that any $n$-bit string can be recovered with high probability from $\exp(\widetilde{O}(n^{1/5}))$ independent random subsequences.