text
stringlengths
89
2.49k
category
stringclasses
19 values
Human Contrast Threshold and Astronomical Visibility: The standard visibility model in light pollution studies is the formula of Hecht (1947), as used e.g. by Schaefer (1990). However it is applicable only to point sources and is shown to be of limited accuracy. A new visibility model is presented for uniform achromatic targets of any size against background luminances ranging from zero to full daylight, produced by a systematic procedure applicable to any appropriate data set (e.g Blackwell (1946)), and based on a simple but previously unrecognized empirical relation between contrast threshold and adaptation luminance. The scotopic luminance correction for variable spectral radiance (colour index) is calculated. For point sources the model is more accurate than Hecht's formula and is verified using telescopic data collected at Mount Wilson by Bowen (1947), enabling the sky brightness at that time to be determined. The result is darker than the calculation by Garstang (2004), implying that light pollution grew more rapidly in subsequent decades than has been supposed. The model is applied to the nebular observations of William Herschel, enabling his visual performance to be quantified. Proposals are made regarding sky quality indicators for public use.
astro-ph_IM
Design and Implementation of the wvrgcal Program: This memo describes the software engineering and technical details of the design and implementation of the wvrgcal program and associated libraries. This program performs off-line correction of atmospheric phase fluctuations in ALMA observations, using the 183 GHz Water Vapour Radiometers (WVRs) installed on the ALMA 12 m dishes. The memo can be used as a guide for detailed study of the source code of the program for purposes of further development or maintenance.
astro-ph_IM
Upgrade of the VERITAS Cherenkov Telescope Array: The VERITAS Cherenkov telescope array has been fully operational since Fall 2007 and has fulfilled or outperformed its design specifications. We are preparing an upgrade program with the goal to lower the energy threshold and improve the sensitivity of VERITAS at all accessible energies. In the baseline program of the upgrade we will relocate one of the four telescopes, replace the photo-sensors by higher efficiency photomultipliers and install a new trigger system. In the enhanced program of the upgrade we foresee, in addition, the construction of a fifth telescope and installation of an active mirror alignment system.
astro-ph_IM
Laboratory gas-phase infrared spectra of two astronomically relevant PAH cations: diindenoperylene, C$_{32}$H$_{16}$$^+$ and dicoronylene, C$_{48}$H$_{20}$$^+$: The first gas-phase infrared spectra of two isolated astronomically relevant and large PAH cations - diindenoperylene (DIP) and dicoronylene (DC) - in the 530$-$1800 cm$^{-1}$ (18.9$-$5.6 $\mu$m) range - are presented. Vibrational band positions are determined for comparison to the aromatic infrared bands (AIBs). The spectra are obtained via infrared multiphoton dissociation (IRMPD) spectroscopy of ions stored in a quadrupole ion trap (QIT) using the intense and tunable radiation of the free electron laser for infrared experiments (FELIX). DIP$^{+}$ shows its main absorption peaks at 737 (13.57), 800 (12.50), 1001 (9.99), 1070 (9.35), 1115 (8.97), 1152 (8.68), 1278 (7.83), 1420 (7.04) and 1550 (6.45) cm$^{-1}$($\mu$m), in good agreement with DFT calculations that are uniformly scaled to take anharmonicities into account. DC$^+$ has its main absorption peaks at 853 (11.72), 876 (11.42), 1032 (9.69), 1168 (8.56), 1300 (7.69), 1427 (7.01) and 1566 (6.39) cm$^{-1}$($\mu$m), that also agree well with the scaled DFT results presented here. The DIP$^+$ and DC$^+$ spectra are compared with the prominent infrared features observed towards NGC 7023. This results both in matches and clear deviations. Moreover, in the 11.0$-$14.0 $\mu$m region, specific bands can be linked to CH out-of-plane (oop) bending modes of different CH edge structures in large PAHs. The molecular origin of these findings and their astronomical relevance are discussed.
astro-ph_IM
Temperature dependence of radiation damage annealing of Silicon Photomultipliers: The last decade has increasingly seen the use of silicon photomultipliers (SiPMs) instead of photomultiplier tubes (PMTs). This is due to various advantages of the former on the latter like its smaller size, lower operating voltage, higher detection efficiency, insensitivity to magnetic fields and mechanical robustness to launch vibrations. All these features make SiPMs ideal for use on space based experiments where the detectors require to be compact, lightweight and capable of surviving launch conditions. A downside with the use of this novel type of detector in space conditions is its susceptibility to radiation damage. In order to understand the lifetime of SiPMs in space, both the damage sustained due to radiation as well as the subsequent recovery, or annealing, from this damage have to be studied. Here we present these studies for three different types of SiPMs from the Hamamatsu S13360 series. Both their behaviour after sustaining radiation equivalent to 2 years in low earth orbit in a typical mission is presented, as well as the recovery of these detectors while stored in different conditions. The storage conditions varied in temperature as well as in operating voltage. The study found that the annealing depends significantly on the temperature of the detectors with those stored at high temperatures recovering significantly faster and at recovering closer to the original performance. Additionally, no significant effect from a reasonable bias voltage on the annealing was observed. Finally the annealing rate as a function of temperature is presented along with various operating strategies for the future SiPM based astrophysical detector POLAR-2 as well as for future SiPM based space borne missions.
astro-ph_IM
Stratospheric Imaging of Polar Mesospheric Clouds: A New Window on Small-Scale Atmospheric Dynamics: Instabilities and turbulence extending to the smallest dynamical scales play important roles in the deposition of energy and momentum by gravity waves throughout the atmosphere. However, these dynamics and their effects have been impossible to quantify to date due to lack of observational guidance. Serendipitous optical images of polar mesospheric clouds at ~82 km obtained by star cameras aboard a cosmology experiment deployed on a stratospheric balloon provide a new observational tool, revealing instability and turbulence structures extending to spatial scales < 20 m. At 82 km, this resolution provides sensitivity extending to the smallest turbulence scale not strongly influenced by viscosity: the "inner scale" of turbulence, $l_0\sim$10($\nu^3$/$\epsilon$)$^{1/4}$. Such images represent a new window into small-scale dynamics that occur throughout the atmosphere but are impossible to observe in such detail at any other altitude. We present a sample of images revealing a range of dynamics features, and employ numerical simulations that resolve these dynamics to guide our interpretation of several observed events.
astro-ph_IM
Cosmological surveys with multi-object spectrographs: Multi-object spectroscopy has been a key technique contributing to the current era of 'precision cosmology'. From the first exploratory surveys of the large-scale structure and evolution of the universe to the current generation of superbly detailed maps spanning a wide range of redshifts, multi-object spectroscopy has been a fundamentally important tool for mapping the rich structure of the cosmic web and extracting cosmological information of increasing variety and precision. This will continue to be true for the foreseeable future, as we seek to map the evolving geometry and structure of the universe over the full extent of cosmic history in order to obtain the most precise and comprehensive measurements of cosmological parameters. Here I briefly summarize the contributions that multi-object spectroscopy has made to cosmology so far, then review the major surveys and instruments currently in play and their prospects for pushing back the cosmological frontier. Finally, I examine some of the next generation of instruments and surveys to explore how the field will develop in coming years, with a particular focus on specialised multi-object spectrographs for cosmology and the capabilities of multi-object spectrographs on the new generation of extremely large telescopes.
astro-ph_IM
Processing Images from Multiple IACTs in the TAIGA Experiment with Convolutional Neural Networks: Extensive air showers created by high-energy particles interacting with the Earth atmosphere can be detected using imaging atmospheric Cherenkov telescopes (IACTs). The IACT images can be analyzed to distinguish between the events caused by gamma rays and by hadrons and to infer the parameters of the event such as the energy of the primary particle. We use convolutional neural networks (CNNs) to analyze Monte Carlo-simulated images from the telescopes of the TAIGA experiment. The analysis includes selection of the images corresponding to the showers caused by gamma rays and estimating the energy of the gamma rays. We compare performance of the CNNs using images from a single telescope and the CNNs using images from two telescopes as inputs.
astro-ph_IM
Equalizing the Pixel Response of the Imaging Photoelectric Polarimeter On-Board the IXPE Mission: The Gas Pixel Detector is a gas detector, sensitive to the polarization of X-rays, currently flying on-board IXPE - the first observatory dedicated to X-ray polarimetry. It detects X-rays and their polarization by imaging the ionization tracks generated by photoelectrons absorbed in the sensitive volume, and then reconstructing the initial direction of the photoelectrons. The primary ionization charge is multiplied and ultimately collected on a finely-pixellated ASIC specifically developed for X-ray polarimetry. The signal of individual pixels is processed independently and gain variations can be substantial, of the order of 20%. Such variations need to be equalized to correctly reconstruct the track shape, and therefore its polarization direction. The method to do such equalization is presented here and is based on the comparison between the mean charge of a pixel with respect to the other pixels for equivalent events. The method is shown to finely equalize the response of the detectors on board IXPE, allowing a better track reconstruction and energy resolution, and can in principle be applied to any imaging detector based on tracks.
astro-ph_IM
An Advanced Atmospheric Dispersion Corrector: The Magellan Visible AO Camera: In addition to the BLINC/MIRAC IR science instruments, the Magellan adaptive secondary AO system will have an EEV CCD47 that can be used both for visible AO science and as a wide-field acquisition camera. The effects of atmospheric dispersion on the elongation of the diffraction limited Magellan adaptive optics system point spread function (PSF) are significant in the near IR. This elongation becomes particularly egregious at visible wavelengths, culminating in a PSF that is 2000\{mu}m long in one direction and diffraction limited (30-60 \{mu}m) in the other over the wavelength band 0.5-1.0\{mu}m for a source at 45\pm zenith angle. The planned Magellan AO system consists of a deformable secondary mirror with 585 actuators. This number of actuators should be sufficient to nyquist sample the atmospheric turbulence and correct images to the diffraction limit at wavelengths as short as 0.7\{mu}m, with useful science being possible as low as 0.5\{mu}m. In order to achieve diffraction limited performance over this broad band, 2000\{mu}m of lateral color must be corrected to better than 10\{mu}m. The traditional atmospheric dispersion corrector (ADC) consists of two identical counter-rotating cemented doublet prisms that correct the primary chromatic aberration. We propose two new ADC designs: the first consisting of two identical counter-rotating prism triplets, and the second consisting of two pairs of cemented counter-rotating prism doublets that use both normal dispersion and anomalous dispersion glass in order to correct both primary and secondary chromatic aberration. The two designs perform 58% and 68%, respectively, better than the traditional two-doublet design. We also present our design for a custom removable wide-field lens that will allow our CCD47 to switch back and forth between an 8.6" FOV for AO science and a 28.5" FOV for acquisition.
astro-ph_IM
Design and Initial Performance of the Prototype for the BEACON Instrument for Detection of Ultrahigh Energy Particles: The Beamforming Elevated Array for COsmic Neutrinos (BEACON) is a planned neutrino telescope designed to detect radio emission from upgoing air showers generated by ultrahigh energy tau neutrino interactions in the Earth. This detection mechanism provides a measurement of the tau flux of cosmic neutrinos. We have installed an 8-channel prototype instrument at high elevation at Barcroft Field Station, which has been running since 2018, and consists of 4 dual-polarized antennas sensitive between 30-80 MHz, whose signals are filtered, amplified, digitized, and saved to disk using a custom data acquisition system (DAQ). The BEACON prototype is at high elevation to maximize effective volume and uses a directional beamforming trigger to improve rejection of anthropogenic background noise at the trigger level. Here we discuss the design, construction, and calibration of the BEACON prototype instrument. We also discuss the radio frequency environment observed by the instrument, and categorize the types of events seen by the instrument, including a likely cosmic ray candidate event.
astro-ph_IM
Detecting and analysing the topology of the cosmic web with spatial clustering algorithms I: Methods: In this paper we explore the use of spatial clustering algorithms as a new computational approach for modeling the cosmic web. We demonstrate that such algorithms are efficient in terms of computing time needed. We explore three distinct spatial methods which we suitably adjust for (i) detecting the topology of the cosmic web and (ii) categorizing various cosmic structures as voids, walls, clusters and superclusters based on a variety of topological and physical criteria such as the physical distance between objects, their masses and local densities. The methods explored are (1) a new spatial method called Gravity Lattice ; (2) a modified version of another spatial clustering algorithm, the ABACUS; and (3) the well known spatial clustering algorithm HDBSCAN. We utilize HDBSCAN in order to detect cosmic structures and categorize them using their overdensity. We demonstrate that the ABACUS method can be combined with the classic DTFE method to obtain similar results in terms of the achieved accuracy with about an order of magnitude less computation time. To further solidify our claims, we draw insights from the computer science domain and compare the quality of the results with and without the application of our method. Finally, we further extend our experiments and verify their effectiveness by showing their ability to scale well with different cosmic web structures that formed at different redshifts.
astro-ph_IM
Seeing Science: The ability to represent scientific data and concepts visually is becoming increasingly important due to the unprecedented exponential growth of computational power during the present digital age. The data sets and simulations scientists in all fields can now create are literally thousands of times as large as those created just 20 years ago. Historically successful methods for data visualization can, and should, be applied to today's huge data sets, but new approaches, also enabled by technology, are needed as well. Increasingly, "modular craftsmanship" will be applied, as relevant functionality from the graphically and technically best tools for a job are combined as-needed, without low-level programming.
astro-ph_IM
How to Scale a Code in the Human Dimension: As scientists' needs for computational techniques and tools grow, they cease to be supportable by software developed in isolation. In many cases, these needs are being met by communities of practice, where software is developed by domain scientists to reach pragmatic goals and satisfy distinct and enumerable scientific goals. We present techniques that have been successful in growing and engaging communities of practice, specifically in the yt and Enzo communities.
astro-ph_IM
Background assessment for the TREX Dark Matter experiment: TREX-DM is conceived to look for low-mass Weakly Interacting Massive Particles (WIMPs) using a gas Time Projection Chamber equipped with micromegas readout planes at the Canfranc Underground Laboratory. The detector can hold in the active volume 20 l of pressurized gas up to 10 bar, corresponding to 0.30 kg of Ar or 0.16 kg of Ne. The micromegas are read with a self-triggered acquisition, allowing for thresholds below 0.4 keV (electron equivalent). A low background level in the lowest energy region is another essential requirement. To assess the expected background, all the relevant sources have been considered, including the measured fluxes of gamma radiation, muons and neutrons at the Canfranc Laboratory, together with the activity of most of the components used in the detector and ancillary systems, obtained in a complete assay program. The background contributions have been simulated by means of a dedicated application based on Geant4 and a custom-made code for the detector response. The background model developed for the detector presently installed in Canfranc points to levels from 1 to 10 counts keV-1 kg-1 d-1 in the region of interest, making TREX-DM competitive in the search for low-mass WIMPs. A roadmap to further decrease it down to 0.1 counts keV-1 kg-1 d-1 is underway.
astro-ph_IM
Neutrino Astronomy - A Review of Future Experiments: Current generation neutrino telescopes cover an energy range from about 10 GeV to beyond $10^9$ GeV. IceCube sets the scale for future experiments to make improvements. Strategies for future upgrades will be discussed in three energy ranges. At the low-energy end, an infill detector to IceCube's DeepCore would add sensitivity in the energy range from a few to a few tens of GeV with the primary goal of measuring the neutrino mass hierarchy. In the central energy range of classical optical neutrino telescopes, next generation detectors are being pursued in the Mediterranean and at Lake Baikal. The KM3NeT detector in its full scale would establish a substantial increase in sensitivity over IceCube. At the highest energies, radio detectors in ice are among the most promising and pursued technologies to increase exposure at $10^9$ GeV by more than an order of magnitude compared to IceCube.
astro-ph_IM
Towards an astronomical foundation model for stars with a Transformer-based model: Rapid strides are currently being made in the field of artificial intelligence using Transformer-based models like Large Language Models (LLMs). The potential of these methods for creating a single, large, versatile model in astronomy has not yet been explored. In this work, we propose a framework for data-driven astronomy that uses the same core techniques and architecture as used by LLMs. Using a variety of observations and labels of stars as an example, we build a Transformer-based model and train it in a self-supervised manner with cross-survey data sets to perform a variety of inference tasks. In particular, we demonstrate that a $\textit{single}$ model can perform both discriminative and generative tasks even if the model was not trained or fine-tuned to do any specific task. For example, on the discriminative task of deriving stellar parameters from Gaia XP spectra, we achieve an accuracy of 47 K in $T_\mathrm{eff}$, 0.11 dex in $\log{g}$, and 0.07 dex in $[\mathrm{M/H}]$, outperforming an expert $\texttt{XGBoost}$ model in the same setting. But the same model can also generate XP spectra from stellar parameters, inpaint unobserved spectral regions, extract empirical stellar loci, and even determine the interstellar extinction curve. Our framework demonstrates that building and training a $\textit{single}$ foundation model without fine-tuning using data and parameters from multiple surveys to predict unmeasured observations and parameters is well within reach. Such "Large Astronomy Models" trained on large quantities of observational data will play a large role in the analysis of current and future large surveys.
astro-ph_IM
Modelling multimodal photometric redshift regression with noisy observations: In this work, we are trying to extent the existing photometric redshift regression models from modeling pure photometric data back to the spectra themselves. To that end, we developed a PCA that is capable of describing the input uncertainty (including missing values) in a dimensionality reduction framework. With this "spectrum generator" at hand, we are capable of treating the redshift regression problem in a fully Bayesian framework, returning a posterior distribution over the redshift. This approach allows therefore to approach the multimodal regression problem in an adequate fashion. In addition, input uncertainty on the magnitudes can be included quite naturally and lastly, the proposed algorithm allows in principle to make predictions outside the training values which makes it a fascinating opportunity for the detection of high-redshifted quasars.
astro-ph_IM
Long-baseline horizontal radio-frequency transmission through polar ice: We report on analysis of englacial radio-frequency (RF) pulser data received over horizontal baselines of 1--5 km, based on broadcasts from two sets of transmitters deployed to depths of up to 1500 meters at the South Pole. First, we analyze data collected usingtwo RF bicone transmitters 1400 meters below the ice surface, and frozen into boreholes drilled for the IceCube experiment in 2011. Additionally, in Dec., 2018, a fat-dipole antenna, fed by one of three high-voltage (~1 kV), fast (~(1-5 ns)) signal generators was lowered into the 1700-m deep icehole drilled for the South Pole Ice Core Experiment (SPICE), approximately 3 km from the geographic South Pole. Signals from transmitters were recorded on the five englacial multi-receiver ARA stations, with receiver depths between 60--200 m. We confirm the long, >1 km RF electric field attenuation length, test our observed signal arrival timing distributions against models, and measure birefringent asymmetries at the 0.15% level.
astro-ph_IM
Accurate, Meshless Methods for Magneto-Hydrodynamics: Recently, we developed a pair of meshless finite-volume Lagrangian methods for hydrodynamics: the 'meshless finite mass' (MFM) and 'meshless finite volume' (MFV) methods. These capture advantages of both smoothed-particle hydrodynamics (SPH) and adaptive mesh-refinement (AMR) schemes. Here, we extend these to include ideal magneto-hydrodynamics (MHD). The MHD equations are second-order consistent and conservative. We augment these with a divergence-cleaning scheme, which maintains div*B~0 to high accuracy. We implement these in the code GIZMO, together with a state-of-the-art implementation of SPH MHD. In every one of a large suite of test problems, the new methods are competitive with moving-mesh and AMR schemes using constrained transport (CT) to ensure div*B=0. They are able to correctly capture the growth and structure of the magneto-rotational instability (MRI), MHD turbulence, and the launching of magnetic jets, in some cases converging more rapidly than AMR codes. Compared to SPH, the MFM/MFV methods exhibit proper convergence at fixed neighbor number, sharper shock capturing, and dramatically reduced noise, div*B errors, and diffusion. Still, 'modern' SPH is able to handle most of our tests, at the cost of much larger kernels and 'by hand' adjustment of artificial diffusion parameters. Compared to AMR, the new meshless methods exhibit enhanced 'grid noise' but reduced advection errors and numerical diffusion, velocity-independent errors, and superior angular momentum conservation and coupling to N-body gravity solvers. As a result they converge more slowly on some problems (involving smooth, slowly-moving flows) but more rapidly on others (involving advection or rotation). In all cases, divergence-control beyond the popular Powell 8-wave approach is necessary, or else all methods we consider will systematically converge to unphysical solutions.
astro-ph_IM
Next-generation telescopes with curved focal surface for ultra-low surface brightness surveys: In spite of major advances in both ground- and space-based instrumentation, the ultra-low-surface brightness universe (ULSB) still remains a largely unexplored volume in observational parameter space. ULSB observations provide unique constraints on a wide variety of objects, from the Zodiacal light all the way to the optical cosmological background radiation, through dust cirri, mass loss shells in giant stars, LSB galaxies and the intracluster light. These surface brightness levels (>28-29 mag arcsec^-2) are observed by maximising the efficiency of the surveys and minimising or removing the systematics arising in the measurement of surface brightness. Based on full-system photon Monte Carlo simulations, we present here the performance of a ground-based telescope aimed at carrying out ULSB observations, with a curved focal surface design. Its off-axis optical design maximises the field of view while minimising the focal ratio. No lenses are used, as their multiple internal scatterings increase the wings of the point spread function (PSF), and the usual requirement of a flat focal plane is relaxed through the use of curved CCD detectors. The telescope has only one unavoidable single refractive surface, the cryostat window, and yet it delivers a PSF with ultra-compact wings, which allows the detection, for a given exposure time, of surface brightness levels nearly three orders of magnitude fainter than any other current telescope.
astro-ph_IM
Systematic biases in low frequency radio interferometric data due to calibration: the LOFAR EoR case: The redshifted 21 cm line of neutral hydrogen is a promising probe of the Epoch of Reionization (EoR). However, its detection requires a thorough understanding and control of the systematic errors. We study two systematic biases observed in the LOFAR EoR residual data after calibration and subtraction of bright discrete foreground sources. The first effect is a suppression in the diffuse foregrounds, which could potentially mean a suppression of the 21 cm signal. The second effect is an excess of noise beyond the thermal noise. The excess noise shows fluctuations on small frequency scales, and hence it can not be easily removed by foreground removal or avoidance methods. Our analysis suggests that sidelobes of residual sources due to the chromatic point spread function and ionospheric scintillation can not be the dominant causes of the excess noise. Rather, both the suppression of diffuse foregrounds and the excess noise can occur due to calibration with an incomplete sky model containing predominantly bright discrete sources. We show that calibrating only on bright sources can cause suppression of other signals and introduce an excess noise in the data. The levels of the suppression and excess noise depend on the relative flux of sources which are not included in the model with respect to the flux of modeled sources. We discuss possible solutions such as using only long baselines to calibrate the interferometric gain solutions as well as simultaneous multi-frequency calibration along with their benefits and shortcomings.
astro-ph_IM
Iris: an Extensible Application for Building and Analyzing Spectral Energy Distributions: Iris is an extensible application that provides astronomers with a user-friendly interface capable of ingesting broad-band data from many different sources in order to build, explore, and model spectral energy distributions (SEDs). Iris takes advantage of the standards defined by the International Virtual Observatory Alliance, but hides the technicalities of such standards by implementing different layers of abstraction on top of them. Such intermediate layers provide hooks that users and developers can exploit in order to extend the capabilities provided by Iris. For instance, custom Python models can be combined in arbitrary ways with the Iris built-in models or with other custom functions. As such, Iris offers a platform for the development and integration of SED data, services, and applications, either from the user's system or from the web. In this paper we describe the built-in features provided by Iris for building and analyzing SEDs. We also explore in some detail the Iris framework and software development kit, showing how astronomers and software developers can plug their code into an integrated SED analysis environment.
astro-ph_IM
A simple and efficient solver for self-gravity in the DISPATCH astrophysical simulation framework: We describe a simple and effective algorithm for solving Poisson's equation in the context of self-gravity within the DISPATCH astrophysical fluid framework. The algorithm leverages the fact that DISPATCH stores multiple time slices and uses asynchronous time-stepping to produce a scheme that does not require any explicit global communication or sub-cycling, only the normal, local communication between patches and the iterative solution to Poisson's equation. We demonstrate that the implementation is suitable for both collections of patches of a single resolution and for hierarchies of adaptively resolved patches. Benchmarks are presented that demonstrate the accuracy, effectiveness and efficiency of the scheme.
astro-ph_IM
The conceptual design of GMagAO-X: visible wavelength high contrast imaging with GMT: We present the conceptual design of GMagAO-X, an extreme adaptive optics system for the 25 m Giant Magellan Telescope (GMT). We are developing GMagAO-X to be available at or shortly after first-light of the GMT, to enable early high contrast exoplanet science in response to the Astro2020 recommendations. A key science goal is the characterization of nearby potentially habitable terrestrial worlds. GMagAO-Xis a woofer-tweeter system, with integrated segment phasing control. The tweeter is a 21,000 actuator segmented deformable mirror, composed of seven 3000 actuator segments. A multi-stage wavefront sensing system provides for bootstrapping, phasing, and high order sensing. The entire instrument is mounted in a rotator to provide gravity invariance. After the main AO system, visible (g to y) and near-IR (Y to H) science channels contain integrated coronagraphic wavefront control systems. The fully corrected and, optionally, coronagraphically filtered beams will then be fed to a suite of focal plane instrumentation including imagers and spectrographs. This will include existing facility instruments at GMT via fiber feeds. To assess the design we have developed an end-to-end frequency-domain modeling framework for assessing the performance of GMagAO-X. The dynamics of the many closed-loop feedback control systems are then modeled. Finally, we employ a frequency-domain model of post-processing algorithms to analyze the final post-processed sensitivity. The CoDR for GMagAO-X was held in September, 2021. Here we present an overview of the science cases, instrument design, expected performance, and concept of operations for GMagAO-X.
astro-ph_IM
REXPACO: an algorithm for high contrast reconstruction of the circumstellar environment by angular differential imaging: Aims. The purpose of this paper is to describe a new post-processing algorithm dedicated to the reconstruction of the spatial distribution of light received from off-axis sources, in particular from circumstellar disks. Methods. Built on the recent PACO algorithm dedicated to the detection of point-like sources, the proposed method is based on the local learning of patch covariances capturing the spatial fluctuations of the stellar leakages. From this statistical modeling, we develop a regularized image reconstruction algorithm (REXPACO) following an inverse problem approach based on a forward image formation model of the off-axis sources in the ADI sequences. Results. Injections of fake circumstellar disks in ADI sequences from the VLT/SPHERE-IRDIS instrument show that both the morphology and the photometry of the disks are better preserved by REXPACO compared to standard postprocessing methods like cADI. In particular, the modeling of the spatial covariances proves usefull in reducing typical ADI artifacts and in better disentangling the signal of these sources from the residual stellar contamination. The application to stars hosting circumstellar disks with various morphologies confirms the ability of REXPACO to produce images of the light distribution with reduced artifacts. Finally, we show how REXPACO can be combined with PACO to disentangle the signal of circumstellar disks from the signal of candidate point-like sources. Conclusions. REXPACO is a novel post-processing algorithm producing numerically deblurred images of the circumstellar environment. It exploits the spatial covariances of the stellar leakages and of the noise to efficiently eliminate this nuisance term.
astro-ph_IM
MegaPipe astrometry for the New Horizons spacecraft: The New Horizons spacecraft, launched by NASA in 2006, will arrive in the Pluto-Charon system on July 14, 2015. There, it will spend a few hours imaging Pluto and its moons. It will then have a small amount of reserve propellant which will be used to direct the probe on to a second, yet to be discovered object in the Kuiper Belt. Data from the MegaPrime camera on CFHT was used to build a precise, high density astrometric reference frame for both the final approach into the Pluto system and the search for the secondary target. Pluto currently lies in the galactic plane. This is a hindrance in that there are potential problems with confusion. However, it is also a benefit, since it allows the use of the UCAC4 astrometric reference catalog, which is normally too sparse for use with MegaCam images. The astrometric accuracy of the final catalogs, as measured by the residuals, is 0.02 arcseconds.
astro-ph_IM
Machine learning techniques to distinguish near-field interference and far-field astrophysical signals in radio telescopes: The CHIME radio telescope operates in the frequency bandwidth of 400 to 800 MHz. The CHIME/FRB collaboration has a data pipeline that analyzes the data in real time, suppresses radio frequency interferences (RFI) and searches for FRBs. However, the RFI removal techniques work best for broadband and narrow FRBs.We wish to create a RFI removal technique that works without making assumptions about the characteristics of the FRB signal. In this thesis we first explore the data of intensity generated by CHIME/FRB backend. After becoming familiar with the structure and organisation of data we present a new novel method for RFI removal using unsupervised machine learning clustering techniques by using multiple beams on CHIME telescope. We are trying to use the analogy of theory of interference for RFI removal by distinguishing near field RFI and far field astrophysical signals in the data. We explored many clustering techniques like K-means,DBSCAN etc but one technique called as HDBSCAN looks particularly promising. Using HDBSCAN clustering technique we have developed the new method for RFI removal. The removal technique upto this point has been developed by us using 3 beams of CHIME telescope. The new novel idea is still in it's incubatory phase and soon we will try to include more beams for our new RFI removal method. We have visually observed that RFI has been been considerably removed from our data. In future we are going to do more calculations to further measure the signal to noise ratio (SNR) of the FRB signal after RFI removal and we will use this technique to compare the SNR measured by current RFI removal technique at CHIME/FRB data pipeline.
astro-ph_IM
An iterative wave-front sensing algorithm for high-contrast imaging systems: Wave-front sensing from focal plane multiple images is a promising technique for high-contrast imaging systems. However, the wave-front error of an optics system can be properly reconstructed only when it is very small. This paper presents an iterative optimization algorithm for the measurement of large static wave-front errors directly from only one focal plane image. We firstly measure the intensity of the pupil image to get the pupil function of the system and acquire the aberrated image on the focal plane with a phase error that is to be measured. Then we induce a dynamic phase to the tested pupil function and calculate the associated intensity of the reconstructed image on the focal plane. The algorithm is to minimize the intensity difference between the reconstructed image and the tested aberrated image on the focal plane, where the induced phase is as the variable of the optimization algorithm. The simulation shows that the wave-front of an optics system can be theoretically reconstructed with a high precision, which indicates that such an iterative algorithm may be an effective way for the wave-front sensing for high-contrast imaging systems.
astro-ph_IM
Analysis Framework for Multi-messenger Astronomy with IceCube: Combining observational data from multiple instruments for multi-messenger astronomy can be challenging due to the complexity of the instrument response functions and likelihood calculation. We introduce a python-based unbinned-likelihood analysis package called i3mla (IceCube Maximum Likelihood Analysis). i3mla is designed to be compatible with the Multi-Mission Maximum Likelihood (3ML) framework, which enables multi-messenger astronomy analyses by combining the likelihood across different instruments. By making it possible to use IceCube data in the 3ML framework, we aim to facilitate the use of neutrino data in multi-messenger astronomy. In this work we illustrate how to use the i3mla package with 3ML and present preliminary sensitivities using the i3mla package and 3ML through a joint-fit with HAWC Public dataset.
astro-ph_IM
A partially dimensionally-split approach to numerical MHD: We modify an existing magnetohydrodynamics algorithm to make it more compatible with a dimensionally-split (DS) framework. It is based on the standard reconstruct-solve-average strategy (using a Riemann solver), and relies on constrained transport to ensure that the magnetic field remains divergence-free (div B = 0). The DS approach, combined with the use of a single, cell-centred grid (for both the fluid quantities and the magnetic field), means that the algorithm can be easily added to existing DS hydrodynamics codes. This makes it particularly useful for mature astrophysical codes, which often model more complicated physical effects on top of an underlying DS hydrodynamics engine, and therefore cannot be restructured easily. Several test problems have been included to demonstrate the accuracy of the algorithm, and illustrative source code has been made freely available online.
astro-ph_IM
Advanced Architectures for Astrophysical Supercomputing: Astronomers have come to rely on the increasing performance of computers to reduce, analyze, simulate and visualize their data. In this environment, faster computation can mean more science outcomes or the opening up of new parameter spaces for investigation. If we are to avoid major issues when implementing codes on advanced architectures, it is important that we have a solid understanding of our algorithms. A recent addition to the high-performance computing scene that highlights this point is the graphics processing unit (GPU). The hardware originally designed for speeding-up graphics rendering in video games is now achieving speed-ups of $O(100\times)$ in general-purpose computation -- performance that cannot be ignored. We are using a generalized approach, based on the analysis of astronomy algorithms, to identify the optimal problem-types and techniques for taking advantage of both current GPU hardware and future developments in computing architectures.
astro-ph_IM
KSIM: simulating KIDSpec, a Microwave Kinetic Inductance Detector spectrograph for the optical/NIR: KIDSpec, the Kinetic Inductance Detector Spectrometer, is a proposed optical to near IR Microwave Kinetic Inductance Detector (MKID) spectrograph. MKIDs are superconducting photon counting detectors which are able to resolve the energy of incoming photons and their time of arrival. KIDSpec will use these detectors to separate incoming spectral orders from a grating, thereby not requiring a cross-disperser. In this paper we present a simulation tool for KIDSpec's potential performance upon construction to optimise a given design. This simulation tool is the KIDSpec Simulator (KSIM), a Python package designed to simulate a variety of KIDSpec and observation parameters. A range of astrophysical objects are simulated: stellar objects, an SDSS observed galaxy, a Seyfert galaxy, and a mock galaxy spectrum from the JAGUAR catalogue. Multiple medium spectral resolution designs for KIDSpec are simulated. The possible impact of MKID energy resolution variance and dead pixels were simulated, with impacts to KIDSpec performance observed using the Reduced Chi-Squared (RCS) value. Using dead pixel percentages from current instruments, the RCS result was found to only increase to 1.21 at worst for one of the designs simulated. SNR comparisons of object simulations between KSIM and X-Shooter's ETC were also simulated. KIDSpec offers a particular improvement over X-Shooter for short and faint observations. For a Seyfert galaxy ($m_{R}=21$) simulation with a 180s exposure, KIDSpec had an average SNR of 4.8, in contrast to 1.5 for X-Shooter. Using KSIM the design of KIDSpec can be optimised to improve the instrument further.
astro-ph_IM
A New Method for Band-limited Imaging with Undersampled Detectors: Since its original use on the Hubble Deep Field, "Drizzle" has become a de facto standard for the combination of images taken by the Hubble Space Tele- scope. However, the drizzle algorithm was developed with small, faint, partially resolved sources in mind, and is not the best possible algorithm for high signal-to-noise unresolved objects. Here, a new method for creating band-limited images from undersampled data is presented. The method uses a drizzled image as a first order approximation and then rapidly converges toward a band-limited image which fits the data given the statistical weighting provided by the drizzled image. The method, named iDrizzle, for iterative Drizzle, effectively eliminates both the small high-frequency artifacts and convolution with an interpolant kernel that can be introduced by drizzling. The method works well in the presence of geometric distortion, and can easily handle cosmic rays, bad pixels, or other missing data. It can combine images taken with random dithers, though the number of dithers required to obtain a good final image depends in part on the quality of the dither placements. iDrizzle may prove most beneficial for producing high-fidelity point spread functions from undersampled images, and could be particularly valuable for future Dark Energy missions such as WFIRST and EUCLID, which will likely attempt to do high precision supernova photometry and lensing experiments with undersampled detectors.
astro-ph_IM
Partially Coherent Optical Modelling of the Ultra-Low-Noise Far-Infrared Imaging Arrays on the SPICA Mission: We have developed a range of theoretical and numerical techniques for modeling the multi-mode, 210-34 micron, ultra-low-noise Transition Edge Sensors that will be used on the SAFARI instrument on the ESA/JAXA cooled-aperture FIR space telescope SPICA. The models include a detailed analysis of the resistive and reactive properties of thin superconducting absorbing films, and a partially coherent mode-matching analysis of patterned films in multi-mode waveguide. The technique allows the natural optical modes, modal responsivities, and Stokes maps of complicated structures comprising patterned films in profiled waveguides and cavities to be determined.
astro-ph_IM
Fully Automated Approaches to Analyze Large-Scale Astronomy Survey Data: Observational astronomy has changed drastically in the last decade: manually driven target-by-target instruments have been replaced by fully automated robotic telescopes. Data acquisition methods have advanced to the point that terabytes of data are flowing in and being stored on a daily basis. At the same time, the vast majority of analysis tools in stellar astrophysics still rely on manual expert interaction. To bridge this gap, we foresee that the next decade will witness a fundamental shift in the approaches to data analysis: case-by-case methods will be replaced by fully automated pipelines that will process the data from their reduction stage, through analysis, to storage. While major effort has been invested in data reduction automation, automated data analysis has mostly been neglected despite the urgent need. Scientific data mining will face serious challenges to identify, understand and eliminate the sources of systematic errors that will arise from this automation. As a special case, we present an artificial intelligence (AI) driven pipeline that is prototyped in the domain of stellar astrophysics (eclipsing binaries in particular), current results and the challenges still ahead.
astro-ph_IM
Automatic morphological classification of galaxy images: We describe an image analysis supervised learning algorithm that can automatically classify galaxy images. The algorithm is first trained using a manually classified images of elliptical, spiral, and edge-on galaxies. A large set of image features is extracted from each image, and the most informative features are selected using Fisher scores. Test images can then be classified using a simple Weighted Nearest Neighbor rule such that the Fisher scores are used as the feature weights. Experimental results show that galaxy images from Galaxy Zoo can be classified automatically to spiral, elliptical and edge-on galaxies with accuracy of ~90% compared to classifications carried out by the author. Full compilable source code of the algorithm is available for free download, and its general-purpose nature makes it suitable for other uses that involve automatic image analysis of celestial objects.
astro-ph_IM
Study of cosmogenic activation above ground of Ar for DarkSide-20k: The production of long-lived radioactive isotopes due to the exposure to cosmic rays on the Earth's surface is an hazard for experiments searching for rare events like the direct detection of galactic dark matter particles. The use of large amounts of liquid Argon is foreseen in different projects, like the DarkSide-20k experiment, intended to look for Weakly Interacting Massive Particles at the Laboratori Nazionali del Gran Sasso. Here, results from the study of the cosmogenic activation of Argon carried out in the context of DarkSide-20k are presented. The induced activity of several isotopes, including 39Ar, and the expected counting rates in the detector have been deduced, considering exposure conditions as realistic as possible.
astro-ph_IM
The analysis of effective galaxies number count for Chinese Space Station Optical Survey(CSS-OS) by image simulation: The Chinese Space Station Optical Survey (CSS-OS) is a mission to explore the vast universe. This mission will equip a 2-meter space telescope to perform a multi-band NUV-optical large area survey (over 40% of the sky) and deep survey (~1% of the sky) for the cosmological and astronomical goals. Galaxy detection is one of the most important methods to achieve scientific goals. In this paper, we evaluate the galaxy number density for CSS-OS in i band (depth, i ~26 for large area survey and ~27 for the deep survey, point source, 5-sigma by the method of image simulation. We also compare galaxies detected by CSS-OS with that of LSST (i~27, point source, 5-sigma. In our simulation, the HUDF galaxy catalogs are used to create mock images due to long enough integration time which meets the completeness requirements of the galaxy analysis for CSS-OS and LSST. The galaxy surface profile and spectrum are produced by the morphological information, photometric redshift and SEDs from the catalogs. The instrumental features and the environmental condition are also considered to produce the mock galaxy images. The galaxies of CSS-OS and LSST are both extracted by SExtractor from the mock i band image and matched with the original catalog. Through the analysis of the extracted galaxies, we find that the effective galaxy number count is ~13 arcmin^-2, ~40 arcmin^-2 and ~42 arcmin^-2 for CSS-OS large area survey, CSS-OS deep survey and LSST, respectively. Moreover, CSS-OS shows the advantage in small galaxy detection with high spatial resolution, especially for the deep survey: about 20% of the galaxies detected by CSS-OS deep survey are not detected by LSST, and they have a small effective radius of re < 0.3".
astro-ph_IM
Matched filter in the low-number count Poisson noise regime: an efficient and effective implementation: The matched filter (MF) is widely used to detect signals hidden within the noise. If the noise is Gaussian, its performances are well-known and describable in an elegant analytical form. The treatment of non-Gaussian noises is often cumbersome as in most cases there is no analytical framework. This is true also for Poisson noise which, especially in the low-number count regime, presents the additional difficulty to be discrete. For this reason in the past methods have been proposed based on heuristic or semi-heuristic arguments. Recently, an analytical form of the MF has been introduced but the computation of the probability of false detection or false alarm (PFA) is based on numerical simulations. To overcome this inefficient and time consuming approach we propose here an effective method to compute the PFA based on the saddle point approximation (SA). We provide the theoretical framework and support our findings by means of numerical simulations. We discuss also the limitations of the MF in practical applications.
astro-ph_IM
Investigating the In-Flight Performance of the UVIT Payload on ASTROSAT: We have studied the performance of the Ultraviolet Imaging Telescope payload on AstroSat and derived a calibration of the FUV and NUV instruments on board. We find that the sensitivity of both the FUV and NUV channels is as expected from ground calibrations, with the FUV effective area about 35% and the NUV effective area about the same as that of GALEX. The point spread function of the instrument is on the order of 1.2-1.6 arcsec. We have found that pixel-to-pixel variations in the sensitivity are less than 10% with spacecraft motion compensating for most of the flat-field variations. We derived a distortion correction but recommend that it be applied post-processing as part of an astrometric solution.
astro-ph_IM
The Importance of Telescope Training in Data Interpretation: In this State of the Profession Consideration, we will discuss the state of hands-on observing within the profession, including: information about professional observing trends; student telescope training, beginning at the undergraduate and graduate levels, as a key to ensuring a base level of technical understanding among astronomers; the role that amateurs can take moving forward; the impact of telescope training on using survey data effectively; and the need for modest investments in new, standard instrumentation at mid-size aperture telescope facilities to ensure their usefulness for the next decade.
astro-ph_IM
Fully-Automated Reduction of Longslit Spectroscopy with the Low Resolution Imaging Spectrometer at Keck Observatory: I present and summarize a software package ("LPipe") for completely automated, end-to-end reduction of both bright and faint sources with the Low-Resolution Imaging Spectrometer (LRIS) at Keck Observatory. It supports all gratings, grisms, and dichroics, and also reduces imaging observations, although it does not include multislit or polarimetric reduction capabilities at present. It is suitable for on-the-fly quicklook reductions at the telescope, for large-scale reductions of archival data-sets, and (in many cases) for science-quality post-run reductions of PI data. To demonstrate its capabilities the pipeline is run in fully-automated mode on all LRIS longslit data in the Keck Observatory Archive acquired during the 12-month period between August 2016 and July 2017. The reduced spectra (of 675 single-object targets, totaling ~200 hours of on-source integration time in each camera), and the pipeline itself, are made publicly available to the community.
astro-ph_IM
Pulsar scattering in space and time: We report on a recent global VLBI experiment in which we study the scatter broadening of pulsars in the spatial and time domain simultaneously. Depending on the distribution of scattering screen(s), geometry predicts that the less spatially broadened parts of the signal arrive earlier than the more broadened parts. This means that over one pulse period the size of the scattering disk should grow from pointlike to the maximum size. An equivalent description is that the pulse profile shows less temporal broadening on the longer baselines. This contribution presents first results that are consistent with the expected expanding rings. We also briefly discuss how the autocorrelations can be used for amplitude calibration. This requires a thorough investigation of the digitisation and the sampler statistics and is not fully solved yet.
astro-ph_IM
Antenna characterization for the HIRAX experiment: The Hydrogen Intensity and Real-time Analysis eXperiment (HIRAX) aims to improve constraints on the dark energy equation of state through measurements of large-scale structure at high redshift ($0.8<z<2.5$), while serving as a state-of-the-art fast radio burst detector. Bright galactic foregrounds contaminate the 400--800~MHz HIRAX frequency band, so meeting the science goals will require precise instrument characterization. In this paper we describe characterization of the HIRAX antenna, focusing on measurements of the antenna beam and antenna noise temperature. Beam measurements of the current HIRAX antenna design were performed in an anechoic chamber and compared to simulations. We report measurement techniques and results, which find a broad and symmetric antenna beam for $\nu <$650MHz, and elevated cross-polarization levels and beam asymmetries for $\nu >$700MHz. Noise temperature measurements of the HIRAX feeds were performed in a custom apparatus built at Yale. In this system, identical loads, one cryogenic and the other at room temperature, are used to take a differential (Y-factor) measurement from which the noise of the system is inferred. Several measurement sets have been conducted using the system, involving CHIME feeds as well as four of the HIRAX active feeds. These measurements give the first noise temperature measurements of the HIRAX feed, revealing a $\sim$60K noise temperature (relative to 30K target) with 40K peak- to-peak frequency-dependent features, and provide the first demonstration of feed repeatability. Both findings inform current and future feed designs.
astro-ph_IM
Exploring the Capabilities of Gibbs Sampling in Pulsar Timing Arrays: We explore the use of Gibbs sampling in estimating the noise properties of individual pulsars and illustrate its effectiveness using the NANOGrav 11-year data set. We find that Gibbs sampling noise modeling (GM) is more efficient than the current standard Bayesian techniques (SM) for single pulsar analyses by yielding model parameter posteriors with average effective-sample-size ratio (GM/SM) of 6 across all parameters and pulsars. Furthermore, the output of GM contains posteriors for the Fourier coefficients that can be used to characterize the underlying red noise process of any pulsar's timing residuals, which are absent in current implementations of SM. Through simulations, we demonstrate the potential for such coefficients to measure the spatial cross-correlations between pulsar pairs produced by a gravitational wave background.
astro-ph_IM
EMC design for the actuators of FAST reflector: The active reflector is one of the three main innovations of the Five-hundred-meter Aperture Spherical radio Telescope (FAST). The deformation of such a huge spherically shaped reflector into different transient parabolic shapes is achieved by using 2225 hydraulic actuators which change the position of the 2225 nodes through the connected down tied cables. For each different tracking process of the telescope, more than 1/3 of these 2225 actuators must be in operation to tune the parabolic aperture accurately to meet the surface error restriction. It means that some of these actuators are inevitably located within the main beam of the receiver, and the Electromagnetic Interference (EMI) from the actuators must be mitigated to ensure the scientific output of the telescope. Based on the threshold level of interference detrimental to radio astronomy presented in ITU-R Recommendation RA.769 and EMI measurements, the shielding efficiency (SE) requirement of each actuator is set to be 80dB in the frequency range from 70MHz to 3GHz. Therefore, Electromagnetic Compatibility (EMC) was taken into account in the actuator design by measures such as power line filters, optical fibers, shielding enclosures and other structural measures. In 2015, all the actuators had been installed at the FAST site. Till now, no apparent EMI from the actuators has been detected by the receiver, which proves the effectiveness of these EMC measures.
astro-ph_IM
Rethinking the modeling of the instrumental response of telescopes with a differentiable optical model: We propose a paradigm shift in the data-driven modeling of the instrumental response field of telescopes. By adding a differentiable optical forward model into the modeling framework, we change the data-driven modeling space from the pixels to the wavefront. This allows to transfer a great deal of complexity from the instrumental response into the forward model while being able to adapt to the observations, remaining data-driven. Our framework allows a way forward to building powerful models that are physically motivated, interpretable, and that do not require special calibration data. We show that for a simplified setting of a space telescope, this framework represents a real performance breakthrough compared to existing data-driven approaches with reconstruction errors decreasing 5 fold at observation resolution and more than 10 fold for a 3x super-resolution. We successfully model chromatic variations of the instrument's response only using noisy broad-band in-focus observations.
astro-ph_IM
Radiative transfer and molecular data for astrochemistry (Review): The estimation of molecular abundances in interstellar clouds from spectroscopic observations requires radiative transfer calculations, which depend on basic molecular input data. This paper reviews recent developments in the fields of molecular data and radiative transfer. The first part is an overview of radiative transfer techniques, along with a "road map" showing which technique should be used in which situation. The second part is a review of measurements and calculations of molecular spectroscopic and collisional data, with a summary of recent collisional calculations and suggested modeling strategies if collision data are unavailable. The paper concludes with an overview of future developments and needs in the areas of radiative transfer and molecular data.
astro-ph_IM
Performance results of HESP physical model: As a continuation to the published work on model based calibration technique with HESP(Hanle Echelle Spectrograph) as a case study, in this paper we present the performance results of the technique. We also describe how the open parameters were chosen in the model for optimization, the glass data accuracy and handling the discrepancies. It is observed through simulations that the discrepancies in glass data can be identified but not quantifiable. So having an accurate glass data is important which is possible to obtain from the glass manufacturers. The model's performance in various aspects is presented using the ThAr calibration frames from HESP during its pre-shipment tests. Accuracy of model predictions and its wave length calibration comparison with conventional empirical fitting, the behaviour of open parameters in optimization, model's ability to track instrumental drifts in the spectrum and the double fibres performance were discussed. It is observed that the optimized model is able to predict to a high accuracy the drifts in the spectrum from environmental fluctuations. It is also observed that the pattern in the spectral drifts across the 2D spectrum which vary from image to image is predictable with the optimized model. We will also discuss the possible science cases where the model can contribute.
astro-ph_IM
Estimating effective wind speed from Gemini Planet Imager's adaptive optics data using covariance maps: The Earth's turbulent atmosphere results in speckled and blurred images of astronomical objects when observed by ground based visible and near-infrared telescopes. Adaptive optics (AO) systems are employed to reduce these atmospheric effects by using wavefront sensors (WFS) and deformable mirrors. Some AO systems are not fast enough to correct for strong, fast, high turbulence wind layers leading to the wind butterfly effect, or wind-driven halo, reducing contrast capabilities in coronagraphic images. Estimating the effective wind speed of the atmosphere allows us to calculate the atmospheric coherence time. This is not only an important parameter to understand for site characterization but could be used to help remove the wind butterfly in post processing. Here we present a method for estimating the atmospheric effective wind speed from spatio-temporal covariance maps generated from pseudo open-loop (POL) WFS data. POL WFS data is used as it aims to reconstruct the full wavefront information when operating in closed-loop. The covariance maps show how different atmospheric turbulent layers traverse the telescope. Our method successfully recovered the effective wind speed from simulated WFS data generated with the soapy python library. The simulated atmospheric turbulence profiles consist of two turbulent layers of ranging strengths and velocities. The method has also been applied to Gemini Planet Imager (GPI) AO WFS data. This gives insight into how the effective wind speed can affect the wind-driven halo seen in the AO image point spread function. In this paper, we will present results from simulated and GPI WFS data.
astro-ph_IM
The Scaling of the RMS with Dwell Time in NANOGrav Pulsars: Pulsar Timing Arrays (PTAs) are collections of well-timed millisecond pulsars that are being used as detectors of gravitational waves (GWs). Given current sensitivity, projected improvements in PTAs and the predicted strength of the GW signals, the detection of GWs with PTAs could occur within the next decade. One way we can improve a PTA is to reduce the measurement noise present in the pulsar timing residuals. If the pulsars included in the array display uncorrelated noise, the root mean square (RMS) of the timing residuals is predicted to scale as $\mathrm{T}^{-1/2}$, where T is the dwell time per observation. In this case, the sensitivity of the array can be increased by increasing T. We studied the 17 pulsars in the five year North American Nanohertz Observatory for Gravitational Waves (NANOGrav) data set to determine if the noise in the timing residuals of the pulsars observed was consistent with this property. For comparison, we performed the same analysis on PSR B1937+21, a pulsar that is known to display red noise. With this method, we find that 15 of the 17 NANOGrav pulsars have timing residuals consistent with the inverse square law. The data also suggest that these 15 pulsars can be observed for up to eight times as long while still exhibiting an RMS that scales as root T.
astro-ph_IM
Citizen Science Astronomy with a Network of Small Telescope: The Launch and Deployment of JWST: We present a coordinated campaign of observations to monitor the brightness of the James Webb Space Telescope (JWST) as it travels toward the second Earth-Sun Lagrange point and unfolds using the network ofUnistellar digital telescopes. Those observations collected by citizen astronomers across the world allowed us to detect specific phases such as the separation from the booster, glare due to a change of orientation after a maneuver, the unfurling of the sunshield, and deployment of the primary mirror. After deployment of the sunshield on January 6 2022, the 6-h lightcurve has a significant amplitude and shows small variations due to the artificial rotation of the space telescope during commissionning. These variations could be due to the deployment of the primary mirror or some changes in orientation of the space telescope. This work illustrates the power of a worldwide array of small telescopes, operated by citizen astronomers, to conduct large scientific campaigns over a long timeframe. In the future, our network and others will continue to monitor JWST to detect potential degradations to the space environment by comparing the evolution of the lightcurve.
astro-ph_IM
A convergent blind deconvolution method for post-adaptive-optics astronomical imaging: In this paper we propose a blind deconvolution method which applies to data perturbed by Poisson noise. The objective function is a generalized Kullback-Leibler divergence, depending on both the unknown object and unknown point spread function (PSF), without the addition of regularization terms; constrained minimization, with suitable convex constraints on both unknowns, is considered. The problem is nonconvex and we propose to solve it by means of an inexact alternating minimization method, whose global convergence to stationary points of the objective function has been recently proved in a general setting. The method is iterative and each iteration, also called outer iteration, consists of alternating an update of the object and the PSF by means of fixed numbers of iterations, also called inner iterations, of the scaled gradient projection (SGP) method. The use of SGP has two advantages: first, it allows to prove global convergence of the blind method; secondly, it allows the introduction of different constraints on the object and the PSF. The specific constraint on the PSF, besides non-negativity and normalization, is an upper bound derived from the so-called Strehl ratio, which is the ratio between the peak value of an aberrated versus a perfect wavefront. Therefore a typical application is the imaging of modern telescopes equipped with adaptive optics systems for partial correction of the aberrations due to atmospheric turbulence. In the paper we describe the algorithm and we recall the results leading to its convergence. Moreover we illustrate its effectiveness by means of numerical experiments whose results indicate that the method, pushed to convergence, is very promising in the reconstruction of non-dense stellar clusters. The case of more complex astronomical targets is also considered, but in this case regularization by early stopping of the outer iterations is required.
astro-ph_IM
The Near-Infrared Spectrograph (NIRSpec) on the James Webb Space Telescope IV. Capabilities and predicted performance for exoplanet characterization: The Near-Inrared Spectrograph (NIRSpec) on the James Webb Space Telescope (JWST) is a very versatile instrument, offering multiobject and integral field spectroscopy with varying spectral resolution ($\sim$30 to $\sim$3000) over a wide wavelength range from 0.6 to 5.3 micron, enabling scientists to study many science themes ranging from the first galaxies to bodies in our own Solar System. In addition to its integral field unit and support for multiobject spectroscopy, NIRSpec features several fixed slits and a wide aperture specifically designed to enable high precision time-series and transit as well as eclipse observations of exoplanets. In this paper we present its capabilities regarding time-series observations, in general, and transit and eclipse spectroscopy of exoplanets in particular. Due to JWST's large collecting area and NIRSpec's excellent throughput, spectral coverage, and detector performance, this mode will allow scientists to characterize the atmosphere of exoplanets with unprecedented sensitivity.
astro-ph_IM
DQSEGDB: A time-interval database for storing gravitational wave observatory metadata: The Data Quality Segment Database (DQSEGDB) software is a database service, backend API, frontend graphical web interface, and client package used by the Laser Interferometer Gravitational-Wave Observatory (LIGO), Virgo, GEO600 and the Kamioka Gravitational wave detector for storing and accessing metadata describing the status of their detectors. The DQSEGDB has been used in the analysis of all published detections of gravitational waves in the advanced detector era. The DQSEGDB currently stores roughly 600 million metadata entries and responds to roughly 600,000 queries per day with an average response time of 0.223 ms.
astro-ph_IM
A Case Study in Astronomical 3-D Printing: The Mysterious Eta Carinae: 3-D printing moves beyond interactive 3-D graphics and provides an excellent tool for both visual and tactile learners, since 3-D printing can now easily communicate complex geometries and full color information. Some limitations of interactive 3-D graphics are also alleviated by 3-D printable models, including issues of limited software support, portability, accessibility, and sustainability. We describe the motivations, methods, and results of our work on using 3-D printing (1) to visualize and understand the Eta Car Homunculus nebula and central binary system and (2) for astronomy outreach and education, specifically, with visually impaired students. One new result we present is the ability to 3-D print full-color models of Eta Car's colliding stellar winds. We also demonstrate how 3-D printing has helped us communicate our improved understanding of the detailed structure of Eta Car's Homunculus nebula and central binary colliding stellar winds, and their links to each other. Attached to this article are full-color 3-D printable files of both a red-blue Homunculus model and the Eta Car colliding stellar winds at orbital phase 1.045. 3-D printing could prove to be vital to how astronomer's reach out and share their work with each other, the public, and new audiences.
astro-ph_IM
A Fourier optics approach to evaluate the astrometric performance of MICADO: We present our investigation into the impact of wavefront errors on high accuracy astrometry using Fourier Optics. MICADO, the upcoming near-IR imaging instrument for the Extremely Large Telescope, will offer capabilities for relative astrometry with an accuracy of 50 micro arcseconds ({\mu}as). Due to the large size of the point spread function (PSF) compared to the astrometric requirement, the detailed shape and position of the PSF on the detector must be well understood. Furthermore, because the atmospheric dispersion corrector of MICADO is a moving component within an otherwise mostly static instrument, it might not be sufficient to perform a simple pre-observation calibration. Therefore, we have built a Fourier Optics framework, allowing us to evaluate the small changes in the centroid position of the PSF as a function of wavefront error. For a complete evaluation, we model both the low order surface form errors, using Zernike polynomials, and the mid- and high-spatial frequencies, using Power Spectral Density analysis. The described work will then make it possible, performing full diffractive beam propagation, to assess the expected astrometric performance of MICADO.
astro-ph_IM
Status of the Medium-Sized Telescope for the Cherenkov Telescope Array: The Cherenkov Telescope Array (CTA), is an international project for the next generation ground- based observatory for gamma-ray astronomy in the energy range from 20 GeV to 300 TeV. The sensitivity in the core energy range will be dominated by up to 40 Medium-Sized Telescopes (MSTs). The MSTs, of Davies-Cotton type with a 12 m diameter reflector are currently in the prototype phase. A full-size mechanical telescope structure has been assembled in Berlin. The telescope is partially equipped with different mirror prototypes, which are currently being tested and evaluated for performances characteristics. A report concentrating on the details of the tele- scope structure, the drive assemblies and the optics of the MST prototype will be given.
astro-ph_IM
The antinucleus annihilation reconstruction algorithm of the GAPS experiment: The General AntiParticle Spectrometer (GAPS) is an Antarctic balloon-borne detector designed to measure low-energy cosmic antinuclei (< 0.25 GeV/n), with a specific focus on antideuterons, as a distinctive signal from dark matter annihilation or decay in the Galactic halo. The instrument consists of a tracker, made up of ten planes of lithium-drifted Silicon Si(Li) detectors, surrounded by a plastic scintillator Time-of-Flight system. GAPS uses a novel particle identification method based on exotic atom capture and decay with the emission of pions, protons, and atomic X-rays from a common annihilation vertex. An important ingredient for the antinuclei identification is the reconstruction of the "annihilation star" topology. A custom antinucleus annihilation reconstruction algorithm, called the "star-finding" algorithm, was developed to reconstruct the annihilation star fully, determining the annihilation vertex position and reconstructing the tracks of the primary and secondary charged particles. The reconstruction algorithm and its performances were studied on simulated data obtained with the Geant4-based GAPS simulation software, which fully reproduced the detector geometry. This custom algorithm was found to have better performance in the vertex resolution and reconstruction efficiency compared with a standard Hough-3D algorithm.
astro-ph_IM
Adaptive Kernel Density Estimation proposal in gravitational wave data analysis: Markov Chain Monte Carlo approach is frequently used within Bayesian framework to sample the target posterior distribution. Its efficiency strongly depends on the proposal used to build the chain. The best jump proposal is the one that closely resembles the unknown target distribution, therefore we suggest an adaptive proposal based on Kernel Density Estimation (KDE). We group parameters of the model according to their correlation and build KDE based on the already accepted points for each group. We adapt the KDE-based proposal until it stabilizes. We argue that such a proposal could be helpful in applications where the data volume is increasing and in the hyper-model sampling. We tested it on several astrophysical datasets (IPTA and LISA) and have shown that in some cases KDE-based proposal also helps to reduce the autocorrelation length of the chains. The efficiency of this proposal is reduces in case of the strong correlations between a large group of parameters.
astro-ph_IM
The LAUE project for broadband gamma-ray focusing lenses: We present the LAUE project devoted to develop an advanced technology for building a high focal length Laue lens for soft gamma--ray astronomy (80-600 keV). The final goal is to develop a focusing optics that can improve the current sensitivity in the above energy band by 2 orders of magnitude.
astro-ph_IM
All-sky Radio SETI: Over the last decade, Aperture Arrays (AA) have successfully replaced parabolic dishes as the technology of choice at low radio frequencies - good examples are the MWA, LWA and LOFAR. Aperture Array based telescopes present several advantages, including sensitivity to the sky over a very wide field-of-view. As digital and data processing systems continue to advance, an all-sky capability is set to emerge, even at GHz frequencies. We argue that assuming SETI events are both rare and transitory in nature, an instrument with a large field-of-view, operating around the so-called water-hole (1-2 GHz), might offer several advantages over contemporary searches. Sir Arthur C. Clarke was the first to recognise the potential importance of an all-sky radio SETI capability, as presented in his book, Imperial Earth. As part of the global SKA (Square Kilometre Array) project, a Mid-Frequency Aperture Array (MFAA) prototype known as MANTIS (Mid- Frequency Aperture Array Transient and Intensity-Mapping System) is now being considered as a precursor for SKA-2. MANTIS can be seen as a first step towards an all-sky radio SETI capability at GHz frequencies. This development has the potential to transform the field of SETI research, in addition to several other scientific programmes.
astro-ph_IM
Scattering efficiencies measurements of soft protons at grazing incidence from an Athena Silicon Pore Optics sample: Soft protons are a potential threat for X-ray missions using grazing incidence optics, as once focused onto the detectors they can contribute to increase the background and possibly induce radiation damage as well. The assessment of these undesired effects is especially relevant for the future ESA X-ray mission Athena, due to its large collecting area. To prevent degradation of the instrumental performance, which ultimately could compromise some of the scientific goals of the mission, the adoption of ad-hoc magnetic diverters is envisaged. Dedicated laboratory measurements are fundamental to understand the mechanisms of proton forward scattering, validate the application of the existing physical models to the Athena case and support the design of the diverters. In this paper we report on scattering efficiency measurements of soft protons impinging at grazing incidence onto a Silicon Pore Optics sample, conducted in the framework of the EXACRAD project. Measurements were taken at two different energies, ~470 keV and ~170 keV, and at four different scattering angles between 0.6 deg and 1.2 deg. The results are generally consistent with previous measurements conducted on eROSITA mirror samples, and as expected the peak of the scattering efficiency is found around the angle of specular reflection.
astro-ph_IM
Mining for Strong Gravitational Lenses with Self-supervised Learning: We employ self-supervised representation learning to distill information from 76 million galaxy images from the Dark Energy Spectroscopic Instrument Legacy Imaging Surveys' Data Release 9. Targeting the identification of new strong gravitational lens candidates, we first create a rapid similarity search tool to discover new strong lenses given only a single labelled example. We then show how training a simple linear classifier on the self-supervised representations, requiring only a few minutes on a CPU, can automatically classify strong lenses with great efficiency. We present 1192 new strong lens candidates that we identified through a brief visual identification campaign, and release an interactive web-based similarity search tool and the top network predictions to facilitate crowd-sourcing rapid discovery of additional strong gravitational lenses and other rare objects: https://github.com/georgestein/ssl-legacysurvey.
astro-ph_IM
The Radar Echo Telescope for Cosmic Rays: Pathfinder Experiment for a Next-Generation Neutrino Observatory: The Radar Echo Telescope for Cosmic Rays (RET-CR) is a recently initiated experiment designed to detect the englacial cascade of a cosmic-ray initiated air shower via in-ice radar, toward the goal of a full-scale, next-generation experiment to detect ultra high energy neutrinos in polar ice. For cosmic rays with a primary energy greater than 10 PeV, roughly 10% of an air-shower's energy reaches the surface of a high elevation ice-sheet ($\gtrsim$2 km) concentrated into a radius of roughly 10 cm. This penetrating shower core creates an in-ice cascade many orders of magnitude more dense than the preceding in-air cascade. This dense cascade can be detected via the radar echo technique, where transmitted radio is reflected from the ionization deposit left in the wake of the cascade. RET-CR will test the radar echo method in nature, with the in-ice cascade of a cosmic-ray initiated air-shower serving as a test beam. We present the projected event rate and sensitivity based upon a three part simulation using CORSIKA, GEANT4, and RadioScatter. RET-CR expects $\sim$1 radar echo event per day.
astro-ph_IM
AstroCloud: A Distributed Cloud Computing and Application Platform for Astronomy: Virtual Observatory (VO) is a data-intensively online astronomical research and education environment, which takes advantages of advanced information technologies to achieve seamless and global access to astronomical information. AstroCloud is a cyber-infrastructure for astronomy research initiated by Chinese Virtual Observatory (China-VO) project, and also a kind of physical distributed platform which integrates lots of tasks such as telescope access proposal management, data archiving, data quality control, data release and open access, cloud based data processing and analysis. It consists of five application channels, i.e. observation, data, tools, cloud and public and is acting as a full lifecycle management system and gateway for astronomical data and telescopes. Physically, the platform is hosted in six cities currently, i.e. Beijing, Nanjing, Shanghai, Kunming, Lijiang and Urumqi, and serving more than 17 thousand users. Achievements from international Virtual Observatories and Cloud Computing are adopted heavily. In the paper, backgrounds of the project, architecture, Cloud Computing environment, key features of the system, current status and future plans are introduced.
astro-ph_IM
Portable Adaptive Optics for Exoplanet Imaging: The Portable Adaptive Optics (PAO) is a low-cost and compact system, designed for 4-meter class telescopes that have no Adaptive Optics (AO), because of the physical space limitation at the Nasmyth or Cassegrain focus and the historically high cost of the conventional AO. The initial scientific observations of the PAO are focused on the direct imaging of exoplanets and sub-stellar companions. This paper discusses the PAO concept and the associated high-contrast imaging performance in our recent observational runs. PAO is delivering a Strehl ratio better than 0.6 in H band under median seeing conditions of 1 arcsec. Combined with our dedicated image rotation and subtraction (IRS) technique and the optimized IRS (O-IRS) algorithm, the averaged contrast ratio for a Vmag (5-9) primary star is 1.3E10-5 and 3.3E10-6 at angular distance of 0.36 arcsec under exposure time of 7 minutes and 2 hours, respectively. PAO has successfully recovered the known exoplanet of \k{appa} And b, in our recent observation at 3.5-meter ARC telescope at Apache Point Observatory. We have performed the associated astrometry and photometry analysis of the recovered kappa And b planet, which gives a projected separation of 0.984 +/- 0.05 arcsec, a position angle of 51.1 +/- 0.5 degrees, and a mass of 10.15 (-1.255) (+2.19) MJup. These results demonstrate that PAO can be used for direct imaging of exoplanets with medium-sized telescopes.
astro-ph_IM
Finding the UV-Visible Path Forward: Proceedings of the Community Workshop to Plan the Future of UV/Visible Space Astrophysics: We present the science cases and technological discussions that came from the workshop entitled "Finding the UV-Visible Path Forward" held at NASA GSFC June 25-26, 2015. The material presented outlines the compelling science that can be enabled by a next generation space-based observatory dedicated for UV-visible science, the technologies that are available to include in that observatory design, and the range of possible alternative launch approaches that could also enable some of the science. The recommendations to the Cosmic Origins Program Analysis Group from the workshop attendees on possible future development directions are outlined.
astro-ph_IM
The possibility of determining open-cluster parameters from BVRI photometry: In the last decades we witnessed an increase in studies of open clusters of the Galaxy, especially because of the good determination for a wide range of values of parameters such as age, distance, reddening, and proper motion. The reliable determination of the parameters strongly depends on the photometry available and especially on the U filter, which is used to obtain the color excess E(B-V) through the color-color diagram (U-B) by (B-V) by fitting a zero age main-sequence. Owing to the difficulty of performing photometry in the U band, many authors have tried to obtain E(B-V) without the filter. But because of the near linearity of the color-color diagrams that use the other bands, combined with the fact that most fitting procedures are highly subjective (many done "by eye") the reliability of those results has always been questioned. Our group has recently developed, a tool that performs isochrone fitting in open-cluster photometric data with a global optimization algorithm, which removes the need to visually perform the fits and thus removes most of the related subjectivity. Here we apply our method to a set of synthetic clusters and two observed open clusters (Trumpler 1 and Melotte 105) using only photometry for the BVRI bands. Our results show that, considering the cluster structural variance caused only by photometric and Poisson sampling errors, our method is able to recover the synthetic cluster parameters with errors of less than 10% for a wide range of ages, distances, and reddening, which clearly demonstrates its potential. The results obtained for Trumpler 1 and Melotte 105 also agree well with previous literature values.
astro-ph_IM
A Simple Proposal for Radial 3D Needlets: We present here a simple construction of a wavelet system for the three-dimensional ball, which we label \emph{Radial 3D Needlets}. The construction envisages a data collection environment where an observer located at the centre of the ball is surrounded by concentric spheres with the same pixelization at different radial distances, for any given resolution. The system is then obtained by weighting the projector operator built on the corresponding set of eigenfunctions, and performing a discretization step which turns out to be computationally very convenient. The resulting wavelets can be shown to have very good localization properties in the real and harmonic domain; their implementation is computationally very convenient, and they allow for exact reconstruction as they form a tight frame systems. Our theoretical results are supported by an extensive numerical analysis.
astro-ph_IM
The upcoming spectroscopic powerhouses at the Isaac Newton Group of Telescopes: The Isaac Newton Group of Telescopes is completing a strategic change for the scientific use of its two telescopes, the 4.2-m William Herschel Telescope (WHT) and the 2.5-m Isaac Newton Telescope (INT). After more than 30 years operating as multi-purpose telescopes, the telescopes will soon complete their shift to nearly-single instrument operation dominated by large surveys. At the WHT, the WEAVE multi-fibre spectrograph is being commissioned in late 2022. Science surveys are expected to launch in 2023. 30% of the available time will be offered in open time. For the INT, construction of HARPS-3, a high-resolution ultra-stable spectrograph for extra-solar planet studies, is underway, with deployment planned for late 2024. The INT itself is being modernised and will operate as a robotic telescope. An average of 40% of the time will be offered as open time. The ING will maintain its student programme. Plans call for moving student work from the INT to the WHT once the INT starts operating robotically.
astro-ph_IM
Audible universe: A multi-disciplinary team recently came together online to discuss the application of sonification in astronomy, focussing on the effective use of sound for scientific discovery and for improving accessibility to astronomy research and education. Here we provide a meeting report.
astro-ph_IM
Hierarchical approach to matched filtering using a reduced basis: Searching for gravitational waves from compact binary coalescence (CBC) is performed by matched filtering the observed strain data from gravitational-wave observatories against a discrete set of waveform templates designed to accurately approximate the expected gravitational-wave signal, and are chosen to efficiently cover a target search region. The computational cost of matched filtering scales with both the number of templates required to cover a parameter space and the in-band duration of the waveform. Both of these factors increase in difficulty as the current observatories improve in sensitivity, especially at low frequencies, and may pose challenges for third-generation observatories. Reducing the cost of matched filtering would make searches of future detector data more tractable. In addition, it would be easier to conduct searches that incorporate the effects of eccentricity, precession or target light sources (e.g. subsolar). We present a hierarchical scheme based on a reduced basis method to decrease the computational cost of conducting a matched-filter based search. Compared to the current methods, we estimate without any loss in sensitivity, a speedup by a factor of ~ 10 for sources with signal-to-noise ratio (SNR) of at least =6.0, and a factor of ~ 6 for SNR of at least 5. Our method is dominated by linear operations which are highly parallelizable. Therefore, we implement our algorithm using graphical processing units (GPUs) and evaluate commercially motivated metrics to demonstrate the efficiency of GPUs in CBC searches. Our scheme can be extended to generic CBC searches and allows for efficient matched filtering using GPUs.
astro-ph_IM
The Influence of Satellite Trails on H.E.S.S. Gamma-Ray Astronomical Observations: The number of satellites launched into low earth orbit has almost tripled (to over 4000) in the last three years due to the increasing commercialisation of space. Satellite constellations with a total of over 400,000 satellites are proposed to be launched in the near future. Many of these satellites are highly reflective, resulting in a high optical brightness that affects ground-based astronomical observations across the electromagnetic spectrum. Despite this, the potential effect of these satellites on Imaging Atmospheric Cherenkov Telescopes (IACTs) has so far been assumed to be negligible due to their nanosecond integration times. This has, however, never been verified. We aim to identify satellite trails in data taken by the High Energy Stereoscopic System (H.E.S.S.) IACT array in Namibia, using Night Sky Background (NSB) data from the CT5 camera installed in 2019. We determine which observation times and pointing directions are affected the most, and evaluate the impact on Hillas parameters used for classification and reconstruction of high-energy Extensive Air Shower events. Finally, we predict how future planned satellite launches will affect gamma-ray observations with IACTs.
astro-ph_IM
A new photopolymer based VPHG for astronomy: The case of SN 2013fj: The spectroscopic studies of near infrared emission arising from supernovae allow to derive crucial quantities that could better characterise physical conditions of the expanding gas, such as the CaII IR HVF spectral feature. For this reason is mandatory to have Diffractive Optical Elements (DOEs) with a spectral coverage in the range 8000 - 10000 Angstroms (for low z sources) combined with a reasonable Signal to Noise Ratio (S/N) and medium-low resolution. In order to cope with all of those requirements we developed a Volume Phase Holographic Grating (VPHG) based on an innovative photosensitive material, developed by Bayer MaterialScience. We demonstrated the capabilities of this new DOE through observation of SN 2013fj as case study at Asiago Copernico Telescope where AFOSC spectrograph is available.
astro-ph_IM
The fiber-fed preslit of GIANO at T.N.G: Giano is a Cryogenic Spectrograph located in T.N.G. (Spain) and commisioned in 2013. It works in the range 950-2500 nm with a resolving power of 50000. This instrument was designed and built for direct feeding from the telescope [2]. However, due to constraints imposed on the telescope interfacing during the pre-commissioning phase, it had to be positioned on the rotating building, far from the telescope focus. Therefore, a new interface to the telescope, based on IR-transmitting ZBLAN fibers with 85\mu m core, was developed.Originally designed to work directly at the $f/11$ nasmyth focus of the telescope, in 2011 it has decided to use a fiber to feed it. The beam from the telescope is focused on a double fiber boundle by a Preslit Optical Bench attached to the Nasmith A interface of the telescope. This Optical Bench contains the fiber feeding system and other important features as a guiding system, a fiber viewer, a fiber feed calibration lamp and a nodding facility between the two fibers. The use of two fibers allow us to have in the echellogram two spectrograms side by side in the same acquisition: one of the star and the other of the sky or simultaneously to have the star and a calibration lamp. Before entering the cryostat the light from the fiber is collectd by a second Preslit Optical Bench attached directly to the Giano cryostat: on this bench the correct f-number to illuminate the cold stop is generated and on the same bench is placed an image slicer to increase the efficiency of the system.
astro-ph_IM
Standard FITS template for simulated astrophysical scenes with the WFIRST coronagraph: The science investigation teams (SITs) for the WFIRST coronagraphic instrument have begun studying the capabilities of the instrument to directly image reflected light off from exoplanets at contrasts down to contrasts of ~10^-9 with respect to the stellar flux. Detection of point sources at these high contrasts requires yield estimates and detailed modeling of the image of the planetary system as it propagates through the telescope optics. While the SITs might generate custom astrophysical scenes, the integrated model, propagated through the internal speckle field, is typically done at JPL. In this white paper, we present a standard file format to ensure a single distribution system between those who produce the raw astrophysical scenes, and JPL modelers who incorporate those scenes into their optical modeling. At its core, our custom file format uses FITS files, and incorporates standards on packaging astrophysical scenes. This includes spectral and astrometric information for planetary and stellar point sources, zodiacal light and extragalactic sources that may appear as contaminants. Adhering to such a uniform data distribution format is necessary, as it ensures seamless work flow between the SITs and modelers at JPL for the goals of understanding limits of the WFIRST coronagraphic instrument.
astro-ph_IM
The performance of the MAGIC telescopes using deep convolutional neural networks with CTLearn: The Major Atmospheric Gamma Imaging Cherenkov (MAGIC) telescope system is located on the Canary Island of La Palma and inspects the very high-energy (VHE, few tens of GeV and above) gamma-ray sky. MAGIC consists of two imaging atmospheric Cherenkov telescopes (IACTs), which capture images of the air showers originating from the absorption of gamma rays and cosmic rays by the atmosphere, through the detection of Cherenkov photons emitted in the shower. The sensitivity of IACTs to gamma-ray sources is mainly determined by the ability to reconstruct the properties (type, energy, and arrival direction) of the primary particle generating the air shower. The state-of-the-art IACT pipeline for shower reconstruction is based on the parameterization of the shower images by extracting geometric and stereoscopic features and machine learning algorithms like random forest or boosted decision trees. In this contribution, we explore deep convolutional neural networks applied directly to the pixelized images of the camera as a promising method for IACT full-event reconstruction and present the performance of the method on observational data using CTLearn, a package for IACT event reconstruction that exploits deep learning.
astro-ph_IM
Polarization loss in reflecting coating: In laser gravitational waves detectors optical loss restricts sensitivity. We discuss polarization scattering as one more possible mechanism of optical losses. Circulated inside interferometer light is polarized and after reflection its plane of polarization can turn a little due to reflecting coating of mirror can have slightly different refraction index along axes $x,\, y$ in plane of mirror surface (optical anisotropy). This anisotropy can be produced during manufacture of coating (elasto-optic effect). This orthogonal polarized light, enhanced in cavity, produces polarization optical loss. Polarization map of mirrors is very important and we propose to measure it. Polarization loss can be important in different precision optical experiments based on usage of polarized light, for example, in quantum speed meter.
astro-ph_IM
LSST Target of Opportunity proposal for locating a core collapse supernova in our galaxy triggered by a neutrino supernova alert: A few times a century, a core collapse supernova (CCSN) occurs in our galaxy. When such galactic CCSNe happen, over 99\% of its gravitational binding energy is released in the form of neutrinos. Over a period of tens of seconds, a powerful neutrino flux is emitted from the collapsing star. When the exploding shock wave finally reaches the surface of the star, optical photons escaping the expanding stellar envelope leave the star and eventually arrive at Earth as a visible brightening. Crucially, although the neutrino signal is prompt, the time to the shock wave breakout can be minutes to many hours later. This means that the neutrino signal will serve as an alert, warning the optical astronomy community the light from the explosion is coming. Quickly identifying the location of the supernova on the sky and disseminating it to the all available ground and spaced-based instruments will be critical to learn as much as possible about the event. Some neutrino experiments can report pointing information for these galactic CCSNe. In particular, the Super-Kamiokande experiment can point to a few degrees for CCSNe near the center of our galaxy. A CCSN located 10 kpc from Earth is expected to result in a pointing resolution on the order of 3 degrees. LSST's field of view (FOV) is well matched to this initial search box. LSSTs depth is also uniquely suited for identifying CCSNe even if they fail or are obscured by the dust of the galactic plane. This is a proposal to, upon receipt of such an alert, prioritize the use of LSST for a full day of observing to continuously monitor a pre-identified region of sky and, by using difference imaging, identify and announce the location of the supernova.
astro-ph_IM
Measuring the Evolution of the NuSTAR Detector Gains: The memo describes the methods used to track the long-term gain variations in the NuSTAR detectors. It builds on the analysis presented in Madsen et al. (2015) using the deployable calibration source to measure the gain drift in the NuSTAR CdZnTe detectors. This is intended to be a live document that is periodically updated as new entries are required in the NuSTAR gain CALDB files. This document covers analysis up through early-2022 and the gain v011 CALDB file released in version 20240226.
astro-ph_IM
Versatile Directional Searches for Gravitational Waves with Pulsar Timing Arrays: By regularly monitoring the most stable millisecond pulsars over many years, pulsar timing arrays (PTAs) are positioned to detect and study correlations in the timing behaviour of those pulsars. Gravitational waves (GWs) from supermassive black hole binaries (SMBHBs) are an exciting potentially detectable source of such correlations. We describe a straight-forward technique by which a PTA can be "phased-up" to form time series of the two polarisation modes of GWs coming from a particular direction of the sky. Our technique requires no assumptions regarding the time-domain behaviour of a GW signal. This method has already been used to place stringent bounds on GWs from individual SMBHBs in circular orbits. Here, we describe the methodology and demonstrate the versatility of the technique in searches for a wide variety of GW signals including bursts with unmodeled waveforms. Using the first six years of data from the Parkes Pulsar Timing Array, we conduct an all-sky search for a detectable excess of GW power from any direction. For the lines of sight to several nearby massive galaxy clusters, we carry out a more detailed search for GW bursts with memory, which are distinct signatures of SMBHB mergers. In all cases, we find that the data are consistent with noise.
astro-ph_IM
The ATLAS All-Sky Stellar Reference Catalog: The Asteroid Terrestrial-impact Last Alert System (ATLAS) observes most of the sky every night in search of dangerous asteroids. Its data are also used to search for photometric variability, where sensitivity to variability is limited by photometric accuracy. Since each exposure spans 7.6 deg corner to corner, variations in atmospheric transparency in excess of 0.01 mag are common, and 0.01 mag photometry cannot be achieved by using a constant flat field calibration image. We therefore have assembled an all-sky reference catalog of approximately one billion stars to m~19 from a variety of sources to calibrate each exposure's astrometry and photometry. Gaia DR2 is the source of astrometry for this ATLAS Refcat2. The sources of g, r, i, z photometry include Pan-STARRS DR1, the ATLAS Pathfinder photometry project, ATLAS re-flattened APASS data, SkyMapper DR1, APASS DR9, the Tycho-2 catalog, and the Yale Bright Star Catalog. We have attempted to make this catalog at least 99% complete to m<19, including the brightest stars in the sky. We believe that the systematic errors are no larger than 5 millimag RMS, although errors are as large as 20 millimag in small patches near the galactic plane.
astro-ph_IM
Data Multiplexing in Radio Interferometric Calibration: New and upcoming radio interferometers will produce unprecedented amounts of data that demand extremely powerful computers for processing. This is a limiting factor due to the large computational power and energy costs involved. Such limitations restrict several key data processing steps in radio interferometry. One such step is calibration where systematic errors in the data are determined and corrected. Accurate calibration is an essential component in reaching many scientific goals in radio astronomy and the use of consensus optimization that exploits the continuity of systematic errors across frequency significantly improves calibration accuracy. In order to reach full consensus, data at all frequencies need to be calibrated simultaneously. In the SKA regime, this can become intractable if the available compute agents do not have the resources to process data from all frequency channels simultaneously. In this paper, we propose a multiplexing scheme that is based on the alternating direction method of multipliers (ADMM) with cyclic updates. With this scheme, it is possible to simultaneously calibrate the full dataset using far fewer compute agents than the number of frequencies at which data are available. We give simulation results to show the feasibility of the proposed multiplexing scheme in simultaneously calibrating a full dataset when a limited number of compute agents are available.
astro-ph_IM
A distributed data warehouse system for astroparticle physics: A distributed data warehouse system is one of the actual issues in the field of astroparticle physics. Famous experiments, such as TAIGA, KASCADE-Grande, produce tens of terabytes of data measured by their instruments. It is critical to have a smart data warehouse system on-site to store the collected data for further distribution effectively. It is also vital to provide scientists with a handy and user-friendly interface to access the collected data with proper permissions not only on-site but also online. The latter case is handy when scientists need to combine data from different experiments for analysis. In this work, we describe an approach to implementing a distributed data warehouse system that allows scientists to acquire just the necessary data from different experiments via the Internet on demand. The implementation is based on CernVM-FS with additional components developed by us to search through the whole available data sets and deliver their subsets to users' computers.
astro-ph_IM
Swift publication statistics: a comparison with other major observatories: Swift is a satellite equipped with gamma-ray, X-ray, and optical-UV instruments aimed at discovering, localizing and collecting data from gamma-ray bursts (GRBs). Launched at the end of 2004, this small-size mission finds about a hundred GRBs per year, totaling more than 700 events as of 2012. In addition to GRBs, Swift observes other energetic events, such as AGNs, novae, and supernovae. Here we look at its success using bibliometric tools; that is the number of papers using Swift data and their impact (i.e., number of citations to those papers). We derived these for the publication years 2005 to 2011, and compared them with the same numbers for other major observatories. Swift provided data for 1101 papers in the interval 2005-2011, with 24 in the first year, to 287 in the last year. In 2011, Swift had more than double the number of publications as Subaru, it overcame Gemini by a large fraction, and reached Keck. It is getting closer to the ~400 publications of the successful high-energy missions XMM-Newton and Chandra, but is still far from the most productive telescopes VLT (over 500) and HST (almost 800). The overall average number of citations per paper, as of November 2012, is 28.3, which is comparable to the others, but lower than Keck (41.8). The science topics covered by Swift publications have changed from the first year, when over 80% of the papers were about GRBs, while in 2011 it was less than 30%.
astro-ph_IM
Status of predictive wavefront control on Keck II adaptive optics bench: on-sky coronagraphic results: The behavior of an adaptive optics (AO) system for ground-based high contrast imaging (HCI) dictates the achievable contrast of the instrument. In conditions where the coherence time of the atmosphere is short compared to the speed of the AO system, the servo-lag error becomes the dominate error term of the AO system. While the AO system measures the wavefront error and subsequently applies a correction (taking a total of 1 to 2 milli-seconds), the atmospheric turbulence above the telescope has changed. In addition to reducing the Strehl ratio, the servo-lag error causes a build-up of speckles along the direction of the dominant wind vector in the coronagraphic image, severely limiting the contrast at small angular separations. One strategy to mitigate this problem is to predict the evolution of the turbulence over the delay. Our predictive wavefront control algorithm minimizes the delay in a mean square sense and has been implemented on the Keck II AO bench. In this paper we report on the latest results of our algorithm and discuss updates to the algorithm itself. We explore how to tune various filter parameters on the basis of both daytime laboratory tests and on-sky tests. We show a reduction in residual-mean-square wavefront error for the predictor compare to the leaky integrator implemented on Keck. Finally, we present contrast improvements for both day time and on-sky tests. Using the L-band vortex coronagraph for Keck's NIRC2 instrument, we find a contrast gain of 2.03 at separation of 3~$\lambda/D$ and up to 3 for larger separations (4-6~$\lambda/D$).
astro-ph_IM
hammurabi X: Simulating Galactic Synchrotron Emission with Random Magnetic Fields: We present version X of the hammurabi package, the HEALPix-based numeric simulator for Galactic polarized emission. Improving on its earlier design, we have fully renewed the framework with modern C++ standards and features. Multi-threading support has been built in to meet the growing computational workload in future research. For the first time, we present precision profiles of hammurabi line-of-sight integral kernel with multi-layer HEALPix shells. In addition to fundamental improvements, this report focuses on simulating polarized synchrotron emission with Gaussian random magnetic fields. Two fast methods are proposed for realizing divergence-free random magnetic fields either on the Galactic scale where a field alignment and strength modulation are imposed, or on a local scale where more physically motivated models like a parameterized magneto-hydrodynamic (MHD) turbulence can be applied. As an example application, we discuss the phenomenological implications of Gaussian random magnetic fields for high Galactic latitude synchrotron foregrounds. In this, we numerically find B/E polarization mode ratios lower than unity based on Gaussian realizations of either MHD turbulent spectra or in spatially aligned magnetic fields.
astro-ph_IM
Modelling astronomical adaptive optics performance with temporally-filtered Wiener reconstruction of slope data: We build on a long-standing tradition in astronomical adaptive optics (AO) of specifying performance metrics and error budgets using linear systems modeling in the spatial-frequency domain. Our goal is to provide a comprehensive tool for the calculation of error budgets in terms of residual temporally filtered phase power spectral densities and variances. In addition, the fast simulation of AO-corrected point spread functions (PSFs) provided by this method can be used as inputs for simulations of science observations with next-generation instruments and telescopes, in particular to predict post-coronagraphic contrast improvements for planet finder systems. We extend the previous results and propose the synthesis of a distributed Kalman filter to mitigate both aniso-servo-lag and aliasing errors whilst minimizing the overall residual variance. We discuss applications to (i) analytic AO-corrected PSF modeling in the spatial-frequency domain, (ii) post-coronagraphic contrast enhancement, (iii) filter optimization for real-time wavefront reconstruction, and (iv) PSF reconstruction from system telemetry. Under perfect knowledge of wind velocities, we show that $\sim$60 nm rms error reduction can be achieved with the distributed Kalman filter embodying anti- aliasing reconstructors on 10 m class high-order AO systems, leading to contrast improvement factors of up to three orders of magnitude at few ${\lambda}/D$ separations ($\sim1-5{\lambda}/D$) for a 0 magnitude star and reaching close to one order of magnitude for a 12 magnitude star.
astro-ph_IM
A Bayesian approach to high fidelity interferometric calibration II: demonstration with simulated data: In a companion paper, we presented BayesCal, a mathematical formalism for mitigating sky-model incompleteness in interferometric calibration. In this paper, we demonstrate the use of BayesCal to calibrate the degenerate gain parameters of full-Stokes simulated observations with a HERA-like hexagonal close-packed redundant array, for three assumed levels of completeness of the a priori known component of the calibration sky model. We compare the BayesCal calibration solutions to those recovered by calibrating the degenerate gain parameters with only the a priori known component of the calibration sky model both with and without imposing physically motivated priors on the gain amplitude solutions and for two choices of baseline length range over which to calibrate. We find that BayesCal provides calibration solutions with up to four orders of magnitude lower power in spurious gain amplitude fluctuations than the calibration solutions derived for the same data set with the alternate approaches, and between $\sim10^7$ and $\sim10^{10}$ times smaller than in the mean degenerate gain amplitude on the full range of spectral scales accessible in the data. Additionally, we find that in the scenarios modelled only BayesCal has sufficiently high fidelity calibration solutions for unbiased recovery of the 21 cm power spectrum on large spectral scales ($k_\parallel \lesssim 0.15~h\mathrm{Mpc}^{-1}$). In all other cases, in the completeness regimes studied, those scales are contaminated.
astro-ph_IM
Point Source Detection and Flux Determination with PGWave: One of the largest uncertainties in the Point Source (PS) studies, at Fermi-LAT energies, is the uncertainty in the diffuse background. In general there are two approaches for PS analysis: background-dependent methods, that include modeling of the diffuse background, and background-independent methods. In this work we study PGWave, which is one of the background-independent methods, based on wavelet filtering to find significant clusters of gamma rays. PGWave is already used in the Fermi-LAT catalog pipeline for finding candidate sources. We test PGWave, not only for source detection, but especially to estimate the flux without the need of a background model. We use Monte Carlo (MC) simulation to study the accuracy of PS detection and estimation of the flux. We present preliminary results of these MC studies.
astro-ph_IM
Is HDF5 a good format to replace UVFITS?: The FITS (Flexible Image Transport System) data format was developed in the late 1970s for storage and exchange of astronomy-related image data. Since then, it has become a standard file format not only for images, but also for radio interferometer data (e.g. UVFITS, FITS-IDI). But is FITS the right format for next-generation telescopes to adopt? The newer Hierarchical Data Format (HDF5) file format offers considerable advantages over FITS, but has yet to gain widespread adoption within radio astronomy. One of the major holdbacks is that HDF5 is not well supported by data reduction software packages. Here, we present a comparison of FITS, HDF5, and the MeasurementSet (MS) format for storage of interferometric data. In addition, we present a tool for converting between formats. We show that the underlying data model of FITS can be ported to HDF5, a first step toward achieving wider HDF5 support.
astro-ph_IM
2HOT: An Improved Parallel Hashed Oct-Tree N-Body Algorithm for Cosmological Simulation: We report on improvements made over the past two decades to our adaptive treecode N-body method (HOT). A mathematical and computational approach to the cosmological N-body problem is described, with performance and scalability measured up to 256k ($2^{18}$) processors. We present error analysis and scientific application results from a series of more than ten 69 billion ($4096^3$) particle cosmological simulations, accounting for $4 \times 10^{20}$ floating point operations. These results include the first simulations using the new constraints on the standard model of cosmology from the Planck satellite. Our simulations set a new standard for accuracy and scientific throughput, while meeting or exceeding the computational efficiency of the latest generation of hybrid TreePM N-body methods.
astro-ph_IM
Apertif, Phased Array Feeds for the Westerbork Synthesis Radio Telescope: We describe the APERture Tile In Focus (Apertif) system, a phased array feed (PAF) upgrade of the Westerbork Synthesis Radio Telescope which has transformed this telescope into a high-sensitivity, wide field-of-view L-band imaging and transient survey instrument. Using novel PAF technology, up to 40 partially overlapping beams can be formed on the sky simultaneously, significantly increasing the survey speed of the telescope. With this upgraded instrument, an imaging survey covering an area of 2300 deg2 is being performed which will deliver both continuum and spectral line data sets, of which the first data has been publicly released. In addition, a time domain transient and pulsar survey covering 15,000 deg2 is in progress. An overview of the Apertif science drivers, hardware and software of the upgraded telescope is presented, along with its key performance characteristics.
astro-ph_IM
On Point Spread Function modelling: towards optimal interpolation: Point Spread Function (PSF) modeling is a central part of any astronomy data analysis relying on measuring the shapes of objects. It is especially crucial for weak gravitational lensing, in order to beat down systematics and allow one to reach the full potential of weak lensing in measuring dark energy. A PSF modeling pipeline is made of two main steps: the first one is to assess its shape on stars, and the second is to interpolate it at any desired position (usually galaxies). We focus on the second part, and compare different interpolation schemes, including polynomial interpolation, radial basis functions, Delaunay triangulation and Kriging. For that purpose, we develop simulations of PSF fields, in which stars are built from a set of basis functions defined from a Principal Components Analysis of a real ground-based image. We find that Kriging gives the most reliable interpolation, significantly better than the traditionally used polynomial interpolation. We also note that although a Kriging interpolation on individual images is enough to control systematics at the level necessary for current weak lensing surveys, more elaborate techniques will have to be developed to reach future ambitious surveys' requirements.
astro-ph_IM
Two modified ILC methods to detect point sources in Cosmic Microwave Background maps: We propose two detection techniques that take advantage of a small sky area approximation and are based on modifications of the "internal linear combination" (ILC) method, an approach widely used in Cosmology for the separation of the various components that contribute to the microwave background. The main advantage of the proposed approach, especially in handling multi-frequency maps of the same region, is that it does not require the "a priori" knowledge of the spatial power-spectrum of either the CMB and/or the Galactic foreground. Hence, it is more robust, easier and more intuitive to use. The performance of the proposed algorithms is tested with numerical experiments that mimic the physical scenario expected for high Galactic latitude observations with the Atacama Large Millimeter/submillimeter Array (ALMA).
astro-ph_IM
Three recipes for improving the image quality with optical long-baseline interferometers: BFMC, LFF, \& DPSC: We present here three recipes for getting better images with optical interferometers. Two of them, Low- Frequencies Filling and Brute-Force Monte Carlo were used in our participation to the Interferometry Beauty Contest this year and can be applied to classical imaging using V 2 and closure phases. These two addition to image reconstruction provide a way of having more reliable images. The last recipe is similar in its principle as the self-calibration technique used in radio-interferometry. We call it also self-calibration, but it uses the wavelength-differential phase as a proxy of the object phase to build-up a full-featured complex visibility set of the observed object. This technique needs a first image-reconstruction run with an available software, using closure-phases and squared visibilities only. We used it for two scientific papers with great success. We discuss here the pros and cons of such imaging technique.
astro-ph_IM
NEBULAR: A simple synthesis code for the hydrogen and helium nebular spectrum: NEBULAR is a lightweight code to synthesize the spectrum of an ideal, mixed hydrogen and helium gas in ionization equilibrium, over a useful range of densities, temperatures and wavelengths. Free-free, free-bound and two-photon continua are included as well as parts of the HI, HeI and HeII line series. NEBULAR interpolates over publicly available data tables; it can be used to easily extract information from these tables without prior knowledge about their data structure. The resulting spectra can be used to e.g. determine equivalent line widths, constrain the contribution of the nebular continuum to a bandpass, and for educational purposes. NEBULAR can resample the spectrum on a user-defined wavelength grid for direct comparison with an observed spectrum; however, it can not be used to fit an observed spectrum.
astro-ph_IM
Prospects for a radio air-shower detector at South Pole: IceCube is currently not only the largest neutrino telescope but also one of the world's most competitive instruments for studying cosmic rays in the PeV to EeV regime where the transition from galactic to extra-galactic sources should occur. Further augmenting this observatory with an array of radio sensors in the 10-100 MHz regime will additionally permit observation of the geomagnetic radio emission from the air shower. Yielding complementary information on the shower development a triple-technology array consisting of radio sensors, the ground sampling stations of IceTop and the in-ice optical modules of IceCube, should significantly improve the understanding of cosmic rays, as well as enhance many aspects of the physics reach of the observatory. Here we present first results from two exploratory setups deployed at the South Pole. Noise measurements from data taken in two consecutive seasons show a very good agreement of the predicted and observed response of the antennas designed specifically for this purpose. The radio background is found to be highly dominated by galactic noise with a striking absence of anthropogenic radio emitters in the frequency band from 25-300 MHz. Motivated by the excellent suitability of the location, we present first performance studies of a proposed Radio Air-Shower Test Array (RASTA) using detailed MonteCarlo simulation and discuss the prospects for its installation.
astro-ph_IM