text
stringlengths
89
2.49k
category
stringclasses
19 values
Safeguarding Old and New Journal Tables for the VO: Status for Extragalactic and Radio Data: Independent of established data centers, and partly for my own research, since 1989 I have been collecting the tabular data from over 2600 articles concerned with radio sources and extragalactic objects in general. Optical character recognition (OCR) was used to recover tables from 740 papers. Tables from only 41 percent of the 2600 articles are available in the CDS or CATS catalog collections, and only slightly better coverage is estimated for the NED database. This fraction is not better for articles published electronically since 2001. Both object databases (NED, SIMBAD, LEDA) as well as catalog browsers (VizieR, CATS) need to be consulted to obtain the most complete information on astronomical objects. More human resources at the data centers and better collaboration between authors, referees, editors, publishers, and data centers are required to improve data coverage and accessibility. The current efforts within the Virtual Observatory (VO) project, to provide retrieval and analysis tools for different types of published and archival data stored at various sites, should be balanced by an equal effort to recover and include large amounts of published data not currently available in this way.
astro-ph_IM
Wavelet transforms of microlensing data: Denoising, extracting intrinsic pulsations, and planetary signals: Wavelets are waveform functions that describe transient and unstable variations, such as noises. In this work, we study the advantages of discrete and continuous wavelet transforms (DWT and CWT) of microlensing data to denoise them and extract their planetary signals and intrinsic pulsations hidden by noises. We first generate synthetic microlensing data and apply wavelet denoising to them. For these simulated microlensing data with ideally Gaussian nosies based on the OGLE photometric accuracy, denoising with DWT reduces standard deviations of data from real models by $0.044$-$0.048$ mag. The efficiency to regenerate real models and planetary signals with denoised data strongly depends on the observing cadence and decreases from $37\%$ to $0.01\%$ by worsening cadence from $15$ min to $6$ hrs. We then apply denoising on $100$ microlensing events discovered by the OGLE group. On average, wavelet denoising for these data improves standard deviations and $\chi^{2}_{\rm n}$ of data with respect to the best-fitted models by $0.023$ mag, and $1.16$, respectively. The best-performing wavelets (based on either the highest signal-to-noise ratio's peak ($\rm{SNR}_{\rm{max}}$), or the highest Pearson's correlation, or the lowest Root Mean Squared Error (RMSE) for denoised data) are from 'Symlet', and 'Biorthogonal' wavelets families in simulated, and OGLE data, respectively. In some denoised data, intrinsic stellar pulsations or small planetary-like deviations appear which were covered with noises in raw data. However, through DWT denoising rather flattened and wide planetary signals could be reconstructed than sharp signals. CWT and 3D frequency-power-time maps could advise about the existence of sharp signals.
astro-ph_IM
The DESI Experiment Part II: Instrument Design: DESI (Dark Energy Spectropic Instrument) is a Stage IV ground-based dark energy experiment that will study baryon acoustic oscillations and the growth of structure through redshift-space distortions with a wide-area galaxy and quasar redshift survey. The DESI instrument is a robotically-actuated, fiber-fed spectrograph capable of taking up to 5,000 simultaneous spectra over a wavelength range from 360 nm to 980 nm. The fibers feed ten three-arm spectrographs with resolution $R= \lambda/\Delta\lambda$ between 2000 and 5500, depending on wavelength. The DESI instrument will be used to conduct a five-year survey designed to cover 14,000 deg$^2$. This powerful instrument will be installed at prime focus on the 4-m Mayall telescope in Kitt Peak, Arizona, along with a new optical corrector, which will provide a three-degree diameter field of view. The DESI collaboration will also deliver a spectroscopic pipeline and data management system to reduce and archive all data for eventual public use.
astro-ph_IM
$\texttt{GWFAST}$: a Fisher information matrix Python code for third-generation gravitational-wave detectors: We introduce $\texttt{GWFAST}$, a Fisher information matrix $\texttt{Python}$ code that allows easy and efficient estimation of signal-to-noise ratios and parameter measurement errors for large catalogs of resolved sources observed by networks of gravitational-wave detectors. In particular, $\texttt{GWFAST}$ includes the effects of the Earth's motion during the evolution of the signal, supports parallel computation, and relies on automatic differentiation rather than on finite differences techniques, which allows the computation of derivatives with accuracy close to machine precision. We also release the library $\texttt{WF4Py}$ implementing state-of-the-art gravitational-wave waveforms in $\texttt{Python}$. In this paper we provide a documentation of $\texttt{GWFAST}$ and $\texttt{WF4Py}$ with practical examples and tests of performance and reliability. In a companion paper we present forecasts for the detection capabilities of the second and third generation of ground-based gravitational-wave detectors, obtained with $\texttt{GWFAST}$.
astro-ph_IM
A fast new catadioptric design for fiber-fed spectrographs: The next generation of massively multiplexed multi-object spectrographs (DESpec, SUMIRE, BigBOSS, 4MOST, HECTOR) demand fast, efficient and affordable spectrographs, with higher resolutions (R = 3000-5000) than current designs. Beam-size is a (relatively) free parameter in the design, but the properties of VPH gratings are such that, for fixed resolution and wavelength coverage, the effect on beam-size on overall VPH efficiency is very small. For alltransmissive cameras, this suggests modest beam-sizes (say 80-150mm) to minimize costs; while for catadioptric (Schmidt-type) cameras, much larger beam-sizes (say 250mm+) are preferred to improve image quality and to minimize obstruction losses. Schmidt designs have benefits in terms of image quality, camera speed and scattered light performance, and recent advances such as MRF technology mean that the required aspherics are no longer a prohibitive cost or risk. A new Schmidt/Maksutov-derived design is presented, which differs from previous designs in having the detector package outside the camera, and adjacent to the spectrograph pupil. The telescope pupil already contains a hole at its center, because of the obstruction from the telescope top-end. With a 250mm beam, it is possible to largely hide a 6cm \times 6cm detector package and its dewar within this hole. This means that the design achieves a very high efficiency, competitive with transmissive designs. The optics are excellent, as least as good as classic Schmidt designs, allowing F/1.25 or even faster cameras. The principal hardware has been costed at $300K per arm, making the design affordable.
astro-ph_IM
Three-dimensional extinction mapping using Gaussian random fields: We present a scheme for using stellar catalogues to map the three-dimensional distributions of extinction and dust within our Galaxy. Extinction is modelled as a Gaussian random field, whose covariance function is set by a simple physical model of the ISM that assumes a Kolmogorov-like power spectrum of turbulent fluctuations. As extinction is modelled as a random field, the spatial resolution of the resulting maps is set naturally by the data available; there is no need to impose any spatial binning. We verify the validity of our scheme by testing it on simulated extinction fields and show that its precision is significantly improved over previous dust-mapping efforts. The approach we describe here can make use of any photometric, spectroscopic or astrometric data; it is not limited to any particular survey. Consequently, it can be applied to a wide range of data from both existing and future surveys.
astro-ph_IM
INTEGRAL/IBIS nine-year Galactic Hard X-Ray Survey: Context. The INTEGRAL observatory operating in a hard X-ray/gamma domain has gathered a large observational data set over nine years starting in 2003. Most of the observing time was dedicated to the Galactic source population study, making possible the deepest Galactic survey in hard X-rays ever compiled. Aims. We aim to perform a Galactic survey that can be used as the basis of Galactic source population studies, and perform mapping of the Milky Way in hard X-rays over the maximum exposure available at |b|<17.5 deg. Methods. We used sky reconstruction algorithms especially developed for the high quality imaging of INTEGRAL/IBIS data. Results. We present sky images, sensitivity maps, and catalogs of detected sources in the three energy bands 17-60, 17-35, and 35-80 keV in the Galactic plane at |b|<17.5 deg. The total number of sources in the reference 17-60 keV band includes 402 objects exceeding a 4.7 sigma detection threshold on the nine-year time-averaged map. Among the identified sources with known and tentatively identified natures, 253 are Galactic objects (108 low-mass X-ray binaries, 82 high-mass X-ray binaries, 36 cataclysmic variables, and 27 are of other types), and 115 are extragalactic objects, including 112 active galactic nuclei (AGNs) and 3 galaxy clusters. The sample of Galactic sources with S/N>4.7 sigma has an identification completeness of ~92%, which is valuable for population studies. Since the survey is based on the nine-year sky maps, it is optimized for persistent sources and may be biased against finding transients.
astro-ph_IM
gSeaGen: a GENIE-based code for neutrino telescopes: The gSeaGen code is a GENIE based application to generate neutrino-induced events in an underwater neutrino detector. The gSeaGen code is able to generate events induced by all neutrino flavours, taking into account topological differences between track-type and shower-like events. The neutrino interaction is simulated taking into account the density and the composition of the media surrounding the detector. The main features of gSeaGen will be presented together with some examples of its application within ANTARES and KM3NeT.
astro-ph_IM
Spectro-photometric distances to stars: a general-purpose Bayesian approach: We developed a code that estimates distances to stars using measured spectroscopic and photometric quantities. We employ a Bayesian approach to build the probability distribution function over stellar evolutionary models given these data, delivering estimates of model parameters for each star individually. The code was first tested on simulations, successfully recovering input distances to mock stars with <1% bias.The method-intrinsic random distance uncertainties for typical spectroscopic survey measurements amount to around 10% for dwarf stars and 20\% for giants, and are most sensitive to the quality of $\log g$ measurements. The code was validated by comparing our distance estimates to parallax measurements from the Hipparcos mission for nearby stars (< 300 pc), to asteroseismic distances of CoRoT red giant stars, and to known distances of well-studied open and globular clusters. The external comparisons confirm that our distances are subject to very small systematic biases with respect to the fundamental Hipparcos scale (+0.4 % for dwarfs, and +1.6% for giants). The typical random distance scatter is 18% for dwarfs, and 26% for giants. For the CoRoT-APOGEE sample, the typical random distance scatter is ~15%, both for the nearby and farther data. Our distances are systematically larger than the CoRoT ones by about +9%, which can mostly be attributed to the different choice of priors. The comparison to known distances of star clusters from SEGUE and APOGEE has led to significant systematic differences for many cluster stars, but with opposite signs, and with substantial scatter. Finally, we tested our distances against those previously determined for a high-quality sample of giant stars from the RAVE survey, again finding a small systematic trend of +5% and an rms scatter of 30%.
astro-ph_IM
Smoothed Particle Radiation Hydrodynamics: Two-Moment method with Local Eddington Tensor Closure: We present a new radiative transfer method (SPH-M1RT) that is coupled dynamically with smoothed particle hydrodynamics (SPH). We implement it in the (task-based parallel) SWIFT galaxy simulation code but it can be straightforwardly implemented in other SPH codes. Our moment-based method simultaneously solves the radiation energy and flux equations in SPH, making it adaptive in space and time. We modify the M1 closure relation to stabilize radiation fronts in the optically thin limit. We also introduce anisotropic artificial viscosity and high-order artificial diffusion schemes, which allow the code to handle radiation transport accurately in both the optically thin and optically thick regimes. Non-equilibrium thermo-chemistry is solved using a semi-implicit sub-cycling technique. The computational cost of our method is independent of the number of sources and can be lowered further by using the reduced speed of light approximation. We demonstrate the robustness of our method by applying it to a set of standard tests from the cosmological radiative transfer comparison project of Iliev et al. The SPH-M1RT scheme is well-suited for modelling situations in which numerous sources emit ionising radiation, such as cosmological simulations of galaxy formation or simulations of the interstellar medium.
astro-ph_IM
Dealing with missing data in the MICROSCOPE space mission: An adaptation of inpainting to handle colored-noise data: The MICROSCOPE space mission, launched on April 25, 2016, aims to test the weak equivalence principle (WEP) with a 10^-15 precision. To reach this performance requires an accurate and robust data analysis method, especially since the possible WEP violation signal will be dominated by a strongly colored noise. An important complication is brought by the fact that some values will be missing -therefore, the measured time series will not be strictly regularly sampled. Those missing values induce a spectral leakage that significantly increases the noise in Fourier space, where the WEP violation signal is looked for, thereby complicating scientific returns. Recently, we developed an inpainting algorithm to correct the MICROSCOPE data for missing values. This code has been integrated in the official MICROSCOPE data processing pipeline because it enables us to significantly measure an equivalence principle violation (EPV) signal in a model-independent way, in the inertial satellite configuration. In this work, we present several improvements to the method that may allow us now to reach the MICROSCOPE requirements for both inertial and spin satellite configurations. The main improvement has been obtained using a prior on the power spectrum of the colored-noise that can be directly derived from the incomplete data. We show that after reconstructing missing values with this new algorithm, a least-squares fit may allow us to significantly measure an EPV signal with a 0.96x10^-15 precision in the inertial mode and 1.2x10^-15 precision in the spin mode. Although, the inpainting method presented in this paper has been optimized to the MICROSCOPE data, it remains sufficiently general to be used in the general context of missing data in time series dominated by an unknown colored-noise. The improved inpainting software, called ICON, is freely available at http://www.cosmostat.org/software/icon.
astro-ph_IM
Data Release 2 of S-PLUS: accurate template-fitting based photometry covering $\sim$1000 square degrees in 12 optical filters: The Southern Photometric Local Universe Survey (S-PLUS) is an ongoing survey of $\sim$9300 deg$^2$ in the southern sky in a 12-band photometric system. This paper presents the second data release (DR2) of S-PLUS, consisting of 514 tiles covering an area of 950 deg$^2$. The data has been fully calibrated using a new photometric calibration technique suitable for the new generation of wide-field multi-filter surveys. This technique consists of a $\chi^2$ minimisation to fit synthetic stellar templates to already calibrated data from other surveys, eliminating the need for standard stars and reducing the survey duration by $\sim$15\%. We compare the template-predicted and S-PLUS instrumental magnitudes to derive the photometric zero-points (ZPs). We show that these ZPs can be further refined by fitting the stellar templates to the 12 S-PLUS magnitudes, which better constrain the models by adding the narrow-band information. We use the STRIPE82 region to estimate ZP errors, which are $\lesssim10$ mmags for filters J0410, J0430, $g$, J0515, $r$, J0660, $i$, J0861 and $z$; $\lesssim 15$ mmags for filter J0378; and $\lesssim 25$ mmags for filters $u$ and J0395. We describe the complete data flow of the S-PLUS/DR2 from observations to the final catalogues and present a brief characterisation of the data. We show that, for a minimum signal-to-noise threshold of 3, the photometric depths of the DR2 range from 19.9 mag to 21.3 mag (measured in Petrosian apertures), depending on the filter. The S-PLUS DR2 can be accessed from the website: https://splus.cloud}{https://splus.cloud.
astro-ph_IM
Fast and Automated Peak Bagging with DIAMONDS (FAMED): Stars of low and intermediate mass that exhibit oscillations may show tens of detectable oscillation modes each. Oscillation modes are a powerful to constrain the internal structure and rotational dynamics of the star, hence tool allowing one to obtain an accurate stellar age. The tens of thousands of solar-like oscillators that have been discovered thus far are representative of the large diversity of fundamental stellar properties and evolutionary stages available. Because of the wide range of oscillation features that can be recognized in such stars, it is particularly challenging to properly characterize the oscillation modes in detail, especially in light of large stellar samples. Overcoming this issue requires an automated approach, which has to be fast, reliable, and flexible at the same time. In addition, this approach should not only be capable of extracting the oscillation mode properties of frequency, linewidth, and amplitude from stars in different evolutionary stages, but also able to assign a correct mode identification for each of the modes extracted. Here we present the new freely available pipeline FAMED (Fast and AutoMated pEak bagging with DIAMONDS), which is capable of performing an automated and detailed asteroseismic analysis in stars ranging from the main sequence up to the core-Helium-burning phase of stellar evolution. This, therefore, includes subgiant stars, stars evolving along the red giant branch (RGB), and stars likely evolving toward the early asymptotic giant branch. In this paper, we additionally show how FAMED can detect rotation from dipolar oscillation modes in main sequence, subgiant, low-luminosity RGB, and core-Helium-burning stars. FAMED can be downloaded from its public GitHub repository (https://github.com/EnricoCorsaro/FAMED).
astro-ph_IM
Searching for changing-state AGNs in massive datasets -- I: applying deep learning and anomaly detection techniques to find AGNs with anomalous variability behaviours: The classic classification scheme for Active Galactic Nuclei (AGNs) was recently challenged by the discovery of the so-called changing-state (changing-look) AGNs (CSAGNs). The physical mechanism behind this phenomenon is still a matter of open debate and the samples are too small and of serendipitous nature to provide robust answers. In order to tackle this problem, we need to design methods that are able to detect AGN right in the act of changing-state. Here we present an anomaly detection (AD) technique designed to identify AGN light curves with anomalous behaviors in massive datasets. The main aim of this technique is to identify CSAGN at different stages of the transition, but it can also be used for more general purposes, such as cleaning massive datasets for AGN variability analyses. We used light curves from the Zwicky Transient Facility data release 5 (ZTF DR5), containing a sample of 230,451 AGNs of different classes. The ZTF DR5 light curves were modeled with a Variational Recurrent Autoencoder (VRAE) architecture, that allowed us to obtain a set of attributes from the VRAE latent space that describes the general behaviour of our sample. These attributes were then used as features for an Isolation Forest (IF) algorithm, that is an anomaly detector for a "one class" kind of problem. We used the VRAE reconstruction errors and the IF anomaly score to select a sample of 8,809 anomalies. These anomalies are dominated by bogus candidates, but we were able to identify 75 promising CSAGN candidates.
astro-ph_IM
Simulation of ionizing radiation in cell phone camera image sensors: The Distributed Electronic Cosmic-ray Observatory (DECO) is a cell phone app that uses a cell phone camera image sensor to detect cosmic-ray particles and particles from radioactive decay. Images recorded by DECO are classified by a convolutional neural network (CNN) according to their morphology. In this project, we develop a GEANT4-derived simulation of particle interactions inside the CMOS sensor using the Allpix$^2$ modular framework. We simulate muons, electrons, and photons with energy range 10 keV to 100 GeV, and their deposited energy agrees well with expectations. Simulated events are recorded and processed in a similar way as data images taken by DECO, and the result shows both similar image morphology with data events and good quantitative data-Monte Carlo agreement.
astro-ph_IM
Design and optimization of dihedral angle offsets for the next generation lunar retro-reflectors: Lunar laser ranging (LLR) to the Apollo retro-reflectors, which features the most long-lasting experiment in testing General Relativity theories, has remained operational over the past four decades. To date, with significant improvement of ground observatory conditions, the bottleneck of LLR accuracy lies in the retro-reflectors. A new generation of large aperture retro-reflectors with intended dihedral angle offsets have been suggested and implemented based on NASA's recent lunar projects to reduce its ranging uncertainty to be less than 1.0 mm. The technique relies on the retro-reflector's ability to offset its relative angular velocity with regard to a ground LLR observatory (LLRO), so that the LLR accuracy can be ensured along with the larger area of beam reflection. In deployment, solid corner-cube reflectors (CCRs) based on empirical successes of the Apollo 11 and 15 arrays have been selected for the next generation lunar reflectors (NGLRs) due to their stability against heat and dust problems on the Moon. In this work, we present the optical effects in designing the new retro-reflectors given various sets of intended diheral angle offsets (DAOs), and support the design principles with the measurements of of two manufactured NGLRs.
astro-ph_IM
Atmospheric effects on extensive air showers observed with the Surface Detector of the Pierre Auger Observatory: Atmospheric parameters, such as pressure (P), temperature (T) and density, affect the development of extensive air showers initiated by energetic cosmic rays. We have studied the impact of atmospheric variations on extensive air showers by means of the surface detector of the Pierre Auger Observatory. The rate of events shows a ~10% seasonal modulation and ~2% diurnal one. We find that the observed behaviour is explained by a model including the effects associated with the variations of pressure and density. The former affects the longitudinal development of air showers while the latter influences the Moliere radius and hence the lateral distribution of the shower particles. The model is validated with full simulations of extensive air showers using atmospheric profiles measured at the site of the Pierre Auger Observatory.
astro-ph_IM
Prototype Schwarzschild-Couder Telescope for the Cherenkov Telescope Array: Commissioning Status of the Optical System: The Cherenkov Telescope Array (CTA), with more than 100 telescopes, will be the largest ever ground-based gamma-ray observatory and is expected to greatly improve on both gamma-ray detection sensitivity and energy coverage compared to current-generation detectors. The 9.7-m Schwarzschild-Couder telescope (SCT) is one of the two candidates for the medium size telescope (MST) design for CTA. The novel aplanatic dual-mirror SCT design offers a wide field-of-view with a compact plate scale, allowing for a large number of camera pixels that improves the angular resolution and reduce the night sky background noise per pixel compared to the traditional single-mirror Davies-Cotton (DC) design of ground-based gamma-ray telescopes. The production, installation, and the alignment of the segmented aspherical mirrors are the main challenges for the realization of the SCT optical system. In this contribution, we report on the commissioning status, the alignment procedures, and initial alignment results during the initial commissioning phase of the optical system of the prototype SCT.
astro-ph_IM
PulsarX: a new pulsar searching package -I. A high performance folding program for pulsar surveys: Pulsar surveys with modern radio telescopes are becoming increasingly computationally demanding. This is particularly true for wide field-of-view pulsar surveys with radio interferometers, and those conducted in real or quasi-real time. These demands result in data analysis bottlenecks that can limit the parameter space covered by the surveys and diminish their scientific return. In this paper, we address the computational challenge of `candidate folding' in pulsar searching, presenting a novel, efficient approach designed to optimise the simultaneous folding of large numbers of pulsar candidates. We provide a complete folding pipeline appropriate for large-scale pulsar surveys including radio frequency interference (RFI) mitigation, dedispersion, folding and parameter optimization. By leveraging the Fast Discrete Dispersion Measure Transform (FDMT) algorithm proposed by Zackay et al. (2017), we have developed an optimized, and cache-friendly implementation that we term the pruned FDMT (pFDMT). The pFDMT approach efficiently reuses intermediate processing results and prunes the unused computation paths, resulting in a significant reduction in arithmetic operations. In addition, we propose a novel folding algorithm based on the Tikhonov-regularised least squares method (TLSM) that can improve the time resolution of the pulsar profile. We present the performance of its real-world application as an integral part of two major pulsar search projects conducted with the MeerKAT telescope: the MPIfR-MeerKAT Galactic Plane Survey (MMGPS) and the Transients and Pulsars with MeerKAT (TRAPUM) project. In our processing, for approximately 500 candidates, the theoretical number of dedispersion operations can be reduced by a factor of around 50 when compared to brute-force dedispersion, which scales with the number of candidates.
astro-ph_IM
A new ray-tracing scheme for 3D diffuse radiation transfer on highly parallel architectures: We present a new numerical scheme to solve the transfer of diffuse radiation on three-dimensional mesh grids which is efficient on processors with highly parallel architecture such as recently popular GPUs and CPUs with multi- and many-core architectures. The scheme is based on the ray-tracing method and the computational cost is proportional to $N_{\rm m}^{5/3}$ where $N_{\rm m}$ is the number of mesh grids, and is devised to compute the radiation transfer along each light-ray completely in parallel with appropriate grouping of the light-rays. We find that the performance of our scheme scales well with the number of adopted CPU cores and GPUs, and also that our scheme is nicely parallelized on a multi-node system by adopting the multiple wave front scheme, and the performance scales well with the amount of the computational resources. As numerical tests to validate our scheme and to give a physical criterion for the angular resolution of our ray-tracing scheme, we perform several numerical simulations of the photo-ionization of neutral hydrogen gas by ionizing radiation sources without the "on-the-spot" approximation, in which the transfer of diffuse radiation by radiative recombination is incorporated in a self-consistent manner.
astro-ph_IM
HAWC response to atmospheric electricity activity: The HAWC Gamma Ray observatory consists of 300 water Cherenkov detectors (WCD) instrumented with four photo multipliers tubes (PMT) per WCD. HAWC is located between two of the highest mountains in Mexico. The high altitude (4100 m asl), the relatively short distance to the Gulf of Mexico (~100 km), the large detecting area (22 000 m$^2$) and its high sensitivity, make HAWC a good instrument to explore the acceleration of particles due to the electric fields existing inside storm clouds. In particular, the scaler system of HAWC records the output of each one of the 1200 PMTs as well as the 2, 3, and 4-fold multiplicities (logic AND in a time window of 30 ns) of each WCD with a sampling rate of 40 Hz. Using the scaler data, we have identified 20 enhancements of the observed rate during periods when storm clouds were over HAWC but without cloud-earth discharges. These enhancements can be produced by electrons with energy of tens of MeV, accelerated by the electric fields of tens of kV/m measured at the site during the storm periods. In this work, we present the recorded data, the method of analysis and our preliminary conclusions on the electron acceleration by the electric fields inside the clouds.
astro-ph_IM
Measurements of Charge Transfer Efficiency in a Proton-irradiated Swept Charge Device: Charge Coupled Devices (CCDs) have been successfully used in several low energy X-ray astronomical satellite over the past two decades. Their high energy resolution and high spatial resolution make them an perfect tool for low energy astronomy, such as formation of galaxy clusters and environment of black holes. The Low Energy X-ray Telescope (LE) group is developing Swept Charge Device (SCD) for the Hard X-ray Modulation Telescope (HXMT) satellite. SCD is a special low energy X-ray CCD, which could be read out a thousand times faster than traditional CCDs, simultaneously keeping excellent energy resolution. A test method for measuring the charge transfer efficiency (CTE) of a prototype SCD has been set up. Studies of the charge transfer inefficiency (CTI) have been performed at a temperature range of operation, with a proton-irradiated SCD.
astro-ph_IM
First Impressions: Early-Time Classification of Supernovae using Host Galaxy Information and Shallow Learning: Substantial effort has been devoted to the characterization of transient phenomena from photometric information. Automated approaches to this problem have taken advantage of complete phase-coverage of an event, limiting their use for triggering rapid follow-up of ongoing phenomena. In this work, we introduce a neural network with a single recurrent layer designed explicitly for early photometric classification of supernovae. Our algorithm leverages transfer learning to account for model misspecification, host galaxy photometry to solve the data scarcity problem soon after discovery, and a custom weighted loss to prioritize accurate early classification. We first train our algorithm using state-of-the-art transient and host galaxy simulations, then adapt its weights and validate it on the spectroscopically-confirmed SNe Ia, SNe II, and SNe Ib/c from the Zwicky Transient Facility Bright Transient Survey. On observed data, our method achieves an overall accuracy of $82 \pm 2$% within 3 days of an event's discovery, and an accuracy of $87 \pm 5$% within 30 days of discovery. At both early and late phases, our method achieves comparable or superior results to the leading classification algorithms with a simpler network architecture. These results help pave the way for rapid photometric and spectroscopic follow-up of scientifically-valuable transients discovered in massive synoptic surveys.
astro-ph_IM
What could KIDSpec, a new MKID spectrograph, do on the ELT?: Microwave Kinetic Inductance Detectors (MKIDs) are beginning to become more prominent in astronomical instrumentation, due to their sensitivity, low noise, high pixel count for superconducting detectors, and inherent energy and time resolving capability. The Kinetic Inductance Detector Spectrometer (KIDSpec) will take advantage of these features, KIDSpec is a medium resolution MKID spectrograph for the optical/near infrared. KIDSpec will contribute to many science areas particularly those involving short and/or faint observations. When short period binary systems are found, typical CCD detectors will struggle to characterise these systems due to the very short exposures required, causing errors as large as the estimated parameter itself. The KIDSpec Simulator (KSIM) has been developed to investigate how much KIDSpec could improve on this. KIDSpec was simulated on an ELT class telescope to find the extent of its potential, and it was found that KIDSpec could observe a $m_{V}\approx{24}$ with an SNR of 5 for a 10s exposure at 1420 spectral resolution. This would mean that KIDSpec on an ELT class telescope could spectroscopically follow up on any LSST photometric discoveries of LISA verification sources.
astro-ph_IM
A Digital Broadband Beamforming Architecture for 2-PAD: We describe an hierarchical, frequency-domain beamforming architecture for synthesising a sky beam from the wideband antenna feeds of digital aperture arrays. The development of densely-packed, all-digital aperture arrays is an important area of research required for the Square Kilometre Array (SKA) radio telescope. The design of real-time signal processing systems for digital aperture arrays is currently a central challenge in pathfinder projects worldwide. In particular, this work describes a specific implementation of the beamforming architecture to the 2-Polarisation All-Digital (2-PAD) aperture array demonstrator.
astro-ph_IM
Even simpler modeling of quadruply lensed quasars (and random quartets) using Witt's hyperbola: Witt (1996) has shown that for an elliptical potential, the four images of a quadruply lensed quasar lie on a rectangular hyperbola that passes through the unlensed quasar position and the center of the potential as well. Wynne and Schechter (2018) have shown that, for the singular isothermal elliptical potential (SIEP), the four images also lie on an `amplitude' ellipse centered on the quasar position with axes parallel to the hyperbola's asymptotes. Witt's hyperbola arises from equating the directions of both sides of the lens equation. The amplitude ellipse derives from equating the magnitudes. One can model any four points as an SIEP in three steps. 1. Find the rectangular hyperbola that passes through the points. 2. Find the aligned ellipse that also passes through them. 3. Find the hyperbola with asymptotes parallel to those of the first that passes through the center of the ellipse and the pair of images closest to each other. The second hyperbola and the ellipse give an SIEP that predicts the positions of the two remaining images where the curves intersect. Pinning the model to the closest pair guarantees a four image model. Such models permit rapid discrimination between gravitationally lensed quasars and random quartets of stars.
astro-ph_IM
A Cyber Infrastructure for the SKA Telescope Manager: The Square Kilometre Array Telescope Manager (SKA TM) will be responsible for assisting the SKA Operations and Observation Management, carrying out System diagnosis and collecting Monitoring & Control data from the SKA sub-systems and components. To provide adequate compute resources, scalability, operation continuity and high availability, as well as strict Quality of Service, the TM cyber-infrastructure (embodied in the Local Infrastructure - LINFRA) consists of COTS hardware and infrastructural software (for example: server monitoring software, host operating system, virtualization software, device firmware), providing a specially tailored Infrastructure as a Service (IaaS) and Platform as a Service (PaaS) solution. The TM infrastructure provides services in the form of computational power, software defined networking, power, storage abstractions, and high level, state of the art IaaS and PaaS management interfaces. This cyber platform will be tailored to each of the two SKA Phase 1 telescopes (SKA_MID in South Africa and SKA_LOW in Australia) instances, each presenting different computational and storage infrastructures and conditioned by location. This cyber platform will provide a compute model enabling TM to manage the deployment and execution of its multiple components (observation scheduler, proposal submission tools, M&C components, Forensic tools and several Databases, etc). In this sense, the TM LINFRA is primarily focused towards the provision of isolated instances, mostly resorting to virtualization technologies, while defaulting to bare hardware if specifically required due to performance, security, availability, or other requirement.
astro-ph_IM
Mid-band gravitational wave detection with precision atomic sensors: We assess the science reach and technical feasibility of a satellite mission based on precision atomic sensors configured to detect gravitational radiation. Conceptual advances in the past three years indicate that a two-satellite constellation with science payloads consisting of atomic sensors based on laser cooled atomic Sr can achieve scientifically interesting gravitational wave strain sensitivities in a frequency band between the LISA and LIGO detectors, roughly 30 mHz to 10 Hz. The discovery potential of the proposed instrument ranges from from observation of new astrophysical sources (e.g. black hole and neutron star binaries) to searches for cosmological sources of stochastic gravitational radiation and searches for dark matter.
astro-ph_IM
Focus Demo: CANFAR+Skytree: A Cloud Computing and Data Mining System for Astronomy: This is a companion Focus Demonstration article to the CANFAR+Skytree poster (Ball 2012), demonstrating the usage of the Skytree machine learning software on the Canadian Advanced Network for Astronomical Research (CANFAR) cloud computing system. CANFAR+Skytree is the world's first cloud computing system for data mining in astronomy.
astro-ph_IM
Study of hadron and gamma-ray acceptance of the MAGIC telescopes: towards an improved background estimation: The MAGIC telescopes are an array of two imaging atmospheric Cherenkov telescopes (IACTs) studying the gamma ray sky at very high-energies (VHE; E>100 GeV). The observations are performed in stereoscopic mode, with both telescopes pointing at the same position in the sky. The MAGIC field of view (FoV) acceptance for hadrons and gamma rays has a complex shape, which depends on several parameters such as the azimuth and zenith angle of the observations. In the standard MAGIC analysis, the strategy adopted for estimating this acceptance is not optimal in the case of complex FoVs. In this contribution we present the results of systematic studies intended to characterise the acceptance for the entire FoV. These studies open the possibility to apply improved background estimation methods to the MAGIC data, useful to investigate the morphology of extended or multiple sources.
astro-ph_IM
Demonstrating the Concept of Parallax with James Webb Space Telescope: We measured the parallax of the James Webb Space Telescope based on near simultaneous observations using the Lulin One-meter Telescope and the GROWTH India Telescope, separated at a distance of ~4214 km. This serves a great demonstration for the concept of parallax commonly taught in introductory astronomy courses.
astro-ph_IM
High Cadence Optical Transient Searches using Drift Scan Imaging III: Development of an Inexpensive Drive Control System and Characterisation and Correction of Drive System Periodic Errors: In order to further develop and implement novel drift scan imaging experiments to undertake wide field, high time resolution surveys for millisecond optical transients, an appropriate telescope drive system is required. This paper describes the development of a simple and inexpensive hardware and software system to monitor, characterise, and correct the primary category of telescope drive errors, periodic errors due to imperfections in the drive and gear chain. A model for the periodic errors is generated from direct measurements of the telescope drive shaft rotation, verified by comparison to astronomical measurements of the periodic errors. The predictive model is generated and applied in real-time in the form of corrections to the drive rate. A demonstration of the system shows that that inherent periodic errors of peak-to-peak amplitude ~100'' are reduced to below the seeing limit of ~3''. This demonstration allowed an estimate of the uncertainties on the transient sensitivity timescales of the prototype survey of Tingay & Joubert (2021), with the nominal timescale sensitivity of 21 ms revised to be in the range of 20 - 22 ms, which does not significantly affect the results of the experiment. The correction system will be adopted into the final version of high cadence imaging experiment, which is currently under construction. The correction system is inexpensive (<$A100) and composed of readily available hardware, and is readily adaptable to other applications. Design details and codes are therefore made publicly available.
astro-ph_IM
Electric sail control mode for amplified transverse thrust: The electric solar wind sail produces thrust by centrifugally spanned high voltage tethers interacting with the solar wind protons. The sail attitude can be controlled and attitude maneuvers are possible by tether voltage modulation synchronous with the sail rotation. Especially, the sail can be inclined with respect to the solar wind direction to obtain transverse thrust to change the osculating orbit angular momentum. Such an inclination has to be maintained by a continual control voltage modulation. Consequently, the tether voltage available for the thrust is less than the maximum voltage provided by the power system. Using a spherical pendulum as a model for a single rotating tether, we derive analytical estimations for the control efficiency for two separate sail control modes. One is a continuous control modulation that corresponds to strictly planar tether tip motion. The other is an on-off modulation with the tether tip moving along a closed loop on a saddle surface. The novel on-off mode is introduced here to both amplify the transverse thrust and reduce the power consumption. During the rotation cycle, the maximum voltage is applied to the tether only over two thrusting arcs when most of the transverse thrust is produced. In addition to the transverse thrust, we obtain the thrusting angle and electric power consumption for the two control modes. It is concluded that while the thrusting angle is about half of the sail inclination for the continuous modulation it approximately equals to the inclination angle for the on-off modulation. The efficiency of the on-off mode is emphasized when power consumption is considered, and the on-off mode can be used to improve the propulsive acceleration through the reduced power system mass.
astro-ph_IM
Current status of Shanghai VLBI correlator: Shanghai Astronomical Observatory has upgraded its DiFX cluster to 420 CPU cores and a 432-TB storage system at the end of 2014. An international network connection for the raw data transfer has also been established. The routine operations for IVS sessions including CRF, AOV, and APSG series began in early 2015. In addition to the IVS observations, the correlator is dedicated to astrophysical and astrometric programs with the Chinese VLBI Network and international joint VLBI observations. It also worked with the new-built Tianma 65-m radio telescope and successfully found fringes as high as at X/Ka and Q bands in late 2015. A more powerful platform is planned for the high data rate and massive data correlation tasks in the future.
astro-ph_IM
The SFXC software correlator for Very Long Baseline Interferometry: Algorithms and Implementation: In this paper a description is given of the SFXC software correlator, developed and maintained at the Joint Institute for VLBI in Europe (JIVE). The software is designed to run on generic Linux-based computing clusters. The correlation algorithm is explained in detail, as are some of the novel modes that software correlation has enabled, such as wide-field VLBI imaging through the use of multiple phase centres and pulsar gating and binning. This is followed by an overview of the software architecture. Finally, the performance of the correlator as a function of number of CPU cores, telescopes and spectral channels is shown.
astro-ph_IM
An optical test bench for the precision characterization of absolute quantum efficiency for the TESS CCD detectors: The Transiting Exoplanet Survey Satellite (TESS) will search for planets transiting bright stars with Ic<13. TESS has been selected by NASA for launch in 2018 as an Astrophysics Explorer mission, and is expected to discover a thousand or more planets that are smaller in size than Neptune. TESS will employ four wide-field optical charge-coupled device (CCD) cameras with a band-pass of 650 nm-1050 nm to detect temporary drops in brightness of stars due to planetary transits. The 1050 nm limit is set by the quantum efficiency (QE) of the CCDs. The detector assembly consists of four back-illuminated MIT Lincoln Laboratory CCID-80 devices. Each CCID-80 device consists of 2048x2048 imaging array and 2048x2048 frame store regions. Very precise on-ground calibration and characterization of CCD detectors will significantly assist in the analysis of the science data obtained in space. The characterization of the absolute QE of the CCD detectors is a crucial part of the characterization process because QE affects the performance of the CCD significantly over the redder wavelengths at which TESS will be operating. An optical test bench with significantly high photometric stability has been developed to perform precise QE measurements. The design of the test setup along with key hardware, methodology, and results from the test campaign are presented.
astro-ph_IM
Summary of the 3rd BINA Workshop: BINA-3 has been the third workshop of this series involving scientists from India and Belgium aimed at fostering future joint research in the view of cutting-edge observatories and advances in theory. BINA-3 was held at the Graphic Era Hill University, 22-24 March 2023 at Bhimtal (near Nainital), Uttarakhand, India. A major event was the inauguration of the International Liquid-Mirror Telescope (ILMT), the first liquid mirror telescope devoted exclusively to astronomy. BINA-3 provided impressive highlights encompassing topics of both general astrophysics and solar physics. Research results and future projects have been featured through invited and contributed talks, and poster presentations.
astro-ph_IM
Simulation of ultra-high energy photon propagation with PRESHOWER 2.0: In this paper we describe a new release of the PRESHOWER program, a tool for Monte Carlo simulation of propagation of ultra-high energy photons in the magnetic field of the Earth. The PRESHOWER program is designed to calculate magnetic pair production and bremsstrahlung and should be used together with other programs to simulate extensive air showers induced by photons. The main new features of the PRESHOWER code include a much faster algorithm applied in the procedures of simulating the processes of gamma conversion and bremsstrahlung, update of the geomagnetic field model, and a minor correction. The new simulation procedure increases the flexibility of the code so that it can also be applied to other magnetic field configurations such as, for example, encountered in the vicinity of the sun or neutron stars.
astro-ph_IM
Gaia astrometry for stars with too few observations - a Bayesian approach: Gaia's astrometric solution aims to determine at least five parameters for each star, together with appropriate estimates of their uncertainties and correlations. This requires at least five distinct observations per star. In the early data reductions the number of observations may be insufficient for a five-parameter solution, and even after the full mission many stars will remain under-observed, including faint stars at the detection limit and transient objects. In such cases it is reasonable to determine only the two position parameters. Their formal uncertainties would however grossly underestimate the actual errors, due to the neglected parallax and proper motion. We aim to develop a recipe to calculate sensible formal uncertainties that can be used in all cases of under-observed stars. Prior information about the typical ranges of stellar parallaxes and proper motions is incorporated in the astrometric solution by means of Bayes' rule. Numerical simulations based on the Gaia Universe Model Snapshot (GUMS) are used to investigate how the prior influences the actual errors and formal uncertainties when different amounts of Gaia observations are available. We develop a criterion for the optimum choice of priors, apply it to a wide range of cases, and derive a global approximation of the optimum prior as a function of magnitude and galactic coordinates. The feasibility of the Bayesian approach is demonstrated through global astrometric solutions of simulated Gaia observations. With an appropriate prior it is possible to derive sensible positions with realistic error estimates for any number of available observations. Even though this recipe works also for well-observed stars it should not be used where a good five-parameter astrometric solution can be obtained without a prior. Parallaxes and proper motions from a solution using priors are always biased and should not be used.
astro-ph_IM
Visible astro-comb filtered by a passively-stabilized Fabry-Perot cavity: We demonstrate a compact 29.3 GHz visible astro-comb covering the spectrum from 560nm to 700nm. A 837 MHz Yb:fiber laser frequency comb phase locked to a Rb clock served as the seed comb to ensure the frequency stability and high side mode suppression ratio. After the visible super-continuum generation, a cavity-length-fixed Fabry-Perot cavity made by ultra-low expansion glass was utilized to filter the comb teeth for eliminating the rapid active dithering. The mirrors were home-made complementary chirped mirrors pair with zero dispersion and high reflection to guarantee no mode skipping. These filtered comb teeth were clearly resolved in an astronomical spectrograph of 49,000 resolution, exhibiting sharp linetype, zero noise floor, and uniform exposure amplitude.
astro-ph_IM
Characterization Of Inpaint Residuals In Interferometric Measurements of the Epoch Of Reionization: Radio Frequency Interference (RFI) is one of the systematic challenges preventing 21cm interferometric instruments from detecting the Epoch of Reionization. To mitigate the effects of RFI on data analysis pipelines, numerous inpaint techniques have been developed to restore RFI corrupted data. We examine the qualitative and quantitative errors introduced into the visibilities and power spectrum due to inpainting. We perform our analysis on simulated data as well as real data from the Hydrogen Epoch of Reionization Array (HERA) Phase 1 upper limits. We also introduce a convolutional neural network that capable of inpainting RFI corrupted data in interferometric instruments. We train our network on simulated data and show that our network is capable at inpainting real data without requiring to be retrained. We find that techniques that incorporate high wavenumbers in delay space in their modeling are best suited for inpainting over narrowband RFI. We also show that with our fiducial parameters Discrete Prolate Spheroidal Sequences (DPSS) and CLEAN provide the best performance for intermittent ``narrowband'' RFI while Gaussian Progress Regression (GPR) and Least Squares Spectral Analysis (LSSA) provide the best performance for larger RFI gaps. However we caution that these qualitative conclusions are sensitive to the chosen hyperparameters of each inpainting technique. We find these results to be consistent in both simulated and real visibilities. We show that all inpainting techniques reliably reproduce foreground dominated modes in the power spectrum. Since the inpainting techniques should not be capable of reproducing noise realizations, we find that the largest errors occur in the noise dominated delay modes. We show that in the future, as the noise level of the data comes down, CLEAN and DPSS are most capable of reproducing the fine frequency structure in the visibilities of HERA data.
astro-ph_IM
Simultaneous analysis of large INTEGRAL/SPI datasets: optimizing the computation of the solution and its variance using sparse matrix algorithms: Nowadays, analyzing and reducing the ever larger astronomical datasets is becoming a crucial challenge, especially for long cumulated observation times. The INTEGRAL/SPI X-gamma-ray spectrometer is an instrument for which it is essential to process many exposures at the same time in order to increase the low signal-to-noise ratio of the weakest sources. In this context, the conventional methods for data reduction are inefficient and sometimes not feasible at all. Processing several years of data simultaneously requires computing not only the solution of a large system of equations, but also the associated uncertainties. We aim at reducing the computation time and the memory usage. Since the SPI transfer function is sparse, we have used some popular methods for the solution of large sparse linear systems; we briefly review these methods. We use the Multifrontal Massively Parallel Solver (MUMPS) to compute the solution of the system of equations. We also need to compute the variance of the solution, which amounts to computing selected entries of the inverse of the sparse matrix corresponding to our linear system. This can be achieved through one of the latest features of the MUMPS software that has been partly motivated by this work. In this paper we provide a brief presentation of this feature and evaluate its effectiveness on astrophysical problems requiring the processing of large datasets simultaneously, such as the study of the entire emission of the Galaxy. We used these algorithms to solve the large sparse systems arising from SPI data processing and to obtain both their solutions and the associated variances. In conclusion, thanks to these newly developed tools, processing large datasets arising from SPI is now feasible with both a reasonable execution time and a low memory usage.
astro-ph_IM
Why should we keep measuring zenital dependence of muon flux? Results obtained at Campinas (SP) BR: The zenital dependence of muon flux which reaches the earth's surface is well known as proportional to cos^n(\theta). Generally, for practical purposes and simplicity in calculations, n is taken as 2. However, compilations of measurements show dependence on the geographical location of the experiments as well as the muons energy range. Since analytical solutions appear to be increasingly less necessary because of the higher accessibility to low cost computational power, accurate and precise determination of the value of the exponent n, under different conditions, can be useful in the necessary calculations to estimate signals and backgrounds, either for terrestrial and underground experiments. In this work we discuss a method for measuring n using a simple muon telescope and the results obtained for measurements taken at Campinas (SP), Brazil. After validation of the method, we intend to extend the measurements for different geographic locations due to the simplicity of the method, and thus collect more values of n that currently exist in compilations of general data on cosmic rays.
astro-ph_IM
Optical calibration of large format adaptive mirrors: Adaptive (or deformable) mirrors are widely used as wavefront correctors in adaptive optics systems. The optical calibration of an adaptive mirror is a fundamental step during its life-cycle: the process is in facts required to compute a set of known commands to operate the adaptive optics system, to compensate alignment and non common-path aberrations, to run chopped or field-stabilized acquisitions. In this work we present the sequence of operations for the optical calibration of adaptive mirrors, with a specific focus on large aperture systems such as the adaptive secondaries. Such systems will be one of the core components of the extremely large telescopes. Beyond presenting the optical procedures, we discuss in detail the actors, their functional requirements and the mutual interactions. A specific emphasys is put on automation, through a clear identification of inputs, outputs and quality indicators for each step: due to a high degrees-of-freedom count (thousands of actuators), an automated approach is preferable to constraint the cost and schedule. In the end we present some algorithms for the evaluation of the measurement noise; this point is particularly important since the calibration setup is typically a large facility in an industrial environment, where the noise level may be a major show-stopper.
astro-ph_IM
Numerical Strategies of Computing the Luminosity Distance: We propose two efficient numerical methods of evaluating the luminosity distance in the spatially flat {\Lambda}CDM universe. The first method is based on the Carlson symmetric form of elliptic integrals, which is highly accurate and can replace numerical quadratures. The second method, using a modified version of Hermite interpolation, is less accurate but involves only basic numerical operations and can be easily implemented. We compare our methods with other numerical approximation schemes and explore their respective features and limitations. Possible extensions of these methods to other cosmological models are also discussed.
astro-ph_IM
Investigation of Residual Blaze Functions in Slit-Based Echelle Spectrograph: We have studied the Residual Blaze Functions (RBF) resulting from division of individual echelle orders by extracted flat-field in spectra obtained by slit-fed OES spectrograph of 2m telescope of Ond\v{r}ejov observatory, Czech Republic. We have eliminated the dependence on target and observation conditions by semiautomatic fitting of global response function, thus getting the instrument-only dependent part, which may be easily incorporated into data reduction pipeline. The improvement of reliability of estimation of continuum on spectra of targets with wide and shallow lines is noticeable and the merging of all orders into the one long spectrum gives much more reliable results.
astro-ph_IM
FACT - The First G-APD Cherenkov Telescope: Status and Results: The First G-APD Cherenkov telescope (FACT) is the first telescope using silicon photon detectors (G-APD aka. SiPM). It is built on the mount of the HEGRA CT3 telescope, still located at the Observatorio del Roque de los Muchachos, and it is successfully in operation since Oct. 2011. The use of Silicon devices promises a higher photon detection efficiency, more robustness and higher precision than photo-multiplier tubes. The FACT collaboration is investigating with which precision these devices can be operated on the long-term. Currently, the telescope is successfully operated from remote and robotic operation is under development. During the past months of operation, the foreseen monitoring program of the brightest known TeV blazars has been carried out, and first physics results have been obtained including a strong flare of Mrk501. An instantaneous flare alert system is already in a testing phase. This presentation will give an overview of the project and summarize its goals, status and first results.
astro-ph_IM
Dishing up the Data: A Decade of Space Missions: The past decade has seen Parkes once again involved in a wide range of space tracking activities that have added to its illustrious legacy. This contribution is a personal recollection of those tracking efforts - both real and celluloid. We begin in a light-hearted vein with some behind-the-scenes views of the popular film, "The DISH", and then turn to more serious contributions; discussing the vital role of the telescope in alleviating the great "traffic jam" at Mars in 2003/04 and salvaging the Doppler Wind Experiment as the Huygens probe descended though the atmosphere of Saturn's largest moon, Titan, in mid-decade. We cap off the decade with a discussion of the search for the missing Apollo 11 slow-scan TV tapes.
astro-ph_IM
High Contrast and High Angular Imaging at Subaru Telescope: Adaptive Optics projects at Subaru Telescope span a wide field of capabilities ranging from ground-layer adaptive optics (GLAO) providing partial correction over a 20 arcmin FOV to extreme adaptive optics (ExAO) for exoplanet imaging. We describe in this paper current and upcoming narrow field-of-view capabilities provided by the Subaru Extreme Adaptive Optics Adaptive Optics (SCExAO) system and its instrument modules, as well as the upcoming 3000-actuator upgrade of the Nasmyth AO system.
astro-ph_IM
POLARIX: a pathfinder mission of X-ray polarimetry: Since the birth of X-ray astronomy, spectral, spatial and timing observation improved dramatically, procuring a wealth of information on the majority of the classes of the celestial sources. Polarimetry, instead, remained basically unprobed. X-ray polarimetry promises to provide additional information procuring two new observable quantities, the degree and the angle of polarization. POLARIX is a mission dedicated to X-ray polarimetry. It exploits the polarimetric response of a Gas Pixel Detector, combined with position sensitivity, that, at the focus of a telescope, results in a huge increase of sensitivity. Three Gas Pixel Detectors are coupled with three X-ray optics which are the heritage of JET-X mission. POLARIX will measure time resolved X-ray polarization with an angular resolution of about 20 arcsec in a field of view of 15 arcmin $\times$ 15 arcmin and with an energy resolution of 20 % at 6 keV. The Minimum Detectable Polarization is 12 % for a source having a flux of 1 mCrab and 10^5 s of observing time. The satellite will be placed in an equatorial orbit of 505 km of altitude by a Vega launcher.The telemetry down-link station will be Malindi. The pointing of POLARIX satellite will be gyroless and it will perform a double pointing during the earth occultation of one source, so maximizing the scientific return. POLARIX data are for 75 % open to the community while 25 % + SVP (Science Verification Phase, 1 month of operation) is dedicated to a core program activity open to the contribution of associated scientists. The planned duration of the mission is one year plus three months of commissioning and SVP, suitable to perform most of the basic science within the reach of this instrument.
astro-ph_IM
S-ACF: A selective estimator for the autocorrelation function of irregularly sampled time series: We present a generalised estimator for the autocorrelation function, S-ACF, which is an extended version of the standard estimator of the autocorrelation function (ACF). S-ACF is a versatile definition that can robustly and efficiently extract periodicity and signal shape information from a time series, independent of the time sampling and with minimal assumptions about the underlying process. Calculating the autocorrelation of irregularly sampled time series becomes possible by generalising the lag of the standard estimator of the ACF to a real parameter and introducing the notion of selection and weight functions. We show that the S-ACF reduces to the standard ACF estimator for regularly sampled time series. Using a large number of synthetic time series we demonstrate that the performance of the S-ACF is as good or better than commonly used Gaussian and rectangular kernel estimators, and is comparable to a combination of interpolation and the standard estimator. We apply the S-ACF to astrophysical data by extracting rotation periods for the spotted star KIC 5110407, and compare our results to Gaussian process (GP) regression and Lomb-Scargle (LS) periodograms. We find that the S-ACF periods typically agree better with those from GP regression than from LS periodograms, especially in cases where there is evolution in the signal shape. The S-ACF has a wide range of potential applications and should be useful in quantitative science disciplines where irregularly sampled time series occur. A Python implementation of the S-ACF is available under the MIT license.
astro-ph_IM
Optimising LSST Observing Strategy for Weak Lensing Systematics: The LSST survey will provide unprecedented statistical power for measurements of dark energy. Consequently, controlling systematic uncertainties is becoming more important than ever. The LSST observing strategy will affect the statistical uncertainty and systematics control for many science cases; here, we focus on weak lensing systematics. The fact that the LSST observing strategy involves hundreds of visits to the same sky area provides new opportunities for systematics mitigation. We explore these opportunities by testing how different dithering strategies (pointing offsets and rotational angle of the camera in different exposures) affect additive weak lensing shear systematics on a baseline operational simulation, using the $\rho-$statistics formalism. Some dithering strategies improve systematics control at the end of the survey by a factor of up to $\sim 3-4$ better than others. We find that a random translational dithering strategy, applied with random rotational dithering at every filter change, is the most effective of those strategies tested in this work at averaging down systematics. Adopting this dithering algorithm, we explore the effect of varying the area of the survey footprint, exposure time, number of exposures in a visit, and exposure to the Galactic plane. We find that any change that increases the average number of exposures (in filters relevant to weak lensing) reduces the additive shear systematics. Some ways to achieve this increase may not be favorable for the weak lensing statistical constraining power or for other probes, and we explore the relative trade-offs between these options given constraints on the overall survey parameters.
astro-ph_IM
Astronomy and the new SI: In 2019 the International System of units (SI) conceptually re-invented itself. This was necessary because quantum-electronic devices had become so precise that the old SI could no longer calibrate them. The new system defines values of fundamental constants (including $c,h,k,e$ but not $G$) and allows units to be realized from the defined constants through any applicable equation of physics. In this new and more abstract SI, units can take on new guises --- for example, the kilogram is at present best implemented as a derived electrical unit. Relevant to astronomy, however, is that several formerly non-SI units, such as electron-volts, light-seconds, and what we may call "gravity seconds" $GM/c^3$, can now be interpreted not as themselves units, but as shorthand for volts and seconds being used with particular equations of physics. Moreover, the classical astronomical units have exact and rather convenient equivalents in the new SI: zero AB magnitude amounts to $\simeq5\times10^{10}$ photons $\rm m^{-2}\,s^{-1}$ per logarithmic frequency or wavelength interval, $\rm 1\,au\simeq 500$ light-seconds, $\rm 1\,pc\simeq 10^8$ light-seconds, while a solar mass $\simeq5$ gravity-seconds. As a result, the unit conversions ubiquitous in astrophysics can now be eliminated, without introducing other problems, as the old-style SI would have done. We review a variety of astrophysical processes illustrating the simplifications possible with the new-style SI, with special attention to gravitational dynamics, where care is needed to avoid propagating the uncertainty in $G$. Well-known systems (GPS satellites, GW170817, and the M87 black hole) are used as examples wherever possible.
astro-ph_IM
Astro2020 Science White Paper: Science Platforms for Resolved Stellar Populations in the Next Decade: Over the past decade, research in resolved stellar populations has made great strides in exploring the nature of dark matter, in unraveling the star formation, chemical enrichment, and dynamical histories of the Milky Way and nearby galaxies, and in probing fundamental physics from general relativity to the structure of stars. Large surveys have been particularly important to the biggest of these discoveries. In the coming decade, current and planned surveys will push these research areas still further through a large variety of discovery spaces, giving us unprecedented views into the low surface brightness Universe, the high surface brightness Universe, the 3D motions of stars, the time domain, and the chemical abundances of stellar populations. These discovery spaces will be opened by a diverse range of facilities, including the continuing Gaia mission, imaging machines like LSST and WFIRST, massively multiplexed spectroscopic platforms like DESI, Subaru-PFS, and MSE, and telescopes with high sensitivity and spatial resolution like JWST, the ELTs, and LUVOIR. We do not know which of these facilities will prove most critical for resolved stellar populations research in the next decade. We can predict, however, that their chance of success will be maximized by granting use of the data to broad communities, that many scientific discoveries will draw on a combination of data from them, and that advances in computing will enable increasingly sophisticated analyses of the large and complex datasets that they will produce. We recommend that Astro2020 1) acknowledge the critical role that data archives will play for stellar populations and other science in the next decade, 2) recognize the opportunity that advances in computing will bring for survey data analysis, and 3) consider investments in Science Platform technology to bring these opportunities to fruition.
astro-ph_IM
The Zwicky Transient Facility: The Zwicky Transient Facility (ZTF) is a next-generation optical synoptic survey that builds on the experience and infrastructure of the Palomar Transient Factory (PTF). Using a new 47 deg$^2$ survey camera, ZTF will survey more than an order of magnitude faster than PTF to discover rare transients and variables. I describe the survey and the camera design. Searches for young supernovae, fast transients, counterparts to gravitational-wave detections, and rare variables will benefit from ZTF's high cadence, wide area survey.
astro-ph_IM
Reconstructing inclined extensive air showers from radio measurements: We present a reconstruction algorithm for extensive air showers with zenith angles between 65$^\circ$ and 85$^\circ$ measured with radio antennas in the 30-80 MHz band. Our algorithm is based on a signal model derived from CoREAS simulations which explicitly takes into account the asymmetries introduced by the superposition of charge-excess and geomagnetic radiation as well as by early-late effects. We exploit correlations among fit parameters to reduce the dimensionality and thus ensure stability of the fit procedure. Our approach reaches a reconstruction efficiency near 100% with an intrinsic resolution for the reconstruction of the electromagnetic energy of well below 5\%. It can be employed in upcoming large-scale radio detection arrays using the 30-80 MHz band, in particular the AugerPrime Radio detector of the Pierre Auger Observatory, and can likely be adapted to experiments such as GRAND operating at higher frequencies.
astro-ph_IM
PolarLight: a CubeSat X-ray Polarimeter based on the Gas Pixel Detector: The gas pixel detector (GPD) is designed and developed for high-sensitivity astronomical X-ray polarimetry, which is a new window about to open in a few years. Due to the small mass, low power, and compact geometry of the GPD, we propose a CubeSat mission Polarimeter Light (PolarLight) to demonstrate and test the technology directly in space. There is no optics but a collimator to constrain the field of view to 2.3 degrees. Filled with pure dimethyl ether (DME) at 0.8 atm and sealed by a beryllium window of 100 micron thick, with a sensitive area of about 1.4 mm by 1.4 mm, PolarLight allows us to observe the brightest X-ray sources on the sky, with a count rate of, e.g., ~0.2 counts/s from the Crab nebula. The PolarLight is 1U in size and mounted in a 6U CubeSat, which was launched into a low Earth Sun-synchronous orbit on October 29, 2018, and is currently under test. More launches with improved designs are planned in 2019. These tests will help increase the technology readiness for future missions such as the enhanced X-ray Timing and Polarimetry (eXTP), better understand the orbital background, and may help constrain the physics with observations of the brightest objects.
astro-ph_IM
The Jiao Tong University Spectroscopic Telescope Project: The Jiao Tong University Spectroscopic Telescope (JUST) is a 4.4-meter f/6.0 segmentedmirror telescope dedicated to spectroscopic observations. The JUST primary mirror is composed of 18 hexagonal segments, each with a diameter of 1.1 m. JUST provides two Nasmyth platforms for placing science instruments. One Nasmyth focus fits a field of view of 10 arcmin and the other has an extended field of view of 1.2 deg with correction optics. A tertiary mirror is used to switch between the two Nasmyth foci. JUST will be installed at a site at Lenghu in Qinghai Province, China, and will conduct spectroscopic observations with three types of instruments to explore the dark universe, trace the dynamic universe, and search for exoplanets: (1) a multi-fiber (2000 fibers) medium-resolution spectrometer (R=4000-5000) to spectroscopically map galaxies and large-scale structure; (2) an integral field unit (IFU) array of 500 optical fibers and/or a long-slit spectrograph dedicated to fast follow-ups of transient sources for multimessenger astronomy; (3) a high-resolution spectrometer (R~100000) designed to identify Jupiter analogs and Earth-like planets, with the capability to characterize the atmospheres of hot exoplanets.
astro-ph_IM
Towards a data-driven model of the sky from low Earth orbit as observed by the Hubble Space Telescope: The sky observed by space telescopes in Low Earth Orbit (LEO) can be dominated by stray light from multiple sources including the Earth, Sun and Moon. This stray light presents a significant challenge to missions that aim to make a secure measurement of the Extragalactic Background Light (EBL). In this work we quantify the impact of stray light on sky observations made by the Hubble Space Telescope (HST) Advanced Camera for Surveys. By selecting on orbital parameters we successfully isolate images with sky that contain minimal and high levels of Earthshine. In addition, we find weather observations from CERES satellites correlates with the observed HST sky surface brightness indicating the value of incorporating such data to characterise the sky. Finally we present a machine learning model of the sky trained on the data used in this work to predict the total observed sky surface brightness. We demonstrate that our initial model is able to predict the total sky brightness under a range of conditions to within 3.9% of the true measured sky. Moreover, we find that the model matches the stray light-free observations better than current physical Zodiacal light models.
astro-ph_IM
Lattice Boltzmann Method for Electromagnetic Wave Propagation: We present a new Lattice Boltzmann (LB) formulation to solve the Maxwell equations for electromagnetic (EM) waves propagating in a heterogeneous medium. By using a pseudo-vector discrete Boltzmann distribution, the scheme is shown to reproduce the continuum Maxwell equations. The technique compares well with a pseudo-spectral method at solving for two-dimensional wave propagation in a heterogeneous medium, which by design contains substantial contrasts in the refractive index. The extension to three dimensions follows naturally and, owing to the recognized efficiency of LB schemes for parallel computation in irregular geometries, it gives a powerful method to numerically simulate a wide range of problems involving EM wave propagation in complex media.
astro-ph_IM
Characterizing Variable Stars in a Single Night with LSST: Stars exhibit a bewildering variety of variable behaviors ranging from explosive magnetic flares to stochastically changing accretion to periodic pulsations or rotations. The principal LSST surveys will have cadences too sparse and irregular to capture most of these phenomena. A novel idea is proposed here to observe a single Galactic field, rich in unobscured stars, in a continuous sequence of $\sim 15$ second exposures for one long winter night in a single photometric band. The result will be a unique dataset of $\sim 1$ million regularly spaced stellar lightcurves. The lightcurves will gives a particularly comprehensive collection of dM star variability. A powerful array of statistical procedures can be applied to the ensemble of lightcurves from the long-standing fields of time series analysis, signal processing and econometrics. Dozens of `features' describing the variability can be extracted and subject to machine learning classification, giving a unique authoritative objective classification of rapidly variable stars. The most effective features can then inform the wider LSST community on the best approaches to variable star identification and classification from the sparse, irregular cadences that dominate the LSST project.
astro-ph_IM
Tails: Chasing Comets with the Zwicky Transient Facility and Deep Learning: We present Tails, an open-source deep-learning framework for the identification and localization of comets in the image data of the Zwicky Transient Facility (ZTF), a robotic optical time-domain survey currently in operation at the Palomar Observatory in California, USA. Tails employs a custom EfficientDet-based architecture and is capable of finding comets in single images in near real time, rather than requiring multiple epochs as with traditional methods. The system achieves state-of-the-art performance with 99% recall, 0.01% false positive rate, and 1-2 pixel root mean square error in the predicted position. We report the initial results of the Tails efficiency evaluation in a production setting on the data of the ZTF Twilight survey, including the first AI-assisted discovery of a comet (C/2020 T2) and the recovery of a comet (P/2016 J3 = P/2021 A3).
astro-ph_IM
Extreme-value modelling for the significance assessment of periodogram peaks: I propose a new procedure to estimate the False Alarm Probability, the measure of significance for peaks of periodograms. The key element of the new procedure is the use of generalized extreme-value distributions, the limiting distribution for maxima of variables from most continuous distributions. This technique allows reliable extrapolation to the very high probability levels required by multiple hypothesis testing, and enables the derivation of confidence intervals of the estimated levels. The estimates are stable against deviations from distributional assumptions, which are otherwise usually made either about the observations themselves or about the theoretical univariate distribution of the periodogram. The quality and the performance of the procedure is demonstrated on simulations and on two multimode variable stars from Sloan Digital Sky Survey Stripe 82.
astro-ph_IM
HCF (HREXI Calibration Facility): Mapping out sub-pixel level responses from high resolution Cadmium Zinc Telluride (CZT) imaging X-ray detectors: The High Resolution Energetic X-Ray Imager (HREXI) CZT detector development program at Harvard is aimed at developing tiled arrays of finely pixelated CZT detectors for use in wide-field coded aperture 3-200 keV X-ray telescopes. A pixel size of $\simeq$ 600 $\mu m$ has already been achieved in the ProtoEXIST2 (P2) detector plane with CZT read out by the NuSTAR ASIC. This paves the way for even smaller 300 $\mu m$ pixels in the next generation HREXI detectors. This article describes a new HREXI calibration facility (HCF) which enables a high resolution sub-pixel level (100 $\mu m$) 2D scan of a 256 $cm^2$ tiled array of 2 $\times$ 2 cm CZT detectors illuminated by a bright X-ray AmpTek Mini-X tube source at timescales of around a day. HCF is a significant improvement from the previous apparatus used for scanning these detectors which took $\simeq$ 3 weeks to complete a 1D scan of a similar detector plane. Moreover, HCF has the capability to scan a large tiled array of CZT detectors ($32cm \times 32cm$) at 100 $\mu m$ resolution in the 10 - 50 keV energy range which was not possible previously. This paper describes the design, construction, and implementation of HCF for the calibration of the P2 detector plane.
astro-ph_IM
Disentangled Representation Learning for Astronomical Chemical Tagging: Modern astronomical surveys are observing spectral data for millions of stars. These spectra contain chemical information that can be used to trace the Galaxy's formation and chemical enrichment history. However, extracting the information from spectra, and making precise and accurate chemical abundance measurements are challenging. Here, we present a data-driven method for isolating the chemical factors of variation in stellar spectra from those of other parameters (i.e. \teff, \logg, \feh). This enables us to build a spectral projection for each star with these parameters removed. We do this with no ab initio knowledge of elemental abundances themselves, and hence bypass the uncertainties and systematics associated with modeling that rely on synthetic stellar spectra. To remove known non-chemical factors of variation, we develop and implement a neural network architecture that learns a disentangled spectral representation. We simulate our recovery of chemically identical stars using the disentangled spectra in a synthetic APOGEE-like dataset. We show that this recovery declines as a function of the signal to noise ratio, but that our neural network architecture outperforms simpler modeling choices. Our work demonstrates the feasibility of data-driven abundance-free chemical tagging.
astro-ph_IM
Design and modeling of a moderate-resolution astronomic spectrograph with volume-phase holographic gratings: We present an optical design of astronomic spectrograph based on a cascade of volume-phase holographic gratings. The cascade consists of three gratings. Each of them provides moderately high spectral resolution in a narrow range of 83 nm. Thus the spectrum image represents three lines covering region 430-680 nm. Two versions of the scheme are described: a full-scale one with estimated resolving power of 5300-7900 and a small-sized one intended for creation of a lab prototype, which provides the resolving power of 1500-3000. Diffraction efficiency modeling confirms that the system throughput can reach 75 %, while stray light caused by the gratings crosstalk is negligible. We also propose a design of image slicer and focal reducer allowing to couple the instrument with the 6-m telescope. Finally, we present concept of the opto-mechanical design.
astro-ph_IM
Using Virtual Observatory with Python: querying remote astronomical databases: This tutorial is devoted to extending an existing catalogue with data taken elsewhere, either from CDS Vizier or Simbad database. As an example, we used the so-called 'Spectroscopic Survey of Stars in the Solar Neighborhood' (aka. S4N, Allende Prieto et al. 2004) in order to retrieve all objects with available data for the set of fundamental stellar parameters effective temperature, surface gravity and metallicity. Then for each object in this dataset we query Simbad database to retrieve the projected rotational velocity. This combines Vizier and Simbad queries made using Python astroquery module. The tutorial covers remote database access, filtering tables with arbitrary criteria, creating and writing your own tables, and basics of plotting in Python.
astro-ph_IM
Synthesizing carbon nanotubes in space: Context. As the 4th most abundant element in the universe, carbon (C) is widespread in the interstellar medium (ISM) in various allotropic forms (e.g., fullerenes have been identified unambiguously in many astronomical environments, the presence of polycyclic aromatic hydrocarbon molecules in space has been commonly admitted, and presolar graphite as well as nanodiamonds have been identified in meteorites). As stable allotropes of these species, whether carbon nanotubes (CNTs) and their hydrogenated counterparts are also present in the ISM or not is unknown. Aims. We explore the possible routes for the formation of CNTs in the ISM and calculate their fingerprint vibrational spectral features in the infrared (IR). Methods. We study the hydrogen-abstraction/acetylene-addition (HACA) mechanism and investigate the synthesis of nanotubes using density functional theory (DFT). The IR vibrational spectra of CNTs and hydrogenated nanotubes (HNTs), as well as their cations, have also been obtained with DFT. Results. We find that CNTs could be synthesized in space through a feasible formation pathway. CNTs and cationic CNTs, as well as their hydrogenated counterparts, exhibit intense vibrational transitions in the IR. Their possible presence in the ISM could be investigated by comparing the calculated vibrational spectra with astronomical observations made by the Infrared Space Observatory, Spitzer Space Telescope, and particularly the upcoming James Webb Space Telescope.
astro-ph_IM
Statistical framework for estimating GNSS bias: We present a statistical framework for estimating global navigation satellite system (GNSS) non-ionospheric differential time delay bias. The biases are estimated by examining differences of measured line integrated electron densities (TEC) that are scaled to equivalent vertical integrated densities. The spatio-temporal variability, instrumentation dependent errors, and errors due to inaccurate ionospheric altitude profile assumptions are modeled as structure functions. These structure functions determine how the TEC differences are weighted in the linear least-squares minimization procedure, which is used to produce the bias estimates. A method for automatic detection and removal of outlier measurements that do not fit into a model of receiver bias is also described. The same statistical framework can be used for a single receiver station, but it also scales to a large global network of receivers. In addition to the Global Positioning System (GPS), the method is also applicable to other dual frequency GNSS systems, such as GLONASS (Globalnaya Navigazionnaya Sputnikovaya Sistema). The use of the framework is demonstrated in practice through several examples. A specific implementation of the methods presented here are used to compute GPS receiver biases for measurements in the MIT Haystack Madrigal distributed database system. Results of the new algorithm are compared with the current MIT Haystack Observatory MAPGPS bias determination algorithm. The new method is found to produce estimates of receiver bias that have reduced day-to-day variability and more consistent coincident vertical TEC values.
astro-ph_IM
Atmospheric transparency in the optical and near IR range above the Shatdzhatmaz summit: The study of atmospheric extinction based on the MASS data has been carried out using the classical photometric pairs method. The extinction in V band can be estimated at 0.m 19. The water vapour content has been derived from GPS measurements. The median value of PWV for clear nights is equal to 7.7 mm.
astro-ph_IM
Non-LTE radiation hydrodynamics in PLUTO: Modeling the dynamics of most astrophysical structures requires an adequate description of the radiation-matter interaction. Several numerical (magneto)hydrodynamics codes were upgraded with a radiation module to fulfill this request. However, those among them that use either the flux-limited diffusion (FLD) or the M1 radiation moment approaches are restricted to the local thermodynamic equilibrium (LTE). This assumption may be not valid in some astrophysical cases. We present an upgraded version of the LTE radiation-hydrodynamics module implemented in the PLUTO code, originally developed by Kolb et al. (2013), which we have extended to handle non-LTE regimes. Starting from the general frequency-integrated comoving-frame equations of radiation hydrodynamics (RHD), we have justified all the assumptions made to obtain the non-LTE equations actually implemented in the module, under the FLD approximation. An operator-split method is employed, with two substeps: the hydrodynamic part is solved with an explicit method by the solvers already available in PLUTO, the non-LTE radiation diffusion and energy exchange part is solved with an implicit method. The module is implemented in the PLUTO environment. It uses databases of radiative quantities that can be provided independently by the user: the radiative power loss, the Planck and Rosseland mean opacities. Our implementation has been validated through different tests, in particular radiative shock tests. The agreement with the semi-analytical solutions (when available) is good, with a maximum error of 7%. Moreover, we have proved that non-LTE approach is of paramount importance to properly model accretion shock structures. Our radiation FLD module represents a step toward the general non-LTE RHD modeling. The module is available, under request, for the community.
astro-ph_IM
Timing Calibration of the NuSTAR X-ray Telescope: The Nuclear Spectroscopic Telescope Array (NuSTAR) mission is the first focusing X-ray telescope in the hard X-ray (3-79 keV) band. Among the phenomena that can be studied in this energy band, some require high time resolution and stability: rotation-powered and accreting millisecond pulsars, fast variability from black holes and neutron stars, X-ray bursts, and more. Moreover, a good alignment of the timestamps of X-ray photons to UTC is key for multi-instrument studies of fast astrophysical processes. In this Paper, we describe the timing calibration of the NuSTAR mission. In particular, we present a method to correct the temperature-dependent frequency response of the on-board temperature-compensated crystal oscillator. Together with measurements of the spacecraft clock offsets obtained during downlinks passes, this allows a precise characterization of the behavior of the oscillator. The calibrated NuSTAR event timestamps for a typical observation are shown to be accurate to a precision of ~65 microsec.
astro-ph_IM
Solar diameter with 2012 Venus transit: The role of Venus and Mercury transits is crucial to know the past history of the solar diameter. Through the W parameter, the logarithmic derivative of the radius with respect to the luminosity, the past values of the solar luminosity can be recovered. The black drop phenomenon affects the evaluation of the instants of internal and external contacts between the planetary disk and the solar limb. With these observed instants compared with the ephemerides the value of the solar diameter is recovered. The black drop and seeing effects are overcome with two fitting circles, to Venus and to the Sun, drawn in the undistorted part of the image. The corrections of ephemerides due to the atmospheric refraction will also be taken into account. The forthcoming transit of Venus will allow an accuracy on the diameter of the Sun better than 0.01 arcsec, with good images of the ingress and of the egress taken each second. Chinese solar observatories are in the optimal conditions to obtain valuable data for the measurement of the solar diameter with the Venus transit of 5/6 June 2012 with an unprecedented accuracy, and with absolute calibration given by the ephemerides.
astro-ph_IM
Simons Observatory Large Aperture Telescope Receiver Design Overview: The Simons Observatory (SO) will make precision temperature and polarization measurements of the cosmic microwave background (CMB) using a series of telescopes which will cover angular scales between one arcminute and tens of degrees and sample frequencies between 27 and 270 GHz. Here we present the current design of the large aperture telescope receiver (LATR), a 2.4 m diameter cryostat that will be mounted on the SO 6 m telescope and will be the largest CMB receiver to date. The cryostat size was chosen to take advantage of the large focal plane area having high Strehl ratios, which is inherent to the Cross-Dragone telescope design. The LATR will be able to accommodate thirteen optics tubes, each having a 36 cm diameter aperture and illuminating several thousand transition-edge sensor (TES) bolometers. This set of equipment will provide an opportunity to make measurements with unparalleled sensitivity. However, the size and complexity of the LATR also pose numerous technical challenges. In the following paper, we present the design of the LATR and include how we address these challenges. The solutions we develop in the process of designing the LATR will be informative for the general CMB community, and for future CMB experiments like CMB-S4.
astro-ph_IM
Digital Signal Processing in Cosmology: We address the problem of discretizing continuous cosmological signals such as a galaxy distribution for further processing with Fast Fourier techniques. Discretizing, in particular representing continuous signals by discrete sets of sample points, introduces an enormous loss of information, which has to be understood in detail if one wants to make inference from the discretely sampled signal towards actual natural physical quantities. We therefore review the mathematics of discretizing signals and the application of Fast Fourier Transforms to demonstrate how the interpretation of the processed data can be affected by these procedures. It is also a well known fact that any practical sampling method introduces sampling artifacts and false information in the form of aliasing. These sampling artifacts, especially aliasing, make further processing of the sampled signal difficult. For this reason we introduce a fast and efficient supersampling method, frequently applied in 3D computer graphics, to cosmological applications such as matter power spectrum estimation. This method consists of two filtering steps which allow for a much better approximation of the ideal sampling procedure, while at the same time being computationally very efficient.Thus, it provides discretely sampled signals which are greately cleaned from aliasing contributions.
astro-ph_IM
A Frequency Selective Surface based focal plane receiver for the OLIMPO balloon-borne telescope: We describe here a focal plane array of Cold-Electron Bolometer (CEB) detectors integrated in a Frequency Selective Surface (FSS) for the 350 GHz detection band of the OLIMPO balloon-borne telescope. In our architecture, the two terminal CEB has been integrated in the periodic unit cell of the FSS structure and is impedance matched to the embedding impedance seen by it and provides a resonant interaction with the incident sub-mm radiation. The detector array has been designed to operate in background noise limited condition for incident powers of 20 pW to 80 pW, making it possible to use the same pixel in both photometric and spectrometric configurations. We present high frequency and dc simulations of our system, together with fabrication details. The frequency response of the FSS array, optical response measurements with hot/cold load in front of optical window and with variable temperature black body source inside cryostat are presented. A comparison of the optical response to the CEB model and estimations of Noise Equivalent power (NEP) is also presented.
astro-ph_IM
Efficient Catalog Matching with Dropout Detection: Not only source catalogs are extracted from astronomy observations. Their sky coverage is always carefully recorded and used in statistical analyses, such as correlation and luminosity function studies. Here we present a novel method for catalog matching, which inherently builds on the coverage information for better performance and completeness. A modified version of the Zones Algorithm is introduced for matching partially overlapping observations, where irrelevant parts of the data are excluded up front for efficiency. Our design enables searches to focus on specific areas on the sky to further speed up the process. Another important advantage of the new method over traditional techniques is its ability to quickly detect dropouts, i.e., the missing components that are in the observed regions of the celestial sphere but did not reach the detection limit in some observations. These often provide invaluable insight into the spectral energy distribution of the matched sources but rarely available in traditional associations.
astro-ph_IM
Reproducibility of the First Image of a Black Hole in the Galaxy M87 from the Event Horizon Telescope (EHT) Collaboration: This paper presents an interdisciplinary effort aiming to develop and share sustainable knowledge necessary to analyze, understand, and use published scientific results to advance reproducibility in multi-messenger astrophysics. Specifically, we target the breakthrough work associated with the generation of the first image of a black hole, called M87. The image was computed by the Event Horizon Telescope Collaboration. Based on the artifacts made available by EHT, we deliver documentation, code, and a computational environment to reproduce the first image of a black hole. Our deliverables support new discovery in multi-messenger astrophysics by providing all the necessary tools for generalizing methods and findings from the EHT use case. Challenges encountered during the reproducibility of EHT results are reported. The result of our effort is an open-source, containerized software package that enables the public to reproduce the first image of a black hole in the galaxy M87.
astro-ph_IM
Rapid search for massive black hole binary coalescences using deep learning: The coalescences of massive black hole binaries are one of the main targets of space-based gravitational wave observatories. Such gravitational wave sources are expected to be accompanied by electromagnetic emissions. Low latency detection of the massive black hole mergers provides a start point for a global-fit analysis to explore the large parameter space of signals simultaneously being present in the data but at great computational cost. To alleviate this issue, we present a deep learning method for rapidly searching for signals of massive black hole binaries in gravitational wave data. Our model is capable of processing a year of data, simulated from the LISA data challenge, in only several seconds, while identifying all coalescences of massive black hole binaries with no false alarms. We further demonstrate that the model shows robust resistance to a wide range of generalization cases, including various waveform families and updated instrumental configurations. This method offers an effective approach that combines advances in artificial intelligence to open a new pathway for space-based gravitational wave observations.
astro-ph_IM
Experimental study of clusters in dense granular gas and implications for the particle stopping time in protoplanetary disks: In protoplanetary disks, zones of dense particle configuration promote planet formation. Solid particles in dense clouds alter their motion through collective effects and back reaction to the gas. The effect of particle-gas feedback with ambient solid-to-gas ratios $\epsilon > 1$ on the stopping time of particles is investigated. In experiments on board the International Space Station we studied the evolution of a dense granular gas while interacting with air. We observed diffusion of clusters released at the onset of an experiment but also the formation of new dynamical clusters. The solid-to-gas mass ratio outside the cluster varied in the range of about $\epsilon_{\rm avg} \sim 2.5 - 60$. We find that the concept of gas drag in a viscous medium still holds, even if the medium is strongly dominated in mass by solids. However, a collective factor has to be used, depending on $\epsilon_{\rm avg} $, i.e. the drag force is reduced by a factor 18 at the highest mass ratios. Therefore, flocks of grains in protoplanetary disks move faster and collide faster than their constituents might suggest.
astro-ph_IM
Performance update of an event-type based analysis for the Cherenkov Telescope Array: The Cherenkov Telescope Array (CTA) will be the next-generation observatory in the field of very-high-energy (20 GeV to 300 TeV) gamma-ray astroparticle physics. The traditional approach to data analysis in this field is to apply quality cuts, optimized using Monte Carlo simulations, on the data acquired to maximize sensitivity. Subsequent steps of the analysis typically use the surviving events to calculate one set of instrument response functions (IRFs) to physically interpret the results. However, an alternative approach is the use of event types, as implemented in experiments such as the Fermi-LAT. This approach divides events into sub-samples based on their reconstruction quality, and a set of IRFs is calculated for each sub-sample. The sub-samples are then combined in a joint analysis, treating them as independent observations. In previous works we demonstrated that event types, classified using Machine Learning methods according to their expected angular reconstruction quality, have the potential to significantly improve the CTA angular and energy resolution of a point-like source analysis. Now, we validated the production of event-type wise full-enclosure IRFs, ready to be used with science tools (such as Gammapy and ctools). We will report on the impact of using such an event-type classification on CTA high-level performance, compared to the traditional procedure.
astro-ph_IM
Polarized wavelets and curvelets on the sphere: The statistics of the temperature anisotropies in the primordial cosmic microwave background radiation field provide a wealth of information for cosmology and for estimating cosmological parameters. An even more acute inference should stem from the study of maps of the polarization state of the CMB radiation. Measuring the extremely weak CMB polarization signal requires very sensitive instruments. The full-sky maps of both temperature and polarization anisotropies of the CMB to be delivered by the upcoming Planck Surveyor satellite experiment are hence being awaited with excitement. Multiscale methods, such as isotropic wavelets, steerable wavelets, or curvelets, have been proposed in the past to analyze the CMB temperature map. In this paper, we contribute to enlarging the set of available transforms for polarized data on the sphere. We describe a set of new multiscale decompositions for polarized data on the sphere, including decimated and undecimated Q-U or E-B wavelet transforms and Q-U or E-B curvelets. The proposed transforms are invertible and so allow for applications in data restoration and denoising.
astro-ph_IM
HEIDI: An Automated Process for the Identification and Extraction of Photometric Light Curves from Astronomical Images: The production of photometric light curves from astronomical images is a very time-consuming task. Larger data sets improve the resolution of the light curve, however, the time requirement scales with data volume. The data analysis is often made more difficult by factors such as a lack of suitable calibration sources and the need to correct for variations in observing conditions from one image to another. Often these variations are unpredictable and corrections are based on experience and intuition. The High Efficiency Image Detection & Identification (HEIDI) pipeline software rapidly processes sets of astronomical images. HEIDI automatically selects multiple sources for calibrating the images using an algorithm that provides a reliable means of correcting for variations between images in a time series. The algorithm takes into account that some sources may intrinsically vary on short time scales and excludes these from being used as calibration sources. HEIDI processes a set of images from an entire night of observation, analyses the variations in brightness of the target objects and produces a light curve all in a matter of minutes. HEIDI has been tested on three different time series of asteroid 939 Isberga and has produced consistent high quality photometric light curves in a fraction of the usual processing time. The software can also be used for other transient sources, e.g. gamma-ray burst optical afterglows. HEIDI is implemented in Python and processes time series astronomical images with minimal user interaction. HEIDI processes up to 1000 images per run in the standard configuration. This limit can be easily increased. HEIDI is not telescope-dependent and will process images even in the case that no telescope specifications are provided. HEIDI has been tested on various Linux . HEIDI is very portable and extremely versatile with minimal hardware requirements.
astro-ph_IM
Exploring a search for long-duration transient gravitational waves associated with magnetar bursts: Soft gamma repeaters and anomalous X-ray pulsars are thought to be magnetars, neutron stars with strong magnetic fields of order $\mathord{\sim} 10^{13}$--$10^{15} \, \mathrm{gauss}$. These objects emit intermittent bursts of hard X-rays and soft gamma rays. Quasiperiodic oscillations in the X-ray tails of giant flares imply the existence of neutron star oscillation modes which could emit gravitational waves powered by the magnetar's magnetic energy reservoir. We describe a method to search for transient gravitational-wave signals associated with magnetar bursts with durations of 10s to 1000s of seconds. The sensitivity of this method is estimated by adding simulated waveforms to data from the sixth science run of Laser Interferometer Gravitational-wave Observatory (LIGO). We find a search sensitivity in terms of the root sum square strain amplitude of $h_{\mathrm{rss}} = 1.3 \times 10^{-21} \, \mathrm{Hz}^{-1/2}$ for a half sine-Gaussian waveform with a central frequency $f_0 = 150 \, \mathrm{Hz}$ and a characteristic time $\tau = 400 \, \mathrm{s}$. This corresponds to a gravitational wave energy of $E_{\mathrm{GW}} = 4.3 \times 10^{46} \, \mathrm{erg}$, the same order of magnitude as the 2004 giant flare which had an estimated electromagnetic energy of $E_{\mathrm{EM}} = \mathord{\sim} 1.7 \times 10^{46} (d/ 8.7 \, \mathrm{kpc})^2 \, \mathrm{erg}$, where $d$ is the distance to SGR 1806-20. We present an extrapolation of these results to Advanced LIGO, estimating a sensitivity to a gravitational wave energy of $E_{\mathrm{GW}} = 3.2 \times 10^{43} \, \mathrm{erg}$ for a magnetar at a distance of $1.6 \, \mathrm{kpc}$. These results suggest this search method can probe significantly below the energy budgets for magnetar burst emission mechanisms such as crust cracking and hydrodynamic deformation.
astro-ph_IM
The EPOCH Project: I. Periodic variable stars in the EROS-2 LMC database: The EPOCH (EROS-2 periodic variable star classification using machine learning) project aims to detect periodic variable stars in the EROS-2 light curve database. In this paper, we present the first result of the classification of periodic variable stars in the EROS-2 LMC database. To classify these variables, we first built a training set by compiling known variables in the Large Magellanic Cloud area from the OGLE and MACHO surveys. We crossmatched these variables with the EROS-2 sources and extracted 22 variability features from 28 392 light curves of the corresponding EROS-2 sources. We then used the random forest method to classify the EROS-2 sources in the training set. We designed the model to separate not only $\delta$ Scuti stars, RR Lyraes, Cepheids, eclipsing binaries, and long-period variables, the superclasses, but also their subclasses, such as RRab, RRc, RRd, and RRe for RR Lyraes, and similarly for the other variable types. The model trained using only the superclasses shows 99% recall and precision, while the model trained on all subclasses shows 87% recall and precision. We applied the trained model to the entire EROS-2 LMC database, which contains about 29 million sources, and found 117 234 periodic variable candidates. Out of these 117 234 periodic variables, 55 285 have not been discovered by either OGLE or MACHO variability studies. This set comprises 1 906 $\delta$ Scuti stars, 6 607 RR Lyraes, 638 Cepheids, 178 Type II Cepheids, 34 562 eclipsing binaries, and 11 394 long-period variables. A catalog of these EROS-2 LMC periodic variable stars will be available online at http://stardb.yonsei.ac.kr and at the CDS website (http://vizier.u-strasbg.fr/viz-bin/VizieR).
astro-ph_IM
Towards time symmetric N-body integration: Computational efficiency demands discretised, hierarchically organised, and individually adaptive time-step sizes (known as the block-step scheme) for the time integration of N-body models. However, most existing N-body codes adapt individual step sizes in a way that violates time symmetry (and symplecticity), resulting in artificial secular dissipation (and often secular growth of energy errors). Using single-orbit integrations, I investigate various possibilities to reduce or eliminate irreversibility from the time stepping scheme. Significant improvements over the standard approach are possible at little extra effort. However, in order to reduce irreversible step-size changes to negligible amounts, such as suitable for long-term integrations of planetary systems, more computational effort is needed, while exact time reversibility appears elusive for discretised individual step sizes.
astro-ph_IM
Search for Continuous Gravitational Wave Signals in Pulsar Timing Residuals: A New Scalable Approach with Diffusive Nested Sampling: Detecting continuous nanohertz gravitational waves (GWs) generated by individual close binaries of supermassive black holes (CB-SMBHs) is one of the primary objectives of pulsar timing arrays (PTAs). The detection sensitivity is slated to increase significantly as the number of well-timed millisecond pulsars will increase by more than an order of magnitude with the advent of next-generation radio telescopes. Currently, the Bayesian analysis pipeline using parallel tempering Markov chain Monte Carlo has been applied in multiple studies for CB-SMBH searches, but it may be challenged by the high dimensionality of the parameter space for future large-scale PTAs. One solution is to reduce the dimensionality by maximizing or marginalizing over uninformative parameters semi-analytically, but it is not clear whether this approach can be extended to more complex signal models without making overly simplified assumptions. Recently, the method of diffusive nested (DNest) sampling shown the capability of coping with high dimensionality and multimodality effectively in Bayesian analysis. In this paper, we apply DNest to search for continuous GWs in simulated pulsar timing residuals and find that it performs well in terms of accuracy, robustness, and efficiency for a PTA including $\mathcal{O}(10^2)$ pulsars. DNest also allows a simultaneous search of multiple sources elegantly, which demonstrates its scalability and general applicability. Our results show that it is convenient and also high beneficial to include DNest in current toolboxes of PTA analysis.
astro-ph_IM
The POLARBEAR-2 and Simons Array Focal Plane Fabrication Status: We present on the status of POLARBEAR-2 A (PB2-A) focal plane fabrication. The PB2-A is the first of three telescopes in the Simon Array (SA), which is an array of three cosmic microwave background (CMB) polarization sensitive telescopes located at the POLARBEAR (PB) site in Northern Chile. As the successor to the PB experiment, each telescope and receiver combination is named as PB2-A, PB2-B, and PB2-C. PB2-A and -B will have nearly identical receivers operating at 90 and 150 GHz while PB2-C will house a receiver operating at 220 and 270 GHz. Each receiver contains a focal plane consisting of seven close-hex packed lenslet coupled sinuous antenna transition edge sensor bolometer arrays. Each array contains 271 di-chroic optical pixels each of which have four TES bolometers for a total of 7588 detectors per receiver. We have produced a set of two types of candidate arrays for PB2-A. The first we call Version 11 (V11) and uses a silicon oxide (SiOx) for the transmission lines and cross-over process for orthogonal polarizations. The second we call Version 13 (V13) and uses silicon nitride (SiNx) for the transmission lines and cross-under process for orthogonal polarizations. We have produced enough of each type of array to fully populate the focal plane of the PB2-A receiver. The average wirebond yield for V11 and V13 arrays is 93.2% and 95.6% respectively. The V11 arrays had a superconducting transition temperature (Tc) of 452 +/- 15 mK, a normal resistance (Rn) of 1.25 +/- 0.20 Ohms, and saturations powers of 5.2 +/- 1.0 pW and 13 +/- 1.2 pW for the 90 and 150 GHz bands respectively. The V13 arrays had a superconducting transition temperature (Tc) of 456 +/-6 mK, a normal resistance (Rn) of 1.1 +/- 0.2 Ohms, and saturations powers of 10.8 +/- 1.8 pW and 22.9 +/- 2.6 pW for the 90 and 150 GHz bands respectively.
astro-ph_IM
A Method To Characterize the Wide-Angle Point Spread Function of Astronomical Images: Uncertainty in the wide-angle Point Spread Function (PSF) at large angles (tens of arcseconds and beyond) is one of the dominant sources of error in a number of important quantities in observational astronomy. Examples include the stellar mass and shape of galactic halos and the maximum extent of starlight in the disks of nearby galaxies. However, modeling the wide-angle PSF has long been a challenge in astronomical imaging. In this paper, we present a self-consistent method to model the wide-angle PSF in images. Scattered light from multiple bright stars is fitted simultaneously with a background model to characterize the extended wing of the PSF using a Bayesian framework operating on pixel-by-pixel level. The method is demonstrated using our software elderflower and is applied to data from the Dragonfly Telephoto Array to model its PSF out to 20-25 arcminutes. We compare the wide-angle PSF of Dragonfly to that of a number of other telescopes, including the SDSS PSF, and show that on scales of arcminutes the scattered light in the Dragonfly PSF is markedly lower than that of other wide-field imaging telescopes. The energy in the wings of the Dragonfly point-spread function is sufficiently low that optical cleanliness plays an important role in defining the PSF. This component of the PSF can be modelled accurately, highlighting the power of our self-contained approach.
astro-ph_IM
Systematic Serendipity: A Test of Unsupervised Machine Learning as a Method for Anomaly Detection: Advances in astronomy are often driven by serendipitous discoveries. As survey astronomy continues to grow, the size and complexity of astronomical databases will increase, and the ability of astronomers to manually scour data and make such discoveries decreases. In this work, we introduce a machine learning-based method to identify anomalies in large datasets to facilitate such discoveries, and apply this method to long cadence lightcurves from NASA's Kepler Mission. Our method clusters data based on density, identifying anomalies as data that lie outside of dense regions. This work serves as a proof-of-concept case study and we test our method on four quarters of the Kepler long cadence lightcurves. We use Kepler's most notorious anomaly, Boyajian's Star (KIC 8462852), as a rare `ground truth' for testing outlier identification to verify that objects of genuine scientific interest are included among the identified anomalies. We evaluate the method's ability to identify known anomalies by identifying unusual behavior in Boyajian's Star, we report the full list of identified anomalies for these quarters, and present a sample subset of identified outliers that includes unusual phenomena, objects that are rare in the Kepler field, and data artifacts. By identifying <4% of each quarter as outlying data, we demonstrate that this anomaly detection method can create a more targeted approach in searching for rare and novel phenomena.
astro-ph_IM
The Radio Detector of the Pierre Auger Observatory -- status and expected performance: As part of the ongoing AugerPrime upgrade of the Pierre Auger Observatory, we are deploying short aperiodic loaded loop antennas measuring radio signals from extensive air showers in the 30-80 MHz band on each of the 1,660 surface detector stations. This new Radio Detector of the Observatory allows us to measure the energy in the electromagnetic cascade of inclined air showers with zenith angles larger than $\sim 65^\circ$. The water-Cherenkov detectors, in turn, perform a virtually pure measurement of the muon component of inclined air showers. The combination of both thus extends the mass-composition sensitivity of the upgraded Observatory to high zenith angles and therefore enlarges the sky coverage of mass-sensitive measurements at the highest energies while at the same time allowing us to cross-check the performance of the established detectors with an additional measurement technique. In this contribution, we outline the concept and design of the Radio Detector, report on its current status and initial results from the first deployed stations, and illustrate its expected performance with a detailed, end-to-end simulation study.
astro-ph_IM
Radio Weak Lensing Shear Measurement in the Visibility Domain - II. Source Extraction: This paper extends the method introduced in Rivi et al. (2016b) to measure galaxy ellipticities in the visibility domain for radio weak lensing surveys. In that paper we focused on the development and testing of the method for the simple case of individual galaxies located at the phase centre, and proposed to extend it to the realistic case of many sources in the field of view by isolating visibilities of each source with a faceting technique. In this second paper we present a detailed algorithm for source extraction in the visibility domain and show its effectiveness as a function of the source number density by running simulations of SKA1-MID observations in the band 950-1150 MHz and comparing original and measured values of galaxies' ellipticities. Shear measurements from a realistic population of 10^4 galaxies randomly located in a field of view of 1 deg^2 (i.e. the source density expected for the current radio weak lensing survey proposal with SKA1) are also performed. At SNR >= 10, the multiplicative bias is only a factor 1.5 worse than what found when analysing individual sources, and is still comparable to the bias values reported for similar measurement methods at optical wavelengths. The additive bias is unchanged from the case of individual sources, but is significantly larger than typically found in optical surveys. This bias depends on the shape of the uv coverage and we suggest that a uv-plane weighting scheme to produce a more isotropic shape could reduce and control additive bias.
astro-ph_IM
The ARCADE 2 Instrument: The second generation Absolute Radiometer for Cosmology, Astrophysics, and Diffuse Emission (ARCADE 2) instrument is a balloon-borne experiment to measure the radiometric temperature of the cosmic microwave background and Galactic and extra-Galactic emission at six frequencies from 3 to 90 GHz. ARCADE 2 utilizes a double-nulled design where emission from the sky is compared to that from an external cryogenic full-aperture blackbody calibrator by cryogenic switching radiometers containing internal blackbody reference loads. In order to further minimize sources of systematic error, ARCADE 2 features a cold fully open aperture with all radiometrically active components maintained at near 2.7 K without windows or other warm objects, achieved through a novel thermal design. We discuss the design and performance of the ARCADE 2 instrument in its 2005 and 2006 flights.
astro-ph_IM
Science Platforms for Heliophysics Data Analysis: We recommend that NASA maintain and fund science platforms that enable interactive and scalable data analysis in order to maximize the scientific return of data collected from space-based instruments.
astro-ph_IM
MuSCAT2: four-color Simultaneous Camera for the 1.52-m Telescopio Carlos Sánchez: We report the development of a 4-color simultaneous camera for the 1.52~m Telescopio Carlos S\'anchez (TCS) in the Teide Observatory, Canaries, Spain. The new instrument, named MuSCAT2, has a capability of 4-color simultaneous imaging in $g$ (400--550 nm), $r$ (550--700 nm), $i$ (700--820 nm), and $z_s$ (820--920 nm) bands. MuSCAT2 equips four 1024$\times$1024 pixel CCDs, having a field of view of 7.4$\times$7.4 arcmin$^2$ with a pixel scale of 0.44 arcsec per pixel. The principal purpose of MuSCAT2 is to perform high-precision multi-color exoplanet transit photometry. We have demonstrated photometric precisions of 0.057%, 0.050%, 0.060%, and 0.076% as root-mean-square residuals of 60~s binning in $g$, $r$, $i$ and $z_s$ bands, respectively, for a G0 V star WASP-12 ($V=11.57\pm0.16$). MuSCAT2 has started science operations since January 2018, with over 250 telescope nights per year. MuSCAT2 is expected to become a reference tool for exoplanet transit observations, and will substantially contribute to the follow-up of the TESS and PLATO space missions.
astro-ph_IM
Rejection criteria based on outliers in the KiDS photometric redshifts and PDF distributions derived by machine learning: The Probability Density Function (PDF) provides an estimate of the photometric redshift (zphot) prediction error. It is crucial for current and future sky surveys, characterized by strict requirements on the zphot precision, reliability and completeness. The present work stands on the assumption that properly defined rejection criteria, capable of identifying and rejecting potential outliers, can increase the precision of zphot estimates and of their cumulative PDF, without sacrificing much in terms of completeness of the sample. We provide a way to assess rejection through proper cuts on the shape descriptors of a PDF, such as the width and the height of the maximum PDF's peak. In this work we tested these rejection criteria to galaxies with photometry extracted from the Kilo Degree Survey (KiDS) ESO Data Release 4, proving that such approach could lead to significant improvements to the zphot quality: e.g., for the clipped sample showing the best trade-off between precision and completeness, we achieve a reduction in outliers fraction of $\simeq 75\%$ and an improvement of $\simeq 6\%$ for NMAD, with respect to the original data set, preserving the $\simeq 93\%$ of its content.
astro-ph_IM
The Effects of Improper Lighting on Professional Astronomical Observations: Europe and a number of countries in the world are investing significant amounts of public money to operate and maintain large, ground-based astronomical facilities. Even larger projects are under development to observe the faintest and most remote astrophysical sources in the universe. As of today, on the planet there are very few sites that satisfy all the demanding criteria for such sensitive and expensive equipment, including a low level of light pollution. Because of the uncontrolled growth of incorrect illumination, even these protected and usually remote sites are at risk. Although the reasons for intelligent lighting reside in energy saving and environmental effects, the impact on scientific research cannot be neglected or underestimated, because of its high cultural value for the progress of the whole mankind. After setting the stage, in this paper I review the effects of improper lighting on professional optical and near-UV astronomical data, and discuss the possible solutions to both preserve the night sky natural darkness and produce an efficient and cost-effective illumination.
astro-ph_IM
The surface detector array of the Telescope Array experiment: The Telescope Array (TA) experiment, located in the western desert of Utah,USA, is designed for observation of extensive air showers from extremely high energy cosmic rays. The experiment has a surface detector array surrounded by three fluorescence detectors to enable simultaneous detection of shower particles at ground level and fluorescence photons along the shower track. The TA surface detectors and fluorescence detectors started full hybrid observation in March, 2008. In this article we describe the design and technical features of the TA surface detector.
astro-ph_IM
Soft proton scattering at grazing incidence from X-ray mirrors: analysis of experimental data in the framework of the non-elastic approximation: Astronomical X-ray observatories with grazing incidence optics face the problem of pseudo-focusing of low energy protons from the mirrors towards the focal plane. Those protons constitute a variable, unpredictable component of the non X-ray background that strongly affects astronomical observations and a correct estimation of their flux at the focal plane is then essential. For this reason, we investigate how they are scattered from the mirror surfaces when impacting with grazing angles. We compare the non-elastic model of reflectivity of particles at grazing incidence proposed by Remizovich et al. (1980) with the few available experimental measurements of proton scattering from X-ray mirrors. We develop a semi-empirical analytical model based on the fit of those experimental data with the Remizovich solution. We conclude that the scattering probability weakly depends on the energy of the impinging protons and that the relative energy losses are necessary to correctly model the data. The model we propose assumes no dependence on the incident energy and can be implemented in particle transport simulation codes to generate, for instance, proton response matrices for specific X-ray missions. Further laboratory measurements at lower energies and on other mirror samples, such as ATHENA Silicon Pore Optics, will improve the resolution of the model and will allow us to build the proper proton response matrices for a wider sample of X-ray observatories.
astro-ph_IM
Studying the Impact of Optical Aberrations on Diffraction-Limited Radial Velocity Instruments: Spectrographs nominally contain a degree of quasi-static optical aberrations resulting from the quality of manufactured component surfaces, imperfect alignment, design residuals, thermal effects, and other other associated phenomena involved in the design and construction process. Aberrations that change over time can mimic the line centroid motion of a Doppler shift, introducing radial velocity (RV) uncertainty that increases time-series variability. Even when instrument drifts are tracked using a precise wavelength calibration source, barycentric motion of the Earth leads to a wavelength shift of stellar light causing a translation of the spectrum across the focal plane array by many pixels. The wavelength shift allows absorption lines to experience different optical propagation paths and aberrations over observing epochs. We use physical optics propagation simulations to study the impact of aberrations on precise Doppler measurements made by diffraction-limited, high-resolution spectrographs. We quantify the uncertainties that cross-correlation techniques introduce in the presence of aberrations and barycentric RV shifts. We find that aberrations which shift the PSF photo-center in the dispersion direction, in particular primary horizontal coma and trefoil, are the most concerning. To maintain aberration-induced RV errors less than 10 cm/s, phase errors for these particular aberrations must be held well below 0.05 waves at the instrument operating wavelength. Our simulations further show that wavelength calibration only partially compensates for instrumental drifts, owing to a behavioral difference between how cross-correlation techniques handle aberrations between starlight versus calibration light. Identifying subtle physical effects that influence RV errors will help ensure that diffraction-limited planet-finding spectrographs are able to reach their full scientific potential.
astro-ph_IM