text
stringlengths 6
128k
|
---|
We consider a relativistic particle model in an enlarged relativistic phase
space M^{18} = (X_\mu, P_\mu, \eta_\alpha, \oeta_\dalpha, \sigma_\alpha,
\osigma_\dalpha, e, \phi), which is derived from the free two-twistor dynamics.
The spin sector variables (\eta_\alpha, \oeta_\dalpha, \sigma_\alpha,\
osigma_\dalpha) satisfy two second class constraints and account for the
relativistic spin structure, and the pair (e,\phi) describes the electric
charge sector. After introducing the Liouville one-form on M^{18}, derived by a
non-linear transformation of the canonical Liouville one-form on the
two-twistor space, we analyze the dynamics described by the first and second
class constraints. We use a composite orthogonal basis in four-momentum space
to obtain the scalars defining the invariant spin projections. The
first-quantized theory provides a consistent set of wave equations, determining
the mass, spin, invariant spin projection and electric charge of the
relativistic particle. The wavefunction provides a generating functional for
free, massive higher spin fields.
|
The high efficiency of converting kinetic energy into gamma-rays estimated
with late-time afterglows in Gamma-Ray Burst (GRB) phenomenon challenges the
commonly accepted internal-shock model. However, the efficiency is still highly
uncertain because it is sensitive to many effects. In this Letter we study the
sideways expansion effect of jets on estimating the efficiency. We find that
this effect is considerable, reducing the efficiency by a factor of $\sim0.5$
for typical parameters, when the afterglow data $\sim 10$ hr after the GRB
trigger are used to derive the kinetic energy. For a more dense circumburst
medium, this effect is more significant. As samples, taking this effect into
account, we specifically calculate the efficiency of two bursts whose
parameters were well constrained. Almost the same results are derived. This
suggests that the sideways expansion effect should be considered when the GRB
efficiency is estimated with the late afterglow data.
|
We classify the trace anomaly for parity-invariant non-relativistic
Schr\"odinger theories in 2+1 dimensions coupled to background Newton-Cartan
gravity. The general anomaly structure looks very different from the one in the
z=2 Lifshitz theories. The type A content of the anomaly is remarkably
identical to that of the relativistic 3+1 dimensional case, suggesting the
conjecture that an a-theorem should exist also in the Newton-Cartan context.
Erratum: due to an overcounting of the number of linearly-independent terms
in the basis, the type A anomaly disappears if Frobenius condition is imposed.
See appended erratum for details. This crucial mistake was pointed out to us in
arXiv:1601.06795.
|
We develop an online learning method for prediction, which is important in
problems with large and/or streaming data sets. We formulate the learning
approach using a covariance-fitting methodology, and show that the resulting
predictor has desirable computational and distribution-free properties: It is
implemented online with a runtime that scales linearly in the number of
samples; has a constant memory requirement; avoids local minima problems; and
prunes away redundant feature dimensions without relying on restrictive
assumptions on the data distribution. In conjunction with the split conformal
approach, it also produces distribution-free prediction confidence intervals in
a computationally efficient manner. The method is demonstrated on both real and
synthetic datasets.
|
In this talk, I will concentrate on $Q^2$-dependence of deep inelastic sum
rules. I will first give a modern definition of deep-inelastic sum rules and
then discuss physical origins of their scaling violation at finite $Q^2$.
Following this, I discuss a few well-known examples, in particular, the Bjorken
sum rule, which is at the center of interest of this symposium. Finally, I
consider the $Q^2 \to 0$ limit of sum rules using low-energy theorems. I think
this can motivate some interesting CEBAF physics.
|
We study the pi N -> pi pi N reaction around the N(1440) mass-shell energy.
Considering the total cross sections and invariant mass distributions, we
discuss the role of N(1440) and its decay processes. The calculation is
performed by extending our previous approach [Phys. Rev. C 69, 025206 (2004)]
to this reaction, in which only the nucleon and Delta(1232) were considered as
intermediate baryon states. The characteristics observed in the recent data for
the pi- p -> pi0 pi0 n reaction obtained by Crystal Ball Collaboration (CBC),
can be understood as a strong interference between the two decay processes:
N(1440) -> pi Delta(1232) and N(1440) -> N(pi pi)_S. It is also found that the
scalar-isoscalar pi pi rescattering effect in the NN*(pi pi)_S vertex, which
corresponds to the propagation of sigma meson, seems to be necessary for
explain ing the several observables of the pi N -> pi pi N reaction: the large
asymmetric shape in the pi0-pi0 invariant mass distributions of the pi- p ->
pi0 pi0 n reaction and the pi+ p -> pi+ pi+ n total cross section.
|
Let $A/\overline{\mathbb{F}}\_p$ and $A'/\overline{\mathbb{F}}\_p$ be
supersingular principally polarized abelian varieties of dimension $g>1$. For
any prime $\ell \ne p$, we give an algorithm that finds a path $\phi \colon A
\rightarrow A'$ in the $(\ell, \dots , \ell)$-isogeny graph in
$\widetilde{O}(p^{g-1})$ group operations on a classical computer, and
$\widetilde{O}(\sqrt{p^{g-1}})$ calls to the Grover oracle on a quantum
computer. The idea is to find paths from $A$ and $A'$ to nodes that correspond
to products of lower dimensional abelian varieties, and to recurse down in
dimension until an elliptic path-finding algorithm (such as Delfs--Galbraith)
can be invoked to connect the paths in dimension $g=1$. In the general case
where $A$ and $A'$ are any two nodes in the graph, this algorithm presents an
asymptotic improvement over all of the algorithms in the current literature. In
the special case where $A$ and $A'$ are a known and relatively small number of
steps away from each other (as is the case in higher dimensional analogues of
SIDH), it gives an asymptotic improvement over the quantum claw finding
algorithms and an asymptotic improvement over the classical van
Oorschot--Wiener algorithm.
|
We consider the estimation of a signal from the knowledge of its noisy linear
random Gaussian projections, a problem relevant in compressed sensing, sparse
superposition codes or code division multiple access just to cite few. There
has been a number of works considering the mutual information for this problem
using the heuristic replica method from statistical physics. Here we put these
considerations on a firm rigorous basis. First, we show, using a Guerra-type
interpolation, that the replica formula yields an upper bound to the exact
mutual information. Secondly, for many relevant practical cases, we present a
converse lower bound via a method that uses spatial coupling, state evolution
analysis and the I-MMSE theorem. This yields, in particular, a single letter
formula for the mutual information and the minimal-mean-square error for random
Gaussian linear estimation of all discrete bounded signals.
|
Brains have evolved a diverse set of neurons with varying morphologies,
physiological properties and rich dynamics that impact their processing of
temporal information. By contrast, most neural network models include a
homogeneous set of units that only vary in terms of their spatial parameters
(weights and biases). To investigate the importance of temporal parameters to
neural function, we trained spiking neural networks on tasks of varying
temporal complexity, with different subsets of parameters held constant. We
find that in a tightly resource constrained setting, adapting conduction delays
is essential to solve all test conditions, and indeed that it is possible to
solve these tasks using only temporal parameters (delays and time constants)
with weights held constant. In the most complex spatio-temporal task we
studied, we found that an adaptable bursting parameter was essential. More
generally, allowing for adaptation of both temporal and spatial parameters
increases network robustness to noise, an important feature for both biological
brains and neuromorphic computing systems. In summary, our findings highlight
how rich and adaptable dynamics are key to solving temporally structured tasks
at a low neural resource cost, which may be part of the reason why biological
neurons vary so dramatically in their physiological properties.
|
Floquet engineering is a convenient strategy to induce nonequilibrium
phenomena in molecular and solid-state systems, or to dramatically alter the
physicochemical properties of matter, bypassing costly and time-consuming
synthetic modifications. In this article, we investigate theoretically some
interesting consequences of the fact that an originally achiral molecular
system can exhibit nonzero circular dichroism (CD) when it is driven with
elliptically polarized light. More specifically, we consider an isotropic
ensemble of small cyclic molecular aggregates in solution whose local
low-frequency vibrational modes are driven by an elliptically polarized
continuous-wave infrared pump. We attribute the origin of a nonzero CD signal
to time-reversal symmetry breaking due to an excitonic Aharonov-Bohm (AB) phase
arising from a time-varying laser electric field, together with coherent
interchromophoric exciton hopping. The obtained Floquet engineered excitonic AB
phases are far more tunable than analogous magnetically-induced electronic AB
phases in nanoscale rings, highlighting a virtually unexplored potential that
Floquet engineered AB phases have in the coherent control of molecular
processes and simultaneously introducing new analogues of magneto-optical
effects in molecular systems which bypass the use of strong magnetic fields.
|
We observed the HH 211 jet in the submillimeter continuum and the CO(3-2) and
SiO(8-7) transitions with the Submillimeter Array. The continuum source
detected at the center of the outflow shows an elongated morphology,
perpendicular to the direction of the outflow axis. The high-velocity emission
of both molecules shows a knotty and highly collimated structure. The SiO(8-7)
emission at the base of the outflow, close to the driving source, spans a wide
range of velocities, from -20 up to 40 km s^{-1}. This suggests that a
wide-angle wind may be the driving mechanism of the HH 211 outflow. For
distances greater than 5" (1500 AU) from the driving source, emission from both
transitions follows a Hubble-law behavior, with SiO(8-7) reaching higher
velocities than CO(3-2), and being located upstream of the CO(3-2) knots. This
indicates that the SiO(8-7) emission is likely tracing entrained gas very close
to the primary jet, while the CO(3-2) is tracing less dense entrained gas. From
the SiO(5-4) data of Hirano et al. we find that the SiO(8-7)/SiO(5-4)
brightness temperature ratio along the jet decreases for knots far from the
driving source. This is consistent with the density decreasing along the jet,
from (3-10)x10^6 cm^{-3} at 500 AU to (0.8-4)x10^6 cm^{-3} at 5000 AU from the
driving source.
|
The barrel time-of-flight (TOF) detector for the PANDA experiment at FAIR in
Darmstadt is planned as a scintillator tile hodoscope (SciTil) using 8000 small
scintillator tiles. It will provide fast event timing for a software trigger in
the otherwise trigger-less data acquisition scheme of PANDA, relative timing in
a multiple track event topology as well as additional particle identification
in the low momentum region. The goal is to achieve a time resolution of sigma ~
100 ps. We have conducted measurements using organic scintillators coupled to
Silicon Photomultipliers (SiPM). The results are encouraging such that we are
confident to reach the required time resolution.
|
Graphene has raised high expectations as a low-loss plasmonic material in
which the plasmon properties can be controlled via electrostatic doping. Here,
we analyze realistic configurations, which produce inhomogeneous doping, in
contrast to what has been so far assumed in the study of plasmons in
nanostructured graphene. Specifically, we investigate backgated ribbons,
co-planar ribbon pairs placed at opposite potentials, and individual ribbons
subject to a uniform electric field. Plasmons in backgated ribbons and ribbon
pairs are similar to those of uniformly doped ribbons, provided the Fermi
energy is appropriately scaled to compensate for finite-size effects such as
the divergence of the carrier density at the edges. In contrast, the plasmons
of a ribbon exposed to a uniform field exhibit distinct dispersion and spatial
profiles that considerably differ from uniformly doped ribbons. Our results
provide a road map to understand graphene plasmons under realistic
electrostatic doping conditions.
|
Model predictive control (MPC) anticipates future events to take appropriate
control actions. Nonlinear MPC (NMPC) deals with nonlinear models and/or
constraints. A Continuation/GMRES Method for NMPC, suggested by T. Ohtsuka in
2004, uses the GMRES iterative algorithm to solve a forward difference
approximation $Ax=b$ of the original NMPC equations on every time step. We have
previously proposed accelerating the GMRES and MINRES convergence by
preconditioning the coefficient matrix $A$. We now suggest simplifying the
construction of the preconditioner, by approximately solving a forward
recursion for the state and a backward recursion for the costate, or simply
reusing previously computed solutions.
|
Alloys of Bi$_2$Te$_3$ and Sb$_2$Te$_3$ ((Bi$_{1-x}$Sb$_x$)$_2$Te$_3$) have
played an essential role in the exploration of topological surface states,
allowing us to study phenomena that would otherwise be obscured by bulk
contributions to conductivity. Thin films of these alloys have been
particularly important for tuning the energy of the Fermi level, a key step in
observing spin-polarized surface currents and the quantum anomalous Hall
effect. Previous studies reported the chemical tuning of the Fermi level to the
Dirac point by controlling the Sb:Bi composition ratio, but the optimum ratio
varies widely across various studies with no consensus. In this work, we use
scanning tunneling microscopy and Landau level spectroscopy, in combination
with X-ray photoemission spectroscopy to isolate the effects of growth factors
such as temperature and composition, and to provide a microscopic picture of
the role that disorder and composition play in determining the carrier density
of epitaxially grown (Bi,Sb)$_2$Te$_3$ thin films. Using Landau level
spectroscopy, we determine that the ideal Sb concentration to place the Fermi
energy to within a few meV of the Dirac point is $x\sim 0.7$. However, we find
that the post- growth annealing temperature can have a drastic impact on
microscopic structure as well as carrier density. In particular, we find that
when films are post-growth annealed at high temperature, better crystallinity
and surface roughness are achieved; but this also produces a larger Te defect
density, adding n-type carriers. This work provides key information necessary
for optimizing thin film quality in this fundamentally and technologically
important class of materials.
|
We investigate the behavior of an $N$-component quantum rotor coupled to a
bosonic dissipative bath having a sub-Ohmic spectral density $J(\omega) \propto
\omega^s$ with $s<1$. With increasing dissipation strength, this system
undergoes a quantum phase transition from a delocalized phase to a localized
phase. We determine the exact critical behavior of this transition in the
large-$N$ limit. For $1>s>1/2$, we find nontrivial critical behavior
corresponding to an interacting renormalization group fixed point while we find
mean-field behavior for $s<1/2$. The results agree with those of the
corresponding long-range interacting classical model. The quantum-to-classical
mapping is therefore valid for the sub-Ohmic rotor model.
|
Nanoscale photothermal sources find important applications in theranostics,
imaging, and catalysis. In this context, graphene offers a unique suite of
optical, electrical, and thermal properties, which we exploit to show
self-consistent active photothermal modulation of its nanoscale response. In
particular, we predict the existence of plasmons confined to the optical
landscape tailored by continuous-wave external-light pumping of homogeneous
graphene. This result relies on the high electron temperatures achievable in
optically pumped clean graphene while its lattice remains near ambient
temperature. Our study opens a new avenue toward the active optical control of
the nanophotonic response in graphene with potential application in
photothermal devices.
|
Molecular D/H ratios are frequently used to probe the chemical past of Solar
System volatiles. Yet it is unclear which parts of the Solar Nebula hosted an
active deuterium fractionation chemistry. To address this question, we present
0".2-0".4 ALMA observations of DCO+ and DCN 2-1, 3-2 and 4-3 towards the nearby
protoplanetary disk around TW Hya, taken as part of the TW Hya Rosetta Stone
project, augmented with archival data. DCO+ is characterized by an excitation
temperature of ~40 K across the 70 au radius pebble disk, indicative of
emission from a warm, elevated molecular layer. Tentatively, DCN is present at
even higher temperatures. Both DCO+ and DCN present substantial emission
cavities in the inner disk, while in the outer disk the DCO+ and DCN
morphologies diverge: most DCN emission originates from a narrow ring peaking
around 30~au, with some additional diffuse DCN emission present at larger
radii, while DCO+ is present in a broad structured ring that extends past the
pebble disk. Based on parametric disk abundance models, these emission patterns
can be explained by a near-constant DCN abundance exterior to the cavity, and
an increasing DCO+ abundance with radius. There appears to be an active
deuterium fractionation chemistry in multiple disk regions around TW Hya, but
not in the cold planetesimal-forming midplane and in the inner disk. More
observations are needed to explore whether deuterium fractionation is actually
absent in these latter regions, and if its absence is a common feature, or
something peculiar to the old TW Hya disk.
|
A non-perturbative lattice regularization of chiral fermions and bosons with
anomaly-free symmetry $G$ in 1+1D spacetime is proposed. More precisely, we ask
"whether there is a local short-range quantum Hamiltonian with a finite Hilbert
space for a finite system realizing onsite symmetry $G$ defined on a 1D spatial
lattice, such that its low energy physics produces a 1+1D anomaly-free chiral
matter theory of symmetry $G$?" In particular, we propose that the
3$_L$-5$_R$-4$_L$-0$_R$ U(1) chiral fermion theory, with two left-moving
fermions of charge-3 and 4, and two right-moving fermions of charge-5 and 0 at
low energy, can be put on a 1D spatial lattice where the U(1) symmetry is
realized as an onsite symmetry, if we include properly designed multi-fermion
interactions with intermediate strength. In general, we propose that any 1+1D
U(1)-anomaly-free chiral matter theory can be defined as a finite system on a
1D lattice with onsite symmetry by using a quantum Hamiltonian with continuous
time, but without suffering from Nielsen-Ninomiya theorem's fermion-doubling,
if we include properly-designed interactions between matter fields. We propose
how to design such interactions by looking for extra symmetries via
bosonization/fermionization. We comment on the new ingredients and the
differences of ours compared to Ginsparg-Wilson fermion, Eichten-Preskill, and
Chen-Giedt-Poppitz (CGP) models, and suggest modifying CGP model to have
successful mirror-decoupling. We show a topological non-perturbative proof of
the equivalence between $G$-symmetric 't Hooft anomaly cancellation conditions
and $G$-symmetric gapping rules (e.g. Haldane's stability conditions for
Luttinger liquid) for multi-U(1) symmetry. We expect our result holds
universally regardless of spatial Hamiltonian or Lagrangian/spacetime path
integral formulation. Numerical tests are demanding tasks but highly desirable
for future work.
|
In this paper, we bring forth a novel approach of video text detection using
Fourier-Laplacian filtering in the frequency domain that includes a
verification technique using Hidden Markov Model (HMM). The proposed approach
deals with the text region appearing not only in horizontal or vertical
directions, but also in any other oblique or curved orientation in the image.
Until now only a few methods have been proposed that look into curved text
detection in video frames, wherein lies our novelty. In our approach, we first
apply Fourier-Laplacian transform on the image followed by an ideal
Laplacian-Gaussian filtering. Thereafter K-means clustering is employed to
obtain the asserted text areas depending on a maximum difference map. Next, the
obtained connected components (CC) are skeletonized to distinguish various text
strings. Complex components are disintegrated into simpler ones according to a
junction removal algorithm followed by a concatenation performed on possible
combination of the disjoint skeletons to obtain the corresponding text area.
Finally these text hypotheses are verified using HMM-based text/non-text
classification system. False positives are thus eliminated giving us a robust
text detection performance. We have tested our framework in multi-oriented text
lines in four scripts, namely, English, Chinese, Devanagari and Bengali, in
video frames and scene texts. The results obtained show that proposed approach
surpasses existing methods on text detection.
|
Magnetic resonance spectroscopy (MRS) is an important technique in biomedical
research and it has the unique capability to give a non-invasive access to the
biochemical content (metabolites) of scanned organs. In the literature, the
quantification (the extraction of the potential biomarkers from the MRS
signals) involves the resolution of an inverse problem based on a parametric
model of the metabolite signal. However, poor signal-to-noise ratio (SNR),
presence of the macromolecule signal or high correlation between metabolite
spectral patterns can cause high uncertainties for most of the metabolites,
which is one of the main reasons that prevents use of MRS in clinical routine.
In this paper, quantification of metabolites in MR Spectroscopic imaging using
deep learning is proposed. A regression framework based on the Convolutional
Neural Networks (CNN) is introduced for an accurate estimation of spectral
parameters. The proposed model learns the spectral features from a large-scale
simulated data set with different variations of human brain spectra and SNRs.
Experimental results demonstrate the accuracy of the proposed method, compared
to state of the art standard quantification method (QUEST), on concentration of
20 metabolites and the macromolecule.
|
In the limit of extremely rapid mass transfer, the response of a donor star
in an interacting binary becomes asymptotically one of adiabatic expansion. We
survey here adiabatic mass loss from Population I stars of mass 0.10 Msun to
100 Msun from the zero age main sequence to the base of the giant branch, or to
central hydrogen exhaustion for lower main sequence stars. For intermediate-
and high-mass stars, dynamical mass transfer is preceded by an extended phase
of thermal time scale mass transfer as the star is stripped of most of its
envelope mass. The critical mass ratio qad above which this delayed dynamical
instability occurs increases with advancing evolutionary age of the donor star,
by ever-increasing factors for more massive donors. Most intermediate- or
high-mass binaries with nondegenerate accretors probably evolve into contact
before manifesting this instability. As they approach the base of the giant
branch, however, and begin developing a convective envelope, qad plummets
dramatically among intermediate-mass stars, to values of order unity, and a
prompt dynamical instability occurs. Among low-mass stars, the prompt
instability prevails throughout main sequence evolution, with q_ad declining
with decreasing mass. Our calculated qad agree well with the behavior of
timedependent models by Chen & Han (2003) of intermediate-mass stars initiating
mass transfer in the Hertzsprung gap. Application of our results to cataclysmic
variables, as systems which must be stable against rapid mass transfer, nicely
circumscribes the range in qad as a function of orbital period in which they
are found. These results are intended to advance the verisimilitude of
population synthesis models of close binary evolution.
|
Visual data in autonomous driving perception, such as camera image and LiDAR
point cloud, can be interpreted as a mixture of two aspects: semantic feature
and geometric structure. Semantics come from the appearance and context of
objects to the sensor, while geometric structure is the actual 3D shape of
point clouds. Most detectors on LiDAR point clouds focus only on analyzing the
geometric structure of objects in real 3D space. Unlike previous works, we
propose to learn both semantic feature and geometric structure via a unified
multi-view framework. Our method exploits the nature of LiDAR scans -- 2D range
images, and applies well-studied 2D convolutions to extract semantic features.
By fusing semantic and geometric features, our method outperforms
state-of-the-art approaches in all categories by a large margin. The
methodology of combining semantic and geometric features provides a unique
perspective of looking at the problems in real-world 3D point cloud detection.
|
Previous results reported in the robotics literature show the relationship
between time-delay control (TDC) and proportional-integral-derivative control
(PID). In this paper, we show that incremental nonlinear dynamic inversion
(INDI) - more familiar in the aerospace community - are in fact equivalent to
TDC. This leads to a meaningful and systematic method for PI(D)-control tuning
of robust nonlinear flight control systems via INDI. We considered a
reformulation of the plant dynamics inversion which removes effector blending
models from the resulting control law, resulting in robust model-free control
laws like PI(D)-control.
|
We study the collapse of a spherically symmetric dust distribution in
$d$-dimensional AdS spacetime. We investigate the role of dimensionality, and
the presence of a negative cosmological constant, in determining the formation
of trapped surfaces and the end state of gravitational collapse. We obtain the
self-similar solution for the case of zero cosmological constant, and show that
one cannot construct a self-similar solution when a cosmological constant is
included.
|
Structured recursion schemes such as folds and unfolds have been widely used
for structuring both functional programs and program semantics. In this
context, it has been customary to implement denotational semantics as folds
over an inductive data type to ensure termination and compositionality.
Separately, operational models can be given by unfolds, and naturally not all
operational models coincide with a given denotational semantics in a meaningful
way. To ensure these semantics are coherent it is important to consider the
property of full abstraction which relates the denotational and the operational
model. In this paper, we show how to engineer a compositional semantics such
that full abstraction comes for free. We do this by using distributive laws
from which we generate both the operational and the denotational model. The
distributive laws ensure the semantics are fully abstract at the type level,
thus relieving the programmer from the burden of the proofs.
|
We delve into pseudo-labeling for semi-supervised monocular 3D object
detection (SSM3OD) and discover two primary issues: a misalignment between the
prediction quality of 3D and 2D attributes and the tendency of depth
supervision derived from pseudo-labels to be noisy, leading to significant
optimization conflicts with other reliable forms of supervision. We introduce a
novel decoupled pseudo-labeling (DPL) approach for SSM3OD. Our approach
features a Decoupled Pseudo-label Generation (DPG) module, designed to
efficiently generate pseudo-labels by separately processing 2D and 3D
attributes. This module incorporates a unique homography-based method for
identifying dependable pseudo-labels in BEV space, specifically for 3D
attributes. Additionally, we present a DepthGradient Projection (DGP) module to
mitigate optimization conflicts caused by noisy depth supervision of
pseudo-labels, effectively decoupling the depth gradient and removing
conflicting gradients. This dual decoupling strategy-at both the pseudo-label
generation and gradient levels-significantly improves the utilization of
pseudo-labels in SSM3OD. Our comprehensive experiments on the KITTI benchmark
demonstrate the superiority of our method over existing approaches.
|
For a simple model of chaotic dynamical systems with a large number of
degrees of freedom, we find that there is an ensemble of unstable periodic
orbits (UPOs) with the special property that the expectation values of
macroscopic quantities can be calculated using only one UPO sampled from the
ensemble. Evidence to support this conclusion is obtained by generating the
ensemble by Monte Carlo calculation for a statistical mechanical model
described by a space-time Hamiltonian that is expressed in terms of Floquet
exponents of UPOs. This result allows us to interpret the recent interesting
discovery that statistical properties of turbulence can be obtained from only
one UPO [G. Kawahara and S. Kida, J. Fluid Mech. {\bf 449}, 291 (2001); S. Kato
and M. Yamada, Phys. Rev. E {\bf 68}, 025302(R)(2003)].
|
Satellite and shipboard data reveal the intermittent vertical information
transport mechanism of turbulence and internal waves that mixes the ocean,
atmosphere, planets and stars.
|
We report evidence of Re and Mo segregation (up to 2.6 at.% and 1 at.%) along
with Cr and Co to the dislocations inside of {\gamma}' precipitates in a second
generation Ni-based single crystal superalloy, after creep deformation at
750{\deg}C under an applied stress of 800 MPa. The observed segregation effects
can be rationalized through bridging the solute partitioning behavior across
the {\gamma}/{\gamma}' interface and the pipe diffusion mechanism along the
core of the dislocation line. This understanding can provide new insights
enabling improved alloy design.
|
The dynamics of amorphous granular matter with frictional interactions cannot
be derived in general from a Hamiltonian and therefore displays oscillatory
instabilities stemming from the onset of complex eigenvalues in the stability
matrix. These instabilities were discovered in the context of one and two
dimensional systems, while the three dimensional case was never studied in
detail. Here we fill this gap by deriving and demonstrating the presence of
oscillatory instabilities in a three dimensional granular packing. We study
binary assemblies of spheres of two sizes interacting via classical Hertz and
Mindlin force laws for the longitudinal and tangent interactions, respectively.
We formulate analytically the stability matrix in 3D and observe that a couple
of complex eigenvalues emerges at the onset of the instability as in the case
of frictional disks in two-dimensions. The dynamics then shows oscillatory
exponential growth in the Mean-Square-Displacement, followed by a catastrophic
event. The generality of these results for any choice of forces that break the
symplectic Hamiltonian symmetry is discussed.
|
Estimating 3D articulated shapes like animal bodies from monocular images is
inherently challenging due to the ambiguities of camera viewpoint, pose,
texture, lighting, etc. We propose ARTIC3D, a self-supervised framework to
reconstruct per-instance 3D shapes from a sparse image collection in-the-wild.
Specifically, ARTIC3D is built upon a skeleton-based surface representation and
is further guided by 2D diffusion priors from Stable Diffusion. First, we
enhance the input images with occlusions/truncation via 2D diffusion to obtain
cleaner mask estimates and semantic features. Second, we perform
diffusion-guided 3D optimization to estimate shape and texture that are of
high-fidelity and faithful to input images. We also propose a novel technique
to calculate more stable image-level gradients via diffusion models compared to
existing alternatives. Finally, we produce realistic animations by fine-tuning
the rendered shape and texture under rigid part transformations. Extensive
evaluations on multiple existing datasets as well as newly introduced noisy web
image collections with occlusions and truncation demonstrate that ARTIC3D
outputs are more robust to noisy images, higher quality in terms of shape and
texture details, and more realistic when animated. Project page:
https://chhankyao.github.io/artic3d/
|
Despite the popularisation of machine learning models, more often than not,
they still operate as black boxes with no insight into what is happening inside
the model. There exist a few methods that allow to visualise and explain why a
model has made a certain prediction. Those methods, however, allow
visualisation of the link between the input and output of the model without
presenting how the model learns to represent the data used to train the model
as whole. In this paper, a method that addresses that issue is proposed, with a
focus on visualising multi-dimensional time-series data. Experiments on a
high-frequency stock market dataset show that the method provides fast and
discernible visualisations. Large datasets can be visualised quickly and on one
plot, which makes it easy for a user to compare the learned representations of
the data. The developed method successfully combines known techniques to
provide an insight into the inner workings of time-series classification
models.
|
An experimental technique for the light lambda-hypernuclei structure studies
by using (negative kaon, neutral pion) reaction and Neutral Meson Spectrometer
(NMS) developed at the BNL has been described. Position dependence calibration
of the BGO conversion planes of the NMS was invented as the solution to the
crucial constraint of the high resolution in single-lambda spectroscopy.
Position parameters of the BGO crystal rods were fitted out from the
out-of-kaon beam measured data as obtained by the original small highly
collimated Co-60 source method, based on the coincidences between top and
bottom signals of each BGO rod.
|
A study of the reinforcement effect of a soft polymer matrix by hard
nanometric filler particles is presented. In the main part of this article, the
structure of the silica filler in the matrix is studied by Small Angle Neutron
Scattering (SANS), and stress-strain isotherms are measured to characterize the
rheological properties of the composites. Our analysis allows us to quantify
the degree of aggregation of the silica in the matrix, which is studied as a
function of pH (4-10), silica volume fraction (3-15%) and silica bead size
(average radius 78 A and 96 A). Rheological properties of the samples are
represented in terms of the strain-dependent reinforcement factor, which
highlights the contribution of the filler. Combining the structural information
with a quantitative analysis of the reinforcement factor, the aggregate size
and compacity (10%-40%) as a function of volume fraction and pH can be deduced.
In a second, more explorative study, the grafting of polymer chains on
nanosilica beads for future reinforcement applications is followed by SANS. The
structure of the silica and the polymer are measured separately by contrast
variation, using deuterated material. The aggregation of the silica beads in
solution is found to decrease during polymerization, reaching a rather low
final aggregation number (less than ten).
|
Decision making is the cognitive process of selecting a course of action
among multiple alternatives. As the decision maker belongs to a complex
microenvironment (which contains multiple decision makers), has to make a
decision where multiple options are present which often leads to a phenomenon
known as the "paradox of choices". The latter refers to the case where too many
options can lead to negative outcomes, such as increased uncertainty, decision
paralysis, and frustration. Here, we employ an entropy driven mechanism within
a statistical physics framework to explain the premises of the paradox. In
turn, we focus on the emergence of a collective "paradox of choice", in the
case of interacting decision-making agents, quantified as the decision
synchronization time. Our findings reveal a trade-off between synchronization
time and the sensing radius, indicating the optimal conditions for information
transfer among group members, which significantly depends the individual
sensitivity parameters. Interestingly, when agents sense their microenvironment
in a biased way or their decisions are influenced by their past choices, then
the collective "paradox of choice" does not occur. In a nutshell, our theory
offers a low-dimensional and unified statistical explanation of the "paradox of
choice" at the individual and at the collective level.
|
Many AI synthesis problems such as planning or scheduling may be modelized as
constraint satisfaction problems (CSP). A CSP is typically defined as the
problem of finding any consistent labeling for a fixed set of variables
satisfying all given constraints between these variables. However, for many
real tasks such as job-shop scheduling, time-table scheduling, design?, all
these constraints have not the same significance and have not to be necessarily
satisfied. A first distinction can be made between hard constraints, which
every solution should satisfy and soft constraints, whose satisfaction has not
to be certain. In this paper, we formalize the notion of possibilistic
constraint satisfaction problems that allows the modeling of uncertainly
satisfied constraints. We use a possibility distribution over labelings to
represent respective possibilities of each labeling. Necessity-valued
constraints allow a simple expression of the respective certainty degrees of
each constraint. The main advantage of our approach is its integration in the
CSP technical framework. Most classical techniques, such as Backtracking (BT),
arcconsistency enforcing (AC) or Forward Checking have been extended to handle
possibilistics CSP and are effectively implemented. The utility of our approach
is demonstrated on a simple design problem.
|
We consider the problem of intelligent and efficient task allocation
mechanism in large-scale mobile edge computing (MEC), which can reduce delay
and energy consumption in a parallel and distributed optimization. In this
paper, we study the joint optimization model to consider cooperative task
management mechanism among mobile terminals (MT), macro cell base station
(MBS), and multiple small cell base station (SBS) for large-scale MEC
applications. We propose a parallel multi-block Alternating Direction Method of
Multipliers (ADMM) based method to model both requirements of low delay and low
energy consumption in the MEC system which formulates the task allocation under
those requirements as a nonlinear 0-1 integer programming problem. To solve the
optimization problem, we develop an efficient combination of conjugate
gradient, Newton and linear search techniques based algorithm with Logarithmic
Smoothing (for global variables updating) and the Cyclic Block coordinate
Gradient Projection (CBGP, for local variables updating) methods, which can
guarantee convergence and reduce computational complexity with a good
scalability. Numerical results demonstrate the effectiveness of the proposed
mechanism and it can effectively reduce delay and energy consumption for a
large-scale MEC system.
|
Being able to predict stock prices might be the unspoken wish of stock
investors. Although stock prices are complicated to predict, there are many
theories about what affects their movements, including interest rates, news and
social media. With the help of Machine Learning, complex patterns in data can
be identified beyond the human intellect. In this thesis, a Machine Learning
model for time series forecasting is created and tested to predict stock
prices. The model is based on a neural network with several layers of LSTM and
fully connected layers. It is trained with historical stock values, technical
indicators and Twitter attribute information retrieved, extracted and
calculated from posts on the social media platform Twitter. These attributes
are sentiment score, favourites, followers, retweets and if an account is
verified. To collect data from Twitter, Twitter's API is used. Sentiment
analysis is conducted with VADER. The results show that by adding more Twitter
attributes, the MSE between the predicted prices and the actual prices improved
by 3%. With technical analysis taken into account, MSE decreases from 0.1617 to
0.1437, which is an improvement of around 11%. The restrictions of this study
include that the selected stock has to be publicly listed on the stock market
and popular on Twitter and among individual investors. Besides, the stock
markets' opening hours differ from Twitter, which constantly available. It may
therefore introduce noises in the model.
|
Clifford's geometric algebra has enjoyed phenomenal development over the last
60 years by mathematicians, theoretical physicists, engineers and computer
scientists in robotics, artificial intelligence and data analysis, introducing
a myriad of different and often confusing notations. The geometric algebra of
Euclidean 3-space, the natural generalization of both the well-known
Gibbs-Heaviside vector algebra, and Hamilton's quaternions, is used here to
study spheroidal domains, spheroidal-graphic projections, the Laplace equation
and its Lie algebra of symmetries. The Cauchy-Kovalevska extension and the
Cauchy kernel function are treated in a unified way. The concept of a
quasi-monogenic family of functions is introduced and studied.
|
Let $p$ be a prime number, and $G$ a compact $p$-adic Lie group. We recall
that the Iwasawa algebra $\Lambda(G)$ is defined to be the completed group ring
of $G$ over the ring of $p$-adic integers. Interesting examples of finitely
generated modules over $\Lambda(G),$ in which $G$ is the image of Galois in the
automorphism group of a $p$-adic Galois representation, abound in arithmetic
geometry. The study of such $\Lambda(G)$-modules arising from arithmetic
geometry can be thought of as a natural generalization of Iwasawa theory. One
of the cornerstones of classical Iwasawa theory is the fact that, when $G$ is
the additive group of $p$-adic integers, a good structure theory for finitely
generated $\Lambda(G)$-modules is known, up to pseudo-isomorphism. The aim of
the present paper is to extend as much as possible of this commutative
structure theory to the non-commuta tive case.
|
In this article we formulate several conjectures concerning the lowest
eigenvalue of a Dirac operator with an external electrostatic potential. The
latter describes a relativistic quantum electron moving in the field of some
(pointwise or extended) nuclei. The main question we ask is whether the
eigenvalue is minimal when the nuclear charge is concentrated at one single
point. This well-known property in nonrelativistic quantum mechanics has
escaped all attempts of proof in the relativistic case.
|
We study the response to perturbation of non-Poisson dichotomous fluctuations
that generate super-diffusion. We adopt the Liouville perspective and with it a
quantum-like approach based on splitting the density distribution into a
symmetric and an anti-symmetric component. To accomodate the equilibrium
condition behind the stationary correlation function, we study the time
evolution of the anti-symmetric component, while keeping the symmetric
component at equilibrium. For any realistic form of the perturbed distribution
density we expect a breakdown of the Onsager principle, namely, of the property
that the subsequent regression of the perturbation to equilibrium is identical
to the corresponding equilibrium correlation function. We find the directions
to follow for the calculation of higher-order correlation functions, an
unsettled problem, which has been addressed in the past by means of
approximations yielding quite different physical effects.
|
The Burnside Problem asks whether a finitely generated group of exponent n is
finite.
We present a solution for 2-generator groups of prime power exponent. Results
of P. Hall and G. Higman extends the finiteness conclusion to groups having
composite exponents. Our main result, called the Generalized Burnside Theorem,
is a solvability theorem that applies to a family of groups called GB
(Generalized Burnside) groups that contain infinite as well as finite groups.
The final section discusses the extension to k-generator groups although
details are left for another time.
|
We develop Bayesian models for density regression with emphasis on discrete
outcomes. The problem of density regression is approached by considering
methods for multivariate density estimation of mixed scale variables, and
obtaining conditional densities from the multivariate ones. The approach to
multivariate mixed scale outcome density estimation that we describe represents
discrete variables, either responses or covariates, as discretised versions of
continuous latent variables. We present and compare several models for
obtaining these thresholds in the challenging context of count data analysis
where the response may be over- and/or under-dispersed in some of the regions
of the covariate space. We utilise a nonparametric mixture of multivariate
Gaussians to model the directly observed and the latent continuous variables.
The paper presents a Markov chain Monte Carlo algorithm for posterior sampling,
sufficient conditions for weak consistency, and illustrations on density, mean
and quantile regression utilizing simulated and real datasets.
|
We introduce and formalize a notion of "a priori knowledge" about a quantum
system, and show some properties about this form of knowledge. Finally, we show
that the Kochen-Specker theorem follows directly from this study. This version
is a draft version, the bibliography in particular is extremely scarce.
Comments welcome.
|
Asymmetric features in exoplanet transit light curves are often interpreted
as a gravity darkening effect especially if there is spectroscopic evidence of
a spin-orbit misalignment. Since other processes can also lead to light curve
asymmetries this may lead to inaccurate gravity darkening parameters. Here we
investigate the case of non-radial pulsations as possible sources of asymmetry
and likely source of misinterpreted parameters through simulations. We obtained
a series of simulated transit light curves of a hypothetical exoplanet-star
system: a host star with no gravity darkening exhibiting small amplitude
pulsations, and a typical hot Jupiter in a circular, edge-on orbit. A number of
scenarios of pulsations of various amplitudes were considered, and a proper
account of the obscuring effect of transits on all the surface intensity
components was made. The magnitude of amplitude and phase modulations of
nonradial pulsations during transits was also also investigated. We then fitted
both a non-gravity-darkened, and a gravity-darkened, free spin-orbit axis model
on the data. The Akaike and Bayesian Information Criteria were used for an
objective selection of the most plausible model. We then explored the
dependence of the parameter deviations on the pulsation properties, in order to
identify configurations that can lead to falsely misaligned solutions.
Low-amplitude pulsations in general do not affect the determination of the
system parameters beyond their noise nature. However, frequencies close to
multiples of the orbital frequency are found to cause distortions leading to
solutions with a side tilted stellar rotational axis, they are therefore
preferable to clean beforehand for the sake of a correct analysis.
Additionally, for cases with higher-amplitude pulsations, it is recommended to
pre-process and clean the pulsations before analysis.
|
Detection of hydroacoustic transmissions is a key enabling technology in
applications such as depth measurements, detection of objects, and undersea
mapping. To cope with the long channel delay spread and the low signal-to-noise
ratio, hydroacoustic signals are constructed with a large time-bandwidth
product, $N$. A promising detector for hydroacoustic signals is the normalized
matched filter (NMF). For the NMF, the detection threshold depends only on $N$,
thereby obviating the need to estimate the characteristics of the sea ambient
noise which are time-varying and hard to estimate. While previous works
analyzed the characteristics of the normalized matched filter (NMF), for
hydroacoustic signals with large $N$ values the expressions available are
computationally complicated to evaluate. Specifically for hydroacoustic signals
of large $N$ values, this paper presents approximations for the probability
distribution of the NMF. These approximations are found extremely accurate in
numerical simulations. We also outline a computationally efficient method to
calculate the receiver operating characteristic (ROC) which is required to
determine the detection threshold. Results from an experiment conducted in the
Mediterranean sea at depth of 900~m agree with the analysis.
|
A linearized tensor renormalization group (LTRG) algorithm is proposed to
calculate the thermodynamic properties of one-dimensional quantum lattice
models, that is incorporated with the infinite time-evolving block decimation
technique, and allows for treating directly the two-dimensional transfer-matrix
tensor network. To illustrate its feasibility, the thermodynamic quantities of
the quantum XY spin chain are calculated accurately by the LTRG, and the
precision is shown to be comparable with (even better than) the transfer matrix
renormalization group (TMRG) method. Unlike the TMRG scheme that can only deal
with the infinite chains, the present LTRG algorithm could treat both finite
and infinite systems, and may be readily extended to boson and fermion quantum
lattice models.
|
We observed the Crab pulsar in October 2008 at the Copernico Telescope in
Asiago - Cima Ekar with the optical photon counter Aqueye (the Asiago Quantum
Eye) which has the best temporal resolution and accuracy ever achieved in the
optical domain (hundreds of picoseconds). Our goal was to perform a detailed
analysis of the optical period and phase drift of the main peak of the Crab
pulsar and compare it with the Jodrell Bank ephemerides. We determined the
position of the main peak using the steepest zero of the cross-correlation
function between the pulsar signal and an accurate optical template. The pulsar
rotational period and period derivative have been measured with great accuracy
using observations covering only a 2 day time interval. The error on the period
is 1.7 ps, limited only by the statistical uncertainty. Both the rotational
frequency and its first derivative are in agreement with those from the Jodrell
Bank radio ephemerides archive. We also found evidence of the optical peak
leading the radio one by ~230 microseconds. The distribution of phase-residuals
of the whole dataset is slightly wider than that of a synthetic signal
generated as a sequence of pulses distributed in time with the probability
proportional to the pulse shape, such as the average count rate and background
level are those of the Crab pulsar observed with Aqueye. The counting
statistics and quality of the data allowed us to determine the pulsar period
and period derivative with great accuracy in 2 days only. The time of arrival
of the optical peak of the Crab pulsar leads the radio one in agreement with
what recently reported in the literature. The distribution of the phase
residuals can be approximated with a Gaussian and is consistent with being
completely caused by photon noise (for the best data sets).
|
We use an analytical forward model based on perturbation theory to predict
the neutral hydrogen (HI) overdensity maps at low redshifts. We investigate its
performance by comparing it directly at the field level to the simulated HI
from the IllustrisTNG simulation TNG300-1 ($L=205\ h^{-1}$ Mpc), in both real
and redshift space. We demonstrate that HI is a biased tracer of the underlying
matter field and find that the cubic bias model describes the simulated HI
power spectrum to within 1% up to $k=0.4 \;(0.3) \,h\,{\rm Mpc}^{-1}$ in real
(redshift) space at redshifts $z=0,1$. Looking at counts in cells, we find an
excellent agreement between the theory and simulations for cells as small as 5
$h^{-1}$ Mpc. These results are in line with expectations from perturbation
theory and they imply that a perturbative description of the HI field is
sufficiently accurate given the characteristics of upcoming 21cm intensity
mapping surveys. Additionally, we study the statistical properties of the model
error - the difference between the truth and the model. We show that on large
scales this error is nearly Gaussian and that it has a flat power spectrum,
with amplitude significantly lower than the standard noise inferred from the HI
power spectrum. We explain the origin of this discrepancy, discuss its
implications for the HI power spectrum Fisher matrix forecasts and argue that
it motivates the HI field-level cosmological inference. On small scales in
redshift space we use the difference between the model and the truth as a proxy
for the Fingers-of-God effect. This allows us to estimate the nonlinear
velocity dispersion of HI and show that it is smaller than for the typical
spectroscopic galaxy samples at the same redshift. Finally, we provide a simple
prescription based on the perturbative forward model which can be used to
efficiently generate accurate HI mock data, in real and redshift space.
|
We revisit the Hall effect and magnetoresistivity by incorporating the
anisotropic scattering caused by apical oxygen vacancies in overdoped La-based
cuprates. The theoretical calculations within the Fermi liquid picture agree
well with a handful of anomalous magneto-transport data, better than the
results using an isotropic scattering rate alone. In particular, we obtain the
upturn of Hall coefficient $R_H$ with decreasing temperature $T$, the initial
drop of $R_H$ in magnetic field $B$ in all overdoped regimes, the linear
resistivity $\rho$ versus $B$ near the van Hove doping level, the temperature
dependence of the magnetoresistivity ratio, and the violation of Kohler's law.
These results suggest that many of the anomalous transport behaviors in
overdoped La$_{2-x}$Sr$_x$CuO$_4$ could actually be understood within the Fermi
liquid picture.
|
Let $H$ be a subgroup of $\pi_{1}(X,x_{0})$. In this paper, we extend the
concept of $X$ being SLT space to $H$-SLT space at $x_0$. First, we show that
the fibers of the endpoint projection $p_{H}:\tilde{X}_{H}\rightarrow X$ are
topological group when $X$ is $H$-SLT space at $x_0$ and $H$ is a normal
subgroup. Also, we show that under these conditions the concepts of
homotopically path Hausdorff relative to $H$ and homotopically Hausdorff
relative to $H$ coincide. Moreover, among other things, we show that the
endpoint projection map $p_{H}$ has the unique path lifting property if and
only if $H$ is a closed normal subgroup of $\pi_{1}^{qtop}(X,x_{0})$ when $X$
is SLT at $x_{0}$. Second, we present conditions under which the whisker
topology is agree with the quotient of compact-open topology on
$\tilde{X}_{H}$. Also, we study the relationship between open subsets of
$\pi_{1}^{wh}(X,x_{0})$ and $\pi_{1}^{qtop}(X,x_{0})$.
|
The non-classical, nonmonotonic inference relation associated with the answer
set semantics for logic programs gives rise to a relationship of 'strong
equivalence' between logical programs that can be verified in 3-valued Goedel
logic, G3, the strongest non-classical intermediate propositional logic
(Lifschitz, Pearce and Valverde, 2001). In this paper we will show that KC (the
logic obtained by adding axiom ~A v ~~A to intuitionistic logic), is the
weakest intermediate logic for which strongly equivalent logic programs, in a
language allowing negations, are logically equivalent.
|
We describe symmetries of the braid monodromy decomposition for a class of
plane curves defined over reals including the real curves with no real points
and proving new divisibility relations for Alexander invariants of such curves.
|
We study feedback control over erasure channels with packet-dropouts. To
achieve robustness with respect to packet-dropouts, the controller transmits
data packets containing plant input predictions, which minimize a finite
horizon cost function. To reduce the data size of packets, we propose to adopt
sparsity-promoting optimizations, namely, ell-1-ell-2 and ell-2-constrained
ell-0 optimizations, for which efficient algorithms exist. We derive sufficient
conditions on design parameters, which guarantee (practical) stability of the
resulting feedback control systems when the number of consecutive
packet-dropouts is bounded.
|
Suppose a finite dimensional semisimple Lie algebra $\mathfrak g$ acts by
derivations on a finite dimensional associative or Lie algebra $A$ over a field
of characteristic $0$. We prove the $\mathfrak g$-invariant analogs of
Wedderburn - Mal'cev and Levi theorems, and the analog of Amitsur's conjecture
on asymptotic behavior for codimensions of polynomial identities with
derivations of $A$. It turns out that for associative algebras the differential
PI-exponent coincides with the ordinary one. Also we prove the analog of
Amitsur's conjecture for finite dimensional associative algebras with an action
of a reductive affine algebraic group by automorphisms and anti-automorphisms
or graded by an arbitrary Abelian group. In addition, we provide criteria for
$G$-, $H$- and graded simplicity in terms of codimensions.
|
This is the third part in a series of papers in which we introduce and
develop a natural, general tensor category theory for suitable module
categories for a vertex (operator) algebra. In this paper (Part III), we
introduce and study intertwining maps and tensor product bifunctors.
|
In a recent work arXiv:2008.08601, Halverson, Maiti and Stoner proposed a
description of neural networks in terms of a Wilsonian effective field theory.
The infinite-width limit is mapped to a free field theory, while finite $N$
corrections are taken into account by interactions (non-Gaussian terms in the
action). In this paper, we study two related aspects of this correspondence.
First, we comment on the concepts of locality and power-counting in this
context. Indeed, these usual space-time notions may not hold for neural
networks (since inputs can be arbitrary), however, the renormalization group
provides natural notions of locality and scaling. Moreover, we comment on
several subtleties, for example, that data components may not have a
permutation symmetry: in that case, we argue that random tensor field theories
could provide a natural generalization. Second, we improve the perturbative
Wilsonian renormalization from arXiv:2008.08601 by providing an analysis in
terms of the nonperturbative renormalization group using the Wetterich-Morris
equation. An important difference with usual nonperturbative RG analysis is
that only the effective (IR) 2-point function is known, which requires setting
the problem with care. Our aim is to provide a useful formalism to investigate
neural networks behavior beyond the large-width limit (i.e.~far from Gaussian
limit) in a nonperturbative fashion. A major result of our analysis is that
changing the standard deviation of the neural network weight distribution can
be interpreted as a renormalization flow in the space of networks. We focus on
translations invariant kernels and provide preliminary numerical results.
|
In this work, we report a lattice calculation of $x$-dependent valence pion
generalized parton distributions (GPDs) at zero skewness with multiple values
of the momentum transfer $-t$. The calculations are based on an $N_f=2+1$ gauge
ensemble of highly improved staggered quarks with Wilson-Clover valence
fermion. The lattice spacing is 0.04 fm, and the pion valence mass is tuned to
be 300 MeV. We determine the Lorentz-invariant amplitudes of the quasi-GPD
matrix elements for both symmetric and asymmetric momenta transfers with
similar values and show the equivalence of both frames. Then, focusing on the
asymmetric frame, we utilize a hybrid scheme to renormalize the quasi-GPD
matrix elements obtained from the lattice calculations. After the Fourier
transforms, the quasi-GPDs are then matched to the light-cone GPDs within the
framework of large momentum effective theory with improved matching, including
the next-to-next-to-leading order perturbative corrections, and leading
renormalon and renormalization group resummations. We also present the
3-dimensional image of the pion in impact-parameter space through the Fourier
transform of the momentum transfer $-t$.
|
We show a detailed investigation of the split Kondo effect in a carbon
nanotube quantum dot with multiple gate electrodes. It is found that the
splitting decreases for increasing magnetic field, to result in a recovered
zero-bias Kondo resonance at finite magnetic field. Surprisingly, in the same
charge state, but under different gate-configurations, the splitting does not
disappear for any value of the magnetic field, but we observe an avoided
crossing of two high-conductance lines. We think that our observations can be
understood in terms of a two-impurity Kondo effect with two spins coupled
antiferromagnetically. The exchange coupling between the two spins can be
influenced by a local gate, and the non-recovery of the Kondo resonance for
certain gate configurations is explained by the existence of a small
antisymmetric contribution to the exchange interaction between the two spins.
|
The world is going through a challenging phase due to the disastrous effect
caused by the COVID-19 pandemic on the healthcare system and the economy. The
rate of spreading, post-COVID-19 symptoms, and the occurrence of new strands of
COVID-19 have put the healthcare systems in disruption across the globe. Due to
this, the task of accurately screening COVID-19 cases has become of utmost
priority. Since the virus infects the respiratory system, Chest X-Ray is an
imaging modality that is adopted extensively for the initial screening. We have
performed a comprehensive study that uses CXR images to identify COVID-19 cases
and realized the necessity of having a more generalizable model. We utilize
MobileNetV2 architecture as the feature extractor and integrate it into Capsule
Networks to construct a fully automated and lightweight model termed as
MobileCaps. MobileCaps is trained and evaluated on the publicly available
dataset with the model ensembling and Bayesian optimization strategies to
efficiently classify CXR images of patients with COVID-19 from non-COVID-19
pneumonia and healthy cases. The proposed model is further evaluated on two
additional RT-PCR confirmed datasets to demonstrate the generalizability. We
also introduce MobileCaps-S and leverage it for performing severity assessment
of CXR images of COVID-19 based on the Radiographic Assessment of Lung Edema
(RALE) scoring technique. Our classification model achieved an overall recall
of 91.60, 94.60, 92.20, and a precision of 98.50, 88.21, 92.62 for COVID-19,
non-COVID-19 pneumonia, and healthy cases, respectively. Further, the severity
assessment model attained an R$^2$ coefficient of 70.51. Owing to the fact that
the proposed models have fewer trainable parameters than the state-of-the-art
models reported in the literature, we believe our models will go a long way in
aiding healthcare systems in the battle against the pandemic.
|
Virtual sensing techniques have gained traction in applications to the
structural health monitoring of monopile-based offshore wind turbines, as the
strain response below the mudline, which is a primary indicator of fatigue
damage accumulation, is impractical to measure directly with physical
instrumentation. The Gaussian process latent force model (GPFLM) is a
generalized Bayesian virtual sensing technique which combines a physics-driven
model of the structure with a data-driven model of latent variables of the
system to extrapolate unmeasured strain states. In the GPLFM, modeling of
unknown sources of excitation as a Gaussian process (GP) serves to facilitate
strain estimation by providing a complete stochastic characterization of the
covariance relationship between input forces and states, using properties of
the GP covariance kernel as well as correlation information supplied by the
mechanical model. It is shown that posterior inference of the latent inputs and
states is performed by Gaussian process regression of measured accelerations,
computed efficiently using Kalman filtering and Rauch-Tung-Striebel smoothing
in an augmented state-space model. While the GPLFM has been previously
demonstrated in numerical studies to improve upon other virtual sensing
techniques in terms of accuracy, robustness, and numerical stability, this work
provides one of the first cases of in-situ validation of the GPLFM. The
predicted strain response by the GPLFM is compared to subsoil strain data
collected from an operating offshore wind turbine in the Westermeerwind Park in
the Netherlands.
|
We discuss a simple model for D-brane transport in non-abelian GLSMs. The
model is the elliptic curve version of a non-abelian GLSM introduced by Hori
and Tong and has gauge group U(2). It has two geometric phases, both of which
describe the same elliptic curve, once realised as a codimension five complete
intersection in G(2,5) and once as a determinantal variety. The determinantal
phase is strongly coupled with unbroken SU(2). There are two singular points in
the moduli space where the theory has a Coulomb branch. Using grade restriction
rules, we show how to transport B-branes between the two phases along paths
avoiding the singular points. With the help of the GLSM hemisphere partition
function we compute analytic continuation matrices and monodromy matrices,
confirming results obtained by different methods.
|
We classify all one-class genera of admissible lattice chains of length at
least 2 in hermitian spaces over number fields.
|
The ongoing integration of renewable generation and distributed energy
resources introduces new challenges to distribution network operation. Due to
the increasing volatility and uncertainty, distribution system operators (DSOs)
are seeking concepts to enable more active management and control. Flexibility
markets (FMs) offer a platform for economically efficient trading of
electricity flexibility between DSOs and other participants. The integration of
cyber, physical and market domains of multiple participants makes FMs a system
of cyber-physical systems (CPSs). While cross-domain integration sets the
foundation for efficient deployment of flexibility, it introduces new physical
and cyber vulnerabilities to participants. This work systematically formulates
threat scenarios for the CPSs of FMs, revealing several remaining security
challenges across all domains. Based on the threat scenarios, unresolved
monitoring requirements for secure participation of DSOs in FMs are identified,
providing the basis for future works that address these gaps with new technical
concepts.
|
Subsequent to our recent report of SDW type transition at 190 K and
antiferromagnetic order below 20 K in EuFe2As2, we have studied the effect of
K-doping on the SDW transition at high temperature and AF order at low
temperature. 50% K doping suppresses the SDW transition and in turn gives rise
to high-temperature superconductivity below T_c = 32 K, as observed in the
electrical resistivity, AC susceptibility as well as magnetization. A well
defined anomaly in the specific heat provides additional evidence for bulk
superconductivity.
|
This paper is a continuation of the work on unbounded Toeplitz-like operators
$T_\Om$ with rational matrix symbol $\Om$ initiated in Groenewald et. al
(Complex Anal. Oper. Theory 15, 1(2021)), where a Wiener-Hopf type
factorization of $\Om$ is obtained and used to determine when $T_\Om$ is
Fredholm and compute the Fredholm index in case $T_\Om$ is Fredholm. Due to the
high level of non-uniqueness and complicated form of the Wiener-Hopf type
factorization, it does not appear useful in determining when $T_\Om$ is
invertible. In the present paper we use state space methods to characterize
invertibility of $T_\Om$ in terms of the existence of a stabilizing solution of
an associated nonsymmetric discrete algebraic Riccati equation, which in turn
leads to a pseudo-canonical factorization of $\Om$ and concrete formulas of
$T_\Om^{-1}$.
|
The role of disorder on physical systems has been widely studied in the
macroscopic and microscopic world. While static disorder is well understood in
many cases, the impact of time-dependent disorder on quantum gases is still
poorly investigated. In our experimental setup, we introduce and characterize a
method capable of producing time-controlled optical-speckle disorder.
Experimentally, coherent light illuminates a combination of a static and a
rotating diffuser, thereby collecting a spatially varying phase due to the
diffusers' structure and a temporally variable phase due to the relative
rotation. Controlling the rotation of the diffuser allows changing the speckle
realization or, for future work, the characteristic time scale of the change of
the speckle pattern, i.e. the correlation time, matching typical time scales of
the quantum gases investigated. We characterize the speckle pattern ex-situ by
measuring its intensity distribution cross-correlating different intensity
patterns. In-situ, we observe its impact on a molecular Bose-Einstein
condensate (BEC) and cross-correlate the density distributions of BECs probed
in different speckle realizations. As one diffuser rotates relative to the
other around the common optical axis, we trace the optical speckle's intensity
cross-correlations and the quantum gas' density cross-correlations. Our results
show comparable outcomes for both measurement methods. The setup allows us to
tune the disorder potential adapted to the characteristics of the quantum gas.
These studies pave the way for investigating nonequilibrium physics in
interacting quantum gases using controlled dynamical-disorder potentials.
|
Emerging storage systems with new flash exhibit ultra-low latency (ULL) that
can address performance disparities between DRAM and conventional solid state
drives (SSDs) in the memory hierarchy. Considering the advanced low-latency
characteristics, different types of I/O completion methods (polling/hybrid) and
storage stack architecture (SPDK) are proposed. While these new techniques are
expected to take costly software interventions off the critical path in
ULL-applied systems, unfortunately no study exists to quantitatively analyze
system-level characteristics and challenges of combining such newly-introduced
techniques with real ULL SSDs. In this work, we comprehensively perform
empirical evaluations with 800GB ULL SSD prototypes and characterize ULL
behaviors by considering a wide range of I/O path parameters, such as different
queues and access patterns. We then analyze the efficiencies and challenges of
the polled-mode and hybrid polling I/O completion methods (added into Linux
kernels 4.4 and 4.10, respectively) and compare them with the efficiencies of a
conventional interrupt-based I/O path. In addition, we revisit the common
expectations of SPDK by examining all the system resources and parameters.
Finally, we demonstrate the challenges of ULL SSDs in a real SPDK-enabled
server-client system. Based on the performance behaviors that this study
uncovers, we also discuss several system implications, which are required to
take a full advantage of ULL SSD in the future.
|
We compute the scattering amplitudes of four massless states for chiral
(closed) bosonic and type II superstrings using the Kawai-Lewellen-Tye ($KLT$)
factorization method. The amplitude in the chiral bosonic case is identical to
a field theory amplitude corresponding to the spin-$2$ tachyon, massless
gravitational sector and massive spin-2 tardyon states of the spectrum. Chiral
type II superstrings amplitude only possess poles associated with the massless
gravitational sector. We briefly discuss the extension of the calculation to
heterotic superstrings.
|
Given a graph $G$, a total labeling on $G$ is called edge-antimagic total
(respectively, vertex-antimagic total) if all edge-weights (respectively,
vertex-weights) are pairwise distinct. If a labeling on $G$ is simultaneously
edge-antimagic total and vertex-antimagic total, it is called a totally
antimagic total labeling. A graph that admits totally antimagic total labeling
is called a totally antimagic total graph. In this paper, we prove that
ladders, prisms and generalised Pertersen graphs are totally antimagic total
graphs. We also show that the chain graph of totally antimagic total graphs is
a totally antimagic total graph.
|
This paper outlines a comprehensive model to increase system efficiency,
preserve network bandwidth, monitor incoming and outgoing packets, ensure the
security of confidential files and reduce power wastage in an organization.
This model illustrates the use and potential application of a Network Analysis
Tool (NAT) in a multi-computer set-up of any scale. The model is designed to
run in the background and not hamper any currently executing applications,
while using minimum system resources. It was developed as open source software,
using VB. Net, with a view to overcoming limitations of legacy systems and
financial restrictions in small-to mid-level organizations like businesses and
educational institutes. It is fully-customizable and serves as a simple and
open-source alternative to existing software. The NAT relies on simple
client-server architecture and uses remote access to monitor and maintain the
computers on a network, for example logging off a user or shutting down a
computer after a certain "idle" time, enabling and disabling applications,
troubleshooting and so on. The NAT was tested in a laboratory and resultant
data is presented, along with the results of a survey that was conducted among
users.
|
In this paper we study the behavior of the solution to the dbar-Neumann
problem for (0,1)-forms on a bi-disc in C^2. We show singularities which arise
at the distinguished boundary are of logarithmic and arctangent type.
|
An impact parameter dependent unintegrated gluon distribution is constructed
as a solution of a nonlinear evolution equation with realistic Glauber--Gribov
input. Photon--jet correlations in the proton fragmentation region of pA
collisions are proposed as a direct probe of the nuclear unintegrated glue.
|
This paper is about alerting acoustic event detection and sound source
localisation in an urban scenario. Specifically, we are interested in spotting
the presence of horns, and sirens of emergency vehicles. In order to obtain a
reliable system able to operate robustly despite the presence of traffic noise,
which can be copious, unstructured and unpredictable, we propose to treat the
spectrograms of incoming stereo signals as images, and apply semantic
segmentation, based on a Unet architecture, to extract the target sound from
the background noise. In a multi-task learning scheme, together with signal
denoising, we perform acoustic event classification to identify the nature of
the alerting sound. Lastly, we use the denoised signals to localise the
acoustic source on the horizon plane, by regressing the direction of arrival of
the sound through a CNN architecture. Our experimental evaluation shows an
average classification rate of 94%, and a median absolute error on the
localisation of 7.5{\deg} when operating on audio frames of 0.5s, and of
2.5{\deg} when operating on frames of 2.5s. The system offers excellent
performance in particularly challenging scenarios, where the noise level is
remarkably high.
|
Humans frequently make decisions with the aid of artificially intelligent
(AI) systems. A common pattern is for the AI to recommend an action to the
human who retains control over the final decision. Researchers have identified
ensuring that a human has appropriate reliance on an AI as a critical component
of achieving complementary performance. We argue that the current definition of
appropriate reliance used in such research lacks formal statistical grounding
and can lead to contradictions. We propose a formal definition of reliance,
based on statistical decision theory, which separates the concepts of reliance
as the probability the decision-maker follows the AI's recommendation from
challenges a human may face in differentiating the signals and forming accurate
beliefs about the situation. Our definition gives rise to a framework that can
be used to guide the design and interpretation of studies on human-AI
complementarity and reliance. Using recent AI-advised decision making studies
from literature, we demonstrate how our framework can be used to separate the
loss due to mis-reliance from the loss due to not accurately differentiating
the signals. We evaluate these losses by comparing to a baseline and a
benchmark for complementary performance defined by the expected payoff achieved
by a rational decision-maker facing the same decision task as the behavioral
decision-makers.
|
In this paper, we define Tribonacci and Tribonacci-Lucas matrix sequences and
investigate their properties.
|
In Part V of this study, we presented an original Lagrangian approach for
computing the dynamic characteristics along stationary rays, by solving the
linear, second-order Jacobi differential equation, considering four sets of
initial conditions as the basic solutions. We then focused on the computation
of the geometric spreading and identification of caustics, where only the two
point-source basic solutions with their corresponding initial conditions are
required. Solutions of the Jacobi equation represent the normal shift vectors
of the paraxial rays and define the geometry of the ray tube with respect to
the stationary central ray. Rather than the Lagrangian approach, the dynamic
characteristics are traditionally computed with the Hamiltonian approach,
formulated normally in terms of two first-order differential equations, where
the solution variables are the paraxial shifts and paraxial slowness changes
along the ray. In this part (Part VI), we compare and relate the two
approaches. We first combine the two first-order Hamiltonian dynamic equations,
eliminating the paraxial variations of the slowness vector. This leads to a
second-order differential equation in terms of the Hamiltonian shift alone,
whose ray-normal counterpart coincides with the normal shift of the
corresponding Lagrangian solution, while the ray-tangent component does not
affect the Jacobian and the geometric spreading. Comparing the proposed
Lagrangian approach to the dynamic ray tracing with the "classical" Hamiltonian
approach, we demonstrate that they are fully compatible for general anisotropy.
We then derive the two-way relationships between the Hamiltonian's and
Lagrangian's Hessians, which are the core computational elements of dynamic ray
theory. Finally, we demonstrate the relationships between these two types of
the Hessians numerically, for a general triclinic medium.
|
We discuss (i) the evaluation of the expectation values of four-quark
operators assuming that the heavy quark expansion for $b$ sector converges at
the third order in $1/m_Q$, and (ii) the estimation of the duality breaking
short distance nonperturbative corrections to the parton decay rate. We finally
point out the implications of the result obtained for the assumption of
quark-hadron duality in heavy quark expansion.
|
Composite Higgs models with a fermionic UV completion can contain additional
colored states beside the usual top-partners. We focus here on a model which
contains in addition SU(3) color octet top partners as well as color singlet
ones. The latter can in principle serve as a dark matter candidate. We consider
a particular composite Higgs model which contains SU(3) color octet top
partners besides the usually considered triplet representations. Moreover,
color singlet top partners are present as well which can in principle serve as
dark matter candidates. We investigate the LHC phenomenology of these unusual
top partners. Some of these states could at first glance be confused with
gluinos predicted in supersymmetric models.
|
We investigate whether sophisticated volatility estimation improves the
out-of-sample performance of mean-variance portfolio strategies relative to the
naive 1/N strategy. The portfolio strategies rely solely upon second moments.
Using a diverse group of econometric and portfolio models across multiple
datasets, most models achieve higher Sharpe ratios and lower portfolio
volatility that are statistically and economically significant relative to the
naive rule, even after controlling for turnover costs. Our results suggest
benefits to employing more sophisticated econometric models than the sample
covariance matrix, and that mean-variance strategies often outperform the naive
portfolio across multiple datasets and assessment criteria.
|
The nonperturbative approach to soft high-energy hadron-hadron scattering,
based on the analytic continuation of Euclidean Wilson-loop correlation
functions, makes possible the investigation of the problem of the asymptotic
energy dependence of hadron-hadron total cross sections by means of lattice
calculations. In this contribution we compare the lattice numerical results to
analytic results obtained with various nonperturbative techniques. We also
discuss the possibility to obtain indications of the rise of hadron-hadron
total cross sections with energy directly from the lattice data.
|
Understanding the similarity of the numerous released large language models
(LLMs) has many uses, e.g., simplifying model selection, detecting illegal
model reuse, and advancing our understanding of what makes LLMs perform well.
In this work, we measure the similarity of representations of a set of LLMs
with 7B parameters. Our results suggest that some LLMs are substantially
different from others. We identify challenges of using representational
similarity measures that suggest the need of careful study of similarity scores
to avoid false conclusions.
|
In this paper, we consider the transmit design for multi-input multi-output
(MIMO) wiretap channel including a malicious jammer. We first transform the
system model into the traditional three-node wiretap channel by whitening the
interference at the legitimate user. Additionally, the eavesdropper channel
state information (ECSI) may be fully or statistically known, even unknown to
the transmitter. Hence, some strategies are proposed in terms of different
levels of ECSI available to the transmitter in our paper. For the case of
unknown ECSI, a target rate for the legitimate user is first specified. And
then an inverse water-filling algorithm is put forward to find the optimal
power allocation for each information symbol, with a stepwise search being used
to adjust the spatial dimension allocated to artificial noise (AN) such that
the target rate is achievable. As for the case of statistical ECSI, several
simulated channels are randomly generated according to the distribution of
ECSI. We show that the ergodic secrecy capacity can be approximated as the
average secrecy capacity of these simulated channels. Through maximizing this
average secrecy capacity, we can obtain a feasible power and spatial dimension
allocation scheme by using one dimension search. Finally, numerical results
reveal the effectiveness and computational efficiency of our algorithms.
|
In this paper, we study the embedded feature selection problem in linear
Support Vector Machines (SVMs), in which a cardinality constraint is employed,
leading to a fully explainable selection model. The problem is NP-hard due to
the presence of the cardinality constraint, even though the original linear SVM
amounts to a problem solvable in polynomial time. To handle the hard problem,
we first introduce two mixed-integer formulations for which novel SDP
relaxations are proposed. Exploiting the sparsity pattern of the relaxations,
we decompose the problems and obtain equivalent relaxations in a much smaller
cone, making the conic approaches scalable. To make the best usage of the
decomposed relaxations, we propose heuristics using the information of its
optimal solution. Moreover, an exact procedure is proposed by solving a
sequence of mixed-integer decomposed SDPs. Numerical results on classical
benchmarking datasets are reported, showing the efficiency and effectiveness of
our approach.
|
Social influence cannot be identified from purely observational data on
social networks, because such influence is generically confounded with latent
homophily, i.e., with a node's network partners being informative about the
node's attributes and therefore its behavior. If the network grows according to
either a latent community (stochastic block) model, or a continuous latent
space model, then latent homophilous attributes can be consistently estimated
from the global pattern of social ties. We show that, for common versions of
those two network models, these estimates are so informative that controlling
for estimated attributes allows for asymptotically unbiased and consistent
estimation of social-influence effects in linear models. In particular, the
bias shrinks at a rate which directly reflects how much information the network
provides about the latent attributes. These are the first results on the
consistent non-experimental estimation of social-influence effects in the
presence of latent homophily, and we discuss the prospects for generalizing
them.
|
With the rise of Deep Learning (DL), our world braces for AI in every edge
device, creating an urgent need for edge-AI SoCs. This SoC hardware needs to
support high throughput, reliable and secure AI processing at Ultra Low Power
(ULP), with a very short time to market. With its strong legacy in edge
solutions and open processing platforms, the EU is well-positioned to become a
leader in this SoC market. However, this requires AI edge processing to become
at least 100 times more energy-efficient, while offering sufficient flexibility
and scalability to deal with AI as a fast-moving target. Since the design space
of these complex SoCs is huge, advanced tooling is needed to make their design
tractable. The CONVOLVE project (currently in Inital stage) addresses these
roadblocks. It takes a holistic approach with innovations at all levels of the
design hierarchy. Starting with an overview of SOTA DL processing support and
our project methodology, this paper presents 8 important design choices largely
impacting the energy efficiency and flexibility of DL hardware. Finding good
solutions is key to making smart-edge computing a reality.
|
The abundance of H2 in molecular clouds, relative to the commonly used tracer
CO, has only been measured toward a few embedded stars, which may be surrounded
by atypical gas. We present observations of near-infrared absorption by H2, CO,
and dust toward stars behind molecular clouds, providing a representative
sample of these molecules in cold molecular gas, primarily in the Taurus
Molecular Cloud. We find N_H2/A_V ~ 1.0x10^21 cm^-2, N_CO/A_V ~ 1.5x10^17 cm^-2
(1.8x10^17 including solid CO), and N_H2/N_CO ~ 6000. The measured N_H2/N_CO
ratio is consistent with that toward embedded stars in various molecular
clouds, but both are less than that derived from mm-wave observations of CO and
star counts. The difference apparently results from the higher directly
measured N_CO/A_V ratio.
|
We consider homogeneous STIT tessellations Y in the \ell-dimensional
Euclidean space and show the triviality of the tail \sigma-algebra. This is a
sharpening of the mixing result by Lachi\`eze-Rey.
|
We study projections in the bidual of a $C^*$-algebra $B$ that are null with
respect to a subalgebra $A$, that is projections $p\in B^{**}$ satisfying
$|\phi|(p)=0$ for every $\phi\in B^*$ annihilating $A$. In the separable case,
$A$-null projections are precisely the peak projections in the bidual of $A$ at
which the subalgebra $A$ interpolates the entire $C^*$-algebra $B$. These are
analogues of null sets in classical function theory, on which several profound
results rely. This motivates the development of a noncommutative variant, which
we use to find appropriate `quantized' versions of some of these classical
facts. Through a delicate generalization of a theorem of Varopoulos, we show
that, roughly speaking, sufficiently regular interpolation projections are null
precisely when their atomic parts are. As an application, we give alternative
proofs and sharpenings of some recent peak-interpolation results of Davidson
and Hartz for algebras on Hilbert function spaces, also illuminating thereby
how earlier noncommutative peak-interpolation theory may be applied. In another
direction, given a convex subset of the state space of $B$, we characterize
when the associated Riesz projection is null. This is then applied to various
important topics in noncommutative function theory, such as the F.& M. Riesz
property, the existence of Lebesgue decompositions, the description of Henkin
functionals, and Arveson's noncommutative Hardy spaces (maximal subdiagonal
algebras).
|
The distribution of transversely polarized quarks inside a transversely
polarized nucleon, known as transversity, encodes a basic piece of information
on the nucleon structure, sharing the same status with the more familiar
unpolarized and helicity distributions. I will review its properties and
discuss different ways to access it, with highlights and limitations. Recent
phenomenological extractions and perspectives are also presented.
|
Analog and digital quantum simulators can efficiently simulate quantum
many-body systems that appear in natural phenomena. However, experimental
limitations of near-term devices still make it challenging to perform the
entire process of quantum simulation. The purification-based quantum simulation
methods can alleviate the limitations in experiments such as the cooling
temperature and noise from the environment, while this method has the drawback
that it requires global entangled measurement with a prohibitively large number
of measurements that scales exponentially with the system size. In this Letter,
we propose that we can overcome these problems by restricting the entangled
measurements to the vicinity of the local observables to be measured, when the
locality of the system can be exploited. We provide theoretical guarantees that
the global purification operation can be replaced with local operations under
some conditions, in particular for the task of cooling and error mitigation. We
furthermore give a numerical verification that the localized purification is
valid even when conditions are not satisfied. Our method bridges the
fundamental concept of locality with quantum simulators, and therefore expected
to open a path to unexplored quantum many-body phenomena.
|
With the advances in customized hardware for quantum annealing and
digital/CMOS Annealing, Quadratic Unconstrained Binary Optimization (QUBO)
models have received growing attention in the optimization literature.
Motivated by an existing general-purpose approach that derives QUBO models from
binary linear programs (BLP), we propose a novel Multilevel Constraint
Transformation Scheme (MLCTS) that derives QUBO models with fewer ancillary
binary variables. We formulate sufficient conditions for the existence of a
compact QUBO formulation (i.e., in the original BLP decision space) in terms of
constraint levelness and demonstrate the flexibility and applicability of MLCTS
on synthetic examples and several well-known combinatorial optimization
problems, i.e., the Maximum 2-Satisfiability Problem, the Linear Ordering
Problem, the Community Detection Problem, and the Maximum Independence Set
Problem. For a proof-of-concept, we compare the performance of two QUBO models
for the latter problem on both a general-purpose software-based solver and a
hardware-based QUBO solver. The MLCTS-derived models demonstrate significantly
better performance for both solvers, in particular, solving up to seven times
more instances with the hardware-based approach.
|
We measure the rest-frame B-band luminosity function of red-sequence galaxies
(RSLF) of five intermediate-redshift (0.5 < z < 0.9), high-mass (sigma > 950
km/s) clusters. Cluster galaxies are identified through photometric redshifts
based on imaging in seven bands (five broad, and two narrow) using the WIYN
3.5m telescope. The luminosity functions are well-fit down to M_B^*+3 for all
of the clusters out to a radius of R_200. For comparison, the luminosity
functions for a sample of 59 low redshift clusters selected from the SDSS are
measured as well. There is a brightening trend (M_B^* increases by 0.7 mags by
z=0.75) with redshift comparable to what is seen in the field for similarly
defined galaxies, although there is a hint that the cluster red-sequence
brightening is more rapid in the past (z>0.5), and relatively shallow at more
recent times. Contrary to other claims, we find little evidence for evolution
of the faint end slope. Previous indications of evolution may be due to
limitations in measurement technique, bias in the sample selection, and cluster
to cluster variation. As seen in both the low and high redshift sample, a
significant amount of variation in luminosity functions parameters alpha and
M^* exists between individual clusters.
|
For the innovation of spintronic technologies, Dirac materials, in which the
low-energy excitation is described as relativistic Dirac fermions, are one of
the most promising systems, because of the fascinating magnetotransport
associated with the extremely high mobility. To incorporate Dirac fermions into
spintronic applications, their quantum transport phenomena are desired to be
manipulated to a large extent by magnetic order in a solid. We here report a
bulk half-integer quantum Hall effect in a layered antiferromagnet EuMnBi$_2$,
in which field-controllable Eu magnetic order significantly suppresses the
interlayer coupling between the Bi layers with Dirac fermions. In addition to
the high mobility more than 10,000 cm$^2$/Vs, Landau level splittings
presumably due to the lifting of spin and valley degeneracy are noticeable even
in a bulk magnet. These results will pave a route to the engineering of
magnetically functionalized Dirac materials.
|
This paper presents a framework that supports the implementation of parallel
solutions for the widespread parametric maximum flow computational routines
used in image segmentation algorithms. The framework is based on supergraphs, a
special construction combining several image graphs into a larger one, and
works on various architectures (multi-core or GPU), either locally or remotely
in a cluster of computing nodes. The framework can also be used for performance
evaluation of parallel implementations of maximum flow algorithms. We present
the case study of a state-of-the-art image segmentation algorithm based on
graph cuts, Constrained Parametric Min-Cut (CPMC), that uses the parallel
framework to solve parametric maximum flow problems, based on a GPU
implementation of the well-known push-relabel algorithm. Our results indicate
that real-time implementations based on the proposed techniques are possible.
|
We present a method of generation of exact and explicit forms of one-sided,
heavy-tailed Levy stable probability distributions g_{\alpha}(x), 0 \leq x <
\infty, 0 < \alpha < 1. We demonstrate that the knowledge of one such a
distribution g_{\alpha}(x) suffices to obtain exactly g_{\alpha^{p}}(x), p=2,
3,... Similarly, from known g_{\alpha}(x) and g_{\beta}(x), 0 < \alpha, \beta <
1, we obtain g_{\alpha \beta}(x). The method is based on the construction of
the integral operator, called Levy transform, which implements the above
operations. For \alpha rational, \alpha = l/k with l < k, we reproduce in this
manner many of the recently obtained exact results for g_{l/k}(x). This
approach can be also recast as an application of the Efros theorem for
generalized Laplace convolutions. It relies solely on efficient definite
integration.
|
A theorem of Graber, Harris, and Starr states that a rationally connected
fibration over a curve has a section. We study an analogous question in
symplectic geometry. Namely, given a rationally connected fibration over a
curve, can one find a section which gives a non-zero Gromov-Witten invariant?
We observe that for any fibration, the existence of a section which gives a
non-zero Gromov-Witten invariant only depends on the generic fiber, i.e. a
variety defined over the function field of a curve. Some examples of rationally
connected fibrations with this property are given, including all rational
surface fibrations. We also prove some results, which says that in certain
cases we can "lift" Gromov-Witten invariants of the base to the total space of
a rationally connected fibration.
|
We investigate hybrid structures based on a bilayer quantum spin Hall system
in proximity to an s-wave superconductor as a platform to mimic time-reversal
symmetric topological superconductors. In this bilayer setup, the induced
pairing can be of intra- or inter-layer type, and domain walls of those
different types of pairing potentials host Kramers partners (time-reversal
conjugate pairs) of Majorana bound states. Interestingly, we discover that such
topological interfaces providing Majorana bound states can also be achieved in
an otherwise homogeneous system by a spatially dependent inter-layer gate
voltage. This gate voltage causes the relative electron densities of the two
layers to vary accordingly which suppresses the inter-layer pairing in regions
with strong gate voltage. We identify particular transport signatures
(zero-bias anomalies) in a five-terminal setup that are uniquely related to the
presence of Kramers pairs of Majorana bound states.
|
Subsets and Splits