Dataset Viewer
Auto-converted to Parquet
id
stringlengths
9
14
submitter
stringlengths
1
64
authors
stringlengths
4
9.62k
title
stringlengths
4
343
comments
stringlengths
1
609
journal-ref
stringlengths
4
404
doi
stringlengths
12
153
report-no
stringlengths
2
254
categories
stringlengths
5
112
license
stringclasses
9 values
orig_abstract
stringlengths
14
3.76k
versions
listlengths
1
60
update_date
stringlengths
10
10
authors_parsed
sequencelengths
1
535
abstract
stringlengths
11
3.75k
1802.00554
Andrew Lensen
Andrew Lensen, Bing Xue, and Mengjie Zhang
Generating Redundant Features with Unsupervised Multi-Tree Genetic Programming
16 pages, preprint for EuroGP '18
null
10.1007/978-3-319-77553-1_6
null
cs.NE cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Recently, feature selection has become an increasingly important area of research due to the surge in high-dimensional datasets in all areas of modern life. A plethora of feature selection algorithms have been proposed, but it is difficult to truly analyse the quality of a given algorithm. Ideally, an algorithm would be evaluated by measuring how well it removes known bad features. Acquiring datasets with such features is inherently difficult, and so a common technique is to add synthetic bad features to an existing dataset. While adding noisy features is an easy task, it is very difficult to automatically add complex, redundant features. This work proposes one of the first approaches to generating redundant features, using a novel genetic programming approach. Initial experiments show that our proposed method can automatically create difficult, redundant features which have the potential to be used for creating high-quality feature selection benchmark datasets.
[ { "created": "Fri, 2 Feb 2018 04:19:04 GMT", "version": "v1" }, { "created": "Tue, 20 Mar 2018 06:35:56 GMT", "version": "v2" } ]
2019-10-24
[ [ "Lensen", "Andrew", "" ], [ "Xue", "Bing", "" ], [ "Zhang", "Mengjie", "" ] ]
Recently, feature selection has become an increasingly important area of research due to the surge in high-dimensional datasets in all areas of modern life. A plethora of feature selection algorithms have been proposed, but it is difficult to truly analyse the quality of a given algorithm. Ideally, an algorithm would be evaluated by measuring how well it removes known bad features. Acquiring datasets with such features is inherently difficult, and so a common technique is to add synthetic bad features to an existing dataset. While adding noisy features is an easy task, it is very difficult to automatically add complex, redundant features. This work proposes one of the first approaches to generating redundant features, using a novel genetic programming approach. Initial experiments show that our proposed method can automatically create difficult, redundant features which have the potential to be used for creating high-quality feature selection benchmark datasets.
hep-th/0209192
Mark A. Stern
Mark A. Stern
Quantum Mechanical Mirror Symmetry, D Branes, and B Fields
22 pages
null
null
Duke-CGTP-02-08
hep-th
null
We construct quantum mechanical models which mimic many features of string theory. We use these models to gain improved descriptions of B fields and gerbes. We examine analogs of T duality, D branes, and mirror symmetry and derive quantum mechanical analogs of standard phenomena, such as the noncommutative geometry induced by a B field.
[ { "created": "Mon, 23 Sep 2002 19:31:32 GMT", "version": "v1" } ]
2007-05-23
[ [ "Stern", "Mark A.", "" ] ]
We construct quantum mechanical models which mimic many features of string theory. We use these models to gain improved descriptions of B fields and gerbes. We examine analogs of T duality, D branes, and mirror symmetry and derive quantum mechanical analogs of standard phenomena, such as the noncommutative geometry induced by a B field.
2007.04028
Amartya Sanyal
Amartya Sanyal, Puneet K Dokania, Varun Kanade, Philip H.S. Torr
How benign is benign overfitting?
null
null
null
null
cs.LG stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We investigate two causes for adversarial vulnerability in deep neural networks: bad data and (poorly) trained models. When trained with SGD, deep neural networks essentially achieve zero training error, even in the presence of label noise, while also exhibiting good generalization on natural test data, something referred to as benign overfitting [2, 10]. However, these models are vulnerable to adversarial attacks. We identify label noise as one of the causes for adversarial vulnerability, and provide theoretical and empirical evidence in support of this. Surprisingly, we find several instances of label noise in datasets such as MNIST and CIFAR, and that robustly trained models incur training error on some of these, i.e. they don't fit the noise. However, removing noisy labels alone does not suffice to achieve adversarial robustness. Standard training procedures bias neural networks towards learning "simple" classification boundaries, which may be less robust than more complex ones. We observe that adversarial training does produce more complex decision boundaries. We conjecture that in part the need for complex decision boundaries arises from sub-optimal representation learning. By means of simple toy examples, we show theoretically how the choice of representation can drastically affect adversarial robustness.
[ { "created": "Wed, 8 Jul 2020 11:07:10 GMT", "version": "v1" } ]
2020-07-09
[ [ "Sanyal", "Amartya", "" ], [ "Dokania", "Puneet K", "" ], [ "Kanade", "Varun", "" ], [ "Torr", "Philip H. S.", "" ] ]
We investigate two causes for adversarial vulnerability in deep neural networks: bad data and (poorly) trained models. When trained with SGD, deep neural networks essentially achieve zero training error, even in the presence of label noise, while also exhibiting good generalization on natural test data, something referred to as benign overfitting [2, 10]. However, these models are vulnerable to adversarial attacks. We identify label noise as one of the causes for adversarial vulnerability, and provide theoretical and empirical evidence in support of this. Surprisingly, we find several instances of label noise in datasets such as MNIST and CIFAR, and that robustly trained models incur training error on some of these, i.e. they don't fit the noise. However, removing noisy labels alone does not suffice to achieve adversarial robustness. Standard training procedures bias neural networks towards learning "simple" classification boundaries, which may be less robust than more complex ones. We observe that adversarial training does produce more complex decision boundaries. We conjecture that in part the need for complex decision boundaries arises from sub-optimal representation learning. By means of simple toy examples, we show theoretically how the choice of representation can drastically affect adversarial robustness.
hep-th/9303053
null
Alan R. White
Analytic Multi-Regge Theory and the Pomeron in QCD : II. Gauge Theory Analysis
149 pages
Int.J.Mod.Phys.A8:4755-4896,1993
10.1142/S0217751X93001910
ANL-HEP-PR-93-16
hep-th hep-ph
null
The high-energy Regge behavior of gauge theories is studied via the formalism of Analytic Multi-Regge Theory. Perturbative results for spontaneously-broken theories are first organised into reggeon diagrams. Unbroken gauge theories are studied via a reggeon diagram infra-red analysis of symmetry restoration. Massless fermions play a crucial role and the case of QCD involves the Super-Critical Pomeron as an essential intermediate stage. An introductory review of the build up of transverse momentum diagrams and reggeon diagrams from leading log calculations in gauge theories is presented first. It is then shown that the results closely reproduce the general structure for multi-regge amplitudes derived in Part I of the article, allowing the construction of general reggeon diagrams for spontaneously-broken theories. Next it is argued that, with a transverse-momentum cut-off, unbroken gauge theories can be reached through an infra-red limiting process which successively decouples fundamental representation Higgs fields . The first infra-red limit studied is the restoration of SU(2) gauge symmetry. The analysis is dominated by the exponentiation of divergences imposed by Reggeon Unitarity and the contribution of massless quarks ...
[ { "created": "Tue, 9 Mar 1993 14:33:25 GMT", "version": "v1" }, { "created": "Wed, 10 Mar 1993 10:12:48 GMT", "version": "v2" } ]
2014-11-18
[ [ "White", "Alan R.", "" ] ]
The high-energy Regge behavior of gauge theories is studied via the formalism of Analytic Multi-Regge Theory. Perturbative results for spontaneously-broken theories are first organised into reggeon diagrams. Unbroken gauge theories are studied via a reggeon diagram infra-red analysis of symmetry restoration. Massless fermions play a crucial role and the case of QCD involves the Super-Critical Pomeron as an essential intermediate stage. An introductory review of the build up of transverse momentum diagrams and reggeon diagrams from leading log calculations in gauge theories is presented first. It is then shown that the results closely reproduce the general structure for multi-regge amplitudes derived in Part I of the article, allowing the construction of general reggeon diagrams for spontaneously-broken theories. Next it is argued that, with a transverse-momentum cut-off, unbroken gauge theories can be reached through an infra-red limiting process which successively decouples fundamental representation Higgs fields . The first infra-red limit studied is the restoration of SU(2) gauge symmetry. The analysis is dominated by the exponentiation of divergences imposed by Reggeon Unitarity and the contribution of massless quarks ...
hep-th/9404121
Miguel Navarro
M. Navarro, J. Guerrero and V. Aldaya
Optics, Mechanics and Quantization of Reparametrization Systems
15 pages, Latex
J.Math.Phys. 35 (1994) 6407-6417
10.1063/1.530682
null
hep-th
null
In this paper we regard the dynamics obtained from Fermat principle as begin the classical theory of light. We (first-)quantize the action and show how close we can get to the Maxwell theory. We show that Quantum Geometric Optics is not a theory of fields in curved space. Considering Classical Mechanics to be on the same footing, we show the parallelism between Quantum Mechanics and Quantum Geometric Optics. We show that, due to the reparametrization invariance of the classical theories, the dynamics of the quantum theories is given by a Hamiltonian constraint. Some implications of the above analogy in the quantization of true reparameterization invariant systems are discussed.
[ { "created": "Wed, 20 Apr 1994 09:26:47 GMT", "version": "v1" } ]
2009-10-28
[ [ "Navarro", "M.", "" ], [ "Guerrero", "J.", "" ], [ "Aldaya", "V.", "" ] ]
In this paper we regard the dynamics obtained from Fermat principle as begin the classical theory of light. We (first-)quantize the action and show how close we can get to the Maxwell theory. We show that Quantum Geometric Optics is not a theory of fields in curved space. Considering Classical Mechanics to be on the same footing, we show the parallelism between Quantum Mechanics and Quantum Geometric Optics. We show that, due to the reparametrization invariance of the classical theories, the dynamics of the quantum theories is given by a Hamiltonian constraint. Some implications of the above analogy in the quantization of true reparameterization invariant systems are discussed.
1804.05661
Thameur Dhieb
Thameur Dhieb, Sourour Njah, Houcine Boubaker, Wael Ouarda, Mounir Ben Ayed, and Adel M. Alimi
An Extended Beta-Elliptic Model and Fuzzy Elementary Perceptual Codes for Online Multilingual Writer Identification using Deep Neural Network
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Actually, the ability to identify the documents authors provides more chances for using these documents for various purposes. In this paper, we present a new effective biometric writer identification system from online handwriting. The system consists of the preprocessing and the segmentation of online handwriting into a sequence of Beta strokes in a first step. Then, from each stroke, we extract a set of static and dynamic features from new proposed model that we called Extended Beta-Elliptic model and from the Fuzzy Elementary Perceptual Codes. Next, all the segments which are composed of N consecutive strokes are categorized into groups and subgroups according to their position and their geometric characteristics. Finally, Deep Neural Network is used as classifier. Experimental results reveal that the proposed system achieves interesting results as compared to those of the existing writer identification systems on Latin and Arabic scripts.
[ { "created": "Mon, 16 Apr 2018 13:27:11 GMT", "version": "v1" }, { "created": "Sun, 20 May 2018 16:10:40 GMT", "version": "v2" }, { "created": "Wed, 30 May 2018 19:14:28 GMT", "version": "v3" }, { "created": "Sat, 10 Nov 2018 11:28:36 GMT", "version": "v4" } ]
2018-11-13
[ [ "Dhieb", "Thameur", "" ], [ "Njah", "Sourour", "" ], [ "Boubaker", "Houcine", "" ], [ "Ouarda", "Wael", "" ], [ "Ayed", "Mounir Ben", "" ], [ "Alimi", "Adel M.", "" ] ]
Actually, the ability to identify the documents authors provides more chances for using these documents for various purposes. In this paper, we present a new effective biometric writer identification system from online handwriting. The system consists of the preprocessing and the segmentation of online handwriting into a sequence of Beta strokes in a first step. Then, from each stroke, we extract a set of static and dynamic features from new proposed model that we called Extended Beta-Elliptic model and from the Fuzzy Elementary Perceptual Codes. Next, all the segments which are composed of N consecutive strokes are categorized into groups and subgroups according to their position and their geometric characteristics. Finally, Deep Neural Network is used as classifier. Experimental results reveal that the proposed system achieves interesting results as compared to those of the existing writer identification systems on Latin and Arabic scripts.
1810.01758
Kaveh Dehghanpour
Qianzhi Zhang and Kaveh Dehghanpour and Zhaoyu Wang and Qiuhua Huang
A Learning-based Power Management for Networked Microgrids Under Incomplete Information
null
null
10.1109/TSG.2019.2933502
null
cs.SY
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper presents an approximate Reinforcement Learning (RL) methodology for bi-level power management of networked Microgrids (MG) in electric distribution systems. In practice, the cooperative agent can have limited or no knowledge of the MG asset behavior and detailed models behind the Point of Common Coupling (PCC). This makes the distribution systems unobservable and impedes conventional optimization solutions for the constrained MG power management problem. To tackle this challenge, we have proposed a bi-level RL framework in a price-based environment. At the higher level, a cooperative agent performs function approximation to predict the behavior of entities under incomplete information of MG parametric models; while at the lower level, each MG provides power-flow-constrained optimal response to price signals. The function approximation scheme is then used within an adaptive RL framework to optimize the price signal as the system load and solar generation change over time. Numerical experiments have verified that, compared to previous works in the literature, the proposed privacy-preserving learning model has better adaptability and enhanced computational speed.
[ { "created": "Mon, 1 Oct 2018 20:38:50 GMT", "version": "v1" }, { "created": "Tue, 5 Mar 2019 16:06:44 GMT", "version": "v2" }, { "created": "Fri, 14 Jun 2019 00:21:24 GMT", "version": "v3" } ]
2019-08-09
[ [ "Zhang", "Qianzhi", "" ], [ "Dehghanpour", "Kaveh", "" ], [ "Wang", "Zhaoyu", "" ], [ "Huang", "Qiuhua", "" ] ]
This paper presents an approximate Reinforcement Learning (RL) methodology for bi-level power management of networked Microgrids (MG) in electric distribution systems. In practice, the cooperative agent can have limited or no knowledge of the MG asset behavior and detailed models behind the Point of Common Coupling (PCC). This makes the distribution systems unobservable and impedes conventional optimization solutions for the constrained MG power management problem. To tackle this challenge, we have proposed a bi-level RL framework in a price-based environment. At the higher level, a cooperative agent performs function approximation to predict the behavior of entities under incomplete information of MG parametric models; while at the lower level, each MG provides power-flow-constrained optimal response to price signals. The function approximation scheme is then used within an adaptive RL framework to optimize the price signal as the system load and solar generation change over time. Numerical experiments have verified that, compared to previous works in the literature, the proposed privacy-preserving learning model has better adaptability and enhanced computational speed.
0905.3602
Fainan Hanif
Muhammad Fainan Hanif and Peter J. Smith
Level Crossing Rates of Interference in Cognitive Radio Networks
submitted to the IEEE Transactions on Wireless Communications
IEEE Transactions on Wireless Communications, vol.9, no.4, pp.1283-1287, 2010
10.1109/TWC.2010.04.090749
null
cs.IT math.IT
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The future deployment of cognitive radios is critically dependent on the fact that the incumbent primary user system must remain as oblivious as possible to their presence. This in turn heavily relies on the fluctuations of the interfering cognitive radio signals. In this letter we compute the level crossing rates of the cumulative interference created by the cognitive radios. We derive analytical formulae for the level crossing rates in Rayleigh and Rician fast fading conditions. We approximate Rayleigh and Rician level crossing rates using fluctuation rates of gamma and scaled noncentral $\chi^2$ processes respectively. The analytical results and the approximations used in their derivations are verified by Monte Carlo simulations and the analysis is applied to a particular CR allocation strategy.
[ { "created": "Fri, 22 May 2009 04:06:43 GMT", "version": "v1" } ]
2013-01-03
[ [ "Hanif", "Muhammad Fainan", "" ], [ "Smith", "Peter J.", "" ] ]
The future deployment of cognitive radios is critically dependent on the fact that the incumbent primary user system must remain as oblivious as possible to their presence. This in turn heavily relies on the fluctuations of the interfering cognitive radio signals. In this letter we compute the level crossing rates of the cumulative interference created by the cognitive radios. We derive analytical formulae for the level crossing rates in Rayleigh and Rician fast fading conditions. We approximate Rayleigh and Rician level crossing rates using fluctuation rates of gamma and scaled noncentral $\chi^2$ processes respectively. The analytical results and the approximations used in their derivations are verified by Monte Carlo simulations and the analysis is applied to a particular CR allocation strategy.
1308.2277
Simon Childs
S.J. Childs
An Improved Temporal Formulation of Pupal Transpiration in Glossina
33 pages, 27 figures, 3 tables. arXiv admin note: text overlap with arXiv:0901.2470
Mathematical Biosciences, 262: 214-229, 2015
null
null
q-bio.OT
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The temporal aspect of a model of pupal dehydration is improved upon. The observed dependence of pupal transpiration on time is attributed to an alternation between two, essential modes, for which the deposition of a thin, pupal skin inside the puparium and its subsequent demise are thought to be responsible. For each mode of transpiration, the results of the Bursell (1958) investigation into pupal dehydration are used as a rudimentary data set. These data are generalised to all temperatures and humidities by invoking the property of multiplicative separability. The problem, then, is that as the temperature varies with time, so does the metabolism and the developmental stages to which the model data pertain, must necessarily warp. The puparial-duration formula of Phelps and Burrows (1969) and Hargrove (2004) is exploited to facilitate a mapping between the constant-temperature time domain of the data and that of some, more general case at hand. The resulting, Glossina morsitans model is extrapolated to other species using their relative surface areas, their relative protected and unprotected transpiration rates and their different fourth instar excretions (drawing, to a lesser extent, from the data of Buxton and Lewis, 1934). In this way the problem of pupal dehydration is formulated as a series of integrals and the consequent survival can be predicted. The discovery of a distinct definition for hygrophilic species, within the formulation, prompts the investigation of the hypothetical effect of a two-day heat wave on pupae. This leads to the conclusion that the classification of species as hygrophilic, mesophilic and xerophilic is largely true only in so much as their third and fourth instars are and, possibly, the hours shortly before eclosion.
[ { "created": "Sat, 10 Aug 2013 05:14:22 GMT", "version": "v1" }, { "created": "Fri, 9 May 2014 19:11:14 GMT", "version": "v2" }, { "created": "Tue, 19 May 2015 14:59:27 GMT", "version": "v3" } ]
2015-05-20
[ [ "Childs", "S. J.", "" ] ]
The temporal aspect of a model of pupal dehydration is improved upon. The observed dependence of pupal transpiration on time is attributed to an alternation between two, essential modes, for which the deposition of a thin, pupal skin inside the puparium and its subsequent demise are thought to be responsible. For each mode of transpiration, the results of the Bursell (1958) investigation into pupal dehydration are used as a rudimentary data set. These data are generalised to all temperatures and humidities by invoking the property of multiplicative separability. The problem, then, is that as the temperature varies with time, so does the metabolism and the developmental stages to which the model data pertain, must necessarily warp. The puparial-duration formula of Phelps and Burrows (1969) and Hargrove (2004) is exploited to facilitate a mapping between the constant-temperature time domain of the data and that of some, more general case at hand. The resulting, Glossina morsitans model is extrapolated to other species using their relative surface areas, their relative protected and unprotected transpiration rates and their different fourth instar excretions (drawing, to a lesser extent, from the data of Buxton and Lewis, 1934). In this way the problem of pupal dehydration is formulated as a series of integrals and the consequent survival can be predicted. The discovery of a distinct definition for hygrophilic species, within the formulation, prompts the investigation of the hypothetical effect of a two-day heat wave on pupae. This leads to the conclusion that the classification of species as hygrophilic, mesophilic and xerophilic is largely true only in so much as their third and fourth instars are and, possibly, the hours shortly before eclosion.
2308.09443
Veronique Bruyere
Thomas Brihaye and V\'eronique Bruy\`ere and Gaspard Reghem
Quantitative Reachability Stackelberg-Pareto Synthesis is NEXPTIME-Complete
null
null
null
null
cs.GT
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we deepen the study of two-player Stackelberg games played on graphs in which Player $0$ announces a strategy and Player $1$, having several objectives, responds rationally by following plays providing him Pareto-optimal payoffs given the strategy of Player $0$. The Stackelberg-Pareto synthesis problem, asking whether Player $0$ can announce a strategy which satisfies his objective, whatever the rational response of Player $1$, has been recently investigated for $\omega$-regular objectives. We solve this problem for weighted graph games and quantitative reachability objectives such that Player $0$ wants to reach his target set with a total cost less than some given upper bound. We show that it is NEXPTIME-complete, as for Boolean reachability objectives.
[ { "created": "Fri, 18 Aug 2023 10:17:18 GMT", "version": "v1" } ]
2023-08-21
[ [ "Brihaye", "Thomas", "" ], [ "Bruyère", "Véronique", "" ], [ "Reghem", "Gaspard", "" ] ]
In this paper, we deepen the study of two-player Stackelberg games played on graphs in which Player $0$ announces a strategy and Player $1$, having several objectives, responds rationally by following plays providing him Pareto-optimal payoffs given the strategy of Player $0$. The Stackelberg-Pareto synthesis problem, asking whether Player $0$ can announce a strategy which satisfies his objective, whatever the rational response of Player $1$, has been recently investigated for $\omega$-regular objectives. We solve this problem for weighted graph games and quantitative reachability objectives such that Player $0$ wants to reach his target set with a total cost less than some given upper bound. We show that it is NEXPTIME-complete, as for Boolean reachability objectives.
2012.02628
Jonathan Cohen
Jonathan D. Cohen
A Mitigation Score for COVID-19
15 pages, 12 figures
null
null
null
q-bio.OT
http://creativecommons.org/licenses/by/4.0/
This note describes a simple score to indicate the effectiveness of mitigation against infections of COVID-19 as observed by new case counts. The score includes normalization, making comparisons across jurisdictions possible. The smoothing employed provides robustness in the face of reporting vagaries while retaining salient features of evolution, enabling a clearer picture for decision makers and the public.
[ { "created": "Wed, 2 Dec 2020 21:25:50 GMT", "version": "v1" } ]
2020-12-07
[ [ "Cohen", "Jonathan D.", "" ] ]
This note describes a simple score to indicate the effectiveness of mitigation against infections of COVID-19 as observed by new case counts. The score includes normalization, making comparisons across jurisdictions possible. The smoothing employed provides robustness in the face of reporting vagaries while retaining salient features of evolution, enabling a clearer picture for decision makers and the public.
1504.04496
Jose Edelstein
Xian O. Camanho, Jose D. Edelstein, Andres Gomberoff, J. Anibal Sierra-Garcia
On AdS to dS transitions in higher-curvature gravity
12 pages, 3 figures; v2: comments and references added
null
null
null
hep-th gr-qc
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We study the possible existence of gravitational phase transitions from AdS to dS geometries in the context of higher-curvature gravities. We use Lanczos-Gauss-Bonnet (LGB) theory with a positive cosmological constant as a toy model. This theory has two maximally symmetric vacua with positive (dS) and negative (AdS) constant curvature. We show that a phase transition from the AdS vacuum to a dS black hole geometry takes place when the temperature reaches a critical value. The transition is produced by nucleation of bubbles of the new phase that expand afterwards. We claim that this phenomenon is not particular to the model under study, and shall also be part of generic gravitational theories with higher-curvature terms.
[ { "created": "Fri, 17 Apr 2015 12:58:57 GMT", "version": "v1" }, { "created": "Mon, 27 Jul 2015 00:04:57 GMT", "version": "v2" } ]
2015-07-28
[ [ "Camanho", "Xian O.", "" ], [ "Edelstein", "Jose D.", "" ], [ "Gomberoff", "Andres", "" ], [ "Sierra-Garcia", "J. Anibal", "" ] ]
We study the possible existence of gravitational phase transitions from AdS to dS geometries in the context of higher-curvature gravities. We use Lanczos-Gauss-Bonnet (LGB) theory with a positive cosmological constant as a toy model. This theory has two maximally symmetric vacua with positive (dS) and negative (AdS) constant curvature. We show that a phase transition from the AdS vacuum to a dS black hole geometry takes place when the temperature reaches a critical value. The transition is produced by nucleation of bubbles of the new phase that expand afterwards. We claim that this phenomenon is not particular to the model under study, and shall also be part of generic gravitational theories with higher-curvature terms.
hep-th/9505126
Matthias Heyssler
Matthias Heyssler (Department of Physics, Durham) Alex C. Kalloniatis (Max-Planck-Institut f\"ur Kernphysik,Heidelberg)
Costituent Quark Picture out of QCD in two dimensions - on the Light-Cone
13 pages, uses elsart.sty 2 Postscript figures, uses epsf.sty 'elsart.sty' and 'elsart12.sty' are available via anonymous-ftp at ftp://ftp.tex.ac.uk/tex-archive/macros/latex/contrib/supported/elsevier
Phys.Lett. B354 (1995) 453-459
10.1016/0370-2693(95)00654-4
MPIH-V25-1994
hep-th hep-ph
null
Using DLCQ as a nonperturbative method, we test Fock-space truncations in ${\rm QCD}_{1+1}$ by studying the mass spectra of hadrons in colour SU(2) and SU(3) at finite harmonic resolution $K$. We include $q\bar q q\bar q$ states for mesons and up to $qqq q\bar q$ states for baryons. With this truncation, we give `predictions' for the masses of the first five states where finite $K$ effects are minimal.
[ { "created": "Sat, 20 May 1995 01:54:21 GMT", "version": "v1" } ]
2009-10-28
[ [ "Heyssler", "Matthias", "", "Department of Physics, Durham" ], [ "Kalloniatis", "Alex C.", "", "Max-Planck-Institut für Kernphysik,Heidelberg" ] ]
Using DLCQ as a nonperturbative method, we test Fock-space truncations in ${\rm QCD}_{1+1}$ by studying the mass spectra of hadrons in colour SU(2) and SU(3) at finite harmonic resolution $K$. We include $q\bar q q\bar q$ states for mesons and up to $qqq q\bar q$ states for baryons. With this truncation, we give `predictions' for the masses of the first five states where finite $K$ effects are minimal.
1701.02294
Marc P. Bellon
Marc P. Bellon
Alien Calculus and non perturbative effects in Quantum Field Theory
4 pages, double-column
Front. Phys. (2016) 11: 113201
10.1007/s11467-016-0580-7
null
hep-th hep-ph
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In many domains of physics, methods are needed to deal with non-perturbative aspects. I want here to argue that a good approach is to work on the Borel transforms of the quantities of interest, the singularities of which give non-perturbative contributions. These singularities in many cases can be largely determined by using the alien calculus developed by Jean \'Ecalle. My main example will be the two point function of a massless theory given as a solution of a renormalization group equation.
[ { "created": "Mon, 9 Jan 2017 18:33:58 GMT", "version": "v1" } ]
2017-01-10
[ [ "Bellon", "Marc P.", "" ] ]
In many domains of physics, methods are needed to deal with non-perturbative aspects. I want here to argue that a good approach is to work on the Borel transforms of the quantities of interest, the singularities of which give non-perturbative contributions. These singularities in many cases can be largely determined by using the alien calculus developed by Jean \'Ecalle. My main example will be the two point function of a massless theory given as a solution of a renormalization group equation.
hep-th/9804204
H. W. Braden
H. W. Braden, V. Varela
The Bogomolny Equations and Solutions for Einstein-Yang-Mills-Dilaton- $\sigma$ Models
24 pages LaTex, 1 Figure, revised text for publication
Phys.Rev.D58:124020,1998
10.1103/PhysRevD.58.124020
MS-98-006
hep-th gr-qc
null
We derive Bogomolny equations for an Einstein-Yang-Mills-dilaton-$\sigma$ model (EYMD-$\sigma$) on a static spacetime, showing that the Einstein equations are satisfied if and only if the associated (conformally scaled) three-metric is flat. These are precisely the static metrics for which super-covariantly constant spinors exist. We study some general properties of these equations and then consider the problem of obtaining axially symmetric solutions for the gauge group SU(2).
[ { "created": "Thu, 30 Apr 1998 16:27:45 GMT", "version": "v1" }, { "created": "Fri, 11 Sep 1998 14:16:21 GMT", "version": "v2" } ]
2008-11-26
[ [ "Braden", "H. W.", "" ], [ "Varela", "V.", "" ] ]
We derive Bogomolny equations for an Einstein-Yang-Mills-dilaton-$\sigma$ model (EYMD-$\sigma$) on a static spacetime, showing that the Einstein equations are satisfied if and only if the associated (conformally scaled) three-metric is flat. These are precisely the static metrics for which super-covariantly constant spinors exist. We study some general properties of these equations and then consider the problem of obtaining axially symmetric solutions for the gauge group SU(2).
hep-th/0408134
Michael Walker
M.L.Walker
Three point SUSY Ward identities without Ghosts
20 pages, no figures, typos corrected
JHEP0412:011,2004
10.1088/1126-6708/2004/12/011
null
hep-th
null
We utilise a non-local gauge transform which renders the entire action of SUSY QED invariant and respects the SUSY algebra modulo the gauge-fixing condition, to derive two- and three-point ghost-free SUSY Ward identities in SUSY QED. We use the cluster decomposition principle to find the Green's function Ward identities and then takes linear combinations of the latter to derive identities for the proper functions.
[ { "created": "Wed, 18 Aug 2004 12:30:50 GMT", "version": "v1" }, { "created": "Fri, 10 Sep 2004 01:29:41 GMT", "version": "v2" } ]
2008-11-26
[ [ "Walker", "M. L.", "" ] ]
We utilise a non-local gauge transform which renders the entire action of SUSY QED invariant and respects the SUSY algebra modulo the gauge-fixing condition, to derive two- and three-point ghost-free SUSY Ward identities in SUSY QED. We use the cluster decomposition principle to find the Green's function Ward identities and then takes linear combinations of the latter to derive identities for the proper functions.
1506.09215
Simon Lacoste-Julien
Jean-Baptiste Alayrac, Piotr Bojanowski, Nishant Agrawal, Josef Sivic, Ivan Laptev, Simon Lacoste-Julien
Unsupervised Learning from Narrated Instruction Videos
Appears in: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2016). 21 pages
null
null
null
cs.CV cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We address the problem of automatically learning the main steps to complete a certain task, such as changing a car tire, from a set of narrated instruction videos. The contributions of this paper are three-fold. First, we develop a new unsupervised learning approach that takes advantage of the complementary nature of the input video and the associated narration. The method solves two clustering problems, one in text and one in video, applied one after each other and linked by joint constraints to obtain a single coherent sequence of steps in both modalities. Second, we collect and annotate a new challenging dataset of real-world instruction videos from the Internet. The dataset contains about 800,000 frames for five different tasks that include complex interactions between people and objects, and are captured in a variety of indoor and outdoor settings. Third, we experimentally demonstrate that the proposed method can automatically discover, in an unsupervised manner, the main steps to achieve the task and locate the steps in the input videos.
[ { "created": "Tue, 30 Jun 2015 19:55:37 GMT", "version": "v1" }, { "created": "Thu, 2 Jul 2015 16:43:36 GMT", "version": "v2" }, { "created": "Mon, 30 Nov 2015 18:10:53 GMT", "version": "v3" }, { "created": "Tue, 28 Jun 2016 18:43:37 GMT", "version": "v4" } ]
2016-06-29
[ [ "Alayrac", "Jean-Baptiste", "" ], [ "Bojanowski", "Piotr", "" ], [ "Agrawal", "Nishant", "" ], [ "Sivic", "Josef", "" ], [ "Laptev", "Ivan", "" ], [ "Lacoste-Julien", "Simon", "" ] ]
We address the problem of automatically learning the main steps to complete a certain task, such as changing a car tire, from a set of narrated instruction videos. The contributions of this paper are three-fold. First, we develop a new unsupervised learning approach that takes advantage of the complementary nature of the input video and the associated narration. The method solves two clustering problems, one in text and one in video, applied one after each other and linked by joint constraints to obtain a single coherent sequence of steps in both modalities. Second, we collect and annotate a new challenging dataset of real-world instruction videos from the Internet. The dataset contains about 800,000 frames for five different tasks that include complex interactions between people and objects, and are captured in a variety of indoor and outdoor settings. Third, we experimentally demonstrate that the proposed method can automatically discover, in an unsupervised manner, the main steps to achieve the task and locate the steps in the input videos.
1802.01089
Eduard Paul Enoiu
Raluca Marinescu, Predrag Filipovikj, Eduard Paul Enoiu, Jonatan Larsson, Cristina Seceleanu
An Energy-aware Mutation Testing Framework for EAST-ADL Architectural Models
Version submitted to the 29th Nordic Workshop on Programming Theory
null
null
null
cs.SE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Early design artifacts of embedded systems, such as architectural models, represent convenient abstractions for reasoning about a system's structure and functionality. One such example is the Electronic Architecture and Software Tools-Architecture Description Language (EAST-ADL), a domain-specific architectural language that targets the automotive industry. EAST-ADL is used to represent both hardware and software elements, as well as related extra-functional information (e.g., timing properties, triggering information, resource consumption). Testing architectural models is an important activity in engineering large-scale industrial systems, which sparks a growing research interest. The main contributions of this paper are: (i) an approach for creating energy-related mutants for EAST-ADL architectural models, (ii) a method for overcoming the equivalent mutant problem (i.e., the problem of finding a test case which can distinguish the observable behavior of a mutant from the original one), (iii) a test generation approach based on UPPAAL Statistical Model Checker (SMC), and (iv) a test selection criteria based on mutation analysis using our MATS tool.
[ { "created": "Sun, 4 Feb 2018 08:21:21 GMT", "version": "v1" } ]
2018-02-06
[ [ "Marinescu", "Raluca", "" ], [ "Filipovikj", "Predrag", "" ], [ "Enoiu", "Eduard Paul", "" ], [ "Larsson", "Jonatan", "" ], [ "Seceleanu", "Cristina", "" ] ]
Early design artifacts of embedded systems, such as architectural models, represent convenient abstractions for reasoning about a system's structure and functionality. One such example is the Electronic Architecture and Software Tools-Architecture Description Language (EAST-ADL), a domain-specific architectural language that targets the automotive industry. EAST-ADL is used to represent both hardware and software elements, as well as related extra-functional information (e.g., timing properties, triggering information, resource consumption). Testing architectural models is an important activity in engineering large-scale industrial systems, which sparks a growing research interest. The main contributions of this paper are: (i) an approach for creating energy-related mutants for EAST-ADL architectural models, (ii) a method for overcoming the equivalent mutant problem (i.e., the problem of finding a test case which can distinguish the observable behavior of a mutant from the original one), (iii) a test generation approach based on UPPAAL Statistical Model Checker (SMC), and (iv) a test selection criteria based on mutation analysis using our MATS tool.
hep-th/9208072
Francois Gieres
Francois Gieres and Stefan Theisen
Superconformally covariant operators and super W algebras
23 pages, LATEX, MPI-Ph/92-66 and KA-THEP-7/92
J.Math.Phys. 34 (1993) 5964-5985
10.1063/1.530243
null
hep-th
null
We study superdifferential operators of order $2n+1$ which are covariant with respect to superconformal changes of coordinates on a compact super Riemann surface. We show that all such operators arise from super M\"obius covariant ones. A canonical matrix representation is presented and applications to classical super W algebras are discussed.
[ { "created": "Thu, 27 Aug 1992 21:49:16 GMT", "version": "v1" } ]
2009-10-22
[ [ "Gieres", "Francois", "" ], [ "Theisen", "Stefan", "" ] ]
We study superdifferential operators of order $2n+1$ which are covariant with respect to superconformal changes of coordinates on a compact super Riemann surface. We show that all such operators arise from super M\"obius covariant ones. A canonical matrix representation is presented and applications to classical super W algebras are discussed.
hep-th/9508075
Fosco Cesar Daniel
D. G. Barcy, C. D. Fosco and L. E. Oxman
On bosonization in $3$ dimensions
11 pages, Latex, omitted references added, typos corrected
Phys.Lett. B375 (1996) 267-272
10.1016/0370-2693(96)00224-9
null
hep-th cond-mat
null
A recently proposed path-integral bosonization scheme for massive fermions in $3$ dimensions is extended by keeping the full momentum-dependence of the one-loop vacuum polarization tensor. This makes it possible to discuss both the massive and massless fermion cases on an equal footing, and moreover the results it yields for massless fermions are consistent with the ones of another, seemingly different, canonical quantization approach to the problem of bosonization for a massless fermionic field in $3$ dimensions.
[ { "created": "Wed, 16 Aug 1995 14:33:30 GMT", "version": "v1" }, { "created": "Tue, 22 Aug 1995 15:56:03 GMT", "version": "v2" } ]
2009-10-28
[ [ "Barcy", "D. G.", "" ], [ "Fosco", "C. D.", "" ], [ "Oxman", "L. E.", "" ] ]
A recently proposed path-integral bosonization scheme for massive fermions in $3$ dimensions is extended by keeping the full momentum-dependence of the one-loop vacuum polarization tensor. This makes it possible to discuss both the massive and massless fermion cases on an equal footing, and moreover the results it yields for massless fermions are consistent with the ones of another, seemingly different, canonical quantization approach to the problem of bosonization for a massless fermionic field in $3$ dimensions.
1210.2255
Yu-tin Huang
Yu-tin Huang and Henrik Johansson
Equivalent D=3 Supergravity Amplitudes from Double Copies of Three-Algebra and Two-Algebra Gauge Theories
5 pages, published version in PRL
null
10.1103/PhysRevLett.110.171601
CERN-PH-TH/2012-254, MCTP-12-22, Saclay--IPhT--T12/076
hep-th
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We show that three-dimensional supergravity amplitudes can be obtained as double copies of either three-algebra super-Chern-Simons matter theory or that of two-algebra super-Yang-Mills theory, when either theory is organized to display the color-kinematics duality. We prove that only helicity-conserving four-dimensional gravity amplitudes have nonvanishing descendants when reduced to three dimensions; implying the vanishing of odd-multiplicity S-matrix elements, in agreement with Chern-Simons matter theory. We explicitly verify the double-copy correspondence at four and six points for N=12,10,8 supergravity theories and discuss its validity for all multiplicity.
[ { "created": "Mon, 8 Oct 2012 12:14:22 GMT", "version": "v1" }, { "created": "Mon, 22 Apr 2013 02:00:46 GMT", "version": "v2" } ]
2013-05-01
[ [ "Huang", "Yu-tin", "" ], [ "Johansson", "Henrik", "" ] ]
We show that three-dimensional supergravity amplitudes can be obtained as double copies of either three-algebra super-Chern-Simons matter theory or that of two-algebra super-Yang-Mills theory, when either theory is organized to display the color-kinematics duality. We prove that only helicity-conserving four-dimensional gravity amplitudes have nonvanishing descendants when reduced to three dimensions; implying the vanishing of odd-multiplicity S-matrix elements, in agreement with Chern-Simons matter theory. We explicitly verify the double-copy correspondence at four and six points for N=12,10,8 supergravity theories and discuss its validity for all multiplicity.
1905.12945
Paolo Di Lillo
Paolo Di Lillo, Stefano Chiaverini, Gianluca Antonelli
Handling robot constraints within a Set-Based Multi-Task Priority Inverse Kinematics Framework
null
Int;Conf;Rob;Aut (2019)
10.1109/ICRA.2019.8793625
null
cs.RO
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Set-Based Multi-Task Priority is a recent framework to handle inverse kinematics for redundant structures. Both equality tasks, i.e., control objectives to be driven to a desired value, and set-bases tasks, i.e., control objectives to be satisfied with a set/range of values can be addressed in a rigorous manner within a priority framework. In addition, optimization tasks, driven by the gradient of a proper function, may be considered as well, usually as lower priority tasks. In this paper the proper design of the tasks, their priority and the use of a Set-Based Multi-Task Priority framework is proposed in order to handle several constraints simultaneously in real-time. It is shown that safety related tasks such as, e.g., joint limits or kinematic singularity, may be properly handled by consider them both at an higher priority as set-based task and at a lower within a proper optimization functional. Experimental results on a 7DOF Jaco$^2$
[ { "created": "Thu, 30 May 2019 10:19:04 GMT", "version": "v1" } ]
2020-07-08
[ [ "Di Lillo", "Paolo", "" ], [ "Chiaverini", "Stefano", "" ], [ "Antonelli", "Gianluca", "" ] ]
Set-Based Multi-Task Priority is a recent framework to handle inverse kinematics for redundant structures. Both equality tasks, i.e., control objectives to be driven to a desired value, and set-bases tasks, i.e., control objectives to be satisfied with a set/range of values can be addressed in a rigorous manner within a priority framework. In addition, optimization tasks, driven by the gradient of a proper function, may be considered as well, usually as lower priority tasks. In this paper the proper design of the tasks, their priority and the use of a Set-Based Multi-Task Priority framework is proposed in order to handle several constraints simultaneously in real-time. It is shown that safety related tasks such as, e.g., joint limits or kinematic singularity, may be properly handled by consider them both at an higher priority as set-based task and at a lower within a proper optimization functional. Experimental results on a 7DOF Jaco$^2$
2404.16734
Enguerrand Prebet
Enguerrand Prebet, Andr\'e Platzer
Uniform Substitution for Differential Refinement Logic
IJCAR 2024
Automated Reasoning, 12th International Joint Conference, IJCAR 2024
10.1007/978-3-031-63501-4_11
null
cs.LO
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper introduces a uniform substitution calculus for differential refinement logic dRL. The logic dRL extends the differential dynamic logic dL such that one can simultaneously reason about properties of and relations between hybrid systems. Refinements are useful e.g. for simplifying proofs by relating a concrete hybrid system to an abstract one from which the property can be proved more easily. Uniform substitution is the key to parsimonious prover microkernels. It enables the verbatim use of single axiom formulas instead of axiom schemata with soundness-critical side conditions scattered across the proof calculus. The uniform substitution rule can then be used to instantiate all axioms soundly. Access to differential variables in dRL enables more control over the notion of refinement, which is shown to be decidable on a fragment of hybrid programs.
[ { "created": "Thu, 25 Apr 2024 16:43:25 GMT", "version": "v1" }, { "created": "Fri, 31 May 2024 09:00:20 GMT", "version": "v2" } ]
2024-07-11
[ [ "Prebet", "Enguerrand", "" ], [ "Platzer", "André", "" ] ]
This paper introduces a uniform substitution calculus for differential refinement logic dRL. The logic dRL extends the differential dynamic logic dL such that one can simultaneously reason about properties of and relations between hybrid systems. Refinements are useful e.g. for simplifying proofs by relating a concrete hybrid system to an abstract one from which the property can be proved more easily. Uniform substitution is the key to parsimonious prover microkernels. It enables the verbatim use of single axiom formulas instead of axiom schemata with soundness-critical side conditions scattered across the proof calculus. The uniform substitution rule can then be used to instantiate all axioms soundly. Access to differential variables in dRL enables more control over the notion of refinement, which is shown to be decidable on a fragment of hybrid programs.
2212.09588
Mingzhu Cai
Mingzhu Cai, Siqi Bao, Xin Tian, Huang He, Fan Wang, Hua Wu
Query Enhanced Knowledge-Intensive Conversation via Unsupervised Joint Modeling
Accepted for publication at ACL2023
null
null
null
cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we propose an unsupervised query enhanced approach for knowledge-intensive conversations, namely QKConv. There are three modules in QKConv: a query generator, an off-the-shelf knowledge selector, and a response generator. QKConv is optimized through joint training, which produces the response by exploring multiple candidate queries and leveraging corresponding selected knowledge. The joint training solely relies on the dialogue context and target response, getting exempt from extra query annotations or knowledge provenances. To evaluate the effectiveness of the proposed QKConv, we conduct experiments on three representative knowledge-intensive conversation datasets: conversational question-answering, task-oriented dialogue, and knowledge-grounded conversation. Experimental results reveal that QKConv performs better than all unsupervised methods across three datasets and achieves competitive performance compared to supervised methods.
[ { "created": "Mon, 19 Dec 2022 16:21:05 GMT", "version": "v1" }, { "created": "Fri, 26 May 2023 11:02:13 GMT", "version": "v2" } ]
2023-05-29
[ [ "Cai", "Mingzhu", "" ], [ "Bao", "Siqi", "" ], [ "Tian", "Xin", "" ], [ "He", "Huang", "" ], [ "Wang", "Fan", "" ], [ "Wu", "Hua", "" ] ]
In this paper, we propose an unsupervised query enhanced approach for knowledge-intensive conversations, namely QKConv. There are three modules in QKConv: a query generator, an off-the-shelf knowledge selector, and a response generator. QKConv is optimized through joint training, which produces the response by exploring multiple candidate queries and leveraging corresponding selected knowledge. The joint training solely relies on the dialogue context and target response, getting exempt from extra query annotations or knowledge provenances. To evaluate the effectiveness of the proposed QKConv, we conduct experiments on three representative knowledge-intensive conversation datasets: conversational question-answering, task-oriented dialogue, and knowledge-grounded conversation. Experimental results reveal that QKConv performs better than all unsupervised methods across three datasets and achieves competitive performance compared to supervised methods.
2311.07559
Conrad Zimmerman
Conrad Zimmerman, Jenna DiVincenzo, Jonathan Aldrich
Sound Gradual Verification with Symbolic Execution
Supplementary material; to be published by Principles of Programming Languages 2024
null
10.1145/3632927
null
cs.PL
http://creativecommons.org/licenses/by/4.0/
Gradual verification, which supports explicitly partial specifications and verifies them with a combination of static and dynamic checks, makes verification more incremental and provides earlier feedback to developers. While an abstract, weakest precondition-based approach to gradual verification was previously proven sound, the approach did not provide sufficient guidance for implementation and optimization of the required run-time checks. More recently, gradual verification was implemented using symbolic execution techniques, but the soundness of the approach (as with related static checkers based on implicit dynamic frames) was an open question. This paper puts practical gradual verification on a sound footing with a formalization of symbolic execution, optimized run-time check generation, and run time execution. We prove our approach is sound; our proof also covers a core subset of the Viper tool, for which we are aware of no previous soundness result. Our formalization enabled us to find a soundness bug in an implemented gradual verification tool and describe the fix necessary to make it sound.
[ { "created": "Mon, 13 Nov 2023 18:52:27 GMT", "version": "v1" } ]
2023-11-14
[ [ "Zimmerman", "Conrad", "" ], [ "DiVincenzo", "Jenna", "" ], [ "Aldrich", "Jonathan", "" ] ]
Gradual verification, which supports explicitly partial specifications and verifies them with a combination of static and dynamic checks, makes verification more incremental and provides earlier feedback to developers. While an abstract, weakest precondition-based approach to gradual verification was previously proven sound, the approach did not provide sufficient guidance for implementation and optimization of the required run-time checks. More recently, gradual verification was implemented using symbolic execution techniques, but the soundness of the approach (as with related static checkers based on implicit dynamic frames) was an open question. This paper puts practical gradual verification on a sound footing with a formalization of symbolic execution, optimized run-time check generation, and run time execution. We prove our approach is sound; our proof also covers a core subset of the Viper tool, for which we are aware of no previous soundness result. Our formalization enabled us to find a soundness bug in an implemented gradual verification tool and describe the fix necessary to make it sound.
1611.03525
Antonia Micol Frassino
Antonia M. Frassino, Robert B. Mann, Fil Simovic
Critical points in Lovelock Black Holes
8 pages, 4 figures. Contribution to the Proceedings of the Second Karl Schwarzschild Meeting, Frankfurt, 20-24 July 2015
null
null
null
hep-th gr-qc
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We review some of the results obtained by introducing a thermodynamic pressure via the cosmological constant in a class of higher curvature theories known as Lovelock gravity. In particular, we focus on a specific relation between the higher-order Lovelock couplings that introduces a peculiar isolated critical point for hyperbolic black holes characterized by non-standard critical exponents.
[ { "created": "Thu, 10 Nov 2016 21:47:19 GMT", "version": "v1" } ]
2016-11-14
[ [ "Frassino", "Antonia M.", "" ], [ "Mann", "Robert B.", "" ], [ "Simovic", "Fil", "" ] ]
We review some of the results obtained by introducing a thermodynamic pressure via the cosmological constant in a class of higher curvature theories known as Lovelock gravity. In particular, we focus on a specific relation between the higher-order Lovelock couplings that introduces a peculiar isolated critical point for hyperbolic black holes characterized by non-standard critical exponents.
1501.03775
Cesar Ag\'on
Cesar A. Agon and Howard J. Schnitzer
Holographic Mutual Information at small separations
13 pages, 1 figure, some references added
null
null
BRX-TH-6291
hep-th
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The holographic mutual information for the small separation of two circles and two strips in 2+1 dimensional space-time is considered based on the known exact minimal surfaces spanning the boundaries on AdS4. The results suggest a universality for the leading term in the short-distance expansion of holographic mutual information. A conjecture for a similar result for d > 2 is also presented, as well as comments about the analogous expansion in conformal field theory.
[ { "created": "Thu, 15 Jan 2015 18:55:42 GMT", "version": "v1" }, { "created": "Tue, 10 Feb 2015 19:48:24 GMT", "version": "v2" }, { "created": "Fri, 19 Aug 2016 22:10:57 GMT", "version": "v3" } ]
2016-08-23
[ [ "Agon", "Cesar A.", "" ], [ "Schnitzer", "Howard J.", "" ] ]
The holographic mutual information for the small separation of two circles and two strips in 2+1 dimensional space-time is considered based on the known exact minimal surfaces spanning the boundaries on AdS4. The results suggest a universality for the leading term in the short-distance expansion of holographic mutual information. A conjecture for a similar result for d > 2 is also presented, as well as comments about the analogous expansion in conformal field theory.
2107.05746
Binghui Peng
Thomas Chen, Xi Chen, Binghui Peng, Mihalis Yannakakis
Computational Hardness of the Hylland-Zeckhauser Scheme
null
null
null
null
cs.GT cs.DS
http://creativecommons.org/licenses/by/4.0/
We study the complexity of the classic Hylland-Zeckhauser scheme [HZ'79] for one-sided matching markets. We show that the problem of finding an $\epsilon$-approximate equilibrium in the HZ scheme is PPAD-hard, and this holds even when $\epsilon$ is polynomially small and when each agent has no more than four distinct utility values. Our hardness result, when combined with the PPAD membership result of [VY'21], resolves the approximation complexity of the HZ scheme. We also show that the problem of approximating the optimal social welfare (the weight of the matching) achievable by HZ equilibria within a certain constant factor is NP-hard.
[ { "created": "Mon, 12 Jul 2021 21:31:28 GMT", "version": "v1" } ]
2021-07-14
[ [ "Chen", "Thomas", "" ], [ "Chen", "Xi", "" ], [ "Peng", "Binghui", "" ], [ "Yannakakis", "Mihalis", "" ] ]
We study the complexity of the classic Hylland-Zeckhauser scheme [HZ'79] for one-sided matching markets. We show that the problem of finding an $\epsilon$-approximate equilibrium in the HZ scheme is PPAD-hard, and this holds even when $\epsilon$ is polynomially small and when each agent has no more than four distinct utility values. Our hardness result, when combined with the PPAD membership result of [VY'21], resolves the approximation complexity of the HZ scheme. We also show that the problem of approximating the optimal social welfare (the weight of the matching) achievable by HZ equilibria within a certain constant factor is NP-hard.
2107.00353
Jeonghyun Byun
Jeonghyun Byun, Dongjae Lee, Hoseong Seo, Inkyu Jang, Jeongjun Choi, H. Jin Kim
Stability and Robustness Analysis of Plug-Pulling using an Aerial Manipulator
to be presented in 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 2021
null
null
null
cs.RO cs.SY eess.SY
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, an autonomous aerial manipulation task of pulling a plug out of an electric socket is conducted, where maintaining the stability and robustness is challenging due to sudden disappearance of a large interaction force. The abrupt change in the dynamical model before and after the separation of the plug can cause destabilization or mission failure. To accomplish aerial plug-pulling, we employ the concept of hybrid automata to divide the task into three operative modes, i.e, wire-pulling, stabilizing, and free-flight. Also, a strategy for trajectory generation and a design of disturbance-observer-based controllers for each operative mode are presented. Furthermore, the theory of hybrid automata is used to prove the stability and robustness during the mode transition. We validate the proposed trajectory generation and control method by an actual wire-pulling experiment with a multirotor-based aerial manipulator.
[ { "created": "Thu, 1 Jul 2021 10:36:27 GMT", "version": "v1" }, { "created": "Tue, 6 Jul 2021 03:05:29 GMT", "version": "v2" } ]
2021-07-07
[ [ "Byun", "Jeonghyun", "" ], [ "Lee", "Dongjae", "" ], [ "Seo", "Hoseong", "" ], [ "Jang", "Inkyu", "" ], [ "Choi", "Jeongjun", "" ], [ "Kim", "H. Jin", "" ] ]
In this paper, an autonomous aerial manipulation task of pulling a plug out of an electric socket is conducted, where maintaining the stability and robustness is challenging due to sudden disappearance of a large interaction force. The abrupt change in the dynamical model before and after the separation of the plug can cause destabilization or mission failure. To accomplish aerial plug-pulling, we employ the concept of hybrid automata to divide the task into three operative modes, i.e, wire-pulling, stabilizing, and free-flight. Also, a strategy for trajectory generation and a design of disturbance-observer-based controllers for each operative mode are presented. Furthermore, the theory of hybrid automata is used to prove the stability and robustness during the mode transition. We validate the proposed trajectory generation and control method by an actual wire-pulling experiment with a multirotor-based aerial manipulator.
2305.06472
Huan Wang
Yan-Fu Li, Huan Wang, Muxia Sun
ChatGPT-Like Large-Scale Foundation Models for Prognostics and Health Management: A Survey and Roadmaps
55 pages, 10 figures
null
null
null
cs.LG cs.AI cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Prognostics and health management (PHM) technology plays a critical role in industrial production and equipment maintenance by identifying and predicting possible equipment failures and damages, thereby allowing necessary maintenance measures to be taken to enhance equipment service life and reliability while reducing production costs and downtime. In recent years, PHM technology based on artificial intelligence (AI) has made remarkable achievements in the context of the industrial IoT and big data, and it is widely used in various industries, such as railway, energy, and aviation, for condition monitoring, fault prediction, and health management. The emergence of large-scale foundation models (LSF-Models) such as ChatGPT and DALLE-E marks the entry of AI into a new era of AI-2.0 from AI-1.0, where deep models have rapidly evolved from a research paradigm of single-modal, single-task, and limited-data to a multi-modal, multi-task, massive data, and super-large model paradigm. ChatGPT represents a landmark achievement in this research paradigm, offering hope for general artificial intelligence due to its highly intelligent natural language understanding ability. However, the PHM field lacks a consensus on how to respond to this significant change in the AI field, and a systematic review and roadmap is required to elucidate future development directions. To fill this gap, this paper systematically expounds on the key components and latest developments of LSF-Models. Then, we systematically answered how to build the LSF-Model applicable to PHM tasks and outlined the challenges and future development roadmaps for this research paradigm.
[ { "created": "Wed, 10 May 2023 21:37:44 GMT", "version": "v1" }, { "created": "Fri, 12 May 2023 10:41:35 GMT", "version": "v2" } ]
2023-05-15
[ [ "Li", "Yan-Fu", "" ], [ "Wang", "Huan", "" ], [ "Sun", "Muxia", "" ] ]
Prognostics and health management (PHM) technology plays a critical role in industrial production and equipment maintenance by identifying and predicting possible equipment failures and damages, thereby allowing necessary maintenance measures to be taken to enhance equipment service life and reliability while reducing production costs and downtime. In recent years, PHM technology based on artificial intelligence (AI) has made remarkable achievements in the context of the industrial IoT and big data, and it is widely used in various industries, such as railway, energy, and aviation, for condition monitoring, fault prediction, and health management. The emergence of large-scale foundation models (LSF-Models) such as ChatGPT and DALLE-E marks the entry of AI into a new era of AI-2.0 from AI-1.0, where deep models have rapidly evolved from a research paradigm of single-modal, single-task, and limited-data to a multi-modal, multi-task, massive data, and super-large model paradigm. ChatGPT represents a landmark achievement in this research paradigm, offering hope for general artificial intelligence due to its highly intelligent natural language understanding ability. However, the PHM field lacks a consensus on how to respond to this significant change in the AI field, and a systematic review and roadmap is required to elucidate future development directions. To fill this gap, this paper systematically expounds on the key components and latest developments of LSF-Models. Then, we systematically answered how to build the LSF-Model applicable to PHM tasks and outlined the challenges and future development roadmaps for this research paradigm.
1903.04205
Stefanos Leonardos Mr.
Vitalik Buterin and Daniel Reijsbergen and Stefanos Leonardos and Georgios Piliouras
Incentives in Ethereum's Hybrid Casper Protocol
Conference version: IEEE International Conference on Blockchain and Cryptocurrency (2019)
International Journal of Network Management, Vol. 30(5), pp.:e2098, (2020)
10.1002/nem.2098
null
cs.CR cs.DC cs.GT
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We present an overview of hybrid Casper the Friendly Finality Gadget (FFG): a Proof-of-Stake checkpointing protocol overlaid onto Ethereum's Proof-of-Work blockchain. We describe its core functionalities and reward scheme, and explore its properties. Our findings indicate that Casper's implemented incentives mechanism ensures liveness, while providing safety guarantees that improve over standard Proof-of-Work protocols. Based on a minimal-impact implementation of the protocol as a smart contract on the blockchain, we discuss additional issues related to parametrisation, funding, throughput and network overhead and detect potential limitations.
[ { "created": "Mon, 11 Mar 2019 10:33:13 GMT", "version": "v1" }, { "created": "Sat, 21 Mar 2020 09:29:36 GMT", "version": "v2" }, { "created": "Sun, 18 Jul 2021 13:03:19 GMT", "version": "v3" } ]
2021-07-20
[ [ "Buterin", "Vitalik", "" ], [ "Reijsbergen", "Daniel", "" ], [ "Leonardos", "Stefanos", "" ], [ "Piliouras", "Georgios", "" ] ]
We present an overview of hybrid Casper the Friendly Finality Gadget (FFG): a Proof-of-Stake checkpointing protocol overlaid onto Ethereum's Proof-of-Work blockchain. We describe its core functionalities and reward scheme, and explore its properties. Our findings indicate that Casper's implemented incentives mechanism ensures liveness, while providing safety guarantees that improve over standard Proof-of-Work protocols. Based on a minimal-impact implementation of the protocol as a smart contract on the blockchain, we discuss additional issues related to parametrisation, funding, throughput and network overhead and detect potential limitations.
2311.06973
Behrouz Azimian
Behrouz Azimian, Shiva Moshtagh, Anamitra Pal, Shanshan Ma
Analytical Verification of Performance of Deep Neural Network Based Time-Synchronized Distribution System State Estimation
8 pages, in Journal of Modern Power Systems and Clean Energy, 2023
null
10.35833/MPCE.2023.000432
null
cs.LG cs.SY eess.SY
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Recently, we demonstrated success of a time-synchronized state estimator using deep neural networks (DNNs) for real-time unobservable distribution systems. In this letter, we provide analytical bounds on the performance of that state estimator as a function of perturbations in the input measurements. It has already been shown that evaluating performance based on only the test dataset might not effectively indicate a trained DNN's ability to handle input perturbations. As such, we analytically verify robustness and trustworthiness of DNNs to input perturbations by treating them as mixed-integer linear programming (MILP) problems. The ability of batch normalization in addressing the scalability limitations of the MILP formulation is also highlighted. The framework is validated by performing time-synchronized distribution system state estimation for a modified IEEE 34-node system and a real-world large distribution system, both of which are incompletely observed by micro-phasor measurement units.
[ { "created": "Sun, 12 Nov 2023 22:01:34 GMT", "version": "v1" }, { "created": "Tue, 6 Feb 2024 21:40:46 GMT", "version": "v2" }, { "created": "Tue, 20 Feb 2024 23:37:08 GMT", "version": "v3" }, { "created": "Thu, 22 Feb 2024 16:33:10 GMT", "version": "v4" } ]
2024-02-23
[ [ "Azimian", "Behrouz", "" ], [ "Moshtagh", "Shiva", "" ], [ "Pal", "Anamitra", "" ], [ "Ma", "Shanshan", "" ] ]
Recently, we demonstrated success of a time-synchronized state estimator using deep neural networks (DNNs) for real-time unobservable distribution systems. In this letter, we provide analytical bounds on the performance of that state estimator as a function of perturbations in the input measurements. It has already been shown that evaluating performance based on only the test dataset might not effectively indicate a trained DNN's ability to handle input perturbations. As such, we analytically verify robustness and trustworthiness of DNNs to input perturbations by treating them as mixed-integer linear programming (MILP) problems. The ability of batch normalization in addressing the scalability limitations of the MILP formulation is also highlighted. The framework is validated by performing time-synchronized distribution system state estimation for a modified IEEE 34-node system and a real-world large distribution system, both of which are incompletely observed by micro-phasor measurement units.
1007.0338
Dirk Kreimer
Spencer Bloch and Dirk Kreimer
Feynman amplitudes and Landau singularities for 1-loop graphs
31p
null
null
IHES M/10/20
hep-th math-ph math.AG math.MP
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We use mixed Hodge structures to investigate Feynman amplitudes as functions of external momenta and masses.
[ { "created": "Fri, 2 Jul 2010 11:12:47 GMT", "version": "v1" }, { "created": "Mon, 26 Jul 2010 18:01:48 GMT", "version": "v2" } ]
2010-07-27
[ [ "Bloch", "Spencer", "" ], [ "Kreimer", "Dirk", "" ] ]
We use mixed Hodge structures to investigate Feynman amplitudes as functions of external momenta and masses.
hep-th/0109133
Cai Rong-gen
Rong-Gen Cai
Gauss-Bonnet Black Holes in AdS Spaces
Revtex, 17 pages with 9 eps figures, v2: section II removed and references added, the version to appear in PRD
Phys.Rev.D65:084014,2002
10.1103/PhysRevD.65.084014
null
hep-th gr-qc
null
We study thermodynamic properties and phase structures of topological black holes in Einstein theory with a Gauss-Bonnet term and a negative cosmological constant. The event horizon of these topological black holes can be a hypersurface with positive, zero or negative constant curvature. When the horizon is a zero curvature hypersurface, the thermodynamic properties of black holes are completely the same as those of black holes without the Gauss-Bonnet term, although the two black hole solutions are quite different. When the horizon is a negative constant curvature hypersurface, the thermodynamic properties of the Gauss-Bonnet black holes are qualitatively similar to those of black holes without the Gauss-Bonnet term. When the event horizon is a hypersurface with positive constant curvature, we find that the thermodynamic properties and phase structures of black holes drastically depend on the spacetime dimension $d$ and the coefficient of the Gauss-Bonnet term: when $d\ge 6$, the properties of black hole are also qualitatively similar to the case without the Gauss-Bonnet term, but when $d=5$, a new phase of locally stable small black hole occurs under a critical value of the Gauss-Bonnet coefficient, and beyond the critical value, the black holes are always thermodynamically stable. However, the locally stable small black hole is not globally preferred, instead a thermal anti-de Sitter space is globally preferred. We find that there is a minimal horizon radius, below which the Hawking-Page phase transition will not occur since for these black holes the thermal anti de Sitter space is always globally preferred.
[ { "created": "Tue, 18 Sep 2001 08:43:03 GMT", "version": "v1" }, { "created": "Sat, 12 Jan 2002 03:57:23 GMT", "version": "v2" } ]
2011-05-05
[ [ "Cai", "Rong-Gen", "" ] ]
We study thermodynamic properties and phase structures of topological black holes in Einstein theory with a Gauss-Bonnet term and a negative cosmological constant. The event horizon of these topological black holes can be a hypersurface with positive, zero or negative constant curvature. When the horizon is a zero curvature hypersurface, the thermodynamic properties of black holes are completely the same as those of black holes without the Gauss-Bonnet term, although the two black hole solutions are quite different. When the horizon is a negative constant curvature hypersurface, the thermodynamic properties of the Gauss-Bonnet black holes are qualitatively similar to those of black holes without the Gauss-Bonnet term. When the event horizon is a hypersurface with positive constant curvature, we find that the thermodynamic properties and phase structures of black holes drastically depend on the spacetime dimension $d$ and the coefficient of the Gauss-Bonnet term: when $d\ge 6$, the properties of black hole are also qualitatively similar to the case without the Gauss-Bonnet term, but when $d=5$, a new phase of locally stable small black hole occurs under a critical value of the Gauss-Bonnet coefficient, and beyond the critical value, the black holes are always thermodynamically stable. However, the locally stable small black hole is not globally preferred, instead a thermal anti-de Sitter space is globally preferred. We find that there is a minimal horizon radius, below which the Hawking-Page phase transition will not occur since for these black holes the thermal anti de Sitter space is always globally preferred.
2310.04131
Ilya Lvovich Shapiro
Ilya L. Shapiro
Antisymmetric Tensor Field and Cheshire Cat Smile of the Local Conformal Symmetry
In v2 discussions were extended and added new references, especially to the previous works about conformal operators acting on antisymmetric tensor fields. Version to be submitted for publication. 20 pages, no figures
null
null
null
hep-th gr-qc
http://creativecommons.org/licenses/by/4.0/
The conformal version of the antisymmetric second-order tensor field in four spacetime dimensions does not have gauge invariance extensively discussed in the literature for more than half a century. Our first observation is that, when coupled to fermions, only the conformal version provides renormalizability of the theory at the one-loop level. General considerations are supported by the derivation of one-loop divergences in the fermionic sector, indicating good chances for asymptotic freedom. The arguments concerning one-loop renormalizability remain valid in the presence of self-interactions and the masses for both fermion and antisymmetric tensor fields. In the flat spacetime limit, regardless of the conformal symmetry has gone, there is an expectation to meet renormalizability in all loop orders.
[ { "created": "Fri, 6 Oct 2023 10:00:58 GMT", "version": "v1" }, { "created": "Fri, 20 Oct 2023 23:50:37 GMT", "version": "v2" } ]
2023-10-24
[ [ "Shapiro", "Ilya L.", "" ] ]
The conformal version of the antisymmetric second-order tensor field in four spacetime dimensions does not have gauge invariance extensively discussed in the literature for more than half a century. Our first observation is that, when coupled to fermions, only the conformal version provides renormalizability of the theory at the one-loop level. General considerations are supported by the derivation of one-loop divergences in the fermionic sector, indicating good chances for asymptotic freedom. The arguments concerning one-loop renormalizability remain valid in the presence of self-interactions and the masses for both fermion and antisymmetric tensor fields. In the flat spacetime limit, regardless of the conformal symmetry has gone, there is an expectation to meet renormalizability in all loop orders.
2301.12623
Hanlin Gu
Hanlin Gu, Jiahuan Luo, Yan Kang, Lixin Fan and Qiang Yang
FedPass: Privacy-Preserving Vertical Federated Deep Learning with Adaptive Obfuscation
6 figures, 9 tables
null
null
null
cs.DC cs.CR cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Vertical federated learning (VFL) allows an active party with labeled feature to leverage auxiliary features from the passive parties to improve model performance. Concerns about the private feature and label leakage in both the training and inference phases of VFL have drawn wide research attention. In this paper, we propose a general privacy-preserving vertical federated deep learning framework called FedPass, which leverages adaptive obfuscation to protect the feature and label simultaneously. Strong privacy-preserving capabilities about private features and labels are theoretically proved (in Theorems 1 and 2). Extensive experimental result s with different datasets and network architectures also justify the superiority of FedPass against existing methods in light of its near-optimal trade-off between privacy and model performance.
[ { "created": "Mon, 30 Jan 2023 02:36:23 GMT", "version": "v1" }, { "created": "Tue, 31 Jan 2023 07:40:05 GMT", "version": "v2" } ]
2023-02-01
[ [ "Gu", "Hanlin", "" ], [ "Luo", "Jiahuan", "" ], [ "Kang", "Yan", "" ], [ "Fan", "Lixin", "" ], [ "Yang", "Qiang", "" ] ]
Vertical federated learning (VFL) allows an active party with labeled feature to leverage auxiliary features from the passive parties to improve model performance. Concerns about the private feature and label leakage in both the training and inference phases of VFL have drawn wide research attention. In this paper, we propose a general privacy-preserving vertical federated deep learning framework called FedPass, which leverages adaptive obfuscation to protect the feature and label simultaneously. Strong privacy-preserving capabilities about private features and labels are theoretically proved (in Theorems 1 and 2). Extensive experimental result s with different datasets and network architectures also justify the superiority of FedPass against existing methods in light of its near-optimal trade-off between privacy and model performance.
0708.0433
Karol Kozlowski Kajetan
K. K. Kozlowski
On the emptiness formation probability of the open XXZ spin-$\tf{1}{2}$ chain
18 pages
J.Stat.Mech.0802:P02006,2008
10.1088/1742-5468/2008/02/P02006
null
hep-th
null
This paper is devoted to the study of the emptiness formation probability $\tau\pa{m}$ of the open XXZ chain. We derive a closed form for $\tau\pa{m}$ at $\Delta=\tf{1}{2}$ when the boundary field vanishes. Moreover we obtain its leading asymptotics for an arbitrary boundary field at the free fermion point. Finally, we compute the first term of the asymptotics of $\ln\pa{\tau\pa{m}}$ in the whole massless regime $-1<\Delta<1$.
[ { "created": "Thu, 2 Aug 2007 22:51:16 GMT", "version": "v1" } ]
2008-11-26
[ [ "Kozlowski", "K. K.", "" ] ]
This paper is devoted to the study of the emptiness formation probability $\tau\pa{m}$ of the open XXZ chain. We derive a closed form for $\tau\pa{m}$ at $\Delta=\tf{1}{2}$ when the boundary field vanishes. Moreover we obtain its leading asymptotics for an arbitrary boundary field at the free fermion point. Finally, we compute the first term of the asymptotics of $\ln\pa{\tau\pa{m}}$ in the whole massless regime $-1<\Delta<1$.
hep-th/0006011
Valentin Khoze
N. Michael Davies, Timothy J. Hollowood and Valentin V. Khoze
Monopoles, affine algebras and the gluino condensate
minor changes, 23 pages, no figures
J.Math.Phys.44:3640-3656,2003
10.1063/1.1586477
null
hep-th hep-ph
null
We examine the low-energy dynamics of four-dimensional supersymmetric gauge theories and calculate the values of the gluino condensate for all simple gauge groups. By initially compactifying the theory on a cylinder we are able to perform calculations in a controlled weakly-coupled way for small radius. The dominant contributions to the path integral on the cylinder arise from magnetic monopoles which play the role of instanton constituents. We find that the semi-classically generated superpotential of the theory is the affine Toda potential for an associated twisted affine algebra. We determine the supersymmetric vacua and calculate the values of the gluino condensate. The number of supersymmetric vacua is equal to c_2, the dual Coxeter number, and in each vacuum the monopoles carry a fraction 1/c_2 of topological charge. As the results are independent of the radius of the circle, they are also valid in the strong coupling regime where the theory becomes decompactified. In this way we obtain values for the gluino condensate which for the classical gauge groups agree with previously known ``weak coupling instanton'' expressions (but not with the ``strong coupling instanton'' calculations). This detailed agreement provides further evidence in favour of the recently advocated resolution of the the gluino condensate puzzle. We also make explicit predictions for the gluino condensate for the exceptional groups.
[ { "created": "Thu, 1 Jun 2000 21:48:19 GMT", "version": "v1" }, { "created": "Mon, 5 Jun 2000 16:07:31 GMT", "version": "v2" } ]
2010-11-19
[ [ "Davies", "N. Michael", "" ], [ "Hollowood", "Timothy J.", "" ], [ "Khoze", "Valentin V.", "" ] ]
We examine the low-energy dynamics of four-dimensional supersymmetric gauge theories and calculate the values of the gluino condensate for all simple gauge groups. By initially compactifying the theory on a cylinder we are able to perform calculations in a controlled weakly-coupled way for small radius. The dominant contributions to the path integral on the cylinder arise from magnetic monopoles which play the role of instanton constituents. We find that the semi-classically generated superpotential of the theory is the affine Toda potential for an associated twisted affine algebra. We determine the supersymmetric vacua and calculate the values of the gluino condensate. The number of supersymmetric vacua is equal to c_2, the dual Coxeter number, and in each vacuum the monopoles carry a fraction 1/c_2 of topological charge. As the results are independent of the radius of the circle, they are also valid in the strong coupling regime where the theory becomes decompactified. In this way we obtain values for the gluino condensate which for the classical gauge groups agree with previously known ``weak coupling instanton'' expressions (but not with the ``strong coupling instanton'' calculations). This detailed agreement provides further evidence in favour of the recently advocated resolution of the the gluino condensate puzzle. We also make explicit predictions for the gluino condensate for the exceptional groups.
hep-th/0005008
Hiroaki Kanno
Tohru Eguchi and Hiroaki Kanno
Five-Dimensional Gauge Theories and Local Mirror Symmetry
18 pages, Latex, minor changes, a version to appear in Nucl.Phys.B
Nucl.Phys.B586:331-345,2000
10.1016/S0550-3213(00)00375-8
UT-882
hep-th
null
We study the dynamics of 5-dimensional gauge theory on $M_4\times S^1$ by compactifying type II/M theory on degenerate Calabi-Yau manifolds. We use the local mirror symmetry and shall show that the prepotential of the 5-dimensional SU(2) gauge theory without matter is given exactly by that of the type II string theory compactified on the local ${\bf F}_2$, i.e. Hirzebruch surface ${\bf F}_2$ lying inside a non-compact Calabi-Yau manifold. It is shown that our result reproduces the Seiberg-Witten theory at the 4-dimensional limit $R\to 0$ ($R$ denotes the radius of $S^1$) and also the result of the uncompactified 5-dimensional theory at $R\to \infty$. We also discuss SU(2) gauge theory with $1\le N_f\le 4$ matter in vector representations and show that they are described by the geometry of the local ${\bf F}_2$ blown up at $N_f$ points.
[ { "created": "Mon, 1 May 2000 07:05:19 GMT", "version": "v1" }, { "created": "Mon, 19 Jun 2000 04:05:07 GMT", "version": "v2" } ]
2010-11-19
[ [ "Eguchi", "Tohru", "" ], [ "Kanno", "Hiroaki", "" ] ]
We study the dynamics of 5-dimensional gauge theory on $M_4\times S^1$ by compactifying type II/M theory on degenerate Calabi-Yau manifolds. We use the local mirror symmetry and shall show that the prepotential of the 5-dimensional SU(2) gauge theory without matter is given exactly by that of the type II string theory compactified on the local ${\bf F}_2$, i.e. Hirzebruch surface ${\bf F}_2$ lying inside a non-compact Calabi-Yau manifold. It is shown that our result reproduces the Seiberg-Witten theory at the 4-dimensional limit $R\to 0$ ($R$ denotes the radius of $S^1$) and also the result of the uncompactified 5-dimensional theory at $R\to \infty$. We also discuss SU(2) gauge theory with $1\le N_f\le 4$ matter in vector representations and show that they are described by the geometry of the local ${\bf F}_2$ blown up at $N_f$ points.
1806.02702
Taejin Lee
Taejin Lee
Four-Graviton Scattering and String Path Integral in the Proper-time gauge
9 pages, 1 figure, new references and comments added
null
null
null
hep-th gr-qc
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We evaluate the four-closed-string scattering amplitude, using the Polyakov string path integral in the proper-time gauge. By identifying the Fock space representation of the four-closed-string-vertex, we obtain a field theoretic expression of the closed string scattering amplitudes. In the zero-slope limit, the four-closed-string scattering amplitude reduces to the four-graviton-scattering amplitude of Einstein's gravity. However, at a finite slope, the four-graviton scattering amplitude in the proper-time gauge differs not only from that of Einstein gravity, but also significantly differs from the conventional one obtained by using the vertex operator technique in string theory. This discrepancy is mainly due to the presence of closed string tachyon poles in the four-graviton-scattering amplitude, which are missing in previous works. Because the tachyon poles in the scattering amplitude considerably alter the short distance behavior of gravitational interaction, they may be important in understanding problems associated with the perturbative theory of quantum gravity and the dark matter within the framework of string theory.
[ { "created": "Thu, 7 Jun 2018 14:27:45 GMT", "version": "v1" }, { "created": "Mon, 17 Jun 2019 11:13:54 GMT", "version": "v2" } ]
2019-06-18
[ [ "Lee", "Taejin", "" ] ]
We evaluate the four-closed-string scattering amplitude, using the Polyakov string path integral in the proper-time gauge. By identifying the Fock space representation of the four-closed-string-vertex, we obtain a field theoretic expression of the closed string scattering amplitudes. In the zero-slope limit, the four-closed-string scattering amplitude reduces to the four-graviton-scattering amplitude of Einstein's gravity. However, at a finite slope, the four-graviton scattering amplitude in the proper-time gauge differs not only from that of Einstein gravity, but also significantly differs from the conventional one obtained by using the vertex operator technique in string theory. This discrepancy is mainly due to the presence of closed string tachyon poles in the four-graviton-scattering amplitude, which are missing in previous works. Because the tachyon poles in the scattering amplitude considerably alter the short distance behavior of gravitational interaction, they may be important in understanding problems associated with the perturbative theory of quantum gravity and the dark matter within the framework of string theory.
1902.06275
Mladen Kova\v{c}evi\'c
Mladen Kova\v{c}evi\'c
Zero-Error Capacity of Duplication Channels
8 pages (double-column), 4 figures. Accepted for publication in IEEE Transactions on Communications
IEEE Trans. Commun., vol. 67, no. 10, pp. 6735-6742, Oct. 2019
10.1109/TCOMM.2019.2931342
null
cs.IT cs.DM math.IT
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper is concerned with the problem of error-free communication over the i.i.d. duplication channel which acts on a transmitted sequence $ x_1 \cdots x_n $ by inserting a random number of copies of each symbol $ x_i $ next to the original symbol. The random variables representing the numbers of inserted copies at each position $ i $ are independent and take values in $ \{0, 1, \ldots, r\} $, where $ r $ is a fixed parameter. A more general model in which blocks of $ \ell $ consecutive symbols are being duplicated, and which is inspired by DNA-based data storage systems wherein the stored molecules are subject to tandem-duplication mutations, is also analyzed. A construction of optimal codes correcting all patterns of errors of this type is described, and the zero-error capacity of the duplication channel---the largest rate at which information can be transmitted through it in an error-free manner---is determined for each $ \ell $ and $ r $.
[ { "created": "Sun, 17 Feb 2019 15:26:07 GMT", "version": "v1" }, { "created": "Thu, 21 Feb 2019 09:01:19 GMT", "version": "v2" }, { "created": "Thu, 6 Jun 2019 08:06:04 GMT", "version": "v3" }, { "created": "Tue, 23 Jul 2019 07:53:24 GMT", "version": "v4" } ]
2020-08-13
[ [ "Kovačević", "Mladen", "" ] ]
This paper is concerned with the problem of error-free communication over the i.i.d. duplication channel which acts on a transmitted sequence $ x_1 \cdots x_n $ by inserting a random number of copies of each symbol $ x_i $ next to the original symbol. The random variables representing the numbers of inserted copies at each position $ i $ are independent and take values in $ \{0, 1, \ldots, r\} $, where $ r $ is a fixed parameter. A more general model in which blocks of $ \ell $ consecutive symbols are being duplicated, and which is inspired by DNA-based data storage systems wherein the stored molecules are subject to tandem-duplication mutations, is also analyzed. A construction of optimal codes correcting all patterns of errors of this type is described, and the zero-error capacity of the duplication channel---the largest rate at which information can be transmitted through it in an error-free manner---is determined for each $ \ell $ and $ r $.
1701.00245
Poul Olesen
Poul Olesen
Non-Abelian bootstrap of primordial magnetism
6 pages. Magnetic energy formula for the flat expanding universe added. Misprint corrected. Some clarifying remarks on energy minimization added
null
null
null
hep-th astro-ph.HE hep-ph math-ph math.MP
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We point out that a primordial magnetic field can be generated in the electroweak phase transition by a non-Abelian bootstrap, where the field is generated by currents of W's, which in turn are extracted from the vacuum by the magnetic field. This magnetic field is produced as a vortex condensate at the electroweak phase transition. It becomes stringy as a consequence of the dynamical evolution due to magnetohydrodynamics.
[ { "created": "Sun, 1 Jan 2017 13:53:40 GMT", "version": "v1" }, { "created": "Wed, 11 Jan 2017 14:16:58 GMT", "version": "v2" }, { "created": "Wed, 25 Jan 2017 14:49:06 GMT", "version": "v3" } ]
2017-01-27
[ [ "Olesen", "Poul", "" ] ]
We point out that a primordial magnetic field can be generated in the electroweak phase transition by a non-Abelian bootstrap, where the field is generated by currents of W's, which in turn are extracted from the vacuum by the magnetic field. This magnetic field is produced as a vortex condensate at the electroweak phase transition. It becomes stringy as a consequence of the dynamical evolution due to magnetohydrodynamics.
2106.13249
Tan Zhi-Xuan
Arwa Alanqary, Gloria Z. Lin, Joie Le, Tan Zhi-Xuan, Vikash K. Mansinghka, Joshua B. Tenenbaum
Modeling the Mistakes of Boundedly Rational Agents Within a Bayesian Theory of Mind
Accepted to CogSci 2021. 6 pages, 5 figures. (Appendix: 1 page, 1 figure)
null
null
null
cs.AI q-bio.NC
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
When inferring the goals that others are trying to achieve, people intuitively understand that others might make mistakes along the way. This is crucial for activities such as teaching, offering assistance, and deciding between blame or forgiveness. However, Bayesian models of theory of mind have generally not accounted for these mistakes, instead modeling agents as mostly optimal in achieving their goals. As a result, they are unable to explain phenomena like locking oneself out of one's house, or losing a game of chess. Here, we extend the Bayesian Theory of Mind framework to model boundedly rational agents who may have mistaken goals, plans, and actions. We formalize this by modeling agents as probabilistic programs, where goals may be confused with semantically similar states, plans may be misguided due to resource-bounded planning, and actions may be unintended due to execution errors. We present experiments eliciting human goal inferences in two domains: (i) a gridworld puzzle with gems locked behind doors, and (ii) a block-stacking domain. Our model better explains human inferences than alternatives, while generalizing across domains. These findings indicate the importance of modeling others as bounded agents, in order to account for the full richness of human intuitive psychology.
[ { "created": "Thu, 24 Jun 2021 18:00:03 GMT", "version": "v1" } ]
2021-06-28
[ [ "Alanqary", "Arwa", "" ], [ "Lin", "Gloria Z.", "" ], [ "Le", "Joie", "" ], [ "Zhi-Xuan", "Tan", "" ], [ "Mansinghka", "Vikash K.", "" ], [ "Tenenbaum", "Joshua B.", "" ] ]
When inferring the goals that others are trying to achieve, people intuitively understand that others might make mistakes along the way. This is crucial for activities such as teaching, offering assistance, and deciding between blame or forgiveness. However, Bayesian models of theory of mind have generally not accounted for these mistakes, instead modeling agents as mostly optimal in achieving their goals. As a result, they are unable to explain phenomena like locking oneself out of one's house, or losing a game of chess. Here, we extend the Bayesian Theory of Mind framework to model boundedly rational agents who may have mistaken goals, plans, and actions. We formalize this by modeling agents as probabilistic programs, where goals may be confused with semantically similar states, plans may be misguided due to resource-bounded planning, and actions may be unintended due to execution errors. We present experiments eliciting human goal inferences in two domains: (i) a gridworld puzzle with gems locked behind doors, and (ii) a block-stacking domain. Our model better explains human inferences than alternatives, while generalizing across domains. These findings indicate the importance of modeling others as bounded agents, in order to account for the full richness of human intuitive psychology.
2201.03363
Anders Sundnes L{\o}vlie
Anders Sundnes L{\o}vlie, Astrid Waagstein and Peter Hyldg{\aa}rd
The Scientific Evidence Indicator for Popular Science News
null
NordMedia 2019. Communication, creativity and imagination: challenging the field. Malm\"o University, Sweden, 21st August 2019
null
null
cs.DL cs.HC
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
To what extent can news media help in providing more credible information about science? This is the core challenge for the Science Evidence Indicator (SEI) project, a collaboration between the Danish popular news website videnskab.dk and the authors of this paper. Looking specifically at medical science news, we aim to provide a transparent assessment of the scientific sources behind a story. This entails identifying some of the criteria that scientists use to assess research, and making it accessible and understandable for readers. We address the following research question: How can we communicate the quality of scientific publications in health science to a non-expert audience? Our goal is to make the assessments understandable for the youngest part of the website's target audience: high school students from age 16 and upwards.
[ { "created": "Mon, 10 Jan 2022 14:28:54 GMT", "version": "v1" } ]
2022-01-11
[ [ "Løvlie", "Anders Sundnes", "" ], [ "Waagstein", "Astrid", "" ], [ "Hyldgård", "Peter", "" ] ]
To what extent can news media help in providing more credible information about science? This is the core challenge for the Science Evidence Indicator (SEI) project, a collaboration between the Danish popular news website videnskab.dk and the authors of this paper. Looking specifically at medical science news, we aim to provide a transparent assessment of the scientific sources behind a story. This entails identifying some of the criteria that scientists use to assess research, and making it accessible and understandable for readers. We address the following research question: How can we communicate the quality of scientific publications in health science to a non-expert audience? Our goal is to make the assessments understandable for the youngest part of the website's target audience: high school students from age 16 and upwards.
2109.05396
Vaibhav B Sinha
Fu Li, C. Gregory Plaxton, Vaibhav B. Sinha
The Obnoxious Facility Location Game with Dichotomous Preferences
34 pages. This is an extended version of a paper presented at the 22nd Italian Conference on Theoretical Computer Science in September 2021
Theoretical Computer Science 961 (2023) 113930
10.1016/j.tcs.2023.113930
null
cs.GT
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We consider a facility location game in which $n$ agents reside at known locations on a path, and $k$ heterogeneous facilities are to be constructed on the path. Each agent is adversely affected by some subset of the facilities, and is unaffected by the others. We design two classes of mechanisms for choosing the facility locations given the reported agent preferences: utilitarian mechanisms that strive to maximize social welfare (i.e., to be efficient), and egalitarian mechanisms that strive to maximize the minimum welfare. For the utilitarian objective, we present a weakly group-strategyproof efficient mechanism for up to three facilities, we give strongly group-strategyproof mechanisms that achieve approximation ratios of $5/3$ and $2$ for $k=1$ and $k > 1$, respectively, and we prove that no strongly group-strategyproof mechanism achieves an approximation ratio less than $5/3$ for the case of a single facility. For the egalitarian objective, we present a strategyproof egalitarian mechanism for arbitrary $k$, and we prove that no weakly group-strategyproof mechanism achieves a $o(\sqrt{n})$ approximation ratio for two facilities. We extend our egalitarian results to the case where the agents are located on a cycle, and we extend our first egalitarian result to the case where the agents are located in the unit square.
[ { "created": "Sun, 12 Sep 2021 01:15:14 GMT", "version": "v1" }, { "created": "Sun, 21 May 2023 20:05:40 GMT", "version": "v2" } ]
2023-05-23
[ [ "Li", "Fu", "" ], [ "Plaxton", "C. Gregory", "" ], [ "Sinha", "Vaibhav B.", "" ] ]
We consider a facility location game in which $n$ agents reside at known locations on a path, and $k$ heterogeneous facilities are to be constructed on the path. Each agent is adversely affected by some subset of the facilities, and is unaffected by the others. We design two classes of mechanisms for choosing the facility locations given the reported agent preferences: utilitarian mechanisms that strive to maximize social welfare (i.e., to be efficient), and egalitarian mechanisms that strive to maximize the minimum welfare. For the utilitarian objective, we present a weakly group-strategyproof efficient mechanism for up to three facilities, we give strongly group-strategyproof mechanisms that achieve approximation ratios of $5/3$ and $2$ for $k=1$ and $k > 1$, respectively, and we prove that no strongly group-strategyproof mechanism achieves an approximation ratio less than $5/3$ for the case of a single facility. For the egalitarian objective, we present a strategyproof egalitarian mechanism for arbitrary $k$, and we prove that no weakly group-strategyproof mechanism achieves a $o(\sqrt{n})$ approximation ratio for two facilities. We extend our egalitarian results to the case where the agents are located on a cycle, and we extend our first egalitarian result to the case where the agents are located in the unit square.
hep-th/0309118
Bogdan Kulik
Z. Guralnik and B. Kulik
Properties of Chiral Wilson Loops
15 pages, two pictures, some references added
JHEP 0401 (2004) 065
10.1088/1126-6708/2004/01/065
null
hep-th
null
We study a class of Wilson Loops in N =4, D=4 Yang-Mills theory belonging to the chiral ring of a N=2, d=1 subalgebra. We show that the expectation value of these loops is independent of their shape. Using properties of the chiral ring, we also show that the expectation value is identically 1. We find the same result for chiral loops in maximally supersymmetric Yang-Mills theory in three, five and six dimensions. In seven dimensions, a generalized Konishi anomaly gives an equation for chiral loops which closely resembles the loop equations of the three dimensional Chern-Simons theory.
[ { "created": "Thu, 11 Sep 2003 11:57:36 GMT", "version": "v1" }, { "created": "Fri, 19 Sep 2003 12:09:33 GMT", "version": "v2" } ]
2009-11-10
[ [ "Guralnik", "Z.", "" ], [ "Kulik", "B.", "" ] ]
We study a class of Wilson Loops in N =4, D=4 Yang-Mills theory belonging to the chiral ring of a N=2, d=1 subalgebra. We show that the expectation value of these loops is independent of their shape. Using properties of the chiral ring, we also show that the expectation value is identically 1. We find the same result for chiral loops in maximally supersymmetric Yang-Mills theory in three, five and six dimensions. In seven dimensions, a generalized Konishi anomaly gives an equation for chiral loops which closely resembles the loop equations of the three dimensional Chern-Simons theory.
2402.07066
Guanyang Wang
Prathamesh Dharangutte, Jie Gao, Ruobin Gong, Guanyang Wang
Differentially Private Range Queries with Correlated Input Perturbation
26 pages, 8 figures
null
null
null
cs.CR cs.LG stat.ME
http://creativecommons.org/licenses/by/4.0/
This work proposes a class of locally differentially private mechanisms for linear queries, in particular range queries, that leverages correlated input perturbation to simultaneously achieve unbiasedness, consistency, statistical transparency, and control over utility requirements in terms of accuracy targets expressed either in certain query margins or as implied by the hierarchical database structure. The proposed Cascade Sampling algorithm instantiates the mechanism exactly and efficiently. Our bounds show that we obtain near-optimal utility while being empirically competitive against output perturbation methods.
[ { "created": "Sat, 10 Feb 2024 23:42:05 GMT", "version": "v1" } ]
2024-02-13
[ [ "Dharangutte", "Prathamesh", "" ], [ "Gao", "Jie", "" ], [ "Gong", "Ruobin", "" ], [ "Wang", "Guanyang", "" ] ]
This work proposes a class of locally differentially private mechanisms for linear queries, in particular range queries, that leverages correlated input perturbation to simultaneously achieve unbiasedness, consistency, statistical transparency, and control over utility requirements in terms of accuracy targets expressed either in certain query margins or as implied by the hierarchical database structure. The proposed Cascade Sampling algorithm instantiates the mechanism exactly and efficiently. Our bounds show that we obtain near-optimal utility while being empirically competitive against output perturbation methods.
1803.10753
Milena \v{C}uki\'c Dr
Milena B. Cukic, Mirjana M. Platisa, Aleksandar Kalauzi, Joji Oommen, Milos R. Ljubisavljevic
The comparison of Higuchi fractal dimension and Sample Entropy analysis of sEMG: effects of muscle contraction intensity and TMS
21 pages, 3 Figures
null
null
null
q-bio.NC
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The aim of the study was to examine how the complexity of surface electromyogram (sEMG) signal, estimated by Higuchi fractal dimension (HFD) and Sample Entropy (SampEn), change depending on muscle contraction intensity and external perturbation of the corticospinal activity during muscle contraction induced by single-pulse Transcranial Magnetic Stimulation (spTMS). HFD and SampEn were computed from sEMG signal recorded at three various levels of voluntary contraction before and after spTMS. After spTMS, both HFD and SampEn decreased at medium compared to the mild contraction. SampEn increased, while HFD did not change significantly at strong compared to medium contraction. spTMS significantly decreased both parameters at all contraction levels. When same parameters were computed from the mathematically generated sine-wave calibration curves, the results show that SampEn has better accuracy at lower (0-40 Hz) and HFD at higher (60-120 Hz) frequencies. Changes in the sEMG complexity associated with increased muscle contraction intensity cannot be accurately depicted by a single complexity measure. Examination of sEMG should entail both SampEn and HFD as they provide complementary information about different frequency components of sEMG. Further studies are needed to explain the implication of changes in nonlinear parameters and their relation to underlying sEMG physiological processes.
[ { "created": "Wed, 28 Mar 2018 17:44:44 GMT", "version": "v1" } ]
2018-03-29
[ [ "Cukic", "Milena B.", "" ], [ "Platisa", "Mirjana M.", "" ], [ "Kalauzi", "Aleksandar", "" ], [ "Oommen", "Joji", "" ], [ "Ljubisavljevic", "Milos R.", "" ] ]
The aim of the study was to examine how the complexity of surface electromyogram (sEMG) signal, estimated by Higuchi fractal dimension (HFD) and Sample Entropy (SampEn), change depending on muscle contraction intensity and external perturbation of the corticospinal activity during muscle contraction induced by single-pulse Transcranial Magnetic Stimulation (spTMS). HFD and SampEn were computed from sEMG signal recorded at three various levels of voluntary contraction before and after spTMS. After spTMS, both HFD and SampEn decreased at medium compared to the mild contraction. SampEn increased, while HFD did not change significantly at strong compared to medium contraction. spTMS significantly decreased both parameters at all contraction levels. When same parameters were computed from the mathematically generated sine-wave calibration curves, the results show that SampEn has better accuracy at lower (0-40 Hz) and HFD at higher (60-120 Hz) frequencies. Changes in the sEMG complexity associated with increased muscle contraction intensity cannot be accurately depicted by a single complexity measure. Examination of sEMG should entail both SampEn and HFD as they provide complementary information about different frequency components of sEMG. Further studies are needed to explain the implication of changes in nonlinear parameters and their relation to underlying sEMG physiological processes.
1806.06738
Antoaneta Serguieva
Tooba Faisal, Nicolas Courtois and Antoaneta Serguieva
The Evolution of Embedding Metadata in Blockchain Transactions
9 pages, 6 figures, 1 table, 2018 International Joint Conference on Neural Networks
null
null
null
cs.CR
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The use of blockchains is growing every day, and their utility has greatly expanded from sending and receiving crypto-coins to smart-contracts and decentralized autonomous organizations. Modern blockchains underpin a variety of applications: from designing a global identity to improving satellite connectivity. In our research we look at the ability of blockchains to store metadata in an increasing volume of transactions and with evolving focus of utilization. We further show that basic approaches to improving blockchain privacy also rely on embedding metadata. This paper identifies and classifies real-life blockchain transactions embedding metadata of a number of major protocols running essentially over the bitcoin blockchain. The empirical analysis here presents the evolution of metadata utilization in the recent years, and the discussion suggests steps towards preventing criminal use. Metadata are relevant to any blockchain, and our analysis considers primarily bitcoin as a case study. The paper concludes that simultaneously with both expanding legitimate utilization of embedded metadata and expanding blockchain functionality, the applied research on improving anonymity and security must also attempt to protect against blockchain abuse.
[ { "created": "Mon, 18 Jun 2018 14:38:02 GMT", "version": "v1" } ]
2018-06-19
[ [ "Faisal", "Tooba", "" ], [ "Courtois", "Nicolas", "" ], [ "Serguieva", "Antoaneta", "" ] ]
The use of blockchains is growing every day, and their utility has greatly expanded from sending and receiving crypto-coins to smart-contracts and decentralized autonomous organizations. Modern blockchains underpin a variety of applications: from designing a global identity to improving satellite connectivity. In our research we look at the ability of blockchains to store metadata in an increasing volume of transactions and with evolving focus of utilization. We further show that basic approaches to improving blockchain privacy also rely on embedding metadata. This paper identifies and classifies real-life blockchain transactions embedding metadata of a number of major protocols running essentially over the bitcoin blockchain. The empirical analysis here presents the evolution of metadata utilization in the recent years, and the discussion suggests steps towards preventing criminal use. Metadata are relevant to any blockchain, and our analysis considers primarily bitcoin as a case study. The paper concludes that simultaneously with both expanding legitimate utilization of embedded metadata and expanding blockchain functionality, the applied research on improving anonymity and security must also attempt to protect against blockchain abuse.
1602.03418
Swami Sankaranarayanan
Swami Sankaranarayanan, Azadeh Alavi, Rama Chellappa
Triplet Similarity Embedding for Face Verification
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this work, we present an unconstrained face verification algorithm and evaluate it on the recently released IJB-A dataset that aims to push the boundaries of face verification methods. The proposed algorithm couples a deep CNN-based approach with a low-dimensional discriminative embedding learnt using triplet similarity constraints in a large margin fashion. Aside from yielding performance improvement, this embedding provides significant advantages in terms of memory and post-processing operations like hashing and visualization. Experiments on the IJB-A dataset show that the proposed algorithm outperforms state of the art methods in verification and identification metrics, while requiring less training time.
[ { "created": "Wed, 10 Feb 2016 15:48:47 GMT", "version": "v1" }, { "created": "Sun, 13 Mar 2016 18:06:34 GMT", "version": "v2" } ]
2016-03-15
[ [ "Sankaranarayanan", "Swami", "" ], [ "Alavi", "Azadeh", "" ], [ "Chellappa", "Rama", "" ] ]
In this work, we present an unconstrained face verification algorithm and evaluate it on the recently released IJB-A dataset that aims to push the boundaries of face verification methods. The proposed algorithm couples a deep CNN-based approach with a low-dimensional discriminative embedding learnt using triplet similarity constraints in a large margin fashion. Aside from yielding performance improvement, this embedding provides significant advantages in terms of memory and post-processing operations like hashing and visualization. Experiments on the IJB-A dataset show that the proposed algorithm outperforms state of the art methods in verification and identification metrics, while requiring less training time.
1810.13098
Chao Li
Chao Li, Zhun Sun, Jinshi Yu, Ming Hou and Qibin Zhao
Low-Rank Embedding of Kernels in Convolutional Neural Networks under Random Shuffling
null
null
null
null
cs.LG stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Although the convolutional neural networks (CNNs) have become popular for various image processing and computer vision task recently, it remains a challenging problem to reduce the storage cost of the parameters for resource-limited platforms. In the previous studies, tensor decomposition (TD) has achieved promising compression performance by embedding the kernel of a convolutional layer into a low-rank subspace. However the employment of TD is naively on the kernel or its specified variants. Unlike the conventional approaches, this paper shows that the kernel can be embedded into more general or even random low-rank subspaces. We demonstrate this by compressing the convolutional layers via randomly-shuffled tensor decomposition (RsTD) for a standard classification task using CIFAR-10. In addition, we analyze how the spatial similarity of the training data influences the low-rank structure of the kernels. The experimental results show that the CNN can be significantly compressed even if the kernels are randomly shuffled. Furthermore, the RsTD-based method yields more stable classification accuracy than the conventional TD-based methods in a large range of compression ratios.
[ { "created": "Wed, 31 Oct 2018 04:05:54 GMT", "version": "v1" } ]
2018-11-01
[ [ "Li", "Chao", "" ], [ "Sun", "Zhun", "" ], [ "Yu", "Jinshi", "" ], [ "Hou", "Ming", "" ], [ "Zhao", "Qibin", "" ] ]
Although the convolutional neural networks (CNNs) have become popular for various image processing and computer vision task recently, it remains a challenging problem to reduce the storage cost of the parameters for resource-limited platforms. In the previous studies, tensor decomposition (TD) has achieved promising compression performance by embedding the kernel of a convolutional layer into a low-rank subspace. However the employment of TD is naively on the kernel or its specified variants. Unlike the conventional approaches, this paper shows that the kernel can be embedded into more general or even random low-rank subspaces. We demonstrate this by compressing the convolutional layers via randomly-shuffled tensor decomposition (RsTD) for a standard classification task using CIFAR-10. In addition, we analyze how the spatial similarity of the training data influences the low-rank structure of the kernels. The experimental results show that the CNN can be significantly compressed even if the kernels are randomly shuffled. Furthermore, the RsTD-based method yields more stable classification accuracy than the conventional TD-based methods in a large range of compression ratios.
1912.09515
Charles Rabideau
Marine De Clerck, Charles Rabideau, Niklas Tanger
Caustics bounding entanglement wedges
32 pages, 8 figures. Typos fixed and minor clarifications added
null
10.1007/JHEP06(2020)166
null
hep-th
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We study the caustics on the boundaries of entanglement wedges in the context of holography in asymptotically AdS$_3$ spacetimes. These entanglement wedges play an important role in our understanding of the emergence of bulk locality. A procedure was proposed by Sanches and Weinberg [arXiv:1703.07780] for identifying boundary operators which are local in the bulk, which also applies to certain regions that lie beyond the reach of HRT surfaces by taking advantage of the lightsheets which bound entanglement wedges. We identify the caustics which terminate these lightsheets in conical deficit and BTZ black hole spacetimes and find that in some examples these caustics lead to a sharp corner in the entanglement wedge. The unexpected shape of these entanglement wedges leads, in those cases, to a breakdown of this procedure. Many of the properties of the rich variety of caustics possible in higher dimensions remains to be explored which, as this work demonstrates, could lead to more unexpected features in the shapes of entanglement wedges.
[ { "created": "Thu, 19 Dec 2019 19:42:39 GMT", "version": "v1" }, { "created": "Wed, 20 May 2020 14:46:05 GMT", "version": "v2" } ]
2020-07-15
[ [ "De Clerck", "Marine", "" ], [ "Rabideau", "Charles", "" ], [ "Tanger", "Niklas", "" ] ]
We study the caustics on the boundaries of entanglement wedges in the context of holography in asymptotically AdS$_3$ spacetimes. These entanglement wedges play an important role in our understanding of the emergence of bulk locality. A procedure was proposed by Sanches and Weinberg [arXiv:1703.07780] for identifying boundary operators which are local in the bulk, which also applies to certain regions that lie beyond the reach of HRT surfaces by taking advantage of the lightsheets which bound entanglement wedges. We identify the caustics which terminate these lightsheets in conical deficit and BTZ black hole spacetimes and find that in some examples these caustics lead to a sharp corner in the entanglement wedge. The unexpected shape of these entanglement wedges leads, in those cases, to a breakdown of this procedure. Many of the properties of the rich variety of caustics possible in higher dimensions remains to be explored which, as this work demonstrates, could lead to more unexpected features in the shapes of entanglement wedges.
2205.05837
Peter Tsimiklis
Florian Girelli, Matteo Laudonio, Adrian Tanasa, Panagiotis Tsimiklis
Group field theory on 2-groups
null
null
null
null
hep-th
http://creativecommons.org/licenses/by-sa/4.0/
Group field theories are quantum field theories built on groups. They can be seen as a tool to generate topological state-sums or quantum gravity models. For four dimensional manifolds, different arguments have pointed towards 2-groups (such as crossed modules) as the relevant symmetry structure to probe four dimensional topological features. Here, we introduce a group field theory built on crossed modules which generate a four dimensional topological model, as we prove that the Feynman diagram amplitudes can be related by Pachner moves. This model is presumably the dual version of the Yetter-Mackaay model.
[ { "created": "Thu, 12 May 2022 02:09:48 GMT", "version": "v1" } ]
2022-05-13
[ [ "Girelli", "Florian", "" ], [ "Laudonio", "Matteo", "" ], [ "Tanasa", "Adrian", "" ], [ "Tsimiklis", "Panagiotis", "" ] ]
Group field theories are quantum field theories built on groups. They can be seen as a tool to generate topological state-sums or quantum gravity models. For four dimensional manifolds, different arguments have pointed towards 2-groups (such as crossed modules) as the relevant symmetry structure to probe four dimensional topological features. Here, we introduce a group field theory built on crossed modules which generate a four dimensional topological model, as we prove that the Feynman diagram amplitudes can be related by Pachner moves. This model is presumably the dual version of the Yetter-Mackaay model.
2208.09237
Wei Zhang
Wei Zhang, Jason K. Wong, Xumeng Wang, Youcheng Gong, Rongchen Zhu, Kai Liu, Zihan Yan, Siwei Tan, Huamin Qu, Siming Chen, and Wei Chen
CohortVA: A Visual Analytic System for Interactive Exploration of Cohorts based on Historical Data
null
null
null
null
cs.HC
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In history research, cohort analysis seeks to identify social structures and figure mobilities by studying the group-based behavior of historical figures. Prior works mainly employ automatic data mining approaches, lacking effective visual explanation. In this paper, we present CohortVA, an interactive visual analytic approach that enables historians to incorporate expertise and insight into the iterative exploration process. The kernel of CohortVA is a novel identification model that generates candidate cohorts and constructs cohort features by means of pre-built knowledge graphs constructed from large-scale history databases. We propose a set of coordinated views to illustrate identified cohorts and features coupled with historical events and figure profiles. Two case studies and interviews with historians demonstrate that CohortVA can greatly enhance the capabilities of cohort identifications, figure authentications, and hypothesis generation.
[ { "created": "Fri, 19 Aug 2022 09:27:42 GMT", "version": "v1" } ]
2022-08-22
[ [ "Zhang", "Wei", "" ], [ "Wong", "Jason K.", "" ], [ "Wang", "Xumeng", "" ], [ "Gong", "Youcheng", "" ], [ "Zhu", "Rongchen", "" ], [ "Liu", "Kai", "" ], [ "Yan", "Zihan", "" ], [ "Tan", "Siwei", "" ], [ "Qu", "Huamin", "" ], [ "Chen", "Siming", "" ], [ "Chen", "Wei", "" ] ]
In history research, cohort analysis seeks to identify social structures and figure mobilities by studying the group-based behavior of historical figures. Prior works mainly employ automatic data mining approaches, lacking effective visual explanation. In this paper, we present CohortVA, an interactive visual analytic approach that enables historians to incorporate expertise and insight into the iterative exploration process. The kernel of CohortVA is a novel identification model that generates candidate cohorts and constructs cohort features by means of pre-built knowledge graphs constructed from large-scale history databases. We propose a set of coordinated views to illustrate identified cohorts and features coupled with historical events and figure profiles. Two case studies and interviews with historians demonstrate that CohortVA can greatly enhance the capabilities of cohort identifications, figure authentications, and hypothesis generation.
1704.04332
Dieyan Liang
Dieyan Liang, Hong Shen
Point Sweep Coverage on Path
null
null
null
null
cs.DS
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
An important application of wireless sensor networks is the deployment of mobile sensors to periodically monitor (cover) a set of points of interest (PoIs). The problem of Point Sweep Coverage is to deploy fewest sensors to periodically cover the set of PoIs. For PoIs in a Eulerian graph, this problem is known NP-Hard even if all sensors are with uniform velocity. In this paper, we study the problem when PoIs are on a line and prove that the decision version of the problem is NP-Complete if the sensors are with different velocities. We first formulate the problem of Max-PoI sweep coverage on path (MPSCP) to find the maximum number of PoIs covered by a given set of sensors, and then show it is NP-Hard. We also extend it to the weighted case, Max-Weight sweep coverage on path (MWSCP) problem to maximum the sum of the weight of PoIs covered. For sensors with uniform velocity, we give a polynomial-time optimal solution to MWSCP. For sensors with constant kinds of velocities, we present a $\frac{1}{2}$-approximation algorithm. For the general case of arbitrary velocities, we propose two algorithms. One is a $\frac{1}{2\alpha}$-approximation algorithm family scheme, where integer $\alpha\ge2$ is the tradeoff factor to balance the time complexity and approximation ratio. The other is a $\frac{1}{2}(1-1/e)$-approximation algorithm by randomized analysis.
[ { "created": "Fri, 14 Apr 2017 02:24:56 GMT", "version": "v1" }, { "created": "Tue, 27 Jun 2017 04:45:46 GMT", "version": "v2" } ]
2017-06-28
[ [ "Liang", "Dieyan", "" ], [ "Shen", "Hong", "" ] ]
An important application of wireless sensor networks is the deployment of mobile sensors to periodically monitor (cover) a set of points of interest (PoIs). The problem of Point Sweep Coverage is to deploy fewest sensors to periodically cover the set of PoIs. For PoIs in a Eulerian graph, this problem is known NP-Hard even if all sensors are with uniform velocity. In this paper, we study the problem when PoIs are on a line and prove that the decision version of the problem is NP-Complete if the sensors are with different velocities. We first formulate the problem of Max-PoI sweep coverage on path (MPSCP) to find the maximum number of PoIs covered by a given set of sensors, and then show it is NP-Hard. We also extend it to the weighted case, Max-Weight sweep coverage on path (MWSCP) problem to maximum the sum of the weight of PoIs covered. For sensors with uniform velocity, we give a polynomial-time optimal solution to MWSCP. For sensors with constant kinds of velocities, we present a $\frac{1}{2}$-approximation algorithm. For the general case of arbitrary velocities, we propose two algorithms. One is a $\frac{1}{2\alpha}$-approximation algorithm family scheme, where integer $\alpha\ge2$ is the tradeoff factor to balance the time complexity and approximation ratio. The other is a $\frac{1}{2}(1-1/e)$-approximation algorithm by randomized analysis.
1909.04436
Martin Shepperd
Martin Shepperd, Yuchen Guo, Ning Li, Mahir Arzoky, Andrea Capiluppi, Steve Counsell, Giuseppe Destefanis, Stephen Swift, Allan Tucker, and Leila Yousefi
The Prevalence of Errors in Machine Learning Experiments
20th International Conference on Intelligent Data Engineering and Automated Learning (IDEAL), 14--16 November 2019
null
null
null
cs.LG cs.AI stat.AP stat.ML
http://creativecommons.org/licenses/by/4.0/
Context: Conducting experiments is central to research machine learning research to benchmark, evaluate and compare learning algorithms. Consequently it is important we conduct reliable, trustworthy experiments. Objective: We investigate the incidence of errors in a sample of machine learning experiments in the domain of software defect prediction. Our focus is simple arithmetical and statistical errors. Method: We analyse 49 papers describing 2456 individual experimental results from a previously undertaken systematic review comparing supervised and unsupervised defect prediction classifiers. We extract the confusion matrices and test for relevant constraints, e.g., the marginal probabilities must sum to one. We also check for multiple statistical significance testing errors. Results: We find that a total of 22 out of 49 papers contain demonstrable errors. Of these 7 were statistical and 16 related to confusion matrix inconsistency (one paper contained both classes of error). Conclusions: Whilst some errors may be of a relatively trivial nature, e.g., transcription errors their presence does not engender confidence. We strongly urge researchers to follow open science principles so errors can be more easily be detected and corrected, thus as a community reduce this worryingly high error rate with our computational experiments.
[ { "created": "Tue, 10 Sep 2019 12:32:00 GMT", "version": "v1" } ]
2019-09-11
[ [ "Shepperd", "Martin", "" ], [ "Guo", "Yuchen", "" ], [ "Li", "Ning", "" ], [ "Arzoky", "Mahir", "" ], [ "Capiluppi", "Andrea", "" ], [ "Counsell", "Steve", "" ], [ "Destefanis", "Giuseppe", "" ], [ "Swift", "Stephen", "" ], [ "Tucker", "Allan", "" ], [ "Yousefi", "Leila", "" ] ]
Context: Conducting experiments is central to research machine learning research to benchmark, evaluate and compare learning algorithms. Consequently it is important we conduct reliable, trustworthy experiments. Objective: We investigate the incidence of errors in a sample of machine learning experiments in the domain of software defect prediction. Our focus is simple arithmetical and statistical errors. Method: We analyse 49 papers describing 2456 individual experimental results from a previously undertaken systematic review comparing supervised and unsupervised defect prediction classifiers. We extract the confusion matrices and test for relevant constraints, e.g., the marginal probabilities must sum to one. We also check for multiple statistical significance testing errors. Results: We find that a total of 22 out of 49 papers contain demonstrable errors. Of these 7 were statistical and 16 related to confusion matrix inconsistency (one paper contained both classes of error). Conclusions: Whilst some errors may be of a relatively trivial nature, e.g., transcription errors their presence does not engender confidence. We strongly urge researchers to follow open science principles so errors can be more easily be detected and corrected, thus as a community reduce this worryingly high error rate with our computational experiments.
1711.02690
David Kutasov
Meseret Asrat, Amit Giveon, Nissan Itzhaki and David Kutasov
Holography Beyond AdS
16 pages; v2: reference updated
null
10.1016/j.nuclphysb.2018.05.005
null
hep-th
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We continue our study of string theory in a background that interpolates between $AdS_3$ in the infrared and a linear dilaton spacetime $R^{1,1}\times R_\phi$ in the UV. This background corresponds via holography to a $CFT_2$ deformed by a certain irrelevant operator of dimension $(2,2)$. We show that for two point functions of local operators in the infrared CFT, conformal perturbation theory in this irrelevant operator has a finite radius of convergence in momentum space, and one can use it to flow up the renormalization group. The spectral density develops an imaginary part above a certain critical value of the spectral parameter; this appears to be related to the non-locality of the theory. In position space, conformal perturbation theory has a vanishing radius of convergence; the leading non-perturbative effect is an imaginary part of the two point function.
[ { "created": "Tue, 7 Nov 2017 19:14:44 GMT", "version": "v1" }, { "created": "Mon, 27 Nov 2017 21:52:04 GMT", "version": "v2" } ]
2018-07-04
[ [ "Asrat", "Meseret", "" ], [ "Giveon", "Amit", "" ], [ "Itzhaki", "Nissan", "" ], [ "Kutasov", "David", "" ] ]
We continue our study of string theory in a background that interpolates between $AdS_3$ in the infrared and a linear dilaton spacetime $R^{1,1}\times R_\phi$ in the UV. This background corresponds via holography to a $CFT_2$ deformed by a certain irrelevant operator of dimension $(2,2)$. We show that for two point functions of local operators in the infrared CFT, conformal perturbation theory in this irrelevant operator has a finite radius of convergence in momentum space, and one can use it to flow up the renormalization group. The spectral density develops an imaginary part above a certain critical value of the spectral parameter; this appears to be related to the non-locality of the theory. In position space, conformal perturbation theory has a vanishing radius of convergence; the leading non-perturbative effect is an imaginary part of the two point function.
1805.11203
Xiang Zhang
Xiang Zhang, Philip A. Chou, Ming-Ting Sun, Maolong Tang, Shanshe Wang, Siwei Ma, Wen Gao
Surface Light Field Compression using a Point Cloud Codec
null
null
null
null
cs.MM
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Light field (LF) representations aim to provide photo-realistic, free-viewpoint viewing experiences. However, the most popular LF representations are images from multiple views. Multi-view image-based representations generally need to restrict the range or degrees of freedom of the viewing experience to what can be interpolated in the image domain, essentially because they lack explicit geometry information. We present a new surface light field (SLF) representation based on explicit geometry, and a method for SLF compression. First, we map the multi-view images of a scene onto a 3D geometric point cloud. The color of each point in the point cloud is a function of viewing direction known as a view map. We represent each view map efficiently in a B-Spline wavelet basis. This representation is capable of modeling diverse surface materials and complex lighting conditions in a highly scalable and adaptive manner. The coefficients of the B-Spline wavelet representation are then compressed spatially. To increase the spatial correlation and thus improve compression efficiency, we introduce a smoothing term to make the coefficients more similar across the 3D space. We compress the coefficients spatially using existing point cloud compression (PCC) methods. On the decoder side, the scene is rendered efficiently from any viewing direction by reconstructing the view map at each point. In contrast to multi-view image-based LF approaches, our method supports photo-realistic rendering of real-world scenes from arbitrary viewpoints, i.e., with an unlimited six degrees of freedom (6DOF). In terms of rate and distortion, experimental results show that our method achieves superior performance with lighter decoder complexity compared with a reference image-plus-geometry compression (IGC) scheme, indicating its potential in practical virtual and augmented reality applications.
[ { "created": "Tue, 29 May 2018 00:08:30 GMT", "version": "v1" } ]
2018-05-30
[ [ "Zhang", "Xiang", "" ], [ "Chou", "Philip A.", "" ], [ "Sun", "Ming-Ting", "" ], [ "Tang", "Maolong", "" ], [ "Wang", "Shanshe", "" ], [ "Ma", "Siwei", "" ], [ "Gao", "Wen", "" ] ]
Light field (LF) representations aim to provide photo-realistic, free-viewpoint viewing experiences. However, the most popular LF representations are images from multiple views. Multi-view image-based representations generally need to restrict the range or degrees of freedom of the viewing experience to what can be interpolated in the image domain, essentially because they lack explicit geometry information. We present a new surface light field (SLF) representation based on explicit geometry, and a method for SLF compression. First, we map the multi-view images of a scene onto a 3D geometric point cloud. The color of each point in the point cloud is a function of viewing direction known as a view map. We represent each view map efficiently in a B-Spline wavelet basis. This representation is capable of modeling diverse surface materials and complex lighting conditions in a highly scalable and adaptive manner. The coefficients of the B-Spline wavelet representation are then compressed spatially. To increase the spatial correlation and thus improve compression efficiency, we introduce a smoothing term to make the coefficients more similar across the 3D space. We compress the coefficients spatially using existing point cloud compression (PCC) methods. On the decoder side, the scene is rendered efficiently from any viewing direction by reconstructing the view map at each point. In contrast to multi-view image-based LF approaches, our method supports photo-realistic rendering of real-world scenes from arbitrary viewpoints, i.e., with an unlimited six degrees of freedom (6DOF). In terms of rate and distortion, experimental results show that our method achieves superior performance with lighter decoder complexity compared with a reference image-plus-geometry compression (IGC) scheme, indicating its potential in practical virtual and augmented reality applications.
1710.04134
Eric Sillekens
Eric Sillekens, Daniel Semrau, Gabriele Liga, Nikita A. Shevchenko, Zhe Li, Alex Alvarado, Polina Bayvel, Robert. I. Killey, Domani\c{c} Lavery
A Simple Nonlinearity-Tailored Probabilistic Shaping Distribution for Square QAM
null
null
10.1364/OFC.2018.M3C.4
null
cs.IT math.IT
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
A new probabilistic shaping distribution that outperforms Maxwell-Boltzmann is studied for the nonlinear fiber channel. Additional gains of 0.1 bit/symbol MI or 0.2 dB SNR for both DP-256QAM and DP-1024QAM are reported after 200 km nonlinear fiber transmission.
[ { "created": "Wed, 11 Oct 2017 15:42:46 GMT", "version": "v1" }, { "created": "Thu, 12 Oct 2017 07:42:01 GMT", "version": "v2" } ]
2018-11-06
[ [ "Sillekens", "Eric", "" ], [ "Semrau", "Daniel", "" ], [ "Liga", "Gabriele", "" ], [ "Shevchenko", "Nikita A.", "" ], [ "Li", "Zhe", "" ], [ "Alvarado", "Alex", "" ], [ "Bayvel", "Polina", "" ], [ "Killey", "Robert. I.", "" ], [ "Lavery", "Domaniç", "" ] ]
A new probabilistic shaping distribution that outperforms Maxwell-Boltzmann is studied for the nonlinear fiber channel. Additional gains of 0.1 bit/symbol MI or 0.2 dB SNR for both DP-256QAM and DP-1024QAM are reported after 200 km nonlinear fiber transmission.
2011.08967
Nuno Crokidakis
Marcelo A. Pires, Nuno Crokidakis, Silvio M. Duarte Queir\'os
Diffusion plays an unusual role in ecological quasi-neutral competition in metapopulations
6 figures, 10 pages, to appear in Nonlinear Dynamics
Nonlinear Dynamics 103, 1219 (2021)
10.1007/s11071-020-06105-4
null
q-bio.PE physics.soc-ph
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We investigate the phenomenology emerging from a 2-species dynamics under the scenario of a quasi-neutral competition within a metapopulation framework. We employ stochastic and deterministic approaches, namely spatially-constrained individual-based Monte Carlo simulations and coupled mean-field ODEs. Our results show the multifold interplay between competition, birth-death dynamics and spatial constraints induces a nonmonotonic relation between the ecological majority-minority switching and the diffusion between patches. This means that diffusion can set off birth-death ratios and enhance the preservation of a species.
[ { "created": "Tue, 17 Nov 2020 22:04:20 GMT", "version": "v1" } ]
2021-02-05
[ [ "Pires", "Marcelo A.", "" ], [ "Crokidakis", "Nuno", "" ], [ "Queirós", "Silvio M. Duarte", "" ] ]
We investigate the phenomenology emerging from a 2-species dynamics under the scenario of a quasi-neutral competition within a metapopulation framework. We employ stochastic and deterministic approaches, namely spatially-constrained individual-based Monte Carlo simulations and coupled mean-field ODEs. Our results show the multifold interplay between competition, birth-death dynamics and spatial constraints induces a nonmonotonic relation between the ecological majority-minority switching and the diffusion between patches. This means that diffusion can set off birth-death ratios and enhance the preservation of a species.
2402.11297
Husein Zolkepli
Husein Zolkepli, Aisyah Razak, Kamarul Adha, Ariff Nazhan
MMMModal -- Multi-Images Multi-Audio Multi-turn Multi-Modal
null
null
null
null
cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Our contribution introduces a groundbreaking multimodal large language model designed to comprehend multi-images, multi-audio, and multi-images-multi-audio within a single multiturn session. Leveraging state-of-the-art models, we utilize the SigLIP encoder for visual inputs and the Whisper Encoder for audio inputs. Notably, this multimodal large language model is bilingual, proficient in understanding both English and Malay simultaneously. We proudly unveil two versions of this model: TinyLlama with 1.1B parameters, and Mistral with 7B parameters. With its ability to navigate diverse modalities and languages, our model represents a significant advancement for the Malaysian context and beyond. All models released at https://huggingface.co/collections/mesolitica/multimodal-malaysian-llm-65c6f893e03f78fa9e5c8859
[ { "created": "Sat, 17 Feb 2024 14:37:38 GMT", "version": "v1" } ]
2024-02-20
[ [ "Zolkepli", "Husein", "" ], [ "Razak", "Aisyah", "" ], [ "Adha", "Kamarul", "" ], [ "Nazhan", "Ariff", "" ] ]
Our contribution introduces a groundbreaking multimodal large language model designed to comprehend multi-images, multi-audio, and multi-images-multi-audio within a single multiturn session. Leveraging state-of-the-art models, we utilize the SigLIP encoder for visual inputs and the Whisper Encoder for audio inputs. Notably, this multimodal large language model is bilingual, proficient in understanding both English and Malay simultaneously. We proudly unveil two versions of this model: TinyLlama with 1.1B parameters, and Mistral with 7B parameters. With its ability to navigate diverse modalities and languages, our model represents a significant advancement for the Malaysian context and beyond. All models released at https://huggingface.co/collections/mesolitica/multimodal-malaysian-llm-65c6f893e03f78fa9e5c8859
0712.0350
Jerzy Lukierski
Marcin Daszkiewicz, Jerzy Lukierski and Mariusz Woronowicz
Quantization of kappa-deformed free fields and kappa-deformed oscillators
9 pages. Talk presented at Supersymmetry and Quantum Supersymmetry 2007 (SQS'07) Conference (Dubna, 30.07-4.08.2007) and IV-th Central European Seminar "Commutative and Noncommutative Quantum Fields" (Vienna, 30.11-2.12.2007). To be published in the proceedings of SQS'07 (2008)
null
null
null
hep-th
null
We describe the deformed E.T. quantization rules for kappa-deformed free quantum fields, and relate these rules with the kappa-deformed algebra of field oscillators.
[ { "created": "Mon, 3 Dec 2007 17:15:55 GMT", "version": "v1" } ]
2007-12-04
[ [ "Daszkiewicz", "Marcin", "" ], [ "Lukierski", "Jerzy", "" ], [ "Woronowicz", "Mariusz", "" ] ]
We describe the deformed E.T. quantization rules for kappa-deformed free quantum fields, and relate these rules with the kappa-deformed algebra of field oscillators.
2207.05764
Monica Jinwoo Kang
Monica Jinwoo Kang, Craig Lawrie, Ki-Hong Lee, Matteo Sacchi, and Jaewon Song
Higgs, Coulomb, and Hall-Littlewood
49 pages + appendices + references, 20 figures, and 5 tables
Physical Review D 106, no.10, 106021 (2022)
10.1103/PhysRevD.106.106021
CALT-TH-2022-024; DESY-22-110
hep-th
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The Higgs branch of 4d $\mathcal{N}=2$ SCFTs can be analyzed via the Hilbert series of the Higgs branch or, in special cases, by computing the Hall-Littlewood index. For any class $\mathcal{S}$ theory corresponding to a genus-zero Riemann surface, they are conjectured to be identical. We present several families of counterexamples. We find that for any class $\mathcal{S}$ theory with four or more $\mathbb{Z}_2$-twisted punctures, they do not match. We construct 3d mirrors for such theories and analyze their Coulomb branch Hilbert series to compute the Higgs branch Hilbert series of the 4d theory. We further construct $a=c$ theories in class $\mathcal{S}$ using the twisted punctures, and these theories, which includes the $\hat{D}_4(SU(2n+1))$ theories, have Hall--Littlewood index different from the Hilbert series of the Higgs branch. We conjecture that this is the case for all $a=c$ theories with non-empty Higgs branch, including $\mathcal{N}\ge 3$ SCFTs.
[ { "created": "Tue, 12 Jul 2022 18:00:01 GMT", "version": "v1" } ]
2023-02-09
[ [ "Kang", "Monica Jinwoo", "" ], [ "Lawrie", "Craig", "" ], [ "Lee", "Ki-Hong", "" ], [ "Sacchi", "Matteo", "" ], [ "Song", "Jaewon", "" ] ]
The Higgs branch of 4d $\mathcal{N}=2$ SCFTs can be analyzed via the Hilbert series of the Higgs branch or, in special cases, by computing the Hall-Littlewood index. For any class $\mathcal{S}$ theory corresponding to a genus-zero Riemann surface, they are conjectured to be identical. We present several families of counterexamples. We find that for any class $\mathcal{S}$ theory with four or more $\mathbb{Z}_2$-twisted punctures, they do not match. We construct 3d mirrors for such theories and analyze their Coulomb branch Hilbert series to compute the Higgs branch Hilbert series of the 4d theory. We further construct $a=c$ theories in class $\mathcal{S}$ using the twisted punctures, and these theories, which includes the $\hat{D}_4(SU(2n+1))$ theories, have Hall--Littlewood index different from the Hilbert series of the Higgs branch. We conjecture that this is the case for all $a=c$ theories with non-empty Higgs branch, including $\mathcal{N}\ge 3$ SCFTs.
1405.5513
Seyed Aidin Sajedi
Fahimeh Abdollahi and Seyed Aidin Sajedi
Correlation of multiple sclerosis (MS) incidence trends with solar and geomagnetic indices: time to revise the method of reporting MS epidemiological data
Single PDF, 8 pages, 3 figures
Iranian Journal of Neurology 2014; 13(2):64-69
null
null
q-bio.TO q-bio.NC
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Background: Recently, we introduced solar related geomagnetic disturbances (GMD) as a potential environmental risk factor for multiple sclerosis (MS). The aim of this study was to test probable correlation between solar activities and GMD with long-term variations of MS incidence. Methods: After a systematic review, we studied the association between alterations in solar wind velocity (Vsw) and planetary A index (Ap, a GMD index) with MS incidence in Tehran and western Greece, during the 23rd solar cycle (1996-2008), by an ecological-correlational study. Results: We found moderate to strong correlations among MS incidence of Tehran with Vsw (Rs=0.665, p=0.013), with one year delay, and also with Ap (Rs=0.864, p=0.001) with 2 year delay. There were very strong correlations among MS incidence data of Greece with Vsw (R=0.906, p<0.001) and with Ap (R=0.844, p=0.001), both with one year lag. Conclusion: It is the first time that a hypothesis has introduced an environmental factor that may describe MS incidence alterations; however, it should be reminded that correlation does not mean necessarily the existence of a causal relationship. Important message of these findings for researchers is to provide MS incidence reports with higher resolution for consecutive years, based on the time of disease onset and relapses, not just the time of diagnosis. Then, it would be possible to further investigate the validity of GMD hypothesis or any other probable environmental risk factors. Keywords: Correlation analysis, Multiple sclerosis, Incidence, Geomagnetic disturbance, Geomagnetic activity, Solar wind velocity, Environmental risk factor.
[ { "created": "Sun, 18 May 2014 21:28:05 GMT", "version": "v1" } ]
2014-05-22
[ [ "Abdollahi", "Fahimeh", "" ], [ "Sajedi", "Seyed Aidin", "" ] ]
Background: Recently, we introduced solar related geomagnetic disturbances (GMD) as a potential environmental risk factor for multiple sclerosis (MS). The aim of this study was to test probable correlation between solar activities and GMD with long-term variations of MS incidence. Methods: After a systematic review, we studied the association between alterations in solar wind velocity (Vsw) and planetary A index (Ap, a GMD index) with MS incidence in Tehran and western Greece, during the 23rd solar cycle (1996-2008), by an ecological-correlational study. Results: We found moderate to strong correlations among MS incidence of Tehran with Vsw (Rs=0.665, p=0.013), with one year delay, and also with Ap (Rs=0.864, p=0.001) with 2 year delay. There were very strong correlations among MS incidence data of Greece with Vsw (R=0.906, p<0.001) and with Ap (R=0.844, p=0.001), both with one year lag. Conclusion: It is the first time that a hypothesis has introduced an environmental factor that may describe MS incidence alterations; however, it should be reminded that correlation does not mean necessarily the existence of a causal relationship. Important message of these findings for researchers is to provide MS incidence reports with higher resolution for consecutive years, based on the time of disease onset and relapses, not just the time of diagnosis. Then, it would be possible to further investigate the validity of GMD hypothesis or any other probable environmental risk factors. Keywords: Correlation analysis, Multiple sclerosis, Incidence, Geomagnetic disturbance, Geomagnetic activity, Solar wind velocity, Environmental risk factor.
1901.01877
Siyao Li
Siyao Li, Daniela Tuninetti and Natasha Devroye
On the Capacity Region of the Layered Packet Erasure Broadcast Channel with Feedback
6 pages, 1 figure, submitted to 2019 IEEE International Conference on Communications (ICC)
null
null
null
cs.IT math.IT
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, the capacity region of the Layered Packet Erasure Broadcast Channel (LPE-BC) with Channel Output Feedback (COF) available at the transmitter is investigated. The LPE-BC is a high-SNR approximation of the fading Gaussian BC recently proposed by Tse and Yates, who characterized the capacity region for any number of users and any number of layers when there is no COF. This paper derives capacity inner and outer bounds for the LPE-BC with COF for the case of two users and any number of layers. The inner bounds generalize past results for the two-user erasure BC, which is a special case of the LPE-BC with COF with only one layer. The novelty lies in the use of \emph{inter-user \& inter-layer network coding} retransmissions (for those packets that have only been received by the unintended user), where each random linear combination may involve packets intended for any user originally sent on any of the layers. Analytical and numerical examples show that the proposed outer bound is optimal for some LPE-BCs.
[ { "created": "Mon, 7 Jan 2019 15:22:27 GMT", "version": "v1" } ]
2019-01-08
[ [ "Li", "Siyao", "" ], [ "Tuninetti", "Daniela", "" ], [ "Devroye", "Natasha", "" ] ]
In this paper, the capacity region of the Layered Packet Erasure Broadcast Channel (LPE-BC) with Channel Output Feedback (COF) available at the transmitter is investigated. The LPE-BC is a high-SNR approximation of the fading Gaussian BC recently proposed by Tse and Yates, who characterized the capacity region for any number of users and any number of layers when there is no COF. This paper derives capacity inner and outer bounds for the LPE-BC with COF for the case of two users and any number of layers. The inner bounds generalize past results for the two-user erasure BC, which is a special case of the LPE-BC with COF with only one layer. The novelty lies in the use of \emph{inter-user \& inter-layer network coding} retransmissions (for those packets that have only been received by the unintended user), where each random linear combination may involve packets intended for any user originally sent on any of the layers. Analytical and numerical examples show that the proposed outer bound is optimal for some LPE-BCs.
q-bio/0509013
Atul Narang
Jason T. Noel, Brenton Cox, Atul Narang
Identification of the growth-limiting step in continuous cultures from initial rates measured in response to substrate-excess conditions
12 pages
null
null
null
q-bio.MN
null
When steady state chemostat cultures are abruptly exposed to substrate-excess conditions, they exhibit long lags before adjusting to the new environment. The identity of the rate-limiting step for this slow response can be inferred from the initial yields and specific growth rates measured by exposing steady state cultures at various dilution rates to substrate-excess conditions. We measured these parameters for glucose-limited cultures of E. coli ML308 growing at various dilution rates between 0.03 and 0.6 1/hr. In all the cases, the initial yields were 20-30% less than the steady state yields. The decline of the yield implies that overflow metabolism is triggered in response to excess glucose. It is therefore unlikely that the initial response of the cells is limited by substrate uptake. The initial specific growth rates of cultures growing at low dilution rates (D = 0.03, 0.05, 0.075, 0.1, 0.3 1/hr) were significantly higher than the steady state specific growth rates. However, the increment in the specific growth rate decreased with the dilution rate, and at D=0.6 1/hr, there was no improvement in the specific growth rate. The initial specific growth rates varied hyperbolically with the dilution, decreasing sharply at dilution rates below 0.1 1/hr and saturating at D=0.6 1/hr. This is consistent with a picture in which the initial response is limited by the activity of glutamate dehydrogenase.
[ { "created": "Mon, 12 Sep 2005 20:22:46 GMT", "version": "v1" } ]
2007-05-23
[ [ "Noel", "Jason T.", "" ], [ "Cox", "Brenton", "" ], [ "Narang", "Atul", "" ] ]
When steady state chemostat cultures are abruptly exposed to substrate-excess conditions, they exhibit long lags before adjusting to the new environment. The identity of the rate-limiting step for this slow response can be inferred from the initial yields and specific growth rates measured by exposing steady state cultures at various dilution rates to substrate-excess conditions. We measured these parameters for glucose-limited cultures of E. coli ML308 growing at various dilution rates between 0.03 and 0.6 1/hr. In all the cases, the initial yields were 20-30% less than the steady state yields. The decline of the yield implies that overflow metabolism is triggered in response to excess glucose. It is therefore unlikely that the initial response of the cells is limited by substrate uptake. The initial specific growth rates of cultures growing at low dilution rates (D = 0.03, 0.05, 0.075, 0.1, 0.3 1/hr) were significantly higher than the steady state specific growth rates. However, the increment in the specific growth rate decreased with the dilution rate, and at D=0.6 1/hr, there was no improvement in the specific growth rate. The initial specific growth rates varied hyperbolically with the dilution, decreasing sharply at dilution rates below 0.1 1/hr and saturating at D=0.6 1/hr. This is consistent with a picture in which the initial response is limited by the activity of glutamate dehydrogenase.
2002.02100
S.H. Shabbeer Basha
S.H. Shabbeer Basha, Viswanath Pulabaigari, Snehasis Mukherjee
An Information-rich Sampling Technique over Spatio-Temporal CNN for Classification of Human Actions in Videos
null
null
null
null
cs.CV cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We propose a novel scheme for human action recognition in videos, using a 3-dimensional Convolutional Neural Network (3D CNN) based classifier. Traditionally in deep learning based human activity recognition approaches, either a few random frames or every $k^{th}$ frame of the video is considered for training the 3D CNN, where $k$ is a small positive integer, like 4, 5, or 6. This kind of sampling reduces the volume of the input data, which speeds-up training of the network and also avoids over-fitting to some extent, thus enhancing the performance of the 3D CNN model. In the proposed video sampling technique, consecutive $k$ frames of a video are aggregated into a single frame by computing a Gaussian-weighted summation of the $k$ frames. The resulting frame (aggregated frame) preserves the information in a better way than the conventional approaches and experimentally shown to perform better. In this paper, a 3D CNN architecture is proposed to extract the spatio-temporal features and follows Long Short-Term Memory (LSTM) to recognize human actions. The proposed 3D CNN architecture is capable of handling the videos where the camera is placed at a distance from the performer. Experiments are performed with KTH and WEIZMANN human actions datasets, whereby it is shown to produce comparable results with the state-of-the-art techniques.
[ { "created": "Thu, 6 Feb 2020 05:07:41 GMT", "version": "v1" }, { "created": "Fri, 7 Feb 2020 06:42:20 GMT", "version": "v2" } ]
2020-02-10
[ [ "Basha", "S. H. Shabbeer", "" ], [ "Pulabaigari", "Viswanath", "" ], [ "Mukherjee", "Snehasis", "" ] ]
We propose a novel scheme for human action recognition in videos, using a 3-dimensional Convolutional Neural Network (3D CNN) based classifier. Traditionally in deep learning based human activity recognition approaches, either a few random frames or every $k^{th}$ frame of the video is considered for training the 3D CNN, where $k$ is a small positive integer, like 4, 5, or 6. This kind of sampling reduces the volume of the input data, which speeds-up training of the network and also avoids over-fitting to some extent, thus enhancing the performance of the 3D CNN model. In the proposed video sampling technique, consecutive $k$ frames of a video are aggregated into a single frame by computing a Gaussian-weighted summation of the $k$ frames. The resulting frame (aggregated frame) preserves the information in a better way than the conventional approaches and experimentally shown to perform better. In this paper, a 3D CNN architecture is proposed to extract the spatio-temporal features and follows Long Short-Term Memory (LSTM) to recognize human actions. The proposed 3D CNN architecture is capable of handling the videos where the camera is placed at a distance from the performer. Experiments are performed with KTH and WEIZMANN human actions datasets, whereby it is shown to produce comparable results with the state-of-the-art techniques.
2404.12984
Marek Wodzinski
Mateusz Daniol, Daria Hemmerling, Jakub Sikora, Pawel Jemiolo, Marek Wodzinski, Magdalena Wojcik-Pedziwiatr
Eye-tracking in Mixed Reality for Diagnosis of Neurodegenerative Diseases
null
null
null
null
cs.HC cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Parkinson's disease ranks as the second most prevalent neurodegenerative disorder globally. This research aims to develop a system leveraging Mixed Reality capabilities for tracking and assessing eye movements. In this paper, we present a medical scenario and outline the development of an application designed to capture eye-tracking signals through Mixed Reality technology for the evaluation of neurodegenerative diseases. Additionally, we introduce a pipeline for extracting clinically relevant features from eye-gaze analysis, describing the capabilities of the proposed system from a medical perspective. The study involved a cohort of healthy control individuals and patients suffering from Parkinson's disease, showcasing the feasibility and potential of the proposed technology for non-intrusive monitoring of eye movement patterns for the diagnosis of neurodegenerative diseases. Clinical relevance - Developing a non-invasive biomarker for Parkinson's disease is urgently needed to accurately detect the disease's onset. This would allow for the timely introduction of neuroprotective treatment at the earliest stage and enable the continuous monitoring of intervention outcomes. The ability to detect subtle changes in eye movements allows for early diagnosis, offering a critical window for intervention before more pronounced symptoms emerge. Eye tracking provides objective and quantifiable biomarkers, ensuring reliable assessments of disease progression and cognitive function. The eye gaze analysis using Mixed Reality glasses is wireless, facilitating convenient assessments in both home and hospital settings. The approach offers the advantage of utilizing hardware that requires no additional specialized attachments, enabling examinations through personal eyewear.
[ { "created": "Fri, 19 Apr 2024 16:34:15 GMT", "version": "v1" }, { "created": "Mon, 3 Jun 2024 10:45:42 GMT", "version": "v2" } ]
2024-06-04
[ [ "Daniol", "Mateusz", "" ], [ "Hemmerling", "Daria", "" ], [ "Sikora", "Jakub", "" ], [ "Jemiolo", "Pawel", "" ], [ "Wodzinski", "Marek", "" ], [ "Wojcik-Pedziwiatr", "Magdalena", "" ] ]
Parkinson's disease ranks as the second most prevalent neurodegenerative disorder globally. This research aims to develop a system leveraging Mixed Reality capabilities for tracking and assessing eye movements. In this paper, we present a medical scenario and outline the development of an application designed to capture eye-tracking signals through Mixed Reality technology for the evaluation of neurodegenerative diseases. Additionally, we introduce a pipeline for extracting clinically relevant features from eye-gaze analysis, describing the capabilities of the proposed system from a medical perspective. The study involved a cohort of healthy control individuals and patients suffering from Parkinson's disease, showcasing the feasibility and potential of the proposed technology for non-intrusive monitoring of eye movement patterns for the diagnosis of neurodegenerative diseases. Clinical relevance - Developing a non-invasive biomarker for Parkinson's disease is urgently needed to accurately detect the disease's onset. This would allow for the timely introduction of neuroprotective treatment at the earliest stage and enable the continuous monitoring of intervention outcomes. The ability to detect subtle changes in eye movements allows for early diagnosis, offering a critical window for intervention before more pronounced symptoms emerge. Eye tracking provides objective and quantifiable biomarkers, ensuring reliable assessments of disease progression and cognitive function. The eye gaze analysis using Mixed Reality glasses is wireless, facilitating convenient assessments in both home and hospital settings. The approach offers the advantage of utilizing hardware that requires no additional specialized attachments, enabling examinations through personal eyewear.
2109.06780
Danijar Hafner
Danijar Hafner
Benchmarking the Spectrum of Agent Capabilities
Published at ICLR 2022. Website: https://danijar.com/crafter
null
null
null
cs.AI cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Evaluating the general abilities of intelligent agents requires complex simulation environments. Existing benchmarks typically evaluate only one narrow task per environment, requiring researchers to perform expensive training runs on many different environments. We introduce Crafter, an open world survival game with visual inputs that evaluates a wide range of general abilities within a single environment. Agents either learn from the provided reward signal or through intrinsic objectives and are evaluated by semantically meaningful achievements that can be unlocked during each episode, such as discovering resources and crafting tools. Consistently unlocking all achievements requires strong generalization, deep exploration, and long-term reasoning. We experimentally verify that Crafter is of appropriate difficulty to drive future research and provide baselines scores of reward agents and unsupervised agents. Furthermore, we observe sophisticated behaviors emerging from maximizing the reward signal, such as building tunnel systems, bridges, houses, and plantations. We hope that Crafter will accelerate research progress by quickly evaluating a wide spectrum of abilities.
[ { "created": "Tue, 14 Sep 2021 15:49:31 GMT", "version": "v1" }, { "created": "Sat, 12 Feb 2022 20:02:13 GMT", "version": "v2" } ]
2022-02-15
[ [ "Hafner", "Danijar", "" ] ]
Evaluating the general abilities of intelligent agents requires complex simulation environments. Existing benchmarks typically evaluate only one narrow task per environment, requiring researchers to perform expensive training runs on many different environments. We introduce Crafter, an open world survival game with visual inputs that evaluates a wide range of general abilities within a single environment. Agents either learn from the provided reward signal or through intrinsic objectives and are evaluated by semantically meaningful achievements that can be unlocked during each episode, such as discovering resources and crafting tools. Consistently unlocking all achievements requires strong generalization, deep exploration, and long-term reasoning. We experimentally verify that Crafter is of appropriate difficulty to drive future research and provide baselines scores of reward agents and unsupervised agents. Furthermore, we observe sophisticated behaviors emerging from maximizing the reward signal, such as building tunnel systems, bridges, houses, and plantations. We hope that Crafter will accelerate research progress by quickly evaluating a wide spectrum of abilities.
1108.1751
Joel Oren
Siavosh Benabbas, Hyun Chul Lee, Joel Oren, Yuli Ye
Efficient Sum-Based Hierarchical Smoothing Under \ell_1-Norm
null
null
null
null
cs.DS
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We introduce a new regression problem which we call the Sum-Based Hierarchical Smoothing problem. Given a directed acyclic graph and a non-negative value, called target value, for each vertex in the graph, we wish to find non-negative values for the vertices satisfying a certain constraint while minimizing the distance of these assigned values and the target values in the lp-norm. The constraint is that the value assigned to each vertex should be no less than the sum of the values assigned to its children. We motivate this problem with applications in information retrieval and web mining. While our problem can be solved in polynomial time using linear programming, given the input size in these applications such a solution might be too slow. We mainly study the \ell_1-norm case restricting the underlying graphs to rooted trees. For this case we provide an efficient algorithm, running in O(n^2) time. While the algorithm is purely combinatorial, its proof of correctness is an elegant use of linear programming duality. We believe that our approach may be applicable to similar problems, where comparable hierarchical constraints are involved, e.g. considering the average of the values assigned to the children of each vertex. While similar in flavor to other smoothing problems like Isotonic Regression (see for example [Angelov et al. SODA'06]), our problem is arguably richer and theoretically more challenging.
[ { "created": "Mon, 8 Aug 2011 17:07:06 GMT", "version": "v1" } ]
2011-08-09
[ [ "Benabbas", "Siavosh", "" ], [ "Lee", "Hyun Chul", "" ], [ "Oren", "Joel", "" ], [ "Ye", "Yuli", "" ] ]
We introduce a new regression problem which we call the Sum-Based Hierarchical Smoothing problem. Given a directed acyclic graph and a non-negative value, called target value, for each vertex in the graph, we wish to find non-negative values for the vertices satisfying a certain constraint while minimizing the distance of these assigned values and the target values in the lp-norm. The constraint is that the value assigned to each vertex should be no less than the sum of the values assigned to its children. We motivate this problem with applications in information retrieval and web mining. While our problem can be solved in polynomial time using linear programming, given the input size in these applications such a solution might be too slow. We mainly study the \ell_1-norm case restricting the underlying graphs to rooted trees. For this case we provide an efficient algorithm, running in O(n^2) time. While the algorithm is purely combinatorial, its proof of correctness is an elegant use of linear programming duality. We believe that our approach may be applicable to similar problems, where comparable hierarchical constraints are involved, e.g. considering the average of the values assigned to the children of each vertex. While similar in flavor to other smoothing problems like Isotonic Regression (see for example [Angelov et al. SODA'06]), our problem is arguably richer and theoretically more challenging.
1306.4974
Gautam Mandal
Pawel Caputa, Gautam Mandal and Ritam Sinha
Dynamical entanglement entropy with angular momentum and U(1) charge
22 pages, 4 figures; (v2) many comments added for better clarity; typos fixed; references added
null
10.1007/JHEP11(2013)052
TIFR/TH/13-16, WITS-CTP-116
hep-th cond-mat.stat-mech
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We consider time-dependent entanglement entropy (EE) for a 1+1 dimensional CFT in the presence of angular momentum and U(1) charge. The EE saturates, irrespective of the initial state, to the grand canonical entropy after a time large compared with the length of the entangling interval. We reproduce the CFT results from an AdS dual consisting of a spinning BTZ black hole and a flat U(1) connection. The apparent discrepancy that the holographic EE does not a priori depend on the U(1) charge while the CFT EE does, is resolved by the charge-dependent shift between the bulk and boundary stress tensors. We show that for small entangling intervals, the entanglement entropy obeys the first law of thermodynamics, as conjectured recently. The saturation of the EE in the field theory is shown to follow from a version of quantum ergodicity; the derivation indicates that it should hold for conformal as well as massive theories in any number of dimensions.
[ { "created": "Thu, 20 Jun 2013 19:49:23 GMT", "version": "v1" }, { "created": "Sun, 25 Aug 2013 12:26:54 GMT", "version": "v2" } ]
2015-06-16
[ [ "Caputa", "Pawel", "" ], [ "Mandal", "Gautam", "" ], [ "Sinha", "Ritam", "" ] ]
We consider time-dependent entanglement entropy (EE) for a 1+1 dimensional CFT in the presence of angular momentum and U(1) charge. The EE saturates, irrespective of the initial state, to the grand canonical entropy after a time large compared with the length of the entangling interval. We reproduce the CFT results from an AdS dual consisting of a spinning BTZ black hole and a flat U(1) connection. The apparent discrepancy that the holographic EE does not a priori depend on the U(1) charge while the CFT EE does, is resolved by the charge-dependent shift between the bulk and boundary stress tensors. We show that for small entangling intervals, the entanglement entropy obeys the first law of thermodynamics, as conjectured recently. The saturation of the EE in the field theory is shown to follow from a version of quantum ergodicity; the derivation indicates that it should hold for conformal as well as massive theories in any number of dimensions.
2302.05803
Mohammadjavad Ghorbanalivakili
Jungwon Kang, Mohammadjavad Ghorbanalivakili, Gunho Sohn, David Beach, and Veronica Marin
TPE-Net: Track Point Extraction and Association Network for Rail Path Proposal Generation
7 pages, 6 figures, and 1 table Jungwon Kang and Mohammadjavad Ghorbanalivakili have equal contribution
null
10.1109/CASE56687.2023.10260541
null
cs.CV cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
One essential feature of an autonomous train is minimizing collision risks with third-party objects. To estimate the risk, the control system must identify topological information of all the rail routes ahead on which the train can possibly move, especially within merging or diverging rails. This way, the train can figure out the status of potential obstacles with respect to its route and hence, make a timely decision. Numerous studies have successfully extracted all rail tracks as a whole within forward-looking images without considering element instances. Still, some image-based methods have employed hard-coded prior knowledge of railway geometry on 3D data to associate left-right rails and generate rail route instances. However, we propose a rail path extraction pipeline in which left-right rail pixels of each rail route instance are extracted and associated through a fully convolutional encoder-decoder architecture called TPE-Net. Two different regression branches for TPE-Net are proposed to regress the locations of center points of each rail route, along with their corresponding left-right pixels. Extracted rail pixels are then spatially clustered to generate topological information of all the possible train routes (ego-paths), discarding non-ego-path ones. Experimental results on a challenging, publicly released benchmark show true-positive-pixel level average precision and recall of 0.9207 and 0.8721, respectively, at about 12 frames per second. Even though our evaluation results are not higher than the SOTA, the proposed regression pipeline performs remarkably in extracting the correspondences by looking once at the image. It generates strong rail route hypotheses without reliance on camera parameters, 3D data, and geometrical constraints.
[ { "created": "Sat, 11 Feb 2023 22:49:06 GMT", "version": "v1" } ]
2024-07-26
[ [ "Kang", "Jungwon", "" ], [ "Ghorbanalivakili", "Mohammadjavad", "" ], [ "Sohn", "Gunho", "" ], [ "Beach", "David", "" ], [ "Marin", "Veronica", "" ] ]
One essential feature of an autonomous train is minimizing collision risks with third-party objects. To estimate the risk, the control system must identify topological information of all the rail routes ahead on which the train can possibly move, especially within merging or diverging rails. This way, the train can figure out the status of potential obstacles with respect to its route and hence, make a timely decision. Numerous studies have successfully extracted all rail tracks as a whole within forward-looking images without considering element instances. Still, some image-based methods have employed hard-coded prior knowledge of railway geometry on 3D data to associate left-right rails and generate rail route instances. However, we propose a rail path extraction pipeline in which left-right rail pixels of each rail route instance are extracted and associated through a fully convolutional encoder-decoder architecture called TPE-Net. Two different regression branches for TPE-Net are proposed to regress the locations of center points of each rail route, along with their corresponding left-right pixels. Extracted rail pixels are then spatially clustered to generate topological information of all the possible train routes (ego-paths), discarding non-ego-path ones. Experimental results on a challenging, publicly released benchmark show true-positive-pixel level average precision and recall of 0.9207 and 0.8721, respectively, at about 12 frames per second. Even though our evaluation results are not higher than the SOTA, the proposed regression pipeline performs remarkably in extracting the correspondences by looking once at the image. It generates strong rail route hypotheses without reliance on camera parameters, 3D data, and geometrical constraints.
hep-th/0105319
Kluson Josef
J. Kluson
Some Remarks About Berkovits' Superstring Field Theory
14 pages, Introduction part and some typos corrected, ref. added
JHEP 0106:045,2001
10.1088/1126-6708/2001/06/045
null
hep-th
null
In this short note we would like to discuss general solutions of the Berkovits superstring field theory, in particular the string field action for fluctuation around such a solution. We will find that fluctuations obey the same equation of motion as the original field with the new BRST operator. Then we will argue that the superstring field theory action for fluctuation field has the same form as the original one.
[ { "created": "Thu, 31 May 2001 17:24:20 GMT", "version": "v1" }, { "created": "Tue, 5 Jun 2001 07:32:02 GMT", "version": "v2" } ]
2010-02-03
[ [ "Kluson", "J.", "" ] ]
In this short note we would like to discuss general solutions of the Berkovits superstring field theory, in particular the string field action for fluctuation around such a solution. We will find that fluctuations obey the same equation of motion as the original field with the new BRST operator. Then we will argue that the superstring field theory action for fluctuation field has the same form as the original one.
hep-th/0607243
Pietro Antonio Grassi
P.A. Grassi and M. Marescotti
Flux Vacua and Supermanifolds
Latex, no figures, 35 pp, misprints and minor changes
JHEP 0701:068,2007
10.1088/1126-6708/2007/01/068
DISTA-UPO-06, DFTT-18/2006
hep-th
null
As been recently pointed out, physically relevant models derived from string theory require the presence of non-vanishing form fluxes besides the usual geometrical constraints. In the case of NS-NS fluxes, the Generalized Complex Geometry encodes these informations in a beautiful geometrical structure. On the other hand, the R-R fluxes call for supergeometry as the underlying mathematical framework. In this context, we analyze the possibility of constructing interesting supermanifolds recasting the geometrical data and RR fluxes. To characterize these supermanifolds we have been guided by the fact topological strings on supermanifolds require the super-Ricci flatness of the target space. This can be achieved by adding to a given bosonic manifold enough anticommuting coordinates and new constraints on the bosonic sub-manifold. We study these constraints at the linear and non-linear level for a pure geometrical setting and in the presence of p-form field strengths. We find that certain spaces admit several super-extensions and we give a parameterization in a simple case of d bosonic coordinates and two fermionic coordinates. In addition, we comment on the role of the RR field in the construction of the super-metric. We give several examples based on supergroup manifolds and coset supermanifolds.
[ { "created": "Sun, 30 Jul 2006 10:26:32 GMT", "version": "v1" }, { "created": "Thu, 2 Nov 2006 11:21:35 GMT", "version": "v2" }, { "created": "Fri, 11 Apr 2008 09:54:32 GMT", "version": "v3" } ]
2010-10-27
[ [ "Grassi", "P. A.", "" ], [ "Marescotti", "M.", "" ] ]
As been recently pointed out, physically relevant models derived from string theory require the presence of non-vanishing form fluxes besides the usual geometrical constraints. In the case of NS-NS fluxes, the Generalized Complex Geometry encodes these informations in a beautiful geometrical structure. On the other hand, the R-R fluxes call for supergeometry as the underlying mathematical framework. In this context, we analyze the possibility of constructing interesting supermanifolds recasting the geometrical data and RR fluxes. To characterize these supermanifolds we have been guided by the fact topological strings on supermanifolds require the super-Ricci flatness of the target space. This can be achieved by adding to a given bosonic manifold enough anticommuting coordinates and new constraints on the bosonic sub-manifold. We study these constraints at the linear and non-linear level for a pure geometrical setting and in the presence of p-form field strengths. We find that certain spaces admit several super-extensions and we give a parameterization in a simple case of d bosonic coordinates and two fermionic coordinates. In addition, we comment on the role of the RR field in the construction of the super-metric. We give several examples based on supergroup manifolds and coset supermanifolds.
1401.8173
Daniel Zaragoza
Daniel Zaragoza
Modeling TCP Throughput with Random Packet Drops
null
null
null
null
cs.NI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The present report deals with the modeling of the long-term throughput, a.k.a., send rate, of the Transmission Control Protocol (TCP) under the following assumptions. (i) We consider a single 'infinite source' using a network path from sender to receiver. (ii) Each TCP packet is randomly dropped with probability p; independently of previous drops or any other event/parameter. (iii) The - never changing - receiver window limits the amount of outstanding data. (iv) The receiver acknowledges every packet. (v) The TCP modeled here conforms to the publicly available standards (RFCs) as concerns congestion control. We validate and determine the limits of the different models proposed here using packet-level simulations. The contributions of the present work are the following: (a) We determine three regimes, and their conditions of applicability, depending on p: Linear law regime, square root law regime, and timeout regime. (b) As concerns the relationship between the linear and square root regimes, we give additional insights relatively to previously published work. (c) We give the exact equations governing the TCP send rate in any regime. (d) From the exact equation and under the further condition that the path is not saturated, we give and discuss approximations for the send rate of the NewReno variant of TCP. A by-product of these calculations is the distribution of the sender window, independently of any timing or saturation consideration. (e) These approximations give results that are accurate to a few percent when compared to simulation results. Detailed comparison and sources of errors between theory and simulations are also discussed.
[ { "created": "Fri, 31 Jan 2014 14:20:22 GMT", "version": "v1" } ]
2014-02-03
[ [ "Zaragoza", "Daniel", "" ] ]
The present report deals with the modeling of the long-term throughput, a.k.a., send rate, of the Transmission Control Protocol (TCP) under the following assumptions. (i) We consider a single 'infinite source' using a network path from sender to receiver. (ii) Each TCP packet is randomly dropped with probability p; independently of previous drops or any other event/parameter. (iii) The - never changing - receiver window limits the amount of outstanding data. (iv) The receiver acknowledges every packet. (v) The TCP modeled here conforms to the publicly available standards (RFCs) as concerns congestion control. We validate and determine the limits of the different models proposed here using packet-level simulations. The contributions of the present work are the following: (a) We determine three regimes, and their conditions of applicability, depending on p: Linear law regime, square root law regime, and timeout regime. (b) As concerns the relationship between the linear and square root regimes, we give additional insights relatively to previously published work. (c) We give the exact equations governing the TCP send rate in any regime. (d) From the exact equation and under the further condition that the path is not saturated, we give and discuss approximations for the send rate of the NewReno variant of TCP. A by-product of these calculations is the distribution of the sender window, independently of any timing or saturation consideration. (e) These approximations give results that are accurate to a few percent when compared to simulation results. Detailed comparison and sources of errors between theory and simulations are also discussed.
End of preview. Expand in Data Studio
README.md exists but content is empty.
Downloads last month
8

Models trained or fine-tuned on LLMsForHepth/hep-th_qbio_cs

Collection including LLMsForHepth/hep-th_qbio_cs