doc_id
stringlengths
2
10
revision_depth
stringclasses
5 values
before_revision
stringlengths
3
309k
after_revision
stringlengths
5
309k
edit_actions
list
sents_char_pos
sequence
0705.0053
1
We show that the mutual fund theorems of Merton (1971) extend to the problem of optimal investment to minimize the probability of lifetime ruin. We obtain four such theorems by considering a financial market both with and without a riskless asset and by considering both constant and random consumption.
We show that the mutual fund theorems of Merton (1971) extend to the problem of optimal investment to minimize the probability of lifetime ruin. We obtain two such theorems by considering a financial market both with and without a riskless asset for random consumption. The striking result is that we obtain two-fund theorems despite the additional source of randomness from consumption.
[ { "type": "R", "before": "four", "after": "two", "start_char_pos": 155, "end_char_pos": 159 }, { "type": "R", "before": "and by considering both constant and random", "after": "for random consumption. The striking result is that we obtain two-fund theorems despite the additional source of randomness from", "start_char_pos": 247, "end_char_pos": 290 } ]
[ 0, 144 ]
0705.1606
1
The self-propelled motion of microscopic bodies immersed in a fluid medium is studied using molecular dynamics simulation. The advantage of the atomistic approach is that the detailed level of description allows complete freedom in specifying the swimmer design and its coupling with the surrounding fluid. A series of two-dimensional swimming bodies employing a variety of propulsion mechanisms -- motivated by biological and microrobotic designs -- is investigated, including the use of moving limbs, changing body shapes and fluid jets. The efficiencies of the mechanisms and the nature of the induced, time-dependent flow fields are found to differ widely among swimmers .
The self-propelled motion of microscopic bodies immersed in a fluid medium is studied using molecular dynamics simulation. The advantage of the atomistic approach is that the detailed level of description allows complete freedom in specifying the swimmer design and its coupling with the surrounding fluid. A series of two-dimensional swimming bodies employing a variety of propulsion mechanisms -- motivated by biological and microrobotic designs -- is investigated, including the use of moving limbs, changing body shapes and fluid jets. The swimming efficiency and the nature of the induced, time-dependent flow fields are found to differ widely among body designs and propulsion mechanisms .
[ { "type": "R", "before": "efficiencies of the mechanisms", "after": "swimming efficiency", "start_char_pos": 544, "end_char_pos": 574 }, { "type": "R", "before": "swimmers", "after": "body designs and propulsion mechanisms", "start_char_pos": 666, "end_char_pos": 674 } ]
[ 0, 122, 306, 539 ]
0705.1949
1
We employ perturbation analysis technique to study trading strategies for multi-asset portfolio and obtain optimal trading methods for wealth maximization under arbitrary utility functions .
We employ perturbation analysis technique to study multi-asset portfolio optimisation with transaction cost. We allow for correlations in risky assets and obtain optimal trading methods for general utility functions. Our analytical results are supported by numerical simulations in the context of the Long Term Growth Model .
[ { "type": "D", "before": "trading strategies for", "after": null, "start_char_pos": 51, "end_char_pos": 73 }, { "type": "A", "before": null, "after": "optimisation with transaction cost. We allow for correlations in risky assets", "start_char_pos": 96, "end_char_pos": 96 }, { "type": "R", "before": "wealth maximization under arbitrary utility functions", "after": "general utility functions. Our analytical results are supported by numerical simulations in the context of the Long Term Growth Model", "start_char_pos": 136, "end_char_pos": 189 } ]
[ 0 ]
0706.0482
1
We perform a stability analysis for the utility maximization problem in a general semimartingale model where both liquid and illiquid assets (random endowments) are present. Small misspecifications of preferences (as modeled via expected utility), as well as views of the world or the market model (as modeled via subjective probabilities) are considered. Simple sufficient conditions are given for the problem to be well-posed, in the sense that optimal wealths and marginal utility-based prices are continuous functionals of the inputs .
We perform a stability analysis for the utility maximization problem in a general semimartingale model where both liquid and illiquid assets (random endowments) are present. Small misspecifications of preferences (as modeled via expected utility), as well as views of the world or the market model (as modeled via subjective probabilities) are considered. Simple sufficient conditions are given for the problem to be well-posed, in the sense the optimal wealth and the marginal utility-based prices are continuous functionals of preferences and probabilistic views .
[ { "type": "R", "before": "that optimal wealths and", "after": "the optimal wealth and the", "start_char_pos": 442, "end_char_pos": 466 }, { "type": "R", "before": "the inputs", "after": "preferences and probabilistic views", "start_char_pos": 527, "end_char_pos": 537 } ]
[ 0, 173, 355 ]
0706.2383
1
The utilization of multiple post translational modifications (PTMs) in regulating a biological response is ubiquitous in cell signaling. If each PTM contributes an additional, equivalent binding site, then one consequence of an increase in the number of PTMs may be to increase the probability that, upon disassociation, a ligand immediately rebinds to its receptor. How such effects may influence cell signaling systems has been less studied. Here, a self-consistent integral equation formalism for ligand rebinding in conjunction with Monte Carlo simulations are employed to further investigate the effects of multiple, equivalent binding sites on shaping biological responses. Multiple regimes that characterize qualitatively different physics due to the differential prevalence of rebinding effects and their relation to systems-level properties are predicted and studied. Calculations suggest that when ligand rebinding contributes significantly to the dose response, a purely allovalent model can influence the binding curves nonlinearly but other mechanistic ingredients are required to achieve high degrees of biochemical cooperativity. It is our hope that these calculations motivate experiments that can further unravel the many functional consequences of multi-site phosphorylation.
the paper is withdrawn for the moment
[ { "type": "R", "before": "The utilization of multiple post translational modifications (PTMs) in regulating a biological response is ubiquitous in cell signaling. If each PTM contributes an additional, equivalent binding site, then one consequence of an increase in the number of PTMs may be to increase the probability that, upon disassociation, a ligand immediately rebinds to its receptor. How such effects may influence cell signaling systems has been less studied. Here, a self-consistent integral equation formalism for ligand rebinding in conjunction with Monte Carlo simulations are employed to further investigate the effects of multiple, equivalent binding sites on shaping biological responses. Multiple regimes that characterize qualitatively different physics due to the differential prevalence of rebinding effects and their relation to systems-level properties are predicted and studied. Calculations suggest that when ligand rebinding contributes significantly to the dose response, a purely allovalent model can influence the binding curves nonlinearly but other mechanistic ingredients are required to achieve high degrees of biochemical cooperativity. It is our hope that these calculations motivate experiments that can further unravel the many functional consequences of multi-site phosphorylation.", "after": "the paper is withdrawn for the moment", "start_char_pos": 0, "end_char_pos": 1293 } ]
[ 0, 136, 366, 443, 679, 876, 1144 ]
0706.2383
2
the paper is withdrawn for the moment
The utilization of multiple phosphorylation sites in regulating a biological response is ubiquitous in cell signaling. If each site contributes an additional, equivalent binding site, then one consequence of an increase in the number of phosphorylations may be to increase the probability that, upon disassociation, a ligand immediately rebinds to its receptor. How such effects may influence cell signaling systems has been less studied. Here, a self-consistent integral equation formalism for ligand rebinding in conjunction with Monte Carlo simulations are employed to further investigate the effects of multiple, equivalent binding sites on shaping biological responses. Multiple regimes that characterize qualitatively different physics due to the differential prevalence of rebinding effects and their relation to systems-level properties are predicted and studied. Calculations suggest that when ligand rebinding contributes significantly to the dose response, a purely allovalent model can influence the binding curves nonlinearly but other mechanistic ingredients are required to achieve high degrees of biochemical cooperativity. It is our hope that these calculations motivate experiments that can further unravel the many functional consequences of multi-site phosphorylation.
[ { "type": "R", "before": "the paper is withdrawn for the moment", "after": "The utilization of multiple phosphorylation sites in regulating a biological response is ubiquitous in cell signaling. If each site contributes an additional, equivalent binding site, then one consequence of an increase in the number of phosphorylations may be to increase the probability that, upon disassociation, a ligand immediately rebinds to its receptor. How such effects may influence cell signaling systems has been less studied. Here, a self-consistent integral equation formalism for ligand rebinding in conjunction with Monte Carlo simulations are employed to further investigate the effects of multiple, equivalent binding sites on shaping biological responses. Multiple regimes that characterize qualitatively different physics due to the differential prevalence of rebinding effects and their relation to systems-level properties are predicted and studied. Calculations suggest that when ligand rebinding contributes significantly to the dose response, a purely allovalent model can influence the binding curves nonlinearly but other mechanistic ingredients are required to achieve high degrees of biochemical cooperativity. It is our hope that these calculations motivate experiments that can further unravel the many functional consequences of multi-site phosphorylation.", "start_char_pos": 0, "end_char_pos": 37 } ]
[ 0 ]
0706.2383
3
The utilization of multiple phosphorylation sites in regulating a biological response is ubiquitous in cell signaling. If each site contributes an additional, equivalent binding site, then one consequence of an increase in the number of phosphorylations may be to increase the probability that, upon disassociation, a ligand immediately rebinds to its receptor. How such effects may influence cell signaling systems has been less studied. Here, a self-consistent integral equation formalism for ligand rebinding in conjunction with Monte Carlo simulations are employed to further investigate the effects of multiple, equivalent binding sites on shaping biological responses. Multiple regimes that characterize qualitatively different physics due to the differential prevalence of rebinding effects and their relation to systems-level properties are predictedand studied . Calculations suggest that when ligand rebinding contributes significantly to the dose response, a purely allovalent model can influence the binding curves nonlinearly but other mechanistic ingredients are required to achieve high degrees of biochemical cooperativity. It is our hope that these calculations motivate experiments that can further unravel the many functional consequences of multi-site phosphorylation .
The utilization of multiple phosphorylation sites in regulating a biological response is ubiquitous in cell signaling. If each site contributes an additional, equivalent binding site, then one consequence of an increase in the number of phosphorylations may be to increase the probability that, upon disassociation, a ligand immediately rebinds to its receptor. How such effects may influence cell signaling systems has been less studied. Here, a self-consistent integral equation formalism for ligand rebinding , in conjunction with Monte Carlo simulations , is employed to further investigate the effects of multiple, equivalent binding sites on shaping biological responses. Multiple regimes that characterize qualitatively different physics due to the differential prevalence of rebinding effects are predicted . Calculations suggest that when ligand rebinding contributes significantly to the dose response, a purely allovalent model can influence the binding curves nonlinearly . The model also predicts that ligand rebinding in itself appears insufficient to generative a highly cooperative biological response .
[ { "type": "A", "before": null, "after": ",", "start_char_pos": 512, "end_char_pos": 512 }, { "type": "R", "before": "are", "after": ", is", "start_char_pos": 557, "end_char_pos": 560 }, { "type": "R", "before": "and their relation to systems-level properties are predictedand studied", "after": "are predicted", "start_char_pos": 799, "end_char_pos": 870 }, { "type": "R", "before": "but other mechanistic ingredients are required to achieve high degrees of biochemical cooperativity. It is our hope that these calculations motivate experiments that can further unravel the many functional consequences of multi-site phosphorylation", "after": ". The model also predicts that ligand rebinding in itself appears insufficient to generative a highly cooperative biological response", "start_char_pos": 1040, "end_char_pos": 1288 } ]
[ 0, 118, 361, 438, 675, 1140 ]
0707.0026
1
A probabilistic structure on sequential dynamical systems is introduced here, the new model will be called Probabilistic Sequential Network, PSN. The morphisms of Probabilistic Sequential Networks are defined using two algebraic conditions , whose imply that the distribution of probabilities in the systems are close . It is proved here that two homomorphic Probabilistic Sequential Networks have the same equilibrium or steady state probabilities . Additionally, the proof of the set of PSN with its morphisms form the category PSN, having the category of sequential dynamical systems SDS, as a full subcategory is given. Several examples of morphisms, subsystems and simulations are given.
A probabilistic structure on sequential dynamical systems is introduced here, the new model will be called Probabilistic Sequential Network, PSN. The morphisms of Probabilistic Sequential Networks are defined using two algebraic conditions . It is proved here that two homomorphic Probabilistic Sequential Networks have the same equilibrium or steady state probabilities if the morphism is either an epimorphism or a monomorphism . Additionally, the proof of the set of PSN with its morphisms form the category PSN, having the category of sequential dynamical systems SDS, as a full subcategory is given. Several examples of morphisms, subsystems and simulations are given.
[ { "type": "D", "before": ", whose imply that the distribution of probabilities in the systems are close", "after": null, "start_char_pos": 240, "end_char_pos": 317 }, { "type": "A", "before": null, "after": "if the morphism is either an epimorphism or a monomorphism", "start_char_pos": 449, "end_char_pos": 449 } ]
[ 0, 145, 319, 451, 624 ]
0707.2076
1
We propose a biologically motivated quantity, twinness, to evaluate local similarity between nodes in a network. The twinness of a pair of nodes is the number of connected, labelled subgraphs of size n in which the two nodes possess identical neighbours. The graph animal algorithm is used to estimate twinness for each pair of nodes (for subgraph sizes n=4 to n=12) in four different protein interaction networks (PINs). These include an Escherichia coli PIN and three Saccharomyces cerevisiae PINs - each obtained using state-of-the-art high throughput methods. In almost all cases, the average twinness of node pairs is vastly higher than expected from a null model obtained by switching links. For all n, we observe a difference in the ratio of type A twins (which are unlinked pairs) to type B twins (which are linked pairs) distinguishing the prokaryote E. Coli from the eukaryote S. Cerevisiae . Interaction similarity is expected due to gene duplication, and whole genome duplication paralogues in S. Cerevisiae have been reported to co-cluster into the same complexes. Indeed, we find that these paralogues are over-represented as twins compared to pairs chosen at random. These results indicate that twinness can detect ancestral relationships from currently available PIN data.
We propose a biologically motivated quantity, twinness, to evaluate local similarity between nodes in a network. The twinness of a pair of nodes is the number of connected, labeled subgraphs of size n in which the two nodes possess identical neighbours. The graph animal algorithm is used to estimate twinness for each pair of nodes (for subgraph sizes n=4 to n=12) in four different protein interaction networks (PINs). These include an Escherichia coli PIN and three Saccharomyces cerevisiae PINs -- each obtained using state-of-the-art high throughput methods. In almost all cases, the average twinness of node pairs is vastly higher than expected from a null model obtained by switching links. For all n, we observe a difference in the ratio of type A twins (which are unlinked pairs) to type B twins (which are linked pairs) distinguishing the prokaryote E. coli from the eukaryote S. cerevisiae . Interaction similarity is expected due to gene duplication, and whole genome duplication paralogues in S. cerevisiae have been reported to co-cluster into the same complexes. Indeed, we find that these paralogous proteins are over-represented as twins compared to pairs chosen at random. These results indicate that twinness can detect ancestral relationships from currently available PIN data.
[ { "type": "R", "before": "labelled", "after": "labeled", "start_char_pos": 173, "end_char_pos": 181 }, { "type": "R", "before": "-", "after": "--", "start_char_pos": 500, "end_char_pos": 501 }, { "type": "R", "before": "Coli", "after": "coli", "start_char_pos": 863, "end_char_pos": 867 }, { "type": "R", "before": "Cerevisiae", "after": "cerevisiae", "start_char_pos": 890, "end_char_pos": 900 }, { "type": "R", "before": "Cerevisiae", "after": "cerevisiae", "start_char_pos": 1009, "end_char_pos": 1019 }, { "type": "R", "before": "paralogues", "after": "paralogous proteins", "start_char_pos": 1105, "end_char_pos": 1115 } ]
[ 0, 112, 254, 421, 563, 697, 902, 1077, 1181 ]
0707.3621
1
We calculate analytically the critical connectivity K_c of Random Threshold Networks (RTN) for homogeneous and inhomogeneous thresholds, and confirm the results by numerical simulations. We find a super-linear increase of K_c with the (average) absolute threshold |h|, which approaches K_c(|h|) \sim |h|^{\alpha with \alpha \approx } 2 for large |h|, and show that this asymptotic scaling is universal for RTN with Poissonian distributed connectivity and threshold distributions with a variance that grows slower than |h|^{\alpha . Interestingly, we find that inhomogeneous distribution of thresholds leads to increased propagation of perturbations for sparsely connected networks, while for densely connected networks damage is reduced . Further, damage propagation in RTN with in-degree distributions that exhibit a scale-free tail k_{in^{\gamma} is studied; we find that a decrease of \gamma can lead to a transition from supercritical (chaotic) to subcritical (ordered) dynamics} . Last, local correlations between node thresholds and in-degree are introduced. Here, numerical simulations show that even weak (anti-)correlations can lead to a transition from ordered to chaotic dynamics, and vice versa. Interestingly, in this case the annealed approximation fails to predict the dynamical behavior for sparse connectivities , suggesting that even weak topological correlations can strongly limit its applicability for finite N .
We calculate analytically the critical connectivity K_c of Random Threshold Networks (RTN) for homogeneous and inhomogeneous thresholds, and confirm the results by numerical simulations. We find a super-linear increase of K_c with the (average) absolute threshold |h|, which approaches K_c(|h|) \sim with \alpha \approx } h^2/( 2 \ln{|h| for large |h|, and show that this asymptotic scaling is universal for RTN with Poissonian distributed connectivity and threshold distributions with a variance that grows slower than h^2 . Interestingly, we find that inhomogeneous distribution of thresholds leads to increased propagation of perturbations for sparsely connected networks, while for densely connected networks damage is reduced ^{\gamma} is studied; we find that a decrease of \gamma can lead to a transition from supercritical (chaotic) to subcritical (ordered) dynamics} ; the cross-over point yields a novel, characteristic connectivity K_d, that has no counterpart in Boolean networks . Last, local correlations between node thresholds and in-degree are introduced. Here, numerical simulations show that even weak (anti-)correlations can lead to a transition from ordered to chaotic dynamics, and vice versa. It is shown that the naive mean-field assumption typical for the annealed approximation leads to false predictions in this case, since correlations between thresholds and out-degree that emerge as a side-effect strongly modify damage propagation behavior .
[ { "type": "D", "before": "|h|^{\\alpha", "after": null, "start_char_pos": 300, "end_char_pos": 311 }, { "type": "A", "before": null, "after": "h^2/(", "start_char_pos": 334, "end_char_pos": 334 }, { "type": "A", "before": null, "after": "\\ln{|h|", "start_char_pos": 337, "end_char_pos": 337 }, { "type": "R", "before": "|h|^{\\alpha", "after": "h^2", "start_char_pos": 520, "end_char_pos": 531 }, { "type": "D", "before": ". Further, damage propagation in RTN with in-degree distributions that exhibit a scale-free tail k_{in", "after": null, "start_char_pos": 739, "end_char_pos": 841 }, { "type": "A", "before": null, "after": "; the cross-over point yields a novel, characteristic connectivity K_d, that has no counterpart in Boolean networks", "start_char_pos": 986, "end_char_pos": 986 }, { "type": "D", "before": "Interestingly, in this case the annealed approximation fails to predict the dynamical behavior for sparse connectivities", "after": null, "start_char_pos": 1211, "end_char_pos": 1331 }, { "type": "R", "before": ", suggesting that even weak topological correlations can strongly limit its applicability for finite N", "after": "It is shown that the naive mean-field assumption typical for the annealed approximation leads to false predictions in this case, since correlations between thresholds and out-degree that emerge as a side-effect strongly modify damage propagation behavior", "start_char_pos": 1332, "end_char_pos": 1434 } ]
[ 0, 186, 533, 740, 862, 1067, 1210 ]
0708.0046
1
The Markowitz mean-variance optimizing framework has served as the basis for modern portfolio theory for more than 50 years. However, efforts to translate this theoretical foundation into a viable portfolio construction algorithm have been plagued by technical difficulties stemming from the instability of the original optimization problemwith respect to the available data. In this paper we address these issues of estimation error by regularizing the Markowitz objective function through the addition of an \ell_1 penalty . This penalty stabilizes the optimization problem, encourages sparse portfolios, and facilitates treatment of transaction costs in a transparent way . We implement this methodology using the Fama and French 48 industry portfolios as our securities . Using only a modest amount of training data, we construct portfolios whose out-of-sample performance, as measured by Sharpe ratio, is consistently and significantly better than that of the na\"{i . In addition to their excellent performance, these portfolios have only a small number of active positions, a highly desirable attribute for real life applications. We conclude by discussing a collection of portfolio construction problems which can be naturally translated into optimizations involving \ell_1 penalties and which can thus be tackled by algorithms similar to those discussed here .
We consider the problem of portfolio selection within the classical Markowitz mean-variance optimizing framework , which has served as the basis for modern portfolio theory for more than 50 years. To stabilize the problem, we propose to add to the Markowitz objective function a penalty which is proportional to the sum of the absolute values of the portfolio weights ( \ell_1 penalty ) . This penalty stabilizes the optimization problem, automatically encourages sparse portfolios, and facilitates an effective treatment of transaction costs . We implement our methodology using as our securities two sets of portfolios constructed by Fama and French: the 48 industry portfolios and 100 portfolios formed on size and book-to-market . In addition to their excellent performance, these portfolios have only a small number of active positions, a desirable feature for small investors, for whom the fixed overhead portion of the transaction cost is not negligible .
[ { "type": "R", "before": "The", "after": "We consider the problem of portfolio selection within the classical", "start_char_pos": 0, "end_char_pos": 3 }, { "type": "A", "before": null, "after": ", which", "start_char_pos": 49, "end_char_pos": 49 }, { "type": "R", "before": "However, efforts to translate this theoretical foundation into a viable portfolio construction algorithm have been plagued by technical difficulties stemming from the instability of the original optimization problemwith respect to the available data. In this paper we address these issues of estimation error by regularizing the", "after": "To stabilize the problem, we propose to add to the", "start_char_pos": 126, "end_char_pos": 454 }, { "type": "R", "before": "through the addition of an", "after": "a penalty which is proportional to the sum of the absolute values of the portfolio weights (", "start_char_pos": 484, "end_char_pos": 510 }, { "type": "A", "before": null, "after": ")", "start_char_pos": 526, "end_char_pos": 526 }, { "type": "A", "before": null, "after": "automatically", "start_char_pos": 579, "end_char_pos": 579 }, { "type": "A", "before": null, "after": "an effective", "start_char_pos": 626, "end_char_pos": 626 }, { "type": "D", "before": "in a transparent way", "after": null, "start_char_pos": 658, "end_char_pos": 678 }, { "type": "R", "before": "this methodology using the Fama and French 48 industry portfolios", "after": "our methodology using", "start_char_pos": 694, "end_char_pos": 759 }, { "type": "R", "before": ". Using only a modest amount of training data, we construct portfolios whose out-of-sample performance, as measured by Sharpe ratio, is consistently and significantly better than that of the na\\\"{i", "after": "two sets of portfolios constructed by Fama and French: the 48 industry portfolios and 100 portfolios formed on size and book-to-market", "start_char_pos": 778, "end_char_pos": 975 }, { "type": "R", "before": "highly desirable attribute for real life applications. We conclude by discussing a collection of portfolio construction problems which can be naturally translated into optimizations involving \\ell_1 penalties and which can thus be tackled by algorithms similar to those discussed here", "after": "desirable feature for small investors, for whom the fixed overhead portion of the transaction cost is not negligible", "start_char_pos": 1087, "end_char_pos": 1371 } ]
[ 0, 125, 376, 528, 680, 977, 1141 ]
0708.0046
2
We consider the problem of portfolio selection within the classical Markowitz mean-variance optimizing framework, which has served as the basis for modern portfolio theory for more than 50 years. To stabilize the problem, we propose to add to the Markowitz objective function a penalty which is proportional to the sum of the absolute values of the portfolio weights (\ell_1 penalty) . This penalty stabilizes the optimization problem, automatically encourages sparse portfolios , and facilitates an effective treatment of transaction costs. We implement our methodology using as our securities two sets of portfolios constructed by Fama and French : the 48 industry portfolios and 100 portfolios formed on size and book-to-market. In addition to their excellent performance, these portfolios have only a small number of active positions, a desirable feature for small investors, for whom the fixed overhead portion of the transaction cost is not negligible .
We consider the problem of portfolio selection within the classical Markowitz mean-variance framework, reformulated as a constrained least-squares regression problem. We propose to add to the objective function a penalty proportional to the sum of the absolute values of the portfolio weights . This penalty regularizes (stabilizes) the optimization problem, encourages sparse portfolios (i.e. portfolios with only few active positions), and allows to account for transaction costs. Our approach recovers as special cases the no-short-positions portfolios, but does allow for short positions in limited number. We implement this methodology on two benchmark data sets constructed by Fama and French . Using only a modest amount of training data, we construct portfolios whose out-of-sample performance, as measured by Sharpe ratio, is consistently and significantly better than that of the naive evenly-weighted portfolio which constitutes, as shown in recent literature, a very tough benchmark .
[ { "type": "R", "before": "optimizing framework, which has served as the basis for modern portfolio theory for more than 50 years. To stabilize the problem, we", "after": "framework, reformulated as a constrained least-squares regression problem. We", "start_char_pos": 92, "end_char_pos": 224 }, { "type": "D", "before": "Markowitz", "after": null, "start_char_pos": 247, "end_char_pos": 256 }, { "type": "D", "before": "which is", "after": null, "start_char_pos": 286, "end_char_pos": 294 }, { "type": "D", "before": "(\\ell_1 penalty)", "after": null, "start_char_pos": 367, "end_char_pos": 383 }, { "type": "R", "before": "stabilizes", "after": "regularizes (stabilizes)", "start_char_pos": 399, "end_char_pos": 409 }, { "type": "D", "before": "automatically", "after": null, "start_char_pos": 436, "end_char_pos": 449 }, { "type": "R", "before": ", and facilitates an effective treatment of", "after": "(i.e. portfolios with only few active positions), and allows to account for", "start_char_pos": 479, "end_char_pos": 522 }, { "type": "R", "before": "We implement our methodology using as our securities two sets of portfolios", "after": "Our approach recovers as special cases the no-short-positions portfolios, but does allow for short positions in limited number. We implement this methodology on two benchmark data sets", "start_char_pos": 542, "end_char_pos": 617 }, { "type": "R", "before": ": the 48 industry portfolios and 100 portfolios formed on size and book-to-market. In addition to their excellent performance, these portfolios have only a small number of active positions, a desirable feature for small investors, for whom the fixed overhead portion of the transaction cost is not negligible", "after": ". Using only a modest amount of training data, we construct portfolios whose out-of-sample performance, as measured by Sharpe ratio, is consistently and significantly better than that of the naive evenly-weighted portfolio which constitutes, as shown in recent literature, a very tough benchmark", "start_char_pos": 649, "end_char_pos": 957 } ]
[ 0, 195, 385, 541, 731 ]
0708.0370
1
Biochemical reaction systems with a low to moderate number of molecules are typically modeled as discrete jump Markov processes. These systems are oftentimes simulated using the Gillespie Algorithm or the Next Reaction Method , which are exact simulation methods . In this paper we make explicit use of the fact that the initiation times of each reaction are given by the firing times of an independent, unit Poisson process with integrated propensity function. We use this representation to develop a modified version of the Next Reaction Method . We then demonstrate how this modified Next Reaction Method is the appropriate method to use to simulate systems in which the rate constants are functions of time (that is, systems that are time dependent Markov Processes). Finally, we extend our modified Next Reaction Method to systems with delays and compare its efficiency with those of the recent algorithms designed for systems with delays.
Chemical reaction systems with a low to moderate number of molecules are typically modeled as discrete jump Markov processes. These systems are oftentimes simulated with methods that produce statistically exact sample paths such as the Gillespie Algorithm or the Next Reaction Method . In this paper we make explicit use of the fact that the initiation times of the reactions can be represented as the firing times of independent, unit rate Poisson processes with internal times given by integrated propensity functions. Using this representation we derive a modified Next Reaction Method and, in a way that achieves efficiency over existing approaches for exact simulation, extend it to systems with time dependent propensities as well as to systems with delays.
[ { "type": "R", "before": "Biochemical", "after": "Chemical", "start_char_pos": 0, "end_char_pos": 11 }, { "type": "R", "before": "using", "after": "with methods that produce statistically exact sample paths such as", "start_char_pos": 168, "end_char_pos": 173 }, { "type": "D", "before": ", which are exact simulation methods", "after": null, "start_char_pos": 226, "end_char_pos": 262 }, { "type": "R", "before": "each reaction are given by the", "after": "the reactions can be represented as the", "start_char_pos": 341, "end_char_pos": 371 }, { "type": "D", "before": "an", "after": null, "start_char_pos": 388, "end_char_pos": 390 }, { "type": "R", "before": "Poisson process with integrated propensity function. We use this representation to develop a modified version of the", "after": "rate Poisson processes with internal times given by integrated propensity functions. Using this representation we derive a modified", "start_char_pos": 409, "end_char_pos": 525 }, { "type": "R", "before": ". We then demonstrate how this modified Next Reaction Method is the appropriate method to use to simulate systems in which the rate constants are functions of time (that is, systems that are time dependent Markov Processes). Finally, we extend our modified Next Reaction Method", "after": "and, in a way that achieves efficiency over existing approaches for exact simulation, extend it", "start_char_pos": 547, "end_char_pos": 824 }, { "type": "R", "before": "delays and compare its efficiency with those of the recent algorithms designed for", "after": "time dependent propensities as well as to", "start_char_pos": 841, "end_char_pos": 923 } ]
[ 0, 128, 264, 461, 548, 771 ]
0708.0377
1
By explicitly representing well-stirred chemical reaction systems with independent, unit Poisson processes we develop a new adaptive tau-leaping procedure. The procedure developed is novel in that we enforce any leap condition via a post-leap check as opposed to performing a pre-leap tau selection. Further, we perform the post-leap check in such a way that the statistics of the sample paths generated will not be skewed by the rejection of a leap. By performing a post-leap check to ensure accuracy, the method developed in this paper is guaranteed to never produce negative population values . The efficiency of the method developed here is demonstrated on a model of a decaying dimer .
By explicitly representing the reaction times of discrete chemical systems as the firing times of independent, unit rate Poisson processes we develop a new adaptive tau-leaping procedure. The procedure developed is novel in that accuracy is guaranteed by performing post-leap checks. Because the representation we use separates the randomness of the model from the state of the system, we are able to perform the post-leap checks in such a way that the statistics of the sample paths generated will not be skewed by the rejections of leaps. Further, since any leap condition is ensured with a probability of one, the simulation method naturally avoids negative population values .
[ { "type": "R", "before": "well-stirred chemical reaction systems with", "after": "the reaction times of discrete chemical systems as the firing times of", "start_char_pos": 27, "end_char_pos": 70 }, { "type": "A", "before": null, "after": "rate", "start_char_pos": 89, "end_char_pos": 89 }, { "type": "R", "before": "we enforce any leap condition via a", "after": "accuracy is guaranteed by performing", "start_char_pos": 198, "end_char_pos": 233 }, { "type": "R", "before": "check as opposed to performing a pre-leap tau selection. Further, we", "after": "checks. Because the representation we use separates the randomness of the model from the state of the system, we are able to", "start_char_pos": 244, "end_char_pos": 312 }, { "type": "R", "before": "check", "after": "checks", "start_char_pos": 335, "end_char_pos": 340 }, { "type": "R", "before": "rejection of a leap. By performing a post-leap check to ensure accuracy, the method developed in this paper is guaranteed to never produce", "after": "rejections of leaps. Further, since any leap condition is ensured with a probability of one, the simulation method naturally avoids", "start_char_pos": 431, "end_char_pos": 569 }, { "type": "D", "before": ". The efficiency of the method developed here is demonstrated on a model of a decaying dimer", "after": null, "start_char_pos": 597, "end_char_pos": 689 } ]
[ 0, 156, 300, 451, 598 ]
0708.0377
2
By explicitly representing the reaction times of discrete chemical systems as the firing times of independent, unit rate Poisson processes we develop a new adaptive tau-leaping procedure. The procedure developed is novel in that accuracy is guaranteed by performing post-leap checks. Because the representation we use separates the randomness of the model from the state of the system, we are able to perform the post-leap checks in such a way that the statistics of the sample paths generated will not be skewed by the rejections of leaps. Further, since any leap condition is ensured with a probability of one, the simulation method naturally avoids negative population values .
By explicitly representing the reaction times of discrete chemical systems as the firing times of independent, unit rate Poisson processes , we develop a new adaptive tau-leaping procedure. The procedure developed is novel in that accuracy is guaranteed by performing postleap checks. Because the representation we use separates the randomness of the model from the state of the system, we are able to perform the postleap checks in such a way that the statistics of the sample paths generated will not be biased by the rejections of leaps. Further, since any leap condition is ensured with a probability of one, the simulation method naturally avoids negative population values
[ { "type": "A", "before": null, "after": ",", "start_char_pos": 139, "end_char_pos": 139 }, { "type": "R", "before": "post-leap", "after": "postleap", "start_char_pos": 267, "end_char_pos": 276 }, { "type": "R", "before": "post-leap", "after": "postleap", "start_char_pos": 414, "end_char_pos": 423 }, { "type": "R", "before": "skewed", "after": "biased", "start_char_pos": 507, "end_char_pos": 513 }, { "type": "D", "before": ".", "after": null, "start_char_pos": 680, "end_char_pos": 681 } ]
[ 0, 188, 284, 541 ]
0708.0544
1
Continuous time random walks are a well suited tool for the description of market behaviour at the smallest scale: the tick-to-tick evolution. We will apply this kind of market model to the valuation of perpetual American options: derivatives with no maturity that can be exercised at any time. Our approach leads to option prices that fulfil the financial formulas when canonical assumptions on the dynamics governing the process are made, but it is still suitable for considering more exotic market conditions.
Continuous-time random walks are a well suited tool for the description of market behaviour at the smallest scale: the tick-to-tick evolution. We will apply this kind of market model to the valuation of perpetual American options: derivatives with no maturity that can be exercised at any time. Our approach leads to option prices that fulfil financial formulas when canonical assumptions on the dynamics governing the process are made, but it is still suitable for more exotic market conditions.
[ { "type": "R", "before": "Continuous time", "after": "Continuous-time", "start_char_pos": 0, "end_char_pos": 15 }, { "type": "D", "before": "the", "after": null, "start_char_pos": 343, "end_char_pos": 346 }, { "type": "D", "before": "considering", "after": null, "start_char_pos": 470, "end_char_pos": 481 } ]
[ 0, 142, 294 ]
0708.0703
1
Neurons in brains are subject to various kinds of noises. Beside of the synaptic noise, the stochastic opening and closing of ion channels represents an intrinsic source of noise that affects the signal processing properties of the neuron. Here we investigated the response of a stochastic Hodgkin-Huxley neuron model to transient input pulses. We found that the response (firing or no firing), as well as the response time, is dependent on the state of the neuron at the moment when input pulse is applied. The state-dependent properties of the response is studied with phase plane analysis method. Using a simple pulse detectionscenario, we demonstrated channel noise enable the neuron to detect subthreshold signals. A simple neuronal network which can marvelously enhance the pulses detecting ability was also proposed. The phenomena of intrinsic stochastic resonance is found both in single neuron level and network level. In single neuron level, the detection ability of the neuron was optimized versus the ion channel patch size(i. e., channel noise intensity). Whereas in network level, the detection ability of the network was optimized versus the number of neurons involved in .
Neurons are subject to various kinds of noise. In addition to synaptic noise, the stochastic opening and closing of ion channels represents an intrinsic source of noise that affects the signal processing properties of the neuron. In this paper, we studied the response of a stochastic Hodgkin-Huxley neuron to transient input subthreshold pulses. It was found that the average response time decreases but variance increases as the amplitude of channel noise increases. In the case of single pulse detection, we show that channel noise enables one neuron to detect the subthreshold signals and an optimal membrane area (or channel noise intensity) exists for a single neuron to achieve optimal performance. However, the detection ability of a single neuron is limited by large errors. Here, we test a simple neuronal network that can enhance the pulse detecting abilities of neurons and find dozens of neurons can perfectly detect subthreshold pulses. The phenomenon of intrinsic stochastic resonance is also found both at the level of single neurons and at the level of networks. At the network level, the detection ability of networks can be optimized for the number of neurons comprising the network .
[ { "type": "D", "before": "in brains", "after": null, "start_char_pos": 8, "end_char_pos": 17 }, { "type": "R", "before": "noises. Beside of the", "after": "noise. In addition to", "start_char_pos": 50, "end_char_pos": 71 }, { "type": "R", "before": "Here we investigated", "after": "In this paper, we studied", "start_char_pos": 240, "end_char_pos": 260 }, { "type": "D", "before": "model", "after": null, "start_char_pos": 312, "end_char_pos": 317 }, { "type": "R", "before": "pulses. We", "after": "subthreshold pulses. It was", "start_char_pos": 337, "end_char_pos": 347 }, { "type": "R", "before": "response (firing or no firing), as well as the response time, is dependent on the state of the neuron at the moment when input pulse is applied. The state-dependent properties of the response is studied with phase plane analysis method. Using a simple pulse detectionscenario, we demonstrated channel noise enable the", "after": "average response time decreases but variance increases as the amplitude of channel noise increases. In the case of single pulse detection, we show that channel noise enables one", "start_char_pos": 363, "end_char_pos": 680 }, { "type": "R", "before": "subthreshold signals. A", "after": "the subthreshold signals and an optimal membrane area (or channel noise intensity) exists for a single neuron to achieve optimal performance. However, the detection ability of a single neuron is limited by large errors. Here, we test a", "start_char_pos": 698, "end_char_pos": 721 }, { "type": "R", "before": "which can marvelously enhance the pulses detecting ability was also proposed. The phenomena", "after": "that can enhance the pulse detecting abilities of neurons and find dozens of neurons can perfectly detect subthreshold pulses. The phenomenon", "start_char_pos": 746, "end_char_pos": 837 }, { "type": "R", "before": "found both in single neuron level and network level. In single neuron level, the detection ability of the neuron was optimized versus the ion channel patch size(i. e., channel noise intensity). Whereas in", "after": "also found both at the level of single neurons and at the level of networks. At the", "start_char_pos": 875, "end_char_pos": 1079 }, { "type": "R", "before": "the network was optimized versus", "after": "networks can be optimized for", "start_char_pos": 1120, "end_char_pos": 1152 }, { "type": "R", "before": "involved in", "after": "comprising the network", "start_char_pos": 1175, "end_char_pos": 1186 } ]
[ 0, 57, 239, 344, 507, 599, 719, 823, 927, 1068 ]
0708.2244
1
Random Threshold Networks (RTN) are widely used in modeling biological systems, such as neural and gene regulatory networks. RTNs with binary elements and random topology have been found to display a phase transition to an ordered phase at low connectivities Physica A 310, 245 (2002)%DIFDELCMD < ]%%% . Here we show that this phase transition is the result of a seemingly minor detail in the update rule, which acts like a biasing external field for a paramagnetic system. For unbiased update rules, there is no phase transition for RTNs of nonzero connectivity. We also show that RTNs with constant number of inputs, K, which have been claimed to be ordered for K=1 Phys. Lett. A 129, 157 (1988)%DIFDELCMD < ]%%% , are actually critical .
%DIFDELCMD < ]%%% %DIFDELCMD < ]%%% This paper has been withdrawn .
[ { "type": "D", "before": "Random Threshold Networks (RTN) are widely used in modeling biological systems, such as neural and gene regulatory networks. RTNs with binary elements and random topology have been found to display a phase transition to an ordered phase at low connectivities", "after": null, "start_char_pos": 0, "end_char_pos": 258 }, { "type": "D", "before": "Physica A 310, 245 (2002)", "after": null, "start_char_pos": 259, "end_char_pos": 284 }, { "type": "D", "before": ". Here we show that this phase transition is the result of a seemingly minor detail in the update rule, which acts like a biasing external field for a paramagnetic system. For unbiased update rules, there is no phase transition for RTNs of nonzero connectivity. We also show that RTNs with constant number of inputs, K, which have been claimed to be ordered for K=1", "after": null, "start_char_pos": 302, "end_char_pos": 667 }, { "type": "D", "before": "Phys. Lett. A 129, 157 (1988)", "after": null, "start_char_pos": 668, "end_char_pos": 697 }, { "type": "R", "before": ", are actually critical", "after": "This paper has been withdrawn", "start_char_pos": 715, "end_char_pos": 738 } ]
[ 0, 124, 473, 563, 673 ]
0708.2707
1
Cooperation plays a key role in the evolution of complex systems. However, the level of cooperation extensively varies with the topology of agent networks in the widely used models of repeated games. Here we show that cooperation remains rather stable by applying long-term learning + innovative strategy adoption rules on a variety of random, regular, small-word, scale-free and modular networks in repeated, multi-agent games. Furthermore, we found that while long-term learning promotes cooperation, innovation makes the level of cooperation less dependent on the actual network topology. Our results demonstrate that long-term learning and innovation , when acting together, extend the range of network topologies enabling the development of cooperation at a wider range of costs and temptations. Learning and innovation help to preserve cooperation during network URLanization , and may be key mechanisms promoting the evolution of URLanizing, complex systems.
Cooperation plays a key role in the evolution of complex systems. However, the level of cooperation extensively varies with the topology of agent networks in the widely used models of repeated games. Here we show that cooperation remains rather stable by applying the reinforcement learning strategy adoption rule, Q-learning on a variety of random, regular, small-word, scale-free and modular network models in repeated, multi-agent Prisoners Dilemma and Hawk-Dove games. Furthermore, we found that using the above model systems other long-term learning strategy adoption rules also promote cooperation, while introducing a low level of noise (as a model of innovation) to the strategy adoption rules makes the level of cooperation less dependent on the actual network topology. Our results demonstrate that long-term learning and random elements in the strategy adoption rules , when acting together, extend the range of network topologies enabling the development of cooperation at a wider range of costs and temptations. These results suggest that a balanced duo of learning and innovation may help to preserve cooperation during the URLanization of real-world networks , and may play a prominent role in the evolution of URLanizing, complex systems.
[ { "type": "R", "before": "long-term learning + innovative strategy adoption rules", "after": "the reinforcement learning strategy adoption rule, Q-learning", "start_char_pos": 264, "end_char_pos": 319 }, { "type": "R", "before": "networks", "after": "network models", "start_char_pos": 388, "end_char_pos": 396 }, { "type": "A", "before": null, "after": "Prisoners Dilemma and Hawk-Dove", "start_char_pos": 422, "end_char_pos": 422 }, { "type": "R", "before": "while", "after": "using the above model systems other", "start_char_pos": 457, "end_char_pos": 462 }, { "type": "R", "before": "promotes cooperation, innovation", "after": "strategy adoption rules also promote cooperation, while introducing a low level of noise (as a model of innovation) to the strategy adoption rules", "start_char_pos": 482, "end_char_pos": 514 }, { "type": "R", "before": "innovation", "after": "random elements in the strategy adoption rules", "start_char_pos": 645, "end_char_pos": 655 }, { "type": "R", "before": "Learning and innovation", "after": "These results suggest that a balanced duo of learning and innovation may", "start_char_pos": 802, "end_char_pos": 825 }, { "type": "R", "before": "network URLanization", "after": "the URLanization of real-world networks", "start_char_pos": 862, "end_char_pos": 882 }, { "type": "R", "before": "be key mechanisms promoting", "after": "play a prominent role in", "start_char_pos": 893, "end_char_pos": 920 } ]
[ 0, 65, 199, 429, 592, 801 ]
0708.2953
1
The numbers of each sort of molecule present in an equilibrium chemical reaction mixture are Poisson distributed . We ask when the same is true of the steady state of a nonequilibrium reaction network and obtain a nearly complete answer. In particular, we show that networks with certain topological features must have a Poisson distribution, whatever the reaction rates. Such driven systems also obey an analog of the fluctuation-dissipation theorem and can equilibrate when in contact with each other . Our results may be relevant to biological systems and to the larger question of how equilibrium ideas might extend to nonequilibrium systems.
In an equilibrium chemical reaction mixture , the number of molecules present obeys a Poisson distribution . We ask when the same is true of the steady state of a nonequilibrium reaction network and obtain an essentially complete answer. In particular, we show that networks with certain topological features must have a Poisson distribution, whatever the reaction rates. Such driven systems also obey an analog of the fluctuation-dissipation theorem . Our results may be relevant to biological systems and to the larger question of how equilibrium concepts might apply to nonequilibrium systems.
[ { "type": "R", "before": "The numbers of each sort of molecule present in", "after": "In", "start_char_pos": 0, "end_char_pos": 47 }, { "type": "R", "before": "are Poisson distributed", "after": ", the number of molecules present obeys a Poisson distribution", "start_char_pos": 89, "end_char_pos": 112 }, { "type": "R", "before": "a nearly", "after": "an essentially", "start_char_pos": 212, "end_char_pos": 220 }, { "type": "D", "before": "and can equilibrate when in contact with each other", "after": null, "start_char_pos": 451, "end_char_pos": 502 }, { "type": "R", "before": "ideas might extend", "after": "concepts might apply", "start_char_pos": 601, "end_char_pos": 619 } ]
[ 0, 114, 237, 371, 504 ]
0709.0561
1
Using our recent contribution, the " partitioned leaping algorithm" [L. A. Harris and P. Clancy, J. Chem. Phys. 125, 144107 (2006)], we investigate the effects of stochasticity in two model biochemical reaction networks. By considering situations both where "leaping" proves beneficial and where it does not, we gain valuable insight that can aid in and advance the use of leaping methods in computational biology. In particular, we identify reaction subnetworks with small populations and large rate constants as a major bottleneck for leaping. We also demonstrate the use of "model reduction" in conjunction with leaping to circumvent this problem and expose a current need in this area: a model-reduction hierarchy analogous to that on which the leaping methods are based D. T. Gillespie, J. Chem. Phys. 113, 297 (2000); 115, 1716 (2001)%DIFDELCMD < ]%%% . In situations where leaping is seen to perform well, we emphasize the significant advantages to using these methods over more traditional simulation approaches. We show how leaping allows for statistical investigations beyond the reach of "exact-stochastic" methods and demonstrate how one can uncover, via such investigations, subtle stochastic effects that would be ignored in deterministic treatments. We then show, using examples taken from the literature, that conditions identified here as lending well to leaping do, in fact, arise in natural biological systems. This leads us to discuss and demonstrate how the statistical information garnered from a leaping study can be used to infer mechanistic details of a biological process and motivate experiments . Overall, we demonstrate to biologists the great potential that leaping methods hold for real-world biological research while highlighting to algorithmists areas still in need of improvement .
" Leaping" methods show great promise for significantly accelerating stochastic simulations of complex biochemical reaction networks. However, few practical applications of leaping have appeared in the literature to date. Here, we address this issue using the " partitioned leaping algorithm" (PLA) [L. A. Harris and P. Clancy, J. Chem. Phys. 125, 144107 (2006)], a recently-introduced multiscale leaping approach. We use the PLA to investigate stochastic effects in two model biochemical reaction networks. %DIFDELCMD < ]%%% The networks that we consider are simple enough so as to be accessible to our intuition but sufficiently complex so as to test the limits of the leaping approach. We demonstrate how the PLA allows us to quantify subtle effects of stochasticity that would be difficult to ascertain otherwise as well as not-so-subtle behaviors that would strain commonly-used "exact" stochastic methods. We also illustrate prime bottlenecks that can hinder the approach and discuss possible strategies for overcoming them . Overall, our aim is to aid and motivate future applications of leaping by providing stark illustrations of the benefits of the method while also elucidating the obstacles that one can expect to encounter .
[ { "type": "D", "before": "Using our recent contribution, the", "after": null, "start_char_pos": 0, "end_char_pos": 34 }, { "type": "A", "before": null, "after": "Leaping\" methods show great promise for significantly accelerating stochastic simulations of complex biochemical reaction networks. However, few practical applications of leaping have appeared in the literature to date. Here, we address this issue using the \"", "start_char_pos": 37, "end_char_pos": 37 }, { "type": "A", "before": null, "after": "(PLA)", "start_char_pos": 69, "end_char_pos": 69 }, { "type": "R", "before": "we investigate the effects of stochasticity", "after": "a recently-introduced multiscale leaping approach. We use the PLA to investigate stochastic effects", "start_char_pos": 135, "end_char_pos": 178 }, { "type": "D", "before": "By considering situations both where \"leaping\" proves beneficial and where it does not, we gain valuable insight that can aid in and advance the use of leaping methods in computational biology. In particular, we identify reaction subnetworks with small populations and large rate constants as a major bottleneck for leaping. We also demonstrate the use of \"model reduction\" in conjunction with leaping to circumvent this problem and expose a current need in this area: a model-reduction hierarchy analogous to that on which the leaping methods are based", "after": null, "start_char_pos": 223, "end_char_pos": 776 }, { "type": "D", "before": "D. T. Gillespie, J. Chem. Phys. 113, 297 (2000); 115, 1716 (2001)", "after": null, "start_char_pos": 777, "end_char_pos": 842 }, { "type": "R", "before": ". In situations where leaping is seen to perform well, we emphasize the significant advantages to using these methods over more traditional simulation approaches. We show how leaping allows for statistical investigations beyond the reach of \"exact-stochastic\" methods and demonstrate how one can uncover, via such investigations, subtle stochastic effects", "after": "The networks that we consider are simple enough so as to be accessible to our intuition but sufficiently complex so as to test the limits of the leaping approach. We demonstrate how the PLA allows us to quantify subtle effects of stochasticity", "start_char_pos": 860, "end_char_pos": 1215 }, { "type": "R", "before": "ignored in deterministic treatments. We then show, using examples taken from the literature, that conditions identified here as lending well to leaping do, in fact, arise in natural biological systems. This leads us to discuss and demonstrate how the statistical information garnered from a leaping study can be used to infer mechanistic details of a biological process and motivate experiments", "after": "difficult to ascertain otherwise as well as not-so-subtle behaviors that would strain commonly-used \"exact\" stochastic methods. We also illustrate prime bottlenecks that can hinder the approach and discuss possible strategies for overcoming them", "start_char_pos": 1230, "end_char_pos": 1624 }, { "type": "R", "before": "we demonstrate to biologists the great potential that leaping methods hold for real-world biological research while highlighting to algorithmists areas still in need of improvement", "after": "our aim is to aid and motivate future applications of leaping by providing stark illustrations of the benefits of the method while also elucidating the obstacles that one can expect to encounter", "start_char_pos": 1636, "end_char_pos": 1816 } ]
[ 0, 107, 222, 416, 547, 802, 825, 1022, 1266, 1431, 1626 ]
0709.0561
2
"Leaping" methods show great promise for significantly accelerating stochastic simulations of complex biochemical reaction networks. However, few practical applications of leaping have appeared in the literature to date. Here, we address this issue using the "partitioned leaping algorithm" (PLA) [L.A. Harris and P. Clancy, J. Chem. Phys. 125, 144107 (2006)], a recently-introduced multiscale leaping approach. We use the PLA to investigate stochastic effects in two model biochemical reaction networks. The networks that we consider are simple enough so as to be accessible to our intuition but sufficiently complex so as to test the limits of the leaping approach . We demonstrate how the PLA allows us to quantify subtle effects of stochasticity that would be difficult to ascertain otherwise as well as not-so-subtle behaviors that would strain commonly-used "exact" stochastic methods. We also illustrate prime bottlenecks that can hinder the approach and discuss possible strategies for overcoming them. Overall, our aim is to aid and motivate future applications of leaping by providing stark illustrations of the benefits of the method while also elucidating the obstacles that one can expect to encounter .
"Leaping" methods show great promise for significantly accelerating stochastic simulations of complex biochemical reaction networks. However, few practical applications of leaping have appeared in the literature to date. Here, we address this issue using the "partitioned leaping algorithm" (PLA) [L.A. Harris and P. Clancy, J. Chem. Phys. 125, 144107 (2006)], a recently-introduced multiscale leaping approach. We use the PLA to investigate stochastic effects in two model biochemical reaction networks. The networks that we consider are simple enough so as to be accessible to our intuition but sufficiently complex so as to be generally representative of real biological systems . We demonstrate how the PLA allows us to quantify subtle effects of stochasticity in these systems that would be difficult to ascertain otherwise as well as not-so-subtle behaviors that would strain commonly-used "exact" stochastic methods. We also illustrate bottlenecks that can hinder the approach and exemplify and discuss possible strategies for overcoming them. Overall, our aim is to aid and motivate future applications of leaping by providing stark illustrations of the benefits of the method while at the same time elucidating obstacles that are often encountered in practice .
[ { "type": "R", "before": "test the limits of the leaping approach", "after": "be generally representative of real biological systems", "start_char_pos": 627, "end_char_pos": 666 }, { "type": "A", "before": null, "after": "in these systems", "start_char_pos": 750, "end_char_pos": 750 }, { "type": "D", "before": "prime", "after": null, "start_char_pos": 912, "end_char_pos": 917 }, { "type": "A", "before": null, "after": "exemplify and", "start_char_pos": 963, "end_char_pos": 963 }, { "type": "R", "before": "also elucidating the obstacles that one can expect to encounter", "after": "at the same time elucidating obstacles that are often encountered in practice", "start_char_pos": 1153, "end_char_pos": 1216 } ]
[ 0, 132, 220, 333, 411, 504, 668, 892, 1012 ]
0709.3291
1
The major objective of this paper is to introduce a novel method for the functional annotation of genes. Our method is based on a graph theoretical measure we call joint betweenness, which is an extension of the well known betweenness centrality measure, involving pairs of genes. We apply our method to the transcriptional regulatory network of yeast to, first, provide a large scale proof of concept of our method and, second, make predictions about the biological function of previously unknown genes .
This paper has been withdrawn .
[ { "type": "R", "before": "The major objective of this paper is to introduce a novel method for the functional annotation of genes. Our method is based on a graph theoretical measure we call joint betweenness, which is an extension of the well known betweenness centrality measure, involving pairs of genes. We apply our method to the transcriptional regulatory network of yeast to, first, provide a large scale proof of concept of our method and, second, make predictions about the biological function of previously unknown genes", "after": "This paper has been withdrawn", "start_char_pos": 0, "end_char_pos": 503 } ]
[ 0, 104, 280 ]
0709.3606
1
We find that discrete noise of inhibiting (signal) molecules can greatly delay the extinction of a regulated component in a prototypical biochemical regulatory network. We calculate analytically the probability distribution of the metastable state of the regulated component and show that deterministic reaction rate equations may fail in predicting the average number of regulated molecules even when this number is large, and the time is short compared to the mean extinction time.
We find that discrete noise of inhibiting (signal) molecules can greatly delay the extinction of plasmids in a plasmid replication system: a prototypical biochemical regulatory network. We calculate the probability distribution of the metastable state of the plasmids and show on this example that the reaction rate equations may fail in predicting the average number of regulated molecules even when this number is large, and the time is much shorter than the mean extinction time.
[ { "type": "R", "before": "a regulated component in a", "after": "plasmids in a plasmid replication system: a", "start_char_pos": 97, "end_char_pos": 123 }, { "type": "D", "before": "analytically", "after": null, "start_char_pos": 182, "end_char_pos": 194 }, { "type": "R", "before": "regulated component and show that deterministic", "after": "plasmids and show on this example that the", "start_char_pos": 255, "end_char_pos": 302 }, { "type": "R", "before": "short compared to", "after": "much shorter than", "start_char_pos": 440, "end_char_pos": 457 } ]
[ 0, 168 ]
0709.3675
1
Microtubules (MTs), nano-tubes which act as struts in the scaffolding of eucaryotic cells, also serve as tracks for intracellular molecular motor transport in eucaryotic cells. In many biological processes, the length distribution of the MTs is controlled by a class of motor proteins called depolymerases (DPs). A DP diffuses or walks along a MT to reach one of the tips of the MT andthen begins depolymerizing the MT itself . We develop a quantitative model that captures both these processes within a single theoretical framework. We show that the action of both diffusing and walking DPs leads to a length-dependent depolymerization of the MT; however, their effects on the length distribution are different. Under experimental conditions, the diffusing DP produces a non-monotonic distribution of lengths as opposed to a nearly exponential distribution in the case of the walking DP. The former also shows a minimum in the r.m.s fluctuation to mean ratio of the length against the DP concentration, while the latter does not. Our findings imply that a diffusing DP, like MCAK, might be used by the cell for precise length control during cell division, while a walking DP, like Kip3p, would be more useful for fast depolymerization .
In many intracellular processes, the length distribution of microtubules is controlled by a class of motor proteins , called depolymerases. Experiments have shown that, following binding to the surface of a microtubule, depolymerases are transported to the microtubule tip(s) by diffusion or directed walk and, then, depolymerize the microtubule from the tip(s) after accumulating there . We develop a quantitative model to study the depolymerizing action of such a generic motor protein, and its possible effects on the length distribution of microtubules. We show that, when the motor protein concentration in solution exceeds a critical value, a steady state is reached where the distribution is non-monotonic provided the motor processivity is sufficiently small. Our findings suggest that such motor proteins have a central role in ensuring precise control of MT lengths .
[ { "type": "R", "before": "Microtubules (MTs), nano-tubes which act as struts in the scaffolding of eucaryotic cells, also serve as tracks for intracellular molecular motor transport in eucaryotic cells. In many biological", "after": "In many intracellular", "start_char_pos": 0, "end_char_pos": 195 }, { "type": "R", "before": "the MTs", "after": "microtubules", "start_char_pos": 234, "end_char_pos": 241 }, { "type": "R", "before": "called depolymerases (DPs). A DP diffuses or walks along a MT to reach one of the tips of", "after": ", called depolymerases. Experiments have shown that, following binding to", "start_char_pos": 285, "end_char_pos": 374 }, { "type": "R", "before": "MT andthen begins depolymerizing the MT itself", "after": "surface of a microtubule, depolymerases are transported to the microtubule tip(s) by diffusion or directed walk and, then, depolymerize the microtubule from the tip(s) after accumulating there", "start_char_pos": 379, "end_char_pos": 425 }, { "type": "R", "before": "that captures both these processes within a single theoretical framework. We show that the action of both diffusing and walking DPs leads to a length-dependent depolymerization of the MT; however, their", "after": "to study the depolymerizing action of such a generic motor protein, and its possible", "start_char_pos": 460, "end_char_pos": 662 }, { "type": "R", "before": "are different. Under experimental conditions, the diffusing DP produces a", "after": "of microtubules. We show that, when the motor protein concentration in solution exceeds a critical value, a steady state is reached where the distribution is", "start_char_pos": 698, "end_char_pos": 771 }, { "type": "R", "before": "distribution of lengths as opposed to a nearly exponential distribution in the case of the walking DP. The former also shows a minimum in the r.m.s fluctuation to mean ratio of the length against the DP concentration, while the latter does not. Our findings imply that a diffusing DP, like MCAK, might be used by the cell for precise length control during cell division, while a walking DP, like Kip3p, would be more useful for fast depolymerization", "after": "provided the motor processivity is sufficiently small. Our findings suggest that such motor proteins have a central role in ensuring precise control of MT lengths", "start_char_pos": 786, "end_char_pos": 1235 } ]
[ 0, 176, 312, 427, 533, 647, 712, 888, 1030 ]
0709.3675
2
In many intracellular processes, the length distribution of microtubules is controlled by a class of motor proteins, called depolymerases . Experiments have shown that, following binding to the surface of a microtubule, depolymerases are transported to the microtubule tip(s) by diffusion or directed walk and, then, depolymerize the microtubule from the tip(s) after accumulating there. We develop a quantitative model to study the depolymerizing action of such a generic motor protein, and its possible effects on the length distribution of microtubules. We show that, when the motor protein concentration in solution exceeds a critical value, a steady state is reached where the distribution is non-monotonic provided the motor processivity is sufficiently small . Our findings suggest that such motor proteins have a central role in ensuring precise control of MT lengths .
In many intracellular processes, the length distribution of microtubules is controlled by depolymerizing motor proteins . Experiments have shown that, following non-specific binding to the surface of a microtubule, depolymerizers are transported to the microtubule tip(s) by diffusion or directed walk and, then, depolymerize the microtubule from the tip(s) after accumulating there. We develop a quantitative model to study the depolymerizing action of such a generic motor protein, and its possible effects on the length distribution of microtubules. We show that, when the motor protein concentration in solution exceeds a critical value, a steady state is reached where the length distribution is, in general, non-monotonic with a single peak. However, for highly processive motors and large motor densities, this distribution effectively becomes an exponential decay . Our findings suggest that such motor proteins may be selectively used by the cell to ensure precise control of MT lengths . The model is also used to analyze experimental observations of motor-induced depolymerization .
[ { "type": "R", "before": "a class of motor proteins, called depolymerases", "after": "depolymerizing motor proteins", "start_char_pos": 90, "end_char_pos": 137 }, { "type": "A", "before": null, "after": "non-specific", "start_char_pos": 179, "end_char_pos": 179 }, { "type": "R", "before": "depolymerases", "after": "depolymerizers", "start_char_pos": 221, "end_char_pos": 234 }, { "type": "R", "before": "distribution is", "after": "length distribution is, in general,", "start_char_pos": 683, "end_char_pos": 698 }, { "type": "R", "before": "provided the motor processivity is sufficiently small", "after": "with a single peak. However, for highly processive motors and large motor densities, this distribution effectively becomes an exponential decay", "start_char_pos": 713, "end_char_pos": 766 }, { "type": "R", "before": "have a central role in ensuring", "after": "may be selectively used by the cell to ensure", "start_char_pos": 815, "end_char_pos": 846 }, { "type": "A", "before": null, "after": ". The model is also used to analyze experimental observations of motor-induced depolymerization", "start_char_pos": 877, "end_char_pos": 877 } ]
[ 0, 388, 557, 768 ]
0710.0892
1
We develop a network-based algorithm to predict and verify indirect regulatory interactions in a large-scale genetic regulatory network. Indirect regulations are important as they constitute the majority of experimental data. Our approach is based on the network topology and can be easily incorporated using a matrix formalism. The essence of the method is to extend the transitivity of indirect regulations ( i.e. A regulates B and B regulates C implies A regulates C) to longer cascades and effectively take care of the signs of the regulations. This algorithm is tailored for large and heavily interconnected networks, which are of growing importance due to the accruement of data from high-throughput experiments. We apply the algorithm to the regulatory networks of Homo sapiens, Saccharomyces cerevisiae, Arabidopsis thaliana and Drosophila melanogaster , resulting in novel predictions with calibrated reliability .
We develop a matrix-based approach to predict and verify indirect interactions in gene and protein regulatory networks. It is based on the approximate transitivity of indirect regulations ( e.g. A regulates B and B regulates C often implies that A regulates C) and optimally takes into account the length of a cascade and signs of intermediate interactions. Our method is at its most powerful when applied to large and densely interconnected networks. It successfully predicts both the yet unknown indirect regulations, as well as the sign (activation or repression) of already known ones. The reliability of sign predictions was calibrated using the gold-standard sets of positive and negative interactions. We fine-tuned the parameters of our algorithm by maximizing the area under the Receiver Operating Characteristic (ROC) curve. We then applied the optimized algorithm to large literature-derived networks of all direct and indirect regulatory interactions in several URLanisms ( Homo sapiens, Saccharomyces cerevisiae, Arabidopsis thaliana and Drosophila melanogaster ) .
[ { "type": "R", "before": "network-based algorithm", "after": "matrix-based approach", "start_char_pos": 13, "end_char_pos": 36 }, { "type": "R", "before": "regulatory interactions in a large-scale genetic regulatory network. Indirect regulations are important as they constitute the majority of experimental data. Our approach", "after": "interactions in gene and protein regulatory networks. It", "start_char_pos": 68, "end_char_pos": 238 }, { "type": "R", "before": "network topology and can be easily incorporated using a matrix formalism. The essence of the method is to extend the", "after": "approximate", "start_char_pos": 255, "end_char_pos": 371 }, { "type": "R", "before": "i.e.", "after": "e.g.", "start_char_pos": 411, "end_char_pos": 415 }, { "type": "R", "before": "implies", "after": "often implies that", "start_char_pos": 448, "end_char_pos": 455 }, { "type": "R", "before": "to longer cascades and effectively take care of the signs of the regulations. This algorithm is tailored for large and heavily interconnected networks, which are of growing importance due to the accruement of data from high-throughput experiments. We apply the algorithm to the regulatory networks of", "after": "and optimally takes into account the length of a cascade and signs of intermediate interactions. Our method is at its most powerful when applied to large and densely interconnected networks. It successfully predicts both the yet unknown indirect regulations, as well as the sign (activation or repression) of already known ones. The reliability of sign predictions was calibrated using the gold-standard sets of positive and negative interactions. We fine-tuned the parameters of our algorithm by maximizing the area under the Receiver Operating Characteristic (ROC) curve. We then applied the optimized algorithm to large literature-derived networks of all direct and indirect regulatory interactions in several URLanisms (", "start_char_pos": 471, "end_char_pos": 771 }, { "type": "R", "before": ", resulting in novel predictions with calibrated reliability", "after": ")", "start_char_pos": 861, "end_char_pos": 921 } ]
[ 0, 136, 225, 328, 415, 548, 718 ]
0710.3959
1
The t copula is often used in risk management as it allows for modelling tail dependence between risks and it is simple to simulate and calibrate. However the use of a standard t copula is often criticized due to its restriction of having a single parameter for the degrees of freedom (dof) that may limit its capability to model the tail dependence structure in a multivariate case. To overcome this problem, grouped t copula was proposed recently where risks are grouped a priori in such a way that each group has a standard t copula with its specific dof parameter. In this paper we propose the use of a new t copula, generalizing grouped t copula to have each group consisting of one risk only, so that a priori grouping is not required. The characteristics of this copula in the bivariate case are described . We explain simulation and calibration procedures and provide examples .
The t copula is often used in risk management as it allows for modelling tail dependence between risks and it is simple to simulate and calibrate. However , the use of a standard t copula is often criticized due to its restriction of having a single parameter for the degrees of freedom (dof) that may limit its capability to model the tail dependence structure in a multivariate case. To overcome this problem, grouped t copula was proposed recently , where risks are grouped a priori in such a way that each group has a standard t copula with its specific dof parameter. In this paper we propose the use of a grouped t copula , where each group consists of one risk factor only, so that a priori grouping is not required. The copula characteristics in the bivariate case are studied . We explain simulation and calibration procedures , including a simulation study on finite sample properties of the maximum likelihood estimators and Kendall's tau approximation. This new copula can be significantly different from the standard t copula in terms of risk measures such as tail dependence, value at risk and expected shortfall. Keywords: grouped t copula, tail dependence, risk management .
[ { "type": "A", "before": null, "after": ",", "start_char_pos": 155, "end_char_pos": 155 }, { "type": "A", "before": null, "after": ",", "start_char_pos": 450, "end_char_pos": 450 }, { "type": "D", "before": "new t copula, generalizing", "after": null, "start_char_pos": 609, "end_char_pos": 635 }, { "type": "R", "before": "to have each group consisting", "after": ", where each group consists", "start_char_pos": 653, "end_char_pos": 682 }, { "type": "A", "before": null, "after": "factor", "start_char_pos": 695, "end_char_pos": 695 }, { "type": "R", "before": "characteristics of this copula", "after": "copula characteristics", "start_char_pos": 749, "end_char_pos": 779 }, { "type": "R", "before": "described", "after": "studied", "start_char_pos": 806, "end_char_pos": 815 }, { "type": "R", "before": "and provide examples", "after": ", including a simulation study on finite sample properties of the maximum likelihood estimators and Kendall's tau approximation. This new copula can be significantly different from the standard t copula in terms of risk measures such as tail dependence, value at risk and expected shortfall. Keywords: grouped t copula, tail dependence, risk management", "start_char_pos": 867, "end_char_pos": 887 } ]
[ 0, 146, 384, 570, 744, 817 ]
0710.4127
1
Stimulated by the long term goal of identifying gene targets for corn crop improvement, we perturbed a simple plant cell system by the addition of abscisic acid, a well-characterized plant hormone, and measured the gene expression with transcriptional microarrays for 150 minutes, every 10 minutes after treatment. Parameters of biochemistry-based models were inferred for 25 downstream (regulated) genes of interest using a thoroughly studied approximation of the posterior distribution, accounting for model uncertainty through Bayesian model averaging. The four causal gene interactions thereby identified as probable under a first-order difference model present testable hypotheses for future experimentation that could conceivably build knowledge of the complex biological system. The small scale of this result compared to many of the reconstructed networks reported in recent literature reflects a cautious approach to systems biology, but without a priori skepticism. The number of putative genes represented on the microarrays used is on the order of the size of the genome; a modification of the methodology for cases in which the number of measured genes is a small fraction of the genome size is also provided. For purposes of inferring biological networks, the main advantage of microarray technology over low-throughput methods may be that the representation of a large enough portion of the genome enables computation of an informative upper bound of how much confidence one may reasonably place in causal relationships between genes on the basis of observed data.
Motivation: Measurements of gene expression over time enable the reconstruction of transcriptional networks. However, Bayesian networks and many other current reconstruction methods rely on assumptions that conflict with the differential equations that describe transcriptional kinetics. Practical approximations of kinetic models would enable inferring causal relationships between genes from expression data of microarray, tag-based and conventional platforms, but conclusions are sensitive to the assumptions made. Results: The representation of a sufficiently large portion of genome enables computation of an upper bound on how much confidence one may place in influences between genes on the basis of expression data. Information about which genes encode transcription factors is not necessary but may be incorporated if available. The methodology is generalized to cover cases in which expression measurements are missing for many of the genes that might control the transcription of the genes of interest. The assumption that the gene expression level is roughly proportional to the rate of translation led to better empirical performance than did either the assumption that the gene expression level is roughly proportional to the protein level or the Bayesian model average of both assumptions. Availability: URL points to R code implementing the methods (R Development Core Team 2004). Supplementary information: URL
[ { "type": "R", "before": "Stimulated by the long term goal of identifying gene targets for corn crop improvement, we perturbed a simple plant cell system by the addition of abscisic acid, a well-characterized plant hormone, and measured the gene expression with transcriptional microarrays for 150 minutes, every 10 minutes after treatment. Parameters of biochemistry-based models were inferred for 25 downstream (regulated) genes of interest using a thoroughly studied approximation of the posterior distribution, accounting for model uncertainty through Bayesian model averaging. The four causal gene interactions thereby identified as probable under a first-order difference model present testable hypotheses for future experimentation that could conceivably build knowledge of the complex biological system. The small scale of this result compared to many of the reconstructed networks reported in recent literature reflects a cautious approach to systems biology, but without a priori skepticism. The number of putative genes represented on the microarrays used is on the order of the size of the genome; a modification of the methodology for cases in which the number of measured genes is a small fraction of the genome size is also provided. For purposes of inferring biological networks, the main advantage of microarray technology over low-throughput methods may be that the", "after": "Motivation: Measurements of gene expression over time enable the reconstruction of transcriptional networks. However, Bayesian networks and many other current reconstruction methods rely on assumptions that conflict with the differential equations that describe transcriptional kinetics. Practical approximations of kinetic models would enable inferring causal relationships between genes from expression data of microarray, tag-based and conventional platforms, but conclusions are sensitive to the assumptions made. Results: The", "start_char_pos": 0, "end_char_pos": 1357 }, { "type": "R", "before": "large enough portion of the", "after": "sufficiently large portion of", "start_char_pos": 1378, "end_char_pos": 1405 }, { "type": "R", "before": "informative upper bound of", "after": "upper bound on", "start_char_pos": 1439, "end_char_pos": 1465 }, { "type": "R", "before": "reasonably place in causal relationships", "after": "place in influences", "start_char_pos": 1494, "end_char_pos": 1534 }, { "type": "R", "before": "observed data.", "after": "expression data. Information about which genes encode transcription factors is not necessary but may be incorporated if available. The methodology is generalized to cover cases in which expression measurements are missing for many of the genes that might control the transcription of the genes of interest. The assumption that the gene expression level is roughly proportional to the rate of translation led to better empirical performance than did either the assumption that the gene expression level is roughly proportional to the protein level or the Bayesian model average of both assumptions. Availability: URL points to R code implementing the methods (R Development Core Team 2004). Supplementary information: URL", "start_char_pos": 1565, "end_char_pos": 1579 } ]
[ 0, 314, 555, 785, 975, 1083, 1222 ]
0711.0916
1
We present a theoretical investigation of the folding of small proteins assisted by chaperones. We describe the proteins in the framework of an effective potential model which contains the Ramachandran angles as degrees of freedom {\it . The cage of chaperonins is modeled by an external confining potential which is also able to take into account hydrophobic and hydrophilic effects inside the cavity. Using the Wang-Landau algorithm Phys. Rev. Lett.%DIFDELCMD < {\bf %%% 86 , 2050 (2001) we determine the density of states g(E ) and analyze in detail the thermodynamical properties of the confined proteins for different sizes of the cage . We show how the confinement through the chaperon dramatically reduces the phase space available for the protein leading to a much faster folding process. Slightly hydrophobic cages seem to make the native structure more stable . However, not any confining potential helps folding . If the inner walls of the cage are strongly hydrophobic, a denaturation process is induced, in which the proteins partially unfold and stick to the walls .
We present a theoretical study of the folding of small proteins inside confining potentials. Proteins are described in the framework of an effective potential model which contains the Ramachandran angles as degrees of freedom and does not need any{\it a priori information about the native state. Hydrogen bonds, dipole-dipole- and hydrophobic interactions are taken explicitly into account. An interesting feature displayed by this potential is the presence of some intermediates between the unfolded and native states. We consider different types of confining potentials in order to study the structural properties of proteins folding inside cages with repulsive or attractive walls . Using the Wang-Landau algorithm %DIFDELCMD < {\bf %%% we determine the density of states (DOS ) and analyze in detail the thermodynamical properties of the confined proteins for different sizes of the cages . We show that confinement dramatically reduces the phase space available to the protein and that the presence of intermediate states can be controlled by varying the properties of the confining potential. Cages with strongly attractive walls lead to the disappearance of the intermediate states and to a two-state folding into a less stable configuration. However, cages with slightly attractive walls make the native structure more stable than in the case of pure repulsive potentials, and the folding process occurs through intermediate configurations. In order to test the metastable states we analyze the free energy landscapes as a function of the configurational energy and of the end-to-end distance as an order parameter .
[ { "type": "R", "before": "investigation", "after": "study", "start_char_pos": 25, "end_char_pos": 38 }, { "type": "R", "before": "assisted by chaperones. We describe the proteins", "after": "inside confining potentials. Proteins are described", "start_char_pos": 72, "end_char_pos": 120 }, { "type": "A", "before": null, "after": "and does not need any", "start_char_pos": 231, "end_char_pos": 231 }, { "type": "A", "before": null, "after": "a priori", "start_char_pos": 236, "end_char_pos": 236 }, { "type": "A", "before": null, "after": "information about the native state. Hydrogen bonds, dipole-dipole- and hydrophobic interactions are taken explicitly into account. An interesting feature displayed by this potential is the presence of some intermediates between the unfolded and native states. We consider different types of confining potentials in order to study the structural properties of proteins folding inside cages with repulsive or attractive walls", "start_char_pos": 237, "end_char_pos": 237 }, { "type": "D", "before": "The cage of chaperonins is modeled by an external confining potential which is also able to take into account hydrophobic and hydrophilic effects inside the cavity.", "after": null, "start_char_pos": 240, "end_char_pos": 404 }, { "type": "D", "before": "Phys. Rev. Lett.", "after": null, "start_char_pos": 437, "end_char_pos": 453 }, { "type": "D", "before": "86", "after": null, "start_char_pos": 475, "end_char_pos": 477 }, { "type": "D", "before": ", 2050 (2001)", "after": null, "start_char_pos": 478, "end_char_pos": 491 }, { "type": "R", "before": "g(E", "after": "(DOS", "start_char_pos": 527, "end_char_pos": 530 }, { "type": "R", "before": "cage", "after": "cages", "start_char_pos": 638, "end_char_pos": 642 }, { "type": "R", "before": "how the confinement through the chaperon", "after": "that confinement", "start_char_pos": 653, "end_char_pos": 693 }, { "type": "R", "before": "for the protein leading to a much faster folding process. Slightly hydrophobic cages seem to", "after": "to the protein and that the presence of intermediate states can be controlled by varying the properties of the confining potential. Cages with strongly attractive walls lead to the disappearance of the intermediate states and to a two-state folding into a less stable configuration. However, cages with slightly attractive walls", "start_char_pos": 741, "end_char_pos": 833 }, { "type": "R", "before": ". However, not any confining potential helps folding . If the inner walls of the cage are strongly hydrophobic, a denaturation process is induced, in which the proteins partially unfold and stick to the walls", "after": "than in the case of pure repulsive potentials, and the folding process occurs through intermediate configurations. In order to test the metastable states we analyze the free energy landscapes as a function of the configurational energy and of the end-to-end distance as an order parameter", "start_char_pos": 872, "end_char_pos": 1080 } ]
[ 0, 95, 239, 404, 442, 453, 644, 798, 873, 926 ]
0711.2624
1
In this paper we will develop an original approach, based in the use of renewal equations, for obtaining pricing expressions for financial instruments whose underlying asset can be solely described through a simple continuous-time random walk (CTRW). This enhances the potential use of CTRW techniques in finance. We solve these equations for different contract specifications in a particular but exemplifying case. We recover the celebrated results for the Wiener process under certain limits.
In this paper we will develop an original approach, based in the use of renewal equations, for obtaining pricing expressions for financial instruments whose underlying asset can be solely described through a simple continuous-time random walk (CTRW). This enhances the potential use of CTRW techniques in finance. We solve these equations for typical contract specifications, in a particular but exemplifying case. We also show how a formal general solution can be found for more exotic derivatives, and we compare prices for alternative models of the underlying. Finally, we recover the celebrated results for the Wiener process under certain limits.
[ { "type": "R", "before": "different contract specifications", "after": "typical contract specifications,", "start_char_pos": 343, "end_char_pos": 376 }, { "type": "A", "before": null, "after": "also show how a formal general solution can be found for more exotic derivatives, and we compare prices for alternative models of the underlying. Finally, we", "start_char_pos": 419, "end_char_pos": 419 } ]
[ 0, 250, 313, 415 ]
0711.2624
2
In this paper we will develop an original approach, based in the use of renewal equations, for obtaining pricing expressions for financial instruments whose underlying asset can be solely described through a simple continuous-time random walk (CTRW) . This enhances the potential use of CTRW techniques in finance. We solve these equations for typical contract specifications, in a particular but exemplifying case. We also show how a formal general solution can be found for more exotic derivatives, and we compare prices for alternative models of the underlying. Finally, we recover the celebrated results for the Wiener process under certain limits.
In this paper we will develop a methodology for obtaining pricing expressions for financial instruments whose underlying asset can be described through a simple continuous-time random walk (CTRW) market model. Our approach is very natural to the issue because it is based in the use of renewal equations, and therefore it enhances the potential use of CTRW techniques in finance. We solve these equations for typical contract specifications, in a particular but exemplifying case. We also show how a formal general solution can be found for more exotic derivatives, and we compare prices for alternative models of the underlying. Finally, we recover the celebrated results for the Wiener process under certain limits.
[ { "type": "R", "before": "an original approach, based in the use of renewal equations,", "after": "a methodology", "start_char_pos": 30, "end_char_pos": 90 }, { "type": "D", "before": "solely", "after": null, "start_char_pos": 181, "end_char_pos": 187 }, { "type": "R", "before": ". This", "after": "market model. Our approach is very natural to the issue because it is based in the use of renewal equations, and therefore it", "start_char_pos": 250, "end_char_pos": 256 } ]
[ 0, 314, 415, 564 ]
0711.2799
1
Boolean networks have been the object of much attention, especially since S. Kauffman proposed them in the 1960's as models for gene regulatory networks. These systems are characterized by being defined on a Boolean state space and by simultaneous updating at discrete time steps. Of particular importance for biological applications are networks in which the indegree for each variable is bounded by a fixed constant, as was stressed by Kauffman in his original papers. An important question is which conditions on the network topology can rule out exponentially long periodic orbits in the system. In this paper, we consider systems with positive feedback interconnections among all variables , which in a continuous setting guarantees a very stable dynamics. We present a construction that shows that for an arbitrary constant 0<c<2 and sufficiently large n there exist n-dimensional Boolean networks with this property in which both the indegree and outdegree of each for each variable is bounded by two and which nevertheless contain periodic orbits of length at least c^n. In Part II of this paper we will prove an inverse result showing that any system with such a dynamic behavior must in a sense be similar to the example described.
Boolean networks have been the object of much attention, especially since S. Kauffman proposed them in the 1960's as models for gene regulatory networks. These systems are characterized by being defined on a Boolean state space and by simultaneous updating at discrete time steps. Of particular importance for biological applications are networks in which the indegree for each variable is bounded by a fixed constant, as was stressed by Kauffman in his original papers. An important question is which conditions on the network topology can rule out exponentially long periodic orbits in the system. In this paper, we consider systems with positive feedback interconnections among all variables (known as cooperative systems) , which in a continuous setting guarantees a very stable dynamics. We show that for an arbitrary constant 0<c<2 and sufficiently large n there exist n-dimensional cooperative Boolean networks in which both the indegree and outdegree of each variable is bounded by two , and which nevertheless contain periodic orbits of length at least c^n. In Part II of this paper we will prove an inverse result showing that any system with such a dynamic behavior must in a sense be similar to the example described.
[ { "type": "A", "before": null, "after": "(known as cooperative systems)", "start_char_pos": 695, "end_char_pos": 695 }, { "type": "R", "before": "present a construction that shows that", "after": "show that", "start_char_pos": 766, "end_char_pos": 804 }, { "type": "R", "before": "Boolean networks with this property", "after": "cooperative Boolean networks", "start_char_pos": 888, "end_char_pos": 923 }, { "type": "D", "before": "for each", "after": null, "start_char_pos": 973, "end_char_pos": 981 }, { "type": "A", "before": null, "after": ",", "start_char_pos": 1009, "end_char_pos": 1009 } ]
[ 0, 153, 280, 470, 599, 762, 1080 ]
0711.3812
1
Experiments in recent years have vividly demonstrated that gene expression can Hobe highly stochastic. How protein concentration fluctuations affect the growth rate of a population of cells, is, however, a wide open question. We present a mathematical model that makes it possible to quantify the effect of protein concentration fluctuations on the growth rate of a population of genetically identical cells. The model predicts that the population's growth rate depends on how the growth rate of a single cell varies with protein concentration, the variance of the protein concentration fluctuations, and the correlation time of these fluctuations. Only when the correlation time of the fluctuations is long compared to the cell cycle time, do they affect the population's growth rate. The model also predicts that when the average concentration of a protein is close to the value that maximizes the growth rate, fluctuations in its concentration always reduce the growth rate. However, when the average protein concentration deviates sufficiently from the optimal level, fluctuations can enhance the growth rate of the population, even when the growth rate of a cell depends linearly on the protein concentration. We also apply our model to perform a cost-benefit analysis of gene regulatory control. Our analysis predicts that the optimal expression level of a gene regulatory protein is determined by the trade-off between the cost of synthesizing the regulatory protein and the benefit of minimizing the fluctuations in the expression of its target gene. We discuss possible experiments that could test our predictions.
Experiments in recent years have vividly demonstrated that gene expression can be highly stochastic. How protein concentration fluctuations affect the growth rate of a population of cells, is, however, a wide open question. We present a mathematical model that makes it possible to quantify the effect of protein concentration fluctuations on the growth rate of a population of genetically identical cells. The model predicts that the population's growth rate depends on how the growth rate of a single cell varies with protein concentration, the variance of the protein concentration fluctuations, and the correlation time of these fluctuations. The model also predicts that when the average concentration of a protein is close to the value that maximizes the growth rate, fluctuations in its concentration always reduce the growth rate. However, when the average protein concentration deviates sufficiently from the optimal level, fluctuations can enhance the growth rate of the population, even when the growth rate of a cell depends linearly on the protein concentration. The model also shows that the ensemble or population average of a quantity, such as the average protein expression level or its variance, is in general not equal to its time average as obtained from tracing a single cell and its descendants. We apply our model to perform a cost-benefit analysis of gene regulatory control. Our analysis predicts that the optimal expression level of a gene regulatory protein is determined by the trade-off between the cost of synthesizing the regulatory protein and the benefit of minimizing the fluctuations in the expression of its target gene. We discuss possible experiments that could test our predictions.
[ { "type": "R", "before": "Hobe", "after": "be", "start_char_pos": 79, "end_char_pos": 83 }, { "type": "D", "before": "Only when the correlation time of the fluctuations is long compared to the cell cycle time, do they affect the population's growth rate.", "after": null, "start_char_pos": 649, "end_char_pos": 785 }, { "type": "R", "before": "We also", "after": "The model also shows that the ensemble or population average of a quantity, such as the average protein expression level or its variance, is in general not equal to its time average as obtained from tracing a single cell and its descendants. We", "start_char_pos": 1215, "end_char_pos": 1222 } ]
[ 0, 102, 225, 408, 648, 785, 977, 1214, 1301, 1558 ]
0712.2771
1
We investigate the use of Kelly's strategy in the construction of an optimal portfolio of assets. With asset prices undergoing a multiplicative random process , we derive approximate analytical results for the optimal investment fractions under various constraints . We show that , when returns and volatilities of the assets are small and borrowing is forbidden , the Kelly-optimal portfolio lies on Markowitz Efficient Frontier. When short positions are also forbidden, only a small fraction of the available assets is included in the Kelly-optimal portfolio. This phenomenon, that we call condensation, is explored in detail.
We investigate the use of Kelly's strategy in the construction of an optimal portfolio of assets. For lognormally distributed asset returns , we derive approximate analytical results for the optimal investment fractions in various settings . We show that when mean returns and volatilities of the assets are small and there is no risk-free asset , the Kelly-optimal portfolio lies on Markowitz Efficient Frontier. Since in the investigated case the Kelly approach forbids short positions and borrowing, often only a small fraction of the available assets is included in the Kelly-optimal portfolio. This phenomenon, that we call condensation, is explored in detail.
[ { "type": "R", "before": "With asset prices undergoing a multiplicative random process", "after": "For lognormally distributed asset returns", "start_char_pos": 98, "end_char_pos": 158 }, { "type": "R", "before": "under various constraints", "after": "in various settings", "start_char_pos": 239, "end_char_pos": 264 }, { "type": "R", "before": ", when", "after": "when mean", "start_char_pos": 280, "end_char_pos": 286 }, { "type": "R", "before": "borrowing is forbidden", "after": "there is no risk-free asset", "start_char_pos": 340, "end_char_pos": 362 }, { "type": "R", "before": "When short positions are also forbidden,", "after": "Since in the investigated case the Kelly approach forbids short positions and borrowing, often", "start_char_pos": 431, "end_char_pos": 471 } ]
[ 0, 97, 266, 430, 561 ]
0712.2773
1
Database replication is difficult but indispensable. We report on our experiences building and deploying middleware-based replication systems both as commercial products and research systems. We identify gaps that still separate academic research from industrial practiceand thus thwart potential technology transfer from academia to the field. We structure our analysis along three axes -- performance, availability and management -- and outline unmet database replication challenges at several levels. We hope to both motivate and aid researchers in bridging these gaps between theory and practice . We sift through the last decade of open-source, academic, and commercial database replication systems and combine this material with case studies resulting from developing and deploying real systems at many customer sites . We propose two agendas, one for academic research and one for industrial R&D, which we expect can bridge the gaps within 5-10 years .
The need for high availability and performance in data management systems has been fueling a long running interest in database replication from both academia and industry. However, academic groups often attack replication problems in isolation, overlooking the need for completeness in their solutions, while commercial teams take a holistic approach that often misses opportunities for fundamental innovation. This has created over time a gap between academic research and industrial practice. This paper aims to characterize the gap along three axes : performance, availability , and administration. We build on our own experience developing and deploying replication systems in commercial and academic settings, as well as on a large body of prior related work . We sift through representative examples from the last decade of open-source, academic, and commercial database replication systems and combine this material with case studies from real systems deployed at Fortune 500 customers . We propose two agendas, one for academic research and one for industrial R&D, which we believe can bridge the gap within 5-10 years . This way, we hope to both motivate and help researchers in making the theory and practice of middleware-based database replication more relevant to each other .
[ { "type": "R", "before": "Database replication is difficult but indispensable. We report on our experiences building and deploying middleware-based replication systems both as commercial products and research systems. We identify gaps that still separate academic research from industrial practiceand thus thwart potential technology transfer from academia to the field. We structure our analysis", "after": "The need for high availability and performance in data management systems has been fueling a long running interest in database replication from both academia and industry. However, academic groups often attack replication problems in isolation, overlooking the need for completeness in their solutions, while commercial teams take a holistic approach that often misses opportunities for fundamental innovation. This has created over time a gap between academic research and industrial practice. This paper aims to characterize the gap", "start_char_pos": 0, "end_char_pos": 370 }, { "type": "R", "before": "--", "after": ":", "start_char_pos": 388, "end_char_pos": 390 }, { "type": "R", "before": "and management -- and outline unmet database replication challenges at several levels. We hope to both motivate and aid researchers in bridging these gaps between theory and practice", "after": ", and administration. We build on our own experience developing and deploying replication systems in commercial and academic settings, as well as on a large body of prior related work", "start_char_pos": 417, "end_char_pos": 599 }, { "type": "A", "before": null, "after": "representative examples from", "start_char_pos": 618, "end_char_pos": 618 }, { "type": "R", "before": "resulting from developing and deploying real systems at many customer sites", "after": "from real systems deployed at Fortune 500 customers", "start_char_pos": 749, "end_char_pos": 824 }, { "type": "R", "before": "expect", "after": "believe", "start_char_pos": 914, "end_char_pos": 920 }, { "type": "R", "before": "gaps", "after": "gap", "start_char_pos": 936, "end_char_pos": 940 }, { "type": "A", "before": null, "after": ". This way, we hope to both motivate and help researchers in making the theory and practice of middleware-based database replication more relevant to each other", "start_char_pos": 959, "end_char_pos": 959 } ]
[ 0, 52, 191, 344, 503, 601, 826 ]
0712.3485
1
Using Malliavin calculus techniques, we derive an analytical formula for the price of European options, for any model including local volatility and jump Poisson process. We show that the accuracy of the formula depends on the smoothness of the payoff . Our approach relies on an asymptotic expansion related to small diffusion and small jump frequency . As a consequence, the calibration of such model becomes very fast .
Using Malliavin calculus techniques, we derive an analytical formula for the price of European options, for any model including local volatility and Poisson jump process. We show that the accuracy of the formula depends on the smoothness of the payoff function . Our approach relies on an asymptotic expansion related to small diffusion and small jump frequency /size. Our formula has excellent accuracy (the error on implied Black-Scholes volatilities for call option is smaller than 2 bp for various strikes and maturities). Additionally, model calibration becomes very rapid .
[ { "type": "R", "before": "jump Poisson", "after": "Poisson jump", "start_char_pos": 149, "end_char_pos": 161 }, { "type": "A", "before": null, "after": "function", "start_char_pos": 252, "end_char_pos": 252 }, { "type": "R", "before": ". As a consequence, the calibration of such model becomes very fast", "after": "/size. Our formula has excellent accuracy (the error on implied Black-Scholes volatilities for call option is smaller than 2 bp for various strikes and maturities). Additionally, model calibration becomes very rapid", "start_char_pos": 354, "end_char_pos": 421 } ]
[ 0, 170, 254, 355 ]
0712.4304
1
Kinesins are biomolecular motors which move on cylindrical nano-tubes called microtubules . A normal microtubule consists of more than one protofilament on which the equispaced motor binding sites form a periodic array. The collective movement of the kinesins on a microtubule is, therefore, analogous to vehicular traffic on multi-lane highways where each protofilament is the analogue of a single lane. We extend a recent model of the traffic of single-headed kinesin KIF1A [{\it Phys. Rev. E {\bf 75}, 041905 (2007)}] by incorporating processes which correspond to shifting of the motor proteins from one protofilament to another. On the basis of analytical treatment of our model, we predict the effects of lane changing on the flux of the KIF1A motors . Our quantitative predictions can be tested, in principle, by carrying out {\it in-vitro} experiments with fluorescently labelled KIF1A molecules.
KIF1A kinesins are single-headed motor proteins which move on cylindrical nano-tubes called microtubules (MT) . A normal MT consists of 13 protofilaments on which the equispaced motor binding sites form a periodic array. The collective movement of the kinesins on a MT is, therefore, analogous to vehicular traffic on multi-lane highways where each protofilament is the analogue of a single lane. Does lane-changing increase or decrease the motor flux per lane? We address this fundamental question here by appropriately extending a recent model [{\it Phys. Rev. E {\bf 75}, 041905 (2007)}] . By carrying out analytical calculations and computer simulations of this extended model, we predict that the flux per lane can increase or decrease with the increasing rate of lane changing , depending on the concentrations of motors and the rate of hydrolysis of ATP, the ``fuel'' molecules. Our predictions can be tested, in principle, by carrying out {\it in-vitro} experiments with fluorescently labelled KIF1A molecules.
[ { "type": "R", "before": "Kinesins are biomolecular motors", "after": "KIF1A kinesins are single-headed motor proteins", "start_char_pos": 0, "end_char_pos": 32 }, { "type": "A", "before": null, "after": "(MT)", "start_char_pos": 90, "end_char_pos": 90 }, { "type": "R", "before": "microtubule consists of more than one protofilament", "after": "MT consists of 13 protofilaments", "start_char_pos": 102, "end_char_pos": 153 }, { "type": "R", "before": "microtubule", "after": "MT", "start_char_pos": 266, "end_char_pos": 277 }, { "type": "R", "before": "We extend", "after": "Does lane-changing increase or decrease the motor flux per lane? We address this fundamental question here by appropriately extending", "start_char_pos": 406, "end_char_pos": 415 }, { "type": "D", "before": "of the traffic of single-headed kinesin KIF1A", "after": null, "start_char_pos": 431, "end_char_pos": 476 }, { "type": "R", "before": "by incorporating processes which correspond to shifting of the motor proteins from one protofilament to another. On the basis of analytical treatment of our", "after": ". By carrying out analytical calculations and computer simulations of this extended", "start_char_pos": 522, "end_char_pos": 678 }, { "type": "R", "before": "the effects", "after": "that the flux per lane can increase or decrease with the increasing rate", "start_char_pos": 697, "end_char_pos": 708 }, { "type": "R", "before": "on the flux of the KIF1A motors . Our quantitative", "after": ", depending on the concentrations of motors and the rate of hydrolysis of ATP, the ``fuel'' molecules. Our", "start_char_pos": 726, "end_char_pos": 776 } ]
[ 0, 92, 220, 405, 488, 634, 759 ]
0801.0718
1
In \mbox{%DIFAUXCMD Gua the notion of stickiness for stochastic processes was introduced. It was also shown that stickiness implies absense of arbitrage in a market with proportional transaction costs. In this paper, we investigate the notion of stickiness further. In particular, we show that stickiness is invariant under composition with continuous functions. We also prove a time change result on stickiness. As an application we provide sufficient conditions for continuous semimartingales to be sticky (A counter example show that not all semi-martingales are sticky). As a result, our paper provides an extended class of stochastic processes that are consistent with the no arbitrage property in a market with friction .
In 2 the notion of stickiness for stochastic processes was introduced. It was also shown that stickiness implies absense of arbitrage in a market with proportional transaction costs. In this paper, we investigate the notion of stickiness further. In particular, we give examples of processes that are not semimartingales but are sticky .
[ { "type": "R", "before": "\\mbox{%DIFAUXCMD Gua", "after": "2", "start_char_pos": 3, "end_char_pos": 23 }, { "type": "R", "before": "show that stickiness is invariant under composition with continuous functions. We also prove a time change result on stickiness. As an application we provide sufficient conditions for continuous semimartingales to be sticky (A counter example show that not all semi-martingales are sticky). As a result, our paper provides an extended class of stochastic", "after": "give examples of", "start_char_pos": 284, "end_char_pos": 638 }, { "type": "R", "before": "consistent with the no arbitrage property in a market with friction", "after": "not semimartingales but are sticky", "start_char_pos": 658, "end_char_pos": 725 } ]
[ 0, 89, 201, 265, 362, 412, 574 ]
0801.1480
1
In this paper, we propose a thermodynamic mechanism for the formation of transcriptional foci via the joint agglomeration of DNA-looping proteins and protein-binding domains on DNA: The competition between the gain in protein-DNA binding free energy and the entropy loss due to DNA looping is argued to result in an effective attraction between loops. A mean-field description can be solved analytically via a mapping to a restricted random-graph ensemble . It shows the emergence of protein clusters containing a finite fraction of all looping proteins. If the entropy loss due to a single DNA loop is high enough, this transition is found to be of first order.
In this paper, we propose a thermodynamic mechanism for the formation of transcriptional foci via the joint agglomeration of DNA-looping proteins and protein-binding domains on DNA: The competition between the gain in protein-DNA binding free energy and the entropy loss due to DNA looping is argued to result in an effective attraction between loops. A mean-field approximation can be described analytically via a mapping to a restricted random-graph ensemble having local degree constraints and global constraints on the number of connected components . It shows the emergence of protein clusters containing a finite fraction of all looping proteins. If the entropy loss due to a single DNA loop is high enough, this transition is found to be of first order.
[ { "type": "R", "before": "description can be solved", "after": "approximation can be described", "start_char_pos": 365, "end_char_pos": 390 }, { "type": "A", "before": null, "after": "having local degree constraints and global constraints on the number of connected components", "start_char_pos": 456, "end_char_pos": 456 } ]
[ 0, 351, 458, 555 ]
0801.4592
1
It is well-known that power control can affect the wireless network capacity. However, recent works show conflicting results: network capacity may increase or decrease with higher transmission power under different scenarios. In this work, we want to explore this paradoxand provide fundamental understanding on power control . Specifically, we want to explore the following questions: (1)Theoretically, should we increase or decrease transmission power to maximize network capacity? (2) Theoretically, how much network capacity gain can we achieve when using power control? (3) Under realistic situations, how do power control, link scheduling and routing interact with each other? Under which scenarios can we expect a large capacity gain by using higher transmission power? To answer these questions, firstly, we prove that the optimal network capacity is a non-decreasing function of transmission power. Secondly, we prove that the optimal network capacity can be increased unlimitedly by higher transmission power in some network configurations. However, when nodes are distributed uniformly, the gain of optimal network capacity by higher transmission power is upper-bounded by a positive constant. Thirdly, we discuss why network capacity in practice may increase or decrease with higher transmission power under different scenarios using carrier sensing and the minimum hop-count routing. Extensive simulations are carried out to verify our analysis . This work provides a deeper understanding on how power control can affect network capacity. Besides the theoretical contributions, it offers some design intuitions to wireless network researchers .
Recent works show conflicting results: network capacity may increase or decrease with higher transmission power under different scenarios. In this work, we want to understand this paradox . Specifically, we address the following questions: (1)Theoretically, should we increase or decrease transmission power to maximize network capacity? (2) Theoretically, how much network capacity gain can we achieve by power control? (3) Under realistic situations, how do power control, link scheduling and routing interact with each other? Under which scenarios can we expect a large capacity gain by using higher transmission power? To answer these questions, firstly, we prove that the optimal network capacity is a non-decreasing function of transmission power. Secondly, we prove that the optimal network capacity can be increased unlimitedly by higher transmission power in some network configurations. However, when nodes are distributed uniformly, the gain of optimal network capacity by higher transmission power is upper-bounded by a positive constant. Thirdly, we discuss why network capacity in practice may increase or decrease with higher transmission power under different scenarios using carrier sensing and the minimum hop-count routing. Extensive simulations are carried out to verify our analysis .
[ { "type": "R", "before": "It is well-known that power control can affect the wireless network capacity. However, recent", "after": "Recent", "start_char_pos": 0, "end_char_pos": 93 }, { "type": "R", "before": "explore this paradoxand provide fundamental understanding on power control", "after": "understand this paradox", "start_char_pos": 251, "end_char_pos": 325 }, { "type": "R", "before": "want to explore", "after": "address", "start_char_pos": 345, "end_char_pos": 360 }, { "type": "R", "before": "when using", "after": "by", "start_char_pos": 549, "end_char_pos": 559 }, { "type": "D", "before": ". This work provides a deeper understanding on how power control can affect network capacity. Besides the theoretical contributions, it offers some design intuitions to wireless network researchers", "after": null, "start_char_pos": 1458, "end_char_pos": 1655 } ]
[ 0, 77, 225, 327, 483, 574, 682, 776, 907, 1050, 1204, 1396, 1459, 1551 ]
0802.1288
1
In this work we introduce Heath-Jarrow-Morton (HJM) interest rate models driven by fractional Brownian motions. By using support arguments we prove that the resulting model is arbitrage-free under proportional transaction costs in the same spirit of Guasoni et al (2006 , 2007) . In particular, we obtain a drift condition which is similar in nature to the classical HJM no-arbitrage drift restriction. The second part of this paper deals with consistency problems related to the fractional HJM dynamics. We give a fairly complete characterization of finite-dimensional invariant manifolds for HJM models with fractional Brownian motion by means of Nagumo-type conditions. As an application, we investigate consistency of Nelson-Siegel family with respect to Ho-Lee and Hull-White models. It turns out that similar to the Brownian case such family does not go well with the fractional HJM dynamics with deterministic volatility. In fact, there is no nontrivial fractional interest rate model consistent with the Nelson-Siegel family.
In this work we introduce Heath-Jarrow-Morton (HJM) interest rate models driven by fractional Brownian motions. By using support arguments we prove that the resulting model is arbitrage free under proportional transaction costs in the same spirit of Guasoni Math. Finance 16 (2006 ) 569-582 . In particular, we obtain a drift condition which is similar in nature to the classical HJM no-arbitrage drift restriction. The second part of this paper deals with consistency problems related to the fractional HJM dynamics. We give a fairly complete characterization of finite-dimensional invariant manifolds for HJM models with fractional Brownian motion by means of Nagumo-type conditions. As an application, we investigate consistency of Nelson-Siegel family with respect to Ho-Lee and Hull-White models. It turns out that similar to the Brownian case such a family does not go well with the fractional HJM dynamics with deterministic volatility. In fact, there is no nontrivial fractional interest rate model consistent with the Nelson-Siegel family.
[ { "type": "R", "before": "arbitrage-free", "after": "arbitrage free", "start_char_pos": 176, "end_char_pos": 190 }, { "type": "R", "before": "et al", "after": "Math. Finance 16", "start_char_pos": 258, "end_char_pos": 263 }, { "type": "R", "before": ", 2007)", "after": ") 569-582", "start_char_pos": 270, "end_char_pos": 277 }, { "type": "A", "before": null, "after": "a", "start_char_pos": 841, "end_char_pos": 841 } ]
[ 0, 111, 279, 402, 504, 672, 788, 929 ]
0802.2004
1
We show that a simple model reproduces very closely the evolution of the GDP in constant dollars of many countries during the times of recession and recovery. A theoretical analysis illustrates how an optimal dynamical policy reduces both recession duration and severity, and increases the value of GDP at all times. We propose a criterion to distinguish a posteriori a dynamical policy from a static one .
We show that a simple model reproduces very closely the evolution of the gross domestic product (GDP) in current and constant dollars of many countries during the times of recession and recovery. A theoretical analysis illustrates how an optimal dynamical policy reduces both recession duration and severity, and increases the value of GDP at all times. We propose a criterion to distinguish a posteriori a dynamical policy from a static one . Finally, we predict that Ukraine will recover its current and constant dollar 1990 GDPs in 2009 and Moldova in 2015 in current dollars and 2009 in constant dollars .
[ { "type": "R", "before": "GDP in", "after": "gross domestic product (GDP) in current and", "start_char_pos": 73, "end_char_pos": 79 }, { "type": "A", "before": null, "after": ". Finally, we predict that Ukraine will recover its current and constant dollar 1990 GDPs in 2009 and Moldova in 2015 in current dollars and 2009 in constant dollars", "start_char_pos": 405, "end_char_pos": 405 } ]
[ 0, 158, 316 ]
0802.2004
2
We show that a simple model reproduces very closely the evolution of the gross domestic product (GDP) in current and constant dollars of many countries during the times of recession and recovery. A theoretical analysis illustrates how an optimal dynamical policy reduces both recession duration and severity, and increases the value of GDP at all times. We propose a criterion to distinguish a posteriori a dynamical policy from a static one. Finally, we predict that Ukraine will recover its current and constant dollar 1990 GDPs in 2009 and Moldova in 2015 in current dollars and 2009 in constant dollars .
We show that a simple and intuitive three-parameter equation fits remarkably well the evolution of the gross domestic product (GDP) in current and constant dollars of many countries during the times of recession and recovery. We then argue that it can be used to detect shocks and discuss its predictive power. Finally, a two-sector theoretical model of recession and recovery illustrates how the severity and length of recession depends on the dynamics of transfer rate between the growing and failing parts of the economy .
[ { "type": "R", "before": "model reproduces very closely", "after": "and intuitive three-parameter equation fits remarkably well", "start_char_pos": 22, "end_char_pos": 51 }, { "type": "R", "before": "A theoretical analysis illustrates how an optimal dynamical policy reduces both recession duration and severity, and increases the value of GDP at all times. We propose a criterion to distinguish a posteriori a dynamical policy from a static one. Finally, we predict that Ukraine will recover its current and constant dollar 1990 GDPs in 2009 and Moldova in 2015 in current dollars and 2009 in constant dollars", "after": "We then argue that it can be used to detect shocks and discuss its predictive power. Finally, a two-sector theoretical model of recession and recovery illustrates how the severity and length of recession depends on the dynamics of transfer rate between the growing and failing parts of the economy", "start_char_pos": 196, "end_char_pos": 606 } ]
[ 0, 195, 353, 442 ]
0802.2004
3
We show that a simple and intuitive three-parameter equation fits remarkably well the evolution of the gross domestic product (GDP) in current and constant dollars of many countries during the times of recession and recovery. We then argue that it can be used to detect shocks and discuss its predictive power. Finally, a two-sector theoretical model of recession and recovery illustrates how the severity and length of recession depends on the dynamics of transfer rate between the growing and failing parts of the economy.
We show that a simple and intuitive three-parameter equation fits remarkably well the evolution of the gross domestic product (GDP) in current and constant dollars of many countries during times of recession and recovery. We then argue that this equation is the response function of the economy to isolated shocks, hence that it can be used to detect large and small shocks, including those which do not lead to a recession; we also discuss its predictive power. Finally, a two-sector toy model of recession and recovery illustrates how the severity and length of recession depends on the dynamics of transfer rate between the growing and failing parts of the economy.
[ { "type": "D", "before": "the", "after": null, "start_char_pos": 189, "end_char_pos": 192 }, { "type": "A", "before": null, "after": "this equation is the response function of the economy to isolated shocks, hence that", "start_char_pos": 245, "end_char_pos": 245 }, { "type": "R", "before": "shocks and", "after": "large and small shocks, including those which do not lead to a recession; we also", "start_char_pos": 271, "end_char_pos": 281 }, { "type": "R", "before": "theoretical", "after": "toy", "start_char_pos": 334, "end_char_pos": 345 } ]
[ 0, 225, 311 ]
0802.2172
1
In that paper, we solve dynamically a partial hedging problem for an American contingent claim: assuming superhedging is not feasible, we explain in this context the notion of efficient hedging by introducing a risk minimization criterion: we consider here the problem of minimizing the conditional expected loss for a given convex and non decreasing loss function. To solve this problem, we provide a connection between the dynamic convex risk functional introduced and the solutionof a quadratic RBSDE (Reflected Backward Stochastic Differential Equations): this is achieved by studying the properties of specific non linear expectations .
In that paper, we provide a new characterization of the solutions of specific reflected backward stochastic differential equations (or RBSDEs) whose driver g is convex and has quadratic growth in its second variable: this is done by introducing the extended notion of g-Snell enveloppe. Then, in a second step, we relate this representation to a specific class of dynamic monetary concave functionals already introduced in a discrete time setting. This connection implies that the solution, characterized by means of non linear expectations , has again the time consistency property .
[ { "type": "D", "before": "solve dynamically a partial hedging problem for an American contingent claim: assuming superhedging is not feasible, we explain in this context the notion of efficient hedging by introducing a risk minimization criterion: we consider here the problem of minimizing the conditional expected loss for a given convex and non decreasing loss function. To solve this problem, we", "after": null, "start_char_pos": 18, "end_char_pos": 391 }, { "type": "R", "before": "connection between the dynamic convex risk functional introduced and the solutionof a quadratic RBSDE (Reflected Backward Stochastic Differential Equations): this is achieved by studying the properties of specific", "after": "new characterization of the solutions of specific reflected backward stochastic differential equations (or RBSDEs) whose driver g is convex and has quadratic growth in its second variable: this is done by introducing the extended notion of g-Snell enveloppe. Then, in a second step, we relate this representation to a specific class of dynamic monetary concave functionals already introduced in a discrete time setting. This connection implies that the solution, characterized by means of", "start_char_pos": 402, "end_char_pos": 615 }, { "type": "A", "before": null, "after": ", has again the time consistency property", "start_char_pos": 640, "end_char_pos": 640 } ]
[ 0, 365 ]
0802.2512
1
Many networks exhibit the small-world property of the neighborhood connectivity being higher than in comparable random networks. However, the standard measure of local neighborhood clustering is typically not defined if a node has one or no neighbor. In such cases, local clustering has traditionally been set to zero and the node was included in the global clustering coefficient. Such a procedure leads to under-estimation of the neighborhood clustering in sparse networks. We propose to include \theta as the proportion of nodes with one or no neighbors to estimate the contribution of these cases to the mean clustering value. Furthermore, we provide a formula for estimating a clustering coefficient that excludes these undefined cases. We find that the definition of the clustering coefficient has a major effect when comparing different networks. For metabolic networks of URLanisms, relations changed for 58\% of the comparisons when a different definition was applied .
Many networks exhibit the small-world property of the neighborhood connectivity being higher than in comparable random networks. However, the standard measure of local neighborhood clustering is typically not defined if a node has one or no neighbor. In such cases, local clustering has traditionally been set to zero and the node was included in the global clustering coefficient. Such a procedure leads to under-estimation of the neighborhood clustering in sparse networks. We propose to include \theta as the proportion of leafs and isolated nodes to estimate the contribution of these cases to the mean clustering value. Furthermore, we provide a formula for estimating a clustering coefficient that excludes these undefined cases. This novel definition leads to values which are up to 140\% higher than the traditional values for the observed networks indicating that neighborhood connectivity is normally underestimated. We find that the definition of the clustering coefficient has a major effect when comparing different networks. For metabolic networks of URLanisms, relations changed for 58\% of the comparisons when a different definition was applied . Values for the clustering coefficient vary with the ratio of isolated and leaf nodes which is critical for comparing networks. We suggest including ratio and definition in publications to enable comparisons across different studies .
[ { "type": "R", "before": "nodes with one or no neighbors", "after": "leafs and isolated nodes", "start_char_pos": 526, "end_char_pos": 556 }, { "type": "A", "before": null, "after": "This novel definition leads to values which are up to 140\\% higher than the traditional values for the observed networks indicating that neighborhood connectivity is normally underestimated.", "start_char_pos": 742, "end_char_pos": 742 }, { "type": "A", "before": null, "after": ". Values for the clustering coefficient vary with the ratio of isolated and leaf nodes which is critical for comparing networks. We suggest including ratio and definition in publications to enable comparisons across different studies", "start_char_pos": 978, "end_char_pos": 978 } ]
[ 0, 128, 250, 381, 475, 630, 741, 854 ]
0802.2512
2
Many networks exhibit the small-world property of the neighborhood connectivity being higher than in comparable random networks. However, the standard measure of local neighborhood clustering is typically not defined if a node has one or no neighbor . In such cases, local clustering has traditionally been set to zero and the node was included in the global clustering coefficient. Such a procedure leads to under-estimation of the neighborhood clustering in sparse networks. We propose to include \theta as the proportion of leafs and isolated nodes to estimate the contribution of these cases to the mean clustering value. Furthermore, we provide a formula for estimating a clustering coefficient that excludes these undefined cases . This novel definition leads to values which are up to 140\% higher than the traditional values for the observed networks indicating that neighborhood connectivity is normally underestimated. We find that the definition of the clustering coefficient has a major effect when comparing different networks. For metabolic networks of URLanisms, relations changed for 58\% of the comparisons when a different definition was applied. Values for the clustering coefficient vary with the ratio of isolated and leaf nodes which is critical for comparing networks. We suggest including ratio and definition in publications to enable comparisons across different studies .
Many networks exhibit the small-world property of the neighborhood connectivity being higher than in comparable random networks. However, the standard measure of local neighborhood clustering is typically not defined if a node has one or no neighbors . In such cases, local clustering has traditionally been set to zero and this value influenced the global clustering coefficient. Such a procedure leads to underestimation of the neighborhood clustering in sparse networks. We propose to include \theta as the proportion of leafs and isolated nodes to estimate the contribution of these cases and provide a formula for estimating a clustering coefficient excluding these cases from the Watts and Strogatz (1998 Nature 393 440-2) definition of the clustering coefficient. Excluding leafs and isolated nodes leads to values which are up to 140\% higher than the traditional values for the observed networks indicating that neighborhood connectivity is normally underestimated. We find that the definition of the clustering coefficient has a major effect when comparing different networks. For metabolic networks of URLanisms, relations changed for 58\% of the comparisons when a different definition was applied. We also show that the definition influences small-world features and that the classification can change from non-small-world to small-world network. We discuss the use of an alternative measure, disconnectedness D, which is less influenced by leafs and isolated nodes .
[ { "type": "R", "before": "neighbor", "after": "neighbors", "start_char_pos": 241, "end_char_pos": 249 }, { "type": "R", "before": "the node was included in the", "after": "this value influenced the", "start_char_pos": 323, "end_char_pos": 351 }, { "type": "R", "before": "under-estimation", "after": "underestimation", "start_char_pos": 409, "end_char_pos": 425 }, { "type": "R", "before": "to the mean clustering value. Furthermore, we", "after": "and", "start_char_pos": 596, "end_char_pos": 641 }, { "type": "R", "before": "that excludes these undefined cases . This novel definition", "after": "excluding these cases from the Watts and Strogatz (1998 Nature 393 440-2) definition of the clustering coefficient. Excluding leafs and isolated nodes", "start_char_pos": 700, "end_char_pos": 759 }, { "type": "R", "before": "Values for the clustering coefficient vary with the ratio of isolated and leaf nodes which is critical for comparing networks. We suggest including ratio and definition in publications to enable comparisons across different studies", "after": "We also show that the definition influences small-world features and that the classification can change from non-small-world to small-world network. We discuss the use of an alternative measure, disconnectedness D, which is less influenced by leafs and isolated nodes", "start_char_pos": 1165, "end_char_pos": 1396 } ]
[ 0, 128, 251, 382, 476, 625, 737, 928, 1040, 1164, 1291 ]
0802.3679
1
In Black-Scholes delta-hedging method generalization, a "mirror-diffusion" inverse stochastic process is introduced with condition determined by the underlying price variance and payoff function. The process reduces an expected option value at maturity under equivalent martingale measure back to the current time. The normalized ksi-returns , correspondent to the kernel function in the found general solution and not dependent explicitly on time, were used for verification of the one-parameter model inherent efficiency, i.e. self-calibration using only historical volatility data. The model minimizes implied volatility bias (for 2004-2007 S& P100 index options) and theoretically yields skews correspondent to practical term structure for interest rate derivatives. It allows increasing the number of stock price distribution parameters.
The proposed model modifies option pricing formulas for the basic case of log-normal probability distribution providing correspondence to formulated criteria of efficiency and completeness. The model is self-calibrating by historic volatility data; it maintains the constant expected value at maturity of the hedged instantaneously self-financing portfolio. The payoff variance dependent on random stock price at maturity obtained under an equivalent martingale measure is taken as a condition for introduced "mirror-time" derivative diffusion discount process. Introduced ksi-return distribution , correspondent to the found general solution of backward drift-diffusion equation and normalized by theoretical diffusion coefficient, does not contain so-called "long tails" and unbiased for considered 2004-2007 S& P 100 index data. The model theoretically yields skews correspondent to practical term structure for interest rate derivatives. The method allows increasing the number of asset price probability distribution parameters.
[ { "type": "R", "before": "In Black-Scholes delta-hedging method generalization, a \"mirror-diffusion\" inverse stochastic process is introduced with condition determined by the underlying price variance and payoff function. The process reduces an expected option", "after": "The proposed model modifies option pricing formulas for the basic case of log-normal probability distribution providing correspondence to formulated criteria of efficiency and completeness. The model is self-calibrating by historic volatility data; it maintains the constant expected", "start_char_pos": 0, "end_char_pos": 234 }, { "type": "R", "before": "under", "after": "of the hedged instantaneously self-financing portfolio. The payoff variance dependent on random stock price at maturity obtained under an", "start_char_pos": 253, "end_char_pos": 258 }, { "type": "R", "before": "back to the current time. The normalized ksi-returns", "after": "is taken as a condition for introduced \"mirror-time\" derivative diffusion discount process. Introduced ksi-return distribution", "start_char_pos": 289, "end_char_pos": 341 }, { "type": "D", "before": "kernel function in the", "after": null, "start_char_pos": 365, "end_char_pos": 387 }, { "type": "R", "before": "and not dependent explicitly on time, were used for verification of the one-parameter model inherent efficiency, i.e. self-calibration using only historical volatility data. The model minimizes implied volatility bias (for", "after": "of backward drift-diffusion equation and normalized by theoretical diffusion coefficient, does not contain so-called \"long tails\" and unbiased for considered", "start_char_pos": 411, "end_char_pos": 633 }, { "type": "R", "before": "P100 index options) and", "after": "P 100 index data. The model", "start_char_pos": 647, "end_char_pos": 670 }, { "type": "R", "before": "It", "after": "The method", "start_char_pos": 771, "end_char_pos": 773 }, { "type": "R", "before": "stock price", "after": "asset price probability", "start_char_pos": 806, "end_char_pos": 817 } ]
[ 0, 195, 314, 584, 770 ]
0802.4411
1
We present a positivity preserving numerical scheme for the pathwise solution of nonlinear stochastic differential equations driven by a multi-dimensional Wiener process and governed by non-commutative linear and non-Lipschitz vector fields. This strong order one scheme uses: (i) Strang exponential splitting, an approximation that decomposes the stochastic flow separately into the drift flow, and the pure diffusion flow governed by the diffusion vector fields; (ii) an implicit Euler method to approximate the drift flow; and (iii) an implicit Milstein method to approximate the pure diffusion flow. The separate approximations for the drift and pure diffusion flows preserve positivity. Therefore the Strang exponential splitting approximation does also. We demonstrate the efficacy of our method by applying it to the Heston model and a variance curve model , and compare it against well-established positivity preserving schemes .
For nonlinear stochastic differential systems, we develop strong fully implicit positivity preserving numerical methods in the case that the zero boundary is non-attracting. These methods are implicit in the diffusion vector fields. They thus apply to a restricted class, namely those with sublinear form. This however, still includes most Langevin derived processes typical of volatility models in finance and molecular simulation in physics. When the zero boundary is attracting and attainable, we specialize to a prototypical model, namely the mean-reverting Cox--Ingersoll--Ross process. We thus consider the non-central chi-squared transition density with fractional degrees of freedom. We prove that we can sample from this density by simulating Poisson distributed sums of powers of generalized Gaussian random variables. Further we prove that Marsaglia's polar method extends to the generalized Gaussian distribution, providing an exact and efficient method for generalized Gaussian sampling. We apply our methods to a variance curve model and the Heston model .
[ { "type": "R", "before": "We present a positivity preserving numerical scheme for the pathwise solution of", "after": "For", "start_char_pos": 0, "end_char_pos": 80 }, { "type": "R", "before": "equations driven by a multi-dimensional Wiener process and governed by non-commutative linear and non-Lipschitz vector fields. This strong order one scheme uses: (i) Strang exponential splitting, an approximation that decomposes the stochastic flow separately into the drift flow, and the pure diffusion flow governed by the diffusion vector fields; (ii) an implicit Euler method to approximate the drift flow; and (iii) an implicit Milstein method to approximate the pure diffusion flow. The separate approximations for the drift and pure diffusion flows preserve positivity. Therefore the Strang exponential splitting approximation does also. We demonstrate the efficacy of our method by applying it to the Heston model and", "after": "systems, we develop strong fully implicit positivity preserving numerical methods in the case that the zero boundary is non-attracting. These methods are implicit in the diffusion vector fields. They thus apply to a restricted class, namely those with sublinear form. This however, still includes most Langevin derived processes typical of volatility models in finance and molecular simulation in physics. When the zero boundary is attracting and attainable, we specialize to a prototypical model, namely the mean-reverting Cox--Ingersoll--Ross process. We thus consider the non-central chi-squared transition density with fractional degrees of freedom. We prove that we can sample from this density by simulating Poisson distributed sums of powers of generalized Gaussian random variables. Further we prove that Marsaglia's polar method extends to the generalized Gaussian distribution, providing an exact and efficient method for generalized Gaussian sampling. We apply our methods to", "start_char_pos": 115, "end_char_pos": 840 }, { "type": "R", "before": ", and compare it against well-established positivity preserving schemes", "after": "and the Heston model", "start_char_pos": 864, "end_char_pos": 935 } ]
[ 0, 241, 464, 525, 603, 691, 759 ]
0802.4411
2
For nonlinear stochastic differential systems, we develop strong fully implicit positivity preserving numerical methods in the case that the zero boundary is non-attracting. These methods are implicit in the diffusion vector fields. They thus apply to a restricted class, namely those with sublinear form. This however, still includes most Langevin derived processes typical of volatility models in finance and molecular simulation in physics. When the zero boundary is attracting and attainable, we specialize to a prototypical model, namely the mean-reverting Cox--Ingersoll--Ross process. We thus consider the non-central chi-squared transition density with fractional degrees of freedom. We prove that we can sample from this density by simulating Poisson distributed sums of powers of generalized Gaussian random variables. Further we prove that Marsaglia's polar method extends to the generalized Gaussian distribution, providing an exact and efficient method for generalized Gaussian sampling. We apply our methods to a variance curve model and the Heston model .
In the Heston stochastic volatility model, the transition probability of the variance process can be represented by a non-central chi-square density. We focus on the case when the number of degrees of freedom is small and the zero boundary is attracting and attainable, typical in foreign exchange markets. We prove a new representation for this density based on sums of powers of generalized Gaussian random variables. Further we prove Marsaglia's polar method extends to this distribution, providing an exact method for generalized Gaussian sampling. The advantages are that for the mean-reverting square-root process in the Heston model and Cox-Ingersoll-Ross model, we can generate samples from the true transition density simply, efficiently and robustly .
[ { "type": "R", "before": "For nonlinear stochastic differential systems, we develop strong fully implicit positivity preserving numerical methods in the case that the zero boundary is non-attracting. These methods are implicit in the diffusion vector fields. They thus apply to a restricted class, namely those with sublinear form. This however, still includes most Langevin derived processes typical of volatility models in finance and molecular simulation in physics. When", "after": "In the Heston stochastic volatility model,", "start_char_pos": 0, "end_char_pos": 448 }, { "type": "A", "before": null, "after": "transition probability of the variance process can be represented by a non-central chi-square density. We focus on the case when the number of degrees of freedom is small and the", "start_char_pos": 453, "end_char_pos": 453 }, { "type": "R", "before": "we specialize to a prototypical model, namely the mean-reverting Cox--Ingersoll--Ross process. We thus consider the non-central chi-squared transition density with fractional degrees of freedom. We prove that we can sample from this density by simulating Poisson distributed", "after": "typical in foreign exchange markets. We prove a new representation for this density based on", "start_char_pos": 498, "end_char_pos": 772 }, { "type": "D", "before": "that", "after": null, "start_char_pos": 847, "end_char_pos": 851 }, { "type": "R", "before": "the generalized Gaussian", "after": "this", "start_char_pos": 888, "end_char_pos": 912 }, { "type": "D", "before": "and efficient", "after": null, "start_char_pos": 946, "end_char_pos": 959 }, { "type": "R", "before": "We apply our methods to a variance curve model and the Heston model", "after": "The advantages are that for the mean-reverting square-root process in the Heston model and Cox-Ingersoll-Ross model, we can generate samples from the true transition density simply, efficiently and robustly", "start_char_pos": 1002, "end_char_pos": 1069 } ]
[ 0, 173, 232, 305, 443, 592, 692, 829, 1001 ]
0802.4411
3
In the Heston stochastic volatility model, the transition probability of the variance process can be represented by a non-central chi-square density. We focus on the case when the number of degrees of freedom is small and the zero boundary is attracting and attainable, typical in foreign exchange markets. We prove a new representation for this density based on sums of powers of generalized Gaussian random variables. Further we prove Marsaglia's polar method extends to this distribution, providing an exact method for generalized Gaussian sampling . The advantages are that for the mean-reverting square-root process in the Heston model and Cox-Ingersoll-Ross model, we can generate samples from the true transition density simply, efficiently and robustly .
The transition probability of a Cox-Ingersoll-Ross process can be represented by a non-central chi-square density. First we prove a new representation for the central chi-square density based on sums of powers of generalized Gaussian random variables. Second we prove Marsaglia's polar method extends to this distribution, providing a simple, exact, robust and efficient acceptance-rejection method for generalized Gaussian sampling and thus central chi-square sampling. Third we derive a simple, high-accuracy, robust and efficient direct inversion method for generalized Gaussian sampling based on the Beasley-Springer-Moro method. Indeed the accuracy of the approximation to the inverse cumulative distribution function is to the tenth decimal place. We then apply our methods to non-central chi-square variance sampling in the Heston model . We focus on the case when the number of degrees of freedom is small and the zero boundary is attracting and attainable, typical in foreign exchange markets. Using the additivity property of the chi-square distribution, our methods apply in all parameter regimes .
[ { "type": "R", "before": "In the Heston stochastic volatility model, the", "after": "The", "start_char_pos": 0, "end_char_pos": 46 }, { "type": "R", "before": "the variance", "after": "a Cox-Ingersoll-Ross", "start_char_pos": 73, "end_char_pos": 85 }, { "type": "R", "before": "We focus on the case when the number of degrees of freedom is small and the zero boundary is attracting and attainable, typical in foreign exchange markets. We", "after": "First we", "start_char_pos": 150, "end_char_pos": 309 }, { "type": "R", "before": "this", "after": "the central chi-square", "start_char_pos": 341, "end_char_pos": 345 }, { "type": "R", "before": "Further", "after": "Second", "start_char_pos": 420, "end_char_pos": 427 }, { "type": "R", "before": "an exact", "after": "a simple, exact, robust and efficient acceptance-rejection", "start_char_pos": 502, "end_char_pos": 510 }, { "type": "R", "before": ". The advantages are that for the mean-reverting square-root process", "after": "and thus central chi-square sampling. Third we derive a simple, high-accuracy, robust and efficient direct inversion method for generalized Gaussian sampling based on the Beasley-Springer-Moro method. Indeed the accuracy of the approximation to the inverse cumulative distribution function is to the tenth decimal place. We then apply our methods to non-central chi-square variance sampling", "start_char_pos": 552, "end_char_pos": 620 }, { "type": "R", "before": "and Cox-Ingersoll-Ross model, we can generate samples from the true transition density simply, efficiently and robustly", "after": ". We focus on the case when the number of degrees of freedom is small and the zero boundary is attracting and attainable, typical in foreign exchange markets. Using the additivity property of the chi-square distribution, our methods apply in all parameter regimes", "start_char_pos": 641, "end_char_pos": 760 } ]
[ 0, 149, 306, 419, 553 ]
0803.0483
1
Translocation of a single stranded DNA through genetically engineered \alpha-hemolysin channels with positively charged walls is studied. It is predicted that transport properties of such channels are dramatically different from the neutral wild type \alpha-hemolysin channel. We assume that the wall charges compensate the fraction x of the bare charge q_{b} of the DNA piece residing in the channel. Our prediction are as follows (i) At small concentration of salt the blocked ion current decreases with x. (ii) The effective charge q of DNA piece grows with x and at q_{b x= 1. (iii) The rate of DNA capture by the channel exponentially grows with x. Our theory is also applicable to translocation of a double stranded DNA in narrow solid state nanopores with positively charged walls.
Translocation of a single stranded DNA through genetically engineered \alpha-hemolysin channels with positively charged walls is studied. It is predicted that transport properties of such channels are dramatically different from neutral wild type \alpha-hemolysin channel. We assume that the wall charges compensate the fraction x of the bare charge q_{b} of the DNA piece residing in the channel. Our prediction are as follows (i) At small concentration of salt the blocked ion current decreases with x. (ii) The effective charge q of DNA piece , which is very small at x = 0 (neutral channel) grows with x and at x= 1 reaches q_{b (iii) The rate of DNA capture by the channel exponentially grows with x. Our theory is also applicable to translocation of a double stranded DNA in narrow solid state nanopores with positively charged walls.
[ { "type": "D", "before": "the", "after": null, "start_char_pos": 229, "end_char_pos": 232 }, { "type": "A", "before": null, "after": ", which is very small at x = 0 (neutral channel)", "start_char_pos": 550, "end_char_pos": 550 }, { "type": "D", "before": "q_{b", "after": null, "start_char_pos": 571, "end_char_pos": 575 }, { "type": "R", "before": "1.", "after": "1 reaches q_{b", "start_char_pos": 579, "end_char_pos": 581 } ]
[ 0, 137, 276, 401 ]
0803.1364
1
In the Kelly game (Kelly, 1956 ), a gambler is allowed to invest a part of the wealth in each turn. With a certain probability this investment is doubled, and otherwise it is lost. Motivated by the complexity of real investments, we propose several modifications of this game to investigate the influence of diversification and limited information on investment performance. Analytical and numerical results obtained from these toy games are well related to their real-life counterparts .
Financial markets, with their vast range of different investment opportunities, can be seen as a system of many different simultaneous games with diverse and often unknown levels of risk and reward. We introduce generalizations to the classic Kelly investment game Kelly ( 1956 ) that incorporates these features, and use them to investigate the influence of diversification and limited information on Kelly-optimal portfolios. In particular we present approximate formulas for optimizing diversified portfolios and exact results for optimal investment in unknown games where the only available information is past outcomes .
[ { "type": "R", "before": "In the Kelly game (Kelly,", "after": "Financial markets, with their vast range of different investment opportunities, can be seen as a system of many different simultaneous games with diverse and often unknown levels of risk and reward. We introduce generalizations to the classic Kelly investment game", "start_char_pos": 0, "end_char_pos": 25 }, { "type": "A", "before": null, "after": "Kelly (", "start_char_pos": 26, "end_char_pos": 26 }, { "type": "R", "before": "), a gambler is allowed to invest a part of the wealth in each turn. With a certain probability this investment is doubled, and otherwise it is lost. Motivated by the complexity of real investments, we propose several modifications of this game", "after": ")", "start_char_pos": 32, "end_char_pos": 276 }, { "type": "A", "before": null, "after": "that incorporates these features, and use them", "start_char_pos": 277, "end_char_pos": 277 }, { "type": "R", "before": "investment performance. Analytical and numerical results obtained from these toy games are well related to their real-life counterparts", "after": "Kelly-optimal portfolios. In particular we present approximate formulas for optimizing diversified portfolios and exact results for optimal investment in unknown games where the only available information is past outcomes", "start_char_pos": 353, "end_char_pos": 488 } ]
[ 0, 100, 181, 376 ]
0803.1558
1
We introduce some new quantitative measures of fluctuations in the process of synthesis of proteins from a single messenger RNA (mRNA) template. We calculate the statistical distributions of these fluctuating quantities and extract the strength of the corresponding translational noise. For these calculations{\bf we use a model that captures both the mechano-chemistry of each individual ribosome as well as their steric interactions in ribosome traffic on the same mRNA track. By comparing our results for a specific gene of the%DIFDELCMD < {\it %%% Escherichia coli bacteria with those for the corresponding homogeneous mRNA template, we demonstrate the effects of the sequence inhomogeneities of real genes on the fluctuations and noise . We also suggest {\it in-vitro} laboratory experimentsfor testing our theoretical predictions .
Proteins are polymerized by cyclic machines called ribosome which use their messenger RNA (mRNA) track also as the corresponding template and the process is called translation. We explore, in depth and detail, the stochastic nature of the translation. We compute various distributions associated with the translation process; one of them, namely dwell time distribution, has been measured in recent single ribosome experiments (Wen et al. Nature{\bf 452 , 598 (2008)). The form of this distribution predicted by our theory is consistent with that extracted from the experimental data. For our quantitative calculations, we use a model that captures both the mechano-chemistry of each individual ribosome as well as their steric interactions %DIFDELCMD < {\it %%% . We also demonstrate the effects of the sequence inhomogeneities of real genes on the fluctuations and noise in translation. In principle, our new predictions can be tested by carrying out {\it in-vitro} experiments .
[ { "type": "R", "before": "We introduce some new quantitative measures of fluctuations in the process of synthesis of proteins from a single", "after": "Proteins are polymerized by cyclic machines called ribosome which use their", "start_char_pos": 0, "end_char_pos": 113 }, { "type": "R", "before": "template. We calculate the statistical distributions of these fluctuating quantities and extract the strength of", "after": "track also as the corresponding template and the process is called translation. We explore, in depth and detail, the stochastic nature of", "start_char_pos": 135, "end_char_pos": 247 }, { "type": "R", "before": "corresponding translational noise. For these calculations", "after": "translation. We compute various distributions associated with the translation process; one of them, namely dwell time distribution, has been measured in recent single ribosome experiments (Wen et al. Nature", "start_char_pos": 252, "end_char_pos": 309 }, { "type": "A", "before": null, "after": "452", "start_char_pos": 314, "end_char_pos": 314 }, { "type": "A", "before": null, "after": ", 598 (2008)). The form of this distribution predicted by our theory is consistent with that extracted from the experimental data. For our quantitative calculations,", "start_char_pos": 315, "end_char_pos": 315 }, { "type": "D", "before": "in ribosome traffic on the same mRNA track. By comparing our results for a specific gene of the", "after": null, "start_char_pos": 437, "end_char_pos": 532 }, { "type": "D", "before": "Escherichia coli", "after": null, "start_char_pos": 554, "end_char_pos": 570 }, { "type": "R", "before": "bacteria with those for the corresponding homogeneous mRNA template, we", "after": ". We also", "start_char_pos": 571, "end_char_pos": 642 }, { "type": "R", "before": ". We also suggest", "after": "in translation. In principle, our new predictions can be tested by carrying out", "start_char_pos": 743, "end_char_pos": 760 }, { "type": "R", "before": "laboratory experimentsfor testing our theoretical predictions", "after": "experiments", "start_char_pos": 776, "end_char_pos": 837 } ]
[ 0, 144, 286, 480, 744 ]
0803.1890
1
A financial market model where agents can only trade using realistic buy-and-hold strategies is considered. Minimal assumptions are made on the nature of the asset-price process - in particular, the semimartingale property is not assumed. Via a natural assumption of limited opportunities for unlimited resulting wealth from trading, coined the No-Unbounded-Profit-with-Bounded-Risk (NUPBR) condition, we establish that asset-prices have be semimartingales, as well as a weakened version of the Fundamental Theorem of Asset Pricing that involves supermartingale deflators rather than equivalent martingale measures. Further, the utility maximization problem is considered and it is shown that using only buy-and-hold strategies, optimal utilities and wealth processes resulting from continuous trading can be approximated arbitrarily well .
A financial market model where agents trade using realistic combinations of buy-and-hold strategies is considered. Minimal assumptions are made on the asset-price process - in particular, the semimartingale property is not assumed. Via a natural assumption of limited opportunities for unlimited resulting wealth from trading, coined the "No Unbounded Profit with Bounded Risk" condition, we establish that asset-prices have to be semimartingales. In a slightly more specialized case, we extend the previous result in a weakened version of the Fundamental Theorem of Asset Pricing that involves supermartingale deflators rather than Equivalent Martingale Measures .
[ { "type": "D", "before": "can only", "after": null, "start_char_pos": 38, "end_char_pos": 46 }, { "type": "A", "before": null, "after": "combinations of", "start_char_pos": 69, "end_char_pos": 69 }, { "type": "D", "before": "nature of the", "after": null, "start_char_pos": 145, "end_char_pos": 158 }, { "type": "R", "before": "No-Unbounded-Profit-with-Bounded-Risk (NUPBR)", "after": "\"No Unbounded Profit with Bounded Risk\"", "start_char_pos": 346, "end_char_pos": 391 }, { "type": "R", "before": "be semimartingales, as well as", "after": "to be semimartingales. In a slightly more specialized case, we extend the previous result in", "start_char_pos": 439, "end_char_pos": 469 }, { "type": "R", "before": "equivalent martingale measures. Further, the utility maximization problem is considered and it is shown that using only buy-and-hold strategies, optimal utilities and wealth processes resulting from continuous trading can be approximated arbitrarily well", "after": "Equivalent Martingale Measures", "start_char_pos": 585, "end_char_pos": 839 } ]
[ 0, 108, 239, 616 ]
0803.1890
2
A financial market model where agents trade using realistic combinations of buy-and-hold strategies is considered. Minimal assumptions are made on the asset-price process - in particular, the semimartingale property is not assumed. Via a natural assumption of limited opportunities for unlimited resulting wealth from trading, coined the "No Unbounded Profit with Bounded Risk" condition , we establish that asset-prices have to be semimartingales. In a slightly more specialized case, we extend the previous result in a weakened version of the Fundamental Theorem of Asset Pricing that involves supermartingale deflators rather than Equivalent Martingale Measures.
A financial market model where agents trade using realistic combinations of buy-and-hold strategies is considered. Minimal assumptions are made on the discounted asset-price process - in particular, the semimartingale property is not assumed. Via a natural market viability assumption , namely, absence of arbitrages of the first kind, we establish that discounted asset-prices have to be semimartingales. In a slightly more specialized case, we extend the previous result in a weakened version of the Fundamental Theorem of Asset Pricing that involves strictly positive supermartingale deflators rather than Equivalent Martingale Measures.
[ { "type": "A", "before": null, "after": "discounted", "start_char_pos": 151, "end_char_pos": 151 }, { "type": "R", "before": "assumption of limited opportunities for unlimited resulting wealth from trading, coined the \"No Unbounded Profit with Bounded Risk\" condition", "after": "market viability assumption", "start_char_pos": 247, "end_char_pos": 388 }, { "type": "A", "before": null, "after": "namely, absence of arbitrages of the first kind,", "start_char_pos": 391, "end_char_pos": 391 }, { "type": "A", "before": null, "after": "discounted", "start_char_pos": 410, "end_char_pos": 410 }, { "type": "A", "before": null, "after": "strictly positive", "start_char_pos": 599, "end_char_pos": 599 } ]
[ 0, 114, 232, 451 ]
0804.0682
1
We address the problem of proteasomal protein translocation . Proteasomes are important for all aspects of the cellular metabolism but the mechanism of protein transport remains unknown. We introduce a new stochastic model of the proteasomal transport. We account for the protein translocation and the topology of the positioning of cleavage centers of a proteasome from first principles. We show by test examples and by a comparison with experimental data that our model allows reconstruction of the translocation rates from mass spectroscopy data on digestion patterns and can be used to investigate the properties of transport in different experimental set-ups. Finally we design an experimental set-up for a synthetic polypeptide with a periodic sequence of amino acids which enables more reliable determination of translocation rates .
We address the problem of proteasomal protein translocation and introduce a new stochastic model of the proteasomal digestion (cleavage) of proteins. In this model we account for the protein translocation and the positioning of cleavage sites of a proteasome from first principles. We show by test examples and by processing experimental data that our model allows reconstruction of the translocation and cleavage rates from mass spectroscopy data on digestion patterns and can be used to investigate the properties of transport in different experimental set-ups. Detailed investigation with this model will enable theoretical quantitative prediction of the proteasomal activity .
[ { "type": "R", "before": ". Proteasomes are important for all aspects of the cellular metabolism but the mechanism of protein transport remains unknown. We", "after": "and", "start_char_pos": 60, "end_char_pos": 189 }, { "type": "R", "before": "transport. We", "after": "digestion (cleavage) of proteins. In this model we", "start_char_pos": 242, "end_char_pos": 255 }, { "type": "D", "before": "topology of the", "after": null, "start_char_pos": 302, "end_char_pos": 317 }, { "type": "R", "before": "centers", "after": "sites", "start_char_pos": 342, "end_char_pos": 349 }, { "type": "R", "before": "a comparison with", "after": "processing", "start_char_pos": 421, "end_char_pos": 438 }, { "type": "A", "before": null, "after": "and cleavage", "start_char_pos": 515, "end_char_pos": 515 }, { "type": "R", "before": "Finally we design an experimental set-up for a synthetic polypeptide with a periodic sequence of amino acids which enables more reliable determination of translocation rates", "after": "Detailed investigation with this model will enable theoretical quantitative prediction of the proteasomal activity", "start_char_pos": 666, "end_char_pos": 839 } ]
[ 0, 186, 252, 388, 665 ]
0804.1034
1
DNA inversion is an important mechanism by which bacteria and bacteriophage switch reversibly between alternative phenotypic states. In such switches, the orientation of a short DNA element is flipped by a site-specific recombinase enzyme. We propose a simple model for a DNA inversion switch in which recombinase production is dependent on the switch state (orientational control). Our model is inspired by the fim switch in Escherichia coli. We present an exact analytical solution of the chemical master equation for the model switch, as well as stochastic simulations. Orientational control causes the switch to deviate from Poissonian behaviour: the distribution of times in the on state shows a peak and successive flip times are correlated . This model should provide a framework for understanding how DNA inversion switch architecture determines switch flipping statistics, with relevance to the role of such switches in microbial population dynamics .
DNA inversion is an important mechanism by which bacteria and bacteriophage switch reversibly between phenotypic states. In such switches, the orientation of a short DNA element is flipped by a site-specific recombinase enzyme. We propose a simple model for a DNA inversion switch in which recombinase production is dependent on the switch state (orientational control). Our model is inspired by the fim switch in Escherichia coli. We present an exact analytical solution of the chemical master equation for the model switch, as well as stochastic simulations. Orientational control causes the switch to deviate from Poissonian behaviour: the distribution of times in the on state shows a peak and successive flip times are correlated .
[ { "type": "D", "before": "alternative", "after": null, "start_char_pos": 102, "end_char_pos": 113 }, { "type": "D", "before": ". This model should provide a framework for understanding how DNA inversion switch architecture determines switch flipping statistics, with relevance to the role of such switches in microbial population dynamics", "after": null, "start_char_pos": 747, "end_char_pos": 958 } ]
[ 0, 132, 239, 382, 443, 572, 748 ]
0804.1227
1
{\it Transcription} is the process whereby RNA molecules are polymerized by molecular machines, called RNA polymerase (RNAP), using the corresponding DNA as the template. Recent {\it in-vivo} experiments with single cells have established that transcription takes place in ``bursts'' or ``pulses'' . In this letter we present a model that captures not only the mechano-chemistry of individual RNAPs and their steric interactions but also the switching of the gene between the ON and OFF states. This model accounts for the statistical properties of the transcriptional bursts. It also shows how the quantitative features of the distributions of these bursts can be tuned by controlling the appropriate steps of operation of the RNAP machines.
{\it Transcription} is the process whereby RNA molecules are polymerized by molecular machines, called RNA polymerase (RNAP), using the corresponding DNA as the template. Recent {\it in-vivo} experiments with single cells have established that transcription takes place in "bursts" or "pulses" . In this letter we present a model that captures not only the mechano-chemistry of individual RNAPs and their steric interactions but also the switching of the gene between the ON and OFF states. This model accounts for the statistical properties of the transcriptional bursts. It also shows how the quantitative features of the distributions of these bursts can be tuned by controlling the appropriate steps of operation of the RNAP machines.
[ { "type": "R", "before": "``bursts'' or ``pulses''", "after": "\"bursts\" or \"pulses\"", "start_char_pos": 273, "end_char_pos": 297 } ]
[ 0, 170, 494, 576 ]
0804.1720
1
Charge transport modification exhibiting an increase of conductance activated by an external green light in Bacteriorhodopsin is correlated to its conformational change. A theoretical model based on a map of the protein structure into a resistor network is implemented to account for a sequential tunneling mechanism of charge transfer through neighbouring amino-acids . The model is validated by comparison with current-voltage experiments and provides for the potential barriers involved in the charge transfer an average height of 69 meV over an interacting radius of 6 \aa . The predictability of the model is also tested on bovine rhodopsin, the prototype of the G protein coupled receptor (GPCR) family also sensitive to the light, with results exhibiting the opposite behaviour of a decrease of conductance in the presence of light.
When moving from native to light activated bacteriorhodopsin, modification of charge transport consisting of an increase of conductance is correlated to the protein conformational change. A theoretical model based on a map of the protein tertiary structure into a resistor network is implemented to account for a sequential tunneling mechanism of charge transfer through neighbouring amino acids . The model is validated by comparison with current-voltage experiments . The predictability of the model is further tested on bovine rhodopsin, a G-protein coupled receptor (GPCR) also sensitive to light. In this case, results show an opposite behaviour with a decrease of conductance in the presence of light.
[ { "type": "R", "before": "Charge transport modification exhibiting", "after": "When moving from native to light activated bacteriorhodopsin, modification of charge transport consisting of", "start_char_pos": 0, "end_char_pos": 40 }, { "type": "D", "before": "activated by an external green light in Bacteriorhodopsin", "after": null, "start_char_pos": 68, "end_char_pos": 125 }, { "type": "R", "before": "its", "after": "the protein", "start_char_pos": 143, "end_char_pos": 146 }, { "type": "A", "before": null, "after": "tertiary", "start_char_pos": 220, "end_char_pos": 220 }, { "type": "R", "before": "amino-acids", "after": "amino acids", "start_char_pos": 358, "end_char_pos": 369 }, { "type": "D", "before": "and provides for the potential barriers involved in the charge transfer an average height of 69 meV over an interacting radius of 6 \\aa", "after": null, "start_char_pos": 442, "end_char_pos": 577 }, { "type": "R", "before": "also", "after": "further", "start_char_pos": 615, "end_char_pos": 619 }, { "type": "R", "before": "the prototype of the G protein", "after": "a G-protein", "start_char_pos": 648, "end_char_pos": 678 }, { "type": "D", "before": "family", "after": null, "start_char_pos": 703, "end_char_pos": 709 }, { "type": "R", "before": "the light, with results exhibiting the opposite behaviour of", "after": "light. In this case, results show an opposite behaviour with", "start_char_pos": 728, "end_char_pos": 788 } ]
[ 0, 169, 371, 579 ]
0804.2912
1
The numeraire portfolio in a financial market is the unique positive wealth process that makes all other nonnegative wealth processes supermartingales , when deflated by it . The numeraire portfolio depends on market characteristics, which include: (a) the information flow available to acting agents, given by a filtration; (b) the statistical evolution of the asset prices and, more generally, the states of nature, given by a probability measure; and (c) possible restrictions that acting agents might be facing on available investment strategies, modeled by a constraints set. In a financial market with continuous-path asset prices, the stable behavior of the numeraire portfolio is established when each of the aforementioned market parameters is changed in an infinitesimal way.
The numeraire portfolio in a financial market is the unique positive wealth process that makes all other nonnegative wealth processes , when deflated by it , supermartingales . The numeraire portfolio depends on market characteristics, which include: (a) the information flow available to acting agents, given by a filtration; (b) the statistical evolution of the asset prices and, more generally, the states of nature, given by a probability measure; and (c) possible restrictions that acting agents might be facing on available investment strategies, modeled by a constraints set. In a financial market with continuous-path asset prices, we establish the stable behavior of the numeraire portfolio when each of the aforementioned market parameters is changed in an infinitesimal way.
[ { "type": "D", "before": "supermartingales", "after": null, "start_char_pos": 134, "end_char_pos": 150 }, { "type": "A", "before": null, "after": ", supermartingales", "start_char_pos": 173, "end_char_pos": 173 }, { "type": "A", "before": null, "after": "we establish", "start_char_pos": 639, "end_char_pos": 639 }, { "type": "D", "before": "is established", "after": null, "start_char_pos": 687, "end_char_pos": 701 } ]
[ 0, 175, 325, 450, 581 ]
0804.3939
1
Recent works on the emergence of homochirality introduce the concept of recycling, rather than the traditional open-flow system described by Frank. This approach has been criticized by Blackmond et al. They claimed that such systems are thermodynamically impossible, except in the cases where non-microreversible reactions are introduced, like in photochemical reactions, or under the influence of physical actions (e.g. by crystal crushing). This point of view reveals misunderstandings about this model of a recycled system, overlooks the possibility of energy exchanges that could take place in prebiotic systems, and leads the authors to unawarely remove the activation reaction and energy source from their `` non-equilibrium '' models. It is especially important to understand what are the concepts behind the notion of recycled systems, and of activation reactions. These points are fundamental to comprehending how chemical systems -- and especially prebiotic chemical systems -- can be maintained in non-equilibrium steady states, and how free energy can be used and exchanged between systems. The proposed approach aims at the decomposition of the problem, avoiding to embrace the whole system at the same time.
The question of the onset of the homochirality on prebiotic Earth still remains a fundamental question in the quest for the origin of life. Recent works in this field introduce the concept of recycling, rather than the traditional open-flow system described by Frank. This approach has been criticized by Blackmond et al. They claimed that such systems are thermodynamically impossible, except in the cases where non-microreversible reactions are introduced, like in photochemical reactions, or under the influence of physical actions (e.g. by crystal crushing). This point of view reveals misunderstandings about this model of a recycled system, overlooks the possibility of energy exchanges that could take place in prebiotic systems, and leads the authors to unawarely remove the activation reaction and energy source from their " non-equilibrium " models. It is especially important to understand what are the concepts behind the notion of recycled systems, and of activation reactions. These points are fundamental to comprehending how chemical systems -- and especially prebiotic chemical systems -- can be maintained in non-equilibrium steady states, and how free energy can be used and exchanged between systems. The proposed approach aims at the decomposition of the problem, avoiding to embrace the whole system at the same time.
[ { "type": "R", "before": "Recent works on the emergence of homochirality", "after": "The question of the onset of the homochirality on prebiotic Earth still remains a fundamental question in the quest for the origin of life. Recent works in this field", "start_char_pos": 0, "end_char_pos": 46 }, { "type": "R", "before": "``", "after": "\"", "start_char_pos": 712, "end_char_pos": 714 }, { "type": "R", "before": "''", "after": "\"", "start_char_pos": 731, "end_char_pos": 733 } ]
[ 0, 147, 201, 442, 741, 872, 1102 ]
0804.4039
1
In this work we model multicore platforms as chip multiprocessors made up from two different types cores; each type of cores has different computation capabilities. In this simple asymmetric multicore model we study the problem of executing parallel programs that are represented as DAGs or chains of tasks . We assume 2 (essential) speeds in our proposed platform. We relax this assumption and consider many speedsin our energy considerations. This problem has been considered previously in the field of parallel computing and scheduling theory for the general case where multiprocessor platforms have K distinct speeds. Our model is a bridge between the assumptions in these fields and recent advances in multicore processors technology . We provide a refined analysis of a recent scheduling method. Based on this analysis, we specialize the scheduling policy and provide an algorithm of (3 o(1)) expected approximation factor. Note that this improves the previous best factor (6 for two speeds) .
In this work we study the problem of scheduling tasks with dependencies in multiprocessor architectures where processors have different speeds. We present the preemptive algorithm "Save-Energy" that given a schedule of tasks it post processes it to improve the energy efficiency without any deterioration of the makespan. In terms of time efficiency, we show that preemptive scheduling in an asymmetric system can achieve the same or better optimal makespan than in a symmetric system. Motivited by real multiprocessor systems, we investigate architectures that exhibit limited asymmetry: there are two essentially different speeds . Interestingly, this special case has not been studied in the field of parallel computing and scheduling theory ; only the general case was studied where processors have K essentially different speeds. We present the non-preemptive algorithm ``Remnants'' that achieves almost optimal makespan . We provide a refined analysis of a recent scheduling method. Based on this analysis, we specialize the scheduling policy and provide an algorithm of (3 + o(1)) expected approximation factor. Note that this improves the previous best factor (6 for two speeds) . We believe that our work will convince researchers to revisit this well studied scheduling problem for these simple, yet realistic, asymmetric multiprocessor architectures .
[ { "type": "D", "before": "model multicore platforms as chip multiprocessors made up from two different types cores; each type of cores has different computation capabilities. In this simple asymmetric multicore model we", "after": null, "start_char_pos": 16, "end_char_pos": 209 }, { "type": "R", "before": "executing parallel programs that are represented as DAGs or chains of tasks", "after": "scheduling tasks with dependencies in multiprocessor architectures where processors have different speeds. We present the preemptive algorithm \"Save-Energy\" that given a schedule of tasks it post processes it to improve the energy efficiency without any deterioration of the makespan. In terms of time efficiency, we show that preemptive scheduling in an asymmetric system can achieve the same or better optimal makespan than in a symmetric system. Motivited by real multiprocessor systems, we investigate architectures that exhibit limited asymmetry: there are two essentially different speeds", "start_char_pos": 231, "end_char_pos": 306 }, { "type": "R", "before": "We assume 2 (essential) speeds in our proposed platform. We relax this assumption and consider many speedsin our energy considerations. This problem has been considered previously", "after": "Interestingly, this special case has not been studied", "start_char_pos": 309, "end_char_pos": 488 }, { "type": "R", "before": "for", "after": "; only", "start_char_pos": 546, "end_char_pos": 549 }, { "type": "R", "before": "where multiprocessor platforms have K distinct speeds. Our model is a bridge between the assumptions in these fields and recent advances in multicore processors technology", "after": "was studied where processors have K essentially different speeds. We present the non-preemptive algorithm ``Remnants'' that achieves almost optimal makespan", "start_char_pos": 567, "end_char_pos": 738 }, { "type": "A", "before": null, "after": "+", "start_char_pos": 893, "end_char_pos": 893 }, { "type": "A", "before": null, "after": ". We believe that our work will convince researchers to revisit this well studied scheduling problem for these simple, yet realistic, asymmetric multiprocessor architectures", "start_char_pos": 999, "end_char_pos": 999 } ]
[ 0, 105, 164, 308, 365, 444, 621, 740, 801, 930 ]
0805.0540
1
Stochastic differential equations are commonly used to describe the dynamics of diffusive systems. Since their introduction, they have also been considered natural candidates for the description of the dynamics of financial returns. In this context we analyze the problem of the analytical characterization of the still unknown probability distribution of returns of the exponential Ornstein-Uhlenbeck model with stochastic volatility , where the price is driven by a Geometric Brownian motion, whose diffusion coefficient is expressed as an exponential function of an hidden variable Y governed by an Ornstein-Uhlenbeck process. We derive closed-form expressions for the probability distribution and its characteristic function in two limit cases. In the first one the fluctuations of Y are larger than the volatility normal level, while the second one corresponds to the assumption of a small stationary value for the variance of Y. Theoretical results are tested numerically by intensive use of Monte Carlo simulations. The effectiveness of the analytical predictions is checked via a careful analysis of the parameters involved in the numerical implementation of the Euler-Maruyama scheme . Both analytical and numerical results we obtain can be exploited for useful applications in stock option pricing and risk management .
We analyze the problem of the analytical characterization of the probability distribution of financial returns in the exponential Ornstein-Uhlenbeck model with stochastic volatility . In this model the prices are driven by a Geometric Brownian motion, whose diffusion coefficient is expressed through an exponential function of an hidden variable Y governed by a mean-reverting process. We derive closed-form expressions for the probability distribution and its characteristic function in two limit cases. In the first one the fluctuations of Y are larger than the volatility normal level, while the second one corresponds to the assumption of a small stationary value for the variance of Y. Theoretical results are tested numerically by intensive use of Monte Carlo simulations. The effectiveness of the analytical predictions is checked via a careful analysis of the parameters involved in the numerical implementation of the Euler-Maruyama scheme and is tested on a data set of financial indexes. In particular, we discuss results for the German DAX30 and Dow Jones Euro Stoxx 50, finding a good agreement between the empirical data and the theoretical description .
[ { "type": "R", "before": "Stochastic differential equations are commonly used to describe the dynamics of diffusive systems. Since their introduction, they have also been considered natural candidates for the description of the dynamics of financial returns. In this context we", "after": "We", "start_char_pos": 0, "end_char_pos": 251 }, { "type": "D", "before": "still unknown", "after": null, "start_char_pos": 314, "end_char_pos": 327 }, { "type": "R", "before": "returns of", "after": "financial returns in", "start_char_pos": 356, "end_char_pos": 366 }, { "type": "R", "before": ", where the price is", "after": ". In this model the prices are", "start_char_pos": 435, "end_char_pos": 455 }, { "type": "R", "before": "as", "after": "through", "start_char_pos": 536, "end_char_pos": 538 }, { "type": "R", "before": "an Ornstein-Uhlenbeck", "after": "a mean-reverting", "start_char_pos": 599, "end_char_pos": 620 }, { "type": "R", "before": ". Both analytical and numerical results we obtain can be exploited for useful applications in stock option pricing and risk management", "after": "and is tested on a data set of financial indexes. In particular, we discuss results for the German DAX30 and Dow Jones Euro Stoxx 50, finding a good agreement between the empirical data and the theoretical description", "start_char_pos": 1193, "end_char_pos": 1327 } ]
[ 0, 98, 232, 629, 748, 934, 1022, 1194 ]
0805.1293
1
Iterative Logic Arrays (ILAs) are ideal as VLSI sub-systems because of their regular structure and its close resemblance with FPGAs (Field Programmable Gate Arrays). AND-EXOR based circuits are of interest in the design of very low power circuits where energy loss implied by high frequency switching is of much consideration. This paper examines the testability of AND-EXOR based Iterative Logic Arrays (ILAs) . For certain ILAs it is possible to find a test set whose size remains constant irrespective of the size of the ILA, while for others it varies with array size. Former type of ILAs is known as Constant-Testable ( C-Testable ) . It has been shown that AND-EXOR based Logic Arrays are C-Testable and size of test set is equal to number of entries in cell truth table . The test generation problem has been shown to be related to certain properties of cycles in a set of graphs derived from cell truth table. By careful analysis of these cycles an efficient test generation technique that can be easily converted to an ATPG program has been presented for both 1D and 2D ILAs. How this property of ILAs can be used for testing FPGAs has also been discussed.Keywords - C-Testable, Iterative Logic Arrays, Bijective, Test Generation, ATPG
Iterative Logic Arrays (ILAs) are ideal as VLSI sub-systems because of their regular structure and its close resemblance with FPGAs (Field Programmable Gate Arrays). Reversible circuits are of interest in the design of very low power circuits where energy loss implied by high frequency switching is not of much consideration. Reversibility is essential for Quantum Computing. This paper examines the testability of Reversible Iterative Logic Arrays (ILAs) composed of reversible k-CNOT gates . For certain ILAs it is possible to find a test set whose size remains constant irrespective of the size of the ILA, while for others it varies with array size. Former type of ILAs is known as Constant-Testable , i.e. C-Testable . It has been shown that Reversible Logic Arrays are C-Testable and size of test set is equal to number of entries in cells truth table implying that the reversible ILAs are also Optimal-Testable, i.e. O-Testable . Uniform-Testability, i.e. U-Testability has been defined and Reversible Heterogeneous ILAs have been characterized as U-Testable. The test generation problem has been shown to be related to certain properties of cycles in a set of graphs derived from cell truth table. By careful analysis of these cycles an efficient test generation technique that can be easily converted to an ATPG program has been presented for both 1-D and 2D ILAs. The same algorithms can be easily extended for n-Dimensional Reversible ILAs.
[ { "type": "R", "before": "AND-EXOR based", "after": "Reversible", "start_char_pos": 166, "end_char_pos": 180 }, { "type": "A", "before": null, "after": "not", "start_char_pos": 304, "end_char_pos": 304 }, { "type": "A", "before": null, "after": "Reversibility is essential for Quantum Computing.", "start_char_pos": 328, "end_char_pos": 328 }, { "type": "R", "before": "AND-EXOR based", "after": "Reversible", "start_char_pos": 368, "end_char_pos": 382 }, { "type": "A", "before": null, "after": "composed of reversible k-CNOT gates", "start_char_pos": 413, "end_char_pos": 413 }, { "type": "R", "before": "(", "after": ", i.e.", "start_char_pos": 626, "end_char_pos": 627 }, { "type": "D", "before": ")", "after": null, "start_char_pos": 639, "end_char_pos": 640 }, { "type": "R", "before": "AND-EXOR based", "after": "Reversible", "start_char_pos": 666, "end_char_pos": 680 }, { "type": "R", "before": "cell truth table", "after": "cells truth table implying that the reversible ILAs are also Optimal-Testable, i.e. O-Testable", "start_char_pos": 763, "end_char_pos": 779 }, { "type": "A", "before": null, "after": "Uniform-Testability, i.e. U-Testability has been defined and Reversible Heterogeneous ILAs have been characterized as U-Testable.", "start_char_pos": 782, "end_char_pos": 782 }, { "type": "R", "before": "1D", "after": "1-D", "start_char_pos": 1073, "end_char_pos": 1075 }, { "type": "R", "before": "How this property of ILAs can be used for testing FPGAs has also been discussed.Keywords - C-Testable, Iterative Logic Arrays, Bijective, Test Generation, ATPG", "after": "The same algorithms can be easily extended for n-Dimensional Reversible ILAs.", "start_char_pos": 1089, "end_char_pos": 1248 } ]
[ 0, 165, 327, 415, 575, 642, 781, 921, 1088, 1169 ]
0806.1170
1
Based on analogies with statistical physics and complexity theory, we have developed in the last decade an approach that diagnoses bubbles as transient super-exponential regimes. Here, We present an analysis of oil prices in US dollars and in other major currencies that diagnoses unsustainable faster-than-exponential behavior. This provides evidence that the recent oil price run-up has been amplified by speculative behavior of the type found during a bubble-like expansion.
We present an analysis of oil prices in US and in other major currencies that diagnoses unsustainable faster-than-exponential behavior. This supports the hypothesis that the recent oil price run-up has been amplified by speculative behavior of the type found during a bubble-like expansion. We also attempt to unravel the information hidden in the oil supply-demand data reported by two leading agencies, the US Energy Information Administration (EIA) and the International Energy Agency (IEA). We suggest that the found increasing discrepancy between the EIA and IEA figures provides a measure of the estimation errors. Rather than a clear transition to a supply restricted regime, we interpret the discrepancy between the IEA and EIA as a signature of uncertainty, and there is no better fuel than uncertainty to promote speculation!
[ { "type": "D", "before": "Based on analogies with statistical physics and complexity theory, we have developed in the last decade an approach that diagnoses bubbles as transient super-exponential regimes. Here,", "after": null, "start_char_pos": 0, "end_char_pos": 184 }, { "type": "D", "before": "dollars", "after": null, "start_char_pos": 228, "end_char_pos": 235 }, { "type": "R", "before": "provides evidence", "after": "supports the hypothesis", "start_char_pos": 334, "end_char_pos": 351 }, { "type": "A", "before": null, "after": "We also attempt to unravel the information hidden in the oil supply-demand data reported by two leading agencies, the US Energy Information Administration (EIA) and the International Energy Agency (IEA). We suggest that the found increasing discrepancy between the EIA and IEA figures provides a measure of the estimation errors. Rather than a clear transition to a supply restricted regime, we interpret the discrepancy between the IEA and EIA as a signature of uncertainty, and there is no better fuel than uncertainty to promote speculation!", "start_char_pos": 478, "end_char_pos": 478 } ]
[ 0, 178, 328 ]
0806.2763
1
One approach to studying the URLanization of biochemistry is to use statistical graph theory. Even in such a heavily simplified method, which disregards most of the dynamic aspects of biochemistry, one is faced with fundamental questions, such as how the chemical reaction systems should be reduced to a graph so that as much information as possible about the original network is retained. Vertices may represent either chemical substances, or reactions , or both . A common practice in the analysis of metabolic networks is to preprocess the network by deleting so called currency metabolites -- abundant molecular species that react with a large number of metabolites. One of several reasons to do this is that they do not put any constraint on the reactions, and thus do not give much information of the higher order functionality of the system. In this paper we evaluate different network representations of metabolism and describe their structure with respect to both modularity and currency metabolites . We find that a "substance network," where all metabolites participating in a reaction are connected, is a good trade-off between high functional overlap (between network modules and observed functions ) and a reasonable number of currency metabolites.
One approach to studying the URLanization of biochemistry is to use statistical graph theory. Even in such a heavily simplified method, which disregards most of the dynamic aspects of biochemistry, one is faced with fundamental questions, such as how the chemical reaction systems should be reduced to a graph retaining as much functional information as possible from the original reaction system. In such graph representations, should the edges go between substrates and products, or substrates and substrates , or both ? Should vertices represent substances or reactions? Different definitions encode different information about the reaction system. In this paper we evaluate four different graph representations of metabolism , applied to data from URLanisms and databases. The graph representations are evaluated by comparing the overlap between clusters (network modules) and annotated functions, and also by comparing the set of identified currency metabolites with those that other authors have identified using qualitative biological arguments . We find that a "substance network," where all metabolites participating in a reaction are connected, is relatively better than others, evaluated both with respect to the functional overlap between modules and functions and to the number and identity of identified currency metabolites.
[ { "type": "R", "before": "so that as much", "after": "retaining as much functional", "start_char_pos": 310, "end_char_pos": 325 }, { "type": "R", "before": "about the original network is retained. Vertices may represent either chemical substances, or reactions", "after": "from the original reaction system. In such graph representations, should the edges go between substrates and products, or substrates and substrates", "start_char_pos": 350, "end_char_pos": 453 }, { "type": "R", "before": ". A common practice in the analysis of metabolic networks is to preprocess the network by deleting so called currency metabolites -- abundant molecular species that react with a large number of metabolites. One of several reasons to do this is that they do not put any constraint on the reactions, and thus do not give much information of the higher order functionality of the", "after": "? Should vertices represent substances or reactions? Different definitions encode different information about the reaction", "start_char_pos": 464, "end_char_pos": 840 }, { "type": "R", "before": "different network", "after": "four different graph", "start_char_pos": 875, "end_char_pos": 892 }, { "type": "R", "before": "and describe their structure with respect to both modularity and currency metabolites", "after": ", applied to data from URLanisms and databases. The graph representations are evaluated by comparing the overlap between clusters (network modules) and annotated functions, and also by comparing the set of identified currency metabolites with those that other authors have identified using qualitative biological arguments", "start_char_pos": 923, "end_char_pos": 1008 }, { "type": "R", "before": "a good trade-off between high functional overlap (between network modules and observed functions ) and a reasonable number of", "after": "relatively better than others, evaluated both with respect to the functional overlap between modules and functions and to the number and identity of identified", "start_char_pos": 1115, "end_char_pos": 1240 } ]
[ 0, 93, 389, 465, 670, 848, 1010 ]
0806.4061
1
In this article we study an optimal stopping/optimal control problem which models the decision facing a risk averse agent over when to sell an asset. The market is incomplete so that the asset exposure cannot be hedged. In addition to the decision over when to sell, the agent has to choose a control strategy which corresponds to a feasible wealth process. We formulate this problem as one involving the choice of a stopping time and a martingale. We conjecture the form of the solution and verify that the candidate solution is equal to the value function. The interesting features of the solution are that it is available in a very explicit form, that for some parameter values the optimal strategy is more sophisticated than might originally be expected, and that although the set-up is based on continuous diffusions, the optimal martingale may involve a jump process. One interpretation of the solution is that it is optimal for the risk averse agent to gamble.
In this article we study an optimal stopping/optimal control problem which models the decision facing a risk-averse agent over when to sell an asset. The market is incomplete so that the asset exposure cannot be hedged. In addition to the decision over when to sell, the agent has to choose a control strategy which corresponds to a feasible wealth process. We formulate this problem as one involving the choice of a stopping time and a martingale. We conjecture the form of the solution and verify that the candidate solution is equal to the value function. The interesting features of the solution are that it is available in a very explicit form, that for some parameter values the optimal strategy is more sophisticated than might originally be expected, and that although the setup is based on continuous diffusions, the optimal martingale may involve a jump process. One interpretation of the solution is that it is optimal for the risk-averse agent to gamble.
[ { "type": "R", "before": "risk averse", "after": "risk-averse", "start_char_pos": 104, "end_char_pos": 115 }, { "type": "R", "before": "set-up", "after": "setup", "start_char_pos": 781, "end_char_pos": 787 }, { "type": "R", "before": "risk averse", "after": "risk-averse", "start_char_pos": 939, "end_char_pos": 950 } ]
[ 0, 149, 219, 357, 448, 558, 873 ]
0806.4125
1
This paper is a survey of recent progress in the theory of ruin for risk processes that earn investment return on invested assets .
This survey treats the problem of ruin in a risk model when assets earn investment income. In addition to a general presentation of the problem, topics covered are a presentation of the relevant integro-differential equations, exact and numerical solutions, asymptotic results, bounds on the ruin probability and also the possibility of minimizing the ruin probability by investment and possibly reinsurance control. The main emphasis is on continuous time models, but discrete time models are also covered. A fairly extensive list of references is provided, particularly of papers published after 1998. For more references to papers published before that, the reader can consult 47 .
[ { "type": "R", "before": "paper is a survey of recent progress in the theory of ruin for risk processes that earn investment return on invested assets", "after": "survey treats the problem of ruin in a risk model when assets earn investment income. In addition to a general presentation of the problem, topics covered are a presentation of the relevant integro-differential equations, exact and numerical solutions, asymptotic results, bounds on the ruin probability and also the possibility of minimizing the ruin probability by investment and possibly reinsurance control. The main emphasis is on continuous time models, but discrete time models are also covered. A fairly extensive list of references is provided, particularly of papers published after 1998. For more references to papers published before that, the reader can consult", "start_char_pos": 5, "end_char_pos": 129 }, { "type": "A", "before": null, "after": "47", "start_char_pos": 130, "end_char_pos": 130 } ]
[ 0 ]
0807.0308
1
Active centres and hot spots of proteins have a paramount importance in enzyme action, protein complex formation and drug design. Recently a number of publications successfully applied the analysis of residue networks to predict active centres in proteins. Most real-world networks show a number of properties, such as small-worldness or scale-free degree distribution, which are rather general features of networks from molecules to the society . Based on extensive analogies I propose that the existing findings and methodology enable us to detect active centres in cells, social networks and ecosystems. Members of these active centres are creative elements of the respective networks, which may help them to survive unprecedented, novel challenges, and play a key role in the development, survival and evolvability of complex systems.
Active centres and hot spots of proteins have a paramount importance in enzyme action, protein complex formation and drug design. Recently a number of publications successfully applied the analysis of residue networks to predict active centres in proteins. Most real-world networks show a number of properties, such as small-worldness or scale-free degree distribution, which are rather general features of networks , from molecules to society at large. Using analogy I propose that existing findings and methodology already enable us to detect active centres in cells, and can be expanded to social networks and ecosystems. Members of these active centres are termed here as creative elements of their respective networks, which may help them to survive unprecedented, novel challenges, and play a key role in the development, survival and evolvability of complex systems.
[ { "type": "A", "before": null, "after": ",", "start_char_pos": 416, "end_char_pos": 416 }, { "type": "R", "before": "the society . Based on extensive analogies", "after": "society at large. Using analogy", "start_char_pos": 435, "end_char_pos": 477 }, { "type": "D", "before": "the", "after": null, "start_char_pos": 493, "end_char_pos": 496 }, { "type": "A", "before": null, "after": "already", "start_char_pos": 531, "end_char_pos": 531 }, { "type": "A", "before": null, "after": "and can be expanded to", "start_char_pos": 577, "end_char_pos": 577 }, { "type": "A", "before": null, "after": "termed here as", "start_char_pos": 646, "end_char_pos": 646 }, { "type": "R", "before": "the", "after": "their", "start_char_pos": 668, "end_char_pos": 671 } ]
[ 0, 129, 256, 448, 609 ]
0807.1823
1
We address two issues that puzzled moral and socio-economic thinking since antiquity. The first of the puzzles is the emergence and selection fitness of altruist behavior in a world of self-reproducing individuals (or capital,etc.; memes in general). The second is the sustainability of growth and the survival in a stochastic world where along with large gains individuals incur often very large and even total losses. We show that the solution for each of these puzzles lies within the other one. It is known that in a multiplicative random process even if in the mathematical evaluation of the expected gain the wins overwhelm the losses, one is likely to be faced with extinction.For example if the probability of a total loss event is arbitrarily small but finite, the measure of histories with non-vanishing gains approaches 0 for asymptotic times >. Thus the optimistic theoretical expectation is dominated by event chains whose probability is too small to happen in reality. We find that in those situations, there is a way to insure self survival. The individuals can insure continuous, unlimited growth by sharing their gains/losses after each step of the random process. We compute the minimal necessary sharing group size Ncrit . Usually, in order to make "sharing" a stable profitable strategy, one has to introduce a credible punishment (/reward) mechanism that can be trusted by the individuals. However, such coercion is incompatible with the concept of real altruism. In a group of Ncrit individuals , defection by one individual leads automatically, thus credibly, to the ruin of the entire group (including the defector) . Consequently sharing becomes the only stable survivalstrategy without the need of any additional enforcing mechanism. This implies the survival and thriving of "altruistic genes" for very long evolutionary periods .
Most of real life systems have a random component: the multitude of endogenous and exogenous factors influencing them result in stochastic fluctuations of the parameters influencing their dynamics. While in the physical systems the random component is usually additive (e.g. molecules can absorb or emit spontaneously energy quanta), in biological, social, economic and financial systems the noise is mainly multiplicative. The special properties of multiplicative noise as opposed to the additive noise have been noticed for a long while. Even though apparently and formally the difference between free additive vs. multiplicative random walks consists in just a move from normal to log-normal distributions, in practice the implications are much more far reaching. In the present paper we study in some detail the difference between the emergence of cooperation in additive vs. multiplicative random environments. While in an additive context the emergence and survival of cooperation requires special conditions (especially some level of reward, punishment, reciprocity), we find that in the multiplicative random context the emergence of cooperation is much more natural and effective. In fact given a set of individuals that play independently and repeatedly a game where they can gain by a factor a>1 with probability p or loose by a factor b<1 with probability q=1-p, one can show that for a^p b^q < 1 < pa +q b the cooperation makes the difference between extinction and (thriving) survival. We study quantitatively the various implications of this observation and its applications in various contexts .
[ { "type": "R", "before": "We address two issues that puzzled moral and socio-economic thinking since antiquity. The first of the puzzles is the emergence and selection fitness of altruist behavior in a world of self-reproducing individuals (or capital,etc.; memes in general). The second is the sustainability of growth and the survival in a stochastic world where along with large gains individuals incur often very large and even total losses. We show that the solution for each of these puzzles lies within the other one. It is known that in a multiplicative random process even if in the mathematical evaluation of the expected gain the wins overwhelm the losses, one is likely to be faced with extinction.For example if the probability of a total loss event is arbitrarily small but finite,", "after": "Most of real life systems have a random component: the multitude of endogenous and exogenous factors influencing them result in stochastic fluctuations of", "start_char_pos": 0, "end_char_pos": 769 }, { "type": "R", "before": "measure of histories with non-vanishing gains approaches 0 for asymptotic times >. Thus the optimistic theoretical expectation is dominated by event chains whose probability is too small to happen in reality. We find that in those situations, there is a way to insure self survival. The individuals can insure continuous, unlimited growth by sharing their gains/losses after each step of the random process. We compute the minimal necessary sharing group size Ncrit . Usually, in order to make \"sharing\" a stable profitable strategy, one has to introduce a credible punishment (/reward) mechanism that can be trusted by the individuals. However, such coercion is incompatible with the concept of real altruism. In a group of Ncrit individuals , defection by one individual leads automatically, thus credibly, to the ruin of the entire group (including the defector) . Consequently sharing becomes the only stable survivalstrategy without the need of any additional enforcing mechanism. This implies the survival and thriving of \"altruistic genes\" for very long evolutionary periods", "after": "parameters influencing their dynamics. While in the physical systems the random component is usually additive (e.g. molecules can absorb or emit spontaneously energy quanta), in biological, social, economic and financial systems the noise is mainly multiplicative. The special properties of multiplicative noise as opposed to the additive noise have been noticed for a long while. Even though apparently and formally the difference between free additive vs. multiplicative random walks consists in just a move from normal to log-normal distributions, in practice the implications are much more far reaching. In the present paper we study in some detail the difference between the emergence of cooperation in additive vs. multiplicative random environments. While in an additive context the emergence and survival of cooperation requires special conditions (especially some level of reward, punishment, reciprocity), we find that in the multiplicative random context the emergence of cooperation is much more natural and effective. In fact given a set of individuals that play independently and repeatedly a game where they can gain by a factor a>1 with probability p or loose by a factor b<1 with probability q=1-p, one can show that for a^p b^q < 1 < pa +q b the cooperation makes the difference between extinction and (thriving) survival. We study quantitatively the various implications of this observation and its applications in various contexts", "start_char_pos": 774, "end_char_pos": 1855 } ]
[ 0, 85, 231, 250, 419, 498, 684, 856, 982, 1056, 1181, 1241, 1410, 1484, 1641, 1759 ]
0807.1823
2
Most of real life systems have a random component: the multitude of endogenous and exogenous factors influencing them result in stochastic fluctuations of the parameters influencing their dynamics. While in the physical systems the random component is usually additive (e.g. molecules can absorb or emit spontaneously energy quanta), in biological, social, economic and financial systems the noise is mainly multiplicative . The special properties of multiplicative noise as opposed to the additive noise have been noticed for a long while. Even though apparently and formally the difference between free additive vs. multiplicative random walks consists in just a move from normal to log-normal distributions, in practice the implications are much more far reaching. In the present paper we study in some detail the difference between the emergence of cooperation in additive vs. multiplicative random environments. While in an additive context the emergence and survival of cooperation requires special conditions (especially some level of reward, punishment, reciprocity), we find that in the multiplicative random context the emergence of cooperation is much more natural and effective. In fact given a set of individuals that play independently and repeatedly a game where they can gain by a factor a>1 with probability p or loose by a factor b<1 with probability q=1-p, one can show that for a^p b^q < 1 < pa +q b the cooperation makes the difference between extinction and (thriving) survival. We study quantitatively the various implications of this observation and its applications in various contexts.
Most real life systems have a random component: the multitude of endogenous and exogenous factors influencing them result in stochastic fluctuations of the parameters determining their dynamics. These empirical systems are in many cases subject to noise of multiplicative nature . The special properties of multiplicative noise as opposed to additive noise have been noticed for a long while. Even though apparently and formally the difference between free additive vs. multiplicative random walks consists in just a move from normal to log-normal distributions, in practice the implications are much more far reaching. While in an additive context the emergence and survival of cooperation requires special conditions (especially some level of reward, punishment, reciprocity), we find that in the multiplicative random context the emergence of cooperation is much more natural and effective. We study the various implications of this observation and its applications in various contexts.
[ { "type": "D", "before": "of", "after": null, "start_char_pos": 5, "end_char_pos": 7 }, { "type": "R", "before": "influencing", "after": "determining", "start_char_pos": 170, "end_char_pos": 181 }, { "type": "R", "before": "While in the physical systems the random component is usually additive (e.g. molecules can absorb or emit spontaneously energy quanta), in biological, social, economic and financial systems the noise is mainly multiplicative", "after": "These empirical systems are in many cases subject to noise of multiplicative nature", "start_char_pos": 198, "end_char_pos": 422 }, { "type": "D", "before": "the", "after": null, "start_char_pos": 486, "end_char_pos": 489 }, { "type": "D", "before": "In the present paper we study in some detail the difference between the emergence of cooperation in additive vs. multiplicative random environments.", "after": null, "start_char_pos": 768, "end_char_pos": 916 }, { "type": "R", "before": "In fact given a set of individuals that play independently and repeatedly a game where they can gain by a factor a>1 with probability p or loose by a factor b<1 with probability q=1-p, one can show that for a^p b^q < 1 < pa +q b the cooperation makes the difference between extinction and (thriving) survival. We study quantitatively", "after": "We study", "start_char_pos": 1191, "end_char_pos": 1524 } ]
[ 0, 197, 424, 540, 767, 916, 1190, 1500 ]
0808.0321
1
Constraint-based modelling of metabolic networks often gives rise to a linear programming problem. Mathematically, there always exists a dual to any linear programming problem, with dual variables known as shadow prices. Here we analyse the structure of the dual problem and the properties of the shadow prices, for typical constraint-based metabolic models. We apply our results to a representative set of recent genome-scale metabolic reconstructions. The dual problem is found to have an elegant structure. The shadow prices are to be interpreted as yield coefficients Varma and Palsson, J. theor. Biol. v165, 477; 503 (1993)%DIFDELCMD < ]%%% . In combination with reaction fluxes , they can be used to decorate the metabolic model with a yield flux network. The yield flux network shows a kind of gauge invariance with respect to rescaling of the stoichiometric matrix, and is arguably more fundamental than either the reaction fluxes or shadow prices. Complementary slackness implies that the yield flux network also satisfies a conservation law, which can be used to explain why the shadow prices are strongly correlated with measures of molecular complexity such as molecular weight and atom count. For the genome-scale models , the shadow prices have a broad distribution and the overall pattern reflects the URLanisation of the metabolism .
%DIFDELCMD < ]%%% A metabolic model can be represented as bipartite graph comprising linked reaction and metabolite nodes. Here it is shown how a network of conserved fluxes can be assigned to the edges of such a graph by combining the reaction fluxes with a conserved metabolite property such as molecular weight. A similar flux network can be constructed by combining the primal and dual solutions to the linear programming problem that typically arises in constraint-based modelling. Such constructions may help with the visualisation of flux distributions in complex metabolic networks. The analysis also explains the strong correlation observed between metabolite shadow prices (the dual linear programming variables) and conserved metabolite properties. The methods were applied to recent metabolic models for Escherichia coli, Saccharomyces cerevisiae, and Methanosarcina barkeri. Detailed results are reported for E. coli; similar results were found for the URLanisms .
[ { "type": "D", "before": "Constraint-based modelling of metabolic networks often gives rise to a linear programming problem. Mathematically, there always exists a dual to any linear programming problem, with dual variables known as shadow prices. Here we analyse the structure of the dual problem and the properties of the shadow prices, for typical constraint-based metabolic models. We apply our results to a representative set of recent genome-scale metabolic reconstructions. The dual problem is found to have an elegant structure. The shadow prices are to be interpreted as yield coefficients", "after": null, "start_char_pos": 0, "end_char_pos": 571 }, { "type": "D", "before": "Varma and Palsson, J. theor. Biol. v165, 477; 503 (1993)", "after": null, "start_char_pos": 572, "end_char_pos": 628 }, { "type": "R", "before": ". In combination with reaction fluxes , they can be used to decorate the metabolic model with a yield flux network. The yield flux network shows a kind of gauge invariance with respect to rescaling of the stoichiometric matrix, and is arguably more fundamental than either the reaction fluxes or shadow prices. Complementary slackness implies that", "after": "A metabolic model can be represented as bipartite graph comprising linked reaction and metabolite nodes. Here it is shown how a network of conserved fluxes can be assigned to", "start_char_pos": 646, "end_char_pos": 993 }, { "type": "R", "before": "yield flux network also satisfies a conservation law, which can be used to explain why the shadow prices are strongly correlated with measures of molecular complexity such as molecular weight and atom count. For the genome-scale models , the shadow prices have a broad distribution and the overall pattern reflects the URLanisation of the metabolism", "after": "edges of such a graph by combining the reaction fluxes with a conserved metabolite property such as molecular weight. A similar flux network can be constructed by combining the primal and dual solutions to the linear programming problem that typically arises in constraint-based modelling. Such constructions may help with the visualisation of flux distributions in complex metabolic networks. The analysis also explains the strong correlation observed between metabolite shadow prices (the dual linear programming variables) and conserved metabolite properties. The methods were applied to recent metabolic models for Escherichia coli, Saccharomyces cerevisiae, and Methanosarcina barkeri. Detailed results are reported for E. coli; similar results were found for the URLanisms", "start_char_pos": 998, "end_char_pos": 1347 } ]
[ 0, 98, 220, 358, 453, 509, 600, 617, 761, 956, 1205 ]
0808.1431
1
The universal scalability law (USL) of computational capacity is a rational function C_p = P(p)/Q(p) with P(p) a linear polynomial and Q(p) a second-degree polynomial in the number of physical processors p, that has long been used for statistical modeling and prediction of computer system performance. We prove that C_p is equivalent to the synchronous throughput bound for a machine-repairman with state-dependent service rate. Simpler rational functions, such as Amdahl's law and Gustafson speedup, are corollaries of this queue-theoretic bound. C_p is both necessary and sufficient for modeling all practical characteristics of computational scalability.
The universal scalability law of computational capacity is a rational function C_p = P(p)/Q(p) with P(p) a linear polynomial and Q(p) a second-degree polynomial in the number of physical processors p, that has been long used for statistical modeling and prediction of computer system performance. We prove that C_p is equivalent to the synchronous throughput bound for a machine-repairman with state-dependent service rate. Simpler rational functions, such as Amdahl's law and Gustafson speedup, are corollaries of this queue-theoretic bound. C_p is further shown to be both necessary and sufficient for modeling all practical characteristics of computational scalability.
[ { "type": "D", "before": "(USL)", "after": null, "start_char_pos": 30, "end_char_pos": 35 }, { "type": "R", "before": "long been", "after": "been long", "start_char_pos": 216, "end_char_pos": 225 }, { "type": "A", "before": null, "after": "further shown to be", "start_char_pos": 556, "end_char_pos": 556 } ]
[ 0, 302, 429, 548 ]
0808.3339
1
By analyzing ] financial time series of exceptional events such as bubbles and crashes, we confirm the existence of nonlinear potential forces in the markets. The shape of the potential function changes rather slowly and smoothly, thereby enabling a statistical prediction of market prices for a limited time range in the mathematical sense. This predictability does not ensure profits, but is expected to be applicable for averting catastrophes .
Basic peculiarities of market price fluctuations are known to be well described by a recently developed random walk model in a temporally deforming quadric potential force whose center is given by a moving average of past price traces Physica A 370, pp91-97, 2006]. By analyzing high-frequency financial time series of exceptional events such as bubbles and crashes, we confirm the appearance of nonlinear potential force in the markets. We show statistical significance of its existence by applying the information criterion. This new time series analysis is expected to be applied widely for detecting a non-stationary symptom in random phenomena .
[ { "type": "R", "before": "By analyzing", "after": "Basic peculiarities of market price fluctuations are known to be well described by a recently developed random walk model in a temporally deforming quadric potential force whose center is given by a moving average of past price traces", "start_char_pos": 0, "end_char_pos": 12 }, { "type": "A", "before": null, "after": "Physica A 370, pp91-97, 2006", "start_char_pos": 13, "end_char_pos": 13 }, { "type": "A", "before": null, "after": ". By analyzing high-frequency", "start_char_pos": 14, "end_char_pos": 14 }, { "type": "R", "before": "existence", "after": "appearance", "start_char_pos": 103, "end_char_pos": 112 }, { "type": "R", "before": "forces", "after": "force", "start_char_pos": 136, "end_char_pos": 142 }, { "type": "R", "before": "The shape of the potential function changes rather slowly and smoothly, thereby enabling a statistical prediction of market prices for a limited time range in the mathematical sense. This predictability does not ensure profits, but", "after": "We show statistical significance of its existence by applying the information criterion. This new time series analysis", "start_char_pos": 159, "end_char_pos": 390 }, { "type": "R", "before": "applicable for averting catastrophes", "after": "applied widely for detecting a non-stationary symptom in random phenomena", "start_char_pos": 409, "end_char_pos": 445 } ]
[ 0, 158, 341 ]
0809.0241
1
In this paper we propose a novel Bayesian methodology for Value-at-Risk computation based on parametric Product Partition Models. Value-at-Risk is a standard tool to measure and control the market risk of an asset or a portfolio, and it is also required for regulatory purposes. Its popularity is partly due to the fact that it is an easily-understood measure of risk. The use of Product Partition Models allows us to remain in a normal setting , and to obtain a closed-form expression for Value-at-Risk computation. We present and compare two different scenarios: a product partition structure on the vector of means and a product partition structure on the vector of variances. We apply our methodology to an Italian stock market data set from Mib30. The numerical results we present clearly show that Product Partition Models can be successfully exploited in order to quantify market risk exposure. Value-at-Risk estimates we obtain are in full agreement with Maximum Likelihood approaches, but our methodology provides richer information about the clustering structure of the data and the presence of outlying points.
In this paper we propose a novel Bayesian methodology for Value-at-Risk computation based on parametric Product Partition Models. Value-at-Risk is a standard tool to measure and control the market risk of an asset or a portfolio, and it is also required for regulatory purposes. Its popularity is partly due to the fact that it is an easily understood measure of risk. The use of Product Partition Models allows us to remain in a Normal setting even in presence of outlying points , and to obtain a closed-form expression for Value-at-Risk computation. We present and compare two different scenarios: a product partition structure on the vector of means and a product partition structure on the vector of variances. We apply our methodology to an Italian stock market data set from Mib30. The numerical results clearly show that Product Partition Models can be successfully exploited in order to quantify market risk exposure. The obtained Value-at-Risk estimates are in full agreement with Maximum Likelihood approaches, but our methodology provides richer information about the clustering structure of the data and the presence of outlying points.
[ { "type": "R", "before": "easily-understood", "after": "easily understood", "start_char_pos": 334, "end_char_pos": 351 }, { "type": "R", "before": "normal setting", "after": "Normal setting even in presence of outlying points", "start_char_pos": 430, "end_char_pos": 444 }, { "type": "D", "before": "we present", "after": null, "start_char_pos": 775, "end_char_pos": 785 }, { "type": "A", "before": null, "after": "The obtained", "start_char_pos": 902, "end_char_pos": 902 }, { "type": "D", "before": "we obtain", "after": null, "start_char_pos": 927, "end_char_pos": 936 } ]
[ 0, 129, 278, 368, 516, 679, 752, 901 ]
0809.0739
1
We propose a mathematical framework for the study of a family of random fields - called forward performances - which arise as numerical representation of certain rational preference relations in mathematical finance. Their spatial structure corresponds to that of utility functions, while the temporal one reflects a Nisio-type semigroup property, referred to as self-generation. In the setting of semimartingale financial markets, we provide a dual formulation of self-generation in addition to the original one, and show equivalence between the two, thus giving a dual characterization of forward performances. Then, we focus on random fields with a log-affine structure and provide necessary and sufficient conditions for self-generation in that case. Finally, we illustrate our methods in financial markets driven by It\^o-processes, where we obtain an explicit parametrization of all log-affine forward performances.
We propose a mathematical framework for the study of a family of random fields - called forward performances - which arise as numerical representation of certain rational preference relations in mathematical finance. Their spatial structure corresponds to that of utility functions, while the temporal one reflects a Nisio-type semigroup property, referred to as self-generation. In the setting of semimartingale financial markets, we provide a dual formulation of self-generation in addition to the original one, and show equivalence between the two, thus giving a dual characterization of forward performances. Then, we focus on random fields with a exponential structure and provide necessary and sufficient conditions for self-generation in that case. Finally, we illustrate our methods in financial markets driven by It\^o-processes, where we obtain an explicit parametrization of all log-affine forward performances.
[ { "type": "R", "before": "log-affine", "after": "exponential", "start_char_pos": 652, "end_char_pos": 662 } ]
[ 0, 216, 379, 612, 754 ]
0809.0739
2
We propose a mathematical framework for the study of a family of random fields - called forward performances - which arise as numerical representation of certain rational preference relations in mathematical finance. Their spatial structure corresponds to that of utility functions, while the temporal one reflects a Nisio-type semigroup property, referred to as self-generation. In the setting of semimartingale financial markets, we provide a dual formulation of self-generation in addition to the original one, and show equivalence between the two, thus giving a dual characterization of forward performances. Then , we focus on random fields with a exponential structure and provide necessary and sufficient conditions for self-generation in that case. Finally, we illustrate our methods in financial markets driven by It\^o-processes, where we obtain an explicit parametrization of all log-affine forward performances.
We propose a mathematical framework for the study of a family of random fields--called forward performances--which arise as numerical representation of certain rational preference relations in mathematical finance. Their spatial structure corresponds to that of utility functions, while the temporal one reflects a Nisio-type semigroup property, referred to as self-generation. In the setting of semimartingale financial markets, we provide a dual formulation of self-generation in addition to the original one, and show equivalence between the two, thus giving a dual characterization of forward performances. Then we focus on random fields with an exponential structure and provide necessary and sufficient conditions for self-generation in that case. Finally, we illustrate our methods in financial markets driven by It\^o-processes, where we obtain an explicit parametrization of all exponential forward performances.
[ { "type": "R", "before": "fields - called forward performances - which", "after": "fields--called forward performances--which", "start_char_pos": 72, "end_char_pos": 116 }, { "type": "D", "before": ",", "after": null, "start_char_pos": 618, "end_char_pos": 619 }, { "type": "R", "before": "a", "after": "an", "start_char_pos": 651, "end_char_pos": 652 }, { "type": "R", "before": "log-affine", "after": "exponential", "start_char_pos": 891, "end_char_pos": 901 } ]
[ 0, 216, 379, 612, 756 ]
0809.1516
1
We construct an estimation and de-noising procedure for an input signal perturbed by a continuous-time Gaussian noise, using the local and occupation times of Gaussian processes. The method relies on the almost-sure minimization of a Stein Unbiased Risk Estimator (SURE ) obtained through integration by parts on Gaussian space, and applied to shrinkage estimators which are constructed by soft and hard thresholding .
Using integration by parts on Gaussian space we construct a Stein Unbiased Risk Estimator (SURE) for the drift of Gaussian processes using their local and occupation times . By almost-sure minimization of the SURE risk of shrinkage estimators we derive an estimation and de-noising procedure for an input signal perturbed by a continuous-time Gaussian noise .
[ { "type": "R", "before": "We construct an estimation and de-noising procedure for an input signal perturbed by a continuous-time Gaussian noise, using the", "after": "Using integration by parts on Gaussian space we construct a Stein Unbiased Risk Estimator (SURE) for the drift of Gaussian processes using their", "start_char_pos": 0, "end_char_pos": 128 }, { "type": "R", "before": "of Gaussian processes. The method relies on the", "after": ". By", "start_char_pos": 156, "end_char_pos": 203 }, { "type": "R", "before": "a Stein Unbiased Risk Estimator (SURE ) obtained through integration by parts on Gaussian space, and applied to shrinkage estimators which are constructed by soft and hard thresholding", "after": "the SURE risk of shrinkage estimators we derive an estimation and de-noising procedure for an input signal perturbed by a continuous-time Gaussian noise", "start_char_pos": 232, "end_char_pos": 416 } ]
[ 0, 178 ]
0809.1985
1
Affine term structure models have gained a lot of attention in the finance literature, which is due to their analytic tractability and statistical flexibility. The aim of this article is to present both , theoretical foundations and empirical aspects . Starting from the first short rate models , namely the Vasicek and the Cox-Ingersoll-Ross ones, we then give an overview of some properties of affine processes and explain their relation to affine term structure models. Pricing and estimation techniques are eventually mentioned, showing how the analytic tractability of affine models can be exploited for practical purposes.
Affine term structure models have gained significant attention in the finance literature, mainly due to their analytical tractability and statistical flexibility. The aim of this article is to present both theoretical foundations as well as empirical aspects of the affine model class . Starting from the original one-factor short-rate models of Vasicek and Coxet al, we provide an overview of the properties of regular affine processes and explain their relationship to affine term structure models. Methods for securities pricing and for parameter estimation are also discussed, demonstrating how the analytical tractability of affine models can be exploited for practical purposes.
[ { "type": "R", "before": "a lot of", "after": "significant", "start_char_pos": 41, "end_char_pos": 49 }, { "type": "R", "before": "which is", "after": "mainly", "start_char_pos": 87, "end_char_pos": 95 }, { "type": "R", "before": "analytic", "after": "analytical", "start_char_pos": 109, "end_char_pos": 117 }, { "type": "R", "before": ", theoretical foundations and empirical aspects", "after": "theoretical foundations as well as empirical aspects of the affine model class", "start_char_pos": 203, "end_char_pos": 250 }, { "type": "R", "before": "first short rate models , namely the", "after": "original one-factor short-rate models of", "start_char_pos": 271, "end_char_pos": 307 }, { "type": "R", "before": "the Cox-Ingersoll-Ross ones, we then give", "after": "Cox", "start_char_pos": 320, "end_char_pos": 361 }, { "type": "A", "before": null, "after": "et al,", "start_char_pos": 361, "end_char_pos": 361 }, { "type": "A", "before": null, "after": "we provide", "start_char_pos": 362, "end_char_pos": 362 }, { "type": "R", "before": "some properties of", "after": "the properties of regular", "start_char_pos": 378, "end_char_pos": 396 }, { "type": "R", "before": "relation", "after": "relationship", "start_char_pos": 432, "end_char_pos": 440 }, { "type": "R", "before": "Pricing and estimation techniques are eventually mentioned, showing how the analytic", "after": "Methods for securities pricing and for parameter estimation are also discussed, demonstrating how the analytical", "start_char_pos": 474, "end_char_pos": 558 } ]
[ 0, 159, 473 ]
0809.3405
1
We discuss the valuation problem for a broad spectrum of plain vanilla and path-dependent options in a general modeling framework, and specifically for L\'evy driven models. Among the derivatives which we consider are digitals, double digitals, asset-or-nothing options, self-quantos, lookback and one-touch options. Extensions to the multivariate case, i.e. basket options and options on the minimum or maximum of several assets are considered as well. As a link to credit risk we derive a valuation formula for equity default swaps. The key idea of this Fourier, resp. Laplace transform based approach is the clear separation of two ingredients, namely the payoff function and the underlying process . The latter can be the driving process itself, but also the supremum or infimum or another process derived from it. For L\'evy processes the analytically extended characteristic functions of the supremum and the infimum process are derived .
The aim of this article is to provide a systematic analysis of the conditions such that Fourier transform valuation formulas are valid in a general framework; i.e. when the option has an arbitrary payoff function and depends on the path of the asset price process. An interplay between the conditions on the payoff function and the process arises naturally. We also extend these results to the multi-dimensional case, and discuss the calculation of Greeks by Fourier transform methods. As an application, we price options on the minimum of two assets in L\'evy and stochastic volatility models .
[ { "type": "R", "before": "We discuss the valuation problem for a broad spectrum of plain vanilla and path-dependent options", "after": "The aim of this article is to provide a systematic analysis of the conditions such that Fourier transform valuation formulas are valid", "start_char_pos": 0, "end_char_pos": 97 }, { "type": "R", "before": "modeling framework, and specifically for L\\'evy driven models. Among the derivatives which we consider are digitals, double digitals, asset-or-nothing options, self-quantos, lookback and one-touch options. Extensions to the multivariate case,", "after": "framework;", "start_char_pos": 111, "end_char_pos": 353 }, { "type": "R", "before": "basket options and options on the minimum or maximum of several assets are considered as well. As a link to credit risk we derive a valuation formula for equity default swaps. The key idea of this Fourier, resp. Laplace transform based approach is the clear separation of two ingredients, namely", "after": "when the option has an arbitrary payoff function and depends on the path of the asset price process. An interplay between the conditions on", "start_char_pos": 359, "end_char_pos": 654 }, { "type": "R", "before": "underlying process . The latter can be the driving process itself, but also the supremum or infimum or another process derived from it. For L\\'evy processes the analytically extended characteristic functions of the supremum and the infimum process are derived", "after": "process arises naturally. We also extend these results to the multi-dimensional case, and discuss the calculation of Greeks by Fourier transform methods. As an application, we price options on the minimum of two assets in L\\'evy and stochastic volatility models", "start_char_pos": 683, "end_char_pos": 942 } ]
[ 0, 173, 316, 453, 534, 570, 703, 818 ]
0809.4139
1
One of the key socioeconomic phenomena to explain is the distribution of wealth. Bouchaud and Mezard have proposed an interesting model of economy [Bouchaud and Mezard (2000)] based on trade and investments of agents. In the mean-field approximation, the model produces a stationary wealth distribution with a power-law tail. In this paper we examine characteristic time scales of the model and show that for any finite number of agents, the validity of the mean-field result is time-limited and the model in fact has no stationary wealth distribution. Further analysis suggests that for heterogeneous agents, the limitations are even stronger. We conclude with general implications of the presented results.
One of the key socioeconomic phenomena to explain is the distribution of wealth. Bouchaud and M\'ezard have proposed an interesting model of economy [Bouchaud and M\'ezard (2000)] based on trade and investments of agents. In the mean-field approximation, the model produces a stationary wealth distribution with a power-law tail. In this paper we examine characteristic time scales of the model and show that for any finite number of agents, the validity of the mean-field result is time-limited and the model in fact has no stationary wealth distribution. Further analysis suggests that for heterogeneous agents, the limitations are even stronger. We conclude with general implications of the presented results.
[ { "type": "R", "before": "Mezard", "after": "M\\'ezard", "start_char_pos": 94, "end_char_pos": 100 }, { "type": "R", "before": "Mezard", "after": "M\\'ezard", "start_char_pos": 161, "end_char_pos": 167 } ]
[ 0, 80, 217, 325, 552, 644 ]
0810.2358
1
Certain short polycations, such as TAT and polyarginine, rapidly pass through cell plasma membranes by an unknown mechanism called transduction. These cell-penetrating peptides (CPPs) when fused to biologically active peptides promise to be medically useful . I offer a simple model in which phosphatidylserines (and possibly other anions) transduce CPPs. The model is consistent with the empirical upper limit on the cargo peptide of about 35 amino acids. More importantly, the model fits experimental data on how the transduction of a polyarginine-fluorophore into mouse C2C12 myoblasts depends on the number of arginines in the CPP and on the CPP concentration .
Certain short polycations, such as TAT and polyarginine, rapidly pass through the plasma membranes of mammalian cells by an unknown mechanism called transduction. These cell-penetrating peptides (CPPs) promise to be medically useful when fused to biologically active peptides . I offer a simple model in which phosphatidylserines transduce CPPs. The model is consistent with the empirical upper limit on the cargo peptide of about 50 amino acids. More importantly, it fits experimental data on how the transduction of a polyarginine-fluorophore into mouse C2C12 myoblasts depends on the number of arginines in the CPP and on the CPP concentration . The model makes two testable predictions .
[ { "type": "R", "before": "cell plasma membranes", "after": "the plasma membranes of mammalian cells", "start_char_pos": 78, "end_char_pos": 99 }, { "type": "A", "before": null, "after": "promise to be medically useful", "start_char_pos": 184, "end_char_pos": 184 }, { "type": "D", "before": "promise to be medically useful", "after": null, "start_char_pos": 228, "end_char_pos": 258 }, { "type": "D", "before": "(and possibly other anions)", "after": null, "start_char_pos": 313, "end_char_pos": 340 }, { "type": "R", "before": "35", "after": "50", "start_char_pos": 442, "end_char_pos": 444 }, { "type": "R", "before": "the model", "after": "it", "start_char_pos": 476, "end_char_pos": 485 }, { "type": "A", "before": null, "after": ". The model makes two testable predictions", "start_char_pos": 665, "end_char_pos": 665 } ]
[ 0, 144, 260, 356, 457 ]
0810.2358
2
Certain short polycations, such as TAT and polyarginine, rapidly pass through the plasma membranes of mammalian cells by an unknown mechanism called transduction . These cell-penetrating peptides (CPPs) promise to be medically useful when fused to biologically active peptides. I offer a simple model in which phosphatidylserines transduce CPPs . The model is consistent with the empirical upper limit on the cargo peptide of about 50 amino acids . More importantly, it fits experimental data on how the transduction of a polyarginine-fluorophore into mouse C2C12 myoblasts depends on the number of arginines in the CPP and on the CPP concentration. The model makes two testable predictions.
Certain short polycations, such as TAT and polyarginine, rapidly pass through the plasma membranes of mammalian cells by an unknown mechanism called transduction as well as by endocytosis and macropinocytosis . These cell-penetrating peptides (CPPs) promise to be medically useful when fused to biologically active peptides. I offer a simple model in which one or more CPPs and the phosphatidylserines of the inner leaflet form a kind of capacitor with a voltage in excess of 180 mV, high enough to create a molecular electropore . The model is consistent with an empirical upper limit on the cargo peptide of 40--60 amino acids and with experimental data on how the transduction of a polyarginine-fluorophore into mouse C2C12 myoblasts depends on the number of arginines in the CPP and on the CPP concentration. The model makes three testable predictions.
[ { "type": "A", "before": null, "after": "as well as by endocytosis and macropinocytosis", "start_char_pos": 162, "end_char_pos": 162 }, { "type": "R", "before": "phosphatidylserines transduce CPPs", "after": "one or more CPPs and the phosphatidylserines of the inner leaflet form a kind of capacitor with a voltage in excess of 180 mV, high enough to create a molecular electropore", "start_char_pos": 311, "end_char_pos": 345 }, { "type": "R", "before": "the", "after": "an", "start_char_pos": 377, "end_char_pos": 380 }, { "type": "R", "before": "about 50 amino acids . More importantly, it fits", "after": "40--60 amino acids and with", "start_char_pos": 427, "end_char_pos": 475 }, { "type": "R", "before": "two", "after": "three", "start_char_pos": 667, "end_char_pos": 670 } ]
[ 0, 164, 278, 347, 449, 650 ]
0810.3280
1
Sequence comparison and alignment has had an enormous impact on our understanding of evolution, biology, and disease. It is likely that comparison and alignment of biological networks will similarly revolutionize our understanding of biological systems . Existing network alignments are based largely upon information external to the networks, such as sequence, because purely topological alignment of large, noisy networks is a theoretically challenging problem that has yet to be solved . In this paper, we present the first alignments of large networks based solely upon detailed network topology, and demonstrate that both species phylogeny (the ``tree of life'') and detailed biological function of individual proteins can be extracted from topological comparisons of biological networks. Our phylogenetic trees provide a new and independent source of phylogenetic information. Our alignments are large, dense, and contiguous, suggesting functional predictions of hundreds of currently unannotated proteins. Finally, our alignment of the protein-protein interaction network of two very different species--yeast and human--indicate that even distant species share a surprising amount of network topology with each other, suggesting that there may be broad similarities in internal cellular wiring across all life on Earth.
Sequence comparison and alignment has had an enormous impact on our understanding of evolution, biology, and disease. Comparison and alignment of biological networks will likely have a similar impact . Existing network alignments use information external to the networks, such as sequence, because no good algorithm for purely topological alignment has yet been devised . In this paper, we present by far the most complete topological alignments of biological networks to date. We demonstrate that both species phylogeny and detailed biological function of individual proteins can be extracted from a solely topological comparison of networks. Topological alignments provide an independent new source of phylogenetic information. Our alignment of the protein-protein interaction networks of two very different species--yeast and human--indicate that even distant species share a surprising amount of network topology with each other, suggesting broad similarities in internal cellular wiring across all life on Earth.
[ { "type": "R", "before": "It is likely that comparison", "after": "Comparison", "start_char_pos": 118, "end_char_pos": 146 }, { "type": "R", "before": "similarly revolutionize our understanding of biological systems", "after": "likely have a similar impact", "start_char_pos": 189, "end_char_pos": 252 }, { "type": "R", "before": "are based largely upon", "after": "use", "start_char_pos": 283, "end_char_pos": 305 }, { "type": "A", "before": null, "after": "no good algorithm for", "start_char_pos": 370, "end_char_pos": 370 }, { "type": "R", "before": "of large, noisy networks is a theoretically challenging problem that has yet to be solved", "after": "has yet been devised", "start_char_pos": 400, "end_char_pos": 489 }, { "type": "R", "before": "the first alignments of large networks based solely upon detailed network topology, and", "after": "by far the most complete topological alignments of biological networks to date. We", "start_char_pos": 518, "end_char_pos": 605 }, { "type": "D", "before": "(the ``tree of life'')", "after": null, "start_char_pos": 646, "end_char_pos": 668 }, { "type": "R", "before": "topological comparisons of biological networks. Our phylogenetic trees provide a new and independent", "after": "a solely topological comparison of networks. Topological alignments provide an independent new", "start_char_pos": 747, "end_char_pos": 847 }, { "type": "D", "before": "alignments are large, dense, and contiguous, suggesting functional predictions of hundreds of currently unannotated proteins. Finally, our", "after": null, "start_char_pos": 888, "end_char_pos": 1026 }, { "type": "R", "before": "network", "after": "networks", "start_char_pos": 1072, "end_char_pos": 1079 }, { "type": "D", "before": "that there may be", "after": null, "start_char_pos": 1237, "end_char_pos": 1254 } ]
[ 0, 117, 491, 794, 883, 1013 ]
0810.3280
2
Sequence comparison and alignment has had an enormous impact on our understanding of evolution, biology, and disease. Comparison and alignment of biological networks will likely have a similar impact. Existing network alignments use information external to the networks, such as sequence, because no good algorithm for purely topological alignment has yet been devised. In this paper, we present by far the most complete topological alignments of biological networks to date. We demonstrate that both species phylogeny and detailed biological function of individual proteins can be extracted from a solely topological comparison of networks. Topological alignments provide an independent new source of phylogenetic information. Our alignment of the protein-protein interaction networks of two very different species--yeast and human--indicate that even distant species share a surprising amount of network topology with each other, suggesting broad similarities in internal cellular wiring across all life on Earth.
Sequence comparison and alignment has had an enormous impact on our understanding of evolution, biology, and disease. Comparison and alignment of biological networks will likely have a similar impact. Existing network alignments use information external to the networks, such as sequence, because no good algorithm for purely topological alignment has yet been devised. In this paper, we present a novel algorithm based solely on network topology, that can be used to align any two networks. We apply it to biological networks to produce by far the most complete topological alignments of biological networks to date. We demonstrate that both species phylogeny and detailed biological function of individual proteins can be extracted from our alignments. Topology-based alignments have the potential to provide a completely new, independent source of phylogenetic information. Our alignment of the protein-protein interaction networks of two very different species--yeast and human--indicate that even distant species share a surprising amount of network topology with each other, suggesting broad similarities in internal cellular wiring across all life on Earth.
[ { "type": "A", "before": null, "after": "a novel algorithm based solely on network topology, that can be used to align any two networks. We apply it to biological networks to produce", "start_char_pos": 396, "end_char_pos": 396 }, { "type": "R", "before": "a solely topological comparison of networks. Topological alignments provide an independent new", "after": "our alignments. Topology-based alignments have the potential to provide a completely new, independent", "start_char_pos": 598, "end_char_pos": 692 } ]
[ 0, 117, 200, 369, 476, 642, 728 ]
0810.5146
1
It turns out that a slightly generalised Margrabe formula exhibits symmetry properties leading to semi-static hedges of rather general options in the bivariate Black-Scholes economy . In order to increase the liquidity of the used hedging instruments for currency options, the duality principle can be used to set up the hedges in a foreign market by using only European vanilla options sometimes along with a risk-less bond. Since the semi-static hedges in the Black-Scholes economy are exact, a closed form valuation formula of a certain weighted barrier swap-option can be easily derived.
It turns out that in the bivariate Black-Scholes economy Margrabe type options exhibit symmetry properties leading to semi-static hedges of rather general barrier options. Some of the results are extended to variants obtained by means of Brownian subordination . In order to increase the liquidity of the hedging instruments for certain currency options, the duality principle can be applied to set up hedges in a foreign market by using only European vanilla options sometimes along with a risk-less bond. Since the semi-static hedges in the Black-Scholes economy are exact, closed form valuation formulas for certain barrier options can be easily derived.
[ { "type": "R", "before": "a slightly generalised Margrabe formula exhibits", "after": "in the bivariate Black-Scholes economy Margrabe type options exhibit", "start_char_pos": 18, "end_char_pos": 66 }, { "type": "R", "before": "options in the bivariate Black-Scholes economy", "after": "barrier options. Some of the results are extended to variants obtained by means of Brownian subordination", "start_char_pos": 135, "end_char_pos": 181 }, { "type": "D", "before": "used", "after": null, "start_char_pos": 226, "end_char_pos": 230 }, { "type": "A", "before": null, "after": "certain", "start_char_pos": 255, "end_char_pos": 255 }, { "type": "R", "before": "used", "after": "applied", "start_char_pos": 303, "end_char_pos": 307 }, { "type": "D", "before": "the", "after": null, "start_char_pos": 318, "end_char_pos": 321 }, { "type": "D", "before": "a", "after": null, "start_char_pos": 496, "end_char_pos": 497 }, { "type": "R", "before": "formula of a certain weighted barrier swap-option", "after": "formulas for certain barrier options", "start_char_pos": 520, "end_char_pos": 569 } ]
[ 0, 183, 426 ]
0811.1296
1
The problem of inhibiting viral DNA ejection from bacteriophages by Mg^{+2} counterions is studied. Experimentally, it is known that Mg^{+2 has a strong and non-monotonic effect on the amount of DNA ejected. There exists an optimal Mg^{+2 concentration at which the least DNA is ejected from the virus. At lower or higher concentrations, more DNA is ejected from the capsid. We propose that this phenomenon is the result of DNA overcharging by Mg^{+2} multivalent counterions. As Mg^{+2} concentration increases from zero, DNA net charge changes from negative to positive. The optimal inhibition corresponds to the Mg^{+2} concentration where DNA is neutral. At lower/higher concentrations, DNA genome is charged. It prefers to be in solution to lower its electrostatic self-energy, which consequently leads to an increase in DNA ejection. Our theory fits experimental data well. The strength of DNA-DNA short range attraction, mediated by Mg^{+2}, is found to be -0.003 k_BT per nucleotide base.
The problem of inhibiting viral DNA ejection from bacteriophages by multivalent counterions, especially Mg^{+2} counterions , is studied. Experimentally, it is known that MgSO_4 salt has a strong and non-monotonic effect on the amount of DNA ejected. There exists an optimal concentration at which the least DNA is ejected from the virus. At lower or higher concentrations, more DNA is ejected from the capsid. We propose that this phenomenon is the result of DNA overcharging by Mg^{+2} multivalent counterions. As Mg^{+2} concentration increases from zero, DNA net charge changes from negative to positive. The optimal inhibition corresponds to the Mg^{+2} concentration where DNA is neutral. At lower/higher concentrations, DNA genome is charged. It prefers to be in solution to lower its electrostatic self-energy, which consequently leads to an increase in DNA ejection. Our theory fits experimental data well. The strength of DNA-DNA short range attraction, mediated by Mg^{+2}, is found to be -0.003 k_BT per nucleotide base.
[ { "type": "A", "before": null, "after": "multivalent counterions, especially", "start_char_pos": 68, "end_char_pos": 68 }, { "type": "A", "before": null, "after": ",", "start_char_pos": 89, "end_char_pos": 89 }, { "type": "R", "before": "Mg^{+2", "after": "MgSO_4 salt", "start_char_pos": 135, "end_char_pos": 141 }, { "type": "D", "before": "Mg^{+2", "after": null, "start_char_pos": 234, "end_char_pos": 240 } ]
[ 0, 101, 209, 304, 376, 478, 574, 660, 715, 841, 881 ]
0811.1296
2
The problem of inhibiting viral DNA ejection from bacteriophages by multivalent counterions, especially Mg^{+2} counterions, is studied. Experimentally, it is known that MgSO_4 salt has a strong and non-monotonic effect on the amount of DNA ejected. There exists an optimal concentration at which the least DNA is ejected from the virus. At lower or higher concentrations, more DNA is ejected from the capsid. We propose that this phenomenon is the result of DNA overcharging by Mg^{+2} multivalent counterions. As Mg^{+2} concentration increases from zero, DNA net charge changes from negative to positive. The optimal inhibition corresponds to the Mg^{+2} concentration where DNA is neutral. At lower/higher concentrations, DNA genome is charged. It prefers to be in solution to lower its electrostatic self-energy, which consequently leads to an increase in DNA ejection. Our theory fits experimental datawell. The strength of DNA-DNA short range attraction , mediated by Mg^{+2}, is found to be -0.004 k_BT per nucleotide base .
The problem of inhibiting viral DNA ejection from bacteriophages by multivalent counterions, specifically Mg^{+2} counterions, is studied. Experimentally, it is known that MgSO_4 salt has a strong and non-monotonic effect on the amount of DNA ejected. There exists an optimal concentration at which the minimum amount of DNA is ejected from the virus. At lower or higher concentrations, more DNA is ejected from the capsid. We propose that this phenomenon is the result of DNA overcharging by Mg^{+2} multivalent counterions. As Mg^{+2} concentration increases from zero, the net charge of DNA changes from negative to positive. The optimal inhibition corresponds to the Mg^{+2} concentration where DNA is neutral. At lower/higher concentrations, DNA genome is charged. It prefers to be in solution to lower its electrostatic self-energy, which consequently leads to an increase in DNA ejection. By fitting our theory to available experimental data, the strength of DNA-DNA short range attraction energies , mediated by Mg^{+2}, is found to be -0.004 k_BT per nucleotide base . This and other fitted parameters agree well with known values from other experiments and computer simulations. The parameters are also in aggreement qualitatively with values for tri- and tetra-valent counterions .
[ { "type": "R", "before": "especially", "after": "specifically", "start_char_pos": 93, "end_char_pos": 103 }, { "type": "R", "before": "least", "after": "minimum amount of", "start_char_pos": 301, "end_char_pos": 306 }, { "type": "R", "before": "DNA net charge", "after": "the net charge of DNA", "start_char_pos": 558, "end_char_pos": 572 }, { "type": "R", "before": "Our theory fits experimental datawell. The", "after": "By fitting our theory to available experimental data, the", "start_char_pos": 875, "end_char_pos": 917 }, { "type": "A", "before": null, "after": "energies", "start_char_pos": 961, "end_char_pos": 961 }, { "type": "A", "before": null, "after": ". This and other fitted parameters agree well with known values from other experiments and computer simulations. The parameters are also in aggreement qualitatively with values for tri- and tetra-valent counterions", "start_char_pos": 1032, "end_char_pos": 1032 } ]
[ 0, 136, 249, 337, 409, 511, 607, 693, 748, 874, 913 ]
0811.2209
1
Complex systems are often modeled as Boolean networks in attempts to capture their logical structure and reveal its dynamical consequences. Approximating the dynamics of continuous variables by discrete values and Boolean logic gates may, however, introduce dynamical possibilities that are not accessible to the original system. We show that large random networks of variables coupled through continuous transfer functions often fail to exhibit the complex dynamics of corresponding Boolean models in the disordered (chaotic) regime, even when each individual function appears to be a good candidate for Boolean idealization. A simple criterion identifies continuous systems that exhibit the full dynamical range of their Boolean counterparts. Transfer functions inferred from the literature on transcriptional regulation of genes do not satisfy the criterion .
Complex systems are often modeled as Boolean networks in attempts to capture their logical structure and reveal its dynamical consequences. Approximating the dynamics of continuous variables by discrete values and Boolean logic gates may, however, introduce dynamical possibilities that are not accessible to the original system. We show that large random networks of variables coupled through continuous transfer functions often fail to exhibit the complex dynamics of corresponding Boolean models in the disordered (chaotic) regime, even when each individual function appears to be a good candidate for Boolean idealization. A suitably modified Boolean theory explains the behavior of systems in which information does not propagate faithfully down certain chains of nodes. Model networks incorporating calculated or directly measured transfer functions reported in the literature on transcriptional regulation of genes are described by the modified theory .
[ { "type": "R", "before": "simple criterion identifies continuous systems that exhibit the full dynamical range of their Boolean counterparts. Transfer functions inferred from", "after": "suitably modified Boolean theory explains the behavior of systems in which information does not propagate faithfully down certain chains of nodes. Model networks incorporating calculated or directly measured transfer functions reported in", "start_char_pos": 629, "end_char_pos": 777 }, { "type": "R", "before": "do not satisfy the criterion", "after": "are described by the modified theory", "start_char_pos": 832, "end_char_pos": 860 } ]
[ 0, 139, 329, 626, 744 ]
0811.3130
1
The creation and annihilation of money and its economic effects are reviewed. Economic values appear "in the mind" of the market participants; e. g., by pretending, maintaining and achieving a particular price for a certain asset. Upon its creation by banks, this kind of " value phantasy " is converted into " real money " often in terms of buyer's debtaccompanied by a simultaneous payment of fiat money to the seller. This money is then multiplied on the money market and is competing against other money supplies for the traded assets, goods and services, where it may cause dilution, inflation and reallocation of resources .
General options for monetary systems, in particular the creation of fiat money, as well as its appropriation in various setups is discussed. We identify private banks as the main source of money through monetization. The asset value is inevitably determined by subjective beliefs and fantasies loosely bound by market constraints. One may imagine such a monetary system as being " suspended in belief. " Inevitably, the interest levied by banks in return for money created via monetizing debt systematically reallocates resources towards the financial institutions, and away from industrial and manual production, farming and labor. Unfortunately, the alternatives appear to be even more troublesome than the present state of affairs. Any system based on interest-free fiat money creation, in order to avoid hyperinflation through excessive borrowing of " free" debt, has either to rely on unjustifiable privileges or chance. And any system based on commodity money such as the gold or silver standard instead of fiat money creation is heavily depending on the quantity of commodities, and also incapable of waging or defending against war through the effective monetization of future loot or loss .
[ { "type": "R", "before": "The creation and annihilation of money and its economic effects are reviewed. Economic values appear \"in the mind\" of the market participants; e. g., by pretending, maintaining and achieving a particular price for a certain asset. Upon its creation by banks, this kind of", "after": "General options for monetary systems, in particular the creation of fiat money, as well as its appropriation in various setups is discussed. We identify private banks as the main source of money through monetization. The asset value is inevitably determined by subjective beliefs and fantasies loosely bound by market constraints. One may imagine such a monetary system as being", "start_char_pos": 0, "end_char_pos": 271 }, { "type": "R", "before": "value phantasy", "after": "suspended in belief.", "start_char_pos": 274, "end_char_pos": 288 }, { "type": "R", "before": "is converted into", "after": "Inevitably, the interest levied by banks in return for money created via monetizing debt systematically reallocates resources towards the financial institutions, and away from industrial and manual production, farming and labor. Unfortunately, the alternatives appear to be even more troublesome than the present state of affairs. Any system based on interest-free fiat money creation, in order to avoid hyperinflation through excessive borrowing of", "start_char_pos": 291, "end_char_pos": 308 }, { "type": "R", "before": "real money \" often in terms of buyer's debtaccompanied by a simultaneous payment of fiat money to the seller. This money is then multiplied on the money market and is competing against other money supplies for the traded assets, goods and services, where it may cause dilution, inflation and reallocation of resources", "after": "free\" debt, has either to rely on unjustifiable privileges or chance. And any system based on commodity money such as the gold or silver standard instead of fiat money creation is heavily depending on the quantity of commodities, and also incapable of waging or defending against war through the effective monetization of future loot or loss", "start_char_pos": 311, "end_char_pos": 628 } ]
[ 0, 77, 142, 230, 420 ]
0811.3130
2
General options for monetary systems, in particular the creation of fiat money, as well as its appropriation in various setups is discussed. We identify private banks as the main source of money through monetization . The asset value is inevitably determined by subjective beliefs and fantasies loosely bound by market constraints. One may imagine such a monetary system as being "suspended in belief." Inevitably, the interest levied by banks in return for money created via monetizing debt systematically reallocates resources towards the financial institutions, and away from industrial and manual production, farming and labor. Unfortunately, the alternatives appear to be even more troublesome than the present state of affairs. Any system based on interest-free fiat money creation, in order to avoid hyperinflation through excessive borrowing of "free" debt, has either to rely on unjustifiable privileges or chance. And any system based on commodity money such as the gold or silver standard instead of fiat money creation is heavily depending on the quantity of commodities, and also incapable of waging or defending against war through the effective monetization of future loot or loss.
General options for monetary systems, in particular the creation of fiat money, as well as its appropriation in various setups is discussed. We identify private banks as the main source of money through monetization of assets and future profits . The asset value is determined by subjective beliefs and fantasies loosely bound by market constraints. One may imagine such a monetary system as being "suspended in belief." The interest levied by banks in return for money created via monetizing future profits systematically reallocates resources towards the financial institutions, and away from industrial and manual production, farming and labor. Unfortunately, the alternatives appear to be even more troublesome than the present state of affairs. Any system based on interest-free fiat money creation, in order to avoid hyperinflation through excessive borrowing of "free" debt, has either to rely on unjustifiable privileges or chance. And any system based on commodity instead of fiat money is heavily depending on the quantity of commodities, and also incapable of waging or defending against war through the effective monetization of future loot or loss.
[ { "type": "A", "before": null, "after": "of assets and future profits", "start_char_pos": 216, "end_char_pos": 216 }, { "type": "D", "before": "inevitably", "after": null, "start_char_pos": 238, "end_char_pos": 248 }, { "type": "R", "before": "Inevitably, the", "after": "The", "start_char_pos": 404, "end_char_pos": 419 }, { "type": "R", "before": "debt", "after": "future profits", "start_char_pos": 488, "end_char_pos": 492 }, { "type": "D", "before": "money such as the gold or silver standard", "after": null, "start_char_pos": 959, "end_char_pos": 1000 }, { "type": "D", "before": "creation", "after": null, "start_char_pos": 1023, "end_char_pos": 1031 } ]
[ 0, 140, 218, 332, 403, 632, 734, 924 ]
0811.3130
3
General options for monetary systems, in particular the creation of fiat money, as well as its appropriation in various setups is discussed. We identify private banks as the main source of money through monetization of assets and future profits. The asset value is determined by subjective beliefs and fantasies loosely bound by market constraints. One may imagine such a monetary system as being "suspended in belief ." The interest levied by banks in return for money created via monetizing future profits systematically reallocates resources towards the financial institutions, and away from industrial and manual production, farming and labor. Unfortunately, the alternatives appear to be even more troublesome than the present state of affairs. Any system based on interest-free fiat money creation, in order to avoid hyperinflation through excessive borrowing of "free" debt, has either to rely on unjustifiable privileges or chance. And any system based on commodity instead of fiat money is heavily depending on the quantity of commodities, and also incapable of waging or defending against war through the effective monetization of future loot or loss.
General options for monetary systems, in particular the creation of fiat money, as well as its appropriation in various setups are discussed. We identify private banks as the main source of money through monetization of assets and future profits. The asset value is determined by subjective confidence, beliefs and fantasies loosely bound by market constraints. One may imagine such a monetary system as being "suspended in thought ." Interest levied by banks in return for money created via monetizing future profits systematically reallocates resources toward the financial institutions, and away from industrial and manual production, farming and labor. Unfortunately, the alternatives appear to be even more troublesome than the present state of affairs. Any system based on interest-free fiat money creation, in order to avoid hyperinflation through excessive borrowing of "free" debt, has either to rely on unjustifiable privileges or chance. And any system based on commodity instead of fiat money is heavily depending on the quantity of commodities, and also incapable of waging or defending against war through the effective monetization of future loot or loss.
[ { "type": "R", "before": "is", "after": "are", "start_char_pos": 127, "end_char_pos": 129 }, { "type": "A", "before": null, "after": "confidence,", "start_char_pos": 290, "end_char_pos": 290 }, { "type": "R", "before": "belief", "after": "thought", "start_char_pos": 412, "end_char_pos": 418 }, { "type": "R", "before": "The interest", "after": "Interest", "start_char_pos": 422, "end_char_pos": 434 }, { "type": "R", "before": "towards", "after": "toward", "start_char_pos": 546, "end_char_pos": 553 } ]
[ 0, 140, 245, 349, 648, 750, 940 ]
0811.3130
4
General options for monetary systems, in particular the creation of fiat money, as well as its appropriation in various setups are discussed. We identify private banks as the main source of money through monetization of assets and future profits. The asset value is determined by subjective confidence, beliefs and fantasies loosely bound by market constraints. One may imagine such a monetary system as being "suspended in thought . " Interest levied by banks in return for money created via monetizing future profits systematically reallocates resources toward the financial institutions, and away from industrial and manual production, farming and labor. Unfortunately, the alternatives appear to be even more troublesome than the present state of affairs. Any system based on interest-free fiat money creation, in order to avoid hyperinflation through excessive borrowing of "free" debt, has either to rely on unjustifiable privileges or chance. And any system based on commodity instead of fiat money is heavily depending on the quantity of commodities, and also incapable of waging or defending against war through the effective monetization of future loot or loss.
Because of their provable indeterminateness, certain economic values and thus prices are fixed by subjective confidence, beliefs and fantasies loosely bound by market constraints. One may imagine such a monetary system as being "suspended in thought ". Monetization of " (derivative) promises" of any kind, as well as interest levied by banks in return for money created via monetizing future profits systematically reallocates resources toward the financial institutions, and away from industrial and manual production, farming and labor. Unfortunately, the alternatives appear to be even more troublesome than the present state of affairs. Any system based on interest-free fiat money creation, in order to avoid hyperinflation through excessive borrowing of "free" debt, has either to rely on unjustifiable privileges or chance. And any system based on commodity instead of fiat money is heavily depending on the quantity of commodities, and also incapable of waging or defending against war through the effective monetization of future loot or loss.
[ { "type": "R", "before": "General options for monetary systems, in particular the creation of fiat money, as well as its appropriation in various setups are discussed. We identify private banks as the main source of money through monetization of assets and future profits. The asset value is determined", "after": "Because of their provable indeterminateness, certain economic values and thus prices are fixed", "start_char_pos": 0, "end_char_pos": 276 }, { "type": "R", "before": ".", "after": "\". Monetization of", "start_char_pos": 432, "end_char_pos": 433 }, { "type": "R", "before": "Interest", "after": "(derivative) promises\" of any kind, as well as interest", "start_char_pos": 436, "end_char_pos": 444 } ]
[ 0, 141, 246, 361, 657, 759, 949 ]
0811.3130
5
Because of their provable indeterminateness, certain economic values and thus prices are fixed by subjective confidence, beliefs and fantasies loosely bound by market constraints. One may imagine such a monetary system as being " suspended in thought " . Monetization of " (derivative) promises " of any kind, as well as interest levied by banks in return for money created via monetizing future profits systematically reallocates resources toward the financial institutions, and away from industrial and manual production, farming and labor. Unfortunately, the alternatives appear to be even more troublesome than the present state of affairs. Any system based on interest-free fiat money creation , in order to avoid hyperinflation through excessive borrowing of "free" debt, has either to rely on unjustifiable privileges or chance. And any system based on commodity instead of fiat money is heavily depending on the quantity of commodities, and also incapable of waging or defending against war through the effective monetization of future loot or loss .
This review is about the convenience, the benefits, as well as the destructive capacities of money. It deals with various aspects of money creation, with its value, and its appropriation. All sorts of money tend to get corrupted by eventually creating too much of them. In the long run, this renders money worthless and deprives people holding it. This misuse of money creation is inevitable and should come as no surprise. Abusive money creation comes in various forms. In the present fiat money system " suspended in free thought " and sustained merely by our belief in and our conditioning to it, money is conveniently created out of " thin air " by excessive government spending and speculative credit creation. Alas, any too tight money supply could ruin an economy by inviting all sorts of unfriendly takeovers, including wars or competition. Therefore the ambivalence of money as benefactor and destroyer should be accepted as destiny .
[ { "type": "R", "before": "Because of their provable indeterminateness, certain economic values and thus prices are fixed by subjective confidence, beliefs and fantasies loosely bound by market constraints. One may imagine such a monetary system as being", "after": "This review is about the convenience, the benefits, as well as the destructive capacities of money. It deals with various aspects of money creation, with its value, and its appropriation. All sorts of money tend to get corrupted by eventually creating too much of them. In the long run, this renders money worthless and deprives people holding it. This misuse of money creation is inevitable and should come as no surprise. Abusive money creation comes in various forms. In the present fiat money system", "start_char_pos": 0, "end_char_pos": 227 }, { "type": "R", "before": "suspended in thought", "after": "suspended in free thought", "start_char_pos": 230, "end_char_pos": 250 }, { "type": "R", "before": ". Monetization of", "after": "and sustained merely by our belief in and our conditioning to it, money is conveniently created out of", "start_char_pos": 253, "end_char_pos": 270 }, { "type": "R", "before": "(derivative) promises", "after": "thin air", "start_char_pos": 273, "end_char_pos": 294 }, { "type": "R", "before": "of any kind, as well as interest levied by banks in return for money created via monetizing future profits systematically reallocates resources toward the financial institutions, and away from industrial and manual production, farming and labor. Unfortunately, the alternatives appear to be even more troublesome than the present state of affairs. Any system based on interest-free fiat money creation , in order to avoid hyperinflation through excessive borrowing of \"free\" debt, has either to rely on unjustifiable privileges or chance. And any system based on commodity instead of fiat money is heavily depending on the quantity of commodities, and also incapable of waging or defending against war through the effective monetization of future loot or loss", "after": "by excessive government spending and speculative credit creation. Alas, any too tight money supply could ruin an economy by inviting all sorts of unfriendly takeovers, including wars or competition. Therefore the ambivalence of money as benefactor and destroyer should be accepted as destiny", "start_char_pos": 297, "end_char_pos": 1056 } ]
[ 0, 179, 542, 644, 835 ]
0812.0033
1
A financial market model with general semimartingale asset-price processes and where agents can only trade using no-short-sale strategies is considered. We show that wealth processes using continuous trading can be approximated very closely by wealth processes using simple combinations of buy-and-hold trading. This approximation is based on controlling the proportions of wealth invested in the assets. As an application, the utility maximization problem is considered and it is shown that optimal utilities and wealth processes resulting from continuous trading can be approximated arbitrarily well by the use of simple combinations of buy-and-hold strategies.
A financial market model with general semimartingale asset-price processes and where agents can only trade using no-short-sales strategies is considered. We show that wealth processes using continuous trading can be approximated very closely by wealth processes using simple combinations of buy-and-hold trading. This approximation is based on controlling the proportions of wealth invested in the assets. As an application, the utility maximization problem is considered and it is shown that optimal expected utilities and wealth processes resulting from continuous trading can be approximated arbitrarily well by the use of simple combinations of buy-and-hold strategies.
[ { "type": "R", "before": "no-short-sale", "after": "no-short-sales", "start_char_pos": 113, "end_char_pos": 126 }, { "type": "A", "before": null, "after": "expected", "start_char_pos": 500, "end_char_pos": 500 } ]
[ 0, 152, 311, 404 ]
0812.0325
1
The analysis of network topology and dynamics is increasingly used for the description of the structure, function and evolution of complex systems. Here we summarize key aspects of the evolvability and robustness of the hierarchical network-set of macromolecules, URLanisms, and ecosystems. Listing the costs and benefits of cooperation as a necessary behaviour to build this network hierarchy, we outline the major hypothesis of the paper: the emergence of hierarchical complexity needs cooperation leading to the ageing of the constituent networks. Local cooperation in a stable environment may lead to over-optimization developing an always-old network, which ages slowly , and dies in an apoptosis-like process. Global cooperation by exploring a rapidly changing environment may cause an occasional over-perturbation exhausting system-resources, causing rapid degradation, ageing and death of an otherwise forever-young network in a necrosis-like process. Giving a number of examples we explain how local and global cooperation can both evoke and help successful ageing. Finally, we show how various forms of cooperation and consequent ageing emerge as key elements in all major steps of evolution from the formation of protocells to the establishment of the globalized, modern human society . Thus, ageing emerges as a price of complexity, which is going hand-in-hand with cooperation enhancing each other in a successful community .
The network concept is increasingly used for the description of complex systems. Here we summarize key aspects of the evolvability and robustness of the hierarchical network-set of macromolecules, URLanisms, and ecosystems. Listing the costs and benefits of cooperation as a necessary behaviour to build this network hierarchy, we outline the major hypothesis of the paper: the emergence of hierarchical complexity needs cooperation leading to the ageing (i.e. gradual deterioration) of the constituent networks. A stable environment develops cooperation leading to over-optimization , and forming an " always-old " network, which accumulates damage , and dies in an apoptosis-like process. A rapidly changing environment develops competition forming a "forever-young" network, which may suffer an occasional over-perturbation exhausting system-resources, and causing death in a necrosis-like process. Giving a number of examples we demonstrate how cooperation evokes the gradual accumulation of damage typical to ageing. Finally, we show how various forms of cooperation and consequent ageing emerge as key elements in all major steps of evolution from the formation of protocells to the establishment of the globalized, modern human society .
[ { "type": "R", "before": "analysis of network topology and dynamics", "after": "network concept", "start_char_pos": 4, "end_char_pos": 45 }, { "type": "D", "before": "the structure, function and evolution of", "after": null, "start_char_pos": 90, "end_char_pos": 130 }, { "type": "A", "before": null, "after": "(i.e. gradual deterioration)", "start_char_pos": 522, "end_char_pos": 522 }, { "type": "R", "before": "Local cooperation in a stable environment may lead", "after": "A stable environment develops cooperation leading", "start_char_pos": 552, "end_char_pos": 602 }, { "type": "R", "before": "developing an", "after": ", and forming an \"", "start_char_pos": 624, "end_char_pos": 637 }, { "type": "A", "before": null, "after": "\"", "start_char_pos": 649, "end_char_pos": 649 }, { "type": "R", "before": "ages slowly", "after": "accumulates damage", "start_char_pos": 665, "end_char_pos": 676 }, { "type": "R", "before": "Global cooperation by exploring a", "after": "A", "start_char_pos": 718, "end_char_pos": 751 }, { "type": "R", "before": "may cause", "after": "develops competition forming a \"forever-young\" network, which may suffer", "start_char_pos": 781, "end_char_pos": 790 }, { "type": "R", "before": "causing rapid degradation, ageing and death of an otherwise forever-young network", "after": "and causing death", "start_char_pos": 852, "end_char_pos": 933 }, { "type": "R", "before": "explain how local and global cooperation can both evoke and help successful", "after": "demonstrate how cooperation evokes the gradual accumulation of damage typical to", "start_char_pos": 993, "end_char_pos": 1068 }, { "type": "D", "before": ". Thus, ageing emerges as a price of complexity, which is going hand-in-hand with cooperation enhancing each other in a successful community", "after": null, "start_char_pos": 1298, "end_char_pos": 1438 } ]
[ 0, 147, 290, 551, 717, 961, 1076, 1299 ]
0812.2066
1
Actin networks in certain URLanisms exhibit a complex pattern-forming dynamics that starts with the appearance of static `spots ' of actin on the cell cortex. Spots soon become mobile, and eventually give rise to , and coexist with, traveling waves of actin. Waves confer motility upon the cell by impinging on its periphery. Here we describe a possible physical mechanism for this distinctive set of dynamic transformations . Starting from the observation that excitable reaction-diffusion models of chemical systems can describe both localized spots and traveling waves, we augment one such model with a variable describing the spatial orientation of its chief constituent . We consider this constituent to be a caricature of polymerized actin, and regard its orientation as a measure of local actin fiber alignment. This spatial anisotropy, which is not possessed by conventional chemical species, profoundly affects localized structures to drive a transformation at fixed parameter values from static spots to moving spots to waves.
Actin networks in certain URLanisms exhibit a complex pattern-forming dynamics that starts with the appearance of static spots of actin on the cell cortex. Spots soon become mobile, executing persistent random walks, and eventually give rise to traveling waves of actin. Here we describe a possible physical mechanism for this distinctive set of dynamic transformations , by equipping an excitable reaction-diffusion model with a field describing the spatial orientation of its chief constituent (which we consider to be actin). The interplay of anisotropic actin growth and spatial inhibition drives a transformation at fixed parameter values from static spots to moving spots to waves.
[ { "type": "R", "before": "`spots '", "after": "spots", "start_char_pos": 121, "end_char_pos": 129 }, { "type": "A", "before": null, "after": "executing persistent random walks,", "start_char_pos": 185, "end_char_pos": 185 }, { "type": "D", "before": ", and coexist with,", "after": null, "start_char_pos": 214, "end_char_pos": 233 }, { "type": "D", "before": "Waves confer motility upon the cell by impinging on its periphery.", "after": null, "start_char_pos": 260, "end_char_pos": 326 }, { "type": "R", "before": ". Starting from the observation that", "after": ", by equipping an", "start_char_pos": 426, "end_char_pos": 462 }, { "type": "D", "before": "models of chemical systems can describe both localized spots and traveling waves, we augment one such", "after": null, "start_char_pos": 492, "end_char_pos": 593 }, { "type": "R", "before": "variable", "after": "field", "start_char_pos": 607, "end_char_pos": 615 }, { "type": "R", "before": ". We consider this constituent to be a caricature of polymerized actin, and regard its orientation as a measure of local actin fiber alignment. This spatial anisotropy, which is not possessed by conventional chemical species, profoundly affects localized structures to drive", "after": "(which we consider to be actin). The interplay of anisotropic actin growth and spatial inhibition drives", "start_char_pos": 676, "end_char_pos": 950 } ]
[ 0, 158, 259, 326, 677, 819 ]
0812.2440
1
We extend the model of liquidity risk of Cetin et al. 5 to allow for price impacts. Starting from simple principles, we show that the impact of a trade on prices is directly proportional to the size of the transaction and the amount of liquidityof the asset. This leads to a new characterization of self-financing trading strategies and a sufficient condition for no arbitrage. We show that, with the use of volatility swaps, contingent claims whose payoffs depend on the value of the asset can be approximately replicated . The replicating costs of such payoffs are obtained from the solutions of BSDEs with
We extend a linear version of the liquidity risk model of Cetin et al. (2004) to allow for price impacts. We show that the impact of a market order on prices depends on the size of the transaction and the level of liquidity. We obtain a simple characterization of self-financing trading strategies and a sufficient condition for no arbitrage. We consider a stochastic volatility model in which the volatility is partly correlated with the liquidity process and show that, with the use of variance swaps, contingent claims whose payoffs depend on the value of the asset can be approximately replicated in this setting . The replicating costs of such payoffs are obtained from the solutions of BSDEs with quadratic growth and analytical properties of these solutions are investigated.
[ { "type": "R", "before": "the model of liquidity risk", "after": "a linear version of the liquidity risk model", "start_char_pos": 10, "end_char_pos": 37 }, { "type": "R", "before": "5", "after": "(2004)", "start_char_pos": 54, "end_char_pos": 55 }, { "type": "R", "before": "Starting from simple principles, we", "after": "We", "start_char_pos": 84, "end_char_pos": 119 }, { "type": "R", "before": "trade on prices is directly proportional to", "after": "market order on prices depends on", "start_char_pos": 146, "end_char_pos": 189 }, { "type": "R", "before": "amount of liquidityof the asset. This leads to a new", "after": "level of liquidity. We obtain a simple", "start_char_pos": 226, "end_char_pos": 278 }, { "type": "A", "before": null, "after": "consider a stochastic volatility model in which the volatility is partly correlated with the liquidity process and", "start_char_pos": 381, "end_char_pos": 381 }, { "type": "R", "before": "volatility", "after": "variance", "start_char_pos": 409, "end_char_pos": 419 }, { "type": "A", "before": null, "after": "in this setting", "start_char_pos": 524, "end_char_pos": 524 }, { "type": "A", "before": null, "after": "quadratic growth and analytical properties of these solutions are investigated.", "start_char_pos": 611, "end_char_pos": 611 } ]
[ 0, 83, 258, 377, 526 ]