Preferencje help
Widoczny [Schowaj] Abstrakt
Liczba wyników
Powiadomienia systemowe
  • Sesja wygasła!

Znaleziono wyników: 5

Liczba wyników na stronie
first rewind previous Strona / 1 next fast forward last
Wyniki wyszukiwania
Wyszukiwano:
w słowach kluczowych:  Bellman equation
help Sortuj według:

help Ogranicz wyniki do:
first rewind previous Strona / 1 next fast forward last
1
EN
In the paper we solve a system of Bellman equations for finite horizon continuous time terminal utility maximization problem with general càdlàg bid and ask prices.We assume that we have a restricted number of transactions at time moments we choose. The main result of the paper says that we can find a regular version of solutions to the system of Bellman equations, which enables us to find the form of nearly optimal strategies.
2
Content available remote The Bruss-Robertson Inequality: Elaborations, Extensions, and Applications
EN
The Bruss-Robertson inequality gives a bound on the maximal number of elements of a random sample whose sum is less than a specified value. The extension of that inequality which is given here neither requires the independence of the summands nor requires the equality of their marginal distributions. A review is also given of the applications of the Bruss-Robertson inequality, especially the applications to problems of combinatorial optimization such as the sequential knapsack problem and the sequential monotone subsequence selection problem.
PL
Nierówność Bruss-Robertson szacuje maksymalną liczbę elementów w próbie, której suma jest ograniczona przez zadaną liczbę. Uogólnienia tej nierówności podane w tej pracy nie wymagają założenia niezależności składników sumy ani tego, by były o tym samym rozkładzie. Podano także przegląd zastosowań nierówności Brussa-Robertsona, a zwłaszcza zastosowania do problemów kombinatorycznych, takich jak sekwencyjny problem upakowania i wybór monotonicznego podciągu.
3
Content available remote The machine learning approach: analysis of experimental results
EN
The article analyses a reinforcement learning method in which the subject of learning is defined. The essence of this method is the selection of activities by a try and fail process and awarding deferred rewards. Theoretical analyses were supplemented by the practical studies, with reference to implementation of the Sarsa( Lambda) algorithm, with replacing eligibility traces and the Epsilon greedy policy.
4
Content available remote Adaptive Machine Reinforcement Learning
EN
In this article is defined a reinforcement learning method, in which a subject of learning is analyzed. The essence of this method is the selection of activities by a try and fail process and awarding deferred rewards. If an environment is characterized by the Markov property, then step-by-step dynamics will enable forecasting of subsequent conditions and awarding subsequent rewards on the basis of the present known conditions and actions, relatively to the Markov decision making process. The relationship between the present conditions and values and the potential future conditions is defined by the Bellman equation. The article discusses also a method of temporal difference learning, mechanism of eligibility traces, as well as their algorithms TD(0) and TD(Lambda). Theoretical analyses were supplemented by the practical studies, with reference to all implementation of the Sarsa(Lambda) algorithm, with replacing eligibility traces and the Epsilon greedy policy.
5
Content available remote Discrete time portfolio selection with proportional transaction costs
EN
In the paper discrete time portfolio selection with maximization of a discounted satisfaction functional is studied. In Section 2 the case without transaction costs is considered and explicit solutions for special satisfaction functions are given. In Section 3 the problem with proportional transaction costs is investigated and optimal strategies are characterized.
first rewind previous Strona / 1 next fast forward last
JavaScript jest wyłączony w Twojej przeglądarce internetowej. Włącz go, a następnie odśwież stronę, aby móc w pełni z niej korzystać.